| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.983957219251337, | |
| "eval_steps": 50, | |
| "global_step": 279, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.053475935828877004, | |
| "grad_norm": 75.08688576643534, | |
| "learning_rate": 5e-07, | |
| "logits/chosen": -2.740462064743042, | |
| "logits/rejected": -2.7282073497772217, | |
| "logps/chosen": -257.7225646972656, | |
| "logps/rejected": -215.97402954101562, | |
| "loss": 0.6903, | |
| "rewards/accuracies": 0.3125, | |
| "rewards/chosen": 0.011177328415215015, | |
| "rewards/margins": 0.0018132120603695512, | |
| "rewards/rejected": 0.009364116936922073, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.10695187165775401, | |
| "grad_norm": 55.24172880066687, | |
| "learning_rate": 1e-06, | |
| "logits/chosen": -2.706225872039795, | |
| "logits/rejected": -2.695012092590332, | |
| "logps/chosen": -241.26431274414062, | |
| "logps/rejected": -213.05960083007812, | |
| "loss": 0.6648, | |
| "rewards/accuracies": 0.668749988079071, | |
| "rewards/chosen": 0.5042427182197571, | |
| "rewards/margins": 0.1558745801448822, | |
| "rewards/rejected": 0.3483680784702301, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.16042780748663102, | |
| "grad_norm": 55.35417452063743, | |
| "learning_rate": 9.991477798614637e-07, | |
| "logits/chosen": -2.567713737487793, | |
| "logits/rejected": -2.5728023052215576, | |
| "logps/chosen": -252.4267578125, | |
| "logps/rejected": -216.68722534179688, | |
| "loss": 0.6277, | |
| "rewards/accuracies": 0.699999988079071, | |
| "rewards/chosen": 1.2536556720733643, | |
| "rewards/margins": 0.6526123881340027, | |
| "rewards/rejected": 0.6010432839393616, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.21390374331550802, | |
| "grad_norm": 41.54015316305311, | |
| "learning_rate": 9.965940245625131e-07, | |
| "logits/chosen": -2.49863862991333, | |
| "logits/rejected": -2.4633045196533203, | |
| "logps/chosen": -239.6151580810547, | |
| "logps/rejected": -216.32748413085938, | |
| "loss": 0.5937, | |
| "rewards/accuracies": 0.731249988079071, | |
| "rewards/chosen": 1.429532766342163, | |
| "rewards/margins": 1.0765190124511719, | |
| "rewards/rejected": 0.3530138432979584, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.26737967914438504, | |
| "grad_norm": 43.266733978086314, | |
| "learning_rate": 9.923474395499264e-07, | |
| "logits/chosen": -2.3930270671844482, | |
| "logits/rejected": -2.391124725341797, | |
| "logps/chosen": -231.81448364257812, | |
| "logps/rejected": -213.6909942626953, | |
| "loss": 0.5968, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": 1.252290964126587, | |
| "rewards/margins": 0.7059992551803589, | |
| "rewards/rejected": 0.5462917685508728, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.32085561497326204, | |
| "grad_norm": 49.04579174218152, | |
| "learning_rate": 9.86422500924775e-07, | |
| "logits/chosen": -2.431389093399048, | |
| "logits/rejected": -2.420288562774658, | |
| "logps/chosen": -233.90640258789062, | |
| "logps/rejected": -232.1009521484375, | |
| "loss": 0.591, | |
| "rewards/accuracies": 0.675000011920929, | |
| "rewards/chosen": 1.3160102367401123, | |
| "rewards/margins": 1.2142870426177979, | |
| "rewards/rejected": 0.10172319412231445, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.37433155080213903, | |
| "grad_norm": 45.76490813889267, | |
| "learning_rate": 9.788394060951227e-07, | |
| "logits/chosen": -2.5551607608795166, | |
| "logits/rejected": -2.560732126235962, | |
| "logps/chosen": -243.9866180419922, | |
| "logps/rejected": -210.0946044921875, | |
| "loss": 0.5951, | |
| "rewards/accuracies": 0.6312500238418579, | |
| "rewards/chosen": 0.9597422480583191, | |
| "rewards/margins": 0.7127350568771362, | |
| "rewards/rejected": 0.2470073401927948, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.42780748663101603, | |
| "grad_norm": 35.943159004433745, | |
| "learning_rate": 9.696240049254742e-07, | |
| "logits/chosen": -2.607058048248291, | |
| "logits/rejected": -2.6059868335723877, | |
| "logps/chosen": -245.34237670898438, | |
| "logps/rejected": -215.133056640625, | |
| "loss": 0.5946, | |
| "rewards/accuracies": 0.7124999761581421, | |
| "rewards/chosen": 0.8869549036026001, | |
| "rewards/margins": 1.1748888492584229, | |
| "rewards/rejected": -0.2879341244697571, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.48128342245989303, | |
| "grad_norm": 43.25794768968881, | |
| "learning_rate": 9.588077116176756e-07, | |
| "logits/chosen": -2.5988757610321045, | |
| "logits/rejected": -2.587991952896118, | |
| "logps/chosen": -279.42425537109375, | |
| "logps/rejected": -214.3015594482422, | |
| "loss": 0.6027, | |
| "rewards/accuracies": 0.7124999761581421, | |
| "rewards/chosen": 1.1399929523468018, | |
| "rewards/margins": 1.116516351699829, | |
| "rewards/rejected": 0.023476576432585716, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.5347593582887701, | |
| "grad_norm": 42.38890964644049, | |
| "learning_rate": 9.464273976236516e-07, | |
| "logits/chosen": -2.5599303245544434, | |
| "logits/rejected": -2.567230701446533, | |
| "logps/chosen": -254.52859497070312, | |
| "logps/rejected": -213.7172088623047, | |
| "loss": 0.6034, | |
| "rewards/accuracies": 0.737500011920929, | |
| "rewards/chosen": 1.0623753070831299, | |
| "rewards/margins": 1.170297622680664, | |
| "rewards/rejected": -0.10792229324579239, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.5347593582887701, | |
| "eval_logits/chosen": -2.5372207164764404, | |
| "eval_logits/rejected": -2.5304667949676514, | |
| "eval_logps/chosen": -232.14488220214844, | |
| "eval_logps/rejected": -213.0785675048828, | |
| "eval_loss": 0.5868557095527649, | |
| "eval_rewards/accuracies": 0.7291666865348816, | |
| "eval_rewards/chosen": 0.7400359511375427, | |
| "eval_rewards/margins": 0.881316602230072, | |
| "eval_rewards/rejected": -0.14128059148788452, | |
| "eval_runtime": 178.1115, | |
| "eval_samples_per_second": 14.934, | |
| "eval_steps_per_second": 0.236, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.5882352941176471, | |
| "grad_norm": 35.35460256124788, | |
| "learning_rate": 9.325252659550308e-07, | |
| "logits/chosen": -2.5456783771514893, | |
| "logits/rejected": -2.5366146564483643, | |
| "logps/chosen": -250.4248504638672, | |
| "logps/rejected": -213.6925048828125, | |
| "loss": 0.5728, | |
| "rewards/accuracies": 0.71875, | |
| "rewards/chosen": 0.9326599836349487, | |
| "rewards/margins": 1.1307684183120728, | |
| "rewards/rejected": -0.19810837507247925, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.6417112299465241, | |
| "grad_norm": 36.50601945986719, | |
| "learning_rate": 9.171487073181197e-07, | |
| "logits/chosen": -2.497873544692993, | |
| "logits/rejected": -2.4702892303466797, | |
| "logps/chosen": -253.4021453857422, | |
| "logps/rejected": -218.5993194580078, | |
| "loss": 0.5838, | |
| "rewards/accuracies": 0.737500011920929, | |
| "rewards/chosen": 0.9917839169502258, | |
| "rewards/margins": 1.002171516418457, | |
| "rewards/rejected": -0.010387664660811424, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.6951871657754011, | |
| "grad_norm": 38.76671127630058, | |
| "learning_rate": 9.003501385646448e-07, | |
| "logits/chosen": -2.4642693996429443, | |
| "logits/rejected": -2.444017171859741, | |
| "logps/chosen": -231.67507934570312, | |
| "logps/rejected": -207.4264678955078, | |
| "loss": 0.5857, | |
| "rewards/accuracies": 0.737500011920929, | |
| "rewards/chosen": 0.6439246535301208, | |
| "rewards/margins": 1.2459630966186523, | |
| "rewards/rejected": -0.6020383834838867, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.7486631016042781, | |
| "grad_norm": 58.07607723312624, | |
| "learning_rate": 8.821868240089676e-07, | |
| "logits/chosen": -2.4684462547302246, | |
| "logits/rejected": -2.4463274478912354, | |
| "logps/chosen": -249.26455688476562, | |
| "logps/rejected": -222.1261749267578, | |
| "loss": 0.587, | |
| "rewards/accuracies": 0.6812499761581421, | |
| "rewards/chosen": 0.7888789772987366, | |
| "rewards/margins": 1.0215551853179932, | |
| "rewards/rejected": -0.23267626762390137, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.8021390374331551, | |
| "grad_norm": 36.09293428060366, | |
| "learning_rate": 8.62720680220876e-07, | |
| "logits/chosen": -2.501340866088867, | |
| "logits/rejected": -2.4918720722198486, | |
| "logps/chosen": -225.94638061523438, | |
| "logps/rejected": -218.7044219970703, | |
| "loss": 0.5926, | |
| "rewards/accuracies": 0.7124999761581421, | |
| "rewards/chosen": 0.8816758394241333, | |
| "rewards/margins": 0.8755629658699036, | |
| "rewards/rejected": 0.006112849805504084, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.8556149732620321, | |
| "grad_norm": 35.74459897866162, | |
| "learning_rate": 8.420180649593929e-07, | |
| "logits/chosen": -2.487574815750122, | |
| "logits/rejected": -2.478224277496338, | |
| "logps/chosen": -232.3713836669922, | |
| "logps/rejected": -208.5014190673828, | |
| "loss": 0.5303, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": 0.8995177149772644, | |
| "rewards/margins": 1.0680756568908691, | |
| "rewards/rejected": -0.16855797171592712, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.9090909090909091, | |
| "grad_norm": 34.953324196432796, | |
| "learning_rate": 8.201495509671036e-07, | |
| "logits/chosen": -2.4268527030944824, | |
| "logits/rejected": -2.394615888595581, | |
| "logps/chosen": -247.49609375, | |
| "logps/rejected": -214.3662567138672, | |
| "loss": 0.5561, | |
| "rewards/accuracies": 0.7250000238418579, | |
| "rewards/chosen": 0.6183642148971558, | |
| "rewards/margins": 1.220983862876892, | |
| "rewards/rejected": -0.6026195883750916, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.9625668449197861, | |
| "grad_norm": 31.39573691027677, | |
| "learning_rate": 7.971896853961042e-07, | |
| "logits/chosen": -2.2767372131347656, | |
| "logits/rejected": -2.252058506011963, | |
| "logps/chosen": -261.1527099609375, | |
| "logps/rejected": -220.43667602539062, | |
| "loss": 0.4894, | |
| "rewards/accuracies": 0.862500011920929, | |
| "rewards/chosen": 0.6145283579826355, | |
| "rewards/margins": 2.0016326904296875, | |
| "rewards/rejected": -1.3871045112609863, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.0160427807486632, | |
| "grad_norm": 17.75039945985475, | |
| "learning_rate": 7.732167356856654e-07, | |
| "logits/chosen": -2.166287660598755, | |
| "logits/rejected": -2.1111903190612793, | |
| "logps/chosen": -252.02944946289062, | |
| "logps/rejected": -232.24282836914062, | |
| "loss": 0.4711, | |
| "rewards/accuracies": 0.831250011920929, | |
| "rewards/chosen": 0.22497756779193878, | |
| "rewards/margins": 1.9439351558685303, | |
| "rewards/rejected": -1.7189576625823975, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 1.0695187165775402, | |
| "grad_norm": 23.02216164814203, | |
| "learning_rate": 7.48312422757881e-07, | |
| "logits/chosen": -2.144843339920044, | |
| "logits/rejected": -2.077284336090088, | |
| "logps/chosen": -268.06707763671875, | |
| "logps/rejected": -234.98910522460938, | |
| "loss": 0.2659, | |
| "rewards/accuracies": 0.9125000238418579, | |
| "rewards/chosen": 1.0067119598388672, | |
| "rewards/margins": 2.941673994064331, | |
| "rewards/rejected": -1.9349620342254639, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.0695187165775402, | |
| "eval_logits/chosen": -2.1482443809509277, | |
| "eval_logits/rejected": -2.1145682334899902, | |
| "eval_logps/chosen": -228.6596221923828, | |
| "eval_logps/rejected": -215.86151123046875, | |
| "eval_loss": 0.5500321388244629, | |
| "eval_rewards/accuracies": 0.7678571343421936, | |
| "eval_rewards/chosen": 1.0885626077651978, | |
| "eval_rewards/margins": 1.508140206336975, | |
| "eval_rewards/rejected": -0.419577419757843, | |
| "eval_runtime": 178.0144, | |
| "eval_samples_per_second": 14.943, | |
| "eval_steps_per_second": 0.236, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.1229946524064172, | |
| "grad_norm": 16.09113711061208, | |
| "learning_rate": 7.225616424408044e-07, | |
| "logits/chosen": -2.157158374786377, | |
| "logits/rejected": -2.128385066986084, | |
| "logps/chosen": -237.67849731445312, | |
| "logps/rejected": -231.83877563476562, | |
| "loss": 0.2083, | |
| "rewards/accuracies": 0.925000011920929, | |
| "rewards/chosen": 2.0537843704223633, | |
| "rewards/margins": 3.18843412399292, | |
| "rewards/rejected": -1.134649634361267, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 1.1764705882352942, | |
| "grad_norm": 17.181507055460564, | |
| "learning_rate": 6.96052176068713e-07, | |
| "logits/chosen": -2.2130496501922607, | |
| "logits/rejected": -2.1648192405700684, | |
| "logps/chosen": -225.992431640625, | |
| "logps/rejected": -225.93643188476562, | |
| "loss": 0.2286, | |
| "rewards/accuracies": 0.9312499761581421, | |
| "rewards/chosen": 1.768799066543579, | |
| "rewards/margins": 2.9698195457458496, | |
| "rewards/rejected": -1.2010207176208496, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.2299465240641712, | |
| "grad_norm": 23.325905689437562, | |
| "learning_rate": 6.688743912460229e-07, | |
| "logits/chosen": -2.257979154586792, | |
| "logits/rejected": -2.24532151222229, | |
| "logps/chosen": -245.7477569580078, | |
| "logps/rejected": -220.7982177734375, | |
| "loss": 0.2429, | |
| "rewards/accuracies": 0.90625, | |
| "rewards/chosen": 1.9697185754776, | |
| "rewards/margins": 3.129720687866211, | |
| "rewards/rejected": -1.1600019931793213, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 1.2834224598930482, | |
| "grad_norm": 19.723516785510316, | |
| "learning_rate": 6.411209337949213e-07, | |
| "logits/chosen": -2.244814395904541, | |
| "logits/rejected": -2.2148585319519043, | |
| "logps/chosen": -234.61392211914062, | |
| "logps/rejected": -258.7725830078125, | |
| "loss": 0.2529, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 1.9211610555648804, | |
| "rewards/margins": 4.399371147155762, | |
| "rewards/rejected": -2.478209972381592, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.3368983957219251, | |
| "grad_norm": 21.739425991143058, | |
| "learning_rate": 6.128864119368233e-07, | |
| "logits/chosen": -2.119013547897339, | |
| "logits/rejected": -2.083933115005493, | |
| "logps/chosen": -228.73965454101562, | |
| "logps/rejected": -208.091552734375, | |
| "loss": 0.2611, | |
| "rewards/accuracies": 0.8999999761581421, | |
| "rewards/chosen": 1.3465216159820557, | |
| "rewards/margins": 3.1292195320129395, | |
| "rewards/rejected": -1.7826979160308838, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.3903743315508021, | |
| "grad_norm": 22.912282028009347, | |
| "learning_rate": 5.842670737842467e-07, | |
| "logits/chosen": -2.0183815956115723, | |
| "logits/rejected": -1.987648367881775, | |
| "logps/chosen": -251.6590118408203, | |
| "logps/rejected": -229.6842803955078, | |
| "loss": 0.2933, | |
| "rewards/accuracies": 0.893750011920929, | |
| "rewards/chosen": 1.9206100702285767, | |
| "rewards/margins": 3.7730064392089844, | |
| "rewards/rejected": -1.8523967266082764, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.4438502673796791, | |
| "grad_norm": 22.11760289550141, | |
| "learning_rate": 5.553604792424922e-07, | |
| "logits/chosen": -2.0348455905914307, | |
| "logits/rejected": -1.9738250970840454, | |
| "logps/chosen": -245.8058319091797, | |
| "logps/rejected": -270.13043212890625, | |
| "loss": 0.2246, | |
| "rewards/accuracies": 0.918749988079071, | |
| "rewards/chosen": 1.6829955577850342, | |
| "rewards/margins": 3.753596782684326, | |
| "rewards/rejected": -2.070600986480713, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.4973262032085561, | |
| "grad_norm": 28.32892420101821, | |
| "learning_rate": 5.262651674395798e-07, | |
| "logits/chosen": -2.038222312927246, | |
| "logits/rejected": -1.9750888347625732, | |
| "logps/chosen": -240.39102172851562, | |
| "logps/rejected": -231.63534545898438, | |
| "loss": 0.2528, | |
| "rewards/accuracies": 0.9312499761581421, | |
| "rewards/chosen": 1.2107157707214355, | |
| "rewards/margins": 3.3977603912353516, | |
| "rewards/rejected": -2.187044143676758, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.5508021390374331, | |
| "grad_norm": 23.418588451486066, | |
| "learning_rate": 4.970803208181314e-07, | |
| "logits/chosen": -2.0812251567840576, | |
| "logits/rejected": -2.0191798210144043, | |
| "logps/chosen": -231.6563720703125, | |
| "logps/rejected": -240.78842163085938, | |
| "loss": 0.292, | |
| "rewards/accuracies": 0.9312499761581421, | |
| "rewards/chosen": 1.538320779800415, | |
| "rewards/margins": 3.8071465492248535, | |
| "rewards/rejected": -2.2688257694244385, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.6042780748663101, | |
| "grad_norm": 19.90761377348951, | |
| "learning_rate": 4.679054270342702e-07, | |
| "logits/chosen": -2.124783992767334, | |
| "logits/rejected": -2.077249050140381, | |
| "logps/chosen": -235.0903778076172, | |
| "logps/rejected": -243.20028686523438, | |
| "loss": 0.2599, | |
| "rewards/accuracies": 0.8999999761581421, | |
| "rewards/chosen": 1.3953964710235596, | |
| "rewards/margins": 3.141361951828003, | |
| "rewards/rejected": -1.7459657192230225, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.6042780748663101, | |
| "eval_logits/chosen": -2.1441855430603027, | |
| "eval_logits/rejected": -2.112197160720825, | |
| "eval_logps/chosen": -231.94912719726562, | |
| "eval_logps/rejected": -222.50692749023438, | |
| "eval_loss": 0.546504020690918, | |
| "eval_rewards/accuracies": 0.7678571343421936, | |
| "eval_rewards/chosen": 0.7596126198768616, | |
| "eval_rewards/margins": 1.8437296152114868, | |
| "eval_rewards/rejected": -1.0841171741485596, | |
| "eval_runtime": 177.8643, | |
| "eval_samples_per_second": 14.955, | |
| "eval_steps_per_second": 0.236, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.6577540106951871, | |
| "grad_norm": 21.93947550108777, | |
| "learning_rate": 4.3883993981608567e-07, | |
| "logits/chosen": -2.181997776031494, | |
| "logits/rejected": -2.123137950897217, | |
| "logps/chosen": -242.07229614257812, | |
| "logps/rejected": -237.5500030517578, | |
| "loss": 0.2643, | |
| "rewards/accuracies": 0.9312499761581421, | |
| "rewards/chosen": 1.5722686052322388, | |
| "rewards/margins": 3.503065824508667, | |
| "rewards/rejected": -1.9307969808578491, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 1.7112299465240641, | |
| "grad_norm": 24.144023300435187, | |
| "learning_rate": 4.0998293993775234e-07, | |
| "logits/chosen": -2.226807117462158, | |
| "logits/rejected": -2.199307680130005, | |
| "logps/chosen": -237.28543090820312, | |
| "logps/rejected": -238.6820831298828, | |
| "loss": 0.2942, | |
| "rewards/accuracies": 0.918749988079071, | |
| "rewards/chosen": 1.480787992477417, | |
| "rewards/margins": 3.468074083328247, | |
| "rewards/rejected": -1.987285852432251, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.7647058823529411, | |
| "grad_norm": 24.407740221287682, | |
| "learning_rate": 3.814327974650066e-07, | |
| "logits/chosen": -2.258931875228882, | |
| "logits/rejected": -2.2550344467163086, | |
| "logps/chosen": -239.78543090820312, | |
| "logps/rejected": -235.02456665039062, | |
| "loss": 0.2401, | |
| "rewards/accuracies": 0.9312499761581421, | |
| "rewards/chosen": 1.147258996963501, | |
| "rewards/margins": 3.696079969406128, | |
| "rewards/rejected": -2.548820972442627, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 1.8181818181818183, | |
| "grad_norm": 22.96884736732694, | |
| "learning_rate": 3.532868364233416e-07, | |
| "logits/chosen": -2.2801005840301514, | |
| "logits/rejected": -2.2604260444641113, | |
| "logps/chosen": -256.41278076171875, | |
| "logps/rejected": -237.5870361328125, | |
| "loss": 0.2635, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": 0.8354307413101196, | |
| "rewards/margins": 3.1207101345062256, | |
| "rewards/rejected": -2.2852795124053955, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.8716577540106951, | |
| "grad_norm": 21.468713981869463, | |
| "learning_rate": 3.256410030320304e-07, | |
| "logits/chosen": -2.2765324115753174, | |
| "logits/rejected": -2.2345592975616455, | |
| "logps/chosen": -226.8909912109375, | |
| "logps/rejected": -206.7959442138672, | |
| "loss": 0.2622, | |
| "rewards/accuracies": 0.893750011920929, | |
| "rewards/chosen": 0.8324386477470398, | |
| "rewards/margins": 3.093674659729004, | |
| "rewards/rejected": -2.2612357139587402, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 1.9251336898395723, | |
| "grad_norm": 25.530423897408046, | |
| "learning_rate": 2.985895386349233e-07, | |
| "logits/chosen": -2.2696175575256348, | |
| "logits/rejected": -2.244253635406494, | |
| "logps/chosen": -264.6595458984375, | |
| "logps/rejected": -227.76455688476562, | |
| "loss": 0.296, | |
| "rewards/accuracies": 0.90625, | |
| "rewards/chosen": 1.391035795211792, | |
| "rewards/margins": 3.3874428272247314, | |
| "rewards/rejected": -1.99640691280365, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.9786096256684491, | |
| "grad_norm": 23.57510964659792, | |
| "learning_rate": 2.7222465844296514e-07, | |
| "logits/chosen": -2.2398269176483154, | |
| "logits/rejected": -2.2095816135406494, | |
| "logps/chosen": -236.76220703125, | |
| "logps/rejected": -239.87649536132812, | |
| "loss": 0.3228, | |
| "rewards/accuracies": 0.9125000238418579, | |
| "rewards/chosen": 1.6798350811004639, | |
| "rewards/margins": 3.8687007427215576, | |
| "rewards/rejected": -2.1888651847839355, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 2.0320855614973263, | |
| "grad_norm": 11.237636994206616, | |
| "learning_rate": 2.466362371835544e-07, | |
| "logits/chosen": -2.2077956199645996, | |
| "logits/rejected": -2.1770200729370117, | |
| "logps/chosen": -253.8419647216797, | |
| "logps/rejected": -241.77560424804688, | |
| "loss": 0.2084, | |
| "rewards/accuracies": 0.9437500238418579, | |
| "rewards/chosen": 1.6418447494506836, | |
| "rewards/margins": 4.010140419006348, | |
| "rewards/rejected": -2.368295431137085, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.085561497326203, | |
| "grad_norm": 14.896733636093304, | |
| "learning_rate": 2.2191150272833386e-07, | |
| "logits/chosen": -2.168065309524536, | |
| "logits/rejected": -2.139242649078369, | |
| "logps/chosen": -239.03555297851562, | |
| "logps/rejected": -262.7664794921875, | |
| "loss": 0.1457, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": 1.6537681818008423, | |
| "rewards/margins": 4.138239860534668, | |
| "rewards/rejected": -2.484471559524536, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 2.1390374331550803, | |
| "grad_norm": 15.573127323302417, | |
| "learning_rate": 1.9813473874379395e-07, | |
| "logits/chosen": -2.1352179050445557, | |
| "logits/rejected": -2.1201109886169434, | |
| "logps/chosen": -243.68276977539062, | |
| "logps/rejected": -257.8447265625, | |
| "loss": 0.1366, | |
| "rewards/accuracies": 0.9750000238418579, | |
| "rewards/chosen": 1.748744249343872, | |
| "rewards/margins": 4.428500652313232, | |
| "rewards/rejected": -2.6797564029693604, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 2.1390374331550803, | |
| "eval_logits/chosen": -2.1154227256774902, | |
| "eval_logits/rejected": -2.0823864936828613, | |
| "eval_logps/chosen": -233.65536499023438, | |
| "eval_logps/rejected": -225.89039611816406, | |
| "eval_loss": 0.5221620202064514, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 0.5889881253242493, | |
| "eval_rewards/margins": 2.0114519596099854, | |
| "eval_rewards/rejected": -1.4224637746810913, | |
| "eval_runtime": 177.8889, | |
| "eval_samples_per_second": 14.953, | |
| "eval_steps_per_second": 0.236, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 2.192513368983957, | |
| "grad_norm": 18.629951176458935, | |
| "learning_rate": 1.7538699737832237e-07, | |
| "logits/chosen": -2.113478899002075, | |
| "logits/rejected": -2.069943904876709, | |
| "logps/chosen": -237.72598266601562, | |
| "logps/rejected": -261.28759765625, | |
| "loss": 0.1322, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 1.763622522354126, | |
| "rewards/margins": 4.295041084289551, | |
| "rewards/rejected": -2.531418561935425, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 2.2459893048128343, | |
| "grad_norm": 14.516798660616674, | |
| "learning_rate": 1.5374582296511053e-07, | |
| "logits/chosen": -2.103269577026367, | |
| "logits/rejected": -2.035609245300293, | |
| "logps/chosen": -238.05905151367188, | |
| "logps/rejected": -222.5261688232422, | |
| "loss": 0.146, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": 1.79998779296875, | |
| "rewards/margins": 4.091170310974121, | |
| "rewards/rejected": -2.29118275642395, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 2.299465240641711, | |
| "grad_norm": 17.56968018814267, | |
| "learning_rate": 1.3328498768278418e-07, | |
| "logits/chosen": -2.057720184326172, | |
| "logits/rejected": -2.012430191040039, | |
| "logps/chosen": -234.6757049560547, | |
| "logps/rejected": -250.71371459960938, | |
| "loss": 0.1554, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 1.705297827720642, | |
| "rewards/margins": 4.06239128112793, | |
| "rewards/rejected": -2.357093334197998, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 2.3529411764705883, | |
| "grad_norm": 12.58961314545, | |
| "learning_rate": 1.1407424007485928e-07, | |
| "logits/chosen": -2.0317530632019043, | |
| "logits/rejected": -1.9697539806365967, | |
| "logps/chosen": -247.4468536376953, | |
| "logps/rejected": -237.97738647460938, | |
| "loss": 0.1268, | |
| "rewards/accuracies": 0.9624999761581421, | |
| "rewards/chosen": 2.1162328720092773, | |
| "rewards/margins": 4.372066020965576, | |
| "rewards/rejected": -2.2558329105377197, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 2.406417112299465, | |
| "grad_norm": 21.013389395031766, | |
| "learning_rate": 9.617906728528679e-08, | |
| "logits/chosen": -2.0227417945861816, | |
| "logits/rejected": -1.967592477798462, | |
| "logps/chosen": -222.5628662109375, | |
| "logps/rejected": -241.11947631835938, | |
| "loss": 0.1401, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 1.6054117679595947, | |
| "rewards/margins": 3.956458330154419, | |
| "rewards/rejected": -2.3510468006134033, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 2.4598930481283423, | |
| "grad_norm": 15.57246636151293, | |
| "learning_rate": 7.966047182060226e-08, | |
| "logits/chosen": -1.9939205646514893, | |
| "logits/rejected": -1.947488784790039, | |
| "logps/chosen": -233.0928192138672, | |
| "logps/rejected": -235.1240692138672, | |
| "loss": 0.1625, | |
| "rewards/accuracies": 0.8999999761581421, | |
| "rewards/chosen": 1.5731914043426514, | |
| "rewards/margins": 3.795226573944092, | |
| "rewards/rejected": -2.2220349311828613, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.5133689839572195, | |
| "grad_norm": 13.432171621332934, | |
| "learning_rate": 6.457476359966684e-08, | |
| "logits/chosen": -1.992582082748413, | |
| "logits/rejected": -1.917047142982483, | |
| "logps/chosen": -227.5784912109375, | |
| "logps/rejected": -232.44485473632812, | |
| "loss": 0.1397, | |
| "rewards/accuracies": 0.925000011920929, | |
| "rewards/chosen": 1.6688787937164307, | |
| "rewards/margins": 3.8221230506896973, | |
| "rewards/rejected": -2.1532444953918457, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 2.5668449197860963, | |
| "grad_norm": 17.803144360723795, | |
| "learning_rate": 5.097336799988067e-08, | |
| "logits/chosen": -1.9870818853378296, | |
| "logits/rejected": -1.9163825511932373, | |
| "logps/chosen": -236.98666381835938, | |
| "logps/rejected": -237.57760620117188, | |
| "loss": 0.1576, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 1.9307695627212524, | |
| "rewards/margins": 4.332042217254639, | |
| "rewards/rejected": -2.4012725353240967, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 2.620320855614973, | |
| "grad_norm": 17.516307037134144, | |
| "learning_rate": 3.8902650554212826e-08, | |
| "logits/chosen": -1.971273422241211, | |
| "logits/rejected": -1.9711790084838867, | |
| "logps/chosen": -243.27542114257812, | |
| "logps/rejected": -231.23574829101562, | |
| "loss": 0.1618, | |
| "rewards/accuracies": 0.9312499761581421, | |
| "rewards/chosen": 2.017270565032959, | |
| "rewards/margins": 4.455116271972656, | |
| "rewards/rejected": -2.4378464221954346, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 2.6737967914438503, | |
| "grad_norm": 19.66674156907619, | |
| "learning_rate": 2.8403758896638707e-08, | |
| "logits/chosen": -1.9871242046356201, | |
| "logits/rejected": -1.9288272857666016, | |
| "logps/chosen": -238.3242645263672, | |
| "logps/rejected": -238.7816162109375, | |
| "loss": 0.1488, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 1.674863576889038, | |
| "rewards/margins": 4.010067939758301, | |
| "rewards/rejected": -2.335204839706421, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.6737967914438503, | |
| "eval_logits/chosen": -1.97805917263031, | |
| "eval_logits/rejected": -1.932228446006775, | |
| "eval_logps/chosen": -233.5652618408203, | |
| "eval_logps/rejected": -226.9253387451172, | |
| "eval_loss": 0.5411336421966553, | |
| "eval_rewards/accuracies": 0.7767857313156128, | |
| "eval_rewards/chosen": 0.5979987382888794, | |
| "eval_rewards/margins": 2.1239590644836426, | |
| "eval_rewards/rejected": -1.5259599685668945, | |
| "eval_runtime": 178.7366, | |
| "eval_samples_per_second": 14.882, | |
| "eval_steps_per_second": 0.235, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.7272727272727275, | |
| "grad_norm": 20.207169205524593, | |
| "learning_rate": 1.951248249476961e-08, | |
| "logits/chosen": -1.9819082021713257, | |
| "logits/rejected": -1.9226843118667603, | |
| "logps/chosen": -231.9635009765625, | |
| "logps/rejected": -247.98446655273438, | |
| "loss": 0.1533, | |
| "rewards/accuracies": 0.925000011920929, | |
| "rewards/chosen": 1.6299917697906494, | |
| "rewards/margins": 4.123189449310303, | |
| "rewards/rejected": -2.4931979179382324, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 2.7807486631016043, | |
| "grad_norm": 16.578647534749194, | |
| "learning_rate": 1.2259130647833626e-08, | |
| "logits/chosen": -1.9796741008758545, | |
| "logits/rejected": -1.9408137798309326, | |
| "logps/chosen": -235.86093139648438, | |
| "logps/rejected": -228.4602813720703, | |
| "loss": 0.1239, | |
| "rewards/accuracies": 0.956250011920929, | |
| "rewards/chosen": 1.7039449214935303, | |
| "rewards/margins": 4.0229291915893555, | |
| "rewards/rejected": -2.3189845085144043, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 2.834224598930481, | |
| "grad_norm": 16.969502053181888, | |
| "learning_rate": 6.668429165893996e-09, | |
| "logits/chosen": -1.9918749332427979, | |
| "logits/rejected": -1.9296365976333618, | |
| "logps/chosen": -262.6291809082031, | |
| "logps/rejected": -232.1147918701172, | |
| "loss": 0.1353, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": 1.8621528148651123, | |
| "rewards/margins": 4.288115501403809, | |
| "rewards/rejected": -2.4259629249572754, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 2.8877005347593583, | |
| "grad_norm": 14.34594308225596, | |
| "learning_rate": 2.759436082516664e-09, | |
| "logits/chosen": -1.991707444190979, | |
| "logits/rejected": -1.9247620105743408, | |
| "logps/chosen": -240.2772674560547, | |
| "logps/rejected": -234.63717651367188, | |
| "loss": 0.1532, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 1.6390937566757202, | |
| "rewards/margins": 4.325857162475586, | |
| "rewards/rejected": -2.686763286590576, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.9411764705882355, | |
| "grad_norm": 14.078396993164128, | |
| "learning_rate": 5.454766882097006e-10, | |
| "logits/chosen": -1.9781386852264404, | |
| "logits/rejected": -1.933526635169983, | |
| "logps/chosen": -232.15744018554688, | |
| "logps/rejected": -263.4436340332031, | |
| "loss": 0.1405, | |
| "rewards/accuracies": 0.90625, | |
| "rewards/chosen": 1.8876819610595703, | |
| "rewards/margins": 4.5024919509887695, | |
| "rewards/rejected": -2.6148104667663574, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 2.983957219251337, | |
| "step": 279, | |
| "total_flos": 3289753017384960.0, | |
| "train_loss": 0.3334879027045329, | |
| "train_runtime": 10390.4498, | |
| "train_samples_per_second": 6.91, | |
| "train_steps_per_second": 0.027 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 279, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3289753017384960.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |