| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 3.0, |
| "eval_steps": 500, |
| "global_step": 3750, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.04, |
| "grad_norm": 24.803234100341797, |
| "learning_rate": 1.3066666666666667e-06, |
| "logits/chosen": 2.3097052574157715, |
| "logits/rejected": 2.3757848739624023, |
| "logps/chosen": -67.16219329833984, |
| "logps/rejected": -80.8543930053711, |
| "loss": 1.0777, |
| "nll_loss": 0.41877487301826477, |
| "rewards/accuracies": 0.7237499952316284, |
| "rewards/chosen": -6.716219425201416, |
| "rewards/margins": 1.3692210912704468, |
| "rewards/rejected": -8.085439682006836, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 12.049943923950195, |
| "learning_rate": 2.64e-06, |
| "logits/chosen": 2.2844557762145996, |
| "logits/rejected": 2.364481210708618, |
| "logps/chosen": -61.06942367553711, |
| "logps/rejected": -75.95172119140625, |
| "loss": 1.0254, |
| "nll_loss": 0.38137227296829224, |
| "rewards/accuracies": 0.7174999713897705, |
| "rewards/chosen": -6.106942176818848, |
| "rewards/margins": 1.4882298707962036, |
| "rewards/rejected": -7.595171928405762, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 19.429025650024414, |
| "learning_rate": 3.973333333333333e-06, |
| "logits/chosen": 2.2412264347076416, |
| "logits/rejected": 2.2957522869110107, |
| "logps/chosen": -56.071502685546875, |
| "logps/rejected": -67.70003509521484, |
| "loss": 1.0618, |
| "nll_loss": 0.3372814655303955, |
| "rewards/accuracies": 0.6650000214576721, |
| "rewards/chosen": -5.607150077819824, |
| "rewards/margins": 1.1628532409667969, |
| "rewards/rejected": -6.770003795623779, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 14.493091583251953, |
| "learning_rate": 5.306666666666667e-06, |
| "logits/chosen": 2.1775848865509033, |
| "logits/rejected": 2.25844144821167, |
| "logps/chosen": -50.50484848022461, |
| "logps/rejected": -64.37080383300781, |
| "loss": 0.9298, |
| "nll_loss": 0.3056701719760895, |
| "rewards/accuracies": 0.7237499952316284, |
| "rewards/chosen": -5.050484657287598, |
| "rewards/margins": 1.3865959644317627, |
| "rewards/rejected": -6.437079906463623, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 17.336822509765625, |
| "learning_rate": 6.640000000000001e-06, |
| "logits/chosen": 2.203662872314453, |
| "logits/rejected": 2.28377103805542, |
| "logps/chosen": -51.80202102661133, |
| "logps/rejected": -65.30438232421875, |
| "loss": 0.9291, |
| "nll_loss": 0.2998887896537781, |
| "rewards/accuracies": 0.7149999737739563, |
| "rewards/chosen": -5.180202007293701, |
| "rewards/margins": 1.3502373695373535, |
| "rewards/rejected": -6.530439376831055, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 25.542375564575195, |
| "learning_rate": 7.973333333333334e-06, |
| "logits/chosen": 2.2555084228515625, |
| "logits/rejected": 2.3327550888061523, |
| "logps/chosen": -49.52461242675781, |
| "logps/rejected": -64.92198181152344, |
| "loss": 0.8288, |
| "nll_loss": 0.29204535484313965, |
| "rewards/accuracies": 0.7674999833106995, |
| "rewards/chosen": -4.9524617195129395, |
| "rewards/margins": 1.5397369861602783, |
| "rewards/rejected": -6.492198467254639, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 10.780914306640625, |
| "learning_rate": 9.306666666666667e-06, |
| "logits/chosen": 2.2995264530181885, |
| "logits/rejected": 2.3702523708343506, |
| "logps/chosen": -49.0943489074707, |
| "logps/rejected": -62.10791015625, |
| "loss": 0.9434, |
| "nll_loss": 0.29649874567985535, |
| "rewards/accuracies": 0.7250000238418579, |
| "rewards/chosen": -4.909435272216797, |
| "rewards/margins": 1.3013556003570557, |
| "rewards/rejected": -6.210791110992432, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 11.368802070617676, |
| "learning_rate": 9.998752338940612e-06, |
| "logits/chosen": 2.4777705669403076, |
| "logits/rejected": 2.5543289184570312, |
| "logps/chosen": -46.3427848815918, |
| "logps/rejected": -58.78908157348633, |
| "loss": 0.8661, |
| "nll_loss": 0.27519530057907104, |
| "rewards/accuracies": 0.7137500047683716, |
| "rewards/chosen": -4.634278297424316, |
| "rewards/margins": 1.2446298599243164, |
| "rewards/rejected": -5.878908634185791, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 13.878002166748047, |
| "learning_rate": 9.988142751731797e-06, |
| "logits/chosen": 2.4681615829467773, |
| "logits/rejected": 2.540001153945923, |
| "logps/chosen": -46.92790985107422, |
| "logps/rejected": -63.343467712402344, |
| "loss": 0.8081, |
| "nll_loss": 0.2872486710548401, |
| "rewards/accuracies": 0.7412499785423279, |
| "rewards/chosen": -4.692790985107422, |
| "rewards/margins": 1.6415550708770752, |
| "rewards/rejected": -6.334346294403076, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 14.774677276611328, |
| "learning_rate": 9.966729958067638e-06, |
| "logits/chosen": 2.509334087371826, |
| "logits/rejected": 2.5635790824890137, |
| "logps/chosen": -48.37749481201172, |
| "logps/rejected": -64.00433349609375, |
| "loss": 0.873, |
| "nll_loss": 0.2907007336616516, |
| "rewards/accuracies": 0.7200000286102295, |
| "rewards/chosen": -4.837749481201172, |
| "rewards/margins": 1.5626837015151978, |
| "rewards/rejected": -6.400432586669922, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 18.610092163085938, |
| "learning_rate": 9.934560333291077e-06, |
| "logits/chosen": 2.5441677570343018, |
| "logits/rejected": 2.613149404525757, |
| "logps/chosen": -48.265201568603516, |
| "logps/rejected": -63.49806213378906, |
| "loss": 0.8581, |
| "nll_loss": 0.2854183316230774, |
| "rewards/accuracies": 0.7287499904632568, |
| "rewards/chosen": -4.826520919799805, |
| "rewards/margins": 1.5232864618301392, |
| "rewards/rejected": -6.349807262420654, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 13.354905128479004, |
| "learning_rate": 9.89170354964594e-06, |
| "logits/chosen": 2.646564245223999, |
| "logits/rejected": 2.700756549835205, |
| "logps/chosen": -45.61009216308594, |
| "logps/rejected": -60.2086067199707, |
| "loss": 0.8363, |
| "nll_loss": 0.27716541290283203, |
| "rewards/accuracies": 0.7300000190734863, |
| "rewards/chosen": -4.561009883880615, |
| "rewards/margins": 1.4598517417907715, |
| "rewards/rejected": -6.0208611488342285, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 5.81559419631958, |
| "learning_rate": 9.83825242538238e-06, |
| "logits/chosen": 2.519411087036133, |
| "logits/rejected": 2.5865354537963867, |
| "logps/chosen": -46.41543960571289, |
| "logps/rejected": -63.667991638183594, |
| "loss": 0.7762, |
| "nll_loss": 0.27636775374412537, |
| "rewards/accuracies": 0.7587500214576721, |
| "rewards/chosen": -4.641544342041016, |
| "rewards/margins": 1.7252554893493652, |
| "rewards/rejected": -6.366799831390381, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 9.440147399902344, |
| "learning_rate": 9.774322723733216e-06, |
| "logits/chosen": 2.4503791332244873, |
| "logits/rejected": 2.5173845291137695, |
| "logps/chosen": -44.43014144897461, |
| "logps/rejected": -62.06157302856445, |
| "loss": 0.7447, |
| "nll_loss": 0.2715470492839813, |
| "rewards/accuracies": 0.7724999785423279, |
| "rewards/chosen": -4.443014621734619, |
| "rewards/margins": 1.7631423473358154, |
| "rewards/rejected": -6.206157207489014, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 9.054909706115723, |
| "learning_rate": 9.700052902196541e-06, |
| "logits/chosen": 2.3488733768463135, |
| "logits/rejected": 2.4224016666412354, |
| "logps/chosen": -45.02783203125, |
| "logps/rejected": -63.32554626464844, |
| "loss": 0.7256, |
| "nll_loss": 0.27446305751800537, |
| "rewards/accuracies": 0.8075000047683716, |
| "rewards/chosen": -4.502783298492432, |
| "rewards/margins": 1.8297713994979858, |
| "rewards/rejected": -6.332554817199707, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 16.773876190185547, |
| "learning_rate": 9.615603812667618e-06, |
| "logits/chosen": 2.4701356887817383, |
| "logits/rejected": 2.5356545448303223, |
| "logps/chosen": -44.635948181152344, |
| "logps/rejected": -62.55714416503906, |
| "loss": 0.7836, |
| "nll_loss": 0.2690128684043884, |
| "rewards/accuracies": 0.7524999976158142, |
| "rewards/chosen": -4.463594913482666, |
| "rewards/margins": 1.7921198606491089, |
| "rewards/rejected": -6.2557148933410645, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 12.325571060180664, |
| "learning_rate": 9.521158353069494e-06, |
| "logits/chosen": 2.450385332107544, |
| "logits/rejected": 2.5140254497528076, |
| "logps/chosen": -47.131587982177734, |
| "logps/rejected": -65.35118865966797, |
| "loss": 0.7507, |
| "nll_loss": 0.27970045804977417, |
| "rewards/accuracies": 0.7762500047683716, |
| "rewards/chosen": -4.713159561157227, |
| "rewards/margins": 1.8219600915908813, |
| "rewards/rejected": -6.535119533538818, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 18.231292724609375, |
| "learning_rate": 9.416921071236821e-06, |
| "logits/chosen": 2.4923791885375977, |
| "logits/rejected": 2.5595548152923584, |
| "logps/chosen": -46.485965728759766, |
| "logps/rejected": -63.454063415527344, |
| "loss": 0.8048, |
| "nll_loss": 0.27803030610084534, |
| "rewards/accuracies": 0.7437499761581421, |
| "rewards/chosen": -4.648596286773682, |
| "rewards/margins": 1.6968092918395996, |
| "rewards/rejected": -6.345406532287598, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 18.762359619140625, |
| "learning_rate": 9.303117721910801e-06, |
| "logits/chosen": 2.571411371231079, |
| "logits/rejected": 2.6245839595794678, |
| "logps/chosen": -44.32810974121094, |
| "logps/rejected": -61.370662689208984, |
| "loss": 0.7592, |
| "nll_loss": 0.2649703919887543, |
| "rewards/accuracies": 0.7662500143051147, |
| "rewards/chosen": -4.4328107833862305, |
| "rewards/margins": 1.7042557001113892, |
| "rewards/rejected": -6.137066841125488, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 13.24950122833252, |
| "learning_rate": 9.179994777804677e-06, |
| "logits/chosen": 2.5039403438568115, |
| "logits/rejected": 2.5774261951446533, |
| "logps/chosen": -45.20160675048828, |
| "logps/rejected": -62.881683349609375, |
| "loss": 0.7717, |
| "nll_loss": 0.2698582708835602, |
| "rewards/accuracies": 0.7787500023841858, |
| "rewards/chosen": -4.520160675048828, |
| "rewards/margins": 1.7680076360702515, |
| "rewards/rejected": -6.288168430328369, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 12.429143905639648, |
| "learning_rate": 9.04781889579873e-06, |
| "logits/chosen": 2.5070714950561523, |
| "logits/rejected": 2.5675339698791504, |
| "logps/chosen": -45.09945297241211, |
| "logps/rejected": -64.37616729736328, |
| "loss": 0.7095, |
| "nll_loss": 0.2686724364757538, |
| "rewards/accuracies": 0.7912499904632568, |
| "rewards/chosen": -4.509944915771484, |
| "rewards/margins": 1.9276715517044067, |
| "rewards/rejected": -6.437616348266602, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 13.30483341217041, |
| "learning_rate": 8.906876339420851e-06, |
| "logits/chosen": 2.388920783996582, |
| "logits/rejected": 2.4621076583862305, |
| "logps/chosen": -42.8404541015625, |
| "logps/rejected": -62.490779876708984, |
| "loss": 0.6913, |
| "nll_loss": 0.25781476497650146, |
| "rewards/accuracies": 0.8112499713897705, |
| "rewards/chosen": -4.284045219421387, |
| "rewards/margins": 1.9650330543518066, |
| "rewards/rejected": -6.249078750610352, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 13.460766792297363, |
| "learning_rate": 8.757472358863481e-06, |
| "logits/chosen": 2.4224061965942383, |
| "logits/rejected": 2.5014472007751465, |
| "logps/chosen": -45.438575744628906, |
| "logps/rejected": -63.852115631103516, |
| "loss": 0.7312, |
| "nll_loss": 0.26547786593437195, |
| "rewards/accuracies": 0.7749999761581421, |
| "rewards/chosen": -4.543857574462891, |
| "rewards/margins": 1.841354250907898, |
| "rewards/rejected": -6.385211944580078, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 9.995412826538086, |
| "learning_rate": 8.599930529879669e-06, |
| "logits/chosen": 2.289586067199707, |
| "logits/rejected": 2.3660268783569336, |
| "logps/chosen": -43.745357513427734, |
| "logps/rejected": -62.561038970947266, |
| "loss": 0.6825, |
| "nll_loss": 0.2639794647693634, |
| "rewards/accuracies": 0.8100000023841858, |
| "rewards/chosen": -4.37453556060791, |
| "rewards/margins": 1.8815685510635376, |
| "rewards/rejected": -6.256104946136475, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 9.7256498336792, |
| "learning_rate": 8.434592052990044e-06, |
| "logits/chosen": 2.1411895751953125, |
| "logits/rejected": 2.216475248336792, |
| "logps/chosen": -41.174503326416016, |
| "logps/rejected": -59.75578308105469, |
| "loss": 0.687, |
| "nll_loss": 0.25248458981513977, |
| "rewards/accuracies": 0.7912499904632568, |
| "rewards/chosen": -4.11745023727417, |
| "rewards/margins": 1.8581281900405884, |
| "rewards/rejected": -5.9755778312683105, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 5.8602166175842285, |
| "learning_rate": 8.261815014518465e-06, |
| "logits/chosen": 2.094317674636841, |
| "logits/rejected": 2.1707794666290283, |
| "logps/chosen": -42.47966003417969, |
| "logps/rejected": -67.64102172851562, |
| "loss": 0.5271, |
| "nll_loss": 0.25495222210884094, |
| "rewards/accuracies": 0.8837500214576721, |
| "rewards/chosen": -4.247966289520264, |
| "rewards/margins": 2.516136884689331, |
| "rewards/rejected": -6.764101982116699, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 7.209846019744873, |
| "learning_rate": 8.081973611056784e-06, |
| "logits/chosen": 1.899792194366455, |
| "logits/rejected": 1.9656882286071777, |
| "logps/chosen": -41.67923355102539, |
| "logps/rejected": -67.45558166503906, |
| "loss": 0.5335, |
| "nll_loss": 0.24939106404781342, |
| "rewards/accuracies": 0.8774999976158142, |
| "rewards/chosen": -4.167923450469971, |
| "rewards/margins": 2.577634811401367, |
| "rewards/rejected": -6.745559215545654, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 12.841861724853516, |
| "learning_rate": 7.89545733903834e-06, |
| "logits/chosen": 1.8675593137741089, |
| "logits/rejected": 1.9421709775924683, |
| "logps/chosen": -43.389644622802734, |
| "logps/rejected": -68.71646881103516, |
| "loss": 0.5398, |
| "nll_loss": 0.2613371014595032, |
| "rewards/accuracies": 0.8974999785423279, |
| "rewards/chosen": -4.338963985443115, |
| "rewards/margins": 2.5326828956604004, |
| "rewards/rejected": -6.871647834777832, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 13.226041793823242, |
| "learning_rate": 7.702670151175435e-06, |
| "logits/chosen": 1.8003557920455933, |
| "logits/rejected": 1.877779245376587, |
| "logps/chosen": -42.74818801879883, |
| "logps/rejected": -67.07624053955078, |
| "loss": 0.5569, |
| "nll_loss": 0.25505152344703674, |
| "rewards/accuracies": 0.8899999856948853, |
| "rewards/chosen": -4.274819374084473, |
| "rewards/margins": 2.4328057765960693, |
| "rewards/rejected": -6.707624435424805, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 6.528555393218994, |
| "learning_rate": 7.5040295815877e-06, |
| "logits/chosen": 1.8589718341827393, |
| "logits/rejected": 1.9315310716629028, |
| "logps/chosen": -42.12541198730469, |
| "logps/rejected": -68.34070587158203, |
| "loss": 0.5316, |
| "nll_loss": 0.25116249918937683, |
| "rewards/accuracies": 0.8849999904632568, |
| "rewards/chosen": -4.212540626525879, |
| "rewards/margins": 2.621530055999756, |
| "rewards/rejected": -6.834071636199951, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 8.806782722473145, |
| "learning_rate": 7.299965841516164e-06, |
| "logits/chosen": 1.896395206451416, |
| "logits/rejected": 1.9738116264343262, |
| "logps/chosen": -42.76396560668945, |
| "logps/rejected": -68.06053161621094, |
| "loss": 0.5404, |
| "nll_loss": 0.2575341463088989, |
| "rewards/accuracies": 0.8799999952316284, |
| "rewards/chosen": -4.27639627456665, |
| "rewards/margins": 2.5296568870544434, |
| "rewards/rejected": -6.8060526847839355, |
| "step": 1550 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 13.923650741577148, |
| "learning_rate": 7.090920887581507e-06, |
| "logits/chosen": 1.881974220275879, |
| "logits/rejected": 1.9411964416503906, |
| "logps/chosen": -41.719696044921875, |
| "logps/rejected": -66.24463653564453, |
| "loss": 0.5483, |
| "nll_loss": 0.24970975518226624, |
| "rewards/accuracies": 0.8762500286102295, |
| "rewards/chosen": -4.171969413757324, |
| "rewards/margins": 2.452493667602539, |
| "rewards/rejected": -6.62446403503418, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 11.483983993530273, |
| "learning_rate": 6.877347464604446e-06, |
| "logits/chosen": 1.8503919839859009, |
| "logits/rejected": 1.9176833629608154, |
| "logps/chosen": -42.25181579589844, |
| "logps/rejected": -67.97564697265625, |
| "loss": 0.541, |
| "nll_loss": 0.25359034538269043, |
| "rewards/accuracies": 0.8812500238418579, |
| "rewards/chosen": -4.225182056427002, |
| "rewards/margins": 2.572382926940918, |
| "rewards/rejected": -6.797564506530762, |
| "step": 1650 |
| }, |
| { |
| "epoch": 1.3599999999999999, |
| "grad_norm": 9.980610847473145, |
| "learning_rate": 6.659708125061242e-06, |
| "logits/chosen": 1.7879056930541992, |
| "logits/rejected": 1.8512917757034302, |
| "logps/chosen": -41.724449157714844, |
| "logps/rejected": -68.18392181396484, |
| "loss": 0.5127, |
| "nll_loss": 0.2496194839477539, |
| "rewards/accuracies": 0.9075000286102295, |
| "rewards/chosen": -4.172445297241211, |
| "rewards/margins": 2.645946741104126, |
| "rewards/rejected": -6.818392276763916, |
| "step": 1700 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 8.7877779006958, |
| "learning_rate": 6.438474227298065e-06, |
| "logits/chosen": 1.813609004020691, |
| "logits/rejected": 1.8792080879211426, |
| "logps/chosen": -42.19902420043945, |
| "logps/rejected": -67.10396575927734, |
| "loss": 0.5475, |
| "nll_loss": 0.2563365399837494, |
| "rewards/accuracies": 0.8887500166893005, |
| "rewards/chosen": -4.219902038574219, |
| "rewards/margins": 2.4904944896698, |
| "rewards/rejected": -6.710396766662598, |
| "step": 1750 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 9.004395484924316, |
| "learning_rate": 6.2141249146737545e-06, |
| "logits/chosen": 1.8060498237609863, |
| "logits/rejected": 1.8766037225723267, |
| "logps/chosen": -40.13544845581055, |
| "logps/rejected": -64.01081085205078, |
| "loss": 0.5417, |
| "nll_loss": 0.23975929617881775, |
| "rewards/accuracies": 0.8712499737739563, |
| "rewards/chosen": -4.013545036315918, |
| "rewards/margins": 2.3875370025634766, |
| "rewards/rejected": -6.401081562042236, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 14.659512519836426, |
| "learning_rate": 5.987146077842015e-06, |
| "logits/chosen": 1.7913141250610352, |
| "logits/rejected": 1.8706274032592773, |
| "logps/chosen": -41.8923454284668, |
| "logps/rejected": -68.07263946533203, |
| "loss": 0.5291, |
| "nll_loss": 0.24965395033359528, |
| "rewards/accuracies": 0.8812500238418579, |
| "rewards/chosen": -4.189234733581543, |
| "rewards/margins": 2.6180291175842285, |
| "rewards/rejected": -6.8072638511657715, |
| "step": 1850 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 6.758883953094482, |
| "learning_rate": 5.7580293024204455e-06, |
| "logits/chosen": 1.7456320524215698, |
| "logits/rejected": 1.8188166618347168, |
| "logps/chosen": -40.59235763549805, |
| "logps/rejected": -67.43293762207031, |
| "loss": 0.4978, |
| "nll_loss": 0.2416204810142517, |
| "rewards/accuracies": 0.8924999833106995, |
| "rewards/chosen": -4.0592360496521, |
| "rewards/margins": 2.6840572357177734, |
| "rewards/rejected": -6.743292808532715, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 11.839475631713867, |
| "learning_rate": 5.5272708043255605e-06, |
| "logits/chosen": 1.771327257156372, |
| "logits/rejected": 1.8334521055221558, |
| "logps/chosen": -40.4869499206543, |
| "logps/rejected": -64.32723999023438, |
| "loss": 0.5586, |
| "nll_loss": 0.24806198477745056, |
| "rewards/accuracies": 0.8675000071525574, |
| "rewards/chosen": -4.048694610595703, |
| "rewards/margins": 2.3840293884277344, |
| "rewards/rejected": -6.4327239990234375, |
| "step": 1950 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 12.160818099975586, |
| "learning_rate": 5.295370355079615e-06, |
| "logits/chosen": 1.7191228866577148, |
| "logits/rejected": 1.7949659824371338, |
| "logps/chosen": -40.17674255371094, |
| "logps/rejected": -67.11283874511719, |
| "loss": 0.5123, |
| "nll_loss": 0.24404479563236237, |
| "rewards/accuracies": 0.8774999976158142, |
| "rewards/chosen": -4.017674446105957, |
| "rewards/margins": 2.6936092376708984, |
| "rewards/rejected": -6.711284160614014, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.6400000000000001, |
| "grad_norm": 12.399029731750488, |
| "learning_rate": 5.062830199416764e-06, |
| "logits/chosen": 1.7645864486694336, |
| "logits/rejected": 1.8290934562683105, |
| "logps/chosen": -41.4688720703125, |
| "logps/rejected": -67.27618408203125, |
| "loss": 0.5257, |
| "nll_loss": 0.24901245534420013, |
| "rewards/accuracies": 0.8949999809265137, |
| "rewards/chosen": -4.146886825561523, |
| "rewards/margins": 2.580731153488159, |
| "rewards/rejected": -6.727618217468262, |
| "step": 2050 |
| }, |
| { |
| "epoch": 1.6800000000000002, |
| "grad_norm": 13.637982368469238, |
| "learning_rate": 4.8301539675328205e-06, |
| "logits/chosen": 1.6826503276824951, |
| "logits/rejected": 1.7500274181365967, |
| "logps/chosen": -40.96113586425781, |
| "logps/rejected": -67.78974914550781, |
| "loss": 0.513, |
| "nll_loss": 0.2408064305782318, |
| "rewards/accuracies": 0.8987500071525574, |
| "rewards/chosen": -4.096114158630371, |
| "rewards/margins": 2.6828606128692627, |
| "rewards/rejected": -6.778974533081055, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 5.442174911499023, |
| "learning_rate": 4.597845584334387e-06, |
| "logits/chosen": 1.6728637218475342, |
| "logits/rejected": 1.7252342700958252, |
| "logps/chosen": -42.251564025878906, |
| "logps/rejected": -69.24467468261719, |
| "loss": 0.5304, |
| "nll_loss": 0.25177592039108276, |
| "rewards/accuracies": 0.8899999856948853, |
| "rewards/chosen": -4.225156307220459, |
| "rewards/margins": 2.699312210083008, |
| "rewards/rejected": -6.92446756362915, |
| "step": 2150 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 10.488734245300293, |
| "learning_rate": 4.366408178049728e-06, |
| "logits/chosen": 1.6670646667480469, |
| "logits/rejected": 1.7333457469940186, |
| "logps/chosen": -39.98529052734375, |
| "logps/rejected": -65.87380981445312, |
| "loss": 0.5198, |
| "nll_loss": 0.2439391314983368, |
| "rewards/accuracies": 0.8762500286102295, |
| "rewards/chosen": -3.9985289573669434, |
| "rewards/margins": 2.588852643966675, |
| "rewards/rejected": -6.587381362915039, |
| "step": 2200 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 11.035846710205078, |
| "learning_rate": 4.136342990565055e-06, |
| "logits/chosen": 1.6644293069839478, |
| "logits/rejected": 1.7320383787155151, |
| "logps/chosen": -40.47782897949219, |
| "logps/rejected": -69.36231994628906, |
| "loss": 0.5059, |
| "nll_loss": 0.2478472888469696, |
| "rewards/accuracies": 0.8949999809265137, |
| "rewards/chosen": -4.047782897949219, |
| "rewards/margins": 2.888448476791382, |
| "rewards/rejected": -6.93623161315918, |
| "step": 2250 |
| }, |
| { |
| "epoch": 1.8399999999999999, |
| "grad_norm": 14.5468168258667, |
| "learning_rate": 3.908148291846225e-06, |
| "logits/chosen": 1.6440070867538452, |
| "logits/rejected": 1.6980115175247192, |
| "logps/chosen": -37.658931732177734, |
| "logps/rejected": -64.18938446044922, |
| "loss": 0.4975, |
| "nll_loss": 0.23423399031162262, |
| "rewards/accuracies": 0.8899999856948853, |
| "rewards/chosen": -3.7658934593200684, |
| "rewards/margins": 2.6530449390411377, |
| "rewards/rejected": -6.418938636779785, |
| "step": 2300 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 15.06531810760498, |
| "learning_rate": 3.6823183007969375e-06, |
| "logits/chosen": 1.633602261543274, |
| "logits/rejected": 1.69327974319458, |
| "logps/chosen": -38.722007751464844, |
| "logps/rejected": -65.98956298828125, |
| "loss": 0.512, |
| "nll_loss": 0.23453904688358307, |
| "rewards/accuracies": 0.8837500214576721, |
| "rewards/chosen": -3.8722009658813477, |
| "rewards/margins": 2.726755380630493, |
| "rewards/rejected": -6.598956108093262, |
| "step": 2350 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 9.41374397277832, |
| "learning_rate": 3.4593421148906523e-06, |
| "logits/chosen": 1.6280794143676758, |
| "logits/rejected": 1.694667935371399, |
| "logps/chosen": -41.9329833984375, |
| "logps/rejected": -68.24996948242188, |
| "loss": 0.5314, |
| "nll_loss": 0.24245284497737885, |
| "rewards/accuracies": 0.8887500166893005, |
| "rewards/chosen": -4.19329833984375, |
| "rewards/margins": 2.6316983699798584, |
| "rewards/rejected": -6.8249969482421875, |
| "step": 2400 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 10.686447143554688, |
| "learning_rate": 3.239702650894364e-06, |
| "logits/chosen": 1.648103952407837, |
| "logits/rejected": 1.7120572328567505, |
| "logps/chosen": -39.89656066894531, |
| "logps/rejected": -67.84294128417969, |
| "loss": 0.5132, |
| "nll_loss": 0.23985882103443146, |
| "rewards/accuracies": 0.8799999952316284, |
| "rewards/chosen": -3.989656448364258, |
| "rewards/margins": 2.794638156890869, |
| "rewards/rejected": -6.784294605255127, |
| "step": 2450 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 6.338383197784424, |
| "learning_rate": 3.023875598978419e-06, |
| "logits/chosen": 1.6715092658996582, |
| "logits/rejected": 1.7344988584518433, |
| "logps/chosen": -39.51519012451172, |
| "logps/rejected": -68.46276092529297, |
| "loss": 0.4834, |
| "nll_loss": 0.24019798636436462, |
| "rewards/accuracies": 0.9112499952316284, |
| "rewards/chosen": -3.951519250869751, |
| "rewards/margins": 2.8947577476501465, |
| "rewards/rejected": -6.846276760101318, |
| "step": 2500 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 5.144768238067627, |
| "learning_rate": 2.812328392477536e-06, |
| "logits/chosen": 1.5945950746536255, |
| "logits/rejected": 1.6544932126998901, |
| "logps/chosen": -37.963443756103516, |
| "logps/rejected": -69.92556762695312, |
| "loss": 0.3913, |
| "nll_loss": 0.22887782752513885, |
| "rewards/accuracies": 0.9712499976158142, |
| "rewards/chosen": -3.796344041824341, |
| "rewards/margins": 3.196213722229004, |
| "rewards/rejected": -6.992558002471924, |
| "step": 2550 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 5.904083251953125, |
| "learning_rate": 2.6055191955342886e-06, |
| "logits/chosen": 1.5594576597213745, |
| "logits/rejected": 1.6052813529968262, |
| "logps/chosen": -39.600093841552734, |
| "logps/rejected": -73.07125854492188, |
| "loss": 0.398, |
| "nll_loss": 0.23794788122177124, |
| "rewards/accuracies": 0.9662500023841858, |
| "rewards/chosen": -3.960009813308716, |
| "rewards/margins": 3.3471157550811768, |
| "rewards/rejected": -7.307126045227051, |
| "step": 2600 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 5.723136901855469, |
| "learning_rate": 2.403895910817593e-06, |
| "logits/chosen": 1.532933235168457, |
| "logits/rejected": 1.5803064107894897, |
| "logps/chosen": -38.91565704345703, |
| "logps/rejected": -70.781982421875, |
| "loss": 0.4101, |
| "nll_loss": 0.23784016072750092, |
| "rewards/accuracies": 0.9574999809265137, |
| "rewards/chosen": -3.8915653228759766, |
| "rewards/margins": 3.186633348464966, |
| "rewards/rejected": -7.07819938659668, |
| "step": 2650 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 7.6983561515808105, |
| "learning_rate": 2.2078952094652705e-06, |
| "logits/chosen": 1.4740697145462036, |
| "logits/rejected": 1.5456255674362183, |
| "logps/chosen": -39.75840377807617, |
| "logps/rejected": -75.84871673583984, |
| "loss": 0.3908, |
| "nll_loss": 0.23151804506778717, |
| "rewards/accuracies": 0.9612500071525574, |
| "rewards/chosen": -3.9758400917053223, |
| "rewards/margins": 3.6090316772460938, |
| "rewards/rejected": -7.584871292114258, |
| "step": 2700 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 7.529278755187988, |
| "learning_rate": 2.017941585351591e-06, |
| "logits/chosen": 1.495976209640503, |
| "logits/rejected": 1.5389071702957153, |
| "logps/chosen": -39.2768440246582, |
| "logps/rejected": -71.7361831665039, |
| "loss": 0.4138, |
| "nll_loss": 0.24059143662452698, |
| "rewards/accuracies": 0.9574999809265137, |
| "rewards/chosen": -3.9276845455169678, |
| "rewards/margins": 3.245933771133423, |
| "rewards/rejected": -7.173618793487549, |
| "step": 2750 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 6.076731204986572, |
| "learning_rate": 1.8344464357280722e-06, |
| "logits/chosen": 1.4751191139221191, |
| "logits/rejected": 1.5268608331680298, |
| "logps/chosen": -38.4973258972168, |
| "logps/rejected": -71.8733901977539, |
| "loss": 0.3943, |
| "nll_loss": 0.22699041664600372, |
| "rewards/accuracies": 0.9700000286102295, |
| "rewards/chosen": -3.8497328758239746, |
| "rewards/margins": 3.3376071453094482, |
| "rewards/rejected": -7.187338829040527, |
| "step": 2800 |
| }, |
| { |
| "epoch": 2.2800000000000002, |
| "grad_norm": 5.921169757843018, |
| "learning_rate": 1.6578071702286396e-06, |
| "logits/chosen": 1.4693129062652588, |
| "logits/rejected": 1.5106064081192017, |
| "logps/chosen": -38.8640022277832, |
| "logps/rejected": -72.8025131225586, |
| "loss": 0.4103, |
| "nll_loss": 0.2377665936946869, |
| "rewards/accuracies": 0.949999988079071, |
| "rewards/chosen": -3.8864006996154785, |
| "rewards/margins": 3.393850088119507, |
| "rewards/rejected": -7.280250072479248, |
| "step": 2850 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 5.996314525604248, |
| "learning_rate": 1.4884063501688539e-06, |
| "logits/chosen": 1.4711755514144897, |
| "logits/rejected": 1.537732720375061, |
| "logps/chosen": -38.81364440917969, |
| "logps/rejected": -73.80380249023438, |
| "loss": 0.3908, |
| "nll_loss": 0.22893303632736206, |
| "rewards/accuracies": 0.9737499952316284, |
| "rewards/chosen": -3.881364345550537, |
| "rewards/margins": 3.499016523361206, |
| "rewards/rejected": -7.380380153656006, |
| "step": 2900 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 10.235599517822266, |
| "learning_rate": 1.3266108600032928e-06, |
| "logits/chosen": 1.469245195388794, |
| "logits/rejected": 1.5233672857284546, |
| "logps/chosen": -39.08128356933594, |
| "logps/rejected": -70.8739242553711, |
| "loss": 0.4124, |
| "nll_loss": 0.2312445044517517, |
| "rewards/accuracies": 0.9512500166893005, |
| "rewards/chosen": -3.9081289768218994, |
| "rewards/margins": 3.1792635917663574, |
| "rewards/rejected": -7.087392807006836, |
| "step": 2950 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 8.700451850891113, |
| "learning_rate": 1.1727711127355118e-06, |
| "logits/chosen": 1.445876955986023, |
| "logits/rejected": 1.494391679763794, |
| "logps/chosen": -38.116580963134766, |
| "logps/rejected": -70.0914535522461, |
| "loss": 0.3982, |
| "nll_loss": 0.22633756697177887, |
| "rewards/accuracies": 0.9574999809265137, |
| "rewards/chosen": -3.8116586208343506, |
| "rewards/margins": 3.1974875926971436, |
| "rewards/rejected": -7.009146213531494, |
| "step": 3000 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 9.24150562286377, |
| "learning_rate": 1.0272202910015083e-06, |
| "logits/chosen": 1.4509037733078003, |
| "logits/rejected": 1.5117911100387573, |
| "logps/chosen": -39.70808410644531, |
| "logps/rejected": -74.1922836303711, |
| "loss": 0.4038, |
| "nll_loss": 0.23606754839420319, |
| "rewards/accuracies": 0.9662500023841858, |
| "rewards/chosen": -3.970808744430542, |
| "rewards/margins": 3.4484190940856934, |
| "rewards/rejected": -7.419227123260498, |
| "step": 3050 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 10.763381958007812, |
| "learning_rate": 8.902736254703347e-07, |
| "logits/chosen": 1.4675711393356323, |
| "logits/rejected": 1.5065741539001465, |
| "logps/chosen": -36.786865234375, |
| "logps/rejected": -67.48006439208984, |
| "loss": 0.405, |
| "nll_loss": 0.22343981266021729, |
| "rewards/accuracies": 0.9624999761581421, |
| "rewards/chosen": -3.6786866188049316, |
| "rewards/margins": 3.0693204402923584, |
| "rewards/rejected": -6.748007297515869, |
| "step": 3100 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 5.758388996124268, |
| "learning_rate": 7.622277121246513e-07, |
| "logits/chosen": 1.483445405960083, |
| "logits/rejected": 1.5389055013656616, |
| "logps/chosen": -38.62077331542969, |
| "logps/rejected": -71.79267883300781, |
| "loss": 0.4121, |
| "nll_loss": 0.23168590664863586, |
| "rewards/accuracies": 0.9587500095367432, |
| "rewards/chosen": -3.8620777130126953, |
| "rewards/margins": 3.3171918392181396, |
| "rewards/rejected": -7.179268836975098, |
| "step": 3150 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 7.307446002960205, |
| "learning_rate": 6.433598698998766e-07, |
| "logits/chosen": 1.4688174724578857, |
| "logits/rejected": 1.5306994915008545, |
| "logps/chosen": -39.013912200927734, |
| "logps/rejected": -71.06527709960938, |
| "loss": 0.416, |
| "nll_loss": 0.23500658571720123, |
| "rewards/accuracies": 0.9424999952316284, |
| "rewards/chosen": -3.9013912677764893, |
| "rewards/margins": 3.2051360607147217, |
| "rewards/rejected": -7.106527328491211, |
| "step": 3200 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 6.810940742492676, |
| "learning_rate": 5.339275400731331e-07, |
| "logits/chosen": 1.4731531143188477, |
| "logits/rejected": 1.5175625085830688, |
| "logps/chosen": -39.05414581298828, |
| "logps/rejected": -69.38569641113281, |
| "loss": 0.4207, |
| "nll_loss": 0.2375660389661789, |
| "rewards/accuracies": 0.949999988079071, |
| "rewards/chosen": -3.9054150581359863, |
| "rewards/margins": 3.0331552028656006, |
| "rewards/rejected": -6.93856954574585, |
| "step": 3250 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 5.935739517211914, |
| "learning_rate": 4.3416772870275295e-07, |
| "logits/chosen": 1.447837233543396, |
| "logits/rejected": 1.5085737705230713, |
| "logps/chosen": -38.072574615478516, |
| "logps/rejected": -71.93220520019531, |
| "loss": 0.3981, |
| "nll_loss": 0.22540073096752167, |
| "rewards/accuracies": 0.9537500143051147, |
| "rewards/chosen": -3.807257652282715, |
| "rewards/margins": 3.385963201522827, |
| "rewards/rejected": -7.193220138549805, |
| "step": 3300 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 6.746799468994141, |
| "learning_rate": 3.442964933259474e-07, |
| "logits/chosen": 1.4588629007339478, |
| "logits/rejected": 1.517850637435913, |
| "logps/chosen": -38.37602615356445, |
| "logps/rejected": -72.57811737060547, |
| "loss": 0.3998, |
| "nll_loss": 0.23044270277023315, |
| "rewards/accuracies": 0.9587500095367432, |
| "rewards/chosen": -3.8376026153564453, |
| "rewards/margins": 3.4202094078063965, |
| "rewards/rejected": -7.257812023162842, |
| "step": 3350 |
| }, |
| { |
| "epoch": 2.7199999999999998, |
| "grad_norm": 7.958431243896484, |
| "learning_rate": 2.6450847502627883e-07, |
| "logits/chosen": 1.4416861534118652, |
| "logits/rejected": 1.5023283958435059, |
| "logps/chosen": -39.2740364074707, |
| "logps/rejected": -73.73572540283203, |
| "loss": 0.4036, |
| "nll_loss": 0.22917690873146057, |
| "rewards/accuracies": 0.9662500023841858, |
| "rewards/chosen": -3.927403450012207, |
| "rewards/margins": 3.446169376373291, |
| "rewards/rejected": -7.373574256896973, |
| "step": 3400 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 8.157621383666992, |
| "learning_rate": 1.9497647688442478e-07, |
| "logits/chosen": 1.4702584743499756, |
| "logits/rejected": 1.5254310369491577, |
| "logps/chosen": -38.0818977355957, |
| "logps/rejected": -70.6051254272461, |
| "loss": 0.4055, |
| "nll_loss": 0.23146916925907135, |
| "rewards/accuracies": 0.9549999833106995, |
| "rewards/chosen": -3.808190107345581, |
| "rewards/margins": 3.252322196960449, |
| "rewards/rejected": -7.060512065887451, |
| "step": 3450 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 6.671606540679932, |
| "learning_rate": 1.358510897251808e-07, |
| "logits/chosen": 1.4628223180770874, |
| "logits/rejected": 1.525000810623169, |
| "logps/chosen": -37.097007751464844, |
| "logps/rejected": -72.08160400390625, |
| "loss": 0.3906, |
| "nll_loss": 0.22892680764198303, |
| "rewards/accuracies": 0.9700000286102295, |
| "rewards/chosen": -3.7097012996673584, |
| "rewards/margins": 3.498459577560425, |
| "rewards/rejected": -7.208160400390625, |
| "step": 3500 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 8.076554298400879, |
| "learning_rate": 8.726036597126619e-08, |
| "logits/chosen": 1.470629096031189, |
| "logits/rejected": 1.535922646522522, |
| "logps/chosen": -36.950157165527344, |
| "logps/rejected": -68.89533996582031, |
| "loss": 0.395, |
| "nll_loss": 0.22603529691696167, |
| "rewards/accuracies": 0.9549999833106995, |
| "rewards/chosen": -3.6950161457061768, |
| "rewards/margins": 3.1945183277130127, |
| "rewards/rejected": -6.8895344734191895, |
| "step": 3550 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 6.550802707672119, |
| "learning_rate": 4.93095423102935e-08, |
| "logits/chosen": 1.4719560146331787, |
| "logits/rejected": 1.5248959064483643, |
| "logps/chosen": -40.710941314697266, |
| "logps/rejected": -74.42230987548828, |
| "loss": 0.4132, |
| "nll_loss": 0.24109821021556854, |
| "rewards/accuracies": 0.9524999856948853, |
| "rewards/chosen": -4.071094512939453, |
| "rewards/margins": 3.371136426925659, |
| "rewards/rejected": -7.442230701446533, |
| "step": 3600 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 6.743546485900879, |
| "learning_rate": 2.2080811775535006e-08, |
| "logits/chosen": 1.4744410514831543, |
| "logits/rejected": 1.5339540243148804, |
| "logps/chosen": -39.543033599853516, |
| "logps/rejected": -74.6586685180664, |
| "loss": 0.3964, |
| "nll_loss": 0.23712339997291565, |
| "rewards/accuracies": 0.9674999713897705, |
| "rewards/chosen": -3.954303026199341, |
| "rewards/margins": 3.511564254760742, |
| "rewards/rejected": -7.46586799621582, |
| "step": 3650 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 11.845396041870117, |
| "learning_rate": 5.633145734114665e-09, |
| "logits/chosen": 1.4899775981903076, |
| "logits/rejected": 1.533370018005371, |
| "logps/chosen": -38.53838348388672, |
| "logps/rejected": -71.89134979248047, |
| "loss": 0.4058, |
| "nll_loss": 0.23659667372703552, |
| "rewards/accuracies": 0.9674999713897705, |
| "rewards/chosen": -3.853839159011841, |
| "rewards/margins": 3.3352959156036377, |
| "rewards/rejected": -7.18913459777832, |
| "step": 3700 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 9.659139633178711, |
| "learning_rate": 2.1661681620654963e-12, |
| "logits/chosen": 1.4696215391159058, |
| "logits/rejected": 1.526814579963684, |
| "logps/chosen": -38.22508239746094, |
| "logps/rejected": -72.54460144042969, |
| "loss": 0.3938, |
| "nll_loss": 0.22828102111816406, |
| "rewards/accuracies": 0.9637500047683716, |
| "rewards/chosen": -3.8225083351135254, |
| "rewards/margins": 3.4319519996643066, |
| "rewards/rejected": -7.254461288452148, |
| "step": 3750 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 3750, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 200, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|