| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9995638901003053, |
| "eval_steps": 100, |
| "global_step": 573, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0017444395987788923, |
| "grad_norm": 29.747415913018244, |
| "learning_rate": 8.620689655172413e-09, |
| "logits/chosen": -2.9494614601135254, |
| "logits/rejected": -2.9794743061065674, |
| "logps/chosen": -1.7114274501800537, |
| "logps/rejected": -1.7004308700561523, |
| "loss": 1.5625, |
| "rewards/accuracies": 0.0, |
| "rewards/chosen": 0.0, |
| "rewards/margins": 0.0, |
| "rewards/rejected": 0.0, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.01744439598778892, |
| "grad_norm": 25.590469762768826, |
| "learning_rate": 8.620689655172414e-08, |
| "logits/chosen": -2.272578477859497, |
| "logits/rejected": -2.341148614883423, |
| "logps/chosen": -1.5401281118392944, |
| "logps/rejected": -1.708099126815796, |
| "loss": 1.5627, |
| "rewards/accuracies": 0.4513888955116272, |
| "rewards/chosen": 0.000227387499762699, |
| "rewards/margins": 0.0002251576370326802, |
| "rewards/rejected": 2.2298513613350224e-06, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.03488879197557784, |
| "grad_norm": 25.207047594807733, |
| "learning_rate": 1.7241379310344828e-07, |
| "logits/chosen": -2.485241413116455, |
| "logits/rejected": -2.5959410667419434, |
| "logps/chosen": -1.570024013519287, |
| "logps/rejected": -1.7283653020858765, |
| "loss": 1.5607, |
| "rewards/accuracies": 0.5, |
| "rewards/chosen": -0.00038363353814929724, |
| "rewards/margins": -9.226792826666497e-06, |
| "rewards/rejected": -0.0003744067216757685, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.05233318796336677, |
| "grad_norm": 31.534894880772583, |
| "learning_rate": 2.586206896551724e-07, |
| "logits/chosen": -2.1619415283203125, |
| "logits/rejected": -2.2504525184631348, |
| "logps/chosen": -1.6232503652572632, |
| "logps/rejected": -1.7250868082046509, |
| "loss": 1.5432, |
| "rewards/accuracies": 0.637499988079071, |
| "rewards/chosen": -0.0028999350033700466, |
| "rewards/margins": 0.0026171673089265823, |
| "rewards/rejected": -0.005517102312296629, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.06977758395115569, |
| "grad_norm": 26.41926451522702, |
| "learning_rate": 3.4482758620689656e-07, |
| "logits/chosen": -2.274902582168579, |
| "logits/rejected": -2.4139533042907715, |
| "logps/chosen": -1.5764671564102173, |
| "logps/rejected": -1.6893203258514404, |
| "loss": 1.509, |
| "rewards/accuracies": 0.7124999761581421, |
| "rewards/chosen": -0.007333276327699423, |
| "rewards/margins": 0.00940671842545271, |
| "rewards/rejected": -0.016739998012781143, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.08722197993894461, |
| "grad_norm": 24.178200461155868, |
| "learning_rate": 4.310344827586206e-07, |
| "logits/chosen": -2.3129639625549316, |
| "logits/rejected": -2.319011926651001, |
| "logps/chosen": -1.5989129543304443, |
| "logps/rejected": -1.7941827774047852, |
| "loss": 1.4196, |
| "rewards/accuracies": 0.737500011920929, |
| "rewards/chosen": -0.026371877640485764, |
| "rewards/margins": 0.03227262943983078, |
| "rewards/rejected": -0.058644503355026245, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.10466637592673354, |
| "grad_norm": 38.93662482016604, |
| "learning_rate": 4.999813941333237e-07, |
| "logits/chosen": -2.455597400665283, |
| "logits/rejected": -2.459132194519043, |
| "logps/chosen": -1.9030494689941406, |
| "logps/rejected": -2.0835158824920654, |
| "loss": 1.3361, |
| "rewards/accuracies": 0.699999988079071, |
| "rewards/chosen": -0.10062487423419952, |
| "rewards/margins": 0.051896799355745316, |
| "rewards/rejected": -0.15252165496349335, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.12211077191452246, |
| "grad_norm": 35.7973285976575, |
| "learning_rate": 4.99330479543236e-07, |
| "logits/chosen": -2.394988536834717, |
| "logits/rejected": -2.4527833461761475, |
| "logps/chosen": -1.8048909902572632, |
| "logps/rejected": -2.0864245891571045, |
| "loss": 1.3198, |
| "rewards/accuracies": 0.668749988079071, |
| "rewards/chosen": -0.11194689571857452, |
| "rewards/margins": 0.057884473353624344, |
| "rewards/rejected": -0.16983136534690857, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.13955516790231137, |
| "grad_norm": 40.16495399833959, |
| "learning_rate": 4.977520391931686e-07, |
| "logits/chosen": -2.266960620880127, |
| "logits/rejected": -2.373418092727661, |
| "logps/chosen": -1.8538503646850586, |
| "logps/rejected": -2.1406590938568115, |
| "loss": 1.2319, |
| "rewards/accuracies": 0.699999988079071, |
| "rewards/chosen": -0.1073283776640892, |
| "rewards/margins": 0.09412004053592682, |
| "rewards/rejected": -0.2014484405517578, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.1569995638901003, |
| "grad_norm": 46.075075778371485, |
| "learning_rate": 4.952519449848961e-07, |
| "logits/chosen": -2.337573289871216, |
| "logits/rejected": -2.3125760555267334, |
| "logps/chosen": -2.0118062496185303, |
| "logps/rejected": -2.2410480976104736, |
| "loss": 1.187, |
| "rewards/accuracies": 0.7562500238418579, |
| "rewards/chosen": -0.15836110711097717, |
| "rewards/margins": 0.09043124318122864, |
| "rewards/rejected": -0.248792365193367, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.17444395987788922, |
| "grad_norm": 35.52679424110663, |
| "learning_rate": 4.9183949743308e-07, |
| "logits/chosen": -2.266244411468506, |
| "logits/rejected": -2.32613205909729, |
| "logps/chosen": -1.9573999643325806, |
| "logps/rejected": -2.245410919189453, |
| "loss": 1.1749, |
| "rewards/accuracies": 0.7250000238418579, |
| "rewards/chosen": -0.15664055943489075, |
| "rewards/margins": 0.1040625348687172, |
| "rewards/rejected": -0.26070311665534973, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.17444395987788922, |
| "eval_logits/chosen": -2.5133261680603027, |
| "eval_logits/rejected": -2.567593812942505, |
| "eval_logps/chosen": -2.063753604888916, |
| "eval_logps/rejected": -2.446528673171997, |
| "eval_loss": 1.0763452053070068, |
| "eval_rewards/accuracies": 0.7891566157341003, |
| "eval_rewards/chosen": -0.17315217852592468, |
| "eval_rewards/margins": 0.1388266682624817, |
| "eval_rewards/rejected": -0.31197884678840637, |
| "eval_runtime": 115.6498, |
| "eval_samples_per_second": 22.853, |
| "eval_steps_per_second": 0.718, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.19188835586567815, |
| "grad_norm": 34.94980526994547, |
| "learning_rate": 4.875273910667434e-07, |
| "logits/chosen": -2.5348660945892334, |
| "logits/rejected": -2.541141986846924, |
| "logps/chosen": -2.0890326499938965, |
| "logps/rejected": -2.563629627227783, |
| "loss": 1.0846, |
| "rewards/accuracies": 0.75, |
| "rewards/chosen": -0.21275702118873596, |
| "rewards/margins": 0.15415263175964355, |
| "rewards/rejected": -0.3669096529483795, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.2093327518534671, |
| "grad_norm": 36.111192447827186, |
| "learning_rate": 4.823316672047889e-07, |
| "logits/chosen": -2.482753276824951, |
| "logits/rejected": -2.542470932006836, |
| "logps/chosen": -2.19625186920166, |
| "logps/rejected": -2.7020630836486816, |
| "loss": 1.0846, |
| "rewards/accuracies": 0.75, |
| "rewards/chosen": -0.20946833491325378, |
| "rewards/margins": 0.1580672562122345, |
| "rewards/rejected": -0.3675355613231659, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.226777147841256, |
| "grad_norm": 35.9198962249818, |
| "learning_rate": 4.762716542812394e-07, |
| "logits/chosen": -2.5084481239318848, |
| "logits/rejected": -2.484107732772827, |
| "logps/chosen": -2.347628116607666, |
| "logps/rejected": -2.8013808727264404, |
| "loss": 1.0397, |
| "rewards/accuracies": 0.768750011920929, |
| "rewards/chosen": -0.2825542390346527, |
| "rewards/margins": 0.14921075105667114, |
| "rewards/rejected": -0.43176499009132385, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.24422154382904493, |
| "grad_norm": 33.812969950120994, |
| "learning_rate": 4.693698959421934e-07, |
| "logits/chosen": -2.4390926361083984, |
| "logits/rejected": -2.5850162506103516, |
| "logps/chosen": -2.2307400703430176, |
| "logps/rejected": -2.76652193069458, |
| "loss": 1.1052, |
| "rewards/accuracies": 0.699999988079071, |
| "rewards/chosen": -0.2658195495605469, |
| "rewards/margins": 0.15902426838874817, |
| "rewards/rejected": -0.42484384775161743, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.26166593981683384, |
| "grad_norm": 38.15404463326893, |
| "learning_rate": 4.616520671819811e-07, |
| "logits/chosen": -2.5228819847106934, |
| "logits/rejected": -2.5446176528930664, |
| "logps/chosen": -2.410109043121338, |
| "logps/rejected": -2.782252550125122, |
| "loss": 1.0492, |
| "rewards/accuracies": 0.7437499761581421, |
| "rewards/chosen": -0.31940045952796936, |
| "rewards/margins": 0.1365361362695694, |
| "rewards/rejected": -0.45593661069869995, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.27911033580462274, |
| "grad_norm": 37.67274567595706, |
| "learning_rate": 4.531468788304991e-07, |
| "logits/chosen": -2.496253252029419, |
| "logits/rejected": -2.5734314918518066, |
| "logps/chosen": -2.237454414367676, |
| "logps/rejected": -2.7140257358551025, |
| "loss": 1.0272, |
| "rewards/accuracies": 0.7749999761581421, |
| "rewards/chosen": -0.27850186824798584, |
| "rewards/margins": 0.15557342767715454, |
| "rewards/rejected": -0.4340752959251404, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.2965547317924117, |
| "grad_norm": 49.34543198418806, |
| "learning_rate": 4.438859707470375e-07, |
| "logits/chosen": -2.530522108078003, |
| "logits/rejected": -2.5276732444763184, |
| "logps/chosen": -2.332202434539795, |
| "logps/rejected": -2.7523207664489746, |
| "loss": 1.0392, |
| "rewards/accuracies": 0.78125, |
| "rewards/chosen": -0.2833098769187927, |
| "rewards/margins": 0.1833934336900711, |
| "rewards/rejected": -0.466703325510025, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.3139991277802006, |
| "grad_norm": 44.17730811384066, |
| "learning_rate": 4.3390379411792524e-07, |
| "logits/chosen": -2.4329633712768555, |
| "logits/rejected": -2.5291526317596436, |
| "logps/chosen": -2.313528537750244, |
| "logps/rejected": -2.7951102256774902, |
| "loss": 1.0006, |
| "rewards/accuracies": 0.78125, |
| "rewards/chosen": -0.3000316917896271, |
| "rewards/margins": 0.158560112118721, |
| "rewards/rejected": -0.4585917890071869, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.3314435237679895, |
| "grad_norm": 47.45322605710491, |
| "learning_rate": 4.23237483295854e-07, |
| "logits/chosen": -2.573887586593628, |
| "logits/rejected": -2.7052321434020996, |
| "logps/chosen": -2.364560127258301, |
| "logps/rejected": -2.8203182220458984, |
| "loss": 1.0715, |
| "rewards/accuracies": 0.793749988079071, |
| "rewards/chosen": -0.3148408532142639, |
| "rewards/margins": 0.16490553319454193, |
| "rewards/rejected": -0.47974634170532227, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.34888791975577843, |
| "grad_norm": 36.54817057924313, |
| "learning_rate": 4.119267176576475e-07, |
| "logits/chosen": -2.688988208770752, |
| "logits/rejected": -2.7872579097747803, |
| "logps/chosen": -2.3317620754241943, |
| "logps/rejected": -2.948866844177246, |
| "loss": 0.9802, |
| "rewards/accuracies": 0.793749988079071, |
| "rewards/chosen": -0.3193766474723816, |
| "rewards/margins": 0.1907302290201187, |
| "rewards/rejected": -0.5101069211959839, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.34888791975577843, |
| "eval_logits/chosen": -2.7230396270751953, |
| "eval_logits/rejected": -2.7872889041900635, |
| "eval_logps/chosen": -2.4268717765808105, |
| "eval_logps/rejected": -2.992184638977051, |
| "eval_loss": 0.9500909447669983, |
| "eval_rewards/accuracies": 0.8012048006057739, |
| "eval_rewards/chosen": -0.31839942932128906, |
| "eval_rewards/margins": 0.2118416577577591, |
| "eval_rewards/rejected": -0.530241072177887, |
| "eval_runtime": 115.5249, |
| "eval_samples_per_second": 22.878, |
| "eval_steps_per_second": 0.718, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.3663323157435674, |
| "grad_norm": 28.997542149982813, |
| "learning_rate": 4.000135739943734e-07, |
| "logits/chosen": -2.5338735580444336, |
| "logits/rejected": -2.5965075492858887, |
| "logps/chosen": -2.3301658630371094, |
| "logps/rejected": -2.923480749130249, |
| "loss": 1.017, |
| "rewards/accuracies": 0.768750011920929, |
| "rewards/chosen": -0.2939096689224243, |
| "rewards/margins": 0.18346476554870605, |
| "rewards/rejected": -0.477374404668808, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.3837767117313563, |
| "grad_norm": 79.37314946467663, |
| "learning_rate": 3.875423699829168e-07, |
| "logits/chosen": -2.5550951957702637, |
| "logits/rejected": -2.769721508026123, |
| "logps/chosen": -2.377946615219116, |
| "logps/rejected": -3.049376964569092, |
| "loss": 0.9655, |
| "rewards/accuracies": 0.824999988079071, |
| "rewards/chosen": -0.33743661642074585, |
| "rewards/margins": 0.20777985453605652, |
| "rewards/rejected": -0.5452165007591248, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.4012211077191452, |
| "grad_norm": 38.56972950284146, |
| "learning_rate": 3.7455949932131175e-07, |
| "logits/chosen": -2.853245258331299, |
| "logits/rejected": -2.7923905849456787, |
| "logps/chosen": -2.5535457134246826, |
| "logps/rejected": -3.186497688293457, |
| "loss": 1.0501, |
| "rewards/accuracies": 0.7562500238418579, |
| "rewards/chosen": -0.37545865774154663, |
| "rewards/margins": 0.21732258796691895, |
| "rewards/rejected": -0.5927812457084656, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.4186655037069342, |
| "grad_norm": 34.32127443417028, |
| "learning_rate": 3.611132591411369e-07, |
| "logits/chosen": -2.845720052719116, |
| "logits/rejected": -2.912646770477295, |
| "logps/chosen": -2.471998691558838, |
| "logps/rejected": -2.9336795806884766, |
| "loss": 0.9871, |
| "rewards/accuracies": 0.78125, |
| "rewards/chosen": -0.3210769295692444, |
| "rewards/margins": 0.18005797266960144, |
| "rewards/rejected": -0.5011348724365234, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.4361098996947231, |
| "grad_norm": 36.8304920546943, |
| "learning_rate": 3.4725367033901473e-07, |
| "logits/chosen": -2.7427520751953125, |
| "logits/rejected": -2.9484190940856934, |
| "logps/chosen": -2.521329879760742, |
| "logps/rejected": -3.1583688259124756, |
| "loss": 1.018, |
| "rewards/accuracies": 0.78125, |
| "rewards/chosen": -0.3658570647239685, |
| "rewards/margins": 0.20647919178009033, |
| "rewards/rejected": -0.5723362565040588, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.453554295682512, |
| "grad_norm": 29.383883913798257, |
| "learning_rate": 3.3303229149558967e-07, |
| "logits/chosen": -2.8211348056793213, |
| "logits/rejected": -2.8875656127929688, |
| "logps/chosen": -2.307373523712158, |
| "logps/rejected": -2.855837821960449, |
| "loss": 0.9869, |
| "rewards/accuracies": 0.8125, |
| "rewards/chosen": -0.29735279083251953, |
| "rewards/margins": 0.1880948543548584, |
| "rewards/rejected": -0.48544764518737793, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.4709986916703009, |
| "grad_norm": 37.77204908034358, |
| "learning_rate": 3.185020270742225e-07, |
| "logits/chosen": -2.7716403007507324, |
| "logits/rejected": -2.8079323768615723, |
| "logps/chosen": -2.386676549911499, |
| "logps/rejected": -2.7951130867004395, |
| "loss": 0.9871, |
| "rewards/accuracies": 0.71875, |
| "rewards/chosen": -0.32528942823410034, |
| "rewards/margins": 0.15507872402668, |
| "rewards/rejected": -0.48036813735961914, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.48844308765808986, |
| "grad_norm": 28.328010250614053, |
| "learning_rate": 3.0371693061291146e-07, |
| "logits/chosen": -2.780062437057495, |
| "logits/rejected": -2.9711029529571533, |
| "logps/chosen": -2.551729679107666, |
| "logps/rejected": -3.2051796913146973, |
| "loss": 1.0056, |
| "rewards/accuracies": 0.768750011920929, |
| "rewards/chosen": -0.40202435851097107, |
| "rewards/margins": 0.1922958493232727, |
| "rewards/rejected": -0.5943201780319214, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.5058874836458788, |
| "grad_norm": 32.14871116699808, |
| "learning_rate": 2.8873200364158106e-07, |
| "logits/chosen": -2.680164337158203, |
| "logits/rejected": -2.803938388824463, |
| "logps/chosen": -2.43742036819458, |
| "logps/rejected": -2.9839630126953125, |
| "loss": 0.9837, |
| "rewards/accuracies": 0.831250011920929, |
| "rewards/chosen": -0.32910409569740295, |
| "rewards/margins": 0.2097328007221222, |
| "rewards/rejected": -0.5388368368148804, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.5233318796336677, |
| "grad_norm": 40.52094893481022, |
| "learning_rate": 2.7360299107277767e-07, |
| "logits/chosen": -2.7708568572998047, |
| "logits/rejected": -2.821788787841797, |
| "logps/chosen": -2.4744372367858887, |
| "logps/rejected": -3.086942672729492, |
| "loss": 0.9548, |
| "rewards/accuracies": 0.793749988079071, |
| "rewards/chosen": -0.3568132221698761, |
| "rewards/margins": 0.20053787529468536, |
| "rewards/rejected": -0.5573510527610779, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.5233318796336677, |
| "eval_logits/chosen": -2.808656692504883, |
| "eval_logits/rejected": -2.8788304328918457, |
| "eval_logps/chosen": -2.571028709411621, |
| "eval_logps/rejected": -3.17358660697937, |
| "eval_loss": 0.9136135578155518, |
| "eval_rewards/accuracies": 0.8162650465965271, |
| "eval_rewards/chosen": -0.3760621249675751, |
| "eval_rewards/margins": 0.22673983871936798, |
| "eval_rewards/rejected": -0.6028019785881042, |
| "eval_runtime": 115.45, |
| "eval_samples_per_second": 22.893, |
| "eval_steps_per_second": 0.719, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.5407762756214566, |
| "grad_norm": 39.89301839500131, |
| "learning_rate": 2.5838617382693414e-07, |
| "logits/chosen": -2.948072671890259, |
| "logits/rejected": -2.9336767196655273, |
| "logps/chosen": -2.5350139141082764, |
| "logps/rejected": -3.0460660457611084, |
| "loss": 1.0184, |
| "rewards/accuracies": 0.78125, |
| "rewards/chosen": -0.37652358412742615, |
| "rewards/margins": 0.18868541717529297, |
| "rewards/rejected": -0.5652090311050415, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.5582206716092455, |
| "grad_norm": 36.83353696515594, |
| "learning_rate": 2.4313815946364877e-07, |
| "logits/chosen": -2.8387434482574463, |
| "logits/rejected": -2.978883743286133, |
| "logps/chosen": -2.4741597175598145, |
| "logps/rejected": -3.0683841705322266, |
| "loss": 0.9653, |
| "rewards/accuracies": 0.7437499761581421, |
| "rewards/chosen": -0.3760759234428406, |
| "rewards/margins": 0.1771194189786911, |
| "rewards/rejected": -0.5531953573226929, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.5756650675970345, |
| "grad_norm": 30.76657514726095, |
| "learning_rate": 2.2791567159784566e-07, |
| "logits/chosen": -2.762192487716675, |
| "logits/rejected": -2.7734150886535645, |
| "logps/chosen": -2.4779891967773438, |
| "logps/rejected": -2.9213876724243164, |
| "loss": 0.9814, |
| "rewards/accuracies": 0.7875000238418579, |
| "rewards/chosen": -0.3409067988395691, |
| "rewards/margins": 0.18150492012500763, |
| "rewards/rejected": -0.5224117040634155, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.5931094635848234, |
| "grad_norm": 32.87393386922799, |
| "learning_rate": 2.12775338884202e-07, |
| "logits/chosen": -2.886357307434082, |
| "logits/rejected": -2.9018540382385254, |
| "logps/chosen": -2.4813778400421143, |
| "logps/rejected": -2.9721169471740723, |
| "loss": 0.9893, |
| "rewards/accuracies": 0.793749988079071, |
| "rewards/chosen": -0.3222178816795349, |
| "rewards/margins": 0.20408673584461212, |
| "rewards/rejected": -0.5263046622276306, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.6105538595726123, |
| "grad_norm": 30.494923835102345, |
| "learning_rate": 1.977734843548328e-07, |
| "logits/chosen": -2.824707508087158, |
| "logits/rejected": -2.79854154586792, |
| "logps/chosen": -2.3759117126464844, |
| "logps/rejected": -2.975604295730591, |
| "loss": 0.9863, |
| "rewards/accuracies": 0.824999988079071, |
| "rewards/chosen": -0.3144580125808716, |
| "rewards/margins": 0.21930566430091858, |
| "rewards/rejected": -0.5337637066841125, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.6279982555604012, |
| "grad_norm": 26.8742332382294, |
| "learning_rate": 1.8296591589391226e-07, |
| "logits/chosen": -2.6628756523132324, |
| "logits/rejected": -2.7284369468688965, |
| "logps/chosen": -2.4052791595458984, |
| "logps/rejected": -2.870692253112793, |
| "loss": 0.9646, |
| "rewards/accuracies": 0.7250000238418579, |
| "rewards/chosen": -0.310857355594635, |
| "rewards/margins": 0.15339362621307373, |
| "rewards/rejected": -0.46425098180770874, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.6454426515481901, |
| "grad_norm": 35.52686655329713, |
| "learning_rate": 1.684077186286792e-07, |
| "logits/chosen": -2.50087308883667, |
| "logits/rejected": -2.6514532566070557, |
| "logps/chosen": -2.2871201038360596, |
| "logps/rejected": -2.937791585922241, |
| "loss": 0.9877, |
| "rewards/accuracies": 0.7875000238418579, |
| "rewards/chosen": -0.3152330219745636, |
| "rewards/margins": 0.2090420424938202, |
| "rewards/rejected": -0.5242750644683838, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.662887047535979, |
| "grad_norm": 30.399222288686712, |
| "learning_rate": 1.5415305000914585e-07, |
| "logits/chosen": -2.7632899284362793, |
| "logits/rejected": -2.886943817138672, |
| "logps/chosen": -2.4425692558288574, |
| "logps/rejected": -3.0246245861053467, |
| "loss": 0.9226, |
| "rewards/accuracies": 0.793749988079071, |
| "rewards/chosen": -0.33751311898231506, |
| "rewards/margins": 0.1871272623538971, |
| "rewards/rejected": -0.5246403813362122, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.680331443523768, |
| "grad_norm": 33.76165947663638, |
| "learning_rate": 1.4025493833882643e-07, |
| "logits/chosen": -2.622499465942383, |
| "logits/rejected": -2.6712939739227295, |
| "logps/chosen": -2.521423816680908, |
| "logps/rejected": -3.01672625541687, |
| "loss": 0.9625, |
| "rewards/accuracies": 0.75, |
| "rewards/chosen": -0.39204803109169006, |
| "rewards/margins": 0.18184103071689606, |
| "rewards/rejected": -0.5738890767097473, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.6977758395115569, |
| "grad_norm": 39.46957147631396, |
| "learning_rate": 1.267650855059656e-07, |
| "logits/chosen": -2.7575411796569824, |
| "logits/rejected": -2.8080406188964844, |
| "logps/chosen": -2.511035203933716, |
| "logps/rejected": -2.9791462421417236, |
| "loss": 0.9834, |
| "rewards/accuracies": 0.78125, |
| "rewards/chosen": -0.3435077667236328, |
| "rewards/margins": 0.1896781176328659, |
| "rewards/rejected": -0.5331858396530151, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.6977758395115569, |
| "eval_logits/chosen": -2.8667454719543457, |
| "eval_logits/rejected": -2.937051773071289, |
| "eval_logps/chosen": -2.4769790172576904, |
| "eval_logps/rejected": -3.0509371757507324, |
| "eval_loss": 0.9041184186935425, |
| "eval_rewards/accuracies": 0.8042168617248535, |
| "eval_rewards/chosen": -0.3384423553943634, |
| "eval_rewards/margins": 0.21529993414878845, |
| "eval_rewards/rejected": -0.5537422895431519, |
| "eval_runtime": 115.5756, |
| "eval_samples_per_second": 22.868, |
| "eval_steps_per_second": 0.718, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.7152202354993459, |
| "grad_norm": 35.58611489394968, |
| "learning_rate": 1.1373367464911796e-07, |
| "logits/chosen": -2.7775580883026123, |
| "logits/rejected": -2.8066816329956055, |
| "logps/chosen": -2.44441294670105, |
| "logps/rejected": -2.940340042114258, |
| "loss": 0.966, |
| "rewards/accuracies": 0.800000011920929, |
| "rewards/chosen": -0.3467921018600464, |
| "rewards/margins": 0.17160984873771667, |
| "rewards/rejected": -0.5184019804000854, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.7326646314871348, |
| "grad_norm": 31.573311664278364, |
| "learning_rate": 1.0120918347257668e-07, |
| "logits/chosen": -2.912554979324341, |
| "logits/rejected": -2.9200565814971924, |
| "logps/chosen": -2.5111441612243652, |
| "logps/rejected": -3.0291829109191895, |
| "loss": 1.0018, |
| "rewards/accuracies": 0.800000011920929, |
| "rewards/chosen": -0.3495512306690216, |
| "rewards/margins": 0.18085847795009613, |
| "rewards/rejected": -0.5304096937179565, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.7501090274749237, |
| "grad_norm": 38.092625083892315, |
| "learning_rate": 8.923820390612991e-08, |
| "logits/chosen": -2.8513288497924805, |
| "logits/rejected": -2.847677707672119, |
| "logps/chosen": -2.44001841545105, |
| "logps/rejected": -3.063814640045166, |
| "loss": 0.9975, |
| "rewards/accuracies": 0.84375, |
| "rewards/chosen": -0.3544333577156067, |
| "rewards/margins": 0.22024652361869812, |
| "rewards/rejected": -0.5746799111366272, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.7675534234627126, |
| "grad_norm": 30.479406895571362, |
| "learning_rate": 7.786526878002125e-08, |
| "logits/chosen": -2.714656114578247, |
| "logits/rejected": -2.838855743408203, |
| "logps/chosen": -2.3737549781799316, |
| "logps/rejected": -3.0006814002990723, |
| "loss": 0.9624, |
| "rewards/accuracies": 0.793749988079071, |
| "rewards/chosen": -0.32796674966812134, |
| "rewards/margins": 0.20443764328956604, |
| "rewards/rejected": -0.532404363155365, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.7849978194505015, |
| "grad_norm": 33.130716501269866, |
| "learning_rate": 6.713268615989654e-08, |
| "logits/chosen": -2.9494526386260986, |
| "logits/rejected": -2.9998550415039062, |
| "logps/chosen": -2.509775161743164, |
| "logps/rejected": -3.015613555908203, |
| "loss": 0.9668, |
| "rewards/accuracies": 0.762499988079071, |
| "rewards/chosen": -0.36944177746772766, |
| "rewards/margins": 0.1895218938589096, |
| "rewards/rejected": -0.5589635968208313, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.8024422154382904, |
| "grad_norm": 33.25889406980063, |
| "learning_rate": 5.7080381958020976e-08, |
| "logits/chosen": -2.674802303314209, |
| "logits/rejected": -2.8434243202209473, |
| "logps/chosen": -2.5848803520202637, |
| "logps/rejected": -3.1880083084106445, |
| "loss": 0.9423, |
| "rewards/accuracies": 0.862500011920929, |
| "rewards/chosen": -0.376645028591156, |
| "rewards/margins": 0.20410259068012238, |
| "rewards/rejected": -0.5807476043701172, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.8198866114260793, |
| "grad_norm": 30.106792425198947, |
| "learning_rate": 4.774575140626316e-08, |
| "logits/chosen": -2.7849488258361816, |
| "logits/rejected": -2.903566360473633, |
| "logps/chosen": -2.3999593257904053, |
| "logps/rejected": -3.0301947593688965, |
| "loss": 0.9476, |
| "rewards/accuracies": 0.8062499761581421, |
| "rewards/chosen": -0.3394695520401001, |
| "rewards/margins": 0.21237297356128693, |
| "rewards/rejected": -0.5518425703048706, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.8373310074138683, |
| "grad_norm": 37.88465481105631, |
| "learning_rate": 3.9163519943375965e-08, |
| "logits/chosen": -2.8227665424346924, |
| "logits/rejected": -2.9443516731262207, |
| "logps/chosen": -2.4296436309814453, |
| "logps/rejected": -2.984886646270752, |
| "loss": 0.8745, |
| "rewards/accuracies": 0.768750011920929, |
| "rewards/chosen": -0.35570722818374634, |
| "rewards/margins": 0.16281530261039734, |
| "rewards/rejected": -0.5185225605964661, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.8547754034016573, |
| "grad_norm": 30.701274278534907, |
| "learning_rate": 3.136561403408422e-08, |
| "logits/chosen": -2.8463587760925293, |
| "logits/rejected": -3.0236496925354004, |
| "logps/chosen": -2.5174646377563477, |
| "logps/rejected": -3.2128849029541016, |
| "loss": 0.9445, |
| "rewards/accuracies": 0.8062499761581421, |
| "rewards/chosen": -0.37872380018234253, |
| "rewards/margins": 0.22148160636425018, |
| "rewards/rejected": -0.6002054810523987, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.8722197993894462, |
| "grad_norm": 32.269326624165465, |
| "learning_rate": 2.4381042400538593e-08, |
| "logits/chosen": -2.871164560317993, |
| "logits/rejected": -3.0147228240966797, |
| "logps/chosen": -2.5578746795654297, |
| "logps/rejected": -3.273369312286377, |
| "loss": 0.9967, |
| "rewards/accuracies": 0.800000011920929, |
| "rewards/chosen": -0.4019395709037781, |
| "rewards/margins": 0.21136029064655304, |
| "rewards/rejected": -0.6132999062538147, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.8722197993894462, |
| "eval_logits/chosen": -2.959242105484009, |
| "eval_logits/rejected": -3.029346227645874, |
| "eval_logps/chosen": -2.568408966064453, |
| "eval_logps/rejected": -3.185546398162842, |
| "eval_loss": 0.8938218951225281, |
| "eval_rewards/accuracies": 0.7891566157341003, |
| "eval_rewards/chosen": -0.37501412630081177, |
| "eval_rewards/margins": 0.23257188498973846, |
| "eval_rewards/rejected": -0.6075860857963562, |
| "eval_runtime": 115.5545, |
| "eval_samples_per_second": 22.872, |
| "eval_steps_per_second": 0.718, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.8896641953772351, |
| "grad_norm": 35.20455083141279, |
| "learning_rate": 1.8235788107963945e-08, |
| "logits/chosen": -2.887869358062744, |
| "logits/rejected": -2.820462703704834, |
| "logps/chosen": -2.587712526321411, |
| "logps/rejected": -3.1114964485168457, |
| "loss": 0.9135, |
| "rewards/accuracies": 0.762499988079071, |
| "rewards/chosen": -0.37796396017074585, |
| "rewards/margins": 0.18830379843711853, |
| "rewards/rejected": -0.5662677884101868, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.907108591365024, |
| "grad_norm": 29.121513812900695, |
| "learning_rate": 1.2952711905950376e-08, |
| "logits/chosen": -2.8955085277557373, |
| "logits/rejected": -2.908362627029419, |
| "logps/chosen": -2.5724124908447266, |
| "logps/rejected": -3.0688400268554688, |
| "loss": 0.9963, |
| "rewards/accuracies": 0.731249988079071, |
| "rewards/chosen": -0.39740028977394104, |
| "rewards/margins": 0.17423763871192932, |
| "rewards/rejected": -0.5716378688812256, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.9245529873528129, |
| "grad_norm": 30.97739103991955, |
| "learning_rate": 8.55146718496283e-09, |
| "logits/chosen": -2.919581890106201, |
| "logits/rejected": -2.8886361122131348, |
| "logps/chosen": -2.599856376647949, |
| "logps/rejected": -2.9770760536193848, |
| "loss": 0.9551, |
| "rewards/accuracies": 0.768750011920929, |
| "rewards/chosen": -0.3779660165309906, |
| "rewards/margins": 0.17554500699043274, |
| "rewards/rejected": -0.5535110235214233, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.9419973833406018, |
| "grad_norm": 44.940688671381224, |
| "learning_rate": 5.048426864438182e-09, |
| "logits/chosen": -2.870760440826416, |
| "logits/rejected": -2.946399688720703, |
| "logps/chosen": -2.518998146057129, |
| "logps/rejected": -3.1112277507781982, |
| "loss": 0.9636, |
| "rewards/accuracies": 0.737500011920929, |
| "rewards/chosen": -0.3970429301261902, |
| "rewards/margins": 0.18053510785102844, |
| "rewards/rejected": -0.577578067779541, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.9594417793283908, |
| "grad_norm": 33.602758621496534, |
| "learning_rate": 2.456622484449139e-09, |
| "logits/chosen": -2.9532032012939453, |
| "logits/rejected": -3.0146090984344482, |
| "logps/chosen": -2.5181241035461426, |
| "logps/rejected": -3.0724215507507324, |
| "loss": 0.9918, |
| "rewards/accuracies": 0.8187500238418579, |
| "rewards/chosen": -0.35534483194351196, |
| "rewards/margins": 0.210611030459404, |
| "rewards/rejected": -0.5659558176994324, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.9768861753161797, |
| "grad_norm": 40.276290803183436, |
| "learning_rate": 7.85695727519492e-10, |
| "logits/chosen": -2.8332512378692627, |
| "logits/rejected": -2.8909554481506348, |
| "logps/chosen": -2.6798324584960938, |
| "logps/rejected": -3.1653189659118652, |
| "loss": 0.957, |
| "rewards/accuracies": 0.7124999761581421, |
| "rewards/chosen": -0.3789713382720947, |
| "rewards/margins": 0.16679120063781738, |
| "rewards/rejected": -0.5457625389099121, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.9943305713039686, |
| "grad_norm": 44.66764267907785, |
| "learning_rate": 4.1862550931942575e-11, |
| "logits/chosen": -2.9107937812805176, |
| "logits/rejected": -2.905510425567627, |
| "logps/chosen": -2.4948418140411377, |
| "logps/rejected": -2.9094674587249756, |
| "loss": 0.965, |
| "rewards/accuracies": 0.7875000238418579, |
| "rewards/chosen": -0.3739563822746277, |
| "rewards/margins": 0.17874310910701752, |
| "rewards/rejected": -0.5526994466781616, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.9995638901003053, |
| "step": 573, |
| "total_flos": 0.0, |
| "train_loss": 1.0578667073141634, |
| "train_runtime": 8389.7514, |
| "train_samples_per_second": 8.743, |
| "train_steps_per_second": 0.068 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 573, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": false, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|