| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.847063169560399, | |
| "eval_steps": 1000, | |
| "global_step": 10000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01847063169560399, | |
| "grad_norm": 1.6779712438583374, | |
| "learning_rate": 6.6e-05, | |
| "loss": 1.6332, | |
| "mean_token_accuracy": 0.5568195793032646, | |
| "num_tokens": 121637.0, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.03694126339120798, | |
| "grad_norm": 1.073460340499878, | |
| "learning_rate": 0.00013266666666666667, | |
| "loss": 1.1787, | |
| "mean_token_accuracy": 0.6344019943475723, | |
| "num_tokens": 246217.0, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.05541189508681197, | |
| "grad_norm": 1.4162907600402832, | |
| "learning_rate": 0.00019933333333333334, | |
| "loss": 1.1193, | |
| "mean_token_accuracy": 0.6494812744855881, | |
| "num_tokens": 369036.0, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.07388252678241596, | |
| "grad_norm": 1.3762811422348022, | |
| "learning_rate": 0.000173421993904824, | |
| "loss": 1.1156, | |
| "mean_token_accuracy": 0.650577632188797, | |
| "num_tokens": 492952.0, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.09235315847801995, | |
| "grad_norm": 0.9165902137756348, | |
| "learning_rate": 0.0001550744859491231, | |
| "loss": 1.0855, | |
| "mean_token_accuracy": 0.6567107254266739, | |
| "num_tokens": 614495.0, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.11082379017362394, | |
| "grad_norm": 0.6837517619132996, | |
| "learning_rate": 0.00014153935488632152, | |
| "loss": 1.0733, | |
| "mean_token_accuracy": 0.6587592279911041, | |
| "num_tokens": 740435.0, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.12929442186922793, | |
| "grad_norm": 0.8884134292602539, | |
| "learning_rate": 0.00013102435641608367, | |
| "loss": 1.0538, | |
| "mean_token_accuracy": 0.665734943151474, | |
| "num_tokens": 863456.0, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.1477650535648319, | |
| "grad_norm": 1.0362532138824463, | |
| "learning_rate": 0.00012255110553085002, | |
| "loss": 1.0675, | |
| "mean_token_accuracy": 0.6605714571475982, | |
| "num_tokens": 988234.0, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.16623568526043592, | |
| "grad_norm": 0.7236880660057068, | |
| "learning_rate": 0.00011553425737574005, | |
| "loss": 1.0327, | |
| "mean_token_accuracy": 0.6679404038190841, | |
| "num_tokens": 1112814.0, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.1847063169560399, | |
| "grad_norm": 0.8120320439338684, | |
| "learning_rate": 0.0001095993248702382, | |
| "loss": 1.0307, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.1847063169560399, | |
| "eval_loss": 1.040740728378296, | |
| "eval_mean_token_accuracy": 0.6653746280670166, | |
| "eval_num_tokens": 1237371.0, | |
| "eval_runtime": 19.1225, | |
| "eval_samples_per_second": 52.295, | |
| "eval_steps_per_second": 6.537, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.20317694865164387, | |
| "grad_norm": 0.9145399332046509, | |
| "learning_rate": 0.00010449410169212441, | |
| "loss": 1.0233, | |
| "mean_token_accuracy": 0.6706056842207908, | |
| "num_tokens": 1359398.0, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.22164758034724788, | |
| "grad_norm": 0.9274869561195374, | |
| "learning_rate": 0.00010004169272643103, | |
| "loss": 1.0225, | |
| "mean_token_accuracy": 0.6711669725179672, | |
| "num_tokens": 1483345.0, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.24011821204285186, | |
| "grad_norm": 0.8308613300323486, | |
| "learning_rate": 9.611386626644256e-05, | |
| "loss": 1.0128, | |
| "mean_token_accuracy": 0.6720483464002609, | |
| "num_tokens": 1607214.0, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.25858884373845586, | |
| "grad_norm": 0.8858506679534912, | |
| "learning_rate": 9.261509270476351e-05, | |
| "loss": 1.0159, | |
| "mean_token_accuracy": 0.6725166749954223, | |
| "num_tokens": 1730229.0, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.27705947543405984, | |
| "grad_norm": 1.0183970928192139, | |
| "learning_rate": 8.9472548255098e-05, | |
| "loss": 1.0056, | |
| "mean_token_accuracy": 0.6739540088176728, | |
| "num_tokens": 1853944.0, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.2955301071296638, | |
| "grad_norm": 0.8774070143699646, | |
| "learning_rate": 8.662961636484199e-05, | |
| "loss": 1.0046, | |
| "mean_token_accuracy": 0.6739509409666061, | |
| "num_tokens": 1980091.0, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.3140007388252678, | |
| "grad_norm": 0.8079282641410828, | |
| "learning_rate": 8.40415267738742e-05, | |
| "loss": 0.9995, | |
| "mean_token_accuracy": 0.6781503772735595, | |
| "num_tokens": 2104193.0, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.33247137052087183, | |
| "grad_norm": 0.6845090389251709, | |
| "learning_rate": 8.167234800792304e-05, | |
| "loss": 1.0119, | |
| "mean_token_accuracy": 0.671802139878273, | |
| "num_tokens": 2228642.0, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.3509420022164758, | |
| "grad_norm": 0.7049577832221985, | |
| "learning_rate": 7.949286335171643e-05, | |
| "loss": 1.0254, | |
| "mean_token_accuracy": 0.666619479060173, | |
| "num_tokens": 2352090.0, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.3694126339120798, | |
| "grad_norm": 0.7556240558624268, | |
| "learning_rate": 7.747903910575024e-05, | |
| "loss": 1.0133, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.3694126339120798, | |
| "eval_loss": 0.9894875288009644, | |
| "eval_mean_token_accuracy": 0.6793018989562988, | |
| "eval_num_tokens": 2475429.0, | |
| "eval_runtime": 19.1453, | |
| "eval_samples_per_second": 52.232, | |
| "eval_steps_per_second": 6.529, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.38788326560768377, | |
| "grad_norm": 0.9750986695289612, | |
| "learning_rate": 7.561089934060305e-05, | |
| "loss": 0.9941, | |
| "mean_token_accuracy": 0.6748151290416717, | |
| "num_tokens": 2600104.0, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.40635389730328775, | |
| "grad_norm": 0.7754963040351868, | |
| "learning_rate": 7.387168551531178e-05, | |
| "loss": 1.0045, | |
| "mean_token_accuracy": 0.6771643495559693, | |
| "num_tokens": 2724139.0, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.4248245289988918, | |
| "grad_norm": 0.6899317502975464, | |
| "learning_rate": 7.224721947627513e-05, | |
| "loss": 0.9987, | |
| "mean_token_accuracy": 0.6749349737167358, | |
| "num_tokens": 2847524.0, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.44329516069449576, | |
| "grad_norm": 0.6299494504928589, | |
| "learning_rate": 7.07254141150883e-05, | |
| "loss": 0.9945, | |
| "mean_token_accuracy": 0.6736445528268814, | |
| "num_tokens": 2972109.0, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.46176579239009974, | |
| "grad_norm": 1.4914665222167969, | |
| "learning_rate": 6.929589286752371e-05, | |
| "loss": 0.9911, | |
| "mean_token_accuracy": 0.6778055882453918, | |
| "num_tokens": 3094757.0, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.4802364240857037, | |
| "grad_norm": 0.8584954142570496, | |
| "learning_rate": 6.794969055356698e-05, | |
| "loss": 0.9822, | |
| "mean_token_accuracy": 0.6790356373786927, | |
| "num_tokens": 3217566.0, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.4987070557813077, | |
| "grad_norm": 0.8136039972305298, | |
| "learning_rate": 6.667901577609308e-05, | |
| "loss": 0.996, | |
| "mean_token_accuracy": 0.6747825038433075, | |
| "num_tokens": 3342407.0, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.5171776874769117, | |
| "grad_norm": 0.7587074637413025, | |
| "learning_rate": 6.547706044716512e-05, | |
| "loss": 0.9825, | |
| "mean_token_accuracy": 0.6811250519752502, | |
| "num_tokens": 3467856.0, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.5356483191725157, | |
| "grad_norm": 0.6996003985404968, | |
| "learning_rate": 6.433784577873342e-05, | |
| "loss": 0.9715, | |
| "mean_token_accuracy": 0.6821656405925751, | |
| "num_tokens": 3589644.0, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.5541189508681197, | |
| "grad_norm": 0.9794346690177917, | |
| "learning_rate": 6.325609676486509e-05, | |
| "loss": 0.9815, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.5541189508681197, | |
| "eval_loss": 0.9679494500160217, | |
| "eval_mean_token_accuracy": 0.6865276889801025, | |
| "eval_num_tokens": 3712997.0, | |
| "eval_runtime": 19.1716, | |
| "eval_samples_per_second": 52.16, | |
| "eval_steps_per_second": 6.52, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.5725895825637237, | |
| "grad_norm": 0.8178902864456177, | |
| "learning_rate": 6.22271391287055e-05, | |
| "loss": 0.9764, | |
| "mean_token_accuracy": 0.681132504940033, | |
| "num_tokens": 3834991.0, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.5910602142593276, | |
| "grad_norm": 0.9663840532302856, | |
| "learning_rate": 6.12468141320462e-05, | |
| "loss": 0.9902, | |
| "mean_token_accuracy": 0.6795903497934341, | |
| "num_tokens": 3956462.0, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.6095308459549317, | |
| "grad_norm": 0.8016713857650757, | |
| "learning_rate": 6.03114077000131e-05, | |
| "loss": 0.9787, | |
| "mean_token_accuracy": 0.680009834766388, | |
| "num_tokens": 4080142.0, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.6280014776505356, | |
| "grad_norm": 0.718947172164917, | |
| "learning_rate": 5.9417591102230663e-05, | |
| "loss": 0.9921, | |
| "mean_token_accuracy": 0.6754868066310883, | |
| "num_tokens": 4203046.0, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.6464721093461396, | |
| "grad_norm": 0.6251752376556396, | |
| "learning_rate": 5.856237102757652e-05, | |
| "loss": 0.9614, | |
| "mean_token_accuracy": 0.6870092713832855, | |
| "num_tokens": 4326548.0, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.6649427410417437, | |
| "grad_norm": 0.9118708372116089, | |
| "learning_rate": 5.7743047343661814e-05, | |
| "loss": 0.968, | |
| "mean_token_accuracy": 0.6833401465415955, | |
| "num_tokens": 4449485.0, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.6834133727373476, | |
| "grad_norm": 0.9887173771858215, | |
| "learning_rate": 5.6957177181117404e-05, | |
| "loss": 0.9675, | |
| "mean_token_accuracy": 0.6842365640401841, | |
| "num_tokens": 4572271.0, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.7018840044329516, | |
| "grad_norm": 1.0056270360946655, | |
| "learning_rate": 5.620254425309578e-05, | |
| "loss": 0.9921, | |
| "mean_token_accuracy": 0.6806173902750016, | |
| "num_tokens": 4694319.0, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.7203546361285555, | |
| "grad_norm": 0.6377823352813721, | |
| "learning_rate": 5.547713253139649e-05, | |
| "loss": 0.9678, | |
| "mean_token_accuracy": 0.6819530457258225, | |
| "num_tokens": 4818201.0, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.7388252678241596, | |
| "grad_norm": 0.9543529748916626, | |
| "learning_rate": 5.477910356647767e-05, | |
| "loss": 0.9731, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.7388252678241596, | |
| "eval_loss": 0.951152503490448, | |
| "eval_mean_token_accuracy": 0.6907490072250366, | |
| "eval_num_tokens": 4940870.0, | |
| "eval_runtime": 19.1874, | |
| "eval_samples_per_second": 52.117, | |
| "eval_steps_per_second": 6.515, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.7572958995197636, | |
| "grad_norm": 0.9113274216651917, | |
| "learning_rate": 5.410677686985887e-05, | |
| "loss": 0.9524, | |
| "mean_token_accuracy": 0.6853961995244027, | |
| "num_tokens": 5065279.0, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.7757665312153675, | |
| "grad_norm": 0.7836439609527588, | |
| "learning_rate": 5.345861288192786e-05, | |
| "loss": 0.9759, | |
| "mean_token_accuracy": 0.6834134519100189, | |
| "num_tokens": 5190358.0, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.7942371629109716, | |
| "grad_norm": 0.8732810020446777, | |
| "learning_rate": 5.283319813188472e-05, | |
| "loss": 0.9675, | |
| "mean_token_accuracy": 0.6815478146076203, | |
| "num_tokens": 5314968.0, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.8127077946065755, | |
| "grad_norm": 0.7674530148506165, | |
| "learning_rate": 5.222923226400155e-05, | |
| "loss": 0.9523, | |
| "mean_token_accuracy": 0.686726200580597, | |
| "num_tokens": 5438772.0, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.8311784263021795, | |
| "grad_norm": 0.8453629612922668, | |
| "learning_rate": 5.164551665900703e-05, | |
| "loss": 0.9633, | |
| "mean_token_accuracy": 0.6856222760677337, | |
| "num_tokens": 5560918.0, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.8496490579977836, | |
| "grad_norm": 1.1269229650497437, | |
| "learning_rate": 5.1080944423879696e-05, | |
| "loss": 0.9581, | |
| "mean_token_accuracy": 0.6846688747406006, | |
| "num_tokens": 5682900.0, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.8681196896933875, | |
| "grad_norm": 0.7400866746902466, | |
| "learning_rate": 5.053449155971992e-05, | |
| "loss": 0.9593, | |
| "mean_token_accuracy": 0.6853216868638993, | |
| "num_tokens": 5807988.0, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.8865903213889915, | |
| "grad_norm": 0.8142501711845398, | |
| "learning_rate": 5.0005209147276734e-05, | |
| "loss": 0.9536, | |
| "mean_token_accuracy": 0.6875155121088028, | |
| "num_tokens": 5931817.0, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.9050609530845954, | |
| "grad_norm": 0.694760262966156, | |
| "learning_rate": 4.949221641439499e-05, | |
| "loss": 0.958, | |
| "mean_token_accuracy": 0.6865524923801423, | |
| "num_tokens": 6055690.0, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.9235315847801995, | |
| "grad_norm": 0.7349626421928406, | |
| "learning_rate": 4.899469457011854e-05, | |
| "loss": 0.9675, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.9235315847801995, | |
| "eval_loss": 0.9417961835861206, | |
| "eval_mean_token_accuracy": 0.6922818598747253, | |
| "eval_num_tokens": 6179818.0, | |
| "eval_runtime": 19.1775, | |
| "eval_samples_per_second": 52.144, | |
| "eval_steps_per_second": 6.518, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.9420022164758035, | |
| "grad_norm": 0.8031660914421082, | |
| "learning_rate": 4.851188130722481e-05, | |
| "loss": 0.9568, | |
| "mean_token_accuracy": 0.6855992400646209, | |
| "num_tokens": 6304542.0, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.9604728481714074, | |
| "grad_norm": 0.8529698252677917, | |
| "learning_rate": 4.804306588920635e-05, | |
| "loss": 0.9631, | |
| "mean_token_accuracy": 0.685149707198143, | |
| "num_tokens": 6426250.0, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.9789434798670115, | |
| "grad_norm": 0.7942742705345154, | |
| "learning_rate": 4.758758474966023e-05, | |
| "loss": 0.9579, | |
| "mean_token_accuracy": 0.6849185460805893, | |
| "num_tokens": 6549483.0, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.9974141115626154, | |
| "grad_norm": 0.8027817010879517, | |
| "learning_rate": 4.7144817542100825e-05, | |
| "loss": 0.9658, | |
| "mean_token_accuracy": 0.684761552810669, | |
| "num_tokens": 6672929.0, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.0158847432582194, | |
| "grad_norm": 0.4089404344558716, | |
| "learning_rate": 4.671418358670517e-05, | |
| "loss": 0.8773, | |
| "mean_token_accuracy": 0.7085312277078628, | |
| "num_tokens": 6800786.0, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.0343553749538235, | |
| "grad_norm": 0.39767757058143616, | |
| "learning_rate": 4.6295138667698956e-05, | |
| "loss": 0.8759, | |
| "mean_token_accuracy": 0.7036858803033829, | |
| "num_tokens": 6924011.0, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.0528260066494275, | |
| "grad_norm": 0.42695361375808716, | |
| "learning_rate": 4.5887172141209994e-05, | |
| "loss": 0.8642, | |
| "mean_token_accuracy": 0.711685739159584, | |
| "num_tokens": 7045607.0, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.0712966383450313, | |
| "grad_norm": 0.45569705963134766, | |
| "learning_rate": 4.548980431863551e-05, | |
| "loss": 0.8588, | |
| "mean_token_accuracy": 0.7119369316101074, | |
| "num_tokens": 7169759.0, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 1.0897672700406353, | |
| "grad_norm": 0.4741804301738739, | |
| "learning_rate": 4.510258409503273e-05, | |
| "loss": 0.8668, | |
| "mean_token_accuracy": 0.7101493185758591, | |
| "num_tokens": 7294070.0, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 1.1082379017362394, | |
| "grad_norm": 0.46725088357925415, | |
| "learning_rate": 4.472508679587051e-05, | |
| "loss": 0.8689, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.1082379017362394, | |
| "eval_loss": 0.9376741051673889, | |
| "eval_mean_token_accuracy": 0.6943356208801269, | |
| "eval_num_tokens": 7418784.0, | |
| "eval_runtime": 19.087, | |
| "eval_samples_per_second": 52.392, | |
| "eval_steps_per_second": 6.549, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.1267085334318434, | |
| "grad_norm": 0.4105652868747711, | |
| "learning_rate": 4.435691221877225e-05, | |
| "loss": 0.8678, | |
| "mean_token_accuracy": 0.7071201595664024, | |
| "num_tokens": 7542724.0, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 1.1451791651274474, | |
| "grad_norm": 0.4177459180355072, | |
| "learning_rate": 4.399768284971994e-05, | |
| "loss": 0.8614, | |
| "mean_token_accuracy": 0.7109738218784333, | |
| "num_tokens": 7665854.0, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 1.1636497968230513, | |
| "grad_norm": 0.47561731934547424, | |
| "learning_rate": 4.364704223564281e-05, | |
| "loss": 0.8657, | |
| "mean_token_accuracy": 0.7084752279520035, | |
| "num_tokens": 7792262.0, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 1.1821204285186553, | |
| "grad_norm": 0.4582621157169342, | |
| "learning_rate": 4.330465349744206e-05, | |
| "loss": 0.8647, | |
| "mean_token_accuracy": 0.7089388114213944, | |
| "num_tokens": 7915414.0, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 1.2005910602142593, | |
| "grad_norm": 0.4826809763908386, | |
| "learning_rate": 4.2970197969350315e-05, | |
| "loss": 0.8597, | |
| "mean_token_accuracy": 0.7096016359329224, | |
| "num_tokens": 8039995.0, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 1.2190616919098634, | |
| "grad_norm": 0.5360994935035706, | |
| "learning_rate": 4.264337395213374e-05, | |
| "loss": 0.8764, | |
| "mean_token_accuracy": 0.7060012793540955, | |
| "num_tokens": 8163184.0, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 1.2375323236054674, | |
| "grad_norm": 0.4621296525001526, | |
| "learning_rate": 4.232389556904849e-05, | |
| "loss": 0.8707, | |
| "mean_token_accuracy": 0.7078948348760605, | |
| "num_tokens": 8287025.0, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 1.2560029553010712, | |
| "grad_norm": 0.5010101199150085, | |
| "learning_rate": 4.201149171469091e-05, | |
| "loss": 0.8698, | |
| "mean_token_accuracy": 0.7087471377849579, | |
| "num_tokens": 8411205.0, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 1.2744735869966752, | |
| "grad_norm": 0.4199393093585968, | |
| "learning_rate": 4.170590508795705e-05, | |
| "loss": 0.8707, | |
| "mean_token_accuracy": 0.7078831547498703, | |
| "num_tokens": 8535583.0, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 1.2929442186922793, | |
| "grad_norm": 0.4775237739086151, | |
| "learning_rate": 4.1406891301271574e-05, | |
| "loss": 0.8688, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.2929442186922793, | |
| "eval_loss": 0.9409459233283997, | |
| "eval_mean_token_accuracy": 0.6942330374717712, | |
| "eval_num_tokens": 8659234.0, | |
| "eval_runtime": 19.0974, | |
| "eval_samples_per_second": 52.363, | |
| "eval_steps_per_second": 6.545, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.3114148503878833, | |
| "grad_norm": 0.4445763826370239, | |
| "learning_rate": 4.111421805907759e-05, | |
| "loss": 0.8648, | |
| "mean_token_accuracy": 0.7086653360724449, | |
| "num_tokens": 8781788.0, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 1.3298854820834873, | |
| "grad_norm": 0.4124562442302704, | |
| "learning_rate": 4.082766439931165e-05, | |
| "loss": 0.872, | |
| "mean_token_accuracy": 0.7063954091072082, | |
| "num_tokens": 8905822.0, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 1.3483561137790914, | |
| "grad_norm": 0.4572960138320923, | |
| "learning_rate": 4.054701999223518e-05, | |
| "loss": 0.8684, | |
| "mean_token_accuracy": 0.7076165169477463, | |
| "num_tokens": 9030043.0, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 1.3668267454746952, | |
| "grad_norm": 0.4722088873386383, | |
| "learning_rate": 4.0272084491566247e-05, | |
| "loss": 0.8637, | |
| "mean_token_accuracy": 0.7081982636451721, | |
| "num_tokens": 9152573.0, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 1.3852973771702992, | |
| "grad_norm": 0.42168575525283813, | |
| "learning_rate": 4.000266693336297e-05, | |
| "loss": 0.8709, | |
| "mean_token_accuracy": 0.706563394665718, | |
| "num_tokens": 9277125.0, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.4037680088659032, | |
| "grad_norm": 0.4825509488582611, | |
| "learning_rate": 3.973858517856019e-05, | |
| "loss": 0.8759, | |
| "mean_token_accuracy": 0.7081832242012024, | |
| "num_tokens": 9400961.0, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 1.4222386405615073, | |
| "grad_norm": 0.525234043598175, | |
| "learning_rate": 3.947966539546186e-05, | |
| "loss": 0.8712, | |
| "mean_token_accuracy": 0.7071002054214478, | |
| "num_tokens": 9524702.0, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 1.440709272257111, | |
| "grad_norm": 0.4405602514743805, | |
| "learning_rate": 3.922574157884801e-05, | |
| "loss": 0.8605, | |
| "mean_token_accuracy": 0.7087134742736816, | |
| "num_tokens": 9646681.0, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.4591799039527151, | |
| "grad_norm": 0.4512486457824707, | |
| "learning_rate": 3.8976655102673755e-05, | |
| "loss": 0.8775, | |
| "mean_token_accuracy": 0.7052119243144989, | |
| "num_tokens": 9768962.0, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 1.4776505356483192, | |
| "grad_norm": 0.5277122855186462, | |
| "learning_rate": 3.873225430362181e-05, | |
| "loss": 0.8564, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.4776505356483192, | |
| "eval_loss": 0.9360803961753845, | |
| "eval_mean_token_accuracy": 0.6951758770942688, | |
| "eval_num_tokens": 9892735.0, | |
| "eval_runtime": 19.0757, | |
| "eval_samples_per_second": 52.423, | |
| "eval_steps_per_second": 6.553, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.4961211673439232, | |
| "grad_norm": 0.4877372682094574, | |
| "learning_rate": 3.8492394093024636e-05, | |
| "loss": 0.8595, | |
| "mean_token_accuracy": 0.7094873589277267, | |
| "num_tokens": 10015371.0, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.5145917990395272, | |
| "grad_norm": 0.4754094183444977, | |
| "learning_rate": 3.825693559490006e-05, | |
| "loss": 0.8574, | |
| "mean_token_accuracy": 0.710657302737236, | |
| "num_tokens": 10137597.0, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 1.5330624307351313, | |
| "grad_norm": 0.5434764623641968, | |
| "learning_rate": 3.8025745808048846e-05, | |
| "loss": 0.8553, | |
| "mean_token_accuracy": 0.7120581448078156, | |
| "num_tokens": 10260281.0, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 1.551533062430735, | |
| "grad_norm": 0.44720467925071716, | |
| "learning_rate": 3.779869729034645e-05, | |
| "loss": 0.863, | |
| "mean_token_accuracy": 0.7101831078529358, | |
| "num_tokens": 10383827.0, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.5700036941263391, | |
| "grad_norm": 0.4419324994087219, | |
| "learning_rate": 3.7575667863526335e-05, | |
| "loss": 0.8643, | |
| "mean_token_accuracy": 0.7083204621076584, | |
| "num_tokens": 10507419.0, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.5884743258219431, | |
| "grad_norm": 0.46732428669929504, | |
| "learning_rate": 3.735654033690154e-05, | |
| "loss": 0.8626, | |
| "mean_token_accuracy": 0.7100337427854538, | |
| "num_tokens": 10631205.0, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 1.606944957517547, | |
| "grad_norm": 0.5493962168693542, | |
| "learning_rate": 3.7141202248604964e-05, | |
| "loss": 0.878, | |
| "mean_token_accuracy": 0.7033761155605316, | |
| "num_tokens": 10756036.0, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.625415589213151, | |
| "grad_norm": 0.48263365030288696, | |
| "learning_rate": 3.6929545623050815e-05, | |
| "loss": 0.8598, | |
| "mean_token_accuracy": 0.7087458032369613, | |
| "num_tokens": 10880395.0, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 1.643886220908755, | |
| "grad_norm": 0.4675855338573456, | |
| "learning_rate": 3.6721466743428706e-05, | |
| "loss": 0.8629, | |
| "mean_token_accuracy": 0.7102112692594528, | |
| "num_tokens": 11004754.0, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 1.662356852604359, | |
| "grad_norm": 0.4457898736000061, | |
| "learning_rate": 3.6516865938141736e-05, | |
| "loss": 0.8583, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.662356852604359, | |
| "eval_loss": 0.9331590533256531, | |
| "eval_mean_token_accuracy": 0.6950131769180298, | |
| "eval_num_tokens": 11128075.0, | |
| "eval_runtime": 19.0567, | |
| "eval_samples_per_second": 52.475, | |
| "eval_steps_per_second": 6.559, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.680827484299963, | |
| "grad_norm": 0.5188403725624084, | |
| "learning_rate": 3.6315647380189556e-05, | |
| "loss": 0.8593, | |
| "mean_token_accuracy": 0.7113982263207436, | |
| "num_tokens": 11250290.0, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 1.6992981159955671, | |
| "grad_norm": 0.46835747361183167, | |
| "learning_rate": 3.611771889857922e-05, | |
| "loss": 0.8585, | |
| "mean_token_accuracy": 0.7080501514673233, | |
| "num_tokens": 11372150.0, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 1.7177687476911712, | |
| "grad_norm": 0.5107430815696716, | |
| "learning_rate": 3.592299180092082e-05, | |
| "loss": 0.8742, | |
| "mean_token_accuracy": 0.7073276859521865, | |
| "num_tokens": 11495701.0, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 1.736239379386775, | |
| "grad_norm": 0.44301339983940125, | |
| "learning_rate": 3.573138070643225e-05, | |
| "loss": 0.867, | |
| "mean_token_accuracy": 0.7066339915990829, | |
| "num_tokens": 11618337.0, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 1.754710011082379, | |
| "grad_norm": 0.5350750684738159, | |
| "learning_rate": 3.554280338863896e-05, | |
| "loss": 0.8759, | |
| "mean_token_accuracy": 0.7075987154245377, | |
| "num_tokens": 11740955.0, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 1.773180642777983, | |
| "grad_norm": 0.48507505655288696, | |
| "learning_rate": 3.535718062711045e-05, | |
| "loss": 0.8691, | |
| "mean_token_accuracy": 0.708899194598198, | |
| "num_tokens": 11864251.0, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 1.7916512744735869, | |
| "grad_norm": 0.5979758501052856, | |
| "learning_rate": 3.517443606762636e-05, | |
| "loss": 0.8714, | |
| "mean_token_accuracy": 0.7103540396690369, | |
| "num_tokens": 11987337.0, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 1.8101219061691909, | |
| "grad_norm": 0.4601200222969055, | |
| "learning_rate": 3.499449609021135e-05, | |
| "loss": 0.862, | |
| "mean_token_accuracy": 0.7084872847795487, | |
| "num_tokens": 12110351.0, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 1.828592537864795, | |
| "grad_norm": 0.4829343557357788, | |
| "learning_rate": 3.4817289684521056e-05, | |
| "loss": 0.8586, | |
| "mean_token_accuracy": 0.7082627350091935, | |
| "num_tokens": 12233622.0, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 1.847063169560399, | |
| "grad_norm": 0.6211116313934326, | |
| "learning_rate": 3.4642748332099756e-05, | |
| "loss": 0.8562, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.847063169560399, | |
| "eval_loss": 0.9294362664222717, | |
| "eval_mean_token_accuracy": 0.6975332975387574, | |
| "eval_num_tokens": 12355693.0, | |
| "eval_runtime": 19.0345, | |
| "eval_samples_per_second": 52.536, | |
| "eval_steps_per_second": 6.567, | |
| "step": 10000 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 10000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 7.133230870488023e+17, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |