| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.9977761304670127, | |
| "eval_steps": 500, | |
| "global_step": 2022, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.014825796886582653, | |
| "grad_norm": 1.3803142660082024, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6862, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.029651593773165306, | |
| "grad_norm": 1.790122423429332, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6305, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04447739065974796, | |
| "grad_norm": 0.8497603342076308, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5946, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.05930318754633061, | |
| "grad_norm": 0.7732974567839591, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5858, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.07412898443291327, | |
| "grad_norm": 0.8204168744024732, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5662, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.08895478131949593, | |
| "grad_norm": 0.7530005874020861, | |
| "learning_rate": 5e-06, | |
| "loss": 0.564, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.10378057820607858, | |
| "grad_norm": 0.6979206993642474, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5536, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.11860637509266123, | |
| "grad_norm": 0.9504756484437652, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5447, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.1334321719792439, | |
| "grad_norm": 0.8937944128313379, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5466, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.14825796886582654, | |
| "grad_norm": 0.6670560083366487, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5354, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.16308376575240918, | |
| "grad_norm": 0.7364925811772932, | |
| "learning_rate": 5e-06, | |
| "loss": 0.535, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.17790956263899185, | |
| "grad_norm": 0.7183776415658619, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5338, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.1927353595255745, | |
| "grad_norm": 0.5967155900573925, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5301, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.20756115641215717, | |
| "grad_norm": 0.5747353912092298, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5263, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.2223869532987398, | |
| "grad_norm": 0.6159997079939371, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5248, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.23721275018532245, | |
| "grad_norm": 0.5680164570830968, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5279, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.2520385470719051, | |
| "grad_norm": 0.5992456838263306, | |
| "learning_rate": 5e-06, | |
| "loss": 0.526, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.2668643439584878, | |
| "grad_norm": 0.5930479122165856, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5224, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.28169014084507044, | |
| "grad_norm": 0.5912410054870685, | |
| "learning_rate": 5e-06, | |
| "loss": 0.52, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.2965159377316531, | |
| "grad_norm": 0.5894696404162991, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5179, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.3113417346182357, | |
| "grad_norm": 0.5643703980877148, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5182, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.32616753150481836, | |
| "grad_norm": 0.5861129018729444, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5104, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.34099332839140106, | |
| "grad_norm": 0.6236624375674511, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5131, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.3558191252779837, | |
| "grad_norm": 0.5812434696304949, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5191, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.37064492216456635, | |
| "grad_norm": 0.7181156036216183, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5093, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.385470719051149, | |
| "grad_norm": 0.587369525728773, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5161, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.40029651593773163, | |
| "grad_norm": 0.5460966546871288, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5178, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.41512231282431433, | |
| "grad_norm": 0.5876447069011888, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5142, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.429948109710897, | |
| "grad_norm": 0.816294324192534, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5125, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.4447739065974796, | |
| "grad_norm": 0.5492628600148616, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5078, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.45959970348406226, | |
| "grad_norm": 0.579955709352068, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5045, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.4744255003706449, | |
| "grad_norm": 0.6279455964566482, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5047, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.4892512972572276, | |
| "grad_norm": 0.6390417903003635, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5067, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.5040770941438102, | |
| "grad_norm": 0.5816721190754486, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5135, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.5189028910303929, | |
| "grad_norm": 0.6458696214772466, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5034, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.5337286879169756, | |
| "grad_norm": 0.6957712465947605, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5048, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.5485544848035582, | |
| "grad_norm": 0.543098433920077, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4991, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.5633802816901409, | |
| "grad_norm": 0.7663819453316695, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5025, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.5782060785767235, | |
| "grad_norm": 0.5194777261335899, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5072, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.5930318754633062, | |
| "grad_norm": 0.8363946362832739, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4948, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.6078576723498889, | |
| "grad_norm": 0.5430505650241441, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5024, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.6226834692364714, | |
| "grad_norm": 0.5574129318838634, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5034, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.6375092661230541, | |
| "grad_norm": 0.701668057920261, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4912, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.6523350630096367, | |
| "grad_norm": 0.5952225478810055, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4898, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.6671608598962194, | |
| "grad_norm": 0.585331803107272, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5027, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.6819866567828021, | |
| "grad_norm": 0.5906175115867641, | |
| "learning_rate": 5e-06, | |
| "loss": 0.49, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.6968124536693847, | |
| "grad_norm": 0.5782512807193211, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4962, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.7116382505559674, | |
| "grad_norm": 0.528401047427795, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4979, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.72646404744255, | |
| "grad_norm": 0.7128769222059271, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4982, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.7412898443291327, | |
| "grad_norm": 0.5767857437742224, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4891, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.7561156412157154, | |
| "grad_norm": 0.6930280814580314, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5006, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.770941438102298, | |
| "grad_norm": 0.6530314579756452, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4966, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.7857672349888807, | |
| "grad_norm": 0.6639683423874188, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4962, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.8005930318754633, | |
| "grad_norm": 0.5556537170587046, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4947, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.815418828762046, | |
| "grad_norm": 0.5740824829726571, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4919, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.8302446256486287, | |
| "grad_norm": 0.5307405473088794, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4864, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.8450704225352113, | |
| "grad_norm": 0.5249601710096029, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4926, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.859896219421794, | |
| "grad_norm": 0.5263542668413514, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4877, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.8747220163083765, | |
| "grad_norm": 0.5027236096291062, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4903, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.8895478131949592, | |
| "grad_norm": 0.687053549139694, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4945, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.9043736100815419, | |
| "grad_norm": 0.6064169430923966, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4898, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.9191994069681245, | |
| "grad_norm": 0.5127672913798731, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4916, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.9340252038547072, | |
| "grad_norm": 0.5911308230908845, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4876, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.9488510007412898, | |
| "grad_norm": 0.5318485612638623, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4911, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.9636767976278725, | |
| "grad_norm": 0.599135799828323, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4907, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.9785025945144552, | |
| "grad_norm": 0.5017158299240659, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4864, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.9933283914010378, | |
| "grad_norm": 0.638905966589537, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4886, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.9992587101556709, | |
| "eval_loss": 0.489400178194046, | |
| "eval_runtime": 119.2986, | |
| "eval_samples_per_second": 152.332, | |
| "eval_steps_per_second": 0.595, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.0081541882876204, | |
| "grad_norm": 0.8019081623682851, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4775, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.0229799851742032, | |
| "grad_norm": 0.5833948483649493, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4344, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.0378057820607858, | |
| "grad_norm": 0.5717232078953663, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4372, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.0526315789473684, | |
| "grad_norm": 0.5320074952877888, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4376, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.0674573758339512, | |
| "grad_norm": 0.5726271946361784, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4314, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.0822831727205338, | |
| "grad_norm": 0.596509516956136, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4347, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.0971089696071163, | |
| "grad_norm": 0.5390515390075241, | |
| "learning_rate": 5e-06, | |
| "loss": 0.434, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.111934766493699, | |
| "grad_norm": 0.6129914776744078, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4354, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.1267605633802817, | |
| "grad_norm": 0.5592764678336093, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4368, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.1415863602668643, | |
| "grad_norm": 0.7060047642916599, | |
| "learning_rate": 5e-06, | |
| "loss": 0.443, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.156412157153447, | |
| "grad_norm": 0.5940229109864722, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4433, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.1712379540400297, | |
| "grad_norm": 0.5491622519034774, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4304, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.1860637509266123, | |
| "grad_norm": 0.6079761033641661, | |
| "learning_rate": 5e-06, | |
| "loss": 0.438, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.200889547813195, | |
| "grad_norm": 0.5693457065422755, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4403, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.2157153446997775, | |
| "grad_norm": 0.6505910795677263, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4375, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.2305411415863603, | |
| "grad_norm": 0.6268213064250889, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4373, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.2453669384729429, | |
| "grad_norm": 0.5919676508645508, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4306, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.2601927353595257, | |
| "grad_norm": 0.5398299551449609, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4353, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.2750185322461083, | |
| "grad_norm": 0.5424618353666834, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4376, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.2898443291326909, | |
| "grad_norm": 0.5634395154305121, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4361, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.3046701260192735, | |
| "grad_norm": 0.5785455634852342, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4367, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.3194959229058563, | |
| "grad_norm": 0.620601502244543, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4367, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.3343217197924389, | |
| "grad_norm": 0.530395060539937, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4347, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.3491475166790214, | |
| "grad_norm": 0.5770452107287607, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4391, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.3639733135656043, | |
| "grad_norm": 0.5858170710249173, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4343, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.3787991104521868, | |
| "grad_norm": 0.5500756623276715, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4346, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.3936249073387694, | |
| "grad_norm": 0.6255449782390206, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4389, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.408450704225352, | |
| "grad_norm": 0.5460524509865728, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4369, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.4232765011119348, | |
| "grad_norm": 0.5924885663735421, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4338, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.4381022979985174, | |
| "grad_norm": 0.5764645405905721, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4367, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.4529280948851, | |
| "grad_norm": 0.5850117598054916, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4362, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.4677538917716828, | |
| "grad_norm": 0.6081832575580323, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4406, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.4825796886582654, | |
| "grad_norm": 0.543363234220218, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4351, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.497405485544848, | |
| "grad_norm": 0.5448664659996519, | |
| "learning_rate": 5e-06, | |
| "loss": 0.436, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.5122312824314306, | |
| "grad_norm": 0.65441524183388, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4372, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.5270570793180134, | |
| "grad_norm": 0.5280772751509661, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4357, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.541882876204596, | |
| "grad_norm": 0.5723443909631686, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4318, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.5567086730911788, | |
| "grad_norm": 0.5085846011938958, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4323, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.5715344699777614, | |
| "grad_norm": 0.638285906027132, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4415, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.586360266864344, | |
| "grad_norm": 0.571218693857694, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4346, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.6011860637509265, | |
| "grad_norm": 0.517292661892654, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4382, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.6160118606375091, | |
| "grad_norm": 0.5102023179424974, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4323, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.630837657524092, | |
| "grad_norm": 0.5721112226520383, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4378, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.6456634544106745, | |
| "grad_norm": 0.5551329559644929, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4348, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.6604892512972573, | |
| "grad_norm": 0.6119221260699137, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4274, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.67531504818384, | |
| "grad_norm": 0.5333241050316515, | |
| "learning_rate": 5e-06, | |
| "loss": 0.432, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.6901408450704225, | |
| "grad_norm": 0.564328179152952, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4391, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.704966641957005, | |
| "grad_norm": 0.5254085440281446, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4359, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.7197924388435877, | |
| "grad_norm": 0.5996709499991413, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4355, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.7346182357301705, | |
| "grad_norm": 0.5493385200435505, | |
| "learning_rate": 5e-06, | |
| "loss": 0.432, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.7494440326167533, | |
| "grad_norm": 0.5906833738744632, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4367, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.7642698295033359, | |
| "grad_norm": 0.5341688856567031, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4298, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.7790956263899185, | |
| "grad_norm": 0.533463029816806, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4344, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.793921423276501, | |
| "grad_norm": 0.5606747279800302, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4362, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.8087472201630836, | |
| "grad_norm": 0.5183566067624092, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4271, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.8235730170496665, | |
| "grad_norm": 0.5326764319923035, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4306, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.838398813936249, | |
| "grad_norm": 0.5967407273470395, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4344, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.8532246108228319, | |
| "grad_norm": 0.569200724691502, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4353, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.8680504077094144, | |
| "grad_norm": 0.5364248806496852, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4364, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.882876204595997, | |
| "grad_norm": 0.5385402357985632, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4379, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.8977020014825796, | |
| "grad_norm": 0.5223344068630751, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4347, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.9125277983691622, | |
| "grad_norm": 0.5617283138858877, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4319, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.927353595255745, | |
| "grad_norm": 0.5893151487107228, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4313, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.9421793921423276, | |
| "grad_norm": 0.5366729507399915, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4399, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.9570051890289104, | |
| "grad_norm": 0.5521187635061144, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4346, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.971830985915493, | |
| "grad_norm": 0.5736814101430955, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4344, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.9866567828020756, | |
| "grad_norm": 0.6153323220981107, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4288, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.47887247800827026, | |
| "eval_runtime": 117.4023, | |
| "eval_samples_per_second": 154.793, | |
| "eval_steps_per_second": 0.605, | |
| "step": 1349 | |
| }, | |
| { | |
| "epoch": 2.001482579688658, | |
| "grad_norm": 0.8057512126070844, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4465, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.0163083765752408, | |
| "grad_norm": 0.6985686766657238, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3758, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.031134173461824, | |
| "grad_norm": 0.7516592382333935, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3739, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.0459599703484064, | |
| "grad_norm": 0.6560426338450871, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3733, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.060785767234989, | |
| "grad_norm": 0.6554981025712345, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3738, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.0756115641215716, | |
| "grad_norm": 0.5777290379829684, | |
| "learning_rate": 5e-06, | |
| "loss": 0.373, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.090437361008154, | |
| "grad_norm": 0.6039864607065941, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3726, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.1052631578947367, | |
| "grad_norm": 0.5737027873462299, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3745, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.1200889547813193, | |
| "grad_norm": 0.5962503173170325, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3793, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 2.1349147516679023, | |
| "grad_norm": 0.5896368959935617, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3783, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.149740548554485, | |
| "grad_norm": 0.5817443097816994, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3775, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 2.1645663454410675, | |
| "grad_norm": 0.5749673449125087, | |
| "learning_rate": 5e-06, | |
| "loss": 0.378, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.17939214232765, | |
| "grad_norm": 0.581781908254146, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3694, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 2.1942179392142327, | |
| "grad_norm": 0.5991853455621899, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3821, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.2090437361008153, | |
| "grad_norm": 0.5786028266566305, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3761, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 2.223869532987398, | |
| "grad_norm": 0.6524631416486719, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3733, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.238695329873981, | |
| "grad_norm": 0.66274701181455, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3812, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 2.2535211267605635, | |
| "grad_norm": 0.5890654308630106, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3785, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.268346923647146, | |
| "grad_norm": 0.5995580394178166, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3766, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 2.2831727205337287, | |
| "grad_norm": 0.5621523912385117, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3819, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.2979985174203112, | |
| "grad_norm": 0.5631391589331654, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3831, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 2.312824314306894, | |
| "grad_norm": 0.598394877682484, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3776, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.327650111193477, | |
| "grad_norm": 0.5751635376845493, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3743, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 2.3424759080800595, | |
| "grad_norm": 0.7431168749670549, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3806, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.357301704966642, | |
| "grad_norm": 0.6148778429363163, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3792, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 2.3721275018532246, | |
| "grad_norm": 0.5904751763294604, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3779, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.386953298739807, | |
| "grad_norm": 0.6160494878206139, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3795, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 2.40177909562639, | |
| "grad_norm": 0.6601509703827891, | |
| "learning_rate": 5e-06, | |
| "loss": 0.381, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.4166048925129724, | |
| "grad_norm": 0.5557483892389672, | |
| "learning_rate": 5e-06, | |
| "loss": 0.38, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 2.431430689399555, | |
| "grad_norm": 0.5514157650195483, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3785, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.446256486286138, | |
| "grad_norm": 0.5896166088948557, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3785, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 2.4610822831727206, | |
| "grad_norm": 0.6187824694109818, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3866, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.475908080059303, | |
| "grad_norm": 0.6196885650354601, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3773, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 2.4907338769458858, | |
| "grad_norm": 0.6497541102453935, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3792, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.5055596738324684, | |
| "grad_norm": 0.6221830439300937, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3814, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 2.5203854707190514, | |
| "grad_norm": 0.5575748670805571, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3788, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 2.535211267605634, | |
| "grad_norm": 0.5820255267213036, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3797, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 2.5500370644922166, | |
| "grad_norm": 0.5778840920723647, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3836, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 2.564862861378799, | |
| "grad_norm": 0.6369372874220104, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3822, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 2.5796886582653817, | |
| "grad_norm": 0.5836272270823646, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3825, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 2.5945144551519643, | |
| "grad_norm": 0.5893118830611698, | |
| "learning_rate": 5e-06, | |
| "loss": 0.382, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 2.609340252038547, | |
| "grad_norm": 0.691740548026418, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3828, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 2.6241660489251295, | |
| "grad_norm": 0.6016891230542475, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3841, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 2.6389918458117125, | |
| "grad_norm": 0.5709726000324739, | |
| "learning_rate": 5e-06, | |
| "loss": 0.38, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 2.653817642698295, | |
| "grad_norm": 0.5707418370568915, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3819, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 2.6686434395848777, | |
| "grad_norm": 0.6121781418312778, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3775, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.6834692364714603, | |
| "grad_norm": 0.5845205444282813, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3794, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 2.698295033358043, | |
| "grad_norm": 0.6090204118467554, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3749, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 2.713120830244626, | |
| "grad_norm": 0.6311725348745423, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3875, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 2.7279466271312085, | |
| "grad_norm": 0.5855235496792751, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3798, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 2.742772424017791, | |
| "grad_norm": 0.6360580241341309, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3795, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 2.7575982209043737, | |
| "grad_norm": 0.6949332357026632, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3833, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 2.7724240177909563, | |
| "grad_norm": 0.5661995725449918, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3815, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 2.787249814677539, | |
| "grad_norm": 0.6181579391717933, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3814, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 2.8020756115641214, | |
| "grad_norm": 0.621591189180585, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3839, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 2.816901408450704, | |
| "grad_norm": 0.6102808235648141, | |
| "learning_rate": 5e-06, | |
| "loss": 0.386, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 2.8317272053372866, | |
| "grad_norm": 0.5570401466362744, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3829, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 2.8465530022238696, | |
| "grad_norm": 0.605864801280795, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3791, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 2.8613787991104522, | |
| "grad_norm": 0.5737361891170868, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3831, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 2.876204595997035, | |
| "grad_norm": 0.5867020788912005, | |
| "learning_rate": 5e-06, | |
| "loss": 0.38, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 2.8910303928836174, | |
| "grad_norm": 0.557534237372008, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3825, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 2.9058561897702, | |
| "grad_norm": 0.5964701402385094, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3852, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 2.920681986656783, | |
| "grad_norm": 0.600588952238815, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3876, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 2.9355077835433656, | |
| "grad_norm": 0.5765877577022089, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3851, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 2.950333580429948, | |
| "grad_norm": 0.6187803453792526, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3818, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 2.965159377316531, | |
| "grad_norm": 0.5518073320285379, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3834, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 2.9799851742031134, | |
| "grad_norm": 0.5343617840090896, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3772, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 2.994810971089696, | |
| "grad_norm": 0.5629835017462416, | |
| "learning_rate": 5e-06, | |
| "loss": 0.382, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 2.9977761304670127, | |
| "eval_loss": 0.4834764897823334, | |
| "eval_runtime": 117.0978, | |
| "eval_samples_per_second": 155.195, | |
| "eval_steps_per_second": 0.606, | |
| "step": 2022 | |
| }, | |
| { | |
| "epoch": 2.9977761304670127, | |
| "step": 2022, | |
| "total_flos": 3386087791656960.0, | |
| "train_loss": 0.4438282892800698, | |
| "train_runtime": 17391.1323, | |
| "train_samples_per_second": 59.561, | |
| "train_steps_per_second": 0.116 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2022, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3386087791656960.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |