| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.21384084894817032, | |
| "eval_steps": 500, | |
| "global_step": 2000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0010692042447408515, | |
| "grad_norm": 0.4522557854652405, | |
| "learning_rate": 0.00018, | |
| "loss": 2.4398, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.002138408489481703, | |
| "grad_norm": 0.5223538875579834, | |
| "learning_rate": 0.0001990954773869347, | |
| "loss": 2.1227, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.003207612734222555, | |
| "grad_norm": 0.47437742352485657, | |
| "learning_rate": 0.00019809045226130653, | |
| "loss": 1.9881, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.004276816978963406, | |
| "grad_norm": 0.3975315988063812, | |
| "learning_rate": 0.0001970854271356784, | |
| "loss": 2.0059, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0053460212237042585, | |
| "grad_norm": 0.29887351393699646, | |
| "learning_rate": 0.00019608040201005026, | |
| "loss": 1.9811, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.00641522546844511, | |
| "grad_norm": 0.3441367745399475, | |
| "learning_rate": 0.00019507537688442212, | |
| "loss": 1.9828, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.007484429713185962, | |
| "grad_norm": 0.2983023226261139, | |
| "learning_rate": 0.00019407035175879398, | |
| "loss": 2.0323, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.008553633957926812, | |
| "grad_norm": 0.30667832493782043, | |
| "learning_rate": 0.00019306532663316582, | |
| "loss": 2.0422, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.009622838202667664, | |
| "grad_norm": 0.3337627947330475, | |
| "learning_rate": 0.0001920603015075377, | |
| "loss": 1.9896, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.010692042447408517, | |
| "grad_norm": 0.3218510150909424, | |
| "learning_rate": 0.00019105527638190955, | |
| "loss": 1.976, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.011761246692149369, | |
| "grad_norm": 0.3417668044567108, | |
| "learning_rate": 0.00019005025125628142, | |
| "loss": 1.9818, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.01283045093689022, | |
| "grad_norm": 0.2790803015232086, | |
| "learning_rate": 0.00018904522613065328, | |
| "loss": 2.0031, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.013899655181631072, | |
| "grad_norm": 0.4081987142562866, | |
| "learning_rate": 0.00018804020100502512, | |
| "loss": 1.9905, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.014968859426371923, | |
| "grad_norm": 0.36898791790008545, | |
| "learning_rate": 0.00018703517587939698, | |
| "loss": 2.0049, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.016038063671112775, | |
| "grad_norm": 0.29053226113319397, | |
| "learning_rate": 0.00018603015075376885, | |
| "loss": 1.9797, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.017107267915853624, | |
| "grad_norm": 0.25266537070274353, | |
| "learning_rate": 0.00018502512562814071, | |
| "loss": 1.9509, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.018176472160594478, | |
| "grad_norm": 0.2833998501300812, | |
| "learning_rate": 0.00018402010050251258, | |
| "loss": 1.9574, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.019245676405335328, | |
| "grad_norm": 0.5657384395599365, | |
| "learning_rate": 0.00018301507537688442, | |
| "loss": 2.0581, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.02031488065007618, | |
| "grad_norm": 0.29206541180610657, | |
| "learning_rate": 0.00018201005025125628, | |
| "loss": 1.8847, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.021384084894817034, | |
| "grad_norm": 0.33557596802711487, | |
| "learning_rate": 0.00018100502512562815, | |
| "loss": 2.0308, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.022453289139557884, | |
| "grad_norm": 0.3242787718772888, | |
| "learning_rate": 0.00018, | |
| "loss": 2.0498, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.023522493384298737, | |
| "grad_norm": 0.37068045139312744, | |
| "learning_rate": 0.00017899497487437188, | |
| "loss": 1.9788, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.024591697629039587, | |
| "grad_norm": 0.3749759793281555, | |
| "learning_rate": 0.00017798994974874371, | |
| "loss": 1.9289, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.02566090187378044, | |
| "grad_norm": 0.3888619840145111, | |
| "learning_rate": 0.00017698492462311558, | |
| "loss": 1.9179, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.02673010611852129, | |
| "grad_norm": 0.377247154712677, | |
| "learning_rate": 0.00017597989949748744, | |
| "loss": 2.0299, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.027799310363262143, | |
| "grad_norm": 0.49286213517189026, | |
| "learning_rate": 0.0001749748743718593, | |
| "loss": 1.9895, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.028868514608002993, | |
| "grad_norm": 0.30507227778434753, | |
| "learning_rate": 0.00017396984924623117, | |
| "loss": 1.9596, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.029937718852743846, | |
| "grad_norm": 0.3705558180809021, | |
| "learning_rate": 0.000172964824120603, | |
| "loss": 2.0357, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.031006923097484696, | |
| "grad_norm": 0.4542909562587738, | |
| "learning_rate": 0.00017195979899497488, | |
| "loss": 1.9323, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.03207612734222555, | |
| "grad_norm": 0.3561468720436096, | |
| "learning_rate": 0.00017095477386934674, | |
| "loss": 1.9695, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.0331453315869664, | |
| "grad_norm": 0.39809998869895935, | |
| "learning_rate": 0.0001699497487437186, | |
| "loss": 2.01, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.03421453583170725, | |
| "grad_norm": 0.279765784740448, | |
| "learning_rate": 0.00016894472361809047, | |
| "loss": 1.9361, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.035283740076448106, | |
| "grad_norm": 0.30650126934051514, | |
| "learning_rate": 0.0001679396984924623, | |
| "loss": 1.9411, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.036352944321188956, | |
| "grad_norm": 0.33605650067329407, | |
| "learning_rate": 0.00016693467336683417, | |
| "loss": 1.9908, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.037422148565929805, | |
| "grad_norm": 0.3010708689689636, | |
| "learning_rate": 0.00016592964824120604, | |
| "loss": 1.9326, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.038491352810670655, | |
| "grad_norm": 0.33158665895462036, | |
| "learning_rate": 0.0001649246231155779, | |
| "loss": 1.9362, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.03956055705541151, | |
| "grad_norm": 0.2630426585674286, | |
| "learning_rate": 0.00016391959798994977, | |
| "loss": 1.9345, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.04062976130015236, | |
| "grad_norm": 0.2744496762752533, | |
| "learning_rate": 0.0001629145728643216, | |
| "loss": 1.916, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.04169896554489321, | |
| "grad_norm": 0.33794867992401123, | |
| "learning_rate": 0.00016190954773869347, | |
| "loss": 1.9082, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.04276816978963407, | |
| "grad_norm": 0.2850714921951294, | |
| "learning_rate": 0.00016090452261306533, | |
| "loss": 1.9752, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.04383737403437492, | |
| "grad_norm": 0.34638258814811707, | |
| "learning_rate": 0.0001598994974874372, | |
| "loss": 1.8934, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.04490657827911577, | |
| "grad_norm": 0.2859291434288025, | |
| "learning_rate": 0.00015889447236180906, | |
| "loss": 1.9893, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.04597578252385662, | |
| "grad_norm": 0.31361544132232666, | |
| "learning_rate": 0.0001578894472361809, | |
| "loss": 1.9233, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.047044986768597474, | |
| "grad_norm": 0.3515094816684723, | |
| "learning_rate": 0.00015688442211055277, | |
| "loss": 1.9668, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.048114191013338324, | |
| "grad_norm": 0.3527173399925232, | |
| "learning_rate": 0.00015587939698492463, | |
| "loss": 1.9771, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.049183395258079174, | |
| "grad_norm": 0.289266973733902, | |
| "learning_rate": 0.0001548743718592965, | |
| "loss": 2.0094, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.050252599502820024, | |
| "grad_norm": 0.30293336510658264, | |
| "learning_rate": 0.00015386934673366836, | |
| "loss": 1.9822, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.05132180374756088, | |
| "grad_norm": 0.31669914722442627, | |
| "learning_rate": 0.0001528643216080402, | |
| "loss": 1.9341, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.05239100799230173, | |
| "grad_norm": 0.3566587567329407, | |
| "learning_rate": 0.00015185929648241206, | |
| "loss": 1.8382, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.05346021223704258, | |
| "grad_norm": 0.2622707188129425, | |
| "learning_rate": 0.00015085427135678393, | |
| "loss": 1.9165, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.05452941648178343, | |
| "grad_norm": 0.41455262899398804, | |
| "learning_rate": 0.0001498492462311558, | |
| "loss": 1.9656, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.05559862072652429, | |
| "grad_norm": 0.37541866302490234, | |
| "learning_rate": 0.00014884422110552766, | |
| "loss": 1.9291, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.056667824971265136, | |
| "grad_norm": 0.36062073707580566, | |
| "learning_rate": 0.0001478391959798995, | |
| "loss": 1.9518, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.057737029216005986, | |
| "grad_norm": 0.30928629636764526, | |
| "learning_rate": 0.00014683417085427136, | |
| "loss": 1.8933, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.058806233460746836, | |
| "grad_norm": 0.29138654470443726, | |
| "learning_rate": 0.00014582914572864323, | |
| "loss": 1.9848, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.05987543770548769, | |
| "grad_norm": 0.259957492351532, | |
| "learning_rate": 0.0001448241206030151, | |
| "loss": 1.8802, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.06094464195022854, | |
| "grad_norm": 0.40919119119644165, | |
| "learning_rate": 0.00014381909547738696, | |
| "loss": 1.9503, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.06201384619496939, | |
| "grad_norm": 0.323599249124527, | |
| "learning_rate": 0.0001428140703517588, | |
| "loss": 1.9125, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.06308305043971024, | |
| "grad_norm": 0.31641989946365356, | |
| "learning_rate": 0.00014180904522613066, | |
| "loss": 1.8805, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.0641522546844511, | |
| "grad_norm": 0.4122176766395569, | |
| "learning_rate": 0.00014080402010050252, | |
| "loss": 1.8804, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.06522145892919196, | |
| "grad_norm": 0.3238036334514618, | |
| "learning_rate": 0.0001397989949748744, | |
| "loss": 1.9645, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.0662906631739328, | |
| "grad_norm": 0.3590157926082611, | |
| "learning_rate": 0.00013879396984924625, | |
| "loss": 1.9316, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.06735986741867366, | |
| "grad_norm": 0.3035522997379303, | |
| "learning_rate": 0.0001377889447236181, | |
| "loss": 1.9626, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.0684290716634145, | |
| "grad_norm": 0.3481757640838623, | |
| "learning_rate": 0.00013678391959798996, | |
| "loss": 1.9469, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.06949827590815535, | |
| "grad_norm": 0.41229742765426636, | |
| "learning_rate": 0.0001357788944723618, | |
| "loss": 2.0198, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.07056748015289621, | |
| "grad_norm": 0.5381227731704712, | |
| "learning_rate": 0.00013477386934673368, | |
| "loss": 1.9417, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.07163668439763705, | |
| "grad_norm": 0.3313769996166229, | |
| "learning_rate": 0.00013376884422110555, | |
| "loss": 1.9537, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.07270588864237791, | |
| "grad_norm": 0.3597472310066223, | |
| "learning_rate": 0.0001327638190954774, | |
| "loss": 1.9603, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.07377509288711877, | |
| "grad_norm": 0.29230740666389465, | |
| "learning_rate": 0.00013175879396984925, | |
| "loss": 1.9189, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.07484429713185961, | |
| "grad_norm": 0.31095099449157715, | |
| "learning_rate": 0.0001307537688442211, | |
| "loss": 1.8667, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.07591350137660047, | |
| "grad_norm": 0.6215882897377014, | |
| "learning_rate": 0.00012974874371859298, | |
| "loss": 1.9415, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.07698270562134131, | |
| "grad_norm": 0.3483082056045532, | |
| "learning_rate": 0.00012874371859296485, | |
| "loss": 1.9062, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.07805190986608217, | |
| "grad_norm": 0.3564540147781372, | |
| "learning_rate": 0.00012773869346733668, | |
| "loss": 1.9511, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.07912111411082302, | |
| "grad_norm": 0.2901301980018616, | |
| "learning_rate": 0.00012673366834170855, | |
| "loss": 1.9828, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.08019031835556387, | |
| "grad_norm": 0.35523131489753723, | |
| "learning_rate": 0.0001257286432160804, | |
| "loss": 1.9864, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.08125952260030472, | |
| "grad_norm": 0.3558770716190338, | |
| "learning_rate": 0.00012472361809045228, | |
| "loss": 1.8961, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.08232872684504558, | |
| "grad_norm": 0.3421788811683655, | |
| "learning_rate": 0.00012371859296482414, | |
| "loss": 1.9287, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.08339793108978642, | |
| "grad_norm": 0.2824752926826477, | |
| "learning_rate": 0.00012271356783919598, | |
| "loss": 1.8427, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.08446713533452728, | |
| "grad_norm": 0.3791171908378601, | |
| "learning_rate": 0.00012170854271356785, | |
| "loss": 1.9233, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.08553633957926814, | |
| "grad_norm": 0.3605143129825592, | |
| "learning_rate": 0.0001207035175879397, | |
| "loss": 1.9573, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.08660554382400898, | |
| "grad_norm": 0.4379942715167999, | |
| "learning_rate": 0.00011969849246231158, | |
| "loss": 1.9844, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.08767474806874984, | |
| "grad_norm": 0.2933856248855591, | |
| "learning_rate": 0.00011869346733668343, | |
| "loss": 1.8231, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.08874395231349068, | |
| "grad_norm": 0.31451016664505005, | |
| "learning_rate": 0.00011768844221105528, | |
| "loss": 1.9715, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.08981315655823154, | |
| "grad_norm": 0.3226332366466522, | |
| "learning_rate": 0.00011668341708542714, | |
| "loss": 1.9178, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.09088236080297239, | |
| "grad_norm": 0.28982308506965637, | |
| "learning_rate": 0.000115678391959799, | |
| "loss": 2.0103, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.09195156504771324, | |
| "grad_norm": 0.30275505781173706, | |
| "learning_rate": 0.00011467336683417087, | |
| "loss": 1.9218, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.09302076929245409, | |
| "grad_norm": 0.3631080687046051, | |
| "learning_rate": 0.00011366834170854272, | |
| "loss": 1.9509, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.09408997353719495, | |
| "grad_norm": 0.3824511468410492, | |
| "learning_rate": 0.00011266331658291458, | |
| "loss": 1.9624, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.09515917778193579, | |
| "grad_norm": 0.24786897003650665, | |
| "learning_rate": 0.00011165829145728644, | |
| "loss": 1.8977, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.09622838202667665, | |
| "grad_norm": 0.39904719591140747, | |
| "learning_rate": 0.00011065326633165829, | |
| "loss": 1.9995, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.09729758627141749, | |
| "grad_norm": 0.3084559738636017, | |
| "learning_rate": 0.00010964824120603017, | |
| "loss": 1.9172, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.09836679051615835, | |
| "grad_norm": 0.35854995250701904, | |
| "learning_rate": 0.00010864321608040202, | |
| "loss": 1.9198, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.0994359947608992, | |
| "grad_norm": 0.29974690079689026, | |
| "learning_rate": 0.00010763819095477387, | |
| "loss": 1.9342, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.10050519900564005, | |
| "grad_norm": 0.3411110043525696, | |
| "learning_rate": 0.00010663316582914574, | |
| "loss": 1.9233, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.1015744032503809, | |
| "grad_norm": 0.3226073980331421, | |
| "learning_rate": 0.00010562814070351759, | |
| "loss": 1.8911, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.10264360749512176, | |
| "grad_norm": 0.35765019059181213, | |
| "learning_rate": 0.00010462311557788944, | |
| "loss": 1.9626, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.1037128117398626, | |
| "grad_norm": 0.3223406672477722, | |
| "learning_rate": 0.00010361809045226132, | |
| "loss": 1.8816, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.10478201598460346, | |
| "grad_norm": 0.3049243986606598, | |
| "learning_rate": 0.00010261306532663317, | |
| "loss": 1.8941, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.10585122022934432, | |
| "grad_norm": 0.2669491469860077, | |
| "learning_rate": 0.00010160804020100503, | |
| "loss": 1.8701, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.10692042447408516, | |
| "grad_norm": 0.3321566581726074, | |
| "learning_rate": 0.00010060301507537689, | |
| "loss": 1.9848, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.10798962871882602, | |
| "grad_norm": 0.3456732928752899, | |
| "learning_rate": 9.959798994974875e-05, | |
| "loss": 1.9494, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.10905883296356686, | |
| "grad_norm": 0.32657039165496826, | |
| "learning_rate": 9.85929648241206e-05, | |
| "loss": 1.9775, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.11012803720830772, | |
| "grad_norm": 0.3481104373931885, | |
| "learning_rate": 9.758793969849247e-05, | |
| "loss": 1.9383, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.11119724145304857, | |
| "grad_norm": 0.3254976272583008, | |
| "learning_rate": 9.658291457286432e-05, | |
| "loss": 1.9553, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.11226644569778942, | |
| "grad_norm": 0.365232914686203, | |
| "learning_rate": 9.55778894472362e-05, | |
| "loss": 1.935, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.11333564994253027, | |
| "grad_norm": 0.3333910405635834, | |
| "learning_rate": 9.457286432160805e-05, | |
| "loss": 1.9184, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.11440485418727113, | |
| "grad_norm": 0.3671784996986389, | |
| "learning_rate": 9.35678391959799e-05, | |
| "loss": 1.9303, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.11547405843201197, | |
| "grad_norm": 0.2980143427848816, | |
| "learning_rate": 9.256281407035176e-05, | |
| "loss": 1.9606, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.11654326267675283, | |
| "grad_norm": 0.29698073863983154, | |
| "learning_rate": 9.155778894472362e-05, | |
| "loss": 1.9324, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.11761246692149367, | |
| "grad_norm": 0.35693129897117615, | |
| "learning_rate": 9.055276381909548e-05, | |
| "loss": 1.9567, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.11868167116623453, | |
| "grad_norm": 0.39428821206092834, | |
| "learning_rate": 8.954773869346734e-05, | |
| "loss": 1.9712, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.11975087541097539, | |
| "grad_norm": 0.4319991171360016, | |
| "learning_rate": 8.85427135678392e-05, | |
| "loss": 1.9229, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.12082007965571623, | |
| "grad_norm": 0.271451860666275, | |
| "learning_rate": 8.753768844221106e-05, | |
| "loss": 1.9374, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.12188928390045708, | |
| "grad_norm": 0.3350280523300171, | |
| "learning_rate": 8.653266331658291e-05, | |
| "loss": 1.97, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.12295848814519794, | |
| "grad_norm": 0.23576690256595612, | |
| "learning_rate": 8.552763819095478e-05, | |
| "loss": 1.9165, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.12402769238993878, | |
| "grad_norm": 0.3180292844772339, | |
| "learning_rate": 8.452261306532664e-05, | |
| "loss": 1.9629, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.12509689663467963, | |
| "grad_norm": 0.35042503476142883, | |
| "learning_rate": 8.351758793969849e-05, | |
| "loss": 1.9137, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.12616610087942048, | |
| "grad_norm": 0.33998608589172363, | |
| "learning_rate": 8.251256281407036e-05, | |
| "loss": 1.8604, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.12723530512416134, | |
| "grad_norm": 0.3942660093307495, | |
| "learning_rate": 8.150753768844221e-05, | |
| "loss": 1.9758, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.1283045093689022, | |
| "grad_norm": 0.34661826491355896, | |
| "learning_rate": 8.050251256281407e-05, | |
| "loss": 1.9638, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.12937371361364305, | |
| "grad_norm": 0.31751224398612976, | |
| "learning_rate": 7.949748743718594e-05, | |
| "loss": 1.8879, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.1304429178583839, | |
| "grad_norm": 0.3244868516921997, | |
| "learning_rate": 7.849246231155779e-05, | |
| "loss": 1.9034, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.13151212210312474, | |
| "grad_norm": 0.28967052698135376, | |
| "learning_rate": 7.748743718592966e-05, | |
| "loss": 1.8447, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.1325813263478656, | |
| "grad_norm": 0.23402956128120422, | |
| "learning_rate": 7.64824120603015e-05, | |
| "loss": 1.8901, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.13365053059260645, | |
| "grad_norm": 0.29249387979507446, | |
| "learning_rate": 7.547738693467337e-05, | |
| "loss": 1.958, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.1347197348373473, | |
| "grad_norm": 0.3257281184196472, | |
| "learning_rate": 7.447236180904524e-05, | |
| "loss": 1.9411, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.13578893908208817, | |
| "grad_norm": 0.2677409052848816, | |
| "learning_rate": 7.346733668341709e-05, | |
| "loss": 1.9618, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.136858143326829, | |
| "grad_norm": 0.5253011584281921, | |
| "learning_rate": 7.246231155778895e-05, | |
| "loss": 1.888, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.13792734757156985, | |
| "grad_norm": 0.3806459307670593, | |
| "learning_rate": 7.14572864321608e-05, | |
| "loss": 1.9266, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.1389965518163107, | |
| "grad_norm": 0.4092906415462494, | |
| "learning_rate": 7.045226130653267e-05, | |
| "loss": 1.9766, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.14006575606105157, | |
| "grad_norm": 0.2884989082813263, | |
| "learning_rate": 6.944723618090453e-05, | |
| "loss": 1.946, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.14113496030579242, | |
| "grad_norm": 0.3107132315635681, | |
| "learning_rate": 6.844221105527638e-05, | |
| "loss": 1.8267, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.14220416455053325, | |
| "grad_norm": 0.29931432008743286, | |
| "learning_rate": 6.743718592964824e-05, | |
| "loss": 1.9226, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.1432733687952741, | |
| "grad_norm": 0.32327061891555786, | |
| "learning_rate": 6.64321608040201e-05, | |
| "loss": 1.9389, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.14434257304001497, | |
| "grad_norm": 0.3812110126018524, | |
| "learning_rate": 6.542713567839197e-05, | |
| "loss": 1.9137, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.14541177728475582, | |
| "grad_norm": 0.3208372890949249, | |
| "learning_rate": 6.442211055276383e-05, | |
| "loss": 1.8635, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.14648098152949668, | |
| "grad_norm": 0.3378104567527771, | |
| "learning_rate": 6.341708542713568e-05, | |
| "loss": 1.9773, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.14755018577423754, | |
| "grad_norm": 0.32731178402900696, | |
| "learning_rate": 6.241206030150753e-05, | |
| "loss": 1.9294, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.14861939001897836, | |
| "grad_norm": 0.30876997113227844, | |
| "learning_rate": 6.14070351758794e-05, | |
| "loss": 1.9457, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.14968859426371922, | |
| "grad_norm": 0.2618562579154968, | |
| "learning_rate": 6.0402010050251256e-05, | |
| "loss": 1.9014, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.15075779850846008, | |
| "grad_norm": 0.3791179656982422, | |
| "learning_rate": 5.939698492462312e-05, | |
| "loss": 1.897, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.15182700275320093, | |
| "grad_norm": 0.31716200709342957, | |
| "learning_rate": 5.839195979899498e-05, | |
| "loss": 1.947, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.1528962069979418, | |
| "grad_norm": 0.2968918979167938, | |
| "learning_rate": 5.738693467336683e-05, | |
| "loss": 1.9895, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.15396541124268262, | |
| "grad_norm": 0.3310339152812958, | |
| "learning_rate": 5.63819095477387e-05, | |
| "loss": 1.9455, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.15503461548742348, | |
| "grad_norm": 0.30880674719810486, | |
| "learning_rate": 5.537688442211055e-05, | |
| "loss": 1.9152, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.15610381973216433, | |
| "grad_norm": 0.28629955649375916, | |
| "learning_rate": 5.437185929648242e-05, | |
| "loss": 1.8665, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.1571730239769052, | |
| "grad_norm": 0.3241690695285797, | |
| "learning_rate": 5.3366834170854276e-05, | |
| "loss": 1.9451, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.15824222822164605, | |
| "grad_norm": 0.26445063948631287, | |
| "learning_rate": 5.236180904522613e-05, | |
| "loss": 1.897, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.1593114324663869, | |
| "grad_norm": 0.3061549961566925, | |
| "learning_rate": 5.135678391959799e-05, | |
| "loss": 1.8208, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.16038063671112773, | |
| "grad_norm": 0.334157794713974, | |
| "learning_rate": 5.035175879396985e-05, | |
| "loss": 1.854, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.1614498409558686, | |
| "grad_norm": 0.2809840142726898, | |
| "learning_rate": 4.934673366834171e-05, | |
| "loss": 1.9399, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.16251904520060945, | |
| "grad_norm": 0.331537663936615, | |
| "learning_rate": 4.834170854271357e-05, | |
| "loss": 1.9757, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.1635882494453503, | |
| "grad_norm": 0.30003657937049866, | |
| "learning_rate": 4.733668341708543e-05, | |
| "loss": 1.9077, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.16465745369009116, | |
| "grad_norm": 0.29070165753364563, | |
| "learning_rate": 4.633165829145729e-05, | |
| "loss": 1.9441, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.165726657934832, | |
| "grad_norm": 0.3907763659954071, | |
| "learning_rate": 4.532663316582915e-05, | |
| "loss": 1.8893, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.16679586217957285, | |
| "grad_norm": 0.3012675940990448, | |
| "learning_rate": 4.4321608040201005e-05, | |
| "loss": 1.9539, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.1678650664243137, | |
| "grad_norm": 0.28272444009780884, | |
| "learning_rate": 4.331658291457287e-05, | |
| "loss": 1.9571, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.16893427066905456, | |
| "grad_norm": 0.3389994502067566, | |
| "learning_rate": 4.231155778894473e-05, | |
| "loss": 1.8685, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.17000347491379542, | |
| "grad_norm": 0.3380429744720459, | |
| "learning_rate": 4.1306532663316586e-05, | |
| "loss": 1.9519, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.17107267915853627, | |
| "grad_norm": 0.3569534122943878, | |
| "learning_rate": 4.0301507537688444e-05, | |
| "loss": 1.8919, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.1721418834032771, | |
| "grad_norm": 0.32496699690818787, | |
| "learning_rate": 3.92964824120603e-05, | |
| "loss": 1.8853, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.17321108764801796, | |
| "grad_norm": 0.3131369650363922, | |
| "learning_rate": 3.829145728643217e-05, | |
| "loss": 1.8952, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.17428029189275882, | |
| "grad_norm": 0.3050549626350403, | |
| "learning_rate": 3.7286432160804025e-05, | |
| "loss": 1.85, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.17534949613749967, | |
| "grad_norm": 0.2770765423774719, | |
| "learning_rate": 3.628140703517588e-05, | |
| "loss": 1.9345, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.17641870038224053, | |
| "grad_norm": 0.40778982639312744, | |
| "learning_rate": 3.527638190954774e-05, | |
| "loss": 1.885, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.17748790462698136, | |
| "grad_norm": 0.305587500333786, | |
| "learning_rate": 3.42713567839196e-05, | |
| "loss": 1.9261, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.17855710887172221, | |
| "grad_norm": 0.2813640534877777, | |
| "learning_rate": 3.3266331658291464e-05, | |
| "loss": 1.981, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.17962631311646307, | |
| "grad_norm": 0.2698950469493866, | |
| "learning_rate": 3.2261306532663315e-05, | |
| "loss": 1.9081, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.18069551736120393, | |
| "grad_norm": 0.3598436713218689, | |
| "learning_rate": 3.125628140703517e-05, | |
| "loss": 1.9129, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.18176472160594478, | |
| "grad_norm": 0.25739961862564087, | |
| "learning_rate": 3.0251256281407038e-05, | |
| "loss": 1.9376, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.1828339258506856, | |
| "grad_norm": 0.30413052439689636, | |
| "learning_rate": 2.9246231155778896e-05, | |
| "loss": 1.9186, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.18390313009542647, | |
| "grad_norm": 0.2563576400279999, | |
| "learning_rate": 2.8241206030150757e-05, | |
| "loss": 1.847, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.18497233434016733, | |
| "grad_norm": 0.3041762113571167, | |
| "learning_rate": 2.7236180904522612e-05, | |
| "loss": 1.9606, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.18604153858490818, | |
| "grad_norm": 0.283974826335907, | |
| "learning_rate": 2.6231155778894474e-05, | |
| "loss": 1.9263, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.18711074282964904, | |
| "grad_norm": 0.3330858051776886, | |
| "learning_rate": 2.522613065326633e-05, | |
| "loss": 1.8791, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.1881799470743899, | |
| "grad_norm": 0.31441086530685425, | |
| "learning_rate": 2.422110552763819e-05, | |
| "loss": 1.9351, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.18924915131913073, | |
| "grad_norm": 0.31107625365257263, | |
| "learning_rate": 2.321608040201005e-05, | |
| "loss": 1.9231, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.19031835556387158, | |
| "grad_norm": 0.34114930033683777, | |
| "learning_rate": 2.2211055276381913e-05, | |
| "loss": 1.9559, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.19138755980861244, | |
| "grad_norm": 0.23545467853546143, | |
| "learning_rate": 2.120603015075377e-05, | |
| "loss": 1.849, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.1924567640533533, | |
| "grad_norm": 0.29872679710388184, | |
| "learning_rate": 2.020100502512563e-05, | |
| "loss": 1.9186, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.19352596829809415, | |
| "grad_norm": 0.34200844168663025, | |
| "learning_rate": 1.9195979899497487e-05, | |
| "loss": 1.9165, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.19459517254283498, | |
| "grad_norm": 0.3605504035949707, | |
| "learning_rate": 1.8190954773869348e-05, | |
| "loss": 1.9204, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.19566437678757584, | |
| "grad_norm": 0.3210300803184509, | |
| "learning_rate": 1.7185929648241206e-05, | |
| "loss": 1.9172, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.1967335810323167, | |
| "grad_norm": 0.31484687328338623, | |
| "learning_rate": 1.6180904522613068e-05, | |
| "loss": 1.9356, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.19780278527705755, | |
| "grad_norm": 0.335658460855484, | |
| "learning_rate": 1.5175879396984927e-05, | |
| "loss": 1.9379, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.1988719895217984, | |
| "grad_norm": 0.36707040667533875, | |
| "learning_rate": 1.4170854271356784e-05, | |
| "loss": 1.9189, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.19994119376653927, | |
| "grad_norm": 0.2470160871744156, | |
| "learning_rate": 1.3165829145728645e-05, | |
| "loss": 1.9095, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.2010103980112801, | |
| "grad_norm": 0.24876108765602112, | |
| "learning_rate": 1.2160804020100503e-05, | |
| "loss": 1.8907, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.20207960225602095, | |
| "grad_norm": 0.31706857681274414, | |
| "learning_rate": 1.1155778894472363e-05, | |
| "loss": 1.9588, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.2031488065007618, | |
| "grad_norm": 0.27313029766082764, | |
| "learning_rate": 1.0150753768844223e-05, | |
| "loss": 1.936, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.20421801074550266, | |
| "grad_norm": 0.2927481532096863, | |
| "learning_rate": 9.14572864321608e-06, | |
| "loss": 1.904, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.20528721499024352, | |
| "grad_norm": 0.261416494846344, | |
| "learning_rate": 8.14070351758794e-06, | |
| "loss": 1.9233, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.20635641923498435, | |
| "grad_norm": 0.2947026193141937, | |
| "learning_rate": 7.1356783919597995e-06, | |
| "loss": 1.8992, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.2074256234797252, | |
| "grad_norm": 0.31104397773742676, | |
| "learning_rate": 6.130653266331659e-06, | |
| "loss": 1.9726, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.20849482772446606, | |
| "grad_norm": 0.2600247263908386, | |
| "learning_rate": 5.125628140703518e-06, | |
| "loss": 2.02, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.20956403196920692, | |
| "grad_norm": 0.30064964294433594, | |
| "learning_rate": 4.120603015075377e-06, | |
| "loss": 1.8599, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.21063323621394778, | |
| "grad_norm": 0.2797416150569916, | |
| "learning_rate": 3.1155778894472364e-06, | |
| "loss": 1.9514, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.21170244045868863, | |
| "grad_norm": 0.34935057163238525, | |
| "learning_rate": 2.1105527638190953e-06, | |
| "loss": 1.8716, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.21277164470342946, | |
| "grad_norm": 0.2744124233722687, | |
| "learning_rate": 1.1055276381909548e-06, | |
| "loss": 1.9384, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.21384084894817032, | |
| "grad_norm": 0.2645925283432007, | |
| "learning_rate": 1.0050251256281409e-07, | |
| "loss": 1.9418, | |
| "step": 2000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 50, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 6.9738373840896e+17, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |