| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9990645463049579, | |
| "eval_steps": 500, | |
| "global_step": 534, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0018709073900841909, | |
| "grad_norm": 0.5139586586877397, | |
| "learning_rate": 1.8518518518518518e-07, | |
| "loss": 1.4813, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.009354536950420954, | |
| "grad_norm": 0.5084917331083768, | |
| "learning_rate": 9.259259259259259e-07, | |
| "loss": 1.4869, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.018709073900841908, | |
| "grad_norm": 0.5086882763520935, | |
| "learning_rate": 1.8518518518518519e-06, | |
| "loss": 1.4912, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02806361085126286, | |
| "grad_norm": 0.49639126734512623, | |
| "learning_rate": 2.7777777777777783e-06, | |
| "loss": 1.4813, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.037418147801683815, | |
| "grad_norm": 0.5064168113795362, | |
| "learning_rate": 3.7037037037037037e-06, | |
| "loss": 1.4849, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04677268475210477, | |
| "grad_norm": 0.474054321279664, | |
| "learning_rate": 4.62962962962963e-06, | |
| "loss": 1.4776, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.05612722170252572, | |
| "grad_norm": 0.40667394555209146, | |
| "learning_rate": 5.555555555555557e-06, | |
| "loss": 1.4756, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06548175865294668, | |
| "grad_norm": 0.2821916256658075, | |
| "learning_rate": 6.481481481481482e-06, | |
| "loss": 1.4585, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.07483629560336763, | |
| "grad_norm": 0.2369804283848295, | |
| "learning_rate": 7.4074074074074075e-06, | |
| "loss": 1.4529, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.08419083255378859, | |
| "grad_norm": 0.23752287185591942, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 1.4356, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.09354536950420954, | |
| "grad_norm": 0.233703618586824, | |
| "learning_rate": 9.25925925925926e-06, | |
| "loss": 1.4195, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.1028999064546305, | |
| "grad_norm": 0.21334448181657395, | |
| "learning_rate": 9.999892908320647e-06, | |
| "loss": 1.4009, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.11225444340505145, | |
| "grad_norm": 0.21146314339445296, | |
| "learning_rate": 9.996145181203616e-06, | |
| "loss": 1.3904, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.1216089803554724, | |
| "grad_norm": 0.20882987047024354, | |
| "learning_rate": 9.98704745668676e-06, | |
| "loss": 1.3718, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.13096351730589337, | |
| "grad_norm": 0.21207151944716085, | |
| "learning_rate": 9.972609476841368e-06, | |
| "loss": 1.3591, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.1403180542563143, | |
| "grad_norm": 0.22054679673319932, | |
| "learning_rate": 9.952846702217886e-06, | |
| "loss": 1.342, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.14967259120673526, | |
| "grad_norm": 0.27057760894999794, | |
| "learning_rate": 9.92778029529039e-06, | |
| "loss": 1.3213, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.15902712815715622, | |
| "grad_norm": 0.29764481463424675, | |
| "learning_rate": 9.897437097795257e-06, | |
| "loss": 1.2966, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.16838166510757718, | |
| "grad_norm": 0.3056628343357791, | |
| "learning_rate": 9.861849601988384e-06, | |
| "loss": 1.2748, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.17773620205799812, | |
| "grad_norm": 0.3099000095239784, | |
| "learning_rate": 9.821055915851647e-06, | |
| "loss": 1.2542, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.18709073900841908, | |
| "grad_norm": 0.2668477842603494, | |
| "learning_rate": 9.775099722285934e-06, | |
| "loss": 1.2217, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.19644527595884004, | |
| "grad_norm": 0.2595557434588695, | |
| "learning_rate": 9.72403023233439e-06, | |
| "loss": 1.2122, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.205799812909261, | |
| "grad_norm": 0.2099058809011832, | |
| "learning_rate": 9.667902132486009e-06, | |
| "loss": 1.1948, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.21515434985968196, | |
| "grad_norm": 0.20126316643616676, | |
| "learning_rate": 9.606775526115963e-06, | |
| "loss": 1.1774, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.2245088868101029, | |
| "grad_norm": 0.17345905255518454, | |
| "learning_rate": 9.540715869125407e-06, | |
| "loss": 1.1681, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.23386342376052385, | |
| "grad_norm": 0.25572107838360714, | |
| "learning_rate": 9.469793899849663e-06, | |
| "loss": 1.1716, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.2432179607109448, | |
| "grad_norm": 0.15018874182675532, | |
| "learning_rate": 9.394085563309827e-06, | |
| "loss": 1.1569, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.25257249766136575, | |
| "grad_norm": 0.15353642479045185, | |
| "learning_rate": 9.31367192988896e-06, | |
| "loss": 1.148, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.26192703461178674, | |
| "grad_norm": 0.16296485857684445, | |
| "learning_rate": 9.228639108519867e-06, | |
| "loss": 1.1504, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.27128157156220767, | |
| "grad_norm": 0.14482935533275312, | |
| "learning_rate": 9.139078154477512e-06, | |
| "loss": 1.1423, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.2806361085126286, | |
| "grad_norm": 0.1566349158870611, | |
| "learning_rate": 9.045084971874738e-06, | |
| "loss": 1.137, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.2899906454630496, | |
| "grad_norm": 0.14554164847096335, | |
| "learning_rate": 8.94676021096575e-06, | |
| "loss": 1.1363, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.2993451824134705, | |
| "grad_norm": 0.1508011037252777, | |
| "learning_rate": 8.844209160367298e-06, | |
| "loss": 1.1304, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3086997193638915, | |
| "grad_norm": 0.14045125665102134, | |
| "learning_rate": 8.737541634312985e-06, | |
| "loss": 1.1332, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.31805425631431244, | |
| "grad_norm": 0.14175186310915225, | |
| "learning_rate": 8.626871855061438e-06, | |
| "loss": 1.1341, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.3274087932647334, | |
| "grad_norm": 0.14507855935792988, | |
| "learning_rate": 8.51231833058426e-06, | |
| "loss": 1.1302, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.33676333021515437, | |
| "grad_norm": 0.1444068343697475, | |
| "learning_rate": 8.39400372766471e-06, | |
| "loss": 1.1333, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.3461178671655753, | |
| "grad_norm": 0.14780446060701052, | |
| "learning_rate": 8.272054740543053e-06, | |
| "loss": 1.1324, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.35547240411599623, | |
| "grad_norm": 0.14569418896265066, | |
| "learning_rate": 8.146601955249187e-06, | |
| "loss": 1.1173, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.3648269410664172, | |
| "grad_norm": 0.14203531449915593, | |
| "learning_rate": 8.017779709767857e-06, | |
| "loss": 1.115, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.37418147801683815, | |
| "grad_norm": 0.13692219577509893, | |
| "learning_rate": 7.88572595018617e-06, | |
| "loss": 1.1135, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.38353601496725914, | |
| "grad_norm": 0.14078263930837534, | |
| "learning_rate": 7.750582082977468e-06, | |
| "loss": 1.1159, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.3928905519176801, | |
| "grad_norm": 0.15520280646196877, | |
| "learning_rate": 7.612492823579744e-06, | |
| "loss": 1.1162, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.402245088868101, | |
| "grad_norm": 0.14752002744879342, | |
| "learning_rate": 7.471606041430724e-06, | |
| "loss": 1.1131, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.411599625818522, | |
| "grad_norm": 0.1509280644564414, | |
| "learning_rate": 7.328072601625558e-06, | |
| "loss": 1.1118, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.42095416276894293, | |
| "grad_norm": 0.1616348908118278, | |
| "learning_rate": 7.18204620336671e-06, | |
| "loss": 1.1044, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.4303086997193639, | |
| "grad_norm": 0.16062290714439983, | |
| "learning_rate": 7.033683215379002e-06, | |
| "loss": 1.1049, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.43966323666978485, | |
| "grad_norm": 0.14900564691058804, | |
| "learning_rate": 6.883142508466054e-06, | |
| "loss": 1.1072, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.4490177736202058, | |
| "grad_norm": 0.14583649841999052, | |
| "learning_rate": 6.730585285387465e-06, | |
| "loss": 1.1062, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.4583723105706268, | |
| "grad_norm": 0.15656333466816236, | |
| "learning_rate": 6.57617490823885e-06, | |
| "loss": 1.1038, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.4677268475210477, | |
| "grad_norm": 0.16642188631242563, | |
| "learning_rate": 6.420076723519615e-06, | |
| "loss": 1.112, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.47708138447146864, | |
| "grad_norm": 0.266837548538648, | |
| "learning_rate": 6.26245788507579e-06, | |
| "loss": 1.1104, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.4864359214218896, | |
| "grad_norm": 0.15891960204556085, | |
| "learning_rate": 6.103487175107508e-06, | |
| "loss": 1.1013, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.49579045837231056, | |
| "grad_norm": 0.15199420768584646, | |
| "learning_rate": 5.943334823432777e-06, | |
| "loss": 1.0965, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.5051449953227315, | |
| "grad_norm": 0.1649301263523591, | |
| "learning_rate": 5.782172325201155e-06, | |
| "loss": 1.1015, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.5144995322731525, | |
| "grad_norm": 0.17068364714449247, | |
| "learning_rate": 5.620172257252427e-06, | |
| "loss": 1.1048, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.5238540692235735, | |
| "grad_norm": 0.1560964872722369, | |
| "learning_rate": 5.457508093317013e-06, | |
| "loss": 1.1048, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.5332086061739943, | |
| "grad_norm": 0.15004408686462048, | |
| "learning_rate": 5.294354018255945e-06, | |
| "loss": 1.1001, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.5425631431244153, | |
| "grad_norm": 0.1563742845945151, | |
| "learning_rate": 5.130884741539367e-06, | |
| "loss": 1.0959, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.5519176800748363, | |
| "grad_norm": 0.15549310467355393, | |
| "learning_rate": 4.967275310163241e-06, | |
| "loss": 1.097, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.5612722170252572, | |
| "grad_norm": 0.17033773254724513, | |
| "learning_rate": 4.803700921204659e-06, | |
| "loss": 1.0991, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.5706267539756782, | |
| "grad_norm": 0.16793233585056505, | |
| "learning_rate": 4.640336734216403e-06, | |
| "loss": 1.1011, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.5799812909260992, | |
| "grad_norm": 0.15858153553414958, | |
| "learning_rate": 4.477357683661734e-06, | |
| "loss": 1.0953, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.5893358278765201, | |
| "grad_norm": 0.18459161802082374, | |
| "learning_rate": 4.314938291590161e-06, | |
| "loss": 1.0905, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.598690364826941, | |
| "grad_norm": 0.16567092468891204, | |
| "learning_rate": 4.1532524807548776e-06, | |
| "loss": 1.0906, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.608044901777362, | |
| "grad_norm": 0.16035972271379115, | |
| "learning_rate": 3.992473388371914e-06, | |
| "loss": 1.0924, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.617399438727783, | |
| "grad_norm": 0.18108290101376773, | |
| "learning_rate": 3.832773180720475e-06, | |
| "loss": 1.092, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.6267539756782039, | |
| "grad_norm": 0.16657617125834437, | |
| "learning_rate": 3.6743228687829596e-06, | |
| "loss": 1.0887, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.6361085126286249, | |
| "grad_norm": 0.1783833220121851, | |
| "learning_rate": 3.517292125122146e-06, | |
| "loss": 1.0938, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.6454630495790459, | |
| "grad_norm": 0.15395137633162537, | |
| "learning_rate": 3.3618491021915334e-06, | |
| "loss": 1.0902, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.6548175865294668, | |
| "grad_norm": 0.16256299187537024, | |
| "learning_rate": 3.2081602522734987e-06, | |
| "loss": 1.0875, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.6641721234798877, | |
| "grad_norm": 0.15781917797031997, | |
| "learning_rate": 3.056390149238022e-06, | |
| "loss": 1.0897, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.6735266604303087, | |
| "grad_norm": 0.15450725652248593, | |
| "learning_rate": 2.906701312312861e-06, | |
| "loss": 1.0883, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.6828811973807296, | |
| "grad_norm": 0.16092272691343973, | |
| "learning_rate": 2.759254032053888e-06, | |
| "loss": 1.0908, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.6922357343311506, | |
| "grad_norm": 0.15881876352311045, | |
| "learning_rate": 2.614206198701958e-06, | |
| "loss": 1.0964, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.7015902712815716, | |
| "grad_norm": 0.16123816798728421, | |
| "learning_rate": 2.471713133110078e-06, | |
| "loss": 1.0826, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.7109448082319925, | |
| "grad_norm": 0.17326937796971717, | |
| "learning_rate": 2.3319274204219427e-06, | |
| "loss": 1.0857, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.7202993451824135, | |
| "grad_norm": 0.15824306754559836, | |
| "learning_rate": 2.1949987466799524e-06, | |
| "loss": 1.0926, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.7296538821328344, | |
| "grad_norm": 0.15676499958123513, | |
| "learning_rate": 2.061073738537635e-06, | |
| "loss": 1.091, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7390084190832554, | |
| "grad_norm": 0.15899394910820866, | |
| "learning_rate": 1.9302958062481673e-06, | |
| "loss": 1.0845, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.7483629560336763, | |
| "grad_norm": 0.16579475502144433, | |
| "learning_rate": 1.8028049900970768e-06, | |
| "loss": 1.0854, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.7577174929840973, | |
| "grad_norm": 0.15812690411171398, | |
| "learning_rate": 1.6787378104435931e-06, | |
| "loss": 1.0867, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.7670720299345183, | |
| "grad_norm": 0.1562749678781465, | |
| "learning_rate": 1.5582271215312294e-06, | |
| "loss": 1.0869, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.7764265668849392, | |
| "grad_norm": 0.1573904723461426, | |
| "learning_rate": 1.4414019692241437e-06, | |
| "loss": 1.0968, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.7857811038353602, | |
| "grad_norm": 0.1663255000536237, | |
| "learning_rate": 1.3283874528215735e-06, | |
| "loss": 1.0896, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.7951356407857811, | |
| "grad_norm": 0.1565901532094871, | |
| "learning_rate": 1.2193045910983864e-06, | |
| "loss": 1.0892, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.804490177736202, | |
| "grad_norm": 0.15886041285660657, | |
| "learning_rate": 1.1142701927151456e-06, | |
| "loss": 1.086, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.813844714686623, | |
| "grad_norm": 0.1653805801993096, | |
| "learning_rate": 1.013396731136465e-06, | |
| "loss": 1.0846, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.823199251637044, | |
| "grad_norm": 0.1659489588893349, | |
| "learning_rate": 9.167922241916055e-07, | |
| "loss": 1.0882, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.8325537885874649, | |
| "grad_norm": 0.14935331942375346, | |
| "learning_rate": 8.245601184062851e-07, | |
| "loss": 1.0919, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.8419083255378859, | |
| "grad_norm": 0.1612238398597313, | |
| "learning_rate": 7.367991782295392e-07, | |
| "loss": 1.0956, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.8512628624883068, | |
| "grad_norm": 0.16096449999893725, | |
| "learning_rate": 6.536033802742814e-07, | |
| "loss": 1.0892, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.8606173994387278, | |
| "grad_norm": 0.18591172254707147, | |
| "learning_rate": 5.750618126847912e-07, | |
| "loss": 1.0814, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.8699719363891487, | |
| "grad_norm": 0.16230740670761631, | |
| "learning_rate": 5.012585797388936e-07, | |
| "loss": 1.0863, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.8793264733395697, | |
| "grad_norm": 0.16399177374939766, | |
| "learning_rate": 4.322727117869951e-07, | |
| "loss": 1.0836, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.8886810102899907, | |
| "grad_norm": 0.19550575898363698, | |
| "learning_rate": 3.6817808062440953e-07, | |
| "loss": 1.091, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.8980355472404116, | |
| "grad_norm": 0.15359012389254986, | |
| "learning_rate": 3.0904332038757977e-07, | |
| "loss": 1.0937, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.9073900841908326, | |
| "grad_norm": 0.1606805442454799, | |
| "learning_rate": 2.5493175405893076e-07, | |
| "loss": 1.0848, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.9167446211412535, | |
| "grad_norm": 0.16036880753390018, | |
| "learning_rate": 2.0590132565903475e-07, | |
| "loss": 1.0904, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.9260991580916744, | |
| "grad_norm": 0.16046720313774515, | |
| "learning_rate": 1.6200453819870122e-07, | |
| "loss": 1.0839, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.9354536950420954, | |
| "grad_norm": 0.16013022562616708, | |
| "learning_rate": 1.232883974574367e-07, | |
| "loss": 1.0876, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.9448082319925164, | |
| "grad_norm": 0.1680369869804664, | |
| "learning_rate": 8.979436164848088e-08, | |
| "loss": 1.0854, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.9541627689429373, | |
| "grad_norm": 0.1538401535295636, | |
| "learning_rate": 6.15582970243117e-08, | |
| "loss": 1.0904, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.9635173058933583, | |
| "grad_norm": 0.1590652953665931, | |
| "learning_rate": 3.861043947016474e-08, | |
| "loss": 1.0783, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.9728718428437793, | |
| "grad_norm": 0.15600780093245614, | |
| "learning_rate": 2.097536212669171e-08, | |
| "loss": 1.0857, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.9822263797942001, | |
| "grad_norm": 0.1590897713521175, | |
| "learning_rate": 8.671949076420883e-09, | |
| "loss": 1.0859, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.9915809167446211, | |
| "grad_norm": 0.15861216238412335, | |
| "learning_rate": 1.7133751222137007e-09, | |
| "loss": 1.0845, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.9990645463049579, | |
| "eval_runtime": 3.2564, | |
| "eval_samples_per_second": 3.071, | |
| "eval_steps_per_second": 0.921, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.9990645463049579, | |
| "step": 534, | |
| "total_flos": 2.2276586722453094e+17, | |
| "train_loss": 1.1599709373734863, | |
| "train_runtime": 17757.3527, | |
| "train_samples_per_second": 1.925, | |
| "train_steps_per_second": 0.03 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 534, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.2276586722453094e+17, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |