{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.5837623762376238, "eval_steps": 500, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007920792079207921, "grad_norm": 3.6492598056793213, "learning_rate": 0.00019928684627575278, "loss": 3.5835, "step": 10 }, { "epoch": 0.015841584158415842, "grad_norm": 0.15779662132263184, "learning_rate": 0.0001984944532488114, "loss": 0.3867, "step": 20 }, { "epoch": 0.023762376237623763, "grad_norm": 0.0943402647972107, "learning_rate": 0.00019770206022187006, "loss": 0.3435, "step": 30 }, { "epoch": 0.031683168316831684, "grad_norm": 0.07766730338335037, "learning_rate": 0.0001969096671949287, "loss": 0.3724, "step": 40 }, { "epoch": 0.039603960396039604, "grad_norm": 0.044251784682273865, "learning_rate": 0.00019611727416798732, "loss": 0.2735, "step": 50 }, { "epoch": 0.047524752475247525, "grad_norm": 0.04691075533628464, "learning_rate": 0.00019532488114104598, "loss": 0.3015, "step": 60 }, { "epoch": 0.055445544554455446, "grad_norm": 0.04267169535160065, "learning_rate": 0.0001945324881141046, "loss": 0.2531, "step": 70 }, { "epoch": 0.06336633663366337, "grad_norm": 0.03702476993203163, "learning_rate": 0.00019374009508716324, "loss": 0.2919, "step": 80 }, { "epoch": 0.07128712871287128, "grad_norm": 0.04517766833305359, "learning_rate": 0.00019294770206022187, "loss": 0.2666, "step": 90 }, { "epoch": 0.07920792079207921, "grad_norm": 0.03789166733622551, "learning_rate": 0.00019215530903328053, "loss": 0.2768, "step": 100 }, { "epoch": 0.08712871287128712, "grad_norm": 0.0474526472389698, "learning_rate": 0.00019136291600633915, "loss": 0.3106, "step": 110 }, { "epoch": 0.09504950495049505, "grad_norm": 0.040468480437994, "learning_rate": 0.00019057052297939778, "loss": 0.2724, "step": 120 }, { "epoch": 0.10297029702970296, "grad_norm": 0.036530930548906326, "learning_rate": 0.00018977812995245644, "loss": 0.2419, "step": 130 }, { "epoch": 0.11089108910891089, "grad_norm": 0.026719143614172935, "learning_rate": 0.00018898573692551507, "loss": 0.251, "step": 140 }, { "epoch": 0.1188118811881188, "grad_norm": 0.037144389003515244, "learning_rate": 0.0001881933438985737, "loss": 0.2759, "step": 150 }, { "epoch": 0.12673267326732673, "grad_norm": 0.04505979269742966, "learning_rate": 0.00018740095087163233, "loss": 0.2502, "step": 160 }, { "epoch": 0.13465346534653466, "grad_norm": 0.030406810343265533, "learning_rate": 0.00018660855784469099, "loss": 0.2571, "step": 170 }, { "epoch": 0.14257425742574256, "grad_norm": 0.04877385124564171, "learning_rate": 0.00018581616481774962, "loss": 0.2649, "step": 180 }, { "epoch": 0.1504950495049505, "grad_norm": 0.04424108564853668, "learning_rate": 0.00018502377179080824, "loss": 0.2891, "step": 190 }, { "epoch": 0.15841584158415842, "grad_norm": 0.036704324185848236, "learning_rate": 0.0001842313787638669, "loss": 0.3115, "step": 200 }, { "epoch": 0.16633663366336635, "grad_norm": 0.03781713545322418, "learning_rate": 0.00018343898573692553, "loss": 0.2741, "step": 210 }, { "epoch": 0.17425742574257425, "grad_norm": 0.034683957695961, "learning_rate": 0.00018264659270998416, "loss": 0.2595, "step": 220 }, { "epoch": 0.18217821782178217, "grad_norm": 0.03879937529563904, "learning_rate": 0.0001818541996830428, "loss": 0.2565, "step": 230 }, { "epoch": 0.1900990099009901, "grad_norm": 0.030427804216742516, "learning_rate": 0.00018106180665610145, "loss": 0.2716, "step": 240 }, { "epoch": 0.19801980198019803, "grad_norm": 0.05185231566429138, "learning_rate": 0.00018026941362916008, "loss": 0.2864, "step": 250 }, { "epoch": 0.20594059405940593, "grad_norm": 0.04268000274896622, "learning_rate": 0.0001794770206022187, "loss": 0.3219, "step": 260 }, { "epoch": 0.21386138613861386, "grad_norm": 0.033408161252737045, "learning_rate": 0.00017868462757527736, "loss": 0.2582, "step": 270 }, { "epoch": 0.22178217821782178, "grad_norm": 0.03039587289094925, "learning_rate": 0.000177892234548336, "loss": 0.2687, "step": 280 }, { "epoch": 0.2297029702970297, "grad_norm": 0.037947878241539, "learning_rate": 0.00017709984152139462, "loss": 0.2966, "step": 290 }, { "epoch": 0.2376237623762376, "grad_norm": 0.03255503624677658, "learning_rate": 0.00017630744849445328, "loss": 0.2648, "step": 300 }, { "epoch": 0.24554455445544554, "grad_norm": 0.029127761721611023, "learning_rate": 0.0001755150554675119, "loss": 0.2581, "step": 310 }, { "epoch": 0.25346534653465347, "grad_norm": 0.03634035587310791, "learning_rate": 0.00017472266244057054, "loss": 0.2677, "step": 320 }, { "epoch": 0.2613861386138614, "grad_norm": 0.030241692438721657, "learning_rate": 0.00017393026941362917, "loss": 0.2799, "step": 330 }, { "epoch": 0.2693069306930693, "grad_norm": 0.034741729497909546, "learning_rate": 0.00017313787638668782, "loss": 0.2826, "step": 340 }, { "epoch": 0.27722772277227725, "grad_norm": 0.03765638545155525, "learning_rate": 0.00017234548335974645, "loss": 0.2895, "step": 350 }, { "epoch": 0.2851485148514851, "grad_norm": 0.026430374011397362, "learning_rate": 0.00017155309033280508, "loss": 0.2595, "step": 360 }, { "epoch": 0.29306930693069305, "grad_norm": 0.03160027414560318, "learning_rate": 0.00017076069730586374, "loss": 0.271, "step": 370 }, { "epoch": 0.300990099009901, "grad_norm": 0.03543638437986374, "learning_rate": 0.00016996830427892237, "loss": 0.2659, "step": 380 }, { "epoch": 0.3089108910891089, "grad_norm": 0.03766461834311485, "learning_rate": 0.000169175911251981, "loss": 0.2729, "step": 390 }, { "epoch": 0.31683168316831684, "grad_norm": 0.03187921643257141, "learning_rate": 0.00016838351822503963, "loss": 0.2656, "step": 400 }, { "epoch": 0.32475247524752476, "grad_norm": 0.039832450449466705, "learning_rate": 0.00016759112519809828, "loss": 0.2582, "step": 410 }, { "epoch": 0.3326732673267327, "grad_norm": 0.037864796817302704, "learning_rate": 0.0001667987321711569, "loss": 0.2789, "step": 420 }, { "epoch": 0.3405940594059406, "grad_norm": 0.0412084236741066, "learning_rate": 0.00016600633914421551, "loss": 0.2697, "step": 430 }, { "epoch": 0.3485148514851485, "grad_norm": 0.03728514164686203, "learning_rate": 0.00016521394611727417, "loss": 0.2574, "step": 440 }, { "epoch": 0.3564356435643564, "grad_norm": 0.042556311935186386, "learning_rate": 0.0001644215530903328, "loss": 0.3146, "step": 450 }, { "epoch": 0.36435643564356435, "grad_norm": 0.03390824422240257, "learning_rate": 0.00016362916006339143, "loss": 0.267, "step": 460 }, { "epoch": 0.3722772277227723, "grad_norm": 0.04026597738265991, "learning_rate": 0.0001628367670364501, "loss": 0.3195, "step": 470 }, { "epoch": 0.3801980198019802, "grad_norm": 0.04190334305167198, "learning_rate": 0.00016204437400950872, "loss": 0.2559, "step": 480 }, { "epoch": 0.38811881188118813, "grad_norm": 0.029818540439009666, "learning_rate": 0.00016125198098256735, "loss": 0.2916, "step": 490 }, { "epoch": 0.39603960396039606, "grad_norm": 0.02525998093187809, "learning_rate": 0.000160459587955626, "loss": 0.2422, "step": 500 }, { "epoch": 0.403960396039604, "grad_norm": 0.03909703716635704, "learning_rate": 0.00015966719492868463, "loss": 0.2621, "step": 510 }, { "epoch": 0.41188118811881186, "grad_norm": 0.034918032586574554, "learning_rate": 0.00015887480190174326, "loss": 0.2557, "step": 520 }, { "epoch": 0.4198019801980198, "grad_norm": 0.0317477211356163, "learning_rate": 0.0001580824088748019, "loss": 0.2712, "step": 530 }, { "epoch": 0.4277227722772277, "grad_norm": 0.036918554455041885, "learning_rate": 0.00015729001584786055, "loss": 0.2379, "step": 540 }, { "epoch": 0.43564356435643564, "grad_norm": 0.029061611741781235, "learning_rate": 0.00015649762282091918, "loss": 0.2673, "step": 550 }, { "epoch": 0.44356435643564357, "grad_norm": 0.03352155163884163, "learning_rate": 0.0001557052297939778, "loss": 0.329, "step": 560 }, { "epoch": 0.4514851485148515, "grad_norm": 0.04538818076252937, "learning_rate": 0.00015491283676703646, "loss": 0.2628, "step": 570 }, { "epoch": 0.4594059405940594, "grad_norm": 0.03284861892461777, "learning_rate": 0.0001541204437400951, "loss": 0.3, "step": 580 }, { "epoch": 0.46732673267326735, "grad_norm": 0.03831164538860321, "learning_rate": 0.00015332805071315372, "loss": 0.3217, "step": 590 }, { "epoch": 0.4752475247524752, "grad_norm": 0.038299866020679474, "learning_rate": 0.00015253565768621235, "loss": 0.2921, "step": 600 }, { "epoch": 0.48316831683168315, "grad_norm": 0.039746932685375214, "learning_rate": 0.000151743264659271, "loss": 0.2629, "step": 610 }, { "epoch": 0.4910891089108911, "grad_norm": 0.029749492183327675, "learning_rate": 0.00015095087163232964, "loss": 0.2698, "step": 620 }, { "epoch": 0.499009900990099, "grad_norm": 0.032644934952259064, "learning_rate": 0.00015015847860538827, "loss": 0.249, "step": 630 }, { "epoch": 0.5069306930693069, "grad_norm": 0.03892878070473671, "learning_rate": 0.00014936608557844692, "loss": 0.256, "step": 640 }, { "epoch": 0.5148514851485149, "grad_norm": 0.0364990159869194, "learning_rate": 0.00014857369255150555, "loss": 0.257, "step": 650 }, { "epoch": 0.5227722772277228, "grad_norm": 0.03153720125555992, "learning_rate": 0.00014778129952456418, "loss": 0.2549, "step": 660 }, { "epoch": 0.5306930693069307, "grad_norm": 0.03544802591204643, "learning_rate": 0.0001469889064976228, "loss": 0.2918, "step": 670 }, { "epoch": 0.5386138613861386, "grad_norm": 0.03145139664411545, "learning_rate": 0.00014619651347068147, "loss": 0.2597, "step": 680 }, { "epoch": 0.5465346534653466, "grad_norm": 0.03319784253835678, "learning_rate": 0.0001454041204437401, "loss": 0.2877, "step": 690 }, { "epoch": 0.5544554455445545, "grad_norm": 0.026824399828910828, "learning_rate": 0.00014461172741679873, "loss": 0.2516, "step": 700 }, { "epoch": 0.5623762376237624, "grad_norm": 0.02784889005124569, "learning_rate": 0.00014381933438985738, "loss": 0.2495, "step": 710 }, { "epoch": 0.5702970297029702, "grad_norm": 0.035639721900224686, "learning_rate": 0.000143026941362916, "loss": 0.2875, "step": 720 }, { "epoch": 0.5782178217821782, "grad_norm": 0.02496819756925106, "learning_rate": 0.00014223454833597464, "loss": 0.2536, "step": 730 }, { "epoch": 0.5861386138613861, "grad_norm": 0.03329946845769882, "learning_rate": 0.0001414421553090333, "loss": 0.2606, "step": 740 }, { "epoch": 0.594059405940594, "grad_norm": 0.03752067685127258, "learning_rate": 0.00014064976228209193, "loss": 0.2412, "step": 750 }, { "epoch": 0.601980198019802, "grad_norm": 0.029111869633197784, "learning_rate": 0.00013985736925515056, "loss": 0.2624, "step": 760 }, { "epoch": 0.6099009900990099, "grad_norm": 0.04291738569736481, "learning_rate": 0.0001390649762282092, "loss": 0.2546, "step": 770 }, { "epoch": 0.6178217821782178, "grad_norm": 0.028964821249246597, "learning_rate": 0.00013827258320126784, "loss": 0.2662, "step": 780 }, { "epoch": 0.6257425742574257, "grad_norm": 0.03424842283129692, "learning_rate": 0.00013748019017432647, "loss": 0.2434, "step": 790 }, { "epoch": 0.6336633663366337, "grad_norm": 0.02667427435517311, "learning_rate": 0.0001366877971473851, "loss": 0.2329, "step": 800 }, { "epoch": 0.6415841584158416, "grad_norm": 0.02951924316585064, "learning_rate": 0.00013589540412044376, "loss": 0.2643, "step": 810 }, { "epoch": 0.6495049504950495, "grad_norm": 0.023334722965955734, "learning_rate": 0.0001351030110935024, "loss": 0.2453, "step": 820 }, { "epoch": 0.6574257425742575, "grad_norm": 0.03401855751872063, "learning_rate": 0.00013431061806656102, "loss": 0.2707, "step": 830 }, { "epoch": 0.6653465346534654, "grad_norm": 0.033355824649333954, "learning_rate": 0.00013351822503961965, "loss": 0.2874, "step": 840 }, { "epoch": 0.6732673267326733, "grad_norm": 0.027616815641522408, "learning_rate": 0.0001327258320126783, "loss": 0.2527, "step": 850 }, { "epoch": 0.6811881188118812, "grad_norm": 0.03240259364247322, "learning_rate": 0.00013193343898573693, "loss": 0.2907, "step": 860 }, { "epoch": 0.689108910891089, "grad_norm": 0.02760745771229267, "learning_rate": 0.00013114104595879556, "loss": 0.2452, "step": 870 }, { "epoch": 0.697029702970297, "grad_norm": 0.034452375024557114, "learning_rate": 0.00013034865293185422, "loss": 0.2401, "step": 880 }, { "epoch": 0.7049504950495049, "grad_norm": 0.04038127139210701, "learning_rate": 0.00012955625990491285, "loss": 0.2586, "step": 890 }, { "epoch": 0.7128712871287128, "grad_norm": 0.03635507449507713, "learning_rate": 0.00012876386687797148, "loss": 0.2752, "step": 900 }, { "epoch": 0.7207920792079208, "grad_norm": 0.030857287347316742, "learning_rate": 0.0001279714738510301, "loss": 0.2694, "step": 910 }, { "epoch": 0.7287128712871287, "grad_norm": 0.035111233592033386, "learning_rate": 0.00012717908082408876, "loss": 0.2569, "step": 920 }, { "epoch": 0.7366336633663366, "grad_norm": 0.025900471955537796, "learning_rate": 0.0001263866877971474, "loss": 0.2561, "step": 930 }, { "epoch": 0.7445544554455445, "grad_norm": 0.035240888595581055, "learning_rate": 0.00012559429477020602, "loss": 0.2869, "step": 940 }, { "epoch": 0.7524752475247525, "grad_norm": 0.03573041409254074, "learning_rate": 0.00012480190174326468, "loss": 0.2934, "step": 950 }, { "epoch": 0.7603960396039604, "grad_norm": 0.03271041065454483, "learning_rate": 0.0001240095087163233, "loss": 0.24, "step": 960 }, { "epoch": 0.7683168316831683, "grad_norm": 0.03942486643791199, "learning_rate": 0.00012321711568938194, "loss": 0.2798, "step": 970 }, { "epoch": 0.7762376237623763, "grad_norm": 0.029904989525675774, "learning_rate": 0.0001224247226624406, "loss": 0.2456, "step": 980 }, { "epoch": 0.7841584158415842, "grad_norm": 0.036968715488910675, "learning_rate": 0.00012163232963549923, "loss": 0.2683, "step": 990 }, { "epoch": 0.7920792079207921, "grad_norm": 0.02765348181128502, "learning_rate": 0.00012083993660855785, "loss": 0.255, "step": 1000 }, { "epoch": 0.8, "grad_norm": 0.02997688390314579, "learning_rate": 0.0001200475435816165, "loss": 0.2705, "step": 1010 }, { "epoch": 0.807920792079208, "grad_norm": 0.03708113357424736, "learning_rate": 0.00011925515055467513, "loss": 0.251, "step": 1020 }, { "epoch": 0.8158415841584158, "grad_norm": 0.03513709455728531, "learning_rate": 0.00011846275752773377, "loss": 0.2886, "step": 1030 }, { "epoch": 0.8237623762376237, "grad_norm": 0.044748757034540176, "learning_rate": 0.00011767036450079241, "loss": 0.2695, "step": 1040 }, { "epoch": 0.8316831683168316, "grad_norm": 0.03215530887246132, "learning_rate": 0.00011687797147385104, "loss": 0.2674, "step": 1050 }, { "epoch": 0.8396039603960396, "grad_norm": 0.0335354283452034, "learning_rate": 0.00011608557844690969, "loss": 0.261, "step": 1060 }, { "epoch": 0.8475247524752475, "grad_norm": 0.03107100911438465, "learning_rate": 0.00011529318541996832, "loss": 0.2453, "step": 1070 }, { "epoch": 0.8554455445544554, "grad_norm": 0.02802272140979767, "learning_rate": 0.00011450079239302696, "loss": 0.2557, "step": 1080 }, { "epoch": 0.8633663366336634, "grad_norm": 0.03208250179886818, "learning_rate": 0.00011370839936608559, "loss": 0.2871, "step": 1090 }, { "epoch": 0.8712871287128713, "grad_norm": 0.03218165040016174, "learning_rate": 0.00011291600633914423, "loss": 0.2274, "step": 1100 }, { "epoch": 0.8792079207920792, "grad_norm": 0.03343340381979942, "learning_rate": 0.00011212361331220287, "loss": 0.284, "step": 1110 }, { "epoch": 0.8871287128712871, "grad_norm": 0.0337323360145092, "learning_rate": 0.0001113312202852615, "loss": 0.2442, "step": 1120 }, { "epoch": 0.8950495049504951, "grad_norm": 0.03257569298148155, "learning_rate": 0.00011053882725832012, "loss": 0.2639, "step": 1130 }, { "epoch": 0.902970297029703, "grad_norm": 0.035853032022714615, "learning_rate": 0.00010974643423137876, "loss": 0.2767, "step": 1140 }, { "epoch": 0.9108910891089109, "grad_norm": 0.03382259979844093, "learning_rate": 0.00010895404120443739, "loss": 0.2789, "step": 1150 }, { "epoch": 0.9188118811881189, "grad_norm": 0.033024683594703674, "learning_rate": 0.00010816164817749603, "loss": 0.2724, "step": 1160 }, { "epoch": 0.9267326732673268, "grad_norm": 0.029246920719742775, "learning_rate": 0.00010736925515055466, "loss": 0.224, "step": 1170 }, { "epoch": 0.9346534653465347, "grad_norm": 0.03842251002788544, "learning_rate": 0.00010657686212361331, "loss": 0.2733, "step": 1180 }, { "epoch": 0.9425742574257425, "grad_norm": 0.027735207229852676, "learning_rate": 0.00010578446909667195, "loss": 0.2592, "step": 1190 }, { "epoch": 0.9504950495049505, "grad_norm": 0.03715645521879196, "learning_rate": 0.00010499207606973058, "loss": 0.2495, "step": 1200 }, { "epoch": 0.9584158415841584, "grad_norm": 0.04229114204645157, "learning_rate": 0.00010419968304278922, "loss": 0.2926, "step": 1210 }, { "epoch": 0.9663366336633663, "grad_norm": 0.0291130393743515, "learning_rate": 0.00010340729001584785, "loss": 0.2273, "step": 1220 }, { "epoch": 0.9742574257425742, "grad_norm": 0.03738139569759369, "learning_rate": 0.0001026148969889065, "loss": 0.2657, "step": 1230 }, { "epoch": 0.9821782178217822, "grad_norm": 0.02908489853143692, "learning_rate": 0.00010182250396196514, "loss": 0.2354, "step": 1240 }, { "epoch": 0.9900990099009901, "grad_norm": 0.03095426596701145, "learning_rate": 0.00010103011093502377, "loss": 0.2518, "step": 1250 }, { "epoch": 0.998019801980198, "grad_norm": 0.029647527262568474, "learning_rate": 0.00010023771790808241, "loss": 0.2573, "step": 1260 }, { "epoch": 1.0055445544554455, "grad_norm": 0.02596193552017212, "learning_rate": 9.944532488114105e-05, "loss": 0.2485, "step": 1270 }, { "epoch": 1.0134653465346535, "grad_norm": 0.0323275551199913, "learning_rate": 9.86529318541997e-05, "loss": 0.2596, "step": 1280 }, { "epoch": 1.0213861386138614, "grad_norm": 0.03321196511387825, "learning_rate": 9.786053882725833e-05, "loss": 0.2619, "step": 1290 }, { "epoch": 1.0293069306930693, "grad_norm": 0.03337240219116211, "learning_rate": 9.706814580031696e-05, "loss": 0.2638, "step": 1300 }, { "epoch": 1.0372277227722773, "grad_norm": 0.03522242233157158, "learning_rate": 9.62757527733756e-05, "loss": 0.2398, "step": 1310 }, { "epoch": 1.0451485148514852, "grad_norm": 0.03158261254429817, "learning_rate": 9.548335974643423e-05, "loss": 0.2626, "step": 1320 }, { "epoch": 1.0530693069306931, "grad_norm": 0.025289108976721764, "learning_rate": 9.469096671949287e-05, "loss": 0.2706, "step": 1330 }, { "epoch": 1.060990099009901, "grad_norm": 0.030833808705210686, "learning_rate": 9.38985736925515e-05, "loss": 0.2246, "step": 1340 }, { "epoch": 1.068910891089109, "grad_norm": 0.03905413672327995, "learning_rate": 9.310618066561014e-05, "loss": 0.2578, "step": 1350 }, { "epoch": 1.076831683168317, "grad_norm": 0.03793856501579285, "learning_rate": 9.231378763866879e-05, "loss": 0.266, "step": 1360 }, { "epoch": 1.0847524752475248, "grad_norm": 0.039849117398262024, "learning_rate": 9.152139461172742e-05, "loss": 0.2591, "step": 1370 }, { "epoch": 1.0926732673267328, "grad_norm": 0.03838738054037094, "learning_rate": 9.072900158478606e-05, "loss": 0.2553, "step": 1380 }, { "epoch": 1.1005940594059407, "grad_norm": 0.032800108194351196, "learning_rate": 8.993660855784469e-05, "loss": 0.2328, "step": 1390 }, { "epoch": 1.1085148514851486, "grad_norm": 0.03500501811504364, "learning_rate": 8.914421553090333e-05, "loss": 0.251, "step": 1400 }, { "epoch": 1.1164356435643565, "grad_norm": 0.03320365026593208, "learning_rate": 8.835182250396196e-05, "loss": 0.274, "step": 1410 }, { "epoch": 1.1243564356435645, "grad_norm": 0.033811185508966446, "learning_rate": 8.75594294770206e-05, "loss": 0.2525, "step": 1420 }, { "epoch": 1.1322772277227724, "grad_norm": 0.03873720020055771, "learning_rate": 8.676703645007925e-05, "loss": 0.2767, "step": 1430 }, { "epoch": 1.1401980198019803, "grad_norm": 0.030054964125156403, "learning_rate": 8.597464342313788e-05, "loss": 0.257, "step": 1440 }, { "epoch": 1.148118811881188, "grad_norm": 0.03347312659025192, "learning_rate": 8.518225039619652e-05, "loss": 0.2583, "step": 1450 }, { "epoch": 1.156039603960396, "grad_norm": 0.023901082575321198, "learning_rate": 8.438985736925515e-05, "loss": 0.2286, "step": 1460 }, { "epoch": 1.1639603960396039, "grad_norm": 0.030662311241030693, "learning_rate": 8.359746434231379e-05, "loss": 0.272, "step": 1470 }, { "epoch": 1.1718811881188118, "grad_norm": 0.035689935088157654, "learning_rate": 8.280507131537243e-05, "loss": 0.2593, "step": 1480 }, { "epoch": 1.1798019801980197, "grad_norm": 0.04119367152452469, "learning_rate": 8.201267828843106e-05, "loss": 0.2778, "step": 1490 }, { "epoch": 1.1877227722772277, "grad_norm": 0.033894915133714676, "learning_rate": 8.122028526148971e-05, "loss": 0.2727, "step": 1500 }, { "epoch": 1.1956435643564356, "grad_norm": 0.031137267127633095, "learning_rate": 8.042789223454834e-05, "loss": 0.2734, "step": 1510 }, { "epoch": 1.2035643564356435, "grad_norm": 0.038331665098667145, "learning_rate": 7.963549920760698e-05, "loss": 0.2925, "step": 1520 }, { "epoch": 1.2114851485148515, "grad_norm": 0.0408506877720356, "learning_rate": 7.884310618066561e-05, "loss": 0.2921, "step": 1530 }, { "epoch": 1.2194059405940594, "grad_norm": 0.04389080032706261, "learning_rate": 7.805071315372425e-05, "loss": 0.2538, "step": 1540 }, { "epoch": 1.2273267326732673, "grad_norm": 0.03198599815368652, "learning_rate": 7.72583201267829e-05, "loss": 0.2759, "step": 1550 }, { "epoch": 1.2352475247524752, "grad_norm": 0.03177543357014656, "learning_rate": 7.646592709984152e-05, "loss": 0.2514, "step": 1560 }, { "epoch": 1.2431683168316832, "grad_norm": 0.02572786435484886, "learning_rate": 7.567353407290017e-05, "loss": 0.2293, "step": 1570 }, { "epoch": 1.251089108910891, "grad_norm": 0.03768354281783104, "learning_rate": 7.48811410459588e-05, "loss": 0.2521, "step": 1580 }, { "epoch": 1.259009900990099, "grad_norm": 0.03194349259138107, "learning_rate": 7.408874801901744e-05, "loss": 0.2912, "step": 1590 }, { "epoch": 1.266930693069307, "grad_norm": 0.02686484344303608, "learning_rate": 7.329635499207608e-05, "loss": 0.261, "step": 1600 }, { "epoch": 1.2748514851485149, "grad_norm": 0.03775911033153534, "learning_rate": 7.250396196513471e-05, "loss": 0.2744, "step": 1610 }, { "epoch": 1.2827722772277228, "grad_norm": 0.03530183061957359, "learning_rate": 7.171156893819336e-05, "loss": 0.2622, "step": 1620 }, { "epoch": 1.2906930693069307, "grad_norm": 0.03804347291588783, "learning_rate": 7.091917591125199e-05, "loss": 0.2877, "step": 1630 }, { "epoch": 1.2986138613861387, "grad_norm": 0.03272108733654022, "learning_rate": 7.012678288431063e-05, "loss": 0.2576, "step": 1640 }, { "epoch": 1.3065346534653466, "grad_norm": 0.03325280174612999, "learning_rate": 6.933438985736926e-05, "loss": 0.2714, "step": 1650 }, { "epoch": 1.3144554455445545, "grad_norm": 0.026619790121912956, "learning_rate": 6.854199683042789e-05, "loss": 0.2279, "step": 1660 }, { "epoch": 1.3223762376237624, "grad_norm": 0.039562638849020004, "learning_rate": 6.774960380348653e-05, "loss": 0.2867, "step": 1670 }, { "epoch": 1.3302970297029704, "grad_norm": 0.03468257933855057, "learning_rate": 6.695721077654516e-05, "loss": 0.2634, "step": 1680 }, { "epoch": 1.3382178217821783, "grad_norm": 0.025986766442656517, "learning_rate": 6.61648177496038e-05, "loss": 0.224, "step": 1690 }, { "epoch": 1.346138613861386, "grad_norm": 0.03160301595926285, "learning_rate": 6.537242472266245e-05, "loss": 0.2952, "step": 1700 }, { "epoch": 1.354059405940594, "grad_norm": 0.03343851491808891, "learning_rate": 6.458003169572108e-05, "loss": 0.2447, "step": 1710 }, { "epoch": 1.3619801980198019, "grad_norm": 0.04069166257977486, "learning_rate": 6.378763866877972e-05, "loss": 0.2356, "step": 1720 }, { "epoch": 1.3699009900990098, "grad_norm": 0.031248316168785095, "learning_rate": 6.299524564183835e-05, "loss": 0.2275, "step": 1730 }, { "epoch": 1.3778217821782177, "grad_norm": 0.04714437574148178, "learning_rate": 6.220285261489699e-05, "loss": 0.2949, "step": 1740 }, { "epoch": 1.3857425742574256, "grad_norm": 0.03378450125455856, "learning_rate": 6.141045958795562e-05, "loss": 0.2812, "step": 1750 }, { "epoch": 1.3936633663366336, "grad_norm": 0.03825616091489792, "learning_rate": 6.061806656101426e-05, "loss": 0.2662, "step": 1760 }, { "epoch": 1.4015841584158415, "grad_norm": 0.03776633366942406, "learning_rate": 5.98256735340729e-05, "loss": 0.2986, "step": 1770 }, { "epoch": 1.4095049504950494, "grad_norm": 0.030764272436499596, "learning_rate": 5.9033280507131536e-05, "loss": 0.2462, "step": 1780 }, { "epoch": 1.4174257425742574, "grad_norm": 0.03300099819898605, "learning_rate": 5.824088748019018e-05, "loss": 0.2423, "step": 1790 }, { "epoch": 1.4253465346534653, "grad_norm": 0.03446972742676735, "learning_rate": 5.7448494453248815e-05, "loss": 0.2428, "step": 1800 }, { "epoch": 1.4332673267326732, "grad_norm": 0.032901886850595474, "learning_rate": 5.665610142630745e-05, "loss": 0.2436, "step": 1810 }, { "epoch": 1.4411881188118811, "grad_norm": 0.04104743152856827, "learning_rate": 5.586370839936609e-05, "loss": 0.2625, "step": 1820 }, { "epoch": 1.449108910891089, "grad_norm": 0.033070165663957596, "learning_rate": 5.5071315372424724e-05, "loss": 0.2588, "step": 1830 }, { "epoch": 1.457029702970297, "grad_norm": 0.03634243831038475, "learning_rate": 5.427892234548336e-05, "loss": 0.26, "step": 1840 }, { "epoch": 1.464950495049505, "grad_norm": 0.033768270164728165, "learning_rate": 5.3486529318542e-05, "loss": 0.2467, "step": 1850 }, { "epoch": 1.4728712871287128, "grad_norm": 0.036840010434389114, "learning_rate": 5.269413629160064e-05, "loss": 0.2563, "step": 1860 }, { "epoch": 1.4807920792079208, "grad_norm": 0.03726380318403244, "learning_rate": 5.1901743264659275e-05, "loss": 0.298, "step": 1870 }, { "epoch": 1.4887128712871287, "grad_norm": 0.032144658267498016, "learning_rate": 5.110935023771791e-05, "loss": 0.275, "step": 1880 }, { "epoch": 1.4966336633663366, "grad_norm": 0.027517810463905334, "learning_rate": 5.031695721077655e-05, "loss": 0.2675, "step": 1890 }, { "epoch": 1.5045544554455446, "grad_norm": 0.0347113199532032, "learning_rate": 4.9524564183835184e-05, "loss": 0.2453, "step": 1900 }, { "epoch": 1.5124752475247525, "grad_norm": 0.03786664828658104, "learning_rate": 4.873217115689383e-05, "loss": 0.2783, "step": 1910 }, { "epoch": 1.5203960396039604, "grad_norm": 0.03823668137192726, "learning_rate": 4.793977812995246e-05, "loss": 0.2248, "step": 1920 }, { "epoch": 1.5283168316831683, "grad_norm": 0.04581904783844948, "learning_rate": 4.714738510301109e-05, "loss": 0.2662, "step": 1930 }, { "epoch": 1.5362376237623763, "grad_norm": 0.03468246012926102, "learning_rate": 4.635499207606973e-05, "loss": 0.244, "step": 1940 }, { "epoch": 1.5441584158415842, "grad_norm": 0.033897291868925095, "learning_rate": 4.5562599049128365e-05, "loss": 0.2751, "step": 1950 }, { "epoch": 1.5520792079207921, "grad_norm": 0.03185650333762169, "learning_rate": 4.477020602218701e-05, "loss": 0.296, "step": 1960 }, { "epoch": 1.56, "grad_norm": 0.032408442348241806, "learning_rate": 4.3977812995245645e-05, "loss": 0.2943, "step": 1970 }, { "epoch": 1.567920792079208, "grad_norm": 0.02897750772535801, "learning_rate": 4.318541996830428e-05, "loss": 0.2473, "step": 1980 }, { "epoch": 1.575841584158416, "grad_norm": 0.03331100195646286, "learning_rate": 4.239302694136292e-05, "loss": 0.2875, "step": 1990 }, { "epoch": 1.5837623762376238, "grad_norm": 0.02858470194041729, "learning_rate": 4.160063391442155e-05, "loss": 0.247, "step": 2000 } ], "logging_steps": 10, "max_steps": 2524, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.0863914532562534e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }