{ "best_metric": 1.1406679153442383, "best_model_checkpoint": "/Users/bbunzeck/Documents/lexdec-llamas/large-char/checkpoint-9063", "epoch": 0.9999130726340234, "eval_steps": 1007, "global_step": 10065, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009934556111614737, "grad_norm": 5.29207706451416, "learning_rate": 1.4999999999999999e-05, "loss": 4.4002, "step": 10 }, { "epoch": 0.0019869112223229474, "grad_norm": 2.1869823932647705, "learning_rate": 2.9999999999999997e-05, "loss": 3.6401, "step": 20 }, { "epoch": 0.0029803668334844216, "grad_norm": 0.9984068870544434, "learning_rate": 4.4999999999999996e-05, "loss": 3.2167, "step": 30 }, { "epoch": 0.003973822444645895, "grad_norm": 0.700920820236206, "learning_rate": 5.9999999999999995e-05, "loss": 3.1034, "step": 40 }, { "epoch": 0.004967278055807369, "grad_norm": 2.1026079654693604, "learning_rate": 7.5e-05, "loss": 3.0336, "step": 50 }, { "epoch": 0.005960733666968843, "grad_norm": 1.2076551914215088, "learning_rate": 8.999999999999999e-05, "loss": 2.8545, "step": 60 }, { "epoch": 0.006954189278130317, "grad_norm": 1.447243332862854, "learning_rate": 0.00010499999999999999, "loss": 2.6664, "step": 70 }, { "epoch": 0.00794764488929179, "grad_norm": 1.2562823295593262, "learning_rate": 0.00011999999999999999, "loss": 2.5406, "step": 80 }, { "epoch": 0.008941100500453264, "grad_norm": 1.3105968236923218, "learning_rate": 0.000135, "loss": 2.4273, "step": 90 }, { "epoch": 0.009934556111614737, "grad_norm": 1.9283251762390137, "learning_rate": 0.00015, "loss": 2.3703, "step": 100 }, { "epoch": 0.010928011722776212, "grad_norm": 1.5582048892974854, "learning_rate": 0.000165, "loss": 2.3019, "step": 110 }, { "epoch": 0.011921467333937686, "grad_norm": 1.2038098573684692, "learning_rate": 0.00017999999999999998, "loss": 2.2016, "step": 120 }, { "epoch": 0.01291492294509916, "grad_norm": 1.2954462766647339, "learning_rate": 0.000195, "loss": 2.1344, "step": 130 }, { "epoch": 0.013908378556260634, "grad_norm": 1.1450471878051758, "learning_rate": 0.00020999999999999998, "loss": 2.0852, "step": 140 }, { "epoch": 0.014901834167422107, "grad_norm": 1.3502197265625, "learning_rate": 0.000225, "loss": 2.03, "step": 150 }, { "epoch": 0.01589528977858358, "grad_norm": 1.293748140335083, "learning_rate": 0.00023999999999999998, "loss": 1.9633, "step": 160 }, { "epoch": 0.016888745389745056, "grad_norm": 1.2430918216705322, "learning_rate": 0.00025499999999999996, "loss": 1.9261, "step": 170 }, { "epoch": 0.01788220100090653, "grad_norm": 1.6550723314285278, "learning_rate": 0.00027, "loss": 1.8994, "step": 180 }, { "epoch": 0.018875656612068, "grad_norm": 1.2588621377944946, "learning_rate": 0.000285, "loss": 1.8689, "step": 190 }, { "epoch": 0.019869112223229474, "grad_norm": 1.072903037071228, "learning_rate": 0.0003, "loss": 1.8291, "step": 200 }, { "epoch": 0.02086256783439095, "grad_norm": 1.5793360471725464, "learning_rate": 0.0002999992393822385, "loss": 1.8128, "step": 210 }, { "epoch": 0.021856023445552424, "grad_norm": 0.9704577922821045, "learning_rate": 0.000299996957536668, "loss": 1.7878, "step": 220 }, { "epoch": 0.022849479056713896, "grad_norm": 0.8479828238487244, "learning_rate": 0.0002999931544864298, "loss": 1.763, "step": 230 }, { "epoch": 0.023842934667875373, "grad_norm": 0.8488192558288574, "learning_rate": 0.00029998783027009295, "loss": 1.7213, "step": 240 }, { "epoch": 0.024836390279036846, "grad_norm": 0.7466472387313843, "learning_rate": 0.0002999809849416534, "loss": 1.7309, "step": 250 }, { "epoch": 0.02582984589019832, "grad_norm": 0.9153957366943359, "learning_rate": 0.0002999726185705334, "loss": 1.6989, "step": 260 }, { "epoch": 0.02682330150135979, "grad_norm": 0.9594451189041138, "learning_rate": 0.00029996273124158115, "loss": 1.7, "step": 270 }, { "epoch": 0.027816757112521268, "grad_norm": 0.8583829998970032, "learning_rate": 0.0002999513230550697, "loss": 1.7048, "step": 280 }, { "epoch": 0.02881021272368274, "grad_norm": 0.8427692651748657, "learning_rate": 0.00029993839412669593, "loss": 1.6674, "step": 290 }, { "epoch": 0.029803668334844213, "grad_norm": 0.8587358593940735, "learning_rate": 0.0002999239445875795, "loss": 1.6609, "step": 300 }, { "epoch": 0.030797123946005686, "grad_norm": 0.7885432839393616, "learning_rate": 0.00029990797458426146, "loss": 1.6317, "step": 310 }, { "epoch": 0.03179057955716716, "grad_norm": 0.8649609684944153, "learning_rate": 0.0002998904842787027, "loss": 1.6259, "step": 320 }, { "epoch": 0.03278403516832863, "grad_norm": 0.757290244102478, "learning_rate": 0.00029987147384828234, "loss": 1.6038, "step": 330 }, { "epoch": 0.03377749077949011, "grad_norm": 0.8383753299713135, "learning_rate": 0.000299850943485796, "loss": 1.5934, "step": 340 }, { "epoch": 0.034770946390651584, "grad_norm": 0.723807692527771, "learning_rate": 0.00029982889339945384, "loss": 1.5714, "step": 350 }, { "epoch": 0.03576440200181306, "grad_norm": 0.7404069304466248, "learning_rate": 0.00029980532381287836, "loss": 1.576, "step": 360 }, { "epoch": 0.03675785761297453, "grad_norm": 0.7834609150886536, "learning_rate": 0.0002997802349651021, "loss": 1.5961, "step": 370 }, { "epoch": 0.037751313224136, "grad_norm": 0.681057870388031, "learning_rate": 0.00029975362711056545, "loss": 1.5537, "step": 380 }, { "epoch": 0.038744768835297476, "grad_norm": 0.6808239221572876, "learning_rate": 0.00029972550051911377, "loss": 1.565, "step": 390 }, { "epoch": 0.03973822444645895, "grad_norm": 0.7648847699165344, "learning_rate": 0.00029969585547599494, "loss": 1.5384, "step": 400 }, { "epoch": 0.04073168005762043, "grad_norm": 0.6745840311050415, "learning_rate": 0.00029966469228185613, "loss": 1.536, "step": 410 }, { "epoch": 0.0417251356687819, "grad_norm": 0.7665774822235107, "learning_rate": 0.0002996320112527412, "loss": 1.5349, "step": 420 }, { "epoch": 0.042718591279943374, "grad_norm": 0.6684510111808777, "learning_rate": 0.00029959781272008695, "loss": 1.5285, "step": 430 }, { "epoch": 0.04371204689110485, "grad_norm": 0.6902784705162048, "learning_rate": 0.0002995620970307203, "loss": 1.526, "step": 440 }, { "epoch": 0.04470550250226632, "grad_norm": 0.7477157115936279, "learning_rate": 0.00029952486454685437, "loss": 1.5277, "step": 450 }, { "epoch": 0.04569895811342779, "grad_norm": 0.8565313220024109, "learning_rate": 0.00029948611564608506, "loss": 1.486, "step": 460 }, { "epoch": 0.046692413724589266, "grad_norm": 0.6772172451019287, "learning_rate": 0.000299445850721387, "loss": 1.5142, "step": 470 }, { "epoch": 0.047685869335750745, "grad_norm": 0.6629778146743774, "learning_rate": 0.0002994040701811098, "loss": 1.4998, "step": 480 }, { "epoch": 0.04867932494691222, "grad_norm": 0.6849949955940247, "learning_rate": 0.00029936077444897373, "loss": 1.5, "step": 490 }, { "epoch": 0.04967278055807369, "grad_norm": 0.6776939630508423, "learning_rate": 0.0002993159639640655, "loss": 1.5044, "step": 500 }, { "epoch": 0.050666236169235164, "grad_norm": 0.7044647336006165, "learning_rate": 0.00029926963918083376, "loss": 1.4885, "step": 510 }, { "epoch": 0.05165969178039664, "grad_norm": 0.8003182411193848, "learning_rate": 0.00029922180056908456, "loss": 1.4668, "step": 520 }, { "epoch": 0.05265314739155811, "grad_norm": 0.6357293725013733, "learning_rate": 0.00029917244861397653, "loss": 1.4843, "step": 530 }, { "epoch": 0.05364660300271958, "grad_norm": 0.6657519936561584, "learning_rate": 0.000299121583816016, "loss": 1.4871, "step": 540 }, { "epoch": 0.054640058613881055, "grad_norm": 0.6451624035835266, "learning_rate": 0.0002990692066910519, "loss": 1.4561, "step": 550 }, { "epoch": 0.055633514225042535, "grad_norm": 0.669553279876709, "learning_rate": 0.00029901531777027046, "loss": 1.4723, "step": 560 }, { "epoch": 0.05662696983620401, "grad_norm": 0.6627147197723389, "learning_rate": 0.00029895991760019007, "loss": 1.4591, "step": 570 }, { "epoch": 0.05762042544736548, "grad_norm": 0.7017959952354431, "learning_rate": 0.00029890300674265526, "loss": 1.4747, "step": 580 }, { "epoch": 0.058613881058526954, "grad_norm": 0.6519443392753601, "learning_rate": 0.0002988445857748316, "loss": 1.4642, "step": 590 }, { "epoch": 0.05960733666968843, "grad_norm": 0.6378771066665649, "learning_rate": 0.0002987846552891995, "loss": 1.4427, "step": 600 }, { "epoch": 0.0606007922808499, "grad_norm": 0.6792002320289612, "learning_rate": 0.000298723215893548, "loss": 1.4413, "step": 610 }, { "epoch": 0.06159424789201137, "grad_norm": 0.6617796421051025, "learning_rate": 0.00029866026821096924, "loss": 1.4423, "step": 620 }, { "epoch": 0.06258770350317285, "grad_norm": 0.6430485844612122, "learning_rate": 0.0002985958128798514, "loss": 1.426, "step": 630 }, { "epoch": 0.06358115911433432, "grad_norm": 0.6020046472549438, "learning_rate": 0.00029852985055387284, "loss": 1.4353, "step": 640 }, { "epoch": 0.06457461472549579, "grad_norm": 0.6419572234153748, "learning_rate": 0.00029846238190199504, "loss": 1.4392, "step": 650 }, { "epoch": 0.06556807033665726, "grad_norm": 0.6080176830291748, "learning_rate": 0.00029839340760845614, "loss": 1.3977, "step": 660 }, { "epoch": 0.06656152594781875, "grad_norm": 0.6436944603919983, "learning_rate": 0.00029832292837276376, "loss": 1.4243, "step": 670 }, { "epoch": 0.06755498155898022, "grad_norm": 0.602425754070282, "learning_rate": 0.00029825094490968794, "loss": 1.4259, "step": 680 }, { "epoch": 0.0685484371701417, "grad_norm": 0.6178025603294373, "learning_rate": 0.00029817745794925415, "loss": 1.4354, "step": 690 }, { "epoch": 0.06954189278130317, "grad_norm": 0.63409423828125, "learning_rate": 0.00029810246823673537, "loss": 1.4314, "step": 700 }, { "epoch": 0.07053534839246464, "grad_norm": 0.6901246905326843, "learning_rate": 0.0002980259765326452, "loss": 1.4197, "step": 710 }, { "epoch": 0.07152880400362611, "grad_norm": 0.6050143837928772, "learning_rate": 0.00029794798361272964, "loss": 1.4255, "step": 720 }, { "epoch": 0.07252225961478759, "grad_norm": 0.59127277135849, "learning_rate": 0.0002978684902679592, "loss": 1.4323, "step": 730 }, { "epoch": 0.07351571522594906, "grad_norm": 0.5614879131317139, "learning_rate": 0.00029778749730452133, "loss": 1.3949, "step": 740 }, { "epoch": 0.07450917083711053, "grad_norm": 0.5925185680389404, "learning_rate": 0.0002977050055438118, "loss": 1.4169, "step": 750 }, { "epoch": 0.075502626448272, "grad_norm": 0.6346535682678223, "learning_rate": 0.0002976210158224266, "loss": 1.4183, "step": 760 }, { "epoch": 0.07649608205943348, "grad_norm": 0.6336431503295898, "learning_rate": 0.00029753552899215344, "loss": 1.3799, "step": 770 }, { "epoch": 0.07748953767059495, "grad_norm": 0.6319008469581604, "learning_rate": 0.0002974485459199629, "loss": 1.3991, "step": 780 }, { "epoch": 0.07848299328175642, "grad_norm": 0.5974861979484558, "learning_rate": 0.000297360067488, "loss": 1.4196, "step": 790 }, { "epoch": 0.0794764488929179, "grad_norm": 0.6162097454071045, "learning_rate": 0.00029727009459357487, "loss": 1.3995, "step": 800 }, { "epoch": 0.08046990450407937, "grad_norm": 0.5667980313301086, "learning_rate": 0.00029717862814915405, "loss": 1.3789, "step": 810 }, { "epoch": 0.08146336011524086, "grad_norm": 0.6212806105613708, "learning_rate": 0.00029708566908235077, "loss": 1.3819, "step": 820 }, { "epoch": 0.08245681572640233, "grad_norm": 0.5819615721702576, "learning_rate": 0.00029699121833591605, "loss": 1.3975, "step": 830 }, { "epoch": 0.0834502713375638, "grad_norm": 0.5672417879104614, "learning_rate": 0.0002968952768677287, "loss": 1.3884, "step": 840 }, { "epoch": 0.08444372694872528, "grad_norm": 0.567499041557312, "learning_rate": 0.00029679784565078584, "loss": 1.3972, "step": 850 }, { "epoch": 0.08543718255988675, "grad_norm": 0.7018792629241943, "learning_rate": 0.00029669892567319307, "loss": 1.3806, "step": 860 }, { "epoch": 0.08643063817104822, "grad_norm": 0.593984067440033, "learning_rate": 0.0002965985179381542, "loss": 1.3832, "step": 870 }, { "epoch": 0.0874240937822097, "grad_norm": 0.568536102771759, "learning_rate": 0.00029649662346396133, "loss": 1.3846, "step": 880 }, { "epoch": 0.08841754939337117, "grad_norm": 0.5485777258872986, "learning_rate": 0.0002963932432839844, "loss": 1.3981, "step": 890 }, { "epoch": 0.08941100500453264, "grad_norm": 0.6345034241676331, "learning_rate": 0.0002962883784466608, "loss": 1.3647, "step": 900 }, { "epoch": 0.09040446061569411, "grad_norm": 0.6001213192939758, "learning_rate": 0.00029618203001548465, "loss": 1.3908, "step": 910 }, { "epoch": 0.09139791622685559, "grad_norm": 0.5769600868225098, "learning_rate": 0.000296074199068996, "loss": 1.388, "step": 920 }, { "epoch": 0.09239137183801706, "grad_norm": 0.5454570651054382, "learning_rate": 0.0002959648867007699, "loss": 1.3648, "step": 930 }, { "epoch": 0.09338482744917853, "grad_norm": 0.5193580389022827, "learning_rate": 0.0002958540940194055, "loss": 1.3651, "step": 940 }, { "epoch": 0.09437828306034, "grad_norm": 0.6213797926902771, "learning_rate": 0.0002957418221485145, "loss": 1.3795, "step": 950 }, { "epoch": 0.09537173867150149, "grad_norm": 0.5813576579093933, "learning_rate": 0.00029562807222671, "loss": 1.3668, "step": 960 }, { "epoch": 0.09636519428266296, "grad_norm": 0.6071997880935669, "learning_rate": 0.00029551284540759473, "loss": 1.3474, "step": 970 }, { "epoch": 0.09735864989382444, "grad_norm": 0.6366572380065918, "learning_rate": 0.0002953961428597496, "loss": 1.3742, "step": 980 }, { "epoch": 0.09835210550498591, "grad_norm": 0.5855928659439087, "learning_rate": 0.00029527796576672173, "loss": 1.3807, "step": 990 }, { "epoch": 0.09934556111614738, "grad_norm": 0.5801404714584351, "learning_rate": 0.00029515831532701234, "loss": 1.364, "step": 1000 }, { "epoch": 0.1000409800439604, "eval_loss": 1.337802767753601, "eval_runtime": 217.2125, "eval_samples_per_second": 230.189, "eval_steps_per_second": 28.774, "step": 1007 }, { "epoch": 0.10033901672730886, "grad_norm": 0.6000391244888306, "learning_rate": 0.00029503719275406477, "loss": 1.3675, "step": 1010 }, { "epoch": 0.10133247233847033, "grad_norm": 0.6123185753822327, "learning_rate": 0.0002949145992762521, "loss": 1.3735, "step": 1020 }, { "epoch": 0.1023259279496318, "grad_norm": 0.610657811164856, "learning_rate": 0.0002947905361368647, "loss": 1.3546, "step": 1030 }, { "epoch": 0.10331938356079327, "grad_norm": 0.6239680051803589, "learning_rate": 0.00029466500459409756, "loss": 1.3486, "step": 1040 }, { "epoch": 0.10431283917195475, "grad_norm": 0.5398300886154175, "learning_rate": 0.0002945380059210377, "loss": 1.3575, "step": 1050 }, { "epoch": 0.10530629478311622, "grad_norm": 0.6671786904335022, "learning_rate": 0.000294409541405651, "loss": 1.3435, "step": 1060 }, { "epoch": 0.10629975039427769, "grad_norm": 0.6025234460830688, "learning_rate": 0.0002942796123507693, "loss": 1.3498, "step": 1070 }, { "epoch": 0.10729320600543917, "grad_norm": 0.5941970348358154, "learning_rate": 0.0002941482200740774, "loss": 1.355, "step": 1080 }, { "epoch": 0.10828666161660064, "grad_norm": 0.5551139712333679, "learning_rate": 0.00029401536590809913, "loss": 1.3345, "step": 1090 }, { "epoch": 0.10928011722776211, "grad_norm": 0.5619410276412964, "learning_rate": 0.00029388105120018446, "loss": 1.3361, "step": 1100 }, { "epoch": 0.1102735728389236, "grad_norm": 0.5853624939918518, "learning_rate": 0.00029374527731249533, "loss": 1.3438, "step": 1110 }, { "epoch": 0.11126702845008507, "grad_norm": 0.5613996386528015, "learning_rate": 0.00029360804562199224, "loss": 1.3356, "step": 1120 }, { "epoch": 0.11226048406124654, "grad_norm": 0.539417028427124, "learning_rate": 0.0002934693575204199, "loss": 1.3412, "step": 1130 }, { "epoch": 0.11325393967240802, "grad_norm": 0.570012092590332, "learning_rate": 0.00029332921441429345, "loss": 1.3558, "step": 1140 }, { "epoch": 0.11424739528356949, "grad_norm": 0.5662525296211243, "learning_rate": 0.0002931876177248841, "loss": 1.3532, "step": 1150 }, { "epoch": 0.11524085089473096, "grad_norm": 0.6034231781959534, "learning_rate": 0.0002930445688882045, "loss": 1.3469, "step": 1160 }, { "epoch": 0.11623430650589243, "grad_norm": 0.5469877123832703, "learning_rate": 0.00029290006935499453, "loss": 1.3246, "step": 1170 }, { "epoch": 0.11722776211705391, "grad_norm": 0.5636057257652283, "learning_rate": 0.00029275412059070636, "loss": 1.3527, "step": 1180 }, { "epoch": 0.11822121772821538, "grad_norm": 0.5818510055541992, "learning_rate": 0.0002926067240754896, "loss": 1.308, "step": 1190 }, { "epoch": 0.11921467333937685, "grad_norm": 0.5315884947776794, "learning_rate": 0.0002924578813041764, "loss": 1.341, "step": 1200 }, { "epoch": 0.12020812895053833, "grad_norm": 0.5662302374839783, "learning_rate": 0.000292307593786266, "loss": 1.3364, "step": 1210 }, { "epoch": 0.1212015845616998, "grad_norm": 0.597668468952179, "learning_rate": 0.00029215586304590996, "loss": 1.339, "step": 1220 }, { "epoch": 0.12219504017286127, "grad_norm": 0.5625566840171814, "learning_rate": 0.00029200269062189623, "loss": 1.3503, "step": 1230 }, { "epoch": 0.12318849578402274, "grad_norm": 0.5155960321426392, "learning_rate": 0.0002918480780676336, "loss": 1.3166, "step": 1240 }, { "epoch": 0.12418195139518422, "grad_norm": 0.5308235883712769, "learning_rate": 0.00029169202695113617, "loss": 1.3372, "step": 1250 }, { "epoch": 0.1251754070063457, "grad_norm": 0.5910816192626953, "learning_rate": 0.00029153453885500736, "loss": 1.3414, "step": 1260 }, { "epoch": 0.12616886261750718, "grad_norm": 0.5780258178710938, "learning_rate": 0.00029137561537642367, "loss": 1.3277, "step": 1270 }, { "epoch": 0.12716231822866864, "grad_norm": 0.6031139492988586, "learning_rate": 0.00029121525812711874, "loss": 1.3304, "step": 1280 }, { "epoch": 0.12815577383983012, "grad_norm": 0.5696874856948853, "learning_rate": 0.00029105346873336684, "loss": 1.3399, "step": 1290 }, { "epoch": 0.12914922945099158, "grad_norm": 0.5428373217582703, "learning_rate": 0.0002908902488359664, "loss": 1.3514, "step": 1300 }, { "epoch": 0.13014268506215307, "grad_norm": 0.5869764089584351, "learning_rate": 0.0002907256000902236, "loss": 1.3248, "step": 1310 }, { "epoch": 0.13113614067331453, "grad_norm": 0.5231671929359436, "learning_rate": 0.0002905595241659351, "loss": 1.3241, "step": 1320 }, { "epoch": 0.13212959628447601, "grad_norm": 0.5357515215873718, "learning_rate": 0.0002903920227473717, "loss": 1.3163, "step": 1330 }, { "epoch": 0.1331230518956375, "grad_norm": 0.5605716705322266, "learning_rate": 0.00029022309753326065, "loss": 1.3366, "step": 1340 }, { "epoch": 0.13411650750679896, "grad_norm": 0.5430563688278198, "learning_rate": 0.0002900527502367689, "loss": 1.3148, "step": 1350 }, { "epoch": 0.13510996311796045, "grad_norm": 0.571535050868988, "learning_rate": 0.0002898809825854855, "loss": 1.3227, "step": 1360 }, { "epoch": 0.1361034187291219, "grad_norm": 0.5518633723258972, "learning_rate": 0.00028970779632140416, "loss": 1.3149, "step": 1370 }, { "epoch": 0.1370968743402834, "grad_norm": 0.5351448059082031, "learning_rate": 0.0002895331932009055, "loss": 1.3057, "step": 1380 }, { "epoch": 0.13809032995144485, "grad_norm": 0.5526089668273926, "learning_rate": 0.0002893571749947393, "loss": 1.3167, "step": 1390 }, { "epoch": 0.13908378556260634, "grad_norm": 0.5857049226760864, "learning_rate": 0.0002891797434880066, "loss": 1.3246, "step": 1400 }, { "epoch": 0.1400772411737678, "grad_norm": 0.5625345706939697, "learning_rate": 0.0002890009004801415, "loss": 1.3059, "step": 1410 }, { "epoch": 0.14107069678492928, "grad_norm": 0.5215427279472351, "learning_rate": 0.0002888206477848928, "loss": 1.3116, "step": 1420 }, { "epoch": 0.14206415239609074, "grad_norm": 0.5295782685279846, "learning_rate": 0.00028863898723030594, "loss": 1.3154, "step": 1430 }, { "epoch": 0.14305760800725223, "grad_norm": 0.573486328125, "learning_rate": 0.00028845592065870413, "loss": 1.3134, "step": 1440 }, { "epoch": 0.1440510636184137, "grad_norm": 0.5500824451446533, "learning_rate": 0.00028827144992666986, "loss": 1.3151, "step": 1450 }, { "epoch": 0.14504451922957518, "grad_norm": 0.5774082541465759, "learning_rate": 0.000288085576905026, "loss": 1.307, "step": 1460 }, { "epoch": 0.14603797484073663, "grad_norm": 0.5409136414527893, "learning_rate": 0.0002878983034788169, "loss": 1.3201, "step": 1470 }, { "epoch": 0.14703143045189812, "grad_norm": 0.5557056069374084, "learning_rate": 0.00028770963154728905, "loss": 1.3092, "step": 1480 }, { "epoch": 0.1480248860630596, "grad_norm": 0.5294632911682129, "learning_rate": 0.0002875195630238721, "loss": 1.2847, "step": 1490 }, { "epoch": 0.14901834167422107, "grad_norm": 0.5896183848381042, "learning_rate": 0.0002873280998361594, "loss": 1.2891, "step": 1500 }, { "epoch": 0.15001179728538255, "grad_norm": 0.5323734283447266, "learning_rate": 0.00028713524392588827, "loss": 1.2986, "step": 1510 }, { "epoch": 0.151005252896544, "grad_norm": 0.5725658535957336, "learning_rate": 0.0002869409972489204, "loss": 1.309, "step": 1520 }, { "epoch": 0.1519987085077055, "grad_norm": 0.5617366433143616, "learning_rate": 0.00028674536177522206, "loss": 1.297, "step": 1530 }, { "epoch": 0.15299216411886696, "grad_norm": 0.5333068370819092, "learning_rate": 0.00028654833948884423, "loss": 1.302, "step": 1540 }, { "epoch": 0.15398561973002844, "grad_norm": 0.5795490145683289, "learning_rate": 0.0002863499323879022, "loss": 1.29, "step": 1550 }, { "epoch": 0.1549790753411899, "grad_norm": 0.5586980581283569, "learning_rate": 0.0002861501424845555, "loss": 1.2961, "step": 1560 }, { "epoch": 0.1559725309523514, "grad_norm": 0.593062698841095, "learning_rate": 0.00028594897180498745, "loss": 1.3074, "step": 1570 }, { "epoch": 0.15696598656351285, "grad_norm": 0.5487144589424133, "learning_rate": 0.00028574642238938467, "loss": 1.2938, "step": 1580 }, { "epoch": 0.15795944217467434, "grad_norm": 0.49600857496261597, "learning_rate": 0.00028554249629191616, "loss": 1.3077, "step": 1590 }, { "epoch": 0.1589528977858358, "grad_norm": 0.5532371997833252, "learning_rate": 0.00028533719558071285, "loss": 1.2892, "step": 1600 }, { "epoch": 0.15994635339699728, "grad_norm": 0.5463429689407349, "learning_rate": 0.0002851305223378462, "loss": 1.3017, "step": 1610 }, { "epoch": 0.16093980900815874, "grad_norm": 0.7598294019699097, "learning_rate": 0.0002849224786593075, "loss": 1.2844, "step": 1620 }, { "epoch": 0.16193326461932023, "grad_norm": 0.589923083782196, "learning_rate": 0.00028471306665498624, "loss": 1.304, "step": 1630 }, { "epoch": 0.16292672023048171, "grad_norm": 0.5281336307525635, "learning_rate": 0.000284502288448649, "loss": 1.2912, "step": 1640 }, { "epoch": 0.16392017584164317, "grad_norm": 0.5142702460289001, "learning_rate": 0.00028429014617791773, "loss": 1.2966, "step": 1650 }, { "epoch": 0.16491363145280466, "grad_norm": 0.5056828856468201, "learning_rate": 0.0002840766419942481, "loss": 1.2957, "step": 1660 }, { "epoch": 0.16590708706396612, "grad_norm": 0.5111185312271118, "learning_rate": 0.0002838617780629079, "loss": 1.2908, "step": 1670 }, { "epoch": 0.1669005426751276, "grad_norm": 0.5428747534751892, "learning_rate": 0.0002836455565629546, "loss": 1.2999, "step": 1680 }, { "epoch": 0.16789399828628906, "grad_norm": 0.5203731060028076, "learning_rate": 0.0002834279796872138, "loss": 1.2998, "step": 1690 }, { "epoch": 0.16888745389745055, "grad_norm": 0.5578112006187439, "learning_rate": 0.00028320904964225665, "loss": 1.2774, "step": 1700 }, { "epoch": 0.169880909508612, "grad_norm": 0.5637850165367126, "learning_rate": 0.0002829887686483775, "loss": 1.2941, "step": 1710 }, { "epoch": 0.1708743651197735, "grad_norm": 0.5894926190376282, "learning_rate": 0.0002827671389395716, "loss": 1.2879, "step": 1720 }, { "epoch": 0.17186782073093496, "grad_norm": 1.018946647644043, "learning_rate": 0.0002825441627635121, "loss": 1.2683, "step": 1730 }, { "epoch": 0.17286127634209644, "grad_norm": 0.5260887742042542, "learning_rate": 0.00028231984238152766, "loss": 1.2786, "step": 1740 }, { "epoch": 0.1738547319532579, "grad_norm": 0.5494965314865112, "learning_rate": 0.000282094180068579, "loss": 1.3083, "step": 1750 }, { "epoch": 0.1748481875644194, "grad_norm": 0.5133727788925171, "learning_rate": 0.0002818671781132364, "loss": 1.279, "step": 1760 }, { "epoch": 0.17584164317558085, "grad_norm": 0.5044395923614502, "learning_rate": 0.0002816388388176561, "loss": 1.2786, "step": 1770 }, { "epoch": 0.17683509878674233, "grad_norm": 0.5667337775230408, "learning_rate": 0.00028140916449755706, "loss": 1.2694, "step": 1780 }, { "epoch": 0.17782855439790382, "grad_norm": 0.5417622923851013, "learning_rate": 0.0002811781574821975, "loss": 1.2952, "step": 1790 }, { "epoch": 0.17882201000906528, "grad_norm": 0.5346735119819641, "learning_rate": 0.0002809458201143513, "loss": 1.2869, "step": 1800 }, { "epoch": 0.17981546562022677, "grad_norm": 0.475115567445755, "learning_rate": 0.0002807121547502843, "loss": 1.2775, "step": 1810 }, { "epoch": 0.18080892123138823, "grad_norm": 0.5102688670158386, "learning_rate": 0.00028047716375972996, "loss": 1.3103, "step": 1820 }, { "epoch": 0.1818023768425497, "grad_norm": 0.5632643103599548, "learning_rate": 0.000280240849525866, "loss": 1.2923, "step": 1830 }, { "epoch": 0.18279583245371117, "grad_norm": 0.5179861187934875, "learning_rate": 0.0002800032144452898, "loss": 1.2797, "step": 1840 }, { "epoch": 0.18378928806487266, "grad_norm": 0.5114406943321228, "learning_rate": 0.00027976426092799416, "loss": 1.28, "step": 1850 }, { "epoch": 0.18478274367603412, "grad_norm": 0.5344343781471252, "learning_rate": 0.000279523991397343, "loss": 1.2766, "step": 1860 }, { "epoch": 0.1857761992871956, "grad_norm": 0.4935680329799652, "learning_rate": 0.00027928240829004663, "loss": 1.2927, "step": 1870 }, { "epoch": 0.18676965489835706, "grad_norm": 0.5560677647590637, "learning_rate": 0.000279039514056137, "loss": 1.2671, "step": 1880 }, { "epoch": 0.18776311050951855, "grad_norm": 0.4876042902469635, "learning_rate": 0.0002787953111589431, "loss": 1.2711, "step": 1890 }, { "epoch": 0.18875656612068, "grad_norm": 0.5092322826385498, "learning_rate": 0.00027854980207506566, "loss": 1.2919, "step": 1900 }, { "epoch": 0.1897500217318415, "grad_norm": 0.5449466109275818, "learning_rate": 0.00027830298929435235, "loss": 1.2756, "step": 1910 }, { "epoch": 0.19074347734300298, "grad_norm": 0.5372341871261597, "learning_rate": 0.0002780548753198723, "loss": 1.2878, "step": 1920 }, { "epoch": 0.19173693295416444, "grad_norm": 0.5671026706695557, "learning_rate": 0.00027780546266789076, "loss": 1.2665, "step": 1930 }, { "epoch": 0.19273038856532593, "grad_norm": 0.4836546778678894, "learning_rate": 0.00027755475386784364, "loss": 1.2862, "step": 1940 }, { "epoch": 0.1937238441764874, "grad_norm": 0.5355504751205444, "learning_rate": 0.00027730275146231186, "loss": 1.276, "step": 1950 }, { "epoch": 0.19471729978764887, "grad_norm": 0.4880260229110718, "learning_rate": 0.0002770494580069954, "loss": 1.2778, "step": 1960 }, { "epoch": 0.19571075539881033, "grad_norm": 0.49935540556907654, "learning_rate": 0.0002767948760706877, "loss": 1.2787, "step": 1970 }, { "epoch": 0.19670421100997182, "grad_norm": 0.5408512353897095, "learning_rate": 0.00027653900823524925, "loss": 1.2841, "step": 1980 }, { "epoch": 0.19769766662113328, "grad_norm": 0.531233549118042, "learning_rate": 0.00027628185709558176, "loss": 1.2799, "step": 1990 }, { "epoch": 0.19869112223229476, "grad_norm": 0.5099895000457764, "learning_rate": 0.0002760234252596015, "loss": 1.2842, "step": 2000 }, { "epoch": 0.19968457784345622, "grad_norm": 0.5348741412162781, "learning_rate": 0.000275763715348213, "loss": 1.2491, "step": 2010 }, { "epoch": 0.2000819600879208, "eval_loss": 1.2746073007583618, "eval_runtime": 216.9749, "eval_samples_per_second": 230.441, "eval_steps_per_second": 28.805, "step": 2014 }, { "epoch": 0.2006780334546177, "grad_norm": 0.5270881056785583, "learning_rate": 0.00027550272999528265, "loss": 1.2673, "step": 2020 }, { "epoch": 0.20167148906577917, "grad_norm": 0.49110147356987, "learning_rate": 0.0002752404718476116, "loss": 1.2561, "step": 2030 }, { "epoch": 0.20266494467694066, "grad_norm": 0.4829249680042267, "learning_rate": 0.0002749769435649094, "loss": 1.268, "step": 2040 }, { "epoch": 0.20365840028810211, "grad_norm": 0.5357676148414612, "learning_rate": 0.00027471214781976645, "loss": 1.2626, "step": 2050 }, { "epoch": 0.2046518558992636, "grad_norm": 0.5558575391769409, "learning_rate": 0.00027444608729762754, "loss": 1.2562, "step": 2060 }, { "epoch": 0.2056453115104251, "grad_norm": 0.5351165533065796, "learning_rate": 0.000274178764696764, "loss": 1.2741, "step": 2070 }, { "epoch": 0.20663876712158655, "grad_norm": 0.5279297232627869, "learning_rate": 0.0002739101827282468, "loss": 1.2431, "step": 2080 }, { "epoch": 0.20763222273274803, "grad_norm": 0.48787686228752136, "learning_rate": 0.0002736403441159188, "loss": 1.2672, "step": 2090 }, { "epoch": 0.2086256783439095, "grad_norm": 0.519707977771759, "learning_rate": 0.0002733692515963672, "loss": 1.264, "step": 2100 }, { "epoch": 0.20961913395507098, "grad_norm": 0.48795512318611145, "learning_rate": 0.0002730969079188958, "loss": 1.2605, "step": 2110 }, { "epoch": 0.21061258956623244, "grad_norm": 0.5582920908927917, "learning_rate": 0.00027282331584549716, "loss": 1.2645, "step": 2120 }, { "epoch": 0.21160604517739393, "grad_norm": 0.5244210958480835, "learning_rate": 0.00027254847815082444, "loss": 1.2577, "step": 2130 }, { "epoch": 0.21259950078855538, "grad_norm": 0.510850191116333, "learning_rate": 0.0002722723976221634, "loss": 1.2462, "step": 2140 }, { "epoch": 0.21359295639971687, "grad_norm": 0.49397844076156616, "learning_rate": 0.00027199507705940413, "loss": 1.2725, "step": 2150 }, { "epoch": 0.21458641201087833, "grad_norm": 0.53153395652771, "learning_rate": 0.0002717165192750125, "loss": 1.2727, "step": 2160 }, { "epoch": 0.21557986762203982, "grad_norm": 0.5746108293533325, "learning_rate": 0.000271436727094002, "loss": 1.2576, "step": 2170 }, { "epoch": 0.21657332323320128, "grad_norm": 0.5098176002502441, "learning_rate": 0.0002711557033539044, "loss": 1.2503, "step": 2180 }, { "epoch": 0.21756677884436276, "grad_norm": 0.5269674062728882, "learning_rate": 0.0002708734509047419, "loss": 1.269, "step": 2190 }, { "epoch": 0.21856023445552422, "grad_norm": 0.5084241032600403, "learning_rate": 0.0002705899726089973, "loss": 1.249, "step": 2200 }, { "epoch": 0.2195536900666857, "grad_norm": 0.49269863963127136, "learning_rate": 0.00027030527134158585, "loss": 1.2597, "step": 2210 }, { "epoch": 0.2205471456778472, "grad_norm": 0.5180286169052124, "learning_rate": 0.00027001934998982524, "loss": 1.2289, "step": 2220 }, { "epoch": 0.22154060128900865, "grad_norm": 0.5060125589370728, "learning_rate": 0.0002697322114534071, "loss": 1.2557, "step": 2230 }, { "epoch": 0.22253405690017014, "grad_norm": 0.4967941641807556, "learning_rate": 0.0002694438586443669, "loss": 1.2596, "step": 2240 }, { "epoch": 0.2235275125113316, "grad_norm": 0.48990583419799805, "learning_rate": 0.0002691542944870549, "loss": 1.2463, "step": 2250 }, { "epoch": 0.2245209681224931, "grad_norm": 0.5117115378379822, "learning_rate": 0.00026886352191810627, "loss": 1.2314, "step": 2260 }, { "epoch": 0.22551442373365455, "grad_norm": 0.5112183690071106, "learning_rate": 0.0002685715438864116, "loss": 1.2411, "step": 2270 }, { "epoch": 0.22650787934481603, "grad_norm": 0.5363621711730957, "learning_rate": 0.00026827836335308644, "loss": 1.2472, "step": 2280 }, { "epoch": 0.2275013349559775, "grad_norm": 0.4729025661945343, "learning_rate": 0.00026798398329144166, "loss": 1.236, "step": 2290 }, { "epoch": 0.22849479056713898, "grad_norm": 0.5384593605995178, "learning_rate": 0.0002676884066869535, "loss": 1.2667, "step": 2300 }, { "epoch": 0.22948824617830044, "grad_norm": 0.48534199595451355, "learning_rate": 0.0002673916365372327, "loss": 1.2261, "step": 2310 }, { "epoch": 0.23048170178946192, "grad_norm": 0.4830830991268158, "learning_rate": 0.0002670936758519946, "loss": 1.2634, "step": 2320 }, { "epoch": 0.23147515740062338, "grad_norm": 0.5320212244987488, "learning_rate": 0.0002667945276530284, "loss": 1.2364, "step": 2330 }, { "epoch": 0.23246861301178487, "grad_norm": 0.4748252034187317, "learning_rate": 0.0002664941949741665, "loss": 1.2546, "step": 2340 }, { "epoch": 0.23346206862294633, "grad_norm": 0.5805224776268005, "learning_rate": 0.0002661926808612539, "loss": 1.2569, "step": 2350 }, { "epoch": 0.23445552423410781, "grad_norm": 0.4791530966758728, "learning_rate": 0.000265889988372117, "loss": 1.2639, "step": 2360 }, { "epoch": 0.2354489798452693, "grad_norm": 0.45609885454177856, "learning_rate": 0.0002655861205765331, "loss": 1.2589, "step": 2370 }, { "epoch": 0.23644243545643076, "grad_norm": 0.5291593670845032, "learning_rate": 0.00026528108055619856, "loss": 1.2485, "step": 2380 }, { "epoch": 0.23743589106759225, "grad_norm": 0.4872785806655884, "learning_rate": 0.0002649748714046983, "loss": 1.2436, "step": 2390 }, { "epoch": 0.2384293466787537, "grad_norm": 0.4877389371395111, "learning_rate": 0.0002646674962274739, "loss": 1.2463, "step": 2400 }, { "epoch": 0.2394228022899152, "grad_norm": 0.5236742496490479, "learning_rate": 0.0002643589581417922, "loss": 1.2537, "step": 2410 }, { "epoch": 0.24041625790107665, "grad_norm": 0.5331168174743652, "learning_rate": 0.0002640492602767139, "loss": 1.2531, "step": 2420 }, { "epoch": 0.24140971351223814, "grad_norm": 0.5398425459861755, "learning_rate": 0.00026373840577306165, "loss": 1.2413, "step": 2430 }, { "epoch": 0.2424031691233996, "grad_norm": 0.5574949979782104, "learning_rate": 0.00026342639778338813, "loss": 1.244, "step": 2440 }, { "epoch": 0.24339662473456108, "grad_norm": 0.4968748390674591, "learning_rate": 0.0002631132394719443, "loss": 1.2412, "step": 2450 }, { "epoch": 0.24439008034572254, "grad_norm": 0.5098783373832703, "learning_rate": 0.0002627989340146471, "loss": 1.2671, "step": 2460 }, { "epoch": 0.24538353595688403, "grad_norm": 0.4852953255176544, "learning_rate": 0.0002624834845990475, "loss": 1.2457, "step": 2470 }, { "epoch": 0.2463769915680455, "grad_norm": 0.5067562460899353, "learning_rate": 0.00026216689442429774, "loss": 1.242, "step": 2480 }, { "epoch": 0.24737044717920698, "grad_norm": 0.46974197030067444, "learning_rate": 0.0002618491667011193, "loss": 1.2425, "step": 2490 }, { "epoch": 0.24836390279036843, "grad_norm": 0.48239001631736755, "learning_rate": 0.0002615303046517703, "loss": 1.2468, "step": 2500 }, { "epoch": 0.24935735840152992, "grad_norm": 0.46368446946144104, "learning_rate": 0.00026121031151001245, "loss": 1.2401, "step": 2510 }, { "epoch": 0.2503508140126914, "grad_norm": 0.4512680470943451, "learning_rate": 0.00026088919052107867, "loss": 1.2537, "step": 2520 }, { "epoch": 0.25134426962385287, "grad_norm": 0.5069119334220886, "learning_rate": 0.00026056694494164, "loss": 1.2514, "step": 2530 }, { "epoch": 0.25233772523501435, "grad_norm": 0.48463037610054016, "learning_rate": 0.00026024357803977263, "loss": 1.238, "step": 2540 }, { "epoch": 0.25333118084617584, "grad_norm": 0.5035899877548218, "learning_rate": 0.0002599190930949247, "loss": 1.2601, "step": 2550 }, { "epoch": 0.25432463645733727, "grad_norm": 0.5174955725669861, "learning_rate": 0.00025959349339788295, "loss": 1.251, "step": 2560 }, { "epoch": 0.25531809206849876, "grad_norm": 0.48764464259147644, "learning_rate": 0.00025926678225073963, "loss": 1.2406, "step": 2570 }, { "epoch": 0.25631154767966025, "grad_norm": 0.48704472184181213, "learning_rate": 0.00025893896296685866, "loss": 1.2393, "step": 2580 }, { "epoch": 0.25730500329082173, "grad_norm": 0.5210717916488647, "learning_rate": 0.00025861003887084246, "loss": 1.2467, "step": 2590 }, { "epoch": 0.25829845890198316, "grad_norm": 0.49823907017707825, "learning_rate": 0.0002582800132984977, "loss": 1.243, "step": 2600 }, { "epoch": 0.25929191451314465, "grad_norm": 0.5364810228347778, "learning_rate": 0.0002579488895968019, "loss": 1.2313, "step": 2610 }, { "epoch": 0.26028537012430614, "grad_norm": 0.45561811327934265, "learning_rate": 0.0002576166711238694, "loss": 1.2194, "step": 2620 }, { "epoch": 0.2612788257354676, "grad_norm": 0.4552469849586487, "learning_rate": 0.0002572833612489169, "loss": 1.2601, "step": 2630 }, { "epoch": 0.26227228134662905, "grad_norm": 0.542340099811554, "learning_rate": 0.00025694896335223024, "loss": 1.2461, "step": 2640 }, { "epoch": 0.26326573695779054, "grad_norm": 0.5542205572128296, "learning_rate": 0.00025661348082512884, "loss": 1.2374, "step": 2650 }, { "epoch": 0.26425919256895203, "grad_norm": 0.4834555685520172, "learning_rate": 0.0002562769170699324, "loss": 1.2396, "step": 2660 }, { "epoch": 0.2652526481801135, "grad_norm": 0.5201283693313599, "learning_rate": 0.00025593927549992594, "loss": 1.2332, "step": 2670 }, { "epoch": 0.266246103791275, "grad_norm": 0.5460206270217896, "learning_rate": 0.000255600559539325, "loss": 1.2342, "step": 2680 }, { "epoch": 0.26723955940243643, "grad_norm": 0.46429523825645447, "learning_rate": 0.0002552607726232413, "loss": 1.2316, "step": 2690 }, { "epoch": 0.2682330150135979, "grad_norm": 0.5969310998916626, "learning_rate": 0.0002549199181976477, "loss": 1.2501, "step": 2700 }, { "epoch": 0.2692264706247594, "grad_norm": 0.48196089267730713, "learning_rate": 0.00025457799971934325, "loss": 1.2389, "step": 2710 }, { "epoch": 0.2702199262359209, "grad_norm": 0.5023167133331299, "learning_rate": 0.0002542350206559182, "loss": 1.2322, "step": 2720 }, { "epoch": 0.2712133818470823, "grad_norm": 0.47286665439605713, "learning_rate": 0.0002538909844857187, "loss": 1.2618, "step": 2730 }, { "epoch": 0.2722068374582438, "grad_norm": 0.5069117546081543, "learning_rate": 0.00025354589469781186, "loss": 1.244, "step": 2740 }, { "epoch": 0.2732002930694053, "grad_norm": 0.4682158827781677, "learning_rate": 0.0002531997547919499, "loss": 1.2455, "step": 2750 }, { "epoch": 0.2741937486805668, "grad_norm": 0.5212481617927551, "learning_rate": 0.0002528525682785349, "loss": 1.2362, "step": 2760 }, { "epoch": 0.2751872042917282, "grad_norm": 0.5417184233665466, "learning_rate": 0.0002525043386785833, "loss": 1.2382, "step": 2770 }, { "epoch": 0.2761806599028897, "grad_norm": 0.46628710627555847, "learning_rate": 0.00025215506952369003, "loss": 1.2395, "step": 2780 }, { "epoch": 0.2771741155140512, "grad_norm": 0.48039111495018005, "learning_rate": 0.0002518047643559927, "loss": 1.2459, "step": 2790 }, { "epoch": 0.2781675711252127, "grad_norm": 0.4942474961280823, "learning_rate": 0.00025145342672813574, "loss": 1.2441, "step": 2800 }, { "epoch": 0.27916102673637416, "grad_norm": 0.4979240894317627, "learning_rate": 0.00025110106020323437, "loss": 1.2372, "step": 2810 }, { "epoch": 0.2801544823475356, "grad_norm": 0.4905546009540558, "learning_rate": 0.0002507476683548384, "loss": 1.2351, "step": 2820 }, { "epoch": 0.2811479379586971, "grad_norm": 0.4817952513694763, "learning_rate": 0.00025039325476689607, "loss": 1.2204, "step": 2830 }, { "epoch": 0.28214139356985857, "grad_norm": 0.4718853533267975, "learning_rate": 0.00025003782303371767, "loss": 1.2331, "step": 2840 }, { "epoch": 0.28313484918102005, "grad_norm": 0.48936033248901367, "learning_rate": 0.00024968137675993897, "loss": 1.2367, "step": 2850 }, { "epoch": 0.2841283047921815, "grad_norm": 0.503688633441925, "learning_rate": 0.00024932391956048497, "loss": 1.2189, "step": 2860 }, { "epoch": 0.28512176040334297, "grad_norm": 0.4779193699359894, "learning_rate": 0.00024896545506053285, "loss": 1.2485, "step": 2870 }, { "epoch": 0.28611521601450446, "grad_norm": 0.4970162808895111, "learning_rate": 0.00024860598689547555, "loss": 1.2167, "step": 2880 }, { "epoch": 0.28710867162566595, "grad_norm": 0.5190041065216064, "learning_rate": 0.0002482455187108846, "loss": 1.196, "step": 2890 }, { "epoch": 0.2881021272368274, "grad_norm": 0.5404960513114929, "learning_rate": 0.00024788405416247347, "loss": 1.2235, "step": 2900 }, { "epoch": 0.28909558284798886, "grad_norm": 0.4441899359226227, "learning_rate": 0.00024752159691606015, "loss": 1.2471, "step": 2910 }, { "epoch": 0.29008903845915035, "grad_norm": 0.48505038022994995, "learning_rate": 0.0002471581506475303, "loss": 1.2312, "step": 2920 }, { "epoch": 0.29108249407031184, "grad_norm": 0.49762651324272156, "learning_rate": 0.00024679371904279973, "loss": 1.2261, "step": 2930 }, { "epoch": 0.29207594968147327, "grad_norm": 0.5253481268882751, "learning_rate": 0.0002464283057977771, "loss": 1.2405, "step": 2940 }, { "epoch": 0.29306940529263475, "grad_norm": 0.4968172311782837, "learning_rate": 0.00024606191461832655, "loss": 1.2225, "step": 2950 }, { "epoch": 0.29406286090379624, "grad_norm": 0.4646175503730774, "learning_rate": 0.00024569454922022987, "loss": 1.2109, "step": 2960 }, { "epoch": 0.29505631651495773, "grad_norm": 0.49472692608833313, "learning_rate": 0.00024532621332914897, "loss": 1.2191, "step": 2970 }, { "epoch": 0.2960497721261192, "grad_norm": 0.49351736903190613, "learning_rate": 0.0002449569106805882, "loss": 1.216, "step": 2980 }, { "epoch": 0.29704322773728065, "grad_norm": 0.49852266907691956, "learning_rate": 0.00024458664501985624, "loss": 1.2301, "step": 2990 }, { "epoch": 0.29803668334844213, "grad_norm": 0.5016636848449707, "learning_rate": 0.0002442154201020283, "loss": 1.2289, "step": 3000 }, { "epoch": 0.2990301389596036, "grad_norm": 0.4916824996471405, "learning_rate": 0.00024384323969190789, "loss": 1.2274, "step": 3010 }, { "epoch": 0.3000235945707651, "grad_norm": 0.46829256415367126, "learning_rate": 0.0002434701075639888, "loss": 1.2118, "step": 3020 }, { "epoch": 0.30012294013188123, "eval_loss": 1.2400528192520142, "eval_runtime": 217.1452, "eval_samples_per_second": 230.261, "eval_steps_per_second": 28.783, "step": 3021 }, { "epoch": 0.30101705018192654, "grad_norm": 0.44890084862709045, "learning_rate": 0.0002430960275024166, "loss": 1.2103, "step": 3030 }, { "epoch": 0.302010505793088, "grad_norm": 0.4859289526939392, "learning_rate": 0.00024272100330095055, "loss": 1.2367, "step": 3040 }, { "epoch": 0.3030039614042495, "grad_norm": 0.4968155026435852, "learning_rate": 0.00024234503876292485, "loss": 1.235, "step": 3050 }, { "epoch": 0.303997417015411, "grad_norm": 0.5458197593688965, "learning_rate": 0.00024196813770121026, "loss": 1.2294, "step": 3060 }, { "epoch": 0.30499087262657243, "grad_norm": 0.4604020118713379, "learning_rate": 0.00024159030393817538, "loss": 1.222, "step": 3070 }, { "epoch": 0.3059843282377339, "grad_norm": 0.5177088379859924, "learning_rate": 0.0002412115413056477, "loss": 1.2068, "step": 3080 }, { "epoch": 0.3069777838488954, "grad_norm": 0.47766199707984924, "learning_rate": 0.00024083185364487522, "loss": 1.2256, "step": 3090 }, { "epoch": 0.3079712394600569, "grad_norm": 0.5014012455940247, "learning_rate": 0.0002404512448064868, "loss": 1.1924, "step": 3100 }, { "epoch": 0.3089646950712184, "grad_norm": 0.4730978310108185, "learning_rate": 0.00024006971865045377, "loss": 1.2349, "step": 3110 }, { "epoch": 0.3099581506823798, "grad_norm": 0.45509788393974304, "learning_rate": 0.0002396872790460504, "loss": 1.2177, "step": 3120 }, { "epoch": 0.3109516062935413, "grad_norm": 0.4922787845134735, "learning_rate": 0.00023930392987181473, "loss": 1.2151, "step": 3130 }, { "epoch": 0.3119450619047028, "grad_norm": 0.4871610403060913, "learning_rate": 0.00023891967501550937, "loss": 1.2133, "step": 3140 }, { "epoch": 0.31293851751586427, "grad_norm": 0.4459008276462555, "learning_rate": 0.00023853451837408185, "loss": 1.2097, "step": 3150 }, { "epoch": 0.3139319731270257, "grad_norm": 0.41748374700546265, "learning_rate": 0.00023814846385362525, "loss": 1.2236, "step": 3160 }, { "epoch": 0.3149254287381872, "grad_norm": 0.5209752321243286, "learning_rate": 0.00023776151536933864, "loss": 1.2015, "step": 3170 }, { "epoch": 0.31591888434934867, "grad_norm": 0.4768379032611847, "learning_rate": 0.00023737367684548716, "loss": 1.2233, "step": 3180 }, { "epoch": 0.31691233996051016, "grad_norm": 0.46641460061073303, "learning_rate": 0.00023698495221536245, "loss": 1.2144, "step": 3190 }, { "epoch": 0.3179057955716716, "grad_norm": 0.49821555614471436, "learning_rate": 0.00023659534542124261, "loss": 1.2003, "step": 3200 }, { "epoch": 0.3188992511828331, "grad_norm": 0.5253980159759521, "learning_rate": 0.0002362048604143522, "loss": 1.2165, "step": 3210 }, { "epoch": 0.31989270679399456, "grad_norm": 0.5121644735336304, "learning_rate": 0.00023581350115482245, "loss": 1.2076, "step": 3220 }, { "epoch": 0.32088616240515605, "grad_norm": 0.499426007270813, "learning_rate": 0.00023542127161165063, "loss": 1.2294, "step": 3230 }, { "epoch": 0.3218796180163175, "grad_norm": 0.48408859968185425, "learning_rate": 0.00023502817576266023, "loss": 1.2156, "step": 3240 }, { "epoch": 0.32287307362747897, "grad_norm": 0.4970872402191162, "learning_rate": 0.00023463421759446033, "loss": 1.2057, "step": 3250 }, { "epoch": 0.32386652923864045, "grad_norm": 0.5192016363143921, "learning_rate": 0.00023423940110240538, "loss": 1.234, "step": 3260 }, { "epoch": 0.32485998484980194, "grad_norm": 0.47482579946517944, "learning_rate": 0.00023384373029055445, "loss": 1.2087, "step": 3270 }, { "epoch": 0.32585344046096343, "grad_norm": 0.4967539608478546, "learning_rate": 0.00023344720917163093, "loss": 1.203, "step": 3280 }, { "epoch": 0.32684689607212486, "grad_norm": 0.46486181020736694, "learning_rate": 0.00023304984176698157, "loss": 1.2044, "step": 3290 }, { "epoch": 0.32784035168328635, "grad_norm": 0.45617178082466125, "learning_rate": 0.00023265163210653569, "loss": 1.2189, "step": 3300 }, { "epoch": 0.32883380729444783, "grad_norm": 0.46393707394599915, "learning_rate": 0.00023225258422876455, "loss": 1.2091, "step": 3310 }, { "epoch": 0.3298272629056093, "grad_norm": 0.4585002064704895, "learning_rate": 0.00023185270218064021, "loss": 1.2031, "step": 3320 }, { "epoch": 0.33082071851677075, "grad_norm": 0.4875582158565521, "learning_rate": 0.00023145199001759452, "loss": 1.2075, "step": 3330 }, { "epoch": 0.33181417412793224, "grad_norm": 0.483347624540329, "learning_rate": 0.00023105045180347788, "loss": 1.2256, "step": 3340 }, { "epoch": 0.3328076297390937, "grad_norm": 0.5154065489768982, "learning_rate": 0.00023064809161051842, "loss": 1.2059, "step": 3350 }, { "epoch": 0.3338010853502552, "grad_norm": 0.5070950984954834, "learning_rate": 0.00023024491351928019, "loss": 1.2033, "step": 3360 }, { "epoch": 0.33479454096141664, "grad_norm": 0.5440912842750549, "learning_rate": 0.00022984092161862206, "loss": 1.2197, "step": 3370 }, { "epoch": 0.33578799657257813, "grad_norm": 0.5029067993164062, "learning_rate": 0.00022943612000565625, "loss": 1.1932, "step": 3380 }, { "epoch": 0.3367814521837396, "grad_norm": 0.5002272129058838, "learning_rate": 0.00022903051278570677, "loss": 1.2088, "step": 3390 }, { "epoch": 0.3377749077949011, "grad_norm": 0.46260136365890503, "learning_rate": 0.00022862410407226762, "loss": 1.2287, "step": 3400 }, { "epoch": 0.3387683634060626, "grad_norm": 0.558641791343689, "learning_rate": 0.00022821689798696132, "loss": 1.2198, "step": 3410 }, { "epoch": 0.339761819017224, "grad_norm": 0.4847974181175232, "learning_rate": 0.00022780889865949689, "loss": 1.2059, "step": 3420 }, { "epoch": 0.3407552746283855, "grad_norm": 0.4504203796386719, "learning_rate": 0.00022740011022762827, "loss": 1.2089, "step": 3430 }, { "epoch": 0.341748730239547, "grad_norm": 0.5143907070159912, "learning_rate": 0.00022699053683711187, "loss": 1.2274, "step": 3440 }, { "epoch": 0.3427421858507085, "grad_norm": 0.4598551094532013, "learning_rate": 0.00022658018264166502, "loss": 1.2136, "step": 3450 }, { "epoch": 0.3437356414618699, "grad_norm": 0.45654499530792236, "learning_rate": 0.00022616905180292363, "loss": 1.1994, "step": 3460 }, { "epoch": 0.3447290970730314, "grad_norm": 0.48295536637306213, "learning_rate": 0.00022575714849039984, "loss": 1.21, "step": 3470 }, { "epoch": 0.3457225526841929, "grad_norm": 0.5002532005310059, "learning_rate": 0.00022534447688144006, "loss": 1.2229, "step": 3480 }, { "epoch": 0.34671600829535437, "grad_norm": 0.45252901315689087, "learning_rate": 0.0002249310411611824, "loss": 1.1866, "step": 3490 }, { "epoch": 0.3477094639065158, "grad_norm": 0.46162939071655273, "learning_rate": 0.00022451684552251414, "loss": 1.2159, "step": 3500 }, { "epoch": 0.3487029195176773, "grad_norm": 0.46276578307151794, "learning_rate": 0.00022410189416602943, "loss": 1.2128, "step": 3510 }, { "epoch": 0.3496963751288388, "grad_norm": 0.5181559920310974, "learning_rate": 0.00022368619129998655, "loss": 1.2188, "step": 3520 }, { "epoch": 0.35068983074000026, "grad_norm": 0.4349026381969452, "learning_rate": 0.00022326974114026535, "loss": 1.2045, "step": 3530 }, { "epoch": 0.3516832863511617, "grad_norm": 0.45742759108543396, "learning_rate": 0.00022285254791032423, "loss": 1.2024, "step": 3540 }, { "epoch": 0.3526767419623232, "grad_norm": 0.48607686161994934, "learning_rate": 0.0002224346158411577, "loss": 1.201, "step": 3550 }, { "epoch": 0.35367019757348467, "grad_norm": 0.4831947982311249, "learning_rate": 0.00022201594917125306, "loss": 1.1895, "step": 3560 }, { "epoch": 0.35466365318464615, "grad_norm": 0.5034534335136414, "learning_rate": 0.00022159655214654787, "loss": 1.2008, "step": 3570 }, { "epoch": 0.35565710879580764, "grad_norm": 0.47319284081459045, "learning_rate": 0.00022117642902038629, "loss": 1.1967, "step": 3580 }, { "epoch": 0.3566505644069691, "grad_norm": 0.5058645009994507, "learning_rate": 0.00022075558405347659, "loss": 1.2149, "step": 3590 }, { "epoch": 0.35764402001813056, "grad_norm": 0.4877967834472656, "learning_rate": 0.0002203340215138475, "loss": 1.2252, "step": 3600 }, { "epoch": 0.35863747562929205, "grad_norm": 0.4703255593776703, "learning_rate": 0.00021991174567680512, "loss": 1.195, "step": 3610 }, { "epoch": 0.35963093124045353, "grad_norm": 0.4860426187515259, "learning_rate": 0.00021948876082488936, "loss": 1.1959, "step": 3620 }, { "epoch": 0.36062438685161496, "grad_norm": 0.46194490790367126, "learning_rate": 0.0002190650712478309, "loss": 1.2074, "step": 3630 }, { "epoch": 0.36161784246277645, "grad_norm": 0.44126129150390625, "learning_rate": 0.0002186406812425073, "loss": 1.1912, "step": 3640 }, { "epoch": 0.36261129807393794, "grad_norm": 0.48812511563301086, "learning_rate": 0.00021821559511289948, "loss": 1.2068, "step": 3650 }, { "epoch": 0.3636047536850994, "grad_norm": 0.4943760931491852, "learning_rate": 0.0002177898171700483, "loss": 1.1923, "step": 3660 }, { "epoch": 0.36459820929626086, "grad_norm": 0.4492897391319275, "learning_rate": 0.00021736335173201066, "loss": 1.1859, "step": 3670 }, { "epoch": 0.36559166490742234, "grad_norm": 0.45745688676834106, "learning_rate": 0.00021693620312381568, "loss": 1.1978, "step": 3680 }, { "epoch": 0.36658512051858383, "grad_norm": 0.5050104856491089, "learning_rate": 0.00021650837567742094, "loss": 1.2172, "step": 3690 }, { "epoch": 0.3675785761297453, "grad_norm": 0.5013807415962219, "learning_rate": 0.00021607987373166855, "loss": 1.2039, "step": 3700 }, { "epoch": 0.3685720317409068, "grad_norm": 0.4632636308670044, "learning_rate": 0.00021565070163224095, "loss": 1.2005, "step": 3710 }, { "epoch": 0.36956548735206823, "grad_norm": 0.46821290254592896, "learning_rate": 0.0002152208637316172, "loss": 1.2227, "step": 3720 }, { "epoch": 0.3705589429632297, "grad_norm": 0.4609948396682739, "learning_rate": 0.00021479036438902847, "loss": 1.2018, "step": 3730 }, { "epoch": 0.3715523985743912, "grad_norm": 0.5113796591758728, "learning_rate": 0.00021435920797041404, "loss": 1.1925, "step": 3740 }, { "epoch": 0.3725458541855527, "grad_norm": 0.47701483964920044, "learning_rate": 0.00021392739884837697, "loss": 1.1992, "step": 3750 }, { "epoch": 0.3735393097967141, "grad_norm": 0.4520336389541626, "learning_rate": 0.00021349494140213986, "loss": 1.202, "step": 3760 }, { "epoch": 0.3745327654078756, "grad_norm": 0.4652255177497864, "learning_rate": 0.00021306184001750012, "loss": 1.1829, "step": 3770 }, { "epoch": 0.3755262210190371, "grad_norm": 0.4625130295753479, "learning_rate": 0.00021262809908678584, "loss": 1.1927, "step": 3780 }, { "epoch": 0.3765196766301986, "grad_norm": 0.44085580110549927, "learning_rate": 0.00021219372300881117, "loss": 1.2191, "step": 3790 }, { "epoch": 0.37751313224136, "grad_norm": 0.47911298274993896, "learning_rate": 0.0002117587161888315, "loss": 1.1941, "step": 3800 }, { "epoch": 0.3785065878525215, "grad_norm": 0.49052807688713074, "learning_rate": 0.00021132308303849906, "loss": 1.214, "step": 3810 }, { "epoch": 0.379500043463683, "grad_norm": 0.48691338300704956, "learning_rate": 0.000210886827975818, "loss": 1.2108, "step": 3820 }, { "epoch": 0.3804934990748445, "grad_norm": 0.44401276111602783, "learning_rate": 0.0002104499554250996, "loss": 1.1792, "step": 3830 }, { "epoch": 0.38148695468600596, "grad_norm": 0.4931083917617798, "learning_rate": 0.00021001246981691758, "loss": 1.2012, "step": 3840 }, { "epoch": 0.3824804102971674, "grad_norm": 0.472159206867218, "learning_rate": 0.0002095743755880628, "loss": 1.2112, "step": 3850 }, { "epoch": 0.3834738659083289, "grad_norm": 0.50935298204422, "learning_rate": 0.0002091356771814987, "loss": 1.206, "step": 3860 }, { "epoch": 0.38446732151949037, "grad_norm": 0.5085238814353943, "learning_rate": 0.0002086963790463159, "loss": 1.2066, "step": 3870 }, { "epoch": 0.38546077713065185, "grad_norm": 0.47772541642189026, "learning_rate": 0.0002082564856376873, "loss": 1.1908, "step": 3880 }, { "epoch": 0.3864542327418133, "grad_norm": 0.5096874833106995, "learning_rate": 0.0002078160014168227, "loss": 1.197, "step": 3890 }, { "epoch": 0.3874476883529748, "grad_norm": 0.5250852704048157, "learning_rate": 0.00020737493085092382, "loss": 1.1967, "step": 3900 }, { "epoch": 0.38844114396413626, "grad_norm": 0.45944342017173767, "learning_rate": 0.00020693327841313867, "loss": 1.2116, "step": 3910 }, { "epoch": 0.38943459957529775, "grad_norm": 0.4569138288497925, "learning_rate": 0.00020649104858251648, "loss": 1.2057, "step": 3920 }, { "epoch": 0.3904280551864592, "grad_norm": 0.4284946322441101, "learning_rate": 0.00020604824584396204, "loss": 1.1949, "step": 3930 }, { "epoch": 0.39142151079762066, "grad_norm": 0.4908120036125183, "learning_rate": 0.0002056048746881905, "loss": 1.1973, "step": 3940 }, { "epoch": 0.39241496640878215, "grad_norm": 0.49430763721466064, "learning_rate": 0.00020516093961168137, "loss": 1.2042, "step": 3950 }, { "epoch": 0.39340842201994364, "grad_norm": 0.45789289474487305, "learning_rate": 0.00020471644511663352, "loss": 1.1794, "step": 3960 }, { "epoch": 0.39440187763110507, "grad_norm": 0.5000009536743164, "learning_rate": 0.00020427139571091893, "loss": 1.1859, "step": 3970 }, { "epoch": 0.39539533324226656, "grad_norm": 0.4725204110145569, "learning_rate": 0.00020382579590803748, "loss": 1.1818, "step": 3980 }, { "epoch": 0.39638878885342804, "grad_norm": 0.4737090766429901, "learning_rate": 0.00020337965022707084, "loss": 1.1922, "step": 3990 }, { "epoch": 0.39738224446458953, "grad_norm": 0.44001784920692444, "learning_rate": 0.00020293296319263664, "loss": 1.1988, "step": 4000 }, { "epoch": 0.398375700075751, "grad_norm": 0.48931002616882324, "learning_rate": 0.00020248573933484286, "loss": 1.1952, "step": 4010 }, { "epoch": 0.39936915568691245, "grad_norm": 0.48461276292800903, "learning_rate": 0.00020203798318924152, "loss": 1.1918, "step": 4020 }, { "epoch": 0.4001639201758416, "eval_loss": 1.209017038345337, "eval_runtime": 217.2474, "eval_samples_per_second": 230.152, "eval_steps_per_second": 28.769, "step": 4028 }, { "epoch": 0.40036261129807393, "grad_norm": 0.4797476530075073, "learning_rate": 0.00020158969929678306, "loss": 1.1786, "step": 4030 }, { "epoch": 0.4013560669092354, "grad_norm": 0.46265390515327454, "learning_rate": 0.00020114089220377005, "loss": 1.2022, "step": 4040 }, { "epoch": 0.4023495225203969, "grad_norm": 0.4856548607349396, "learning_rate": 0.00020069156646181107, "loss": 1.2066, "step": 4050 }, { "epoch": 0.40334297813155834, "grad_norm": 0.5205573439598083, "learning_rate": 0.0002002417266277746, "loss": 1.2003, "step": 4060 }, { "epoch": 0.4043364337427198, "grad_norm": 0.484475314617157, "learning_rate": 0.00019979137726374295, "loss": 1.1849, "step": 4070 }, { "epoch": 0.4053298893538813, "grad_norm": 0.5231068730354309, "learning_rate": 0.0001993405229369658, "loss": 1.1948, "step": 4080 }, { "epoch": 0.4063233449650428, "grad_norm": 0.46499723196029663, "learning_rate": 0.0001988891682198139, "loss": 1.1968, "step": 4090 }, { "epoch": 0.40731680057620423, "grad_norm": 0.4803023934364319, "learning_rate": 0.00019843731768973274, "loss": 1.1899, "step": 4100 }, { "epoch": 0.4083102561873657, "grad_norm": 0.43668457865715027, "learning_rate": 0.00019798497592919624, "loss": 1.1801, "step": 4110 }, { "epoch": 0.4093037117985272, "grad_norm": 0.45928719639778137, "learning_rate": 0.00019753214752566006, "loss": 1.1942, "step": 4120 }, { "epoch": 0.4102971674096887, "grad_norm": 0.4647465646266937, "learning_rate": 0.0001970788370715152, "loss": 1.1839, "step": 4130 }, { "epoch": 0.4112906230208502, "grad_norm": 0.43589547276496887, "learning_rate": 0.0001966250491640415, "loss": 1.1944, "step": 4140 }, { "epoch": 0.4122840786320116, "grad_norm": 0.5093525052070618, "learning_rate": 0.0001961707884053608, "loss": 1.1657, "step": 4150 }, { "epoch": 0.4132775342431731, "grad_norm": 0.4544735848903656, "learning_rate": 0.0001957160594023905, "loss": 1.1793, "step": 4160 }, { "epoch": 0.4142709898543346, "grad_norm": 0.43822965025901794, "learning_rate": 0.00019526086676679663, "loss": 1.1679, "step": 4170 }, { "epoch": 0.41526444546549607, "grad_norm": 0.43124911189079285, "learning_rate": 0.00019480521511494727, "loss": 1.1797, "step": 4180 }, { "epoch": 0.4162579010766575, "grad_norm": 0.44309306144714355, "learning_rate": 0.0001943491090678656, "loss": 1.1674, "step": 4190 }, { "epoch": 0.417251356687819, "grad_norm": 0.4919864237308502, "learning_rate": 0.00019389255325118305, "loss": 1.1823, "step": 4200 }, { "epoch": 0.4182448122989805, "grad_norm": 0.4437265396118164, "learning_rate": 0.0001934355522950926, "loss": 1.1908, "step": 4210 }, { "epoch": 0.41923826791014196, "grad_norm": 0.5036196112632751, "learning_rate": 0.00019297811083430138, "loss": 1.1932, "step": 4220 }, { "epoch": 0.4202317235213034, "grad_norm": 0.4650508463382721, "learning_rate": 0.0001925202335079841, "loss": 1.1797, "step": 4230 }, { "epoch": 0.4212251791324649, "grad_norm": 0.44443807005882263, "learning_rate": 0.0001920619249597357, "loss": 1.2067, "step": 4240 }, { "epoch": 0.42221863474362636, "grad_norm": 0.49494221806526184, "learning_rate": 0.00019160318983752465, "loss": 1.2038, "step": 4250 }, { "epoch": 0.42321209035478785, "grad_norm": 0.48809465765953064, "learning_rate": 0.00019114403279364524, "loss": 1.1701, "step": 4260 }, { "epoch": 0.4242055459659493, "grad_norm": 0.4628167748451233, "learning_rate": 0.00019068445848467088, "loss": 1.1821, "step": 4270 }, { "epoch": 0.42519900157711077, "grad_norm": 0.5052282214164734, "learning_rate": 0.00019022447157140667, "loss": 1.166, "step": 4280 }, { "epoch": 0.42619245718827226, "grad_norm": 0.45449918508529663, "learning_rate": 0.00018976407671884217, "loss": 1.172, "step": 4290 }, { "epoch": 0.42718591279943374, "grad_norm": 0.463618665933609, "learning_rate": 0.000189303278596104, "loss": 1.1616, "step": 4300 }, { "epoch": 0.42817936841059523, "grad_norm": 0.48953506350517273, "learning_rate": 0.00018884208187640875, "loss": 1.2022, "step": 4310 }, { "epoch": 0.42917282402175666, "grad_norm": 0.5115376114845276, "learning_rate": 0.00018838049123701524, "loss": 1.18, "step": 4320 }, { "epoch": 0.43016627963291815, "grad_norm": 0.44180265069007874, "learning_rate": 0.00018791851135917739, "loss": 1.1851, "step": 4330 }, { "epoch": 0.43115973524407963, "grad_norm": 0.46057870984077454, "learning_rate": 0.00018745614692809644, "loss": 1.1856, "step": 4340 }, { "epoch": 0.4321531908552411, "grad_norm": 0.4646226167678833, "learning_rate": 0.0001869934026328738, "loss": 1.1756, "step": 4350 }, { "epoch": 0.43314664646640255, "grad_norm": 0.4839836657047272, "learning_rate": 0.00018653028316646317, "loss": 1.203, "step": 4360 }, { "epoch": 0.43414010207756404, "grad_norm": 0.44381341338157654, "learning_rate": 0.00018606679322562307, "loss": 1.1633, "step": 4370 }, { "epoch": 0.4351335576887255, "grad_norm": 0.48917341232299805, "learning_rate": 0.00018560293751086927, "loss": 1.1794, "step": 4380 }, { "epoch": 0.436127013299887, "grad_norm": 0.46939659118652344, "learning_rate": 0.00018513872072642708, "loss": 1.1754, "step": 4390 }, { "epoch": 0.43712046891104844, "grad_norm": 0.5408668518066406, "learning_rate": 0.00018467414758018358, "loss": 1.1794, "step": 4400 }, { "epoch": 0.43811392452220993, "grad_norm": 0.4510866105556488, "learning_rate": 0.0001842092227836398, "loss": 1.1839, "step": 4410 }, { "epoch": 0.4391073801333714, "grad_norm": 0.45456522703170776, "learning_rate": 0.00018374395105186342, "loss": 1.2092, "step": 4420 }, { "epoch": 0.4401008357445329, "grad_norm": 0.49070245027542114, "learning_rate": 0.00018327833710344012, "loss": 1.1782, "step": 4430 }, { "epoch": 0.4410942913556944, "grad_norm": 0.49562186002731323, "learning_rate": 0.00018281238566042655, "loss": 1.1865, "step": 4440 }, { "epoch": 0.4420877469668558, "grad_norm": 0.4716269373893738, "learning_rate": 0.00018234610144830189, "loss": 1.1852, "step": 4450 }, { "epoch": 0.4430812025780173, "grad_norm": 0.4782949388027191, "learning_rate": 0.00018187948919592023, "loss": 1.17, "step": 4460 }, { "epoch": 0.4440746581891788, "grad_norm": 0.6574309468269348, "learning_rate": 0.00018141255363546244, "loss": 1.1777, "step": 4470 }, { "epoch": 0.4450681138003403, "grad_norm": 0.4680328965187073, "learning_rate": 0.00018094529950238829, "loss": 1.1595, "step": 4480 }, { "epoch": 0.4460615694115017, "grad_norm": 0.44611451029777527, "learning_rate": 0.00018047773153538835, "loss": 1.1956, "step": 4490 }, { "epoch": 0.4470550250226632, "grad_norm": 0.4821056127548218, "learning_rate": 0.00018000985447633592, "loss": 1.1744, "step": 4500 }, { "epoch": 0.4480484806338247, "grad_norm": 0.4614666700363159, "learning_rate": 0.00017954167307023902, "loss": 1.1776, "step": 4510 }, { "epoch": 0.4490419362449862, "grad_norm": 0.4979506731033325, "learning_rate": 0.00017907319206519228, "loss": 1.1666, "step": 4520 }, { "epoch": 0.4500353918561476, "grad_norm": 0.4544410705566406, "learning_rate": 0.00017860441621232864, "loss": 1.189, "step": 4530 }, { "epoch": 0.4510288474673091, "grad_norm": 0.5314710736274719, "learning_rate": 0.00017813535026577128, "loss": 1.1726, "step": 4540 }, { "epoch": 0.4520223030784706, "grad_norm": 0.4279493987560272, "learning_rate": 0.0001776659989825854, "loss": 1.1875, "step": 4550 }, { "epoch": 0.45301575868963206, "grad_norm": 0.4388754963874817, "learning_rate": 0.00017719636712273005, "loss": 1.1734, "step": 4560 }, { "epoch": 0.4540092143007935, "grad_norm": 0.490273118019104, "learning_rate": 0.0001767264594490095, "loss": 1.1733, "step": 4570 }, { "epoch": 0.455002669911955, "grad_norm": 0.40806838870048523, "learning_rate": 0.00017625628072702562, "loss": 1.1559, "step": 4580 }, { "epoch": 0.45599612552311647, "grad_norm": 0.4387802183628082, "learning_rate": 0.00017578583572512877, "loss": 1.1799, "step": 4590 }, { "epoch": 0.45698958113427796, "grad_norm": 0.44947245717048645, "learning_rate": 0.00017531512921436997, "loss": 1.1874, "step": 4600 }, { "epoch": 0.45798303674543944, "grad_norm": 0.5053061842918396, "learning_rate": 0.00017484416596845226, "loss": 1.1804, "step": 4610 }, { "epoch": 0.4589764923566009, "grad_norm": 0.4740687310695648, "learning_rate": 0.00017437295076368252, "loss": 1.1537, "step": 4620 }, { "epoch": 0.45996994796776236, "grad_norm": 0.4325508177280426, "learning_rate": 0.0001739014883789228, "loss": 1.1573, "step": 4630 }, { "epoch": 0.46096340357892385, "grad_norm": 0.478100061416626, "learning_rate": 0.00017342978359554187, "loss": 1.1819, "step": 4640 }, { "epoch": 0.46195685919008533, "grad_norm": 0.4926578104496002, "learning_rate": 0.00017295784119736698, "loss": 1.1886, "step": 4650 }, { "epoch": 0.46295031480124677, "grad_norm": 0.5400477051734924, "learning_rate": 0.000172485665970635, "loss": 1.1698, "step": 4660 }, { "epoch": 0.46394377041240825, "grad_norm": 0.4423806965351105, "learning_rate": 0.00017201326270394418, "loss": 1.1771, "step": 4670 }, { "epoch": 0.46493722602356974, "grad_norm": 0.4588819742202759, "learning_rate": 0.00017154063618820533, "loss": 1.1817, "step": 4680 }, { "epoch": 0.4659306816347312, "grad_norm": 0.5755475759506226, "learning_rate": 0.00017106779121659344, "loss": 1.1763, "step": 4690 }, { "epoch": 0.46692413724589266, "grad_norm": 0.45726439356803894, "learning_rate": 0.00017059473258449894, "loss": 1.1706, "step": 4700 }, { "epoch": 0.46791759285705414, "grad_norm": 0.4874899089336395, "learning_rate": 0.00017012146508947916, "loss": 1.1842, "step": 4710 }, { "epoch": 0.46891104846821563, "grad_norm": 0.4870886206626892, "learning_rate": 0.00016964799353120963, "loss": 1.1556, "step": 4720 }, { "epoch": 0.4699045040793771, "grad_norm": 0.4287513494491577, "learning_rate": 0.00016917432271143534, "loss": 1.1887, "step": 4730 }, { "epoch": 0.4708979596905386, "grad_norm": 0.47464653849601746, "learning_rate": 0.00016870045743392207, "loss": 1.1768, "step": 4740 }, { "epoch": 0.47189141530170003, "grad_norm": 0.48040926456451416, "learning_rate": 0.0001682264025044079, "loss": 1.1398, "step": 4750 }, { "epoch": 0.4728848709128615, "grad_norm": 0.4692235589027405, "learning_rate": 0.00016775216273055406, "loss": 1.1681, "step": 4760 }, { "epoch": 0.473878326524023, "grad_norm": 0.4798252582550049, "learning_rate": 0.00016727774292189657, "loss": 1.1564, "step": 4770 }, { "epoch": 0.4748717821351845, "grad_norm": 0.4889027774333954, "learning_rate": 0.0001668031478897971, "loss": 1.1723, "step": 4780 }, { "epoch": 0.4758652377463459, "grad_norm": 0.498568058013916, "learning_rate": 0.00016632838244739454, "loss": 1.1648, "step": 4790 }, { "epoch": 0.4768586933575074, "grad_norm": 0.46477189660072327, "learning_rate": 0.00016585345140955593, "loss": 1.1548, "step": 4800 }, { "epoch": 0.4778521489686689, "grad_norm": 0.46384197473526, "learning_rate": 0.00016537835959282767, "loss": 1.1611, "step": 4810 }, { "epoch": 0.4788456045798304, "grad_norm": 0.4606354832649231, "learning_rate": 0.00016490311181538672, "loss": 1.1592, "step": 4820 }, { "epoch": 0.4798390601909918, "grad_norm": 0.46360450983047485, "learning_rate": 0.00016442771289699185, "loss": 1.1653, "step": 4830 }, { "epoch": 0.4808325158021533, "grad_norm": 0.47855210304260254, "learning_rate": 0.00016395216765893443, "loss": 1.1784, "step": 4840 }, { "epoch": 0.4818259714133148, "grad_norm": 0.4936244487762451, "learning_rate": 0.00016347648092398997, "loss": 1.1697, "step": 4850 }, { "epoch": 0.4828194270244763, "grad_norm": 0.5215084552764893, "learning_rate": 0.00016300065751636878, "loss": 1.1623, "step": 4860 }, { "epoch": 0.4838128826356377, "grad_norm": 0.49761301279067993, "learning_rate": 0.00016252470226166735, "loss": 1.1845, "step": 4870 }, { "epoch": 0.4848063382467992, "grad_norm": 0.47753623127937317, "learning_rate": 0.0001620486199868193, "loss": 1.1761, "step": 4880 }, { "epoch": 0.4857997938579607, "grad_norm": 0.4606369137763977, "learning_rate": 0.0001615724155200464, "loss": 1.1824, "step": 4890 }, { "epoch": 0.48679324946912217, "grad_norm": 0.4668729305267334, "learning_rate": 0.00016109609369080964, "loss": 1.1553, "step": 4900 }, { "epoch": 0.48778670508028366, "grad_norm": 0.47536182403564453, "learning_rate": 0.00016061965932976033, "loss": 1.1663, "step": 4910 }, { "epoch": 0.4887801606914451, "grad_norm": 0.4314727783203125, "learning_rate": 0.00016014311726869092, "loss": 1.1698, "step": 4920 }, { "epoch": 0.4897736163026066, "grad_norm": 0.44343844056129456, "learning_rate": 0.00015966647234048618, "loss": 1.1678, "step": 4930 }, { "epoch": 0.49076707191376806, "grad_norm": 0.45859286189079285, "learning_rate": 0.000159189729379074, "loss": 1.1873, "step": 4940 }, { "epoch": 0.49176052752492955, "grad_norm": 0.43602892756462097, "learning_rate": 0.0001587128932193767, "loss": 1.157, "step": 4950 }, { "epoch": 0.492753983136091, "grad_norm": 0.48735079169273376, "learning_rate": 0.00015823596869726152, "loss": 1.1682, "step": 4960 }, { "epoch": 0.49374743874725247, "grad_norm": 0.4638786017894745, "learning_rate": 0.0001577589606494921, "loss": 1.166, "step": 4970 }, { "epoch": 0.49474089435841395, "grad_norm": 0.44135597348213196, "learning_rate": 0.0001572818739136789, "loss": 1.1742, "step": 4980 }, { "epoch": 0.49573434996957544, "grad_norm": 0.4518062174320221, "learning_rate": 0.00015680471332823055, "loss": 1.1699, "step": 4990 }, { "epoch": 0.49672780558073687, "grad_norm": 0.4678098261356354, "learning_rate": 0.00015632748373230462, "loss": 1.1634, "step": 5000 }, { "epoch": 0.49772126119189836, "grad_norm": 0.4754001498222351, "learning_rate": 0.00015585018996575853, "loss": 1.1536, "step": 5010 }, { "epoch": 0.49871471680305984, "grad_norm": 0.5086048245429993, "learning_rate": 0.00015537283686910047, "loss": 1.1584, "step": 5020 }, { "epoch": 0.49970817241422133, "grad_norm": 0.4387681484222412, "learning_rate": 0.00015489542928344037, "loss": 1.1691, "step": 5030 }, { "epoch": 0.500204900219802, "eval_loss": 1.1872472763061523, "eval_runtime": 217.2564, "eval_samples_per_second": 230.143, "eval_steps_per_second": 28.768, "step": 5035 }, { "epoch": 0.5007016280253828, "grad_norm": 0.5286032557487488, "learning_rate": 0.00015441797205044083, "loss": 1.1789, "step": 5040 }, { "epoch": 0.5016950836365442, "grad_norm": 0.4659021198749542, "learning_rate": 0.00015394047001226775, "loss": 1.1713, "step": 5050 }, { "epoch": 0.5026885392477057, "grad_norm": 0.442771852016449, "learning_rate": 0.0001534629280115416, "loss": 1.1679, "step": 5060 }, { "epoch": 0.5036819948588672, "grad_norm": 0.5038478374481201, "learning_rate": 0.00015298535089128813, "loss": 1.1657, "step": 5070 }, { "epoch": 0.5046754504700287, "grad_norm": 0.4490121304988861, "learning_rate": 0.00015250774349488916, "loss": 1.1552, "step": 5080 }, { "epoch": 0.5056689060811902, "grad_norm": 0.41231900453567505, "learning_rate": 0.00015203011066603357, "loss": 1.1688, "step": 5090 }, { "epoch": 0.5066623616923517, "grad_norm": 0.42963293194770813, "learning_rate": 0.0001515524572486683, "loss": 1.1657, "step": 5100 }, { "epoch": 0.5076558173035131, "grad_norm": 0.4359903931617737, "learning_rate": 0.00015107478808694887, "loss": 1.1612, "step": 5110 }, { "epoch": 0.5086492729146745, "grad_norm": 0.48204314708709717, "learning_rate": 0.0001505971080251907, "loss": 1.1583, "step": 5120 }, { "epoch": 0.509642728525836, "grad_norm": 0.4897457957267761, "learning_rate": 0.00015011942190781954, "loss": 1.1604, "step": 5130 }, { "epoch": 0.5106361841369975, "grad_norm": 0.5152519345283508, "learning_rate": 0.0001496417345793228, "loss": 1.16, "step": 5140 }, { "epoch": 0.511629639748159, "grad_norm": 0.48295149207115173, "learning_rate": 0.00014916405088419998, "loss": 1.1748, "step": 5150 }, { "epoch": 0.5126230953593205, "grad_norm": 0.46634209156036377, "learning_rate": 0.0001486863756669137, "loss": 1.1541, "step": 5160 }, { "epoch": 0.513616550970482, "grad_norm": 0.4615635871887207, "learning_rate": 0.00014820871377184075, "loss": 1.1675, "step": 5170 }, { "epoch": 0.5146100065816435, "grad_norm": 0.4817815124988556, "learning_rate": 0.00014773107004322278, "loss": 1.1546, "step": 5180 }, { "epoch": 0.515603462192805, "grad_norm": 0.4719342887401581, "learning_rate": 0.0001472534493251171, "loss": 1.156, "step": 5190 }, { "epoch": 0.5165969178039663, "grad_norm": 0.4903907775878906, "learning_rate": 0.00014677585646134778, "loss": 1.1622, "step": 5200 }, { "epoch": 0.5175903734151278, "grad_norm": 0.4558243155479431, "learning_rate": 0.00014629829629545632, "loss": 1.1388, "step": 5210 }, { "epoch": 0.5185838290262893, "grad_norm": 0.48934248089790344, "learning_rate": 0.00014582077367065278, "loss": 1.1584, "step": 5220 }, { "epoch": 0.5195772846374508, "grad_norm": 0.4721735417842865, "learning_rate": 0.00014534329342976612, "loss": 1.1593, "step": 5230 }, { "epoch": 0.5205707402486123, "grad_norm": 0.45554596185684204, "learning_rate": 0.00014486586041519587, "loss": 1.1659, "step": 5240 }, { "epoch": 0.5215641958597738, "grad_norm": 0.4855104684829712, "learning_rate": 0.00014438847946886236, "loss": 1.1833, "step": 5250 }, { "epoch": 0.5225576514709352, "grad_norm": 0.45857903361320496, "learning_rate": 0.00014391115543215801, "loss": 1.1389, "step": 5260 }, { "epoch": 0.5235511070820967, "grad_norm": 0.4469977021217346, "learning_rate": 0.00014343389314589797, "loss": 1.1607, "step": 5270 }, { "epoch": 0.5245445626932581, "grad_norm": 0.5102470517158508, "learning_rate": 0.00014295669745027125, "loss": 1.1543, "step": 5280 }, { "epoch": 0.5255380183044196, "grad_norm": 0.44836586713790894, "learning_rate": 0.00014247957318479148, "loss": 1.1553, "step": 5290 }, { "epoch": 0.5265314739155811, "grad_norm": 0.47462692856788635, "learning_rate": 0.0001420025251882478, "loss": 1.1671, "step": 5300 }, { "epoch": 0.5275249295267426, "grad_norm": 0.4809962511062622, "learning_rate": 0.00014152555829865593, "loss": 1.1536, "step": 5310 }, { "epoch": 0.5285183851379041, "grad_norm": 0.49753010272979736, "learning_rate": 0.00014104867735320912, "loss": 1.1645, "step": 5320 }, { "epoch": 0.5295118407490655, "grad_norm": 0.4475755989551544, "learning_rate": 0.0001405718871882289, "loss": 1.1403, "step": 5330 }, { "epoch": 0.530505296360227, "grad_norm": 0.45260560512542725, "learning_rate": 0.00014009519263911616, "loss": 1.1516, "step": 5340 }, { "epoch": 0.5314987519713885, "grad_norm": 0.4384668469429016, "learning_rate": 0.00013961859854030208, "loss": 1.1622, "step": 5350 }, { "epoch": 0.53249220758255, "grad_norm": 0.45300155878067017, "learning_rate": 0.00013914210972519928, "loss": 1.167, "step": 5360 }, { "epoch": 0.5334856631937114, "grad_norm": 0.42206814885139465, "learning_rate": 0.00013866573102615237, "loss": 1.1656, "step": 5370 }, { "epoch": 0.5344791188048729, "grad_norm": 0.428107887506485, "learning_rate": 0.0001381894672743894, "loss": 1.1582, "step": 5380 }, { "epoch": 0.5354725744160344, "grad_norm": 0.452441930770874, "learning_rate": 0.0001377133232999726, "loss": 1.1541, "step": 5390 }, { "epoch": 0.5364660300271958, "grad_norm": 0.4755808711051941, "learning_rate": 0.0001372373039317495, "loss": 1.1601, "step": 5400 }, { "epoch": 0.5374594856383573, "grad_norm": 0.4153245687484741, "learning_rate": 0.0001367614139973039, "loss": 1.1545, "step": 5410 }, { "epoch": 0.5384529412495188, "grad_norm": 0.545102059841156, "learning_rate": 0.000136285658322907, "loss": 1.162, "step": 5420 }, { "epoch": 0.5394463968606803, "grad_norm": 0.41430729627609253, "learning_rate": 0.00013581004173346834, "loss": 1.1573, "step": 5430 }, { "epoch": 0.5404398524718418, "grad_norm": 0.43141236901283264, "learning_rate": 0.00013533456905248683, "loss": 1.148, "step": 5440 }, { "epoch": 0.5414333080830032, "grad_norm": 0.5155332088470459, "learning_rate": 0.00013485924510200215, "loss": 1.1574, "step": 5450 }, { "epoch": 0.5424267636941646, "grad_norm": 0.45795363187789917, "learning_rate": 0.00013438407470254543, "loss": 1.1485, "step": 5460 }, { "epoch": 0.5434202193053261, "grad_norm": 0.42748311161994934, "learning_rate": 0.0001339090626730906, "loss": 1.1608, "step": 5470 }, { "epoch": 0.5444136749164876, "grad_norm": 0.47087302803993225, "learning_rate": 0.00013343421383100554, "loss": 1.1383, "step": 5480 }, { "epoch": 0.5454071305276491, "grad_norm": 0.44623327255249023, "learning_rate": 0.000132959532992003, "loss": 1.1434, "step": 5490 }, { "epoch": 0.5464005861388106, "grad_norm": 0.4788418114185333, "learning_rate": 0.0001324850249700921, "loss": 1.162, "step": 5500 }, { "epoch": 0.5473940417499721, "grad_norm": 0.4729274809360504, "learning_rate": 0.0001320106945775292, "loss": 1.1502, "step": 5510 }, { "epoch": 0.5483874973611336, "grad_norm": 0.46386897563934326, "learning_rate": 0.0001315365466247692, "loss": 1.1205, "step": 5520 }, { "epoch": 0.5493809529722951, "grad_norm": 0.481486976146698, "learning_rate": 0.0001310625859204169, "loss": 1.1629, "step": 5530 }, { "epoch": 0.5503744085834564, "grad_norm": 0.47394269704818726, "learning_rate": 0.00013058881727117798, "loss": 1.1538, "step": 5540 }, { "epoch": 0.5513678641946179, "grad_norm": 0.49581393599510193, "learning_rate": 0.00013011524548181048, "loss": 1.1592, "step": 5550 }, { "epoch": 0.5523613198057794, "grad_norm": 0.4401605725288391, "learning_rate": 0.00012964187535507594, "loss": 1.1313, "step": 5560 }, { "epoch": 0.5533547754169409, "grad_norm": 0.4444689154624939, "learning_rate": 0.00012916871169169061, "loss": 1.1543, "step": 5570 }, { "epoch": 0.5543482310281024, "grad_norm": 0.45363694429397583, "learning_rate": 0.00012869575929027704, "loss": 1.1467, "step": 5580 }, { "epoch": 0.5553416866392639, "grad_norm": 0.444588303565979, "learning_rate": 0.0001282230229473152, "loss": 1.1617, "step": 5590 }, { "epoch": 0.5563351422504254, "grad_norm": 0.43078240752220154, "learning_rate": 0.00012775050745709383, "loss": 1.1503, "step": 5600 }, { "epoch": 0.5573285978615868, "grad_norm": 0.48163166642189026, "learning_rate": 0.00012727821761166196, "loss": 1.1515, "step": 5610 }, { "epoch": 0.5583220534727483, "grad_norm": 0.5182563662528992, "learning_rate": 0.00012680615820078017, "loss": 1.1501, "step": 5620 }, { "epoch": 0.5593155090839097, "grad_norm": 0.4423312842845917, "learning_rate": 0.0001263343340118722, "loss": 1.1455, "step": 5630 }, { "epoch": 0.5603089646950712, "grad_norm": 0.46226614713668823, "learning_rate": 0.00012586274982997596, "loss": 1.1618, "step": 5640 }, { "epoch": 0.5613024203062327, "grad_norm": 0.4315178096294403, "learning_rate": 0.00012539141043769566, "loss": 1.1513, "step": 5650 }, { "epoch": 0.5622958759173942, "grad_norm": 0.4481825530529022, "learning_rate": 0.00012492032061515274, "loss": 1.1332, "step": 5660 }, { "epoch": 0.5632893315285556, "grad_norm": 0.47609102725982666, "learning_rate": 0.00012444948513993774, "loss": 1.1482, "step": 5670 }, { "epoch": 0.5642827871397171, "grad_norm": 0.45403629541397095, "learning_rate": 0.00012397890878706162, "loss": 1.1515, "step": 5680 }, { "epoch": 0.5652762427508786, "grad_norm": 0.49668949842453003, "learning_rate": 0.00012350859632890752, "loss": 1.1629, "step": 5690 }, { "epoch": 0.5662696983620401, "grad_norm": 0.4693937301635742, "learning_rate": 0.00012303855253518222, "loss": 1.1386, "step": 5700 }, { "epoch": 0.5672631539732015, "grad_norm": 0.5180835127830505, "learning_rate": 0.00012256878217286774, "loss": 1.1454, "step": 5710 }, { "epoch": 0.568256609584363, "grad_norm": 0.4567181169986725, "learning_rate": 0.0001220992900061733, "loss": 1.1535, "step": 5720 }, { "epoch": 0.5692500651955245, "grad_norm": 0.45975708961486816, "learning_rate": 0.00012163008079648655, "loss": 1.158, "step": 5730 }, { "epoch": 0.5702435208066859, "grad_norm": 0.4463796615600586, "learning_rate": 0.0001211611593023256, "loss": 1.1439, "step": 5740 }, { "epoch": 0.5712369764178474, "grad_norm": 0.46401357650756836, "learning_rate": 0.00012069253027929076, "loss": 1.1325, "step": 5750 }, { "epoch": 0.5722304320290089, "grad_norm": 0.4709807336330414, "learning_rate": 0.00012022419848001607, "loss": 1.1458, "step": 5760 }, { "epoch": 0.5732238876401704, "grad_norm": 0.43508750200271606, "learning_rate": 0.0001197561686541214, "loss": 1.1312, "step": 5770 }, { "epoch": 0.5742173432513319, "grad_norm": 0.47728458046913147, "learning_rate": 0.00011928844554816401, "loss": 1.1529, "step": 5780 }, { "epoch": 0.5752107988624934, "grad_norm": 0.5009467601776123, "learning_rate": 0.00011882103390559055, "loss": 1.144, "step": 5790 }, { "epoch": 0.5762042544736548, "grad_norm": 0.41657334566116333, "learning_rate": 0.00011835393846668907, "loss": 1.1423, "step": 5800 }, { "epoch": 0.5771977100848162, "grad_norm": 0.4635260999202728, "learning_rate": 0.00011788716396854067, "loss": 1.162, "step": 5810 }, { "epoch": 0.5781911656959777, "grad_norm": 0.4591808319091797, "learning_rate": 0.00011742071514497172, "loss": 1.1526, "step": 5820 }, { "epoch": 0.5791846213071392, "grad_norm": 0.4596636891365051, "learning_rate": 0.00011695459672650566, "loss": 1.16, "step": 5830 }, { "epoch": 0.5801780769183007, "grad_norm": 0.49629148840904236, "learning_rate": 0.00011648881344031516, "loss": 1.1517, "step": 5840 }, { "epoch": 0.5811715325294622, "grad_norm": 0.5193811655044556, "learning_rate": 0.00011602337001017399, "loss": 1.1521, "step": 5850 }, { "epoch": 0.5821649881406237, "grad_norm": 0.5228062272071838, "learning_rate": 0.00011555827115640947, "loss": 1.1403, "step": 5860 }, { "epoch": 0.5831584437517852, "grad_norm": 0.47855010628700256, "learning_rate": 0.00011509352159585417, "loss": 1.1249, "step": 5870 }, { "epoch": 0.5841518993629465, "grad_norm": 0.4856327474117279, "learning_rate": 0.00011462912604179843, "loss": 1.1467, "step": 5880 }, { "epoch": 0.585145354974108, "grad_norm": 0.46515578031539917, "learning_rate": 0.00011416508920394228, "loss": 1.1362, "step": 5890 }, { "epoch": 0.5861388105852695, "grad_norm": 0.43773916363716125, "learning_rate": 0.00011370141578834791, "loss": 1.1276, "step": 5900 }, { "epoch": 0.587132266196431, "grad_norm": 0.4612313210964203, "learning_rate": 0.00011323811049739184, "loss": 1.1434, "step": 5910 }, { "epoch": 0.5881257218075925, "grad_norm": 0.4420901834964752, "learning_rate": 0.00011277517802971704, "loss": 1.1363, "step": 5920 }, { "epoch": 0.589119177418754, "grad_norm": 0.47640514373779297, "learning_rate": 0.00011231262308018568, "loss": 1.1415, "step": 5930 }, { "epoch": 0.5901126330299155, "grad_norm": 0.5014554262161255, "learning_rate": 0.00011185045033983126, "loss": 1.1485, "step": 5940 }, { "epoch": 0.5911060886410769, "grad_norm": 0.4807797968387604, "learning_rate": 0.00011138866449581098, "loss": 1.1464, "step": 5950 }, { "epoch": 0.5920995442522384, "grad_norm": 0.4298631250858307, "learning_rate": 0.00011092727023135847, "loss": 1.141, "step": 5960 }, { "epoch": 0.5930929998633998, "grad_norm": 0.4730483889579773, "learning_rate": 0.00011046627222573598, "loss": 1.135, "step": 5970 }, { "epoch": 0.5940864554745613, "grad_norm": 0.5459967851638794, "learning_rate": 0.00011000567515418709, "loss": 1.1385, "step": 5980 }, { "epoch": 0.5950799110857228, "grad_norm": 0.4976285696029663, "learning_rate": 0.00010954548368788932, "loss": 1.1202, "step": 5990 }, { "epoch": 0.5960733666968843, "grad_norm": 0.46126067638397217, "learning_rate": 0.00010908570249390679, "loss": 1.1218, "step": 6000 }, { "epoch": 0.5970668223080458, "grad_norm": 0.44945207238197327, "learning_rate": 0.0001086263362351426, "loss": 1.1402, "step": 6010 }, { "epoch": 0.5980602779192072, "grad_norm": 0.4707368314266205, "learning_rate": 0.00010816738957029202, "loss": 1.1474, "step": 6020 }, { "epoch": 0.5990537335303687, "grad_norm": 0.45081543922424316, "learning_rate": 0.00010770886715379475, "loss": 1.1361, "step": 6030 }, { "epoch": 0.6000471891415302, "grad_norm": 0.4335690140724182, "learning_rate": 0.00010725077363578817, "loss": 1.1266, "step": 6040 }, { "epoch": 0.6002458802637625, "eval_loss": 1.1717205047607422, "eval_runtime": 217.1349, "eval_samples_per_second": 230.272, "eval_steps_per_second": 28.784, "step": 6042 }, { "epoch": 0.6010406447526916, "grad_norm": 0.4617256224155426, "learning_rate": 0.00010679311366205971, "loss": 1.1539, "step": 6050 }, { "epoch": 0.6020341003638531, "grad_norm": 0.4457249045372009, "learning_rate": 0.00010633589187400009, "loss": 1.1466, "step": 6060 }, { "epoch": 0.6030275559750146, "grad_norm": 0.5100376009941101, "learning_rate": 0.00010587911290855625, "loss": 1.1374, "step": 6070 }, { "epoch": 0.604021011586176, "grad_norm": 0.5062493681907654, "learning_rate": 0.00010542278139818404, "loss": 1.1294, "step": 6080 }, { "epoch": 0.6050144671973375, "grad_norm": 0.46003487706184387, "learning_rate": 0.00010496690197080147, "loss": 1.1456, "step": 6090 }, { "epoch": 0.606007922808499, "grad_norm": 0.46181225776672363, "learning_rate": 0.00010451147924974182, "loss": 1.1337, "step": 6100 }, { "epoch": 0.6070013784196605, "grad_norm": 0.441353440284729, "learning_rate": 0.00010405651785370654, "loss": 1.1297, "step": 6110 }, { "epoch": 0.607994834030822, "grad_norm": 0.4799842834472656, "learning_rate": 0.00010360202239671843, "loss": 1.1326, "step": 6120 }, { "epoch": 0.6089882896419835, "grad_norm": 0.4211976230144501, "learning_rate": 0.0001031479974880752, "loss": 1.1404, "step": 6130 }, { "epoch": 0.6099817452531449, "grad_norm": 0.4226428270339966, "learning_rate": 0.0001026944477323022, "loss": 1.1398, "step": 6140 }, { "epoch": 0.6109752008643063, "grad_norm": 0.47160401940345764, "learning_rate": 0.00010224137772910618, "loss": 1.1356, "step": 6150 }, { "epoch": 0.6119686564754678, "grad_norm": 0.46349361538887024, "learning_rate": 0.00010178879207332834, "loss": 1.1294, "step": 6160 }, { "epoch": 0.6129621120866293, "grad_norm": 0.4886729121208191, "learning_rate": 0.0001013366953548978, "loss": 1.142, "step": 6170 }, { "epoch": 0.6139555676977908, "grad_norm": 0.4758968949317932, "learning_rate": 0.00010088509215878528, "loss": 1.1339, "step": 6180 }, { "epoch": 0.6149490233089523, "grad_norm": 0.47533220052719116, "learning_rate": 0.00010043398706495613, "loss": 1.1345, "step": 6190 }, { "epoch": 0.6159424789201138, "grad_norm": 0.5099872946739197, "learning_rate": 9.998338464832434e-05, "loss": 1.1543, "step": 6200 }, { "epoch": 0.6169359345312753, "grad_norm": 0.4612841010093689, "learning_rate": 9.953328947870598e-05, "loss": 1.1223, "step": 6210 }, { "epoch": 0.6179293901424368, "grad_norm": 0.4441303610801697, "learning_rate": 9.908370612077272e-05, "loss": 1.1365, "step": 6220 }, { "epoch": 0.6189228457535981, "grad_norm": 0.4773942232131958, "learning_rate": 9.863463913400582e-05, "loss": 1.1127, "step": 6230 }, { "epoch": 0.6199163013647596, "grad_norm": 0.4970172643661499, "learning_rate": 9.818609307264953e-05, "loss": 1.1174, "step": 6240 }, { "epoch": 0.6209097569759211, "grad_norm": 0.4488612413406372, "learning_rate": 9.773807248566534e-05, "loss": 1.135, "step": 6250 }, { "epoch": 0.6219032125870826, "grad_norm": 0.4947340488433838, "learning_rate": 9.729058191668527e-05, "loss": 1.1314, "step": 6260 }, { "epoch": 0.6228966681982441, "grad_norm": 0.45969218015670776, "learning_rate": 9.684362590396643e-05, "loss": 1.1388, "step": 6270 }, { "epoch": 0.6238901238094056, "grad_norm": 0.4401233494281769, "learning_rate": 9.639720898034454e-05, "loss": 1.1507, "step": 6280 }, { "epoch": 0.624883579420567, "grad_norm": 0.4139381945133209, "learning_rate": 9.59513356731882e-05, "loss": 1.1186, "step": 6290 }, { "epoch": 0.6258770350317285, "grad_norm": 0.47768634557724, "learning_rate": 9.550601050435279e-05, "loss": 1.1327, "step": 6300 }, { "epoch": 0.6268704906428899, "grad_norm": 0.4693679213523865, "learning_rate": 9.506123799013476e-05, "loss": 1.1194, "step": 6310 }, { "epoch": 0.6278639462540514, "grad_norm": 0.47899654507637024, "learning_rate": 9.461702264122584e-05, "loss": 1.1265, "step": 6320 }, { "epoch": 0.6288574018652129, "grad_norm": 0.4823016822338104, "learning_rate": 9.417336896266701e-05, "loss": 1.1275, "step": 6330 }, { "epoch": 0.6298508574763744, "grad_norm": 0.4641190469264984, "learning_rate": 9.373028145380326e-05, "loss": 1.1441, "step": 6340 }, { "epoch": 0.6308443130875359, "grad_norm": 0.4628022313117981, "learning_rate": 9.328776460823764e-05, "loss": 1.1281, "step": 6350 }, { "epoch": 0.6318377686986973, "grad_norm": 0.4776950776576996, "learning_rate": 9.284582291378573e-05, "loss": 1.1374, "step": 6360 }, { "epoch": 0.6328312243098588, "grad_norm": 0.4577087163925171, "learning_rate": 9.240446085243033e-05, "loss": 1.1427, "step": 6370 }, { "epoch": 0.6338246799210203, "grad_norm": 0.44643932580947876, "learning_rate": 9.196368290027572e-05, "loss": 1.1274, "step": 6380 }, { "epoch": 0.6348181355321818, "grad_norm": 0.46955859661102295, "learning_rate": 9.152349352750232e-05, "loss": 1.1528, "step": 6390 }, { "epoch": 0.6358115911433432, "grad_norm": 0.4443439841270447, "learning_rate": 9.108389719832165e-05, "loss": 1.1308, "step": 6400 }, { "epoch": 0.6368050467545047, "grad_norm": 0.482073575258255, "learning_rate": 9.064489837093069e-05, "loss": 1.1355, "step": 6410 }, { "epoch": 0.6377985023656662, "grad_norm": 0.49913904070854187, "learning_rate": 9.020650149746679e-05, "loss": 1.1257, "step": 6420 }, { "epoch": 0.6387919579768276, "grad_norm": 0.468218594789505, "learning_rate": 8.976871102396269e-05, "loss": 1.1463, "step": 6430 }, { "epoch": 0.6397854135879891, "grad_norm": 0.45414066314697266, "learning_rate": 8.93315313903011e-05, "loss": 1.1282, "step": 6440 }, { "epoch": 0.6407788691991506, "grad_norm": 0.5159599781036377, "learning_rate": 8.889496703017006e-05, "loss": 1.1465, "step": 6450 }, { "epoch": 0.6417723248103121, "grad_norm": 0.4611871838569641, "learning_rate": 8.845902237101754e-05, "loss": 1.1271, "step": 6460 }, { "epoch": 0.6427657804214736, "grad_norm": 0.4411086440086365, "learning_rate": 8.80237018340069e-05, "loss": 1.1407, "step": 6470 }, { "epoch": 0.643759236032635, "grad_norm": 0.5023905634880066, "learning_rate": 8.758900983397195e-05, "loss": 1.1351, "step": 6480 }, { "epoch": 0.6447526916437964, "grad_norm": 0.47759824991226196, "learning_rate": 8.71549507793721e-05, "loss": 1.1267, "step": 6490 }, { "epoch": 0.6457461472549579, "grad_norm": 0.46164053678512573, "learning_rate": 8.672152907224765e-05, "loss": 1.136, "step": 6500 }, { "epoch": 0.6467396028661194, "grad_norm": 0.4555439352989197, "learning_rate": 8.628874910817528e-05, "loss": 1.131, "step": 6510 }, { "epoch": 0.6477330584772809, "grad_norm": 0.49815115332603455, "learning_rate": 8.585661527622343e-05, "loss": 1.1535, "step": 6520 }, { "epoch": 0.6487265140884424, "grad_norm": 0.4290812909603119, "learning_rate": 8.542513195890758e-05, "loss": 1.1386, "step": 6530 }, { "epoch": 0.6497199696996039, "grad_norm": 0.4509336054325104, "learning_rate": 8.499430353214605e-05, "loss": 1.147, "step": 6540 }, { "epoch": 0.6507134253107654, "grad_norm": 0.4503090977668762, "learning_rate": 8.456413436521558e-05, "loss": 1.1416, "step": 6550 }, { "epoch": 0.6517068809219269, "grad_norm": 0.5119098424911499, "learning_rate": 8.413462882070699e-05, "loss": 1.1187, "step": 6560 }, { "epoch": 0.6527003365330882, "grad_norm": 0.4846774637699127, "learning_rate": 8.370579125448091e-05, "loss": 1.1266, "step": 6570 }, { "epoch": 0.6536937921442497, "grad_norm": 0.46956613659858704, "learning_rate": 8.327762601562346e-05, "loss": 1.1507, "step": 6580 }, { "epoch": 0.6546872477554112, "grad_norm": 0.4863440990447998, "learning_rate": 8.285013744640266e-05, "loss": 1.1214, "step": 6590 }, { "epoch": 0.6556807033665727, "grad_norm": 0.45145437121391296, "learning_rate": 8.24233298822236e-05, "loss": 1.1216, "step": 6600 }, { "epoch": 0.6566741589777342, "grad_norm": 0.46321019530296326, "learning_rate": 8.199720765158533e-05, "loss": 1.1395, "step": 6610 }, { "epoch": 0.6576676145888957, "grad_norm": 0.4763803780078888, "learning_rate": 8.15717750760362e-05, "loss": 1.1231, "step": 6620 }, { "epoch": 0.6586610702000572, "grad_norm": 0.4704056680202484, "learning_rate": 8.114703647013056e-05, "loss": 1.1189, "step": 6630 }, { "epoch": 0.6596545258112186, "grad_norm": 0.42659151554107666, "learning_rate": 8.072299614138482e-05, "loss": 1.127, "step": 6640 }, { "epoch": 0.6606479814223801, "grad_norm": 0.4490351676940918, "learning_rate": 8.029965839023373e-05, "loss": 1.13, "step": 6650 }, { "epoch": 0.6616414370335415, "grad_norm": 0.4944393038749695, "learning_rate": 7.98770275099868e-05, "loss": 1.104, "step": 6660 }, { "epoch": 0.662634892644703, "grad_norm": 0.4969693720340729, "learning_rate": 7.94551077867846e-05, "loss": 1.1124, "step": 6670 }, { "epoch": 0.6636283482558645, "grad_norm": 0.6347571015357971, "learning_rate": 7.903390349955574e-05, "loss": 1.1362, "step": 6680 }, { "epoch": 0.664621803867026, "grad_norm": 0.48183801770210266, "learning_rate": 7.861341891997308e-05, "loss": 1.1368, "step": 6690 }, { "epoch": 0.6656152594781874, "grad_norm": 0.44424450397491455, "learning_rate": 7.819365831241036e-05, "loss": 1.1216, "step": 6700 }, { "epoch": 0.6666087150893489, "grad_norm": 0.5084063410758972, "learning_rate": 7.77746259338993e-05, "loss": 1.1184, "step": 6710 }, { "epoch": 0.6676021707005104, "grad_norm": 0.4749411940574646, "learning_rate": 7.73563260340862e-05, "loss": 1.1196, "step": 6720 }, { "epoch": 0.6685956263116719, "grad_norm": 0.4693940579891205, "learning_rate": 7.693876285518873e-05, "loss": 1.135, "step": 6730 }, { "epoch": 0.6695890819228333, "grad_norm": 0.46934443712234497, "learning_rate": 7.652194063195329e-05, "loss": 1.1349, "step": 6740 }, { "epoch": 0.6705825375339948, "grad_norm": 0.4727706015110016, "learning_rate": 7.61058635916116e-05, "loss": 1.1158, "step": 6750 }, { "epoch": 0.6715759931451563, "grad_norm": 0.45518192648887634, "learning_rate": 7.569053595383814e-05, "loss": 1.1143, "step": 6760 }, { "epoch": 0.6725694487563177, "grad_norm": 0.5114362835884094, "learning_rate": 7.527596193070742e-05, "loss": 1.1223, "step": 6770 }, { "epoch": 0.6735629043674792, "grad_norm": 0.4581073224544525, "learning_rate": 7.486214572665084e-05, "loss": 1.1209, "step": 6780 }, { "epoch": 0.6745563599786407, "grad_norm": 0.4839348793029785, "learning_rate": 7.444909153841455e-05, "loss": 1.1201, "step": 6790 }, { "epoch": 0.6755498155898022, "grad_norm": 0.4677003026008606, "learning_rate": 7.40368035550165e-05, "loss": 1.111, "step": 6800 }, { "epoch": 0.6765432712009637, "grad_norm": 0.47851434350013733, "learning_rate": 7.362528595770423e-05, "loss": 1.1365, "step": 6810 }, { "epoch": 0.6775367268121252, "grad_norm": 0.5189026594161987, "learning_rate": 7.321454291991239e-05, "loss": 1.1314, "step": 6820 }, { "epoch": 0.6785301824232866, "grad_norm": 0.6284201741218567, "learning_rate": 7.28045786072202e-05, "loss": 1.1249, "step": 6830 }, { "epoch": 0.679523638034448, "grad_norm": 0.43699899315834045, "learning_rate": 7.239539717730956e-05, "loss": 1.1302, "step": 6840 }, { "epoch": 0.6805170936456095, "grad_norm": 0.47337090969085693, "learning_rate": 7.198700277992268e-05, "loss": 1.1182, "step": 6850 }, { "epoch": 0.681510549256771, "grad_norm": 0.47022855281829834, "learning_rate": 7.157939955681996e-05, "loss": 1.1297, "step": 6860 }, { "epoch": 0.6825040048679325, "grad_norm": 0.4596823453903198, "learning_rate": 7.117259164173799e-05, "loss": 1.1275, "step": 6870 }, { "epoch": 0.683497460479094, "grad_norm": 0.44979989528656006, "learning_rate": 7.07665831603479e-05, "loss": 1.1286, "step": 6880 }, { "epoch": 0.6844909160902555, "grad_norm": 0.5144652128219604, "learning_rate": 7.036137823021311e-05, "loss": 1.1467, "step": 6890 }, { "epoch": 0.685484371701417, "grad_norm": 0.48372334241867065, "learning_rate": 6.995698096074797e-05, "loss": 1.1158, "step": 6900 }, { "epoch": 0.6864778273125783, "grad_norm": 0.47437506914138794, "learning_rate": 6.95533954531756e-05, "loss": 1.1125, "step": 6910 }, { "epoch": 0.6874712829237398, "grad_norm": 0.4924662709236145, "learning_rate": 6.91506258004868e-05, "loss": 1.1178, "step": 6920 }, { "epoch": 0.6884647385349013, "grad_norm": 0.46547582745552063, "learning_rate": 6.874867608739833e-05, "loss": 1.1199, "step": 6930 }, { "epoch": 0.6894581941460628, "grad_norm": 0.4472568929195404, "learning_rate": 6.834755039031133e-05, "loss": 1.1346, "step": 6940 }, { "epoch": 0.6904516497572243, "grad_norm": 0.45158204436302185, "learning_rate": 6.794725277727016e-05, "loss": 1.1231, "step": 6950 }, { "epoch": 0.6914451053683858, "grad_norm": 0.45651930570602417, "learning_rate": 6.75477873079212e-05, "loss": 1.1179, "step": 6960 }, { "epoch": 0.6924385609795473, "grad_norm": 0.43802908062934875, "learning_rate": 6.714915803347151e-05, "loss": 1.1132, "step": 6970 }, { "epoch": 0.6934320165907087, "grad_norm": 0.5048275589942932, "learning_rate": 6.675136899664792e-05, "loss": 1.1225, "step": 6980 }, { "epoch": 0.6944254722018702, "grad_norm": 0.4675825238227844, "learning_rate": 6.635442423165571e-05, "loss": 1.1294, "step": 6990 }, { "epoch": 0.6954189278130316, "grad_norm": 0.4738437235355377, "learning_rate": 6.595832776413823e-05, "loss": 1.1366, "step": 7000 }, { "epoch": 0.6964123834241931, "grad_norm": 0.44783490896224976, "learning_rate": 6.55630836111354e-05, "loss": 1.1166, "step": 7010 }, { "epoch": 0.6974058390353546, "grad_norm": 0.4776122570037842, "learning_rate": 6.516869578104371e-05, "loss": 1.1091, "step": 7020 }, { "epoch": 0.6983992946465161, "grad_norm": 0.4984320402145386, "learning_rate": 6.477516827357486e-05, "loss": 1.1122, "step": 7030 }, { "epoch": 0.6993927502576776, "grad_norm": 0.4734194874763489, "learning_rate": 6.438250507971574e-05, "loss": 1.1184, "step": 7040 }, { "epoch": 0.7002868603077229, "eval_loss": 1.158501148223877, "eval_runtime": 217.2336, "eval_samples_per_second": 230.167, "eval_steps_per_second": 28.771, "step": 7049 }, { "epoch": 0.700386205868839, "grad_norm": 0.44331488013267517, "learning_rate": 6.399071018168768e-05, "loss": 1.131, "step": 7050 }, { "epoch": 0.7013796614800005, "grad_norm": 0.4725908041000366, "learning_rate": 6.359978755290616e-05, "loss": 1.1043, "step": 7060 }, { "epoch": 0.702373117091162, "grad_norm": 0.49042266607284546, "learning_rate": 6.320974115794042e-05, "loss": 1.1078, "step": 7070 }, { "epoch": 0.7033665727023234, "grad_norm": 0.48375070095062256, "learning_rate": 6.282057495247326e-05, "loss": 1.1141, "step": 7080 }, { "epoch": 0.7043600283134849, "grad_norm": 0.5035314559936523, "learning_rate": 6.24322928832611e-05, "loss": 1.1205, "step": 7090 }, { "epoch": 0.7053534839246464, "grad_norm": 0.5112337470054626, "learning_rate": 6.204489888809386e-05, "loss": 1.1102, "step": 7100 }, { "epoch": 0.7063469395358078, "grad_norm": 0.4766629934310913, "learning_rate": 6.16583968957548e-05, "loss": 1.1269, "step": 7110 }, { "epoch": 0.7073403951469693, "grad_norm": 0.4522871673107147, "learning_rate": 6.127279082598104e-05, "loss": 1.1266, "step": 7120 }, { "epoch": 0.7083338507581308, "grad_norm": 0.454632043838501, "learning_rate": 6.088808458942365e-05, "loss": 1.1302, "step": 7130 }, { "epoch": 0.7093273063692923, "grad_norm": 0.4837658703327179, "learning_rate": 6.050428208760781e-05, "loss": 1.1296, "step": 7140 }, { "epoch": 0.7103207619804538, "grad_norm": 0.49942177534103394, "learning_rate": 6.012138721289361e-05, "loss": 1.1091, "step": 7150 }, { "epoch": 0.7113142175916153, "grad_norm": 0.5515566468238831, "learning_rate": 5.97394038484362e-05, "loss": 1.1259, "step": 7160 }, { "epoch": 0.7123076732027767, "grad_norm": 0.4293889105319977, "learning_rate": 5.93583358681467e-05, "loss": 1.107, "step": 7170 }, { "epoch": 0.7133011288139381, "grad_norm": 0.41633716225624084, "learning_rate": 5.897818713665282e-05, "loss": 1.0986, "step": 7180 }, { "epoch": 0.7142945844250996, "grad_norm": 0.5082070231437683, "learning_rate": 5.8598961509259504e-05, "loss": 1.119, "step": 7190 }, { "epoch": 0.7152880400362611, "grad_norm": 0.47421905398368835, "learning_rate": 5.8220662831910135e-05, "loss": 1.1287, "step": 7200 }, { "epoch": 0.7162814956474226, "grad_norm": 0.4765404462814331, "learning_rate": 5.784329494114721e-05, "loss": 1.1085, "step": 7210 }, { "epoch": 0.7172749512585841, "grad_norm": 0.46578580141067505, "learning_rate": 5.746686166407369e-05, "loss": 1.1377, "step": 7220 }, { "epoch": 0.7182684068697456, "grad_norm": 0.47334933280944824, "learning_rate": 5.709136681831413e-05, "loss": 1.1177, "step": 7230 }, { "epoch": 0.7192618624809071, "grad_norm": 0.49011778831481934, "learning_rate": 5.671681421197576e-05, "loss": 1.1225, "step": 7240 }, { "epoch": 0.7202553180920686, "grad_norm": 0.43727976083755493, "learning_rate": 5.634320764361013e-05, "loss": 1.1127, "step": 7250 }, { "epoch": 0.7212487737032299, "grad_norm": 0.4456111490726471, "learning_rate": 5.5970550902174486e-05, "loss": 1.1053, "step": 7260 }, { "epoch": 0.7222422293143914, "grad_norm": 0.45397263765335083, "learning_rate": 5.559884776699341e-05, "loss": 1.1067, "step": 7270 }, { "epoch": 0.7232356849255529, "grad_norm": 0.4732288122177124, "learning_rate": 5.522810200772013e-05, "loss": 1.1174, "step": 7280 }, { "epoch": 0.7242291405367144, "grad_norm": 0.4691105782985687, "learning_rate": 5.485831738429888e-05, "loss": 1.115, "step": 7290 }, { "epoch": 0.7252225961478759, "grad_norm": 0.42165741324424744, "learning_rate": 5.448949764692632e-05, "loss": 1.1222, "step": 7300 }, { "epoch": 0.7262160517590374, "grad_norm": 0.44083935022354126, "learning_rate": 5.412164653601378e-05, "loss": 1.129, "step": 7310 }, { "epoch": 0.7272095073701988, "grad_norm": 0.46778151392936707, "learning_rate": 5.375476778214899e-05, "loss": 1.1169, "step": 7320 }, { "epoch": 0.7282029629813603, "grad_norm": 0.4633641541004181, "learning_rate": 5.338886510605862e-05, "loss": 1.1047, "step": 7330 }, { "epoch": 0.7291964185925217, "grad_norm": 0.4749923050403595, "learning_rate": 5.30239422185704e-05, "loss": 1.1272, "step": 7340 }, { "epoch": 0.7301898742036832, "grad_norm": 0.47855105996131897, "learning_rate": 5.266000282057532e-05, "loss": 1.1113, "step": 7350 }, { "epoch": 0.7311833298148447, "grad_norm": 0.4963724911212921, "learning_rate": 5.229705060299026e-05, "loss": 1.1158, "step": 7360 }, { "epoch": 0.7321767854260062, "grad_norm": 0.4988853633403778, "learning_rate": 5.193508924672067e-05, "loss": 1.0987, "step": 7370 }, { "epoch": 0.7331702410371677, "grad_norm": 0.4893016219139099, "learning_rate": 5.157412242262303e-05, "loss": 1.1102, "step": 7380 }, { "epoch": 0.7341636966483291, "grad_norm": 0.4634610712528229, "learning_rate": 5.1214153791467755e-05, "loss": 1.1077, "step": 7390 }, { "epoch": 0.7351571522594906, "grad_norm": 0.4903322458267212, "learning_rate": 5.085518700390187e-05, "loss": 1.1157, "step": 7400 }, { "epoch": 0.7361506078706521, "grad_norm": 0.48547953367233276, "learning_rate": 5.0497225700412334e-05, "loss": 1.1186, "step": 7410 }, { "epoch": 0.7371440634818136, "grad_norm": 0.4877443313598633, "learning_rate": 5.014027351128872e-05, "loss": 1.1081, "step": 7420 }, { "epoch": 0.738137519092975, "grad_norm": 0.4813385307788849, "learning_rate": 4.978433405658679e-05, "loss": 1.1058, "step": 7430 }, { "epoch": 0.7391309747041365, "grad_norm": 0.48355165123939514, "learning_rate": 4.9429410946091416e-05, "loss": 1.111, "step": 7440 }, { "epoch": 0.740124430315298, "grad_norm": 0.4709435701370239, "learning_rate": 4.9075507779280246e-05, "loss": 1.1368, "step": 7450 }, { "epoch": 0.7411178859264594, "grad_norm": 0.4199182093143463, "learning_rate": 4.872262814528707e-05, "loss": 1.1167, "step": 7460 }, { "epoch": 0.7421113415376209, "grad_norm": 0.4348940849304199, "learning_rate": 4.8370775622865535e-05, "loss": 1.1209, "step": 7470 }, { "epoch": 0.7431047971487824, "grad_norm": 0.48564040660858154, "learning_rate": 4.801995378035259e-05, "loss": 1.1043, "step": 7480 }, { "epoch": 0.7440982527599439, "grad_norm": 0.45493337512016296, "learning_rate": 4.767016617563254e-05, "loss": 1.1222, "step": 7490 }, { "epoch": 0.7450917083711054, "grad_norm": 0.4598395824432373, "learning_rate": 4.732141635610098e-05, "loss": 1.1182, "step": 7500 }, { "epoch": 0.7460851639822668, "grad_norm": 0.49085426330566406, "learning_rate": 4.697370785862863e-05, "loss": 1.1018, "step": 7510 }, { "epoch": 0.7470786195934283, "grad_norm": 0.453783243894577, "learning_rate": 4.6627044209525674e-05, "loss": 1.1188, "step": 7520 }, { "epoch": 0.7480720752045897, "grad_norm": 0.4396882653236389, "learning_rate": 4.628142892450575e-05, "loss": 1.1036, "step": 7530 }, { "epoch": 0.7490655308157512, "grad_norm": 0.4846864342689514, "learning_rate": 4.593686550865059e-05, "loss": 1.1215, "step": 7540 }, { "epoch": 0.7500589864269127, "grad_norm": 0.5176936388015747, "learning_rate": 4.5593357456374175e-05, "loss": 1.1189, "step": 7550 }, { "epoch": 0.7510524420380742, "grad_norm": 0.444042444229126, "learning_rate": 4.5250908251387594e-05, "loss": 1.1057, "step": 7560 }, { "epoch": 0.7520458976492357, "grad_norm": 0.4722656011581421, "learning_rate": 4.490952136666338e-05, "loss": 1.1209, "step": 7570 }, { "epoch": 0.7530393532603972, "grad_norm": 0.5398408770561218, "learning_rate": 4.456920026440064e-05, "loss": 1.1152, "step": 7580 }, { "epoch": 0.7540328088715587, "grad_norm": 0.4860037565231323, "learning_rate": 4.422994839598969e-05, "loss": 1.1058, "step": 7590 }, { "epoch": 0.75502626448272, "grad_norm": 0.4634205102920532, "learning_rate": 4.389176920197723e-05, "loss": 1.0958, "step": 7600 }, { "epoch": 0.7560197200938815, "grad_norm": 0.48629575967788696, "learning_rate": 4.355466611203119e-05, "loss": 1.0981, "step": 7610 }, { "epoch": 0.757013175705043, "grad_norm": 0.4982178509235382, "learning_rate": 4.32186425449062e-05, "loss": 1.1066, "step": 7620 }, { "epoch": 0.7580066313162045, "grad_norm": 0.4328030049800873, "learning_rate": 4.288370190840887e-05, "loss": 1.1062, "step": 7630 }, { "epoch": 0.759000086927366, "grad_norm": 0.4465351700782776, "learning_rate": 4.2549847599363196e-05, "loss": 1.1038, "step": 7640 }, { "epoch": 0.7599935425385275, "grad_norm": 0.48122888803482056, "learning_rate": 4.221708300357598e-05, "loss": 1.0967, "step": 7650 }, { "epoch": 0.760986998149689, "grad_norm": 0.4321027100086212, "learning_rate": 4.1885411495802794e-05, "loss": 1.1123, "step": 7660 }, { "epoch": 0.7619804537608504, "grad_norm": 0.47689056396484375, "learning_rate": 4.155483643971349e-05, "loss": 1.1147, "step": 7670 }, { "epoch": 0.7629739093720119, "grad_norm": 0.4920623004436493, "learning_rate": 4.122536118785824e-05, "loss": 1.112, "step": 7680 }, { "epoch": 0.7639673649831733, "grad_norm": 0.4757070243358612, "learning_rate": 4.089698908163329e-05, "loss": 1.1338, "step": 7690 }, { "epoch": 0.7649608205943348, "grad_norm": 0.4824003279209137, "learning_rate": 4.056972345124746e-05, "loss": 1.1306, "step": 7700 }, { "epoch": 0.7659542762054963, "grad_norm": 0.4163155257701874, "learning_rate": 4.0243567615688116e-05, "loss": 1.107, "step": 7710 }, { "epoch": 0.7669477318166578, "grad_norm": 0.47464796900749207, "learning_rate": 3.9918524882687566e-05, "loss": 1.1041, "step": 7720 }, { "epoch": 0.7679411874278192, "grad_norm": 0.5027470588684082, "learning_rate": 3.959459854868942e-05, "loss": 1.115, "step": 7730 }, { "epoch": 0.7689346430389807, "grad_norm": 0.4467754662036896, "learning_rate": 3.9271791898815366e-05, "loss": 1.1251, "step": 7740 }, { "epoch": 0.7699280986501422, "grad_norm": 0.4438583254814148, "learning_rate": 3.895010820683172e-05, "loss": 1.0996, "step": 7750 }, { "epoch": 0.7709215542613037, "grad_norm": 0.47995617985725403, "learning_rate": 3.8629550735116114e-05, "loss": 1.1178, "step": 7760 }, { "epoch": 0.7719150098724651, "grad_norm": 0.4706804156303406, "learning_rate": 3.8310122734624754e-05, "loss": 1.1237, "step": 7770 }, { "epoch": 0.7729084654836266, "grad_norm": 0.448822021484375, "learning_rate": 3.7991827444859e-05, "loss": 1.1222, "step": 7780 }, { "epoch": 0.7739019210947881, "grad_norm": 0.5071418881416321, "learning_rate": 3.767466809383293e-05, "loss": 1.129, "step": 7790 }, { "epoch": 0.7748953767059495, "grad_norm": 0.49738389253616333, "learning_rate": 3.7358647898040364e-05, "loss": 1.1227, "step": 7800 }, { "epoch": 0.775888832317111, "grad_norm": 0.4692402184009552, "learning_rate": 3.7043770062422246e-05, "loss": 1.1085, "step": 7810 }, { "epoch": 0.7768822879282725, "grad_norm": 0.5283324718475342, "learning_rate": 3.6730037780334277e-05, "loss": 1.1146, "step": 7820 }, { "epoch": 0.777875743539434, "grad_norm": 0.4743114411830902, "learning_rate": 3.6417454233514335e-05, "loss": 1.1042, "step": 7830 }, { "epoch": 0.7788691991505955, "grad_norm": 0.4658038914203644, "learning_rate": 3.610602259205044e-05, "loss": 1.1038, "step": 7840 }, { "epoch": 0.779862654761757, "grad_norm": 0.4278145432472229, "learning_rate": 3.5795746014348466e-05, "loss": 1.0945, "step": 7850 }, { "epoch": 0.7808561103729184, "grad_norm": 0.44836124777793884, "learning_rate": 3.54866276471e-05, "loss": 1.1145, "step": 7860 }, { "epoch": 0.7818495659840798, "grad_norm": 0.5174553990364075, "learning_rate": 3.517867062525072e-05, "loss": 1.0976, "step": 7870 }, { "epoch": 0.7828430215952413, "grad_norm": 0.4808596968650818, "learning_rate": 3.487187807196838e-05, "loss": 1.1224, "step": 7880 }, { "epoch": 0.7838364772064028, "grad_norm": 0.4512399435043335, "learning_rate": 3.456625309861116e-05, "loss": 1.1017, "step": 7890 }, { "epoch": 0.7848299328175643, "grad_norm": 0.5241881012916565, "learning_rate": 3.426179880469612e-05, "loss": 1.1113, "step": 7900 }, { "epoch": 0.7858233884287258, "grad_norm": 0.5214290022850037, "learning_rate": 3.3958518277867864e-05, "loss": 1.1136, "step": 7910 }, { "epoch": 0.7868168440398873, "grad_norm": 0.47837206721305847, "learning_rate": 3.365641459386715e-05, "loss": 1.1177, "step": 7920 }, { "epoch": 0.7878102996510488, "grad_norm": 0.48158422112464905, "learning_rate": 3.3355490816499716e-05, "loss": 1.099, "step": 7930 }, { "epoch": 0.7888037552622101, "grad_norm": 0.49961307644844055, "learning_rate": 3.305574999760507e-05, "loss": 1.1128, "step": 7940 }, { "epoch": 0.7897972108733716, "grad_norm": 0.4905932545661926, "learning_rate": 3.275719517702587e-05, "loss": 1.1126, "step": 7950 }, { "epoch": 0.7907906664845331, "grad_norm": 0.5185438394546509, "learning_rate": 3.2459829382576664e-05, "loss": 1.098, "step": 7960 }, { "epoch": 0.7917841220956946, "grad_norm": 0.7136569023132324, "learning_rate": 3.21636556300136e-05, "loss": 1.0941, "step": 7970 }, { "epoch": 0.7927775777068561, "grad_norm": 0.5284348130226135, "learning_rate": 3.186867692300349e-05, "loss": 1.115, "step": 7980 }, { "epoch": 0.7937710333180176, "grad_norm": 0.46627479791641235, "learning_rate": 3.157489625309363e-05, "loss": 1.1049, "step": 7990 }, { "epoch": 0.7947644889291791, "grad_norm": 0.44333553314208984, "learning_rate": 3.12823165996813e-05, "loss": 1.1043, "step": 8000 }, { "epoch": 0.7957579445403405, "grad_norm": 0.45862966775894165, "learning_rate": 3.099094092998359e-05, "loss": 1.0948, "step": 8010 }, { "epoch": 0.796751400151502, "grad_norm": 0.4916597902774811, "learning_rate": 3.070077219900729e-05, "loss": 1.1166, "step": 8020 }, { "epoch": 0.7977448557626634, "grad_norm": 0.49191707372665405, "learning_rate": 3.041181334951887e-05, "loss": 1.1113, "step": 8030 }, { "epoch": 0.7987383113738249, "grad_norm": 0.4618023931980133, "learning_rate": 3.012406731201485e-05, "loss": 1.0965, "step": 8040 }, { "epoch": 0.7997317669849864, "grad_norm": 0.4719119369983673, "learning_rate": 2.9837537004691852e-05, "loss": 1.11, "step": 8050 }, { "epoch": 0.8003278403516833, "eval_loss": 1.1472482681274414, "eval_runtime": 217.1534, "eval_samples_per_second": 230.252, "eval_steps_per_second": 28.781, "step": 8056 }, { "epoch": 0.8007252225961479, "grad_norm": 0.47482019662857056, "learning_rate": 2.9552225333417047e-05, "loss": 1.1232, "step": 8060 }, { "epoch": 0.8017186782073094, "grad_norm": 0.5259802937507629, "learning_rate": 2.9268135191698776e-05, "loss": 1.1066, "step": 8070 }, { "epoch": 0.8027121338184708, "grad_norm": 0.4933863580226898, "learning_rate": 2.8985269460657175e-05, "loss": 1.1193, "step": 8080 }, { "epoch": 0.8037055894296323, "grad_norm": 0.512607991695404, "learning_rate": 2.870363100899491e-05, "loss": 1.1055, "step": 8090 }, { "epoch": 0.8046990450407938, "grad_norm": 0.4950838088989258, "learning_rate": 2.8423222692968052e-05, "loss": 1.1113, "step": 8100 }, { "epoch": 0.8056925006519552, "grad_norm": 0.5034070014953613, "learning_rate": 2.814404735635721e-05, "loss": 1.1177, "step": 8110 }, { "epoch": 0.8066859562631167, "grad_norm": 0.463691771030426, "learning_rate": 2.7866107830438626e-05, "loss": 1.1178, "step": 8120 }, { "epoch": 0.8076794118742782, "grad_norm": 0.4671690762042999, "learning_rate": 2.7589406933955597e-05, "loss": 1.0893, "step": 8130 }, { "epoch": 0.8086728674854397, "grad_norm": 0.4814799726009369, "learning_rate": 2.731394747308956e-05, "loss": 1.1296, "step": 8140 }, { "epoch": 0.8096663230966011, "grad_norm": 0.5085082650184631, "learning_rate": 2.7039732241432016e-05, "loss": 1.1144, "step": 8150 }, { "epoch": 0.8106597787077626, "grad_norm": 0.5503705143928528, "learning_rate": 2.6766764019956006e-05, "loss": 1.0936, "step": 8160 }, { "epoch": 0.8116532343189241, "grad_norm": 0.5416733622550964, "learning_rate": 2.6495045576987828e-05, "loss": 1.0877, "step": 8170 }, { "epoch": 0.8126466899300856, "grad_norm": 0.5096998810768127, "learning_rate": 2.622457966817921e-05, "loss": 1.1084, "step": 8180 }, { "epoch": 0.8136401455412471, "grad_norm": 0.5110299587249756, "learning_rate": 2.5955369036479047e-05, "loss": 1.1143, "step": 8190 }, { "epoch": 0.8146336011524085, "grad_norm": 0.49047693610191345, "learning_rate": 2.568741641210591e-05, "loss": 1.1078, "step": 8200 }, { "epoch": 0.81562705676357, "grad_norm": 0.6099783778190613, "learning_rate": 2.5420724512520142e-05, "loss": 1.104, "step": 8210 }, { "epoch": 0.8166205123747314, "grad_norm": 0.42165687680244446, "learning_rate": 2.5155296042396305e-05, "loss": 1.1026, "step": 8220 }, { "epoch": 0.8176139679858929, "grad_norm": 0.5082433223724365, "learning_rate": 2.489113369359592e-05, "loss": 1.1073, "step": 8230 }, { "epoch": 0.8186074235970544, "grad_norm": 0.522196888923645, "learning_rate": 2.462824014513991e-05, "loss": 1.0992, "step": 8240 }, { "epoch": 0.8196008792082159, "grad_norm": 0.4945440888404846, "learning_rate": 2.4366618063181652e-05, "loss": 1.1087, "step": 8250 }, { "epoch": 0.8205943348193774, "grad_norm": 0.4586648643016815, "learning_rate": 2.4106270100979898e-05, "loss": 1.0966, "step": 8260 }, { "epoch": 0.8215877904305389, "grad_norm": 0.49676162004470825, "learning_rate": 2.3847198898871688e-05, "loss": 1.1034, "step": 8270 }, { "epoch": 0.8225812460417004, "grad_norm": 0.4228706657886505, "learning_rate": 2.3589407084245854e-05, "loss": 1.11, "step": 8280 }, { "epoch": 0.8235747016528617, "grad_norm": 0.43764999508857727, "learning_rate": 2.3332897271516177e-05, "loss": 1.0958, "step": 8290 }, { "epoch": 0.8245681572640232, "grad_norm": 0.4700354337692261, "learning_rate": 2.3077672062094892e-05, "loss": 1.1059, "step": 8300 }, { "epoch": 0.8255616128751847, "grad_norm": 0.44920220971107483, "learning_rate": 2.2823734044366325e-05, "loss": 1.1033, "step": 8310 }, { "epoch": 0.8265550684863462, "grad_norm": 0.519624650478363, "learning_rate": 2.257108579366074e-05, "loss": 1.0999, "step": 8320 }, { "epoch": 0.8275485240975077, "grad_norm": 0.5019779205322266, "learning_rate": 2.2319729872228086e-05, "loss": 1.0867, "step": 8330 }, { "epoch": 0.8285419797086692, "grad_norm": 0.4180760383605957, "learning_rate": 2.2069668829212122e-05, "loss": 1.1194, "step": 8340 }, { "epoch": 0.8295354353198306, "grad_norm": 0.47347915172576904, "learning_rate": 2.1820905200624382e-05, "loss": 1.1144, "step": 8350 }, { "epoch": 0.8305288909309921, "grad_norm": 0.4253215789794922, "learning_rate": 2.1573441509318727e-05, "loss": 1.0972, "step": 8360 }, { "epoch": 0.8315223465421535, "grad_norm": 0.4493556618690491, "learning_rate": 2.1327280264965462e-05, "loss": 1.0999, "step": 8370 }, { "epoch": 0.832515802153315, "grad_norm": 0.486507773399353, "learning_rate": 2.1082423964026208e-05, "loss": 1.0979, "step": 8380 }, { "epoch": 0.8335092577644765, "grad_norm": 0.5007907152175903, "learning_rate": 2.0838875089728236e-05, "loss": 1.1062, "step": 8390 }, { "epoch": 0.834502713375638, "grad_norm": 0.6081424951553345, "learning_rate": 2.059663611203956e-05, "loss": 1.1103, "step": 8400 }, { "epoch": 0.8354961689867995, "grad_norm": 0.4662105441093445, "learning_rate": 2.0355709487643807e-05, "loss": 1.0989, "step": 8410 }, { "epoch": 0.836489624597961, "grad_norm": 0.48797285556793213, "learning_rate": 2.0116097659915225e-05, "loss": 1.0989, "step": 8420 }, { "epoch": 0.8374830802091224, "grad_norm": 0.4916144609451294, "learning_rate": 1.987780305889397e-05, "loss": 1.0876, "step": 8430 }, { "epoch": 0.8384765358202839, "grad_norm": 0.4921468198299408, "learning_rate": 1.9640828101261413e-05, "loss": 1.1062, "step": 8440 }, { "epoch": 0.8394699914314454, "grad_norm": 0.48090696334838867, "learning_rate": 1.9405175190315747e-05, "loss": 1.1114, "step": 8450 }, { "epoch": 0.8404634470426068, "grad_norm": 0.46354711055755615, "learning_rate": 1.917084671594754e-05, "loss": 1.1038, "step": 8460 }, { "epoch": 0.8414569026537683, "grad_norm": 0.5036799907684326, "learning_rate": 1.8937845054615373e-05, "loss": 1.1213, "step": 8470 }, { "epoch": 0.8424503582649298, "grad_norm": 0.509793221950531, "learning_rate": 1.870617256932199e-05, "loss": 1.1071, "step": 8480 }, { "epoch": 0.8434438138760912, "grad_norm": 0.46539217233657837, "learning_rate": 1.8475831609590135e-05, "loss": 1.11, "step": 8490 }, { "epoch": 0.8444372694872527, "grad_norm": 0.4852668046951294, "learning_rate": 1.8246824511438868e-05, "loss": 1.1038, "step": 8500 }, { "epoch": 0.8454307250984142, "grad_norm": 0.5432482361793518, "learning_rate": 1.801915359735967e-05, "loss": 1.0954, "step": 8510 }, { "epoch": 0.8464241807095757, "grad_norm": 0.4479030668735504, "learning_rate": 1.7792821176293097e-05, "loss": 1.1083, "step": 8520 }, { "epoch": 0.8474176363207372, "grad_norm": 0.49862831830978394, "learning_rate": 1.7567829543605282e-05, "loss": 1.1043, "step": 8530 }, { "epoch": 0.8484110919318986, "grad_norm": 0.47839561104774475, "learning_rate": 1.734418098106468e-05, "loss": 1.1001, "step": 8540 }, { "epoch": 0.84940454754306, "grad_norm": 0.49436119198799133, "learning_rate": 1.712187775681883e-05, "loss": 1.0945, "step": 8550 }, { "epoch": 0.8503980031542215, "grad_norm": 0.496619313955307, "learning_rate": 1.6900922125371498e-05, "loss": 1.0862, "step": 8560 }, { "epoch": 0.851391458765383, "grad_norm": 0.4753607511520386, "learning_rate": 1.6681316327559763e-05, "loss": 1.1149, "step": 8570 }, { "epoch": 0.8523849143765445, "grad_norm": 0.5014109015464783, "learning_rate": 1.646306259053116e-05, "loss": 1.0854, "step": 8580 }, { "epoch": 0.853378369987706, "grad_norm": 0.48267966508865356, "learning_rate": 1.624616312772133e-05, "loss": 1.0944, "step": 8590 }, { "epoch": 0.8543718255988675, "grad_norm": 0.49954667687416077, "learning_rate": 1.6030620138831336e-05, "loss": 1.0941, "step": 8600 }, { "epoch": 0.855365281210029, "grad_norm": 0.494306355714798, "learning_rate": 1.5816435809805557e-05, "loss": 1.1007, "step": 8610 }, { "epoch": 0.8563587368211905, "grad_norm": 0.4919048845767975, "learning_rate": 1.5603612312809373e-05, "loss": 1.0906, "step": 8620 }, { "epoch": 0.8573521924323518, "grad_norm": 0.501146674156189, "learning_rate": 1.5392151806207254e-05, "loss": 1.1094, "step": 8630 }, { "epoch": 0.8583456480435133, "grad_norm": 0.49974754452705383, "learning_rate": 1.5182056434540729e-05, "loss": 1.1068, "step": 8640 }, { "epoch": 0.8593391036546748, "grad_norm": 0.4796450436115265, "learning_rate": 1.4973328328506734e-05, "loss": 1.0967, "step": 8650 }, { "epoch": 0.8603325592658363, "grad_norm": 0.4462718963623047, "learning_rate": 1.4765969604935995e-05, "loss": 1.1199, "step": 8660 }, { "epoch": 0.8613260148769978, "grad_norm": 0.48056918382644653, "learning_rate": 1.4559982366771616e-05, "loss": 1.0958, "step": 8670 }, { "epoch": 0.8623194704881593, "grad_norm": 0.49090835452079773, "learning_rate": 1.4355368703047543e-05, "loss": 1.1091, "step": 8680 }, { "epoch": 0.8633129260993208, "grad_norm": 0.4484087824821472, "learning_rate": 1.4152130688867647e-05, "loss": 1.0883, "step": 8690 }, { "epoch": 0.8643063817104822, "grad_norm": 0.47980430722236633, "learning_rate": 1.3950270385384526e-05, "loss": 1.1294, "step": 8700 }, { "epoch": 0.8652998373216437, "grad_norm": 0.44564327597618103, "learning_rate": 1.3749789839778574e-05, "loss": 1.0989, "step": 8710 }, { "epoch": 0.8662932929328051, "grad_norm": 0.5069140791893005, "learning_rate": 1.3550691085237292e-05, "loss": 1.0914, "step": 8720 }, { "epoch": 0.8672867485439666, "grad_norm": 0.46581125259399414, "learning_rate": 1.335297614093469e-05, "loss": 1.0894, "step": 8730 }, { "epoch": 0.8682802041551281, "grad_norm": 0.47419315576553345, "learning_rate": 1.315664701201074e-05, "loss": 1.0856, "step": 8740 }, { "epoch": 0.8692736597662896, "grad_norm": 0.4483776092529297, "learning_rate": 1.2961705689551122e-05, "loss": 1.1067, "step": 8750 }, { "epoch": 0.870267115377451, "grad_norm": 0.49877870082855225, "learning_rate": 1.2768154150566856e-05, "loss": 1.1183, "step": 8760 }, { "epoch": 0.8712605709886125, "grad_norm": 0.4901898205280304, "learning_rate": 1.2575994357974523e-05, "loss": 1.0934, "step": 8770 }, { "epoch": 0.872254026599774, "grad_norm": 0.454645574092865, "learning_rate": 1.2385228260576091e-05, "loss": 1.1092, "step": 8780 }, { "epoch": 0.8732474822109355, "grad_norm": 0.46396562457084656, "learning_rate": 1.2195857793039337e-05, "loss": 1.1136, "step": 8790 }, { "epoch": 0.8742409378220969, "grad_norm": 0.45802921056747437, "learning_rate": 1.2007884875878143e-05, "loss": 1.0872, "step": 8800 }, { "epoch": 0.8752343934332584, "grad_norm": 0.487572580575943, "learning_rate": 1.182131141543301e-05, "loss": 1.1161, "step": 8810 }, { "epoch": 0.8762278490444199, "grad_norm": 0.46634870767593384, "learning_rate": 1.163613930385181e-05, "loss": 1.095, "step": 8820 }, { "epoch": 0.8772213046555813, "grad_norm": 0.4800409972667694, "learning_rate": 1.1452370419070517e-05, "loss": 1.0905, "step": 8830 }, { "epoch": 0.8782147602667428, "grad_norm": 0.4725741446018219, "learning_rate": 1.1270006624794154e-05, "loss": 1.088, "step": 8840 }, { "epoch": 0.8792082158779043, "grad_norm": 0.44453057646751404, "learning_rate": 1.1089049770477892e-05, "loss": 1.1069, "step": 8850 }, { "epoch": 0.8802016714890658, "grad_norm": 0.47712013125419617, "learning_rate": 1.0909501691308387e-05, "loss": 1.1027, "step": 8860 }, { "epoch": 0.8811951271002273, "grad_norm": 0.4958110451698303, "learning_rate": 1.07313642081851e-05, "loss": 1.094, "step": 8870 }, { "epoch": 0.8821885827113888, "grad_norm": 0.4982723593711853, "learning_rate": 1.0554639127701808e-05, "loss": 1.0986, "step": 8880 }, { "epoch": 0.8831820383225502, "grad_norm": 0.4847375154495239, "learning_rate": 1.0379328242128294e-05, "loss": 1.095, "step": 8890 }, { "epoch": 0.8841754939337116, "grad_norm": 0.5036672949790955, "learning_rate": 1.020543332939221e-05, "loss": 1.1015, "step": 8900 }, { "epoch": 0.8851689495448731, "grad_norm": 0.4520683288574219, "learning_rate": 1.003295615306105e-05, "loss": 1.1039, "step": 8910 }, { "epoch": 0.8861624051560346, "grad_norm": 0.45851820707321167, "learning_rate": 9.861898462324158e-06, "loss": 1.0939, "step": 8920 }, { "epoch": 0.8871558607671961, "grad_norm": 0.6486339569091797, "learning_rate": 9.692261991975076e-06, "loss": 1.0876, "step": 8930 }, { "epoch": 0.8881493163783576, "grad_norm": 0.46632036566734314, "learning_rate": 9.524048462394024e-06, "loss": 1.1131, "step": 8940 }, { "epoch": 0.8891427719895191, "grad_norm": 0.45053717494010925, "learning_rate": 9.357259579530285e-06, "loss": 1.1084, "step": 8950 }, { "epoch": 0.8901362276006806, "grad_norm": 0.4486932158470154, "learning_rate": 9.191897034885033e-06, "loss": 1.0832, "step": 8960 }, { "epoch": 0.8911296832118419, "grad_norm": 0.4375132620334625, "learning_rate": 9.027962505494108e-06, "loss": 1.0946, "step": 8970 }, { "epoch": 0.8921231388230034, "grad_norm": 0.49327731132507324, "learning_rate": 8.86545765391104e-06, "loss": 1.1064, "step": 8980 }, { "epoch": 0.8931165944341649, "grad_norm": 0.4718957245349884, "learning_rate": 8.704384128190156e-06, "loss": 1.0874, "step": 8990 }, { "epoch": 0.8941100500453264, "grad_norm": 0.4812639653682709, "learning_rate": 8.544743561869977e-06, "loss": 1.0994, "step": 9000 }, { "epoch": 0.8951035056564879, "grad_norm": 0.5206982493400574, "learning_rate": 8.386537573956431e-06, "loss": 1.1021, "step": 9010 }, { "epoch": 0.8960969612676494, "grad_norm": 0.49871826171875, "learning_rate": 8.229767768906687e-06, "loss": 1.1138, "step": 9020 }, { "epoch": 0.8970904168788109, "grad_norm": 0.4538607895374298, "learning_rate": 8.074435736612712e-06, "loss": 1.0826, "step": 9030 }, { "epoch": 0.8980838724899723, "grad_norm": 0.4945856034755707, "learning_rate": 7.920543052385236e-06, "loss": 1.1204, "step": 9040 }, { "epoch": 0.8990773281011338, "grad_norm": 0.5100029110908508, "learning_rate": 7.768091276937688e-06, "loss": 1.1011, "step": 9050 }, { "epoch": 0.9000707837122952, "grad_norm": 0.4894271194934845, "learning_rate": 7.617081956370413e-06, "loss": 1.0853, "step": 9060 }, { "epoch": 0.9003688203956437, "eval_loss": 1.1406679153442383, "eval_runtime": 216.5409, "eval_samples_per_second": 230.903, "eval_steps_per_second": 28.863, "step": 9063 }, { "epoch": 0.9010642393234567, "grad_norm": 0.511590838432312, "learning_rate": 7.467516622155079e-06, "loss": 1.1079, "step": 9070 }, { "epoch": 0.9020576949346182, "grad_norm": 0.45951762795448303, "learning_rate": 7.319396791118998e-06, "loss": 1.0963, "step": 9080 }, { "epoch": 0.9030511505457797, "grad_norm": 0.4679279625415802, "learning_rate": 7.172723965429811e-06, "loss": 1.0922, "step": 9090 }, { "epoch": 0.9040446061569412, "grad_norm": 0.5013454556465149, "learning_rate": 7.027499632580264e-06, "loss": 1.0782, "step": 9100 }, { "epoch": 0.9050380617681026, "grad_norm": 0.49527832865715027, "learning_rate": 6.883725265373185e-06, "loss": 1.09, "step": 9110 }, { "epoch": 0.9060315173792641, "grad_norm": 0.4694949984550476, "learning_rate": 6.74140232190632e-06, "loss": 1.083, "step": 9120 }, { "epoch": 0.9070249729904256, "grad_norm": 0.4897943139076233, "learning_rate": 6.600532245557833e-06, "loss": 1.0954, "step": 9130 }, { "epoch": 0.908018428601587, "grad_norm": 0.48320844769477844, "learning_rate": 6.461116464971461e-06, "loss": 1.1021, "step": 9140 }, { "epoch": 0.9090118842127485, "grad_norm": 0.498017817735672, "learning_rate": 6.323156394042139e-06, "loss": 1.1142, "step": 9150 }, { "epoch": 0.91000533982391, "grad_norm": 0.4676865041255951, "learning_rate": 6.186653431901628e-06, "loss": 1.0964, "step": 9160 }, { "epoch": 0.9109987954350715, "grad_norm": 0.5362632870674133, "learning_rate": 6.051608962904225e-06, "loss": 1.1146, "step": 9170 }, { "epoch": 0.9119922510462329, "grad_norm": 0.5107495188713074, "learning_rate": 5.91802435661296e-06, "loss": 1.1102, "step": 9180 }, { "epoch": 0.9129857066573944, "grad_norm": 0.47366631031036377, "learning_rate": 5.78590096778544e-06, "loss": 1.0898, "step": 9190 }, { "epoch": 0.9139791622685559, "grad_norm": 0.44392383098602295, "learning_rate": 5.655240136360306e-06, "loss": 1.0917, "step": 9200 }, { "epoch": 0.9149726178797174, "grad_norm": 0.5138424038887024, "learning_rate": 5.5260431874435515e-06, "loss": 1.0922, "step": 9210 }, { "epoch": 0.9159660734908789, "grad_norm": 0.49133333563804626, "learning_rate": 5.398311431295061e-06, "loss": 1.103, "step": 9220 }, { "epoch": 0.9169595291020403, "grad_norm": 0.4637450873851776, "learning_rate": 5.2720461633154354e-06, "loss": 1.1077, "step": 9230 }, { "epoch": 0.9179529847132017, "grad_norm": 0.47091439366340637, "learning_rate": 5.147248664032744e-06, "loss": 1.0943, "step": 9240 }, { "epoch": 0.9189464403243632, "grad_norm": 0.4741147756576538, "learning_rate": 5.023920199089576e-06, "loss": 1.0887, "step": 9250 }, { "epoch": 0.9199398959355247, "grad_norm": 0.48109114170074463, "learning_rate": 4.902062019230191e-06, "loss": 1.0887, "step": 9260 }, { "epoch": 0.9209333515466862, "grad_norm": 0.5040728449821472, "learning_rate": 4.78167536028789e-06, "loss": 1.0824, "step": 9270 }, { "epoch": 0.9219268071578477, "grad_norm": 0.5055690407752991, "learning_rate": 4.662761443172414e-06, "loss": 1.1001, "step": 9280 }, { "epoch": 0.9229202627690092, "grad_norm": 0.4466319978237152, "learning_rate": 4.545321473857616e-06, "loss": 1.0991, "step": 9290 }, { "epoch": 0.9239137183801707, "grad_norm": 0.5149044394493103, "learning_rate": 4.429356643369159e-06, "loss": 1.0929, "step": 9300 }, { "epoch": 0.9249071739913322, "grad_norm": 0.4559183418750763, "learning_rate": 4.314868127772503e-06, "loss": 1.0871, "step": 9310 }, { "epoch": 0.9259006296024935, "grad_norm": 0.4630989134311676, "learning_rate": 4.201857088160987e-06, "loss": 1.0856, "step": 9320 }, { "epoch": 0.926894085213655, "grad_norm": 0.4689970016479492, "learning_rate": 4.090324670643968e-06, "loss": 1.1021, "step": 9330 }, { "epoch": 0.9278875408248165, "grad_norm": 0.4830831289291382, "learning_rate": 3.9802720063353e-06, "loss": 1.1006, "step": 9340 }, { "epoch": 0.928880996435978, "grad_norm": 0.44650521874427795, "learning_rate": 3.8717002113417885e-06, "loss": 1.1021, "step": 9350 }, { "epoch": 0.9298744520471395, "grad_norm": 0.5089627504348755, "learning_rate": 3.764610386751904e-06, "loss": 1.0884, "step": 9360 }, { "epoch": 0.930867907658301, "grad_norm": 0.5018206238746643, "learning_rate": 3.659003618624656e-06, "loss": 1.087, "step": 9370 }, { "epoch": 0.9318613632694625, "grad_norm": 0.6836445331573486, "learning_rate": 3.554880977978436e-06, "loss": 1.1166, "step": 9380 }, { "epoch": 0.9328548188806239, "grad_norm": 0.4803834855556488, "learning_rate": 3.4522435207803567e-06, "loss": 1.0942, "step": 9390 }, { "epoch": 0.9338482744917853, "grad_norm": 0.52688068151474, "learning_rate": 3.3510922879353476e-06, "loss": 1.0967, "step": 9400 }, { "epoch": 0.9348417301029468, "grad_norm": 0.4448314309120178, "learning_rate": 3.2514283052757605e-06, "loss": 1.1083, "step": 9410 }, { "epoch": 0.9358351857141083, "grad_norm": 0.5127049088478088, "learning_rate": 3.1532525835508636e-06, "loss": 1.1008, "step": 9420 }, { "epoch": 0.9368286413252698, "grad_norm": 0.4845719635486603, "learning_rate": 3.0565661184166146e-06, "loss": 1.0877, "step": 9430 }, { "epoch": 0.9378220969364313, "grad_norm": 0.4783158302307129, "learning_rate": 2.9613698904255857e-06, "loss": 1.0992, "step": 9440 }, { "epoch": 0.9388155525475927, "grad_norm": 0.44096460938453674, "learning_rate": 2.867664865017022e-06, "loss": 1.1033, "step": 9450 }, { "epoch": 0.9398090081587542, "grad_norm": 0.47711020708084106, "learning_rate": 2.7754519925069995e-06, "loss": 1.0719, "step": 9460 }, { "epoch": 0.9408024637699157, "grad_norm": 0.46728119254112244, "learning_rate": 2.684732208078816e-06, "loss": 1.0953, "step": 9470 }, { "epoch": 0.9417959193810772, "grad_norm": 0.48205727338790894, "learning_rate": 2.5955064317735474e-06, "loss": 1.1177, "step": 9480 }, { "epoch": 0.9427893749922386, "grad_norm": 0.4576367139816284, "learning_rate": 2.5077755684806745e-06, "loss": 1.11, "step": 9490 }, { "epoch": 0.9437828306034001, "grad_norm": 0.4937973618507385, "learning_rate": 2.421540507928871e-06, "loss": 1.0879, "step": 9500 }, { "epoch": 0.9447762862145616, "grad_norm": 0.4714371860027313, "learning_rate": 2.3368021246770786e-06, "loss": 1.0791, "step": 9510 }, { "epoch": 0.945769741825723, "grad_norm": 0.4823916554450989, "learning_rate": 2.2535612781055478e-06, "loss": 1.0933, "step": 9520 }, { "epoch": 0.9467631974368845, "grad_norm": 0.48076769709587097, "learning_rate": 2.1718188124071613e-06, "loss": 1.0987, "step": 9530 }, { "epoch": 0.947756653048046, "grad_norm": 0.48411691188812256, "learning_rate": 2.0915755565788728e-06, "loss": 1.1189, "step": 9540 }, { "epoch": 0.9487501086592075, "grad_norm": 0.47723978757858276, "learning_rate": 2.0128323244132826e-06, "loss": 1.11, "step": 9550 }, { "epoch": 0.949743564270369, "grad_norm": 0.4722275137901306, "learning_rate": 1.9355899144904263e-06, "loss": 1.1024, "step": 9560 }, { "epoch": 0.9507370198815304, "grad_norm": 0.4917920231819153, "learning_rate": 1.8598491101695978e-06, "loss": 1.1054, "step": 9570 }, { "epoch": 0.9517304754926919, "grad_norm": 0.47444161772727966, "learning_rate": 1.7856106795814729e-06, "loss": 1.1151, "step": 9580 }, { "epoch": 0.9527239311038533, "grad_norm": 0.4620038866996765, "learning_rate": 1.7128753756203484e-06, "loss": 1.0862, "step": 9590 }, { "epoch": 0.9537173867150148, "grad_norm": 0.45393404364585876, "learning_rate": 1.6416439359363487e-06, "loss": 1.0934, "step": 9600 }, { "epoch": 0.9547108423261763, "grad_norm": 0.5473287105560303, "learning_rate": 1.5719170829281646e-06, "loss": 1.0973, "step": 9610 }, { "epoch": 0.9557042979373378, "grad_norm": 0.4900316298007965, "learning_rate": 1.503695523735543e-06, "loss": 1.095, "step": 9620 }, { "epoch": 0.9566977535484993, "grad_norm": 0.49104568362236023, "learning_rate": 1.4369799502322254e-06, "loss": 1.0882, "step": 9630 }, { "epoch": 0.9576912091596608, "grad_norm": 0.4944591820240021, "learning_rate": 1.3717710390188709e-06, "loss": 1.0907, "step": 9640 }, { "epoch": 0.9586846647708223, "grad_norm": 0.46487265825271606, "learning_rate": 1.3080694514162437e-06, "loss": 1.1026, "step": 9650 }, { "epoch": 0.9596781203819836, "grad_norm": 0.49333804845809937, "learning_rate": 1.245875833458454e-06, "loss": 1.0832, "step": 9660 }, { "epoch": 0.9606715759931451, "grad_norm": 0.48843681812286377, "learning_rate": 1.1851908158864443e-06, "loss": 1.0941, "step": 9670 }, { "epoch": 0.9616650316043066, "grad_norm": 0.4628733694553375, "learning_rate": 1.126015014141546e-06, "loss": 1.0861, "step": 9680 }, { "epoch": 0.9626584872154681, "grad_norm": 0.5206993818283081, "learning_rate": 1.0683490283593499e-06, "loss": 1.0882, "step": 9690 }, { "epoch": 0.9636519428266296, "grad_norm": 0.4243178367614746, "learning_rate": 1.0121934433634793e-06, "loss": 1.0828, "step": 9700 }, { "epoch": 0.9646453984377911, "grad_norm": 0.4830005466938019, "learning_rate": 9.575488286597267e-07, "loss": 1.0858, "step": 9710 }, { "epoch": 0.9656388540489526, "grad_norm": 0.475365549325943, "learning_rate": 9.044157384302919e-07, "loss": 1.0938, "step": 9720 }, { "epoch": 0.966632309660114, "grad_norm": 0.5186288356781006, "learning_rate": 8.527947115281542e-07, "loss": 1.1088, "step": 9730 }, { "epoch": 0.9676257652712754, "grad_norm": 0.4813586175441742, "learning_rate": 8.026862714715587e-07, "loss": 1.1051, "step": 9740 }, { "epoch": 0.9686192208824369, "grad_norm": 0.4743388295173645, "learning_rate": 7.540909264387718e-07, "loss": 1.0817, "step": 9750 }, { "epoch": 0.9696126764935984, "grad_norm": 0.5275624394416809, "learning_rate": 7.07009169262901e-07, "loss": 1.0867, "step": 9760 }, { "epoch": 0.9706061321047599, "grad_norm": 0.5017628073692322, "learning_rate": 6.614414774268661e-07, "loss": 1.0886, "step": 9770 }, { "epoch": 0.9715995877159214, "grad_norm": 0.5396377444267273, "learning_rate": 6.173883130586532e-07, "loss": 1.0984, "step": 9780 }, { "epoch": 0.9725930433270829, "grad_norm": 0.4866466522216797, "learning_rate": 5.748501229264679e-07, "loss": 1.0913, "step": 9790 }, { "epoch": 0.9735864989382443, "grad_norm": 0.4783855080604553, "learning_rate": 5.338273384343895e-07, "loss": 1.0894, "step": 9800 }, { "epoch": 0.9745799545494058, "grad_norm": 0.4767221510410309, "learning_rate": 4.94320375617857e-07, "loss": 1.0908, "step": 9810 }, { "epoch": 0.9755734101605673, "grad_norm": 0.506241500377655, "learning_rate": 4.563296351394907e-07, "loss": 1.0976, "step": 9820 }, { "epoch": 0.9765668657717287, "grad_norm": 0.46907883882522583, "learning_rate": 4.1985550228506025e-07, "loss": 1.0968, "step": 9830 }, { "epoch": 0.9775603213828902, "grad_norm": 0.4644452929496765, "learning_rate": 3.8489834695953903e-07, "loss": 1.0773, "step": 9840 }, { "epoch": 0.9785537769940517, "grad_norm": 0.5320615768432617, "learning_rate": 3.5145852368337336e-07, "loss": 1.1059, "step": 9850 }, { "epoch": 0.9795472326052131, "grad_norm": 0.4989818036556244, "learning_rate": 3.195363715888688e-07, "loss": 1.0912, "step": 9860 }, { "epoch": 0.9805406882163746, "grad_norm": 0.4837660789489746, "learning_rate": 2.891322144167929e-07, "loss": 1.097, "step": 9870 }, { "epoch": 0.9815341438275361, "grad_norm": 0.4884801208972931, "learning_rate": 2.6024636051301116e-07, "loss": 1.1015, "step": 9880 }, { "epoch": 0.9825275994386976, "grad_norm": 0.48822057247161865, "learning_rate": 2.328791028254562e-07, "loss": 1.0956, "step": 9890 }, { "epoch": 0.9835210550498591, "grad_norm": 0.49632689356803894, "learning_rate": 2.0703071890108003e-07, "loss": 1.1018, "step": 9900 }, { "epoch": 0.9845145106610206, "grad_norm": 0.5012553334236145, "learning_rate": 1.8270147088308984e-07, "loss": 1.0819, "step": 9910 }, { "epoch": 0.985507966272182, "grad_norm": 0.4779779613018036, "learning_rate": 1.5989160550826663e-07, "loss": 1.1054, "step": 9920 }, { "epoch": 0.9865014218833434, "grad_norm": 0.4931352138519287, "learning_rate": 1.3860135410443395e-07, "loss": 1.1017, "step": 9930 }, { "epoch": 0.9874948774945049, "grad_norm": 0.44229504466056824, "learning_rate": 1.1883093258819309e-07, "loss": 1.0976, "step": 9940 }, { "epoch": 0.9884883331056664, "grad_norm": 0.462110310792923, "learning_rate": 1.0058054146265815e-07, "loss": 1.0927, "step": 9950 }, { "epoch": 0.9894817887168279, "grad_norm": 0.491909921169281, "learning_rate": 8.385036581545768e-08, "loss": 1.0874, "step": 9960 }, { "epoch": 0.9904752443279894, "grad_norm": 0.46828725934028625, "learning_rate": 6.864057531681955e-08, "loss": 1.0923, "step": 9970 }, { "epoch": 0.9914686999391509, "grad_norm": 0.4887199401855469, "learning_rate": 5.49513242179056e-08, "loss": 1.1085, "step": 9980 }, { "epoch": 0.9924621555503124, "grad_norm": 0.4563218057155609, "learning_rate": 4.2782751349246246e-08, "loss": 1.0857, "step": 9990 }, { "epoch": 0.9934556111614737, "grad_norm": 0.4627005159854889, "learning_rate": 3.213498011924165e-08, "loss": 1.1072, "step": 10000 }, { "epoch": 0.9944490667726352, "grad_norm": 0.48577913641929626, "learning_rate": 2.3008118513012653e-08, "loss": 1.0907, "step": 10010 }, { "epoch": 0.9954425223837967, "grad_norm": 0.45495930314064026, "learning_rate": 1.540225909130166e-08, "loss": 1.0997, "step": 10020 }, { "epoch": 0.9964359779949582, "grad_norm": 0.4800977408885956, "learning_rate": 9.317478989423477e-09, "loss": 1.0939, "step": 10030 }, { "epoch": 0.9974294336061197, "grad_norm": 0.5329833030700684, "learning_rate": 4.753839916632474e-09, "loss": 1.1044, "step": 10040 }, { "epoch": 0.9984228892172812, "grad_norm": 0.4718695282936096, "learning_rate": 1.7113881553731945e-09, "loss": 1.1051, "step": 10050 }, { "epoch": 0.9994163448284427, "grad_norm": 0.5231459140777588, "learning_rate": 1.901545608973265e-10, "loss": 1.1028, "step": 10060 } ], "logging_steps": 10, "max_steps": 10065, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1007, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.165714517491712e+16, "train_batch_size": 16, "trial_name": null, "trial_params": null }