{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 3859, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002591344908007256, "grad_norm": 111.01161666281861, "learning_rate": 5.181347150259068e-07, "loss": 12.3076, "step": 1 }, { "epoch": 0.001295672454003628, "grad_norm": 128.99309976149883, "learning_rate": 2.5906735751295338e-06, "loss": 11.8519, "step": 5 }, { "epoch": 0.002591344908007256, "grad_norm": 117.2928726132405, "learning_rate": 5.1813471502590676e-06, "loss": 12.3366, "step": 10 }, { "epoch": 0.0038870173620108835, "grad_norm": 89.9905121675766, "learning_rate": 7.772020725388602e-06, "loss": 11.7319, "step": 15 }, { "epoch": 0.005182689816014512, "grad_norm": 52.319681120581, "learning_rate": 1.0362694300518135e-05, "loss": 8.7838, "step": 20 }, { "epoch": 0.0064783622700181395, "grad_norm": 38.24088557602701, "learning_rate": 1.2953367875647668e-05, "loss": 7.3055, "step": 25 }, { "epoch": 0.007774034724021767, "grad_norm": 25.83067627606677, "learning_rate": 1.5544041450777204e-05, "loss": 5.7221, "step": 30 }, { "epoch": 0.009069707178025395, "grad_norm": 7.762285591416621, "learning_rate": 1.813471502590674e-05, "loss": 4.3566, "step": 35 }, { "epoch": 0.010365379632029024, "grad_norm": 4.409457180338291, "learning_rate": 2.072538860103627e-05, "loss": 3.6268, "step": 40 }, { "epoch": 0.011661052086032651, "grad_norm": 2.601317803014797, "learning_rate": 2.3316062176165805e-05, "loss": 3.3055, "step": 45 }, { "epoch": 0.012956724540036279, "grad_norm": 1.8400291034971308, "learning_rate": 2.5906735751295337e-05, "loss": 3.0095, "step": 50 }, { "epoch": 0.014252396994039906, "grad_norm": 1.4104791862031891, "learning_rate": 2.8497409326424872e-05, "loss": 2.7918, "step": 55 }, { "epoch": 0.015548069448043534, "grad_norm": 1.4652220793947772, "learning_rate": 3.108808290155441e-05, "loss": 2.7691, "step": 60 }, { "epoch": 0.01684374190204716, "grad_norm": 2.1149598221315955, "learning_rate": 3.367875647668394e-05, "loss": 2.6915, "step": 65 }, { "epoch": 0.01813941435605079, "grad_norm": 2.508562821895131, "learning_rate": 3.626943005181348e-05, "loss": 2.5864, "step": 70 }, { "epoch": 0.019435086810054417, "grad_norm": 5.312266133287117, "learning_rate": 3.886010362694301e-05, "loss": 2.3482, "step": 75 }, { "epoch": 0.020730759264058048, "grad_norm": 6.635762805683308, "learning_rate": 4.145077720207254e-05, "loss": 2.0166, "step": 80 }, { "epoch": 0.022026431718061675, "grad_norm": 1.6009933261373699, "learning_rate": 4.404145077720208e-05, "loss": 1.5391, "step": 85 }, { "epoch": 0.023322104172065303, "grad_norm": 1.390865987104561, "learning_rate": 4.663212435233161e-05, "loss": 1.3383, "step": 90 }, { "epoch": 0.02461777662606893, "grad_norm": 1.0505820893868818, "learning_rate": 4.922279792746114e-05, "loss": 1.2646, "step": 95 }, { "epoch": 0.025913449080072558, "grad_norm": 1.0086607725155037, "learning_rate": 5.1813471502590674e-05, "loss": 1.3202, "step": 100 }, { "epoch": 0.027209121534076185, "grad_norm": 5.464697038037798, "learning_rate": 5.440414507772021e-05, "loss": 1.2451, "step": 105 }, { "epoch": 0.028504793988079813, "grad_norm": 0.830545367098088, "learning_rate": 5.6994818652849744e-05, "loss": 1.2681, "step": 110 }, { "epoch": 0.02980046644208344, "grad_norm": 0.7132598923355695, "learning_rate": 5.9585492227979276e-05, "loss": 1.2245, "step": 115 }, { "epoch": 0.031096138896087068, "grad_norm": 0.8138882694476683, "learning_rate": 6.217616580310881e-05, "loss": 1.2305, "step": 120 }, { "epoch": 0.0323918113500907, "grad_norm": 0.8807133122197233, "learning_rate": 6.476683937823834e-05, "loss": 1.2634, "step": 125 }, { "epoch": 0.03368748380409432, "grad_norm": 0.62485091325362, "learning_rate": 6.735751295336788e-05, "loss": 1.3386, "step": 130 }, { "epoch": 0.034983156258097954, "grad_norm": 0.7549667435061143, "learning_rate": 6.994818652849742e-05, "loss": 1.2513, "step": 135 }, { "epoch": 0.03627882871210158, "grad_norm": 0.6420487687876172, "learning_rate": 7.253886010362695e-05, "loss": 1.2101, "step": 140 }, { "epoch": 0.03757450116610521, "grad_norm": 0.7398721962367694, "learning_rate": 7.512953367875648e-05, "loss": 1.2776, "step": 145 }, { "epoch": 0.03887017362010883, "grad_norm": 0.6450506158667421, "learning_rate": 7.772020725388602e-05, "loss": 1.2096, "step": 150 }, { "epoch": 0.040165846074112464, "grad_norm": 0.6506841169994266, "learning_rate": 8.031088082901554e-05, "loss": 1.228, "step": 155 }, { "epoch": 0.041461518528116095, "grad_norm": 0.711359178146542, "learning_rate": 8.290155440414508e-05, "loss": 1.2474, "step": 160 }, { "epoch": 0.04275719098211972, "grad_norm": 1.459357949005984, "learning_rate": 8.549222797927462e-05, "loss": 1.2251, "step": 165 }, { "epoch": 0.04405286343612335, "grad_norm": 0.8207401278130507, "learning_rate": 8.808290155440416e-05, "loss": 1.2625, "step": 170 }, { "epoch": 0.045348535890126974, "grad_norm": 0.6271337633480082, "learning_rate": 9.067357512953368e-05, "loss": 1.2222, "step": 175 }, { "epoch": 0.046644208344130605, "grad_norm": 0.6465301721957318, "learning_rate": 9.326424870466322e-05, "loss": 1.2357, "step": 180 }, { "epoch": 0.04793988079813423, "grad_norm": 0.5864341660869246, "learning_rate": 9.585492227979275e-05, "loss": 1.2422, "step": 185 }, { "epoch": 0.04923555325213786, "grad_norm": 0.5644516524286168, "learning_rate": 9.844559585492228e-05, "loss": 1.1901, "step": 190 }, { "epoch": 0.050531225706141485, "grad_norm": 0.654016063568062, "learning_rate": 0.00010103626943005182, "loss": 1.182, "step": 195 }, { "epoch": 0.051826898160145116, "grad_norm": 0.569272168627719, "learning_rate": 0.00010362694300518135, "loss": 1.1958, "step": 200 }, { "epoch": 0.05312257061414875, "grad_norm": 0.6368680847682564, "learning_rate": 0.00010621761658031089, "loss": 1.2694, "step": 205 }, { "epoch": 0.05441824306815237, "grad_norm": 0.5948244637123268, "learning_rate": 0.00010880829015544042, "loss": 1.1967, "step": 210 }, { "epoch": 0.055713915522156, "grad_norm": 0.5541951524654064, "learning_rate": 0.00011139896373056995, "loss": 1.1941, "step": 215 }, { "epoch": 0.057009587976159626, "grad_norm": 0.5701648418604164, "learning_rate": 0.00011398963730569949, "loss": 1.2443, "step": 220 }, { "epoch": 0.05830526043016326, "grad_norm": 0.4865998338825169, "learning_rate": 0.00011658031088082901, "loss": 1.2106, "step": 225 }, { "epoch": 0.05960093288416688, "grad_norm": 0.5131095922403084, "learning_rate": 0.00011917098445595855, "loss": 1.1278, "step": 230 }, { "epoch": 0.06089660533817051, "grad_norm": 0.6563988864261227, "learning_rate": 0.0001217616580310881, "loss": 1.1991, "step": 235 }, { "epoch": 0.062192277792174136, "grad_norm": 0.5769507376758363, "learning_rate": 0.00012435233160621763, "loss": 1.1674, "step": 240 }, { "epoch": 0.06348795024617776, "grad_norm": 0.5283246008008212, "learning_rate": 0.00012694300518134715, "loss": 1.2367, "step": 245 }, { "epoch": 0.0647836227001814, "grad_norm": 0.5536285028186603, "learning_rate": 0.00012953367875647668, "loss": 1.1649, "step": 250 }, { "epoch": 0.06607929515418502, "grad_norm": 0.5163473385172975, "learning_rate": 0.00013212435233160623, "loss": 1.2867, "step": 255 }, { "epoch": 0.06737496760818865, "grad_norm": 0.5475813868100357, "learning_rate": 0.00013471502590673575, "loss": 1.2152, "step": 260 }, { "epoch": 0.06867064006219228, "grad_norm": 0.511381640001362, "learning_rate": 0.00013730569948186528, "loss": 1.2065, "step": 265 }, { "epoch": 0.06996631251619591, "grad_norm": 0.4982475298525502, "learning_rate": 0.00013989637305699483, "loss": 1.2142, "step": 270 }, { "epoch": 0.07126198497019953, "grad_norm": 0.5433832176160214, "learning_rate": 0.00014248704663212436, "loss": 1.2276, "step": 275 }, { "epoch": 0.07255765742420316, "grad_norm": 0.47834421596861043, "learning_rate": 0.0001450777202072539, "loss": 1.1183, "step": 280 }, { "epoch": 0.0738533298782068, "grad_norm": 2.276620321866631, "learning_rate": 0.0001476683937823834, "loss": 1.2225, "step": 285 }, { "epoch": 0.07514900233221042, "grad_norm": 0.5622928799874418, "learning_rate": 0.00015025906735751296, "loss": 1.142, "step": 290 }, { "epoch": 0.07644467478621404, "grad_norm": 0.48315577966120404, "learning_rate": 0.0001528497409326425, "loss": 1.1481, "step": 295 }, { "epoch": 0.07774034724021767, "grad_norm": 0.5054965652645259, "learning_rate": 0.00015544041450777204, "loss": 1.1334, "step": 300 }, { "epoch": 0.0790360196942213, "grad_norm": 0.5412094676646951, "learning_rate": 0.00015803108808290156, "loss": 1.2215, "step": 305 }, { "epoch": 0.08033169214822493, "grad_norm": 0.489058036506764, "learning_rate": 0.00016062176165803108, "loss": 1.1651, "step": 310 }, { "epoch": 0.08162736460222855, "grad_norm": 0.5225993706207775, "learning_rate": 0.00016321243523316064, "loss": 1.173, "step": 315 }, { "epoch": 0.08292303705623219, "grad_norm": 0.4941052544991184, "learning_rate": 0.00016580310880829016, "loss": 1.1537, "step": 320 }, { "epoch": 0.08421870951023581, "grad_norm": 0.5317455067919804, "learning_rate": 0.0001683937823834197, "loss": 1.1702, "step": 325 }, { "epoch": 0.08551438196423944, "grad_norm": 0.5382129937077951, "learning_rate": 0.00017098445595854924, "loss": 1.1954, "step": 330 }, { "epoch": 0.08681005441824306, "grad_norm": 0.4873382544458068, "learning_rate": 0.00017357512953367876, "loss": 1.1292, "step": 335 }, { "epoch": 0.0881057268722467, "grad_norm": 0.869634547296972, "learning_rate": 0.00017616580310880832, "loss": 1.2342, "step": 340 }, { "epoch": 0.08940139932625032, "grad_norm": 0.47105707267021146, "learning_rate": 0.0001787564766839378, "loss": 1.1903, "step": 345 }, { "epoch": 0.09069707178025395, "grad_norm": 0.483761714583125, "learning_rate": 0.00018134715025906737, "loss": 1.1753, "step": 350 }, { "epoch": 0.09199274423425757, "grad_norm": 0.5094212322859152, "learning_rate": 0.00018393782383419692, "loss": 1.1736, "step": 355 }, { "epoch": 0.09328841668826121, "grad_norm": 0.49665577393117427, "learning_rate": 0.00018652849740932644, "loss": 1.19, "step": 360 }, { "epoch": 0.09458408914226483, "grad_norm": 0.4634306042294334, "learning_rate": 0.00018911917098445597, "loss": 1.2068, "step": 365 }, { "epoch": 0.09587976159626846, "grad_norm": 0.4771955076541271, "learning_rate": 0.0001917098445595855, "loss": 1.1746, "step": 370 }, { "epoch": 0.0971754340502721, "grad_norm": 0.49488098544084536, "learning_rate": 0.00019430051813471504, "loss": 1.1781, "step": 375 }, { "epoch": 0.09847110650427572, "grad_norm": 0.45559447847276796, "learning_rate": 0.00019689119170984457, "loss": 1.207, "step": 380 }, { "epoch": 0.09976677895827935, "grad_norm": 0.530644388105199, "learning_rate": 0.0001994818652849741, "loss": 1.2059, "step": 385 }, { "epoch": 0.10106245141228297, "grad_norm": 0.4262101606219483, "learning_rate": 0.0001999993453944367, "loss": 1.1728, "step": 390 }, { "epoch": 0.10235812386628661, "grad_norm": 0.5019930710713301, "learning_rate": 0.00019999668607402385, "loss": 1.1566, "step": 395 }, { "epoch": 0.10365379632029023, "grad_norm": 0.41212142912399047, "learning_rate": 0.00019999198118027207, "loss": 1.1367, "step": 400 }, { "epoch": 0.10494946877429386, "grad_norm": 0.4880776387603866, "learning_rate": 0.00019998523080942663, "loss": 1.1654, "step": 405 }, { "epoch": 0.1062451412282975, "grad_norm": 0.46286032679323585, "learning_rate": 0.00019997643509957582, "loss": 1.1944, "step": 410 }, { "epoch": 0.10754081368230112, "grad_norm": 0.5382364265983718, "learning_rate": 0.00019996559423064838, "loss": 1.1871, "step": 415 }, { "epoch": 0.10883648613630474, "grad_norm": 0.5437136001752437, "learning_rate": 0.0001999527084244095, "loss": 1.1211, "step": 420 }, { "epoch": 0.11013215859030837, "grad_norm": 0.4683209703143848, "learning_rate": 0.00019993777794445662, "loss": 1.2111, "step": 425 }, { "epoch": 0.111427831044312, "grad_norm": 0.43839848429469386, "learning_rate": 0.00019992080309621371, "loss": 1.1655, "step": 430 }, { "epoch": 0.11272350349831563, "grad_norm": 0.4352884542992481, "learning_rate": 0.00019990178422692528, "loss": 1.1674, "step": 435 }, { "epoch": 0.11401917595231925, "grad_norm": 0.45552967972039005, "learning_rate": 0.00019988072172564918, "loss": 1.2005, "step": 440 }, { "epoch": 0.11531484840632288, "grad_norm": 0.46569070576531457, "learning_rate": 0.0001998576160232485, "loss": 1.1611, "step": 445 }, { "epoch": 0.11661052086032651, "grad_norm": 0.4344558621258368, "learning_rate": 0.00019983246759238305, "loss": 1.1632, "step": 450 }, { "epoch": 0.11790619331433014, "grad_norm": 0.4524631212974695, "learning_rate": 0.00019980527694749952, "loss": 1.2003, "step": 455 }, { "epoch": 0.11920186576833376, "grad_norm": 0.692041740038434, "learning_rate": 0.00019977604464482083, "loss": 1.233, "step": 460 }, { "epoch": 0.1204975382223374, "grad_norm": 0.4499116414766127, "learning_rate": 0.00019974477128233505, "loss": 1.1431, "step": 465 }, { "epoch": 0.12179321067634102, "grad_norm": 0.4755303917497374, "learning_rate": 0.00019971145749978294, "loss": 1.186, "step": 470 }, { "epoch": 0.12308888313034465, "grad_norm": 0.4587427439073901, "learning_rate": 0.00019967610397864493, "loss": 1.1772, "step": 475 }, { "epoch": 0.12438455558434827, "grad_norm": 0.43004396404922146, "learning_rate": 0.0001996387114421272, "loss": 1.1613, "step": 480 }, { "epoch": 0.1256802280383519, "grad_norm": 0.4457323824719563, "learning_rate": 0.0001995992806551468, "loss": 1.2043, "step": 485 }, { "epoch": 0.12697590049235552, "grad_norm": 0.46784030959656325, "learning_rate": 0.00019955781242431622, "loss": 1.1678, "step": 490 }, { "epoch": 0.12827157294635916, "grad_norm": 0.4610982688278206, "learning_rate": 0.00019951430759792654, "loss": 1.2353, "step": 495 }, { "epoch": 0.1295672454003628, "grad_norm": 0.4259792086854904, "learning_rate": 0.0001994687670659305, "loss": 1.1612, "step": 500 }, { "epoch": 0.1308629178543664, "grad_norm": 0.45520840895219594, "learning_rate": 0.00019942119175992383, "loss": 1.099, "step": 505 }, { "epoch": 0.13215859030837004, "grad_norm": 0.4473123797320602, "learning_rate": 0.00019937158265312667, "loss": 1.2141, "step": 510 }, { "epoch": 0.13345426276237368, "grad_norm": 0.4713319110562635, "learning_rate": 0.00019931994076036324, "loss": 1.1454, "step": 515 }, { "epoch": 0.1347499352163773, "grad_norm": 0.5136825790567691, "learning_rate": 0.00019926626713804137, "loss": 1.2187, "step": 520 }, { "epoch": 0.13604560767038093, "grad_norm": 0.4727966110942116, "learning_rate": 0.00019921056288413076, "loss": 1.1988, "step": 525 }, { "epoch": 0.13734128012438457, "grad_norm": 0.4364604236158943, "learning_rate": 0.00019915282913814052, "loss": 1.1685, "step": 530 }, { "epoch": 0.13863695257838818, "grad_norm": 0.48078362848188844, "learning_rate": 0.00019909306708109585, "loss": 1.1773, "step": 535 }, { "epoch": 0.13993262503239182, "grad_norm": 0.45186495929915804, "learning_rate": 0.00019903127793551408, "loss": 1.1409, "step": 540 }, { "epoch": 0.14122829748639543, "grad_norm": 0.4557721517395367, "learning_rate": 0.0001989674629653793, "loss": 1.1243, "step": 545 }, { "epoch": 0.14252396994039906, "grad_norm": 0.47938538143807347, "learning_rate": 0.00019890162347611687, "loss": 1.1644, "step": 550 }, { "epoch": 0.1438196423944027, "grad_norm": 0.4541147395460418, "learning_rate": 0.00019883376081456636, "loss": 1.2578, "step": 555 }, { "epoch": 0.1451153148484063, "grad_norm": 0.4780871852039313, "learning_rate": 0.00019876387636895437, "loss": 1.206, "step": 560 }, { "epoch": 0.14641098730240995, "grad_norm": 0.42628778420968705, "learning_rate": 0.00019869197156886586, "loss": 1.1728, "step": 565 }, { "epoch": 0.1477066597564136, "grad_norm": 0.4303201965277035, "learning_rate": 0.00019861804788521493, "loss": 1.1807, "step": 570 }, { "epoch": 0.1490023322104172, "grad_norm": 0.48120812526739415, "learning_rate": 0.00019854210683021485, "loss": 1.1825, "step": 575 }, { "epoch": 0.15029800466442084, "grad_norm": 0.479571534871543, "learning_rate": 0.0001984641499573472, "loss": 1.1213, "step": 580 }, { "epoch": 0.15159367711842447, "grad_norm": 0.4517330978325847, "learning_rate": 0.0001983841788613297, "loss": 1.1799, "step": 585 }, { "epoch": 0.15288934957242808, "grad_norm": 0.45662716170150547, "learning_rate": 0.00019830219517808404, "loss": 1.2154, "step": 590 }, { "epoch": 0.15418502202643172, "grad_norm": 0.46419741725194397, "learning_rate": 0.00019821820058470215, "loss": 1.2244, "step": 595 }, { "epoch": 0.15548069448043533, "grad_norm": 0.48643943236229664, "learning_rate": 0.00019813219679941203, "loss": 1.1773, "step": 600 }, { "epoch": 0.15677636693443897, "grad_norm": 0.4213189862656036, "learning_rate": 0.00019804418558154243, "loss": 1.1918, "step": 605 }, { "epoch": 0.1580720393884426, "grad_norm": 0.4039792149279076, "learning_rate": 0.00019795416873148703, "loss": 1.1937, "step": 610 }, { "epoch": 0.15936771184244622, "grad_norm": 0.42408763347068035, "learning_rate": 0.00019786214809066753, "loss": 1.1792, "step": 615 }, { "epoch": 0.16066338429644986, "grad_norm": 0.4418941392259797, "learning_rate": 0.00019776812554149603, "loss": 1.0983, "step": 620 }, { "epoch": 0.1619590567504535, "grad_norm": 0.44778137756258257, "learning_rate": 0.00019767210300733647, "loss": 1.1441, "step": 625 }, { "epoch": 0.1632547292044571, "grad_norm": 0.4214693702892699, "learning_rate": 0.0001975740824524653, "loss": 1.126, "step": 630 }, { "epoch": 0.16455040165846074, "grad_norm": 0.431868681760054, "learning_rate": 0.00019747406588203128, "loss": 1.2244, "step": 635 }, { "epoch": 0.16584607411246438, "grad_norm": 0.41521488211670027, "learning_rate": 0.00019737205534201455, "loss": 1.1443, "step": 640 }, { "epoch": 0.167141746566468, "grad_norm": 0.4141916874944073, "learning_rate": 0.00019726805291918464, "loss": 1.2012, "step": 645 }, { "epoch": 0.16843741902047163, "grad_norm": 0.8019720904500388, "learning_rate": 0.0001971620607410579, "loss": 1.1871, "step": 650 }, { "epoch": 0.16973309147447524, "grad_norm": 0.44302123796014203, "learning_rate": 0.00019705408097585393, "loss": 1.217, "step": 655 }, { "epoch": 0.17102876392847888, "grad_norm": 0.44266067140812565, "learning_rate": 0.0001969441158324512, "loss": 1.159, "step": 660 }, { "epoch": 0.17232443638248252, "grad_norm": 0.44647716533971304, "learning_rate": 0.0001968321675603419, "loss": 1.1653, "step": 665 }, { "epoch": 0.17362010883648613, "grad_norm": 0.44623301312377617, "learning_rate": 0.000196718238449586, "loss": 1.2078, "step": 670 }, { "epoch": 0.17491578129048976, "grad_norm": 0.41505140877535734, "learning_rate": 0.00019660233083076416, "loss": 1.1727, "step": 675 }, { "epoch": 0.1762114537444934, "grad_norm": 0.4603289406918021, "learning_rate": 0.00019648444707493035, "loss": 1.2151, "step": 680 }, { "epoch": 0.177507126198497, "grad_norm": 0.4185225256437822, "learning_rate": 0.00019636458959356316, "loss": 1.1708, "step": 685 }, { "epoch": 0.17880279865250065, "grad_norm": 0.4207300329188764, "learning_rate": 0.00019624276083851655, "loss": 1.1821, "step": 690 }, { "epoch": 0.1800984711065043, "grad_norm": 0.44173498052158916, "learning_rate": 0.00019611896330196956, "loss": 1.175, "step": 695 }, { "epoch": 0.1813941435605079, "grad_norm": 0.43648924778909787, "learning_rate": 0.0001959931995163756, "loss": 1.1409, "step": 700 }, { "epoch": 0.18268981601451154, "grad_norm": 0.41607673192874983, "learning_rate": 0.00019586547205441038, "loss": 1.1903, "step": 705 }, { "epoch": 0.18398548846851515, "grad_norm": 0.40413543802190666, "learning_rate": 0.00019573578352891937, "loss": 1.1395, "step": 710 }, { "epoch": 0.18528116092251878, "grad_norm": 0.45031653314971887, "learning_rate": 0.00019560413659286437, "loss": 1.1551, "step": 715 }, { "epoch": 0.18657683337652242, "grad_norm": 0.4308169448866955, "learning_rate": 0.00019547053393926934, "loss": 1.1875, "step": 720 }, { "epoch": 0.18787250583052603, "grad_norm": 0.42706739288180534, "learning_rate": 0.000195334978301165, "loss": 1.1828, "step": 725 }, { "epoch": 0.18916817828452967, "grad_norm": 0.43201723960587324, "learning_rate": 0.00019519747245153333, "loss": 1.1269, "step": 730 }, { "epoch": 0.1904638507385333, "grad_norm": 0.45113312626512714, "learning_rate": 0.0001950580192032505, "loss": 1.228, "step": 735 }, { "epoch": 0.19175952319253692, "grad_norm": 0.44161985380394386, "learning_rate": 0.0001949166214090295, "loss": 1.1242, "step": 740 }, { "epoch": 0.19305519564654056, "grad_norm": 0.46536800518090093, "learning_rate": 0.00019477328196136178, "loss": 1.205, "step": 745 }, { "epoch": 0.1943508681005442, "grad_norm": 0.4304870726759157, "learning_rate": 0.00019462800379245807, "loss": 1.1677, "step": 750 }, { "epoch": 0.1956465405545478, "grad_norm": 0.4500472206954455, "learning_rate": 0.0001944807898741883, "loss": 1.2031, "step": 755 }, { "epoch": 0.19694221300855144, "grad_norm": 0.4270521213996621, "learning_rate": 0.00019433164321802095, "loss": 1.221, "step": 760 }, { "epoch": 0.19823788546255505, "grad_norm": 0.4500335364742716, "learning_rate": 0.00019418056687496135, "loss": 1.2206, "step": 765 }, { "epoch": 0.1995335579165587, "grad_norm": 0.46760662103580936, "learning_rate": 0.00019402756393548936, "loss": 1.2628, "step": 770 }, { "epoch": 0.20082923037056233, "grad_norm": 0.4374227196524263, "learning_rate": 0.00019387263752949598, "loss": 1.1895, "step": 775 }, { "epoch": 0.20212490282456594, "grad_norm": 0.44423980739831237, "learning_rate": 0.00019371579082621952, "loss": 1.1915, "step": 780 }, { "epoch": 0.20342057527856958, "grad_norm": 0.5067585361087695, "learning_rate": 0.00019355702703418063, "loss": 1.1396, "step": 785 }, { "epoch": 0.20471624773257321, "grad_norm": 0.4397713592175906, "learning_rate": 0.0001933963494011168, "loss": 1.1829, "step": 790 }, { "epoch": 0.20601192018657682, "grad_norm": 0.41142670533283804, "learning_rate": 0.0001932337612139157, "loss": 1.1403, "step": 795 }, { "epoch": 0.20730759264058046, "grad_norm": 0.431969506585272, "learning_rate": 0.00019306926579854821, "loss": 1.1219, "step": 800 }, { "epoch": 0.2086032650945841, "grad_norm": 0.44265249198309475, "learning_rate": 0.00019290286652000018, "loss": 1.1897, "step": 805 }, { "epoch": 0.2098989375485877, "grad_norm": 0.4277657316516301, "learning_rate": 0.0001927345667822037, "loss": 1.1746, "step": 810 }, { "epoch": 0.21119461000259135, "grad_norm": 0.40123037504170683, "learning_rate": 0.00019256437002796744, "loss": 1.1775, "step": 815 }, { "epoch": 0.212490282456595, "grad_norm": 0.4250319228234625, "learning_rate": 0.00019239227973890622, "loss": 1.1357, "step": 820 }, { "epoch": 0.2137859549105986, "grad_norm": 0.469142216595277, "learning_rate": 0.0001922182994353697, "loss": 1.1709, "step": 825 }, { "epoch": 0.21508162736460223, "grad_norm": 0.4205186026441914, "learning_rate": 0.0001920424326763706, "loss": 1.1846, "step": 830 }, { "epoch": 0.21637729981860584, "grad_norm": 0.4364780340562057, "learning_rate": 0.00019186468305951165, "loss": 1.1611, "step": 835 }, { "epoch": 0.21767297227260948, "grad_norm": 0.42862545463693674, "learning_rate": 0.00019168505422091214, "loss": 1.2604, "step": 840 }, { "epoch": 0.21896864472661312, "grad_norm": 0.4298577312506422, "learning_rate": 0.00019150354983513346, "loss": 1.1584, "step": 845 }, { "epoch": 0.22026431718061673, "grad_norm": 0.4188255164345959, "learning_rate": 0.00019132017361510396, "loss": 1.1681, "step": 850 }, { "epoch": 0.22155998963462037, "grad_norm": 0.42610008321144127, "learning_rate": 0.00019113492931204304, "loss": 1.2502, "step": 855 }, { "epoch": 0.222855662088624, "grad_norm": 0.41917044829804134, "learning_rate": 0.00019094782071538434, "loss": 1.1441, "step": 860 }, { "epoch": 0.22415133454262762, "grad_norm": 0.41937866972534293, "learning_rate": 0.0001907588516526983, "loss": 1.2056, "step": 865 }, { "epoch": 0.22544700699663126, "grad_norm": 0.47221544899406714, "learning_rate": 0.00019056802598961376, "loss": 1.1754, "step": 870 }, { "epoch": 0.2267426794506349, "grad_norm": 0.45128209262811003, "learning_rate": 0.000190375347629739, "loss": 1.1963, "step": 875 }, { "epoch": 0.2280383519046385, "grad_norm": 0.41254887942116875, "learning_rate": 0.00019018082051458176, "loss": 1.1696, "step": 880 }, { "epoch": 0.22933402435864214, "grad_norm": 0.4336535335825639, "learning_rate": 0.00018998444862346873, "loss": 1.1942, "step": 885 }, { "epoch": 0.23062969681264575, "grad_norm": 0.4203068249880088, "learning_rate": 0.00018978623597346408, "loss": 1.1571, "step": 890 }, { "epoch": 0.2319253692666494, "grad_norm": 0.4504420384371143, "learning_rate": 0.00018958618661928732, "loss": 1.2145, "step": 895 }, { "epoch": 0.23322104172065303, "grad_norm": 1.2852982602777483, "learning_rate": 0.00018938430465323034, "loss": 1.1602, "step": 900 }, { "epoch": 0.23451671417465664, "grad_norm": 0.4456499671770626, "learning_rate": 0.0001891805942050736, "loss": 1.2208, "step": 905 }, { "epoch": 0.23581238662866028, "grad_norm": 0.4218488577208689, "learning_rate": 0.00018897505944200186, "loss": 1.1544, "step": 910 }, { "epoch": 0.2371080590826639, "grad_norm": 0.4509293570117875, "learning_rate": 0.00018876770456851877, "loss": 1.2167, "step": 915 }, { "epoch": 0.23840373153666752, "grad_norm": 0.4354264628463739, "learning_rate": 0.00018855853382636093, "loss": 1.2103, "step": 920 }, { "epoch": 0.23969940399067116, "grad_norm": 0.41330759489370955, "learning_rate": 0.00018834755149441104, "loss": 1.1713, "step": 925 }, { "epoch": 0.2409950764446748, "grad_norm": 0.43531782144147513, "learning_rate": 0.00018813476188861043, "loss": 1.1783, "step": 930 }, { "epoch": 0.2422907488986784, "grad_norm": 0.4260507531562403, "learning_rate": 0.00018792016936187086, "loss": 1.1473, "step": 935 }, { "epoch": 0.24358642135268205, "grad_norm": 0.46161844432613697, "learning_rate": 0.00018770377830398525, "loss": 1.1719, "step": 940 }, { "epoch": 0.24488209380668566, "grad_norm": 0.44108410378556556, "learning_rate": 0.00018748559314153818, "loss": 1.2101, "step": 945 }, { "epoch": 0.2461777662606893, "grad_norm": 0.43144293315888466, "learning_rate": 0.00018726561833781497, "loss": 1.1933, "step": 950 }, { "epoch": 0.24747343871469293, "grad_norm": 0.437185606075459, "learning_rate": 0.00018704385839271074, "loss": 1.1563, "step": 955 }, { "epoch": 0.24876911116869654, "grad_norm": 0.44187547375302677, "learning_rate": 0.00018682031784263814, "loss": 1.1219, "step": 960 }, { "epoch": 0.2500647836227002, "grad_norm": 0.4286003185934626, "learning_rate": 0.00018659500126043456, "loss": 1.1347, "step": 965 }, { "epoch": 0.2513604560767038, "grad_norm": 0.5193197943854574, "learning_rate": 0.00018636791325526872, "loss": 1.2118, "step": 970 }, { "epoch": 0.25265612853070746, "grad_norm": 0.4243801823353042, "learning_rate": 0.00018613905847254624, "loss": 1.2015, "step": 975 }, { "epoch": 0.25395180098471104, "grad_norm": 0.4017619629276402, "learning_rate": 0.0001859084415938147, "loss": 1.1405, "step": 980 }, { "epoch": 0.2552474734387147, "grad_norm": 0.46780432324474447, "learning_rate": 0.00018567606733666775, "loss": 1.1803, "step": 985 }, { "epoch": 0.2565431458927183, "grad_norm": 0.41305152773119796, "learning_rate": 0.00018544194045464886, "loss": 1.138, "step": 990 }, { "epoch": 0.25783881834672195, "grad_norm": 0.4090708833399176, "learning_rate": 0.0001852060657371538, "loss": 1.2036, "step": 995 }, { "epoch": 0.2591344908007256, "grad_norm": 0.43431398541973126, "learning_rate": 0.00018496844800933277, "loss": 1.1328, "step": 1000 }, { "epoch": 0.26043016325472923, "grad_norm": 0.4473287652132595, "learning_rate": 0.0001847290921319918, "loss": 1.183, "step": 1005 }, { "epoch": 0.2617258357087328, "grad_norm": 0.42139497138720455, "learning_rate": 0.00018448800300149314, "loss": 1.1885, "step": 1010 }, { "epoch": 0.26302150816273645, "grad_norm": 0.4192742651155342, "learning_rate": 0.00018424518554965516, "loss": 1.1673, "step": 1015 }, { "epoch": 0.2643171806167401, "grad_norm": 0.4276967282946324, "learning_rate": 0.00018400064474365156, "loss": 1.1442, "step": 1020 }, { "epoch": 0.2656128530707437, "grad_norm": 0.4279959695376702, "learning_rate": 0.00018375438558590967, "loss": 1.1713, "step": 1025 }, { "epoch": 0.26690852552474736, "grad_norm": 0.4138904922875282, "learning_rate": 0.00018350641311400812, "loss": 1.1294, "step": 1030 }, { "epoch": 0.26820419797875095, "grad_norm": 0.43337813322172525, "learning_rate": 0.0001832567324005737, "loss": 1.1421, "step": 1035 }, { "epoch": 0.2694998704327546, "grad_norm": 0.4418444877860619, "learning_rate": 0.00018300534855317783, "loss": 1.1775, "step": 1040 }, { "epoch": 0.2707955428867582, "grad_norm": 0.46191806443855654, "learning_rate": 0.00018275226671423195, "loss": 1.1188, "step": 1045 }, { "epoch": 0.27209121534076186, "grad_norm": 0.4449373885557454, "learning_rate": 0.0001824974920608821, "loss": 1.1445, "step": 1050 }, { "epoch": 0.2733868877947655, "grad_norm": 0.4138240087897747, "learning_rate": 0.0001822410298049035, "loss": 1.1403, "step": 1055 }, { "epoch": 0.27468256024876914, "grad_norm": 0.4410582459345502, "learning_rate": 0.00018198288519259353, "loss": 1.1835, "step": 1060 }, { "epoch": 0.2759782327027727, "grad_norm": 0.4419431606975578, "learning_rate": 0.0001817230635046645, "loss": 1.2308, "step": 1065 }, { "epoch": 0.27727390515677636, "grad_norm": 0.41612794192020086, "learning_rate": 0.0001814615700561358, "loss": 1.21, "step": 1070 }, { "epoch": 0.27856957761078, "grad_norm": 0.4728861496942093, "learning_rate": 0.00018119841019622487, "loss": 1.1687, "step": 1075 }, { "epoch": 0.27986525006478363, "grad_norm": 0.39483035476368555, "learning_rate": 0.0001809335893082381, "loss": 1.2031, "step": 1080 }, { "epoch": 0.28116092251878727, "grad_norm": 0.40167341751112157, "learning_rate": 0.0001806671128094605, "loss": 1.1495, "step": 1085 }, { "epoch": 0.28245659497279085, "grad_norm": 0.42702634032994197, "learning_rate": 0.0001803989861510449, "loss": 1.1885, "step": 1090 }, { "epoch": 0.2837522674267945, "grad_norm": 0.3970503569404702, "learning_rate": 0.00018012921481790054, "loss": 1.1608, "step": 1095 }, { "epoch": 0.28504793988079813, "grad_norm": 0.42637878811973734, "learning_rate": 0.0001798578043285807, "loss": 1.1386, "step": 1100 }, { "epoch": 0.28634361233480177, "grad_norm": 0.39626715470331225, "learning_rate": 0.00017958476023517008, "loss": 1.1355, "step": 1105 }, { "epoch": 0.2876392847888054, "grad_norm": 0.4158526138591002, "learning_rate": 0.00017931008812317089, "loss": 1.1844, "step": 1110 }, { "epoch": 0.28893495724280904, "grad_norm": 0.4110580138866959, "learning_rate": 0.00017903379361138884, "loss": 1.179, "step": 1115 }, { "epoch": 0.2902306296968126, "grad_norm": 0.39558148991812647, "learning_rate": 0.0001787558823518181, "loss": 1.1529, "step": 1120 }, { "epoch": 0.29152630215081626, "grad_norm": 0.43235047146868255, "learning_rate": 0.0001784763600295257, "loss": 1.1253, "step": 1125 }, { "epoch": 0.2928219746048199, "grad_norm": 0.4029217795192826, "learning_rate": 0.00017819523236253524, "loss": 1.1775, "step": 1130 }, { "epoch": 0.29411764705882354, "grad_norm": 0.43737260937528, "learning_rate": 0.0001779125051017099, "loss": 1.1658, "step": 1135 }, { "epoch": 0.2954133195128272, "grad_norm": 0.4223589589109796, "learning_rate": 0.00017762818403063485, "loss": 1.2193, "step": 1140 }, { "epoch": 0.29670899196683076, "grad_norm": 0.42153072570772854, "learning_rate": 0.0001773422749654988, "loss": 1.1381, "step": 1145 }, { "epoch": 0.2980046644208344, "grad_norm": 0.44333235395249593, "learning_rate": 0.0001770547837549752, "loss": 1.2025, "step": 1150 }, { "epoch": 0.29930033687483804, "grad_norm": 0.390430213600016, "learning_rate": 0.0001767657162801025, "loss": 1.2179, "step": 1155 }, { "epoch": 0.3005960093288417, "grad_norm": 0.39596734484692325, "learning_rate": 0.00017647507845416392, "loss": 1.1443, "step": 1160 }, { "epoch": 0.3018916817828453, "grad_norm": 0.3964805626892672, "learning_rate": 0.00017618287622256625, "loss": 1.0953, "step": 1165 }, { "epoch": 0.30318735423684895, "grad_norm": 0.397165519182473, "learning_rate": 0.00017588911556271858, "loss": 1.2468, "step": 1170 }, { "epoch": 0.30448302669085253, "grad_norm": 0.40738383895968944, "learning_rate": 0.00017559380248390982, "loss": 1.2279, "step": 1175 }, { "epoch": 0.30577869914485617, "grad_norm": 0.4322556749554571, "learning_rate": 0.00017529694302718574, "loss": 1.1614, "step": 1180 }, { "epoch": 0.3070743715988598, "grad_norm": 0.3795921376766387, "learning_rate": 0.0001749985432652255, "loss": 1.1514, "step": 1185 }, { "epoch": 0.30837004405286345, "grad_norm": 0.3984980575708157, "learning_rate": 0.00017469860930221734, "loss": 1.2262, "step": 1190 }, { "epoch": 0.3096657165068671, "grad_norm": 0.4239852292999169, "learning_rate": 0.00017439714727373378, "loss": 1.1279, "step": 1195 }, { "epoch": 0.31096138896087067, "grad_norm": 0.4243568518066299, "learning_rate": 0.00017409416334660606, "loss": 1.1482, "step": 1200 }, { "epoch": 0.3122570614148743, "grad_norm": 0.39261780158026205, "learning_rate": 0.00017378966371879803, "loss": 1.1245, "step": 1205 }, { "epoch": 0.31355273386887794, "grad_norm": 0.4129200331115499, "learning_rate": 0.00017348365461927932, "loss": 1.164, "step": 1210 }, { "epoch": 0.3148484063228816, "grad_norm": 0.46767017383779913, "learning_rate": 0.00017317614230789792, "loss": 1.2038, "step": 1215 }, { "epoch": 0.3161440787768852, "grad_norm": 0.4636682224687034, "learning_rate": 0.00017286713307525212, "loss": 1.1798, "step": 1220 }, { "epoch": 0.31743975123088886, "grad_norm": 0.464530514696448, "learning_rate": 0.00017255663324256194, "loss": 1.1569, "step": 1225 }, { "epoch": 0.31873542368489244, "grad_norm": 0.49392901042147025, "learning_rate": 0.00017224464916153963, "loss": 1.1722, "step": 1230 }, { "epoch": 0.3200310961388961, "grad_norm": 0.437281931716071, "learning_rate": 0.00017193118721425986, "loss": 1.1472, "step": 1235 }, { "epoch": 0.3213267685928997, "grad_norm": 0.49973298936543603, "learning_rate": 0.00017161625381302914, "loss": 1.1662, "step": 1240 }, { "epoch": 0.32262244104690335, "grad_norm": 0.9030290188031315, "learning_rate": 0.00017129985540025473, "loss": 1.1751, "step": 1245 }, { "epoch": 0.323918113500907, "grad_norm": 0.46932404302432773, "learning_rate": 0.00017098199844831262, "loss": 1.1753, "step": 1250 }, { "epoch": 0.3252137859549106, "grad_norm": 0.43737397519531773, "learning_rate": 0.0001706626894594154, "loss": 1.1539, "step": 1255 }, { "epoch": 0.3265094584089142, "grad_norm": 0.47678251178263875, "learning_rate": 0.00017034193496547902, "loss": 1.1572, "step": 1260 }, { "epoch": 0.32780513086291785, "grad_norm": 0.48005221081289434, "learning_rate": 0.00017001974152798942, "loss": 1.2371, "step": 1265 }, { "epoch": 0.3291008033169215, "grad_norm": 0.4005197274939468, "learning_rate": 0.000169696115737868, "loss": 1.1375, "step": 1270 }, { "epoch": 0.3303964757709251, "grad_norm": 0.520470386619454, "learning_rate": 0.00016937106421533707, "loss": 1.2201, "step": 1275 }, { "epoch": 0.33169214822492876, "grad_norm": 0.4024597397089223, "learning_rate": 0.00016904459360978427, "loss": 1.16, "step": 1280 }, { "epoch": 0.33298782067893234, "grad_norm": 0.43690159855763916, "learning_rate": 0.00016871671059962655, "loss": 1.2303, "step": 1285 }, { "epoch": 0.334283493132936, "grad_norm": 0.42841787432183803, "learning_rate": 0.00016838742189217366, "loss": 1.1601, "step": 1290 }, { "epoch": 0.3355791655869396, "grad_norm": 0.41150940120312113, "learning_rate": 0.00016805673422349082, "loss": 1.1542, "step": 1295 }, { "epoch": 0.33687483804094326, "grad_norm": 0.4108846561673421, "learning_rate": 0.000167724654358261, "loss": 1.1497, "step": 1300 }, { "epoch": 0.3381705104949469, "grad_norm": 0.4485077675503156, "learning_rate": 0.00016739118908964647, "loss": 1.1979, "step": 1305 }, { "epoch": 0.3394661829489505, "grad_norm": 0.40380367544471263, "learning_rate": 0.00016705634523915, "loss": 1.156, "step": 1310 }, { "epoch": 0.3407618554029541, "grad_norm": 0.4315676369718417, "learning_rate": 0.000166720129656475, "loss": 1.1876, "step": 1315 }, { "epoch": 0.34205752785695775, "grad_norm": 0.42298698967693915, "learning_rate": 0.00016638254921938587, "loss": 1.226, "step": 1320 }, { "epoch": 0.3433532003109614, "grad_norm": 0.40711175836921143, "learning_rate": 0.00016604361083356675, "loss": 1.1865, "step": 1325 }, { "epoch": 0.34464887276496503, "grad_norm": 0.4209373436004817, "learning_rate": 0.0001657033214324807, "loss": 1.1485, "step": 1330 }, { "epoch": 0.34594454521896867, "grad_norm": 0.3973628247221751, "learning_rate": 0.0001653616879772277, "loss": 1.1419, "step": 1335 }, { "epoch": 0.34724021767297225, "grad_norm": 0.40714636825697015, "learning_rate": 0.00016501871745640213, "loss": 1.1607, "step": 1340 }, { "epoch": 0.3485358901269759, "grad_norm": 0.4091289111560899, "learning_rate": 0.00016467441688595015, "loss": 1.169, "step": 1345 }, { "epoch": 0.3498315625809795, "grad_norm": 0.39294191213880614, "learning_rate": 0.0001643287933090258, "loss": 1.1856, "step": 1350 }, { "epoch": 0.35112723503498317, "grad_norm": 0.4153756492643378, "learning_rate": 0.00016398185379584707, "loss": 1.1601, "step": 1355 }, { "epoch": 0.3524229074889868, "grad_norm": 0.39772148869790713, "learning_rate": 0.0001636336054435514, "loss": 1.1402, "step": 1360 }, { "epoch": 0.3537185799429904, "grad_norm": 0.4071165220940827, "learning_rate": 0.00016328405537605032, "loss": 1.1333, "step": 1365 }, { "epoch": 0.355014252396994, "grad_norm": 0.4269465126269822, "learning_rate": 0.00016293321074388375, "loss": 1.1948, "step": 1370 }, { "epoch": 0.35630992485099766, "grad_norm": 0.41865801202656294, "learning_rate": 0.00016258107872407375, "loss": 1.1465, "step": 1375 }, { "epoch": 0.3576055973050013, "grad_norm": 0.40838786927024157, "learning_rate": 0.00016222766651997789, "loss": 1.1695, "step": 1380 }, { "epoch": 0.35890126975900494, "grad_norm": 0.4096176785342024, "learning_rate": 0.0001618729813611414, "loss": 1.1447, "step": 1385 }, { "epoch": 0.3601969422130086, "grad_norm": 0.41309116576243843, "learning_rate": 0.00016151703050314986, "loss": 1.1804, "step": 1390 }, { "epoch": 0.36149261466701216, "grad_norm": 0.4053926841703832, "learning_rate": 0.00016115982122748043, "loss": 1.1471, "step": 1395 }, { "epoch": 0.3627882871210158, "grad_norm": 0.4058497805268484, "learning_rate": 0.00016080136084135297, "loss": 1.1494, "step": 1400 }, { "epoch": 0.36408395957501943, "grad_norm": 0.4216552491807955, "learning_rate": 0.00016044165667758055, "loss": 1.1928, "step": 1405 }, { "epoch": 0.36537963202902307, "grad_norm": 0.4336798036895343, "learning_rate": 0.0001600807160944195, "loss": 1.19, "step": 1410 }, { "epoch": 0.3666753044830267, "grad_norm": 0.42380600181334993, "learning_rate": 0.00015971854647541884, "loss": 1.1674, "step": 1415 }, { "epoch": 0.3679709769370303, "grad_norm": 0.4116489897844781, "learning_rate": 0.00015935515522926927, "loss": 1.1407, "step": 1420 }, { "epoch": 0.36926664939103393, "grad_norm": 0.4051816422975703, "learning_rate": 0.00015899054978965157, "loss": 1.1861, "step": 1425 }, { "epoch": 0.37056232184503757, "grad_norm": 0.4374412539295244, "learning_rate": 0.0001586247376150846, "loss": 1.2273, "step": 1430 }, { "epoch": 0.3718579942990412, "grad_norm": 0.4225494310104097, "learning_rate": 0.00015825772618877263, "loss": 1.2218, "step": 1435 }, { "epoch": 0.37315366675304484, "grad_norm": 0.4202018672382752, "learning_rate": 0.00015788952301845237, "loss": 1.1155, "step": 1440 }, { "epoch": 0.3744493392070485, "grad_norm": 0.4313070413558118, "learning_rate": 0.0001575201356362393, "loss": 1.1551, "step": 1445 }, { "epoch": 0.37574501166105206, "grad_norm": 0.40304574068983823, "learning_rate": 0.00015714957159847367, "loss": 1.1491, "step": 1450 }, { "epoch": 0.3770406841150557, "grad_norm": 0.4413697670634288, "learning_rate": 0.00015677783848556576, "loss": 1.1631, "step": 1455 }, { "epoch": 0.37833635656905934, "grad_norm": 0.4021843049770175, "learning_rate": 0.00015640494390184112, "loss": 1.1304, "step": 1460 }, { "epoch": 0.379632029023063, "grad_norm": 0.41280223940193794, "learning_rate": 0.0001560308954753847, "loss": 1.1458, "step": 1465 }, { "epoch": 0.3809277014770666, "grad_norm": 0.44403414056325813, "learning_rate": 0.00015565570085788495, "loss": 1.2007, "step": 1470 }, { "epoch": 0.3822233739310702, "grad_norm": 0.42327265730452945, "learning_rate": 0.00015527936772447725, "loss": 1.1168, "step": 1475 }, { "epoch": 0.38351904638507384, "grad_norm": 0.41872242471581905, "learning_rate": 0.00015490190377358704, "loss": 1.1551, "step": 1480 }, { "epoch": 0.3848147188390775, "grad_norm": 0.4076213102610575, "learning_rate": 0.00015452331672677206, "loss": 1.0902, "step": 1485 }, { "epoch": 0.3861103912930811, "grad_norm": 0.39807246728274887, "learning_rate": 0.00015414361432856475, "loss": 1.1598, "step": 1490 }, { "epoch": 0.38740606374708475, "grad_norm": 0.40854063250285516, "learning_rate": 0.00015376280434631345, "loss": 1.1806, "step": 1495 }, { "epoch": 0.3887017362010884, "grad_norm": 0.40837052057355316, "learning_rate": 0.00015338089457002382, "loss": 1.0829, "step": 1500 }, { "epoch": 0.38999740865509197, "grad_norm": 0.4123609619777807, "learning_rate": 0.00015299789281219935, "loss": 1.1688, "step": 1505 }, { "epoch": 0.3912930811090956, "grad_norm": 0.4316199953541538, "learning_rate": 0.00015261380690768144, "loss": 1.1543, "step": 1510 }, { "epoch": 0.39258875356309925, "grad_norm": 0.4398633371362292, "learning_rate": 0.00015222864471348943, "loss": 1.1724, "step": 1515 }, { "epoch": 0.3938844260171029, "grad_norm": 0.4081954480961544, "learning_rate": 0.00015184241410865954, "loss": 1.1269, "step": 1520 }, { "epoch": 0.3951800984711065, "grad_norm": 0.3994337234690745, "learning_rate": 0.00015145512299408388, "loss": 1.1598, "step": 1525 }, { "epoch": 0.3964757709251101, "grad_norm": 0.4415117779921472, "learning_rate": 0.00015106677929234877, "loss": 1.0969, "step": 1530 }, { "epoch": 0.39777144337911374, "grad_norm": 0.4033267536569039, "learning_rate": 0.0001506773909475727, "loss": 1.1955, "step": 1535 }, { "epoch": 0.3990671158331174, "grad_norm": 0.39024517506575723, "learning_rate": 0.00015028696592524386, "loss": 1.1562, "step": 1540 }, { "epoch": 0.400362788287121, "grad_norm": 0.4359801226192451, "learning_rate": 0.000149895512212057, "loss": 1.1695, "step": 1545 }, { "epoch": 0.40165846074112466, "grad_norm": 0.3996337334749442, "learning_rate": 0.00014950303781575034, "loss": 1.0982, "step": 1550 }, { "epoch": 0.4029541331951283, "grad_norm": 0.39816310192557475, "learning_rate": 0.00014910955076494152, "loss": 1.1223, "step": 1555 }, { "epoch": 0.4042498056491319, "grad_norm": 0.37894092829518466, "learning_rate": 0.00014871505910896352, "loss": 1.1217, "step": 1560 }, { "epoch": 0.4055454781031355, "grad_norm": 0.4319425134415384, "learning_rate": 0.0001483195709176999, "loss": 1.1966, "step": 1565 }, { "epoch": 0.40684115055713915, "grad_norm": 0.3996173231205214, "learning_rate": 0.00014792309428141978, "loss": 1.1039, "step": 1570 }, { "epoch": 0.4081368230111428, "grad_norm": 0.40718491453474887, "learning_rate": 0.0001475256373106123, "loss": 1.1725, "step": 1575 }, { "epoch": 0.40943249546514643, "grad_norm": 0.4286665586760103, "learning_rate": 0.00014712720813582066, "loss": 1.1443, "step": 1580 }, { "epoch": 0.41072816791915, "grad_norm": 0.4033525905918133, "learning_rate": 0.00014672781490747606, "loss": 1.1742, "step": 1585 }, { "epoch": 0.41202384037315365, "grad_norm": 0.40558326017198004, "learning_rate": 0.00014632746579573052, "loss": 1.2117, "step": 1590 }, { "epoch": 0.4133195128271573, "grad_norm": 0.4056524890982348, "learning_rate": 0.0001459261689902902, "loss": 1.2034, "step": 1595 }, { "epoch": 0.4146151852811609, "grad_norm": 0.44100980335173223, "learning_rate": 0.00014552393270024765, "loss": 1.1544, "step": 1600 }, { "epoch": 0.41591085773516456, "grad_norm": 0.39891755340287666, "learning_rate": 0.00014512076515391375, "loss": 1.1256, "step": 1605 }, { "epoch": 0.4172065301891682, "grad_norm": 0.40682040920269774, "learning_rate": 0.00014471667459864973, "loss": 1.1499, "step": 1610 }, { "epoch": 0.4185022026431718, "grad_norm": 0.44312416626229856, "learning_rate": 0.00014431166930069816, "loss": 1.0977, "step": 1615 }, { "epoch": 0.4197978750971754, "grad_norm": 0.4301392634897964, "learning_rate": 0.00014390575754501402, "loss": 1.139, "step": 1620 }, { "epoch": 0.42109354755117906, "grad_norm": 0.418724699659618, "learning_rate": 0.0001434989476350951, "loss": 1.1592, "step": 1625 }, { "epoch": 0.4223892200051827, "grad_norm": 0.4134020107191313, "learning_rate": 0.00014309124789281226, "loss": 1.2104, "step": 1630 }, { "epoch": 0.42368489245918634, "grad_norm": 0.39580446880805387, "learning_rate": 0.00014268266665823912, "loss": 1.1757, "step": 1635 }, { "epoch": 0.42498056491319, "grad_norm": 0.41472381643526246, "learning_rate": 0.00014227321228948146, "loss": 1.1322, "step": 1640 }, { "epoch": 0.42627623736719356, "grad_norm": 0.39275693329437805, "learning_rate": 0.0001418628931625062, "loss": 1.1742, "step": 1645 }, { "epoch": 0.4275719098211972, "grad_norm": 0.38619824420216714, "learning_rate": 0.00014145171767097027, "loss": 1.1511, "step": 1650 }, { "epoch": 0.42886758227520083, "grad_norm": 0.4254507618246464, "learning_rate": 0.00014103969422604856, "loss": 1.1231, "step": 1655 }, { "epoch": 0.43016325472920447, "grad_norm": 0.40459401347151125, "learning_rate": 0.00014062683125626218, "loss": 1.1384, "step": 1660 }, { "epoch": 0.4314589271832081, "grad_norm": 0.409140169024848, "learning_rate": 0.0001402131372073058, "loss": 1.191, "step": 1665 }, { "epoch": 0.4327545996372117, "grad_norm": 0.389695812394897, "learning_rate": 0.00013979862054187505, "loss": 1.1543, "step": 1670 }, { "epoch": 0.43405027209121533, "grad_norm": 0.42112717245729503, "learning_rate": 0.00013938328973949336, "loss": 1.1715, "step": 1675 }, { "epoch": 0.43534594454521897, "grad_norm": 0.3864671779444786, "learning_rate": 0.0001389671532963384, "loss": 1.1508, "step": 1680 }, { "epoch": 0.4366416169992226, "grad_norm": 0.39052429534303434, "learning_rate": 0.00013855021972506844, "loss": 1.1054, "step": 1685 }, { "epoch": 0.43793728945322624, "grad_norm": 0.404117930960366, "learning_rate": 0.0001381324975546481, "loss": 1.1623, "step": 1690 }, { "epoch": 0.4392329619072299, "grad_norm": 0.45134279442398223, "learning_rate": 0.0001377139953301739, "loss": 1.1301, "step": 1695 }, { "epoch": 0.44052863436123346, "grad_norm": 0.42206784378450607, "learning_rate": 0.00013729472161269946, "loss": 1.18, "step": 1700 }, { "epoch": 0.4418243068152371, "grad_norm": 0.43048779695832334, "learning_rate": 0.00013687468497906044, "loss": 1.1413, "step": 1705 }, { "epoch": 0.44311997926924074, "grad_norm": 0.3902243379067639, "learning_rate": 0.00013645389402169893, "loss": 1.1473, "step": 1710 }, { "epoch": 0.4444156517232444, "grad_norm": 0.3790071383777115, "learning_rate": 0.00013603235734848784, "loss": 1.2114, "step": 1715 }, { "epoch": 0.445711324177248, "grad_norm": 0.39998613746463574, "learning_rate": 0.00013561008358255468, "loss": 1.1193, "step": 1720 }, { "epoch": 0.4470069966312516, "grad_norm": 0.4136780934178077, "learning_rate": 0.0001351870813621054, "loss": 1.1417, "step": 1725 }, { "epoch": 0.44830266908525523, "grad_norm": 0.39313685685153266, "learning_rate": 0.00013476335934024735, "loss": 1.1437, "step": 1730 }, { "epoch": 0.44959834153925887, "grad_norm": 0.41028501570471904, "learning_rate": 0.00013433892618481248, "loss": 1.1032, "step": 1735 }, { "epoch": 0.4508940139932625, "grad_norm": 0.43704478597398877, "learning_rate": 0.00013391379057817995, "loss": 1.1933, "step": 1740 }, { "epoch": 0.45218968644726615, "grad_norm": 0.37937862506331654, "learning_rate": 0.00013348796121709862, "loss": 1.187, "step": 1745 }, { "epoch": 0.4534853589012698, "grad_norm": 0.40181873506259413, "learning_rate": 0.00013306144681250908, "loss": 1.1625, "step": 1750 }, { "epoch": 0.45478103135527337, "grad_norm": 0.4117326535807377, "learning_rate": 0.00013263425608936536, "loss": 1.1875, "step": 1755 }, { "epoch": 0.456076703809277, "grad_norm": 0.39234377730019654, "learning_rate": 0.00013220639778645663, "loss": 1.1888, "step": 1760 }, { "epoch": 0.45737237626328064, "grad_norm": 0.42622878050330665, "learning_rate": 0.0001317778806562283, "loss": 1.151, "step": 1765 }, { "epoch": 0.4586680487172843, "grad_norm": 0.3997439152224071, "learning_rate": 0.000131348713464603, "loss": 1.1271, "step": 1770 }, { "epoch": 0.4599637211712879, "grad_norm": 0.4151543746885238, "learning_rate": 0.0001309189049908014, "loss": 1.1745, "step": 1775 }, { "epoch": 0.4612593936252915, "grad_norm": 0.4136123970052988, "learning_rate": 0.00013048846402716237, "loss": 1.1446, "step": 1780 }, { "epoch": 0.46255506607929514, "grad_norm": 0.3925741414110596, "learning_rate": 0.0001300573993789633, "loss": 1.1186, "step": 1785 }, { "epoch": 0.4638507385332988, "grad_norm": 0.43074300043773284, "learning_rate": 0.00012962571986423993, "loss": 1.2004, "step": 1790 }, { "epoch": 0.4651464109873024, "grad_norm": 0.3948645456920277, "learning_rate": 0.00012919343431360596, "loss": 1.1534, "step": 1795 }, { "epoch": 0.46644208344130605, "grad_norm": 0.4423236206923003, "learning_rate": 0.00012876055157007242, "loss": 1.1509, "step": 1800 }, { "epoch": 0.4677377558953097, "grad_norm": 0.39371508060725335, "learning_rate": 0.00012832708048886679, "loss": 1.1941, "step": 1805 }, { "epoch": 0.4690334283493133, "grad_norm": 0.4156050706970669, "learning_rate": 0.00012789302993725175, "loss": 1.2233, "step": 1810 }, { "epoch": 0.4703291008033169, "grad_norm": 0.3844699731827056, "learning_rate": 0.0001274584087943439, "loss": 1.172, "step": 1815 }, { "epoch": 0.47162477325732055, "grad_norm": 0.392439037665497, "learning_rate": 0.00012702322595093212, "loss": 1.1935, "step": 1820 }, { "epoch": 0.4729204457113242, "grad_norm": 0.42109447627306007, "learning_rate": 0.00012658749030929566, "loss": 1.0821, "step": 1825 }, { "epoch": 0.4742161181653278, "grad_norm": 0.3985407382924968, "learning_rate": 0.00012615121078302202, "loss": 1.1564, "step": 1830 }, { "epoch": 0.4755117906193314, "grad_norm": 0.40190944372077425, "learning_rate": 0.0001257143962968246, "loss": 1.1858, "step": 1835 }, { "epoch": 0.47680746307333505, "grad_norm": 0.45076787480964914, "learning_rate": 0.00012527705578636023, "loss": 1.1514, "step": 1840 }, { "epoch": 0.4781031355273387, "grad_norm": 0.4363861711704788, "learning_rate": 0.0001248391981980462, "loss": 1.133, "step": 1845 }, { "epoch": 0.4793988079813423, "grad_norm": 0.43776230767541446, "learning_rate": 0.00012440083248887754, "loss": 1.2082, "step": 1850 }, { "epoch": 0.48069448043534596, "grad_norm": 0.4127859770012284, "learning_rate": 0.00012396196762624341, "loss": 1.1613, "step": 1855 }, { "epoch": 0.4819901528893496, "grad_norm": 0.3940447846474964, "learning_rate": 0.00012352261258774395, "loss": 1.15, "step": 1860 }, { "epoch": 0.4832858253433532, "grad_norm": 0.4193523558931992, "learning_rate": 0.0001230827763610066, "loss": 1.1382, "step": 1865 }, { "epoch": 0.4845814977973568, "grad_norm": 0.395557633792394, "learning_rate": 0.00012264246794350202, "loss": 1.1678, "step": 1870 }, { "epoch": 0.48587717025136046, "grad_norm": 0.40423444904954847, "learning_rate": 0.00012220169634236038, "loss": 1.157, "step": 1875 }, { "epoch": 0.4871728427053641, "grad_norm": 0.3870746898688869, "learning_rate": 0.00012176047057418682, "loss": 1.1439, "step": 1880 }, { "epoch": 0.48846851515936773, "grad_norm": 0.4458893388021195, "learning_rate": 0.00012131879966487709, "loss": 1.1513, "step": 1885 }, { "epoch": 0.4897641876133713, "grad_norm": 0.38478400725184214, "learning_rate": 0.00012087669264943302, "loss": 1.1333, "step": 1890 }, { "epoch": 0.49105986006737495, "grad_norm": 0.37819480972862285, "learning_rate": 0.00012043415857177751, "loss": 1.1663, "step": 1895 }, { "epoch": 0.4923555325213786, "grad_norm": 0.4125210415708117, "learning_rate": 0.00011999120648456974, "loss": 1.1457, "step": 1900 }, { "epoch": 0.49365120497538223, "grad_norm": 0.3869780886696955, "learning_rate": 0.00011954784544901971, "loss": 1.1536, "step": 1905 }, { "epoch": 0.49494687742938587, "grad_norm": 0.40907576777936727, "learning_rate": 0.00011910408453470316, "loss": 1.1361, "step": 1910 }, { "epoch": 0.4962425498833895, "grad_norm": 0.3737647497815572, "learning_rate": 0.00011865993281937589, "loss": 1.1006, "step": 1915 }, { "epoch": 0.4975382223373931, "grad_norm": 0.41125742423172207, "learning_rate": 0.00011821539938878801, "loss": 1.1641, "step": 1920 }, { "epoch": 0.4988338947913967, "grad_norm": 0.38132649216646763, "learning_rate": 0.00011777049333649826, "loss": 1.1525, "step": 1925 }, { "epoch": 0.5001295672454004, "grad_norm": 0.38834910320923277, "learning_rate": 0.00011732522376368781, "loss": 1.1531, "step": 1930 }, { "epoch": 0.501425239699404, "grad_norm": 0.40314094443067383, "learning_rate": 0.0001168795997789742, "loss": 1.1592, "step": 1935 }, { "epoch": 0.5027209121534076, "grad_norm": 0.4040865102971838, "learning_rate": 0.00011643363049822496, "loss": 1.2547, "step": 1940 }, { "epoch": 0.5040165846074113, "grad_norm": 0.38562943372958325, "learning_rate": 0.00011598732504437107, "loss": 1.1109, "step": 1945 }, { "epoch": 0.5053122570614149, "grad_norm": 0.4250211446665384, "learning_rate": 0.00011554069254722051, "loss": 1.1941, "step": 1950 }, { "epoch": 0.5066079295154186, "grad_norm": 0.4032090808510551, "learning_rate": 0.00011509374214327131, "loss": 1.1572, "step": 1955 }, { "epoch": 0.5079036019694221, "grad_norm": 0.40575976251729745, "learning_rate": 0.00011464648297552478, "loss": 1.1657, "step": 1960 }, { "epoch": 0.5091992744234257, "grad_norm": 0.41052107415559824, "learning_rate": 0.00011419892419329844, "loss": 1.1642, "step": 1965 }, { "epoch": 0.5104949468774294, "grad_norm": 0.3958628275091308, "learning_rate": 0.00011375107495203873, "loss": 1.2116, "step": 1970 }, { "epoch": 0.511790619331433, "grad_norm": 0.38478613819779206, "learning_rate": 0.00011330294441313402, "loss": 1.1451, "step": 1975 }, { "epoch": 0.5130862917854366, "grad_norm": 0.43128674114012594, "learning_rate": 0.00011285454174372692, "loss": 1.1494, "step": 1980 }, { "epoch": 0.5143819642394403, "grad_norm": 0.4424953381805022, "learning_rate": 0.0001124058761165268, "loss": 1.0932, "step": 1985 }, { "epoch": 0.5156776366934439, "grad_norm": 0.4063559419481111, "learning_rate": 0.00011195695670962234, "loss": 1.1137, "step": 1990 }, { "epoch": 0.5169733091474475, "grad_norm": 0.3975356665709952, "learning_rate": 0.00011150779270629353, "loss": 1.1873, "step": 1995 }, { "epoch": 0.5182689816014512, "grad_norm": 0.4185740854748383, "learning_rate": 0.00011105839329482397, "loss": 1.1108, "step": 2000 }, { "epoch": 0.5195646540554548, "grad_norm": 0.38989666093808334, "learning_rate": 0.00011060876766831285, "loss": 1.1663, "step": 2005 }, { "epoch": 0.5208603265094585, "grad_norm": 0.40596909372428913, "learning_rate": 0.00011015892502448692, "loss": 1.1382, "step": 2010 }, { "epoch": 0.522155998963462, "grad_norm": 0.37956800433559007, "learning_rate": 0.00010970887456551234, "loss": 1.104, "step": 2015 }, { "epoch": 0.5234516714174656, "grad_norm": 0.3972204538731939, "learning_rate": 0.00010925862549780637, "loss": 1.1414, "step": 2020 }, { "epoch": 0.5247473438714693, "grad_norm": 0.4266101329792826, "learning_rate": 0.00010880818703184919, "loss": 1.2131, "step": 2025 }, { "epoch": 0.5260430163254729, "grad_norm": 0.4059164191077208, "learning_rate": 0.00010835756838199524, "loss": 1.1005, "step": 2030 }, { "epoch": 0.5273386887794765, "grad_norm": 0.37690318544316265, "learning_rate": 0.00010790677876628501, "loss": 1.1448, "step": 2035 }, { "epoch": 0.5286343612334802, "grad_norm": 0.39147947369844754, "learning_rate": 0.00010745582740625631, "loss": 1.161, "step": 2040 }, { "epoch": 0.5299300336874838, "grad_norm": 0.4061256597825514, "learning_rate": 0.00010700472352675556, "loss": 1.1861, "step": 2045 }, { "epoch": 0.5312257061414875, "grad_norm": 0.3877742069695207, "learning_rate": 0.00010655347635574937, "loss": 1.1816, "step": 2050 }, { "epoch": 0.5325213785954911, "grad_norm": 0.3971133244937986, "learning_rate": 0.00010610209512413536, "loss": 1.2017, "step": 2055 }, { "epoch": 0.5338170510494947, "grad_norm": 0.41551450058475425, "learning_rate": 0.0001056505890655537, "loss": 1.1692, "step": 2060 }, { "epoch": 0.5351127235034984, "grad_norm": 0.43646326285085413, "learning_rate": 0.00010519896741619803, "loss": 1.0993, "step": 2065 }, { "epoch": 0.5364083959575019, "grad_norm": 0.38667765038581386, "learning_rate": 0.00010474723941462658, "loss": 1.1151, "step": 2070 }, { "epoch": 0.5377040684115055, "grad_norm": 0.39156927596261365, "learning_rate": 0.00010429541430157313, "loss": 1.1641, "step": 2075 }, { "epoch": 0.5389997408655092, "grad_norm": 0.39506144218425854, "learning_rate": 0.00010384350131975802, "loss": 1.2444, "step": 2080 }, { "epoch": 0.5402954133195128, "grad_norm": 0.3881326675938924, "learning_rate": 0.00010339150971369906, "loss": 1.129, "step": 2085 }, { "epoch": 0.5415910857735164, "grad_norm": 0.3901794452377504, "learning_rate": 0.00010293944872952248, "loss": 1.1536, "step": 2090 }, { "epoch": 0.5428867582275201, "grad_norm": 0.3852428804649106, "learning_rate": 0.00010248732761477371, "loss": 1.1365, "step": 2095 }, { "epoch": 0.5441824306815237, "grad_norm": 0.4261491575765494, "learning_rate": 0.0001020351556182282, "loss": 1.089, "step": 2100 }, { "epoch": 0.5454781031355274, "grad_norm": 0.4070655687819442, "learning_rate": 0.00010158294198970232, "loss": 1.1591, "step": 2105 }, { "epoch": 0.546773775589531, "grad_norm": 0.4142505000419484, "learning_rate": 0.00010113069597986402, "loss": 1.1403, "step": 2110 }, { "epoch": 0.5480694480435346, "grad_norm": 0.3988427337375973, "learning_rate": 0.0001006784268400437, "loss": 1.1043, "step": 2115 }, { "epoch": 0.5493651204975383, "grad_norm": 0.37926295342271793, "learning_rate": 0.00010022614382204492, "loss": 1.1694, "step": 2120 }, { "epoch": 0.5506607929515418, "grad_norm": 0.3854042883512697, "learning_rate": 9.97738561779551e-05, "loss": 1.1891, "step": 2125 }, { "epoch": 0.5519564654055454, "grad_norm": 0.4035084032954952, "learning_rate": 9.932157315995631e-05, "loss": 1.1563, "step": 2130 }, { "epoch": 0.5532521378595491, "grad_norm": 0.4083098281929601, "learning_rate": 9.8869304020136e-05, "loss": 1.2097, "step": 2135 }, { "epoch": 0.5545478103135527, "grad_norm": 0.3884516778316842, "learning_rate": 9.841705801029769e-05, "loss": 1.1932, "step": 2140 }, { "epoch": 0.5558434827675564, "grad_norm": 0.38224759440627476, "learning_rate": 9.79648443817718e-05, "loss": 1.0978, "step": 2145 }, { "epoch": 0.55713915522156, "grad_norm": 0.46057864945040095, "learning_rate": 9.751267238522631e-05, "loss": 1.1465, "step": 2150 }, { "epoch": 0.5584348276755636, "grad_norm": 0.38627135619303393, "learning_rate": 9.706055127047755e-05, "loss": 1.1533, "step": 2155 }, { "epoch": 0.5597305001295673, "grad_norm": 0.3891179649712941, "learning_rate": 9.660849028630096e-05, "loss": 1.1984, "step": 2160 }, { "epoch": 0.5610261725835709, "grad_norm": 0.39223924604212307, "learning_rate": 9.615649868024199e-05, "loss": 1.1463, "step": 2165 }, { "epoch": 0.5623218450375745, "grad_norm": 0.39471896059848777, "learning_rate": 9.570458569842688e-05, "loss": 1.1487, "step": 2170 }, { "epoch": 0.5636175174915782, "grad_norm": 0.3928874290731547, "learning_rate": 9.525276058537344e-05, "loss": 1.1315, "step": 2175 }, { "epoch": 0.5649131899455817, "grad_norm": 0.4064990206731607, "learning_rate": 9.480103258380198e-05, "loss": 1.1562, "step": 2180 }, { "epoch": 0.5662088623995853, "grad_norm": 0.38415170548201955, "learning_rate": 9.434941093444632e-05, "loss": 1.1811, "step": 2185 }, { "epoch": 0.567504534853589, "grad_norm": 0.3859240132285136, "learning_rate": 9.389790487586465e-05, "loss": 1.2216, "step": 2190 }, { "epoch": 0.5688002073075926, "grad_norm": 0.40615716009377495, "learning_rate": 9.344652364425065e-05, "loss": 1.2206, "step": 2195 }, { "epoch": 0.5700958797615963, "grad_norm": 0.3919524007695092, "learning_rate": 9.299527647324444e-05, "loss": 1.1929, "step": 2200 }, { "epoch": 0.5713915522155999, "grad_norm": 0.3821208159345121, "learning_rate": 9.254417259374374e-05, "loss": 1.0911, "step": 2205 }, { "epoch": 0.5726872246696035, "grad_norm": 0.42022947326779125, "learning_rate": 9.2093221233715e-05, "loss": 1.1504, "step": 2210 }, { "epoch": 0.5739828971236072, "grad_norm": 0.4544847994091584, "learning_rate": 9.164243161800477e-05, "loss": 1.1582, "step": 2215 }, { "epoch": 0.5752785695776108, "grad_norm": 0.41094059546634143, "learning_rate": 9.119181296815085e-05, "loss": 1.1297, "step": 2220 }, { "epoch": 0.5765742420316144, "grad_norm": 0.3805641014764856, "learning_rate": 9.074137450219364e-05, "loss": 1.185, "step": 2225 }, { "epoch": 0.5778699144856181, "grad_norm": 0.39504497212816847, "learning_rate": 9.02911254344877e-05, "loss": 1.1875, "step": 2230 }, { "epoch": 0.5791655869396216, "grad_norm": 0.40451192253045454, "learning_rate": 8.984107497551311e-05, "loss": 1.1368, "step": 2235 }, { "epoch": 0.5804612593936253, "grad_norm": 0.3769782624908954, "learning_rate": 8.939123233168717e-05, "loss": 1.1391, "step": 2240 }, { "epoch": 0.5817569318476289, "grad_norm": 0.4058748960564205, "learning_rate": 8.894160670517606e-05, "loss": 1.1751, "step": 2245 }, { "epoch": 0.5830526043016325, "grad_norm": 0.3881661677768996, "learning_rate": 8.849220729370651e-05, "loss": 1.1215, "step": 2250 }, { "epoch": 0.5843482767556362, "grad_norm": 0.3807361487895863, "learning_rate": 8.804304329037771e-05, "loss": 1.1604, "step": 2255 }, { "epoch": 0.5856439492096398, "grad_norm": 0.3780073193073684, "learning_rate": 8.759412388347321e-05, "loss": 1.1159, "step": 2260 }, { "epoch": 0.5869396216636434, "grad_norm": 0.3910762750334457, "learning_rate": 8.71454582562731e-05, "loss": 1.1283, "step": 2265 }, { "epoch": 0.5882352941176471, "grad_norm": 0.3577377830833829, "learning_rate": 8.669705558686599e-05, "loss": 1.0881, "step": 2270 }, { "epoch": 0.5895309665716507, "grad_norm": 0.38870803678431104, "learning_rate": 8.624892504796128e-05, "loss": 1.1302, "step": 2275 }, { "epoch": 0.5908266390256544, "grad_norm": 0.43127992879460963, "learning_rate": 8.580107580670163e-05, "loss": 1.183, "step": 2280 }, { "epoch": 0.592122311479658, "grad_norm": 0.37408901643032355, "learning_rate": 8.535351702447524e-05, "loss": 1.1624, "step": 2285 }, { "epoch": 0.5934179839336615, "grad_norm": 0.4003666810623533, "learning_rate": 8.49062578567287e-05, "loss": 1.1193, "step": 2290 }, { "epoch": 0.5947136563876652, "grad_norm": 0.3723417367446364, "learning_rate": 8.445930745277953e-05, "loss": 1.146, "step": 2295 }, { "epoch": 0.5960093288416688, "grad_norm": 0.35036357860872624, "learning_rate": 8.401267495562894e-05, "loss": 1.0963, "step": 2300 }, { "epoch": 0.5973050012956724, "grad_norm": 0.3969850486600124, "learning_rate": 8.356636950177509e-05, "loss": 1.1645, "step": 2305 }, { "epoch": 0.5986006737496761, "grad_norm": 0.3908066674173964, "learning_rate": 8.312040022102581e-05, "loss": 1.1573, "step": 2310 }, { "epoch": 0.5998963462036797, "grad_norm": 0.3970139193288793, "learning_rate": 8.26747762363122e-05, "loss": 1.1475, "step": 2315 }, { "epoch": 0.6011920186576833, "grad_norm": 0.3809149209886327, "learning_rate": 8.222950666350176e-05, "loss": 1.0996, "step": 2320 }, { "epoch": 0.602487691111687, "grad_norm": 0.3831688926512033, "learning_rate": 8.1784600611212e-05, "loss": 1.1586, "step": 2325 }, { "epoch": 0.6037833635656906, "grad_norm": 0.37892711696966613, "learning_rate": 8.134006718062417e-05, "loss": 1.1394, "step": 2330 }, { "epoch": 0.6050790360196943, "grad_norm": 0.3796251460883534, "learning_rate": 8.089591546529686e-05, "loss": 1.1342, "step": 2335 }, { "epoch": 0.6063747084736979, "grad_norm": 0.3854947360629934, "learning_rate": 8.04521545509803e-05, "loss": 1.1234, "step": 2340 }, { "epoch": 0.6076703809277014, "grad_norm": 0.39157237980540016, "learning_rate": 8.00087935154303e-05, "loss": 1.1789, "step": 2345 }, { "epoch": 0.6089660533817051, "grad_norm": 0.37973961502373016, "learning_rate": 7.956584142822248e-05, "loss": 1.1725, "step": 2350 }, { "epoch": 0.6102617258357087, "grad_norm": 0.4002316998884078, "learning_rate": 7.912330735056702e-05, "loss": 1.1761, "step": 2355 }, { "epoch": 0.6115573982897123, "grad_norm": 0.3933317900533284, "learning_rate": 7.868120033512294e-05, "loss": 1.1279, "step": 2360 }, { "epoch": 0.612853070743716, "grad_norm": 0.3994518364881897, "learning_rate": 7.82395294258132e-05, "loss": 1.1292, "step": 2365 }, { "epoch": 0.6141487431977196, "grad_norm": 0.3774270174393722, "learning_rate": 7.779830365763963e-05, "loss": 1.1699, "step": 2370 }, { "epoch": 0.6154444156517233, "grad_norm": 0.3936443342897645, "learning_rate": 7.735753205649798e-05, "loss": 1.1691, "step": 2375 }, { "epoch": 0.6167400881057269, "grad_norm": 0.3958990632556032, "learning_rate": 7.691722363899346e-05, "loss": 1.1053, "step": 2380 }, { "epoch": 0.6180357605597305, "grad_norm": 0.40113089490170173, "learning_rate": 7.647738741225605e-05, "loss": 1.0826, "step": 2385 }, { "epoch": 0.6193314330137342, "grad_norm": 0.39930864888851864, "learning_rate": 7.60380323737566e-05, "loss": 1.199, "step": 2390 }, { "epoch": 0.6206271054677378, "grad_norm": 0.3962521144461154, "learning_rate": 7.559916751112248e-05, "loss": 1.1933, "step": 2395 }, { "epoch": 0.6219227779217413, "grad_norm": 0.37888926408124746, "learning_rate": 7.516080180195379e-05, "loss": 1.1713, "step": 2400 }, { "epoch": 0.623218450375745, "grad_norm": 0.3708133442805237, "learning_rate": 7.472294421363982e-05, "loss": 1.2177, "step": 2405 }, { "epoch": 0.6245141228297486, "grad_norm": 0.41709292403420495, "learning_rate": 7.428560370317542e-05, "loss": 1.1678, "step": 2410 }, { "epoch": 0.6258097952837522, "grad_norm": 0.38822425697798224, "learning_rate": 7.3848789216978e-05, "loss": 1.1628, "step": 2415 }, { "epoch": 0.6271054677377559, "grad_norm": 0.38293720555111366, "learning_rate": 7.341250969070435e-05, "loss": 1.1668, "step": 2420 }, { "epoch": 0.6284011401917595, "grad_norm": 0.3982879136320185, "learning_rate": 7.297677404906787e-05, "loss": 1.1808, "step": 2425 }, { "epoch": 0.6296968126457632, "grad_norm": 0.4970736508224388, "learning_rate": 7.254159120565614e-05, "loss": 1.1915, "step": 2430 }, { "epoch": 0.6309924850997668, "grad_norm": 0.3859126967679283, "learning_rate": 7.210697006274829e-05, "loss": 1.1032, "step": 2435 }, { "epoch": 0.6322881575537704, "grad_norm": 0.3804578357151546, "learning_rate": 7.167291951113322e-05, "loss": 1.1567, "step": 2440 }, { "epoch": 0.6335838300077741, "grad_norm": 0.38190993552219266, "learning_rate": 7.123944842992759e-05, "loss": 1.091, "step": 2445 }, { "epoch": 0.6348795024617777, "grad_norm": 0.38091685226387034, "learning_rate": 7.080656568639406e-05, "loss": 1.1457, "step": 2450 }, { "epoch": 0.6361751749157812, "grad_norm": 0.3853362546606581, "learning_rate": 7.037428013576013e-05, "loss": 1.1463, "step": 2455 }, { "epoch": 0.6374708473697849, "grad_norm": 0.36970236476000445, "learning_rate": 6.994260062103674e-05, "loss": 1.1393, "step": 2460 }, { "epoch": 0.6387665198237885, "grad_norm": 0.40400326202643533, "learning_rate": 6.951153597283766e-05, "loss": 1.1295, "step": 2465 }, { "epoch": 0.6400621922777922, "grad_norm": 0.3816874564299913, "learning_rate": 6.908109500919861e-05, "loss": 1.1662, "step": 2470 }, { "epoch": 0.6413578647317958, "grad_norm": 0.3799563658685613, "learning_rate": 6.865128653539699e-05, "loss": 1.1876, "step": 2475 }, { "epoch": 0.6426535371857994, "grad_norm": 0.379580454677419, "learning_rate": 6.822211934377176e-05, "loss": 1.1005, "step": 2480 }, { "epoch": 0.6439492096398031, "grad_norm": 0.36403149625971076, "learning_rate": 6.77936022135434e-05, "loss": 1.0924, "step": 2485 }, { "epoch": 0.6452448820938067, "grad_norm": 0.3864263981944349, "learning_rate": 6.736574391063466e-05, "loss": 1.166, "step": 2490 }, { "epoch": 0.6465405545478103, "grad_norm": 0.4278354770987037, "learning_rate": 6.693855318749096e-05, "loss": 1.1463, "step": 2495 }, { "epoch": 0.647836227001814, "grad_norm": 0.3881716691518243, "learning_rate": 6.651203878290139e-05, "loss": 1.1658, "step": 2500 }, { "epoch": 0.6491318994558176, "grad_norm": 0.3773480140369342, "learning_rate": 6.608620942182011e-05, "loss": 1.1561, "step": 2505 }, { "epoch": 0.6504275719098211, "grad_norm": 0.38362919406181467, "learning_rate": 6.566107381518758e-05, "loss": 1.1553, "step": 2510 }, { "epoch": 0.6517232443638248, "grad_norm": 0.3889927393341413, "learning_rate": 6.523664065975268e-05, "loss": 1.0927, "step": 2515 }, { "epoch": 0.6530189168178284, "grad_norm": 0.3770322611579158, "learning_rate": 6.481291863789461e-05, "loss": 1.1621, "step": 2520 }, { "epoch": 0.6543145892718321, "grad_norm": 0.3583870387014761, "learning_rate": 6.43899164174453e-05, "loss": 1.1688, "step": 2525 }, { "epoch": 0.6556102617258357, "grad_norm": 0.36282391266493386, "learning_rate": 6.396764265151221e-05, "loss": 1.0909, "step": 2530 }, { "epoch": 0.6569059341798393, "grad_norm": 0.40358089248833323, "learning_rate": 6.35461059783011e-05, "loss": 1.185, "step": 2535 }, { "epoch": 0.658201606633843, "grad_norm": 0.3622883465103346, "learning_rate": 6.312531502093958e-05, "loss": 1.1586, "step": 2540 }, { "epoch": 0.6594972790878466, "grad_norm": 0.39189718464338913, "learning_rate": 6.270527838730053e-05, "loss": 1.1592, "step": 2545 }, { "epoch": 0.6607929515418502, "grad_norm": 0.36779638890391525, "learning_rate": 6.228600466982611e-05, "loss": 1.156, "step": 2550 }, { "epoch": 0.6620886239958539, "grad_norm": 0.37597682220271855, "learning_rate": 6.186750244535194e-05, "loss": 1.1353, "step": 2555 }, { "epoch": 0.6633842964498575, "grad_norm": 0.3886787309603002, "learning_rate": 6.144978027493158e-05, "loss": 1.1341, "step": 2560 }, { "epoch": 0.664679968903861, "grad_norm": 0.35942315304405126, "learning_rate": 6.103284670366162e-05, "loss": 1.1597, "step": 2565 }, { "epoch": 0.6659756413578647, "grad_norm": 0.41146010175047676, "learning_rate": 6.061671026050668e-05, "loss": 1.18, "step": 2570 }, { "epoch": 0.6672713138118683, "grad_norm": 0.38590903226923434, "learning_rate": 6.0201379458124964e-05, "loss": 1.1639, "step": 2575 }, { "epoch": 0.668566986265872, "grad_norm": 0.386535230930437, "learning_rate": 5.978686279269421e-05, "loss": 1.0866, "step": 2580 }, { "epoch": 0.6698626587198756, "grad_norm": 0.3877441934910258, "learning_rate": 5.9373168743737864e-05, "loss": 1.1325, "step": 2585 }, { "epoch": 0.6711583311738792, "grad_norm": 0.3990007152806006, "learning_rate": 5.896030577395144e-05, "loss": 1.1188, "step": 2590 }, { "epoch": 0.6724540036278829, "grad_norm": 0.3970387094956777, "learning_rate": 5.854828232902976e-05, "loss": 1.1939, "step": 2595 }, { "epoch": 0.6737496760818865, "grad_norm": 0.39839395255304766, "learning_rate": 5.813710683749379e-05, "loss": 1.2078, "step": 2600 }, { "epoch": 0.6750453485358902, "grad_norm": 0.39807324424891044, "learning_rate": 5.772678771051858e-05, "loss": 1.1513, "step": 2605 }, { "epoch": 0.6763410209898938, "grad_norm": 0.371233691663641, "learning_rate": 5.7317333341760906e-05, "loss": 1.1896, "step": 2610 }, { "epoch": 0.6776366934438974, "grad_norm": 0.3914320864664493, "learning_rate": 5.690875210718778e-05, "loss": 1.144, "step": 2615 }, { "epoch": 0.678932365897901, "grad_norm": 0.38952180289646227, "learning_rate": 5.6501052364904906e-05, "loss": 1.1669, "step": 2620 }, { "epoch": 0.6802280383519046, "grad_norm": 0.3623267889899647, "learning_rate": 5.6094242454986e-05, "loss": 1.179, "step": 2625 }, { "epoch": 0.6815237108059082, "grad_norm": 0.3637529461983289, "learning_rate": 5.568833069930186e-05, "loss": 1.1244, "step": 2630 }, { "epoch": 0.6828193832599119, "grad_norm": 0.387937514120572, "learning_rate": 5.528332540135031e-05, "loss": 1.1699, "step": 2635 }, { "epoch": 0.6841150557139155, "grad_norm": 0.37878126929029915, "learning_rate": 5.487923484608629e-05, "loss": 1.131, "step": 2640 }, { "epoch": 0.6854107281679191, "grad_norm": 0.3903173526956151, "learning_rate": 5.4476067299752385e-05, "loss": 1.1165, "step": 2645 }, { "epoch": 0.6867064006219228, "grad_norm": 0.3857342653067374, "learning_rate": 5.4073831009709805e-05, "loss": 1.0854, "step": 2650 }, { "epoch": 0.6880020730759264, "grad_norm": 0.3661818366505174, "learning_rate": 5.367253420426952e-05, "loss": 1.173, "step": 2655 }, { "epoch": 0.6892977455299301, "grad_norm": 0.3695899768899783, "learning_rate": 5.3272185092524004e-05, "loss": 1.1917, "step": 2660 }, { "epoch": 0.6905934179839337, "grad_norm": 0.3709048127868108, "learning_rate": 5.287279186417938e-05, "loss": 1.1595, "step": 2665 }, { "epoch": 0.6918890904379373, "grad_norm": 0.388715290668864, "learning_rate": 5.2474362689387745e-05, "loss": 1.1681, "step": 2670 }, { "epoch": 0.6931847628919409, "grad_norm": 0.41572003898003435, "learning_rate": 5.207690571858025e-05, "loss": 1.1344, "step": 2675 }, { "epoch": 0.6944804353459445, "grad_norm": 0.3688637831571822, "learning_rate": 5.1680429082300134e-05, "loss": 1.1911, "step": 2680 }, { "epoch": 0.6957761077999481, "grad_norm": 0.3793170468762598, "learning_rate": 5.128494089103652e-05, "loss": 1.1127, "step": 2685 }, { "epoch": 0.6970717802539518, "grad_norm": 0.43134296541622413, "learning_rate": 5.0890449235058525e-05, "loss": 1.1784, "step": 2690 }, { "epoch": 0.6983674527079554, "grad_norm": 0.3635096529171273, "learning_rate": 5.049696218424969e-05, "loss": 1.1428, "step": 2695 }, { "epoch": 0.699663125161959, "grad_norm": 0.3878749986675832, "learning_rate": 5.010448778794303e-05, "loss": 1.1741, "step": 2700 }, { "epoch": 0.7009587976159627, "grad_norm": 0.40169342624996973, "learning_rate": 4.971303407475618e-05, "loss": 1.1599, "step": 2705 }, { "epoch": 0.7022544700699663, "grad_norm": 0.3694514584240007, "learning_rate": 4.932260905242731e-05, "loss": 1.1187, "step": 2710 }, { "epoch": 0.70355014252397, "grad_norm": 0.40107470744868656, "learning_rate": 4.893322070765126e-05, "loss": 1.1087, "step": 2715 }, { "epoch": 0.7048458149779736, "grad_norm": 0.3973309016316779, "learning_rate": 4.8544877005916126e-05, "loss": 1.1353, "step": 2720 }, { "epoch": 0.7061414874319772, "grad_norm": 0.36945482581363814, "learning_rate": 4.815758589134046e-05, "loss": 1.1537, "step": 2725 }, { "epoch": 0.7074371598859808, "grad_norm": 0.39405014852767495, "learning_rate": 4.777135528651058e-05, "loss": 1.1864, "step": 2730 }, { "epoch": 0.7087328323399844, "grad_norm": 0.36438744458294, "learning_rate": 4.738619309231857e-05, "loss": 1.1571, "step": 2735 }, { "epoch": 0.710028504793988, "grad_norm": 0.3718851263620155, "learning_rate": 4.700210718780072e-05, "loss": 1.1375, "step": 2740 }, { "epoch": 0.7113241772479917, "grad_norm": 0.3968100064213781, "learning_rate": 4.6619105429976193e-05, "loss": 1.0992, "step": 2745 }, { "epoch": 0.7126198497019953, "grad_norm": 0.3735407865710053, "learning_rate": 4.623719565368657e-05, "loss": 1.154, "step": 2750 }, { "epoch": 0.713915522155999, "grad_norm": 0.3855769481718261, "learning_rate": 4.585638567143529e-05, "loss": 1.1087, "step": 2755 }, { "epoch": 0.7152111946100026, "grad_norm": 0.39904660177632, "learning_rate": 4.547668327322796e-05, "loss": 1.1557, "step": 2760 }, { "epoch": 0.7165068670640062, "grad_norm": 0.37552561838186715, "learning_rate": 4.5098096226413e-05, "loss": 1.1788, "step": 2765 }, { "epoch": 0.7178025395180099, "grad_norm": 0.39324730815800873, "learning_rate": 4.472063227552274e-05, "loss": 1.1688, "step": 2770 }, { "epoch": 0.7190982119720135, "grad_norm": 0.38301296634310333, "learning_rate": 4.434429914211508e-05, "loss": 1.1711, "step": 2775 }, { "epoch": 0.7203938844260172, "grad_norm": 0.3709443062095975, "learning_rate": 4.396910452461532e-05, "loss": 1.1718, "step": 2780 }, { "epoch": 0.7216895568800207, "grad_norm": 0.39239048686699185, "learning_rate": 4.3595056098158906e-05, "loss": 1.1574, "step": 2785 }, { "epoch": 0.7229852293340243, "grad_norm": 0.41357484414155377, "learning_rate": 4.322216151443428e-05, "loss": 1.1074, "step": 2790 }, { "epoch": 0.724280901788028, "grad_norm": 0.3672227895875153, "learning_rate": 4.2850428401526376e-05, "loss": 1.1028, "step": 2795 }, { "epoch": 0.7255765742420316, "grad_norm": 0.3664872416848814, "learning_rate": 4.2479864363760726e-05, "loss": 1.1508, "step": 2800 }, { "epoch": 0.7268722466960352, "grad_norm": 0.3710822782612391, "learning_rate": 4.211047698154765e-05, "loss": 1.1369, "step": 2805 }, { "epoch": 0.7281679191500389, "grad_norm": 0.38199563385579133, "learning_rate": 4.1742273811227395e-05, "loss": 1.1716, "step": 2810 }, { "epoch": 0.7294635916040425, "grad_norm": 0.36978932404422615, "learning_rate": 4.1375262384915433e-05, "loss": 1.1613, "step": 2815 }, { "epoch": 0.7307592640580461, "grad_norm": 0.37568319756030377, "learning_rate": 4.100945021034843e-05, "loss": 1.1396, "step": 2820 }, { "epoch": 0.7320549365120498, "grad_norm": 0.3802598263999195, "learning_rate": 4.064484477073074e-05, "loss": 1.172, "step": 2825 }, { "epoch": 0.7333506089660534, "grad_norm": 0.3690252573947318, "learning_rate": 4.028145352458118e-05, "loss": 1.1413, "step": 2830 }, { "epoch": 0.7346462814200571, "grad_norm": 0.3942754823686454, "learning_rate": 3.991928390558054e-05, "loss": 1.1069, "step": 2835 }, { "epoch": 0.7359419538740606, "grad_norm": 0.3720821371337478, "learning_rate": 3.95583433224195e-05, "loss": 1.0835, "step": 2840 }, { "epoch": 0.7372376263280642, "grad_norm": 0.3807303246565132, "learning_rate": 3.9198639158647056e-05, "loss": 1.1501, "step": 2845 }, { "epoch": 0.7385332987820679, "grad_norm": 0.36231750839777666, "learning_rate": 3.884017877251959e-05, "loss": 1.0563, "step": 2850 }, { "epoch": 0.7398289712360715, "grad_norm": 0.388874708381432, "learning_rate": 3.8482969496850166e-05, "loss": 1.0889, "step": 2855 }, { "epoch": 0.7411246436900751, "grad_norm": 0.38013479771110215, "learning_rate": 3.812701863885865e-05, "loss": 1.1145, "step": 2860 }, { "epoch": 0.7424203161440788, "grad_norm": 0.3717583300327742, "learning_rate": 3.7772333480022185e-05, "loss": 1.0663, "step": 2865 }, { "epoch": 0.7437159885980824, "grad_norm": 0.36405208418925755, "learning_rate": 3.741892127592625e-05, "loss": 1.1102, "step": 2870 }, { "epoch": 0.745011661052086, "grad_norm": 0.3723066034084609, "learning_rate": 3.706678925611629e-05, "loss": 1.1602, "step": 2875 }, { "epoch": 0.7463073335060897, "grad_norm": 0.3783951434167031, "learning_rate": 3.67159446239497e-05, "loss": 1.1233, "step": 2880 }, { "epoch": 0.7476030059600933, "grad_norm": 0.37951067018123125, "learning_rate": 3.636639455644858e-05, "loss": 1.0545, "step": 2885 }, { "epoch": 0.748898678414097, "grad_norm": 0.37421338215744115, "learning_rate": 3.601814620415296e-05, "loss": 1.1367, "step": 2890 }, { "epoch": 0.7501943508681005, "grad_norm": 0.3603159123651838, "learning_rate": 3.567120669097422e-05, "loss": 1.0909, "step": 2895 }, { "epoch": 0.7514900233221041, "grad_norm": 0.3788231837374063, "learning_rate": 3.532558311404986e-05, "loss": 1.1441, "step": 2900 }, { "epoch": 0.7527856957761078, "grad_norm": 0.3840910679589615, "learning_rate": 3.498128254359788e-05, "loss": 1.1403, "step": 2905 }, { "epoch": 0.7540813682301114, "grad_norm": 0.3753865209395276, "learning_rate": 3.4638312022772335e-05, "loss": 1.1004, "step": 2910 }, { "epoch": 0.755377040684115, "grad_norm": 0.3954242259257984, "learning_rate": 3.4296678567519345e-05, "loss": 1.1347, "step": 2915 }, { "epoch": 0.7566727131381187, "grad_norm": 0.3741932569564154, "learning_rate": 3.3956389166433276e-05, "loss": 1.0908, "step": 2920 }, { "epoch": 0.7579683855921223, "grad_norm": 0.3930950040373874, "learning_rate": 3.361745078061416e-05, "loss": 1.1988, "step": 2925 }, { "epoch": 0.759264058046126, "grad_norm": 0.3685363645101882, "learning_rate": 3.327987034352499e-05, "loss": 1.2067, "step": 2930 }, { "epoch": 0.7605597305001296, "grad_norm": 0.3595878774367224, "learning_rate": 3.294365476085001e-05, "loss": 1.1128, "step": 2935 }, { "epoch": 0.7618554029541332, "grad_norm": 0.3823858203204671, "learning_rate": 3.260881091035356e-05, "loss": 1.1894, "step": 2940 }, { "epoch": 0.7631510754081369, "grad_norm": 0.35550275725905045, "learning_rate": 3.227534564173903e-05, "loss": 1.0744, "step": 2945 }, { "epoch": 0.7644467478621404, "grad_norm": 0.38923682484253824, "learning_rate": 3.1943265776509215e-05, "loss": 1.1599, "step": 2950 }, { "epoch": 0.765742420316144, "grad_norm": 0.3789384247163881, "learning_rate": 3.1612578107826375e-05, "loss": 1.1504, "step": 2955 }, { "epoch": 0.7670380927701477, "grad_norm": 0.3620246203990784, "learning_rate": 3.128328940037345e-05, "loss": 1.0777, "step": 2960 }, { "epoch": 0.7683337652241513, "grad_norm": 0.36815557923240855, "learning_rate": 3.095540639021578e-05, "loss": 1.1174, "step": 2965 }, { "epoch": 0.769629437678155, "grad_norm": 0.3547702710213197, "learning_rate": 3.0628935784662947e-05, "loss": 1.1485, "step": 2970 }, { "epoch": 0.7709251101321586, "grad_norm": 0.3781571691176777, "learning_rate": 3.030388426213202e-05, "loss": 1.1883, "step": 2975 }, { "epoch": 0.7722207825861622, "grad_norm": 0.40077870265330406, "learning_rate": 2.9980258472010624e-05, "loss": 1.1932, "step": 2980 }, { "epoch": 0.7735164550401659, "grad_norm": 0.4044666107500387, "learning_rate": 2.9658065034520978e-05, "loss": 1.1241, "step": 2985 }, { "epoch": 0.7748121274941695, "grad_norm": 0.38599774202480436, "learning_rate": 2.9337310540584662e-05, "loss": 1.1289, "step": 2990 }, { "epoch": 0.7761077999481731, "grad_norm": 0.3937959908311204, "learning_rate": 2.90180015516874e-05, "loss": 1.124, "step": 2995 }, { "epoch": 0.7774034724021768, "grad_norm": 0.3632550787528134, "learning_rate": 2.8700144599745304e-05, "loss": 1.054, "step": 3000 }, { "epoch": 0.7786991448561803, "grad_norm": 0.3622147641579977, "learning_rate": 2.8383746186970885e-05, "loss": 1.1541, "step": 3005 }, { "epoch": 0.7799948173101839, "grad_norm": 0.37112400149549435, "learning_rate": 2.806881278574016e-05, "loss": 1.0959, "step": 3010 }, { "epoch": 0.7812904897641876, "grad_norm": 0.35896620554471, "learning_rate": 2.7755350838460437e-05, "loss": 1.1097, "step": 3015 }, { "epoch": 0.7825861622181912, "grad_norm": 0.3722645573928852, "learning_rate": 2.7443366757438084e-05, "loss": 1.1536, "step": 3020 }, { "epoch": 0.7838818346721949, "grad_norm": 0.3733728824194641, "learning_rate": 2.71328669247479e-05, "loss": 1.0988, "step": 3025 }, { "epoch": 0.7851775071261985, "grad_norm": 0.39726305827044567, "learning_rate": 2.6823857692102115e-05, "loss": 1.1347, "step": 3030 }, { "epoch": 0.7864731795802021, "grad_norm": 0.3757164187600116, "learning_rate": 2.6516345380720685e-05, "loss": 1.1498, "step": 3035 }, { "epoch": 0.7877688520342058, "grad_norm": 0.3795959004315457, "learning_rate": 2.6210336281201996e-05, "loss": 1.1716, "step": 3040 }, { "epoch": 0.7890645244882094, "grad_norm": 0.35714774223397056, "learning_rate": 2.5905836653393955e-05, "loss": 1.1829, "step": 3045 }, { "epoch": 0.790360196942213, "grad_norm": 0.3654895953286083, "learning_rate": 2.5602852726266246e-05, "loss": 1.1138, "step": 3050 }, { "epoch": 0.7916558693962167, "grad_norm": 0.3790228967067868, "learning_rate": 2.53013906977827e-05, "loss": 1.1361, "step": 3055 }, { "epoch": 0.7929515418502202, "grad_norm": 0.3766214688716055, "learning_rate": 2.500145673477452e-05, "loss": 1.1386, "step": 3060 }, { "epoch": 0.7942472143042238, "grad_norm": 0.35689403644408657, "learning_rate": 2.4703056972814298e-05, "loss": 1.1716, "step": 3065 }, { "epoch": 0.7955428867582275, "grad_norm": 0.3837337630770634, "learning_rate": 2.44061975160902e-05, "loss": 1.1253, "step": 3070 }, { "epoch": 0.7968385592122311, "grad_norm": 0.3617908256654209, "learning_rate": 2.4110884437281433e-05, "loss": 1.1669, "step": 3075 }, { "epoch": 0.7981342316662348, "grad_norm": 0.35978473851190906, "learning_rate": 2.381712377743379e-05, "loss": 1.1195, "step": 3080 }, { "epoch": 0.7994299041202384, "grad_norm": 0.36747445883447966, "learning_rate": 2.352492154583611e-05, "loss": 1.1494, "step": 3085 }, { "epoch": 0.800725576574242, "grad_norm": 0.38717115547376924, "learning_rate": 2.323428371989752e-05, "loss": 1.1342, "step": 3090 }, { "epoch": 0.8020212490282457, "grad_norm": 0.39093054065016897, "learning_rate": 2.2945216245024804e-05, "loss": 1.1304, "step": 3095 }, { "epoch": 0.8033169214822493, "grad_norm": 0.368756140981371, "learning_rate": 2.265772503450122e-05, "loss": 1.1902, "step": 3100 }, { "epoch": 0.804612593936253, "grad_norm": 0.3877482016759796, "learning_rate": 2.237181596936515e-05, "loss": 1.2071, "step": 3105 }, { "epoch": 0.8059082663902566, "grad_norm": 0.37792584298436904, "learning_rate": 2.2087494898290084e-05, "loss": 1.1263, "step": 3110 }, { "epoch": 0.8072039388442601, "grad_norm": 0.377175581709425, "learning_rate": 2.1804767637464783e-05, "loss": 1.1409, "step": 3115 }, { "epoch": 0.8084996112982638, "grad_norm": 0.3675212675701305, "learning_rate": 2.152363997047432e-05, "loss": 1.1222, "step": 3120 }, { "epoch": 0.8097952837522674, "grad_norm": 0.38222618034509426, "learning_rate": 2.1244117648181926e-05, "loss": 1.1282, "step": 3125 }, { "epoch": 0.811090956206271, "grad_norm": 0.36558116292433085, "learning_rate": 2.0966206388611177e-05, "loss": 1.1599, "step": 3130 }, { "epoch": 0.8123866286602747, "grad_norm": 0.3727524846387591, "learning_rate": 2.0689911876829127e-05, "loss": 1.1496, "step": 3135 }, { "epoch": 0.8136823011142783, "grad_norm": 0.3628037534722254, "learning_rate": 2.0415239764829976e-05, "loss": 1.1416, "step": 3140 }, { "epoch": 0.8149779735682819, "grad_norm": 0.3761023568840631, "learning_rate": 2.014219567141932e-05, "loss": 1.0828, "step": 3145 }, { "epoch": 0.8162736460222856, "grad_norm": 0.3504341651004141, "learning_rate": 1.9870785182099505e-05, "loss": 1.0956, "step": 3150 }, { "epoch": 0.8175693184762892, "grad_norm": 0.3734885666536526, "learning_rate": 1.960101384895511e-05, "loss": 1.1044, "step": 3155 }, { "epoch": 0.8188649909302929, "grad_norm": 0.386535956936462, "learning_rate": 1.9332887190539516e-05, "loss": 1.2004, "step": 3160 }, { "epoch": 0.8201606633842965, "grad_norm": 0.3646373120550804, "learning_rate": 1.9066410691761937e-05, "loss": 1.1377, "step": 3165 }, { "epoch": 0.8214563358383, "grad_norm": 0.3715486267985764, "learning_rate": 1.8801589803775154e-05, "loss": 1.1128, "step": 3170 }, { "epoch": 0.8227520082923037, "grad_norm": 0.3789134249781072, "learning_rate": 1.8538429943864244e-05, "loss": 1.1429, "step": 3175 }, { "epoch": 0.8240476807463073, "grad_norm": 0.4352073640959862, "learning_rate": 1.8276936495335485e-05, "loss": 1.1167, "step": 3180 }, { "epoch": 0.8253433532003109, "grad_norm": 0.3681435590124036, "learning_rate": 1.8017114807406478e-05, "loss": 1.1063, "step": 3185 }, { "epoch": 0.8266390256543146, "grad_norm": 0.3694513270515317, "learning_rate": 1.775897019509649e-05, "loss": 1.1836, "step": 3190 }, { "epoch": 0.8279346981083182, "grad_norm": 0.388666401561133, "learning_rate": 1.7502507939117897e-05, "loss": 1.1396, "step": 3195 }, { "epoch": 0.8292303705623219, "grad_norm": 0.3687281974613644, "learning_rate": 1.7247733285768098e-05, "loss": 1.1352, "step": 3200 }, { "epoch": 0.8305260430163255, "grad_norm": 0.3733462546424826, "learning_rate": 1.6994651446822153e-05, "loss": 1.125, "step": 3205 }, { "epoch": 0.8318217154703291, "grad_norm": 0.3745095547011446, "learning_rate": 1.6743267599426303e-05, "loss": 1.1645, "step": 3210 }, { "epoch": 0.8331173879243328, "grad_norm": 0.36579848745894494, "learning_rate": 1.649358688599191e-05, "loss": 1.1373, "step": 3215 }, { "epoch": 0.8344130603783364, "grad_norm": 0.37060439276258045, "learning_rate": 1.624561441409034e-05, "loss": 1.0775, "step": 3220 }, { "epoch": 0.83570873283234, "grad_norm": 0.38418584872461575, "learning_rate": 1.5999355256348448e-05, "loss": 1.1678, "step": 3225 }, { "epoch": 0.8370044052863436, "grad_norm": 0.3814911739387983, "learning_rate": 1.5754814450344845e-05, "loss": 1.1772, "step": 3230 }, { "epoch": 0.8383000777403472, "grad_norm": 0.3642771788771785, "learning_rate": 1.5511996998506883e-05, "loss": 1.1279, "step": 3235 }, { "epoch": 0.8395957501943508, "grad_norm": 0.3798091083229991, "learning_rate": 1.527090786800821e-05, "loss": 1.1663, "step": 3240 }, { "epoch": 0.8408914226483545, "grad_norm": 0.3621301040720381, "learning_rate": 1.5031551990667236e-05, "loss": 1.0961, "step": 3245 }, { "epoch": 0.8421870951023581, "grad_norm": 0.3716900347036682, "learning_rate": 1.4793934262846232e-05, "loss": 1.1246, "step": 3250 }, { "epoch": 0.8434827675563618, "grad_norm": 0.3431347363725958, "learning_rate": 1.4558059545351143e-05, "loss": 1.126, "step": 3255 }, { "epoch": 0.8447784400103654, "grad_norm": 0.38516161930376763, "learning_rate": 1.4323932663332251e-05, "loss": 1.1487, "step": 3260 }, { "epoch": 0.846074112464369, "grad_norm": 0.36582214467557017, "learning_rate": 1.4091558406185335e-05, "loss": 1.1464, "step": 3265 }, { "epoch": 0.8473697849183727, "grad_norm": 0.3600375664520519, "learning_rate": 1.3860941527453786e-05, "loss": 1.1331, "step": 3270 }, { "epoch": 0.8486654573723763, "grad_norm": 0.36937163487487523, "learning_rate": 1.3632086744731299e-05, "loss": 1.1, "step": 3275 }, { "epoch": 0.84996112982638, "grad_norm": 0.37811170010742284, "learning_rate": 1.3404998739565433e-05, "loss": 1.1278, "step": 3280 }, { "epoch": 0.8512568022803835, "grad_norm": 0.36422197248748484, "learning_rate": 1.3179682157361872e-05, "loss": 1.1193, "step": 3285 }, { "epoch": 0.8525524747343871, "grad_norm": 0.37165480607350426, "learning_rate": 1.2956141607289262e-05, "loss": 1.1615, "step": 3290 }, { "epoch": 0.8538481471883907, "grad_norm": 0.37701311154160283, "learning_rate": 1.2734381662185035e-05, "loss": 1.1036, "step": 3295 }, { "epoch": 0.8551438196423944, "grad_norm": 0.3815170151959314, "learning_rate": 1.2514406858461847e-05, "loss": 1.1735, "step": 3300 }, { "epoch": 0.856439492096398, "grad_norm": 0.37919701821847407, "learning_rate": 1.2296221696014732e-05, "loss": 1.1837, "step": 3305 }, { "epoch": 0.8577351645504017, "grad_norm": 0.36743928128733283, "learning_rate": 1.2079830638129164e-05, "loss": 1.1148, "step": 3310 }, { "epoch": 0.8590308370044053, "grad_norm": 0.3738619978563127, "learning_rate": 1.1865238111389588e-05, "loss": 1.1604, "step": 3315 }, { "epoch": 0.8603265094584089, "grad_norm": 0.3541904420302459, "learning_rate": 1.1652448505588998e-05, "loss": 1.1544, "step": 3320 }, { "epoch": 0.8616221819124126, "grad_norm": 0.36684430135825935, "learning_rate": 1.1441466173639092e-05, "loss": 1.156, "step": 3325 }, { "epoch": 0.8629178543664162, "grad_norm": 0.37365304474041194, "learning_rate": 1.1232295431481222e-05, "loss": 1.1084, "step": 3330 }, { "epoch": 0.8642135268204199, "grad_norm": 0.3810375689519541, "learning_rate": 1.1024940557998143e-05, "loss": 1.1386, "step": 3335 }, { "epoch": 0.8655091992744234, "grad_norm": 0.3673719354100514, "learning_rate": 1.0819405794926418e-05, "loss": 1.1629, "step": 3340 }, { "epoch": 0.866804871728427, "grad_norm": 0.37342554741406375, "learning_rate": 1.0615695346769693e-05, "loss": 1.1134, "step": 3345 }, { "epoch": 0.8681005441824307, "grad_norm": 0.3635218597663001, "learning_rate": 1.0413813380712701e-05, "loss": 1.1147, "step": 3350 }, { "epoch": 0.8693962166364343, "grad_norm": 0.3570736451827848, "learning_rate": 1.0213764026535921e-05, "loss": 1.1321, "step": 3355 }, { "epoch": 0.8706918890904379, "grad_norm": 0.3891720605828237, "learning_rate": 1.0015551376531296e-05, "loss": 1.1001, "step": 3360 }, { "epoch": 0.8719875615444416, "grad_norm": 0.37097051363645606, "learning_rate": 9.819179485418273e-06, "loss": 1.1241, "step": 3365 }, { "epoch": 0.8732832339984452, "grad_norm": 0.3917978913676709, "learning_rate": 9.624652370261034e-06, "loss": 1.1304, "step": 3370 }, { "epoch": 0.8745789064524488, "grad_norm": 0.37226262222810586, "learning_rate": 9.431974010386258e-06, "loss": 1.1421, "step": 3375 }, { "epoch": 0.8758745789064525, "grad_norm": 0.3751233666218166, "learning_rate": 9.24114834730171e-06, "loss": 1.1022, "step": 3380 }, { "epoch": 0.8771702513604561, "grad_norm": 0.36555054148949856, "learning_rate": 9.052179284615658e-06, "loss": 1.13, "step": 3385 }, { "epoch": 0.8784659238144598, "grad_norm": 0.38148359927958986, "learning_rate": 8.865070687956977e-06, "loss": 1.1894, "step": 3390 }, { "epoch": 0.8797615962684633, "grad_norm": 0.416874595298691, "learning_rate": 8.679826384896061e-06, "loss": 1.1736, "step": 3395 }, { "epoch": 0.8810572687224669, "grad_norm": 0.387504501992165, "learning_rate": 8.496450164866565e-06, "loss": 1.1511, "step": 3400 }, { "epoch": 0.8823529411764706, "grad_norm": 0.34801518476142257, "learning_rate": 8.314945779087858e-06, "loss": 1.1224, "step": 3405 }, { "epoch": 0.8836486136304742, "grad_norm": 0.3758149676256447, "learning_rate": 8.135316940488347e-06, "loss": 1.1448, "step": 3410 }, { "epoch": 0.8849442860844778, "grad_norm": 0.36726096701274225, "learning_rate": 7.9575673236294e-06, "loss": 1.1719, "step": 3415 }, { "epoch": 0.8862399585384815, "grad_norm": 0.3578469295777726, "learning_rate": 7.781700564630311e-06, "loss": 1.1385, "step": 3420 }, { "epoch": 0.8875356309924851, "grad_norm": 0.38263190643853745, "learning_rate": 7.6077202610938205e-06, "loss": 1.177, "step": 3425 }, { "epoch": 0.8888313034464888, "grad_norm": 0.38665333786102024, "learning_rate": 7.435629972032565e-06, "loss": 1.1358, "step": 3430 }, { "epoch": 0.8901269759004924, "grad_norm": 0.36097779607915553, "learning_rate": 7.265433217796302e-06, "loss": 1.1055, "step": 3435 }, { "epoch": 0.891422648354496, "grad_norm": 0.4052378204511826, "learning_rate": 7.0971334799998404e-06, "loss": 1.0912, "step": 3440 }, { "epoch": 0.8927183208084997, "grad_norm": 0.3704412054341248, "learning_rate": 6.930734201451816e-06, "loss": 1.1591, "step": 3445 }, { "epoch": 0.8940139932625032, "grad_norm": 0.37555016415939446, "learning_rate": 6.7662387860843225e-06, "loss": 1.1149, "step": 3450 }, { "epoch": 0.8953096657165068, "grad_norm": 0.3798653467067839, "learning_rate": 6.603650598883226e-06, "loss": 1.0958, "step": 3455 }, { "epoch": 0.8966053381705105, "grad_norm": 0.34342261742501184, "learning_rate": 6.4429729658193714e-06, "loss": 1.0718, "step": 3460 }, { "epoch": 0.8979010106245141, "grad_norm": 0.38053269516746213, "learning_rate": 6.284209173780498e-06, "loss": 1.1133, "step": 3465 }, { "epoch": 0.8991966830785177, "grad_norm": 0.38564284280238426, "learning_rate": 6.127362470504027e-06, "loss": 1.1239, "step": 3470 }, { "epoch": 0.9004923555325214, "grad_norm": 0.3624213790138937, "learning_rate": 5.972436064510667e-06, "loss": 1.1766, "step": 3475 }, { "epoch": 0.901788027986525, "grad_norm": 0.36614157981464224, "learning_rate": 5.819433125038643e-06, "loss": 1.146, "step": 3480 }, { "epoch": 0.9030837004405287, "grad_norm": 0.39479288263797235, "learning_rate": 5.668356781979056e-06, "loss": 1.1669, "step": 3485 }, { "epoch": 0.9043793728945323, "grad_norm": 0.35385937176801885, "learning_rate": 5.519210125811713e-06, "loss": 1.1224, "step": 3490 }, { "epoch": 0.9056750453485359, "grad_norm": 0.3778900407632985, "learning_rate": 5.371996207541941e-06, "loss": 1.1287, "step": 3495 }, { "epoch": 0.9069707178025396, "grad_norm": 0.35939889684365445, "learning_rate": 5.226718038638234e-06, "loss": 1.1274, "step": 3500 }, { "epoch": 0.9082663902565431, "grad_norm": 0.38058001351217496, "learning_rate": 5.083378590970511e-06, "loss": 1.154, "step": 3505 }, { "epoch": 0.9095620627105467, "grad_norm": 0.36339105009188977, "learning_rate": 4.941980796749524e-06, "loss": 1.1234, "step": 3510 }, { "epoch": 0.9108577351645504, "grad_norm": 0.3753439481173232, "learning_rate": 4.802527548466684e-06, "loss": 1.0982, "step": 3515 }, { "epoch": 0.912153407618554, "grad_norm": 0.3547606097647797, "learning_rate": 4.665021698834981e-06, "loss": 1.1405, "step": 3520 }, { "epoch": 0.9134490800725577, "grad_norm": 0.3804573132062168, "learning_rate": 4.529466060730681e-06, "loss": 1.1529, "step": 3525 }, { "epoch": 0.9147447525265613, "grad_norm": 0.3766860837932323, "learning_rate": 4.395863407135614e-06, "loss": 1.0821, "step": 3530 }, { "epoch": 0.9160404249805649, "grad_norm": 0.3760175631598513, "learning_rate": 4.264216471080651e-06, "loss": 1.1451, "step": 3535 }, { "epoch": 0.9173360974345686, "grad_norm": 0.3901693827353405, "learning_rate": 4.134527945589639e-06, "loss": 1.1148, "step": 3540 }, { "epoch": 0.9186317698885722, "grad_norm": 0.4488796382773471, "learning_rate": 4.006800483624396e-06, "loss": 1.1614, "step": 3545 }, { "epoch": 0.9199274423425758, "grad_norm": 0.40897274265302724, "learning_rate": 3.881036698030449e-06, "loss": 1.1744, "step": 3550 }, { "epoch": 0.9212231147965795, "grad_norm": 0.3690887838305789, "learning_rate": 3.7572391614834833e-06, "loss": 1.136, "step": 3555 }, { "epoch": 0.922518787250583, "grad_norm": 0.3688927093580592, "learning_rate": 3.6354104064368566e-06, "loss": 1.0947, "step": 3560 }, { "epoch": 0.9238144597045866, "grad_norm": 0.3826643203341735, "learning_rate": 3.515552925069676e-06, "loss": 1.1757, "step": 3565 }, { "epoch": 0.9251101321585903, "grad_norm": 0.3684225941093978, "learning_rate": 3.397669169235862e-06, "loss": 1.1574, "step": 3570 }, { "epoch": 0.9264058046125939, "grad_norm": 0.3447005114527934, "learning_rate": 3.281761550414042e-06, "loss": 1.0847, "step": 3575 }, { "epoch": 0.9277014770665976, "grad_norm": 0.38643242398136646, "learning_rate": 3.1678324396581137e-06, "loss": 1.2221, "step": 3580 }, { "epoch": 0.9289971495206012, "grad_norm": 0.37779005091381196, "learning_rate": 3.0558841675488393e-06, "loss": 1.1612, "step": 3585 }, { "epoch": 0.9302928219746048, "grad_norm": 0.38097475759248384, "learning_rate": 2.9459190241461043e-06, "loss": 1.1267, "step": 3590 }, { "epoch": 0.9315884944286085, "grad_norm": 0.37618918488273334, "learning_rate": 2.8379392589421237e-06, "loss": 1.1563, "step": 3595 }, { "epoch": 0.9328841668826121, "grad_norm": 0.3621451372937871, "learning_rate": 2.7319470808153892e-06, "loss": 1.1197, "step": 3600 }, { "epoch": 0.9341798393366157, "grad_norm": 0.35758689530964316, "learning_rate": 2.6279446579854707e-06, "loss": 1.1446, "step": 3605 }, { "epoch": 0.9354755117906194, "grad_norm": 0.3803519145314357, "learning_rate": 2.525934117968731e-06, "loss": 1.1823, "step": 3610 }, { "epoch": 0.9367711842446229, "grad_norm": 0.36540749180301496, "learning_rate": 2.4259175475347172e-06, "loss": 1.1748, "step": 3615 }, { "epoch": 0.9380668566986265, "grad_norm": 0.34819215016667493, "learning_rate": 2.3278969926635252e-06, "loss": 1.1036, "step": 3620 }, { "epoch": 0.9393625291526302, "grad_norm": 0.3586884150884424, "learning_rate": 2.2318744585039796e-06, "loss": 1.1192, "step": 3625 }, { "epoch": 0.9406582016066338, "grad_norm": 0.37877938188955895, "learning_rate": 2.1378519093324776e-06, "loss": 1.128, "step": 3630 }, { "epoch": 0.9419538740606375, "grad_norm": 0.3781105512240613, "learning_rate": 2.0458312685129876e-06, "loss": 1.1612, "step": 3635 }, { "epoch": 0.9432495465146411, "grad_norm": 0.37487413050848234, "learning_rate": 1.955814418457591e-06, "loss": 1.1396, "step": 3640 }, { "epoch": 0.9445452189686447, "grad_norm": 0.3840114910436269, "learning_rate": 1.867803200587981e-06, "loss": 1.1129, "step": 3645 }, { "epoch": 0.9458408914226484, "grad_norm": 0.3719048343830175, "learning_rate": 1.7817994152978468e-06, "loss": 1.1157, "step": 3650 }, { "epoch": 0.947136563876652, "grad_norm": 0.3772856287062892, "learning_rate": 1.6978048219159714e-06, "loss": 1.1194, "step": 3655 }, { "epoch": 0.9484322363306557, "grad_norm": 0.3598408324385285, "learning_rate": 1.6158211386703259e-06, "loss": 1.1284, "step": 3660 }, { "epoch": 0.9497279087846593, "grad_norm": 0.365397297934618, "learning_rate": 1.5358500426528422e-06, "loss": 1.1282, "step": 3665 }, { "epoch": 0.9510235812386628, "grad_norm": 0.3659664905751428, "learning_rate": 1.4578931697851406e-06, "loss": 1.1051, "step": 3670 }, { "epoch": 0.9523192536926665, "grad_norm": 0.3549703867179501, "learning_rate": 1.3819521147851123e-06, "loss": 1.1292, "step": 3675 }, { "epoch": 0.9536149261466701, "grad_norm": 0.36856057544037374, "learning_rate": 1.3080284311341674e-06, "loss": 1.141, "step": 3680 }, { "epoch": 0.9549105986006737, "grad_norm": 0.3869169065095209, "learning_rate": 1.236123631045627e-06, "loss": 1.1622, "step": 3685 }, { "epoch": 0.9562062710546774, "grad_norm": 0.35481859423446044, "learning_rate": 1.1662391854336263e-06, "loss": 1.1223, "step": 3690 }, { "epoch": 0.957501943508681, "grad_norm": 0.38700062266195656, "learning_rate": 1.0983765238831377e-06, "loss": 1.1501, "step": 3695 }, { "epoch": 0.9587976159626846, "grad_norm": 0.3672922868247709, "learning_rate": 1.032537034620684e-06, "loss": 1.1419, "step": 3700 }, { "epoch": 0.9600932884166883, "grad_norm": 0.37605980694280633, "learning_rate": 9.687220644859275e-07, "loss": 1.133, "step": 3705 }, { "epoch": 0.9613889608706919, "grad_norm": 0.36877920673269027, "learning_rate": 9.069329189041464e-07, "loss": 1.1608, "step": 3710 }, { "epoch": 0.9626846333246956, "grad_norm": 0.381237649400668, "learning_rate": 8.471708618595142e-07, "loss": 1.132, "step": 3715 }, { "epoch": 0.9639803057786992, "grad_norm": 1.8685424500994194, "learning_rate": 7.894371158692627e-07, "loss": 1.1656, "step": 3720 }, { "epoch": 0.9652759782327027, "grad_norm": 0.38287612615523475, "learning_rate": 7.337328619586359e-07, "loss": 1.1992, "step": 3725 }, { "epoch": 0.9665716506867064, "grad_norm": 0.3838324502538139, "learning_rate": 6.800592396367545e-07, "loss": 1.1816, "step": 3730 }, { "epoch": 0.96786732314071, "grad_norm": 0.36218594513869495, "learning_rate": 6.284173468733334e-07, "loss": 1.1188, "step": 3735 }, { "epoch": 0.9691629955947136, "grad_norm": 0.3605898736094485, "learning_rate": 5.788082400761563e-07, "loss": 1.1342, "step": 3740 }, { "epoch": 0.9704586680487173, "grad_norm": 0.3819284101383782, "learning_rate": 5.312329340695143e-07, "loss": 1.1341, "step": 3745 }, { "epoch": 0.9717543405027209, "grad_norm": 0.371840798865179, "learning_rate": 4.856924020734565e-07, "loss": 1.1315, "step": 3750 }, { "epoch": 0.9730500129567246, "grad_norm": 0.37381875278150484, "learning_rate": 4.4218757568380563e-07, "loss": 1.1129, "step": 3755 }, { "epoch": 0.9743456854107282, "grad_norm": 0.37126771253731844, "learning_rate": 4.007193448532065e-07, "loss": 1.1666, "step": 3760 }, { "epoch": 0.9756413578647318, "grad_norm": 0.36988126308198593, "learning_rate": 3.612885578728298e-07, "loss": 1.1429, "step": 3765 }, { "epoch": 0.9769370303187355, "grad_norm": 0.3778202136491613, "learning_rate": 3.2389602135507457e-07, "loss": 1.1985, "step": 3770 }, { "epoch": 0.9782327027727391, "grad_norm": 0.3544356950249304, "learning_rate": 2.885425002170594e-07, "loss": 1.1469, "step": 3775 }, { "epoch": 0.9795283752267426, "grad_norm": 0.37874103417633515, "learning_rate": 2.5522871766494595e-07, "loss": 1.1101, "step": 3780 }, { "epoch": 0.9808240476807463, "grad_norm": 0.3788020452535586, "learning_rate": 2.2395535517917287e-07, "loss": 1.1895, "step": 3785 }, { "epoch": 0.9821197201347499, "grad_norm": 0.37493085740769366, "learning_rate": 1.947230525005006e-07, "loss": 1.1213, "step": 3790 }, { "epoch": 0.9834153925887535, "grad_norm": 0.36932806609876156, "learning_rate": 1.6753240761693268e-07, "loss": 1.1141, "step": 3795 }, { "epoch": 0.9847110650427572, "grad_norm": 0.3675515005536323, "learning_rate": 1.4238397675150339e-07, "loss": 1.1292, "step": 3800 }, { "epoch": 0.9860067374967608, "grad_norm": 0.3595954681947242, "learning_rate": 1.1927827435084248e-07, "loss": 1.1044, "step": 3805 }, { "epoch": 0.9873024099507645, "grad_norm": 0.3750795386829234, "learning_rate": 9.821577307470575e-08, "loss": 1.1441, "step": 3810 }, { "epoch": 0.9885980824047681, "grad_norm": 0.36749997185699007, "learning_rate": 7.919690378629385e-08, "loss": 1.171, "step": 3815 }, { "epoch": 0.9898937548587717, "grad_norm": 0.3622505687301632, "learning_rate": 6.222205554339277e-08, "loss": 1.1675, "step": 3820 }, { "epoch": 0.9911894273127754, "grad_norm": 0.3589951765733612, "learning_rate": 4.729157559049124e-08, "loss": 1.1451, "step": 3825 }, { "epoch": 0.992485099766779, "grad_norm": 0.372014869272016, "learning_rate": 3.440576935164197e-08, "loss": 1.1383, "step": 3830 }, { "epoch": 0.9937807722207825, "grad_norm": 0.36261244538328924, "learning_rate": 2.3564900424188906e-08, "loss": 1.1706, "step": 3835 }, { "epoch": 0.9950764446747862, "grad_norm": 0.37064889363490416, "learning_rate": 1.4769190573393765e-08, "loss": 1.1344, "step": 3840 }, { "epoch": 0.9963721171287898, "grad_norm": 0.4019334217280328, "learning_rate": 8.018819727928505e-09, "loss": 1.19, "step": 3845 }, { "epoch": 0.9976677895827935, "grad_norm": 0.3532944069166521, "learning_rate": 3.313925976156096e-09, "loss": 1.1704, "step": 3850 }, { "epoch": 0.9989634620367971, "grad_norm": 0.3954932358396543, "learning_rate": 6.546055633105397e-10, "loss": 1.0999, "step": 3855 }, { "epoch": 1.0, "eval_runtime": 3.6908, "eval_samples_per_second": 2.709, "eval_steps_per_second": 0.813, "step": 3859 }, { "epoch": 1.0, "step": 3859, "total_flos": 1.3011067827388416e+16, "train_loss": 1.2484874595598594, "train_runtime": 22644.1998, "train_samples_per_second": 2.726, "train_steps_per_second": 0.17 } ], "logging_steps": 5, "max_steps": 3859, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.3011067827388416e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }