diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5453 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 3859, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0002591344908007256, + "grad_norm": 111.01161666281861, + "learning_rate": 5.181347150259068e-07, + "loss": 12.3076, + "step": 1 + }, + { + "epoch": 0.001295672454003628, + "grad_norm": 128.99309976149883, + "learning_rate": 2.5906735751295338e-06, + "loss": 11.8519, + "step": 5 + }, + { + "epoch": 0.002591344908007256, + "grad_norm": 117.2928726132405, + "learning_rate": 5.1813471502590676e-06, + "loss": 12.3366, + "step": 10 + }, + { + "epoch": 0.0038870173620108835, + "grad_norm": 89.9905121675766, + "learning_rate": 7.772020725388602e-06, + "loss": 11.7319, + "step": 15 + }, + { + "epoch": 0.005182689816014512, + "grad_norm": 52.319681120581, + "learning_rate": 1.0362694300518135e-05, + "loss": 8.7838, + "step": 20 + }, + { + "epoch": 0.0064783622700181395, + "grad_norm": 38.24088557602701, + "learning_rate": 1.2953367875647668e-05, + "loss": 7.3055, + "step": 25 + }, + { + "epoch": 0.007774034724021767, + "grad_norm": 25.83067627606677, + "learning_rate": 1.5544041450777204e-05, + "loss": 5.7221, + "step": 30 + }, + { + "epoch": 0.009069707178025395, + "grad_norm": 7.762285591416621, + "learning_rate": 1.813471502590674e-05, + "loss": 4.3566, + "step": 35 + }, + { + "epoch": 0.010365379632029024, + "grad_norm": 4.409457180338291, + "learning_rate": 2.072538860103627e-05, + "loss": 3.6268, + "step": 40 + }, + { + "epoch": 0.011661052086032651, + "grad_norm": 2.601317803014797, + "learning_rate": 2.3316062176165805e-05, + "loss": 3.3055, + "step": 45 + }, + { + "epoch": 0.012956724540036279, + "grad_norm": 1.8400291034971308, + "learning_rate": 2.5906735751295337e-05, + "loss": 3.0095, + "step": 50 + }, + { + "epoch": 0.014252396994039906, + "grad_norm": 1.4104791862031891, + "learning_rate": 2.8497409326424872e-05, + "loss": 2.7918, + "step": 55 + }, + { + "epoch": 0.015548069448043534, + "grad_norm": 1.4652220793947772, + "learning_rate": 3.108808290155441e-05, + "loss": 2.7691, + "step": 60 + }, + { + "epoch": 0.01684374190204716, + "grad_norm": 2.1149598221315955, + "learning_rate": 3.367875647668394e-05, + "loss": 2.6915, + "step": 65 + }, + { + "epoch": 0.01813941435605079, + "grad_norm": 2.508562821895131, + "learning_rate": 3.626943005181348e-05, + "loss": 2.5864, + "step": 70 + }, + { + "epoch": 0.019435086810054417, + "grad_norm": 5.312266133287117, + "learning_rate": 3.886010362694301e-05, + "loss": 2.3482, + "step": 75 + }, + { + "epoch": 0.020730759264058048, + "grad_norm": 6.635762805683308, + "learning_rate": 4.145077720207254e-05, + "loss": 2.0166, + "step": 80 + }, + { + "epoch": 0.022026431718061675, + "grad_norm": 1.6009933261373699, + "learning_rate": 4.404145077720208e-05, + "loss": 1.5391, + "step": 85 + }, + { + "epoch": 0.023322104172065303, + "grad_norm": 1.390865987104561, + "learning_rate": 4.663212435233161e-05, + "loss": 1.3383, + "step": 90 + }, + { + "epoch": 0.02461777662606893, + "grad_norm": 1.0505820893868818, + "learning_rate": 4.922279792746114e-05, + "loss": 1.2646, + "step": 95 + }, + { + "epoch": 0.025913449080072558, + "grad_norm": 1.0086607725155037, + "learning_rate": 5.1813471502590674e-05, + "loss": 1.3202, + "step": 100 + }, + { + "epoch": 0.027209121534076185, + "grad_norm": 5.464697038037798, + "learning_rate": 5.440414507772021e-05, + "loss": 1.2451, + "step": 105 + }, + { + "epoch": 0.028504793988079813, + "grad_norm": 0.830545367098088, + "learning_rate": 5.6994818652849744e-05, + "loss": 1.2681, + "step": 110 + }, + { + "epoch": 0.02980046644208344, + "grad_norm": 0.7132598923355695, + "learning_rate": 5.9585492227979276e-05, + "loss": 1.2245, + "step": 115 + }, + { + "epoch": 0.031096138896087068, + "grad_norm": 0.8138882694476683, + "learning_rate": 6.217616580310881e-05, + "loss": 1.2305, + "step": 120 + }, + { + "epoch": 0.0323918113500907, + "grad_norm": 0.8807133122197233, + "learning_rate": 6.476683937823834e-05, + "loss": 1.2634, + "step": 125 + }, + { + "epoch": 0.03368748380409432, + "grad_norm": 0.62485091325362, + "learning_rate": 6.735751295336788e-05, + "loss": 1.3386, + "step": 130 + }, + { + "epoch": 0.034983156258097954, + "grad_norm": 0.7549667435061143, + "learning_rate": 6.994818652849742e-05, + "loss": 1.2513, + "step": 135 + }, + { + "epoch": 0.03627882871210158, + "grad_norm": 0.6420487687876172, + "learning_rate": 7.253886010362695e-05, + "loss": 1.2101, + "step": 140 + }, + { + "epoch": 0.03757450116610521, + "grad_norm": 0.7398721962367694, + "learning_rate": 7.512953367875648e-05, + "loss": 1.2776, + "step": 145 + }, + { + "epoch": 0.03887017362010883, + "grad_norm": 0.6450506158667421, + "learning_rate": 7.772020725388602e-05, + "loss": 1.2096, + "step": 150 + }, + { + "epoch": 0.040165846074112464, + "grad_norm": 0.6506841169994266, + "learning_rate": 8.031088082901554e-05, + "loss": 1.228, + "step": 155 + }, + { + "epoch": 0.041461518528116095, + "grad_norm": 0.711359178146542, + "learning_rate": 8.290155440414508e-05, + "loss": 1.2474, + "step": 160 + }, + { + "epoch": 0.04275719098211972, + "grad_norm": 1.459357949005984, + "learning_rate": 8.549222797927462e-05, + "loss": 1.2251, + "step": 165 + }, + { + "epoch": 0.04405286343612335, + "grad_norm": 0.8207401278130507, + "learning_rate": 8.808290155440416e-05, + "loss": 1.2625, + "step": 170 + }, + { + "epoch": 0.045348535890126974, + "grad_norm": 0.6271337633480082, + "learning_rate": 9.067357512953368e-05, + "loss": 1.2222, + "step": 175 + }, + { + "epoch": 0.046644208344130605, + "grad_norm": 0.6465301721957318, + "learning_rate": 9.326424870466322e-05, + "loss": 1.2357, + "step": 180 + }, + { + "epoch": 0.04793988079813423, + "grad_norm": 0.5864341660869246, + "learning_rate": 9.585492227979275e-05, + "loss": 1.2422, + "step": 185 + }, + { + "epoch": 0.04923555325213786, + "grad_norm": 0.5644516524286168, + "learning_rate": 9.844559585492228e-05, + "loss": 1.1901, + "step": 190 + }, + { + "epoch": 0.050531225706141485, + "grad_norm": 0.654016063568062, + "learning_rate": 0.00010103626943005182, + "loss": 1.182, + "step": 195 + }, + { + "epoch": 0.051826898160145116, + "grad_norm": 0.569272168627719, + "learning_rate": 0.00010362694300518135, + "loss": 1.1958, + "step": 200 + }, + { + "epoch": 0.05312257061414875, + "grad_norm": 0.6368680847682564, + "learning_rate": 0.00010621761658031089, + "loss": 1.2694, + "step": 205 + }, + { + "epoch": 0.05441824306815237, + "grad_norm": 0.5948244637123268, + "learning_rate": 0.00010880829015544042, + "loss": 1.1967, + "step": 210 + }, + { + "epoch": 0.055713915522156, + "grad_norm": 0.5541951524654064, + "learning_rate": 0.00011139896373056995, + "loss": 1.1941, + "step": 215 + }, + { + "epoch": 0.057009587976159626, + "grad_norm": 0.5701648418604164, + "learning_rate": 0.00011398963730569949, + "loss": 1.2443, + "step": 220 + }, + { + "epoch": 0.05830526043016326, + "grad_norm": 0.4865998338825169, + "learning_rate": 0.00011658031088082901, + "loss": 1.2106, + "step": 225 + }, + { + "epoch": 0.05960093288416688, + "grad_norm": 0.5131095922403084, + "learning_rate": 0.00011917098445595855, + "loss": 1.1278, + "step": 230 + }, + { + "epoch": 0.06089660533817051, + "grad_norm": 0.6563988864261227, + "learning_rate": 0.0001217616580310881, + "loss": 1.1991, + "step": 235 + }, + { + "epoch": 0.062192277792174136, + "grad_norm": 0.5769507376758363, + "learning_rate": 0.00012435233160621763, + "loss": 1.1674, + "step": 240 + }, + { + "epoch": 0.06348795024617776, + "grad_norm": 0.5283246008008212, + "learning_rate": 0.00012694300518134715, + "loss": 1.2367, + "step": 245 + }, + { + "epoch": 0.0647836227001814, + "grad_norm": 0.5536285028186603, + "learning_rate": 0.00012953367875647668, + "loss": 1.1649, + "step": 250 + }, + { + "epoch": 0.06607929515418502, + "grad_norm": 0.5163473385172975, + "learning_rate": 0.00013212435233160623, + "loss": 1.2867, + "step": 255 + }, + { + "epoch": 0.06737496760818865, + "grad_norm": 0.5475813868100357, + "learning_rate": 0.00013471502590673575, + "loss": 1.2152, + "step": 260 + }, + { + "epoch": 0.06867064006219228, + "grad_norm": 0.511381640001362, + "learning_rate": 0.00013730569948186528, + "loss": 1.2065, + "step": 265 + }, + { + "epoch": 0.06996631251619591, + "grad_norm": 0.4982475298525502, + "learning_rate": 0.00013989637305699483, + "loss": 1.2142, + "step": 270 + }, + { + "epoch": 0.07126198497019953, + "grad_norm": 0.5433832176160214, + "learning_rate": 0.00014248704663212436, + "loss": 1.2276, + "step": 275 + }, + { + "epoch": 0.07255765742420316, + "grad_norm": 0.47834421596861043, + "learning_rate": 0.0001450777202072539, + "loss": 1.1183, + "step": 280 + }, + { + "epoch": 0.0738533298782068, + "grad_norm": 2.276620321866631, + "learning_rate": 0.0001476683937823834, + "loss": 1.2225, + "step": 285 + }, + { + "epoch": 0.07514900233221042, + "grad_norm": 0.5622928799874418, + "learning_rate": 0.00015025906735751296, + "loss": 1.142, + "step": 290 + }, + { + "epoch": 0.07644467478621404, + "grad_norm": 0.48315577966120404, + "learning_rate": 0.0001528497409326425, + "loss": 1.1481, + "step": 295 + }, + { + "epoch": 0.07774034724021767, + "grad_norm": 0.5054965652645259, + "learning_rate": 0.00015544041450777204, + "loss": 1.1334, + "step": 300 + }, + { + "epoch": 0.0790360196942213, + "grad_norm": 0.5412094676646951, + "learning_rate": 0.00015803108808290156, + "loss": 1.2215, + "step": 305 + }, + { + "epoch": 0.08033169214822493, + "grad_norm": 0.489058036506764, + "learning_rate": 0.00016062176165803108, + "loss": 1.1651, + "step": 310 + }, + { + "epoch": 0.08162736460222855, + "grad_norm": 0.5225993706207775, + "learning_rate": 0.00016321243523316064, + "loss": 1.173, + "step": 315 + }, + { + "epoch": 0.08292303705623219, + "grad_norm": 0.4941052544991184, + "learning_rate": 0.00016580310880829016, + "loss": 1.1537, + "step": 320 + }, + { + "epoch": 0.08421870951023581, + "grad_norm": 0.5317455067919804, + "learning_rate": 0.0001683937823834197, + "loss": 1.1702, + "step": 325 + }, + { + "epoch": 0.08551438196423944, + "grad_norm": 0.5382129937077951, + "learning_rate": 0.00017098445595854924, + "loss": 1.1954, + "step": 330 + }, + { + "epoch": 0.08681005441824306, + "grad_norm": 0.4873382544458068, + "learning_rate": 0.00017357512953367876, + "loss": 1.1292, + "step": 335 + }, + { + "epoch": 0.0881057268722467, + "grad_norm": 0.869634547296972, + "learning_rate": 0.00017616580310880832, + "loss": 1.2342, + "step": 340 + }, + { + "epoch": 0.08940139932625032, + "grad_norm": 0.47105707267021146, + "learning_rate": 0.0001787564766839378, + "loss": 1.1903, + "step": 345 + }, + { + "epoch": 0.09069707178025395, + "grad_norm": 0.483761714583125, + "learning_rate": 0.00018134715025906737, + "loss": 1.1753, + "step": 350 + }, + { + "epoch": 0.09199274423425757, + "grad_norm": 0.5094212322859152, + "learning_rate": 0.00018393782383419692, + "loss": 1.1736, + "step": 355 + }, + { + "epoch": 0.09328841668826121, + "grad_norm": 0.49665577393117427, + "learning_rate": 0.00018652849740932644, + "loss": 1.19, + "step": 360 + }, + { + "epoch": 0.09458408914226483, + "grad_norm": 0.4634306042294334, + "learning_rate": 0.00018911917098445597, + "loss": 1.2068, + "step": 365 + }, + { + "epoch": 0.09587976159626846, + "grad_norm": 0.4771955076541271, + "learning_rate": 0.0001917098445595855, + "loss": 1.1746, + "step": 370 + }, + { + "epoch": 0.0971754340502721, + "grad_norm": 0.49488098544084536, + "learning_rate": 0.00019430051813471504, + "loss": 1.1781, + "step": 375 + }, + { + "epoch": 0.09847110650427572, + "grad_norm": 0.45559447847276796, + "learning_rate": 0.00019689119170984457, + "loss": 1.207, + "step": 380 + }, + { + "epoch": 0.09976677895827935, + "grad_norm": 0.530644388105199, + "learning_rate": 0.0001994818652849741, + "loss": 1.2059, + "step": 385 + }, + { + "epoch": 0.10106245141228297, + "grad_norm": 0.4262101606219483, + "learning_rate": 0.0001999993453944367, + "loss": 1.1728, + "step": 390 + }, + { + "epoch": 0.10235812386628661, + "grad_norm": 0.5019930710713301, + "learning_rate": 0.00019999668607402385, + "loss": 1.1566, + "step": 395 + }, + { + "epoch": 0.10365379632029023, + "grad_norm": 0.41212142912399047, + "learning_rate": 0.00019999198118027207, + "loss": 1.1367, + "step": 400 + }, + { + "epoch": 0.10494946877429386, + "grad_norm": 0.4880776387603866, + "learning_rate": 0.00019998523080942663, + "loss": 1.1654, + "step": 405 + }, + { + "epoch": 0.1062451412282975, + "grad_norm": 0.46286032679323585, + "learning_rate": 0.00019997643509957582, + "loss": 1.1944, + "step": 410 + }, + { + "epoch": 0.10754081368230112, + "grad_norm": 0.5382364265983718, + "learning_rate": 0.00019996559423064838, + "loss": 1.1871, + "step": 415 + }, + { + "epoch": 0.10883648613630474, + "grad_norm": 0.5437136001752437, + "learning_rate": 0.0001999527084244095, + "loss": 1.1211, + "step": 420 + }, + { + "epoch": 0.11013215859030837, + "grad_norm": 0.4683209703143848, + "learning_rate": 0.00019993777794445662, + "loss": 1.2111, + "step": 425 + }, + { + "epoch": 0.111427831044312, + "grad_norm": 0.43839848429469386, + "learning_rate": 0.00019992080309621371, + "loss": 1.1655, + "step": 430 + }, + { + "epoch": 0.11272350349831563, + "grad_norm": 0.4352884542992481, + "learning_rate": 0.00019990178422692528, + "loss": 1.1674, + "step": 435 + }, + { + "epoch": 0.11401917595231925, + "grad_norm": 0.45552967972039005, + "learning_rate": 0.00019988072172564918, + "loss": 1.2005, + "step": 440 + }, + { + "epoch": 0.11531484840632288, + "grad_norm": 0.46569070576531457, + "learning_rate": 0.0001998576160232485, + "loss": 1.1611, + "step": 445 + }, + { + "epoch": 0.11661052086032651, + "grad_norm": 0.4344558621258368, + "learning_rate": 0.00019983246759238305, + "loss": 1.1632, + "step": 450 + }, + { + "epoch": 0.11790619331433014, + "grad_norm": 0.4524631212974695, + "learning_rate": 0.00019980527694749952, + "loss": 1.2003, + "step": 455 + }, + { + "epoch": 0.11920186576833376, + "grad_norm": 0.692041740038434, + "learning_rate": 0.00019977604464482083, + "loss": 1.233, + "step": 460 + }, + { + "epoch": 0.1204975382223374, + "grad_norm": 0.4499116414766127, + "learning_rate": 0.00019974477128233505, + "loss": 1.1431, + "step": 465 + }, + { + "epoch": 0.12179321067634102, + "grad_norm": 0.4755303917497374, + "learning_rate": 0.00019971145749978294, + "loss": 1.186, + "step": 470 + }, + { + "epoch": 0.12308888313034465, + "grad_norm": 0.4587427439073901, + "learning_rate": 0.00019967610397864493, + "loss": 1.1772, + "step": 475 + }, + { + "epoch": 0.12438455558434827, + "grad_norm": 0.43004396404922146, + "learning_rate": 0.0001996387114421272, + "loss": 1.1613, + "step": 480 + }, + { + "epoch": 0.1256802280383519, + "grad_norm": 0.4457323824719563, + "learning_rate": 0.0001995992806551468, + "loss": 1.2043, + "step": 485 + }, + { + "epoch": 0.12697590049235552, + "grad_norm": 0.46784030959656325, + "learning_rate": 0.00019955781242431622, + "loss": 1.1678, + "step": 490 + }, + { + "epoch": 0.12827157294635916, + "grad_norm": 0.4610982688278206, + "learning_rate": 0.00019951430759792654, + "loss": 1.2353, + "step": 495 + }, + { + "epoch": 0.1295672454003628, + "grad_norm": 0.4259792086854904, + "learning_rate": 0.0001994687670659305, + "loss": 1.1612, + "step": 500 + }, + { + "epoch": 0.1308629178543664, + "grad_norm": 0.45520840895219594, + "learning_rate": 0.00019942119175992383, + "loss": 1.099, + "step": 505 + }, + { + "epoch": 0.13215859030837004, + "grad_norm": 0.4473123797320602, + "learning_rate": 0.00019937158265312667, + "loss": 1.2141, + "step": 510 + }, + { + "epoch": 0.13345426276237368, + "grad_norm": 0.4713319110562635, + "learning_rate": 0.00019931994076036324, + "loss": 1.1454, + "step": 515 + }, + { + "epoch": 0.1347499352163773, + "grad_norm": 0.5136825790567691, + "learning_rate": 0.00019926626713804137, + "loss": 1.2187, + "step": 520 + }, + { + "epoch": 0.13604560767038093, + "grad_norm": 0.4727966110942116, + "learning_rate": 0.00019921056288413076, + "loss": 1.1988, + "step": 525 + }, + { + "epoch": 0.13734128012438457, + "grad_norm": 0.4364604236158943, + "learning_rate": 0.00019915282913814052, + "loss": 1.1685, + "step": 530 + }, + { + "epoch": 0.13863695257838818, + "grad_norm": 0.48078362848188844, + "learning_rate": 0.00019909306708109585, + "loss": 1.1773, + "step": 535 + }, + { + "epoch": 0.13993262503239182, + "grad_norm": 0.45186495929915804, + "learning_rate": 0.00019903127793551408, + "loss": 1.1409, + "step": 540 + }, + { + "epoch": 0.14122829748639543, + "grad_norm": 0.4557721517395367, + "learning_rate": 0.0001989674629653793, + "loss": 1.1243, + "step": 545 + }, + { + "epoch": 0.14252396994039906, + "grad_norm": 0.47938538143807347, + "learning_rate": 0.00019890162347611687, + "loss": 1.1644, + "step": 550 + }, + { + "epoch": 0.1438196423944027, + "grad_norm": 0.4541147395460418, + "learning_rate": 0.00019883376081456636, + "loss": 1.2578, + "step": 555 + }, + { + "epoch": 0.1451153148484063, + "grad_norm": 0.4780871852039313, + "learning_rate": 0.00019876387636895437, + "loss": 1.206, + "step": 560 + }, + { + "epoch": 0.14641098730240995, + "grad_norm": 0.42628778420968705, + "learning_rate": 0.00019869197156886586, + "loss": 1.1728, + "step": 565 + }, + { + "epoch": 0.1477066597564136, + "grad_norm": 0.4303201965277035, + "learning_rate": 0.00019861804788521493, + "loss": 1.1807, + "step": 570 + }, + { + "epoch": 0.1490023322104172, + "grad_norm": 0.48120812526739415, + "learning_rate": 0.00019854210683021485, + "loss": 1.1825, + "step": 575 + }, + { + "epoch": 0.15029800466442084, + "grad_norm": 0.479571534871543, + "learning_rate": 0.0001984641499573472, + "loss": 1.1213, + "step": 580 + }, + { + "epoch": 0.15159367711842447, + "grad_norm": 0.4517330978325847, + "learning_rate": 0.0001983841788613297, + "loss": 1.1799, + "step": 585 + }, + { + "epoch": 0.15288934957242808, + "grad_norm": 0.45662716170150547, + "learning_rate": 0.00019830219517808404, + "loss": 1.2154, + "step": 590 + }, + { + "epoch": 0.15418502202643172, + "grad_norm": 0.46419741725194397, + "learning_rate": 0.00019821820058470215, + "loss": 1.2244, + "step": 595 + }, + { + "epoch": 0.15548069448043533, + "grad_norm": 0.48643943236229664, + "learning_rate": 0.00019813219679941203, + "loss": 1.1773, + "step": 600 + }, + { + "epoch": 0.15677636693443897, + "grad_norm": 0.4213189862656036, + "learning_rate": 0.00019804418558154243, + "loss": 1.1918, + "step": 605 + }, + { + "epoch": 0.1580720393884426, + "grad_norm": 0.4039792149279076, + "learning_rate": 0.00019795416873148703, + "loss": 1.1937, + "step": 610 + }, + { + "epoch": 0.15936771184244622, + "grad_norm": 0.42408763347068035, + "learning_rate": 0.00019786214809066753, + "loss": 1.1792, + "step": 615 + }, + { + "epoch": 0.16066338429644986, + "grad_norm": 0.4418941392259797, + "learning_rate": 0.00019776812554149603, + "loss": 1.0983, + "step": 620 + }, + { + "epoch": 0.1619590567504535, + "grad_norm": 0.44778137756258257, + "learning_rate": 0.00019767210300733647, + "loss": 1.1441, + "step": 625 + }, + { + "epoch": 0.1632547292044571, + "grad_norm": 0.4214693702892699, + "learning_rate": 0.0001975740824524653, + "loss": 1.126, + "step": 630 + }, + { + "epoch": 0.16455040165846074, + "grad_norm": 0.431868681760054, + "learning_rate": 0.00019747406588203128, + "loss": 1.2244, + "step": 635 + }, + { + "epoch": 0.16584607411246438, + "grad_norm": 0.41521488211670027, + "learning_rate": 0.00019737205534201455, + "loss": 1.1443, + "step": 640 + }, + { + "epoch": 0.167141746566468, + "grad_norm": 0.4141916874944073, + "learning_rate": 0.00019726805291918464, + "loss": 1.2012, + "step": 645 + }, + { + "epoch": 0.16843741902047163, + "grad_norm": 0.8019720904500388, + "learning_rate": 0.0001971620607410579, + "loss": 1.1871, + "step": 650 + }, + { + "epoch": 0.16973309147447524, + "grad_norm": 0.44302123796014203, + "learning_rate": 0.00019705408097585393, + "loss": 1.217, + "step": 655 + }, + { + "epoch": 0.17102876392847888, + "grad_norm": 0.44266067140812565, + "learning_rate": 0.0001969441158324512, + "loss": 1.159, + "step": 660 + }, + { + "epoch": 0.17232443638248252, + "grad_norm": 0.44647716533971304, + "learning_rate": 0.0001968321675603419, + "loss": 1.1653, + "step": 665 + }, + { + "epoch": 0.17362010883648613, + "grad_norm": 0.44623301312377617, + "learning_rate": 0.000196718238449586, + "loss": 1.2078, + "step": 670 + }, + { + "epoch": 0.17491578129048976, + "grad_norm": 0.41505140877535734, + "learning_rate": 0.00019660233083076416, + "loss": 1.1727, + "step": 675 + }, + { + "epoch": 0.1762114537444934, + "grad_norm": 0.4603289406918021, + "learning_rate": 0.00019648444707493035, + "loss": 1.2151, + "step": 680 + }, + { + "epoch": 0.177507126198497, + "grad_norm": 0.4185225256437822, + "learning_rate": 0.00019636458959356316, + "loss": 1.1708, + "step": 685 + }, + { + "epoch": 0.17880279865250065, + "grad_norm": 0.4207300329188764, + "learning_rate": 0.00019624276083851655, + "loss": 1.1821, + "step": 690 + }, + { + "epoch": 0.1800984711065043, + "grad_norm": 0.44173498052158916, + "learning_rate": 0.00019611896330196956, + "loss": 1.175, + "step": 695 + }, + { + "epoch": 0.1813941435605079, + "grad_norm": 0.43648924778909787, + "learning_rate": 0.0001959931995163756, + "loss": 1.1409, + "step": 700 + }, + { + "epoch": 0.18268981601451154, + "grad_norm": 0.41607673192874983, + "learning_rate": 0.00019586547205441038, + "loss": 1.1903, + "step": 705 + }, + { + "epoch": 0.18398548846851515, + "grad_norm": 0.40413543802190666, + "learning_rate": 0.00019573578352891937, + "loss": 1.1395, + "step": 710 + }, + { + "epoch": 0.18528116092251878, + "grad_norm": 0.45031653314971887, + "learning_rate": 0.00019560413659286437, + "loss": 1.1551, + "step": 715 + }, + { + "epoch": 0.18657683337652242, + "grad_norm": 0.4308169448866955, + "learning_rate": 0.00019547053393926934, + "loss": 1.1875, + "step": 720 + }, + { + "epoch": 0.18787250583052603, + "grad_norm": 0.42706739288180534, + "learning_rate": 0.000195334978301165, + "loss": 1.1828, + "step": 725 + }, + { + "epoch": 0.18916817828452967, + "grad_norm": 0.43201723960587324, + "learning_rate": 0.00019519747245153333, + "loss": 1.1269, + "step": 730 + }, + { + "epoch": 0.1904638507385333, + "grad_norm": 0.45113312626512714, + "learning_rate": 0.0001950580192032505, + "loss": 1.228, + "step": 735 + }, + { + "epoch": 0.19175952319253692, + "grad_norm": 0.44161985380394386, + "learning_rate": 0.0001949166214090295, + "loss": 1.1242, + "step": 740 + }, + { + "epoch": 0.19305519564654056, + "grad_norm": 0.46536800518090093, + "learning_rate": 0.00019477328196136178, + "loss": 1.205, + "step": 745 + }, + { + "epoch": 0.1943508681005442, + "grad_norm": 0.4304870726759157, + "learning_rate": 0.00019462800379245807, + "loss": 1.1677, + "step": 750 + }, + { + "epoch": 0.1956465405545478, + "grad_norm": 0.4500472206954455, + "learning_rate": 0.0001944807898741883, + "loss": 1.2031, + "step": 755 + }, + { + "epoch": 0.19694221300855144, + "grad_norm": 0.4270521213996621, + "learning_rate": 0.00019433164321802095, + "loss": 1.221, + "step": 760 + }, + { + "epoch": 0.19823788546255505, + "grad_norm": 0.4500335364742716, + "learning_rate": 0.00019418056687496135, + "loss": 1.2206, + "step": 765 + }, + { + "epoch": 0.1995335579165587, + "grad_norm": 0.46760662103580936, + "learning_rate": 0.00019402756393548936, + "loss": 1.2628, + "step": 770 + }, + { + "epoch": 0.20082923037056233, + "grad_norm": 0.4374227196524263, + "learning_rate": 0.00019387263752949598, + "loss": 1.1895, + "step": 775 + }, + { + "epoch": 0.20212490282456594, + "grad_norm": 0.44423980739831237, + "learning_rate": 0.00019371579082621952, + "loss": 1.1915, + "step": 780 + }, + { + "epoch": 0.20342057527856958, + "grad_norm": 0.5067585361087695, + "learning_rate": 0.00019355702703418063, + "loss": 1.1396, + "step": 785 + }, + { + "epoch": 0.20471624773257321, + "grad_norm": 0.4397713592175906, + "learning_rate": 0.0001933963494011168, + "loss": 1.1829, + "step": 790 + }, + { + "epoch": 0.20601192018657682, + "grad_norm": 0.41142670533283804, + "learning_rate": 0.0001932337612139157, + "loss": 1.1403, + "step": 795 + }, + { + "epoch": 0.20730759264058046, + "grad_norm": 0.431969506585272, + "learning_rate": 0.00019306926579854821, + "loss": 1.1219, + "step": 800 + }, + { + "epoch": 0.2086032650945841, + "grad_norm": 0.44265249198309475, + "learning_rate": 0.00019290286652000018, + "loss": 1.1897, + "step": 805 + }, + { + "epoch": 0.2098989375485877, + "grad_norm": 0.4277657316516301, + "learning_rate": 0.0001927345667822037, + "loss": 1.1746, + "step": 810 + }, + { + "epoch": 0.21119461000259135, + "grad_norm": 0.40123037504170683, + "learning_rate": 0.00019256437002796744, + "loss": 1.1775, + "step": 815 + }, + { + "epoch": 0.212490282456595, + "grad_norm": 0.4250319228234625, + "learning_rate": 0.00019239227973890622, + "loss": 1.1357, + "step": 820 + }, + { + "epoch": 0.2137859549105986, + "grad_norm": 0.469142216595277, + "learning_rate": 0.0001922182994353697, + "loss": 1.1709, + "step": 825 + }, + { + "epoch": 0.21508162736460223, + "grad_norm": 0.4205186026441914, + "learning_rate": 0.0001920424326763706, + "loss": 1.1846, + "step": 830 + }, + { + "epoch": 0.21637729981860584, + "grad_norm": 0.4364780340562057, + "learning_rate": 0.00019186468305951165, + "loss": 1.1611, + "step": 835 + }, + { + "epoch": 0.21767297227260948, + "grad_norm": 0.42862545463693674, + "learning_rate": 0.00019168505422091214, + "loss": 1.2604, + "step": 840 + }, + { + "epoch": 0.21896864472661312, + "grad_norm": 0.4298577312506422, + "learning_rate": 0.00019150354983513346, + "loss": 1.1584, + "step": 845 + }, + { + "epoch": 0.22026431718061673, + "grad_norm": 0.4188255164345959, + "learning_rate": 0.00019132017361510396, + "loss": 1.1681, + "step": 850 + }, + { + "epoch": 0.22155998963462037, + "grad_norm": 0.42610008321144127, + "learning_rate": 0.00019113492931204304, + "loss": 1.2502, + "step": 855 + }, + { + "epoch": 0.222855662088624, + "grad_norm": 0.41917044829804134, + "learning_rate": 0.00019094782071538434, + "loss": 1.1441, + "step": 860 + }, + { + "epoch": 0.22415133454262762, + "grad_norm": 0.41937866972534293, + "learning_rate": 0.0001907588516526983, + "loss": 1.2056, + "step": 865 + }, + { + "epoch": 0.22544700699663126, + "grad_norm": 0.47221544899406714, + "learning_rate": 0.00019056802598961376, + "loss": 1.1754, + "step": 870 + }, + { + "epoch": 0.2267426794506349, + "grad_norm": 0.45128209262811003, + "learning_rate": 0.000190375347629739, + "loss": 1.1963, + "step": 875 + }, + { + "epoch": 0.2280383519046385, + "grad_norm": 0.41254887942116875, + "learning_rate": 0.00019018082051458176, + "loss": 1.1696, + "step": 880 + }, + { + "epoch": 0.22933402435864214, + "grad_norm": 0.4336535335825639, + "learning_rate": 0.00018998444862346873, + "loss": 1.1942, + "step": 885 + }, + { + "epoch": 0.23062969681264575, + "grad_norm": 0.4203068249880088, + "learning_rate": 0.00018978623597346408, + "loss": 1.1571, + "step": 890 + }, + { + "epoch": 0.2319253692666494, + "grad_norm": 0.4504420384371143, + "learning_rate": 0.00018958618661928732, + "loss": 1.2145, + "step": 895 + }, + { + "epoch": 0.23322104172065303, + "grad_norm": 1.2852982602777483, + "learning_rate": 0.00018938430465323034, + "loss": 1.1602, + "step": 900 + }, + { + "epoch": 0.23451671417465664, + "grad_norm": 0.4456499671770626, + "learning_rate": 0.0001891805942050736, + "loss": 1.2208, + "step": 905 + }, + { + "epoch": 0.23581238662866028, + "grad_norm": 0.4218488577208689, + "learning_rate": 0.00018897505944200186, + "loss": 1.1544, + "step": 910 + }, + { + "epoch": 0.2371080590826639, + "grad_norm": 0.4509293570117875, + "learning_rate": 0.00018876770456851877, + "loss": 1.2167, + "step": 915 + }, + { + "epoch": 0.23840373153666752, + "grad_norm": 0.4354264628463739, + "learning_rate": 0.00018855853382636093, + "loss": 1.2103, + "step": 920 + }, + { + "epoch": 0.23969940399067116, + "grad_norm": 0.41330759489370955, + "learning_rate": 0.00018834755149441104, + "loss": 1.1713, + "step": 925 + }, + { + "epoch": 0.2409950764446748, + "grad_norm": 0.43531782144147513, + "learning_rate": 0.00018813476188861043, + "loss": 1.1783, + "step": 930 + }, + { + "epoch": 0.2422907488986784, + "grad_norm": 0.4260507531562403, + "learning_rate": 0.00018792016936187086, + "loss": 1.1473, + "step": 935 + }, + { + "epoch": 0.24358642135268205, + "grad_norm": 0.46161844432613697, + "learning_rate": 0.00018770377830398525, + "loss": 1.1719, + "step": 940 + }, + { + "epoch": 0.24488209380668566, + "grad_norm": 0.44108410378556556, + "learning_rate": 0.00018748559314153818, + "loss": 1.2101, + "step": 945 + }, + { + "epoch": 0.2461777662606893, + "grad_norm": 0.43144293315888466, + "learning_rate": 0.00018726561833781497, + "loss": 1.1933, + "step": 950 + }, + { + "epoch": 0.24747343871469293, + "grad_norm": 0.437185606075459, + "learning_rate": 0.00018704385839271074, + "loss": 1.1563, + "step": 955 + }, + { + "epoch": 0.24876911116869654, + "grad_norm": 0.44187547375302677, + "learning_rate": 0.00018682031784263814, + "loss": 1.1219, + "step": 960 + }, + { + "epoch": 0.2500647836227002, + "grad_norm": 0.4286003185934626, + "learning_rate": 0.00018659500126043456, + "loss": 1.1347, + "step": 965 + }, + { + "epoch": 0.2513604560767038, + "grad_norm": 0.5193197943854574, + "learning_rate": 0.00018636791325526872, + "loss": 1.2118, + "step": 970 + }, + { + "epoch": 0.25265612853070746, + "grad_norm": 0.4243801823353042, + "learning_rate": 0.00018613905847254624, + "loss": 1.2015, + "step": 975 + }, + { + "epoch": 0.25395180098471104, + "grad_norm": 0.4017619629276402, + "learning_rate": 0.0001859084415938147, + "loss": 1.1405, + "step": 980 + }, + { + "epoch": 0.2552474734387147, + "grad_norm": 0.46780432324474447, + "learning_rate": 0.00018567606733666775, + "loss": 1.1803, + "step": 985 + }, + { + "epoch": 0.2565431458927183, + "grad_norm": 0.41305152773119796, + "learning_rate": 0.00018544194045464886, + "loss": 1.138, + "step": 990 + }, + { + "epoch": 0.25783881834672195, + "grad_norm": 0.4090708833399176, + "learning_rate": 0.0001852060657371538, + "loss": 1.2036, + "step": 995 + }, + { + "epoch": 0.2591344908007256, + "grad_norm": 0.43431398541973126, + "learning_rate": 0.00018496844800933277, + "loss": 1.1328, + "step": 1000 + }, + { + "epoch": 0.26043016325472923, + "grad_norm": 0.4473287652132595, + "learning_rate": 0.0001847290921319918, + "loss": 1.183, + "step": 1005 + }, + { + "epoch": 0.2617258357087328, + "grad_norm": 0.42139497138720455, + "learning_rate": 0.00018448800300149314, + "loss": 1.1885, + "step": 1010 + }, + { + "epoch": 0.26302150816273645, + "grad_norm": 0.4192742651155342, + "learning_rate": 0.00018424518554965516, + "loss": 1.1673, + "step": 1015 + }, + { + "epoch": 0.2643171806167401, + "grad_norm": 0.4276967282946324, + "learning_rate": 0.00018400064474365156, + "loss": 1.1442, + "step": 1020 + }, + { + "epoch": 0.2656128530707437, + "grad_norm": 0.4279959695376702, + "learning_rate": 0.00018375438558590967, + "loss": 1.1713, + "step": 1025 + }, + { + "epoch": 0.26690852552474736, + "grad_norm": 0.4138904922875282, + "learning_rate": 0.00018350641311400812, + "loss": 1.1294, + "step": 1030 + }, + { + "epoch": 0.26820419797875095, + "grad_norm": 0.43337813322172525, + "learning_rate": 0.0001832567324005737, + "loss": 1.1421, + "step": 1035 + }, + { + "epoch": 0.2694998704327546, + "grad_norm": 0.4418444877860619, + "learning_rate": 0.00018300534855317783, + "loss": 1.1775, + "step": 1040 + }, + { + "epoch": 0.2707955428867582, + "grad_norm": 0.46191806443855654, + "learning_rate": 0.00018275226671423195, + "loss": 1.1188, + "step": 1045 + }, + { + "epoch": 0.27209121534076186, + "grad_norm": 0.4449373885557454, + "learning_rate": 0.0001824974920608821, + "loss": 1.1445, + "step": 1050 + }, + { + "epoch": 0.2733868877947655, + "grad_norm": 0.4138240087897747, + "learning_rate": 0.0001822410298049035, + "loss": 1.1403, + "step": 1055 + }, + { + "epoch": 0.27468256024876914, + "grad_norm": 0.4410582459345502, + "learning_rate": 0.00018198288519259353, + "loss": 1.1835, + "step": 1060 + }, + { + "epoch": 0.2759782327027727, + "grad_norm": 0.4419431606975578, + "learning_rate": 0.0001817230635046645, + "loss": 1.2308, + "step": 1065 + }, + { + "epoch": 0.27727390515677636, + "grad_norm": 0.41612794192020086, + "learning_rate": 0.0001814615700561358, + "loss": 1.21, + "step": 1070 + }, + { + "epoch": 0.27856957761078, + "grad_norm": 0.4728861496942093, + "learning_rate": 0.00018119841019622487, + "loss": 1.1687, + "step": 1075 + }, + { + "epoch": 0.27986525006478363, + "grad_norm": 0.39483035476368555, + "learning_rate": 0.0001809335893082381, + "loss": 1.2031, + "step": 1080 + }, + { + "epoch": 0.28116092251878727, + "grad_norm": 0.40167341751112157, + "learning_rate": 0.0001806671128094605, + "loss": 1.1495, + "step": 1085 + }, + { + "epoch": 0.28245659497279085, + "grad_norm": 0.42702634032994197, + "learning_rate": 0.0001803989861510449, + "loss": 1.1885, + "step": 1090 + }, + { + "epoch": 0.2837522674267945, + "grad_norm": 0.3970503569404702, + "learning_rate": 0.00018012921481790054, + "loss": 1.1608, + "step": 1095 + }, + { + "epoch": 0.28504793988079813, + "grad_norm": 0.42637878811973734, + "learning_rate": 0.0001798578043285807, + "loss": 1.1386, + "step": 1100 + }, + { + "epoch": 0.28634361233480177, + "grad_norm": 0.39626715470331225, + "learning_rate": 0.00017958476023517008, + "loss": 1.1355, + "step": 1105 + }, + { + "epoch": 0.2876392847888054, + "grad_norm": 0.4158526138591002, + "learning_rate": 0.00017931008812317089, + "loss": 1.1844, + "step": 1110 + }, + { + "epoch": 0.28893495724280904, + "grad_norm": 0.4110580138866959, + "learning_rate": 0.00017903379361138884, + "loss": 1.179, + "step": 1115 + }, + { + "epoch": 0.2902306296968126, + "grad_norm": 0.39558148991812647, + "learning_rate": 0.0001787558823518181, + "loss": 1.1529, + "step": 1120 + }, + { + "epoch": 0.29152630215081626, + "grad_norm": 0.43235047146868255, + "learning_rate": 0.0001784763600295257, + "loss": 1.1253, + "step": 1125 + }, + { + "epoch": 0.2928219746048199, + "grad_norm": 0.4029217795192826, + "learning_rate": 0.00017819523236253524, + "loss": 1.1775, + "step": 1130 + }, + { + "epoch": 0.29411764705882354, + "grad_norm": 0.43737260937528, + "learning_rate": 0.0001779125051017099, + "loss": 1.1658, + "step": 1135 + }, + { + "epoch": 0.2954133195128272, + "grad_norm": 0.4223589589109796, + "learning_rate": 0.00017762818403063485, + "loss": 1.2193, + "step": 1140 + }, + { + "epoch": 0.29670899196683076, + "grad_norm": 0.42153072570772854, + "learning_rate": 0.0001773422749654988, + "loss": 1.1381, + "step": 1145 + }, + { + "epoch": 0.2980046644208344, + "grad_norm": 0.44333235395249593, + "learning_rate": 0.0001770547837549752, + "loss": 1.2025, + "step": 1150 + }, + { + "epoch": 0.29930033687483804, + "grad_norm": 0.390430213600016, + "learning_rate": 0.0001767657162801025, + "loss": 1.2179, + "step": 1155 + }, + { + "epoch": 0.3005960093288417, + "grad_norm": 0.39596734484692325, + "learning_rate": 0.00017647507845416392, + "loss": 1.1443, + "step": 1160 + }, + { + "epoch": 0.3018916817828453, + "grad_norm": 0.3964805626892672, + "learning_rate": 0.00017618287622256625, + "loss": 1.0953, + "step": 1165 + }, + { + "epoch": 0.30318735423684895, + "grad_norm": 0.397165519182473, + "learning_rate": 0.00017588911556271858, + "loss": 1.2468, + "step": 1170 + }, + { + "epoch": 0.30448302669085253, + "grad_norm": 0.40738383895968944, + "learning_rate": 0.00017559380248390982, + "loss": 1.2279, + "step": 1175 + }, + { + "epoch": 0.30577869914485617, + "grad_norm": 0.4322556749554571, + "learning_rate": 0.00017529694302718574, + "loss": 1.1614, + "step": 1180 + }, + { + "epoch": 0.3070743715988598, + "grad_norm": 0.3795921376766387, + "learning_rate": 0.0001749985432652255, + "loss": 1.1514, + "step": 1185 + }, + { + "epoch": 0.30837004405286345, + "grad_norm": 0.3984980575708157, + "learning_rate": 0.00017469860930221734, + "loss": 1.2262, + "step": 1190 + }, + { + "epoch": 0.3096657165068671, + "grad_norm": 0.4239852292999169, + "learning_rate": 0.00017439714727373378, + "loss": 1.1279, + "step": 1195 + }, + { + "epoch": 0.31096138896087067, + "grad_norm": 0.4243568518066299, + "learning_rate": 0.00017409416334660606, + "loss": 1.1482, + "step": 1200 + }, + { + "epoch": 0.3122570614148743, + "grad_norm": 0.39261780158026205, + "learning_rate": 0.00017378966371879803, + "loss": 1.1245, + "step": 1205 + }, + { + "epoch": 0.31355273386887794, + "grad_norm": 0.4129200331115499, + "learning_rate": 0.00017348365461927932, + "loss": 1.164, + "step": 1210 + }, + { + "epoch": 0.3148484063228816, + "grad_norm": 0.46767017383779913, + "learning_rate": 0.00017317614230789792, + "loss": 1.2038, + "step": 1215 + }, + { + "epoch": 0.3161440787768852, + "grad_norm": 0.4636682224687034, + "learning_rate": 0.00017286713307525212, + "loss": 1.1798, + "step": 1220 + }, + { + "epoch": 0.31743975123088886, + "grad_norm": 0.464530514696448, + "learning_rate": 0.00017255663324256194, + "loss": 1.1569, + "step": 1225 + }, + { + "epoch": 0.31873542368489244, + "grad_norm": 0.49392901042147025, + "learning_rate": 0.00017224464916153963, + "loss": 1.1722, + "step": 1230 + }, + { + "epoch": 0.3200310961388961, + "grad_norm": 0.437281931716071, + "learning_rate": 0.00017193118721425986, + "loss": 1.1472, + "step": 1235 + }, + { + "epoch": 0.3213267685928997, + "grad_norm": 0.49973298936543603, + "learning_rate": 0.00017161625381302914, + "loss": 1.1662, + "step": 1240 + }, + { + "epoch": 0.32262244104690335, + "grad_norm": 0.9030290188031315, + "learning_rate": 0.00017129985540025473, + "loss": 1.1751, + "step": 1245 + }, + { + "epoch": 0.323918113500907, + "grad_norm": 0.46932404302432773, + "learning_rate": 0.00017098199844831262, + "loss": 1.1753, + "step": 1250 + }, + { + "epoch": 0.3252137859549106, + "grad_norm": 0.43737397519531773, + "learning_rate": 0.0001706626894594154, + "loss": 1.1539, + "step": 1255 + }, + { + "epoch": 0.3265094584089142, + "grad_norm": 0.47678251178263875, + "learning_rate": 0.00017034193496547902, + "loss": 1.1572, + "step": 1260 + }, + { + "epoch": 0.32780513086291785, + "grad_norm": 0.48005221081289434, + "learning_rate": 0.00017001974152798942, + "loss": 1.2371, + "step": 1265 + }, + { + "epoch": 0.3291008033169215, + "grad_norm": 0.4005197274939468, + "learning_rate": 0.000169696115737868, + "loss": 1.1375, + "step": 1270 + }, + { + "epoch": 0.3303964757709251, + "grad_norm": 0.520470386619454, + "learning_rate": 0.00016937106421533707, + "loss": 1.2201, + "step": 1275 + }, + { + "epoch": 0.33169214822492876, + "grad_norm": 0.4024597397089223, + "learning_rate": 0.00016904459360978427, + "loss": 1.16, + "step": 1280 + }, + { + "epoch": 0.33298782067893234, + "grad_norm": 0.43690159855763916, + "learning_rate": 0.00016871671059962655, + "loss": 1.2303, + "step": 1285 + }, + { + "epoch": 0.334283493132936, + "grad_norm": 0.42841787432183803, + "learning_rate": 0.00016838742189217366, + "loss": 1.1601, + "step": 1290 + }, + { + "epoch": 0.3355791655869396, + "grad_norm": 0.41150940120312113, + "learning_rate": 0.00016805673422349082, + "loss": 1.1542, + "step": 1295 + }, + { + "epoch": 0.33687483804094326, + "grad_norm": 0.4108846561673421, + "learning_rate": 0.000167724654358261, + "loss": 1.1497, + "step": 1300 + }, + { + "epoch": 0.3381705104949469, + "grad_norm": 0.4485077675503156, + "learning_rate": 0.00016739118908964647, + "loss": 1.1979, + "step": 1305 + }, + { + "epoch": 0.3394661829489505, + "grad_norm": 0.40380367544471263, + "learning_rate": 0.00016705634523915, + "loss": 1.156, + "step": 1310 + }, + { + "epoch": 0.3407618554029541, + "grad_norm": 0.4315676369718417, + "learning_rate": 0.000166720129656475, + "loss": 1.1876, + "step": 1315 + }, + { + "epoch": 0.34205752785695775, + "grad_norm": 0.42298698967693915, + "learning_rate": 0.00016638254921938587, + "loss": 1.226, + "step": 1320 + }, + { + "epoch": 0.3433532003109614, + "grad_norm": 0.40711175836921143, + "learning_rate": 0.00016604361083356675, + "loss": 1.1865, + "step": 1325 + }, + { + "epoch": 0.34464887276496503, + "grad_norm": 0.4209373436004817, + "learning_rate": 0.0001657033214324807, + "loss": 1.1485, + "step": 1330 + }, + { + "epoch": 0.34594454521896867, + "grad_norm": 0.3973628247221751, + "learning_rate": 0.0001653616879772277, + "loss": 1.1419, + "step": 1335 + }, + { + "epoch": 0.34724021767297225, + "grad_norm": 0.40714636825697015, + "learning_rate": 0.00016501871745640213, + "loss": 1.1607, + "step": 1340 + }, + { + "epoch": 0.3485358901269759, + "grad_norm": 0.4091289111560899, + "learning_rate": 0.00016467441688595015, + "loss": 1.169, + "step": 1345 + }, + { + "epoch": 0.3498315625809795, + "grad_norm": 0.39294191213880614, + "learning_rate": 0.0001643287933090258, + "loss": 1.1856, + "step": 1350 + }, + { + "epoch": 0.35112723503498317, + "grad_norm": 0.4153756492643378, + "learning_rate": 0.00016398185379584707, + "loss": 1.1601, + "step": 1355 + }, + { + "epoch": 0.3524229074889868, + "grad_norm": 0.39772148869790713, + "learning_rate": 0.0001636336054435514, + "loss": 1.1402, + "step": 1360 + }, + { + "epoch": 0.3537185799429904, + "grad_norm": 0.4071165220940827, + "learning_rate": 0.00016328405537605032, + "loss": 1.1333, + "step": 1365 + }, + { + "epoch": 0.355014252396994, + "grad_norm": 0.4269465126269822, + "learning_rate": 0.00016293321074388375, + "loss": 1.1948, + "step": 1370 + }, + { + "epoch": 0.35630992485099766, + "grad_norm": 0.41865801202656294, + "learning_rate": 0.00016258107872407375, + "loss": 1.1465, + "step": 1375 + }, + { + "epoch": 0.3576055973050013, + "grad_norm": 0.40838786927024157, + "learning_rate": 0.00016222766651997789, + "loss": 1.1695, + "step": 1380 + }, + { + "epoch": 0.35890126975900494, + "grad_norm": 0.4096176785342024, + "learning_rate": 0.0001618729813611414, + "loss": 1.1447, + "step": 1385 + }, + { + "epoch": 0.3601969422130086, + "grad_norm": 0.41309116576243843, + "learning_rate": 0.00016151703050314986, + "loss": 1.1804, + "step": 1390 + }, + { + "epoch": 0.36149261466701216, + "grad_norm": 0.4053926841703832, + "learning_rate": 0.00016115982122748043, + "loss": 1.1471, + "step": 1395 + }, + { + "epoch": 0.3627882871210158, + "grad_norm": 0.4058497805268484, + "learning_rate": 0.00016080136084135297, + "loss": 1.1494, + "step": 1400 + }, + { + "epoch": 0.36408395957501943, + "grad_norm": 0.4216552491807955, + "learning_rate": 0.00016044165667758055, + "loss": 1.1928, + "step": 1405 + }, + { + "epoch": 0.36537963202902307, + "grad_norm": 0.4336798036895343, + "learning_rate": 0.0001600807160944195, + "loss": 1.19, + "step": 1410 + }, + { + "epoch": 0.3666753044830267, + "grad_norm": 0.42380600181334993, + "learning_rate": 0.00015971854647541884, + "loss": 1.1674, + "step": 1415 + }, + { + "epoch": 0.3679709769370303, + "grad_norm": 0.4116489897844781, + "learning_rate": 0.00015935515522926927, + "loss": 1.1407, + "step": 1420 + }, + { + "epoch": 0.36926664939103393, + "grad_norm": 0.4051816422975703, + "learning_rate": 0.00015899054978965157, + "loss": 1.1861, + "step": 1425 + }, + { + "epoch": 0.37056232184503757, + "grad_norm": 0.4374412539295244, + "learning_rate": 0.0001586247376150846, + "loss": 1.2273, + "step": 1430 + }, + { + "epoch": 0.3718579942990412, + "grad_norm": 0.4225494310104097, + "learning_rate": 0.00015825772618877263, + "loss": 1.2218, + "step": 1435 + }, + { + "epoch": 0.37315366675304484, + "grad_norm": 0.4202018672382752, + "learning_rate": 0.00015788952301845237, + "loss": 1.1155, + "step": 1440 + }, + { + "epoch": 0.3744493392070485, + "grad_norm": 0.4313070413558118, + "learning_rate": 0.0001575201356362393, + "loss": 1.1551, + "step": 1445 + }, + { + "epoch": 0.37574501166105206, + "grad_norm": 0.40304574068983823, + "learning_rate": 0.00015714957159847367, + "loss": 1.1491, + "step": 1450 + }, + { + "epoch": 0.3770406841150557, + "grad_norm": 0.4413697670634288, + "learning_rate": 0.00015677783848556576, + "loss": 1.1631, + "step": 1455 + }, + { + "epoch": 0.37833635656905934, + "grad_norm": 0.4021843049770175, + "learning_rate": 0.00015640494390184112, + "loss": 1.1304, + "step": 1460 + }, + { + "epoch": 0.379632029023063, + "grad_norm": 0.41280223940193794, + "learning_rate": 0.0001560308954753847, + "loss": 1.1458, + "step": 1465 + }, + { + "epoch": 0.3809277014770666, + "grad_norm": 0.44403414056325813, + "learning_rate": 0.00015565570085788495, + "loss": 1.2007, + "step": 1470 + }, + { + "epoch": 0.3822233739310702, + "grad_norm": 0.42327265730452945, + "learning_rate": 0.00015527936772447725, + "loss": 1.1168, + "step": 1475 + }, + { + "epoch": 0.38351904638507384, + "grad_norm": 0.41872242471581905, + "learning_rate": 0.00015490190377358704, + "loss": 1.1551, + "step": 1480 + }, + { + "epoch": 0.3848147188390775, + "grad_norm": 0.4076213102610575, + "learning_rate": 0.00015452331672677206, + "loss": 1.0902, + "step": 1485 + }, + { + "epoch": 0.3861103912930811, + "grad_norm": 0.39807246728274887, + "learning_rate": 0.00015414361432856475, + "loss": 1.1598, + "step": 1490 + }, + { + "epoch": 0.38740606374708475, + "grad_norm": 0.40854063250285516, + "learning_rate": 0.00015376280434631345, + "loss": 1.1806, + "step": 1495 + }, + { + "epoch": 0.3887017362010884, + "grad_norm": 0.40837052057355316, + "learning_rate": 0.00015338089457002382, + "loss": 1.0829, + "step": 1500 + }, + { + "epoch": 0.38999740865509197, + "grad_norm": 0.4123609619777807, + "learning_rate": 0.00015299789281219935, + "loss": 1.1688, + "step": 1505 + }, + { + "epoch": 0.3912930811090956, + "grad_norm": 0.4316199953541538, + "learning_rate": 0.00015261380690768144, + "loss": 1.1543, + "step": 1510 + }, + { + "epoch": 0.39258875356309925, + "grad_norm": 0.4398633371362292, + "learning_rate": 0.00015222864471348943, + "loss": 1.1724, + "step": 1515 + }, + { + "epoch": 0.3938844260171029, + "grad_norm": 0.4081954480961544, + "learning_rate": 0.00015184241410865954, + "loss": 1.1269, + "step": 1520 + }, + { + "epoch": 0.3951800984711065, + "grad_norm": 0.3994337234690745, + "learning_rate": 0.00015145512299408388, + "loss": 1.1598, + "step": 1525 + }, + { + "epoch": 0.3964757709251101, + "grad_norm": 0.4415117779921472, + "learning_rate": 0.00015106677929234877, + "loss": 1.0969, + "step": 1530 + }, + { + "epoch": 0.39777144337911374, + "grad_norm": 0.4033267536569039, + "learning_rate": 0.0001506773909475727, + "loss": 1.1955, + "step": 1535 + }, + { + "epoch": 0.3990671158331174, + "grad_norm": 0.39024517506575723, + "learning_rate": 0.00015028696592524386, + "loss": 1.1562, + "step": 1540 + }, + { + "epoch": 0.400362788287121, + "grad_norm": 0.4359801226192451, + "learning_rate": 0.000149895512212057, + "loss": 1.1695, + "step": 1545 + }, + { + "epoch": 0.40165846074112466, + "grad_norm": 0.3996337334749442, + "learning_rate": 0.00014950303781575034, + "loss": 1.0982, + "step": 1550 + }, + { + "epoch": 0.4029541331951283, + "grad_norm": 0.39816310192557475, + "learning_rate": 0.00014910955076494152, + "loss": 1.1223, + "step": 1555 + }, + { + "epoch": 0.4042498056491319, + "grad_norm": 0.37894092829518466, + "learning_rate": 0.00014871505910896352, + "loss": 1.1217, + "step": 1560 + }, + { + "epoch": 0.4055454781031355, + "grad_norm": 0.4319425134415384, + "learning_rate": 0.0001483195709176999, + "loss": 1.1966, + "step": 1565 + }, + { + "epoch": 0.40684115055713915, + "grad_norm": 0.3996173231205214, + "learning_rate": 0.00014792309428141978, + "loss": 1.1039, + "step": 1570 + }, + { + "epoch": 0.4081368230111428, + "grad_norm": 0.40718491453474887, + "learning_rate": 0.0001475256373106123, + "loss": 1.1725, + "step": 1575 + }, + { + "epoch": 0.40943249546514643, + "grad_norm": 0.4286665586760103, + "learning_rate": 0.00014712720813582066, + "loss": 1.1443, + "step": 1580 + }, + { + "epoch": 0.41072816791915, + "grad_norm": 0.4033525905918133, + "learning_rate": 0.00014672781490747606, + "loss": 1.1742, + "step": 1585 + }, + { + "epoch": 0.41202384037315365, + "grad_norm": 0.40558326017198004, + "learning_rate": 0.00014632746579573052, + "loss": 1.2117, + "step": 1590 + }, + { + "epoch": 0.4133195128271573, + "grad_norm": 0.4056524890982348, + "learning_rate": 0.0001459261689902902, + "loss": 1.2034, + "step": 1595 + }, + { + "epoch": 0.4146151852811609, + "grad_norm": 0.44100980335173223, + "learning_rate": 0.00014552393270024765, + "loss": 1.1544, + "step": 1600 + }, + { + "epoch": 0.41591085773516456, + "grad_norm": 0.39891755340287666, + "learning_rate": 0.00014512076515391375, + "loss": 1.1256, + "step": 1605 + }, + { + "epoch": 0.4172065301891682, + "grad_norm": 0.40682040920269774, + "learning_rate": 0.00014471667459864973, + "loss": 1.1499, + "step": 1610 + }, + { + "epoch": 0.4185022026431718, + "grad_norm": 0.44312416626229856, + "learning_rate": 0.00014431166930069816, + "loss": 1.0977, + "step": 1615 + }, + { + "epoch": 0.4197978750971754, + "grad_norm": 0.4301392634897964, + "learning_rate": 0.00014390575754501402, + "loss": 1.139, + "step": 1620 + }, + { + "epoch": 0.42109354755117906, + "grad_norm": 0.418724699659618, + "learning_rate": 0.0001434989476350951, + "loss": 1.1592, + "step": 1625 + }, + { + "epoch": 0.4223892200051827, + "grad_norm": 0.4134020107191313, + "learning_rate": 0.00014309124789281226, + "loss": 1.2104, + "step": 1630 + }, + { + "epoch": 0.42368489245918634, + "grad_norm": 0.39580446880805387, + "learning_rate": 0.00014268266665823912, + "loss": 1.1757, + "step": 1635 + }, + { + "epoch": 0.42498056491319, + "grad_norm": 0.41472381643526246, + "learning_rate": 0.00014227321228948146, + "loss": 1.1322, + "step": 1640 + }, + { + "epoch": 0.42627623736719356, + "grad_norm": 0.39275693329437805, + "learning_rate": 0.0001418628931625062, + "loss": 1.1742, + "step": 1645 + }, + { + "epoch": 0.4275719098211972, + "grad_norm": 0.38619824420216714, + "learning_rate": 0.00014145171767097027, + "loss": 1.1511, + "step": 1650 + }, + { + "epoch": 0.42886758227520083, + "grad_norm": 0.4254507618246464, + "learning_rate": 0.00014103969422604856, + "loss": 1.1231, + "step": 1655 + }, + { + "epoch": 0.43016325472920447, + "grad_norm": 0.40459401347151125, + "learning_rate": 0.00014062683125626218, + "loss": 1.1384, + "step": 1660 + }, + { + "epoch": 0.4314589271832081, + "grad_norm": 0.409140169024848, + "learning_rate": 0.0001402131372073058, + "loss": 1.191, + "step": 1665 + }, + { + "epoch": 0.4327545996372117, + "grad_norm": 0.389695812394897, + "learning_rate": 0.00013979862054187505, + "loss": 1.1543, + "step": 1670 + }, + { + "epoch": 0.43405027209121533, + "grad_norm": 0.42112717245729503, + "learning_rate": 0.00013938328973949336, + "loss": 1.1715, + "step": 1675 + }, + { + "epoch": 0.43534594454521897, + "grad_norm": 0.3864671779444786, + "learning_rate": 0.0001389671532963384, + "loss": 1.1508, + "step": 1680 + }, + { + "epoch": 0.4366416169992226, + "grad_norm": 0.39052429534303434, + "learning_rate": 0.00013855021972506844, + "loss": 1.1054, + "step": 1685 + }, + { + "epoch": 0.43793728945322624, + "grad_norm": 0.404117930960366, + "learning_rate": 0.0001381324975546481, + "loss": 1.1623, + "step": 1690 + }, + { + "epoch": 0.4392329619072299, + "grad_norm": 0.45134279442398223, + "learning_rate": 0.0001377139953301739, + "loss": 1.1301, + "step": 1695 + }, + { + "epoch": 0.44052863436123346, + "grad_norm": 0.42206784378450607, + "learning_rate": 0.00013729472161269946, + "loss": 1.18, + "step": 1700 + }, + { + "epoch": 0.4418243068152371, + "grad_norm": 0.43048779695832334, + "learning_rate": 0.00013687468497906044, + "loss": 1.1413, + "step": 1705 + }, + { + "epoch": 0.44311997926924074, + "grad_norm": 0.3902243379067639, + "learning_rate": 0.00013645389402169893, + "loss": 1.1473, + "step": 1710 + }, + { + "epoch": 0.4444156517232444, + "grad_norm": 0.3790071383777115, + "learning_rate": 0.00013603235734848784, + "loss": 1.2114, + "step": 1715 + }, + { + "epoch": 0.445711324177248, + "grad_norm": 0.39998613746463574, + "learning_rate": 0.00013561008358255468, + "loss": 1.1193, + "step": 1720 + }, + { + "epoch": 0.4470069966312516, + "grad_norm": 0.4136780934178077, + "learning_rate": 0.0001351870813621054, + "loss": 1.1417, + "step": 1725 + }, + { + "epoch": 0.44830266908525523, + "grad_norm": 0.39313685685153266, + "learning_rate": 0.00013476335934024735, + "loss": 1.1437, + "step": 1730 + }, + { + "epoch": 0.44959834153925887, + "grad_norm": 0.41028501570471904, + "learning_rate": 0.00013433892618481248, + "loss": 1.1032, + "step": 1735 + }, + { + "epoch": 0.4508940139932625, + "grad_norm": 0.43704478597398877, + "learning_rate": 0.00013391379057817995, + "loss": 1.1933, + "step": 1740 + }, + { + "epoch": 0.45218968644726615, + "grad_norm": 0.37937862506331654, + "learning_rate": 0.00013348796121709862, + "loss": 1.187, + "step": 1745 + }, + { + "epoch": 0.4534853589012698, + "grad_norm": 0.40181873506259413, + "learning_rate": 0.00013306144681250908, + "loss": 1.1625, + "step": 1750 + }, + { + "epoch": 0.45478103135527337, + "grad_norm": 0.4117326535807377, + "learning_rate": 0.00013263425608936536, + "loss": 1.1875, + "step": 1755 + }, + { + "epoch": 0.456076703809277, + "grad_norm": 0.39234377730019654, + "learning_rate": 0.00013220639778645663, + "loss": 1.1888, + "step": 1760 + }, + { + "epoch": 0.45737237626328064, + "grad_norm": 0.42622878050330665, + "learning_rate": 0.0001317778806562283, + "loss": 1.151, + "step": 1765 + }, + { + "epoch": 0.4586680487172843, + "grad_norm": 0.3997439152224071, + "learning_rate": 0.000131348713464603, + "loss": 1.1271, + "step": 1770 + }, + { + "epoch": 0.4599637211712879, + "grad_norm": 0.4151543746885238, + "learning_rate": 0.0001309189049908014, + "loss": 1.1745, + "step": 1775 + }, + { + "epoch": 0.4612593936252915, + "grad_norm": 0.4136123970052988, + "learning_rate": 0.00013048846402716237, + "loss": 1.1446, + "step": 1780 + }, + { + "epoch": 0.46255506607929514, + "grad_norm": 0.3925741414110596, + "learning_rate": 0.0001300573993789633, + "loss": 1.1186, + "step": 1785 + }, + { + "epoch": 0.4638507385332988, + "grad_norm": 0.43074300043773284, + "learning_rate": 0.00012962571986423993, + "loss": 1.2004, + "step": 1790 + }, + { + "epoch": 0.4651464109873024, + "grad_norm": 0.3948645456920277, + "learning_rate": 0.00012919343431360596, + "loss": 1.1534, + "step": 1795 + }, + { + "epoch": 0.46644208344130605, + "grad_norm": 0.4423236206923003, + "learning_rate": 0.00012876055157007242, + "loss": 1.1509, + "step": 1800 + }, + { + "epoch": 0.4677377558953097, + "grad_norm": 0.39371508060725335, + "learning_rate": 0.00012832708048886679, + "loss": 1.1941, + "step": 1805 + }, + { + "epoch": 0.4690334283493133, + "grad_norm": 0.4156050706970669, + "learning_rate": 0.00012789302993725175, + "loss": 1.2233, + "step": 1810 + }, + { + "epoch": 0.4703291008033169, + "grad_norm": 0.3844699731827056, + "learning_rate": 0.0001274584087943439, + "loss": 1.172, + "step": 1815 + }, + { + "epoch": 0.47162477325732055, + "grad_norm": 0.392439037665497, + "learning_rate": 0.00012702322595093212, + "loss": 1.1935, + "step": 1820 + }, + { + "epoch": 0.4729204457113242, + "grad_norm": 0.42109447627306007, + "learning_rate": 0.00012658749030929566, + "loss": 1.0821, + "step": 1825 + }, + { + "epoch": 0.4742161181653278, + "grad_norm": 0.3985407382924968, + "learning_rate": 0.00012615121078302202, + "loss": 1.1564, + "step": 1830 + }, + { + "epoch": 0.4755117906193314, + "grad_norm": 0.40190944372077425, + "learning_rate": 0.0001257143962968246, + "loss": 1.1858, + "step": 1835 + }, + { + "epoch": 0.47680746307333505, + "grad_norm": 0.45076787480964914, + "learning_rate": 0.00012527705578636023, + "loss": 1.1514, + "step": 1840 + }, + { + "epoch": 0.4781031355273387, + "grad_norm": 0.4363861711704788, + "learning_rate": 0.0001248391981980462, + "loss": 1.133, + "step": 1845 + }, + { + "epoch": 0.4793988079813423, + "grad_norm": 0.43776230767541446, + "learning_rate": 0.00012440083248887754, + "loss": 1.2082, + "step": 1850 + }, + { + "epoch": 0.48069448043534596, + "grad_norm": 0.4127859770012284, + "learning_rate": 0.00012396196762624341, + "loss": 1.1613, + "step": 1855 + }, + { + "epoch": 0.4819901528893496, + "grad_norm": 0.3940447846474964, + "learning_rate": 0.00012352261258774395, + "loss": 1.15, + "step": 1860 + }, + { + "epoch": 0.4832858253433532, + "grad_norm": 0.4193523558931992, + "learning_rate": 0.0001230827763610066, + "loss": 1.1382, + "step": 1865 + }, + { + "epoch": 0.4845814977973568, + "grad_norm": 0.395557633792394, + "learning_rate": 0.00012264246794350202, + "loss": 1.1678, + "step": 1870 + }, + { + "epoch": 0.48587717025136046, + "grad_norm": 0.40423444904954847, + "learning_rate": 0.00012220169634236038, + "loss": 1.157, + "step": 1875 + }, + { + "epoch": 0.4871728427053641, + "grad_norm": 0.3870746898688869, + "learning_rate": 0.00012176047057418682, + "loss": 1.1439, + "step": 1880 + }, + { + "epoch": 0.48846851515936773, + "grad_norm": 0.4458893388021195, + "learning_rate": 0.00012131879966487709, + "loss": 1.1513, + "step": 1885 + }, + { + "epoch": 0.4897641876133713, + "grad_norm": 0.38478400725184214, + "learning_rate": 0.00012087669264943302, + "loss": 1.1333, + "step": 1890 + }, + { + "epoch": 0.49105986006737495, + "grad_norm": 0.37819480972862285, + "learning_rate": 0.00012043415857177751, + "loss": 1.1663, + "step": 1895 + }, + { + "epoch": 0.4923555325213786, + "grad_norm": 0.4125210415708117, + "learning_rate": 0.00011999120648456974, + "loss": 1.1457, + "step": 1900 + }, + { + "epoch": 0.49365120497538223, + "grad_norm": 0.3869780886696955, + "learning_rate": 0.00011954784544901971, + "loss": 1.1536, + "step": 1905 + }, + { + "epoch": 0.49494687742938587, + "grad_norm": 0.40907576777936727, + "learning_rate": 0.00011910408453470316, + "loss": 1.1361, + "step": 1910 + }, + { + "epoch": 0.4962425498833895, + "grad_norm": 0.3737647497815572, + "learning_rate": 0.00011865993281937589, + "loss": 1.1006, + "step": 1915 + }, + { + "epoch": 0.4975382223373931, + "grad_norm": 0.41125742423172207, + "learning_rate": 0.00011821539938878801, + "loss": 1.1641, + "step": 1920 + }, + { + "epoch": 0.4988338947913967, + "grad_norm": 0.38132649216646763, + "learning_rate": 0.00011777049333649826, + "loss": 1.1525, + "step": 1925 + }, + { + "epoch": 0.5001295672454004, + "grad_norm": 0.38834910320923277, + "learning_rate": 0.00011732522376368781, + "loss": 1.1531, + "step": 1930 + }, + { + "epoch": 0.501425239699404, + "grad_norm": 0.40314094443067383, + "learning_rate": 0.0001168795997789742, + "loss": 1.1592, + "step": 1935 + }, + { + "epoch": 0.5027209121534076, + "grad_norm": 0.4040865102971838, + "learning_rate": 0.00011643363049822496, + "loss": 1.2547, + "step": 1940 + }, + { + "epoch": 0.5040165846074113, + "grad_norm": 0.38562943372958325, + "learning_rate": 0.00011598732504437107, + "loss": 1.1109, + "step": 1945 + }, + { + "epoch": 0.5053122570614149, + "grad_norm": 0.4250211446665384, + "learning_rate": 0.00011554069254722051, + "loss": 1.1941, + "step": 1950 + }, + { + "epoch": 0.5066079295154186, + "grad_norm": 0.4032090808510551, + "learning_rate": 0.00011509374214327131, + "loss": 1.1572, + "step": 1955 + }, + { + "epoch": 0.5079036019694221, + "grad_norm": 0.40575976251729745, + "learning_rate": 0.00011464648297552478, + "loss": 1.1657, + "step": 1960 + }, + { + "epoch": 0.5091992744234257, + "grad_norm": 0.41052107415559824, + "learning_rate": 0.00011419892419329844, + "loss": 1.1642, + "step": 1965 + }, + { + "epoch": 0.5104949468774294, + "grad_norm": 0.3958628275091308, + "learning_rate": 0.00011375107495203873, + "loss": 1.2116, + "step": 1970 + }, + { + "epoch": 0.511790619331433, + "grad_norm": 0.38478613819779206, + "learning_rate": 0.00011330294441313402, + "loss": 1.1451, + "step": 1975 + }, + { + "epoch": 0.5130862917854366, + "grad_norm": 0.43128674114012594, + "learning_rate": 0.00011285454174372692, + "loss": 1.1494, + "step": 1980 + }, + { + "epoch": 0.5143819642394403, + "grad_norm": 0.4424953381805022, + "learning_rate": 0.0001124058761165268, + "loss": 1.0932, + "step": 1985 + }, + { + "epoch": 0.5156776366934439, + "grad_norm": 0.4063559419481111, + "learning_rate": 0.00011195695670962234, + "loss": 1.1137, + "step": 1990 + }, + { + "epoch": 0.5169733091474475, + "grad_norm": 0.3975356665709952, + "learning_rate": 0.00011150779270629353, + "loss": 1.1873, + "step": 1995 + }, + { + "epoch": 0.5182689816014512, + "grad_norm": 0.4185740854748383, + "learning_rate": 0.00011105839329482397, + "loss": 1.1108, + "step": 2000 + }, + { + "epoch": 0.5195646540554548, + "grad_norm": 0.38989666093808334, + "learning_rate": 0.00011060876766831285, + "loss": 1.1663, + "step": 2005 + }, + { + "epoch": 0.5208603265094585, + "grad_norm": 0.40596909372428913, + "learning_rate": 0.00011015892502448692, + "loss": 1.1382, + "step": 2010 + }, + { + "epoch": 0.522155998963462, + "grad_norm": 0.37956800433559007, + "learning_rate": 0.00010970887456551234, + "loss": 1.104, + "step": 2015 + }, + { + "epoch": 0.5234516714174656, + "grad_norm": 0.3972204538731939, + "learning_rate": 0.00010925862549780637, + "loss": 1.1414, + "step": 2020 + }, + { + "epoch": 0.5247473438714693, + "grad_norm": 0.4266101329792826, + "learning_rate": 0.00010880818703184919, + "loss": 1.2131, + "step": 2025 + }, + { + "epoch": 0.5260430163254729, + "grad_norm": 0.4059164191077208, + "learning_rate": 0.00010835756838199524, + "loss": 1.1005, + "step": 2030 + }, + { + "epoch": 0.5273386887794765, + "grad_norm": 0.37690318544316265, + "learning_rate": 0.00010790677876628501, + "loss": 1.1448, + "step": 2035 + }, + { + "epoch": 0.5286343612334802, + "grad_norm": 0.39147947369844754, + "learning_rate": 0.00010745582740625631, + "loss": 1.161, + "step": 2040 + }, + { + "epoch": 0.5299300336874838, + "grad_norm": 0.4061256597825514, + "learning_rate": 0.00010700472352675556, + "loss": 1.1861, + "step": 2045 + }, + { + "epoch": 0.5312257061414875, + "grad_norm": 0.3877742069695207, + "learning_rate": 0.00010655347635574937, + "loss": 1.1816, + "step": 2050 + }, + { + "epoch": 0.5325213785954911, + "grad_norm": 0.3971133244937986, + "learning_rate": 0.00010610209512413536, + "loss": 1.2017, + "step": 2055 + }, + { + "epoch": 0.5338170510494947, + "grad_norm": 0.41551450058475425, + "learning_rate": 0.0001056505890655537, + "loss": 1.1692, + "step": 2060 + }, + { + "epoch": 0.5351127235034984, + "grad_norm": 0.43646326285085413, + "learning_rate": 0.00010519896741619803, + "loss": 1.0993, + "step": 2065 + }, + { + "epoch": 0.5364083959575019, + "grad_norm": 0.38667765038581386, + "learning_rate": 0.00010474723941462658, + "loss": 1.1151, + "step": 2070 + }, + { + "epoch": 0.5377040684115055, + "grad_norm": 0.39156927596261365, + "learning_rate": 0.00010429541430157313, + "loss": 1.1641, + "step": 2075 + }, + { + "epoch": 0.5389997408655092, + "grad_norm": 0.39506144218425854, + "learning_rate": 0.00010384350131975802, + "loss": 1.2444, + "step": 2080 + }, + { + "epoch": 0.5402954133195128, + "grad_norm": 0.3881326675938924, + "learning_rate": 0.00010339150971369906, + "loss": 1.129, + "step": 2085 + }, + { + "epoch": 0.5415910857735164, + "grad_norm": 0.3901794452377504, + "learning_rate": 0.00010293944872952248, + "loss": 1.1536, + "step": 2090 + }, + { + "epoch": 0.5428867582275201, + "grad_norm": 0.3852428804649106, + "learning_rate": 0.00010248732761477371, + "loss": 1.1365, + "step": 2095 + }, + { + "epoch": 0.5441824306815237, + "grad_norm": 0.4261491575765494, + "learning_rate": 0.0001020351556182282, + "loss": 1.089, + "step": 2100 + }, + { + "epoch": 0.5454781031355274, + "grad_norm": 0.4070655687819442, + "learning_rate": 0.00010158294198970232, + "loss": 1.1591, + "step": 2105 + }, + { + "epoch": 0.546773775589531, + "grad_norm": 0.4142505000419484, + "learning_rate": 0.00010113069597986402, + "loss": 1.1403, + "step": 2110 + }, + { + "epoch": 0.5480694480435346, + "grad_norm": 0.3988427337375973, + "learning_rate": 0.0001006784268400437, + "loss": 1.1043, + "step": 2115 + }, + { + "epoch": 0.5493651204975383, + "grad_norm": 0.37926295342271793, + "learning_rate": 0.00010022614382204492, + "loss": 1.1694, + "step": 2120 + }, + { + "epoch": 0.5506607929515418, + "grad_norm": 0.3854042883512697, + "learning_rate": 9.97738561779551e-05, + "loss": 1.1891, + "step": 2125 + }, + { + "epoch": 0.5519564654055454, + "grad_norm": 0.4035084032954952, + "learning_rate": 9.932157315995631e-05, + "loss": 1.1563, + "step": 2130 + }, + { + "epoch": 0.5532521378595491, + "grad_norm": 0.4083098281929601, + "learning_rate": 9.8869304020136e-05, + "loss": 1.2097, + "step": 2135 + }, + { + "epoch": 0.5545478103135527, + "grad_norm": 0.3884516778316842, + "learning_rate": 9.841705801029769e-05, + "loss": 1.1932, + "step": 2140 + }, + { + "epoch": 0.5558434827675564, + "grad_norm": 0.38224759440627476, + "learning_rate": 9.79648443817718e-05, + "loss": 1.0978, + "step": 2145 + }, + { + "epoch": 0.55713915522156, + "grad_norm": 0.46057864945040095, + "learning_rate": 9.751267238522631e-05, + "loss": 1.1465, + "step": 2150 + }, + { + "epoch": 0.5584348276755636, + "grad_norm": 0.38627135619303393, + "learning_rate": 9.706055127047755e-05, + "loss": 1.1533, + "step": 2155 + }, + { + "epoch": 0.5597305001295673, + "grad_norm": 0.3891179649712941, + "learning_rate": 9.660849028630096e-05, + "loss": 1.1984, + "step": 2160 + }, + { + "epoch": 0.5610261725835709, + "grad_norm": 0.39223924604212307, + "learning_rate": 9.615649868024199e-05, + "loss": 1.1463, + "step": 2165 + }, + { + "epoch": 0.5623218450375745, + "grad_norm": 0.39471896059848777, + "learning_rate": 9.570458569842688e-05, + "loss": 1.1487, + "step": 2170 + }, + { + "epoch": 0.5636175174915782, + "grad_norm": 0.3928874290731547, + "learning_rate": 9.525276058537344e-05, + "loss": 1.1315, + "step": 2175 + }, + { + "epoch": 0.5649131899455817, + "grad_norm": 0.4064990206731607, + "learning_rate": 9.480103258380198e-05, + "loss": 1.1562, + "step": 2180 + }, + { + "epoch": 0.5662088623995853, + "grad_norm": 0.38415170548201955, + "learning_rate": 9.434941093444632e-05, + "loss": 1.1811, + "step": 2185 + }, + { + "epoch": 0.567504534853589, + "grad_norm": 0.3859240132285136, + "learning_rate": 9.389790487586465e-05, + "loss": 1.2216, + "step": 2190 + }, + { + "epoch": 0.5688002073075926, + "grad_norm": 0.40615716009377495, + "learning_rate": 9.344652364425065e-05, + "loss": 1.2206, + "step": 2195 + }, + { + "epoch": 0.5700958797615963, + "grad_norm": 0.3919524007695092, + "learning_rate": 9.299527647324444e-05, + "loss": 1.1929, + "step": 2200 + }, + { + "epoch": 0.5713915522155999, + "grad_norm": 0.3821208159345121, + "learning_rate": 9.254417259374374e-05, + "loss": 1.0911, + "step": 2205 + }, + { + "epoch": 0.5726872246696035, + "grad_norm": 0.42022947326779125, + "learning_rate": 9.2093221233715e-05, + "loss": 1.1504, + "step": 2210 + }, + { + "epoch": 0.5739828971236072, + "grad_norm": 0.4544847994091584, + "learning_rate": 9.164243161800477e-05, + "loss": 1.1582, + "step": 2215 + }, + { + "epoch": 0.5752785695776108, + "grad_norm": 0.41094059546634143, + "learning_rate": 9.119181296815085e-05, + "loss": 1.1297, + "step": 2220 + }, + { + "epoch": 0.5765742420316144, + "grad_norm": 0.3805641014764856, + "learning_rate": 9.074137450219364e-05, + "loss": 1.185, + "step": 2225 + }, + { + "epoch": 0.5778699144856181, + "grad_norm": 0.39504497212816847, + "learning_rate": 9.02911254344877e-05, + "loss": 1.1875, + "step": 2230 + }, + { + "epoch": 0.5791655869396216, + "grad_norm": 0.40451192253045454, + "learning_rate": 8.984107497551311e-05, + "loss": 1.1368, + "step": 2235 + }, + { + "epoch": 0.5804612593936253, + "grad_norm": 0.3769782624908954, + "learning_rate": 8.939123233168717e-05, + "loss": 1.1391, + "step": 2240 + }, + { + "epoch": 0.5817569318476289, + "grad_norm": 0.4058748960564205, + "learning_rate": 8.894160670517606e-05, + "loss": 1.1751, + "step": 2245 + }, + { + "epoch": 0.5830526043016325, + "grad_norm": 0.3881661677768996, + "learning_rate": 8.849220729370651e-05, + "loss": 1.1215, + "step": 2250 + }, + { + "epoch": 0.5843482767556362, + "grad_norm": 0.3807361487895863, + "learning_rate": 8.804304329037771e-05, + "loss": 1.1604, + "step": 2255 + }, + { + "epoch": 0.5856439492096398, + "grad_norm": 0.3780073193073684, + "learning_rate": 8.759412388347321e-05, + "loss": 1.1159, + "step": 2260 + }, + { + "epoch": 0.5869396216636434, + "grad_norm": 0.3910762750334457, + "learning_rate": 8.71454582562731e-05, + "loss": 1.1283, + "step": 2265 + }, + { + "epoch": 0.5882352941176471, + "grad_norm": 0.3577377830833829, + "learning_rate": 8.669705558686599e-05, + "loss": 1.0881, + "step": 2270 + }, + { + "epoch": 0.5895309665716507, + "grad_norm": 0.38870803678431104, + "learning_rate": 8.624892504796128e-05, + "loss": 1.1302, + "step": 2275 + }, + { + "epoch": 0.5908266390256544, + "grad_norm": 0.43127992879460963, + "learning_rate": 8.580107580670163e-05, + "loss": 1.183, + "step": 2280 + }, + { + "epoch": 0.592122311479658, + "grad_norm": 0.37408901643032355, + "learning_rate": 8.535351702447524e-05, + "loss": 1.1624, + "step": 2285 + }, + { + "epoch": 0.5934179839336615, + "grad_norm": 0.4003666810623533, + "learning_rate": 8.49062578567287e-05, + "loss": 1.1193, + "step": 2290 + }, + { + "epoch": 0.5947136563876652, + "grad_norm": 0.3723417367446364, + "learning_rate": 8.445930745277953e-05, + "loss": 1.146, + "step": 2295 + }, + { + "epoch": 0.5960093288416688, + "grad_norm": 0.35036357860872624, + "learning_rate": 8.401267495562894e-05, + "loss": 1.0963, + "step": 2300 + }, + { + "epoch": 0.5973050012956724, + "grad_norm": 0.3969850486600124, + "learning_rate": 8.356636950177509e-05, + "loss": 1.1645, + "step": 2305 + }, + { + "epoch": 0.5986006737496761, + "grad_norm": 0.3908066674173964, + "learning_rate": 8.312040022102581e-05, + "loss": 1.1573, + "step": 2310 + }, + { + "epoch": 0.5998963462036797, + "grad_norm": 0.3970139193288793, + "learning_rate": 8.26747762363122e-05, + "loss": 1.1475, + "step": 2315 + }, + { + "epoch": 0.6011920186576833, + "grad_norm": 0.3809149209886327, + "learning_rate": 8.222950666350176e-05, + "loss": 1.0996, + "step": 2320 + }, + { + "epoch": 0.602487691111687, + "grad_norm": 0.3831688926512033, + "learning_rate": 8.1784600611212e-05, + "loss": 1.1586, + "step": 2325 + }, + { + "epoch": 0.6037833635656906, + "grad_norm": 0.37892711696966613, + "learning_rate": 8.134006718062417e-05, + "loss": 1.1394, + "step": 2330 + }, + { + "epoch": 0.6050790360196943, + "grad_norm": 0.3796251460883534, + "learning_rate": 8.089591546529686e-05, + "loss": 1.1342, + "step": 2335 + }, + { + "epoch": 0.6063747084736979, + "grad_norm": 0.3854947360629934, + "learning_rate": 8.04521545509803e-05, + "loss": 1.1234, + "step": 2340 + }, + { + "epoch": 0.6076703809277014, + "grad_norm": 0.39157237980540016, + "learning_rate": 8.00087935154303e-05, + "loss": 1.1789, + "step": 2345 + }, + { + "epoch": 0.6089660533817051, + "grad_norm": 0.37973961502373016, + "learning_rate": 7.956584142822248e-05, + "loss": 1.1725, + "step": 2350 + }, + { + "epoch": 0.6102617258357087, + "grad_norm": 0.4002316998884078, + "learning_rate": 7.912330735056702e-05, + "loss": 1.1761, + "step": 2355 + }, + { + "epoch": 0.6115573982897123, + "grad_norm": 0.3933317900533284, + "learning_rate": 7.868120033512294e-05, + "loss": 1.1279, + "step": 2360 + }, + { + "epoch": 0.612853070743716, + "grad_norm": 0.3994518364881897, + "learning_rate": 7.82395294258132e-05, + "loss": 1.1292, + "step": 2365 + }, + { + "epoch": 0.6141487431977196, + "grad_norm": 0.3774270174393722, + "learning_rate": 7.779830365763963e-05, + "loss": 1.1699, + "step": 2370 + }, + { + "epoch": 0.6154444156517233, + "grad_norm": 0.3936443342897645, + "learning_rate": 7.735753205649798e-05, + "loss": 1.1691, + "step": 2375 + }, + { + "epoch": 0.6167400881057269, + "grad_norm": 0.3958990632556032, + "learning_rate": 7.691722363899346e-05, + "loss": 1.1053, + "step": 2380 + }, + { + "epoch": 0.6180357605597305, + "grad_norm": 0.40113089490170173, + "learning_rate": 7.647738741225605e-05, + "loss": 1.0826, + "step": 2385 + }, + { + "epoch": 0.6193314330137342, + "grad_norm": 0.39930864888851864, + "learning_rate": 7.60380323737566e-05, + "loss": 1.199, + "step": 2390 + }, + { + "epoch": 0.6206271054677378, + "grad_norm": 0.3962521144461154, + "learning_rate": 7.559916751112248e-05, + "loss": 1.1933, + "step": 2395 + }, + { + "epoch": 0.6219227779217413, + "grad_norm": 0.37888926408124746, + "learning_rate": 7.516080180195379e-05, + "loss": 1.1713, + "step": 2400 + }, + { + "epoch": 0.623218450375745, + "grad_norm": 0.3708133442805237, + "learning_rate": 7.472294421363982e-05, + "loss": 1.2177, + "step": 2405 + }, + { + "epoch": 0.6245141228297486, + "grad_norm": 0.41709292403420495, + "learning_rate": 7.428560370317542e-05, + "loss": 1.1678, + "step": 2410 + }, + { + "epoch": 0.6258097952837522, + "grad_norm": 0.38822425697798224, + "learning_rate": 7.3848789216978e-05, + "loss": 1.1628, + "step": 2415 + }, + { + "epoch": 0.6271054677377559, + "grad_norm": 0.38293720555111366, + "learning_rate": 7.341250969070435e-05, + "loss": 1.1668, + "step": 2420 + }, + { + "epoch": 0.6284011401917595, + "grad_norm": 0.3982879136320185, + "learning_rate": 7.297677404906787e-05, + "loss": 1.1808, + "step": 2425 + }, + { + "epoch": 0.6296968126457632, + "grad_norm": 0.4970736508224388, + "learning_rate": 7.254159120565614e-05, + "loss": 1.1915, + "step": 2430 + }, + { + "epoch": 0.6309924850997668, + "grad_norm": 0.3859126967679283, + "learning_rate": 7.210697006274829e-05, + "loss": 1.1032, + "step": 2435 + }, + { + "epoch": 0.6322881575537704, + "grad_norm": 0.3804578357151546, + "learning_rate": 7.167291951113322e-05, + "loss": 1.1567, + "step": 2440 + }, + { + "epoch": 0.6335838300077741, + "grad_norm": 0.38190993552219266, + "learning_rate": 7.123944842992759e-05, + "loss": 1.091, + "step": 2445 + }, + { + "epoch": 0.6348795024617777, + "grad_norm": 0.38091685226387034, + "learning_rate": 7.080656568639406e-05, + "loss": 1.1457, + "step": 2450 + }, + { + "epoch": 0.6361751749157812, + "grad_norm": 0.3853362546606581, + "learning_rate": 7.037428013576013e-05, + "loss": 1.1463, + "step": 2455 + }, + { + "epoch": 0.6374708473697849, + "grad_norm": 0.36970236476000445, + "learning_rate": 6.994260062103674e-05, + "loss": 1.1393, + "step": 2460 + }, + { + "epoch": 0.6387665198237885, + "grad_norm": 0.40400326202643533, + "learning_rate": 6.951153597283766e-05, + "loss": 1.1295, + "step": 2465 + }, + { + "epoch": 0.6400621922777922, + "grad_norm": 0.3816874564299913, + "learning_rate": 6.908109500919861e-05, + "loss": 1.1662, + "step": 2470 + }, + { + "epoch": 0.6413578647317958, + "grad_norm": 0.3799563658685613, + "learning_rate": 6.865128653539699e-05, + "loss": 1.1876, + "step": 2475 + }, + { + "epoch": 0.6426535371857994, + "grad_norm": 0.379580454677419, + "learning_rate": 6.822211934377176e-05, + "loss": 1.1005, + "step": 2480 + }, + { + "epoch": 0.6439492096398031, + "grad_norm": 0.36403149625971076, + "learning_rate": 6.77936022135434e-05, + "loss": 1.0924, + "step": 2485 + }, + { + "epoch": 0.6452448820938067, + "grad_norm": 0.3864263981944349, + "learning_rate": 6.736574391063466e-05, + "loss": 1.166, + "step": 2490 + }, + { + "epoch": 0.6465405545478103, + "grad_norm": 0.4278354770987037, + "learning_rate": 6.693855318749096e-05, + "loss": 1.1463, + "step": 2495 + }, + { + "epoch": 0.647836227001814, + "grad_norm": 0.3881716691518243, + "learning_rate": 6.651203878290139e-05, + "loss": 1.1658, + "step": 2500 + }, + { + "epoch": 0.6491318994558176, + "grad_norm": 0.3773480140369342, + "learning_rate": 6.608620942182011e-05, + "loss": 1.1561, + "step": 2505 + }, + { + "epoch": 0.6504275719098211, + "grad_norm": 0.38362919406181467, + "learning_rate": 6.566107381518758e-05, + "loss": 1.1553, + "step": 2510 + }, + { + "epoch": 0.6517232443638248, + "grad_norm": 0.3889927393341413, + "learning_rate": 6.523664065975268e-05, + "loss": 1.0927, + "step": 2515 + }, + { + "epoch": 0.6530189168178284, + "grad_norm": 0.3770322611579158, + "learning_rate": 6.481291863789461e-05, + "loss": 1.1621, + "step": 2520 + }, + { + "epoch": 0.6543145892718321, + "grad_norm": 0.3583870387014761, + "learning_rate": 6.43899164174453e-05, + "loss": 1.1688, + "step": 2525 + }, + { + "epoch": 0.6556102617258357, + "grad_norm": 0.36282391266493386, + "learning_rate": 6.396764265151221e-05, + "loss": 1.0909, + "step": 2530 + }, + { + "epoch": 0.6569059341798393, + "grad_norm": 0.40358089248833323, + "learning_rate": 6.35461059783011e-05, + "loss": 1.185, + "step": 2535 + }, + { + "epoch": 0.658201606633843, + "grad_norm": 0.3622883465103346, + "learning_rate": 6.312531502093958e-05, + "loss": 1.1586, + "step": 2540 + }, + { + "epoch": 0.6594972790878466, + "grad_norm": 0.39189718464338913, + "learning_rate": 6.270527838730053e-05, + "loss": 1.1592, + "step": 2545 + }, + { + "epoch": 0.6607929515418502, + "grad_norm": 0.36779638890391525, + "learning_rate": 6.228600466982611e-05, + "loss": 1.156, + "step": 2550 + }, + { + "epoch": 0.6620886239958539, + "grad_norm": 0.37597682220271855, + "learning_rate": 6.186750244535194e-05, + "loss": 1.1353, + "step": 2555 + }, + { + "epoch": 0.6633842964498575, + "grad_norm": 0.3886787309603002, + "learning_rate": 6.144978027493158e-05, + "loss": 1.1341, + "step": 2560 + }, + { + "epoch": 0.664679968903861, + "grad_norm": 0.35942315304405126, + "learning_rate": 6.103284670366162e-05, + "loss": 1.1597, + "step": 2565 + }, + { + "epoch": 0.6659756413578647, + "grad_norm": 0.41146010175047676, + "learning_rate": 6.061671026050668e-05, + "loss": 1.18, + "step": 2570 + }, + { + "epoch": 0.6672713138118683, + "grad_norm": 0.38590903226923434, + "learning_rate": 6.0201379458124964e-05, + "loss": 1.1639, + "step": 2575 + }, + { + "epoch": 0.668566986265872, + "grad_norm": 0.386535230930437, + "learning_rate": 5.978686279269421e-05, + "loss": 1.0866, + "step": 2580 + }, + { + "epoch": 0.6698626587198756, + "grad_norm": 0.3877441934910258, + "learning_rate": 5.9373168743737864e-05, + "loss": 1.1325, + "step": 2585 + }, + { + "epoch": 0.6711583311738792, + "grad_norm": 0.3990007152806006, + "learning_rate": 5.896030577395144e-05, + "loss": 1.1188, + "step": 2590 + }, + { + "epoch": 0.6724540036278829, + "grad_norm": 0.3970387094956777, + "learning_rate": 5.854828232902976e-05, + "loss": 1.1939, + "step": 2595 + }, + { + "epoch": 0.6737496760818865, + "grad_norm": 0.39839395255304766, + "learning_rate": 5.813710683749379e-05, + "loss": 1.2078, + "step": 2600 + }, + { + "epoch": 0.6750453485358902, + "grad_norm": 0.39807324424891044, + "learning_rate": 5.772678771051858e-05, + "loss": 1.1513, + "step": 2605 + }, + { + "epoch": 0.6763410209898938, + "grad_norm": 0.371233691663641, + "learning_rate": 5.7317333341760906e-05, + "loss": 1.1896, + "step": 2610 + }, + { + "epoch": 0.6776366934438974, + "grad_norm": 0.3914320864664493, + "learning_rate": 5.690875210718778e-05, + "loss": 1.144, + "step": 2615 + }, + { + "epoch": 0.678932365897901, + "grad_norm": 0.38952180289646227, + "learning_rate": 5.6501052364904906e-05, + "loss": 1.1669, + "step": 2620 + }, + { + "epoch": 0.6802280383519046, + "grad_norm": 0.3623267889899647, + "learning_rate": 5.6094242454986e-05, + "loss": 1.179, + "step": 2625 + }, + { + "epoch": 0.6815237108059082, + "grad_norm": 0.3637529461983289, + "learning_rate": 5.568833069930186e-05, + "loss": 1.1244, + "step": 2630 + }, + { + "epoch": 0.6828193832599119, + "grad_norm": 0.387937514120572, + "learning_rate": 5.528332540135031e-05, + "loss": 1.1699, + "step": 2635 + }, + { + "epoch": 0.6841150557139155, + "grad_norm": 0.37878126929029915, + "learning_rate": 5.487923484608629e-05, + "loss": 1.131, + "step": 2640 + }, + { + "epoch": 0.6854107281679191, + "grad_norm": 0.3903173526956151, + "learning_rate": 5.4476067299752385e-05, + "loss": 1.1165, + "step": 2645 + }, + { + "epoch": 0.6867064006219228, + "grad_norm": 0.3857342653067374, + "learning_rate": 5.4073831009709805e-05, + "loss": 1.0854, + "step": 2650 + }, + { + "epoch": 0.6880020730759264, + "grad_norm": 0.3661818366505174, + "learning_rate": 5.367253420426952e-05, + "loss": 1.173, + "step": 2655 + }, + { + "epoch": 0.6892977455299301, + "grad_norm": 0.3695899768899783, + "learning_rate": 5.3272185092524004e-05, + "loss": 1.1917, + "step": 2660 + }, + { + "epoch": 0.6905934179839337, + "grad_norm": 0.3709048127868108, + "learning_rate": 5.287279186417938e-05, + "loss": 1.1595, + "step": 2665 + }, + { + "epoch": 0.6918890904379373, + "grad_norm": 0.388715290668864, + "learning_rate": 5.2474362689387745e-05, + "loss": 1.1681, + "step": 2670 + }, + { + "epoch": 0.6931847628919409, + "grad_norm": 0.41572003898003435, + "learning_rate": 5.207690571858025e-05, + "loss": 1.1344, + "step": 2675 + }, + { + "epoch": 0.6944804353459445, + "grad_norm": 0.3688637831571822, + "learning_rate": 5.1680429082300134e-05, + "loss": 1.1911, + "step": 2680 + }, + { + "epoch": 0.6957761077999481, + "grad_norm": 0.3793170468762598, + "learning_rate": 5.128494089103652e-05, + "loss": 1.1127, + "step": 2685 + }, + { + "epoch": 0.6970717802539518, + "grad_norm": 0.43134296541622413, + "learning_rate": 5.0890449235058525e-05, + "loss": 1.1784, + "step": 2690 + }, + { + "epoch": 0.6983674527079554, + "grad_norm": 0.3635096529171273, + "learning_rate": 5.049696218424969e-05, + "loss": 1.1428, + "step": 2695 + }, + { + "epoch": 0.699663125161959, + "grad_norm": 0.3878749986675832, + "learning_rate": 5.010448778794303e-05, + "loss": 1.1741, + "step": 2700 + }, + { + "epoch": 0.7009587976159627, + "grad_norm": 0.40169342624996973, + "learning_rate": 4.971303407475618e-05, + "loss": 1.1599, + "step": 2705 + }, + { + "epoch": 0.7022544700699663, + "grad_norm": 0.3694514584240007, + "learning_rate": 4.932260905242731e-05, + "loss": 1.1187, + "step": 2710 + }, + { + "epoch": 0.70355014252397, + "grad_norm": 0.40107470744868656, + "learning_rate": 4.893322070765126e-05, + "loss": 1.1087, + "step": 2715 + }, + { + "epoch": 0.7048458149779736, + "grad_norm": 0.3973309016316779, + "learning_rate": 4.8544877005916126e-05, + "loss": 1.1353, + "step": 2720 + }, + { + "epoch": 0.7061414874319772, + "grad_norm": 0.36945482581363814, + "learning_rate": 4.815758589134046e-05, + "loss": 1.1537, + "step": 2725 + }, + { + "epoch": 0.7074371598859808, + "grad_norm": 0.39405014852767495, + "learning_rate": 4.777135528651058e-05, + "loss": 1.1864, + "step": 2730 + }, + { + "epoch": 0.7087328323399844, + "grad_norm": 0.36438744458294, + "learning_rate": 4.738619309231857e-05, + "loss": 1.1571, + "step": 2735 + }, + { + "epoch": 0.710028504793988, + "grad_norm": 0.3718851263620155, + "learning_rate": 4.700210718780072e-05, + "loss": 1.1375, + "step": 2740 + }, + { + "epoch": 0.7113241772479917, + "grad_norm": 0.3968100064213781, + "learning_rate": 4.6619105429976193e-05, + "loss": 1.0992, + "step": 2745 + }, + { + "epoch": 0.7126198497019953, + "grad_norm": 0.3735407865710053, + "learning_rate": 4.623719565368657e-05, + "loss": 1.154, + "step": 2750 + }, + { + "epoch": 0.713915522155999, + "grad_norm": 0.3855769481718261, + "learning_rate": 4.585638567143529e-05, + "loss": 1.1087, + "step": 2755 + }, + { + "epoch": 0.7152111946100026, + "grad_norm": 0.39904660177632, + "learning_rate": 4.547668327322796e-05, + "loss": 1.1557, + "step": 2760 + }, + { + "epoch": 0.7165068670640062, + "grad_norm": 0.37552561838186715, + "learning_rate": 4.5098096226413e-05, + "loss": 1.1788, + "step": 2765 + }, + { + "epoch": 0.7178025395180099, + "grad_norm": 0.39324730815800873, + "learning_rate": 4.472063227552274e-05, + "loss": 1.1688, + "step": 2770 + }, + { + "epoch": 0.7190982119720135, + "grad_norm": 0.38301296634310333, + "learning_rate": 4.434429914211508e-05, + "loss": 1.1711, + "step": 2775 + }, + { + "epoch": 0.7203938844260172, + "grad_norm": 0.3709443062095975, + "learning_rate": 4.396910452461532e-05, + "loss": 1.1718, + "step": 2780 + }, + { + "epoch": 0.7216895568800207, + "grad_norm": 0.39239048686699185, + "learning_rate": 4.3595056098158906e-05, + "loss": 1.1574, + "step": 2785 + }, + { + "epoch": 0.7229852293340243, + "grad_norm": 0.41357484414155377, + "learning_rate": 4.322216151443428e-05, + "loss": 1.1074, + "step": 2790 + }, + { + "epoch": 0.724280901788028, + "grad_norm": 0.3672227895875153, + "learning_rate": 4.2850428401526376e-05, + "loss": 1.1028, + "step": 2795 + }, + { + "epoch": 0.7255765742420316, + "grad_norm": 0.3664872416848814, + "learning_rate": 4.2479864363760726e-05, + "loss": 1.1508, + "step": 2800 + }, + { + "epoch": 0.7268722466960352, + "grad_norm": 0.3710822782612391, + "learning_rate": 4.211047698154765e-05, + "loss": 1.1369, + "step": 2805 + }, + { + "epoch": 0.7281679191500389, + "grad_norm": 0.38199563385579133, + "learning_rate": 4.1742273811227395e-05, + "loss": 1.1716, + "step": 2810 + }, + { + "epoch": 0.7294635916040425, + "grad_norm": 0.36978932404422615, + "learning_rate": 4.1375262384915433e-05, + "loss": 1.1613, + "step": 2815 + }, + { + "epoch": 0.7307592640580461, + "grad_norm": 0.37568319756030377, + "learning_rate": 4.100945021034843e-05, + "loss": 1.1396, + "step": 2820 + }, + { + "epoch": 0.7320549365120498, + "grad_norm": 0.3802598263999195, + "learning_rate": 4.064484477073074e-05, + "loss": 1.172, + "step": 2825 + }, + { + "epoch": 0.7333506089660534, + "grad_norm": 0.3690252573947318, + "learning_rate": 4.028145352458118e-05, + "loss": 1.1413, + "step": 2830 + }, + { + "epoch": 0.7346462814200571, + "grad_norm": 0.3942754823686454, + "learning_rate": 3.991928390558054e-05, + "loss": 1.1069, + "step": 2835 + }, + { + "epoch": 0.7359419538740606, + "grad_norm": 0.3720821371337478, + "learning_rate": 3.95583433224195e-05, + "loss": 1.0835, + "step": 2840 + }, + { + "epoch": 0.7372376263280642, + "grad_norm": 0.3807303246565132, + "learning_rate": 3.9198639158647056e-05, + "loss": 1.1501, + "step": 2845 + }, + { + "epoch": 0.7385332987820679, + "grad_norm": 0.36231750839777666, + "learning_rate": 3.884017877251959e-05, + "loss": 1.0563, + "step": 2850 + }, + { + "epoch": 0.7398289712360715, + "grad_norm": 0.388874708381432, + "learning_rate": 3.8482969496850166e-05, + "loss": 1.0889, + "step": 2855 + }, + { + "epoch": 0.7411246436900751, + "grad_norm": 0.38013479771110215, + "learning_rate": 3.812701863885865e-05, + "loss": 1.1145, + "step": 2860 + }, + { + "epoch": 0.7424203161440788, + "grad_norm": 0.3717583300327742, + "learning_rate": 3.7772333480022185e-05, + "loss": 1.0663, + "step": 2865 + }, + { + "epoch": 0.7437159885980824, + "grad_norm": 0.36405208418925755, + "learning_rate": 3.741892127592625e-05, + "loss": 1.1102, + "step": 2870 + }, + { + "epoch": 0.745011661052086, + "grad_norm": 0.3723066034084609, + "learning_rate": 3.706678925611629e-05, + "loss": 1.1602, + "step": 2875 + }, + { + "epoch": 0.7463073335060897, + "grad_norm": 0.3783951434167031, + "learning_rate": 3.67159446239497e-05, + "loss": 1.1233, + "step": 2880 + }, + { + "epoch": 0.7476030059600933, + "grad_norm": 0.37951067018123125, + "learning_rate": 3.636639455644858e-05, + "loss": 1.0545, + "step": 2885 + }, + { + "epoch": 0.748898678414097, + "grad_norm": 0.37421338215744115, + "learning_rate": 3.601814620415296e-05, + "loss": 1.1367, + "step": 2890 + }, + { + "epoch": 0.7501943508681005, + "grad_norm": 0.3603159123651838, + "learning_rate": 3.567120669097422e-05, + "loss": 1.0909, + "step": 2895 + }, + { + "epoch": 0.7514900233221041, + "grad_norm": 0.3788231837374063, + "learning_rate": 3.532558311404986e-05, + "loss": 1.1441, + "step": 2900 + }, + { + "epoch": 0.7527856957761078, + "grad_norm": 0.3840910679589615, + "learning_rate": 3.498128254359788e-05, + "loss": 1.1403, + "step": 2905 + }, + { + "epoch": 0.7540813682301114, + "grad_norm": 0.3753865209395276, + "learning_rate": 3.4638312022772335e-05, + "loss": 1.1004, + "step": 2910 + }, + { + "epoch": 0.755377040684115, + "grad_norm": 0.3954242259257984, + "learning_rate": 3.4296678567519345e-05, + "loss": 1.1347, + "step": 2915 + }, + { + "epoch": 0.7566727131381187, + "grad_norm": 0.3741932569564154, + "learning_rate": 3.3956389166433276e-05, + "loss": 1.0908, + "step": 2920 + }, + { + "epoch": 0.7579683855921223, + "grad_norm": 0.3930950040373874, + "learning_rate": 3.361745078061416e-05, + "loss": 1.1988, + "step": 2925 + }, + { + "epoch": 0.759264058046126, + "grad_norm": 0.3685363645101882, + "learning_rate": 3.327987034352499e-05, + "loss": 1.2067, + "step": 2930 + }, + { + "epoch": 0.7605597305001296, + "grad_norm": 0.3595878774367224, + "learning_rate": 3.294365476085001e-05, + "loss": 1.1128, + "step": 2935 + }, + { + "epoch": 0.7618554029541332, + "grad_norm": 0.3823858203204671, + "learning_rate": 3.260881091035356e-05, + "loss": 1.1894, + "step": 2940 + }, + { + "epoch": 0.7631510754081369, + "grad_norm": 0.35550275725905045, + "learning_rate": 3.227534564173903e-05, + "loss": 1.0744, + "step": 2945 + }, + { + "epoch": 0.7644467478621404, + "grad_norm": 0.38923682484253824, + "learning_rate": 3.1943265776509215e-05, + "loss": 1.1599, + "step": 2950 + }, + { + "epoch": 0.765742420316144, + "grad_norm": 0.3789384247163881, + "learning_rate": 3.1612578107826375e-05, + "loss": 1.1504, + "step": 2955 + }, + { + "epoch": 0.7670380927701477, + "grad_norm": 0.3620246203990784, + "learning_rate": 3.128328940037345e-05, + "loss": 1.0777, + "step": 2960 + }, + { + "epoch": 0.7683337652241513, + "grad_norm": 0.36815557923240855, + "learning_rate": 3.095540639021578e-05, + "loss": 1.1174, + "step": 2965 + }, + { + "epoch": 0.769629437678155, + "grad_norm": 0.3547702710213197, + "learning_rate": 3.0628935784662947e-05, + "loss": 1.1485, + "step": 2970 + }, + { + "epoch": 0.7709251101321586, + "grad_norm": 0.3781571691176777, + "learning_rate": 3.030388426213202e-05, + "loss": 1.1883, + "step": 2975 + }, + { + "epoch": 0.7722207825861622, + "grad_norm": 0.40077870265330406, + "learning_rate": 2.9980258472010624e-05, + "loss": 1.1932, + "step": 2980 + }, + { + "epoch": 0.7735164550401659, + "grad_norm": 0.4044666107500387, + "learning_rate": 2.9658065034520978e-05, + "loss": 1.1241, + "step": 2985 + }, + { + "epoch": 0.7748121274941695, + "grad_norm": 0.38599774202480436, + "learning_rate": 2.9337310540584662e-05, + "loss": 1.1289, + "step": 2990 + }, + { + "epoch": 0.7761077999481731, + "grad_norm": 0.3937959908311204, + "learning_rate": 2.90180015516874e-05, + "loss": 1.124, + "step": 2995 + }, + { + "epoch": 0.7774034724021768, + "grad_norm": 0.3632550787528134, + "learning_rate": 2.8700144599745304e-05, + "loss": 1.054, + "step": 3000 + }, + { + "epoch": 0.7786991448561803, + "grad_norm": 0.3622147641579977, + "learning_rate": 2.8383746186970885e-05, + "loss": 1.1541, + "step": 3005 + }, + { + "epoch": 0.7799948173101839, + "grad_norm": 0.37112400149549435, + "learning_rate": 2.806881278574016e-05, + "loss": 1.0959, + "step": 3010 + }, + { + "epoch": 0.7812904897641876, + "grad_norm": 0.35896620554471, + "learning_rate": 2.7755350838460437e-05, + "loss": 1.1097, + "step": 3015 + }, + { + "epoch": 0.7825861622181912, + "grad_norm": 0.3722645573928852, + "learning_rate": 2.7443366757438084e-05, + "loss": 1.1536, + "step": 3020 + }, + { + "epoch": 0.7838818346721949, + "grad_norm": 0.3733728824194641, + "learning_rate": 2.71328669247479e-05, + "loss": 1.0988, + "step": 3025 + }, + { + "epoch": 0.7851775071261985, + "grad_norm": 0.39726305827044567, + "learning_rate": 2.6823857692102115e-05, + "loss": 1.1347, + "step": 3030 + }, + { + "epoch": 0.7864731795802021, + "grad_norm": 0.3757164187600116, + "learning_rate": 2.6516345380720685e-05, + "loss": 1.1498, + "step": 3035 + }, + { + "epoch": 0.7877688520342058, + "grad_norm": 0.3795959004315457, + "learning_rate": 2.6210336281201996e-05, + "loss": 1.1716, + "step": 3040 + }, + { + "epoch": 0.7890645244882094, + "grad_norm": 0.35714774223397056, + "learning_rate": 2.5905836653393955e-05, + "loss": 1.1829, + "step": 3045 + }, + { + "epoch": 0.790360196942213, + "grad_norm": 0.3654895953286083, + "learning_rate": 2.5602852726266246e-05, + "loss": 1.1138, + "step": 3050 + }, + { + "epoch": 0.7916558693962167, + "grad_norm": 0.3790228967067868, + "learning_rate": 2.53013906977827e-05, + "loss": 1.1361, + "step": 3055 + }, + { + "epoch": 0.7929515418502202, + "grad_norm": 0.3766214688716055, + "learning_rate": 2.500145673477452e-05, + "loss": 1.1386, + "step": 3060 + }, + { + "epoch": 0.7942472143042238, + "grad_norm": 0.35689403644408657, + "learning_rate": 2.4703056972814298e-05, + "loss": 1.1716, + "step": 3065 + }, + { + "epoch": 0.7955428867582275, + "grad_norm": 0.3837337630770634, + "learning_rate": 2.44061975160902e-05, + "loss": 1.1253, + "step": 3070 + }, + { + "epoch": 0.7968385592122311, + "grad_norm": 0.3617908256654209, + "learning_rate": 2.4110884437281433e-05, + "loss": 1.1669, + "step": 3075 + }, + { + "epoch": 0.7981342316662348, + "grad_norm": 0.35978473851190906, + "learning_rate": 2.381712377743379e-05, + "loss": 1.1195, + "step": 3080 + }, + { + "epoch": 0.7994299041202384, + "grad_norm": 0.36747445883447966, + "learning_rate": 2.352492154583611e-05, + "loss": 1.1494, + "step": 3085 + }, + { + "epoch": 0.800725576574242, + "grad_norm": 0.38717115547376924, + "learning_rate": 2.323428371989752e-05, + "loss": 1.1342, + "step": 3090 + }, + { + "epoch": 0.8020212490282457, + "grad_norm": 0.39093054065016897, + "learning_rate": 2.2945216245024804e-05, + "loss": 1.1304, + "step": 3095 + }, + { + "epoch": 0.8033169214822493, + "grad_norm": 0.368756140981371, + "learning_rate": 2.265772503450122e-05, + "loss": 1.1902, + "step": 3100 + }, + { + "epoch": 0.804612593936253, + "grad_norm": 0.3877482016759796, + "learning_rate": 2.237181596936515e-05, + "loss": 1.2071, + "step": 3105 + }, + { + "epoch": 0.8059082663902566, + "grad_norm": 0.37792584298436904, + "learning_rate": 2.2087494898290084e-05, + "loss": 1.1263, + "step": 3110 + }, + { + "epoch": 0.8072039388442601, + "grad_norm": 0.377175581709425, + "learning_rate": 2.1804767637464783e-05, + "loss": 1.1409, + "step": 3115 + }, + { + "epoch": 0.8084996112982638, + "grad_norm": 0.3675212675701305, + "learning_rate": 2.152363997047432e-05, + "loss": 1.1222, + "step": 3120 + }, + { + "epoch": 0.8097952837522674, + "grad_norm": 0.38222618034509426, + "learning_rate": 2.1244117648181926e-05, + "loss": 1.1282, + "step": 3125 + }, + { + "epoch": 0.811090956206271, + "grad_norm": 0.36558116292433085, + "learning_rate": 2.0966206388611177e-05, + "loss": 1.1599, + "step": 3130 + }, + { + "epoch": 0.8123866286602747, + "grad_norm": 0.3727524846387591, + "learning_rate": 2.0689911876829127e-05, + "loss": 1.1496, + "step": 3135 + }, + { + "epoch": 0.8136823011142783, + "grad_norm": 0.3628037534722254, + "learning_rate": 2.0415239764829976e-05, + "loss": 1.1416, + "step": 3140 + }, + { + "epoch": 0.8149779735682819, + "grad_norm": 0.3761023568840631, + "learning_rate": 2.014219567141932e-05, + "loss": 1.0828, + "step": 3145 + }, + { + "epoch": 0.8162736460222856, + "grad_norm": 0.3504341651004141, + "learning_rate": 1.9870785182099505e-05, + "loss": 1.0956, + "step": 3150 + }, + { + "epoch": 0.8175693184762892, + "grad_norm": 0.3734885666536526, + "learning_rate": 1.960101384895511e-05, + "loss": 1.1044, + "step": 3155 + }, + { + "epoch": 0.8188649909302929, + "grad_norm": 0.386535956936462, + "learning_rate": 1.9332887190539516e-05, + "loss": 1.2004, + "step": 3160 + }, + { + "epoch": 0.8201606633842965, + "grad_norm": 0.3646373120550804, + "learning_rate": 1.9066410691761937e-05, + "loss": 1.1377, + "step": 3165 + }, + { + "epoch": 0.8214563358383, + "grad_norm": 0.3715486267985764, + "learning_rate": 1.8801589803775154e-05, + "loss": 1.1128, + "step": 3170 + }, + { + "epoch": 0.8227520082923037, + "grad_norm": 0.3789134249781072, + "learning_rate": 1.8538429943864244e-05, + "loss": 1.1429, + "step": 3175 + }, + { + "epoch": 0.8240476807463073, + "grad_norm": 0.4352073640959862, + "learning_rate": 1.8276936495335485e-05, + "loss": 1.1167, + "step": 3180 + }, + { + "epoch": 0.8253433532003109, + "grad_norm": 0.3681435590124036, + "learning_rate": 1.8017114807406478e-05, + "loss": 1.1063, + "step": 3185 + }, + { + "epoch": 0.8266390256543146, + "grad_norm": 0.3694513270515317, + "learning_rate": 1.775897019509649e-05, + "loss": 1.1836, + "step": 3190 + }, + { + "epoch": 0.8279346981083182, + "grad_norm": 0.388666401561133, + "learning_rate": 1.7502507939117897e-05, + "loss": 1.1396, + "step": 3195 + }, + { + "epoch": 0.8292303705623219, + "grad_norm": 0.3687281974613644, + "learning_rate": 1.7247733285768098e-05, + "loss": 1.1352, + "step": 3200 + }, + { + "epoch": 0.8305260430163255, + "grad_norm": 0.3733462546424826, + "learning_rate": 1.6994651446822153e-05, + "loss": 1.125, + "step": 3205 + }, + { + "epoch": 0.8318217154703291, + "grad_norm": 0.3745095547011446, + "learning_rate": 1.6743267599426303e-05, + "loss": 1.1645, + "step": 3210 + }, + { + "epoch": 0.8331173879243328, + "grad_norm": 0.36579848745894494, + "learning_rate": 1.649358688599191e-05, + "loss": 1.1373, + "step": 3215 + }, + { + "epoch": 0.8344130603783364, + "grad_norm": 0.37060439276258045, + "learning_rate": 1.624561441409034e-05, + "loss": 1.0775, + "step": 3220 + }, + { + "epoch": 0.83570873283234, + "grad_norm": 0.38418584872461575, + "learning_rate": 1.5999355256348448e-05, + "loss": 1.1678, + "step": 3225 + }, + { + "epoch": 0.8370044052863436, + "grad_norm": 0.3814911739387983, + "learning_rate": 1.5754814450344845e-05, + "loss": 1.1772, + "step": 3230 + }, + { + "epoch": 0.8383000777403472, + "grad_norm": 0.3642771788771785, + "learning_rate": 1.5511996998506883e-05, + "loss": 1.1279, + "step": 3235 + }, + { + "epoch": 0.8395957501943508, + "grad_norm": 0.3798091083229991, + "learning_rate": 1.527090786800821e-05, + "loss": 1.1663, + "step": 3240 + }, + { + "epoch": 0.8408914226483545, + "grad_norm": 0.3621301040720381, + "learning_rate": 1.5031551990667236e-05, + "loss": 1.0961, + "step": 3245 + }, + { + "epoch": 0.8421870951023581, + "grad_norm": 0.3716900347036682, + "learning_rate": 1.4793934262846232e-05, + "loss": 1.1246, + "step": 3250 + }, + { + "epoch": 0.8434827675563618, + "grad_norm": 0.3431347363725958, + "learning_rate": 1.4558059545351143e-05, + "loss": 1.126, + "step": 3255 + }, + { + "epoch": 0.8447784400103654, + "grad_norm": 0.38516161930376763, + "learning_rate": 1.4323932663332251e-05, + "loss": 1.1487, + "step": 3260 + }, + { + "epoch": 0.846074112464369, + "grad_norm": 0.36582214467557017, + "learning_rate": 1.4091558406185335e-05, + "loss": 1.1464, + "step": 3265 + }, + { + "epoch": 0.8473697849183727, + "grad_norm": 0.3600375664520519, + "learning_rate": 1.3860941527453786e-05, + "loss": 1.1331, + "step": 3270 + }, + { + "epoch": 0.8486654573723763, + "grad_norm": 0.36937163487487523, + "learning_rate": 1.3632086744731299e-05, + "loss": 1.1, + "step": 3275 + }, + { + "epoch": 0.84996112982638, + "grad_norm": 0.37811170010742284, + "learning_rate": 1.3404998739565433e-05, + "loss": 1.1278, + "step": 3280 + }, + { + "epoch": 0.8512568022803835, + "grad_norm": 0.36422197248748484, + "learning_rate": 1.3179682157361872e-05, + "loss": 1.1193, + "step": 3285 + }, + { + "epoch": 0.8525524747343871, + "grad_norm": 0.37165480607350426, + "learning_rate": 1.2956141607289262e-05, + "loss": 1.1615, + "step": 3290 + }, + { + "epoch": 0.8538481471883907, + "grad_norm": 0.37701311154160283, + "learning_rate": 1.2734381662185035e-05, + "loss": 1.1036, + "step": 3295 + }, + { + "epoch": 0.8551438196423944, + "grad_norm": 0.3815170151959314, + "learning_rate": 1.2514406858461847e-05, + "loss": 1.1735, + "step": 3300 + }, + { + "epoch": 0.856439492096398, + "grad_norm": 0.37919701821847407, + "learning_rate": 1.2296221696014732e-05, + "loss": 1.1837, + "step": 3305 + }, + { + "epoch": 0.8577351645504017, + "grad_norm": 0.36743928128733283, + "learning_rate": 1.2079830638129164e-05, + "loss": 1.1148, + "step": 3310 + }, + { + "epoch": 0.8590308370044053, + "grad_norm": 0.3738619978563127, + "learning_rate": 1.1865238111389588e-05, + "loss": 1.1604, + "step": 3315 + }, + { + "epoch": 0.8603265094584089, + "grad_norm": 0.3541904420302459, + "learning_rate": 1.1652448505588998e-05, + "loss": 1.1544, + "step": 3320 + }, + { + "epoch": 0.8616221819124126, + "grad_norm": 0.36684430135825935, + "learning_rate": 1.1441466173639092e-05, + "loss": 1.156, + "step": 3325 + }, + { + "epoch": 0.8629178543664162, + "grad_norm": 0.37365304474041194, + "learning_rate": 1.1232295431481222e-05, + "loss": 1.1084, + "step": 3330 + }, + { + "epoch": 0.8642135268204199, + "grad_norm": 0.3810375689519541, + "learning_rate": 1.1024940557998143e-05, + "loss": 1.1386, + "step": 3335 + }, + { + "epoch": 0.8655091992744234, + "grad_norm": 0.3673719354100514, + "learning_rate": 1.0819405794926418e-05, + "loss": 1.1629, + "step": 3340 + }, + { + "epoch": 0.866804871728427, + "grad_norm": 0.37342554741406375, + "learning_rate": 1.0615695346769693e-05, + "loss": 1.1134, + "step": 3345 + }, + { + "epoch": 0.8681005441824307, + "grad_norm": 0.3635218597663001, + "learning_rate": 1.0413813380712701e-05, + "loss": 1.1147, + "step": 3350 + }, + { + "epoch": 0.8693962166364343, + "grad_norm": 0.3570736451827848, + "learning_rate": 1.0213764026535921e-05, + "loss": 1.1321, + "step": 3355 + }, + { + "epoch": 0.8706918890904379, + "grad_norm": 0.3891720605828237, + "learning_rate": 1.0015551376531296e-05, + "loss": 1.1001, + "step": 3360 + }, + { + "epoch": 0.8719875615444416, + "grad_norm": 0.37097051363645606, + "learning_rate": 9.819179485418273e-06, + "loss": 1.1241, + "step": 3365 + }, + { + "epoch": 0.8732832339984452, + "grad_norm": 0.3917978913676709, + "learning_rate": 9.624652370261034e-06, + "loss": 1.1304, + "step": 3370 + }, + { + "epoch": 0.8745789064524488, + "grad_norm": 0.37226262222810586, + "learning_rate": 9.431974010386258e-06, + "loss": 1.1421, + "step": 3375 + }, + { + "epoch": 0.8758745789064525, + "grad_norm": 0.3751233666218166, + "learning_rate": 9.24114834730171e-06, + "loss": 1.1022, + "step": 3380 + }, + { + "epoch": 0.8771702513604561, + "grad_norm": 0.36555054148949856, + "learning_rate": 9.052179284615658e-06, + "loss": 1.13, + "step": 3385 + }, + { + "epoch": 0.8784659238144598, + "grad_norm": 0.38148359927958986, + "learning_rate": 8.865070687956977e-06, + "loss": 1.1894, + "step": 3390 + }, + { + "epoch": 0.8797615962684633, + "grad_norm": 0.416874595298691, + "learning_rate": 8.679826384896061e-06, + "loss": 1.1736, + "step": 3395 + }, + { + "epoch": 0.8810572687224669, + "grad_norm": 0.387504501992165, + "learning_rate": 8.496450164866565e-06, + "loss": 1.1511, + "step": 3400 + }, + { + "epoch": 0.8823529411764706, + "grad_norm": 0.34801518476142257, + "learning_rate": 8.314945779087858e-06, + "loss": 1.1224, + "step": 3405 + }, + { + "epoch": 0.8836486136304742, + "grad_norm": 0.3758149676256447, + "learning_rate": 8.135316940488347e-06, + "loss": 1.1448, + "step": 3410 + }, + { + "epoch": 0.8849442860844778, + "grad_norm": 0.36726096701274225, + "learning_rate": 7.9575673236294e-06, + "loss": 1.1719, + "step": 3415 + }, + { + "epoch": 0.8862399585384815, + "grad_norm": 0.3578469295777726, + "learning_rate": 7.781700564630311e-06, + "loss": 1.1385, + "step": 3420 + }, + { + "epoch": 0.8875356309924851, + "grad_norm": 0.38263190643853745, + "learning_rate": 7.6077202610938205e-06, + "loss": 1.177, + "step": 3425 + }, + { + "epoch": 0.8888313034464888, + "grad_norm": 0.38665333786102024, + "learning_rate": 7.435629972032565e-06, + "loss": 1.1358, + "step": 3430 + }, + { + "epoch": 0.8901269759004924, + "grad_norm": 0.36097779607915553, + "learning_rate": 7.265433217796302e-06, + "loss": 1.1055, + "step": 3435 + }, + { + "epoch": 0.891422648354496, + "grad_norm": 0.4052378204511826, + "learning_rate": 7.0971334799998404e-06, + "loss": 1.0912, + "step": 3440 + }, + { + "epoch": 0.8927183208084997, + "grad_norm": 0.3704412054341248, + "learning_rate": 6.930734201451816e-06, + "loss": 1.1591, + "step": 3445 + }, + { + "epoch": 0.8940139932625032, + "grad_norm": 0.37555016415939446, + "learning_rate": 6.7662387860843225e-06, + "loss": 1.1149, + "step": 3450 + }, + { + "epoch": 0.8953096657165068, + "grad_norm": 0.3798653467067839, + "learning_rate": 6.603650598883226e-06, + "loss": 1.0958, + "step": 3455 + }, + { + "epoch": 0.8966053381705105, + "grad_norm": 0.34342261742501184, + "learning_rate": 6.4429729658193714e-06, + "loss": 1.0718, + "step": 3460 + }, + { + "epoch": 0.8979010106245141, + "grad_norm": 0.38053269516746213, + "learning_rate": 6.284209173780498e-06, + "loss": 1.1133, + "step": 3465 + }, + { + "epoch": 0.8991966830785177, + "grad_norm": 0.38564284280238426, + "learning_rate": 6.127362470504027e-06, + "loss": 1.1239, + "step": 3470 + }, + { + "epoch": 0.9004923555325214, + "grad_norm": 0.3624213790138937, + "learning_rate": 5.972436064510667e-06, + "loss": 1.1766, + "step": 3475 + }, + { + "epoch": 0.901788027986525, + "grad_norm": 0.36614157981464224, + "learning_rate": 5.819433125038643e-06, + "loss": 1.146, + "step": 3480 + }, + { + "epoch": 0.9030837004405287, + "grad_norm": 0.39479288263797235, + "learning_rate": 5.668356781979056e-06, + "loss": 1.1669, + "step": 3485 + }, + { + "epoch": 0.9043793728945323, + "grad_norm": 0.35385937176801885, + "learning_rate": 5.519210125811713e-06, + "loss": 1.1224, + "step": 3490 + }, + { + "epoch": 0.9056750453485359, + "grad_norm": 0.3778900407632985, + "learning_rate": 5.371996207541941e-06, + "loss": 1.1287, + "step": 3495 + }, + { + "epoch": 0.9069707178025396, + "grad_norm": 0.35939889684365445, + "learning_rate": 5.226718038638234e-06, + "loss": 1.1274, + "step": 3500 + }, + { + "epoch": 0.9082663902565431, + "grad_norm": 0.38058001351217496, + "learning_rate": 5.083378590970511e-06, + "loss": 1.154, + "step": 3505 + }, + { + "epoch": 0.9095620627105467, + "grad_norm": 0.36339105009188977, + "learning_rate": 4.941980796749524e-06, + "loss": 1.1234, + "step": 3510 + }, + { + "epoch": 0.9108577351645504, + "grad_norm": 0.3753439481173232, + "learning_rate": 4.802527548466684e-06, + "loss": 1.0982, + "step": 3515 + }, + { + "epoch": 0.912153407618554, + "grad_norm": 0.3547606097647797, + "learning_rate": 4.665021698834981e-06, + "loss": 1.1405, + "step": 3520 + }, + { + "epoch": 0.9134490800725577, + "grad_norm": 0.3804573132062168, + "learning_rate": 4.529466060730681e-06, + "loss": 1.1529, + "step": 3525 + }, + { + "epoch": 0.9147447525265613, + "grad_norm": 0.3766860837932323, + "learning_rate": 4.395863407135614e-06, + "loss": 1.0821, + "step": 3530 + }, + { + "epoch": 0.9160404249805649, + "grad_norm": 0.3760175631598513, + "learning_rate": 4.264216471080651e-06, + "loss": 1.1451, + "step": 3535 + }, + { + "epoch": 0.9173360974345686, + "grad_norm": 0.3901693827353405, + "learning_rate": 4.134527945589639e-06, + "loss": 1.1148, + "step": 3540 + }, + { + "epoch": 0.9186317698885722, + "grad_norm": 0.4488796382773471, + "learning_rate": 4.006800483624396e-06, + "loss": 1.1614, + "step": 3545 + }, + { + "epoch": 0.9199274423425758, + "grad_norm": 0.40897274265302724, + "learning_rate": 3.881036698030449e-06, + "loss": 1.1744, + "step": 3550 + }, + { + "epoch": 0.9212231147965795, + "grad_norm": 0.3690887838305789, + "learning_rate": 3.7572391614834833e-06, + "loss": 1.136, + "step": 3555 + }, + { + "epoch": 0.922518787250583, + "grad_norm": 0.3688927093580592, + "learning_rate": 3.6354104064368566e-06, + "loss": 1.0947, + "step": 3560 + }, + { + "epoch": 0.9238144597045866, + "grad_norm": 0.3826643203341735, + "learning_rate": 3.515552925069676e-06, + "loss": 1.1757, + "step": 3565 + }, + { + "epoch": 0.9251101321585903, + "grad_norm": 0.3684225941093978, + "learning_rate": 3.397669169235862e-06, + "loss": 1.1574, + "step": 3570 + }, + { + "epoch": 0.9264058046125939, + "grad_norm": 0.3447005114527934, + "learning_rate": 3.281761550414042e-06, + "loss": 1.0847, + "step": 3575 + }, + { + "epoch": 0.9277014770665976, + "grad_norm": 0.38643242398136646, + "learning_rate": 3.1678324396581137e-06, + "loss": 1.2221, + "step": 3580 + }, + { + "epoch": 0.9289971495206012, + "grad_norm": 0.37779005091381196, + "learning_rate": 3.0558841675488393e-06, + "loss": 1.1612, + "step": 3585 + }, + { + "epoch": 0.9302928219746048, + "grad_norm": 0.38097475759248384, + "learning_rate": 2.9459190241461043e-06, + "loss": 1.1267, + "step": 3590 + }, + { + "epoch": 0.9315884944286085, + "grad_norm": 0.37618918488273334, + "learning_rate": 2.8379392589421237e-06, + "loss": 1.1563, + "step": 3595 + }, + { + "epoch": 0.9328841668826121, + "grad_norm": 0.3621451372937871, + "learning_rate": 2.7319470808153892e-06, + "loss": 1.1197, + "step": 3600 + }, + { + "epoch": 0.9341798393366157, + "grad_norm": 0.35758689530964316, + "learning_rate": 2.6279446579854707e-06, + "loss": 1.1446, + "step": 3605 + }, + { + "epoch": 0.9354755117906194, + "grad_norm": 0.3803519145314357, + "learning_rate": 2.525934117968731e-06, + "loss": 1.1823, + "step": 3610 + }, + { + "epoch": 0.9367711842446229, + "grad_norm": 0.36540749180301496, + "learning_rate": 2.4259175475347172e-06, + "loss": 1.1748, + "step": 3615 + }, + { + "epoch": 0.9380668566986265, + "grad_norm": 0.34819215016667493, + "learning_rate": 2.3278969926635252e-06, + "loss": 1.1036, + "step": 3620 + }, + { + "epoch": 0.9393625291526302, + "grad_norm": 0.3586884150884424, + "learning_rate": 2.2318744585039796e-06, + "loss": 1.1192, + "step": 3625 + }, + { + "epoch": 0.9406582016066338, + "grad_norm": 0.37877938188955895, + "learning_rate": 2.1378519093324776e-06, + "loss": 1.128, + "step": 3630 + }, + { + "epoch": 0.9419538740606375, + "grad_norm": 0.3781105512240613, + "learning_rate": 2.0458312685129876e-06, + "loss": 1.1612, + "step": 3635 + }, + { + "epoch": 0.9432495465146411, + "grad_norm": 0.37487413050848234, + "learning_rate": 1.955814418457591e-06, + "loss": 1.1396, + "step": 3640 + }, + { + "epoch": 0.9445452189686447, + "grad_norm": 0.3840114910436269, + "learning_rate": 1.867803200587981e-06, + "loss": 1.1129, + "step": 3645 + }, + { + "epoch": 0.9458408914226484, + "grad_norm": 0.3719048343830175, + "learning_rate": 1.7817994152978468e-06, + "loss": 1.1157, + "step": 3650 + }, + { + "epoch": 0.947136563876652, + "grad_norm": 0.3772856287062892, + "learning_rate": 1.6978048219159714e-06, + "loss": 1.1194, + "step": 3655 + }, + { + "epoch": 0.9484322363306557, + "grad_norm": 0.3598408324385285, + "learning_rate": 1.6158211386703259e-06, + "loss": 1.1284, + "step": 3660 + }, + { + "epoch": 0.9497279087846593, + "grad_norm": 0.365397297934618, + "learning_rate": 1.5358500426528422e-06, + "loss": 1.1282, + "step": 3665 + }, + { + "epoch": 0.9510235812386628, + "grad_norm": 0.3659664905751428, + "learning_rate": 1.4578931697851406e-06, + "loss": 1.1051, + "step": 3670 + }, + { + "epoch": 0.9523192536926665, + "grad_norm": 0.3549703867179501, + "learning_rate": 1.3819521147851123e-06, + "loss": 1.1292, + "step": 3675 + }, + { + "epoch": 0.9536149261466701, + "grad_norm": 0.36856057544037374, + "learning_rate": 1.3080284311341674e-06, + "loss": 1.141, + "step": 3680 + }, + { + "epoch": 0.9549105986006737, + "grad_norm": 0.3869169065095209, + "learning_rate": 1.236123631045627e-06, + "loss": 1.1622, + "step": 3685 + }, + { + "epoch": 0.9562062710546774, + "grad_norm": 0.35481859423446044, + "learning_rate": 1.1662391854336263e-06, + "loss": 1.1223, + "step": 3690 + }, + { + "epoch": 0.957501943508681, + "grad_norm": 0.38700062266195656, + "learning_rate": 1.0983765238831377e-06, + "loss": 1.1501, + "step": 3695 + }, + { + "epoch": 0.9587976159626846, + "grad_norm": 0.3672922868247709, + "learning_rate": 1.032537034620684e-06, + "loss": 1.1419, + "step": 3700 + }, + { + "epoch": 0.9600932884166883, + "grad_norm": 0.37605980694280633, + "learning_rate": 9.687220644859275e-07, + "loss": 1.133, + "step": 3705 + }, + { + "epoch": 0.9613889608706919, + "grad_norm": 0.36877920673269027, + "learning_rate": 9.069329189041464e-07, + "loss": 1.1608, + "step": 3710 + }, + { + "epoch": 0.9626846333246956, + "grad_norm": 0.381237649400668, + "learning_rate": 8.471708618595142e-07, + "loss": 1.132, + "step": 3715 + }, + { + "epoch": 0.9639803057786992, + "grad_norm": 1.8685424500994194, + "learning_rate": 7.894371158692627e-07, + "loss": 1.1656, + "step": 3720 + }, + { + "epoch": 0.9652759782327027, + "grad_norm": 0.38287612615523475, + "learning_rate": 7.337328619586359e-07, + "loss": 1.1992, + "step": 3725 + }, + { + "epoch": 0.9665716506867064, + "grad_norm": 0.3838324502538139, + "learning_rate": 6.800592396367545e-07, + "loss": 1.1816, + "step": 3730 + }, + { + "epoch": 0.96786732314071, + "grad_norm": 0.36218594513869495, + "learning_rate": 6.284173468733334e-07, + "loss": 1.1188, + "step": 3735 + }, + { + "epoch": 0.9691629955947136, + "grad_norm": 0.3605898736094485, + "learning_rate": 5.788082400761563e-07, + "loss": 1.1342, + "step": 3740 + }, + { + "epoch": 0.9704586680487173, + "grad_norm": 0.3819284101383782, + "learning_rate": 5.312329340695143e-07, + "loss": 1.1341, + "step": 3745 + }, + { + "epoch": 0.9717543405027209, + "grad_norm": 0.371840798865179, + "learning_rate": 4.856924020734565e-07, + "loss": 1.1315, + "step": 3750 + }, + { + "epoch": 0.9730500129567246, + "grad_norm": 0.37381875278150484, + "learning_rate": 4.4218757568380563e-07, + "loss": 1.1129, + "step": 3755 + }, + { + "epoch": 0.9743456854107282, + "grad_norm": 0.37126771253731844, + "learning_rate": 4.007193448532065e-07, + "loss": 1.1666, + "step": 3760 + }, + { + "epoch": 0.9756413578647318, + "grad_norm": 0.36988126308198593, + "learning_rate": 3.612885578728298e-07, + "loss": 1.1429, + "step": 3765 + }, + { + "epoch": 0.9769370303187355, + "grad_norm": 0.3778202136491613, + "learning_rate": 3.2389602135507457e-07, + "loss": 1.1985, + "step": 3770 + }, + { + "epoch": 0.9782327027727391, + "grad_norm": 0.3544356950249304, + "learning_rate": 2.885425002170594e-07, + "loss": 1.1469, + "step": 3775 + }, + { + "epoch": 0.9795283752267426, + "grad_norm": 0.37874103417633515, + "learning_rate": 2.5522871766494595e-07, + "loss": 1.1101, + "step": 3780 + }, + { + "epoch": 0.9808240476807463, + "grad_norm": 0.3788020452535586, + "learning_rate": 2.2395535517917287e-07, + "loss": 1.1895, + "step": 3785 + }, + { + "epoch": 0.9821197201347499, + "grad_norm": 0.37493085740769366, + "learning_rate": 1.947230525005006e-07, + "loss": 1.1213, + "step": 3790 + }, + { + "epoch": 0.9834153925887535, + "grad_norm": 0.36932806609876156, + "learning_rate": 1.6753240761693268e-07, + "loss": 1.1141, + "step": 3795 + }, + { + "epoch": 0.9847110650427572, + "grad_norm": 0.3675515005536323, + "learning_rate": 1.4238397675150339e-07, + "loss": 1.1292, + "step": 3800 + }, + { + "epoch": 0.9860067374967608, + "grad_norm": 0.3595954681947242, + "learning_rate": 1.1927827435084248e-07, + "loss": 1.1044, + "step": 3805 + }, + { + "epoch": 0.9873024099507645, + "grad_norm": 0.3750795386829234, + "learning_rate": 9.821577307470575e-08, + "loss": 1.1441, + "step": 3810 + }, + { + "epoch": 0.9885980824047681, + "grad_norm": 0.36749997185699007, + "learning_rate": 7.919690378629385e-08, + "loss": 1.171, + "step": 3815 + }, + { + "epoch": 0.9898937548587717, + "grad_norm": 0.3622505687301632, + "learning_rate": 6.222205554339277e-08, + "loss": 1.1675, + "step": 3820 + }, + { + "epoch": 0.9911894273127754, + "grad_norm": 0.3589951765733612, + "learning_rate": 4.729157559049124e-08, + "loss": 1.1451, + "step": 3825 + }, + { + "epoch": 0.992485099766779, + "grad_norm": 0.372014869272016, + "learning_rate": 3.440576935164197e-08, + "loss": 1.1383, + "step": 3830 + }, + { + "epoch": 0.9937807722207825, + "grad_norm": 0.36261244538328924, + "learning_rate": 2.3564900424188906e-08, + "loss": 1.1706, + "step": 3835 + }, + { + "epoch": 0.9950764446747862, + "grad_norm": 0.37064889363490416, + "learning_rate": 1.4769190573393765e-08, + "loss": 1.1344, + "step": 3840 + }, + { + "epoch": 0.9963721171287898, + "grad_norm": 0.4019334217280328, + "learning_rate": 8.018819727928505e-09, + "loss": 1.19, + "step": 3845 + }, + { + "epoch": 0.9976677895827935, + "grad_norm": 0.3532944069166521, + "learning_rate": 3.313925976156096e-09, + "loss": 1.1704, + "step": 3850 + }, + { + "epoch": 0.9989634620367971, + "grad_norm": 0.3954932358396543, + "learning_rate": 6.546055633105397e-10, + "loss": 1.0999, + "step": 3855 + }, + { + "epoch": 1.0, + "eval_runtime": 3.6908, + "eval_samples_per_second": 2.709, + "eval_steps_per_second": 0.813, + "step": 3859 + }, + { + "epoch": 1.0, + "step": 3859, + "total_flos": 1.3011067827388416e+16, + "train_loss": 1.2484874595598594, + "train_runtime": 22644.1998, + "train_samples_per_second": 2.726, + "train_steps_per_second": 0.17 + } + ], + "logging_steps": 5, + "max_steps": 3859, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.3011067827388416e+16, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}