{ "best_global_step": 148440, "best_metric": 0.7365977168083191, "best_model_checkpoint": "./output_optimized/checkpoint-148440", "epoch": 5.0, "eval_steps": 500, "global_step": 148440, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0033683643222850983, "grad_norm": 1.9248961210250854, "learning_rate": 2.000808407437348e-06, "loss": 5.0547, "step": 100 }, { "epoch": 0.006736728644570197, "grad_norm": 1.0100284814834595, "learning_rate": 4.021827000808407e-06, "loss": 4.9226, "step": 200 }, { "epoch": 0.010105092966855295, "grad_norm": 0.8452675938606262, "learning_rate": 6.042845594179466e-06, "loss": 4.8032, "step": 300 }, { "epoch": 0.013473457289140393, "grad_norm": 0.878520131111145, "learning_rate": 8.063864187550524e-06, "loss": 4.7193, "step": 400 }, { "epoch": 0.016841821611425493, "grad_norm": 1.0584090948104858, "learning_rate": 1.0084882780921583e-05, "loss": 4.5769, "step": 500 }, { "epoch": 0.02021018593371059, "grad_norm": 1.1221024990081787, "learning_rate": 1.2105901374292643e-05, "loss": 4.3638, "step": 600 }, { "epoch": 0.02357855025599569, "grad_norm": 1.0861926078796387, "learning_rate": 1.41269199676637e-05, "loss": 4.155, "step": 700 }, { "epoch": 0.026946914578280787, "grad_norm": 0.9905880093574524, "learning_rate": 1.6147938561034762e-05, "loss": 3.9701, "step": 800 }, { "epoch": 0.030315278900565887, "grad_norm": 0.9011399149894714, "learning_rate": 1.816895715440582e-05, "loss": 3.7974, "step": 900 }, { "epoch": 0.03368364322285099, "grad_norm": 0.8480072617530823, "learning_rate": 2.0189975747776877e-05, "loss": 3.6336, "step": 1000 }, { "epoch": 0.03705200754513608, "grad_norm": 0.7879598736763, "learning_rate": 2.2210994341147935e-05, "loss": 3.4894, "step": 1100 }, { "epoch": 0.04042037186742118, "grad_norm": 0.7154058814048767, "learning_rate": 2.4232012934518997e-05, "loss": 3.3559, "step": 1200 }, { "epoch": 0.04378873618970628, "grad_norm": 0.6783929467201233, "learning_rate": 2.6253031527890058e-05, "loss": 3.235, "step": 1300 }, { "epoch": 0.04715710051199138, "grad_norm": 0.6816830635070801, "learning_rate": 2.8274050121261112e-05, "loss": 3.1325, "step": 1400 }, { "epoch": 0.05052546483427647, "grad_norm": 0.6334635615348816, "learning_rate": 3.029506871463217e-05, "loss": 3.0338, "step": 1500 }, { "epoch": 0.05389382915656157, "grad_norm": 0.6427187323570251, "learning_rate": 3.2316087308003235e-05, "loss": 2.9431, "step": 1600 }, { "epoch": 0.057262193478846674, "grad_norm": 1.2221870422363281, "learning_rate": 3.433710590137429e-05, "loss": 2.8522, "step": 1700 }, { "epoch": 0.060630557801131774, "grad_norm": 1.4814372062683105, "learning_rate": 3.635812449474535e-05, "loss": 2.7769, "step": 1800 }, { "epoch": 0.06399892212341687, "grad_norm": 1.1986867189407349, "learning_rate": 3.837914308811641e-05, "loss": 2.7072, "step": 1900 }, { "epoch": 0.06736728644570197, "grad_norm": 1.0901182889938354, "learning_rate": 4.0400161681487466e-05, "loss": 2.6491, "step": 2000 }, { "epoch": 0.07073565076798706, "grad_norm": 1.8428642749786377, "learning_rate": 4.2421180274858524e-05, "loss": 2.5944, "step": 2100 }, { "epoch": 0.07410401509027216, "grad_norm": 1.2476333379745483, "learning_rate": 4.444219886822959e-05, "loss": 2.5473, "step": 2200 }, { "epoch": 0.07747237941255726, "grad_norm": 1.3376599550247192, "learning_rate": 4.6463217461600646e-05, "loss": 2.5115, "step": 2300 }, { "epoch": 0.08084074373484236, "grad_norm": 2.028588056564331, "learning_rate": 4.84842360549717e-05, "loss": 2.4653, "step": 2400 }, { "epoch": 0.08420910805712746, "grad_norm": 1.5413949489593506, "learning_rate": 5.0505254648342755e-05, "loss": 2.4203, "step": 2500 }, { "epoch": 0.08757747237941256, "grad_norm": 1.3644788265228271, "learning_rate": 5.252627324171382e-05, "loss": 2.3903, "step": 2600 }, { "epoch": 0.09094583670169766, "grad_norm": 1.130286693572998, "learning_rate": 5.454729183508488e-05, "loss": 2.3527, "step": 2700 }, { "epoch": 0.09431420102398276, "grad_norm": 1.6935234069824219, "learning_rate": 5.6568310428455935e-05, "loss": 2.3229, "step": 2800 }, { "epoch": 0.09768256534626785, "grad_norm": 1.9905304908752441, "learning_rate": 5.8589329021827e-05, "loss": 2.2926, "step": 2900 }, { "epoch": 0.10105092966855295, "grad_norm": 1.4488565921783447, "learning_rate": 6.061034761519806e-05, "loss": 2.2595, "step": 3000 }, { "epoch": 0.10441929399083805, "grad_norm": 1.6210366487503052, "learning_rate": 6.263136620856912e-05, "loss": 2.2317, "step": 3100 }, { "epoch": 0.10778765831312315, "grad_norm": 1.6805219650268555, "learning_rate": 6.465238480194017e-05, "loss": 2.1987, "step": 3200 }, { "epoch": 0.11115602263540825, "grad_norm": 1.6502385139465332, "learning_rate": 6.667340339531123e-05, "loss": 2.181, "step": 3300 }, { "epoch": 0.11452438695769335, "grad_norm": 1.831292986869812, "learning_rate": 6.869442198868228e-05, "loss": 2.1532, "step": 3400 }, { "epoch": 0.11789275127997845, "grad_norm": 1.9188601970672607, "learning_rate": 7.071544058205335e-05, "loss": 2.1127, "step": 3500 }, { "epoch": 0.12126111560226355, "grad_norm": 1.6280624866485596, "learning_rate": 7.273645917542441e-05, "loss": 2.0921, "step": 3600 }, { "epoch": 0.12462947992454863, "grad_norm": 1.5468984842300415, "learning_rate": 7.475747776879546e-05, "loss": 2.0626, "step": 3700 }, { "epoch": 0.12799784424683375, "grad_norm": 1.7183716297149658, "learning_rate": 7.677849636216651e-05, "loss": 2.0302, "step": 3800 }, { "epoch": 0.13136620856911885, "grad_norm": 1.8704299926757812, "learning_rate": 7.879951495553758e-05, "loss": 1.9946, "step": 3900 }, { "epoch": 0.13473457289140395, "grad_norm": 1.719117522239685, "learning_rate": 8.082053354890864e-05, "loss": 1.955, "step": 4000 }, { "epoch": 0.13810293721368902, "grad_norm": 1.8330260515213013, "learning_rate": 8.28415521422797e-05, "loss": 1.9137, "step": 4100 }, { "epoch": 0.14147130153597412, "grad_norm": 2.341217517852783, "learning_rate": 8.486257073565076e-05, "loss": 1.8765, "step": 4200 }, { "epoch": 0.14483966585825922, "grad_norm": 1.792738914489746, "learning_rate": 8.688358932902182e-05, "loss": 1.837, "step": 4300 }, { "epoch": 0.14820803018054432, "grad_norm": 2.237147092819214, "learning_rate": 8.890460792239287e-05, "loss": 1.7995, "step": 4400 }, { "epoch": 0.15157639450282942, "grad_norm": 1.8367396593093872, "learning_rate": 9.092562651576394e-05, "loss": 1.7706, "step": 4500 }, { "epoch": 0.15494475882511452, "grad_norm": 2.0587222576141357, "learning_rate": 9.2946645109135e-05, "loss": 1.7431, "step": 4600 }, { "epoch": 0.15831312314739962, "grad_norm": 1.8988635540008545, "learning_rate": 9.496766370250605e-05, "loss": 1.7179, "step": 4700 }, { "epoch": 0.16168148746968472, "grad_norm": 2.192547559738159, "learning_rate": 9.698868229587712e-05, "loss": 1.6932, "step": 4800 }, { "epoch": 0.16504985179196982, "grad_norm": 1.6913732290267944, "learning_rate": 9.900970088924818e-05, "loss": 1.6705, "step": 4900 }, { "epoch": 0.16841821611425492, "grad_norm": 1.705934762954712, "learning_rate": 0.00010103071948261923, "loss": 1.6509, "step": 5000 }, { "epoch": 0.17178658043654002, "grad_norm": 1.9764398336410522, "learning_rate": 0.0001030517380759903, "loss": 1.6212, "step": 5100 }, { "epoch": 0.17515494475882512, "grad_norm": 2.109279155731201, "learning_rate": 0.00010507275666936134, "loss": 1.6087, "step": 5200 }, { "epoch": 0.17852330908111022, "grad_norm": 2.036076784133911, "learning_rate": 0.0001070937752627324, "loss": 1.5878, "step": 5300 }, { "epoch": 0.18189167340339532, "grad_norm": 2.1286652088165283, "learning_rate": 0.00010911479385610347, "loss": 1.5641, "step": 5400 }, { "epoch": 0.18526003772568042, "grad_norm": 2.04007625579834, "learning_rate": 0.00011113581244947452, "loss": 1.5462, "step": 5500 }, { "epoch": 0.18862840204796552, "grad_norm": 1.9469410181045532, "learning_rate": 0.00011315683104284558, "loss": 1.528, "step": 5600 }, { "epoch": 0.1919967663702506, "grad_norm": 1.9078123569488525, "learning_rate": 0.00011517784963621665, "loss": 1.5101, "step": 5700 }, { "epoch": 0.1953651306925357, "grad_norm": 2.1443777084350586, "learning_rate": 0.0001171988682295877, "loss": 1.4885, "step": 5800 }, { "epoch": 0.1987334950148208, "grad_norm": 1.8993617296218872, "learning_rate": 0.00011921988682295876, "loss": 1.478, "step": 5900 }, { "epoch": 0.2021018593371059, "grad_norm": 1.7812656164169312, "learning_rate": 0.00012124090541632983, "loss": 1.4621, "step": 6000 }, { "epoch": 0.205470223659391, "grad_norm": 1.7858940362930298, "learning_rate": 0.0001232619240097009, "loss": 1.4494, "step": 6100 }, { "epoch": 0.2088385879816761, "grad_norm": 1.9336419105529785, "learning_rate": 0.00012528294260307193, "loss": 1.4393, "step": 6200 }, { "epoch": 0.2122069523039612, "grad_norm": 1.85440993309021, "learning_rate": 0.000127303961196443, "loss": 1.43, "step": 6300 }, { "epoch": 0.2155753166262463, "grad_norm": 1.583737850189209, "learning_rate": 0.00012932497978981406, "loss": 1.4198, "step": 6400 }, { "epoch": 0.2189436809485314, "grad_norm": 2.076510429382324, "learning_rate": 0.00013134599838318512, "loss": 1.4034, "step": 6500 }, { "epoch": 0.2223120452708165, "grad_norm": 1.6790341138839722, "learning_rate": 0.0001333670169765562, "loss": 1.3937, "step": 6600 }, { "epoch": 0.2256804095931016, "grad_norm": 1.9147748947143555, "learning_rate": 0.00013538803556992725, "loss": 1.3818, "step": 6700 }, { "epoch": 0.2290487739153867, "grad_norm": 1.6485368013381958, "learning_rate": 0.0001374090541632983, "loss": 1.3758, "step": 6800 }, { "epoch": 0.2324171382376718, "grad_norm": 2.087151527404785, "learning_rate": 0.00013943007275666935, "loss": 1.3696, "step": 6900 }, { "epoch": 0.2357855025599569, "grad_norm": 1.7565312385559082, "learning_rate": 0.0001414510913500404, "loss": 1.3579, "step": 7000 }, { "epoch": 0.239153866882242, "grad_norm": 1.83383047580719, "learning_rate": 0.00014347210994341146, "loss": 1.3532, "step": 7100 }, { "epoch": 0.2425222312045271, "grad_norm": 1.8977510929107666, "learning_rate": 0.00014549312853678252, "loss": 1.3465, "step": 7200 }, { "epoch": 0.24589059552681217, "grad_norm": 2.1144802570343018, "learning_rate": 0.00014751414713015358, "loss": 1.3389, "step": 7300 }, { "epoch": 0.24925895984909727, "grad_norm": 1.6824164390563965, "learning_rate": 0.00014953516572352465, "loss": 1.3283, "step": 7400 }, { "epoch": 0.2526273241713824, "grad_norm": 1.8628549575805664, "learning_rate": 0.00015155618431689571, "loss": 1.3225, "step": 7500 }, { "epoch": 0.2559956884936675, "grad_norm": 1.651308298110962, "learning_rate": 0.00015357720291026675, "loss": 1.3153, "step": 7600 }, { "epoch": 0.2593640528159526, "grad_norm": 1.6041486263275146, "learning_rate": 0.00015559822150363782, "loss": 1.3056, "step": 7700 }, { "epoch": 0.2627324171382377, "grad_norm": 1.9386959075927734, "learning_rate": 0.00015761924009700888, "loss": 1.3046, "step": 7800 }, { "epoch": 0.2661007814605228, "grad_norm": 1.8345019817352295, "learning_rate": 0.00015964025869037995, "loss": 1.2884, "step": 7900 }, { "epoch": 0.2694691457828079, "grad_norm": 1.4670854806900024, "learning_rate": 0.000161661277283751, "loss": 1.286, "step": 8000 }, { "epoch": 0.27283751010509294, "grad_norm": 1.8858684301376343, "learning_rate": 0.00016368229587712205, "loss": 1.2774, "step": 8100 }, { "epoch": 0.27620587442737804, "grad_norm": 1.8875221014022827, "learning_rate": 0.0001657033144704931, "loss": 1.2736, "step": 8200 }, { "epoch": 0.27957423874966314, "grad_norm": 1.7752630710601807, "learning_rate": 0.00016772433306386418, "loss": 1.2746, "step": 8300 }, { "epoch": 0.28294260307194824, "grad_norm": 1.5532513856887817, "learning_rate": 0.00016974535165723524, "loss": 1.267, "step": 8400 }, { "epoch": 0.28631096739423334, "grad_norm": 1.6465749740600586, "learning_rate": 0.0001717663702506063, "loss": 1.2599, "step": 8500 }, { "epoch": 0.28967933171651844, "grad_norm": 1.5771738290786743, "learning_rate": 0.00017378738884397737, "loss": 1.2557, "step": 8600 }, { "epoch": 0.29304769603880354, "grad_norm": 1.4705991744995117, "learning_rate": 0.0001758084074373484, "loss": 1.252, "step": 8700 }, { "epoch": 0.29641606036108864, "grad_norm": 1.489914059638977, "learning_rate": 0.00017782942603071947, "loss": 1.2425, "step": 8800 }, { "epoch": 0.29978442468337374, "grad_norm": 1.5901821851730347, "learning_rate": 0.00017985044462409054, "loss": 1.2341, "step": 8900 }, { "epoch": 0.30315278900565884, "grad_norm": 1.5143710374832153, "learning_rate": 0.0001818714632174616, "loss": 1.2309, "step": 9000 }, { "epoch": 0.30652115332794394, "grad_norm": 1.5409547090530396, "learning_rate": 0.00018389248181083267, "loss": 1.2258, "step": 9100 }, { "epoch": 0.30988951765022904, "grad_norm": 1.4094816446304321, "learning_rate": 0.00018591350040420368, "loss": 1.2197, "step": 9200 }, { "epoch": 0.31325788197251414, "grad_norm": 1.6748660802841187, "learning_rate": 0.00018793451899757474, "loss": 1.2176, "step": 9300 }, { "epoch": 0.31662624629479924, "grad_norm": 1.587318778038025, "learning_rate": 0.0001899555375909458, "loss": 1.2091, "step": 9400 }, { "epoch": 0.31999461061708434, "grad_norm": 1.5422818660736084, "learning_rate": 0.00019197655618431687, "loss": 1.2079, "step": 9500 }, { "epoch": 0.32336297493936944, "grad_norm": 1.30134117603302, "learning_rate": 0.00019399757477768793, "loss": 1.2017, "step": 9600 }, { "epoch": 0.32673133926165454, "grad_norm": 1.368249773979187, "learning_rate": 0.00019601859337105897, "loss": 1.2026, "step": 9700 }, { "epoch": 0.33009970358393964, "grad_norm": 1.5968406200408936, "learning_rate": 0.00019803961196443004, "loss": 1.1902, "step": 9800 }, { "epoch": 0.33346806790622474, "grad_norm": 1.435455083847046, "learning_rate": 0.0002000606305578011, "loss": 1.1894, "step": 9900 }, { "epoch": 0.33683643222850984, "grad_norm": 1.4132752418518066, "learning_rate": 0.00020208164915117217, "loss": 1.1841, "step": 10000 }, { "epoch": 0.34020479655079494, "grad_norm": 1.5038225650787354, "learning_rate": 0.00020410266774454323, "loss": 1.1859, "step": 10100 }, { "epoch": 0.34357316087308004, "grad_norm": 1.4201886653900146, "learning_rate": 0.0002061236863379143, "loss": 1.1743, "step": 10200 }, { "epoch": 0.34694152519536514, "grad_norm": 1.3544988632202148, "learning_rate": 0.00020814470493128533, "loss": 1.1724, "step": 10300 }, { "epoch": 0.35030988951765024, "grad_norm": 1.6340460777282715, "learning_rate": 0.0002101657235246564, "loss": 1.1679, "step": 10400 }, { "epoch": 0.35367825383993534, "grad_norm": 1.3137534856796265, "learning_rate": 0.00021218674211802746, "loss": 1.1602, "step": 10500 }, { "epoch": 0.35704661816222044, "grad_norm": 1.3838586807250977, "learning_rate": 0.00021420776071139853, "loss": 1.1602, "step": 10600 }, { "epoch": 0.36041498248450554, "grad_norm": 1.46292245388031, "learning_rate": 0.0002162287793047696, "loss": 1.154, "step": 10700 }, { "epoch": 0.36378334680679064, "grad_norm": 1.2189207077026367, "learning_rate": 0.00021824979789814066, "loss": 1.1518, "step": 10800 }, { "epoch": 0.36715171112907574, "grad_norm": 1.3567001819610596, "learning_rate": 0.0002202708164915117, "loss": 1.1437, "step": 10900 }, { "epoch": 0.37052007545136084, "grad_norm": 1.3379132747650146, "learning_rate": 0.00022229183508488276, "loss": 1.1418, "step": 11000 }, { "epoch": 0.37388843977364594, "grad_norm": 1.2323216199874878, "learning_rate": 0.00022431285367825382, "loss": 1.1382, "step": 11100 }, { "epoch": 0.37725680409593104, "grad_norm": 1.9002209901809692, "learning_rate": 0.0002263338722716249, "loss": 1.1323, "step": 11200 }, { "epoch": 0.3806251684182161, "grad_norm": 1.3124207258224487, "learning_rate": 0.00022835489086499595, "loss": 1.1336, "step": 11300 }, { "epoch": 0.3839935327405012, "grad_norm": 1.3445236682891846, "learning_rate": 0.000230375909458367, "loss": 1.1259, "step": 11400 }, { "epoch": 0.3873618970627863, "grad_norm": 1.2994790077209473, "learning_rate": 0.00023239692805173805, "loss": 1.1298, "step": 11500 }, { "epoch": 0.3907302613850714, "grad_norm": 1.2289458513259888, "learning_rate": 0.00023441794664510912, "loss": 1.1226, "step": 11600 }, { "epoch": 0.3940986257073565, "grad_norm": 1.1772109270095825, "learning_rate": 0.00023643896523848018, "loss": 1.1215, "step": 11700 }, { "epoch": 0.3974669900296416, "grad_norm": 1.3592746257781982, "learning_rate": 0.00023845998383185125, "loss": 1.1179, "step": 11800 }, { "epoch": 0.4008353543519267, "grad_norm": 1.1571407318115234, "learning_rate": 0.0002404810024252223, "loss": 1.1124, "step": 11900 }, { "epoch": 0.4042037186742118, "grad_norm": 1.3592592477798462, "learning_rate": 0.00024250202101859335, "loss": 1.1083, "step": 12000 }, { "epoch": 0.4075720829964969, "grad_norm": 1.2856664657592773, "learning_rate": 0.00024452303961196444, "loss": 1.1083, "step": 12100 }, { "epoch": 0.410940447318782, "grad_norm": 1.2512900829315186, "learning_rate": 0.0002465440582053355, "loss": 1.0984, "step": 12200 }, { "epoch": 0.4143088116410671, "grad_norm": 1.239823818206787, "learning_rate": 0.0002485650767987065, "loss": 1.0984, "step": 12300 }, { "epoch": 0.4176771759633522, "grad_norm": 1.4879858493804932, "learning_rate": 0.0002505860953920776, "loss": 1.0962, "step": 12400 }, { "epoch": 0.4210455402856373, "grad_norm": 1.1708803176879883, "learning_rate": 0.00025260711398544865, "loss": 1.0914, "step": 12500 }, { "epoch": 0.4244139046079224, "grad_norm": 1.092463731765747, "learning_rate": 0.0002546281325788197, "loss": 1.089, "step": 12600 }, { "epoch": 0.4277822689302075, "grad_norm": 1.2004188299179077, "learning_rate": 0.0002566491511721908, "loss": 1.0884, "step": 12700 }, { "epoch": 0.4311506332524926, "grad_norm": 1.1339149475097656, "learning_rate": 0.0002586701697655618, "loss": 1.0857, "step": 12800 }, { "epoch": 0.4345189975747777, "grad_norm": 1.3253908157348633, "learning_rate": 0.00026069118835893285, "loss": 1.0821, "step": 12900 }, { "epoch": 0.4378873618970628, "grad_norm": 1.2655895948410034, "learning_rate": 0.00026271220695230394, "loss": 1.0775, "step": 13000 }, { "epoch": 0.4412557262193479, "grad_norm": 1.4968757629394531, "learning_rate": 0.000264733225545675, "loss": 1.0764, "step": 13100 }, { "epoch": 0.444624090541633, "grad_norm": 1.200173020362854, "learning_rate": 0.00026675424413904607, "loss": 1.0768, "step": 13200 }, { "epoch": 0.4479924548639181, "grad_norm": 1.3085741996765137, "learning_rate": 0.0002687752627324171, "loss": 1.0739, "step": 13300 }, { "epoch": 0.4513608191862032, "grad_norm": 1.181569218635559, "learning_rate": 0.00027079628132578815, "loss": 1.0685, "step": 13400 }, { "epoch": 0.4547291835084883, "grad_norm": 1.2471662759780884, "learning_rate": 0.00027281729991915924, "loss": 1.0668, "step": 13500 }, { "epoch": 0.4580975478307734, "grad_norm": 1.0714460611343384, "learning_rate": 0.0002748383185125303, "loss": 1.0687, "step": 13600 }, { "epoch": 0.4614659121530585, "grad_norm": 1.2705806493759155, "learning_rate": 0.00027685933710590137, "loss": 1.0602, "step": 13700 }, { "epoch": 0.4648342764753436, "grad_norm": 1.199216365814209, "learning_rate": 0.0002788803556992724, "loss": 1.0572, "step": 13800 }, { "epoch": 0.4682026407976287, "grad_norm": 1.1781370639801025, "learning_rate": 0.00028090137429264344, "loss": 1.056, "step": 13900 }, { "epoch": 0.4715710051199138, "grad_norm": 1.1787018775939941, "learning_rate": 0.00028292239288601453, "loss": 1.0582, "step": 14000 }, { "epoch": 0.4749393694421989, "grad_norm": 1.0408787727355957, "learning_rate": 0.00028494341147938557, "loss": 1.0509, "step": 14100 }, { "epoch": 0.478307733764484, "grad_norm": 1.2054550647735596, "learning_rate": 0.00028696443007275666, "loss": 1.0496, "step": 14200 }, { "epoch": 0.4816760980867691, "grad_norm": 1.059328317642212, "learning_rate": 0.0002889854486661277, "loss": 1.0484, "step": 14300 }, { "epoch": 0.4850444624090542, "grad_norm": 1.0218919515609741, "learning_rate": 0.0002910064672594988, "loss": 1.0443, "step": 14400 }, { "epoch": 0.4884128267313393, "grad_norm": 1.569550633430481, "learning_rate": 0.00029302748585286983, "loss": 1.0439, "step": 14500 }, { "epoch": 0.49178119105362433, "grad_norm": 1.2660326957702637, "learning_rate": 0.00029504850444624087, "loss": 1.0412, "step": 14600 }, { "epoch": 0.49514955537590943, "grad_norm": 0.9795782566070557, "learning_rate": 0.00029706952303961196, "loss": 1.0362, "step": 14700 }, { "epoch": 0.49851791969819453, "grad_norm": 1.0511739253997803, "learning_rate": 0.000299090541632983, "loss": 1.0379, "step": 14800 }, { "epoch": 0.5018862840204796, "grad_norm": 1.077668309211731, "learning_rate": 0.00029987649330818285, "loss": 1.034, "step": 14900 }, { "epoch": 0.5052546483427648, "grad_norm": 0.9709302186965942, "learning_rate": 0.0002996519356866972, "loss": 1.0297, "step": 15000 }, { "epoch": 0.5086230126650498, "grad_norm": 1.038855791091919, "learning_rate": 0.0002994273780652115, "loss": 1.0295, "step": 15100 }, { "epoch": 0.511991376987335, "grad_norm": 1.1095309257507324, "learning_rate": 0.00029920282044372586, "loss": 1.0241, "step": 15200 }, { "epoch": 0.51535974130962, "grad_norm": 1.0058341026306152, "learning_rate": 0.00029897826282224014, "loss": 1.0233, "step": 15300 }, { "epoch": 0.5187281056319052, "grad_norm": 1.1029912233352661, "learning_rate": 0.00029875370520075447, "loss": 1.0253, "step": 15400 }, { "epoch": 0.5220964699541902, "grad_norm": 1.1280447244644165, "learning_rate": 0.0002985291475792688, "loss": 1.02, "step": 15500 }, { "epoch": 0.5254648342764754, "grad_norm": 1.0095000267028809, "learning_rate": 0.00029830458995778314, "loss": 1.0185, "step": 15600 }, { "epoch": 0.5288331985987604, "grad_norm": 1.071540117263794, "learning_rate": 0.0002980800323362975, "loss": 1.0133, "step": 15700 }, { "epoch": 0.5322015629210456, "grad_norm": 0.9709872007369995, "learning_rate": 0.0002978554747148118, "loss": 1.0189, "step": 15800 }, { "epoch": 0.5355699272433306, "grad_norm": 1.1805214881896973, "learning_rate": 0.0002976309170933261, "loss": 1.0145, "step": 15900 }, { "epoch": 0.5389382915656158, "grad_norm": 1.1302651166915894, "learning_rate": 0.0002974063594718405, "loss": 1.0088, "step": 16000 }, { "epoch": 0.5423066558879008, "grad_norm": 1.280207872390747, "learning_rate": 0.00029718180185035476, "loss": 1.0126, "step": 16100 }, { "epoch": 0.5456750202101859, "grad_norm": 1.024566888809204, "learning_rate": 0.0002969572442288691, "loss": 1.006, "step": 16200 }, { "epoch": 0.549043384532471, "grad_norm": 1.192209243774414, "learning_rate": 0.00029673268660738343, "loss": 1.0076, "step": 16300 }, { "epoch": 0.5524117488547561, "grad_norm": 0.9792165756225586, "learning_rate": 0.0002965081289858977, "loss": 1.005, "step": 16400 }, { "epoch": 0.5557801131770412, "grad_norm": 1.6060813665390015, "learning_rate": 0.0002962835713644121, "loss": 0.9999, "step": 16500 }, { "epoch": 0.5591484774993263, "grad_norm": 0.9630849957466125, "learning_rate": 0.0002960590137429264, "loss": 1.0024, "step": 16600 }, { "epoch": 0.5625168418216114, "grad_norm": 0.9396387934684753, "learning_rate": 0.0002958344561214407, "loss": 1.0004, "step": 16700 }, { "epoch": 0.5658852061438965, "grad_norm": 0.9031047821044922, "learning_rate": 0.00029560989849995506, "loss": 0.9984, "step": 16800 }, { "epoch": 0.5692535704661816, "grad_norm": 1.0625028610229492, "learning_rate": 0.0002953853408784694, "loss": 0.9932, "step": 16900 }, { "epoch": 0.5726219347884667, "grad_norm": 0.9275569319725037, "learning_rate": 0.0002951607832569837, "loss": 0.9925, "step": 17000 }, { "epoch": 0.5759902991107518, "grad_norm": 0.8847247362136841, "learning_rate": 0.00029493622563549806, "loss": 0.9909, "step": 17100 }, { "epoch": 0.5793586634330369, "grad_norm": 0.9581294059753418, "learning_rate": 0.0002947116680140124, "loss": 0.9923, "step": 17200 }, { "epoch": 0.582727027755322, "grad_norm": 1.003164529800415, "learning_rate": 0.0002944871103925267, "loss": 0.9914, "step": 17300 }, { "epoch": 0.5860953920776071, "grad_norm": 1.010026454925537, "learning_rate": 0.000294262552771041, "loss": 0.9854, "step": 17400 }, { "epoch": 0.5894637563998922, "grad_norm": 0.9266247153282166, "learning_rate": 0.00029403799514955535, "loss": 0.9879, "step": 17500 }, { "epoch": 0.5928321207221773, "grad_norm": 0.9909249544143677, "learning_rate": 0.0002938134375280697, "loss": 0.9876, "step": 17600 }, { "epoch": 0.5962004850444624, "grad_norm": 0.8353651165962219, "learning_rate": 0.000293588879906584, "loss": 0.9846, "step": 17700 }, { "epoch": 0.5995688493667475, "grad_norm": 0.9142294526100159, "learning_rate": 0.00029336432228509835, "loss": 0.9877, "step": 17800 }, { "epoch": 0.6029372136890326, "grad_norm": 0.9139926433563232, "learning_rate": 0.00029313976466361263, "loss": 0.9831, "step": 17900 }, { "epoch": 0.6063055780113177, "grad_norm": 0.8871977925300598, "learning_rate": 0.000292915207042127, "loss": 0.9817, "step": 18000 }, { "epoch": 0.6096739423336028, "grad_norm": 0.9932221174240112, "learning_rate": 0.0002926906494206413, "loss": 0.979, "step": 18100 }, { "epoch": 0.6130423066558879, "grad_norm": 0.9240766167640686, "learning_rate": 0.00029246609179915564, "loss": 0.9774, "step": 18200 }, { "epoch": 0.616410670978173, "grad_norm": 0.932101845741272, "learning_rate": 0.00029224153417767, "loss": 0.9797, "step": 18300 }, { "epoch": 0.6197790353004581, "grad_norm": 0.9871794581413269, "learning_rate": 0.00029201697655618426, "loss": 0.9809, "step": 18400 }, { "epoch": 0.6231473996227432, "grad_norm": 0.8771729469299316, "learning_rate": 0.00029179241893469865, "loss": 0.9761, "step": 18500 }, { "epoch": 0.6265157639450283, "grad_norm": 1.103968858718872, "learning_rate": 0.0002915678613132129, "loss": 0.9779, "step": 18600 }, { "epoch": 0.6298841282673134, "grad_norm": 1.0848268270492554, "learning_rate": 0.00029134330369172726, "loss": 0.9722, "step": 18700 }, { "epoch": 0.6332524925895985, "grad_norm": 0.8907010555267334, "learning_rate": 0.0002911187460702416, "loss": 0.9709, "step": 18800 }, { "epoch": 0.6366208569118836, "grad_norm": 0.9081377983093262, "learning_rate": 0.00029089418844875593, "loss": 0.9681, "step": 18900 }, { "epoch": 0.6399892212341687, "grad_norm": 0.873905599117279, "learning_rate": 0.00029066963082727027, "loss": 0.9735, "step": 19000 }, { "epoch": 0.6433575855564538, "grad_norm": 0.9111950397491455, "learning_rate": 0.0002904450732057846, "loss": 0.97, "step": 19100 }, { "epoch": 0.6467259498787389, "grad_norm": 0.9769060611724854, "learning_rate": 0.0002902205155842989, "loss": 0.9688, "step": 19200 }, { "epoch": 0.650094314201024, "grad_norm": 1.0822559595108032, "learning_rate": 0.0002899959579628132, "loss": 0.968, "step": 19300 }, { "epoch": 0.6534626785233091, "grad_norm": 0.8573871850967407, "learning_rate": 0.00028977140034132755, "loss": 0.967, "step": 19400 }, { "epoch": 0.6568310428455941, "grad_norm": 0.989267885684967, "learning_rate": 0.0002895468427198419, "loss": 0.9652, "step": 19500 }, { "epoch": 0.6601994071678793, "grad_norm": 0.9819543361663818, "learning_rate": 0.0002893222850983562, "loss": 0.9597, "step": 19600 }, { "epoch": 0.6635677714901643, "grad_norm": 0.9204864501953125, "learning_rate": 0.00028909772747687056, "loss": 0.9641, "step": 19700 }, { "epoch": 0.6669361358124495, "grad_norm": 1.1161561012268066, "learning_rate": 0.0002888731698553849, "loss": 0.9619, "step": 19800 }, { "epoch": 0.6703045001347345, "grad_norm": 0.8925914764404297, "learning_rate": 0.0002886486122338992, "loss": 0.9599, "step": 19900 }, { "epoch": 0.6736728644570197, "grad_norm": 0.9228368401527405, "learning_rate": 0.0002884240546124135, "loss": 0.9603, "step": 20000 }, { "epoch": 0.6770412287793047, "grad_norm": 0.8357170224189758, "learning_rate": 0.00028819949699092785, "loss": 0.9562, "step": 20100 }, { "epoch": 0.6804095931015899, "grad_norm": 0.9358044266700745, "learning_rate": 0.0002879749393694422, "loss": 0.963, "step": 20200 }, { "epoch": 0.6837779574238749, "grad_norm": 0.83344966173172, "learning_rate": 0.0002877503817479565, "loss": 0.9555, "step": 20300 }, { "epoch": 0.6871463217461601, "grad_norm": 0.8138599395751953, "learning_rate": 0.0002875258241264708, "loss": 0.9572, "step": 20400 }, { "epoch": 0.6905146860684451, "grad_norm": 0.886358380317688, "learning_rate": 0.0002873012665049852, "loss": 0.9525, "step": 20500 }, { "epoch": 0.6938830503907303, "grad_norm": 0.9100881814956665, "learning_rate": 0.00028707670888349947, "loss": 0.9545, "step": 20600 }, { "epoch": 0.6972514147130153, "grad_norm": 0.8902551531791687, "learning_rate": 0.0002868521512620138, "loss": 0.9542, "step": 20700 }, { "epoch": 0.7006197790353005, "grad_norm": 1.24783456325531, "learning_rate": 0.00028662759364052814, "loss": 0.9536, "step": 20800 }, { "epoch": 0.7039881433575855, "grad_norm": 0.831633448600769, "learning_rate": 0.0002864030360190425, "loss": 0.9526, "step": 20900 }, { "epoch": 0.7073565076798707, "grad_norm": 0.8300578594207764, "learning_rate": 0.0002861784783975568, "loss": 0.9478, "step": 21000 }, { "epoch": 0.7107248720021557, "grad_norm": 0.9245336055755615, "learning_rate": 0.0002859539207760711, "loss": 0.9531, "step": 21100 }, { "epoch": 0.7140932363244409, "grad_norm": 0.8717476725578308, "learning_rate": 0.0002857293631545854, "loss": 0.9487, "step": 21200 }, { "epoch": 0.7174616006467259, "grad_norm": 0.8705008625984192, "learning_rate": 0.00028550480553309976, "loss": 0.9486, "step": 21300 }, { "epoch": 0.7208299649690111, "grad_norm": 0.8324209451675415, "learning_rate": 0.0002852802479116141, "loss": 0.9506, "step": 21400 }, { "epoch": 0.7241983292912961, "grad_norm": 0.9107707738876343, "learning_rate": 0.00028505569029012843, "loss": 0.9465, "step": 21500 }, { "epoch": 0.7275666936135813, "grad_norm": 0.8204140067100525, "learning_rate": 0.00028483113266864277, "loss": 0.9462, "step": 21600 }, { "epoch": 0.7309350579358663, "grad_norm": 0.8158605098724365, "learning_rate": 0.00028460657504715705, "loss": 0.9483, "step": 21700 }, { "epoch": 0.7343034222581515, "grad_norm": 0.8872929811477661, "learning_rate": 0.00028438201742567144, "loss": 0.9461, "step": 21800 }, { "epoch": 0.7376717865804365, "grad_norm": 0.8722573518753052, "learning_rate": 0.0002841574598041857, "loss": 0.943, "step": 21900 }, { "epoch": 0.7410401509027217, "grad_norm": 0.8522630929946899, "learning_rate": 0.00028393290218270005, "loss": 0.9449, "step": 22000 }, { "epoch": 0.7444085152250067, "grad_norm": 0.850837767124176, "learning_rate": 0.0002837083445612144, "loss": 0.9441, "step": 22100 }, { "epoch": 0.7477768795472919, "grad_norm": 0.7917930483818054, "learning_rate": 0.0002834837869397287, "loss": 0.9442, "step": 22200 }, { "epoch": 0.7511452438695769, "grad_norm": 0.9168843030929565, "learning_rate": 0.00028325922931824306, "loss": 0.9427, "step": 22300 }, { "epoch": 0.7545136081918621, "grad_norm": 0.9117637276649475, "learning_rate": 0.00028303467169675734, "loss": 0.9436, "step": 22400 }, { "epoch": 0.7578819725141471, "grad_norm": 0.8345003724098206, "learning_rate": 0.0002828101140752717, "loss": 0.9394, "step": 22500 }, { "epoch": 0.7612503368364322, "grad_norm": 0.8290796875953674, "learning_rate": 0.000282585556453786, "loss": 0.938, "step": 22600 }, { "epoch": 0.7646187011587173, "grad_norm": 0.7990386486053467, "learning_rate": 0.00028236099883230034, "loss": 0.9393, "step": 22700 }, { "epoch": 0.7679870654810024, "grad_norm": 0.7871207594871521, "learning_rate": 0.0002821364412108147, "loss": 0.9376, "step": 22800 }, { "epoch": 0.7713554298032875, "grad_norm": 0.7800641059875488, "learning_rate": 0.000281911883589329, "loss": 0.9366, "step": 22900 }, { "epoch": 0.7747237941255726, "grad_norm": 0.8521484136581421, "learning_rate": 0.00028168732596784335, "loss": 0.9377, "step": 23000 }, { "epoch": 0.7780921584478577, "grad_norm": 0.8726188540458679, "learning_rate": 0.00028146276834635763, "loss": 0.9344, "step": 23100 }, { "epoch": 0.7814605227701428, "grad_norm": 0.9499660730361938, "learning_rate": 0.00028123821072487197, "loss": 0.9357, "step": 23200 }, { "epoch": 0.7848288870924279, "grad_norm": 0.8431583642959595, "learning_rate": 0.0002810136531033863, "loss": 0.9373, "step": 23300 }, { "epoch": 0.788197251414713, "grad_norm": 0.8140375018119812, "learning_rate": 0.00028078909548190064, "loss": 0.9352, "step": 23400 }, { "epoch": 0.7915656157369981, "grad_norm": 0.8821849226951599, "learning_rate": 0.00028056453786041497, "loss": 0.9346, "step": 23500 }, { "epoch": 0.7949339800592832, "grad_norm": 1.0091123580932617, "learning_rate": 0.0002803399802389293, "loss": 0.9328, "step": 23600 }, { "epoch": 0.7983023443815683, "grad_norm": 1.0306652784347534, "learning_rate": 0.0002801154226174436, "loss": 0.9332, "step": 23700 }, { "epoch": 0.8016707087038534, "grad_norm": 0.8444597721099854, "learning_rate": 0.000279890864995958, "loss": 0.9372, "step": 23800 }, { "epoch": 0.8050390730261385, "grad_norm": 0.8706274628639221, "learning_rate": 0.00027966630737447226, "loss": 0.9351, "step": 23900 }, { "epoch": 0.8084074373484236, "grad_norm": 0.8325883150100708, "learning_rate": 0.0002794417497529866, "loss": 0.9296, "step": 24000 }, { "epoch": 0.8117758016707087, "grad_norm": 0.7604151964187622, "learning_rate": 0.00027921719213150093, "loss": 0.9292, "step": 24100 }, { "epoch": 0.8151441659929938, "grad_norm": 0.916677713394165, "learning_rate": 0.0002789926345100152, "loss": 0.9274, "step": 24200 }, { "epoch": 0.8185125303152789, "grad_norm": 0.8420447111129761, "learning_rate": 0.0002787680768885296, "loss": 0.9293, "step": 24300 }, { "epoch": 0.821880894637564, "grad_norm": 0.8805976510047913, "learning_rate": 0.0002785435192670439, "loss": 0.9243, "step": 24400 }, { "epoch": 0.8252492589598491, "grad_norm": 0.8013381361961365, "learning_rate": 0.0002783189616455582, "loss": 0.9296, "step": 24500 }, { "epoch": 0.8286176232821342, "grad_norm": 0.8348533511161804, "learning_rate": 0.00027809440402407255, "loss": 0.9256, "step": 24600 }, { "epoch": 0.8319859876044193, "grad_norm": 0.7982239723205566, "learning_rate": 0.0002778698464025869, "loss": 0.9273, "step": 24700 }, { "epoch": 0.8353543519267044, "grad_norm": 0.9021079540252686, "learning_rate": 0.0002776452887811012, "loss": 0.9238, "step": 24800 }, { "epoch": 0.8387227162489895, "grad_norm": 0.7782942652702332, "learning_rate": 0.0002774207311596155, "loss": 0.9248, "step": 24900 }, { "epoch": 0.8420910805712746, "grad_norm": 0.8153879046440125, "learning_rate": 0.00027719617353812984, "loss": 0.9272, "step": 25000 }, { "epoch": 0.8454594448935597, "grad_norm": 0.9402956366539001, "learning_rate": 0.00027697161591664417, "loss": 0.9243, "step": 25100 }, { "epoch": 0.8488278092158448, "grad_norm": 0.7734837532043457, "learning_rate": 0.0002767470582951585, "loss": 0.9262, "step": 25200 }, { "epoch": 0.8521961735381299, "grad_norm": 0.7865928411483765, "learning_rate": 0.00027652250067367284, "loss": 0.9247, "step": 25300 }, { "epoch": 0.855564537860415, "grad_norm": 0.8157060742378235, "learning_rate": 0.0002762979430521872, "loss": 0.9235, "step": 25400 }, { "epoch": 0.8589329021827001, "grad_norm": 0.9254295825958252, "learning_rate": 0.0002760733854307015, "loss": 0.9201, "step": 25500 }, { "epoch": 0.8623012665049852, "grad_norm": 0.7986881136894226, "learning_rate": 0.00027584882780921585, "loss": 0.9215, "step": 25600 }, { "epoch": 0.8656696308272703, "grad_norm": 1.0841712951660156, "learning_rate": 0.00027562427018773013, "loss": 0.9234, "step": 25700 }, { "epoch": 0.8690379951495554, "grad_norm": 0.8564463257789612, "learning_rate": 0.00027539971256624446, "loss": 0.9238, "step": 25800 }, { "epoch": 0.8724063594718404, "grad_norm": 0.7546072602272034, "learning_rate": 0.0002751751549447588, "loss": 0.9195, "step": 25900 }, { "epoch": 0.8757747237941256, "grad_norm": 1.1822130680084229, "learning_rate": 0.00027495059732327313, "loss": 0.9204, "step": 26000 }, { "epoch": 0.8791430881164106, "grad_norm": 0.7953358292579651, "learning_rate": 0.00027472603970178747, "loss": 0.9234, "step": 26100 }, { "epoch": 0.8825114524386958, "grad_norm": 0.7510402798652649, "learning_rate": 0.00027450148208030175, "loss": 0.9193, "step": 26200 }, { "epoch": 0.8858798167609808, "grad_norm": 1.1970155239105225, "learning_rate": 0.00027427692445881614, "loss": 0.9165, "step": 26300 }, { "epoch": 0.889248181083266, "grad_norm": 1.0077903270721436, "learning_rate": 0.0002740523668373304, "loss": 0.9136, "step": 26400 }, { "epoch": 0.892616545405551, "grad_norm": 0.7909373044967651, "learning_rate": 0.00027382780921584476, "loss": 0.9162, "step": 26500 }, { "epoch": 0.8959849097278362, "grad_norm": 0.7484354376792908, "learning_rate": 0.0002736032515943591, "loss": 0.9151, "step": 26600 }, { "epoch": 0.8993532740501212, "grad_norm": 0.7733712792396545, "learning_rate": 0.0002733786939728734, "loss": 0.9171, "step": 26700 }, { "epoch": 0.9027216383724064, "grad_norm": 1.0612813234329224, "learning_rate": 0.00027315413635138776, "loss": 0.9202, "step": 26800 }, { "epoch": 0.9060900026946914, "grad_norm": 0.8826086521148682, "learning_rate": 0.00027292957872990204, "loss": 0.9184, "step": 26900 }, { "epoch": 0.9094583670169766, "grad_norm": 0.7911275029182434, "learning_rate": 0.0002727050211084164, "loss": 0.9144, "step": 27000 }, { "epoch": 0.9128267313392616, "grad_norm": 0.7702584862709045, "learning_rate": 0.0002724804634869307, "loss": 0.9086, "step": 27100 }, { "epoch": 0.9161950956615468, "grad_norm": 0.7692966461181641, "learning_rate": 0.00027225590586544505, "loss": 0.9127, "step": 27200 }, { "epoch": 0.9195634599838318, "grad_norm": 0.7272462844848633, "learning_rate": 0.0002720313482439594, "loss": 0.9129, "step": 27300 }, { "epoch": 0.922931824306117, "grad_norm": 0.7643866539001465, "learning_rate": 0.0002718067906224737, "loss": 0.9127, "step": 27400 }, { "epoch": 0.926300188628402, "grad_norm": 0.8528723120689392, "learning_rate": 0.000271582233000988, "loss": 0.9118, "step": 27500 }, { "epoch": 0.9296685529506872, "grad_norm": 0.9294866323471069, "learning_rate": 0.0002713576753795024, "loss": 0.913, "step": 27600 }, { "epoch": 0.9330369172729722, "grad_norm": 0.7819476127624512, "learning_rate": 0.00027113311775801667, "loss": 0.9094, "step": 27700 }, { "epoch": 0.9364052815952574, "grad_norm": 0.9195557236671448, "learning_rate": 0.000270908560136531, "loss": 0.9102, "step": 27800 }, { "epoch": 0.9397736459175424, "grad_norm": 0.8723948001861572, "learning_rate": 0.00027068400251504534, "loss": 0.9099, "step": 27900 }, { "epoch": 0.9431420102398276, "grad_norm": 0.7643413543701172, "learning_rate": 0.0002704594448935597, "loss": 0.9101, "step": 28000 }, { "epoch": 0.9465103745621126, "grad_norm": 0.8065772652626038, "learning_rate": 0.000270234887272074, "loss": 0.9099, "step": 28100 }, { "epoch": 0.9498787388843978, "grad_norm": 0.7567630410194397, "learning_rate": 0.0002700103296505883, "loss": 0.9092, "step": 28200 }, { "epoch": 0.9532471032066828, "grad_norm": 0.8164196014404297, "learning_rate": 0.0002697857720291026, "loss": 0.9115, "step": 28300 }, { "epoch": 0.956615467528968, "grad_norm": 0.7787773013114929, "learning_rate": 0.00026956121440761696, "loss": 0.9062, "step": 28400 }, { "epoch": 0.959983831851253, "grad_norm": 0.724734902381897, "learning_rate": 0.0002693366567861313, "loss": 0.9078, "step": 28500 }, { "epoch": 0.9633521961735382, "grad_norm": 0.813438355922699, "learning_rate": 0.00026911209916464563, "loss": 0.9075, "step": 28600 }, { "epoch": 0.9667205604958232, "grad_norm": 0.9264329671859741, "learning_rate": 0.00026888754154315997, "loss": 0.9052, "step": 28700 }, { "epoch": 0.9700889248181084, "grad_norm": 0.8521473407745361, "learning_rate": 0.0002686629839216743, "loss": 0.9034, "step": 28800 }, { "epoch": 0.9734572891403934, "grad_norm": 0.7979128360748291, "learning_rate": 0.0002684384263001886, "loss": 0.9052, "step": 28900 }, { "epoch": 0.9768256534626786, "grad_norm": 0.7806681394577026, "learning_rate": 0.0002682138686787029, "loss": 0.9039, "step": 29000 }, { "epoch": 0.9801940177849636, "grad_norm": 0.7266373038291931, "learning_rate": 0.00026798931105721725, "loss": 0.9072, "step": 29100 }, { "epoch": 0.9835623821072487, "grad_norm": 0.8608720302581787, "learning_rate": 0.0002677647534357316, "loss": 0.9062, "step": 29200 }, { "epoch": 0.9869307464295338, "grad_norm": 0.7421156764030457, "learning_rate": 0.0002675401958142459, "loss": 0.9059, "step": 29300 }, { "epoch": 0.9902991107518189, "grad_norm": 0.7319119572639465, "learning_rate": 0.00026731563819276026, "loss": 0.9074, "step": 29400 }, { "epoch": 0.993667475074104, "grad_norm": 1.1928077936172485, "learning_rate": 0.00026709108057127454, "loss": 0.9024, "step": 29500 }, { "epoch": 0.9970358393963891, "grad_norm": 0.7490862011909485, "learning_rate": 0.00026686652294978893, "loss": 0.904, "step": 29600 }, { "epoch": 1.0, "eval_loss": 0.840101957321167, "eval_runtime": 1.8082, "eval_samples_per_second": 2765.256, "eval_steps_per_second": 43.691, "step": 29688 }, { "epoch": 1.0004042037186742, "grad_norm": 0.7771435379981995, "learning_rate": 0.0002666419653283032, "loss": 0.9014, "step": 29700 }, { "epoch": 1.0037725680409593, "grad_norm": 0.7980936765670776, "learning_rate": 0.00026641740770681755, "loss": 0.9045, "step": 29800 }, { "epoch": 1.0071409323632443, "grad_norm": 0.7477487921714783, "learning_rate": 0.0002661928500853319, "loss": 0.904, "step": 29900 }, { "epoch": 1.0105092966855296, "grad_norm": 0.7866721153259277, "learning_rate": 0.00026596829246384616, "loss": 0.905, "step": 30000 }, { "epoch": 1.0138776610078146, "grad_norm": 0.8210276365280151, "learning_rate": 0.00026574373484236055, "loss": 0.9023, "step": 30100 }, { "epoch": 1.0172460253300997, "grad_norm": 0.683873176574707, "learning_rate": 0.00026551917722087483, "loss": 0.899, "step": 30200 }, { "epoch": 1.0206143896523847, "grad_norm": 0.772179901599884, "learning_rate": 0.00026529461959938917, "loss": 0.8972, "step": 30300 }, { "epoch": 1.02398275397467, "grad_norm": 0.9590583443641663, "learning_rate": 0.0002650700619779035, "loss": 0.9048, "step": 30400 }, { "epoch": 1.027351118296955, "grad_norm": 0.8030016422271729, "learning_rate": 0.00026484550435641784, "loss": 0.9007, "step": 30500 }, { "epoch": 1.03071948261924, "grad_norm": 0.7494839429855347, "learning_rate": 0.0002646209467349322, "loss": 0.9002, "step": 30600 }, { "epoch": 1.0340878469415251, "grad_norm": 0.7339987754821777, "learning_rate": 0.00026439638911344645, "loss": 0.8995, "step": 30700 }, { "epoch": 1.0374562112638104, "grad_norm": 0.7874321937561035, "learning_rate": 0.0002641718314919608, "loss": 0.8982, "step": 30800 }, { "epoch": 1.0408245755860954, "grad_norm": 0.6752808690071106, "learning_rate": 0.0002639472738704751, "loss": 0.8986, "step": 30900 }, { "epoch": 1.0441929399083805, "grad_norm": 0.758162796497345, "learning_rate": 0.00026372271624898946, "loss": 0.8963, "step": 31000 }, { "epoch": 1.0475613042306655, "grad_norm": 1.2287030220031738, "learning_rate": 0.0002634981586275038, "loss": 0.8999, "step": 31100 }, { "epoch": 1.0509296685529508, "grad_norm": 0.7714762687683105, "learning_rate": 0.00026327360100601813, "loss": 0.8952, "step": 31200 }, { "epoch": 1.0542980328752358, "grad_norm": 0.7721061706542969, "learning_rate": 0.00026304904338453247, "loss": 0.895, "step": 31300 }, { "epoch": 1.0576663971975209, "grad_norm": 0.7775672078132629, "learning_rate": 0.0002628244857630468, "loss": 0.8973, "step": 31400 }, { "epoch": 1.061034761519806, "grad_norm": 0.8213030695915222, "learning_rate": 0.0002625999281415611, "loss": 0.8958, "step": 31500 }, { "epoch": 1.0644031258420912, "grad_norm": 1.3196107149124146, "learning_rate": 0.0002623753705200754, "loss": 0.8953, "step": 31600 }, { "epoch": 1.0677714901643762, "grad_norm": 0.7078897953033447, "learning_rate": 0.00026215081289858975, "loss": 0.8981, "step": 31700 }, { "epoch": 1.0711398544866613, "grad_norm": 0.7529389262199402, "learning_rate": 0.0002619262552771041, "loss": 0.8963, "step": 31800 }, { "epoch": 1.0745082188089463, "grad_norm": 0.7418708801269531, "learning_rate": 0.0002617016976556184, "loss": 0.8931, "step": 31900 }, { "epoch": 1.0778765831312316, "grad_norm": 0.6806597113609314, "learning_rate": 0.0002614771400341327, "loss": 0.8938, "step": 32000 }, { "epoch": 1.0812449474535166, "grad_norm": 0.8359425067901611, "learning_rate": 0.0002612525824126471, "loss": 0.8909, "step": 32100 }, { "epoch": 1.0846133117758017, "grad_norm": 0.7026517391204834, "learning_rate": 0.0002610280247911614, "loss": 0.8932, "step": 32200 }, { "epoch": 1.0879816760980867, "grad_norm": 0.9777092933654785, "learning_rate": 0.0002608034671696757, "loss": 0.8936, "step": 32300 }, { "epoch": 1.0913500404203718, "grad_norm": 0.7617833018302917, "learning_rate": 0.00026057890954819004, "loss": 0.8957, "step": 32400 }, { "epoch": 1.094718404742657, "grad_norm": 0.7257563471794128, "learning_rate": 0.0002603543519267044, "loss": 0.8947, "step": 32500 }, { "epoch": 1.098086769064942, "grad_norm": 0.6984378695487976, "learning_rate": 0.0002601297943052187, "loss": 0.8933, "step": 32600 }, { "epoch": 1.1014551333872271, "grad_norm": 0.7216737270355225, "learning_rate": 0.000259905236683733, "loss": 0.8912, "step": 32700 }, { "epoch": 1.1048234977095122, "grad_norm": 0.8986912965774536, "learning_rate": 0.00025968067906224733, "loss": 0.8924, "step": 32800 }, { "epoch": 1.1081918620317974, "grad_norm": 0.7360677719116211, "learning_rate": 0.00025945612144076167, "loss": 0.8912, "step": 32900 }, { "epoch": 1.1115602263540825, "grad_norm": 0.7221621870994568, "learning_rate": 0.000259231563819276, "loss": 0.8903, "step": 33000 }, { "epoch": 1.1149285906763675, "grad_norm": 0.8666985034942627, "learning_rate": 0.00025900700619779034, "loss": 0.8869, "step": 33100 }, { "epoch": 1.1182969549986526, "grad_norm": 0.7224547266960144, "learning_rate": 0.00025878244857630467, "loss": 0.8909, "step": 33200 }, { "epoch": 1.1216653193209378, "grad_norm": 1.036876916885376, "learning_rate": 0.000258557890954819, "loss": 0.8931, "step": 33300 }, { "epoch": 1.1250336836432229, "grad_norm": 0.795993447303772, "learning_rate": 0.00025833333333333334, "loss": 0.8853, "step": 33400 }, { "epoch": 1.128402047965508, "grad_norm": 0.7044444680213928, "learning_rate": 0.0002581087757118476, "loss": 0.8913, "step": 33500 }, { "epoch": 1.131770412287793, "grad_norm": 0.7511852979660034, "learning_rate": 0.00025788421809036196, "loss": 0.8886, "step": 33600 }, { "epoch": 1.1351387766100782, "grad_norm": 0.8065422177314758, "learning_rate": 0.0002576596604688763, "loss": 0.8882, "step": 33700 }, { "epoch": 1.1385071409323633, "grad_norm": 0.7372477650642395, "learning_rate": 0.00025743510284739063, "loss": 0.8886, "step": 33800 }, { "epoch": 1.1418755052546483, "grad_norm": 0.7889522314071655, "learning_rate": 0.00025721054522590496, "loss": 0.8886, "step": 33900 }, { "epoch": 1.1452438695769334, "grad_norm": 0.6768883466720581, "learning_rate": 0.00025698598760441925, "loss": 0.8895, "step": 34000 }, { "epoch": 1.1486122338992186, "grad_norm": 0.8599404096603394, "learning_rate": 0.00025676142998293363, "loss": 0.8904, "step": 34100 }, { "epoch": 1.1519805982215037, "grad_norm": 0.716754138469696, "learning_rate": 0.0002565368723614479, "loss": 0.8898, "step": 34200 }, { "epoch": 1.1553489625437887, "grad_norm": 0.7366636991500854, "learning_rate": 0.00025631231473996225, "loss": 0.8872, "step": 34300 }, { "epoch": 1.1587173268660738, "grad_norm": 0.7212886810302734, "learning_rate": 0.0002560877571184766, "loss": 0.8897, "step": 34400 }, { "epoch": 1.162085691188359, "grad_norm": 0.768371045589447, "learning_rate": 0.00025586319949699087, "loss": 0.8871, "step": 34500 }, { "epoch": 1.165454055510644, "grad_norm": 0.7179331183433533, "learning_rate": 0.00025563864187550526, "loss": 0.8864, "step": 34600 }, { "epoch": 1.1688224198329291, "grad_norm": 0.7503130435943604, "learning_rate": 0.00025541408425401954, "loss": 0.8857, "step": 34700 }, { "epoch": 1.1721907841552142, "grad_norm": 0.7497594952583313, "learning_rate": 0.00025518952663253387, "loss": 0.8858, "step": 34800 }, { "epoch": 1.1755591484774994, "grad_norm": 0.6965382695198059, "learning_rate": 0.0002549649690110482, "loss": 0.8867, "step": 34900 }, { "epoch": 1.1789275127997845, "grad_norm": 0.7309035658836365, "learning_rate": 0.00025474041138956254, "loss": 0.8865, "step": 35000 }, { "epoch": 1.1822958771220695, "grad_norm": 0.692935049533844, "learning_rate": 0.0002545158537680769, "loss": 0.8839, "step": 35100 }, { "epoch": 1.1856642414443546, "grad_norm": 0.7206865549087524, "learning_rate": 0.0002542912961465912, "loss": 0.8833, "step": 35200 }, { "epoch": 1.1890326057666396, "grad_norm": 0.718701958656311, "learning_rate": 0.0002540667385251055, "loss": 0.8831, "step": 35300 }, { "epoch": 1.1924009700889249, "grad_norm": 0.797205924987793, "learning_rate": 0.0002538421809036199, "loss": 0.8851, "step": 35400 }, { "epoch": 1.19576933441121, "grad_norm": 0.6977315545082092, "learning_rate": 0.00025361762328213416, "loss": 0.8834, "step": 35500 }, { "epoch": 1.199137698733495, "grad_norm": 0.768139660358429, "learning_rate": 0.0002533930656606485, "loss": 0.8822, "step": 35600 }, { "epoch": 1.2025060630557802, "grad_norm": 0.7986385226249695, "learning_rate": 0.00025316850803916283, "loss": 0.8816, "step": 35700 }, { "epoch": 1.2058744273780653, "grad_norm": 0.7320616841316223, "learning_rate": 0.00025294395041767717, "loss": 0.8827, "step": 35800 }, { "epoch": 1.2092427917003503, "grad_norm": 0.7750969529151917, "learning_rate": 0.0002527193927961915, "loss": 0.8851, "step": 35900 }, { "epoch": 1.2126111560226354, "grad_norm": 0.7603546380996704, "learning_rate": 0.0002524948351747058, "loss": 0.8834, "step": 36000 }, { "epoch": 1.2159795203449204, "grad_norm": 0.6906490325927734, "learning_rate": 0.0002522702775532201, "loss": 0.8835, "step": 36100 }, { "epoch": 1.2193478846672057, "grad_norm": 0.7611385583877563, "learning_rate": 0.00025204571993173446, "loss": 0.8832, "step": 36200 }, { "epoch": 1.2227162489894907, "grad_norm": 0.732187807559967, "learning_rate": 0.0002518211623102488, "loss": 0.8813, "step": 36300 }, { "epoch": 1.2260846133117758, "grad_norm": 0.7710540890693665, "learning_rate": 0.0002515966046887631, "loss": 0.8835, "step": 36400 }, { "epoch": 1.2294529776340608, "grad_norm": 0.7388759255409241, "learning_rate": 0.0002513720470672774, "loss": 0.8846, "step": 36500 }, { "epoch": 1.232821341956346, "grad_norm": 0.7000882029533386, "learning_rate": 0.0002511474894457918, "loss": 0.8818, "step": 36600 }, { "epoch": 1.2361897062786311, "grad_norm": 0.7058309316635132, "learning_rate": 0.0002509229318243061, "loss": 0.8827, "step": 36700 }, { "epoch": 1.2395580706009162, "grad_norm": 0.80439692735672, "learning_rate": 0.0002506983742028204, "loss": 0.8793, "step": 36800 }, { "epoch": 1.2429264349232012, "grad_norm": 0.7056805491447449, "learning_rate": 0.00025047381658133475, "loss": 0.8763, "step": 36900 }, { "epoch": 1.2462947992454865, "grad_norm": 0.7477532625198364, "learning_rate": 0.0002502492589598491, "loss": 0.8803, "step": 37000 }, { "epoch": 1.2496631635677715, "grad_norm": 0.8339030742645264, "learning_rate": 0.0002500247013383634, "loss": 0.8782, "step": 37100 }, { "epoch": 1.2530315278900566, "grad_norm": 0.7052040100097656, "learning_rate": 0.00024980014371687775, "loss": 0.8834, "step": 37200 }, { "epoch": 1.2563998922123416, "grad_norm": 0.7932031750679016, "learning_rate": 0.00024957558609539204, "loss": 0.882, "step": 37300 }, { "epoch": 1.2597682565346267, "grad_norm": 0.7311996817588806, "learning_rate": 0.00024935102847390637, "loss": 0.8802, "step": 37400 }, { "epoch": 1.263136620856912, "grad_norm": 0.7274471521377563, "learning_rate": 0.0002491264708524207, "loss": 0.8801, "step": 37500 }, { "epoch": 1.266504985179197, "grad_norm": 0.6959684491157532, "learning_rate": 0.00024890191323093504, "loss": 0.8742, "step": 37600 }, { "epoch": 1.269873349501482, "grad_norm": 0.7355061173439026, "learning_rate": 0.0002486773556094494, "loss": 0.8763, "step": 37700 }, { "epoch": 1.2732417138237673, "grad_norm": 0.6900234818458557, "learning_rate": 0.00024845279798796366, "loss": 0.879, "step": 37800 }, { "epoch": 1.2766100781460523, "grad_norm": 0.7259871363639832, "learning_rate": 0.00024822824036647805, "loss": 0.8769, "step": 37900 }, { "epoch": 1.2799784424683374, "grad_norm": 0.7086994051933289, "learning_rate": 0.00024800368274499233, "loss": 0.877, "step": 38000 }, { "epoch": 1.2833468067906224, "grad_norm": 0.7197995185852051, "learning_rate": 0.00024777912512350666, "loss": 0.8757, "step": 38100 }, { "epoch": 1.2867151711129075, "grad_norm": 0.7302994728088379, "learning_rate": 0.000247554567502021, "loss": 0.8801, "step": 38200 }, { "epoch": 1.2900835354351927, "grad_norm": 0.6908255815505981, "learning_rate": 0.00024733000988053533, "loss": 0.8766, "step": 38300 }, { "epoch": 1.2934518997574778, "grad_norm": 0.8338357210159302, "learning_rate": 0.00024710545225904967, "loss": 0.874, "step": 38400 }, { "epoch": 1.2968202640797628, "grad_norm": 0.7342631220817566, "learning_rate": 0.00024688089463756395, "loss": 0.8786, "step": 38500 }, { "epoch": 1.300188628402048, "grad_norm": 0.7488550543785095, "learning_rate": 0.0002466563370160783, "loss": 0.8813, "step": 38600 }, { "epoch": 1.3035569927243331, "grad_norm": 0.7316462993621826, "learning_rate": 0.0002464317793945926, "loss": 0.8755, "step": 38700 }, { "epoch": 1.3069253570466182, "grad_norm": 0.7482060790061951, "learning_rate": 0.00024620722177310695, "loss": 0.8737, "step": 38800 }, { "epoch": 1.3102937213689032, "grad_norm": 0.7243971228599548, "learning_rate": 0.0002459826641516213, "loss": 0.8759, "step": 38900 }, { "epoch": 1.3136620856911883, "grad_norm": 0.8776415586471558, "learning_rate": 0.0002457581065301356, "loss": 0.8757, "step": 39000 }, { "epoch": 1.3170304500134735, "grad_norm": 0.6730260848999023, "learning_rate": 0.00024553354890864996, "loss": 0.873, "step": 39100 }, { "epoch": 1.3203988143357586, "grad_norm": 0.6937080025672913, "learning_rate": 0.0002453089912871643, "loss": 0.8737, "step": 39200 }, { "epoch": 1.3237671786580436, "grad_norm": 0.7004138827323914, "learning_rate": 0.0002450844336656786, "loss": 0.8771, "step": 39300 }, { "epoch": 1.3271355429803289, "grad_norm": 0.6719589829444885, "learning_rate": 0.0002448598760441929, "loss": 0.8758, "step": 39400 }, { "epoch": 1.330503907302614, "grad_norm": 0.738394021987915, "learning_rate": 0.00024463531842270725, "loss": 0.8743, "step": 39500 }, { "epoch": 1.333872271624899, "grad_norm": 0.7103344202041626, "learning_rate": 0.0002444107608012216, "loss": 0.8712, "step": 39600 }, { "epoch": 1.337240635947184, "grad_norm": 0.69880610704422, "learning_rate": 0.0002441862031797359, "loss": 0.8737, "step": 39700 }, { "epoch": 1.340609000269469, "grad_norm": 0.8922190070152283, "learning_rate": 0.00024396164555825023, "loss": 0.8718, "step": 39800 }, { "epoch": 1.3439773645917543, "grad_norm": 0.7306973338127136, "learning_rate": 0.00024373708793676456, "loss": 0.8712, "step": 39900 }, { "epoch": 1.3473457289140394, "grad_norm": 0.7311689257621765, "learning_rate": 0.00024351253031527887, "loss": 0.875, "step": 40000 }, { "epoch": 1.3507140932363244, "grad_norm": 0.7093273997306824, "learning_rate": 0.00024328797269379323, "loss": 0.8711, "step": 40100 }, { "epoch": 1.3540824575586097, "grad_norm": 0.6579886078834534, "learning_rate": 0.00024306341507230754, "loss": 0.8773, "step": 40200 }, { "epoch": 1.3574508218808947, "grad_norm": 0.6832658648490906, "learning_rate": 0.00024283885745082185, "loss": 0.8727, "step": 40300 }, { "epoch": 1.3608191862031798, "grad_norm": 0.7372367978096008, "learning_rate": 0.00024261429982933618, "loss": 0.8718, "step": 40400 }, { "epoch": 1.3641875505254648, "grad_norm": 0.691271185874939, "learning_rate": 0.0002423897422078505, "loss": 0.8716, "step": 40500 }, { "epoch": 1.3675559148477499, "grad_norm": 0.7311553359031677, "learning_rate": 0.00024216518458636485, "loss": 0.8711, "step": 40600 }, { "epoch": 1.3709242791700351, "grad_norm": 0.7508808970451355, "learning_rate": 0.00024194062696487916, "loss": 0.8708, "step": 40700 }, { "epoch": 1.3742926434923202, "grad_norm": 0.707360029220581, "learning_rate": 0.0002417160693433935, "loss": 0.8727, "step": 40800 }, { "epoch": 1.3776610078146052, "grad_norm": 0.6759727001190186, "learning_rate": 0.0002414915117219078, "loss": 0.8681, "step": 40900 }, { "epoch": 1.3810293721368903, "grad_norm": 0.6768555045127869, "learning_rate": 0.00024126695410042217, "loss": 0.8709, "step": 41000 }, { "epoch": 1.3843977364591753, "grad_norm": 0.7042515873908997, "learning_rate": 0.00024104239647893647, "loss": 0.8711, "step": 41100 }, { "epoch": 1.3877661007814606, "grad_norm": 0.7361947894096375, "learning_rate": 0.00024081783885745078, "loss": 0.87, "step": 41200 }, { "epoch": 1.3911344651037456, "grad_norm": 0.7217181921005249, "learning_rate": 0.00024059328123596512, "loss": 0.8673, "step": 41300 }, { "epoch": 1.3945028294260307, "grad_norm": 0.6747815608978271, "learning_rate": 0.00024036872361447945, "loss": 0.8686, "step": 41400 }, { "epoch": 1.397871193748316, "grad_norm": 0.8375660181045532, "learning_rate": 0.0002401441659929938, "loss": 0.8711, "step": 41500 }, { "epoch": 1.401239558070601, "grad_norm": 0.7188234925270081, "learning_rate": 0.0002399196083715081, "loss": 0.8687, "step": 41600 }, { "epoch": 1.404607922392886, "grad_norm": 0.7048287987709045, "learning_rate": 0.00023969505075002243, "loss": 0.8663, "step": 41700 }, { "epoch": 1.407976286715171, "grad_norm": 0.703464925289154, "learning_rate": 0.00023947049312853677, "loss": 0.8698, "step": 41800 }, { "epoch": 1.411344651037456, "grad_norm": 0.725428581237793, "learning_rate": 0.0002392459355070511, "loss": 0.8704, "step": 41900 }, { "epoch": 1.4147130153597414, "grad_norm": 0.7408603429794312, "learning_rate": 0.0002390213778855654, "loss": 0.868, "step": 42000 }, { "epoch": 1.4180813796820264, "grad_norm": 0.718657910823822, "learning_rate": 0.00023879682026407974, "loss": 0.8684, "step": 42100 }, { "epoch": 1.4214497440043115, "grad_norm": 0.6897197961807251, "learning_rate": 0.00023857226264259408, "loss": 0.8638, "step": 42200 }, { "epoch": 1.4248181083265967, "grad_norm": 0.7002888917922974, "learning_rate": 0.0002383477050211084, "loss": 0.8694, "step": 42300 }, { "epoch": 1.4281864726488818, "grad_norm": 0.6644707322120667, "learning_rate": 0.00023812314739962272, "loss": 0.8684, "step": 42400 }, { "epoch": 1.4315548369711668, "grad_norm": 0.7583789229393005, "learning_rate": 0.00023789858977813703, "loss": 0.8671, "step": 42500 }, { "epoch": 1.4349232012934519, "grad_norm": 1.1469764709472656, "learning_rate": 0.0002376740321566514, "loss": 0.8659, "step": 42600 }, { "epoch": 1.438291565615737, "grad_norm": 0.7378877997398376, "learning_rate": 0.0002374494745351657, "loss": 0.8668, "step": 42700 }, { "epoch": 1.4416599299380222, "grad_norm": 0.6720541715621948, "learning_rate": 0.00023722491691368004, "loss": 0.867, "step": 42800 }, { "epoch": 1.4450282942603072, "grad_norm": 0.6979005336761475, "learning_rate": 0.00023700035929219435, "loss": 0.865, "step": 42900 }, { "epoch": 1.4483966585825923, "grad_norm": 0.6991161108016968, "learning_rate": 0.0002367758016707087, "loss": 0.8646, "step": 43000 }, { "epoch": 1.4517650229048775, "grad_norm": 0.7050434350967407, "learning_rate": 0.00023655124404922302, "loss": 0.8663, "step": 43100 }, { "epoch": 1.4551333872271626, "grad_norm": 0.733540415763855, "learning_rate": 0.00023632668642773732, "loss": 0.869, "step": 43200 }, { "epoch": 1.4585017515494476, "grad_norm": 0.6488509774208069, "learning_rate": 0.00023610212880625166, "loss": 0.8645, "step": 43300 }, { "epoch": 1.4618701158717327, "grad_norm": 0.760858416557312, "learning_rate": 0.00023587757118476597, "loss": 0.8671, "step": 43400 }, { "epoch": 1.4652384801940177, "grad_norm": 0.6800232529640198, "learning_rate": 0.00023565301356328033, "loss": 0.8647, "step": 43500 }, { "epoch": 1.468606844516303, "grad_norm": 0.6814864873886108, "learning_rate": 0.00023542845594179464, "loss": 0.8679, "step": 43600 }, { "epoch": 1.471975208838588, "grad_norm": 0.6680698394775391, "learning_rate": 0.00023520389832030897, "loss": 0.8638, "step": 43700 }, { "epoch": 1.475343573160873, "grad_norm": 0.7103099226951599, "learning_rate": 0.00023497934069882328, "loss": 0.8625, "step": 43800 }, { "epoch": 1.478711937483158, "grad_norm": 0.6946255564689636, "learning_rate": 0.00023475478307733764, "loss": 0.8636, "step": 43900 }, { "epoch": 1.4820803018054431, "grad_norm": 0.7505577802658081, "learning_rate": 0.00023453022545585195, "loss": 0.8622, "step": 44000 }, { "epoch": 1.4854486661277284, "grad_norm": 0.623587965965271, "learning_rate": 0.00023430566783436626, "loss": 0.8617, "step": 44100 }, { "epoch": 1.4888170304500135, "grad_norm": 0.7254552841186523, "learning_rate": 0.0002340811102128806, "loss": 0.8631, "step": 44200 }, { "epoch": 1.4921853947722985, "grad_norm": 0.7500590682029724, "learning_rate": 0.00023385655259139493, "loss": 0.8665, "step": 44300 }, { "epoch": 1.4955537590945838, "grad_norm": 0.6879542469978333, "learning_rate": 0.00023363199496990926, "loss": 0.8652, "step": 44400 }, { "epoch": 1.4989221234168688, "grad_norm": 1.074501872062683, "learning_rate": 0.00023340743734842357, "loss": 0.8635, "step": 44500 }, { "epoch": 1.5022904877391539, "grad_norm": 0.6877649426460266, "learning_rate": 0.0002331828797269379, "loss": 0.8639, "step": 44600 }, { "epoch": 1.5056588520614391, "grad_norm": 0.6956282258033752, "learning_rate": 0.00023295832210545224, "loss": 0.8625, "step": 44700 }, { "epoch": 1.509027216383724, "grad_norm": 0.6834612488746643, "learning_rate": 0.00023273376448396658, "loss": 0.863, "step": 44800 }, { "epoch": 1.5123955807060092, "grad_norm": 0.6679215431213379, "learning_rate": 0.00023250920686248089, "loss": 0.8604, "step": 44900 }, { "epoch": 1.5157639450282943, "grad_norm": 0.7261891961097717, "learning_rate": 0.00023228464924099525, "loss": 0.8614, "step": 45000 }, { "epoch": 1.5191323093505793, "grad_norm": 0.6932804584503174, "learning_rate": 0.00023206009161950956, "loss": 0.8629, "step": 45100 }, { "epoch": 1.5225006736728646, "grad_norm": 0.9139769673347473, "learning_rate": 0.00023183553399802386, "loss": 0.8592, "step": 45200 }, { "epoch": 1.5258690379951494, "grad_norm": 0.7025532126426697, "learning_rate": 0.0002316109763765382, "loss": 0.8623, "step": 45300 }, { "epoch": 1.5292374023174347, "grad_norm": 0.6872456073760986, "learning_rate": 0.0002313864187550525, "loss": 0.8634, "step": 45400 }, { "epoch": 1.5326057666397197, "grad_norm": 0.6915197968482971, "learning_rate": 0.00023116186113356687, "loss": 0.8604, "step": 45500 }, { "epoch": 1.5359741309620047, "grad_norm": 0.6892699599266052, "learning_rate": 0.00023093730351208118, "loss": 0.8591, "step": 45600 }, { "epoch": 1.53934249528429, "grad_norm": 0.7258453369140625, "learning_rate": 0.00023071274589059551, "loss": 0.8611, "step": 45700 }, { "epoch": 1.542710859606575, "grad_norm": 0.6410971879959106, "learning_rate": 0.00023048818826910982, "loss": 0.8585, "step": 45800 }, { "epoch": 1.54607922392886, "grad_norm": 0.7481923699378967, "learning_rate": 0.00023026363064762418, "loss": 0.8612, "step": 45900 }, { "epoch": 1.5494475882511454, "grad_norm": 0.6729730367660522, "learning_rate": 0.0002300390730261385, "loss": 0.8594, "step": 46000 }, { "epoch": 1.5528159525734302, "grad_norm": 0.6492398977279663, "learning_rate": 0.0002298145154046528, "loss": 0.8596, "step": 46100 }, { "epoch": 1.5561843168957155, "grad_norm": 0.7804999947547913, "learning_rate": 0.00022958995778316714, "loss": 0.8588, "step": 46200 }, { "epoch": 1.5595526812180005, "grad_norm": 0.692258358001709, "learning_rate": 0.00022936540016168144, "loss": 0.8644, "step": 46300 }, { "epoch": 1.5629210455402855, "grad_norm": 0.8399534821510315, "learning_rate": 0.0002291408425401958, "loss": 0.8563, "step": 46400 }, { "epoch": 1.5662894098625708, "grad_norm": 0.7541986107826233, "learning_rate": 0.00022891628491871011, "loss": 0.8597, "step": 46500 }, { "epoch": 1.5696577741848559, "grad_norm": 0.7269881367683411, "learning_rate": 0.00022869172729722445, "loss": 0.8613, "step": 46600 }, { "epoch": 1.573026138507141, "grad_norm": 0.6875913143157959, "learning_rate": 0.00022846716967573876, "loss": 0.8569, "step": 46700 }, { "epoch": 1.5763945028294262, "grad_norm": 0.6820959448814392, "learning_rate": 0.00022824261205425312, "loss": 0.8578, "step": 46800 }, { "epoch": 1.579762867151711, "grad_norm": 0.63446044921875, "learning_rate": 0.00022801805443276743, "loss": 0.8583, "step": 46900 }, { "epoch": 1.5831312314739963, "grad_norm": 0.7082163691520691, "learning_rate": 0.00022779349681128174, "loss": 0.8591, "step": 47000 }, { "epoch": 1.5864995957962813, "grad_norm": 0.6886746883392334, "learning_rate": 0.0002275689391897961, "loss": 0.8569, "step": 47100 }, { "epoch": 1.5898679601185663, "grad_norm": 0.6877068281173706, "learning_rate": 0.0002273443815683104, "loss": 0.8563, "step": 47200 }, { "epoch": 1.5932363244408516, "grad_norm": 0.7213451266288757, "learning_rate": 0.00022711982394682474, "loss": 0.8576, "step": 47300 }, { "epoch": 1.5966046887631367, "grad_norm": 0.6784006357192993, "learning_rate": 0.00022689526632533905, "loss": 0.8561, "step": 47400 }, { "epoch": 1.5999730530854217, "grad_norm": 0.7651084661483765, "learning_rate": 0.0002266707087038534, "loss": 0.8533, "step": 47500 }, { "epoch": 1.603341417407707, "grad_norm": 0.6762063503265381, "learning_rate": 0.00022644615108236772, "loss": 0.8593, "step": 47600 }, { "epoch": 1.6067097817299918, "grad_norm": 0.6897014379501343, "learning_rate": 0.00022622159346088205, "loss": 0.8584, "step": 47700 }, { "epoch": 1.610078146052277, "grad_norm": 0.642955482006073, "learning_rate": 0.00022599703583939636, "loss": 0.8536, "step": 47800 }, { "epoch": 1.613446510374562, "grad_norm": 0.676459014415741, "learning_rate": 0.00022577247821791067, "loss": 0.8561, "step": 47900 }, { "epoch": 1.6168148746968471, "grad_norm": 0.7143009901046753, "learning_rate": 0.00022554792059642503, "loss": 0.8593, "step": 48000 }, { "epoch": 1.6201832390191324, "grad_norm": 0.7367132306098938, "learning_rate": 0.00022532336297493934, "loss": 0.8547, "step": 48100 }, { "epoch": 1.6235516033414175, "grad_norm": 0.6949586868286133, "learning_rate": 0.00022509880535345368, "loss": 0.8577, "step": 48200 }, { "epoch": 1.6269199676637025, "grad_norm": 0.672558605670929, "learning_rate": 0.00022487424773196798, "loss": 0.8526, "step": 48300 }, { "epoch": 1.6302883319859878, "grad_norm": 0.6996020078659058, "learning_rate": 0.00022464969011048235, "loss": 0.8527, "step": 48400 }, { "epoch": 1.6336566963082726, "grad_norm": 0.7007562518119812, "learning_rate": 0.00022442513248899666, "loss": 0.8536, "step": 48500 }, { "epoch": 1.6370250606305579, "grad_norm": 0.7103307247161865, "learning_rate": 0.000224200574867511, "loss": 0.8557, "step": 48600 }, { "epoch": 1.640393424952843, "grad_norm": 0.7483230829238892, "learning_rate": 0.0002239760172460253, "loss": 0.857, "step": 48700 }, { "epoch": 1.643761789275128, "grad_norm": 0.7092981338500977, "learning_rate": 0.00022375145962453966, "loss": 0.8589, "step": 48800 }, { "epoch": 1.6471301535974132, "grad_norm": 0.7270293235778809, "learning_rate": 0.00022352690200305397, "loss": 0.8553, "step": 48900 }, { "epoch": 1.650498517919698, "grad_norm": 0.6916648149490356, "learning_rate": 0.00022330234438156828, "loss": 0.8566, "step": 49000 }, { "epoch": 1.6538668822419833, "grad_norm": 0.6687049269676208, "learning_rate": 0.0002230777867600826, "loss": 0.8532, "step": 49100 }, { "epoch": 1.6572352465642683, "grad_norm": 0.689947783946991, "learning_rate": 0.00022285322913859695, "loss": 0.8556, "step": 49200 }, { "epoch": 1.6606036108865534, "grad_norm": 0.7345608472824097, "learning_rate": 0.00022262867151711128, "loss": 0.8526, "step": 49300 }, { "epoch": 1.6639719752088387, "grad_norm": 0.6420150399208069, "learning_rate": 0.0002224041138956256, "loss": 0.8584, "step": 49400 }, { "epoch": 1.6673403395311237, "grad_norm": 0.7315524816513062, "learning_rate": 0.00022217955627413993, "loss": 0.8516, "step": 49500 }, { "epoch": 1.6707087038534087, "grad_norm": 0.674518883228302, "learning_rate": 0.00022195499865265426, "loss": 0.8564, "step": 49600 }, { "epoch": 1.674077068175694, "grad_norm": 0.6845901608467102, "learning_rate": 0.0002217304410311686, "loss": 0.8493, "step": 49700 }, { "epoch": 1.6774454324979788, "grad_norm": 0.6932186484336853, "learning_rate": 0.0002215058834096829, "loss": 0.8516, "step": 49800 }, { "epoch": 1.680813796820264, "grad_norm": 0.6544727087020874, "learning_rate": 0.0002212813257881972, "loss": 0.8519, "step": 49900 }, { "epoch": 1.6841821611425492, "grad_norm": 0.7253223657608032, "learning_rate": 0.00022105676816671157, "loss": 0.8529, "step": 50000 }, { "epoch": 1.6875505254648342, "grad_norm": 0.6576797962188721, "learning_rate": 0.00022083221054522588, "loss": 0.8528, "step": 50100 }, { "epoch": 1.6909188897871195, "grad_norm": 0.715744137763977, "learning_rate": 0.00022060765292374022, "loss": 0.8537, "step": 50200 }, { "epoch": 1.6942872541094045, "grad_norm": 0.6994728446006775, "learning_rate": 0.00022038309530225453, "loss": 0.852, "step": 50300 }, { "epoch": 1.6976556184316896, "grad_norm": 0.6934739947319031, "learning_rate": 0.0002201585376807689, "loss": 0.8509, "step": 50400 }, { "epoch": 1.7010239827539748, "grad_norm": 0.7031779885292053, "learning_rate": 0.0002199339800592832, "loss": 0.853, "step": 50500 }, { "epoch": 1.7043923470762596, "grad_norm": 0.7014051079750061, "learning_rate": 0.00021970942243779753, "loss": 0.8514, "step": 50600 }, { "epoch": 1.707760711398545, "grad_norm": 0.8519027233123779, "learning_rate": 0.00021948486481631184, "loss": 0.8525, "step": 50700 }, { "epoch": 1.71112907572083, "grad_norm": 0.6523889899253845, "learning_rate": 0.00021926030719482615, "loss": 0.8514, "step": 50800 }, { "epoch": 1.714497440043115, "grad_norm": 0.7346422076225281, "learning_rate": 0.0002190357495733405, "loss": 0.856, "step": 50900 }, { "epoch": 1.7178658043654003, "grad_norm": 0.7833214402198792, "learning_rate": 0.00021881119195185482, "loss": 0.8518, "step": 51000 }, { "epoch": 1.7212341686876853, "grad_norm": 0.599700391292572, "learning_rate": 0.00021858663433036915, "loss": 0.8539, "step": 51100 }, { "epoch": 1.7246025330099704, "grad_norm": 0.7568506002426147, "learning_rate": 0.00021836207670888346, "loss": 0.8504, "step": 51200 }, { "epoch": 1.7279708973322556, "grad_norm": 0.71254563331604, "learning_rate": 0.00021813751908739782, "loss": 0.849, "step": 51300 }, { "epoch": 1.7313392616545404, "grad_norm": 0.6965556144714355, "learning_rate": 0.00021791296146591213, "loss": 0.8506, "step": 51400 }, { "epoch": 1.7347076259768257, "grad_norm": 0.6838847994804382, "learning_rate": 0.00021768840384442647, "loss": 0.8487, "step": 51500 }, { "epoch": 1.7380759902991108, "grad_norm": 0.6968240737915039, "learning_rate": 0.00021746384622294077, "loss": 0.8498, "step": 51600 }, { "epoch": 1.7414443546213958, "grad_norm": 0.6424885988235474, "learning_rate": 0.00021723928860145514, "loss": 0.8495, "step": 51700 }, { "epoch": 1.744812718943681, "grad_norm": 0.6591705083847046, "learning_rate": 0.00021701473097996945, "loss": 0.8474, "step": 51800 }, { "epoch": 1.7481810832659659, "grad_norm": 0.9693089723587036, "learning_rate": 0.00021679017335848375, "loss": 0.8468, "step": 51900 }, { "epoch": 1.7515494475882512, "grad_norm": 0.6744303107261658, "learning_rate": 0.0002165656157369981, "loss": 0.8488, "step": 52000 }, { "epoch": 1.7549178119105362, "grad_norm": 0.7250896096229553, "learning_rate": 0.00021634105811551242, "loss": 0.8501, "step": 52100 }, { "epoch": 1.7582861762328212, "grad_norm": 0.6125505566596985, "learning_rate": 0.00021611650049402676, "loss": 0.8481, "step": 52200 }, { "epoch": 1.7616545405551065, "grad_norm": 0.6853183507919312, "learning_rate": 0.00021589194287254107, "loss": 0.8496, "step": 52300 }, { "epoch": 1.7650229048773916, "grad_norm": 0.6613684892654419, "learning_rate": 0.0002156673852510554, "loss": 0.8496, "step": 52400 }, { "epoch": 1.7683912691996766, "grad_norm": 0.6338353157043457, "learning_rate": 0.00021544282762956974, "loss": 0.8473, "step": 52500 }, { "epoch": 1.7717596335219619, "grad_norm": 0.7311720848083496, "learning_rate": 0.00021521827000808407, "loss": 0.8494, "step": 52600 }, { "epoch": 1.7751279978442467, "grad_norm": 0.8986194133758545, "learning_rate": 0.00021499371238659838, "loss": 0.847, "step": 52700 }, { "epoch": 1.778496362166532, "grad_norm": 0.7145596146583557, "learning_rate": 0.0002147691547651127, "loss": 0.8513, "step": 52800 }, { "epoch": 1.781864726488817, "grad_norm": 0.673674464225769, "learning_rate": 0.00021454459714362705, "loss": 0.8496, "step": 52900 }, { "epoch": 1.785233090811102, "grad_norm": 0.6568505167961121, "learning_rate": 0.00021432003952214136, "loss": 0.8477, "step": 53000 }, { "epoch": 1.7886014551333873, "grad_norm": 0.8661649227142334, "learning_rate": 0.0002140954819006557, "loss": 0.8437, "step": 53100 }, { "epoch": 1.7919698194556724, "grad_norm": 0.636997401714325, "learning_rate": 0.00021387092427917, "loss": 0.8498, "step": 53200 }, { "epoch": 1.7953381837779574, "grad_norm": 0.6258701682090759, "learning_rate": 0.00021364636665768436, "loss": 0.846, "step": 53300 }, { "epoch": 1.7987065481002427, "grad_norm": 0.7167558073997498, "learning_rate": 0.00021342180903619867, "loss": 0.8496, "step": 53400 }, { "epoch": 1.8020749124225275, "grad_norm": 0.725511372089386, "learning_rate": 0.000213197251414713, "loss": 0.8503, "step": 53500 }, { "epoch": 1.8054432767448128, "grad_norm": 0.7038520574569702, "learning_rate": 0.00021297269379322732, "loss": 0.8486, "step": 53600 }, { "epoch": 1.8088116410670978, "grad_norm": 0.6768267154693604, "learning_rate": 0.00021274813617174162, "loss": 0.8475, "step": 53700 }, { "epoch": 1.8121800053893828, "grad_norm": 0.6625707745552063, "learning_rate": 0.000212523578550256, "loss": 0.8466, "step": 53800 }, { "epoch": 1.815548369711668, "grad_norm": 0.6589378118515015, "learning_rate": 0.0002122990209287703, "loss": 0.8449, "step": 53900 }, { "epoch": 1.8189167340339532, "grad_norm": 0.8694218993186951, "learning_rate": 0.00021207446330728463, "loss": 0.8464, "step": 54000 }, { "epoch": 1.8222850983562382, "grad_norm": 0.6879692673683167, "learning_rate": 0.00021184990568579894, "loss": 0.8464, "step": 54100 }, { "epoch": 1.8256534626785235, "grad_norm": 0.7406269907951355, "learning_rate": 0.0002116253480643133, "loss": 0.8451, "step": 54200 }, { "epoch": 1.8290218270008083, "grad_norm": 0.645255982875824, "learning_rate": 0.0002114007904428276, "loss": 0.8459, "step": 54300 }, { "epoch": 1.8323901913230936, "grad_norm": 0.6275530457496643, "learning_rate": 0.00021117623282134194, "loss": 0.848, "step": 54400 }, { "epoch": 1.8357585556453786, "grad_norm": 0.673546314239502, "learning_rate": 0.00021095167519985625, "loss": 0.8421, "step": 54500 }, { "epoch": 1.8391269199676636, "grad_norm": 0.6627029180526733, "learning_rate": 0.00021072711757837061, "loss": 0.8443, "step": 54600 }, { "epoch": 1.842495284289949, "grad_norm": 0.6747744083404541, "learning_rate": 0.00021050255995688492, "loss": 0.8439, "step": 54700 }, { "epoch": 1.845863648612234, "grad_norm": 0.6658967733383179, "learning_rate": 0.00021027800233539923, "loss": 0.8435, "step": 54800 }, { "epoch": 1.849232012934519, "grad_norm": 0.6818722486495972, "learning_rate": 0.00021005344471391357, "loss": 0.8433, "step": 54900 }, { "epoch": 1.852600377256804, "grad_norm": 0.6817068457603455, "learning_rate": 0.0002098288870924279, "loss": 0.8449, "step": 55000 }, { "epoch": 1.855968741579089, "grad_norm": 0.6656964421272278, "learning_rate": 0.00020960432947094224, "loss": 0.8448, "step": 55100 }, { "epoch": 1.8593371059013744, "grad_norm": 0.7307213544845581, "learning_rate": 0.00020937977184945654, "loss": 0.8439, "step": 55200 }, { "epoch": 1.8627054702236594, "grad_norm": 0.7471979856491089, "learning_rate": 0.00020915521422797088, "loss": 0.8461, "step": 55300 }, { "epoch": 1.8660738345459444, "grad_norm": 0.6660134792327881, "learning_rate": 0.00020893065660648521, "loss": 0.849, "step": 55400 }, { "epoch": 1.8694421988682297, "grad_norm": 0.6332405209541321, "learning_rate": 0.00020870609898499955, "loss": 0.8417, "step": 55500 }, { "epoch": 1.8728105631905145, "grad_norm": 0.6369178295135498, "learning_rate": 0.00020848154136351386, "loss": 0.844, "step": 55600 }, { "epoch": 1.8761789275127998, "grad_norm": 0.6878752112388611, "learning_rate": 0.00020825698374202817, "loss": 0.8435, "step": 55700 }, { "epoch": 1.8795472918350848, "grad_norm": 0.7077382206916809, "learning_rate": 0.00020803242612054253, "loss": 0.8448, "step": 55800 }, { "epoch": 1.8829156561573699, "grad_norm": 0.6542516350746155, "learning_rate": 0.00020780786849905684, "loss": 0.8441, "step": 55900 }, { "epoch": 1.8862840204796552, "grad_norm": 0.6779966950416565, "learning_rate": 0.00020758331087757117, "loss": 0.8452, "step": 56000 }, { "epoch": 1.8896523848019402, "grad_norm": 0.7874273657798767, "learning_rate": 0.00020735875325608548, "loss": 0.8408, "step": 56100 }, { "epoch": 1.8930207491242252, "grad_norm": 0.7043356895446777, "learning_rate": 0.00020713419563459984, "loss": 0.8459, "step": 56200 }, { "epoch": 1.8963891134465105, "grad_norm": 0.6597324013710022, "learning_rate": 0.00020690963801311415, "loss": 0.8418, "step": 56300 }, { "epoch": 1.8997574777687953, "grad_norm": 0.6724010705947876, "learning_rate": 0.00020668508039162848, "loss": 0.8459, "step": 56400 }, { "epoch": 1.9031258420910806, "grad_norm": 0.7157804369926453, "learning_rate": 0.0002064605227701428, "loss": 0.8459, "step": 56500 }, { "epoch": 1.9064942064133656, "grad_norm": 0.730675995349884, "learning_rate": 0.0002062359651486571, "loss": 0.8477, "step": 56600 }, { "epoch": 1.9098625707356507, "grad_norm": 0.6512274146080017, "learning_rate": 0.00020601140752717146, "loss": 0.8424, "step": 56700 }, { "epoch": 1.913230935057936, "grad_norm": 0.679719090461731, "learning_rate": 0.00020578684990568577, "loss": 0.8465, "step": 56800 }, { "epoch": 1.916599299380221, "grad_norm": 0.7341735363006592, "learning_rate": 0.0002055622922842001, "loss": 0.8465, "step": 56900 }, { "epoch": 1.919967663702506, "grad_norm": 0.6398690342903137, "learning_rate": 0.00020533773466271441, "loss": 0.8417, "step": 57000 }, { "epoch": 1.9233360280247913, "grad_norm": 0.8629583716392517, "learning_rate": 0.00020511317704122878, "loss": 0.8438, "step": 57100 }, { "epoch": 1.9267043923470761, "grad_norm": 0.6348667740821838, "learning_rate": 0.00020488861941974308, "loss": 0.8411, "step": 57200 }, { "epoch": 1.9300727566693614, "grad_norm": 1.0562591552734375, "learning_rate": 0.00020466406179825742, "loss": 0.8401, "step": 57300 }, { "epoch": 1.9334411209916464, "grad_norm": 0.7126754522323608, "learning_rate": 0.00020443950417677173, "loss": 0.8469, "step": 57400 }, { "epoch": 1.9368094853139315, "grad_norm": 0.624739408493042, "learning_rate": 0.00020421494655528606, "loss": 0.8422, "step": 57500 }, { "epoch": 1.9401778496362168, "grad_norm": 0.6490176916122437, "learning_rate": 0.0002039903889338004, "loss": 0.841, "step": 57600 }, { "epoch": 1.9435462139585018, "grad_norm": 0.6586236357688904, "learning_rate": 0.0002037658313123147, "loss": 0.8412, "step": 57700 }, { "epoch": 1.9469145782807868, "grad_norm": 0.66822749376297, "learning_rate": 0.00020354127369082904, "loss": 0.8418, "step": 57800 }, { "epoch": 1.9502829426030721, "grad_norm": 0.6851320266723633, "learning_rate": 0.00020331671606934338, "loss": 0.8414, "step": 57900 }, { "epoch": 1.953651306925357, "grad_norm": 0.6610788702964783, "learning_rate": 0.0002030921584478577, "loss": 0.8409, "step": 58000 }, { "epoch": 1.9570196712476422, "grad_norm": 0.6205683946609497, "learning_rate": 0.00020286760082637202, "loss": 0.8384, "step": 58100 }, { "epoch": 1.9603880355699272, "grad_norm": 0.6619114875793457, "learning_rate": 0.00020264304320488636, "loss": 0.8404, "step": 58200 }, { "epoch": 1.9637563998922123, "grad_norm": 0.6718551516532898, "learning_rate": 0.0002024184855834007, "loss": 0.8428, "step": 58300 }, { "epoch": 1.9671247642144976, "grad_norm": 0.696954607963562, "learning_rate": 0.00020219392796191503, "loss": 0.8377, "step": 58400 }, { "epoch": 1.9704931285367824, "grad_norm": 0.7444251775741577, "learning_rate": 0.00020196937034042933, "loss": 0.8399, "step": 58500 }, { "epoch": 1.9738614928590676, "grad_norm": 1.1385860443115234, "learning_rate": 0.00020174481271894364, "loss": 0.8424, "step": 58600 }, { "epoch": 1.9772298571813527, "grad_norm": 0.6890325546264648, "learning_rate": 0.000201520255097458, "loss": 0.8441, "step": 58700 }, { "epoch": 1.9805982215036377, "grad_norm": 0.6715498566627502, "learning_rate": 0.0002012956974759723, "loss": 0.8426, "step": 58800 }, { "epoch": 1.983966585825923, "grad_norm": 0.683191180229187, "learning_rate": 0.00020107113985448665, "loss": 0.8408, "step": 58900 }, { "epoch": 1.987334950148208, "grad_norm": 0.7549853324890137, "learning_rate": 0.00020084658223300096, "loss": 0.8406, "step": 59000 }, { "epoch": 1.990703314470493, "grad_norm": 0.6465044617652893, "learning_rate": 0.00020062202461151532, "loss": 0.8398, "step": 59100 }, { "epoch": 1.9940716787927784, "grad_norm": 0.6415732502937317, "learning_rate": 0.00020039746699002963, "loss": 0.8385, "step": 59200 }, { "epoch": 1.9974400431150632, "grad_norm": 0.6739877462387085, "learning_rate": 0.00020017290936854396, "loss": 0.8405, "step": 59300 }, { "epoch": 2.0, "eval_loss": 0.7839618921279907, "eval_runtime": 1.804, "eval_samples_per_second": 2771.644, "eval_steps_per_second": 43.792, "step": 59376 }, { "epoch": 2.0008084074373484, "grad_norm": 0.7029405832290649, "learning_rate": 0.00019994835174705827, "loss": 0.8394, "step": 59400 }, { "epoch": 2.0041767717596337, "grad_norm": 1.100113034248352, "learning_rate": 0.00019972379412557258, "loss": 0.8415, "step": 59500 }, { "epoch": 2.0075451360819185, "grad_norm": 0.6430971026420593, "learning_rate": 0.00019949923650408694, "loss": 0.8368, "step": 59600 }, { "epoch": 2.010913500404204, "grad_norm": 0.6618381142616272, "learning_rate": 0.00019927467888260125, "loss": 0.8374, "step": 59700 }, { "epoch": 2.0142818647264886, "grad_norm": 0.7243971228599548, "learning_rate": 0.00019905012126111558, "loss": 0.84, "step": 59800 }, { "epoch": 2.017650229048774, "grad_norm": 0.6579388380050659, "learning_rate": 0.0001988255636396299, "loss": 0.8395, "step": 59900 }, { "epoch": 2.021018593371059, "grad_norm": 0.6798832416534424, "learning_rate": 0.00019860100601814425, "loss": 0.8378, "step": 60000 }, { "epoch": 2.024386957693344, "grad_norm": 0.7237014174461365, "learning_rate": 0.00019837644839665856, "loss": 0.8403, "step": 60100 }, { "epoch": 2.0277553220156292, "grad_norm": 0.6715816259384155, "learning_rate": 0.0001981518907751729, "loss": 0.8412, "step": 60200 }, { "epoch": 2.0311236863379145, "grad_norm": 0.6749352812767029, "learning_rate": 0.0001979273331536872, "loss": 0.8387, "step": 60300 }, { "epoch": 2.0344920506601993, "grad_norm": 0.6498907804489136, "learning_rate": 0.00019770277553220154, "loss": 0.8365, "step": 60400 }, { "epoch": 2.0378604149824846, "grad_norm": 0.6706451177597046, "learning_rate": 0.00019747821791071588, "loss": 0.8353, "step": 60500 }, { "epoch": 2.0412287793047694, "grad_norm": 0.6508938670158386, "learning_rate": 0.00019725366028923018, "loss": 0.8386, "step": 60600 }, { "epoch": 2.0445971436270547, "grad_norm": 0.6738646626472473, "learning_rate": 0.00019702910266774452, "loss": 0.8357, "step": 60700 }, { "epoch": 2.04796550794934, "grad_norm": 0.6249734163284302, "learning_rate": 0.00019680454504625885, "loss": 0.8381, "step": 60800 }, { "epoch": 2.051333872271625, "grad_norm": 0.72087162733078, "learning_rate": 0.0001965799874247732, "loss": 0.8377, "step": 60900 }, { "epoch": 2.05470223659391, "grad_norm": 0.7066697478294373, "learning_rate": 0.0001963554298032875, "loss": 0.8397, "step": 61000 }, { "epoch": 2.0580706009161953, "grad_norm": 0.7159758806228638, "learning_rate": 0.00019613087218180186, "loss": 0.8382, "step": 61100 }, { "epoch": 2.06143896523848, "grad_norm": 0.7034109830856323, "learning_rate": 0.00019590631456031617, "loss": 0.8386, "step": 61200 }, { "epoch": 2.0648073295607654, "grad_norm": 0.648295521736145, "learning_rate": 0.0001956817569388305, "loss": 0.8353, "step": 61300 }, { "epoch": 2.0681756938830502, "grad_norm": 0.909184992313385, "learning_rate": 0.0001954571993173448, "loss": 0.8366, "step": 61400 }, { "epoch": 2.0715440582053355, "grad_norm": 0.6607633233070374, "learning_rate": 0.00019523264169585912, "loss": 0.8389, "step": 61500 }, { "epoch": 2.0749124225276208, "grad_norm": 0.6451642513275146, "learning_rate": 0.00019500808407437348, "loss": 0.8379, "step": 61600 }, { "epoch": 2.0782807868499056, "grad_norm": 0.6556397080421448, "learning_rate": 0.0001947835264528878, "loss": 0.837, "step": 61700 }, { "epoch": 2.081649151172191, "grad_norm": 0.6268017292022705, "learning_rate": 0.00019455896883140212, "loss": 0.8379, "step": 61800 }, { "epoch": 2.0850175154944757, "grad_norm": 0.6620351672172546, "learning_rate": 0.00019433441120991643, "loss": 0.8377, "step": 61900 }, { "epoch": 2.088385879816761, "grad_norm": 0.6334593296051025, "learning_rate": 0.0001941098535884308, "loss": 0.8367, "step": 62000 }, { "epoch": 2.091754244139046, "grad_norm": 0.6498568058013916, "learning_rate": 0.0001938852959669451, "loss": 0.836, "step": 62100 }, { "epoch": 2.095122608461331, "grad_norm": 0.6664881110191345, "learning_rate": 0.00019366073834545944, "loss": 0.8366, "step": 62200 }, { "epoch": 2.0984909727836163, "grad_norm": 0.6434110999107361, "learning_rate": 0.00019343618072397375, "loss": 0.8339, "step": 62300 }, { "epoch": 2.1018593371059016, "grad_norm": 0.6958553194999695, "learning_rate": 0.00019321162310248805, "loss": 0.8363, "step": 62400 }, { "epoch": 2.1052277014281864, "grad_norm": 0.7187645435333252, "learning_rate": 0.00019298706548100242, "loss": 0.8337, "step": 62500 }, { "epoch": 2.1085960657504716, "grad_norm": 0.6984855532646179, "learning_rate": 0.00019276250785951672, "loss": 0.8363, "step": 62600 }, { "epoch": 2.1119644300727565, "grad_norm": 0.7342332601547241, "learning_rate": 0.00019253795023803106, "loss": 0.8385, "step": 62700 }, { "epoch": 2.1153327943950417, "grad_norm": 0.7250052690505981, "learning_rate": 0.00019231339261654537, "loss": 0.8333, "step": 62800 }, { "epoch": 2.118701158717327, "grad_norm": 0.6984183192253113, "learning_rate": 0.00019208883499505973, "loss": 0.8343, "step": 62900 }, { "epoch": 2.122069523039612, "grad_norm": 1.0108722448349, "learning_rate": 0.00019186427737357404, "loss": 0.836, "step": 63000 }, { "epoch": 2.125437887361897, "grad_norm": 0.6725150942802429, "learning_rate": 0.00019163971975208837, "loss": 0.8372, "step": 63100 }, { "epoch": 2.1288062516841824, "grad_norm": 0.6805692911148071, "learning_rate": 0.0001914151621306027, "loss": 0.8311, "step": 63200 }, { "epoch": 2.132174616006467, "grad_norm": 0.9061579704284668, "learning_rate": 0.00019119060450911702, "loss": 0.8343, "step": 63300 }, { "epoch": 2.1355429803287524, "grad_norm": 0.6339781880378723, "learning_rate": 0.00019096604688763135, "loss": 0.8405, "step": 63400 }, { "epoch": 2.1389113446510373, "grad_norm": 0.6159859895706177, "learning_rate": 0.00019074148926614566, "loss": 0.8352, "step": 63500 }, { "epoch": 2.1422797089733225, "grad_norm": 0.6404457092285156, "learning_rate": 0.00019051693164466002, "loss": 0.835, "step": 63600 }, { "epoch": 2.145648073295608, "grad_norm": 0.619216799736023, "learning_rate": 0.00019029237402317433, "loss": 0.8323, "step": 63700 }, { "epoch": 2.1490164376178926, "grad_norm": 0.6672142744064331, "learning_rate": 0.00019006781640168867, "loss": 0.8318, "step": 63800 }, { "epoch": 2.152384801940178, "grad_norm": 0.6713272929191589, "learning_rate": 0.00018984325878020297, "loss": 0.835, "step": 63900 }, { "epoch": 2.155753166262463, "grad_norm": 0.6657119989395142, "learning_rate": 0.00018961870115871734, "loss": 0.8312, "step": 64000 }, { "epoch": 2.159121530584748, "grad_norm": 0.7010686993598938, "learning_rate": 0.00018939414353723164, "loss": 0.8334, "step": 64100 }, { "epoch": 2.1624898949070332, "grad_norm": 0.67507404088974, "learning_rate": 0.00018916958591574598, "loss": 0.8337, "step": 64200 }, { "epoch": 2.165858259229318, "grad_norm": 0.7008448243141174, "learning_rate": 0.0001889450282942603, "loss": 0.8334, "step": 64300 }, { "epoch": 2.1692266235516033, "grad_norm": 0.634978175163269, "learning_rate": 0.0001887204706727746, "loss": 0.8372, "step": 64400 }, { "epoch": 2.1725949878738886, "grad_norm": 0.7248919010162354, "learning_rate": 0.00018849591305128896, "loss": 0.8355, "step": 64500 }, { "epoch": 2.1759633521961734, "grad_norm": 0.6367628574371338, "learning_rate": 0.00018827135542980327, "loss": 0.8328, "step": 64600 }, { "epoch": 2.1793317165184587, "grad_norm": 0.712471604347229, "learning_rate": 0.0001880467978083176, "loss": 0.8296, "step": 64700 }, { "epoch": 2.1827000808407435, "grad_norm": 0.671159565448761, "learning_rate": 0.0001878222401868319, "loss": 0.8339, "step": 64800 }, { "epoch": 2.186068445163029, "grad_norm": 0.6872281432151794, "learning_rate": 0.00018759768256534627, "loss": 0.8313, "step": 64900 }, { "epoch": 2.189436809485314, "grad_norm": 0.7017882466316223, "learning_rate": 0.00018737312494386058, "loss": 0.8327, "step": 65000 }, { "epoch": 2.192805173807599, "grad_norm": 0.7563288807868958, "learning_rate": 0.00018714856732237491, "loss": 0.8342, "step": 65100 }, { "epoch": 2.196173538129884, "grad_norm": 0.6648709177970886, "learning_rate": 0.00018692400970088922, "loss": 0.8318, "step": 65200 }, { "epoch": 2.1995419024521694, "grad_norm": 0.6673408150672913, "learning_rate": 0.00018669945207940356, "loss": 0.8328, "step": 65300 }, { "epoch": 2.2029102667744542, "grad_norm": 0.8705912232398987, "learning_rate": 0.0001864748944579179, "loss": 0.831, "step": 65400 }, { "epoch": 2.2062786310967395, "grad_norm": 0.6253674030303955, "learning_rate": 0.0001862503368364322, "loss": 0.8322, "step": 65500 }, { "epoch": 2.2096469954190243, "grad_norm": 0.7175179719924927, "learning_rate": 0.00018602577921494654, "loss": 0.8316, "step": 65600 }, { "epoch": 2.2130153597413096, "grad_norm": 0.6952610611915588, "learning_rate": 0.00018580122159346087, "loss": 0.831, "step": 65700 }, { "epoch": 2.216383724063595, "grad_norm": 0.6487058401107788, "learning_rate": 0.0001855766639719752, "loss": 0.8323, "step": 65800 }, { "epoch": 2.2197520883858797, "grad_norm": 0.7087188959121704, "learning_rate": 0.00018535210635048951, "loss": 0.8337, "step": 65900 }, { "epoch": 2.223120452708165, "grad_norm": 0.689513087272644, "learning_rate": 0.00018512754872900385, "loss": 0.834, "step": 66000 }, { "epoch": 2.22648881703045, "grad_norm": 1.3590748310089111, "learning_rate": 0.00018490299110751818, "loss": 0.8328, "step": 66100 }, { "epoch": 2.229857181352735, "grad_norm": 0.640274167060852, "learning_rate": 0.0001846784334860325, "loss": 0.8333, "step": 66200 }, { "epoch": 2.2332255456750203, "grad_norm": 0.794620156288147, "learning_rate": 0.00018445387586454683, "loss": 0.8343, "step": 66300 }, { "epoch": 2.236593909997305, "grad_norm": 0.6511592268943787, "learning_rate": 0.00018422931824306114, "loss": 0.8328, "step": 66400 }, { "epoch": 2.2399622743195904, "grad_norm": 0.6505751609802246, "learning_rate": 0.0001840047606215755, "loss": 0.8302, "step": 66500 }, { "epoch": 2.2433306386418757, "grad_norm": 0.7967960834503174, "learning_rate": 0.0001837802030000898, "loss": 0.83, "step": 66600 }, { "epoch": 2.2466990029641605, "grad_norm": 0.708914577960968, "learning_rate": 0.00018355564537860414, "loss": 0.8305, "step": 66700 }, { "epoch": 2.2500673672864457, "grad_norm": 0.6149790287017822, "learning_rate": 0.00018333108775711845, "loss": 0.8334, "step": 66800 }, { "epoch": 2.2534357316087306, "grad_norm": 0.6834396719932556, "learning_rate": 0.0001831065301356328, "loss": 0.8314, "step": 66900 }, { "epoch": 2.256804095931016, "grad_norm": 0.7162107229232788, "learning_rate": 0.00018288197251414712, "loss": 0.8327, "step": 67000 }, { "epoch": 2.260172460253301, "grad_norm": 0.6464916467666626, "learning_rate": 0.00018265741489266143, "loss": 0.8311, "step": 67100 }, { "epoch": 2.263540824575586, "grad_norm": 0.680050253868103, "learning_rate": 0.00018243285727117576, "loss": 0.8293, "step": 67200 }, { "epoch": 2.266909188897871, "grad_norm": 0.7750843167304993, "learning_rate": 0.00018220829964969007, "loss": 0.8286, "step": 67300 }, { "epoch": 2.2702775532201565, "grad_norm": 0.6611768007278442, "learning_rate": 0.00018198374202820443, "loss": 0.8311, "step": 67400 }, { "epoch": 2.2736459175424413, "grad_norm": 0.7073565125465393, "learning_rate": 0.00018175918440671874, "loss": 0.829, "step": 67500 }, { "epoch": 2.2770142818647265, "grad_norm": 0.6587579846382141, "learning_rate": 0.00018153462678523308, "loss": 0.8301, "step": 67600 }, { "epoch": 2.280382646187012, "grad_norm": 0.6691922545433044, "learning_rate": 0.00018131006916374739, "loss": 0.8291, "step": 67700 }, { "epoch": 2.2837510105092966, "grad_norm": 0.6645218133926392, "learning_rate": 0.00018108551154226175, "loss": 0.8294, "step": 67800 }, { "epoch": 2.287119374831582, "grad_norm": 0.6749672293663025, "learning_rate": 0.00018086095392077606, "loss": 0.8324, "step": 67900 }, { "epoch": 2.2904877391538667, "grad_norm": 0.6658663153648376, "learning_rate": 0.0001806363962992904, "loss": 0.8292, "step": 68000 }, { "epoch": 2.293856103476152, "grad_norm": 0.7315119504928589, "learning_rate": 0.0001804118386778047, "loss": 0.8289, "step": 68100 }, { "epoch": 2.2972244677984373, "grad_norm": 0.6700948476791382, "learning_rate": 0.00018018728105631903, "loss": 0.8266, "step": 68200 }, { "epoch": 2.300592832120722, "grad_norm": 0.6325820088386536, "learning_rate": 0.00017996272343483337, "loss": 0.8312, "step": 68300 }, { "epoch": 2.3039611964430073, "grad_norm": 0.6814998984336853, "learning_rate": 0.00017973816581334768, "loss": 0.8288, "step": 68400 }, { "epoch": 2.307329560765292, "grad_norm": 0.6832602024078369, "learning_rate": 0.000179513608191862, "loss": 0.8278, "step": 68500 }, { "epoch": 2.3106979250875774, "grad_norm": 0.7399811148643494, "learning_rate": 0.00017928905057037635, "loss": 0.8313, "step": 68600 }, { "epoch": 2.3140662894098627, "grad_norm": 0.6988467574119568, "learning_rate": 0.00017906449294889068, "loss": 0.8298, "step": 68700 }, { "epoch": 2.3174346537321475, "grad_norm": 0.6190904974937439, "learning_rate": 0.000178839935327405, "loss": 0.8307, "step": 68800 }, { "epoch": 2.320803018054433, "grad_norm": 0.6427486538887024, "learning_rate": 0.00017861537770591933, "loss": 0.8309, "step": 68900 }, { "epoch": 2.324171382376718, "grad_norm": 0.661953330039978, "learning_rate": 0.00017839082008443366, "loss": 0.8291, "step": 69000 }, { "epoch": 2.327539746699003, "grad_norm": 0.6629980206489563, "learning_rate": 0.00017816626246294797, "loss": 0.8268, "step": 69100 }, { "epoch": 2.330908111021288, "grad_norm": 0.8455718755722046, "learning_rate": 0.0001779417048414623, "loss": 0.8281, "step": 69200 }, { "epoch": 2.3342764753435734, "grad_norm": 0.6195480227470398, "learning_rate": 0.0001777171472199766, "loss": 0.8271, "step": 69300 }, { "epoch": 2.3376448396658582, "grad_norm": 0.6320804357528687, "learning_rate": 0.00017749258959849098, "loss": 0.8272, "step": 69400 }, { "epoch": 2.3410132039881435, "grad_norm": 0.6865086555480957, "learning_rate": 0.00017726803197700528, "loss": 0.83, "step": 69500 }, { "epoch": 2.3443815683104283, "grad_norm": 0.6443759799003601, "learning_rate": 0.00017704347435551962, "loss": 0.828, "step": 69600 }, { "epoch": 2.3477499326327136, "grad_norm": 0.6651887893676758, "learning_rate": 0.00017681891673403393, "loss": 0.8309, "step": 69700 }, { "epoch": 2.351118296954999, "grad_norm": 0.6643931269645691, "learning_rate": 0.0001765943591125483, "loss": 0.8285, "step": 69800 }, { "epoch": 2.3544866612772837, "grad_norm": 0.6697126030921936, "learning_rate": 0.0001763698014910626, "loss": 0.8264, "step": 69900 }, { "epoch": 2.357855025599569, "grad_norm": 0.6785570979118347, "learning_rate": 0.0001761452438695769, "loss": 0.8255, "step": 70000 }, { "epoch": 2.3612233899218538, "grad_norm": 0.6412234306335449, "learning_rate": 0.00017592068624809124, "loss": 0.8271, "step": 70100 }, { "epoch": 2.364591754244139, "grad_norm": 0.7083507180213928, "learning_rate": 0.00017569612862660555, "loss": 0.83, "step": 70200 }, { "epoch": 2.3679601185664243, "grad_norm": 0.6519679427146912, "learning_rate": 0.0001754715710051199, "loss": 0.8249, "step": 70300 }, { "epoch": 2.371328482888709, "grad_norm": 0.7082831263542175, "learning_rate": 0.00017524701338363422, "loss": 0.8287, "step": 70400 }, { "epoch": 2.3746968472109944, "grad_norm": 0.6893306374549866, "learning_rate": 0.00017502245576214855, "loss": 0.8241, "step": 70500 }, { "epoch": 2.378065211533279, "grad_norm": 0.6624453663825989, "learning_rate": 0.00017479789814066286, "loss": 0.8281, "step": 70600 }, { "epoch": 2.3814335758555645, "grad_norm": 0.7071661353111267, "learning_rate": 0.00017457334051917722, "loss": 0.8257, "step": 70700 }, { "epoch": 2.3848019401778497, "grad_norm": 0.6531967520713806, "learning_rate": 0.00017434878289769153, "loss": 0.8253, "step": 70800 }, { "epoch": 2.3881703045001346, "grad_norm": 0.6394172310829163, "learning_rate": 0.00017412422527620587, "loss": 0.8291, "step": 70900 }, { "epoch": 2.39153866882242, "grad_norm": 0.7370265126228333, "learning_rate": 0.00017389966765472018, "loss": 0.8282, "step": 71000 }, { "epoch": 2.394907033144705, "grad_norm": 0.6256112456321716, "learning_rate": 0.0001736751100332345, "loss": 0.8267, "step": 71100 }, { "epoch": 2.39827539746699, "grad_norm": 0.6319020390510559, "learning_rate": 0.00017345055241174885, "loss": 0.8271, "step": 71200 }, { "epoch": 2.401643761789275, "grad_norm": 0.638664186000824, "learning_rate": 0.00017322599479026315, "loss": 0.8259, "step": 71300 }, { "epoch": 2.4050121261115605, "grad_norm": 0.6793828010559082, "learning_rate": 0.0001730014371687775, "loss": 0.8265, "step": 71400 }, { "epoch": 2.4083804904338453, "grad_norm": 0.7026681303977966, "learning_rate": 0.00017277687954729182, "loss": 0.8296, "step": 71500 }, { "epoch": 2.4117488547561305, "grad_norm": 0.6481872200965881, "learning_rate": 0.00017255232192580616, "loss": 0.826, "step": 71600 }, { "epoch": 2.4151172190784154, "grad_norm": 0.68873530626297, "learning_rate": 0.00017232776430432047, "loss": 0.8283, "step": 71700 }, { "epoch": 2.4184855834007006, "grad_norm": 0.6869419813156128, "learning_rate": 0.0001721032066828348, "loss": 0.8251, "step": 71800 }, { "epoch": 2.421853947722986, "grad_norm": 0.6462306380271912, "learning_rate": 0.00017187864906134914, "loss": 0.8265, "step": 71900 }, { "epoch": 2.4252223120452707, "grad_norm": 0.6632818579673767, "learning_rate": 0.00017165409143986345, "loss": 0.8287, "step": 72000 }, { "epoch": 2.428590676367556, "grad_norm": 0.7588053345680237, "learning_rate": 0.00017142953381837778, "loss": 0.8278, "step": 72100 }, { "epoch": 2.431959040689841, "grad_norm": 0.6804139614105225, "learning_rate": 0.0001712049761968921, "loss": 0.8266, "step": 72200 }, { "epoch": 2.435327405012126, "grad_norm": 0.8674142956733704, "learning_rate": 0.00017098041857540645, "loss": 0.8242, "step": 72300 }, { "epoch": 2.4386957693344113, "grad_norm": 1.0071535110473633, "learning_rate": 0.00017075586095392076, "loss": 0.8294, "step": 72400 }, { "epoch": 2.442064133656696, "grad_norm": 0.6555808782577515, "learning_rate": 0.0001705313033324351, "loss": 0.8256, "step": 72500 }, { "epoch": 2.4454324979789814, "grad_norm": 0.6730812788009644, "learning_rate": 0.0001703067457109494, "loss": 0.8254, "step": 72600 }, { "epoch": 2.4488008623012663, "grad_norm": 0.6546606421470642, "learning_rate": 0.00017008218808946377, "loss": 0.8257, "step": 72700 }, { "epoch": 2.4521692266235515, "grad_norm": 0.6847641468048096, "learning_rate": 0.00016985763046797807, "loss": 0.824, "step": 72800 }, { "epoch": 2.455537590945837, "grad_norm": 0.6383925080299377, "learning_rate": 0.00016963307284649238, "loss": 0.8238, "step": 72900 }, { "epoch": 2.4589059552681216, "grad_norm": 0.6601071357727051, "learning_rate": 0.00016940851522500672, "loss": 0.8215, "step": 73000 }, { "epoch": 2.462274319590407, "grad_norm": 0.6254110336303711, "learning_rate": 0.00016918395760352102, "loss": 0.8254, "step": 73100 }, { "epoch": 2.465642683912692, "grad_norm": 0.6367729902267456, "learning_rate": 0.0001689593999820354, "loss": 0.8201, "step": 73200 }, { "epoch": 2.469011048234977, "grad_norm": 0.6448660492897034, "learning_rate": 0.0001687348423605497, "loss": 0.8237, "step": 73300 }, { "epoch": 2.4723794125572622, "grad_norm": 0.6661122441291809, "learning_rate": 0.00016851028473906403, "loss": 0.8257, "step": 73400 }, { "epoch": 2.4757477768795475, "grad_norm": 0.619644820690155, "learning_rate": 0.00016828572711757834, "loss": 0.8238, "step": 73500 }, { "epoch": 2.4791161412018323, "grad_norm": 0.7607592344284058, "learning_rate": 0.0001680611694960927, "loss": 0.8225, "step": 73600 }, { "epoch": 2.4824845055241176, "grad_norm": 0.6733140349388123, "learning_rate": 0.000167836611874607, "loss": 0.8212, "step": 73700 }, { "epoch": 2.4858528698464024, "grad_norm": 0.9612058401107788, "learning_rate": 0.00016761205425312134, "loss": 0.8214, "step": 73800 }, { "epoch": 2.4892212341686877, "grad_norm": 0.618076503276825, "learning_rate": 0.00016738749663163565, "loss": 0.8231, "step": 73900 }, { "epoch": 2.492589598490973, "grad_norm": 0.6606545448303223, "learning_rate": 0.00016716293901015, "loss": 0.8235, "step": 74000 }, { "epoch": 2.4959579628132578, "grad_norm": 0.6311343312263489, "learning_rate": 0.00016693838138866432, "loss": 0.8227, "step": 74100 }, { "epoch": 2.499326327135543, "grad_norm": 0.713347315788269, "learning_rate": 0.00016671382376717863, "loss": 0.825, "step": 74200 }, { "epoch": 2.502694691457828, "grad_norm": 0.658001720905304, "learning_rate": 0.00016648926614569297, "loss": 0.8227, "step": 74300 }, { "epoch": 2.506063055780113, "grad_norm": 0.7130460739135742, "learning_rate": 0.0001662647085242073, "loss": 0.823, "step": 74400 }, { "epoch": 2.5094314201023984, "grad_norm": 0.7092128992080688, "learning_rate": 0.00016604015090272164, "loss": 0.8271, "step": 74500 }, { "epoch": 2.512799784424683, "grad_norm": 0.6681484580039978, "learning_rate": 0.00016581559328123594, "loss": 0.8208, "step": 74600 }, { "epoch": 2.5161681487469685, "grad_norm": 0.649299144744873, "learning_rate": 0.00016559103565975028, "loss": 0.8229, "step": 74700 }, { "epoch": 2.5195365130692533, "grad_norm": 0.6099591255187988, "learning_rate": 0.00016536647803826461, "loss": 0.8221, "step": 74800 }, { "epoch": 2.5229048773915386, "grad_norm": 0.655128538608551, "learning_rate": 0.00016514192041677892, "loss": 0.8216, "step": 74900 }, { "epoch": 2.526273241713824, "grad_norm": 2.5822091102600098, "learning_rate": 0.00016491736279529326, "loss": 0.8213, "step": 75000 }, { "epoch": 2.529641606036109, "grad_norm": 0.6650177240371704, "learning_rate": 0.00016469280517380757, "loss": 0.8279, "step": 75100 }, { "epoch": 2.533009970358394, "grad_norm": 0.9380492568016052, "learning_rate": 0.00016446824755232193, "loss": 0.8245, "step": 75200 }, { "epoch": 2.536378334680679, "grad_norm": 0.6475105881690979, "learning_rate": 0.00016424368993083624, "loss": 0.8233, "step": 75300 }, { "epoch": 2.539746699002964, "grad_norm": 0.6452984809875488, "learning_rate": 0.00016401913230935057, "loss": 0.8218, "step": 75400 }, { "epoch": 2.5431150633252493, "grad_norm": 1.478945255279541, "learning_rate": 0.00016379457468786488, "loss": 0.825, "step": 75500 }, { "epoch": 2.5464834276475345, "grad_norm": 0.6364376544952393, "learning_rate": 0.00016357001706637924, "loss": 0.8215, "step": 75600 }, { "epoch": 2.5498517919698194, "grad_norm": 0.5982120633125305, "learning_rate": 0.00016334545944489355, "loss": 0.8244, "step": 75700 }, { "epoch": 2.5532201562921046, "grad_norm": 0.641855001449585, "learning_rate": 0.00016312090182340786, "loss": 0.8238, "step": 75800 }, { "epoch": 2.5565885206143895, "grad_norm": 0.6564063429832458, "learning_rate": 0.0001628963442019222, "loss": 0.8228, "step": 75900 }, { "epoch": 2.5599568849366747, "grad_norm": 0.6515690684318542, "learning_rate": 0.0001626717865804365, "loss": 0.825, "step": 76000 }, { "epoch": 2.56332524925896, "grad_norm": 0.7063090205192566, "learning_rate": 0.00016244722895895086, "loss": 0.8232, "step": 76100 }, { "epoch": 2.566693613581245, "grad_norm": 1.018594741821289, "learning_rate": 0.00016222267133746517, "loss": 0.8219, "step": 76200 }, { "epoch": 2.57006197790353, "grad_norm": 0.7534065842628479, "learning_rate": 0.0001619981137159795, "loss": 0.8208, "step": 76300 }, { "epoch": 2.573430342225815, "grad_norm": 0.6062216758728027, "learning_rate": 0.00016177355609449382, "loss": 0.8224, "step": 76400 }, { "epoch": 2.5767987065481, "grad_norm": 0.6766044497489929, "learning_rate": 0.00016154899847300818, "loss": 0.823, "step": 76500 }, { "epoch": 2.5801670708703854, "grad_norm": 0.668302595615387, "learning_rate": 0.00016132444085152249, "loss": 0.8211, "step": 76600 }, { "epoch": 2.5835354351926707, "grad_norm": 0.8911457061767578, "learning_rate": 0.0001610998832300368, "loss": 0.8212, "step": 76700 }, { "epoch": 2.5869037995149555, "grad_norm": 0.718221127986908, "learning_rate": 0.00016087532560855113, "loss": 0.8228, "step": 76800 }, { "epoch": 2.590272163837241, "grad_norm": 0.7011673450469971, "learning_rate": 0.00016065076798706546, "loss": 0.8216, "step": 76900 }, { "epoch": 2.5936405281595256, "grad_norm": 0.6552968621253967, "learning_rate": 0.0001604262103655798, "loss": 0.8216, "step": 77000 }, { "epoch": 2.597008892481811, "grad_norm": 0.6576195955276489, "learning_rate": 0.0001602016527440941, "loss": 0.8197, "step": 77100 }, { "epoch": 2.600377256804096, "grad_norm": 0.6853031516075134, "learning_rate": 0.00015997709512260847, "loss": 0.8186, "step": 77200 }, { "epoch": 2.603745621126381, "grad_norm": 0.631230890750885, "learning_rate": 0.00015975253750112278, "loss": 0.8212, "step": 77300 }, { "epoch": 2.6071139854486662, "grad_norm": 0.6358488202095032, "learning_rate": 0.0001595279798796371, "loss": 0.8221, "step": 77400 }, { "epoch": 2.610482349770951, "grad_norm": 1.2744354009628296, "learning_rate": 0.00015930342225815142, "loss": 0.8214, "step": 77500 }, { "epoch": 2.6138507140932363, "grad_norm": 0.9848027229309082, "learning_rate": 0.00015907886463666578, "loss": 0.8238, "step": 77600 }, { "epoch": 2.6172190784155216, "grad_norm": 0.7024006843566895, "learning_rate": 0.0001588543070151801, "loss": 0.8203, "step": 77700 }, { "epoch": 2.6205874427378064, "grad_norm": 0.628036379814148, "learning_rate": 0.0001586297493936944, "loss": 0.8214, "step": 77800 }, { "epoch": 2.6239558070600917, "grad_norm": 1.1920697689056396, "learning_rate": 0.00015840519177220873, "loss": 0.8203, "step": 77900 }, { "epoch": 2.6273241713823765, "grad_norm": 0.6503965258598328, "learning_rate": 0.00015818063415072304, "loss": 0.8213, "step": 78000 }, { "epoch": 2.6306925357046618, "grad_norm": 0.6561105847358704, "learning_rate": 0.0001579560765292374, "loss": 0.8215, "step": 78100 }, { "epoch": 2.634060900026947, "grad_norm": 0.9598469734191895, "learning_rate": 0.0001577315189077517, "loss": 0.8189, "step": 78200 }, { "epoch": 2.637429264349232, "grad_norm": 0.6700890064239502, "learning_rate": 0.00015750696128626605, "loss": 0.8217, "step": 78300 }, { "epoch": 2.640797628671517, "grad_norm": 0.6581085920333862, "learning_rate": 0.00015728240366478036, "loss": 0.8207, "step": 78400 }, { "epoch": 2.644165992993802, "grad_norm": 0.6442667841911316, "learning_rate": 0.00015705784604329472, "loss": 0.8198, "step": 78500 }, { "epoch": 2.647534357316087, "grad_norm": 1.1659986972808838, "learning_rate": 0.00015683328842180903, "loss": 0.8232, "step": 78600 }, { "epoch": 2.6509027216383725, "grad_norm": 0.6638743877410889, "learning_rate": 0.00015660873080032333, "loss": 0.8184, "step": 78700 }, { "epoch": 2.6542710859606578, "grad_norm": 0.649681806564331, "learning_rate": 0.00015638417317883767, "loss": 0.8195, "step": 78800 }, { "epoch": 2.6576394502829426, "grad_norm": 1.0518876314163208, "learning_rate": 0.00015615961555735198, "loss": 0.8222, "step": 78900 }, { "epoch": 2.661007814605228, "grad_norm": 0.6661698222160339, "learning_rate": 0.00015593505793586634, "loss": 0.8192, "step": 79000 }, { "epoch": 2.6643761789275127, "grad_norm": 0.6558882594108582, "learning_rate": 0.00015571050031438065, "loss": 0.8151, "step": 79100 }, { "epoch": 2.667744543249798, "grad_norm": 0.621672511100769, "learning_rate": 0.00015548594269289498, "loss": 0.8185, "step": 79200 }, { "epoch": 2.671112907572083, "grad_norm": 0.6606272459030151, "learning_rate": 0.00015526138507140932, "loss": 0.8178, "step": 79300 }, { "epoch": 2.674481271894368, "grad_norm": 0.685043215751648, "learning_rate": 0.00015503682744992365, "loss": 0.8199, "step": 79400 }, { "epoch": 2.6778496362166533, "grad_norm": 0.6863681077957153, "learning_rate": 0.00015481226982843796, "loss": 0.8209, "step": 79500 }, { "epoch": 2.681218000538938, "grad_norm": 0.6480740308761597, "learning_rate": 0.00015458771220695227, "loss": 0.8209, "step": 79600 }, { "epoch": 2.6845863648612234, "grad_norm": 0.6727776527404785, "learning_rate": 0.00015436315458546663, "loss": 0.8207, "step": 79700 }, { "epoch": 2.6879547291835086, "grad_norm": 0.7330679893493652, "learning_rate": 0.00015413859696398094, "loss": 0.8217, "step": 79800 }, { "epoch": 2.6913230935057935, "grad_norm": 0.7128227949142456, "learning_rate": 0.00015391403934249528, "loss": 0.8229, "step": 79900 }, { "epoch": 2.6946914578280787, "grad_norm": 0.6492688059806824, "learning_rate": 0.00015368948172100958, "loss": 0.8162, "step": 80000 }, { "epoch": 2.6980598221503636, "grad_norm": 0.8806473016738892, "learning_rate": 0.00015346492409952395, "loss": 0.8227, "step": 80100 }, { "epoch": 2.701428186472649, "grad_norm": 0.8799885511398315, "learning_rate": 0.00015324036647803825, "loss": 0.8165, "step": 80200 }, { "epoch": 2.704796550794934, "grad_norm": 0.6979735493659973, "learning_rate": 0.0001530158088565526, "loss": 0.8198, "step": 80300 }, { "epoch": 2.7081649151172194, "grad_norm": 0.6624419689178467, "learning_rate": 0.0001527912512350669, "loss": 0.8194, "step": 80400 }, { "epoch": 2.711533279439504, "grad_norm": 0.7164821624755859, "learning_rate": 0.00015256669361358126, "loss": 0.8193, "step": 80500 }, { "epoch": 2.7149016437617894, "grad_norm": 0.6688589453697205, "learning_rate": 0.00015234213599209557, "loss": 0.8156, "step": 80600 }, { "epoch": 2.7182700080840743, "grad_norm": 0.6661184430122375, "learning_rate": 0.00015211757837060988, "loss": 0.8199, "step": 80700 }, { "epoch": 2.7216383724063595, "grad_norm": 0.6526447534561157, "learning_rate": 0.0001518930207491242, "loss": 0.8187, "step": 80800 }, { "epoch": 2.725006736728645, "grad_norm": 0.6510984897613525, "learning_rate": 0.00015166846312763852, "loss": 0.8189, "step": 80900 }, { "epoch": 2.7283751010509296, "grad_norm": 0.6921165585517883, "learning_rate": 0.00015144390550615288, "loss": 0.8182, "step": 81000 }, { "epoch": 2.731743465373215, "grad_norm": 0.7041354179382324, "learning_rate": 0.0001512193478846672, "loss": 0.8208, "step": 81100 }, { "epoch": 2.7351118296954997, "grad_norm": 0.639445424079895, "learning_rate": 0.00015099479026318152, "loss": 0.8202, "step": 81200 }, { "epoch": 2.738480194017785, "grad_norm": 0.7262235879898071, "learning_rate": 0.00015077023264169583, "loss": 0.8184, "step": 81300 }, { "epoch": 2.7418485583400702, "grad_norm": 0.6470584273338318, "learning_rate": 0.0001505456750202102, "loss": 0.8199, "step": 81400 }, { "epoch": 2.745216922662355, "grad_norm": 0.6758275628089905, "learning_rate": 0.0001503211173987245, "loss": 0.817, "step": 81500 }, { "epoch": 2.7485852869846403, "grad_norm": 0.6550074815750122, "learning_rate": 0.0001500965597772388, "loss": 0.82, "step": 81600 }, { "epoch": 2.751953651306925, "grad_norm": 0.6312419176101685, "learning_rate": 0.00014987200215575315, "loss": 0.8209, "step": 81700 }, { "epoch": 2.7553220156292104, "grad_norm": 0.6456059813499451, "learning_rate": 0.00014964744453426748, "loss": 0.8173, "step": 81800 }, { "epoch": 2.7586903799514957, "grad_norm": 0.6251012086868286, "learning_rate": 0.00014942288691278182, "loss": 0.8147, "step": 81900 }, { "epoch": 2.7620587442737805, "grad_norm": 0.6424401998519897, "learning_rate": 0.00014919832929129613, "loss": 0.8173, "step": 82000 }, { "epoch": 2.7654271085960658, "grad_norm": 0.7199423313140869, "learning_rate": 0.00014897377166981046, "loss": 0.8163, "step": 82100 }, { "epoch": 2.7687954729183506, "grad_norm": 0.6573197841644287, "learning_rate": 0.0001487492140483248, "loss": 0.8145, "step": 82200 }, { "epoch": 2.772163837240636, "grad_norm": 0.6857665777206421, "learning_rate": 0.0001485246564268391, "loss": 0.8188, "step": 82300 }, { "epoch": 2.775532201562921, "grad_norm": 0.6398062705993652, "learning_rate": 0.00014830009880535344, "loss": 0.8173, "step": 82400 }, { "epoch": 2.7789005658852064, "grad_norm": 0.6743867993354797, "learning_rate": 0.00014807554118386777, "loss": 0.8176, "step": 82500 }, { "epoch": 2.782268930207491, "grad_norm": 0.6104719042778015, "learning_rate": 0.0001478509835623821, "loss": 0.8171, "step": 82600 }, { "epoch": 2.7856372945297765, "grad_norm": 0.6518858075141907, "learning_rate": 0.00014762642594089642, "loss": 0.8203, "step": 82700 }, { "epoch": 2.7890056588520613, "grad_norm": 0.7392122745513916, "learning_rate": 0.00014740186831941075, "loss": 0.8171, "step": 82800 }, { "epoch": 2.7923740231743466, "grad_norm": 0.6652575135231018, "learning_rate": 0.0001471773106979251, "loss": 0.8155, "step": 82900 }, { "epoch": 2.795742387496632, "grad_norm": 0.698665201663971, "learning_rate": 0.0001469527530764394, "loss": 0.8202, "step": 83000 }, { "epoch": 2.7991107518189167, "grad_norm": 0.6012236475944519, "learning_rate": 0.00014672819545495373, "loss": 0.817, "step": 83100 }, { "epoch": 2.802479116141202, "grad_norm": 0.6791641116142273, "learning_rate": 0.00014650363783346804, "loss": 0.8185, "step": 83200 }, { "epoch": 2.8058474804634868, "grad_norm": 0.694733738899231, "learning_rate": 0.00014627908021198237, "loss": 0.8161, "step": 83300 }, { "epoch": 2.809215844785772, "grad_norm": 1.395378589630127, "learning_rate": 0.0001460545225904967, "loss": 0.8136, "step": 83400 }, { "epoch": 2.8125842091080573, "grad_norm": 1.1096270084381104, "learning_rate": 0.00014582996496901104, "loss": 0.8187, "step": 83500 }, { "epoch": 2.815952573430342, "grad_norm": 0.6426212787628174, "learning_rate": 0.00014560540734752535, "loss": 0.817, "step": 83600 }, { "epoch": 2.8193209377526274, "grad_norm": 0.6201661229133606, "learning_rate": 0.0001453808497260397, "loss": 0.8166, "step": 83700 }, { "epoch": 2.822689302074912, "grad_norm": 0.6646463871002197, "learning_rate": 0.00014515629210455402, "loss": 0.82, "step": 83800 }, { "epoch": 2.8260576663971975, "grad_norm": 0.6990267038345337, "learning_rate": 0.00014493173448306833, "loss": 0.8169, "step": 83900 }, { "epoch": 2.8294260307194827, "grad_norm": 0.671810507774353, "learning_rate": 0.00014470717686158267, "loss": 0.8167, "step": 84000 }, { "epoch": 2.8327943950417676, "grad_norm": 0.686876118183136, "learning_rate": 0.000144482619240097, "loss": 0.8137, "step": 84100 }, { "epoch": 2.836162759364053, "grad_norm": 0.6434644460678101, "learning_rate": 0.0001442580616186113, "loss": 0.8157, "step": 84200 }, { "epoch": 2.8395311236863376, "grad_norm": 0.652999758720398, "learning_rate": 0.00014403350399712564, "loss": 0.8121, "step": 84300 }, { "epoch": 2.842899488008623, "grad_norm": 0.6468531489372253, "learning_rate": 0.00014380894637563998, "loss": 0.8175, "step": 84400 }, { "epoch": 2.846267852330908, "grad_norm": 0.6018803119659424, "learning_rate": 0.00014358438875415432, "loss": 0.816, "step": 84500 }, { "epoch": 2.8496362166531934, "grad_norm": 0.6775005459785461, "learning_rate": 0.00014335983113266862, "loss": 0.8128, "step": 84600 }, { "epoch": 2.8530045809754783, "grad_norm": 0.664910078048706, "learning_rate": 0.00014313527351118296, "loss": 0.8172, "step": 84700 }, { "epoch": 2.8563729452977635, "grad_norm": 0.6959900259971619, "learning_rate": 0.0001429107158896973, "loss": 0.8135, "step": 84800 }, { "epoch": 2.8597413096200484, "grad_norm": 0.6213033199310303, "learning_rate": 0.0001426861582682116, "loss": 0.8153, "step": 84900 }, { "epoch": 2.8631096739423336, "grad_norm": 0.6437749266624451, "learning_rate": 0.00014246160064672594, "loss": 0.8114, "step": 85000 }, { "epoch": 2.866478038264619, "grad_norm": 0.6763966083526611, "learning_rate": 0.00014223704302524027, "loss": 0.8138, "step": 85100 }, { "epoch": 2.8698464025869037, "grad_norm": 1.10175621509552, "learning_rate": 0.00014201248540375458, "loss": 0.8153, "step": 85200 }, { "epoch": 2.873214766909189, "grad_norm": 0.6517946124076843, "learning_rate": 0.00014178792778226892, "loss": 0.8147, "step": 85300 }, { "epoch": 2.876583131231474, "grad_norm": 0.7234548926353455, "learning_rate": 0.00014156337016078325, "loss": 0.8182, "step": 85400 }, { "epoch": 2.879951495553759, "grad_norm": 0.6792501211166382, "learning_rate": 0.00014133881253929759, "loss": 0.8154, "step": 85500 }, { "epoch": 2.8833198598760443, "grad_norm": 0.6222261786460876, "learning_rate": 0.0001411142549178119, "loss": 0.816, "step": 85600 }, { "epoch": 2.886688224198329, "grad_norm": 0.6346508264541626, "learning_rate": 0.00014088969729632623, "loss": 0.8159, "step": 85700 }, { "epoch": 2.8900565885206144, "grad_norm": 0.7334688305854797, "learning_rate": 0.00014066513967484054, "loss": 0.8166, "step": 85800 }, { "epoch": 2.8934249528428992, "grad_norm": 0.6864719986915588, "learning_rate": 0.00014044058205335487, "loss": 0.819, "step": 85900 }, { "epoch": 2.8967933171651845, "grad_norm": 0.5927285552024841, "learning_rate": 0.0001402160244318692, "loss": 0.8148, "step": 86000 }, { "epoch": 2.90016168148747, "grad_norm": 0.660213828086853, "learning_rate": 0.00013999146681038354, "loss": 0.815, "step": 86100 }, { "epoch": 2.903530045809755, "grad_norm": 0.6887788772583008, "learning_rate": 0.00013976690918889785, "loss": 0.8143, "step": 86200 }, { "epoch": 2.90689841013204, "grad_norm": 1.1385151147842407, "learning_rate": 0.00013954235156741219, "loss": 0.8139, "step": 86300 }, { "epoch": 2.910266774454325, "grad_norm": 0.6464802026748657, "learning_rate": 0.00013931779394592652, "loss": 0.8149, "step": 86400 }, { "epoch": 2.91363513877661, "grad_norm": 0.6867853999137878, "learning_rate": 0.00013909323632444086, "loss": 0.8139, "step": 86500 }, { "epoch": 2.9170035030988952, "grad_norm": 0.8868036866188049, "learning_rate": 0.00013886867870295516, "loss": 0.8139, "step": 86600 }, { "epoch": 2.9203718674211805, "grad_norm": 0.6516538262367249, "learning_rate": 0.0001386441210814695, "loss": 0.8153, "step": 86700 }, { "epoch": 2.9237402317434653, "grad_norm": 0.6790093779563904, "learning_rate": 0.0001384195634599838, "loss": 0.8133, "step": 86800 }, { "epoch": 2.9271085960657506, "grad_norm": 0.6333130598068237, "learning_rate": 0.00013819500583849814, "loss": 0.8117, "step": 86900 }, { "epoch": 2.9304769603880354, "grad_norm": 0.7102107405662537, "learning_rate": 0.00013797044821701248, "loss": 0.8128, "step": 87000 }, { "epoch": 2.9338453247103207, "grad_norm": 0.7193422913551331, "learning_rate": 0.00013774589059552679, "loss": 0.8121, "step": 87100 }, { "epoch": 2.937213689032606, "grad_norm": 0.7304584980010986, "learning_rate": 0.00013752133297404112, "loss": 0.8114, "step": 87200 }, { "epoch": 2.9405820533548908, "grad_norm": 0.8924300074577332, "learning_rate": 0.00013729677535255546, "loss": 0.8137, "step": 87300 }, { "epoch": 2.943950417677176, "grad_norm": 0.6813507080078125, "learning_rate": 0.0001370722177310698, "loss": 0.8109, "step": 87400 }, { "epoch": 2.947318781999461, "grad_norm": 0.6427081823348999, "learning_rate": 0.0001368476601095841, "loss": 0.8125, "step": 87500 }, { "epoch": 2.950687146321746, "grad_norm": 0.6571387052536011, "learning_rate": 0.00013662310248809843, "loss": 0.8132, "step": 87600 }, { "epoch": 2.9540555106440314, "grad_norm": 0.7705689072608948, "learning_rate": 0.00013639854486661277, "loss": 0.8129, "step": 87700 }, { "epoch": 2.957423874966316, "grad_norm": 0.7075904011726379, "learning_rate": 0.00013617398724512708, "loss": 0.8143, "step": 87800 }, { "epoch": 2.9607922392886015, "grad_norm": 0.6926144957542419, "learning_rate": 0.0001359494296236414, "loss": 0.8115, "step": 87900 }, { "epoch": 2.9641606036108863, "grad_norm": 0.7183883190155029, "learning_rate": 0.00013572487200215575, "loss": 0.8122, "step": 88000 }, { "epoch": 2.9675289679331716, "grad_norm": 0.6963924169540405, "learning_rate": 0.00013550031438067006, "loss": 0.8151, "step": 88100 }, { "epoch": 2.970897332255457, "grad_norm": 0.6360912919044495, "learning_rate": 0.0001352757567591844, "loss": 0.8129, "step": 88200 }, { "epoch": 2.974265696577742, "grad_norm": 0.7141982316970825, "learning_rate": 0.00013505119913769873, "loss": 0.8142, "step": 88300 }, { "epoch": 2.977634060900027, "grad_norm": 0.6647577285766602, "learning_rate": 0.00013482664151621306, "loss": 0.811, "step": 88400 }, { "epoch": 2.981002425222312, "grad_norm": 0.6686524152755737, "learning_rate": 0.00013460208389472737, "loss": 0.8149, "step": 88500 }, { "epoch": 2.984370789544597, "grad_norm": 0.66321861743927, "learning_rate": 0.0001343775262732417, "loss": 0.8103, "step": 88600 }, { "epoch": 2.9877391538668823, "grad_norm": 0.622572660446167, "learning_rate": 0.00013415296865175601, "loss": 0.8144, "step": 88700 }, { "epoch": 2.9911075181891675, "grad_norm": 0.667248547077179, "learning_rate": 0.00013392841103027035, "loss": 0.8108, "step": 88800 }, { "epoch": 2.9944758825114524, "grad_norm": 0.6621103882789612, "learning_rate": 0.00013370385340878468, "loss": 0.8106, "step": 88900 }, { "epoch": 2.9978442468337376, "grad_norm": 0.6303636431694031, "learning_rate": 0.00013347929578729902, "loss": 0.8104, "step": 89000 }, { "epoch": 3.0, "eval_loss": 0.760505199432373, "eval_runtime": 1.8138, "eval_samples_per_second": 2756.665, "eval_steps_per_second": 43.555, "step": 89064 }, { "epoch": 3.0012126111560224, "grad_norm": 0.6279102563858032, "learning_rate": 0.00013325473816581333, "loss": 0.813, "step": 89100 }, { "epoch": 3.0045809754783077, "grad_norm": 0.6355727910995483, "learning_rate": 0.00013303018054432766, "loss": 0.811, "step": 89200 }, { "epoch": 3.007949339800593, "grad_norm": 0.6710761189460754, "learning_rate": 0.000132805622922842, "loss": 0.8147, "step": 89300 }, { "epoch": 3.011317704122878, "grad_norm": 0.6306372284889221, "learning_rate": 0.00013258106530135633, "loss": 0.8151, "step": 89400 }, { "epoch": 3.014686068445163, "grad_norm": 0.6803897023200989, "learning_rate": 0.00013235650767987064, "loss": 0.8086, "step": 89500 }, { "epoch": 3.0180544327674483, "grad_norm": 0.6367260217666626, "learning_rate": 0.00013213195005838498, "loss": 0.8119, "step": 89600 }, { "epoch": 3.021422797089733, "grad_norm": 0.664561927318573, "learning_rate": 0.00013190739243689928, "loss": 0.8094, "step": 89700 }, { "epoch": 3.0247911614120184, "grad_norm": 0.6986654996871948, "learning_rate": 0.00013168283481541362, "loss": 0.8116, "step": 89800 }, { "epoch": 3.0281595257343032, "grad_norm": 0.6583049893379211, "learning_rate": 0.00013145827719392795, "loss": 0.8148, "step": 89900 }, { "epoch": 3.0315278900565885, "grad_norm": 0.6716769337654114, "learning_rate": 0.00013123371957244226, "loss": 0.8137, "step": 90000 }, { "epoch": 3.034896254378874, "grad_norm": 0.7089846134185791, "learning_rate": 0.0001310091619509566, "loss": 0.809, "step": 90100 }, { "epoch": 3.0382646187011586, "grad_norm": 0.6873953938484192, "learning_rate": 0.00013078460432947093, "loss": 0.8113, "step": 90200 }, { "epoch": 3.041632983023444, "grad_norm": 0.6583080887794495, "learning_rate": 0.00013056004670798527, "loss": 0.8067, "step": 90300 }, { "epoch": 3.045001347345729, "grad_norm": 0.6893338561058044, "learning_rate": 0.00013033548908649958, "loss": 0.8126, "step": 90400 }, { "epoch": 3.048369711668014, "grad_norm": 0.9407336115837097, "learning_rate": 0.0001301109314650139, "loss": 0.8077, "step": 90500 }, { "epoch": 3.0517380759902992, "grad_norm": 1.1255403757095337, "learning_rate": 0.00012988637384352822, "loss": 0.8104, "step": 90600 }, { "epoch": 3.055106440312584, "grad_norm": 0.6687456369400024, "learning_rate": 0.00012966181622204255, "loss": 0.8122, "step": 90700 }, { "epoch": 3.0584748046348693, "grad_norm": 0.6100497245788574, "learning_rate": 0.0001294372586005569, "loss": 0.8098, "step": 90800 }, { "epoch": 3.0618431689571546, "grad_norm": 0.6621761918067932, "learning_rate": 0.00012921270097907123, "loss": 0.8118, "step": 90900 }, { "epoch": 3.0652115332794394, "grad_norm": 0.8076705932617188, "learning_rate": 0.00012898814335758553, "loss": 0.811, "step": 91000 }, { "epoch": 3.0685798976017247, "grad_norm": 0.6705955266952515, "learning_rate": 0.00012876358573609987, "loss": 0.8116, "step": 91100 }, { "epoch": 3.0719482619240095, "grad_norm": 0.6365945339202881, "learning_rate": 0.0001285390281146142, "loss": 0.8074, "step": 91200 }, { "epoch": 3.0753166262462948, "grad_norm": 0.9314165115356445, "learning_rate": 0.00012831447049312854, "loss": 0.8092, "step": 91300 }, { "epoch": 3.07868499056858, "grad_norm": 0.6733311414718628, "learning_rate": 0.00012808991287164285, "loss": 0.8102, "step": 91400 }, { "epoch": 3.082053354890865, "grad_norm": 0.6910605430603027, "learning_rate": 0.00012786535525015718, "loss": 0.8065, "step": 91500 }, { "epoch": 3.08542171921315, "grad_norm": 0.7043003439903259, "learning_rate": 0.0001276407976286715, "loss": 0.8073, "step": 91600 }, { "epoch": 3.0887900835354354, "grad_norm": 0.6764921545982361, "learning_rate": 0.00012741624000718583, "loss": 0.8089, "step": 91700 }, { "epoch": 3.09215844785772, "grad_norm": 0.6997144818305969, "learning_rate": 0.00012719168238570016, "loss": 0.8088, "step": 91800 }, { "epoch": 3.0955268121800055, "grad_norm": 0.6801837086677551, "learning_rate": 0.0001269671247642145, "loss": 0.8094, "step": 91900 }, { "epoch": 3.0988951765022903, "grad_norm": 0.6339113116264343, "learning_rate": 0.0001267425671427288, "loss": 0.8109, "step": 92000 }, { "epoch": 3.1022635408245756, "grad_norm": 0.6691506505012512, "learning_rate": 0.00012651800952124314, "loss": 0.8135, "step": 92100 }, { "epoch": 3.105631905146861, "grad_norm": 0.6617900133132935, "learning_rate": 0.00012629345189975747, "loss": 0.8106, "step": 92200 }, { "epoch": 3.1090002694691456, "grad_norm": 0.6737276911735535, "learning_rate": 0.0001260688942782718, "loss": 0.8128, "step": 92300 }, { "epoch": 3.112368633791431, "grad_norm": 0.670802652835846, "learning_rate": 0.00012584433665678612, "loss": 0.8106, "step": 92400 }, { "epoch": 3.115736998113716, "grad_norm": 0.626956045627594, "learning_rate": 0.00012561977903530045, "loss": 0.8114, "step": 92500 }, { "epoch": 3.119105362436001, "grad_norm": 0.6243528127670288, "learning_rate": 0.00012539522141381476, "loss": 0.8115, "step": 92600 }, { "epoch": 3.1224737267582863, "grad_norm": 0.6828027367591858, "learning_rate": 0.0001251706637923291, "loss": 0.807, "step": 92700 }, { "epoch": 3.125842091080571, "grad_norm": 0.704557478427887, "learning_rate": 0.00012494610617084343, "loss": 0.8085, "step": 92800 }, { "epoch": 3.1292104554028564, "grad_norm": 0.6188080310821533, "learning_rate": 0.00012472154854935774, "loss": 0.8079, "step": 92900 }, { "epoch": 3.1325788197251416, "grad_norm": 0.6640317440032959, "learning_rate": 0.00012449699092787207, "loss": 0.808, "step": 93000 }, { "epoch": 3.1359471840474265, "grad_norm": 0.6489530801773071, "learning_rate": 0.0001242724333063864, "loss": 0.8089, "step": 93100 }, { "epoch": 3.1393155483697117, "grad_norm": 0.6527587175369263, "learning_rate": 0.00012404787568490074, "loss": 0.809, "step": 93200 }, { "epoch": 3.1426839126919965, "grad_norm": 0.6802580952644348, "learning_rate": 0.00012382331806341508, "loss": 0.8099, "step": 93300 }, { "epoch": 3.146052277014282, "grad_norm": 0.6751042604446411, "learning_rate": 0.0001235987604419294, "loss": 0.8103, "step": 93400 }, { "epoch": 3.149420641336567, "grad_norm": 0.723804235458374, "learning_rate": 0.0001233742028204437, "loss": 0.8043, "step": 93500 }, { "epoch": 3.152789005658852, "grad_norm": 0.768860399723053, "learning_rate": 0.00012314964519895803, "loss": 0.8105, "step": 93600 }, { "epoch": 3.156157369981137, "grad_norm": 0.6464242339134216, "learning_rate": 0.00012292508757747237, "loss": 0.808, "step": 93700 }, { "epoch": 3.1595257343034224, "grad_norm": 0.7125059962272644, "learning_rate": 0.0001227005299559867, "loss": 0.8103, "step": 93800 }, { "epoch": 3.1628940986257073, "grad_norm": 0.6584749221801758, "learning_rate": 0.000122475972334501, "loss": 0.81, "step": 93900 }, { "epoch": 3.1662624629479925, "grad_norm": 0.6689501404762268, "learning_rate": 0.00012225141471301534, "loss": 0.8086, "step": 94000 }, { "epoch": 3.169630827270278, "grad_norm": 0.6383669376373291, "learning_rate": 0.00012202685709152968, "loss": 0.8097, "step": 94100 }, { "epoch": 3.1729991915925626, "grad_norm": 0.6786794066429138, "learning_rate": 0.000121802299470044, "loss": 0.809, "step": 94200 }, { "epoch": 3.176367555914848, "grad_norm": 0.6704023480415344, "learning_rate": 0.00012157774184855834, "loss": 0.8086, "step": 94300 }, { "epoch": 3.1797359202371327, "grad_norm": 0.652862012386322, "learning_rate": 0.00012135318422707266, "loss": 0.808, "step": 94400 }, { "epoch": 3.183104284559418, "grad_norm": 0.8095204830169678, "learning_rate": 0.00012112862660558698, "loss": 0.8066, "step": 94500 }, { "epoch": 3.1864726488817032, "grad_norm": 0.6581931710243225, "learning_rate": 0.0001209040689841013, "loss": 0.8061, "step": 94600 }, { "epoch": 3.189841013203988, "grad_norm": 0.6642458438873291, "learning_rate": 0.00012067951136261564, "loss": 0.8079, "step": 94700 }, { "epoch": 3.1932093775262733, "grad_norm": 0.6264484524726868, "learning_rate": 0.00012045495374112996, "loss": 0.8108, "step": 94800 }, { "epoch": 3.196577741848558, "grad_norm": 0.6631668210029602, "learning_rate": 0.0001202303961196443, "loss": 0.8076, "step": 94900 }, { "epoch": 3.1999461061708434, "grad_norm": 0.636448085308075, "learning_rate": 0.00012000583849815862, "loss": 0.805, "step": 95000 }, { "epoch": 3.2033144704931287, "grad_norm": 0.6331253051757812, "learning_rate": 0.00011978128087667295, "loss": 0.8089, "step": 95100 }, { "epoch": 3.2066828348154135, "grad_norm": 0.6623615026473999, "learning_rate": 0.00011955672325518727, "loss": 0.8089, "step": 95200 }, { "epoch": 3.2100511991376988, "grad_norm": 0.671399712562561, "learning_rate": 0.00011933216563370161, "loss": 0.8054, "step": 95300 }, { "epoch": 3.213419563459984, "grad_norm": 0.6822311282157898, "learning_rate": 0.00011910760801221592, "loss": 0.8057, "step": 95400 }, { "epoch": 3.216787927782269, "grad_norm": 0.6769167184829712, "learning_rate": 0.00011888305039073025, "loss": 0.8065, "step": 95500 }, { "epoch": 3.220156292104554, "grad_norm": 0.6246688961982727, "learning_rate": 0.00011865849276924457, "loss": 0.8063, "step": 95600 }, { "epoch": 3.223524656426839, "grad_norm": 0.6401100754737854, "learning_rate": 0.00011843393514775891, "loss": 0.807, "step": 95700 }, { "epoch": 3.226893020749124, "grad_norm": 0.6910848021507263, "learning_rate": 0.00011820937752627323, "loss": 0.8068, "step": 95800 }, { "epoch": 3.2302613850714095, "grad_norm": 0.7022745013237, "learning_rate": 0.00011798481990478756, "loss": 0.8077, "step": 95900 }, { "epoch": 3.2336297493936943, "grad_norm": 0.7098489999771118, "learning_rate": 0.00011776026228330189, "loss": 0.8079, "step": 96000 }, { "epoch": 3.2369981137159796, "grad_norm": 1.6580332517623901, "learning_rate": 0.00011753570466181622, "loss": 0.8052, "step": 96100 }, { "epoch": 3.240366478038265, "grad_norm": 0.6385944485664368, "learning_rate": 0.00011731114704033054, "loss": 0.8097, "step": 96200 }, { "epoch": 3.2437348423605497, "grad_norm": 0.6737959384918213, "learning_rate": 0.00011708658941884488, "loss": 0.8075, "step": 96300 }, { "epoch": 3.247103206682835, "grad_norm": 0.6414308547973633, "learning_rate": 0.00011686203179735919, "loss": 0.8074, "step": 96400 }, { "epoch": 3.2504715710051197, "grad_norm": 0.6727792024612427, "learning_rate": 0.00011663747417587351, "loss": 0.8082, "step": 96500 }, { "epoch": 3.253839935327405, "grad_norm": 0.6280369162559509, "learning_rate": 0.00011641291655438784, "loss": 0.8098, "step": 96600 }, { "epoch": 3.2572082996496903, "grad_norm": 0.6988112330436707, "learning_rate": 0.00011618835893290216, "loss": 0.8079, "step": 96700 }, { "epoch": 3.260576663971975, "grad_norm": 0.6650980710983276, "learning_rate": 0.0001159638013114165, "loss": 0.8084, "step": 96800 }, { "epoch": 3.2639450282942604, "grad_norm": 0.6886364817619324, "learning_rate": 0.00011573924368993082, "loss": 0.8074, "step": 96900 }, { "epoch": 3.267313392616545, "grad_norm": 0.6584846377372742, "learning_rate": 0.00011551468606844516, "loss": 0.8111, "step": 97000 }, { "epoch": 3.2706817569388305, "grad_norm": 0.9834907650947571, "learning_rate": 0.00011529012844695948, "loss": 0.8088, "step": 97100 }, { "epoch": 3.2740501212611157, "grad_norm": 0.6571055054664612, "learning_rate": 0.00011506557082547381, "loss": 0.8043, "step": 97200 }, { "epoch": 3.2774184855834005, "grad_norm": 0.6906171441078186, "learning_rate": 0.00011484101320398814, "loss": 0.8081, "step": 97300 }, { "epoch": 3.280786849905686, "grad_norm": 0.7476776838302612, "learning_rate": 0.00011461645558250246, "loss": 0.8064, "step": 97400 }, { "epoch": 3.284155214227971, "grad_norm": 0.6860908269882202, "learning_rate": 0.00011439189796101678, "loss": 0.8073, "step": 97500 }, { "epoch": 3.287523578550256, "grad_norm": 0.6590797901153564, "learning_rate": 0.00011416734033953111, "loss": 0.807, "step": 97600 }, { "epoch": 3.290891942872541, "grad_norm": 0.7175418138504028, "learning_rate": 0.00011394278271804544, "loss": 0.8073, "step": 97700 }, { "epoch": 3.2942603071948264, "grad_norm": 0.6721409559249878, "learning_rate": 0.00011371822509655977, "loss": 0.806, "step": 97800 }, { "epoch": 3.2976286715171113, "grad_norm": 0.6370182633399963, "learning_rate": 0.00011349366747507409, "loss": 0.8094, "step": 97900 }, { "epoch": 3.3009970358393965, "grad_norm": 0.6653867959976196, "learning_rate": 0.00011326910985358843, "loss": 0.8021, "step": 98000 }, { "epoch": 3.3043654001616813, "grad_norm": 0.635477602481842, "learning_rate": 0.00011304455223210275, "loss": 0.8074, "step": 98100 }, { "epoch": 3.3077337644839666, "grad_norm": 0.7132477760314941, "learning_rate": 0.00011281999461061708, "loss": 0.8048, "step": 98200 }, { "epoch": 3.311102128806252, "grad_norm": 0.661605954170227, "learning_rate": 0.00011259543698913139, "loss": 0.8065, "step": 98300 }, { "epoch": 3.3144704931285367, "grad_norm": 0.7334872484207153, "learning_rate": 0.00011237087936764573, "loss": 0.8056, "step": 98400 }, { "epoch": 3.317838857450822, "grad_norm": 0.7113956212997437, "learning_rate": 0.00011214632174616005, "loss": 0.8068, "step": 98500 }, { "epoch": 3.321207221773107, "grad_norm": 0.6793413758277893, "learning_rate": 0.00011192176412467438, "loss": 0.8102, "step": 98600 }, { "epoch": 3.324575586095392, "grad_norm": 0.6595569849014282, "learning_rate": 0.0001116972065031887, "loss": 0.8045, "step": 98700 }, { "epoch": 3.3279439504176773, "grad_norm": 0.6264058351516724, "learning_rate": 0.00011147264888170304, "loss": 0.8059, "step": 98800 }, { "epoch": 3.331312314739962, "grad_norm": 0.7037299275398254, "learning_rate": 0.00011124809126021736, "loss": 0.8044, "step": 98900 }, { "epoch": 3.3346806790622474, "grad_norm": 0.6255789995193481, "learning_rate": 0.0001110235336387317, "loss": 0.8081, "step": 99000 }, { "epoch": 3.3380490433845322, "grad_norm": 0.6675742864608765, "learning_rate": 0.00011079897601724602, "loss": 0.8028, "step": 99100 }, { "epoch": 3.3414174077068175, "grad_norm": 0.6799077391624451, "learning_rate": 0.00011057441839576035, "loss": 0.804, "step": 99200 }, { "epoch": 3.3447857720291028, "grad_norm": 0.6606206893920898, "learning_rate": 0.00011034986077427466, "loss": 0.8064, "step": 99300 }, { "epoch": 3.3481541363513876, "grad_norm": 0.7276676893234253, "learning_rate": 0.00011012530315278898, "loss": 0.8074, "step": 99400 }, { "epoch": 3.351522500673673, "grad_norm": 0.6872825026512146, "learning_rate": 0.00010990074553130332, "loss": 0.8056, "step": 99500 }, { "epoch": 3.354890864995958, "grad_norm": 0.6640327572822571, "learning_rate": 0.00010967618790981764, "loss": 0.8078, "step": 99600 }, { "epoch": 3.358259229318243, "grad_norm": 0.9947742819786072, "learning_rate": 0.00010945163028833198, "loss": 0.8018, "step": 99700 }, { "epoch": 3.361627593640528, "grad_norm": 0.6360524296760559, "learning_rate": 0.0001092270726668463, "loss": 0.807, "step": 99800 }, { "epoch": 3.3649959579628135, "grad_norm": 0.7029403448104858, "learning_rate": 0.00010900251504536063, "loss": 0.8059, "step": 99900 }, { "epoch": 3.3683643222850983, "grad_norm": 0.7347244024276733, "learning_rate": 0.00010877795742387497, "loss": 0.8069, "step": 100000 }, { "epoch": 3.3717326866073836, "grad_norm": 0.6576654314994812, "learning_rate": 0.00010855339980238929, "loss": 0.8035, "step": 100100 }, { "epoch": 3.3751010509296684, "grad_norm": 0.6341889500617981, "learning_rate": 0.0001083288421809036, "loss": 0.8065, "step": 100200 }, { "epoch": 3.3784694152519537, "grad_norm": 0.6809447407722473, "learning_rate": 0.00010810428455941793, "loss": 0.8071, "step": 100300 }, { "epoch": 3.381837779574239, "grad_norm": 0.6789582967758179, "learning_rate": 0.00010787972693793226, "loss": 0.8081, "step": 100400 }, { "epoch": 3.3852061438965237, "grad_norm": 0.6672530174255371, "learning_rate": 0.00010765516931644659, "loss": 0.8079, "step": 100500 }, { "epoch": 3.388574508218809, "grad_norm": 0.6484895348548889, "learning_rate": 0.00010743061169496091, "loss": 0.8053, "step": 100600 }, { "epoch": 3.391942872541094, "grad_norm": 0.6628451943397522, "learning_rate": 0.00010720605407347525, "loss": 0.8041, "step": 100700 }, { "epoch": 3.395311236863379, "grad_norm": 0.6461237668991089, "learning_rate": 0.00010698149645198957, "loss": 0.804, "step": 100800 }, { "epoch": 3.3986796011856644, "grad_norm": 0.6682327389717102, "learning_rate": 0.0001067569388305039, "loss": 0.8049, "step": 100900 }, { "epoch": 3.402047965507949, "grad_norm": 0.6611519455909729, "learning_rate": 0.00010653238120901823, "loss": 0.8004, "step": 101000 }, { "epoch": 3.4054163298302345, "grad_norm": 0.6624406576156616, "learning_rate": 0.00010630782358753256, "loss": 0.8039, "step": 101100 }, { "epoch": 3.4087846941525197, "grad_norm": 0.6566054224967957, "learning_rate": 0.00010608326596604687, "loss": 0.8058, "step": 101200 }, { "epoch": 3.4121530584748045, "grad_norm": 0.6751037240028381, "learning_rate": 0.0001058587083445612, "loss": 0.8051, "step": 101300 }, { "epoch": 3.41552142279709, "grad_norm": 0.6572412252426147, "learning_rate": 0.00010563415072307553, "loss": 0.8048, "step": 101400 }, { "epoch": 3.418889787119375, "grad_norm": 0.7792493104934692, "learning_rate": 0.00010540959310158986, "loss": 0.8044, "step": 101500 }, { "epoch": 3.42225815144166, "grad_norm": 0.6584118604660034, "learning_rate": 0.00010518503548010418, "loss": 0.8026, "step": 101600 }, { "epoch": 3.425626515763945, "grad_norm": 0.6414441466331482, "learning_rate": 0.00010496047785861852, "loss": 0.8068, "step": 101700 }, { "epoch": 3.42899488008623, "grad_norm": 0.6849080920219421, "learning_rate": 0.00010473592023713284, "loss": 0.8052, "step": 101800 }, { "epoch": 3.4323632444085153, "grad_norm": 0.6778447031974792, "learning_rate": 0.00010451136261564717, "loss": 0.8069, "step": 101900 }, { "epoch": 3.4357316087308005, "grad_norm": 0.6514096260070801, "learning_rate": 0.0001042868049941615, "loss": 0.8042, "step": 102000 }, { "epoch": 3.4390999730530853, "grad_norm": 0.630409300327301, "learning_rate": 0.00010406224737267583, "loss": 0.8029, "step": 102100 }, { "epoch": 3.4424683373753706, "grad_norm": 0.709867000579834, "learning_rate": 0.00010383768975119014, "loss": 0.8026, "step": 102200 }, { "epoch": 3.4458367016976554, "grad_norm": 0.6706274151802063, "learning_rate": 0.00010361313212970447, "loss": 0.8069, "step": 102300 }, { "epoch": 3.4492050660199407, "grad_norm": 0.6600052118301392, "learning_rate": 0.0001033885745082188, "loss": 0.8028, "step": 102400 }, { "epoch": 3.452573430342226, "grad_norm": 0.6819061040878296, "learning_rate": 0.00010316401688673313, "loss": 0.8063, "step": 102500 }, { "epoch": 3.455941794664511, "grad_norm": 0.6513516902923584, "learning_rate": 0.00010293945926524745, "loss": 0.8008, "step": 102600 }, { "epoch": 3.459310158986796, "grad_norm": 0.623355507850647, "learning_rate": 0.00010271490164376179, "loss": 0.8035, "step": 102700 }, { "epoch": 3.462678523309081, "grad_norm": 0.6720972657203674, "learning_rate": 0.00010249034402227611, "loss": 0.8057, "step": 102800 }, { "epoch": 3.466046887631366, "grad_norm": 0.6593225002288818, "learning_rate": 0.00010226578640079045, "loss": 0.8026, "step": 102900 }, { "epoch": 3.4694152519536514, "grad_norm": 0.6316161751747131, "learning_rate": 0.00010204122877930477, "loss": 0.8019, "step": 103000 }, { "epoch": 3.4727836162759362, "grad_norm": 0.7166170477867126, "learning_rate": 0.00010181667115781907, "loss": 0.8046, "step": 103100 }, { "epoch": 3.4761519805982215, "grad_norm": 0.6723181009292603, "learning_rate": 0.00010159211353633341, "loss": 0.8038, "step": 103200 }, { "epoch": 3.4795203449205068, "grad_norm": 0.6460021734237671, "learning_rate": 0.00010136755591484773, "loss": 0.8049, "step": 103300 }, { "epoch": 3.4828887092427916, "grad_norm": 0.7485737800598145, "learning_rate": 0.00010114299829336207, "loss": 0.805, "step": 103400 }, { "epoch": 3.486257073565077, "grad_norm": 0.67794269323349, "learning_rate": 0.00010091844067187639, "loss": 0.8026, "step": 103500 }, { "epoch": 3.489625437887362, "grad_norm": 0.6387248635292053, "learning_rate": 0.00010069388305039072, "loss": 0.8043, "step": 103600 }, { "epoch": 3.492993802209647, "grad_norm": 0.662192702293396, "learning_rate": 0.00010046932542890505, "loss": 0.8031, "step": 103700 }, { "epoch": 3.496362166531932, "grad_norm": 0.7354533076286316, "learning_rate": 0.00010024476780741938, "loss": 0.8051, "step": 103800 }, { "epoch": 3.499730530854217, "grad_norm": 0.7070392370223999, "learning_rate": 0.0001000202101859337, "loss": 0.8036, "step": 103900 }, { "epoch": 3.5030988951765023, "grad_norm": 0.7134096622467041, "learning_rate": 9.979565256444804e-05, "loss": 0.8034, "step": 104000 }, { "epoch": 3.5064672594987876, "grad_norm": 0.6784800291061401, "learning_rate": 9.957109494296235e-05, "loss": 0.8029, "step": 104100 }, { "epoch": 3.5098356238210724, "grad_norm": 0.6493478417396545, "learning_rate": 9.934653732147668e-05, "loss": 0.8022, "step": 104200 }, { "epoch": 3.5132039881433577, "grad_norm": 0.6267081499099731, "learning_rate": 9.9121979699991e-05, "loss": 0.8027, "step": 104300 }, { "epoch": 3.5165723524656425, "grad_norm": 0.7198253273963928, "learning_rate": 9.889742207850534e-05, "loss": 0.8021, "step": 104400 }, { "epoch": 3.5199407167879277, "grad_norm": 0.7415684461593628, "learning_rate": 9.867286445701966e-05, "loss": 0.8049, "step": 104500 }, { "epoch": 3.523309081110213, "grad_norm": 0.6261735558509827, "learning_rate": 9.8448306835534e-05, "loss": 0.8018, "step": 104600 }, { "epoch": 3.526677445432498, "grad_norm": 0.6972131133079529, "learning_rate": 9.822374921404832e-05, "loss": 0.8046, "step": 104700 }, { "epoch": 3.530045809754783, "grad_norm": 0.657211184501648, "learning_rate": 9.799919159256265e-05, "loss": 0.8049, "step": 104800 }, { "epoch": 3.533414174077068, "grad_norm": 0.6501233577728271, "learning_rate": 9.777463397107697e-05, "loss": 0.8004, "step": 104900 }, { "epoch": 3.536782538399353, "grad_norm": 0.7275915741920471, "learning_rate": 9.75500763495913e-05, "loss": 0.804, "step": 105000 }, { "epoch": 3.5401509027216385, "grad_norm": 0.6765680313110352, "learning_rate": 9.732551872810562e-05, "loss": 0.8044, "step": 105100 }, { "epoch": 3.5435192670439237, "grad_norm": 0.7218645811080933, "learning_rate": 9.710096110661995e-05, "loss": 0.8034, "step": 105200 }, { "epoch": 3.5468876313662085, "grad_norm": 0.713777482509613, "learning_rate": 9.687640348513427e-05, "loss": 0.8034, "step": 105300 }, { "epoch": 3.550255995688494, "grad_norm": 0.6668462157249451, "learning_rate": 9.665184586364861e-05, "loss": 0.8029, "step": 105400 }, { "epoch": 3.5536243600107786, "grad_norm": 0.7213618755340576, "learning_rate": 9.642728824216293e-05, "loss": 0.802, "step": 105500 }, { "epoch": 3.556992724333064, "grad_norm": 0.6867049932479858, "learning_rate": 9.620273062067726e-05, "loss": 0.8027, "step": 105600 }, { "epoch": 3.560361088655349, "grad_norm": 0.6288276314735413, "learning_rate": 9.597817299919159e-05, "loss": 0.8017, "step": 105700 }, { "epoch": 3.563729452977634, "grad_norm": 0.6391323208808899, "learning_rate": 9.575361537770592e-05, "loss": 0.8013, "step": 105800 }, { "epoch": 3.5670978172999193, "grad_norm": 0.6860449314117432, "learning_rate": 9.552905775622024e-05, "loss": 0.7998, "step": 105900 }, { "epoch": 3.570466181622204, "grad_norm": 0.7907363772392273, "learning_rate": 9.530450013473455e-05, "loss": 0.8067, "step": 106000 }, { "epoch": 3.5738345459444893, "grad_norm": 0.6776504516601562, "learning_rate": 9.507994251324889e-05, "loss": 0.8017, "step": 106100 }, { "epoch": 3.5772029102667746, "grad_norm": 0.7182029485702515, "learning_rate": 9.485538489176321e-05, "loss": 0.7997, "step": 106200 }, { "epoch": 3.5805712745890594, "grad_norm": 0.6875755786895752, "learning_rate": 9.463082727027754e-05, "loss": 0.8043, "step": 106300 }, { "epoch": 3.5839396389113447, "grad_norm": 0.6848233938217163, "learning_rate": 9.440626964879187e-05, "loss": 0.8025, "step": 106400 }, { "epoch": 3.5873080032336295, "grad_norm": 0.6334069967269897, "learning_rate": 9.41817120273062e-05, "loss": 0.8032, "step": 106500 }, { "epoch": 3.590676367555915, "grad_norm": 0.6613095998764038, "learning_rate": 9.395715440582052e-05, "loss": 0.8014, "step": 106600 }, { "epoch": 3.5940447318782, "grad_norm": 0.6903994083404541, "learning_rate": 9.373259678433486e-05, "loss": 0.8043, "step": 106700 }, { "epoch": 3.597413096200485, "grad_norm": 0.6296311616897583, "learning_rate": 9.350803916284918e-05, "loss": 0.8003, "step": 106800 }, { "epoch": 3.60078146052277, "grad_norm": 0.7521107196807861, "learning_rate": 9.328348154136351e-05, "loss": 0.7999, "step": 106900 }, { "epoch": 3.604149824845055, "grad_norm": 0.6513036489486694, "learning_rate": 9.305892391987782e-05, "loss": 0.8039, "step": 107000 }, { "epoch": 3.6075181891673402, "grad_norm": 0.6684443950653076, "learning_rate": 9.283436629839216e-05, "loss": 0.8027, "step": 107100 }, { "epoch": 3.6108865534896255, "grad_norm": 0.6627715826034546, "learning_rate": 9.260980867690648e-05, "loss": 0.8025, "step": 107200 }, { "epoch": 3.6142549178119108, "grad_norm": 0.7347738742828369, "learning_rate": 9.238525105542081e-05, "loss": 0.8036, "step": 107300 }, { "epoch": 3.6176232821341956, "grad_norm": 0.7007977366447449, "learning_rate": 9.216069343393514e-05, "loss": 0.8035, "step": 107400 }, { "epoch": 3.620991646456481, "grad_norm": 0.6845256686210632, "learning_rate": 9.193613581244947e-05, "loss": 0.8032, "step": 107500 }, { "epoch": 3.6243600107787657, "grad_norm": 0.7089165449142456, "learning_rate": 9.171157819096379e-05, "loss": 0.8046, "step": 107600 }, { "epoch": 3.627728375101051, "grad_norm": 0.7869235873222351, "learning_rate": 9.148702056947813e-05, "loss": 0.8019, "step": 107700 }, { "epoch": 3.631096739423336, "grad_norm": 0.6592691540718079, "learning_rate": 9.126246294799245e-05, "loss": 0.8012, "step": 107800 }, { "epoch": 3.634465103745621, "grad_norm": 0.6670995354652405, "learning_rate": 9.103790532650677e-05, "loss": 0.8028, "step": 107900 }, { "epoch": 3.6378334680679063, "grad_norm": 0.7031666040420532, "learning_rate": 9.081334770502109e-05, "loss": 0.8017, "step": 108000 }, { "epoch": 3.641201832390191, "grad_norm": 0.7147188782691956, "learning_rate": 9.058879008353543e-05, "loss": 0.7982, "step": 108100 }, { "epoch": 3.6445701967124764, "grad_norm": 0.7295541763305664, "learning_rate": 9.036423246204975e-05, "loss": 0.8005, "step": 108200 }, { "epoch": 3.6479385610347617, "grad_norm": 0.6590719819068909, "learning_rate": 9.013967484056408e-05, "loss": 0.8037, "step": 108300 }, { "epoch": 3.6513069253570465, "grad_norm": 0.6898632049560547, "learning_rate": 8.99151172190784e-05, "loss": 0.8002, "step": 108400 }, { "epoch": 3.6546752896793318, "grad_norm": 0.661747395992279, "learning_rate": 8.969055959759274e-05, "loss": 0.8035, "step": 108500 }, { "epoch": 3.6580436540016166, "grad_norm": 0.7051920294761658, "learning_rate": 8.946600197610706e-05, "loss": 0.801, "step": 108600 }, { "epoch": 3.661412018323902, "grad_norm": 0.6946534514427185, "learning_rate": 8.92414443546214e-05, "loss": 0.803, "step": 108700 }, { "epoch": 3.664780382646187, "grad_norm": 0.649159848690033, "learning_rate": 8.901688673313572e-05, "loss": 0.8003, "step": 108800 }, { "epoch": 3.6681487469684724, "grad_norm": 0.6959517002105713, "learning_rate": 8.879232911165003e-05, "loss": 0.8025, "step": 108900 }, { "epoch": 3.671517111290757, "grad_norm": 0.7102181315422058, "learning_rate": 8.856777149016436e-05, "loss": 0.8003, "step": 109000 }, { "epoch": 3.6748854756130425, "grad_norm": 0.6565383076667786, "learning_rate": 8.83432138686787e-05, "loss": 0.7994, "step": 109100 }, { "epoch": 3.6782538399353273, "grad_norm": 0.6567991375923157, "learning_rate": 8.811865624719302e-05, "loss": 0.802, "step": 109200 }, { "epoch": 3.6816222042576126, "grad_norm": 0.6707866787910461, "learning_rate": 8.789409862570736e-05, "loss": 0.7997, "step": 109300 }, { "epoch": 3.684990568579898, "grad_norm": 0.6689081192016602, "learning_rate": 8.766954100422168e-05, "loss": 0.8005, "step": 109400 }, { "epoch": 3.6883589329021826, "grad_norm": 0.6506887078285217, "learning_rate": 8.744498338273601e-05, "loss": 0.8017, "step": 109500 }, { "epoch": 3.691727297224468, "grad_norm": 0.6316550970077515, "learning_rate": 8.722042576125033e-05, "loss": 0.8001, "step": 109600 }, { "epoch": 3.6950956615467527, "grad_norm": 0.6964483261108398, "learning_rate": 8.699586813976467e-05, "loss": 0.8005, "step": 109700 }, { "epoch": 3.698464025869038, "grad_norm": 0.6844159364700317, "learning_rate": 8.677131051827898e-05, "loss": 0.7971, "step": 109800 }, { "epoch": 3.7018323901913233, "grad_norm": 0.5916749835014343, "learning_rate": 8.65467528967933e-05, "loss": 0.8008, "step": 109900 }, { "epoch": 3.705200754513608, "grad_norm": 0.7180382609367371, "learning_rate": 8.632219527530763e-05, "loss": 0.7967, "step": 110000 }, { "epoch": 3.7085691188358934, "grad_norm": 0.6756430864334106, "learning_rate": 8.609763765382196e-05, "loss": 0.8002, "step": 110100 }, { "epoch": 3.711937483158178, "grad_norm": 0.6347695589065552, "learning_rate": 8.587308003233629e-05, "loss": 0.801, "step": 110200 }, { "epoch": 3.7153058474804634, "grad_norm": 0.712821364402771, "learning_rate": 8.564852241085061e-05, "loss": 0.8, "step": 110300 }, { "epoch": 3.7186742118027487, "grad_norm": 0.7329118251800537, "learning_rate": 8.542396478936495e-05, "loss": 0.7991, "step": 110400 }, { "epoch": 3.7220425761250335, "grad_norm": 0.6399374008178711, "learning_rate": 8.519940716787927e-05, "loss": 0.802, "step": 110500 }, { "epoch": 3.725410940447319, "grad_norm": 0.7076860666275024, "learning_rate": 8.49748495463936e-05, "loss": 0.8004, "step": 110600 }, { "epoch": 3.7287793047696036, "grad_norm": 0.6736636161804199, "learning_rate": 8.475029192490793e-05, "loss": 0.7985, "step": 110700 }, { "epoch": 3.732147669091889, "grad_norm": 0.7174369692802429, "learning_rate": 8.452573430342225e-05, "loss": 0.7997, "step": 110800 }, { "epoch": 3.735516033414174, "grad_norm": 0.6668689250946045, "learning_rate": 8.430117668193657e-05, "loss": 0.7984, "step": 110900 }, { "epoch": 3.7388843977364594, "grad_norm": 0.6991139054298401, "learning_rate": 8.40766190604509e-05, "loss": 0.8011, "step": 111000 }, { "epoch": 3.7422527620587442, "grad_norm": 0.6559997797012329, "learning_rate": 8.385206143896523e-05, "loss": 0.8002, "step": 111100 }, { "epoch": 3.7456211263810295, "grad_norm": 0.6859176754951477, "learning_rate": 8.362750381747956e-05, "loss": 0.799, "step": 111200 }, { "epoch": 3.7489894907033143, "grad_norm": 0.7021101117134094, "learning_rate": 8.340294619599388e-05, "loss": 0.796, "step": 111300 }, { "epoch": 3.7523578550255996, "grad_norm": 0.6929513216018677, "learning_rate": 8.317838857450822e-05, "loss": 0.7977, "step": 111400 }, { "epoch": 3.755726219347885, "grad_norm": 0.6644559502601624, "learning_rate": 8.295383095302254e-05, "loss": 0.7994, "step": 111500 }, { "epoch": 3.7590945836701697, "grad_norm": 0.6280823349952698, "learning_rate": 8.272927333153687e-05, "loss": 0.7974, "step": 111600 }, { "epoch": 3.762462947992455, "grad_norm": 0.699496328830719, "learning_rate": 8.25047157100512e-05, "loss": 0.7995, "step": 111700 }, { "epoch": 3.7658313123147398, "grad_norm": 0.7624558806419373, "learning_rate": 8.228015808856552e-05, "loss": 0.8004, "step": 111800 }, { "epoch": 3.769199676637025, "grad_norm": 0.7288152575492859, "learning_rate": 8.205560046707984e-05, "loss": 0.8029, "step": 111900 }, { "epoch": 3.7725680409593103, "grad_norm": 0.6420552730560303, "learning_rate": 8.183104284559417e-05, "loss": 0.8015, "step": 112000 }, { "epoch": 3.775936405281595, "grad_norm": 0.6827369332313538, "learning_rate": 8.16064852241085e-05, "loss": 0.7997, "step": 112100 }, { "epoch": 3.7793047696038804, "grad_norm": 0.7149909138679504, "learning_rate": 8.138192760262283e-05, "loss": 0.8008, "step": 112200 }, { "epoch": 3.782673133926165, "grad_norm": 0.7044945955276489, "learning_rate": 8.115736998113715e-05, "loss": 0.7974, "step": 112300 }, { "epoch": 3.7860414982484505, "grad_norm": 0.7744246125221252, "learning_rate": 8.093281235965149e-05, "loss": 0.7971, "step": 112400 }, { "epoch": 3.7894098625707358, "grad_norm": 0.6370006203651428, "learning_rate": 8.070825473816581e-05, "loss": 0.7987, "step": 112500 }, { "epoch": 3.792778226893021, "grad_norm": 0.6973426342010498, "learning_rate": 8.048369711668015e-05, "loss": 0.7999, "step": 112600 }, { "epoch": 3.796146591215306, "grad_norm": 0.7414847612380981, "learning_rate": 8.025913949519445e-05, "loss": 0.7979, "step": 112700 }, { "epoch": 3.799514955537591, "grad_norm": 0.6913410425186157, "learning_rate": 8.003458187370878e-05, "loss": 0.796, "step": 112800 }, { "epoch": 3.802883319859876, "grad_norm": 0.6863036155700684, "learning_rate": 7.981002425222311e-05, "loss": 0.7975, "step": 112900 }, { "epoch": 3.806251684182161, "grad_norm": 0.804077684879303, "learning_rate": 7.958546663073743e-05, "loss": 0.7968, "step": 113000 }, { "epoch": 3.8096200485044465, "grad_norm": 1.013036847114563, "learning_rate": 7.936090900925177e-05, "loss": 0.7999, "step": 113100 }, { "epoch": 3.8129884128267313, "grad_norm": 0.7045647501945496, "learning_rate": 7.913635138776609e-05, "loss": 0.7991, "step": 113200 }, { "epoch": 3.8163567771490166, "grad_norm": 0.6505812406539917, "learning_rate": 7.891179376628042e-05, "loss": 0.8006, "step": 113300 }, { "epoch": 3.8197251414713014, "grad_norm": 0.6592190861701965, "learning_rate": 7.868723614479475e-05, "loss": 0.7991, "step": 113400 }, { "epoch": 3.8230935057935866, "grad_norm": 0.6840342879295349, "learning_rate": 7.846267852330908e-05, "loss": 0.798, "step": 113500 }, { "epoch": 3.826461870115872, "grad_norm": 0.6523902416229248, "learning_rate": 7.82381209018234e-05, "loss": 0.7982, "step": 113600 }, { "epoch": 3.8298302344381567, "grad_norm": 0.6831576228141785, "learning_rate": 7.801356328033772e-05, "loss": 0.7988, "step": 113700 }, { "epoch": 3.833198598760442, "grad_norm": 0.6503862738609314, "learning_rate": 7.778900565885205e-05, "loss": 0.7995, "step": 113800 }, { "epoch": 3.836566963082727, "grad_norm": 0.6413772702217102, "learning_rate": 7.756444803736638e-05, "loss": 0.7992, "step": 113900 }, { "epoch": 3.839935327405012, "grad_norm": 0.708017885684967, "learning_rate": 7.73398904158807e-05, "loss": 0.7995, "step": 114000 }, { "epoch": 3.8433036917272974, "grad_norm": 0.686399519443512, "learning_rate": 7.711533279439504e-05, "loss": 0.7974, "step": 114100 }, { "epoch": 3.846672056049582, "grad_norm": 0.6614728569984436, "learning_rate": 7.689077517290936e-05, "loss": 0.7972, "step": 114200 }, { "epoch": 3.8500404203718674, "grad_norm": 1.1526070833206177, "learning_rate": 7.66662175514237e-05, "loss": 0.795, "step": 114300 }, { "epoch": 3.8534087846941523, "grad_norm": 0.6846965551376343, "learning_rate": 7.644165992993802e-05, "loss": 0.799, "step": 114400 }, { "epoch": 3.8567771490164375, "grad_norm": 0.699941098690033, "learning_rate": 7.621710230845235e-05, "loss": 0.796, "step": 114500 }, { "epoch": 3.860145513338723, "grad_norm": 0.7409800291061401, "learning_rate": 7.599254468696666e-05, "loss": 0.7975, "step": 114600 }, { "epoch": 3.863513877661008, "grad_norm": 0.7050840854644775, "learning_rate": 7.5767987065481e-05, "loss": 0.7965, "step": 114700 }, { "epoch": 3.866882241983293, "grad_norm": 0.6288536787033081, "learning_rate": 7.554342944399532e-05, "loss": 0.7987, "step": 114800 }, { "epoch": 3.870250606305578, "grad_norm": 0.69952791929245, "learning_rate": 7.531887182250965e-05, "loss": 0.7983, "step": 114900 }, { "epoch": 3.873618970627863, "grad_norm": 0.7967835664749146, "learning_rate": 7.509431420102397e-05, "loss": 0.7992, "step": 115000 }, { "epoch": 3.8769873349501482, "grad_norm": 0.6878296732902527, "learning_rate": 7.486975657953831e-05, "loss": 0.7953, "step": 115100 }, { "epoch": 3.8803556992724335, "grad_norm": 0.750624418258667, "learning_rate": 7.464519895805263e-05, "loss": 0.7972, "step": 115200 }, { "epoch": 3.8837240635947183, "grad_norm": 0.6286699771881104, "learning_rate": 7.442064133656697e-05, "loss": 0.7979, "step": 115300 }, { "epoch": 3.8870924279170036, "grad_norm": 0.7046734094619751, "learning_rate": 7.419608371508129e-05, "loss": 0.7991, "step": 115400 }, { "epoch": 3.8904607922392884, "grad_norm": 0.6798041462898254, "learning_rate": 7.397152609359561e-05, "loss": 0.8019, "step": 115500 }, { "epoch": 3.8938291565615737, "grad_norm": 0.6814720034599304, "learning_rate": 7.374696847210994e-05, "loss": 0.7992, "step": 115600 }, { "epoch": 3.897197520883859, "grad_norm": 0.6247894167900085, "learning_rate": 7.352241085062427e-05, "loss": 0.7959, "step": 115700 }, { "epoch": 3.900565885206144, "grad_norm": 0.6836827397346497, "learning_rate": 7.329785322913859e-05, "loss": 0.7971, "step": 115800 }, { "epoch": 3.903934249528429, "grad_norm": 0.6945433616638184, "learning_rate": 7.307329560765291e-05, "loss": 0.7947, "step": 115900 }, { "epoch": 3.907302613850714, "grad_norm": 0.7244613766670227, "learning_rate": 7.284873798616724e-05, "loss": 0.7966, "step": 116000 }, { "epoch": 3.910670978172999, "grad_norm": 0.6739228367805481, "learning_rate": 7.262418036468158e-05, "loss": 0.7987, "step": 116100 }, { "epoch": 3.9140393424952844, "grad_norm": 0.7027273774147034, "learning_rate": 7.23996227431959e-05, "loss": 0.7966, "step": 116200 }, { "epoch": 3.9174077068175692, "grad_norm": 0.7074873447418213, "learning_rate": 7.217506512171022e-05, "loss": 0.7953, "step": 116300 }, { "epoch": 3.9207760711398545, "grad_norm": 0.6511521935462952, "learning_rate": 7.195050750022454e-05, "loss": 0.8015, "step": 116400 }, { "epoch": 3.9241444354621393, "grad_norm": 0.7246668934822083, "learning_rate": 7.172594987873888e-05, "loss": 0.7963, "step": 116500 }, { "epoch": 3.9275127997844246, "grad_norm": 0.6860557198524475, "learning_rate": 7.15013922572532e-05, "loss": 0.7973, "step": 116600 }, { "epoch": 3.93088116410671, "grad_norm": 0.7153899669647217, "learning_rate": 7.127683463576754e-05, "loss": 0.7959, "step": 116700 }, { "epoch": 3.934249528428995, "grad_norm": 0.6767284274101257, "learning_rate": 7.105227701428186e-05, "loss": 0.7971, "step": 116800 }, { "epoch": 3.93761789275128, "grad_norm": 0.6651219129562378, "learning_rate": 7.082771939279618e-05, "loss": 0.7959, "step": 116900 }, { "epoch": 3.940986257073565, "grad_norm": 0.6394347548484802, "learning_rate": 7.060316177131051e-05, "loss": 0.7988, "step": 117000 }, { "epoch": 3.94435462139585, "grad_norm": 0.6922232508659363, "learning_rate": 7.037860414982484e-05, "loss": 0.7961, "step": 117100 }, { "epoch": 3.9477229857181353, "grad_norm": 0.6803662776947021, "learning_rate": 7.015404652833917e-05, "loss": 0.7982, "step": 117200 }, { "epoch": 3.9510913500404206, "grad_norm": 0.7751142978668213, "learning_rate": 6.992948890685349e-05, "loss": 0.7968, "step": 117300 }, { "epoch": 3.9544597143627054, "grad_norm": 0.6750118136405945, "learning_rate": 6.970493128536781e-05, "loss": 0.7955, "step": 117400 }, { "epoch": 3.9578280786849906, "grad_norm": 0.6720499992370605, "learning_rate": 6.948037366388215e-05, "loss": 0.7975, "step": 117500 }, { "epoch": 3.9611964430072755, "grad_norm": 0.7234087586402893, "learning_rate": 6.925581604239647e-05, "loss": 0.7954, "step": 117600 }, { "epoch": 3.9645648073295607, "grad_norm": 0.6893304586410522, "learning_rate": 6.903125842091079e-05, "loss": 0.7958, "step": 117700 }, { "epoch": 3.967933171651846, "grad_norm": 0.6690725684165955, "learning_rate": 6.880670079942513e-05, "loss": 0.794, "step": 117800 }, { "epoch": 3.971301535974131, "grad_norm": 0.6747570633888245, "learning_rate": 6.858214317793945e-05, "loss": 0.7962, "step": 117900 }, { "epoch": 3.974669900296416, "grad_norm": 0.7051836848258972, "learning_rate": 6.835758555645378e-05, "loss": 0.7947, "step": 118000 }, { "epoch": 3.978038264618701, "grad_norm": 0.6995989084243774, "learning_rate": 6.81330279349681e-05, "loss": 0.795, "step": 118100 }, { "epoch": 3.981406628940986, "grad_norm": 0.6704210638999939, "learning_rate": 6.790847031348243e-05, "loss": 0.7944, "step": 118200 }, { "epoch": 3.9847749932632714, "grad_norm": 0.6781893372535706, "learning_rate": 6.768391269199676e-05, "loss": 0.7954, "step": 118300 }, { "epoch": 3.9881433575855567, "grad_norm": 0.6674184203147888, "learning_rate": 6.745935507051109e-05, "loss": 0.7984, "step": 118400 }, { "epoch": 3.9915117219078415, "grad_norm": 0.6545355319976807, "learning_rate": 6.723479744902542e-05, "loss": 0.7946, "step": 118500 }, { "epoch": 3.994880086230127, "grad_norm": 0.6851162910461426, "learning_rate": 6.701023982753974e-05, "loss": 0.7954, "step": 118600 }, { "epoch": 3.9982484505524116, "grad_norm": 0.6962630152702332, "learning_rate": 6.678568220605406e-05, "loss": 0.7949, "step": 118700 }, { "epoch": 4.0, "eval_loss": 0.7465363144874573, "eval_runtime": 1.8135, "eval_samples_per_second": 2757.131, "eval_steps_per_second": 43.563, "step": 118752 }, { "epoch": 4.001616814874697, "grad_norm": 0.7494738101959229, "learning_rate": 6.65611245845684e-05, "loss": 0.7985, "step": 118800 }, { "epoch": 4.004985179196982, "grad_norm": 0.6624446511268616, "learning_rate": 6.633656696308272e-05, "loss": 0.7954, "step": 118900 }, { "epoch": 4.008353543519267, "grad_norm": 0.69657963514328, "learning_rate": 6.611200934159706e-05, "loss": 0.7957, "step": 119000 }, { "epoch": 4.011721907841552, "grad_norm": 0.706929087638855, "learning_rate": 6.588745172011138e-05, "loss": 0.7969, "step": 119100 }, { "epoch": 4.015090272163837, "grad_norm": 0.7379885911941528, "learning_rate": 6.56628940986257e-05, "loss": 0.7968, "step": 119200 }, { "epoch": 4.018458636486122, "grad_norm": 0.6987789869308472, "learning_rate": 6.543833647714002e-05, "loss": 0.7959, "step": 119300 }, { "epoch": 4.021827000808408, "grad_norm": 0.6569750308990479, "learning_rate": 6.521377885565436e-05, "loss": 0.7917, "step": 119400 }, { "epoch": 4.025195365130693, "grad_norm": 0.7094253301620483, "learning_rate": 6.498922123416869e-05, "loss": 0.7937, "step": 119500 }, { "epoch": 4.028563729452977, "grad_norm": 0.7103100419044495, "learning_rate": 6.476466361268301e-05, "loss": 0.7988, "step": 119600 }, { "epoch": 4.0319320937752625, "grad_norm": 0.7065451145172119, "learning_rate": 6.454010599119733e-05, "loss": 0.7905, "step": 119700 }, { "epoch": 4.035300458097548, "grad_norm": 0.690153181552887, "learning_rate": 6.431554836971166e-05, "loss": 0.7948, "step": 119800 }, { "epoch": 4.038668822419833, "grad_norm": 0.7387445569038391, "learning_rate": 6.409099074822599e-05, "loss": 0.7944, "step": 119900 }, { "epoch": 4.042037186742118, "grad_norm": 0.6826400756835938, "learning_rate": 6.386643312674031e-05, "loss": 0.7935, "step": 120000 }, { "epoch": 4.045405551064403, "grad_norm": 0.6468490362167358, "learning_rate": 6.364187550525463e-05, "loss": 0.796, "step": 120100 }, { "epoch": 4.048773915386688, "grad_norm": 0.6864830851554871, "learning_rate": 6.341731788376897e-05, "loss": 0.7948, "step": 120200 }, { "epoch": 4.052142279708973, "grad_norm": 0.7382389903068542, "learning_rate": 6.319276026228329e-05, "loss": 0.7943, "step": 120300 }, { "epoch": 4.0555106440312585, "grad_norm": 0.7259221076965332, "learning_rate": 6.296820264079763e-05, "loss": 0.7975, "step": 120400 }, { "epoch": 4.058879008353544, "grad_norm": 0.7078028917312622, "learning_rate": 6.274364501931195e-05, "loss": 0.7969, "step": 120500 }, { "epoch": 4.062247372675829, "grad_norm": 0.6797070503234863, "learning_rate": 6.251908739782627e-05, "loss": 0.7946, "step": 120600 }, { "epoch": 4.065615736998113, "grad_norm": 0.7093040347099304, "learning_rate": 6.22945297763406e-05, "loss": 0.7931, "step": 120700 }, { "epoch": 4.068984101320399, "grad_norm": 0.6741282939910889, "learning_rate": 6.206997215485493e-05, "loss": 0.7975, "step": 120800 }, { "epoch": 4.072352465642684, "grad_norm": 0.6645729541778564, "learning_rate": 6.184541453336926e-05, "loss": 0.7959, "step": 120900 }, { "epoch": 4.075720829964969, "grad_norm": 0.6510828137397766, "learning_rate": 6.162085691188358e-05, "loss": 0.7945, "step": 121000 }, { "epoch": 4.0790891942872545, "grad_norm": 0.7214908599853516, "learning_rate": 6.13962992903979e-05, "loss": 0.7934, "step": 121100 }, { "epoch": 4.082457558609539, "grad_norm": 0.7020803093910217, "learning_rate": 6.117174166891224e-05, "loss": 0.7952, "step": 121200 }, { "epoch": 4.085825922931824, "grad_norm": 0.7014907002449036, "learning_rate": 6.094718404742656e-05, "loss": 0.7932, "step": 121300 }, { "epoch": 4.089194287254109, "grad_norm": 0.6577868461608887, "learning_rate": 6.072262642594089e-05, "loss": 0.7904, "step": 121400 }, { "epoch": 4.092562651576395, "grad_norm": 0.6450611352920532, "learning_rate": 6.049806880445522e-05, "loss": 0.7917, "step": 121500 }, { "epoch": 4.09593101589868, "grad_norm": 0.7224249243736267, "learning_rate": 6.027351118296954e-05, "loss": 0.793, "step": 121600 }, { "epoch": 4.099299380220964, "grad_norm": 0.6740891337394714, "learning_rate": 6.004895356148387e-05, "loss": 0.7929, "step": 121700 }, { "epoch": 4.10266774454325, "grad_norm": 0.7219831347465515, "learning_rate": 5.98243959399982e-05, "loss": 0.7921, "step": 121800 }, { "epoch": 4.106036108865535, "grad_norm": 0.7117908596992493, "learning_rate": 5.9599838318512525e-05, "loss": 0.7936, "step": 121900 }, { "epoch": 4.10940447318782, "grad_norm": 0.8350685238838196, "learning_rate": 5.9375280697026854e-05, "loss": 0.7924, "step": 122000 }, { "epoch": 4.112772837510105, "grad_norm": 0.6902172565460205, "learning_rate": 5.9150723075541175e-05, "loss": 0.7931, "step": 122100 }, { "epoch": 4.116141201832391, "grad_norm": 0.6757128238677979, "learning_rate": 5.8926165454055504e-05, "loss": 0.7954, "step": 122200 }, { "epoch": 4.119509566154675, "grad_norm": 0.6877591013908386, "learning_rate": 5.870160783256983e-05, "loss": 0.7916, "step": 122300 }, { "epoch": 4.12287793047696, "grad_norm": 0.677144467830658, "learning_rate": 5.847705021108416e-05, "loss": 0.7951, "step": 122400 }, { "epoch": 4.1262462947992455, "grad_norm": 0.6870157122612, "learning_rate": 5.825249258959848e-05, "loss": 0.7931, "step": 122500 }, { "epoch": 4.129614659121531, "grad_norm": 0.7329301834106445, "learning_rate": 5.802793496811281e-05, "loss": 0.7947, "step": 122600 }, { "epoch": 4.132983023443816, "grad_norm": 0.7153013348579407, "learning_rate": 5.780337734662714e-05, "loss": 0.7927, "step": 122700 }, { "epoch": 4.1363513877661005, "grad_norm": 0.6681983470916748, "learning_rate": 5.757881972514147e-05, "loss": 0.7966, "step": 122800 }, { "epoch": 4.139719752088386, "grad_norm": 0.6760982275009155, "learning_rate": 5.7354262103655796e-05, "loss": 0.7932, "step": 122900 }, { "epoch": 4.143088116410671, "grad_norm": 0.6279290914535522, "learning_rate": 5.712970448217012e-05, "loss": 0.7936, "step": 123000 }, { "epoch": 4.146456480732956, "grad_norm": 0.6763176918029785, "learning_rate": 5.6905146860684446e-05, "loss": 0.7966, "step": 123100 }, { "epoch": 4.1498248450552415, "grad_norm": 0.6843128204345703, "learning_rate": 5.6680589239198774e-05, "loss": 0.7933, "step": 123200 }, { "epoch": 4.153193209377526, "grad_norm": 0.6466397643089294, "learning_rate": 5.64560316177131e-05, "loss": 0.7955, "step": 123300 }, { "epoch": 4.156561573699811, "grad_norm": 0.6983882188796997, "learning_rate": 5.623147399622743e-05, "loss": 0.7951, "step": 123400 }, { "epoch": 4.159929938022096, "grad_norm": 0.6982767581939697, "learning_rate": 5.600691637474175e-05, "loss": 0.7932, "step": 123500 }, { "epoch": 4.163298302344382, "grad_norm": 0.6864616870880127, "learning_rate": 5.578235875325608e-05, "loss": 0.7926, "step": 123600 }, { "epoch": 4.166666666666667, "grad_norm": 0.7273553609848022, "learning_rate": 5.555780113177041e-05, "loss": 0.7935, "step": 123700 }, { "epoch": 4.170035030988951, "grad_norm": 0.6874902248382568, "learning_rate": 5.533324351028474e-05, "loss": 0.7913, "step": 123800 }, { "epoch": 4.173403395311237, "grad_norm": 0.7229161262512207, "learning_rate": 5.5108685888799067e-05, "loss": 0.7887, "step": 123900 }, { "epoch": 4.176771759633522, "grad_norm": 0.7190577387809753, "learning_rate": 5.488412826731338e-05, "loss": 0.7908, "step": 124000 }, { "epoch": 4.180140123955807, "grad_norm": 0.6708100438117981, "learning_rate": 5.4659570645827717e-05, "loss": 0.7937, "step": 124100 }, { "epoch": 4.183508488278092, "grad_norm": 0.8221725821495056, "learning_rate": 5.4435013024342045e-05, "loss": 0.7923, "step": 124200 }, { "epoch": 4.186876852600378, "grad_norm": 0.7301611304283142, "learning_rate": 5.4210455402856373e-05, "loss": 0.7946, "step": 124300 }, { "epoch": 4.190245216922662, "grad_norm": 0.7508817315101624, "learning_rate": 5.39858977813707e-05, "loss": 0.7921, "step": 124400 }, { "epoch": 4.193613581244947, "grad_norm": 0.7159464359283447, "learning_rate": 5.376134015988502e-05, "loss": 0.7916, "step": 124500 }, { "epoch": 4.196981945567233, "grad_norm": 0.6572859883308411, "learning_rate": 5.3536782538399345e-05, "loss": 0.7932, "step": 124600 }, { "epoch": 4.200350309889518, "grad_norm": 0.8845502138137817, "learning_rate": 5.3312224916913674e-05, "loss": 0.7914, "step": 124700 }, { "epoch": 4.203718674211803, "grad_norm": 0.7034029364585876, "learning_rate": 5.3087667295428e-05, "loss": 0.7925, "step": 124800 }, { "epoch": 4.2070870385340875, "grad_norm": 0.66123366355896, "learning_rate": 5.2863109673942324e-05, "loss": 0.7956, "step": 124900 }, { "epoch": 4.210455402856373, "grad_norm": 0.7376932501792908, "learning_rate": 5.263855205245665e-05, "loss": 0.7891, "step": 125000 }, { "epoch": 4.213823767178658, "grad_norm": 0.7154867649078369, "learning_rate": 5.241399443097098e-05, "loss": 0.7921, "step": 125100 }, { "epoch": 4.217192131500943, "grad_norm": 0.6891977190971375, "learning_rate": 5.218943680948531e-05, "loss": 0.7938, "step": 125200 }, { "epoch": 4.220560495823229, "grad_norm": 0.713810384273529, "learning_rate": 5.196487918799964e-05, "loss": 0.7926, "step": 125300 }, { "epoch": 4.223928860145513, "grad_norm": 0.7049047350883484, "learning_rate": 5.174032156651396e-05, "loss": 0.7955, "step": 125400 }, { "epoch": 4.227297224467798, "grad_norm": 0.7068188786506653, "learning_rate": 5.151576394502829e-05, "loss": 0.7966, "step": 125500 }, { "epoch": 4.2306655887900835, "grad_norm": 0.7107961177825928, "learning_rate": 5.1291206323542616e-05, "loss": 0.7959, "step": 125600 }, { "epoch": 4.234033953112369, "grad_norm": 0.7235453128814697, "learning_rate": 5.1066648702056944e-05, "loss": 0.7904, "step": 125700 }, { "epoch": 4.237402317434654, "grad_norm": 0.7281498908996582, "learning_rate": 5.084209108057127e-05, "loss": 0.7916, "step": 125800 }, { "epoch": 4.240770681756938, "grad_norm": 0.726445198059082, "learning_rate": 5.0617533459085594e-05, "loss": 0.7905, "step": 125900 }, { "epoch": 4.244139046079224, "grad_norm": 0.7289313673973083, "learning_rate": 5.039297583759992e-05, "loss": 0.796, "step": 126000 }, { "epoch": 4.247507410401509, "grad_norm": 0.6762425899505615, "learning_rate": 5.016841821611425e-05, "loss": 0.7898, "step": 126100 }, { "epoch": 4.250875774723794, "grad_norm": 0.7091333866119385, "learning_rate": 4.994386059462858e-05, "loss": 0.7946, "step": 126200 }, { "epoch": 4.2542441390460795, "grad_norm": 0.6724031567573547, "learning_rate": 4.971930297314291e-05, "loss": 0.7931, "step": 126300 }, { "epoch": 4.257612503368365, "grad_norm": 0.7157333493232727, "learning_rate": 4.949474535165723e-05, "loss": 0.7925, "step": 126400 }, { "epoch": 4.260980867690649, "grad_norm": 0.7710690498352051, "learning_rate": 4.927018773017156e-05, "loss": 0.795, "step": 126500 }, { "epoch": 4.264349232012934, "grad_norm": 0.6310470104217529, "learning_rate": 4.9045630108685886e-05, "loss": 0.7931, "step": 126600 }, { "epoch": 4.26771759633522, "grad_norm": 0.6831243634223938, "learning_rate": 4.8821072487200215e-05, "loss": 0.7926, "step": 126700 }, { "epoch": 4.271085960657505, "grad_norm": 0.675401508808136, "learning_rate": 4.859651486571454e-05, "loss": 0.7902, "step": 126800 }, { "epoch": 4.27445432497979, "grad_norm": 0.715557873249054, "learning_rate": 4.8371957244228865e-05, "loss": 0.7932, "step": 126900 }, { "epoch": 4.2778226893020745, "grad_norm": 0.6567670702934265, "learning_rate": 4.814739962274319e-05, "loss": 0.7925, "step": 127000 }, { "epoch": 4.28119105362436, "grad_norm": 0.6738788485527039, "learning_rate": 4.792284200125752e-05, "loss": 0.7932, "step": 127100 }, { "epoch": 4.284559417946645, "grad_norm": 0.6749903559684753, "learning_rate": 4.769828437977185e-05, "loss": 0.792, "step": 127200 }, { "epoch": 4.28792778226893, "grad_norm": 0.6749402284622192, "learning_rate": 4.7473726758286165e-05, "loss": 0.7924, "step": 127300 }, { "epoch": 4.291296146591216, "grad_norm": 0.7213864326477051, "learning_rate": 4.724916913680049e-05, "loss": 0.7923, "step": 127400 }, { "epoch": 4.2946645109135, "grad_norm": 0.6684932112693787, "learning_rate": 4.702461151531482e-05, "loss": 0.7906, "step": 127500 }, { "epoch": 4.298032875235785, "grad_norm": 0.676459014415741, "learning_rate": 4.680005389382916e-05, "loss": 0.791, "step": 127600 }, { "epoch": 4.3014012395580705, "grad_norm": 0.6472882032394409, "learning_rate": 4.6575496272343485e-05, "loss": 0.7893, "step": 127700 }, { "epoch": 4.304769603880356, "grad_norm": 0.6770561337471008, "learning_rate": 4.63509386508578e-05, "loss": 0.7907, "step": 127800 }, { "epoch": 4.308137968202641, "grad_norm": 0.6275637149810791, "learning_rate": 4.612638102937213e-05, "loss": 0.7913, "step": 127900 }, { "epoch": 4.311506332524926, "grad_norm": 0.6540498733520508, "learning_rate": 4.590182340788646e-05, "loss": 0.793, "step": 128000 }, { "epoch": 4.314874696847211, "grad_norm": 0.7203840017318726, "learning_rate": 4.5677265786400785e-05, "loss": 0.7913, "step": 128100 }, { "epoch": 4.318243061169496, "grad_norm": 0.6884647607803345, "learning_rate": 4.5452708164915114e-05, "loss": 0.7928, "step": 128200 }, { "epoch": 4.321611425491781, "grad_norm": 0.8688404560089111, "learning_rate": 4.5228150543429435e-05, "loss": 0.791, "step": 128300 }, { "epoch": 4.3249797898140665, "grad_norm": 0.6714907884597778, "learning_rate": 4.5003592921943764e-05, "loss": 0.7909, "step": 128400 }, { "epoch": 4.328348154136352, "grad_norm": 0.6698452830314636, "learning_rate": 4.477903530045809e-05, "loss": 0.7934, "step": 128500 }, { "epoch": 4.331716518458636, "grad_norm": 0.6693661212921143, "learning_rate": 4.455447767897242e-05, "loss": 0.7917, "step": 128600 }, { "epoch": 4.335084882780921, "grad_norm": 0.6978726983070374, "learning_rate": 4.432992005748675e-05, "loss": 0.7903, "step": 128700 }, { "epoch": 4.338453247103207, "grad_norm": 0.7052512168884277, "learning_rate": 4.410536243600107e-05, "loss": 0.7909, "step": 128800 }, { "epoch": 4.341821611425492, "grad_norm": 0.6648239493370056, "learning_rate": 4.38808048145154e-05, "loss": 0.793, "step": 128900 }, { "epoch": 4.345189975747777, "grad_norm": 0.6616061329841614, "learning_rate": 4.365624719302973e-05, "loss": 0.7915, "step": 129000 }, { "epoch": 4.348558340070062, "grad_norm": 0.7084757089614868, "learning_rate": 4.3431689571544056e-05, "loss": 0.7918, "step": 129100 }, { "epoch": 4.351926704392347, "grad_norm": 0.6875206828117371, "learning_rate": 4.3207131950058384e-05, "loss": 0.7893, "step": 129200 }, { "epoch": 4.355295068714632, "grad_norm": 0.6854655742645264, "learning_rate": 4.2982574328572706e-05, "loss": 0.7859, "step": 129300 }, { "epoch": 4.358663433036917, "grad_norm": 0.7505910992622375, "learning_rate": 4.2758016707087034e-05, "loss": 0.7934, "step": 129400 }, { "epoch": 4.362031797359203, "grad_norm": 0.6831994652748108, "learning_rate": 4.253345908560136e-05, "loss": 0.7941, "step": 129500 }, { "epoch": 4.365400161681487, "grad_norm": 0.7224695086479187, "learning_rate": 4.230890146411569e-05, "loss": 0.7907, "step": 129600 }, { "epoch": 4.368768526003772, "grad_norm": 0.7102277874946594, "learning_rate": 4.208434384263001e-05, "loss": 0.7926, "step": 129700 }, { "epoch": 4.372136890326058, "grad_norm": 0.6760639548301697, "learning_rate": 4.185978622114434e-05, "loss": 0.7949, "step": 129800 }, { "epoch": 4.375505254648343, "grad_norm": 0.6891269087791443, "learning_rate": 4.163522859965867e-05, "loss": 0.7901, "step": 129900 }, { "epoch": 4.378873618970628, "grad_norm": 0.6708014607429504, "learning_rate": 4.1410670978173e-05, "loss": 0.7894, "step": 130000 }, { "epoch": 4.382241983292913, "grad_norm": 0.7219834327697754, "learning_rate": 4.1186113356687327e-05, "loss": 0.7917, "step": 130100 }, { "epoch": 4.385610347615198, "grad_norm": 0.6797767281532288, "learning_rate": 4.096155573520165e-05, "loss": 0.79, "step": 130200 }, { "epoch": 4.388978711937483, "grad_norm": 1.1183526515960693, "learning_rate": 4.073699811371598e-05, "loss": 0.7903, "step": 130300 }, { "epoch": 4.392347076259768, "grad_norm": 0.6638786792755127, "learning_rate": 4.0512440492230305e-05, "loss": 0.7929, "step": 130400 }, { "epoch": 4.3957154405820535, "grad_norm": 0.6631746888160706, "learning_rate": 4.0287882870744633e-05, "loss": 0.7895, "step": 130500 }, { "epoch": 4.399083804904339, "grad_norm": 0.6432344317436218, "learning_rate": 4.006332524925896e-05, "loss": 0.793, "step": 130600 }, { "epoch": 4.402452169226623, "grad_norm": 0.6888932585716248, "learning_rate": 3.983876762777328e-05, "loss": 0.7863, "step": 130700 }, { "epoch": 4.4058205335489085, "grad_norm": 0.6781389713287354, "learning_rate": 3.9614210006287605e-05, "loss": 0.7894, "step": 130800 }, { "epoch": 4.409188897871194, "grad_norm": 0.6817391514778137, "learning_rate": 3.9389652384801934e-05, "loss": 0.7915, "step": 130900 }, { "epoch": 4.412557262193479, "grad_norm": 0.7040595412254333, "learning_rate": 3.916509476331627e-05, "loss": 0.7898, "step": 131000 }, { "epoch": 4.415925626515764, "grad_norm": 0.6829173564910889, "learning_rate": 3.89405371418306e-05, "loss": 0.7911, "step": 131100 }, { "epoch": 4.419293990838049, "grad_norm": 0.7113239765167236, "learning_rate": 3.871597952034491e-05, "loss": 0.7871, "step": 131200 }, { "epoch": 4.422662355160334, "grad_norm": 0.6862262487411499, "learning_rate": 3.849142189885924e-05, "loss": 0.7915, "step": 131300 }, { "epoch": 4.426030719482619, "grad_norm": 0.6750577092170715, "learning_rate": 3.826686427737357e-05, "loss": 0.7925, "step": 131400 }, { "epoch": 4.429399083804904, "grad_norm": 0.7142935395240784, "learning_rate": 3.80423066558879e-05, "loss": 0.7906, "step": 131500 }, { "epoch": 4.43276744812719, "grad_norm": 0.7272329330444336, "learning_rate": 3.7817749034402226e-05, "loss": 0.7898, "step": 131600 }, { "epoch": 4.436135812449475, "grad_norm": 0.7627023458480835, "learning_rate": 3.759319141291655e-05, "loss": 0.7907, "step": 131700 }, { "epoch": 4.439504176771759, "grad_norm": 0.6552876830101013, "learning_rate": 3.7368633791430876e-05, "loss": 0.7901, "step": 131800 }, { "epoch": 4.442872541094045, "grad_norm": 0.7480065822601318, "learning_rate": 3.7144076169945204e-05, "loss": 0.7917, "step": 131900 }, { "epoch": 4.44624090541633, "grad_norm": 0.7101287841796875, "learning_rate": 3.691951854845953e-05, "loss": 0.7909, "step": 132000 }, { "epoch": 4.449609269738615, "grad_norm": 0.7063937187194824, "learning_rate": 3.669496092697386e-05, "loss": 0.7897, "step": 132100 }, { "epoch": 4.4529776340609, "grad_norm": 0.7050164341926575, "learning_rate": 3.647040330548819e-05, "loss": 0.7924, "step": 132200 }, { "epoch": 4.456345998383185, "grad_norm": 0.7028830051422119, "learning_rate": 3.624584568400251e-05, "loss": 0.7894, "step": 132300 }, { "epoch": 4.45971436270547, "grad_norm": 0.7394464015960693, "learning_rate": 3.602128806251684e-05, "loss": 0.7916, "step": 132400 }, { "epoch": 4.463082727027755, "grad_norm": 0.6921053528785706, "learning_rate": 3.579673044103116e-05, "loss": 0.7888, "step": 132500 }, { "epoch": 4.466451091350041, "grad_norm": 0.6709455847740173, "learning_rate": 3.557217281954549e-05, "loss": 0.7915, "step": 132600 }, { "epoch": 4.469819455672326, "grad_norm": 0.7549142241477966, "learning_rate": 3.534761519805982e-05, "loss": 0.7925, "step": 132700 }, { "epoch": 4.47318781999461, "grad_norm": 0.716698944568634, "learning_rate": 3.5123057576574146e-05, "loss": 0.7893, "step": 132800 }, { "epoch": 4.4765561843168955, "grad_norm": 0.6615895628929138, "learning_rate": 3.4898499955088475e-05, "loss": 0.7864, "step": 132900 }, { "epoch": 4.479924548639181, "grad_norm": 0.8180909156799316, "learning_rate": 3.4673942333602796e-05, "loss": 0.7902, "step": 133000 }, { "epoch": 4.483292912961466, "grad_norm": 0.7339180111885071, "learning_rate": 3.4449384712117125e-05, "loss": 0.789, "step": 133100 }, { "epoch": 4.486661277283751, "grad_norm": 0.921172559261322, "learning_rate": 3.422482709063145e-05, "loss": 0.7876, "step": 133200 }, { "epoch": 4.490029641606036, "grad_norm": 0.7154456973075867, "learning_rate": 3.400026946914578e-05, "loss": 0.7911, "step": 133300 }, { "epoch": 4.493398005928321, "grad_norm": 0.9519571661949158, "learning_rate": 3.377571184766011e-05, "loss": 0.7895, "step": 133400 }, { "epoch": 4.496766370250606, "grad_norm": 0.7096812725067139, "learning_rate": 3.355115422617443e-05, "loss": 0.7892, "step": 133500 }, { "epoch": 4.5001347345728915, "grad_norm": 0.6969826817512512, "learning_rate": 3.332659660468876e-05, "loss": 0.7913, "step": 133600 }, { "epoch": 4.503503098895177, "grad_norm": 0.6900309920310974, "learning_rate": 3.310203898320309e-05, "loss": 0.7909, "step": 133700 }, { "epoch": 4.506871463217461, "grad_norm": 0.7131757736206055, "learning_rate": 3.287748136171742e-05, "loss": 0.7932, "step": 133800 }, { "epoch": 4.510239827539746, "grad_norm": 0.6708106398582458, "learning_rate": 3.265292374023174e-05, "loss": 0.7889, "step": 133900 }, { "epoch": 4.513608191862032, "grad_norm": 0.7623175382614136, "learning_rate": 3.242836611874607e-05, "loss": 0.7906, "step": 134000 }, { "epoch": 4.516976556184317, "grad_norm": 0.7173711657524109, "learning_rate": 3.2203808497260395e-05, "loss": 0.7869, "step": 134100 }, { "epoch": 4.520344920506602, "grad_norm": 0.7341943979263306, "learning_rate": 3.197925087577472e-05, "loss": 0.7897, "step": 134200 }, { "epoch": 4.5237132848288875, "grad_norm": 0.7385048866271973, "learning_rate": 3.1754693254289045e-05, "loss": 0.7884, "step": 134300 }, { "epoch": 4.527081649151172, "grad_norm": 0.7369481921195984, "learning_rate": 3.1530135632803374e-05, "loss": 0.7922, "step": 134400 }, { "epoch": 4.530450013473457, "grad_norm": 0.7261357307434082, "learning_rate": 3.13055780113177e-05, "loss": 0.7896, "step": 134500 }, { "epoch": 4.533818377795742, "grad_norm": 0.7368954420089722, "learning_rate": 3.108102038983203e-05, "loss": 0.7877, "step": 134600 }, { "epoch": 4.537186742118028, "grad_norm": 0.6391497850418091, "learning_rate": 3.085646276834635e-05, "loss": 0.7883, "step": 134700 }, { "epoch": 4.540555106440313, "grad_norm": 0.6790878176689148, "learning_rate": 3.063190514686068e-05, "loss": 0.7878, "step": 134800 }, { "epoch": 4.543923470762598, "grad_norm": 0.6800740361213684, "learning_rate": 3.0407347525375006e-05, "loss": 0.7879, "step": 134900 }, { "epoch": 4.5472918350848825, "grad_norm": 0.6817995309829712, "learning_rate": 3.0182789903889334e-05, "loss": 0.7893, "step": 135000 }, { "epoch": 4.550660199407168, "grad_norm": 0.706876814365387, "learning_rate": 2.995823228240366e-05, "loss": 0.7877, "step": 135100 }, { "epoch": 4.554028563729453, "grad_norm": 0.6842048168182373, "learning_rate": 2.9733674660917988e-05, "loss": 0.7887, "step": 135200 }, { "epoch": 4.557396928051738, "grad_norm": 0.8431084752082825, "learning_rate": 2.9509117039432316e-05, "loss": 0.7895, "step": 135300 }, { "epoch": 4.560765292374024, "grad_norm": 0.6965360045433044, "learning_rate": 2.928455941794664e-05, "loss": 0.7857, "step": 135400 }, { "epoch": 4.564133656696308, "grad_norm": 0.675323486328125, "learning_rate": 2.906000179646097e-05, "loss": 0.7917, "step": 135500 }, { "epoch": 4.567502021018593, "grad_norm": 0.6971179842948914, "learning_rate": 2.8835444174975295e-05, "loss": 0.791, "step": 135600 }, { "epoch": 4.5708703853408785, "grad_norm": 0.7096483111381531, "learning_rate": 2.8610886553489623e-05, "loss": 0.7886, "step": 135700 }, { "epoch": 4.574238749663164, "grad_norm": 0.6643834710121155, "learning_rate": 2.838632893200395e-05, "loss": 0.7903, "step": 135800 }, { "epoch": 4.577607113985449, "grad_norm": 0.6781139373779297, "learning_rate": 2.8161771310518276e-05, "loss": 0.7861, "step": 135900 }, { "epoch": 4.580975478307733, "grad_norm": 0.69442218542099, "learning_rate": 2.7937213689032605e-05, "loss": 0.7884, "step": 136000 }, { "epoch": 4.584343842630019, "grad_norm": 0.6849011778831482, "learning_rate": 2.771265606754693e-05, "loss": 0.7892, "step": 136100 }, { "epoch": 4.587712206952304, "grad_norm": 0.7133069634437561, "learning_rate": 2.7488098446061258e-05, "loss": 0.7904, "step": 136200 }, { "epoch": 4.591080571274589, "grad_norm": 0.7414596080780029, "learning_rate": 2.7263540824575583e-05, "loss": 0.7933, "step": 136300 }, { "epoch": 4.5944489355968745, "grad_norm": 0.7594481706619263, "learning_rate": 2.703898320308991e-05, "loss": 0.7901, "step": 136400 }, { "epoch": 4.597817299919159, "grad_norm": 0.7436200976371765, "learning_rate": 2.681442558160424e-05, "loss": 0.7863, "step": 136500 }, { "epoch": 4.601185664241444, "grad_norm": 0.7052866220474243, "learning_rate": 2.6589867960118562e-05, "loss": 0.7899, "step": 136600 }, { "epoch": 4.604554028563729, "grad_norm": 0.720867395401001, "learning_rate": 2.636531033863289e-05, "loss": 0.786, "step": 136700 }, { "epoch": 4.607922392886015, "grad_norm": 0.7140068411827087, "learning_rate": 2.6140752717147215e-05, "loss": 0.7886, "step": 136800 }, { "epoch": 4.6112907572083, "grad_norm": 0.6707795858383179, "learning_rate": 2.5916195095661544e-05, "loss": 0.792, "step": 136900 }, { "epoch": 4.614659121530584, "grad_norm": 0.6954900622367859, "learning_rate": 2.5691637474175872e-05, "loss": 0.785, "step": 137000 }, { "epoch": 4.61802748585287, "grad_norm": 0.695578396320343, "learning_rate": 2.5467079852690197e-05, "loss": 0.7897, "step": 137100 }, { "epoch": 4.621395850175155, "grad_norm": 0.8920716047286987, "learning_rate": 2.5242522231204525e-05, "loss": 0.7912, "step": 137200 }, { "epoch": 4.62476421449744, "grad_norm": 0.6703862547874451, "learning_rate": 2.501796460971885e-05, "loss": 0.79, "step": 137300 }, { "epoch": 4.628132578819725, "grad_norm": 0.7049610018730164, "learning_rate": 2.479340698823318e-05, "loss": 0.7876, "step": 137400 }, { "epoch": 4.63150094314201, "grad_norm": 0.7894124388694763, "learning_rate": 2.4568849366747504e-05, "loss": 0.7897, "step": 137500 }, { "epoch": 4.634869307464295, "grad_norm": 2.398905038833618, "learning_rate": 2.4344291745261832e-05, "loss": 0.7909, "step": 137600 }, { "epoch": 4.63823767178658, "grad_norm": 0.6975995302200317, "learning_rate": 2.411973412377616e-05, "loss": 0.789, "step": 137700 }, { "epoch": 4.641606036108866, "grad_norm": 0.717632532119751, "learning_rate": 2.3895176502290486e-05, "loss": 0.7883, "step": 137800 }, { "epoch": 4.644974400431151, "grad_norm": 0.7279338240623474, "learning_rate": 2.3670618880804814e-05, "loss": 0.787, "step": 137900 }, { "epoch": 4.648342764753436, "grad_norm": 0.6684018969535828, "learning_rate": 2.344606125931914e-05, "loss": 0.7875, "step": 138000 }, { "epoch": 4.6517111290757205, "grad_norm": 0.7201947569847107, "learning_rate": 2.3221503637833468e-05, "loss": 0.791, "step": 138100 }, { "epoch": 4.655079493398006, "grad_norm": 0.71978360414505, "learning_rate": 2.2996946016347796e-05, "loss": 0.7871, "step": 138200 }, { "epoch": 4.658447857720291, "grad_norm": 0.6821112632751465, "learning_rate": 2.2772388394862118e-05, "loss": 0.7888, "step": 138300 }, { "epoch": 4.661816222042576, "grad_norm": 0.7123835682868958, "learning_rate": 2.2547830773376446e-05, "loss": 0.7879, "step": 138400 }, { "epoch": 4.6651845863648616, "grad_norm": 0.6627094745635986, "learning_rate": 2.232327315189077e-05, "loss": 0.7901, "step": 138500 }, { "epoch": 4.668552950687147, "grad_norm": 0.6200422644615173, "learning_rate": 2.20987155304051e-05, "loss": 0.791, "step": 138600 }, { "epoch": 4.671921315009431, "grad_norm": 0.7372182011604309, "learning_rate": 2.1874157908919425e-05, "loss": 0.7846, "step": 138700 }, { "epoch": 4.6752896793317165, "grad_norm": 0.7178613543510437, "learning_rate": 2.1649600287433753e-05, "loss": 0.7903, "step": 138800 }, { "epoch": 4.678658043654002, "grad_norm": 0.7218438386917114, "learning_rate": 2.142504266594808e-05, "loss": 0.7881, "step": 138900 }, { "epoch": 4.682026407976287, "grad_norm": 0.7254891395568848, "learning_rate": 2.1200485044462406e-05, "loss": 0.7882, "step": 139000 }, { "epoch": 4.685394772298572, "grad_norm": 0.6931572556495667, "learning_rate": 2.0975927422976735e-05, "loss": 0.7878, "step": 139100 }, { "epoch": 4.688763136620857, "grad_norm": 0.6938092708587646, "learning_rate": 2.075136980149106e-05, "loss": 0.7859, "step": 139200 }, { "epoch": 4.692131500943142, "grad_norm": 0.7220520973205566, "learning_rate": 2.0526812180005388e-05, "loss": 0.7874, "step": 139300 }, { "epoch": 4.695499865265427, "grad_norm": 0.7422808408737183, "learning_rate": 2.0302254558519717e-05, "loss": 0.7861, "step": 139400 }, { "epoch": 4.698868229587712, "grad_norm": 0.6828125715255737, "learning_rate": 2.007769693703404e-05, "loss": 0.7864, "step": 139500 }, { "epoch": 4.702236593909998, "grad_norm": 0.7483955025672913, "learning_rate": 1.985313931554837e-05, "loss": 0.7847, "step": 139600 }, { "epoch": 4.705604958232282, "grad_norm": 0.7267476916313171, "learning_rate": 1.9628581694062695e-05, "loss": 0.788, "step": 139700 }, { "epoch": 4.708973322554567, "grad_norm": 0.6834740042686462, "learning_rate": 1.9404024072577024e-05, "loss": 0.7845, "step": 139800 }, { "epoch": 4.712341686876853, "grad_norm": 0.7835246324539185, "learning_rate": 1.9179466451091345e-05, "loss": 0.7856, "step": 139900 }, { "epoch": 4.715710051199138, "grad_norm": 0.6620598435401917, "learning_rate": 1.8954908829605674e-05, "loss": 0.7915, "step": 140000 }, { "epoch": 4.719078415521423, "grad_norm": 0.7632565498352051, "learning_rate": 1.8730351208120002e-05, "loss": 0.7878, "step": 140100 }, { "epoch": 4.7224467798437075, "grad_norm": 0.6834028363227844, "learning_rate": 1.8505793586634327e-05, "loss": 0.7852, "step": 140200 }, { "epoch": 4.725815144165993, "grad_norm": 0.8396201729774475, "learning_rate": 1.8281235965148655e-05, "loss": 0.7893, "step": 140300 }, { "epoch": 4.729183508488278, "grad_norm": 0.7341758608818054, "learning_rate": 1.8056678343662984e-05, "loss": 0.7892, "step": 140400 }, { "epoch": 4.732551872810563, "grad_norm": 0.74070143699646, "learning_rate": 1.783212072217731e-05, "loss": 0.7864, "step": 140500 }, { "epoch": 4.735920237132849, "grad_norm": 0.7226638793945312, "learning_rate": 1.7607563100691637e-05, "loss": 0.7861, "step": 140600 }, { "epoch": 4.739288601455133, "grad_norm": 0.6856757402420044, "learning_rate": 1.7383005479205962e-05, "loss": 0.7878, "step": 140700 }, { "epoch": 4.742656965777418, "grad_norm": 0.695364773273468, "learning_rate": 1.7158447857720287e-05, "loss": 0.788, "step": 140800 }, { "epoch": 4.7460253300997035, "grad_norm": 0.7357635498046875, "learning_rate": 1.6933890236234616e-05, "loss": 0.7853, "step": 140900 }, { "epoch": 4.749393694421989, "grad_norm": 0.6697304248809814, "learning_rate": 1.6709332614748944e-05, "loss": 0.7868, "step": 141000 }, { "epoch": 4.752762058744274, "grad_norm": 0.719422459602356, "learning_rate": 1.648477499326327e-05, "loss": 0.7887, "step": 141100 }, { "epoch": 4.756130423066558, "grad_norm": 0.650025486946106, "learning_rate": 1.6260217371777598e-05, "loss": 0.7901, "step": 141200 }, { "epoch": 4.759498787388844, "grad_norm": 0.6734464168548584, "learning_rate": 1.6035659750291923e-05, "loss": 0.7885, "step": 141300 }, { "epoch": 4.762867151711129, "grad_norm": 0.6939591765403748, "learning_rate": 1.581110212880625e-05, "loss": 0.7896, "step": 141400 }, { "epoch": 4.766235516033414, "grad_norm": 0.6683725118637085, "learning_rate": 1.558654450732058e-05, "loss": 0.7909, "step": 141500 }, { "epoch": 4.7696038803556995, "grad_norm": 0.7492696642875671, "learning_rate": 1.5361986885834905e-05, "loss": 0.7856, "step": 141600 }, { "epoch": 4.772972244677985, "grad_norm": 0.6370770931243896, "learning_rate": 1.5137429264349231e-05, "loss": 0.7858, "step": 141700 }, { "epoch": 4.776340609000269, "grad_norm": 0.7032959461212158, "learning_rate": 1.4912871642863558e-05, "loss": 0.7833, "step": 141800 }, { "epoch": 4.779708973322554, "grad_norm": 0.6765040159225464, "learning_rate": 1.4688314021377885e-05, "loss": 0.7861, "step": 141900 }, { "epoch": 4.78307733764484, "grad_norm": 0.6935933232307434, "learning_rate": 1.446375639989221e-05, "loss": 0.7834, "step": 142000 }, { "epoch": 4.786445701967125, "grad_norm": 0.716052770614624, "learning_rate": 1.4239198778406538e-05, "loss": 0.7857, "step": 142100 }, { "epoch": 4.78981406628941, "grad_norm": 0.7378547787666321, "learning_rate": 1.4014641156920865e-05, "loss": 0.7849, "step": 142200 }, { "epoch": 4.793182430611695, "grad_norm": 0.7271299958229065, "learning_rate": 1.3790083535435192e-05, "loss": 0.7877, "step": 142300 }, { "epoch": 4.79655079493398, "grad_norm": 0.6609006524085999, "learning_rate": 1.3565525913949518e-05, "loss": 0.7857, "step": 142400 }, { "epoch": 4.799919159256265, "grad_norm": 0.6987965703010559, "learning_rate": 1.3340968292463845e-05, "loss": 0.7848, "step": 142500 }, { "epoch": 4.80328752357855, "grad_norm": 0.6643743515014648, "learning_rate": 1.3116410670978172e-05, "loss": 0.7872, "step": 142600 }, { "epoch": 4.806655887900836, "grad_norm": 0.642590343952179, "learning_rate": 1.28918530494925e-05, "loss": 0.7858, "step": 142700 }, { "epoch": 4.810024252223121, "grad_norm": 0.6913971304893494, "learning_rate": 1.2667295428006827e-05, "loss": 0.7862, "step": 142800 }, { "epoch": 4.813392616545405, "grad_norm": 0.7790234684944153, "learning_rate": 1.2442737806521152e-05, "loss": 0.7869, "step": 142900 }, { "epoch": 4.8167609808676906, "grad_norm": 0.6860402226448059, "learning_rate": 1.2218180185035479e-05, "loss": 0.7872, "step": 143000 }, { "epoch": 4.820129345189976, "grad_norm": 0.73018479347229, "learning_rate": 1.1993622563549805e-05, "loss": 0.785, "step": 143100 }, { "epoch": 4.823497709512261, "grad_norm": 0.7606022357940674, "learning_rate": 1.1769064942064132e-05, "loss": 0.7835, "step": 143200 }, { "epoch": 4.826866073834546, "grad_norm": 0.7062585949897766, "learning_rate": 1.154450732057846e-05, "loss": 0.7863, "step": 143300 }, { "epoch": 4.830234438156831, "grad_norm": 0.7451142072677612, "learning_rate": 1.1319949699092787e-05, "loss": 0.7853, "step": 143400 }, { "epoch": 4.833602802479116, "grad_norm": 0.6708253622055054, "learning_rate": 1.1095392077607114e-05, "loss": 0.7872, "step": 143500 }, { "epoch": 4.836971166801401, "grad_norm": 0.6949145197868347, "learning_rate": 1.0870834456121439e-05, "loss": 0.7877, "step": 143600 }, { "epoch": 4.8403395311236865, "grad_norm": 0.7620200514793396, "learning_rate": 1.0646276834635766e-05, "loss": 0.784, "step": 143700 }, { "epoch": 4.843707895445972, "grad_norm": 0.6912148594856262, "learning_rate": 1.0421719213150092e-05, "loss": 0.7838, "step": 143800 }, { "epoch": 4.847076259768256, "grad_norm": 0.6649619936943054, "learning_rate": 1.019716159166442e-05, "loss": 0.7847, "step": 143900 }, { "epoch": 4.8504446240905414, "grad_norm": 0.6840892434120178, "learning_rate": 9.972603970178747e-06, "loss": 0.7842, "step": 144000 }, { "epoch": 4.853812988412827, "grad_norm": 0.6763936877250671, "learning_rate": 9.748046348693074e-06, "loss": 0.7884, "step": 144100 }, { "epoch": 4.857181352735112, "grad_norm": 0.7060201168060303, "learning_rate": 9.523488727207401e-06, "loss": 0.7869, "step": 144200 }, { "epoch": 4.860549717057397, "grad_norm": 0.6832892298698425, "learning_rate": 9.298931105721728e-06, "loss": 0.7846, "step": 144300 }, { "epoch": 4.863918081379682, "grad_norm": 0.6947805285453796, "learning_rate": 9.074373484236054e-06, "loss": 0.7876, "step": 144400 }, { "epoch": 4.867286445701967, "grad_norm": 0.6749753355979919, "learning_rate": 8.849815862750381e-06, "loss": 0.7848, "step": 144500 }, { "epoch": 4.870654810024252, "grad_norm": 0.7636469602584839, "learning_rate": 8.625258241264708e-06, "loss": 0.7846, "step": 144600 }, { "epoch": 4.874023174346537, "grad_norm": 0.6952735781669617, "learning_rate": 8.400700619779035e-06, "loss": 0.7889, "step": 144700 }, { "epoch": 4.877391538668823, "grad_norm": 0.6861172318458557, "learning_rate": 8.176142998293361e-06, "loss": 0.787, "step": 144800 }, { "epoch": 4.880759902991107, "grad_norm": 0.7696015238761902, "learning_rate": 7.951585376807688e-06, "loss": 0.7858, "step": 144900 }, { "epoch": 4.884128267313392, "grad_norm": 0.7360905408859253, "learning_rate": 7.727027755322015e-06, "loss": 0.7875, "step": 145000 }, { "epoch": 4.887496631635678, "grad_norm": 0.7402172088623047, "learning_rate": 7.502470133836341e-06, "loss": 0.7901, "step": 145100 }, { "epoch": 4.890864995957963, "grad_norm": 0.7550304532051086, "learning_rate": 7.277912512350668e-06, "loss": 0.7881, "step": 145200 }, { "epoch": 4.894233360280248, "grad_norm": 0.7490784525871277, "learning_rate": 7.053354890864996e-06, "loss": 0.7866, "step": 145300 }, { "epoch": 4.8976017246025325, "grad_norm": 0.675995945930481, "learning_rate": 6.828797269379322e-06, "loss": 0.7868, "step": 145400 }, { "epoch": 4.900970088924818, "grad_norm": 0.6862391829490662, "learning_rate": 6.604239647893648e-06, "loss": 0.7851, "step": 145500 }, { "epoch": 4.904338453247103, "grad_norm": 0.6939913630485535, "learning_rate": 6.379682026407976e-06, "loss": 0.7871, "step": 145600 }, { "epoch": 4.907706817569388, "grad_norm": 0.741671085357666, "learning_rate": 6.155124404922303e-06, "loss": 0.7829, "step": 145700 }, { "epoch": 4.911075181891674, "grad_norm": 0.6528514623641968, "learning_rate": 5.930566783436629e-06, "loss": 0.7869, "step": 145800 }, { "epoch": 4.914443546213959, "grad_norm": 0.6991773843765259, "learning_rate": 5.706009161950956e-06, "loss": 0.7847, "step": 145900 }, { "epoch": 4.917811910536243, "grad_norm": 0.6840393543243408, "learning_rate": 5.481451540465283e-06, "loss": 0.7874, "step": 146000 }, { "epoch": 4.9211802748585285, "grad_norm": 0.6707117557525635, "learning_rate": 5.2568939189796095e-06, "loss": 0.7874, "step": 146100 }, { "epoch": 4.924548639180814, "grad_norm": 0.7077836990356445, "learning_rate": 5.032336297493937e-06, "loss": 0.7858, "step": 146200 }, { "epoch": 4.927917003503099, "grad_norm": 0.6887540817260742, "learning_rate": 4.807778676008263e-06, "loss": 0.7834, "step": 146300 }, { "epoch": 4.931285367825384, "grad_norm": 0.710809051990509, "learning_rate": 4.5832210545225905e-06, "loss": 0.7855, "step": 146400 }, { "epoch": 4.93465373214767, "grad_norm": 0.6486634016036987, "learning_rate": 4.358663433036916e-06, "loss": 0.7833, "step": 146500 }, { "epoch": 4.938022096469954, "grad_norm": 0.7380653619766235, "learning_rate": 4.134105811551244e-06, "loss": 0.7825, "step": 146600 }, { "epoch": 4.941390460792239, "grad_norm": 0.672187328338623, "learning_rate": 3.909548190065571e-06, "loss": 0.7853, "step": 146700 }, { "epoch": 4.9447588251145245, "grad_norm": 0.7624046206474304, "learning_rate": 3.684990568579897e-06, "loss": 0.787, "step": 146800 }, { "epoch": 4.94812718943681, "grad_norm": 0.7582620978355408, "learning_rate": 3.460432947094224e-06, "loss": 0.7852, "step": 146900 }, { "epoch": 4.951495553759095, "grad_norm": 0.7094969153404236, "learning_rate": 3.235875325608551e-06, "loss": 0.7845, "step": 147000 }, { "epoch": 4.954863918081379, "grad_norm": 0.7196788787841797, "learning_rate": 3.0113177041228775e-06, "loss": 0.7844, "step": 147100 }, { "epoch": 4.958232282403665, "grad_norm": 0.6927585005760193, "learning_rate": 2.7867600826372042e-06, "loss": 0.7839, "step": 147200 }, { "epoch": 4.96160064672595, "grad_norm": 0.7253730893135071, "learning_rate": 2.5622024611515314e-06, "loss": 0.7838, "step": 147300 }, { "epoch": 4.964969011048235, "grad_norm": 0.6914359331130981, "learning_rate": 2.337644839665858e-06, "loss": 0.7853, "step": 147400 }, { "epoch": 4.9683373753705204, "grad_norm": 0.7005605101585388, "learning_rate": 2.113087218180185e-06, "loss": 0.7852, "step": 147500 }, { "epoch": 4.971705739692805, "grad_norm": 0.6463395357131958, "learning_rate": 1.8885295966945117e-06, "loss": 0.7868, "step": 147600 }, { "epoch": 4.97507410401509, "grad_norm": 0.6650646328926086, "learning_rate": 1.6639719752088385e-06, "loss": 0.7874, "step": 147700 }, { "epoch": 4.978442468337375, "grad_norm": 0.7527862191200256, "learning_rate": 1.4394143537231652e-06, "loss": 0.7866, "step": 147800 }, { "epoch": 4.981810832659661, "grad_norm": 0.7495971322059631, "learning_rate": 1.2148567322374921e-06, "loss": 0.786, "step": 147900 }, { "epoch": 4.985179196981946, "grad_norm": 0.6791830062866211, "learning_rate": 9.902991107518188e-07, "loss": 0.7863, "step": 148000 }, { "epoch": 4.98854756130423, "grad_norm": 0.6754565834999084, "learning_rate": 7.657414892661456e-07, "loss": 0.7832, "step": 148100 }, { "epoch": 4.9919159256265155, "grad_norm": 0.6369759440422058, "learning_rate": 5.411838677804724e-07, "loss": 0.7807, "step": 148200 }, { "epoch": 4.995284289948801, "grad_norm": 0.7656301259994507, "learning_rate": 3.166262462947992e-07, "loss": 0.7841, "step": 148300 }, { "epoch": 4.998652654271086, "grad_norm": 0.6411105394363403, "learning_rate": 9.206862480912601e-08, "loss": 0.787, "step": 148400 }, { "epoch": 5.0, "eval_loss": 0.7365977168083191, "eval_runtime": 1.819, "eval_samples_per_second": 2748.792, "eval_steps_per_second": 43.431, "step": 148440 } ], "logging_steps": 100, "max_steps": 148440, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8701968384000000.0, "train_batch_size": 32, "trial_name": null, "trial_params": null }