{ "best_metric": 3.3023488521575928, "best_model_checkpoint": "/scratch/cl5625/exceptions/models/100M__1208/checkpoint-90000", "epoch": 10.242587601078167, "eval_steps": 1000, "global_step": 95000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005390835579514825, "grad_norm": 1.3872060775756836, "learning_rate": 0.0003, "loss": 8.7938, "step": 50 }, { "epoch": 0.01078167115902965, "grad_norm": 3.891347885131836, "learning_rate": 0.0006, "loss": 6.992, "step": 100 }, { "epoch": 0.016172506738544475, "grad_norm": 1.848716378211975, "learning_rate": 0.0005996762007555315, "loss": 6.5045, "step": 150 }, { "epoch": 0.0215633423180593, "grad_norm": 1.3053127527236938, "learning_rate": 0.000599352401511063, "loss": 6.2433, "step": 200 }, { "epoch": 0.026954177897574125, "grad_norm": 1.2040534019470215, "learning_rate": 0.0005990286022665946, "loss": 6.0719, "step": 250 }, { "epoch": 0.03234501347708895, "grad_norm": 1.105197548866272, "learning_rate": 0.0005987048030221263, "loss": 5.9606, "step": 300 }, { "epoch": 0.03773584905660377, "grad_norm": 1.502025842666626, "learning_rate": 0.0005983810037776578, "loss": 5.8818, "step": 350 }, { "epoch": 0.0431266846361186, "grad_norm": 2.493607759475708, "learning_rate": 0.0005980572045331894, "loss": 5.8101, "step": 400 }, { "epoch": 0.04851752021563342, "grad_norm": 1.3217182159423828, "learning_rate": 0.0005977334052887209, "loss": 5.7175, "step": 450 }, { "epoch": 0.05390835579514825, "grad_norm": 1.215240240097046, "learning_rate": 0.0005974096060442526, "loss": 5.6639, "step": 500 }, { "epoch": 0.05929919137466307, "grad_norm": 1.3821635246276855, "learning_rate": 0.0005970858067997841, "loss": 5.5744, "step": 550 }, { "epoch": 0.0646900269541779, "grad_norm": 1.378275752067566, "learning_rate": 0.0005967620075553157, "loss": 5.5287, "step": 600 }, { "epoch": 0.07008086253369272, "grad_norm": 1.1444066762924194, "learning_rate": 0.0005964382083108472, "loss": 5.4332, "step": 650 }, { "epoch": 0.07547169811320754, "grad_norm": 1.657812476158142, "learning_rate": 0.0005961144090663788, "loss": 5.4098, "step": 700 }, { "epoch": 0.08086253369272237, "grad_norm": 1.3864067792892456, "learning_rate": 0.0005957906098219104, "loss": 5.3168, "step": 750 }, { "epoch": 0.0862533692722372, "grad_norm": 1.0762931108474731, "learning_rate": 0.0005954668105774419, "loss": 5.2678, "step": 800 }, { "epoch": 0.09164420485175202, "grad_norm": 1.2499072551727295, "learning_rate": 0.0005951430113329735, "loss": 5.2107, "step": 850 }, { "epoch": 0.09703504043126684, "grad_norm": 0.9282752275466919, "learning_rate": 0.0005948192120885051, "loss": 5.1849, "step": 900 }, { "epoch": 0.10242587601078167, "grad_norm": 1.071282982826233, "learning_rate": 0.0005944954128440366, "loss": 5.1275, "step": 950 }, { "epoch": 0.1078167115902965, "grad_norm": 1.3316810131072998, "learning_rate": 0.0005941716135995682, "loss": 5.1151, "step": 1000 }, { "epoch": 0.1078167115902965, "eval_accuracy": 0.22672739349001877, "eval_loss": 5.027144432067871, "eval_runtime": 184.654, "eval_samples_per_second": 97.539, "eval_steps_per_second": 6.098, "step": 1000 }, { "epoch": 0.11320754716981132, "grad_norm": 1.1736186742782593, "learning_rate": 0.0005938478143550997, "loss": 5.0742, "step": 1050 }, { "epoch": 0.11859838274932614, "grad_norm": 1.217553973197937, "learning_rate": 0.0005935240151106314, "loss": 5.0002, "step": 1100 }, { "epoch": 0.12398921832884097, "grad_norm": 1.3116732835769653, "learning_rate": 0.0005932002158661629, "loss": 4.9959, "step": 1150 }, { "epoch": 0.1293800539083558, "grad_norm": 1.10372793674469, "learning_rate": 0.0005928764166216945, "loss": 4.9613, "step": 1200 }, { "epoch": 0.1347708894878706, "grad_norm": 1.0251359939575195, "learning_rate": 0.000592552617377226, "loss": 4.9081, "step": 1250 }, { "epoch": 0.14016172506738545, "grad_norm": 1.1212270259857178, "learning_rate": 0.0005922288181327577, "loss": 4.9098, "step": 1300 }, { "epoch": 0.14555256064690028, "grad_norm": 1.0358482599258423, "learning_rate": 0.0005919050188882893, "loss": 4.8835, "step": 1350 }, { "epoch": 0.1509433962264151, "grad_norm": 1.127200722694397, "learning_rate": 0.0005915812196438207, "loss": 4.85, "step": 1400 }, { "epoch": 0.15633423180592992, "grad_norm": 0.9541453123092651, "learning_rate": 0.0005912574203993524, "loss": 4.8199, "step": 1450 }, { "epoch": 0.16172506738544473, "grad_norm": 0.9905341863632202, "learning_rate": 0.0005909336211548839, "loss": 4.812, "step": 1500 }, { "epoch": 0.16711590296495957, "grad_norm": 0.9883430600166321, "learning_rate": 0.0005906098219104155, "loss": 4.7776, "step": 1550 }, { "epoch": 0.1725067385444744, "grad_norm": 1.14104425907135, "learning_rate": 0.000590286022665947, "loss": 4.7877, "step": 1600 }, { "epoch": 0.1778975741239892, "grad_norm": 0.915554404258728, "learning_rate": 0.0005899622234214787, "loss": 4.7362, "step": 1650 }, { "epoch": 0.18328840970350405, "grad_norm": 0.8896968364715576, "learning_rate": 0.0005896384241770102, "loss": 4.6944, "step": 1700 }, { "epoch": 0.18867924528301888, "grad_norm": 0.7295605540275574, "learning_rate": 0.0005893146249325418, "loss": 4.679, "step": 1750 }, { "epoch": 0.1940700808625337, "grad_norm": 0.8035542368888855, "learning_rate": 0.0005889908256880733, "loss": 4.669, "step": 1800 }, { "epoch": 0.19946091644204852, "grad_norm": 0.9095346927642822, "learning_rate": 0.0005886670264436049, "loss": 4.6768, "step": 1850 }, { "epoch": 0.20485175202156333, "grad_norm": 0.8672581315040588, "learning_rate": 0.0005883432271991365, "loss": 4.6219, "step": 1900 }, { "epoch": 0.21024258760107817, "grad_norm": 0.7349112033843994, "learning_rate": 0.0005880194279546681, "loss": 4.6077, "step": 1950 }, { "epoch": 0.215633423180593, "grad_norm": 0.9716205596923828, "learning_rate": 0.0005876956287101996, "loss": 4.5792, "step": 2000 }, { "epoch": 0.215633423180593, "eval_accuracy": 0.27099464439050674, "eval_loss": 4.504695415496826, "eval_runtime": 184.1142, "eval_samples_per_second": 97.825, "eval_steps_per_second": 6.116, "step": 2000 }, { "epoch": 0.2210242587601078, "grad_norm": 1.203096628189087, "learning_rate": 0.0005873718294657312, "loss": 4.5695, "step": 2050 }, { "epoch": 0.22641509433962265, "grad_norm": 0.9952515959739685, "learning_rate": 0.0005870480302212628, "loss": 4.5412, "step": 2100 }, { "epoch": 0.23180592991913745, "grad_norm": 0.8814882040023804, "learning_rate": 0.0005867242309767943, "loss": 4.5392, "step": 2150 }, { "epoch": 0.2371967654986523, "grad_norm": 1.0191090106964111, "learning_rate": 0.0005864004317323259, "loss": 4.5151, "step": 2200 }, { "epoch": 0.24258760107816713, "grad_norm": 0.8580870628356934, "learning_rate": 0.0005860766324878575, "loss": 4.5014, "step": 2250 }, { "epoch": 0.24797843665768193, "grad_norm": 0.8280946016311646, "learning_rate": 0.000585752833243389, "loss": 4.4755, "step": 2300 }, { "epoch": 0.25336927223719674, "grad_norm": 0.9512032866477966, "learning_rate": 0.0005854290339989206, "loss": 4.4624, "step": 2350 }, { "epoch": 0.2587601078167116, "grad_norm": 1.000420093536377, "learning_rate": 0.0005851052347544521, "loss": 4.4523, "step": 2400 }, { "epoch": 0.2641509433962264, "grad_norm": 0.9130716919898987, "learning_rate": 0.0005847814355099838, "loss": 4.4414, "step": 2450 }, { "epoch": 0.2695417789757412, "grad_norm": 0.8131120800971985, "learning_rate": 0.0005844576362655154, "loss": 4.4556, "step": 2500 }, { "epoch": 0.2749326145552561, "grad_norm": 0.9587961435317993, "learning_rate": 0.0005841338370210469, "loss": 4.4083, "step": 2550 }, { "epoch": 0.2803234501347709, "grad_norm": 0.846928060054779, "learning_rate": 0.0005838100377765785, "loss": 4.3825, "step": 2600 }, { "epoch": 0.2857142857142857, "grad_norm": 0.9020141363143921, "learning_rate": 0.0005834862385321101, "loss": 4.3869, "step": 2650 }, { "epoch": 0.29110512129380056, "grad_norm": 1.0460323095321655, "learning_rate": 0.0005831624392876417, "loss": 4.381, "step": 2700 }, { "epoch": 0.29649595687331537, "grad_norm": 0.8714868426322937, "learning_rate": 0.0005828386400431731, "loss": 4.3618, "step": 2750 }, { "epoch": 0.3018867924528302, "grad_norm": 0.7531731724739075, "learning_rate": 0.0005825148407987048, "loss": 4.361, "step": 2800 }, { "epoch": 0.30727762803234504, "grad_norm": 0.7624960541725159, "learning_rate": 0.0005821910415542363, "loss": 4.3642, "step": 2850 }, { "epoch": 0.31266846361185985, "grad_norm": 0.8426908850669861, "learning_rate": 0.0005818672423097679, "loss": 4.3367, "step": 2900 }, { "epoch": 0.31805929919137466, "grad_norm": 0.9197669625282288, "learning_rate": 0.0005815434430652994, "loss": 4.3496, "step": 2950 }, { "epoch": 0.32345013477088946, "grad_norm": 0.8955463171005249, "learning_rate": 0.0005812196438208311, "loss": 4.2988, "step": 3000 }, { "epoch": 0.32345013477088946, "eval_accuracy": 0.2984499253065723, "eval_loss": 4.238968372344971, "eval_runtime": 185.178, "eval_samples_per_second": 97.263, "eval_steps_per_second": 6.081, "step": 3000 }, { "epoch": 0.3288409703504043, "grad_norm": 0.7810043096542358, "learning_rate": 0.0005808958445763626, "loss": 4.3082, "step": 3050 }, { "epoch": 0.33423180592991913, "grad_norm": 0.7354643940925598, "learning_rate": 0.0005805720453318942, "loss": 4.2858, "step": 3100 }, { "epoch": 0.33962264150943394, "grad_norm": 0.7374374270439148, "learning_rate": 0.0005802482460874257, "loss": 4.2772, "step": 3150 }, { "epoch": 0.3450134770889488, "grad_norm": 0.7399438619613647, "learning_rate": 0.0005799244468429573, "loss": 4.2993, "step": 3200 }, { "epoch": 0.3504043126684636, "grad_norm": 0.7688404321670532, "learning_rate": 0.0005796006475984889, "loss": 4.2841, "step": 3250 }, { "epoch": 0.3557951482479784, "grad_norm": 0.7221906781196594, "learning_rate": 0.0005792768483540205, "loss": 4.2667, "step": 3300 }, { "epoch": 0.3611859838274933, "grad_norm": 0.7419180274009705, "learning_rate": 0.000578953049109552, "loss": 4.2709, "step": 3350 }, { "epoch": 0.3665768194070081, "grad_norm": 0.7906516790390015, "learning_rate": 0.0005786292498650836, "loss": 4.2394, "step": 3400 }, { "epoch": 0.3719676549865229, "grad_norm": 0.7784256339073181, "learning_rate": 0.0005783054506206152, "loss": 4.2519, "step": 3450 }, { "epoch": 0.37735849056603776, "grad_norm": 0.7829201221466064, "learning_rate": 0.0005779816513761467, "loss": 4.2314, "step": 3500 }, { "epoch": 0.38274932614555257, "grad_norm": 0.8310431241989136, "learning_rate": 0.0005776578521316782, "loss": 4.2097, "step": 3550 }, { "epoch": 0.3881401617250674, "grad_norm": 0.8430302739143372, "learning_rate": 0.0005773340528872099, "loss": 4.2306, "step": 3600 }, { "epoch": 0.3935309973045822, "grad_norm": 0.733945906162262, "learning_rate": 0.0005770102536427414, "loss": 4.2221, "step": 3650 }, { "epoch": 0.39892183288409705, "grad_norm": 0.6383342742919922, "learning_rate": 0.000576686454398273, "loss": 4.2028, "step": 3700 }, { "epoch": 0.40431266846361186, "grad_norm": 0.6672252416610718, "learning_rate": 0.0005763626551538045, "loss": 4.1801, "step": 3750 }, { "epoch": 0.40970350404312667, "grad_norm": 0.7545523643493652, "learning_rate": 0.0005760388559093362, "loss": 4.1881, "step": 3800 }, { "epoch": 0.41509433962264153, "grad_norm": 0.7340240478515625, "learning_rate": 0.0005757150566648678, "loss": 4.171, "step": 3850 }, { "epoch": 0.42048517520215634, "grad_norm": 0.7903727889060974, "learning_rate": 0.0005753912574203993, "loss": 4.1784, "step": 3900 }, { "epoch": 0.42587601078167114, "grad_norm": 0.8018081188201904, "learning_rate": 0.0005750674581759309, "loss": 4.1838, "step": 3950 }, { "epoch": 0.431266846361186, "grad_norm": 0.722425103187561, "learning_rate": 0.0005747436589314624, "loss": 4.1617, "step": 4000 }, { "epoch": 0.431266846361186, "eval_accuracy": 0.31149631215800827, "eval_loss": 4.093442440032959, "eval_runtime": 185.3481, "eval_samples_per_second": 97.174, "eval_steps_per_second": 6.075, "step": 4000 }, { "epoch": 0.4366576819407008, "grad_norm": 0.9414203763008118, "learning_rate": 0.0005744198596869941, "loss": 4.15, "step": 4050 }, { "epoch": 0.4420485175202156, "grad_norm": 0.7414451241493225, "learning_rate": 0.0005740960604425255, "loss": 4.1367, "step": 4100 }, { "epoch": 0.4474393530997305, "grad_norm": 0.693946361541748, "learning_rate": 0.0005737722611980572, "loss": 4.1503, "step": 4150 }, { "epoch": 0.4528301886792453, "grad_norm": 0.8669963479042053, "learning_rate": 0.0005734484619535887, "loss": 4.1342, "step": 4200 }, { "epoch": 0.4582210242587601, "grad_norm": 0.717522144317627, "learning_rate": 0.0005731246627091203, "loss": 4.1361, "step": 4250 }, { "epoch": 0.4636118598382749, "grad_norm": 0.5836500525474548, "learning_rate": 0.0005728008634646518, "loss": 4.1184, "step": 4300 }, { "epoch": 0.46900269541778977, "grad_norm": 0.6450395584106445, "learning_rate": 0.0005724770642201835, "loss": 4.1149, "step": 4350 }, { "epoch": 0.4743935309973046, "grad_norm": 0.6113871335983276, "learning_rate": 0.000572153264975715, "loss": 4.1102, "step": 4400 }, { "epoch": 0.4797843665768194, "grad_norm": 0.7142659425735474, "learning_rate": 0.0005718294657312466, "loss": 4.1382, "step": 4450 }, { "epoch": 0.48517520215633425, "grad_norm": 0.6929596066474915, "learning_rate": 0.0005715056664867781, "loss": 4.0957, "step": 4500 }, { "epoch": 0.49056603773584906, "grad_norm": 0.5346873998641968, "learning_rate": 0.0005711818672423097, "loss": 4.1046, "step": 4550 }, { "epoch": 0.49595687331536387, "grad_norm": 0.702147901058197, "learning_rate": 0.0005708580679978413, "loss": 4.1, "step": 4600 }, { "epoch": 0.5013477088948787, "grad_norm": 0.5867359042167664, "learning_rate": 0.0005705342687533729, "loss": 4.0953, "step": 4650 }, { "epoch": 0.5067385444743935, "grad_norm": 0.6331294178962708, "learning_rate": 0.0005702104695089044, "loss": 4.0979, "step": 4700 }, { "epoch": 0.5121293800539084, "grad_norm": 0.6199395656585693, "learning_rate": 0.000569886670264436, "loss": 4.0987, "step": 4750 }, { "epoch": 0.5175202156334232, "grad_norm": 0.6783130764961243, "learning_rate": 0.0005695628710199675, "loss": 4.0776, "step": 4800 }, { "epoch": 0.522911051212938, "grad_norm": 0.6518664956092834, "learning_rate": 0.0005692390717754991, "loss": 4.1048, "step": 4850 }, { "epoch": 0.5283018867924528, "grad_norm": 0.6022597551345825, "learning_rate": 0.0005689152725310306, "loss": 4.0778, "step": 4900 }, { "epoch": 0.5336927223719676, "grad_norm": 0.6618348360061646, "learning_rate": 0.0005685914732865623, "loss": 4.0617, "step": 4950 }, { "epoch": 0.5390835579514824, "grad_norm": 0.5986215472221375, "learning_rate": 0.0005682676740420939, "loss": 4.0655, "step": 5000 }, { "epoch": 0.5390835579514824, "eval_accuracy": 0.32159864036122304, "eval_loss": 3.9903128147125244, "eval_runtime": 185.2705, "eval_samples_per_second": 97.215, "eval_steps_per_second": 6.078, "step": 5000 }, { "epoch": 0.5444743935309974, "grad_norm": 0.5965196490287781, "learning_rate": 0.0005679438747976254, "loss": 4.0629, "step": 5050 }, { "epoch": 0.5498652291105122, "grad_norm": 0.618198812007904, "learning_rate": 0.000567620075553157, "loss": 4.048, "step": 5100 }, { "epoch": 0.555256064690027, "grad_norm": 0.6078091263771057, "learning_rate": 0.0005672962763086886, "loss": 4.0601, "step": 5150 }, { "epoch": 0.5606469002695418, "grad_norm": 0.5713509917259216, "learning_rate": 0.0005669724770642202, "loss": 4.054, "step": 5200 }, { "epoch": 0.5660377358490566, "grad_norm": 0.7706556916236877, "learning_rate": 0.0005666486778197517, "loss": 4.0499, "step": 5250 }, { "epoch": 0.5714285714285714, "grad_norm": 0.6719933748245239, "learning_rate": 0.0005663248785752833, "loss": 4.0337, "step": 5300 }, { "epoch": 0.5768194070080862, "grad_norm": 0.6685216426849365, "learning_rate": 0.0005660010793308148, "loss": 4.0071, "step": 5350 }, { "epoch": 0.5822102425876011, "grad_norm": 0.6304830312728882, "learning_rate": 0.0005656772800863465, "loss": 4.0476, "step": 5400 }, { "epoch": 0.5876010781671159, "grad_norm": 0.6301653385162354, "learning_rate": 0.0005653534808418779, "loss": 4.0282, "step": 5450 }, { "epoch": 0.5929919137466307, "grad_norm": 0.6767484545707703, "learning_rate": 0.0005650296815974096, "loss": 4.0349, "step": 5500 }, { "epoch": 0.5983827493261455, "grad_norm": 0.6710530519485474, "learning_rate": 0.0005647058823529411, "loss": 4.0179, "step": 5550 }, { "epoch": 0.6037735849056604, "grad_norm": 0.5611121654510498, "learning_rate": 0.0005643820831084727, "loss": 4.0202, "step": 5600 }, { "epoch": 0.6091644204851752, "grad_norm": 0.6187557578086853, "learning_rate": 0.0005640582838640042, "loss": 3.9935, "step": 5650 }, { "epoch": 0.6145552560646901, "grad_norm": 0.6690448522567749, "learning_rate": 0.0005637344846195358, "loss": 4.0082, "step": 5700 }, { "epoch": 0.6199460916442049, "grad_norm": 0.6771255135536194, "learning_rate": 0.0005634106853750674, "loss": 4.0026, "step": 5750 }, { "epoch": 0.6253369272237197, "grad_norm": 0.5823601484298706, "learning_rate": 0.000563086886130599, "loss": 4.014, "step": 5800 }, { "epoch": 0.6307277628032345, "grad_norm": 0.6157869100570679, "learning_rate": 0.0005627630868861305, "loss": 3.9952, "step": 5850 }, { "epoch": 0.6361185983827493, "grad_norm": 0.6894795894622803, "learning_rate": 0.0005624392876416621, "loss": 3.9947, "step": 5900 }, { "epoch": 0.6415094339622641, "grad_norm": 0.6236327290534973, "learning_rate": 0.0005621154883971937, "loss": 3.996, "step": 5950 }, { "epoch": 0.6469002695417789, "grad_norm": 0.6344878077507019, "learning_rate": 0.0005617916891527253, "loss": 3.9934, "step": 6000 }, { "epoch": 0.6469002695417789, "eval_accuracy": 0.3280596843351111, "eval_loss": 3.924257278442383, "eval_runtime": 185.4361, "eval_samples_per_second": 97.128, "eval_steps_per_second": 6.072, "step": 6000 }, { "epoch": 0.6522911051212938, "grad_norm": 0.5802189111709595, "learning_rate": 0.0005614743658931463, "loss": 3.9855, "step": 6050 }, { "epoch": 0.6576819407008087, "grad_norm": 0.6164999604225159, "learning_rate": 0.0005611505666486777, "loss": 4.0042, "step": 6100 }, { "epoch": 0.6630727762803235, "grad_norm": 0.6946608424186707, "learning_rate": 0.0005608267674042094, "loss": 3.989, "step": 6150 }, { "epoch": 0.6684636118598383, "grad_norm": 0.6158466935157776, "learning_rate": 0.0005605029681597409, "loss": 3.9871, "step": 6200 }, { "epoch": 0.6738544474393531, "grad_norm": 0.6288596391677856, "learning_rate": 0.0005601791689152725, "loss": 3.9773, "step": 6250 }, { "epoch": 0.6792452830188679, "grad_norm": 0.7042039036750793, "learning_rate": 0.000559855369670804, "loss": 3.9791, "step": 6300 }, { "epoch": 0.6846361185983828, "grad_norm": 0.5582059025764465, "learning_rate": 0.0005595315704263357, "loss": 3.9639, "step": 6350 }, { "epoch": 0.6900269541778976, "grad_norm": 0.6060189604759216, "learning_rate": 0.0005592077711818672, "loss": 3.9775, "step": 6400 }, { "epoch": 0.6954177897574124, "grad_norm": 0.7476151585578918, "learning_rate": 0.0005588839719373988, "loss": 3.9536, "step": 6450 }, { "epoch": 0.7008086253369272, "grad_norm": 0.6221930384635925, "learning_rate": 0.0005585601726929303, "loss": 3.9583, "step": 6500 }, { "epoch": 0.706199460916442, "grad_norm": 0.6130439639091492, "learning_rate": 0.0005582363734484619, "loss": 3.9417, "step": 6550 }, { "epoch": 0.7115902964959568, "grad_norm": 0.5701190233230591, "learning_rate": 0.0005579125742039935, "loss": 3.9488, "step": 6600 }, { "epoch": 0.7169811320754716, "grad_norm": 0.5955173969268799, "learning_rate": 0.0005575887749595251, "loss": 3.9511, "step": 6650 }, { "epoch": 0.7223719676549866, "grad_norm": 0.6578053832054138, "learning_rate": 0.0005572649757150566, "loss": 3.9526, "step": 6700 }, { "epoch": 0.7277628032345014, "grad_norm": 0.7318617105484009, "learning_rate": 0.0005569411764705882, "loss": 3.9511, "step": 6750 }, { "epoch": 0.7331536388140162, "grad_norm": 0.6341660618782043, "learning_rate": 0.0005566173772261198, "loss": 3.918, "step": 6800 }, { "epoch": 0.738544474393531, "grad_norm": 0.5913158059120178, "learning_rate": 0.0005562935779816513, "loss": 3.9336, "step": 6850 }, { "epoch": 0.7439353099730458, "grad_norm": 0.6217501759529114, "learning_rate": 0.0005559697787371828, "loss": 3.9485, "step": 6900 }, { "epoch": 0.7493261455525606, "grad_norm": 0.629990816116333, "learning_rate": 0.0005556459794927145, "loss": 3.9304, "step": 6950 }, { "epoch": 0.7547169811320755, "grad_norm": 0.598209798336029, "learning_rate": 0.000555322180248246, "loss": 3.9432, "step": 7000 }, { "epoch": 0.7547169811320755, "eval_accuracy": 0.33339541034990466, "eval_loss": 3.867119312286377, "eval_runtime": 185.4718, "eval_samples_per_second": 97.109, "eval_steps_per_second": 6.071, "step": 7000 }, { "epoch": 0.7601078167115903, "grad_norm": 0.7586105465888977, "learning_rate": 0.0005549983810037776, "loss": 3.911, "step": 7050 }, { "epoch": 0.7654986522911051, "grad_norm": 0.6012836694717407, "learning_rate": 0.0005546745817593091, "loss": 3.9126, "step": 7100 }, { "epoch": 0.77088948787062, "grad_norm": 0.6788042187690735, "learning_rate": 0.0005543507825148408, "loss": 3.9149, "step": 7150 }, { "epoch": 0.7762803234501348, "grad_norm": 0.60019451379776, "learning_rate": 0.0005540269832703723, "loss": 3.8948, "step": 7200 }, { "epoch": 0.7816711590296496, "grad_norm": 0.5468128323554993, "learning_rate": 0.0005537031840259039, "loss": 3.9228, "step": 7250 }, { "epoch": 0.7870619946091644, "grad_norm": 0.560189425945282, "learning_rate": 0.0005533793847814354, "loss": 3.8974, "step": 7300 }, { "epoch": 0.7924528301886793, "grad_norm": 0.6238011717796326, "learning_rate": 0.000553055585536967, "loss": 3.9169, "step": 7350 }, { "epoch": 0.7978436657681941, "grad_norm": 0.6309143304824829, "learning_rate": 0.0005527317862924987, "loss": 3.9167, "step": 7400 }, { "epoch": 0.8032345013477089, "grad_norm": 0.5579202771186829, "learning_rate": 0.0005524079870480301, "loss": 3.9102, "step": 7450 }, { "epoch": 0.8086253369272237, "grad_norm": 0.637144923210144, "learning_rate": 0.0005520841878035618, "loss": 3.9305, "step": 7500 }, { "epoch": 0.8140161725067385, "grad_norm": 0.560736358165741, "learning_rate": 0.0005517603885590933, "loss": 3.9012, "step": 7550 }, { "epoch": 0.8194070080862533, "grad_norm": 0.5624358057975769, "learning_rate": 0.0005514365893146249, "loss": 3.8973, "step": 7600 }, { "epoch": 0.8247978436657682, "grad_norm": 0.5344393253326416, "learning_rate": 0.0005511127900701564, "loss": 3.8894, "step": 7650 }, { "epoch": 0.8301886792452831, "grad_norm": 0.5487794280052185, "learning_rate": 0.000550788990825688, "loss": 3.9173, "step": 7700 }, { "epoch": 0.8355795148247979, "grad_norm": 0.5416852831840515, "learning_rate": 0.0005504651915812196, "loss": 3.9131, "step": 7750 }, { "epoch": 0.8409703504043127, "grad_norm": 0.6219443678855896, "learning_rate": 0.0005501413923367512, "loss": 3.8783, "step": 7800 }, { "epoch": 0.8463611859838275, "grad_norm": 0.5484825968742371, "learning_rate": 0.0005498175930922827, "loss": 3.8889, "step": 7850 }, { "epoch": 0.8517520215633423, "grad_norm": 0.5506182909011841, "learning_rate": 0.0005494937938478143, "loss": 3.8945, "step": 7900 }, { "epoch": 0.8571428571428571, "grad_norm": 0.5658506155014038, "learning_rate": 0.0005491699946033459, "loss": 3.883, "step": 7950 }, { "epoch": 0.862533692722372, "grad_norm": 0.5695346593856812, "learning_rate": 0.0005488461953588775, "loss": 3.8869, "step": 8000 }, { "epoch": 0.862533692722372, "eval_accuracy": 0.33750498852571176, "eval_loss": 3.818067789077759, "eval_runtime": 185.1439, "eval_samples_per_second": 97.281, "eval_steps_per_second": 6.082, "step": 8000 }, { "epoch": 0.8679245283018868, "grad_norm": 0.5362111926078796, "learning_rate": 0.000548522396114409, "loss": 3.8853, "step": 8050 }, { "epoch": 0.8733153638814016, "grad_norm": 0.6255955696105957, "learning_rate": 0.0005482050728548299, "loss": 3.8955, "step": 8100 }, { "epoch": 0.8787061994609164, "grad_norm": 0.4989166557788849, "learning_rate": 0.0005478812736103615, "loss": 3.8747, "step": 8150 }, { "epoch": 0.8840970350404312, "grad_norm": 0.5338106751441956, "learning_rate": 0.0005475574743658931, "loss": 3.8772, "step": 8200 }, { "epoch": 0.889487870619946, "grad_norm": 0.6587180495262146, "learning_rate": 0.0005472336751214246, "loss": 3.8723, "step": 8250 }, { "epoch": 0.894878706199461, "grad_norm": 0.6544439196586609, "learning_rate": 0.0005469098758769562, "loss": 3.856, "step": 8300 }, { "epoch": 0.9002695417789758, "grad_norm": 0.5896086692810059, "learning_rate": 0.0005465860766324878, "loss": 3.8439, "step": 8350 }, { "epoch": 0.9056603773584906, "grad_norm": 0.579795241355896, "learning_rate": 0.0005462622773880194, "loss": 3.892, "step": 8400 }, { "epoch": 0.9110512129380054, "grad_norm": 0.6916389465332031, "learning_rate": 0.000545938478143551, "loss": 3.8569, "step": 8450 }, { "epoch": 0.9164420485175202, "grad_norm": 0.5650646686553955, "learning_rate": 0.0005456146788990825, "loss": 3.8475, "step": 8500 }, { "epoch": 0.921832884097035, "grad_norm": 0.5492226481437683, "learning_rate": 0.000545290879654614, "loss": 3.86, "step": 8550 }, { "epoch": 0.9272237196765498, "grad_norm": 0.6170901656150818, "learning_rate": 0.0005449670804101457, "loss": 3.86, "step": 8600 }, { "epoch": 0.9326145552560647, "grad_norm": 0.605499804019928, "learning_rate": 0.0005446432811656773, "loss": 3.8616, "step": 8650 }, { "epoch": 0.9380053908355795, "grad_norm": 0.5699788331985474, "learning_rate": 0.0005443194819212088, "loss": 3.8398, "step": 8700 }, { "epoch": 0.9433962264150944, "grad_norm": 0.6459490656852722, "learning_rate": 0.0005439956826767404, "loss": 3.8561, "step": 8750 }, { "epoch": 0.9487870619946092, "grad_norm": 0.5878280401229858, "learning_rate": 0.000543671883432272, "loss": 3.8437, "step": 8800 }, { "epoch": 0.954177897574124, "grad_norm": 0.5819820165634155, "learning_rate": 0.0005433480841878035, "loss": 3.8392, "step": 8850 }, { "epoch": 0.9595687331536388, "grad_norm": 0.6361739635467529, "learning_rate": 0.000543024284943335, "loss": 3.8312, "step": 8900 }, { "epoch": 0.9649595687331537, "grad_norm": 0.6311874389648438, "learning_rate": 0.0005427004856988667, "loss": 3.8481, "step": 8950 }, { "epoch": 0.9703504043126685, "grad_norm": 0.6704027056694031, "learning_rate": 0.0005423766864543982, "loss": 3.848, "step": 9000 }, { "epoch": 0.9703504043126685, "eval_accuracy": 0.3414659295509887, "eval_loss": 3.779182195663452, "eval_runtime": 184.1503, "eval_samples_per_second": 97.806, "eval_steps_per_second": 6.115, "step": 9000 }, { "epoch": 0.9757412398921833, "grad_norm": 0.5311286449432373, "learning_rate": 0.0005420528872099298, "loss": 3.8298, "step": 9050 }, { "epoch": 0.9811320754716981, "grad_norm": 0.5362455248832703, "learning_rate": 0.0005417290879654613, "loss": 3.8289, "step": 9100 }, { "epoch": 0.9865229110512129, "grad_norm": 0.6195308566093445, "learning_rate": 0.000541405288720993, "loss": 3.8465, "step": 9150 }, { "epoch": 0.9919137466307277, "grad_norm": 0.5365903973579407, "learning_rate": 0.0005410814894765245, "loss": 3.8248, "step": 9200 }, { "epoch": 0.9973045822102425, "grad_norm": 0.652845025062561, "learning_rate": 0.0005407576902320561, "loss": 3.8518, "step": 9250 }, { "epoch": 1.0026954177897573, "grad_norm": 0.60700523853302, "learning_rate": 0.0005404338909875876, "loss": 3.7881, "step": 9300 }, { "epoch": 1.0080862533692723, "grad_norm": 0.6487429738044739, "learning_rate": 0.0005401100917431192, "loss": 3.7647, "step": 9350 }, { "epoch": 1.013477088948787, "grad_norm": 0.561632513999939, "learning_rate": 0.0005397862924986508, "loss": 3.7791, "step": 9400 }, { "epoch": 1.0188679245283019, "grad_norm": 0.5459903478622437, "learning_rate": 0.0005394624932541824, "loss": 3.7855, "step": 9450 }, { "epoch": 1.0242587601078168, "grad_norm": 0.5852130055427551, "learning_rate": 0.0005391386940097139, "loss": 3.7552, "step": 9500 }, { "epoch": 1.0296495956873315, "grad_norm": 0.6032885909080505, "learning_rate": 0.0005388148947652455, "loss": 3.7671, "step": 9550 }, { "epoch": 1.0350404312668464, "grad_norm": 0.6321649551391602, "learning_rate": 0.000538491095520777, "loss": 3.795, "step": 9600 }, { "epoch": 1.0404312668463611, "grad_norm": 0.615997314453125, "learning_rate": 0.0005381672962763086, "loss": 3.7544, "step": 9650 }, { "epoch": 1.045822102425876, "grad_norm": 0.6234314441680908, "learning_rate": 0.0005378434970318403, "loss": 3.7772, "step": 9700 }, { "epoch": 1.0512129380053907, "grad_norm": 0.638446569442749, "learning_rate": 0.0005375196977873718, "loss": 3.753, "step": 9750 }, { "epoch": 1.0566037735849056, "grad_norm": 0.6064075827598572, "learning_rate": 0.0005371958985429034, "loss": 3.765, "step": 9800 }, { "epoch": 1.0619946091644206, "grad_norm": 0.572862982749939, "learning_rate": 0.0005368720992984349, "loss": 3.7706, "step": 9850 }, { "epoch": 1.0673854447439353, "grad_norm": 0.6439206600189209, "learning_rate": 0.0005365483000539665, "loss": 3.7649, "step": 9900 }, { "epoch": 1.0727762803234502, "grad_norm": 0.5527343153953552, "learning_rate": 0.0005362245008094981, "loss": 3.7804, "step": 9950 }, { "epoch": 1.0781671159029649, "grad_norm": 0.5652170777320862, "learning_rate": 0.0005359007015650297, "loss": 3.7701, "step": 10000 }, { "epoch": 1.0781671159029649, "eval_accuracy": 0.3446935722364057, "eval_loss": 3.7489497661590576, "eval_runtime": 184.1084, "eval_samples_per_second": 97.828, "eval_steps_per_second": 6.116, "step": 10000 }, { "epoch": 1.0835579514824798, "grad_norm": 0.5648168325424194, "learning_rate": 0.0005355769023205612, "loss": 3.7698, "step": 10050 }, { "epoch": 1.0889487870619945, "grad_norm": 0.6025657057762146, "learning_rate": 0.0005352531030760928, "loss": 3.7595, "step": 10100 }, { "epoch": 1.0943396226415094, "grad_norm": 0.6286342144012451, "learning_rate": 0.0005349357798165137, "loss": 3.7772, "step": 10150 }, { "epoch": 1.0997304582210243, "grad_norm": 0.565532386302948, "learning_rate": 0.0005346119805720453, "loss": 3.7646, "step": 10200 }, { "epoch": 1.105121293800539, "grad_norm": 0.5853555202484131, "learning_rate": 0.0005342881813275768, "loss": 3.7637, "step": 10250 }, { "epoch": 1.110512129380054, "grad_norm": 0.6412824988365173, "learning_rate": 0.0005339708580679978, "loss": 3.7642, "step": 10300 }, { "epoch": 1.1159029649595686, "grad_norm": 0.5646089315414429, "learning_rate": 0.0005336470588235293, "loss": 3.7553, "step": 10350 }, { "epoch": 1.1212938005390836, "grad_norm": 0.6126633882522583, "learning_rate": 0.000533323259579061, "loss": 3.7789, "step": 10400 }, { "epoch": 1.1266846361185983, "grad_norm": 0.5688264966011047, "learning_rate": 0.0005329994603345925, "loss": 3.7638, "step": 10450 }, { "epoch": 1.1320754716981132, "grad_norm": 0.5909513235092163, "learning_rate": 0.0005326756610901241, "loss": 3.7545, "step": 10500 }, { "epoch": 1.137466307277628, "grad_norm": 0.6165957450866699, "learning_rate": 0.0005323518618456556, "loss": 3.7571, "step": 10550 }, { "epoch": 1.1428571428571428, "grad_norm": 0.6512620449066162, "learning_rate": 0.0005320280626011872, "loss": 3.7638, "step": 10600 }, { "epoch": 1.1482479784366577, "grad_norm": 0.5640995502471924, "learning_rate": 0.0005317042633567188, "loss": 3.7548, "step": 10650 }, { "epoch": 1.1536388140161726, "grad_norm": 0.5489684343338013, "learning_rate": 0.0005313804641122504, "loss": 3.745, "step": 10700 }, { "epoch": 1.1590296495956873, "grad_norm": 0.5724853873252869, "learning_rate": 0.0005310566648677819, "loss": 3.7553, "step": 10750 }, { "epoch": 1.1644204851752022, "grad_norm": 0.6125072240829468, "learning_rate": 0.0005307328656233135, "loss": 3.7447, "step": 10800 }, { "epoch": 1.169811320754717, "grad_norm": 0.551490843296051, "learning_rate": 0.000530409066378845, "loss": 3.7589, "step": 10850 }, { "epoch": 1.1752021563342319, "grad_norm": 0.5600878000259399, "learning_rate": 0.0005300852671343766, "loss": 3.7586, "step": 10900 }, { "epoch": 1.1805929919137466, "grad_norm": 0.5831260681152344, "learning_rate": 0.0005297614678899081, "loss": 3.7425, "step": 10950 }, { "epoch": 1.1859838274932615, "grad_norm": 0.6004246473312378, "learning_rate": 0.0005294376686454398, "loss": 3.7519, "step": 11000 }, { "epoch": 1.1859838274932615, "eval_accuracy": 0.3466813768189716, "eval_loss": 3.723628520965576, "eval_runtime": 183.7188, "eval_samples_per_second": 98.036, "eval_steps_per_second": 6.129, "step": 11000 }, { "epoch": 1.1913746630727764, "grad_norm": 0.5393942594528198, "learning_rate": 0.0005291138694009714, "loss": 3.753, "step": 11050 }, { "epoch": 1.196765498652291, "grad_norm": 0.5937897562980652, "learning_rate": 0.0005287900701565029, "loss": 3.7424, "step": 11100 }, { "epoch": 1.202156334231806, "grad_norm": 0.5454901456832886, "learning_rate": 0.0005284662709120345, "loss": 3.7416, "step": 11150 }, { "epoch": 1.2075471698113207, "grad_norm": 0.5742169618606567, "learning_rate": 0.0005281424716675661, "loss": 3.7442, "step": 11200 }, { "epoch": 1.2129380053908356, "grad_norm": 0.5434097051620483, "learning_rate": 0.0005278186724230977, "loss": 3.7337, "step": 11250 }, { "epoch": 1.2183288409703503, "grad_norm": 0.5956228971481323, "learning_rate": 0.0005274948731786292, "loss": 3.7535, "step": 11300 }, { "epoch": 1.2237196765498652, "grad_norm": 0.5586687922477722, "learning_rate": 0.0005271710739341608, "loss": 3.7113, "step": 11350 }, { "epoch": 1.2291105121293802, "grad_norm": 0.5536643266677856, "learning_rate": 0.0005268472746896923, "loss": 3.7524, "step": 11400 }, { "epoch": 1.2345013477088949, "grad_norm": 0.5457449555397034, "learning_rate": 0.000526523475445224, "loss": 3.7376, "step": 11450 }, { "epoch": 1.2398921832884098, "grad_norm": 0.6064246892929077, "learning_rate": 0.0005261996762007554, "loss": 3.7229, "step": 11500 }, { "epoch": 1.2452830188679245, "grad_norm": 0.5753145217895508, "learning_rate": 0.0005258758769562871, "loss": 3.7512, "step": 11550 }, { "epoch": 1.2506738544474394, "grad_norm": 0.6125341057777405, "learning_rate": 0.0005255520777118186, "loss": 3.7342, "step": 11600 }, { "epoch": 1.256064690026954, "grad_norm": 0.589798629283905, "learning_rate": 0.0005252282784673502, "loss": 3.7213, "step": 11650 }, { "epoch": 1.261455525606469, "grad_norm": 0.6119717359542847, "learning_rate": 0.0005249044792228817, "loss": 3.7299, "step": 11700 }, { "epoch": 1.266846361185984, "grad_norm": 0.6549661159515381, "learning_rate": 0.0005245806799784133, "loss": 3.7474, "step": 11750 }, { "epoch": 1.2722371967654986, "grad_norm": 0.5407906174659729, "learning_rate": 0.0005242568807339449, "loss": 3.7444, "step": 11800 }, { "epoch": 1.2776280323450135, "grad_norm": 0.5647597312927246, "learning_rate": 0.0005239330814894765, "loss": 3.744, "step": 11850 }, { "epoch": 1.2830188679245282, "grad_norm": 0.6143915057182312, "learning_rate": 0.000523609282245008, "loss": 3.7301, "step": 11900 }, { "epoch": 1.2884097035040432, "grad_norm": 0.5904483795166016, "learning_rate": 0.0005232854830005396, "loss": 3.7286, "step": 11950 }, { "epoch": 1.2938005390835579, "grad_norm": 0.5413955450057983, "learning_rate": 0.0005229616837560712, "loss": 3.7387, "step": 12000 }, { "epoch": 1.2938005390835579, "eval_accuracy": 0.3488454163855726, "eval_loss": 3.6995937824249268, "eval_runtime": 184.1584, "eval_samples_per_second": 97.802, "eval_steps_per_second": 6.114, "step": 12000 }, { "epoch": 1.2991913746630728, "grad_norm": 0.5844531059265137, "learning_rate": 0.0005226378845116028, "loss": 3.7271, "step": 12050 }, { "epoch": 1.3045822102425877, "grad_norm": 0.692406415939331, "learning_rate": 0.0005223140852671344, "loss": 3.7326, "step": 12100 }, { "epoch": 1.3099730458221024, "grad_norm": 0.531926691532135, "learning_rate": 0.0005219902860226659, "loss": 3.7207, "step": 12150 }, { "epoch": 1.3153638814016173, "grad_norm": 0.5657473206520081, "learning_rate": 0.0005216664867781975, "loss": 3.7375, "step": 12200 }, { "epoch": 1.320754716981132, "grad_norm": 0.5740625858306885, "learning_rate": 0.000521342687533729, "loss": 3.7009, "step": 12250 }, { "epoch": 1.326145552560647, "grad_norm": 0.5976618528366089, "learning_rate": 0.0005210188882892606, "loss": 3.7069, "step": 12300 }, { "epoch": 1.3315363881401616, "grad_norm": 0.5708423256874084, "learning_rate": 0.0005206950890447922, "loss": 3.7345, "step": 12350 }, { "epoch": 1.3369272237196765, "grad_norm": 0.6060131192207336, "learning_rate": 0.0005203712898003238, "loss": 3.7245, "step": 12400 }, { "epoch": 1.3423180592991915, "grad_norm": 0.5739725232124329, "learning_rate": 0.0005200474905558553, "loss": 3.7225, "step": 12450 }, { "epoch": 1.3477088948787062, "grad_norm": 0.5318633913993835, "learning_rate": 0.0005197236913113869, "loss": 3.6992, "step": 12500 }, { "epoch": 1.353099730458221, "grad_norm": 0.5786360502243042, "learning_rate": 0.0005193998920669184, "loss": 3.7161, "step": 12550 }, { "epoch": 1.3584905660377358, "grad_norm": 0.5843545198440552, "learning_rate": 0.0005190760928224501, "loss": 3.7138, "step": 12600 }, { "epoch": 1.3638814016172507, "grad_norm": 0.579613447189331, "learning_rate": 0.0005187522935779816, "loss": 3.7258, "step": 12650 }, { "epoch": 1.3692722371967654, "grad_norm": 0.6323828101158142, "learning_rate": 0.0005184284943335132, "loss": 3.7231, "step": 12700 }, { "epoch": 1.3746630727762803, "grad_norm": 0.5477507710456848, "learning_rate": 0.0005181046950890447, "loss": 3.7115, "step": 12750 }, { "epoch": 1.3800539083557952, "grad_norm": 0.5417460203170776, "learning_rate": 0.0005177808958445764, "loss": 3.7014, "step": 12800 }, { "epoch": 1.38544474393531, "grad_norm": 0.5849714279174805, "learning_rate": 0.0005174635725849973, "loss": 3.7173, "step": 12850 }, { "epoch": 1.3908355795148248, "grad_norm": 0.5853115916252136, "learning_rate": 0.0005171397733405288, "loss": 3.7121, "step": 12900 }, { "epoch": 1.3962264150943398, "grad_norm": 0.5460019111633301, "learning_rate": 0.0005168159740960603, "loss": 3.738, "step": 12950 }, { "epoch": 1.4016172506738545, "grad_norm": 0.6247442960739136, "learning_rate": 0.000516492174851592, "loss": 3.7317, "step": 13000 }, { "epoch": 1.4016172506738545, "eval_accuracy": 0.3512436029254138, "eval_loss": 3.6774609088897705, "eval_runtime": 183.7689, "eval_samples_per_second": 98.009, "eval_steps_per_second": 6.127, "step": 13000 }, { "epoch": 1.4070080862533692, "grad_norm": 0.6057413220405579, "learning_rate": 0.0005161683756071235, "loss": 3.7133, "step": 13050 }, { "epoch": 1.412398921832884, "grad_norm": 0.5644381642341614, "learning_rate": 0.0005158445763626551, "loss": 3.698, "step": 13100 }, { "epoch": 1.417789757412399, "grad_norm": 0.5523656010627747, "learning_rate": 0.0005155207771181866, "loss": 3.6978, "step": 13150 }, { "epoch": 1.4231805929919137, "grad_norm": 0.5388391613960266, "learning_rate": 0.0005151969778737183, "loss": 3.7136, "step": 13200 }, { "epoch": 1.4285714285714286, "grad_norm": 0.6056439876556396, "learning_rate": 0.0005148731786292498, "loss": 3.7094, "step": 13250 }, { "epoch": 1.4339622641509435, "grad_norm": 0.6014870405197144, "learning_rate": 0.0005145493793847814, "loss": 3.7149, "step": 13300 }, { "epoch": 1.4393530997304582, "grad_norm": 0.6494507193565369, "learning_rate": 0.000514225580140313, "loss": 3.7201, "step": 13350 }, { "epoch": 1.444743935309973, "grad_norm": 0.5282674431800842, "learning_rate": 0.0005139017808958445, "loss": 3.712, "step": 13400 }, { "epoch": 1.4501347708894878, "grad_norm": 0.5607653856277466, "learning_rate": 0.0005135779816513762, "loss": 3.7172, "step": 13450 }, { "epoch": 1.4555256064690028, "grad_norm": 0.5750332474708557, "learning_rate": 0.0005132541824069076, "loss": 3.7082, "step": 13500 }, { "epoch": 1.4609164420485174, "grad_norm": 0.5525111556053162, "learning_rate": 0.0005129303831624393, "loss": 3.7189, "step": 13550 }, { "epoch": 1.4663072776280324, "grad_norm": 0.5573806762695312, "learning_rate": 0.0005126065839179708, "loss": 3.7074, "step": 13600 }, { "epoch": 1.4716981132075473, "grad_norm": 0.5327258706092834, "learning_rate": 0.0005122827846735024, "loss": 3.7108, "step": 13650 }, { "epoch": 1.477088948787062, "grad_norm": 0.6094257831573486, "learning_rate": 0.0005119589854290339, "loss": 3.7015, "step": 13700 }, { "epoch": 1.482479784366577, "grad_norm": 0.5354958772659302, "learning_rate": 0.0005116351861845655, "loss": 3.7019, "step": 13750 }, { "epoch": 1.4878706199460916, "grad_norm": 0.5956083536148071, "learning_rate": 0.0005113113869400971, "loss": 3.696, "step": 13800 }, { "epoch": 1.4932614555256065, "grad_norm": 0.5589125156402588, "learning_rate": 0.0005109875876956287, "loss": 3.6968, "step": 13850 }, { "epoch": 1.4986522911051212, "grad_norm": 0.5720492601394653, "learning_rate": 0.0005106637884511602, "loss": 3.6979, "step": 13900 }, { "epoch": 1.5040431266846361, "grad_norm": 0.5775404572486877, "learning_rate": 0.0005103399892066918, "loss": 3.6881, "step": 13950 }, { "epoch": 1.509433962264151, "grad_norm": 0.5249178409576416, "learning_rate": 0.0005100161899622234, "loss": 3.6971, "step": 14000 }, { "epoch": 1.509433962264151, "eval_accuracy": 0.35375946054275814, "eval_loss": 3.6579151153564453, "eval_runtime": 184.1099, "eval_samples_per_second": 97.827, "eval_steps_per_second": 6.116, "step": 14000 }, { "epoch": 1.5148247978436657, "grad_norm": 0.5867551565170288, "learning_rate": 0.000509692390717755, "loss": 3.6801, "step": 14050 }, { "epoch": 1.5202156334231804, "grad_norm": 0.5887311100959778, "learning_rate": 0.0005093685914732865, "loss": 3.6936, "step": 14100 }, { "epoch": 1.5256064690026954, "grad_norm": 0.5415863394737244, "learning_rate": 0.0005090447922288181, "loss": 3.6752, "step": 14150 }, { "epoch": 1.5309973045822103, "grad_norm": 0.5622326731681824, "learning_rate": 0.0005087209929843496, "loss": 3.6959, "step": 14200 }, { "epoch": 1.536388140161725, "grad_norm": 0.5287529230117798, "learning_rate": 0.0005083971937398812, "loss": 3.7028, "step": 14250 }, { "epoch": 1.54177897574124, "grad_norm": 0.5576687455177307, "learning_rate": 0.0005080733944954127, "loss": 3.7062, "step": 14300 }, { "epoch": 1.5471698113207548, "grad_norm": 0.5946956872940063, "learning_rate": 0.0005077495952509444, "loss": 3.7062, "step": 14350 }, { "epoch": 1.5525606469002695, "grad_norm": 0.5189834237098694, "learning_rate": 0.0005074257960064759, "loss": 3.6921, "step": 14400 }, { "epoch": 1.5579514824797842, "grad_norm": 0.5719146728515625, "learning_rate": 0.0005071019967620075, "loss": 3.6974, "step": 14450 }, { "epoch": 1.5633423180592994, "grad_norm": 0.5453420281410217, "learning_rate": 0.000506778197517539, "loss": 3.6744, "step": 14500 }, { "epoch": 1.568733153638814, "grad_norm": 0.5827313661575317, "learning_rate": 0.0005064543982730707, "loss": 3.6753, "step": 14550 }, { "epoch": 1.5741239892183287, "grad_norm": 0.5802609920501709, "learning_rate": 0.0005061305990286023, "loss": 3.6732, "step": 14600 }, { "epoch": 1.5795148247978437, "grad_norm": 0.5037614703178406, "learning_rate": 0.0005058067997841338, "loss": 3.6892, "step": 14650 }, { "epoch": 1.5849056603773586, "grad_norm": 0.5056049823760986, "learning_rate": 0.0005054830005396654, "loss": 3.6858, "step": 14700 }, { "epoch": 1.5902964959568733, "grad_norm": 0.6599529385566711, "learning_rate": 0.0005051592012951969, "loss": 3.6799, "step": 14750 }, { "epoch": 1.595687331536388, "grad_norm": 0.563842236995697, "learning_rate": 0.0005048354020507286, "loss": 3.6988, "step": 14800 }, { "epoch": 1.6010781671159031, "grad_norm": 0.5434414744377136, "learning_rate": 0.00050451160280626, "loss": 3.6767, "step": 14850 }, { "epoch": 1.6064690026954178, "grad_norm": 0.61106938123703, "learning_rate": 0.0005041878035617917, "loss": 3.6979, "step": 14900 }, { "epoch": 1.6118598382749325, "grad_norm": 0.5468453764915466, "learning_rate": 0.0005038640043173232, "loss": 3.6902, "step": 14950 }, { "epoch": 1.6172506738544474, "grad_norm": 0.589815616607666, "learning_rate": 0.0005035402050728548, "loss": 3.6885, "step": 15000 }, { "epoch": 1.6172506738544474, "eval_accuracy": 0.3551121889960484, "eval_loss": 3.6401309967041016, "eval_runtime": 183.883, "eval_samples_per_second": 97.948, "eval_steps_per_second": 6.123, "step": 15000 }, { "epoch": 1.6226415094339623, "grad_norm": 0.5671663284301758, "learning_rate": 0.0005032164058283863, "loss": 3.6831, "step": 15050 }, { "epoch": 1.628032345013477, "grad_norm": 0.6244860887527466, "learning_rate": 0.0005028926065839179, "loss": 3.6613, "step": 15100 }, { "epoch": 1.633423180592992, "grad_norm": 0.5171533823013306, "learning_rate": 0.0005025688073394495, "loss": 3.6778, "step": 15150 }, { "epoch": 1.6388140161725069, "grad_norm": 0.5679621696472168, "learning_rate": 0.0005022450080949811, "loss": 3.6648, "step": 15200 }, { "epoch": 1.6442048517520216, "grad_norm": 0.5312994718551636, "learning_rate": 0.0005019212088505126, "loss": 3.6837, "step": 15250 }, { "epoch": 1.6495956873315363, "grad_norm": 0.5288447737693787, "learning_rate": 0.0005015974096060442, "loss": 3.6684, "step": 15300 }, { "epoch": 1.6549865229110512, "grad_norm": 0.6825982928276062, "learning_rate": 0.0005012736103615758, "loss": 3.671, "step": 15350 }, { "epoch": 1.6603773584905661, "grad_norm": 0.5444086790084839, "learning_rate": 0.0005009498111171074, "loss": 3.7001, "step": 15400 }, { "epoch": 1.6657681940700808, "grad_norm": 0.5970629453659058, "learning_rate": 0.0005006260118726389, "loss": 3.6732, "step": 15450 }, { "epoch": 1.6711590296495957, "grad_norm": 0.6333000063896179, "learning_rate": 0.0005003022126281705, "loss": 3.6567, "step": 15500 }, { "epoch": 1.6765498652291106, "grad_norm": 0.5841792821884155, "learning_rate": 0.000499978413383702, "loss": 3.6866, "step": 15550 }, { "epoch": 1.6819407008086253, "grad_norm": 0.5716648697853088, "learning_rate": 0.0004996546141392336, "loss": 3.6464, "step": 15600 }, { "epoch": 1.68733153638814, "grad_norm": 0.6867700219154358, "learning_rate": 0.0004993308148947651, "loss": 3.667, "step": 15650 }, { "epoch": 1.692722371967655, "grad_norm": 0.5503226518630981, "learning_rate": 0.0004990070156502968, "loss": 3.6594, "step": 15700 }, { "epoch": 1.6981132075471699, "grad_norm": 0.5787608623504639, "learning_rate": 0.0004986832164058284, "loss": 3.6795, "step": 15750 }, { "epoch": 1.7035040431266846, "grad_norm": 0.5347179770469666, "learning_rate": 0.0004983594171613599, "loss": 3.6652, "step": 15800 }, { "epoch": 1.7088948787061995, "grad_norm": 0.5248307585716248, "learning_rate": 0.0004980356179168915, "loss": 3.6812, "step": 15850 }, { "epoch": 1.7142857142857144, "grad_norm": 0.5549546480178833, "learning_rate": 0.000497711818672423, "loss": 3.6636, "step": 15900 }, { "epoch": 1.719676549865229, "grad_norm": 0.6331003308296204, "learning_rate": 0.0004973880194279547, "loss": 3.6594, "step": 15950 }, { "epoch": 1.7250673854447438, "grad_norm": 0.6510477662086487, "learning_rate": 0.0004970642201834862, "loss": 3.6771, "step": 16000 }, { "epoch": 1.7250673854447438, "eval_accuracy": 0.3569796061789159, "eval_loss": 3.619459867477417, "eval_runtime": 183.9465, "eval_samples_per_second": 97.914, "eval_steps_per_second": 6.121, "step": 16000 }, { "epoch": 1.7304582210242587, "grad_norm": 0.6069450378417969, "learning_rate": 0.0004967404209390178, "loss": 3.654, "step": 16050 }, { "epoch": 1.7358490566037736, "grad_norm": 0.5242086052894592, "learning_rate": 0.0004964166216945493, "loss": 3.6722, "step": 16100 }, { "epoch": 1.7412398921832883, "grad_norm": 0.612231433391571, "learning_rate": 0.000496092822450081, "loss": 3.6597, "step": 16150 }, { "epoch": 1.7466307277628033, "grad_norm": 0.6022439002990723, "learning_rate": 0.0004957690232056125, "loss": 3.6801, "step": 16200 }, { "epoch": 1.7520215633423182, "grad_norm": 0.5812779664993286, "learning_rate": 0.0004954452239611441, "loss": 3.6582, "step": 16250 }, { "epoch": 1.7574123989218329, "grad_norm": 0.5933021306991577, "learning_rate": 0.0004951214247166756, "loss": 3.6652, "step": 16300 }, { "epoch": 1.7628032345013476, "grad_norm": 0.5812273621559143, "learning_rate": 0.0004947976254722072, "loss": 3.6697, "step": 16350 }, { "epoch": 1.7681940700808625, "grad_norm": 0.6487030386924744, "learning_rate": 0.0004944738262277387, "loss": 3.6677, "step": 16400 }, { "epoch": 1.7735849056603774, "grad_norm": 0.548700749874115, "learning_rate": 0.0004941500269832703, "loss": 3.658, "step": 16450 }, { "epoch": 1.778975741239892, "grad_norm": 0.6098374724388123, "learning_rate": 0.0004938262277388019, "loss": 3.6542, "step": 16500 }, { "epoch": 1.784366576819407, "grad_norm": 0.6051768660545349, "learning_rate": 0.0004935024284943335, "loss": 3.6566, "step": 16550 }, { "epoch": 1.789757412398922, "grad_norm": 0.6026385426521301, "learning_rate": 0.000493178629249865, "loss": 3.653, "step": 16600 }, { "epoch": 1.7951482479784366, "grad_norm": 0.562525749206543, "learning_rate": 0.0004928548300053966, "loss": 3.6491, "step": 16650 }, { "epoch": 1.8005390835579513, "grad_norm": 0.571699321269989, "learning_rate": 0.0004925310307609282, "loss": 3.6552, "step": 16700 }, { "epoch": 1.8059299191374663, "grad_norm": 0.5961069464683533, "learning_rate": 0.0004922072315164598, "loss": 3.6492, "step": 16750 }, { "epoch": 1.8113207547169812, "grad_norm": 0.5025362372398376, "learning_rate": 0.0004918834322719913, "loss": 3.6478, "step": 16800 }, { "epoch": 1.8167115902964959, "grad_norm": 0.5551483631134033, "learning_rate": 0.0004915661090124123, "loss": 3.6596, "step": 16850 }, { "epoch": 1.8221024258760108, "grad_norm": 0.5412172079086304, "learning_rate": 0.0004912423097679439, "loss": 3.6498, "step": 16900 }, { "epoch": 1.8274932614555257, "grad_norm": 0.6038126945495605, "learning_rate": 0.0004909185105234754, "loss": 3.6542, "step": 16950 }, { "epoch": 1.8328840970350404, "grad_norm": 0.5766414403915405, "learning_rate": 0.000490594711279007, "loss": 3.6503, "step": 17000 }, { "epoch": 1.8328840970350404, "eval_accuracy": 0.35894415904348953, "eval_loss": 3.6033706665039062, "eval_runtime": 184.1497, "eval_samples_per_second": 97.806, "eval_steps_per_second": 6.115, "step": 17000 }, { "epoch": 1.838274932614555, "grad_norm": 0.582373857498169, "learning_rate": 0.0004902709120345385, "loss": 3.641, "step": 17050 }, { "epoch": 1.8436657681940702, "grad_norm": 0.5415629148483276, "learning_rate": 0.00048994711279007, "loss": 3.6456, "step": 17100 }, { "epoch": 1.849056603773585, "grad_norm": 0.5694625973701477, "learning_rate": 0.0004896233135456017, "loss": 3.6568, "step": 17150 }, { "epoch": 1.8544474393530996, "grad_norm": 0.547035276889801, "learning_rate": 0.0004892995143011333, "loss": 3.654, "step": 17200 }, { "epoch": 1.8598382749326146, "grad_norm": 0.6010807752609253, "learning_rate": 0.0004889757150566648, "loss": 3.6348, "step": 17250 }, { "epoch": 1.8652291105121295, "grad_norm": 0.5640713572502136, "learning_rate": 0.0004886519158121964, "loss": 3.6555, "step": 17300 }, { "epoch": 1.8706199460916442, "grad_norm": 0.6125467419624329, "learning_rate": 0.000488328116567728, "loss": 3.656, "step": 17350 }, { "epoch": 1.8760107816711589, "grad_norm": 0.6842520833015442, "learning_rate": 0.0004880043173232595, "loss": 3.6642, "step": 17400 }, { "epoch": 1.881401617250674, "grad_norm": 0.5659988522529602, "learning_rate": 0.0004876805180787911, "loss": 3.6325, "step": 17450 }, { "epoch": 1.8867924528301887, "grad_norm": 0.6425742506980896, "learning_rate": 0.0004873567188343227, "loss": 3.6474, "step": 17500 }, { "epoch": 1.8921832884097034, "grad_norm": 0.5763974785804749, "learning_rate": 0.0004870329195898542, "loss": 3.6383, "step": 17550 }, { "epoch": 1.8975741239892183, "grad_norm": 0.5723247528076172, "learning_rate": 0.00048670912034538583, "loss": 3.6529, "step": 17600 }, { "epoch": 1.9029649595687332, "grad_norm": 0.5686931610107422, "learning_rate": 0.0004863853211009174, "loss": 3.6496, "step": 17650 }, { "epoch": 1.908355795148248, "grad_norm": 0.5761986970901489, "learning_rate": 0.000486061521856449, "loss": 3.6397, "step": 17700 }, { "epoch": 1.9137466307277629, "grad_norm": 0.6086380481719971, "learning_rate": 0.00048573772261198054, "loss": 3.6421, "step": 17750 }, { "epoch": 1.9191374663072778, "grad_norm": 0.5803984999656677, "learning_rate": 0.00048541392336751214, "loss": 3.6658, "step": 17800 }, { "epoch": 1.9245283018867925, "grad_norm": 0.6012025475502014, "learning_rate": 0.0004850901241230437, "loss": 3.6396, "step": 17850 }, { "epoch": 1.9299191374663072, "grad_norm": 0.5470909476280212, "learning_rate": 0.0004847663248785753, "loss": 3.6366, "step": 17900 }, { "epoch": 1.935309973045822, "grad_norm": 0.6374045014381409, "learning_rate": 0.0004844425256341068, "loss": 3.6164, "step": 17950 }, { "epoch": 1.940700808625337, "grad_norm": 0.5931817889213562, "learning_rate": 0.00048411872638963834, "loss": 3.6523, "step": 18000 }, { "epoch": 1.940700808625337, "eval_accuracy": 0.35995310975973477, "eval_loss": 3.5907158851623535, "eval_runtime": 183.7191, "eval_samples_per_second": 98.036, "eval_steps_per_second": 6.129, "step": 18000 }, { "epoch": 1.9460916442048517, "grad_norm": 0.5779642462730408, "learning_rate": 0.00048379492714516995, "loss": 3.6288, "step": 18050 }, { "epoch": 1.9514824797843666, "grad_norm": 0.5749161839485168, "learning_rate": 0.0004834711279007015, "loss": 3.6552, "step": 18100 }, { "epoch": 1.9568733153638815, "grad_norm": 0.5829963684082031, "learning_rate": 0.0004831473286562331, "loss": 3.6656, "step": 18150 }, { "epoch": 1.9622641509433962, "grad_norm": 0.547026515007019, "learning_rate": 0.00048282352941176465, "loss": 3.6443, "step": 18200 }, { "epoch": 1.967654986522911, "grad_norm": 0.5671616792678833, "learning_rate": 0.00048249973016729626, "loss": 3.6491, "step": 18250 }, { "epoch": 1.9730458221024259, "grad_norm": 0.5770370960235596, "learning_rate": 0.0004821759309228278, "loss": 3.6299, "step": 18300 }, { "epoch": 1.9784366576819408, "grad_norm": 0.5550295114517212, "learning_rate": 0.00048185213167835936, "loss": 3.6396, "step": 18350 }, { "epoch": 1.9838274932614555, "grad_norm": 0.5505232214927673, "learning_rate": 0.00048152833243389096, "loss": 3.6541, "step": 18400 }, { "epoch": 1.9892183288409704, "grad_norm": 0.5260711908340454, "learning_rate": 0.0004812045331894225, "loss": 3.6188, "step": 18450 }, { "epoch": 1.9946091644204853, "grad_norm": 0.5258703231811523, "learning_rate": 0.0004808807339449541, "loss": 3.6256, "step": 18500 }, { "epoch": 2.0, "grad_norm": 1.088941216468811, "learning_rate": 0.0004805569347004856, "loss": 3.6356, "step": 18550 }, { "epoch": 2.0053908355795147, "grad_norm": 0.5310294032096863, "learning_rate": 0.0004802331354560173, "loss": 3.542, "step": 18600 }, { "epoch": 2.01078167115903, "grad_norm": 0.5242578387260437, "learning_rate": 0.00047990933621154877, "loss": 3.5532, "step": 18650 }, { "epoch": 2.0161725067385445, "grad_norm": 0.5820230841636658, "learning_rate": 0.0004795855369670804, "loss": 3.5545, "step": 18700 }, { "epoch": 2.0215633423180592, "grad_norm": 0.6573295593261719, "learning_rate": 0.0004792617377226119, "loss": 3.5472, "step": 18750 }, { "epoch": 2.026954177897574, "grad_norm": 0.5828113555908203, "learning_rate": 0.0004789379384781435, "loss": 3.5532, "step": 18800 }, { "epoch": 2.032345013477089, "grad_norm": 0.6123409867286682, "learning_rate": 0.0004786141392336751, "loss": 3.5318, "step": 18850 }, { "epoch": 2.0377358490566038, "grad_norm": 0.5386969447135925, "learning_rate": 0.00047829033998920663, "loss": 3.5529, "step": 18900 }, { "epoch": 2.0431266846361185, "grad_norm": 0.5555720329284668, "learning_rate": 0.00047796654074473824, "loss": 3.5571, "step": 18950 }, { "epoch": 2.0485175202156336, "grad_norm": 0.5521024465560913, "learning_rate": 0.0004776427415002698, "loss": 3.5469, "step": 19000 }, { "epoch": 2.0485175202156336, "eval_accuracy": 0.3615377034756212, "eval_loss": 3.5789687633514404, "eval_runtime": 184.2139, "eval_samples_per_second": 97.772, "eval_steps_per_second": 6.112, "step": 19000 }, { "epoch": 2.0539083557951483, "grad_norm": 0.6329952478408813, "learning_rate": 0.00047732541824069074, "loss": 3.5373, "step": 19050 }, { "epoch": 2.059299191374663, "grad_norm": 0.5793041586875916, "learning_rate": 0.0004770016189962223, "loss": 3.5498, "step": 19100 }, { "epoch": 2.0646900269541777, "grad_norm": 0.5503807067871094, "learning_rate": 0.0004766778197517539, "loss": 3.5437, "step": 19150 }, { "epoch": 2.070080862533693, "grad_norm": 0.5668014883995056, "learning_rate": 0.0004763540205072854, "loss": 3.5563, "step": 19200 }, { "epoch": 2.0754716981132075, "grad_norm": 0.5484731197357178, "learning_rate": 0.00047603022126281705, "loss": 3.554, "step": 19250 }, { "epoch": 2.0808625336927222, "grad_norm": 0.6245366930961609, "learning_rate": 0.00047570642201834855, "loss": 3.5468, "step": 19300 }, { "epoch": 2.0862533692722374, "grad_norm": 0.6225091814994812, "learning_rate": 0.00047538262277388015, "loss": 3.5642, "step": 19350 }, { "epoch": 2.091644204851752, "grad_norm": 0.5625278949737549, "learning_rate": 0.0004750588235294117, "loss": 3.5567, "step": 19400 }, { "epoch": 2.0970350404312668, "grad_norm": 0.632989764213562, "learning_rate": 0.0004747350242849433, "loss": 3.5484, "step": 19450 }, { "epoch": 2.1024258760107815, "grad_norm": 0.5744699239730835, "learning_rate": 0.00047441122504047486, "loss": 3.5617, "step": 19500 }, { "epoch": 2.1078167115902966, "grad_norm": 0.5948736071586609, "learning_rate": 0.0004740874257960064, "loss": 3.542, "step": 19550 }, { "epoch": 2.1132075471698113, "grad_norm": 0.5679342746734619, "learning_rate": 0.000473763626551538, "loss": 3.5556, "step": 19600 }, { "epoch": 2.118598382749326, "grad_norm": 0.5541220307350159, "learning_rate": 0.00047343982730706956, "loss": 3.5516, "step": 19650 }, { "epoch": 2.123989218328841, "grad_norm": 0.544877290725708, "learning_rate": 0.00047311602806260117, "loss": 3.5489, "step": 19700 }, { "epoch": 2.129380053908356, "grad_norm": 0.5709410309791565, "learning_rate": 0.0004727922288181327, "loss": 3.5652, "step": 19750 }, { "epoch": 2.1347708894878705, "grad_norm": 0.5732446908950806, "learning_rate": 0.0004724684295736643, "loss": 3.5745, "step": 19800 }, { "epoch": 2.1401617250673857, "grad_norm": 0.6136924028396606, "learning_rate": 0.0004721446303291959, "loss": 3.5503, "step": 19850 }, { "epoch": 2.1455525606469004, "grad_norm": 0.5857067108154297, "learning_rate": 0.0004718208310847275, "loss": 3.5585, "step": 19900 }, { "epoch": 2.150943396226415, "grad_norm": 0.5760287046432495, "learning_rate": 0.000471497031840259, "loss": 3.5601, "step": 19950 }, { "epoch": 2.1563342318059298, "grad_norm": 0.5563052296638489, "learning_rate": 0.0004711732325957905, "loss": 3.553, "step": 20000 }, { "epoch": 2.1563342318059298, "eval_accuracy": 0.36297746289205085, "eval_loss": 3.5687758922576904, "eval_runtime": 183.8162, "eval_samples_per_second": 97.984, "eval_steps_per_second": 6.126, "step": 20000 }, { "epoch": 2.161725067385445, "grad_norm": 0.6046416759490967, "learning_rate": 0.00047084943335132213, "loss": 3.5431, "step": 20050 }, { "epoch": 2.1671159029649596, "grad_norm": 0.5706889033317566, "learning_rate": 0.0004705256341068537, "loss": 3.5522, "step": 20100 }, { "epoch": 2.1725067385444743, "grad_norm": 0.6285442113876343, "learning_rate": 0.0004702018348623853, "loss": 3.5604, "step": 20150 }, { "epoch": 2.177897574123989, "grad_norm": 0.567588746547699, "learning_rate": 0.00046987803561791684, "loss": 3.5649, "step": 20200 }, { "epoch": 2.183288409703504, "grad_norm": 0.543428897857666, "learning_rate": 0.00046955423637344844, "loss": 3.5729, "step": 20250 }, { "epoch": 2.188679245283019, "grad_norm": 0.6000525951385498, "learning_rate": 0.00046923043712898, "loss": 3.5503, "step": 20300 }, { "epoch": 2.1940700808625335, "grad_norm": 0.6174874305725098, "learning_rate": 0.0004689066378845116, "loss": 3.5539, "step": 20350 }, { "epoch": 2.1994609164420487, "grad_norm": 0.591032087802887, "learning_rate": 0.00046858283864004315, "loss": 3.5589, "step": 20400 }, { "epoch": 2.2048517520215634, "grad_norm": 0.5443971753120422, "learning_rate": 0.0004682590393955747, "loss": 3.551, "step": 20450 }, { "epoch": 2.210242587601078, "grad_norm": 0.6590529084205627, "learning_rate": 0.0004679352401511063, "loss": 3.5497, "step": 20500 }, { "epoch": 2.215633423180593, "grad_norm": 0.592398464679718, "learning_rate": 0.0004676114409066378, "loss": 3.5583, "step": 20550 }, { "epoch": 2.221024258760108, "grad_norm": 0.5806289911270142, "learning_rate": 0.00046728764166216946, "loss": 3.5656, "step": 20600 }, { "epoch": 2.2264150943396226, "grad_norm": 0.5720568299293518, "learning_rate": 0.00046696384241770095, "loss": 3.5663, "step": 20650 }, { "epoch": 2.2318059299191373, "grad_norm": 0.6045317649841309, "learning_rate": 0.00046664004317323256, "loss": 3.5707, "step": 20700 }, { "epoch": 2.2371967654986524, "grad_norm": 0.587020218372345, "learning_rate": 0.0004663162439287641, "loss": 3.5501, "step": 20750 }, { "epoch": 2.242587601078167, "grad_norm": 0.6044445037841797, "learning_rate": 0.00046599244468429566, "loss": 3.5433, "step": 20800 }, { "epoch": 2.247978436657682, "grad_norm": 0.5762318968772888, "learning_rate": 0.00046566864543982726, "loss": 3.5415, "step": 20850 }, { "epoch": 2.2533692722371965, "grad_norm": 0.5729345083236694, "learning_rate": 0.0004653448461953588, "loss": 3.5473, "step": 20900 }, { "epoch": 2.2587601078167117, "grad_norm": 0.619462251663208, "learning_rate": 0.0004650210469508904, "loss": 3.5689, "step": 20950 }, { "epoch": 2.2641509433962264, "grad_norm": 0.5913404226303101, "learning_rate": 0.00046469724770642197, "loss": 3.5623, "step": 21000 }, { "epoch": 2.2641509433962264, "eval_accuracy": 0.3636772961424639, "eval_loss": 3.5585999488830566, "eval_runtime": 183.9266, "eval_samples_per_second": 97.925, "eval_steps_per_second": 6.122, "step": 21000 }, { "epoch": 2.269541778975741, "grad_norm": 0.584413468837738, "learning_rate": 0.0004643734484619536, "loss": 3.5674, "step": 21050 }, { "epoch": 2.274932614555256, "grad_norm": 0.5533130764961243, "learning_rate": 0.0004640496492174851, "loss": 3.5502, "step": 21100 }, { "epoch": 2.280323450134771, "grad_norm": 0.5667862296104431, "learning_rate": 0.0004637323259579061, "loss": 3.5487, "step": 21150 }, { "epoch": 2.2857142857142856, "grad_norm": 0.5526608824729919, "learning_rate": 0.0004634085267134376, "loss": 3.5586, "step": 21200 }, { "epoch": 2.2911051212938007, "grad_norm": 0.6517942547798157, "learning_rate": 0.00046308472746896923, "loss": 3.5757, "step": 21250 }, { "epoch": 2.2964959568733154, "grad_norm": 0.5420659184455872, "learning_rate": 0.00046276092822450073, "loss": 3.5505, "step": 21300 }, { "epoch": 2.30188679245283, "grad_norm": 0.5619015097618103, "learning_rate": 0.00046243712898003233, "loss": 3.5671, "step": 21350 }, { "epoch": 2.3072776280323453, "grad_norm": 0.5901092886924744, "learning_rate": 0.0004621133297355639, "loss": 3.5691, "step": 21400 }, { "epoch": 2.31266846361186, "grad_norm": 0.5045002102851868, "learning_rate": 0.0004617895304910955, "loss": 3.5569, "step": 21450 }, { "epoch": 2.3180592991913747, "grad_norm": 0.6043163537979126, "learning_rate": 0.00046146573124662704, "loss": 3.5506, "step": 21500 }, { "epoch": 2.3234501347708894, "grad_norm": 0.5670344829559326, "learning_rate": 0.00046114193200215864, "loss": 3.5562, "step": 21550 }, { "epoch": 2.3288409703504045, "grad_norm": 0.5510123372077942, "learning_rate": 0.0004608181327576902, "loss": 3.5494, "step": 21600 }, { "epoch": 2.334231805929919, "grad_norm": 0.5462912917137146, "learning_rate": 0.00046049433351322175, "loss": 3.5591, "step": 21650 }, { "epoch": 2.339622641509434, "grad_norm": 0.6254079341888428, "learning_rate": 0.00046017053426875335, "loss": 3.5383, "step": 21700 }, { "epoch": 2.3450134770889486, "grad_norm": 0.5986773371696472, "learning_rate": 0.0004598467350242849, "loss": 3.5331, "step": 21750 }, { "epoch": 2.3504043126684637, "grad_norm": 0.5816046595573425, "learning_rate": 0.0004595229357798165, "loss": 3.5548, "step": 21800 }, { "epoch": 2.3557951482479784, "grad_norm": 0.5853009223937988, "learning_rate": 0.00045919913653534806, "loss": 3.552, "step": 21850 }, { "epoch": 2.361185983827493, "grad_norm": 0.5687203407287598, "learning_rate": 0.00045887533729087966, "loss": 3.5502, "step": 21900 }, { "epoch": 2.3665768194070083, "grad_norm": 0.604446291923523, "learning_rate": 0.00045855153804641116, "loss": 3.5531, "step": 21950 }, { "epoch": 2.371967654986523, "grad_norm": 0.570324182510376, "learning_rate": 0.0004582277388019427, "loss": 3.5464, "step": 22000 }, { "epoch": 2.371967654986523, "eval_accuracy": 0.36487475961906735, "eval_loss": 3.5480477809906006, "eval_runtime": 183.8636, "eval_samples_per_second": 97.958, "eval_steps_per_second": 6.124, "step": 22000 }, { "epoch": 2.3773584905660377, "grad_norm": 0.6100995540618896, "learning_rate": 0.0004579039395574743, "loss": 3.5415, "step": 22050 }, { "epoch": 2.382749326145553, "grad_norm": 0.5746794939041138, "learning_rate": 0.00045758014031300586, "loss": 3.557, "step": 22100 }, { "epoch": 2.3881401617250675, "grad_norm": 0.6097766757011414, "learning_rate": 0.00045725634106853747, "loss": 3.5718, "step": 22150 }, { "epoch": 2.393530997304582, "grad_norm": 0.5994210243225098, "learning_rate": 0.000456932541824069, "loss": 3.5689, "step": 22200 }, { "epoch": 2.398921832884097, "grad_norm": 0.5749094486236572, "learning_rate": 0.0004566087425796006, "loss": 3.538, "step": 22250 }, { "epoch": 2.404312668463612, "grad_norm": 0.549342691898346, "learning_rate": 0.0004562849433351322, "loss": 3.5508, "step": 22300 }, { "epoch": 2.4097035040431267, "grad_norm": 0.5516054630279541, "learning_rate": 0.0004559611440906638, "loss": 3.5465, "step": 22350 }, { "epoch": 2.4150943396226414, "grad_norm": 0.5950953364372253, "learning_rate": 0.00045563734484619533, "loss": 3.5643, "step": 22400 }, { "epoch": 2.420485175202156, "grad_norm": 0.8571392297744751, "learning_rate": 0.0004553135456017269, "loss": 3.5725, "step": 22450 }, { "epoch": 2.4258760107816713, "grad_norm": 0.6473979949951172, "learning_rate": 0.0004549897463572585, "loss": 3.5584, "step": 22500 }, { "epoch": 2.431266846361186, "grad_norm": 0.5623269081115723, "learning_rate": 0.00045466594711279, "loss": 3.5454, "step": 22550 }, { "epoch": 2.4366576819407006, "grad_norm": 0.5944833159446716, "learning_rate": 0.00045434214786832164, "loss": 3.5497, "step": 22600 }, { "epoch": 2.442048517520216, "grad_norm": 0.5693549513816833, "learning_rate": 0.00045401834862385314, "loss": 3.552, "step": 22650 }, { "epoch": 2.4474393530997305, "grad_norm": 0.5360573530197144, "learning_rate": 0.00045369454937938474, "loss": 3.5426, "step": 22700 }, { "epoch": 2.452830188679245, "grad_norm": 0.5875699520111084, "learning_rate": 0.0004533707501349163, "loss": 3.5768, "step": 22750 }, { "epoch": 2.4582210242587603, "grad_norm": 0.5625613331794739, "learning_rate": 0.0004530469508904479, "loss": 3.543, "step": 22800 }, { "epoch": 2.463611859838275, "grad_norm": 0.5869686603546143, "learning_rate": 0.0004527296276308688, "loss": 3.5705, "step": 22850 }, { "epoch": 2.4690026954177897, "grad_norm": 0.5737190842628479, "learning_rate": 0.0004524058283864004, "loss": 3.5454, "step": 22900 }, { "epoch": 2.4743935309973044, "grad_norm": 0.5693169236183167, "learning_rate": 0.00045208202914193195, "loss": 3.5611, "step": 22950 }, { "epoch": 2.4797843665768196, "grad_norm": 0.6664920449256897, "learning_rate": 0.00045175822989746355, "loss": 3.543, "step": 23000 }, { "epoch": 2.4797843665768196, "eval_accuracy": 0.3658964227231869, "eval_loss": 3.5396640300750732, "eval_runtime": 184.0061, "eval_samples_per_second": 97.883, "eval_steps_per_second": 6.119, "step": 23000 }, { "epoch": 2.4851752021563343, "grad_norm": 0.5835683345794678, "learning_rate": 0.0004514344306529951, "loss": 3.5382, "step": 23050 }, { "epoch": 2.490566037735849, "grad_norm": 0.5643031001091003, "learning_rate": 0.0004511106314085267, "loss": 3.537, "step": 23100 }, { "epoch": 2.4959568733153636, "grad_norm": 0.5767573118209839, "learning_rate": 0.00045078683216405826, "loss": 3.5492, "step": 23150 }, { "epoch": 2.501347708894879, "grad_norm": 0.5756224989891052, "learning_rate": 0.00045046303291958976, "loss": 3.5463, "step": 23200 }, { "epoch": 2.5067385444743935, "grad_norm": 0.6499918103218079, "learning_rate": 0.0004501392336751214, "loss": 3.561, "step": 23250 }, { "epoch": 2.512129380053908, "grad_norm": 0.6103466153144836, "learning_rate": 0.0004498154344306529, "loss": 3.5578, "step": 23300 }, { "epoch": 2.5175202156334233, "grad_norm": 0.6318963170051575, "learning_rate": 0.0004494916351861845, "loss": 3.5593, "step": 23350 }, { "epoch": 2.522911051212938, "grad_norm": 0.6232575178146362, "learning_rate": 0.00044916783594171607, "loss": 3.5548, "step": 23400 }, { "epoch": 2.5283018867924527, "grad_norm": 0.5690689086914062, "learning_rate": 0.00044884403669724767, "loss": 3.5452, "step": 23450 }, { "epoch": 2.533692722371968, "grad_norm": 0.6115185618400574, "learning_rate": 0.0004485202374527792, "loss": 3.5415, "step": 23500 }, { "epoch": 2.5390835579514826, "grad_norm": 0.5683715343475342, "learning_rate": 0.00044819643820831083, "loss": 3.5434, "step": 23550 }, { "epoch": 2.5444743935309972, "grad_norm": 0.6168181896209717, "learning_rate": 0.0004478726389638424, "loss": 3.5498, "step": 23600 }, { "epoch": 2.5498652291105124, "grad_norm": 0.6147149205207825, "learning_rate": 0.00044754883971937393, "loss": 3.5289, "step": 23650 }, { "epoch": 2.555256064690027, "grad_norm": 0.5647293925285339, "learning_rate": 0.00044722504047490553, "loss": 3.5448, "step": 23700 }, { "epoch": 2.560646900269542, "grad_norm": 0.6555283665657043, "learning_rate": 0.0004469012412304371, "loss": 3.5521, "step": 23750 }, { "epoch": 2.5660377358490565, "grad_norm": 0.5926530361175537, "learning_rate": 0.0004465774419859687, "loss": 3.5618, "step": 23800 }, { "epoch": 2.571428571428571, "grad_norm": 0.6275022029876709, "learning_rate": 0.00044625364274150024, "loss": 3.5548, "step": 23850 }, { "epoch": 2.5768194070080863, "grad_norm": 0.5570287704467773, "learning_rate": 0.00044592984349703184, "loss": 3.5318, "step": 23900 }, { "epoch": 2.582210242587601, "grad_norm": 0.5836941599845886, "learning_rate": 0.00044560604425256334, "loss": 3.5265, "step": 23950 }, { "epoch": 2.5876010781671157, "grad_norm": 0.546286404132843, "learning_rate": 0.000445282245008095, "loss": 3.5598, "step": 24000 }, { "epoch": 2.5876010781671157, "eval_accuracy": 0.36674695752899866, "eval_loss": 3.528761386871338, "eval_runtime": 184.0106, "eval_samples_per_second": 97.88, "eval_steps_per_second": 6.119, "step": 24000 }, { "epoch": 2.592991913746631, "grad_norm": 0.5817551612854004, "learning_rate": 0.0004449584457636265, "loss": 3.554, "step": 24050 }, { "epoch": 2.5983827493261455, "grad_norm": 0.6376777291297913, "learning_rate": 0.00044463464651915805, "loss": 3.5397, "step": 24100 }, { "epoch": 2.6037735849056602, "grad_norm": 0.6055445671081543, "learning_rate": 0.00044431084727468965, "loss": 3.5507, "step": 24150 }, { "epoch": 2.6091644204851754, "grad_norm": 0.5456017851829529, "learning_rate": 0.0004439870480302212, "loss": 3.526, "step": 24200 }, { "epoch": 2.61455525606469, "grad_norm": 0.6268299221992493, "learning_rate": 0.0004436632487857528, "loss": 3.5493, "step": 24250 }, { "epoch": 2.6199460916442048, "grad_norm": 0.6223170757293701, "learning_rate": 0.00044333944954128436, "loss": 3.5491, "step": 24300 }, { "epoch": 2.62533692722372, "grad_norm": 0.580644428730011, "learning_rate": 0.00044301565029681596, "loss": 3.5434, "step": 24350 }, { "epoch": 2.6307277628032346, "grad_norm": 0.5493316054344177, "learning_rate": 0.0004426918510523475, "loss": 3.5352, "step": 24400 }, { "epoch": 2.6361185983827493, "grad_norm": 0.5446970462799072, "learning_rate": 0.0004423680518078791, "loss": 3.5151, "step": 24450 }, { "epoch": 2.641509433962264, "grad_norm": 0.5663889646530151, "learning_rate": 0.00044204425256341067, "loss": 3.5352, "step": 24500 }, { "epoch": 2.6469002695417787, "grad_norm": 0.5662615895271301, "learning_rate": 0.0004417204533189422, "loss": 3.5313, "step": 24550 }, { "epoch": 2.652291105121294, "grad_norm": 0.5774913430213928, "learning_rate": 0.0004413966540744738, "loss": 3.5433, "step": 24600 }, { "epoch": 2.6576819407008085, "grad_norm": 0.5947896838188171, "learning_rate": 0.0004410728548300053, "loss": 3.5317, "step": 24650 }, { "epoch": 2.6630727762803232, "grad_norm": 0.5944833159446716, "learning_rate": 0.0004407490555855369, "loss": 3.519, "step": 24700 }, { "epoch": 2.6684636118598384, "grad_norm": 0.6091347932815552, "learning_rate": 0.0004404252563410685, "loss": 3.5494, "step": 24750 }, { "epoch": 2.673854447439353, "grad_norm": 0.6208272576332092, "learning_rate": 0.0004401014570966001, "loss": 3.5352, "step": 24800 }, { "epoch": 2.6792452830188678, "grad_norm": 0.5658981204032898, "learning_rate": 0.00043977765785213163, "loss": 3.5444, "step": 24850 }, { "epoch": 2.684636118598383, "grad_norm": 0.632360577583313, "learning_rate": 0.0004394538586076632, "loss": 3.5415, "step": 24900 }, { "epoch": 2.6900269541778976, "grad_norm": 0.6172274351119995, "learning_rate": 0.0004391300593631948, "loss": 3.5341, "step": 24950 }, { "epoch": 2.6954177897574123, "grad_norm": 0.5555505156517029, "learning_rate": 0.00043880626011872634, "loss": 3.533, "step": 25000 }, { "epoch": 2.6954177897574123, "eval_accuracy": 0.3680360153900296, "eval_loss": 3.5185306072235107, "eval_runtime": 183.9436, "eval_samples_per_second": 97.916, "eval_steps_per_second": 6.121, "step": 25000 }, { "epoch": 2.7008086253369274, "grad_norm": 0.6439934968948364, "learning_rate": 0.00043848246087425794, "loss": 3.5493, "step": 25050 }, { "epoch": 2.706199460916442, "grad_norm": 0.6363284587860107, "learning_rate": 0.0004381586616297895, "loss": 3.5325, "step": 25100 }, { "epoch": 2.711590296495957, "grad_norm": 0.6599604487419128, "learning_rate": 0.0004378348623853211, "loss": 3.5416, "step": 25150 }, { "epoch": 2.7169811320754715, "grad_norm": 0.626559317111969, "learning_rate": 0.00043751106314085265, "loss": 3.5485, "step": 25200 }, { "epoch": 2.7223719676549867, "grad_norm": 0.6159871816635132, "learning_rate": 0.00043718726389638425, "loss": 3.5455, "step": 25250 }, { "epoch": 2.7277628032345014, "grad_norm": 0.6169484853744507, "learning_rate": 0.00043686346465191575, "loss": 3.5336, "step": 25300 }, { "epoch": 2.733153638814016, "grad_norm": 0.6067425012588501, "learning_rate": 0.0004365396654074473, "loss": 3.5217, "step": 25350 }, { "epoch": 2.7385444743935308, "grad_norm": 0.5563640594482422, "learning_rate": 0.0004362158661629789, "loss": 3.5439, "step": 25400 }, { "epoch": 2.743935309973046, "grad_norm": 0.7104645371437073, "learning_rate": 0.00043589206691851045, "loss": 3.5473, "step": 25450 }, { "epoch": 2.7493261455525606, "grad_norm": 0.5590559244155884, "learning_rate": 0.00043556826767404206, "loss": 3.5257, "step": 25500 }, { "epoch": 2.7547169811320753, "grad_norm": 0.5929838418960571, "learning_rate": 0.0004352444684295736, "loss": 3.5374, "step": 25550 }, { "epoch": 2.7601078167115904, "grad_norm": 0.6350831985473633, "learning_rate": 0.0004349206691851052, "loss": 3.5381, "step": 25600 }, { "epoch": 2.765498652291105, "grad_norm": 0.7433158755302429, "learning_rate": 0.00043459686994063676, "loss": 3.5282, "step": 25650 }, { "epoch": 2.77088948787062, "grad_norm": 0.6278929114341736, "learning_rate": 0.00043427307069616837, "loss": 3.5368, "step": 25700 }, { "epoch": 2.776280323450135, "grad_norm": 0.6056564450263977, "learning_rate": 0.0004339492714516999, "loss": 3.5388, "step": 25750 }, { "epoch": 2.7816711590296497, "grad_norm": 0.6050384640693665, "learning_rate": 0.00043362547220723147, "loss": 3.5446, "step": 25800 }, { "epoch": 2.7870619946091644, "grad_norm": 0.5863341689109802, "learning_rate": 0.0004333016729627631, "loss": 3.5392, "step": 25850 }, { "epoch": 2.7924528301886795, "grad_norm": 0.6051446795463562, "learning_rate": 0.0004329778737182946, "loss": 3.5314, "step": 25900 }, { "epoch": 2.797843665768194, "grad_norm": 0.5601605772972107, "learning_rate": 0.00043265407447382623, "loss": 3.5512, "step": 25950 }, { "epoch": 2.803234501347709, "grad_norm": 0.6006129384040833, "learning_rate": 0.0004323302752293577, "loss": 3.5272, "step": 26000 }, { "epoch": 2.803234501347709, "eval_accuracy": 0.3687935433238722, "eval_loss": 3.511807441711426, "eval_runtime": 183.9921, "eval_samples_per_second": 97.89, "eval_steps_per_second": 6.12, "step": 26000 }, { "epoch": 2.8086253369272236, "grad_norm": 0.5731135606765747, "learning_rate": 0.00043200647598488933, "loss": 3.5367, "step": 26050 }, { "epoch": 2.8140161725067383, "grad_norm": 0.5950818061828613, "learning_rate": 0.0004316826767404209, "loss": 3.5272, "step": 26100 }, { "epoch": 2.8194070080862534, "grad_norm": 0.6465256810188293, "learning_rate": 0.00043136535348084183, "loss": 3.5273, "step": 26150 }, { "epoch": 2.824797843665768, "grad_norm": 0.6396045684814453, "learning_rate": 0.0004310415542363734, "loss": 3.5309, "step": 26200 }, { "epoch": 2.830188679245283, "grad_norm": 0.575483500957489, "learning_rate": 0.000430717754991905, "loss": 3.5388, "step": 26250 }, { "epoch": 2.835579514824798, "grad_norm": 0.5966452360153198, "learning_rate": 0.00043039395574743654, "loss": 3.523, "step": 26300 }, { "epoch": 2.8409703504043127, "grad_norm": 0.5425668954849243, "learning_rate": 0.00043007015650296814, "loss": 3.5356, "step": 26350 }, { "epoch": 2.8463611859838274, "grad_norm": 0.587317943572998, "learning_rate": 0.0004297463572584997, "loss": 3.5102, "step": 26400 }, { "epoch": 2.8517520215633425, "grad_norm": 0.578007161617279, "learning_rate": 0.0004294225580140313, "loss": 3.542, "step": 26450 }, { "epoch": 2.857142857142857, "grad_norm": 0.5723671913146973, "learning_rate": 0.00042909875876956285, "loss": 3.5329, "step": 26500 }, { "epoch": 2.862533692722372, "grad_norm": 0.5588173866271973, "learning_rate": 0.0004287749595250944, "loss": 3.5404, "step": 26550 }, { "epoch": 2.867924528301887, "grad_norm": 0.5535792112350464, "learning_rate": 0.000428451160280626, "loss": 3.5266, "step": 26600 }, { "epoch": 2.8733153638814017, "grad_norm": 0.573268711566925, "learning_rate": 0.0004281273610361575, "loss": 3.5303, "step": 26650 }, { "epoch": 2.8787061994609164, "grad_norm": 0.5750014185905457, "learning_rate": 0.0004278035617916891, "loss": 3.5032, "step": 26700 }, { "epoch": 2.884097035040431, "grad_norm": 0.6313278079032898, "learning_rate": 0.00042747976254722066, "loss": 3.5281, "step": 26750 }, { "epoch": 2.889487870619946, "grad_norm": 0.5894135236740112, "learning_rate": 0.00042715596330275226, "loss": 3.5066, "step": 26800 }, { "epoch": 2.894878706199461, "grad_norm": 0.5986575484275818, "learning_rate": 0.0004268321640582838, "loss": 3.5441, "step": 26850 }, { "epoch": 2.9002695417789757, "grad_norm": 0.5755992531776428, "learning_rate": 0.0004265083648138154, "loss": 3.5257, "step": 26900 }, { "epoch": 2.9056603773584904, "grad_norm": 0.6677845120429993, "learning_rate": 0.00042618456556934697, "loss": 3.544, "step": 26950 }, { "epoch": 2.9110512129380055, "grad_norm": 0.6155284643173218, "learning_rate": 0.0004258607663248785, "loss": 3.5211, "step": 27000 }, { "epoch": 2.9110512129380055, "eval_accuracy": 0.3694859881779139, "eval_loss": 3.50223445892334, "eval_runtime": 183.9135, "eval_samples_per_second": 97.932, "eval_steps_per_second": 6.122, "step": 27000 }, { "epoch": 2.91644204851752, "grad_norm": 0.6333293318748474, "learning_rate": 0.0004255369670804101, "loss": 3.5293, "step": 27050 }, { "epoch": 2.921832884097035, "grad_norm": 0.5835155248641968, "learning_rate": 0.0004252131678359417, "loss": 3.5198, "step": 27100 }, { "epoch": 2.92722371967655, "grad_norm": 0.5927255749702454, "learning_rate": 0.0004248893685914733, "loss": 3.5244, "step": 27150 }, { "epoch": 2.9326145552560647, "grad_norm": 0.6329957842826843, "learning_rate": 0.00042456556934700483, "loss": 3.5264, "step": 27200 }, { "epoch": 2.9380053908355794, "grad_norm": 0.6049855351448059, "learning_rate": 0.00042424177010253643, "loss": 3.5223, "step": 27250 }, { "epoch": 2.9433962264150946, "grad_norm": 0.5478748679161072, "learning_rate": 0.00042391797085806793, "loss": 3.52, "step": 27300 }, { "epoch": 2.9487870619946093, "grad_norm": 0.599579930305481, "learning_rate": 0.0004235941716135995, "loss": 3.5254, "step": 27350 }, { "epoch": 2.954177897574124, "grad_norm": 0.597079873085022, "learning_rate": 0.0004232703723691311, "loss": 3.5192, "step": 27400 }, { "epoch": 2.9595687331536387, "grad_norm": 0.5469141602516174, "learning_rate": 0.00042294657312466264, "loss": 3.5294, "step": 27450 }, { "epoch": 2.964959568733154, "grad_norm": 0.6046788692474365, "learning_rate": 0.00042262277388019424, "loss": 3.5236, "step": 27500 }, { "epoch": 2.9703504043126685, "grad_norm": 0.5884106159210205, "learning_rate": 0.0004222989746357258, "loss": 3.524, "step": 27550 }, { "epoch": 2.975741239892183, "grad_norm": 0.634127676486969, "learning_rate": 0.0004219751753912574, "loss": 3.5287, "step": 27600 }, { "epoch": 2.981132075471698, "grad_norm": 0.5886887311935425, "learning_rate": 0.00042165137614678895, "loss": 3.5248, "step": 27650 }, { "epoch": 2.986522911051213, "grad_norm": 0.5682351589202881, "learning_rate": 0.00042132757690232055, "loss": 3.5282, "step": 27700 }, { "epoch": 2.9919137466307277, "grad_norm": 0.5956007242202759, "learning_rate": 0.0004210037776578521, "loss": 3.5301, "step": 27750 }, { "epoch": 2.9973045822102424, "grad_norm": 0.6513469815254211, "learning_rate": 0.00042067997841338365, "loss": 3.499, "step": 27800 }, { "epoch": 3.0026954177897576, "grad_norm": 0.5566192865371704, "learning_rate": 0.00042035617916891526, "loss": 3.4543, "step": 27850 }, { "epoch": 3.0080862533692723, "grad_norm": 0.6176077127456665, "learning_rate": 0.0004200323799244468, "loss": 3.423, "step": 27900 }, { "epoch": 3.013477088948787, "grad_norm": 0.6107637286186218, "learning_rate": 0.0004197085806799784, "loss": 3.4323, "step": 27950 }, { "epoch": 3.018867924528302, "grad_norm": 0.6046628355979919, "learning_rate": 0.0004193847814355099, "loss": 3.4424, "step": 28000 }, { "epoch": 3.018867924528302, "eval_accuracy": 0.37057208244450746, "eval_loss": 3.4983136653900146, "eval_runtime": 183.6916, "eval_samples_per_second": 98.05, "eval_steps_per_second": 6.13, "step": 28000 }, { "epoch": 3.024258760107817, "grad_norm": 0.6203792095184326, "learning_rate": 0.0004190609821910415, "loss": 3.4246, "step": 28050 }, { "epoch": 3.0296495956873315, "grad_norm": 0.5849449038505554, "learning_rate": 0.00041873718294657306, "loss": 3.413, "step": 28100 }, { "epoch": 3.035040431266846, "grad_norm": 0.6401404142379761, "learning_rate": 0.00041841338370210467, "loss": 3.4293, "step": 28150 }, { "epoch": 3.0404312668463613, "grad_norm": 0.5675318241119385, "learning_rate": 0.00041809606044252557, "loss": 3.4324, "step": 28200 }, { "epoch": 3.045822102425876, "grad_norm": 0.5723515152931213, "learning_rate": 0.00041777226119805717, "loss": 3.4211, "step": 28250 }, { "epoch": 3.0512129380053907, "grad_norm": 0.5614615678787231, "learning_rate": 0.0004174484619535887, "loss": 3.4325, "step": 28300 }, { "epoch": 3.056603773584906, "grad_norm": 0.5907319784164429, "learning_rate": 0.00041712466270912033, "loss": 3.4343, "step": 28350 }, { "epoch": 3.0619946091644206, "grad_norm": 0.6139299869537354, "learning_rate": 0.0004168008634646519, "loss": 3.4437, "step": 28400 }, { "epoch": 3.0673854447439353, "grad_norm": 0.6088677048683167, "learning_rate": 0.0004164770642201835, "loss": 3.4369, "step": 28450 }, { "epoch": 3.07277628032345, "grad_norm": 0.6208965182304382, "learning_rate": 0.00041615326497571503, "loss": 3.4344, "step": 28500 }, { "epoch": 3.078167115902965, "grad_norm": 0.6358165144920349, "learning_rate": 0.0004158294657312466, "loss": 3.4428, "step": 28550 }, { "epoch": 3.08355795148248, "grad_norm": 0.6254170536994934, "learning_rate": 0.0004155056664867782, "loss": 3.4463, "step": 28600 }, { "epoch": 3.0889487870619945, "grad_norm": 0.585701584815979, "learning_rate": 0.0004151818672423097, "loss": 3.4471, "step": 28650 }, { "epoch": 3.0943396226415096, "grad_norm": 0.6081207394599915, "learning_rate": 0.0004148580679978413, "loss": 3.4386, "step": 28700 }, { "epoch": 3.0997304582210243, "grad_norm": 0.6378210186958313, "learning_rate": 0.00041453426875337284, "loss": 3.4258, "step": 28750 }, { "epoch": 3.105121293800539, "grad_norm": 0.6246366500854492, "learning_rate": 0.00041421046950890445, "loss": 3.4211, "step": 28800 }, { "epoch": 3.1105121293800537, "grad_norm": 0.5723994374275208, "learning_rate": 0.000413886670264436, "loss": 3.4276, "step": 28850 }, { "epoch": 3.115902964959569, "grad_norm": 0.6091293692588806, "learning_rate": 0.0004135628710199676, "loss": 3.4424, "step": 28900 }, { "epoch": 3.1212938005390836, "grad_norm": 0.5650001168251038, "learning_rate": 0.00041323907177549915, "loss": 3.4413, "step": 28950 }, { "epoch": 3.1266846361185983, "grad_norm": 0.6242942214012146, "learning_rate": 0.0004129152725310307, "loss": 3.4427, "step": 29000 }, { "epoch": 3.1266846361185983, "eval_accuracy": 0.37073506177622917, "eval_loss": 3.4975240230560303, "eval_runtime": 183.9334, "eval_samples_per_second": 97.921, "eval_steps_per_second": 6.122, "step": 29000 }, { "epoch": 3.1320754716981134, "grad_norm": 0.5982543230056763, "learning_rate": 0.0004125914732865623, "loss": 3.4407, "step": 29050 }, { "epoch": 3.137466307277628, "grad_norm": 0.584873616695404, "learning_rate": 0.00041226767404209386, "loss": 3.4539, "step": 29100 }, { "epoch": 3.142857142857143, "grad_norm": 0.6070249676704407, "learning_rate": 0.00041194387479762546, "loss": 3.4543, "step": 29150 }, { "epoch": 3.1482479784366575, "grad_norm": 0.6300134658813477, "learning_rate": 0.000411620075553157, "loss": 3.4457, "step": 29200 }, { "epoch": 3.1536388140161726, "grad_norm": 0.6087648868560791, "learning_rate": 0.0004112962763086886, "loss": 3.4461, "step": 29250 }, { "epoch": 3.1590296495956873, "grad_norm": 0.6131951808929443, "learning_rate": 0.0004109724770642201, "loss": 3.4512, "step": 29300 }, { "epoch": 3.164420485175202, "grad_norm": 0.6422128677368164, "learning_rate": 0.00041064867781975177, "loss": 3.4648, "step": 29350 }, { "epoch": 3.169811320754717, "grad_norm": 0.7360274791717529, "learning_rate": 0.00041032487857528327, "loss": 3.4428, "step": 29400 }, { "epoch": 3.175202156334232, "grad_norm": 0.5676162838935852, "learning_rate": 0.0004100010793308148, "loss": 3.4475, "step": 29450 }, { "epoch": 3.1805929919137466, "grad_norm": 0.5553027987480164, "learning_rate": 0.0004096772800863464, "loss": 3.4486, "step": 29500 }, { "epoch": 3.1859838274932613, "grad_norm": 0.6182715892791748, "learning_rate": 0.000409353480841878, "loss": 3.4488, "step": 29550 }, { "epoch": 3.1913746630727764, "grad_norm": 0.5933767557144165, "learning_rate": 0.0004090296815974096, "loss": 3.4744, "step": 29600 }, { "epoch": 3.196765498652291, "grad_norm": 0.5704579949378967, "learning_rate": 0.00040870588235294113, "loss": 3.4456, "step": 29650 }, { "epoch": 3.202156334231806, "grad_norm": 0.6044793128967285, "learning_rate": 0.00040838208310847273, "loss": 3.455, "step": 29700 }, { "epoch": 3.207547169811321, "grad_norm": 0.5789558291435242, "learning_rate": 0.0004080582838640043, "loss": 3.4687, "step": 29750 }, { "epoch": 3.2129380053908356, "grad_norm": 0.6031609177589417, "learning_rate": 0.0004077344846195359, "loss": 3.4433, "step": 29800 }, { "epoch": 3.2183288409703503, "grad_norm": 0.6147695183753967, "learning_rate": 0.00040741068537506744, "loss": 3.4398, "step": 29850 }, { "epoch": 3.223719676549865, "grad_norm": 0.6561545133590698, "learning_rate": 0.000407086886130599, "loss": 3.4364, "step": 29900 }, { "epoch": 3.22911051212938, "grad_norm": 0.6324929594993591, "learning_rate": 0.0004067630868861306, "loss": 3.4486, "step": 29950 }, { "epoch": 3.234501347708895, "grad_norm": 0.6108676195144653, "learning_rate": 0.0004064392876416621, "loss": 3.4427, "step": 30000 }, { "epoch": 3.234501347708895, "eval_accuracy": 0.3712915818676149, "eval_loss": 3.490433931350708, "eval_runtime": 183.7094, "eval_samples_per_second": 98.041, "eval_steps_per_second": 6.129, "step": 30000 }, { "epoch": 3.2398921832884096, "grad_norm": 0.627756655216217, "learning_rate": 0.0004061154883971937, "loss": 3.4489, "step": 30050 }, { "epoch": 3.2452830188679247, "grad_norm": 0.6552205681800842, "learning_rate": 0.00040579168915272525, "loss": 3.4595, "step": 30100 }, { "epoch": 3.2506738544474394, "grad_norm": 0.605854868888855, "learning_rate": 0.00040546788990825685, "loss": 3.4753, "step": 30150 }, { "epoch": 3.256064690026954, "grad_norm": 0.6816122531890869, "learning_rate": 0.0004051440906637884, "loss": 3.4714, "step": 30200 }, { "epoch": 3.2614555256064692, "grad_norm": 0.6047664880752563, "learning_rate": 0.00040482029141931995, "loss": 3.458, "step": 30250 }, { "epoch": 3.266846361185984, "grad_norm": 0.6369194984436035, "learning_rate": 0.00040449649217485156, "loss": 3.4702, "step": 30300 }, { "epoch": 3.2722371967654986, "grad_norm": 0.5737183690071106, "learning_rate": 0.0004041791689152725, "loss": 3.4396, "step": 30350 }, { "epoch": 3.2776280323450133, "grad_norm": 0.6164860725402832, "learning_rate": 0.00040385536967080406, "loss": 3.443, "step": 30400 }, { "epoch": 3.2830188679245285, "grad_norm": 0.5741621255874634, "learning_rate": 0.00040353157042633567, "loss": 3.4596, "step": 30450 }, { "epoch": 3.288409703504043, "grad_norm": 0.6739873886108398, "learning_rate": 0.0004032077711818672, "loss": 3.4586, "step": 30500 }, { "epoch": 3.293800539083558, "grad_norm": 0.6138100624084473, "learning_rate": 0.0004028839719373988, "loss": 3.4622, "step": 30550 }, { "epoch": 3.2991913746630726, "grad_norm": 0.6170074939727783, "learning_rate": 0.00040256017269293037, "loss": 3.4439, "step": 30600 }, { "epoch": 3.3045822102425877, "grad_norm": 0.6086063981056213, "learning_rate": 0.00040223637344846187, "loss": 3.4403, "step": 30650 }, { "epoch": 3.3099730458221024, "grad_norm": 0.6176971793174744, "learning_rate": 0.00040191257420399347, "loss": 3.4583, "step": 30700 }, { "epoch": 3.315363881401617, "grad_norm": 0.6030409336090088, "learning_rate": 0.000401588774959525, "loss": 3.4435, "step": 30750 }, { "epoch": 3.3207547169811322, "grad_norm": 0.5938195586204529, "learning_rate": 0.00040126497571505663, "loss": 3.457, "step": 30800 }, { "epoch": 3.326145552560647, "grad_norm": 0.6070782542228699, "learning_rate": 0.0004009411764705882, "loss": 3.4617, "step": 30850 }, { "epoch": 3.3315363881401616, "grad_norm": 0.6172419190406799, "learning_rate": 0.0004006173772261198, "loss": 3.4423, "step": 30900 }, { "epoch": 3.3369272237196768, "grad_norm": 0.5756070017814636, "learning_rate": 0.00040029357798165133, "loss": 3.4723, "step": 30950 }, { "epoch": 3.3423180592991915, "grad_norm": 0.7206381559371948, "learning_rate": 0.00039996977873718294, "loss": 3.445, "step": 31000 }, { "epoch": 3.3423180592991915, "eval_accuracy": 0.37233475824352175, "eval_loss": 3.4826910495758057, "eval_runtime": 184.0439, "eval_samples_per_second": 97.862, "eval_steps_per_second": 6.118, "step": 31000 }, { "epoch": 3.347708894878706, "grad_norm": 0.6098358631134033, "learning_rate": 0.0003996459794927145, "loss": 3.4583, "step": 31050 }, { "epoch": 3.353099730458221, "grad_norm": 0.625518798828125, "learning_rate": 0.00039932218024824604, "loss": 3.4487, "step": 31100 }, { "epoch": 3.358490566037736, "grad_norm": 0.5937629342079163, "learning_rate": 0.00039899838100377764, "loss": 3.4645, "step": 31150 }, { "epoch": 3.3638814016172507, "grad_norm": 0.5838909149169922, "learning_rate": 0.0003986745817593092, "loss": 3.4557, "step": 31200 }, { "epoch": 3.3692722371967654, "grad_norm": 0.6747406721115112, "learning_rate": 0.0003983507825148408, "loss": 3.4407, "step": 31250 }, { "epoch": 3.37466307277628, "grad_norm": 0.6010386943817139, "learning_rate": 0.00039802698327037235, "loss": 3.4678, "step": 31300 }, { "epoch": 3.3800539083557952, "grad_norm": 0.6182563900947571, "learning_rate": 0.00039770318402590396, "loss": 3.4691, "step": 31350 }, { "epoch": 3.38544474393531, "grad_norm": 0.5988255143165588, "learning_rate": 0.00039737938478143545, "loss": 3.4774, "step": 31400 }, { "epoch": 3.3908355795148246, "grad_norm": 0.5775631666183472, "learning_rate": 0.000397055585536967, "loss": 3.4581, "step": 31450 }, { "epoch": 3.3962264150943398, "grad_norm": 0.6162555813789368, "learning_rate": 0.0003967317862924986, "loss": 3.4486, "step": 31500 }, { "epoch": 3.4016172506738545, "grad_norm": 0.6046192646026611, "learning_rate": 0.00039640798704803016, "loss": 3.4534, "step": 31550 }, { "epoch": 3.407008086253369, "grad_norm": 0.6041988730430603, "learning_rate": 0.00039608418780356176, "loss": 3.4575, "step": 31600 }, { "epoch": 3.4123989218328843, "grad_norm": 0.623914897441864, "learning_rate": 0.0003957603885590933, "loss": 3.4494, "step": 31650 }, { "epoch": 3.417789757412399, "grad_norm": 0.5987055897712708, "learning_rate": 0.0003954365893146249, "loss": 3.4583, "step": 31700 }, { "epoch": 3.4231805929919137, "grad_norm": 0.5991863012313843, "learning_rate": 0.00039511279007015647, "loss": 3.4641, "step": 31750 }, { "epoch": 3.4285714285714284, "grad_norm": 0.6010271310806274, "learning_rate": 0.00039478899082568807, "loss": 3.4433, "step": 31800 }, { "epoch": 3.4339622641509435, "grad_norm": 0.5700473189353943, "learning_rate": 0.0003944651915812196, "loss": 3.4547, "step": 31850 }, { "epoch": 3.439353099730458, "grad_norm": 0.5874118208885193, "learning_rate": 0.0003941413923367512, "loss": 3.47, "step": 31900 }, { "epoch": 3.444743935309973, "grad_norm": 0.5711833834648132, "learning_rate": 0.0003938175930922828, "loss": 3.4649, "step": 31950 }, { "epoch": 3.450134770889488, "grad_norm": 0.6957682967185974, "learning_rate": 0.0003934937938478143, "loss": 3.4717, "step": 32000 }, { "epoch": 3.450134770889488, "eval_accuracy": 0.37300981863551314, "eval_loss": 3.476459264755249, "eval_runtime": 183.7361, "eval_samples_per_second": 98.026, "eval_steps_per_second": 6.128, "step": 32000 }, { "epoch": 3.4555256064690028, "grad_norm": 0.5690134167671204, "learning_rate": 0.0003931699946033459, "loss": 3.4463, "step": 32050 }, { "epoch": 3.4609164420485174, "grad_norm": 0.6350628733634949, "learning_rate": 0.00039284619535887743, "loss": 3.4583, "step": 32100 }, { "epoch": 3.466307277628032, "grad_norm": 0.5968349575996399, "learning_rate": 0.00039252239611440904, "loss": 3.459, "step": 32150 }, { "epoch": 3.4716981132075473, "grad_norm": 0.6233981251716614, "learning_rate": 0.0003921985968699406, "loss": 3.4794, "step": 32200 }, { "epoch": 3.477088948787062, "grad_norm": 0.6959568858146667, "learning_rate": 0.0003918747976254722, "loss": 3.4749, "step": 32250 }, { "epoch": 3.4824797843665767, "grad_norm": 0.6333785653114319, "learning_rate": 0.00039155099838100374, "loss": 3.4438, "step": 32300 }, { "epoch": 3.487870619946092, "grad_norm": 0.6264799237251282, "learning_rate": 0.0003912336751214247, "loss": 3.4413, "step": 32350 }, { "epoch": 3.4932614555256065, "grad_norm": 0.6589958071708679, "learning_rate": 0.00039090987587695624, "loss": 3.4783, "step": 32400 }, { "epoch": 3.498652291105121, "grad_norm": 0.5954399704933167, "learning_rate": 0.00039058607663248785, "loss": 3.4627, "step": 32450 }, { "epoch": 3.5040431266846364, "grad_norm": 0.6438902616500854, "learning_rate": 0.0003902622773880194, "loss": 3.4567, "step": 32500 }, { "epoch": 3.509433962264151, "grad_norm": 0.7091028690338135, "learning_rate": 0.000389938478143551, "loss": 3.4599, "step": 32550 }, { "epoch": 3.5148247978436657, "grad_norm": 0.625196099281311, "learning_rate": 0.00038961467889908255, "loss": 3.4453, "step": 32600 }, { "epoch": 3.5202156334231804, "grad_norm": 0.6431267857551575, "learning_rate": 0.00038929087965461405, "loss": 3.453, "step": 32650 }, { "epoch": 3.525606469002695, "grad_norm": 0.6067655682563782, "learning_rate": 0.00038896708041014566, "loss": 3.4535, "step": 32700 }, { "epoch": 3.5309973045822103, "grad_norm": 0.5781969428062439, "learning_rate": 0.0003886432811656772, "loss": 3.4454, "step": 32750 }, { "epoch": 3.536388140161725, "grad_norm": 0.6490238904953003, "learning_rate": 0.0003883194819212088, "loss": 3.4601, "step": 32800 }, { "epoch": 3.5417789757412397, "grad_norm": 0.6111918091773987, "learning_rate": 0.00038799568267674036, "loss": 3.4692, "step": 32850 }, { "epoch": 3.547169811320755, "grad_norm": 0.6645634174346924, "learning_rate": 0.00038767188343227197, "loss": 3.4725, "step": 32900 }, { "epoch": 3.5525606469002695, "grad_norm": 0.9656093716621399, "learning_rate": 0.0003873480841878035, "loss": 3.4587, "step": 32950 }, { "epoch": 3.557951482479784, "grad_norm": 0.629120409488678, "learning_rate": 0.0003870242849433351, "loss": 3.4426, "step": 33000 }, { "epoch": 3.557951482479784, "eval_accuracy": 0.37355123597549267, "eval_loss": 3.4684126377105713, "eval_runtime": 184.2402, "eval_samples_per_second": 97.758, "eval_steps_per_second": 6.112, "step": 33000 }, { "epoch": 3.5633423180592994, "grad_norm": 0.5992081165313721, "learning_rate": 0.00038670048569886667, "loss": 3.4547, "step": 33050 }, { "epoch": 3.568733153638814, "grad_norm": 0.6156100630760193, "learning_rate": 0.0003863766864543982, "loss": 3.466, "step": 33100 }, { "epoch": 3.5741239892183287, "grad_norm": 0.6006748080253601, "learning_rate": 0.00038605288720992983, "loss": 3.4431, "step": 33150 }, { "epoch": 3.579514824797844, "grad_norm": 0.6171281933784485, "learning_rate": 0.0003857290879654614, "loss": 3.4524, "step": 33200 }, { "epoch": 3.5849056603773586, "grad_norm": 0.781915545463562, "learning_rate": 0.000385405288720993, "loss": 3.4424, "step": 33250 }, { "epoch": 3.5902964959568733, "grad_norm": 0.6024929285049438, "learning_rate": 0.00038508148947652453, "loss": 3.4464, "step": 33300 }, { "epoch": 3.595687331536388, "grad_norm": 0.6088691353797913, "learning_rate": 0.00038475769023205614, "loss": 3.4655, "step": 33350 }, { "epoch": 3.601078167115903, "grad_norm": 0.6233068704605103, "learning_rate": 0.00038443389098758763, "loss": 3.4556, "step": 33400 }, { "epoch": 3.606469002695418, "grad_norm": 0.6492874026298523, "learning_rate": 0.00038411009174311924, "loss": 3.4574, "step": 33450 }, { "epoch": 3.6118598382749325, "grad_norm": 0.588290810585022, "learning_rate": 0.0003837862924986508, "loss": 3.4696, "step": 33500 }, { "epoch": 3.617250673854447, "grad_norm": 0.6026934385299683, "learning_rate": 0.00038346249325418234, "loss": 3.4575, "step": 33550 }, { "epoch": 3.6226415094339623, "grad_norm": 0.635048508644104, "learning_rate": 0.00038313869400971395, "loss": 3.4538, "step": 33600 }, { "epoch": 3.628032345013477, "grad_norm": 0.592736005783081, "learning_rate": 0.0003828148947652455, "loss": 3.4527, "step": 33650 }, { "epoch": 3.6334231805929917, "grad_norm": 0.6240489482879639, "learning_rate": 0.0003824910955207771, "loss": 3.4532, "step": 33700 }, { "epoch": 3.638814016172507, "grad_norm": 0.6334491968154907, "learning_rate": 0.00038216729627630865, "loss": 3.4497, "step": 33750 }, { "epoch": 3.6442048517520216, "grad_norm": 0.6171436309814453, "learning_rate": 0.00038184349703184026, "loss": 3.4446, "step": 33800 }, { "epoch": 3.6495956873315363, "grad_norm": 0.6213189363479614, "learning_rate": 0.0003815196977873718, "loss": 3.4504, "step": 33850 }, { "epoch": 3.6549865229110514, "grad_norm": 0.5768414735794067, "learning_rate": 0.0003811958985429034, "loss": 3.4571, "step": 33900 }, { "epoch": 3.660377358490566, "grad_norm": 0.5723716616630554, "learning_rate": 0.00038087209929843496, "loss": 3.4619, "step": 33950 }, { "epoch": 3.665768194070081, "grad_norm": 0.5816717147827148, "learning_rate": 0.00038054830005396646, "loss": 3.4603, "step": 34000 }, { "epoch": 3.665768194070081, "eval_accuracy": 0.37412351073561156, "eval_loss": 3.4615986347198486, "eval_runtime": 183.9553, "eval_samples_per_second": 97.91, "eval_steps_per_second": 6.121, "step": 34000 }, { "epoch": 3.671159029649596, "grad_norm": 0.5690119862556458, "learning_rate": 0.00038022450080949806, "loss": 3.4572, "step": 34050 }, { "epoch": 3.6765498652291106, "grad_norm": 0.6545220613479614, "learning_rate": 0.0003799007015650296, "loss": 3.4421, "step": 34100 }, { "epoch": 3.6819407008086253, "grad_norm": 0.6318289041519165, "learning_rate": 0.0003795769023205612, "loss": 3.4508, "step": 34150 }, { "epoch": 3.68733153638814, "grad_norm": 0.6190627217292786, "learning_rate": 0.00037925310307609277, "loss": 3.456, "step": 34200 }, { "epoch": 3.6927223719676547, "grad_norm": 0.5709021687507629, "learning_rate": 0.0003789293038316244, "loss": 3.4468, "step": 34250 }, { "epoch": 3.69811320754717, "grad_norm": 0.6180399060249329, "learning_rate": 0.0003786055045871559, "loss": 3.4551, "step": 34300 }, { "epoch": 3.7035040431266846, "grad_norm": 0.6174043416976929, "learning_rate": 0.0003782817053426875, "loss": 3.4556, "step": 34350 }, { "epoch": 3.7088948787061993, "grad_norm": 0.6184794902801514, "learning_rate": 0.0003779579060982191, "loss": 3.4495, "step": 34400 }, { "epoch": 3.7142857142857144, "grad_norm": 0.6114649176597595, "learning_rate": 0.00037763410685375063, "loss": 3.4795, "step": 34450 }, { "epoch": 3.719676549865229, "grad_norm": 0.6328994035720825, "learning_rate": 0.00037731030760928223, "loss": 3.4406, "step": 34500 }, { "epoch": 3.725067385444744, "grad_norm": 0.6460617184638977, "learning_rate": 0.0003769929843497032, "loss": 3.45, "step": 34550 }, { "epoch": 3.730458221024259, "grad_norm": 0.7519969344139099, "learning_rate": 0.00037666918510523474, "loss": 3.4476, "step": 34600 }, { "epoch": 3.7358490566037736, "grad_norm": 0.591691792011261, "learning_rate": 0.00037634538586076634, "loss": 3.4637, "step": 34650 }, { "epoch": 3.7412398921832883, "grad_norm": 0.6448811292648315, "learning_rate": 0.00037602158661629784, "loss": 3.449, "step": 34700 }, { "epoch": 3.7466307277628035, "grad_norm": 0.6556087136268616, "learning_rate": 0.0003756977873718294, "loss": 3.4546, "step": 34750 }, { "epoch": 3.752021563342318, "grad_norm": 0.6165753602981567, "learning_rate": 0.000375373988127361, "loss": 3.4493, "step": 34800 }, { "epoch": 3.757412398921833, "grad_norm": 0.6732285618782043, "learning_rate": 0.00037505018888289254, "loss": 3.4393, "step": 34850 }, { "epoch": 3.7628032345013476, "grad_norm": 0.6393305063247681, "learning_rate": 0.00037472638963842415, "loss": 3.4489, "step": 34900 }, { "epoch": 3.7681940700808623, "grad_norm": 0.6352855563163757, "learning_rate": 0.0003744025903939557, "loss": 3.4633, "step": 34950 }, { "epoch": 3.7735849056603774, "grad_norm": 0.6707683205604553, "learning_rate": 0.0003740787911494873, "loss": 3.4463, "step": 35000 }, { "epoch": 3.7735849056603774, "eval_accuracy": 0.3749070067096418, "eval_loss": 3.4561455249786377, "eval_runtime": 183.9041, "eval_samples_per_second": 97.937, "eval_steps_per_second": 6.123, "step": 35000 }, { "epoch": 3.778975741239892, "grad_norm": 0.6262804269790649, "learning_rate": 0.00037375499190501885, "loss": 3.4567, "step": 35050 }, { "epoch": 3.784366576819407, "grad_norm": 0.61308354139328, "learning_rate": 0.0003734311926605504, "loss": 3.4635, "step": 35100 }, { "epoch": 3.789757412398922, "grad_norm": 0.6734222173690796, "learning_rate": 0.000373107393416082, "loss": 3.4569, "step": 35150 }, { "epoch": 3.7951482479784366, "grad_norm": 0.5994874835014343, "learning_rate": 0.00037278359417161356, "loss": 3.4739, "step": 35200 }, { "epoch": 3.8005390835579513, "grad_norm": 0.6053499579429626, "learning_rate": 0.00037245979492714517, "loss": 3.4575, "step": 35250 }, { "epoch": 3.8059299191374665, "grad_norm": 0.6050474643707275, "learning_rate": 0.0003721359956826767, "loss": 3.4419, "step": 35300 }, { "epoch": 3.811320754716981, "grad_norm": 0.7089765071868896, "learning_rate": 0.0003718121964382083, "loss": 3.471, "step": 35350 }, { "epoch": 3.816711590296496, "grad_norm": 0.6321031451225281, "learning_rate": 0.0003714883971937398, "loss": 3.4344, "step": 35400 }, { "epoch": 3.822102425876011, "grad_norm": 0.655995786190033, "learning_rate": 0.0003711645979492714, "loss": 3.457, "step": 35450 }, { "epoch": 3.8274932614555257, "grad_norm": 0.617946445941925, "learning_rate": 0.00037084079870480297, "loss": 3.4314, "step": 35500 }, { "epoch": 3.8328840970350404, "grad_norm": 0.6026831865310669, "learning_rate": 0.0003705169994603345, "loss": 3.464, "step": 35550 }, { "epoch": 3.838274932614555, "grad_norm": 0.6153815388679504, "learning_rate": 0.00037019320021586613, "loss": 3.4441, "step": 35600 }, { "epoch": 3.8436657681940702, "grad_norm": 0.6427329778671265, "learning_rate": 0.0003698694009713977, "loss": 3.4645, "step": 35650 }, { "epoch": 3.849056603773585, "grad_norm": 0.6545435786247253, "learning_rate": 0.0003695456017269293, "loss": 3.4523, "step": 35700 }, { "epoch": 3.8544474393530996, "grad_norm": 0.6320939064025879, "learning_rate": 0.00036922180248246083, "loss": 3.4408, "step": 35750 }, { "epoch": 3.8598382749326143, "grad_norm": 0.6298137307167053, "learning_rate": 0.00036889800323799244, "loss": 3.46, "step": 35800 }, { "epoch": 3.8652291105121295, "grad_norm": 0.6715940237045288, "learning_rate": 0.000368574203993524, "loss": 3.4544, "step": 35850 }, { "epoch": 3.870619946091644, "grad_norm": 0.5963342189788818, "learning_rate": 0.0003682504047490556, "loss": 3.4677, "step": 35900 }, { "epoch": 3.876010781671159, "grad_norm": 0.6674957275390625, "learning_rate": 0.00036792660550458714, "loss": 3.452, "step": 35950 }, { "epoch": 3.881401617250674, "grad_norm": 0.626230001449585, "learning_rate": 0.00036760280626011864, "loss": 3.4731, "step": 36000 }, { "epoch": 3.881401617250674, "eval_accuracy": 0.37561976965370475, "eval_loss": 3.4497599601745605, "eval_runtime": 184.0301, "eval_samples_per_second": 97.87, "eval_steps_per_second": 6.119, "step": 36000 }, { "epoch": 3.8867924528301887, "grad_norm": 0.6103031039237976, "learning_rate": 0.0003672790070156503, "loss": 3.4362, "step": 36050 }, { "epoch": 3.8921832884097034, "grad_norm": 0.6194986701011658, "learning_rate": 0.0003669552077711818, "loss": 3.4354, "step": 36100 }, { "epoch": 3.8975741239892185, "grad_norm": 0.6245982050895691, "learning_rate": 0.0003666314085267134, "loss": 3.4358, "step": 36150 }, { "epoch": 3.9029649595687332, "grad_norm": 0.5864128470420837, "learning_rate": 0.00036630760928224495, "loss": 3.4409, "step": 36200 }, { "epoch": 3.908355795148248, "grad_norm": 0.6105799674987793, "learning_rate": 0.00036598381003777656, "loss": 3.4513, "step": 36250 }, { "epoch": 3.913746630727763, "grad_norm": 0.6365011930465698, "learning_rate": 0.0003656600107933081, "loss": 3.4527, "step": 36300 }, { "epoch": 3.9191374663072778, "grad_norm": 0.6909796595573425, "learning_rate": 0.0003653362115488397, "loss": 3.4705, "step": 36350 }, { "epoch": 3.9245283018867925, "grad_norm": 0.795727014541626, "learning_rate": 0.00036501241230437126, "loss": 3.4567, "step": 36400 }, { "epoch": 3.929919137466307, "grad_norm": 0.6463427543640137, "learning_rate": 0.0003646886130599028, "loss": 3.4496, "step": 36450 }, { "epoch": 3.935309973045822, "grad_norm": 0.6198216676712036, "learning_rate": 0.0003643648138154344, "loss": 3.4657, "step": 36500 }, { "epoch": 3.940700808625337, "grad_norm": 0.6134477853775024, "learning_rate": 0.00036404101457096597, "loss": 3.4736, "step": 36550 }, { "epoch": 3.9460916442048517, "grad_norm": 0.6083407998085022, "learning_rate": 0.00036371721532649757, "loss": 3.4535, "step": 36600 }, { "epoch": 3.9514824797843664, "grad_norm": 0.617780864238739, "learning_rate": 0.0003633934160820291, "loss": 3.4614, "step": 36650 }, { "epoch": 3.9568733153638815, "grad_norm": 0.620871365070343, "learning_rate": 0.00036306961683756073, "loss": 3.4446, "step": 36700 }, { "epoch": 3.9622641509433962, "grad_norm": 0.684502124786377, "learning_rate": 0.0003627458175930922, "loss": 3.452, "step": 36750 }, { "epoch": 3.967654986522911, "grad_norm": 0.6206174492835999, "learning_rate": 0.0003624220183486238, "loss": 3.4475, "step": 36800 }, { "epoch": 3.973045822102426, "grad_norm": 0.6756080389022827, "learning_rate": 0.0003620982191041554, "loss": 3.4326, "step": 36850 }, { "epoch": 3.9784366576819408, "grad_norm": 0.618306577205658, "learning_rate": 0.00036177441985968693, "loss": 3.4326, "step": 36900 }, { "epoch": 3.9838274932614555, "grad_norm": 0.6362138390541077, "learning_rate": 0.00036145062061521854, "loss": 3.4391, "step": 36950 }, { "epoch": 3.9892183288409706, "grad_norm": 0.6646049618721008, "learning_rate": 0.0003611268213707501, "loss": 3.4391, "step": 37000 }, { "epoch": 3.9892183288409706, "eval_accuracy": 0.3763430719278858, "eval_loss": 3.442918539047241, "eval_runtime": 183.962, "eval_samples_per_second": 97.906, "eval_steps_per_second": 6.121, "step": 37000 }, { "epoch": 3.9946091644204853, "grad_norm": 0.6206597089767456, "learning_rate": 0.0003608030221262817, "loss": 3.4462, "step": 37050 }, { "epoch": 4.0, "grad_norm": 1.2068943977355957, "learning_rate": 0.00036047922288181324, "loss": 3.4468, "step": 37100 }, { "epoch": 4.005390835579515, "grad_norm": 0.6037599444389343, "learning_rate": 0.00036015542363734485, "loss": 3.3481, "step": 37150 }, { "epoch": 4.010781671159029, "grad_norm": 0.6229018568992615, "learning_rate": 0.0003598316243928764, "loss": 3.352, "step": 37200 }, { "epoch": 4.0161725067385445, "grad_norm": 0.6315988302230835, "learning_rate": 0.00035950782514840795, "loss": 3.345, "step": 37250 }, { "epoch": 4.02156334231806, "grad_norm": 0.601076602935791, "learning_rate": 0.00035918402590393955, "loss": 3.3492, "step": 37300 }, { "epoch": 4.026954177897574, "grad_norm": 0.6532601714134216, "learning_rate": 0.00035886022665947105, "loss": 3.3611, "step": 37350 }, { "epoch": 4.032345013477089, "grad_norm": 0.6189490556716919, "learning_rate": 0.0003585364274150027, "loss": 3.3661, "step": 37400 }, { "epoch": 4.037735849056604, "grad_norm": 0.6153668761253357, "learning_rate": 0.0003582126281705342, "loss": 3.3486, "step": 37450 }, { "epoch": 4.0431266846361185, "grad_norm": 0.6020147204399109, "learning_rate": 0.0003578888289260658, "loss": 3.3675, "step": 37500 }, { "epoch": 4.048517520215634, "grad_norm": 0.6158584356307983, "learning_rate": 0.00035756502968159736, "loss": 3.3505, "step": 37550 }, { "epoch": 4.053908355795148, "grad_norm": 0.649570643901825, "learning_rate": 0.00035724123043712896, "loss": 3.3583, "step": 37600 }, { "epoch": 4.059299191374663, "grad_norm": 0.6425063014030457, "learning_rate": 0.0003569174311926605, "loss": 3.3569, "step": 37650 }, { "epoch": 4.064690026954178, "grad_norm": 0.6738588213920593, "learning_rate": 0.00035659363194819206, "loss": 3.3734, "step": 37700 }, { "epoch": 4.070080862533692, "grad_norm": 0.6135583519935608, "learning_rate": 0.00035626983270372367, "loss": 3.3618, "step": 37750 }, { "epoch": 4.0754716981132075, "grad_norm": 0.6319704055786133, "learning_rate": 0.0003559460334592552, "loss": 3.3697, "step": 37800 }, { "epoch": 4.080862533692723, "grad_norm": 0.6221917271614075, "learning_rate": 0.0003556222342147868, "loss": 3.3572, "step": 37850 }, { "epoch": 4.086253369272237, "grad_norm": 0.6539017558097839, "learning_rate": 0.0003552984349703184, "loss": 3.368, "step": 37900 }, { "epoch": 4.091644204851752, "grad_norm": 0.6019672155380249, "learning_rate": 0.00035497463572585, "loss": 3.3568, "step": 37950 }, { "epoch": 4.097035040431267, "grad_norm": 0.6425307989120483, "learning_rate": 0.00035465083648138153, "loss": 3.3773, "step": 38000 }, { "epoch": 4.097035040431267, "eval_accuracy": 0.37657037376919367, "eval_loss": 3.4480812549591064, "eval_runtime": 183.7404, "eval_samples_per_second": 98.024, "eval_steps_per_second": 6.128, "step": 38000 }, { "epoch": 4.1024258760107815, "grad_norm": 0.6474397778511047, "learning_rate": 0.00035432703723691314, "loss": 3.3589, "step": 38050 }, { "epoch": 4.107816711590297, "grad_norm": 0.6612837314605713, "learning_rate": 0.00035400323799244463, "loss": 3.3826, "step": 38100 }, { "epoch": 4.113207547169812, "grad_norm": 0.6642338037490845, "learning_rate": 0.0003536794387479762, "loss": 3.3693, "step": 38150 }, { "epoch": 4.118598382749326, "grad_norm": 0.6186237931251526, "learning_rate": 0.0003533556395035078, "loss": 3.3729, "step": 38200 }, { "epoch": 4.123989218328841, "grad_norm": 0.5964476466178894, "learning_rate": 0.00035303184025903934, "loss": 3.3645, "step": 38250 }, { "epoch": 4.129380053908355, "grad_norm": 0.6504191160202026, "learning_rate": 0.00035270804101457094, "loss": 3.3925, "step": 38300 }, { "epoch": 4.1347708894878705, "grad_norm": 0.6848699450492859, "learning_rate": 0.0003523842417701025, "loss": 3.3713, "step": 38350 }, { "epoch": 4.140161725067386, "grad_norm": 0.6712897419929504, "learning_rate": 0.00035206691851052345, "loss": 3.3645, "step": 38400 }, { "epoch": 4.1455525606469, "grad_norm": 0.6874973773956299, "learning_rate": 0.000351743119266055, "loss": 3.3804, "step": 38450 }, { "epoch": 4.150943396226415, "grad_norm": 0.6249614357948303, "learning_rate": 0.0003514193200215866, "loss": 3.3727, "step": 38500 }, { "epoch": 4.15633423180593, "grad_norm": 0.6266210675239563, "learning_rate": 0.00035109552077711815, "loss": 3.3783, "step": 38550 }, { "epoch": 4.1617250673854445, "grad_norm": 0.6524613499641418, "learning_rate": 0.00035077172153264976, "loss": 3.3802, "step": 38600 }, { "epoch": 4.16711590296496, "grad_norm": 0.6768172383308411, "learning_rate": 0.0003504479222881813, "loss": 3.3788, "step": 38650 }, { "epoch": 4.172506738544475, "grad_norm": 0.7132668495178223, "learning_rate": 0.0003501241230437129, "loss": 3.3795, "step": 38700 }, { "epoch": 4.177897574123989, "grad_norm": 0.6380689144134521, "learning_rate": 0.0003498003237992444, "loss": 3.379, "step": 38750 }, { "epoch": 4.183288409703504, "grad_norm": 0.6313562393188477, "learning_rate": 0.000349476524554776, "loss": 3.3853, "step": 38800 }, { "epoch": 4.188679245283019, "grad_norm": 0.6632422804832458, "learning_rate": 0.00034915272531030756, "loss": 3.3703, "step": 38850 }, { "epoch": 4.1940700808625335, "grad_norm": 0.6951757073402405, "learning_rate": 0.0003488289260658391, "loss": 3.3957, "step": 38900 }, { "epoch": 4.199460916442049, "grad_norm": 0.6744063496589661, "learning_rate": 0.0003485051268213707, "loss": 3.4029, "step": 38950 }, { "epoch": 4.204851752021563, "grad_norm": 0.6867125034332275, "learning_rate": 0.00034818132757690227, "loss": 3.3765, "step": 39000 }, { "epoch": 4.204851752021563, "eval_accuracy": 0.37701585060923304, "eval_loss": 3.4399898052215576, "eval_runtime": 183.7265, "eval_samples_per_second": 98.032, "eval_steps_per_second": 6.129, "step": 39000 }, { "epoch": 4.210242587601078, "grad_norm": 0.7025331854820251, "learning_rate": 0.0003478575283324339, "loss": 3.3783, "step": 39050 }, { "epoch": 4.215633423180593, "grad_norm": 0.620579719543457, "learning_rate": 0.0003475337290879654, "loss": 3.3775, "step": 39100 }, { "epoch": 4.2210242587601075, "grad_norm": 0.6752045154571533, "learning_rate": 0.00034720992984349703, "loss": 3.3801, "step": 39150 }, { "epoch": 4.226415094339623, "grad_norm": 0.6968411803245544, "learning_rate": 0.0003468861305990286, "loss": 3.3906, "step": 39200 }, { "epoch": 4.231805929919138, "grad_norm": 0.6568784117698669, "learning_rate": 0.0003465623313545602, "loss": 3.3789, "step": 39250 }, { "epoch": 4.237196765498652, "grad_norm": 0.6356040239334106, "learning_rate": 0.00034623853211009173, "loss": 3.3842, "step": 39300 }, { "epoch": 4.242587601078167, "grad_norm": 0.658807635307312, "learning_rate": 0.00034591473286562323, "loss": 3.3939, "step": 39350 }, { "epoch": 4.247978436657682, "grad_norm": 0.6783695816993713, "learning_rate": 0.0003455909336211549, "loss": 3.3761, "step": 39400 }, { "epoch": 4.2533692722371965, "grad_norm": 0.692827045917511, "learning_rate": 0.0003452671343766864, "loss": 3.3755, "step": 39450 }, { "epoch": 4.258760107816712, "grad_norm": 0.6296021938323975, "learning_rate": 0.000344943335132218, "loss": 3.3845, "step": 39500 }, { "epoch": 4.264150943396227, "grad_norm": 0.6483592391014099, "learning_rate": 0.00034461953588774954, "loss": 3.3662, "step": 39550 }, { "epoch": 4.269541778975741, "grad_norm": 0.6927255392074585, "learning_rate": 0.00034429573664328115, "loss": 3.3778, "step": 39600 }, { "epoch": 4.274932614555256, "grad_norm": 0.6652512550354004, "learning_rate": 0.0003439719373988127, "loss": 3.3664, "step": 39650 }, { "epoch": 4.280323450134771, "grad_norm": 0.6730897426605225, "learning_rate": 0.00034364813815434425, "loss": 3.3955, "step": 39700 }, { "epoch": 4.285714285714286, "grad_norm": 0.6138418316841125, "learning_rate": 0.00034332433890987585, "loss": 3.3833, "step": 39750 }, { "epoch": 4.291105121293801, "grad_norm": 0.646796464920044, "learning_rate": 0.0003430005396654074, "loss": 3.3601, "step": 39800 }, { "epoch": 4.296495956873315, "grad_norm": 0.6964720487594604, "learning_rate": 0.000342676740420939, "loss": 3.3958, "step": 39850 }, { "epoch": 4.30188679245283, "grad_norm": 0.6758812069892883, "learning_rate": 0.00034235294117647056, "loss": 3.3742, "step": 39900 }, { "epoch": 4.307277628032345, "grad_norm": 0.6758851408958435, "learning_rate": 0.00034202914193200216, "loss": 3.3595, "step": 39950 }, { "epoch": 4.3126684636118595, "grad_norm": 0.5890964269638062, "learning_rate": 0.0003417053426875337, "loss": 3.3936, "step": 40000 }, { "epoch": 4.3126684636118595, "eval_accuracy": 0.37785552012626333, "eval_loss": 3.4354093074798584, "eval_runtime": 184.1833, "eval_samples_per_second": 97.788, "eval_steps_per_second": 6.113, "step": 40000 }, { "epoch": 4.318059299191375, "grad_norm": 0.6625378131866455, "learning_rate": 0.0003413815434430653, "loss": 3.3808, "step": 40050 }, { "epoch": 4.32345013477089, "grad_norm": 0.6554327011108398, "learning_rate": 0.0003410577441985968, "loss": 3.3846, "step": 40100 }, { "epoch": 4.328840970350404, "grad_norm": 0.6583057045936584, "learning_rate": 0.00034073394495412837, "loss": 3.3761, "step": 40150 }, { "epoch": 4.334231805929919, "grad_norm": 0.6082165837287903, "learning_rate": 0.00034041014570965997, "loss": 3.3949, "step": 40200 }, { "epoch": 4.339622641509434, "grad_norm": 0.6874740719795227, "learning_rate": 0.0003400863464651915, "loss": 3.3726, "step": 40250 }, { "epoch": 4.345013477088949, "grad_norm": 0.6384726762771606, "learning_rate": 0.0003397625472207231, "loss": 3.3946, "step": 40300 }, { "epoch": 4.350404312668464, "grad_norm": 0.6966606378555298, "learning_rate": 0.0003394387479762547, "loss": 3.3995, "step": 40350 }, { "epoch": 4.355795148247978, "grad_norm": 0.6895248293876648, "learning_rate": 0.0003391149487317863, "loss": 3.3819, "step": 40400 }, { "epoch": 4.361185983827493, "grad_norm": 0.6699813604354858, "learning_rate": 0.00033879114948731783, "loss": 3.3903, "step": 40450 }, { "epoch": 4.366576819407008, "grad_norm": 0.6132922768592834, "learning_rate": 0.00033846735024284944, "loss": 3.3853, "step": 40500 }, { "epoch": 4.3719676549865225, "grad_norm": 0.6432852745056152, "learning_rate": 0.000338143550998381, "loss": 3.3811, "step": 40550 }, { "epoch": 4.377358490566038, "grad_norm": 0.6538777351379395, "learning_rate": 0.00033781975175391254, "loss": 3.376, "step": 40600 }, { "epoch": 4.382749326145553, "grad_norm": 0.7164791822433472, "learning_rate": 0.00033749595250944414, "loss": 3.3981, "step": 40650 }, { "epoch": 4.388140161725067, "grad_norm": 0.6560450792312622, "learning_rate": 0.00033717215326497564, "loss": 3.3787, "step": 40700 }, { "epoch": 4.393530997304582, "grad_norm": 0.687532901763916, "learning_rate": 0.0003368483540205073, "loss": 3.3732, "step": 40750 }, { "epoch": 4.398921832884097, "grad_norm": 0.7162414789199829, "learning_rate": 0.0003365245547760388, "loss": 3.3951, "step": 40800 }, { "epoch": 4.404312668463612, "grad_norm": 0.643511950969696, "learning_rate": 0.0003362007555315704, "loss": 3.4023, "step": 40850 }, { "epoch": 4.409703504043127, "grad_norm": 0.6941072940826416, "learning_rate": 0.00033587695628710195, "loss": 3.3932, "step": 40900 }, { "epoch": 4.415094339622642, "grad_norm": 0.5936976075172424, "learning_rate": 0.00033555315704263355, "loss": 3.3823, "step": 40950 }, { "epoch": 4.420485175202156, "grad_norm": 0.6859274506568909, "learning_rate": 0.0003352293577981651, "loss": 3.3884, "step": 41000 }, { "epoch": 4.420485175202156, "eval_accuracy": 0.3781056390740123, "eval_loss": 3.4318952560424805, "eval_runtime": 183.8135, "eval_samples_per_second": 97.985, "eval_steps_per_second": 6.126, "step": 41000 }, { "epoch": 4.425876010781671, "grad_norm": 0.6612915992736816, "learning_rate": 0.00033490555855369665, "loss": 3.3839, "step": 41050 }, { "epoch": 4.431266846361186, "grad_norm": 0.6748125553131104, "learning_rate": 0.00033458175930922826, "loss": 3.3767, "step": 41100 }, { "epoch": 4.436657681940701, "grad_norm": 0.6383491158485413, "learning_rate": 0.0003342579600647598, "loss": 3.3884, "step": 41150 }, { "epoch": 4.442048517520216, "grad_norm": 0.672412633895874, "learning_rate": 0.0003339341608202914, "loss": 3.3683, "step": 41200 }, { "epoch": 4.44743935309973, "grad_norm": 0.6618825197219849, "learning_rate": 0.00033361036157582297, "loss": 3.3797, "step": 41250 }, { "epoch": 4.452830188679245, "grad_norm": 0.6703552603721619, "learning_rate": 0.00033328656233135457, "loss": 3.3878, "step": 41300 }, { "epoch": 4.45822102425876, "grad_norm": 0.677817165851593, "learning_rate": 0.0003329627630868861, "loss": 3.4008, "step": 41350 }, { "epoch": 4.463611859838275, "grad_norm": 0.6925399303436279, "learning_rate": 0.0003326389638424177, "loss": 3.3772, "step": 41400 }, { "epoch": 4.46900269541779, "grad_norm": 0.6213698387145996, "learning_rate": 0.0003323151645979492, "loss": 3.3777, "step": 41450 }, { "epoch": 4.474393530997305, "grad_norm": 0.6628020405769348, "learning_rate": 0.00033199136535348077, "loss": 3.4046, "step": 41500 }, { "epoch": 4.479784366576819, "grad_norm": 0.6591078639030457, "learning_rate": 0.0003316675661090124, "loss": 3.4125, "step": 41550 }, { "epoch": 4.485175202156334, "grad_norm": 0.6607985496520996, "learning_rate": 0.00033134376686454393, "loss": 3.3717, "step": 41600 }, { "epoch": 4.490566037735849, "grad_norm": 0.6439804434776306, "learning_rate": 0.00033101996762007553, "loss": 3.3877, "step": 41650 }, { "epoch": 4.495956873315364, "grad_norm": 0.6608968377113342, "learning_rate": 0.0003306961683756071, "loss": 3.3859, "step": 41700 }, { "epoch": 4.501347708894879, "grad_norm": 0.6322570443153381, "learning_rate": 0.0003303723691311387, "loss": 3.4031, "step": 41750 }, { "epoch": 4.506738544474393, "grad_norm": 0.6550046801567078, "learning_rate": 0.00033004856988667024, "loss": 3.4043, "step": 41800 }, { "epoch": 4.512129380053908, "grad_norm": 0.6358398795127869, "learning_rate": 0.0003297247706422018, "loss": 3.3698, "step": 41850 }, { "epoch": 4.517520215633423, "grad_norm": 0.6461077332496643, "learning_rate": 0.0003294009713977334, "loss": 3.3978, "step": 41900 }, { "epoch": 4.5229110512129385, "grad_norm": 0.6892066597938538, "learning_rate": 0.00032907717215326494, "loss": 3.362, "step": 41950 }, { "epoch": 4.528301886792453, "grad_norm": 0.6454676389694214, "learning_rate": 0.00032875337290879655, "loss": 3.3762, "step": 42000 }, { "epoch": 4.528301886792453, "eval_accuracy": 0.3790023513571452, "eval_loss": 3.4271507263183594, "eval_runtime": 184.2053, "eval_samples_per_second": 97.777, "eval_steps_per_second": 6.113, "step": 42000 }, { "epoch": 4.533692722371968, "grad_norm": 0.6774193048477173, "learning_rate": 0.00032842957366432805, "loss": 3.3983, "step": 42050 }, { "epoch": 4.539083557951482, "grad_norm": 0.6757726669311523, "learning_rate": 0.0003281057744198597, "loss": 3.3727, "step": 42100 }, { "epoch": 4.544474393530997, "grad_norm": 0.6542174816131592, "learning_rate": 0.0003277819751753912, "loss": 3.3891, "step": 42150 }, { "epoch": 4.549865229110512, "grad_norm": 0.6432123184204102, "learning_rate": 0.0003274581759309228, "loss": 3.385, "step": 42200 }, { "epoch": 4.555256064690027, "grad_norm": 0.7210422158241272, "learning_rate": 0.00032713437668645436, "loss": 3.411, "step": 42250 }, { "epoch": 4.560646900269542, "grad_norm": 0.773587167263031, "learning_rate": 0.0003268105774419859, "loss": 3.3781, "step": 42300 }, { "epoch": 4.566037735849057, "grad_norm": 0.6960445642471313, "learning_rate": 0.0003264867781975175, "loss": 3.3946, "step": 42350 }, { "epoch": 4.571428571428571, "grad_norm": 0.6441209316253662, "learning_rate": 0.00032616297895304906, "loss": 3.3652, "step": 42400 }, { "epoch": 4.576819407008086, "grad_norm": 0.6879327893257141, "learning_rate": 0.00032584565569347, "loss": 3.3817, "step": 42450 }, { "epoch": 4.5822102425876015, "grad_norm": 0.6669209003448486, "learning_rate": 0.0003255218564490016, "loss": 3.3879, "step": 42500 }, { "epoch": 4.587601078167116, "grad_norm": 0.6996499300003052, "learning_rate": 0.00032519805720453317, "loss": 3.3882, "step": 42550 }, { "epoch": 4.592991913746631, "grad_norm": 0.6694384217262268, "learning_rate": 0.0003248742579600647, "loss": 3.3783, "step": 42600 }, { "epoch": 4.598382749326145, "grad_norm": 0.6417438387870789, "learning_rate": 0.0003245504587155963, "loss": 3.3931, "step": 42650 }, { "epoch": 4.60377358490566, "grad_norm": 0.7074453234672546, "learning_rate": 0.0003242266594711278, "loss": 3.3937, "step": 42700 }, { "epoch": 4.609164420485175, "grad_norm": 0.6900057196617126, "learning_rate": 0.0003239028602266595, "loss": 3.3728, "step": 42750 }, { "epoch": 4.6145552560646905, "grad_norm": 0.6593276262283325, "learning_rate": 0.000323579060982191, "loss": 3.404, "step": 42800 }, { "epoch": 4.619946091644205, "grad_norm": 0.6587552428245544, "learning_rate": 0.0003232552617377226, "loss": 3.3996, "step": 42850 }, { "epoch": 4.62533692722372, "grad_norm": 0.6677352786064148, "learning_rate": 0.00032293146249325413, "loss": 3.3965, "step": 42900 }, { "epoch": 4.630727762803234, "grad_norm": 0.6786492466926575, "learning_rate": 0.00032260766324878574, "loss": 3.3921, "step": 42950 }, { "epoch": 4.636118598382749, "grad_norm": 0.6595161557197571, "learning_rate": 0.0003222838640043173, "loss": 3.375, "step": 43000 }, { "epoch": 4.636118598382749, "eval_accuracy": 0.3788302451828471, "eval_loss": 3.422513008117676, "eval_runtime": 183.7825, "eval_samples_per_second": 98.002, "eval_steps_per_second": 6.127, "step": 43000 }, { "epoch": 4.6415094339622645, "grad_norm": 0.6542896628379822, "learning_rate": 0.00032196006475984884, "loss": 3.397, "step": 43050 }, { "epoch": 4.646900269541779, "grad_norm": 0.6302821040153503, "learning_rate": 0.00032163626551538044, "loss": 3.3931, "step": 43100 }, { "epoch": 4.652291105121294, "grad_norm": 0.6723839640617371, "learning_rate": 0.0003213189422558014, "loss": 3.3764, "step": 43150 }, { "epoch": 4.657681940700809, "grad_norm": 0.7077513933181763, "learning_rate": 0.00032099514301133295, "loss": 3.3969, "step": 43200 }, { "epoch": 4.663072776280323, "grad_norm": 0.6480196118354797, "learning_rate": 0.00032067134376686455, "loss": 3.4047, "step": 43250 }, { "epoch": 4.668463611859838, "grad_norm": 0.6761808395385742, "learning_rate": 0.0003203475445223961, "loss": 3.3932, "step": 43300 }, { "epoch": 4.6738544474393535, "grad_norm": 0.6225863099098206, "learning_rate": 0.0003200237452779277, "loss": 3.3894, "step": 43350 }, { "epoch": 4.679245283018868, "grad_norm": 0.6918447613716125, "learning_rate": 0.00031969994603345926, "loss": 3.3835, "step": 43400 }, { "epoch": 4.684636118598383, "grad_norm": 0.6438918709754944, "learning_rate": 0.00031937614678899075, "loss": 3.3819, "step": 43450 }, { "epoch": 4.690026954177897, "grad_norm": 0.6605272889137268, "learning_rate": 0.00031905234754452236, "loss": 3.3883, "step": 43500 }, { "epoch": 4.695417789757412, "grad_norm": 0.6213443875312805, "learning_rate": 0.0003187285483000539, "loss": 3.3871, "step": 43550 }, { "epoch": 4.7008086253369274, "grad_norm": 0.6892732977867126, "learning_rate": 0.0003184047490555855, "loss": 3.375, "step": 43600 }, { "epoch": 4.706199460916442, "grad_norm": 0.6602584719657898, "learning_rate": 0.00031808094981111706, "loss": 3.3888, "step": 43650 }, { "epoch": 4.711590296495957, "grad_norm": 0.6708425283432007, "learning_rate": 0.00031775715056664867, "loss": 3.3869, "step": 43700 }, { "epoch": 4.716981132075472, "grad_norm": 0.6698873043060303, "learning_rate": 0.0003174333513221802, "loss": 3.373, "step": 43750 }, { "epoch": 4.722371967654986, "grad_norm": 0.6338597536087036, "learning_rate": 0.00031710955207771177, "loss": 3.3879, "step": 43800 }, { "epoch": 4.727762803234501, "grad_norm": 0.7395240664482117, "learning_rate": 0.0003167857528332434, "loss": 3.4032, "step": 43850 }, { "epoch": 4.7331536388140165, "grad_norm": 0.6551410555839539, "learning_rate": 0.0003164619535887749, "loss": 3.394, "step": 43900 }, { "epoch": 4.738544474393531, "grad_norm": 0.7224283814430237, "learning_rate": 0.00031613815434430653, "loss": 3.3859, "step": 43950 }, { "epoch": 4.743935309973046, "grad_norm": 0.6312947273254395, "learning_rate": 0.0003158143550998381, "loss": 3.3747, "step": 44000 }, { "epoch": 4.743935309973046, "eval_accuracy": 0.38003672684913903, "eval_loss": 3.4164884090423584, "eval_runtime": 184.2338, "eval_samples_per_second": 97.762, "eval_steps_per_second": 6.112, "step": 44000 }, { "epoch": 4.74932614555256, "grad_norm": 0.6509761214256287, "learning_rate": 0.0003154905558553697, "loss": 3.3905, "step": 44050 }, { "epoch": 4.754716981132075, "grad_norm": 0.6715309023857117, "learning_rate": 0.0003151667566109012, "loss": 3.3842, "step": 44100 }, { "epoch": 4.7601078167115904, "grad_norm": 0.7263250350952148, "learning_rate": 0.00031484295736643284, "loss": 3.3871, "step": 44150 }, { "epoch": 4.765498652291106, "grad_norm": 0.6736378073692322, "learning_rate": 0.00031451915812196434, "loss": 3.3561, "step": 44200 }, { "epoch": 4.77088948787062, "grad_norm": 0.7403033971786499, "learning_rate": 0.0003141953588774959, "loss": 3.3938, "step": 44250 }, { "epoch": 4.776280323450135, "grad_norm": 0.7216389179229736, "learning_rate": 0.0003138715596330275, "loss": 3.3794, "step": 44300 }, { "epoch": 4.781671159029649, "grad_norm": 0.72726970911026, "learning_rate": 0.00031354776038855904, "loss": 3.3543, "step": 44350 }, { "epoch": 4.787061994609164, "grad_norm": 0.6150906085968018, "learning_rate": 0.00031322396114409065, "loss": 3.3835, "step": 44400 }, { "epoch": 4.7924528301886795, "grad_norm": 0.6653279066085815, "learning_rate": 0.0003129066378845116, "loss": 3.3903, "step": 44450 }, { "epoch": 4.797843665768194, "grad_norm": 0.6354139447212219, "learning_rate": 0.00031258283864004315, "loss": 3.3956, "step": 44500 }, { "epoch": 4.803234501347709, "grad_norm": 0.6651866436004639, "learning_rate": 0.0003122590393955747, "loss": 3.4122, "step": 44550 }, { "epoch": 4.808625336927224, "grad_norm": 0.6413260698318481, "learning_rate": 0.0003119352401511063, "loss": 3.3788, "step": 44600 }, { "epoch": 4.814016172506738, "grad_norm": 0.7094629406929016, "learning_rate": 0.00031161144090663786, "loss": 3.3893, "step": 44650 }, { "epoch": 4.819407008086253, "grad_norm": 0.6571715474128723, "learning_rate": 0.00031128764166216946, "loss": 3.3826, "step": 44700 }, { "epoch": 4.824797843665769, "grad_norm": 0.6378207206726074, "learning_rate": 0.00031096384241770096, "loss": 3.3812, "step": 44750 }, { "epoch": 4.830188679245283, "grad_norm": 0.6692754030227661, "learning_rate": 0.0003106400431732326, "loss": 3.3762, "step": 44800 }, { "epoch": 4.835579514824798, "grad_norm": 0.6558183431625366, "learning_rate": 0.0003103162439287641, "loss": 3.3784, "step": 44850 }, { "epoch": 4.840970350404312, "grad_norm": 0.6741693615913391, "learning_rate": 0.0003099924446842957, "loss": 3.3902, "step": 44900 }, { "epoch": 4.846361185983827, "grad_norm": 0.6589843034744263, "learning_rate": 0.00030966864543982727, "loss": 3.4018, "step": 44950 }, { "epoch": 4.8517520215633425, "grad_norm": 0.6733647584915161, "learning_rate": 0.0003093448461953588, "loss": 3.3985, "step": 45000 }, { "epoch": 4.8517520215633425, "eval_accuracy": 0.38011973765542934, "eval_loss": 3.4101078510284424, "eval_runtime": 183.7639, "eval_samples_per_second": 98.012, "eval_steps_per_second": 6.127, "step": 45000 }, { "epoch": 4.857142857142857, "grad_norm": 0.679538905620575, "learning_rate": 0.0003090210469508904, "loss": 3.3971, "step": 45050 }, { "epoch": 4.862533692722372, "grad_norm": 0.7209023237228394, "learning_rate": 0.00030869724770642197, "loss": 3.391, "step": 45100 }, { "epoch": 4.867924528301887, "grad_norm": 0.6559917330741882, "learning_rate": 0.0003083734484619536, "loss": 3.4099, "step": 45150 }, { "epoch": 4.873315363881401, "grad_norm": 0.6977805495262146, "learning_rate": 0.00030804964921748513, "loss": 3.3935, "step": 45200 }, { "epoch": 4.878706199460916, "grad_norm": 0.7327739000320435, "learning_rate": 0.00030772584997301673, "loss": 3.3777, "step": 45250 }, { "epoch": 4.884097035040432, "grad_norm": 0.6754328608512878, "learning_rate": 0.0003074020507285483, "loss": 3.3802, "step": 45300 }, { "epoch": 4.889487870619946, "grad_norm": 0.6827671527862549, "learning_rate": 0.0003070782514840799, "loss": 3.3702, "step": 45350 }, { "epoch": 4.894878706199461, "grad_norm": 0.6643690466880798, "learning_rate": 0.00030675445223961144, "loss": 3.3821, "step": 45400 }, { "epoch": 4.900269541778976, "grad_norm": 0.707639217376709, "learning_rate": 0.00030643065299514294, "loss": 3.3884, "step": 45450 }, { "epoch": 4.90566037735849, "grad_norm": 0.7274863123893738, "learning_rate": 0.00030610685375067454, "loss": 3.3776, "step": 45500 }, { "epoch": 4.9110512129380055, "grad_norm": 0.6537758708000183, "learning_rate": 0.0003057830545062061, "loss": 3.3664, "step": 45550 }, { "epoch": 4.916442048517521, "grad_norm": 0.6469917297363281, "learning_rate": 0.0003054592552617377, "loss": 3.3919, "step": 45600 }, { "epoch": 4.921832884097035, "grad_norm": 0.6994689702987671, "learning_rate": 0.00030513545601726925, "loss": 3.3893, "step": 45650 }, { "epoch": 4.92722371967655, "grad_norm": 0.667890727519989, "learning_rate": 0.00030481165677280085, "loss": 3.3883, "step": 45700 }, { "epoch": 4.932614555256064, "grad_norm": 0.6717064380645752, "learning_rate": 0.0003044878575283324, "loss": 3.3864, "step": 45750 }, { "epoch": 4.938005390835579, "grad_norm": 0.65635085105896, "learning_rate": 0.000304164058283864, "loss": 3.3935, "step": 45800 }, { "epoch": 4.943396226415095, "grad_norm": 0.6819509863853455, "learning_rate": 0.00030384025903939556, "loss": 3.3761, "step": 45850 }, { "epoch": 4.948787061994609, "grad_norm": 0.6925830245018005, "learning_rate": 0.0003035164597949271, "loss": 3.3798, "step": 45900 }, { "epoch": 4.954177897574124, "grad_norm": 0.6927830576896667, "learning_rate": 0.0003031926605504587, "loss": 3.3879, "step": 45950 }, { "epoch": 4.959568733153639, "grad_norm": 0.637687623500824, "learning_rate": 0.00030286886130599026, "loss": 3.3853, "step": 46000 }, { "epoch": 4.959568733153639, "eval_accuracy": 0.38066202421851136, "eval_loss": 3.406095504760742, "eval_runtime": 184.1141, "eval_samples_per_second": 97.825, "eval_steps_per_second": 6.116, "step": 46000 }, { "epoch": 4.964959568733153, "grad_norm": 0.6471781134605408, "learning_rate": 0.00030254506206152187, "loss": 3.375, "step": 46050 }, { "epoch": 4.9703504043126685, "grad_norm": 0.6780849099159241, "learning_rate": 0.00030222126281705336, "loss": 3.3776, "step": 46100 }, { "epoch": 4.975741239892184, "grad_norm": 1.019982099533081, "learning_rate": 0.000301897463572585, "loss": 3.3775, "step": 46150 }, { "epoch": 4.981132075471698, "grad_norm": 0.6539048552513123, "learning_rate": 0.0003015736643281165, "loss": 3.3683, "step": 46200 }, { "epoch": 4.986522911051213, "grad_norm": 0.6219794750213623, "learning_rate": 0.00030124986508364807, "loss": 3.3791, "step": 46250 }, { "epoch": 4.991913746630727, "grad_norm": 0.6488443613052368, "learning_rate": 0.0003009260658391797, "loss": 3.3952, "step": 46300 }, { "epoch": 4.997304582210242, "grad_norm": 0.6938602328300476, "learning_rate": 0.0003006022665947112, "loss": 3.3779, "step": 46350 }, { "epoch": 5.002695417789758, "grad_norm": 0.6452742218971252, "learning_rate": 0.00030027846735024283, "loss": 3.3349, "step": 46400 }, { "epoch": 5.008086253369272, "grad_norm": 0.6376250982284546, "learning_rate": 0.00029995466810577443, "loss": 3.2813, "step": 46450 }, { "epoch": 5.013477088948787, "grad_norm": 0.7294584512710571, "learning_rate": 0.00029963086886130593, "loss": 3.2785, "step": 46500 }, { "epoch": 5.018867924528302, "grad_norm": 0.661499559879303, "learning_rate": 0.00029930706961683754, "loss": 3.2773, "step": 46550 }, { "epoch": 5.024258760107816, "grad_norm": 0.6674730181694031, "learning_rate": 0.0002989832703723691, "loss": 3.2923, "step": 46600 }, { "epoch": 5.0296495956873315, "grad_norm": 0.678300142288208, "learning_rate": 0.0002986594711279007, "loss": 3.2873, "step": 46650 }, { "epoch": 5.035040431266847, "grad_norm": 0.7316428422927856, "learning_rate": 0.00029833567188343224, "loss": 3.3123, "step": 46700 }, { "epoch": 5.040431266846361, "grad_norm": 0.6575501561164856, "learning_rate": 0.00029801187263896385, "loss": 3.2837, "step": 46750 }, { "epoch": 5.045822102425876, "grad_norm": 0.7176420092582703, "learning_rate": 0.0002976880733944954, "loss": 3.2904, "step": 46800 }, { "epoch": 5.051212938005391, "grad_norm": 0.7104209661483765, "learning_rate": 0.00029736427415002695, "loss": 3.3006, "step": 46850 }, { "epoch": 5.056603773584905, "grad_norm": 0.7049729228019714, "learning_rate": 0.0002970404749055585, "loss": 3.3261, "step": 46900 }, { "epoch": 5.061994609164421, "grad_norm": 0.622776985168457, "learning_rate": 0.0002967166756610901, "loss": 3.2972, "step": 46950 }, { "epoch": 5.067385444743936, "grad_norm": 0.697993814945221, "learning_rate": 0.00029639287641662165, "loss": 3.2829, "step": 47000 }, { "epoch": 5.067385444743936, "eval_accuracy": 0.38110869624031674, "eval_loss": 3.4087939262390137, "eval_runtime": 183.8604, "eval_samples_per_second": 97.96, "eval_steps_per_second": 6.124, "step": 47000 }, { "epoch": 5.07277628032345, "grad_norm": 0.6486660242080688, "learning_rate": 0.00029606907717215326, "loss": 3.3014, "step": 47050 }, { "epoch": 5.078167115902965, "grad_norm": 0.69781494140625, "learning_rate": 0.0002957452779276848, "loss": 3.3001, "step": 47100 }, { "epoch": 5.083557951482479, "grad_norm": 0.6906808614730835, "learning_rate": 0.00029542147868321636, "loss": 3.3136, "step": 47150 }, { "epoch": 5.0889487870619945, "grad_norm": 0.6491060256958008, "learning_rate": 0.00029509767943874796, "loss": 3.3082, "step": 47200 }, { "epoch": 5.09433962264151, "grad_norm": 0.6999182105064392, "learning_rate": 0.0002947738801942795, "loss": 3.3151, "step": 47250 }, { "epoch": 5.099730458221024, "grad_norm": 0.6435338854789734, "learning_rate": 0.0002944500809498111, "loss": 3.3212, "step": 47300 }, { "epoch": 5.105121293800539, "grad_norm": 0.6768415570259094, "learning_rate": 0.00029412628170534267, "loss": 3.2983, "step": 47350 }, { "epoch": 5.110512129380054, "grad_norm": 0.6941831111907959, "learning_rate": 0.0002938024824608742, "loss": 3.3183, "step": 47400 }, { "epoch": 5.115902964959568, "grad_norm": 0.6833462715148926, "learning_rate": 0.00029347868321640577, "loss": 3.3287, "step": 47450 }, { "epoch": 5.121293800539084, "grad_norm": 0.6526753902435303, "learning_rate": 0.0002931548839719374, "loss": 3.3072, "step": 47500 }, { "epoch": 5.126684636118599, "grad_norm": 0.6595203876495361, "learning_rate": 0.0002928310847274689, "loss": 3.3137, "step": 47550 }, { "epoch": 5.132075471698113, "grad_norm": 0.689451277256012, "learning_rate": 0.00029250728548300053, "loss": 3.3211, "step": 47600 }, { "epoch": 5.137466307277628, "grad_norm": 0.6616727709770203, "learning_rate": 0.0002921834862385321, "loss": 3.3066, "step": 47650 }, { "epoch": 5.142857142857143, "grad_norm": 0.6385394930839539, "learning_rate": 0.0002918596869940637, "loss": 3.3176, "step": 47700 }, { "epoch": 5.1482479784366575, "grad_norm": 0.6855962872505188, "learning_rate": 0.0002915358877495952, "loss": 3.3303, "step": 47750 }, { "epoch": 5.153638814016173, "grad_norm": 0.6520223617553711, "learning_rate": 0.0002912120885051268, "loss": 3.319, "step": 47800 }, { "epoch": 5.159029649595688, "grad_norm": 0.7033669948577881, "learning_rate": 0.00029088828926065834, "loss": 3.3262, "step": 47850 }, { "epoch": 5.164420485175202, "grad_norm": 0.7349123358726501, "learning_rate": 0.00029056449001618994, "loss": 3.3109, "step": 47900 }, { "epoch": 5.169811320754717, "grad_norm": 0.7183597683906555, "learning_rate": 0.0002902406907717215, "loss": 3.3172, "step": 47950 }, { "epoch": 5.175202156334231, "grad_norm": 0.690875768661499, "learning_rate": 0.0002899168915272531, "loss": 3.3263, "step": 48000 }, { "epoch": 5.175202156334231, "eval_accuracy": 0.38107121099402075, "eval_loss": 3.410407781600952, "eval_runtime": 184.0011, "eval_samples_per_second": 97.885, "eval_steps_per_second": 6.12, "step": 48000 }, { "epoch": 5.180592991913747, "grad_norm": 0.6746882796287537, "learning_rate": 0.00028959309228278465, "loss": 3.3005, "step": 48050 }, { "epoch": 5.185983827493262, "grad_norm": 0.6469348669052124, "learning_rate": 0.00028926929303831625, "loss": 3.3094, "step": 48100 }, { "epoch": 5.191374663072776, "grad_norm": 0.683721125125885, "learning_rate": 0.0002889454937938478, "loss": 3.318, "step": 48150 }, { "epoch": 5.196765498652291, "grad_norm": 0.782717227935791, "learning_rate": 0.00028862169454937935, "loss": 3.3255, "step": 48200 }, { "epoch": 5.202156334231806, "grad_norm": 0.7064043283462524, "learning_rate": 0.0002882978953049109, "loss": 3.3129, "step": 48250 }, { "epoch": 5.2075471698113205, "grad_norm": 0.7118088006973267, "learning_rate": 0.0002879740960604425, "loss": 3.3106, "step": 48300 }, { "epoch": 5.212938005390836, "grad_norm": 0.6886478066444397, "learning_rate": 0.00028765029681597406, "loss": 3.322, "step": 48350 }, { "epoch": 5.218328840970351, "grad_norm": 0.6945542097091675, "learning_rate": 0.00028732649757150566, "loss": 3.3282, "step": 48400 }, { "epoch": 5.223719676549865, "grad_norm": 0.7024538516998291, "learning_rate": 0.0002870026983270372, "loss": 3.2974, "step": 48450 }, { "epoch": 5.22911051212938, "grad_norm": 0.680157482624054, "learning_rate": 0.00028667889908256877, "loss": 3.3084, "step": 48500 }, { "epoch": 5.234501347708895, "grad_norm": 0.7121500968933105, "learning_rate": 0.00028635509983810037, "loss": 3.3226, "step": 48550 }, { "epoch": 5.2398921832884096, "grad_norm": 0.7295412421226501, "learning_rate": 0.0002860313005936319, "loss": 3.3359, "step": 48600 }, { "epoch": 5.245283018867925, "grad_norm": 0.6911020874977112, "learning_rate": 0.00028570750134916347, "loss": 3.3272, "step": 48650 }, { "epoch": 5.250673854447439, "grad_norm": 0.7227579355239868, "learning_rate": 0.0002853837021046951, "loss": 3.3175, "step": 48700 }, { "epoch": 5.256064690026954, "grad_norm": 0.64582359790802, "learning_rate": 0.00028505990286022663, "loss": 3.317, "step": 48750 }, { "epoch": 5.261455525606469, "grad_norm": 0.7002373337745667, "learning_rate": 0.0002847361036157582, "loss": 3.3165, "step": 48800 }, { "epoch": 5.2668463611859835, "grad_norm": 0.7050011157989502, "learning_rate": 0.0002844123043712898, "loss": 3.3082, "step": 48850 }, { "epoch": 5.272237196765499, "grad_norm": 0.6821650266647339, "learning_rate": 0.00028408850512682133, "loss": 3.3348, "step": 48900 }, { "epoch": 5.277628032345014, "grad_norm": 0.6783719658851624, "learning_rate": 0.00028376470588235294, "loss": 3.3126, "step": 48950 }, { "epoch": 5.283018867924528, "grad_norm": 0.729580283164978, "learning_rate": 0.0002834409066378845, "loss": 3.3098, "step": 49000 }, { "epoch": 5.283018867924528, "eval_accuracy": 0.3817646337240527, "eval_loss": 3.4049880504608154, "eval_runtime": 183.9025, "eval_samples_per_second": 97.938, "eval_steps_per_second": 6.123, "step": 49000 }, { "epoch": 5.288409703504043, "grad_norm": 0.7196259498596191, "learning_rate": 0.00028311710739341604, "loss": 3.3177, "step": 49050 }, { "epoch": 5.293800539083558, "grad_norm": 0.7036062479019165, "learning_rate": 0.00028279330814894764, "loss": 3.3319, "step": 49100 }, { "epoch": 5.2991913746630726, "grad_norm": 0.7958604097366333, "learning_rate": 0.0002824695089044792, "loss": 3.3291, "step": 49150 }, { "epoch": 5.304582210242588, "grad_norm": 0.6777556538581848, "learning_rate": 0.00028214570966001075, "loss": 3.3213, "step": 49200 }, { "epoch": 5.309973045822103, "grad_norm": 0.7265875935554504, "learning_rate": 0.00028182191041554235, "loss": 3.3201, "step": 49250 }, { "epoch": 5.315363881401617, "grad_norm": 0.6683766841888428, "learning_rate": 0.0002814981111710739, "loss": 3.3334, "step": 49300 }, { "epoch": 5.320754716981132, "grad_norm": 0.6723276972770691, "learning_rate": 0.0002811743119266055, "loss": 3.3208, "step": 49350 }, { "epoch": 5.3261455525606465, "grad_norm": 0.674947202205658, "learning_rate": 0.00028085051268213706, "loss": 3.3206, "step": 49400 }, { "epoch": 5.331536388140162, "grad_norm": 0.6870039701461792, "learning_rate": 0.0002805267134376686, "loss": 3.3373, "step": 49450 }, { "epoch": 5.336927223719677, "grad_norm": 0.6902183890342712, "learning_rate": 0.00028020291419320016, "loss": 3.3207, "step": 49500 }, { "epoch": 5.342318059299191, "grad_norm": 0.7033862471580505, "learning_rate": 0.00027987911494873176, "loss": 3.3258, "step": 49550 }, { "epoch": 5.347708894878706, "grad_norm": 0.673235297203064, "learning_rate": 0.0002795553157042633, "loss": 3.3198, "step": 49600 }, { "epoch": 5.353099730458221, "grad_norm": 0.7145218849182129, "learning_rate": 0.0002792315164597949, "loss": 3.3315, "step": 49650 }, { "epoch": 5.3584905660377355, "grad_norm": 0.6835147738456726, "learning_rate": 0.00027891419320021587, "loss": 3.3322, "step": 49700 }, { "epoch": 5.363881401617251, "grad_norm": 0.6946256160736084, "learning_rate": 0.0002785903939557474, "loss": 3.3398, "step": 49750 }, { "epoch": 5.369272237196766, "grad_norm": 0.7527667284011841, "learning_rate": 0.00027826659471127897, "loss": 3.3284, "step": 49800 }, { "epoch": 5.37466307277628, "grad_norm": 0.7169890999794006, "learning_rate": 0.0002779427954668105, "loss": 3.3206, "step": 49850 }, { "epoch": 5.380053908355795, "grad_norm": 0.6716258525848389, "learning_rate": 0.0002776189962223421, "loss": 3.3221, "step": 49900 }, { "epoch": 5.38544474393531, "grad_norm": 0.7162343263626099, "learning_rate": 0.0002772951969778737, "loss": 3.3214, "step": 49950 }, { "epoch": 5.390835579514825, "grad_norm": 0.7313789129257202, "learning_rate": 0.0002769713977334053, "loss": 3.348, "step": 50000 }, { "epoch": 5.390835579514825, "eval_accuracy": 0.3818524252574068, "eval_loss": 3.4028215408325195, "eval_runtime": 183.8802, "eval_samples_per_second": 97.95, "eval_steps_per_second": 6.124, "step": 50000 }, { "epoch": 5.39622641509434, "grad_norm": 0.6996383666992188, "learning_rate": 0.00027664759848893683, "loss": 3.3008, "step": 50050 }, { "epoch": 5.401617250673855, "grad_norm": 0.6793367862701416, "learning_rate": 0.00027632379924446844, "loss": 3.3247, "step": 50100 }, { "epoch": 5.407008086253369, "grad_norm": 0.6811390519142151, "learning_rate": 0.000276, "loss": 3.3223, "step": 50150 }, { "epoch": 5.412398921832884, "grad_norm": 0.7058961391448975, "learning_rate": 0.00027567620075553154, "loss": 3.3379, "step": 50200 }, { "epoch": 5.4177897574123985, "grad_norm": 0.7083661556243896, "learning_rate": 0.0002753524015110631, "loss": 3.3497, "step": 50250 }, { "epoch": 5.423180592991914, "grad_norm": 0.6529145836830139, "learning_rate": 0.0002750286022665947, "loss": 3.3276, "step": 50300 }, { "epoch": 5.428571428571429, "grad_norm": 0.6388300061225891, "learning_rate": 0.00027470480302212624, "loss": 3.3116, "step": 50350 }, { "epoch": 5.433962264150943, "grad_norm": 0.6704314947128296, "learning_rate": 0.00027438100377765785, "loss": 3.3319, "step": 50400 }, { "epoch": 5.439353099730458, "grad_norm": 0.7339666485786438, "learning_rate": 0.0002740572045331894, "loss": 3.3106, "step": 50450 }, { "epoch": 5.444743935309973, "grad_norm": 0.695424497127533, "learning_rate": 0.00027373340528872095, "loss": 3.3386, "step": 50500 }, { "epoch": 5.450134770889488, "grad_norm": 0.700162410736084, "learning_rate": 0.00027340960604425255, "loss": 3.3246, "step": 50550 }, { "epoch": 5.455525606469003, "grad_norm": 0.7001470923423767, "learning_rate": 0.0002730858067997841, "loss": 3.3226, "step": 50600 }, { "epoch": 5.460916442048518, "grad_norm": 0.6577569842338562, "learning_rate": 0.00027276200755531565, "loss": 3.3271, "step": 50650 }, { "epoch": 5.466307277628032, "grad_norm": 0.6836386322975159, "learning_rate": 0.00027243820831084726, "loss": 3.3389, "step": 50700 }, { "epoch": 5.471698113207547, "grad_norm": 0.7137777805328369, "learning_rate": 0.0002721144090663788, "loss": 3.327, "step": 50750 }, { "epoch": 5.4770889487870615, "grad_norm": 0.6904305219650269, "learning_rate": 0.00027179060982191036, "loss": 3.3095, "step": 50800 }, { "epoch": 5.482479784366577, "grad_norm": 0.7418445944786072, "learning_rate": 0.00027146681057744197, "loss": 3.3346, "step": 50850 }, { "epoch": 5.487870619946092, "grad_norm": 0.6904473304748535, "learning_rate": 0.0002711430113329735, "loss": 3.3331, "step": 50900 }, { "epoch": 5.493261455525606, "grad_norm": 0.7091851830482483, "learning_rate": 0.0002708192120885051, "loss": 3.3373, "step": 50950 }, { "epoch": 5.498652291105121, "grad_norm": 0.7276546359062195, "learning_rate": 0.00027049541284403667, "loss": 3.331, "step": 51000 }, { "epoch": 5.498652291105121, "eval_accuracy": 0.3823172423114772, "eval_loss": 3.39615797996521, "eval_runtime": 184.0329, "eval_samples_per_second": 97.868, "eval_steps_per_second": 6.118, "step": 51000 }, { "epoch": 5.504043126684636, "grad_norm": 0.6925030946731567, "learning_rate": 0.0002701716135995683, "loss": 3.3433, "step": 51050 }, { "epoch": 5.509433962264151, "grad_norm": 0.652752161026001, "learning_rate": 0.0002698478143550998, "loss": 3.3233, "step": 51100 }, { "epoch": 5.514824797843666, "grad_norm": 0.7031452655792236, "learning_rate": 0.0002695240151106314, "loss": 3.3256, "step": 51150 }, { "epoch": 5.520215633423181, "grad_norm": 0.7321648597717285, "learning_rate": 0.00026920021586616293, "loss": 3.307, "step": 51200 }, { "epoch": 5.525606469002695, "grad_norm": 0.6757856011390686, "learning_rate": 0.00026887641662169453, "loss": 3.3193, "step": 51250 }, { "epoch": 5.53099730458221, "grad_norm": 0.7370916604995728, "learning_rate": 0.0002685526173772261, "loss": 3.3289, "step": 51300 }, { "epoch": 5.536388140161725, "grad_norm": 0.7094478011131287, "learning_rate": 0.0002682288181327577, "loss": 3.3272, "step": 51350 }, { "epoch": 5.54177897574124, "grad_norm": 0.6774241924285889, "learning_rate": 0.00026790501888828924, "loss": 3.3227, "step": 51400 }, { "epoch": 5.547169811320755, "grad_norm": 0.7134442925453186, "learning_rate": 0.00026758121964382084, "loss": 3.3302, "step": 51450 }, { "epoch": 5.55256064690027, "grad_norm": 0.7070966958999634, "learning_rate": 0.00026725742039935234, "loss": 3.3249, "step": 51500 }, { "epoch": 5.557951482479784, "grad_norm": 0.7182289361953735, "learning_rate": 0.00026693362115488394, "loss": 3.3378, "step": 51550 }, { "epoch": 5.563342318059299, "grad_norm": 0.6908033490180969, "learning_rate": 0.0002666098219104155, "loss": 3.3393, "step": 51600 }, { "epoch": 5.568733153638814, "grad_norm": 0.7248824238777161, "learning_rate": 0.0002662860226659471, "loss": 3.3423, "step": 51650 }, { "epoch": 5.574123989218329, "grad_norm": 0.6812718510627747, "learning_rate": 0.00026596222342147865, "loss": 3.3402, "step": 51700 }, { "epoch": 5.579514824797844, "grad_norm": 0.6986221671104431, "learning_rate": 0.00026563842417701026, "loss": 3.3352, "step": 51750 }, { "epoch": 5.584905660377358, "grad_norm": 0.6742642521858215, "learning_rate": 0.0002653146249325418, "loss": 3.3482, "step": 51800 }, { "epoch": 5.590296495956873, "grad_norm": 0.667993426322937, "learning_rate": 0.00026499082568807336, "loss": 3.3278, "step": 51850 }, { "epoch": 5.595687331536388, "grad_norm": 0.739936888217926, "learning_rate": 0.00026466702644360496, "loss": 3.3371, "step": 51900 }, { "epoch": 5.601078167115903, "grad_norm": 0.6672866344451904, "learning_rate": 0.0002643432271991365, "loss": 3.3375, "step": 51950 }, { "epoch": 5.606469002695418, "grad_norm": 0.7498651742935181, "learning_rate": 0.00026401942795466806, "loss": 3.3412, "step": 52000 }, { "epoch": 5.606469002695418, "eval_accuracy": 0.3828715893451067, "eval_loss": 3.3931350708007812, "eval_runtime": 183.9258, "eval_samples_per_second": 97.925, "eval_steps_per_second": 6.122, "step": 52000 }, { "epoch": 5.611859838274933, "grad_norm": 0.7274742126464844, "learning_rate": 0.00026369562871019967, "loss": 3.3442, "step": 52050 }, { "epoch": 5.617250673854447, "grad_norm": 0.7057594656944275, "learning_rate": 0.0002633783054506206, "loss": 3.3284, "step": 52100 }, { "epoch": 5.622641509433962, "grad_norm": 0.671108067035675, "learning_rate": 0.00026305450620615217, "loss": 3.334, "step": 52150 }, { "epoch": 5.628032345013477, "grad_norm": 0.7190834283828735, "learning_rate": 0.0002627307069616837, "loss": 3.3304, "step": 52200 }, { "epoch": 5.633423180592992, "grad_norm": 0.6813502907752991, "learning_rate": 0.0002624069077172153, "loss": 3.3345, "step": 52250 }, { "epoch": 5.638814016172507, "grad_norm": 0.6833723187446594, "learning_rate": 0.0002620831084727469, "loss": 3.3245, "step": 52300 }, { "epoch": 5.644204851752022, "grad_norm": 0.7639450430870056, "learning_rate": 0.0002617593092282784, "loss": 3.3322, "step": 52350 }, { "epoch": 5.649595687331536, "grad_norm": 0.6970094442367554, "learning_rate": 0.00026143550998381003, "loss": 3.3326, "step": 52400 }, { "epoch": 5.654986522911051, "grad_norm": 0.6890552043914795, "learning_rate": 0.0002611117107393416, "loss": 3.3346, "step": 52450 }, { "epoch": 5.660377358490566, "grad_norm": 0.6932628154754639, "learning_rate": 0.00026078791149487313, "loss": 3.3241, "step": 52500 }, { "epoch": 5.665768194070081, "grad_norm": 0.683397650718689, "learning_rate": 0.00026046411225040474, "loss": 3.3241, "step": 52550 }, { "epoch": 5.671159029649596, "grad_norm": 0.6974855661392212, "learning_rate": 0.0002601403130059363, "loss": 3.3463, "step": 52600 }, { "epoch": 5.67654986522911, "grad_norm": 0.6833411455154419, "learning_rate": 0.0002598165137614679, "loss": 3.3375, "step": 52650 }, { "epoch": 5.681940700808625, "grad_norm": 0.7359110713005066, "learning_rate": 0.00025949271451699944, "loss": 3.3277, "step": 52700 }, { "epoch": 5.6873315363881405, "grad_norm": 0.7000057697296143, "learning_rate": 0.000259168915272531, "loss": 3.3307, "step": 52750 }, { "epoch": 5.692722371967655, "grad_norm": 0.7026528120040894, "learning_rate": 0.0002588451160280626, "loss": 3.3244, "step": 52800 }, { "epoch": 5.69811320754717, "grad_norm": 0.7201850414276123, "learning_rate": 0.00025852131678359415, "loss": 3.3252, "step": 52850 }, { "epoch": 5.703504043126685, "grad_norm": 0.7258884310722351, "learning_rate": 0.0002581975175391257, "loss": 3.3325, "step": 52900 }, { "epoch": 5.708894878706199, "grad_norm": 0.7302329540252686, "learning_rate": 0.0002578737182946573, "loss": 3.3446, "step": 52950 }, { "epoch": 5.714285714285714, "grad_norm": 0.7130410075187683, "learning_rate": 0.00025754991905018885, "loss": 3.3414, "step": 53000 }, { "epoch": 5.714285714285714, "eval_accuracy": 0.38328479627746515, "eval_loss": 3.3882997035980225, "eval_runtime": 183.7964, "eval_samples_per_second": 97.994, "eval_steps_per_second": 6.126, "step": 53000 }, { "epoch": 5.719676549865229, "grad_norm": 0.7381608486175537, "learning_rate": 0.00025722611980572046, "loss": 3.3485, "step": 53050 }, { "epoch": 5.725067385444744, "grad_norm": 0.6873294115066528, "learning_rate": 0.000256902320561252, "loss": 3.3458, "step": 53100 }, { "epoch": 5.730458221024259, "grad_norm": 0.683340847492218, "learning_rate": 0.00025657852131678356, "loss": 3.3562, "step": 53150 }, { "epoch": 5.735849056603773, "grad_norm": 0.7045843005180359, "learning_rate": 0.0002562547220723151, "loss": 3.3257, "step": 53200 }, { "epoch": 5.741239892183288, "grad_norm": 0.7063679695129395, "learning_rate": 0.0002559309228278467, "loss": 3.3495, "step": 53250 }, { "epoch": 5.7466307277628035, "grad_norm": 0.7272491455078125, "learning_rate": 0.00025560712358337827, "loss": 3.3404, "step": 53300 }, { "epoch": 5.752021563342318, "grad_norm": 0.7092030644416809, "learning_rate": 0.00025528332433890987, "loss": 3.3261, "step": 53350 }, { "epoch": 5.757412398921833, "grad_norm": 0.7118086218833923, "learning_rate": 0.0002549595250944414, "loss": 3.3423, "step": 53400 }, { "epoch": 5.762803234501348, "grad_norm": 0.7487428188323975, "learning_rate": 0.000254635725849973, "loss": 3.3416, "step": 53450 }, { "epoch": 5.768194070080862, "grad_norm": 0.665637731552124, "learning_rate": 0.0002543119266055046, "loss": 3.3426, "step": 53500 }, { "epoch": 5.773584905660377, "grad_norm": 0.7153809666633606, "learning_rate": 0.00025398812736103613, "loss": 3.3474, "step": 53550 }, { "epoch": 5.7789757412398925, "grad_norm": 0.7571694850921631, "learning_rate": 0.0002536643281165677, "loss": 3.3268, "step": 53600 }, { "epoch": 5.784366576819407, "grad_norm": 0.7039183974266052, "learning_rate": 0.0002533405288720993, "loss": 3.3451, "step": 53650 }, { "epoch": 5.789757412398922, "grad_norm": 0.7550934553146362, "learning_rate": 0.00025301672962763083, "loss": 3.3295, "step": 53700 }, { "epoch": 5.795148247978437, "grad_norm": 0.7558539509773254, "learning_rate": 0.00025269293038316244, "loss": 3.3295, "step": 53750 }, { "epoch": 5.800539083557951, "grad_norm": 0.6939939856529236, "learning_rate": 0.000252369131138694, "loss": 3.3275, "step": 53800 }, { "epoch": 5.8059299191374665, "grad_norm": 0.7229671478271484, "learning_rate": 0.00025204533189422554, "loss": 3.3375, "step": 53850 }, { "epoch": 5.811320754716981, "grad_norm": 0.7354257702827454, "learning_rate": 0.00025172153264975714, "loss": 3.3143, "step": 53900 }, { "epoch": 5.816711590296496, "grad_norm": 0.7292751669883728, "learning_rate": 0.0002513977334052887, "loss": 3.3298, "step": 53950 }, { "epoch": 5.822102425876011, "grad_norm": 0.7425629496574402, "learning_rate": 0.00025107393416082025, "loss": 3.3438, "step": 54000 }, { "epoch": 5.822102425876011, "eval_accuracy": 0.3835859820824869, "eval_loss": 3.382173776626587, "eval_runtime": 184.3297, "eval_samples_per_second": 97.711, "eval_steps_per_second": 6.109, "step": 54000 }, { "epoch": 5.827493261455525, "grad_norm": 0.7313019037246704, "learning_rate": 0.00025075013491635185, "loss": 3.3422, "step": 54050 }, { "epoch": 5.83288409703504, "grad_norm": 0.8147013783454895, "learning_rate": 0.0002504263356718834, "loss": 3.3338, "step": 54100 }, { "epoch": 5.8382749326145555, "grad_norm": 0.7204222083091736, "learning_rate": 0.000250102536427415, "loss": 3.3189, "step": 54150 }, { "epoch": 5.84366576819407, "grad_norm": 0.6922837495803833, "learning_rate": 0.00024977873718294656, "loss": 3.317, "step": 54200 }, { "epoch": 5.849056603773585, "grad_norm": 0.7175344228744507, "learning_rate": 0.0002494549379384781, "loss": 3.3275, "step": 54250 }, { "epoch": 5.8544474393531, "grad_norm": 0.6663742065429688, "learning_rate": 0.0002491311386940097, "loss": 3.3034, "step": 54300 }, { "epoch": 5.859838274932614, "grad_norm": 0.6701042056083679, "learning_rate": 0.00024880733944954126, "loss": 3.3596, "step": 54350 }, { "epoch": 5.8652291105121295, "grad_norm": 0.7388389706611633, "learning_rate": 0.0002484835402050728, "loss": 3.3137, "step": 54400 }, { "epoch": 5.870619946091644, "grad_norm": 0.7365710735321045, "learning_rate": 0.0002481597409606044, "loss": 3.3216, "step": 54450 }, { "epoch": 5.876010781671159, "grad_norm": 0.710827112197876, "learning_rate": 0.00024783594171613597, "loss": 3.3269, "step": 54500 }, { "epoch": 5.881401617250674, "grad_norm": 0.7242414951324463, "learning_rate": 0.0002475121424716675, "loss": 3.3325, "step": 54550 }, { "epoch": 5.886792452830189, "grad_norm": 0.7195931077003479, "learning_rate": 0.0002471883432271991, "loss": 3.3223, "step": 54600 }, { "epoch": 5.892183288409703, "grad_norm": 0.6709275841712952, "learning_rate": 0.0002468645439827307, "loss": 3.3036, "step": 54650 }, { "epoch": 5.8975741239892185, "grad_norm": 0.6818590760231018, "learning_rate": 0.0002465407447382623, "loss": 3.331, "step": 54700 }, { "epoch": 5.902964959568733, "grad_norm": 0.723003089427948, "learning_rate": 0.00024621694549379383, "loss": 3.3229, "step": 54750 }, { "epoch": 5.908355795148248, "grad_norm": 0.6944183111190796, "learning_rate": 0.00024589314624932543, "loss": 3.351, "step": 54800 }, { "epoch": 5.913746630727763, "grad_norm": 0.7226275205612183, "learning_rate": 0.00024556934700485693, "loss": 3.3272, "step": 54850 }, { "epoch": 5.919137466307277, "grad_norm": 0.6976397633552551, "learning_rate": 0.00024524554776038853, "loss": 3.3186, "step": 54900 }, { "epoch": 5.9245283018867925, "grad_norm": 0.7749685049057007, "learning_rate": 0.0002449217485159201, "loss": 3.3312, "step": 54950 }, { "epoch": 5.929919137466308, "grad_norm": 0.7344666719436646, "learning_rate": 0.0002445979492714517, "loss": 3.3077, "step": 55000 }, { "epoch": 5.929919137466308, "eval_accuracy": 0.3844160901453895, "eval_loss": 3.3785133361816406, "eval_runtime": 183.7531, "eval_samples_per_second": 98.017, "eval_steps_per_second": 6.128, "step": 55000 }, { "epoch": 5.935309973045822, "grad_norm": 0.6995274424552917, "learning_rate": 0.00024427415002698324, "loss": 3.3377, "step": 55050 }, { "epoch": 5.940700808625337, "grad_norm": 0.7057139277458191, "learning_rate": 0.00024395035078251482, "loss": 3.3364, "step": 55100 }, { "epoch": 5.946091644204852, "grad_norm": 0.7679343819618225, "learning_rate": 0.0002436265515380464, "loss": 3.3456, "step": 55150 }, { "epoch": 5.951482479784366, "grad_norm": 0.7506498098373413, "learning_rate": 0.00024330275229357797, "loss": 3.3296, "step": 55200 }, { "epoch": 5.9568733153638815, "grad_norm": 0.7279877662658691, "learning_rate": 0.00024297895304910952, "loss": 3.3316, "step": 55250 }, { "epoch": 5.962264150943396, "grad_norm": 0.686543881893158, "learning_rate": 0.0002426551538046411, "loss": 3.2993, "step": 55300 }, { "epoch": 5.967654986522911, "grad_norm": 0.7167565226554871, "learning_rate": 0.00024233135456017265, "loss": 3.328, "step": 55350 }, { "epoch": 5.973045822102426, "grad_norm": 0.7074522972106934, "learning_rate": 0.00024200755531570423, "loss": 3.3511, "step": 55400 }, { "epoch": 5.97843665768194, "grad_norm": 0.6831624507904053, "learning_rate": 0.0002416837560712358, "loss": 3.3474, "step": 55450 }, { "epoch": 5.9838274932614555, "grad_norm": 0.7267654538154602, "learning_rate": 0.00024135995682676739, "loss": 3.3315, "step": 55500 }, { "epoch": 5.989218328840971, "grad_norm": 0.7426101565361023, "learning_rate": 0.00024103615758229896, "loss": 3.3361, "step": 55550 }, { "epoch": 5.994609164420485, "grad_norm": 0.7580520510673523, "learning_rate": 0.00024071235833783054, "loss": 3.3454, "step": 55600 }, { "epoch": 6.0, "grad_norm": 1.4545377492904663, "learning_rate": 0.00024038855909336212, "loss": 3.3377, "step": 55650 }, { "epoch": 6.005390835579515, "grad_norm": 0.6739314198493958, "learning_rate": 0.00024006475984889364, "loss": 3.2288, "step": 55700 }, { "epoch": 6.010781671159029, "grad_norm": 0.6961607933044434, "learning_rate": 0.00023974096060442522, "loss": 3.2358, "step": 55750 }, { "epoch": 6.0161725067385445, "grad_norm": 0.7214361429214478, "learning_rate": 0.0002394171613599568, "loss": 3.2439, "step": 55800 }, { "epoch": 6.02156334231806, "grad_norm": 0.7168677449226379, "learning_rate": 0.00023909336211548837, "loss": 3.2532, "step": 55850 }, { "epoch": 6.026954177897574, "grad_norm": 0.7242555618286133, "learning_rate": 0.00023876956287101995, "loss": 3.2447, "step": 55900 }, { "epoch": 6.032345013477089, "grad_norm": 0.7259112596511841, "learning_rate": 0.00023844576362655153, "loss": 3.2433, "step": 55950 }, { "epoch": 6.037735849056604, "grad_norm": 0.7300722002983093, "learning_rate": 0.0002381219643820831, "loss": 3.2654, "step": 56000 }, { "epoch": 6.037735849056604, "eval_accuracy": 0.38417759705663673, "eval_loss": 3.380847454071045, "eval_runtime": 183.8731, "eval_samples_per_second": 97.953, "eval_steps_per_second": 6.124, "step": 56000 }, { "epoch": 6.0431266846361185, "grad_norm": 0.7531189322471619, "learning_rate": 0.00023779816513761466, "loss": 3.2651, "step": 56050 }, { "epoch": 6.048517520215634, "grad_norm": 0.7250877022743225, "learning_rate": 0.00023748084187803558, "loss": 3.2445, "step": 56100 }, { "epoch": 6.053908355795148, "grad_norm": 0.7453233599662781, "learning_rate": 0.00023715704263356716, "loss": 3.2522, "step": 56150 }, { "epoch": 6.059299191374663, "grad_norm": 0.7315291166305542, "learning_rate": 0.00023683324338909874, "loss": 3.2403, "step": 56200 }, { "epoch": 6.064690026954178, "grad_norm": 0.7161756157875061, "learning_rate": 0.00023650944414463032, "loss": 3.2553, "step": 56250 }, { "epoch": 6.070080862533692, "grad_norm": 0.7513576745986938, "learning_rate": 0.0002361856449001619, "loss": 3.2478, "step": 56300 }, { "epoch": 6.0754716981132075, "grad_norm": 0.7142834663391113, "learning_rate": 0.00023586184565569347, "loss": 3.2543, "step": 56350 }, { "epoch": 6.080862533692723, "grad_norm": 0.6998046040534973, "learning_rate": 0.00023553804641122502, "loss": 3.2514, "step": 56400 }, { "epoch": 6.086253369272237, "grad_norm": 0.7207847833633423, "learning_rate": 0.00023521424716675657, "loss": 3.2571, "step": 56450 }, { "epoch": 6.091644204851752, "grad_norm": 0.7710952162742615, "learning_rate": 0.00023489044792228815, "loss": 3.2567, "step": 56500 }, { "epoch": 6.097035040431267, "grad_norm": 0.7176562547683716, "learning_rate": 0.00023456664867781973, "loss": 3.2456, "step": 56550 }, { "epoch": 6.1024258760107815, "grad_norm": 0.7144829630851746, "learning_rate": 0.0002342428494333513, "loss": 3.2536, "step": 56600 }, { "epoch": 6.107816711590297, "grad_norm": 0.7680326104164124, "learning_rate": 0.00023391905018888288, "loss": 3.2511, "step": 56650 }, { "epoch": 6.113207547169812, "grad_norm": 0.7204803824424744, "learning_rate": 0.00023359525094441443, "loss": 3.2494, "step": 56700 }, { "epoch": 6.118598382749326, "grad_norm": 0.7161763310432434, "learning_rate": 0.000233271451699946, "loss": 3.2641, "step": 56750 }, { "epoch": 6.123989218328841, "grad_norm": 0.7113860845565796, "learning_rate": 0.0002329476524554776, "loss": 3.2594, "step": 56800 }, { "epoch": 6.129380053908355, "grad_norm": 0.710684061050415, "learning_rate": 0.00023262385321100917, "loss": 3.2647, "step": 56850 }, { "epoch": 6.1347708894878705, "grad_norm": 0.7076157331466675, "learning_rate": 0.00023230005396654072, "loss": 3.2609, "step": 56900 }, { "epoch": 6.140161725067386, "grad_norm": 0.7792293429374695, "learning_rate": 0.0002319762547220723, "loss": 3.2541, "step": 56950 }, { "epoch": 6.1455525606469, "grad_norm": 0.7125359773635864, "learning_rate": 0.00023165245547760387, "loss": 3.2674, "step": 57000 }, { "epoch": 6.1455525606469, "eval_accuracy": 0.38459482414584434, "eval_loss": 3.3812313079833984, "eval_runtime": 184.1175, "eval_samples_per_second": 97.823, "eval_steps_per_second": 6.116, "step": 57000 }, { "epoch": 6.150943396226415, "grad_norm": 0.758549153804779, "learning_rate": 0.00023132865623313542, "loss": 3.2666, "step": 57050 }, { "epoch": 6.15633423180593, "grad_norm": 0.7191378474235535, "learning_rate": 0.000231004856988667, "loss": 3.2799, "step": 57100 }, { "epoch": 6.1617250673854445, "grad_norm": 0.7202212810516357, "learning_rate": 0.00023068105774419858, "loss": 3.2571, "step": 57150 }, { "epoch": 6.16711590296496, "grad_norm": 0.7385558485984802, "learning_rate": 0.00023035725849973016, "loss": 3.2776, "step": 57200 }, { "epoch": 6.172506738544475, "grad_norm": 0.740178644657135, "learning_rate": 0.00023003345925526173, "loss": 3.2691, "step": 57250 }, { "epoch": 6.177897574123989, "grad_norm": 0.741432785987854, "learning_rate": 0.00022970966001079328, "loss": 3.2609, "step": 57300 }, { "epoch": 6.183288409703504, "grad_norm": 0.7559853792190552, "learning_rate": 0.00022938586076632484, "loss": 3.2646, "step": 57350 }, { "epoch": 6.188679245283019, "grad_norm": 0.7264957427978516, "learning_rate": 0.0002290620615218564, "loss": 3.2543, "step": 57400 }, { "epoch": 6.1940700808625335, "grad_norm": 0.728132426738739, "learning_rate": 0.000228738262277388, "loss": 3.2698, "step": 57450 }, { "epoch": 6.199460916442049, "grad_norm": 0.7514779567718506, "learning_rate": 0.00022841446303291957, "loss": 3.2635, "step": 57500 }, { "epoch": 6.204851752021563, "grad_norm": 0.7553060054779053, "learning_rate": 0.00022809066378845115, "loss": 3.2622, "step": 57550 }, { "epoch": 6.210242587601078, "grad_norm": 0.7538262605667114, "learning_rate": 0.0002277733405288721, "loss": 3.2754, "step": 57600 }, { "epoch": 6.215633423180593, "grad_norm": 0.7464041113853455, "learning_rate": 0.00022744954128440365, "loss": 3.2429, "step": 57650 }, { "epoch": 6.2210242587601075, "grad_norm": 0.7480874061584473, "learning_rate": 0.0002271257420399352, "loss": 3.266, "step": 57700 }, { "epoch": 6.226415094339623, "grad_norm": 0.7676622867584229, "learning_rate": 0.00022680194279546678, "loss": 3.2746, "step": 57750 }, { "epoch": 6.231805929919138, "grad_norm": 0.7216105461120605, "learning_rate": 0.00022647814355099835, "loss": 3.2647, "step": 57800 }, { "epoch": 6.237196765498652, "grad_norm": 0.7596701383590698, "learning_rate": 0.00022615434430652993, "loss": 3.2692, "step": 57850 }, { "epoch": 6.242587601078167, "grad_norm": 0.7444921135902405, "learning_rate": 0.0002258305450620615, "loss": 3.2524, "step": 57900 }, { "epoch": 6.247978436657682, "grad_norm": 0.757045567035675, "learning_rate": 0.0002255067458175931, "loss": 3.2445, "step": 57950 }, { "epoch": 6.2533692722371965, "grad_norm": 0.7225556373596191, "learning_rate": 0.00022518294657312467, "loss": 3.2691, "step": 58000 }, { "epoch": 6.2533692722371965, "eval_accuracy": 0.3846125345665581, "eval_loss": 3.3798320293426514, "eval_runtime": 183.8445, "eval_samples_per_second": 97.969, "eval_steps_per_second": 6.125, "step": 58000 }, { "epoch": 6.258760107816712, "grad_norm": 0.7398164868354797, "learning_rate": 0.0002248591473286562, "loss": 3.268, "step": 58050 }, { "epoch": 6.264150943396227, "grad_norm": 0.7138086557388306, "learning_rate": 0.00022453534808418777, "loss": 3.2739, "step": 58100 }, { "epoch": 6.269541778975741, "grad_norm": 0.7250698804855347, "learning_rate": 0.00022421154883971934, "loss": 3.2892, "step": 58150 }, { "epoch": 6.274932614555256, "grad_norm": 0.735781192779541, "learning_rate": 0.00022388774959525092, "loss": 3.2771, "step": 58200 }, { "epoch": 6.280323450134771, "grad_norm": 0.6944800019264221, "learning_rate": 0.0002235639503507825, "loss": 3.2606, "step": 58250 }, { "epoch": 6.285714285714286, "grad_norm": 0.7499034404754639, "learning_rate": 0.00022324015110631408, "loss": 3.2808, "step": 58300 }, { "epoch": 6.291105121293801, "grad_norm": 0.7447680234909058, "learning_rate": 0.00022291635186184565, "loss": 3.2746, "step": 58350 }, { "epoch": 6.296495956873315, "grad_norm": 0.7814144492149353, "learning_rate": 0.0002225925526173772, "loss": 3.2653, "step": 58400 }, { "epoch": 6.30188679245283, "grad_norm": 0.7258203625679016, "learning_rate": 0.00022226875337290878, "loss": 3.2543, "step": 58450 }, { "epoch": 6.307277628032345, "grad_norm": 0.732068657875061, "learning_rate": 0.00022194495412844033, "loss": 3.2562, "step": 58500 }, { "epoch": 6.3126684636118595, "grad_norm": 0.7250475883483887, "learning_rate": 0.0002216211548839719, "loss": 3.274, "step": 58550 }, { "epoch": 6.318059299191375, "grad_norm": 0.725308895111084, "learning_rate": 0.0002212973556395035, "loss": 3.272, "step": 58600 }, { "epoch": 6.32345013477089, "grad_norm": 0.7906202673912048, "learning_rate": 0.00022097355639503507, "loss": 3.2792, "step": 58650 }, { "epoch": 6.328840970350404, "grad_norm": 0.8644693493843079, "learning_rate": 0.00022064975715056664, "loss": 3.2385, "step": 58700 }, { "epoch": 6.334231805929919, "grad_norm": 0.7321583032608032, "learning_rate": 0.0002203259579060982, "loss": 3.276, "step": 58750 }, { "epoch": 6.339622641509434, "grad_norm": 0.8114139437675476, "learning_rate": 0.00022000215866162977, "loss": 3.2713, "step": 58800 }, { "epoch": 6.345013477088949, "grad_norm": 0.7152981758117676, "learning_rate": 0.00021967835941716135, "loss": 3.257, "step": 58850 }, { "epoch": 6.350404312668464, "grad_norm": 0.7789607048034668, "learning_rate": 0.0002193545601726929, "loss": 3.2629, "step": 58900 }, { "epoch": 6.355795148247978, "grad_norm": 0.7857829928398132, "learning_rate": 0.00021903076092822448, "loss": 3.2739, "step": 58950 }, { "epoch": 6.361185983827493, "grad_norm": 0.7703624367713928, "learning_rate": 0.00021870696168375606, "loss": 3.276, "step": 59000 }, { "epoch": 6.361185983827493, "eval_accuracy": 0.38511038209852405, "eval_loss": 3.375075101852417, "eval_runtime": 184.271, "eval_samples_per_second": 97.742, "eval_steps_per_second": 6.111, "step": 59000 }, { "epoch": 6.366576819407008, "grad_norm": 0.7588159441947937, "learning_rate": 0.0002183831624392876, "loss": 3.2779, "step": 59050 }, { "epoch": 6.3719676549865225, "grad_norm": 0.8037658333778381, "learning_rate": 0.00021805936319481918, "loss": 3.2748, "step": 59100 }, { "epoch": 6.377358490566038, "grad_norm": 0.7042297124862671, "learning_rate": 0.00021773556395035076, "loss": 3.2582, "step": 59150 }, { "epoch": 6.382749326145553, "grad_norm": 0.7502013444900513, "learning_rate": 0.00021741176470588234, "loss": 3.2693, "step": 59200 }, { "epoch": 6.388140161725067, "grad_norm": 0.7545161843299866, "learning_rate": 0.00021708796546141392, "loss": 3.2658, "step": 59250 }, { "epoch": 6.393530997304582, "grad_norm": 0.739264190196991, "learning_rate": 0.0002167641662169455, "loss": 3.2854, "step": 59300 }, { "epoch": 6.398921832884097, "grad_norm": 0.8167062997817993, "learning_rate": 0.00021644036697247702, "loss": 3.2723, "step": 59350 }, { "epoch": 6.404312668463612, "grad_norm": 0.7420883178710938, "learning_rate": 0.0002161165677280086, "loss": 3.2852, "step": 59400 }, { "epoch": 6.409703504043127, "grad_norm": 0.7469702363014221, "learning_rate": 0.00021579276848354017, "loss": 3.2801, "step": 59450 }, { "epoch": 6.415094339622642, "grad_norm": 0.796428382396698, "learning_rate": 0.00021546896923907175, "loss": 3.271, "step": 59500 }, { "epoch": 6.420485175202156, "grad_norm": 0.7591307163238525, "learning_rate": 0.00021514516999460333, "loss": 3.2745, "step": 59550 }, { "epoch": 6.425876010781671, "grad_norm": 0.7786388397216797, "learning_rate": 0.0002148213707501349, "loss": 3.2793, "step": 59600 }, { "epoch": 6.431266846361186, "grad_norm": 0.8503186702728271, "learning_rate": 0.00021449757150566648, "loss": 3.2832, "step": 59650 }, { "epoch": 6.436657681940701, "grad_norm": 0.7847234010696411, "learning_rate": 0.00021417377226119806, "loss": 3.2923, "step": 59700 }, { "epoch": 6.442048517520216, "grad_norm": 0.7348524928092957, "learning_rate": 0.0002138499730167296, "loss": 3.2674, "step": 59750 }, { "epoch": 6.44743935309973, "grad_norm": 0.7561467289924622, "learning_rate": 0.00021352617377226116, "loss": 3.2832, "step": 59800 }, { "epoch": 6.452830188679245, "grad_norm": 0.7439683675765991, "learning_rate": 0.00021320237452779274, "loss": 3.2768, "step": 59850 }, { "epoch": 6.45822102425876, "grad_norm": 0.7063082456588745, "learning_rate": 0.00021287857528332432, "loss": 3.2697, "step": 59900 }, { "epoch": 6.463611859838275, "grad_norm": 0.7481423020362854, "learning_rate": 0.0002125547760388559, "loss": 3.2742, "step": 59950 }, { "epoch": 6.46900269541779, "grad_norm": 0.7237406373023987, "learning_rate": 0.00021223097679438747, "loss": 3.2903, "step": 60000 }, { "epoch": 6.46900269541779, "eval_accuracy": 0.38577588103638777, "eval_loss": 3.3692877292633057, "eval_runtime": 183.7708, "eval_samples_per_second": 98.008, "eval_steps_per_second": 6.127, "step": 60000 }, { "epoch": 6.474393530997305, "grad_norm": 1.16843843460083, "learning_rate": 0.00021190717754991905, "loss": 3.2967, "step": 60050 }, { "epoch": 6.479784366576819, "grad_norm": 0.7618022561073303, "learning_rate": 0.0002115833783054506, "loss": 3.2958, "step": 60100 }, { "epoch": 6.485175202156334, "grad_norm": 0.7672574520111084, "learning_rate": 0.00021125957906098218, "loss": 3.2812, "step": 60150 }, { "epoch": 6.490566037735849, "grad_norm": 0.7672653794288635, "learning_rate": 0.00021093577981651373, "loss": 3.2856, "step": 60200 }, { "epoch": 6.495956873315364, "grad_norm": 0.7688764929771423, "learning_rate": 0.0002106119805720453, "loss": 3.2644, "step": 60250 }, { "epoch": 6.501347708894879, "grad_norm": 0.715165376663208, "learning_rate": 0.00021028818132757689, "loss": 3.2857, "step": 60300 }, { "epoch": 6.506738544474393, "grad_norm": 0.7562441229820251, "learning_rate": 0.00020996438208310846, "loss": 3.2896, "step": 60350 }, { "epoch": 6.512129380053908, "grad_norm": 0.6920077204704285, "learning_rate": 0.00020964058283864001, "loss": 3.2756, "step": 60400 }, { "epoch": 6.517520215633423, "grad_norm": 0.6866161227226257, "learning_rate": 0.0002093167835941716, "loss": 3.2792, "step": 60450 }, { "epoch": 6.5229110512129385, "grad_norm": 0.7109043598175049, "learning_rate": 0.00020899298434970317, "loss": 3.2687, "step": 60500 }, { "epoch": 6.528301886792453, "grad_norm": 0.7530231475830078, "learning_rate": 0.00020866918510523475, "loss": 3.2742, "step": 60550 }, { "epoch": 6.533692722371968, "grad_norm": 0.7493744492530823, "learning_rate": 0.00020834538586076632, "loss": 3.2848, "step": 60600 }, { "epoch": 6.539083557951482, "grad_norm": 0.733198344707489, "learning_rate": 0.00020802158661629787, "loss": 3.2755, "step": 60650 }, { "epoch": 6.544474393530997, "grad_norm": 0.9269688129425049, "learning_rate": 0.00020769778737182943, "loss": 3.2978, "step": 60700 }, { "epoch": 6.549865229110512, "grad_norm": 0.7194287776947021, "learning_rate": 0.000207373988127361, "loss": 3.278, "step": 60750 }, { "epoch": 6.555256064690027, "grad_norm": 0.7988015413284302, "learning_rate": 0.00020705018888289258, "loss": 3.2922, "step": 60800 }, { "epoch": 6.560646900269542, "grad_norm": 0.7761523127555847, "learning_rate": 0.00020672638963842416, "loss": 3.2839, "step": 60850 }, { "epoch": 6.566037735849057, "grad_norm": 0.8137022852897644, "learning_rate": 0.00020640259039395574, "loss": 3.2923, "step": 60900 }, { "epoch": 6.571428571428571, "grad_norm": 0.7631632089614868, "learning_rate": 0.00020607879114948731, "loss": 3.2751, "step": 60950 }, { "epoch": 6.576819407008086, "grad_norm": 0.7450467348098755, "learning_rate": 0.0002057549919050189, "loss": 3.2681, "step": 61000 }, { "epoch": 6.576819407008086, "eval_accuracy": 0.3860572920158272, "eval_loss": 3.3661625385284424, "eval_runtime": 183.9813, "eval_samples_per_second": 97.896, "eval_steps_per_second": 6.12, "step": 61000 }, { "epoch": 6.5822102425876015, "grad_norm": 0.7303608059883118, "learning_rate": 0.00020543119266055041, "loss": 3.2785, "step": 61050 }, { "epoch": 6.587601078167116, "grad_norm": 0.7955816984176636, "learning_rate": 0.000205107393416082, "loss": 3.2928, "step": 61100 }, { "epoch": 6.592991913746631, "grad_norm": 0.772139310836792, "learning_rate": 0.00020478359417161357, "loss": 3.2785, "step": 61150 }, { "epoch": 6.598382749326145, "grad_norm": 0.7001410126686096, "learning_rate": 0.00020445979492714515, "loss": 3.2655, "step": 61200 }, { "epoch": 6.60377358490566, "grad_norm": 0.7417465448379517, "learning_rate": 0.00020413599568267673, "loss": 3.2931, "step": 61250 }, { "epoch": 6.609164420485175, "grad_norm": 0.7509415149688721, "learning_rate": 0.0002038121964382083, "loss": 3.2842, "step": 61300 }, { "epoch": 6.6145552560646905, "grad_norm": 0.8277612924575806, "learning_rate": 0.00020348839719373988, "loss": 3.2657, "step": 61350 }, { "epoch": 6.619946091644205, "grad_norm": 0.7039164304733276, "learning_rate": 0.00020316459794927146, "loss": 3.2932, "step": 61400 }, { "epoch": 6.62533692722372, "grad_norm": 0.7653436064720154, "learning_rate": 0.000202840798704803, "loss": 3.2882, "step": 61450 }, { "epoch": 6.630727762803234, "grad_norm": 0.753315806388855, "learning_rate": 0.00020251699946033456, "loss": 3.2836, "step": 61500 }, { "epoch": 6.636118598382749, "grad_norm": 0.7230058312416077, "learning_rate": 0.00020219320021586614, "loss": 3.2888, "step": 61550 }, { "epoch": 6.6415094339622645, "grad_norm": 0.7700191140174866, "learning_rate": 0.00020186940097139771, "loss": 3.293, "step": 61600 }, { "epoch": 6.646900269541779, "grad_norm": 0.7520590424537659, "learning_rate": 0.0002015456017269293, "loss": 3.2894, "step": 61650 }, { "epoch": 6.652291105121294, "grad_norm": 0.7307371497154236, "learning_rate": 0.00020122180248246087, "loss": 3.2735, "step": 61700 }, { "epoch": 6.657681940700809, "grad_norm": 0.7702900171279907, "learning_rate": 0.00020089800323799242, "loss": 3.3107, "step": 61750 }, { "epoch": 6.663072776280323, "grad_norm": 0.7241050601005554, "learning_rate": 0.000200574203993524, "loss": 3.2667, "step": 61800 }, { "epoch": 6.668463611859838, "grad_norm": 0.7302685379981995, "learning_rate": 0.00020025040474905558, "loss": 3.2695, "step": 61850 }, { "epoch": 6.6738544474393535, "grad_norm": 0.7514686584472656, "learning_rate": 0.0001999330814894765, "loss": 3.2696, "step": 61900 }, { "epoch": 6.679245283018868, "grad_norm": 0.7941187024116516, "learning_rate": 0.00019960928224500808, "loss": 3.2697, "step": 61950 }, { "epoch": 6.684636118598383, "grad_norm": 0.7222111821174622, "learning_rate": 0.00019928548300053966, "loss": 3.2793, "step": 62000 }, { "epoch": 6.684636118598383, "eval_accuracy": 0.38628166022916416, "eval_loss": 3.3625741004943848, "eval_runtime": 184.0577, "eval_samples_per_second": 97.855, "eval_steps_per_second": 6.118, "step": 62000 }, { "epoch": 6.690026954177897, "grad_norm": 0.7426719665527344, "learning_rate": 0.00019896168375607123, "loss": 3.2672, "step": 62050 }, { "epoch": 6.695417789757412, "grad_norm": 0.7612356543540955, "learning_rate": 0.00019863788451160278, "loss": 3.2768, "step": 62100 }, { "epoch": 6.7008086253369274, "grad_norm": 0.7817362546920776, "learning_rate": 0.00019831408526713436, "loss": 3.2943, "step": 62150 }, { "epoch": 6.706199460916442, "grad_norm": 0.7813862562179565, "learning_rate": 0.00019799028602266594, "loss": 3.28, "step": 62200 }, { "epoch": 6.711590296495957, "grad_norm": 0.8101115822792053, "learning_rate": 0.0001976664867781975, "loss": 3.2889, "step": 62250 }, { "epoch": 6.716981132075472, "grad_norm": 0.7497881054878235, "learning_rate": 0.00019734916351861844, "loss": 3.2753, "step": 62300 }, { "epoch": 6.722371967654986, "grad_norm": 0.793246865272522, "learning_rate": 0.00019702536427415002, "loss": 3.283, "step": 62350 }, { "epoch": 6.727762803234501, "grad_norm": 0.7845967411994934, "learning_rate": 0.0001967015650296816, "loss": 3.287, "step": 62400 }, { "epoch": 6.7331536388140165, "grad_norm": 0.7628470063209534, "learning_rate": 0.00019637776578521315, "loss": 3.2814, "step": 62450 }, { "epoch": 6.738544474393531, "grad_norm": 0.7873274683952332, "learning_rate": 0.00019605396654074473, "loss": 3.2863, "step": 62500 }, { "epoch": 6.743935309973046, "grad_norm": 0.7793614268302917, "learning_rate": 0.0001957301672962763, "loss": 3.2713, "step": 62550 }, { "epoch": 6.74932614555256, "grad_norm": 0.8013772964477539, "learning_rate": 0.00019540636805180785, "loss": 3.28, "step": 62600 }, { "epoch": 6.754716981132075, "grad_norm": 0.7564038634300232, "learning_rate": 0.00019508256880733943, "loss": 3.2837, "step": 62650 }, { "epoch": 6.7601078167115904, "grad_norm": 0.7453675270080566, "learning_rate": 0.000194758769562871, "loss": 3.2719, "step": 62700 }, { "epoch": 6.765498652291106, "grad_norm": 0.8338208198547363, "learning_rate": 0.00019443497031840256, "loss": 3.2872, "step": 62750 }, { "epoch": 6.77088948787062, "grad_norm": 0.7567179203033447, "learning_rate": 0.00019411117107393414, "loss": 3.2853, "step": 62800 }, { "epoch": 6.776280323450135, "grad_norm": 0.770332396030426, "learning_rate": 0.00019378737182946572, "loss": 3.2586, "step": 62850 }, { "epoch": 6.781671159029649, "grad_norm": 0.7977242469787598, "learning_rate": 0.0001934635725849973, "loss": 3.2788, "step": 62900 }, { "epoch": 6.787061994609164, "grad_norm": 0.7252545952796936, "learning_rate": 0.00019313977334052887, "loss": 3.2967, "step": 62950 }, { "epoch": 6.7924528301886795, "grad_norm": 0.7643560171127319, "learning_rate": 0.00019281597409606042, "loss": 3.2753, "step": 63000 }, { "epoch": 6.7924528301886795, "eval_accuracy": 0.3866451041389036, "eval_loss": 3.3582923412323, "eval_runtime": 183.964, "eval_samples_per_second": 97.905, "eval_steps_per_second": 6.121, "step": 63000 }, { "epoch": 6.797843665768194, "grad_norm": 0.769708514213562, "learning_rate": 0.00019249217485159197, "loss": 3.2593, "step": 63050 }, { "epoch": 6.803234501347709, "grad_norm": 0.7512614130973816, "learning_rate": 0.00019216837560712355, "loss": 3.279, "step": 63100 }, { "epoch": 6.808625336927224, "grad_norm": 0.7702354192733765, "learning_rate": 0.00019184457636265513, "loss": 3.2774, "step": 63150 }, { "epoch": 6.814016172506738, "grad_norm": 0.771395742893219, "learning_rate": 0.0001915207771181867, "loss": 3.2896, "step": 63200 }, { "epoch": 6.819407008086253, "grad_norm": 0.7912437319755554, "learning_rate": 0.00019119697787371828, "loss": 3.272, "step": 63250 }, { "epoch": 6.824797843665769, "grad_norm": 0.7434420585632324, "learning_rate": 0.00019087317862924986, "loss": 3.3104, "step": 63300 }, { "epoch": 6.830188679245283, "grad_norm": 0.775771975517273, "learning_rate": 0.00019054937938478144, "loss": 3.2854, "step": 63350 }, { "epoch": 6.835579514824798, "grad_norm": 0.768345057964325, "learning_rate": 0.00019022558014031302, "loss": 3.2817, "step": 63400 }, { "epoch": 6.840970350404312, "grad_norm": 0.7371838092803955, "learning_rate": 0.00018990178089584454, "loss": 3.2816, "step": 63450 }, { "epoch": 6.846361185983827, "grad_norm": 0.7820178270339966, "learning_rate": 0.00018957798165137612, "loss": 3.2847, "step": 63500 }, { "epoch": 6.8517520215633425, "grad_norm": 0.7918660044670105, "learning_rate": 0.0001892541824069077, "loss": 3.277, "step": 63550 }, { "epoch": 6.857142857142857, "grad_norm": 0.73056960105896, "learning_rate": 0.00018893038316243927, "loss": 3.2855, "step": 63600 }, { "epoch": 6.862533692722372, "grad_norm": 0.7188368439674377, "learning_rate": 0.00018860658391797085, "loss": 3.2765, "step": 63650 }, { "epoch": 6.867924528301887, "grad_norm": 0.7747359275817871, "learning_rate": 0.00018828278467350243, "loss": 3.2962, "step": 63700 }, { "epoch": 6.873315363881401, "grad_norm": 0.7617150545120239, "learning_rate": 0.000187958985429034, "loss": 3.2799, "step": 63750 }, { "epoch": 6.878706199460916, "grad_norm": 0.8125529289245605, "learning_rate": 0.00018763518618456556, "loss": 3.2672, "step": 63800 }, { "epoch": 6.884097035040432, "grad_norm": 0.8084162473678589, "learning_rate": 0.0001873113869400971, "loss": 3.2708, "step": 63850 }, { "epoch": 6.889487870619946, "grad_norm": 0.735432505607605, "learning_rate": 0.00018698758769562868, "loss": 3.2802, "step": 63900 }, { "epoch": 6.894878706199461, "grad_norm": 0.7418496012687683, "learning_rate": 0.00018666378845116026, "loss": 3.2816, "step": 63950 }, { "epoch": 6.900269541778976, "grad_norm": 0.7551790475845337, "learning_rate": 0.00018633998920669184, "loss": 3.2875, "step": 64000 }, { "epoch": 6.900269541778976, "eval_accuracy": 0.3873492835048292, "eval_loss": 3.3529202938079834, "eval_runtime": 183.9905, "eval_samples_per_second": 97.891, "eval_steps_per_second": 6.12, "step": 64000 }, { "epoch": 6.90566037735849, "grad_norm": 0.7530993223190308, "learning_rate": 0.00018601618996222342, "loss": 3.282, "step": 64050 }, { "epoch": 6.9110512129380055, "grad_norm": 0.7252298593521118, "learning_rate": 0.00018569239071775497, "loss": 3.283, "step": 64100 }, { "epoch": 6.916442048517521, "grad_norm": 0.7530275583267212, "learning_rate": 0.00018536859147328655, "loss": 3.2738, "step": 64150 }, { "epoch": 6.921832884097035, "grad_norm": 0.7903400659561157, "learning_rate": 0.00018504479222881812, "loss": 3.2877, "step": 64200 }, { "epoch": 6.92722371967655, "grad_norm": 0.786756157875061, "learning_rate": 0.0001847209929843497, "loss": 3.2815, "step": 64250 }, { "epoch": 6.932614555256064, "grad_norm": 0.8159517645835876, "learning_rate": 0.00018439719373988125, "loss": 3.2693, "step": 64300 }, { "epoch": 6.938005390835579, "grad_norm": 0.7632095217704773, "learning_rate": 0.00018407339449541283, "loss": 3.2753, "step": 64350 }, { "epoch": 6.943396226415095, "grad_norm": 0.7830843329429626, "learning_rate": 0.00018374959525094438, "loss": 3.2897, "step": 64400 }, { "epoch": 6.948787061994609, "grad_norm": 0.8015329241752625, "learning_rate": 0.00018342579600647596, "loss": 3.2668, "step": 64450 }, { "epoch": 6.954177897574124, "grad_norm": 0.8383193612098694, "learning_rate": 0.00018310199676200753, "loss": 3.2993, "step": 64500 }, { "epoch": 6.959568733153639, "grad_norm": 0.7615668177604675, "learning_rate": 0.0001827781975175391, "loss": 3.2865, "step": 64550 }, { "epoch": 6.964959568733153, "grad_norm": 0.7987204194068909, "learning_rate": 0.0001824543982730707, "loss": 3.291, "step": 64600 }, { "epoch": 6.9703504043126685, "grad_norm": 0.7923724055290222, "learning_rate": 0.00018213059902860227, "loss": 3.2708, "step": 64650 }, { "epoch": 6.975741239892184, "grad_norm": 0.8186677098274231, "learning_rate": 0.00018180679978413382, "loss": 3.2715, "step": 64700 }, { "epoch": 6.981132075471698, "grad_norm": 0.8212941288948059, "learning_rate": 0.00018148300053966537, "loss": 3.2677, "step": 64750 }, { "epoch": 6.986522911051213, "grad_norm": 0.8504579067230225, "learning_rate": 0.00018115920129519695, "loss": 3.2813, "step": 64800 }, { "epoch": 6.991913746630727, "grad_norm": 0.7678773403167725, "learning_rate": 0.00018083540205072852, "loss": 3.2945, "step": 64850 }, { "epoch": 6.997304582210242, "grad_norm": 0.7796960473060608, "learning_rate": 0.0001805116028062601, "loss": 3.287, "step": 64900 }, { "epoch": 7.002695417789758, "grad_norm": 0.8160527944564819, "learning_rate": 0.00018018780356179168, "loss": 3.2342, "step": 64950 }, { "epoch": 7.008086253369272, "grad_norm": 0.7837775945663452, "learning_rate": 0.00017986400431732326, "loss": 3.2082, "step": 65000 }, { "epoch": 7.008086253369272, "eval_accuracy": 0.38750389656418927, "eval_loss": 3.3548026084899902, "eval_runtime": 183.8139, "eval_samples_per_second": 97.985, "eval_steps_per_second": 6.126, "step": 65000 }, { "epoch": 7.013477088948787, "grad_norm": 0.8339738845825195, "learning_rate": 0.00017954020507285483, "loss": 3.1856, "step": 65050 }, { "epoch": 7.018867924528302, "grad_norm": 0.8263199329376221, "learning_rate": 0.0001792164058283864, "loss": 3.1704, "step": 65100 }, { "epoch": 7.024258760107816, "grad_norm": 0.7783221006393433, "learning_rate": 0.00017889260658391794, "loss": 3.2089, "step": 65150 }, { "epoch": 7.0296495956873315, "grad_norm": 0.762937068939209, "learning_rate": 0.00017856880733944951, "loss": 3.2022, "step": 65200 }, { "epoch": 7.035040431266847, "grad_norm": 0.7570405602455139, "learning_rate": 0.0001782450080949811, "loss": 3.1999, "step": 65250 }, { "epoch": 7.040431266846361, "grad_norm": 0.833223819732666, "learning_rate": 0.00017792120885051267, "loss": 3.1842, "step": 65300 }, { "epoch": 7.045822102425876, "grad_norm": 0.746315598487854, "learning_rate": 0.00017760388559093362, "loss": 3.1927, "step": 65350 }, { "epoch": 7.051212938005391, "grad_norm": 0.7781280279159546, "learning_rate": 0.0001772800863464652, "loss": 3.1935, "step": 65400 }, { "epoch": 7.056603773584905, "grad_norm": 0.839777946472168, "learning_rate": 0.00017695628710199672, "loss": 3.2116, "step": 65450 }, { "epoch": 7.061994609164421, "grad_norm": 0.8046913146972656, "learning_rate": 0.0001766324878575283, "loss": 3.2046, "step": 65500 }, { "epoch": 7.067385444743936, "grad_norm": 0.7587380409240723, "learning_rate": 0.00017630868861305988, "loss": 3.208, "step": 65550 }, { "epoch": 7.07277628032345, "grad_norm": 0.7880898118019104, "learning_rate": 0.00017598488936859146, "loss": 3.2019, "step": 65600 }, { "epoch": 7.078167115902965, "grad_norm": 0.7449747323989868, "learning_rate": 0.00017566109012412303, "loss": 3.1968, "step": 65650 }, { "epoch": 7.083557951482479, "grad_norm": 0.7823360562324524, "learning_rate": 0.0001753372908796546, "loss": 3.2062, "step": 65700 }, { "epoch": 7.0889487870619945, "grad_norm": 0.78197181224823, "learning_rate": 0.0001750134916351862, "loss": 3.2057, "step": 65750 }, { "epoch": 7.09433962264151, "grad_norm": 0.7566399574279785, "learning_rate": 0.00017468969239071774, "loss": 3.2, "step": 65800 }, { "epoch": 7.099730458221024, "grad_norm": 0.8071035742759705, "learning_rate": 0.00017436589314624932, "loss": 3.1944, "step": 65850 }, { "epoch": 7.105121293800539, "grad_norm": 0.8223758339881897, "learning_rate": 0.00017404209390178087, "loss": 3.1912, "step": 65900 }, { "epoch": 7.110512129380054, "grad_norm": 0.7952433824539185, "learning_rate": 0.00017371829465731244, "loss": 3.222, "step": 65950 }, { "epoch": 7.115902964959568, "grad_norm": 0.7765783667564392, "learning_rate": 0.00017339449541284402, "loss": 3.2291, "step": 66000 }, { "epoch": 7.115902964959568, "eval_accuracy": 0.38722042117988126, "eval_loss": 3.3594112396240234, "eval_runtime": 183.9219, "eval_samples_per_second": 97.927, "eval_steps_per_second": 6.122, "step": 66000 }, { "epoch": 7.121293800539084, "grad_norm": 0.7651994824409485, "learning_rate": 0.0001730706961683756, "loss": 3.1816, "step": 66050 }, { "epoch": 7.126684636118599, "grad_norm": 0.7809786796569824, "learning_rate": 0.00017274689692390715, "loss": 3.2177, "step": 66100 }, { "epoch": 7.132075471698113, "grad_norm": 0.772335410118103, "learning_rate": 0.00017242309767943873, "loss": 3.2218, "step": 66150 }, { "epoch": 7.137466307277628, "grad_norm": 0.8068752884864807, "learning_rate": 0.0001720992984349703, "loss": 3.2222, "step": 66200 }, { "epoch": 7.142857142857143, "grad_norm": 0.7680201530456543, "learning_rate": 0.00017177549919050188, "loss": 3.21, "step": 66250 }, { "epoch": 7.1482479784366575, "grad_norm": 0.7757328748703003, "learning_rate": 0.00017145169994603346, "loss": 3.2183, "step": 66300 }, { "epoch": 7.153638814016173, "grad_norm": 0.7772490978240967, "learning_rate": 0.000171127900701565, "loss": 3.2189, "step": 66350 }, { "epoch": 7.159029649595688, "grad_norm": 0.8381431102752686, "learning_rate": 0.0001708041014570966, "loss": 3.196, "step": 66400 }, { "epoch": 7.164420485175202, "grad_norm": 0.8712314367294312, "learning_rate": 0.00017048030221262814, "loss": 3.2298, "step": 66450 }, { "epoch": 7.169811320754717, "grad_norm": 0.7782764434814453, "learning_rate": 0.00017015650296815972, "loss": 3.2195, "step": 66500 }, { "epoch": 7.175202156334231, "grad_norm": 0.7687488794326782, "learning_rate": 0.0001698327037236913, "loss": 3.2183, "step": 66550 }, { "epoch": 7.180592991913747, "grad_norm": 0.7948805689811707, "learning_rate": 0.00016950890447922287, "loss": 3.2202, "step": 66600 }, { "epoch": 7.185983827493262, "grad_norm": 0.8118099570274353, "learning_rate": 0.00016918510523475445, "loss": 3.2115, "step": 66650 }, { "epoch": 7.191374663072776, "grad_norm": 0.8137215971946716, "learning_rate": 0.00016886130599028603, "loss": 3.2174, "step": 66700 }, { "epoch": 7.196765498652291, "grad_norm": 0.7838942408561707, "learning_rate": 0.00016853750674581755, "loss": 3.2021, "step": 66750 }, { "epoch": 7.202156334231806, "grad_norm": 0.7940311431884766, "learning_rate": 0.00016821370750134913, "loss": 3.2059, "step": 66800 }, { "epoch": 7.2075471698113205, "grad_norm": 0.7829987406730652, "learning_rate": 0.0001678899082568807, "loss": 3.2321, "step": 66850 }, { "epoch": 7.212938005390836, "grad_norm": 0.7903277277946472, "learning_rate": 0.00016756610901241228, "loss": 3.2087, "step": 66900 }, { "epoch": 7.218328840970351, "grad_norm": 0.8051270246505737, "learning_rate": 0.00016724230976794386, "loss": 3.2256, "step": 66950 }, { "epoch": 7.223719676549865, "grad_norm": 0.7584980726242065, "learning_rate": 0.00016691851052347544, "loss": 3.2144, "step": 67000 }, { "epoch": 7.223719676549865, "eval_accuracy": 0.38720227614761626, "eval_loss": 3.356732130050659, "eval_runtime": 184.0554, "eval_samples_per_second": 97.856, "eval_steps_per_second": 6.118, "step": 67000 }, { "epoch": 7.22911051212938, "grad_norm": 0.818000316619873, "learning_rate": 0.00016659471127900702, "loss": 3.2173, "step": 67050 }, { "epoch": 7.234501347708895, "grad_norm": 0.8095887899398804, "learning_rate": 0.0001662709120345386, "loss": 3.2123, "step": 67100 }, { "epoch": 7.2398921832884096, "grad_norm": 0.8081578612327576, "learning_rate": 0.00016594711279007015, "loss": 3.2229, "step": 67150 }, { "epoch": 7.245283018867925, "grad_norm": 0.7935521006584167, "learning_rate": 0.0001656233135456017, "loss": 3.2172, "step": 67200 }, { "epoch": 7.250673854447439, "grad_norm": 0.8998975157737732, "learning_rate": 0.00016529951430113327, "loss": 3.2108, "step": 67250 }, { "epoch": 7.256064690026954, "grad_norm": 0.8090941309928894, "learning_rate": 0.00016497571505666485, "loss": 3.2352, "step": 67300 }, { "epoch": 7.261455525606469, "grad_norm": 0.7772446870803833, "learning_rate": 0.00016465191581219643, "loss": 3.2357, "step": 67350 }, { "epoch": 7.2668463611859835, "grad_norm": 0.8178821206092834, "learning_rate": 0.000164328116567728, "loss": 3.2374, "step": 67400 }, { "epoch": 7.272237196765499, "grad_norm": 0.7874232530593872, "learning_rate": 0.00016400431732325956, "loss": 3.2318, "step": 67450 }, { "epoch": 7.277628032345014, "grad_norm": 0.8284779191017151, "learning_rate": 0.00016368051807879114, "loss": 3.2212, "step": 67500 }, { "epoch": 7.283018867924528, "grad_norm": 0.8559075593948364, "learning_rate": 0.0001633567188343227, "loss": 3.1971, "step": 67550 }, { "epoch": 7.288409703504043, "grad_norm": 0.7753306031227112, "learning_rate": 0.00016303291958985426, "loss": 3.2145, "step": 67600 }, { "epoch": 7.293800539083558, "grad_norm": 0.7762905359268188, "learning_rate": 0.00016270912034538584, "loss": 3.2352, "step": 67650 }, { "epoch": 7.2991913746630726, "grad_norm": 0.7790316343307495, "learning_rate": 0.00016238532110091742, "loss": 3.1933, "step": 67700 }, { "epoch": 7.304582210242588, "grad_norm": 0.8162766695022583, "learning_rate": 0.000162061521856449, "loss": 3.2162, "step": 67750 }, { "epoch": 7.309973045822103, "grad_norm": 0.7991843223571777, "learning_rate": 0.00016173772261198055, "loss": 3.2277, "step": 67800 }, { "epoch": 7.315363881401617, "grad_norm": 0.777722954750061, "learning_rate": 0.00016141392336751212, "loss": 3.2262, "step": 67850 }, { "epoch": 7.320754716981132, "grad_norm": 0.7924916744232178, "learning_rate": 0.0001610901241230437, "loss": 3.221, "step": 67900 }, { "epoch": 7.3261455525606465, "grad_norm": 0.8169491291046143, "learning_rate": 0.00016076632487857528, "loss": 3.2437, "step": 67950 }, { "epoch": 7.331536388140162, "grad_norm": 0.7886592149734497, "learning_rate": 0.00016044252563410686, "loss": 3.2273, "step": 68000 }, { "epoch": 7.331536388140162, "eval_accuracy": 0.3883553005931035, "eval_loss": 3.351078987121582, "eval_runtime": 185.9442, "eval_samples_per_second": 96.862, "eval_steps_per_second": 6.056, "step": 68000 }, { "epoch": 7.336927223719677, "grad_norm": 0.8167215585708618, "learning_rate": 0.0001601187263896384, "loss": 3.2205, "step": 68050 }, { "epoch": 7.342318059299191, "grad_norm": 0.7827270030975342, "learning_rate": 0.00015979492714516996, "loss": 3.213, "step": 68100 }, { "epoch": 7.347708894878706, "grad_norm": 0.8101873397827148, "learning_rate": 0.00015947112790070154, "loss": 3.221, "step": 68150 }, { "epoch": 7.353099730458221, "grad_norm": 0.7950315475463867, "learning_rate": 0.00015914732865623311, "loss": 3.2326, "step": 68200 }, { "epoch": 7.3584905660377355, "grad_norm": 0.7943675518035889, "learning_rate": 0.0001588235294117647, "loss": 3.2038, "step": 68250 }, { "epoch": 7.363881401617251, "grad_norm": 0.7641896605491638, "learning_rate": 0.00015849973016729627, "loss": 3.2036, "step": 68300 }, { "epoch": 7.369272237196766, "grad_norm": 0.8040649890899658, "learning_rate": 0.00015817593092282785, "loss": 3.2418, "step": 68350 }, { "epoch": 7.37466307277628, "grad_norm": 0.8314372897148132, "learning_rate": 0.00015785213167835942, "loss": 3.2144, "step": 68400 }, { "epoch": 7.380053908355795, "grad_norm": 0.8147042393684387, "learning_rate": 0.00015752833243389095, "loss": 3.219, "step": 68450 }, { "epoch": 7.38544474393531, "grad_norm": 0.767492413520813, "learning_rate": 0.00015720453318942253, "loss": 3.2212, "step": 68500 }, { "epoch": 7.390835579514825, "grad_norm": 0.7924861311912537, "learning_rate": 0.0001568807339449541, "loss": 3.2114, "step": 68550 }, { "epoch": 7.39622641509434, "grad_norm": 0.8257951736450195, "learning_rate": 0.00015655693470048568, "loss": 3.228, "step": 68600 }, { "epoch": 7.401617250673855, "grad_norm": 0.7666079998016357, "learning_rate": 0.00015623313545601726, "loss": 3.2288, "step": 68650 }, { "epoch": 7.407008086253369, "grad_norm": 0.8601511716842651, "learning_rate": 0.00015590933621154884, "loss": 3.239, "step": 68700 }, { "epoch": 7.412398921832884, "grad_norm": 0.7925523519515991, "learning_rate": 0.00015558553696708041, "loss": 3.2177, "step": 68750 }, { "epoch": 7.4177897574123985, "grad_norm": 0.818068265914917, "learning_rate": 0.00015526173772261196, "loss": 3.2129, "step": 68800 }, { "epoch": 7.423180592991914, "grad_norm": 0.8102258443832397, "learning_rate": 0.00015493793847814354, "loss": 3.2238, "step": 68850 }, { "epoch": 7.428571428571429, "grad_norm": 0.8156552314758301, "learning_rate": 0.0001546141392336751, "loss": 3.2322, "step": 68900 }, { "epoch": 7.433962264150943, "grad_norm": 0.8134991526603699, "learning_rate": 0.00015429033998920667, "loss": 3.2442, "step": 68950 }, { "epoch": 7.439353099730458, "grad_norm": 0.8857681155204773, "learning_rate": 0.00015397301672962762, "loss": 3.2405, "step": 69000 }, { "epoch": 7.439353099730458, "eval_accuracy": 0.3885801034179917, "eval_loss": 3.3481028079986572, "eval_runtime": 184.4596, "eval_samples_per_second": 97.642, "eval_steps_per_second": 6.104, "step": 69000 }, { "epoch": 7.444743935309973, "grad_norm": 0.8334435224533081, "learning_rate": 0.0001536492174851592, "loss": 3.2515, "step": 69050 }, { "epoch": 7.450134770889488, "grad_norm": 0.8292680978775024, "learning_rate": 0.00015332541824069078, "loss": 3.2223, "step": 69100 }, { "epoch": 7.455525606469003, "grad_norm": 0.819562554359436, "learning_rate": 0.00015300161899622233, "loss": 3.2417, "step": 69150 }, { "epoch": 7.460916442048518, "grad_norm": 0.788324236869812, "learning_rate": 0.00015267781975175388, "loss": 3.2305, "step": 69200 }, { "epoch": 7.466307277628032, "grad_norm": 0.8466244339942932, "learning_rate": 0.00015235402050728546, "loss": 3.2109, "step": 69250 }, { "epoch": 7.471698113207547, "grad_norm": 0.7949193716049194, "learning_rate": 0.00015203022126281703, "loss": 3.2201, "step": 69300 }, { "epoch": 7.4770889487870615, "grad_norm": 0.7899576425552368, "learning_rate": 0.0001517064220183486, "loss": 3.2475, "step": 69350 }, { "epoch": 7.482479784366577, "grad_norm": 0.7817439436912537, "learning_rate": 0.0001513826227738802, "loss": 3.2138, "step": 69400 }, { "epoch": 7.487870619946092, "grad_norm": 0.9032320380210876, "learning_rate": 0.00015105882352941177, "loss": 3.2157, "step": 69450 }, { "epoch": 7.493261455525606, "grad_norm": 0.7862300276756287, "learning_rate": 0.00015073502428494332, "loss": 3.2233, "step": 69500 }, { "epoch": 7.498652291105121, "grad_norm": 0.8258659839630127, "learning_rate": 0.0001504112250404749, "loss": 3.2201, "step": 69550 }, { "epoch": 7.504043126684636, "grad_norm": 0.7886598110198975, "learning_rate": 0.00015008742579600647, "loss": 3.2187, "step": 69600 }, { "epoch": 7.509433962264151, "grad_norm": 0.8414490222930908, "learning_rate": 0.00014976362655153802, "loss": 3.2265, "step": 69650 }, { "epoch": 7.514824797843666, "grad_norm": 0.7889809012413025, "learning_rate": 0.0001494398273070696, "loss": 3.23, "step": 69700 }, { "epoch": 7.520215633423181, "grad_norm": 0.8091080188751221, "learning_rate": 0.00014911602806260118, "loss": 3.2209, "step": 69750 }, { "epoch": 7.525606469002695, "grad_norm": 0.8536641001701355, "learning_rate": 0.00014879222881813273, "loss": 3.2336, "step": 69800 }, { "epoch": 7.53099730458221, "grad_norm": 0.8249064683914185, "learning_rate": 0.0001484684295736643, "loss": 3.2296, "step": 69850 }, { "epoch": 7.536388140161725, "grad_norm": 0.7949578762054443, "learning_rate": 0.00014814463032919589, "loss": 3.2286, "step": 69900 }, { "epoch": 7.54177897574124, "grad_norm": 0.9103390574455261, "learning_rate": 0.00014782083108472744, "loss": 3.2317, "step": 69950 }, { "epoch": 7.547169811320755, "grad_norm": 0.8113278150558472, "learning_rate": 0.00014749703184025901, "loss": 3.2308, "step": 70000 }, { "epoch": 7.547169811320755, "eval_accuracy": 0.38873004440317566, "eval_loss": 3.3429155349731445, "eval_runtime": 184.234, "eval_samples_per_second": 97.762, "eval_steps_per_second": 6.112, "step": 70000 }, { "epoch": 7.55256064690027, "grad_norm": 0.8181444406509399, "learning_rate": 0.0001471732325957906, "loss": 3.2208, "step": 70050 }, { "epoch": 7.557951482479784, "grad_norm": 0.8472036123275757, "learning_rate": 0.00014684943335132217, "loss": 3.2312, "step": 70100 }, { "epoch": 7.563342318059299, "grad_norm": 0.8449604511260986, "learning_rate": 0.00014652563410685375, "loss": 3.2106, "step": 70150 }, { "epoch": 7.568733153638814, "grad_norm": 0.8788800239562988, "learning_rate": 0.0001462018348623853, "loss": 3.219, "step": 70200 }, { "epoch": 7.574123989218329, "grad_norm": 0.7950506806373596, "learning_rate": 0.00014587803561791687, "loss": 3.2318, "step": 70250 }, { "epoch": 7.579514824797844, "grad_norm": 0.7994862794876099, "learning_rate": 0.00014555423637344845, "loss": 3.2356, "step": 70300 }, { "epoch": 7.584905660377358, "grad_norm": 0.8473143577575684, "learning_rate": 0.00014523043712898003, "loss": 3.2211, "step": 70350 }, { "epoch": 7.590296495956873, "grad_norm": 0.7861645817756653, "learning_rate": 0.00014490663788451158, "loss": 3.2259, "step": 70400 }, { "epoch": 7.595687331536388, "grad_norm": 0.81363445520401, "learning_rate": 0.00014458283864004316, "loss": 3.2253, "step": 70450 }, { "epoch": 7.601078167115903, "grad_norm": 0.8435284495353699, "learning_rate": 0.00014425903939557474, "loss": 3.226, "step": 70500 }, { "epoch": 7.606469002695418, "grad_norm": 0.8990793824195862, "learning_rate": 0.00014393524015110631, "loss": 3.2115, "step": 70550 }, { "epoch": 7.611859838274933, "grad_norm": 0.8239708542823792, "learning_rate": 0.00014361144090663786, "loss": 3.2359, "step": 70600 }, { "epoch": 7.617250673854447, "grad_norm": 0.9180291891098022, "learning_rate": 0.00014328764166216944, "loss": 3.2315, "step": 70650 }, { "epoch": 7.622641509433962, "grad_norm": 0.8144366145133972, "learning_rate": 0.00014296384241770102, "loss": 3.2421, "step": 70700 }, { "epoch": 7.628032345013477, "grad_norm": 0.8614203929901123, "learning_rate": 0.0001426400431732326, "loss": 3.2179, "step": 70750 }, { "epoch": 7.633423180592992, "grad_norm": 0.772205114364624, "learning_rate": 0.00014231624392876417, "loss": 3.2513, "step": 70800 }, { "epoch": 7.638814016172507, "grad_norm": 0.804109513759613, "learning_rate": 0.00014199244468429573, "loss": 3.2094, "step": 70850 }, { "epoch": 7.644204851752022, "grad_norm": 0.8096456527709961, "learning_rate": 0.0001416686454398273, "loss": 3.2403, "step": 70900 }, { "epoch": 7.649595687331536, "grad_norm": 0.8947812914848328, "learning_rate": 0.00014134484619535888, "loss": 3.228, "step": 70950 }, { "epoch": 7.654986522911051, "grad_norm": 0.8135775327682495, "learning_rate": 0.00014102104695089043, "loss": 3.2202, "step": 71000 }, { "epoch": 7.654986522911051, "eval_accuracy": 0.3893292650794725, "eval_loss": 3.3402256965637207, "eval_runtime": 183.9805, "eval_samples_per_second": 97.896, "eval_steps_per_second": 6.12, "step": 71000 }, { "epoch": 7.660377358490566, "grad_norm": 0.8074039816856384, "learning_rate": 0.000140697247706422, "loss": 3.2351, "step": 71050 }, { "epoch": 7.665768194070081, "grad_norm": 0.8246932029724121, "learning_rate": 0.0001403734484619536, "loss": 3.2318, "step": 71100 }, { "epoch": 7.671159029649596, "grad_norm": 0.8261672854423523, "learning_rate": 0.00014004964921748514, "loss": 3.2233, "step": 71150 }, { "epoch": 7.67654986522911, "grad_norm": 0.8752914071083069, "learning_rate": 0.00013972584997301671, "loss": 3.2371, "step": 71200 }, { "epoch": 7.681940700808625, "grad_norm": 0.7831204533576965, "learning_rate": 0.0001394020507285483, "loss": 3.2334, "step": 71250 }, { "epoch": 7.6873315363881405, "grad_norm": 0.8146886229515076, "learning_rate": 0.00013907825148407984, "loss": 3.2139, "step": 71300 }, { "epoch": 7.692722371967655, "grad_norm": 0.8081901669502258, "learning_rate": 0.00013875445223961142, "loss": 3.2364, "step": 71350 }, { "epoch": 7.69811320754717, "grad_norm": 0.9263750314712524, "learning_rate": 0.000138430652995143, "loss": 3.2265, "step": 71400 }, { "epoch": 7.703504043126685, "grad_norm": 0.8330522179603577, "learning_rate": 0.00013810685375067455, "loss": 3.223, "step": 71450 }, { "epoch": 7.708894878706199, "grad_norm": 0.8679282665252686, "learning_rate": 0.00013778305450620613, "loss": 3.2363, "step": 71500 }, { "epoch": 7.714285714285714, "grad_norm": 0.8539426326751709, "learning_rate": 0.0001374592552617377, "loss": 3.2308, "step": 71550 }, { "epoch": 7.719676549865229, "grad_norm": 0.8268200755119324, "learning_rate": 0.00013713545601726928, "loss": 3.2322, "step": 71600 }, { "epoch": 7.725067385444744, "grad_norm": 0.7831496000289917, "learning_rate": 0.00013681165677280086, "loss": 3.2434, "step": 71650 }, { "epoch": 7.730458221024259, "grad_norm": 0.7843145728111267, "learning_rate": 0.0001364878575283324, "loss": 3.223, "step": 71700 }, { "epoch": 7.735849056603773, "grad_norm": 0.7985066771507263, "learning_rate": 0.000136164058283864, "loss": 3.2263, "step": 71750 }, { "epoch": 7.741239892183288, "grad_norm": 0.8175166249275208, "learning_rate": 0.00013584025903939557, "loss": 3.2276, "step": 71800 }, { "epoch": 7.7466307277628035, "grad_norm": 0.8650901913642883, "learning_rate": 0.00013551645979492714, "loss": 3.2336, "step": 71850 }, { "epoch": 7.752021563342318, "grad_norm": 0.829919159412384, "learning_rate": 0.00013519913653534807, "loss": 3.2258, "step": 71900 }, { "epoch": 7.757412398921833, "grad_norm": 0.791461706161499, "learning_rate": 0.00013487533729087965, "loss": 3.2179, "step": 71950 }, { "epoch": 7.762803234501348, "grad_norm": 0.8111125826835632, "learning_rate": 0.0001345515380464112, "loss": 3.2477, "step": 72000 }, { "epoch": 7.762803234501348, "eval_accuracy": 0.38987426796474994, "eval_loss": 3.3354620933532715, "eval_runtime": 183.9298, "eval_samples_per_second": 97.923, "eval_steps_per_second": 6.122, "step": 72000 }, { "epoch": 7.768194070080862, "grad_norm": 0.8132084012031555, "learning_rate": 0.00013422773880194277, "loss": 3.2368, "step": 72050 }, { "epoch": 7.773584905660377, "grad_norm": 0.800094723701477, "learning_rate": 0.00013390393955747435, "loss": 3.2144, "step": 72100 }, { "epoch": 7.7789757412398925, "grad_norm": 0.8695988655090332, "learning_rate": 0.00013358014031300593, "loss": 3.2305, "step": 72150 }, { "epoch": 7.784366576819407, "grad_norm": 0.8408200740814209, "learning_rate": 0.0001332563410685375, "loss": 3.2296, "step": 72200 }, { "epoch": 7.789757412398922, "grad_norm": 0.834050714969635, "learning_rate": 0.00013293254182406906, "loss": 3.234, "step": 72250 }, { "epoch": 7.795148247978437, "grad_norm": 0.7629789710044861, "learning_rate": 0.00013260874257960064, "loss": 3.2245, "step": 72300 }, { "epoch": 7.800539083557951, "grad_norm": 0.8096799254417419, "learning_rate": 0.0001322849433351322, "loss": 3.2308, "step": 72350 }, { "epoch": 7.8059299191374665, "grad_norm": 0.8130748867988586, "learning_rate": 0.0001319611440906638, "loss": 3.2271, "step": 72400 }, { "epoch": 7.811320754716981, "grad_norm": 0.853058934211731, "learning_rate": 0.00013163734484619534, "loss": 3.237, "step": 72450 }, { "epoch": 7.816711590296496, "grad_norm": 0.831373393535614, "learning_rate": 0.00013131354560172692, "loss": 3.2437, "step": 72500 }, { "epoch": 7.822102425876011, "grad_norm": 0.8419806957244873, "learning_rate": 0.0001309897463572585, "loss": 3.2297, "step": 72550 }, { "epoch": 7.827493261455525, "grad_norm": 0.8754237294197083, "learning_rate": 0.00013066594711279007, "loss": 3.2413, "step": 72600 }, { "epoch": 7.83288409703504, "grad_norm": 0.8107507824897766, "learning_rate": 0.00013034214786832162, "loss": 3.2333, "step": 72650 }, { "epoch": 7.8382749326145555, "grad_norm": 0.8244099617004395, "learning_rate": 0.0001300183486238532, "loss": 3.2451, "step": 72700 }, { "epoch": 7.84366576819407, "grad_norm": 0.8542118072509766, "learning_rate": 0.00012969454937938478, "loss": 3.2322, "step": 72750 }, { "epoch": 7.849056603773585, "grad_norm": 0.8142945766448975, "learning_rate": 0.00012937075013491636, "loss": 3.233, "step": 72800 }, { "epoch": 7.8544474393531, "grad_norm": 0.7852938771247864, "learning_rate": 0.0001290469508904479, "loss": 3.2352, "step": 72850 }, { "epoch": 7.859838274932614, "grad_norm": 0.8006114959716797, "learning_rate": 0.00012872315164597949, "loss": 3.229, "step": 72900 }, { "epoch": 7.8652291105121295, "grad_norm": 0.8257787823677063, "learning_rate": 0.00012839935240151106, "loss": 3.2434, "step": 72950 }, { "epoch": 7.870619946091644, "grad_norm": 0.8480833768844604, "learning_rate": 0.00012807555315704261, "loss": 3.2472, "step": 73000 }, { "epoch": 7.870619946091644, "eval_accuracy": 0.39002583874325114, "eval_loss": 3.333691120147705, "eval_runtime": 183.9046, "eval_samples_per_second": 97.937, "eval_steps_per_second": 6.123, "step": 73000 }, { "epoch": 7.876010781671159, "grad_norm": 0.9012708067893982, "learning_rate": 0.0001277517539125742, "loss": 3.2418, "step": 73050 }, { "epoch": 7.881401617250674, "grad_norm": 0.8225118517875671, "learning_rate": 0.00012742795466810577, "loss": 3.2268, "step": 73100 }, { "epoch": 7.886792452830189, "grad_norm": 0.8552934527397156, "learning_rate": 0.00012710415542363732, "loss": 3.2262, "step": 73150 }, { "epoch": 7.892183288409703, "grad_norm": 0.8058131337165833, "learning_rate": 0.0001267803561791689, "loss": 3.2375, "step": 73200 }, { "epoch": 7.8975741239892185, "grad_norm": 0.8629428148269653, "learning_rate": 0.00012645655693470048, "loss": 3.2325, "step": 73250 }, { "epoch": 7.902964959568733, "grad_norm": 0.9265134334564209, "learning_rate": 0.00012613275769023203, "loss": 3.2333, "step": 73300 }, { "epoch": 7.908355795148248, "grad_norm": 0.8891592621803284, "learning_rate": 0.0001258089584457636, "loss": 3.2321, "step": 73350 }, { "epoch": 7.913746630727763, "grad_norm": 0.8467879295349121, "learning_rate": 0.00012548515920129518, "loss": 3.2317, "step": 73400 }, { "epoch": 7.919137466307277, "grad_norm": 0.8680068254470825, "learning_rate": 0.00012516135995682676, "loss": 3.231, "step": 73450 }, { "epoch": 7.9245283018867925, "grad_norm": 0.8347758650779724, "learning_rate": 0.0001248375607123583, "loss": 3.2237, "step": 73500 }, { "epoch": 7.929919137466308, "grad_norm": 0.82417893409729, "learning_rate": 0.0001245137614678899, "loss": 3.2242, "step": 73550 }, { "epoch": 7.935309973045822, "grad_norm": 0.8774154782295227, "learning_rate": 0.00012418996222342147, "loss": 3.225, "step": 73600 }, { "epoch": 7.940700808625337, "grad_norm": 0.8639497756958008, "learning_rate": 0.00012386616297895304, "loss": 3.2485, "step": 73650 }, { "epoch": 7.946091644204852, "grad_norm": 0.8347174525260925, "learning_rate": 0.0001235423637344846, "loss": 3.242, "step": 73700 }, { "epoch": 7.951482479784366, "grad_norm": 0.8348175883293152, "learning_rate": 0.00012321856449001617, "loss": 3.2229, "step": 73750 }, { "epoch": 7.9568733153638815, "grad_norm": 0.8157823085784912, "learning_rate": 0.00012289476524554775, "loss": 3.2208, "step": 73800 }, { "epoch": 7.962264150943396, "grad_norm": 0.7920084595680237, "learning_rate": 0.00012257096600107933, "loss": 3.239, "step": 73850 }, { "epoch": 7.967654986522911, "grad_norm": 0.8304669260978699, "learning_rate": 0.0001222471667566109, "loss": 3.231, "step": 73900 }, { "epoch": 7.973045822102426, "grad_norm": 0.8345317840576172, "learning_rate": 0.00012192336751214245, "loss": 3.2325, "step": 73950 }, { "epoch": 7.97843665768194, "grad_norm": 0.8521400094032288, "learning_rate": 0.00012159956826767403, "loss": 3.2422, "step": 74000 }, { "epoch": 7.97843665768194, "eval_accuracy": 0.3904562128318843, "eval_loss": 3.3273274898529053, "eval_runtime": 184.3767, "eval_samples_per_second": 97.686, "eval_steps_per_second": 6.107, "step": 74000 }, { "epoch": 7.9838274932614555, "grad_norm": 0.8273377418518066, "learning_rate": 0.00012127576902320561, "loss": 3.229, "step": 74050 }, { "epoch": 7.989218328840971, "grad_norm": 0.83464515209198, "learning_rate": 0.00012095196977873717, "loss": 3.2182, "step": 74100 }, { "epoch": 7.994609164420485, "grad_norm": 0.8124722838401794, "learning_rate": 0.00012062817053426874, "loss": 3.2395, "step": 74150 }, { "epoch": 8.0, "grad_norm": 1.7434592247009277, "learning_rate": 0.00012030437128980032, "loss": 3.2201, "step": 74200 }, { "epoch": 8.005390835579515, "grad_norm": 0.8386273980140686, "learning_rate": 0.00011998057204533188, "loss": 3.1444, "step": 74250 }, { "epoch": 8.01078167115903, "grad_norm": 0.8330357074737549, "learning_rate": 0.00011965677280086346, "loss": 3.1653, "step": 74300 }, { "epoch": 8.016172506738544, "grad_norm": 0.7729629278182983, "learning_rate": 0.00011933297355639502, "loss": 3.1518, "step": 74350 }, { "epoch": 8.021563342318059, "grad_norm": 0.8658410906791687, "learning_rate": 0.00011900917431192659, "loss": 3.1664, "step": 74400 }, { "epoch": 8.026954177897574, "grad_norm": 0.8363280296325684, "learning_rate": 0.00011868537506745816, "loss": 3.1571, "step": 74450 }, { "epoch": 8.032345013477089, "grad_norm": 0.8109269738197327, "learning_rate": 0.00011836157582298974, "loss": 3.1505, "step": 74500 }, { "epoch": 8.037735849056604, "grad_norm": 0.8341225981712341, "learning_rate": 0.00011803777657852132, "loss": 3.1531, "step": 74550 }, { "epoch": 8.04312668463612, "grad_norm": 0.8043491244316101, "learning_rate": 0.00011771397733405287, "loss": 3.1557, "step": 74600 }, { "epoch": 8.048517520215633, "grad_norm": 0.8187711238861084, "learning_rate": 0.00011739017808958445, "loss": 3.1508, "step": 74650 }, { "epoch": 8.053908355795148, "grad_norm": 0.8134503960609436, "learning_rate": 0.00011706637884511602, "loss": 3.1506, "step": 74700 }, { "epoch": 8.059299191374663, "grad_norm": 0.8754953145980835, "learning_rate": 0.0001167425796006476, "loss": 3.1456, "step": 74750 }, { "epoch": 8.064690026954178, "grad_norm": 0.8755296468734741, "learning_rate": 0.00011641878035617915, "loss": 3.1674, "step": 74800 }, { "epoch": 8.070080862533693, "grad_norm": 0.8251467347145081, "learning_rate": 0.00011609498111171073, "loss": 3.1531, "step": 74850 }, { "epoch": 8.075471698113208, "grad_norm": 0.8318637013435364, "learning_rate": 0.00011577118186724231, "loss": 3.1546, "step": 74900 }, { "epoch": 8.080862533692722, "grad_norm": 0.8319134712219238, "learning_rate": 0.00011544738262277387, "loss": 3.1566, "step": 74950 }, { "epoch": 8.086253369272237, "grad_norm": 0.8469785451889038, "learning_rate": 0.00011512358337830544, "loss": 3.1478, "step": 75000 }, { "epoch": 8.086253369272237, "eval_accuracy": 0.39015133282867687, "eval_loss": 3.333881139755249, "eval_runtime": 183.6708, "eval_samples_per_second": 98.061, "eval_steps_per_second": 6.131, "step": 75000 }, { "epoch": 8.091644204851752, "grad_norm": 0.8197541832923889, "learning_rate": 0.00011479978413383701, "loss": 3.1571, "step": 75050 }, { "epoch": 8.097035040431267, "grad_norm": 0.8593742251396179, "learning_rate": 0.00011447598488936858, "loss": 3.1667, "step": 75100 }, { "epoch": 8.102425876010782, "grad_norm": 0.8601323962211609, "learning_rate": 0.00011415218564490016, "loss": 3.1579, "step": 75150 }, { "epoch": 8.107816711590296, "grad_norm": 0.8454037308692932, "learning_rate": 0.00011382838640043172, "loss": 3.1665, "step": 75200 }, { "epoch": 8.11320754716981, "grad_norm": 0.837957501411438, "learning_rate": 0.00011350458715596328, "loss": 3.1702, "step": 75250 }, { "epoch": 8.118598382749326, "grad_norm": 0.8624327182769775, "learning_rate": 0.00011318078791149486, "loss": 3.1602, "step": 75300 }, { "epoch": 8.123989218328841, "grad_norm": 0.8311221599578857, "learning_rate": 0.00011285698866702644, "loss": 3.171, "step": 75350 }, { "epoch": 8.129380053908356, "grad_norm": 0.8376172780990601, "learning_rate": 0.00011253318942255802, "loss": 3.1492, "step": 75400 }, { "epoch": 8.134770889487871, "grad_norm": 0.8223761916160583, "learning_rate": 0.00011220939017808957, "loss": 3.1569, "step": 75450 }, { "epoch": 8.140161725067385, "grad_norm": 0.8498263359069824, "learning_rate": 0.00011188559093362115, "loss": 3.1573, "step": 75500 }, { "epoch": 8.1455525606469, "grad_norm": 0.8648282289505005, "learning_rate": 0.00011156179168915272, "loss": 3.1564, "step": 75550 }, { "epoch": 8.150943396226415, "grad_norm": 0.8844196200370789, "learning_rate": 0.00011123799244468429, "loss": 3.1778, "step": 75600 }, { "epoch": 8.15633423180593, "grad_norm": 0.8654659390449524, "learning_rate": 0.00011091419320021585, "loss": 3.1603, "step": 75650 }, { "epoch": 8.161725067385445, "grad_norm": 0.8266019821166992, "learning_rate": 0.00011059039395574743, "loss": 3.1665, "step": 75700 }, { "epoch": 8.167115902964959, "grad_norm": 0.8561160564422607, "learning_rate": 0.00011026659471127899, "loss": 3.1696, "step": 75750 }, { "epoch": 8.172506738544474, "grad_norm": 0.8629851937294006, "learning_rate": 0.00010994279546681057, "loss": 3.1561, "step": 75800 }, { "epoch": 8.177897574123989, "grad_norm": 0.8809854388237, "learning_rate": 0.00010961899622234213, "loss": 3.1695, "step": 75850 }, { "epoch": 8.183288409703504, "grad_norm": 0.8539608716964722, "learning_rate": 0.00010930167296276309, "loss": 3.148, "step": 75900 }, { "epoch": 8.18867924528302, "grad_norm": 0.8707504272460938, "learning_rate": 0.00010897787371829465, "loss": 3.1776, "step": 75950 }, { "epoch": 8.194070080862534, "grad_norm": 0.8520840406417847, "learning_rate": 0.00010866055045871559, "loss": 3.1733, "step": 76000 }, { "epoch": 8.194070080862534, "eval_accuracy": 0.39060810956904896, "eval_loss": 3.3335835933685303, "eval_runtime": 184.0123, "eval_samples_per_second": 97.879, "eval_steps_per_second": 6.119, "step": 76000 }, { "epoch": 8.199460916442048, "grad_norm": 0.8671824932098389, "learning_rate": 0.00010833675121424717, "loss": 3.1841, "step": 76050 }, { "epoch": 8.204851752021563, "grad_norm": 0.8446840643882751, "learning_rate": 0.00010801295196977872, "loss": 3.1579, "step": 76100 }, { "epoch": 8.210242587601078, "grad_norm": 0.8217098116874695, "learning_rate": 0.0001076891527253103, "loss": 3.1861, "step": 76150 }, { "epoch": 8.215633423180593, "grad_norm": 0.8734848499298096, "learning_rate": 0.00010736535348084187, "loss": 3.17, "step": 76200 }, { "epoch": 8.221024258760108, "grad_norm": 0.8050230741500854, "learning_rate": 0.00010704155423637345, "loss": 3.164, "step": 76250 }, { "epoch": 8.226415094339623, "grad_norm": 0.826291024684906, "learning_rate": 0.000106717754991905, "loss": 3.1588, "step": 76300 }, { "epoch": 8.231805929919137, "grad_norm": 0.8282190561294556, "learning_rate": 0.00010639395574743658, "loss": 3.1784, "step": 76350 }, { "epoch": 8.237196765498652, "grad_norm": 0.846244215965271, "learning_rate": 0.00010607015650296816, "loss": 3.1863, "step": 76400 }, { "epoch": 8.242587601078167, "grad_norm": 0.8849758505821228, "learning_rate": 0.00010574635725849972, "loss": 3.1791, "step": 76450 }, { "epoch": 8.247978436657682, "grad_norm": 0.8456253409385681, "learning_rate": 0.00010542255801403128, "loss": 3.1703, "step": 76500 }, { "epoch": 8.253369272237197, "grad_norm": 0.8574989438056946, "learning_rate": 0.00010509875876956286, "loss": 3.1736, "step": 76550 }, { "epoch": 8.25876010781671, "grad_norm": 0.8383358120918274, "learning_rate": 0.00010477495952509443, "loss": 3.1941, "step": 76600 }, { "epoch": 8.264150943396226, "grad_norm": 0.8484417796134949, "learning_rate": 0.000104451160280626, "loss": 3.1728, "step": 76650 }, { "epoch": 8.269541778975741, "grad_norm": 0.8441385626792908, "learning_rate": 0.00010412736103615758, "loss": 3.1828, "step": 76700 }, { "epoch": 8.274932614555256, "grad_norm": 0.8040993213653564, "learning_rate": 0.00010380356179168913, "loss": 3.1683, "step": 76750 }, { "epoch": 8.280323450134771, "grad_norm": 0.841343343257904, "learning_rate": 0.00010347976254722071, "loss": 3.1808, "step": 76800 }, { "epoch": 8.285714285714286, "grad_norm": 0.8964943289756775, "learning_rate": 0.00010315596330275229, "loss": 3.1784, "step": 76850 }, { "epoch": 8.2911051212938, "grad_norm": 0.8301113247871399, "learning_rate": 0.00010283216405828387, "loss": 3.158, "step": 76900 }, { "epoch": 8.296495956873315, "grad_norm": 0.8281328082084656, "learning_rate": 0.00010250836481381542, "loss": 3.1641, "step": 76950 }, { "epoch": 8.30188679245283, "grad_norm": 0.8478591442108154, "learning_rate": 0.000102184565569347, "loss": 3.1655, "step": 77000 }, { "epoch": 8.30188679245283, "eval_accuracy": 0.390922007761945, "eval_loss": 3.3310768604278564, "eval_runtime": 184.6299, "eval_samples_per_second": 97.552, "eval_steps_per_second": 6.099, "step": 77000 }, { "epoch": 8.307277628032345, "grad_norm": 0.845905065536499, "learning_rate": 0.00010186076632487857, "loss": 3.1771, "step": 77050 }, { "epoch": 8.31266846361186, "grad_norm": 0.8499204516410828, "learning_rate": 0.00010153696708041014, "loss": 3.1876, "step": 77100 }, { "epoch": 8.318059299191376, "grad_norm": 0.8497974872589111, "learning_rate": 0.0001012131678359417, "loss": 3.1632, "step": 77150 }, { "epoch": 8.323450134770889, "grad_norm": 0.8825196027755737, "learning_rate": 0.00010088936859147328, "loss": 3.1681, "step": 77200 }, { "epoch": 8.328840970350404, "grad_norm": 0.8770312070846558, "learning_rate": 0.00010056556934700485, "loss": 3.1855, "step": 77250 }, { "epoch": 8.33423180592992, "grad_norm": 0.8177728652954102, "learning_rate": 0.00010024177010253642, "loss": 3.1692, "step": 77300 }, { "epoch": 8.339622641509434, "grad_norm": 0.8661665320396423, "learning_rate": 9.9917970858068e-05, "loss": 3.2012, "step": 77350 }, { "epoch": 8.34501347708895, "grad_norm": 0.8462918996810913, "learning_rate": 9.959417161359956e-05, "loss": 3.1861, "step": 77400 }, { "epoch": 8.350404312668463, "grad_norm": 0.8591238260269165, "learning_rate": 9.927037236913112e-05, "loss": 3.1845, "step": 77450 }, { "epoch": 8.355795148247978, "grad_norm": 0.9080222249031067, "learning_rate": 9.89465731246627e-05, "loss": 3.1787, "step": 77500 }, { "epoch": 8.361185983827493, "grad_norm": 0.8710142970085144, "learning_rate": 9.862277388019428e-05, "loss": 3.1737, "step": 77550 }, { "epoch": 8.366576819407008, "grad_norm": 0.8547530770301819, "learning_rate": 9.829897463572583e-05, "loss": 3.1841, "step": 77600 }, { "epoch": 8.371967654986523, "grad_norm": 0.828269362449646, "learning_rate": 9.797517539125741e-05, "loss": 3.1937, "step": 77650 }, { "epoch": 8.377358490566039, "grad_norm": 0.8966538310050964, "learning_rate": 9.765137614678899e-05, "loss": 3.1819, "step": 77700 }, { "epoch": 8.382749326145552, "grad_norm": 0.8636840581893921, "learning_rate": 9.732757690232056e-05, "loss": 3.1875, "step": 77750 }, { "epoch": 8.388140161725067, "grad_norm": 0.85294109582901, "learning_rate": 9.700377765785211e-05, "loss": 3.1726, "step": 77800 }, { "epoch": 8.393530997304582, "grad_norm": 0.8370219469070435, "learning_rate": 9.667997841338369e-05, "loss": 3.1848, "step": 77850 }, { "epoch": 8.398921832884097, "grad_norm": 0.8667328953742981, "learning_rate": 9.635617916891527e-05, "loss": 3.1749, "step": 77900 }, { "epoch": 8.404312668463612, "grad_norm": 0.9079992771148682, "learning_rate": 9.603237992444683e-05, "loss": 3.1722, "step": 77950 }, { "epoch": 8.409703504043126, "grad_norm": 0.831434428691864, "learning_rate": 9.57085806799784e-05, "loss": 3.16, "step": 78000 }, { "epoch": 8.409703504043126, "eval_accuracy": 0.39106336516899165, "eval_loss": 3.3274149894714355, "eval_runtime": 185.0572, "eval_samples_per_second": 97.327, "eval_steps_per_second": 6.085, "step": 78000 }, { "epoch": 8.415094339622641, "grad_norm": 0.8659220933914185, "learning_rate": 9.538478143550998e-05, "loss": 3.1835, "step": 78050 }, { "epoch": 8.420485175202156, "grad_norm": 0.8522933125495911, "learning_rate": 9.506098219104154e-05, "loss": 3.1773, "step": 78100 }, { "epoch": 8.425876010781671, "grad_norm": 0.9333472847938538, "learning_rate": 9.473718294657312e-05, "loss": 3.1707, "step": 78150 }, { "epoch": 8.431266846361186, "grad_norm": 0.849563717842102, "learning_rate": 9.44133837021047e-05, "loss": 3.18, "step": 78200 }, { "epoch": 8.436657681940702, "grad_norm": 0.8461838364601135, "learning_rate": 9.408958445763625e-05, "loss": 3.1864, "step": 78250 }, { "epoch": 8.442048517520215, "grad_norm": 0.8481196165084839, "learning_rate": 9.376578521316782e-05, "loss": 3.1801, "step": 78300 }, { "epoch": 8.44743935309973, "grad_norm": 0.8493364453315735, "learning_rate": 9.34419859686994e-05, "loss": 3.1697, "step": 78350 }, { "epoch": 8.452830188679245, "grad_norm": 0.8151680827140808, "learning_rate": 9.311818672423098e-05, "loss": 3.187, "step": 78400 }, { "epoch": 8.45822102425876, "grad_norm": 0.8653626441955566, "learning_rate": 9.279438747976253e-05, "loss": 3.1659, "step": 78450 }, { "epoch": 8.463611859838275, "grad_norm": 0.8948111534118652, "learning_rate": 9.247058823529411e-05, "loss": 3.1754, "step": 78500 }, { "epoch": 8.46900269541779, "grad_norm": 0.860377311706543, "learning_rate": 9.214678899082568e-05, "loss": 3.1764, "step": 78550 }, { "epoch": 8.474393530997304, "grad_norm": 0.87001633644104, "learning_rate": 9.182298974635726e-05, "loss": 3.186, "step": 78600 }, { "epoch": 8.479784366576819, "grad_norm": 0.8715295195579529, "learning_rate": 9.150566648677819e-05, "loss": 3.1833, "step": 78650 }, { "epoch": 8.485175202156334, "grad_norm": 0.8244383931159973, "learning_rate": 9.118186724230976e-05, "loss": 3.1753, "step": 78700 }, { "epoch": 8.49056603773585, "grad_norm": 0.8781893849372864, "learning_rate": 9.085806799784134e-05, "loss": 3.1659, "step": 78750 }, { "epoch": 8.495956873315365, "grad_norm": 0.8835569620132446, "learning_rate": 9.053426875337289e-05, "loss": 3.1791, "step": 78800 }, { "epoch": 8.501347708894878, "grad_norm": 0.8591119647026062, "learning_rate": 9.021046950890447e-05, "loss": 3.1674, "step": 78850 }, { "epoch": 8.506738544474393, "grad_norm": 0.8816818594932556, "learning_rate": 8.988667026443605e-05, "loss": 3.1713, "step": 78900 }, { "epoch": 8.512129380053908, "grad_norm": 0.8852073550224304, "learning_rate": 8.956287101996761e-05, "loss": 3.193, "step": 78950 }, { "epoch": 8.517520215633423, "grad_norm": 0.8786860704421997, "learning_rate": 8.923907177549918e-05, "loss": 3.1795, "step": 79000 }, { "epoch": 8.517520215633423, "eval_accuracy": 0.3913402127271429, "eval_loss": 3.3251218795776367, "eval_runtime": 183.9651, "eval_samples_per_second": 97.904, "eval_steps_per_second": 6.121, "step": 79000 }, { "epoch": 8.522911051212938, "grad_norm": 0.8540938496589661, "learning_rate": 8.891527253103075e-05, "loss": 3.1695, "step": 79050 }, { "epoch": 8.528301886792454, "grad_norm": 0.826810896396637, "learning_rate": 8.859147328656233e-05, "loss": 3.1638, "step": 79100 }, { "epoch": 8.533692722371967, "grad_norm": 0.8859862089157104, "learning_rate": 8.82676740420939e-05, "loss": 3.1749, "step": 79150 }, { "epoch": 8.539083557951482, "grad_norm": 0.8580690622329712, "learning_rate": 8.794387479762546e-05, "loss": 3.1868, "step": 79200 }, { "epoch": 8.544474393530997, "grad_norm": 0.8796784281730652, "learning_rate": 8.762007555315704e-05, "loss": 3.1744, "step": 79250 }, { "epoch": 8.549865229110512, "grad_norm": 0.8341676592826843, "learning_rate": 8.72962763086886e-05, "loss": 3.1754, "step": 79300 }, { "epoch": 8.555256064690028, "grad_norm": 0.8569247722625732, "learning_rate": 8.697247706422018e-05, "loss": 3.1855, "step": 79350 }, { "epoch": 8.560646900269543, "grad_norm": 0.8699577450752258, "learning_rate": 8.664867781975174e-05, "loss": 3.1847, "step": 79400 }, { "epoch": 8.566037735849056, "grad_norm": 0.8954500555992126, "learning_rate": 8.632487857528331e-05, "loss": 3.171, "step": 79450 }, { "epoch": 8.571428571428571, "grad_norm": 0.874020516872406, "learning_rate": 8.600107933081489e-05, "loss": 3.1874, "step": 79500 }, { "epoch": 8.576819407008086, "grad_norm": 0.8852034211158752, "learning_rate": 8.567728008634646e-05, "loss": 3.1983, "step": 79550 }, { "epoch": 8.582210242587601, "grad_norm": 0.8179734945297241, "learning_rate": 8.535348084187804e-05, "loss": 3.1768, "step": 79600 }, { "epoch": 8.587601078167117, "grad_norm": 0.8448224067687988, "learning_rate": 8.502968159740959e-05, "loss": 3.1825, "step": 79650 }, { "epoch": 8.59299191374663, "grad_norm": 0.8318737745285034, "learning_rate": 8.470588235294117e-05, "loss": 3.1686, "step": 79700 }, { "epoch": 8.598382749326145, "grad_norm": 0.8655438423156738, "learning_rate": 8.438208310847275e-05, "loss": 3.1878, "step": 79750 }, { "epoch": 8.60377358490566, "grad_norm": 0.8982657194137573, "learning_rate": 8.405828386400431e-05, "loss": 3.1873, "step": 79800 }, { "epoch": 8.609164420485175, "grad_norm": 0.9005709886550903, "learning_rate": 8.373448461953587e-05, "loss": 3.1742, "step": 79850 }, { "epoch": 8.61455525606469, "grad_norm": 0.8569704294204712, "learning_rate": 8.341068537506745e-05, "loss": 3.1849, "step": 79900 }, { "epoch": 8.619946091644206, "grad_norm": 0.8516122698783875, "learning_rate": 8.308688613059902e-05, "loss": 3.1693, "step": 79950 }, { "epoch": 8.625336927223719, "grad_norm": 0.8622608184814453, "learning_rate": 8.27630868861306e-05, "loss": 3.1806, "step": 80000 }, { "epoch": 8.625336927223719, "eval_accuracy": 0.3918276295818787, "eval_loss": 3.320551872253418, "eval_runtime": 185.0275, "eval_samples_per_second": 97.342, "eval_steps_per_second": 6.086, "step": 80000 }, { "epoch": 8.630727762803234, "grad_norm": 0.8917423486709595, "learning_rate": 8.243928764166216e-05, "loss": 3.1862, "step": 80050 }, { "epoch": 8.63611859838275, "grad_norm": 0.9212331771850586, "learning_rate": 8.211548839719374e-05, "loss": 3.1743, "step": 80100 }, { "epoch": 8.641509433962264, "grad_norm": 0.8408286571502686, "learning_rate": 8.17916891527253e-05, "loss": 3.1645, "step": 80150 }, { "epoch": 8.64690026954178, "grad_norm": 0.8397675156593323, "learning_rate": 8.146788990825688e-05, "loss": 3.1815, "step": 80200 }, { "epoch": 8.652291105121293, "grad_norm": 0.9127941131591797, "learning_rate": 8.114409066378844e-05, "loss": 3.1774, "step": 80250 }, { "epoch": 8.657681940700808, "grad_norm": 0.8323633074760437, "learning_rate": 8.082029141932e-05, "loss": 3.1848, "step": 80300 }, { "epoch": 8.663072776280323, "grad_norm": 0.8468418121337891, "learning_rate": 8.049649217485158e-05, "loss": 3.1739, "step": 80350 }, { "epoch": 8.668463611859838, "grad_norm": 0.8884438872337341, "learning_rate": 8.017269293038316e-05, "loss": 3.1894, "step": 80400 }, { "epoch": 8.673854447439354, "grad_norm": 0.8243781924247742, "learning_rate": 7.984889368591474e-05, "loss": 3.1719, "step": 80450 }, { "epoch": 8.679245283018869, "grad_norm": 0.8693651556968689, "learning_rate": 7.952509444144629e-05, "loss": 3.1779, "step": 80500 }, { "epoch": 8.684636118598382, "grad_norm": 0.8584123253822327, "learning_rate": 7.920129519697787e-05, "loss": 3.1675, "step": 80550 }, { "epoch": 8.690026954177897, "grad_norm": 0.8708306550979614, "learning_rate": 7.887749595250945e-05, "loss": 3.188, "step": 80600 }, { "epoch": 8.695417789757412, "grad_norm": 0.8800524473190308, "learning_rate": 7.855369670804101e-05, "loss": 3.1725, "step": 80650 }, { "epoch": 8.700808625336927, "grad_norm": 0.8486020565032959, "learning_rate": 7.822989746357257e-05, "loss": 3.1483, "step": 80700 }, { "epoch": 8.706199460916443, "grad_norm": 0.8517580628395081, "learning_rate": 7.790609821910415e-05, "loss": 3.1761, "step": 80750 }, { "epoch": 8.711590296495956, "grad_norm": 0.8633285164833069, "learning_rate": 7.758229897463572e-05, "loss": 3.1863, "step": 80800 }, { "epoch": 8.716981132075471, "grad_norm": 0.8587695956230164, "learning_rate": 7.725849973016729e-05, "loss": 3.169, "step": 80850 }, { "epoch": 8.722371967654986, "grad_norm": 0.919442892074585, "learning_rate": 7.693470048569886e-05, "loss": 3.1797, "step": 80900 }, { "epoch": 8.727762803234501, "grad_norm": 0.8610529899597168, "learning_rate": 7.661090124123042e-05, "loss": 3.1519, "step": 80950 }, { "epoch": 8.733153638814017, "grad_norm": 0.8181192278862, "learning_rate": 7.6287101996762e-05, "loss": 3.1844, "step": 81000 }, { "epoch": 8.733153638814017, "eval_accuracy": 0.3922472470346182, "eval_loss": 3.317126512527466, "eval_runtime": 185.4624, "eval_samples_per_second": 97.114, "eval_steps_per_second": 6.071, "step": 81000 }, { "epoch": 8.738544474393532, "grad_norm": 0.9170516729354858, "learning_rate": 7.596330275229358e-05, "loss": 3.1725, "step": 81050 }, { "epoch": 8.743935309973045, "grad_norm": 0.8559374809265137, "learning_rate": 7.563950350782515e-05, "loss": 3.1941, "step": 81100 }, { "epoch": 8.74932614555256, "grad_norm": 0.8597634434700012, "learning_rate": 7.53157042633567e-05, "loss": 3.1579, "step": 81150 }, { "epoch": 8.754716981132075, "grad_norm": 0.9581350088119507, "learning_rate": 7.499190501888828e-05, "loss": 3.1843, "step": 81200 }, { "epoch": 8.76010781671159, "grad_norm": 0.8512847423553467, "learning_rate": 7.466810577441986e-05, "loss": 3.1959, "step": 81250 }, { "epoch": 8.765498652291106, "grad_norm": 0.8884473443031311, "learning_rate": 7.434430652995142e-05, "loss": 3.1829, "step": 81300 }, { "epoch": 8.77088948787062, "grad_norm": 0.908321738243103, "learning_rate": 7.4020507285483e-05, "loss": 3.1715, "step": 81350 }, { "epoch": 8.776280323450134, "grad_norm": 0.9126222729682922, "learning_rate": 7.369670804101457e-05, "loss": 3.1931, "step": 81400 }, { "epoch": 8.78167115902965, "grad_norm": 0.907088577747345, "learning_rate": 7.337290879654614e-05, "loss": 3.1721, "step": 81450 }, { "epoch": 8.787061994609164, "grad_norm": 0.8581763505935669, "learning_rate": 7.304910955207771e-05, "loss": 3.1772, "step": 81500 }, { "epoch": 8.79245283018868, "grad_norm": 0.85309898853302, "learning_rate": 7.272531030760927e-05, "loss": 3.1777, "step": 81550 }, { "epoch": 8.797843665768195, "grad_norm": 0.8943376541137695, "learning_rate": 7.240798704803021e-05, "loss": 3.175, "step": 81600 }, { "epoch": 8.80323450134771, "grad_norm": 0.9350511431694031, "learning_rate": 7.208418780356179e-05, "loss": 3.2047, "step": 81650 }, { "epoch": 8.808625336927223, "grad_norm": 0.840402364730835, "learning_rate": 7.176038855909337e-05, "loss": 3.1851, "step": 81700 }, { "epoch": 8.814016172506738, "grad_norm": 0.8565220832824707, "learning_rate": 7.143658931462493e-05, "loss": 3.2089, "step": 81750 }, { "epoch": 8.819407008086253, "grad_norm": 0.9116523861885071, "learning_rate": 7.11127900701565e-05, "loss": 3.2072, "step": 81800 }, { "epoch": 8.824797843665769, "grad_norm": 0.8699450492858887, "learning_rate": 7.078899082568807e-05, "loss": 3.1803, "step": 81850 }, { "epoch": 8.830188679245284, "grad_norm": 0.929908812046051, "learning_rate": 7.046519158121964e-05, "loss": 3.162, "step": 81900 }, { "epoch": 8.835579514824797, "grad_norm": 0.8746318221092224, "learning_rate": 7.014139233675121e-05, "loss": 3.1813, "step": 81950 }, { "epoch": 8.840970350404312, "grad_norm": 0.909820556640625, "learning_rate": 6.981759309228278e-05, "loss": 3.1883, "step": 82000 }, { "epoch": 8.840970350404312, "eval_accuracy": 0.39259884777958587, "eval_loss": 3.3136179447174072, "eval_runtime": 183.8821, "eval_samples_per_second": 97.949, "eval_steps_per_second": 6.123, "step": 82000 }, { "epoch": 8.846361185983827, "grad_norm": 0.8548611998558044, "learning_rate": 6.949379384781434e-05, "loss": 3.1702, "step": 82050 }, { "epoch": 8.851752021563343, "grad_norm": 0.8615626692771912, "learning_rate": 6.916999460334592e-05, "loss": 3.1741, "step": 82100 }, { "epoch": 8.857142857142858, "grad_norm": 0.8501192927360535, "learning_rate": 6.884619535887748e-05, "loss": 3.1789, "step": 82150 }, { "epoch": 8.862533692722373, "grad_norm": 0.8494750261306763, "learning_rate": 6.852239611440906e-05, "loss": 3.1689, "step": 82200 }, { "epoch": 8.867924528301886, "grad_norm": 0.8440556526184082, "learning_rate": 6.819859686994062e-05, "loss": 3.1691, "step": 82250 }, { "epoch": 8.873315363881401, "grad_norm": 0.8936507701873779, "learning_rate": 6.78747976254722e-05, "loss": 3.1835, "step": 82300 }, { "epoch": 8.878706199460916, "grad_norm": 0.8939010500907898, "learning_rate": 6.755099838100377e-05, "loss": 3.1793, "step": 82350 }, { "epoch": 8.884097035040432, "grad_norm": 0.8997593522071838, "learning_rate": 6.722719913653534e-05, "loss": 3.17, "step": 82400 }, { "epoch": 8.889487870619947, "grad_norm": 0.9238204956054688, "learning_rate": 6.690339989206692e-05, "loss": 3.1736, "step": 82450 }, { "epoch": 8.89487870619946, "grad_norm": 0.851765513420105, "learning_rate": 6.657960064759849e-05, "loss": 3.1734, "step": 82500 }, { "epoch": 8.900269541778975, "grad_norm": 0.8584917783737183, "learning_rate": 6.625580140313006e-05, "loss": 3.1807, "step": 82550 }, { "epoch": 8.90566037735849, "grad_norm": 0.8543685674667358, "learning_rate": 6.593200215866163e-05, "loss": 3.1884, "step": 82600 }, { "epoch": 8.911051212938006, "grad_norm": 0.854233980178833, "learning_rate": 6.560820291419319e-05, "loss": 3.173, "step": 82650 }, { "epoch": 8.91644204851752, "grad_norm": 0.9199441075325012, "learning_rate": 6.528440366972477e-05, "loss": 3.1865, "step": 82700 }, { "epoch": 8.921832884097036, "grad_norm": 0.9200901389122009, "learning_rate": 6.496060442525633e-05, "loss": 3.1904, "step": 82750 }, { "epoch": 8.92722371967655, "grad_norm": 0.8842583298683167, "learning_rate": 6.46368051807879e-05, "loss": 3.1736, "step": 82800 }, { "epoch": 8.932614555256064, "grad_norm": 0.8270998597145081, "learning_rate": 6.431300593631948e-05, "loss": 3.1883, "step": 82850 }, { "epoch": 8.93800539083558, "grad_norm": 0.9040391445159912, "learning_rate": 6.398920669185104e-05, "loss": 3.1703, "step": 82900 }, { "epoch": 8.943396226415095, "grad_norm": 0.8602606654167175, "learning_rate": 6.366540744738262e-05, "loss": 3.1694, "step": 82950 }, { "epoch": 8.94878706199461, "grad_norm": 0.8608500361442566, "learning_rate": 6.334160820291418e-05, "loss": 3.164, "step": 83000 }, { "epoch": 8.94878706199461, "eval_accuracy": 0.39290916042718405, "eval_loss": 3.311744213104248, "eval_runtime": 184.0949, "eval_samples_per_second": 97.835, "eval_steps_per_second": 6.116, "step": 83000 }, { "epoch": 8.954177897574123, "grad_norm": 0.8918322920799255, "learning_rate": 6.301780895844576e-05, "loss": 3.1759, "step": 83050 }, { "epoch": 8.959568733153638, "grad_norm": 0.8488160967826843, "learning_rate": 6.269400971397732e-05, "loss": 3.1466, "step": 83100 }, { "epoch": 8.964959568733153, "grad_norm": 0.8954882025718689, "learning_rate": 6.23702104695089e-05, "loss": 3.1558, "step": 83150 }, { "epoch": 8.970350404312669, "grad_norm": 0.8527268767356873, "learning_rate": 6.204641122504047e-05, "loss": 3.1796, "step": 83200 }, { "epoch": 8.975741239892184, "grad_norm": 0.8429299592971802, "learning_rate": 6.172261198057204e-05, "loss": 3.1788, "step": 83250 }, { "epoch": 8.981132075471699, "grad_norm": 0.8879923224449158, "learning_rate": 6.139881273610362e-05, "loss": 3.1768, "step": 83300 }, { "epoch": 8.986522911051212, "grad_norm": 0.8896589279174805, "learning_rate": 6.107501349163518e-05, "loss": 3.1725, "step": 83350 }, { "epoch": 8.991913746630727, "grad_norm": 0.9053753614425659, "learning_rate": 6.0751214247166755e-05, "loss": 3.1803, "step": 83400 }, { "epoch": 8.997304582210242, "grad_norm": 0.8551260828971863, "learning_rate": 6.042741500269832e-05, "loss": 3.1864, "step": 83450 }, { "epoch": 9.002695417789758, "grad_norm": 0.8562919497489929, "learning_rate": 6.01036157582299e-05, "loss": 3.1486, "step": 83500 }, { "epoch": 9.008086253369273, "grad_norm": 0.8919478058815002, "learning_rate": 5.978629249865083e-05, "loss": 3.108, "step": 83550 }, { "epoch": 9.013477088948788, "grad_norm": 0.9198151230812073, "learning_rate": 5.94624932541824e-05, "loss": 3.1126, "step": 83600 }, { "epoch": 9.018867924528301, "grad_norm": 0.8580641150474548, "learning_rate": 5.913869400971397e-05, "loss": 3.1151, "step": 83650 }, { "epoch": 9.024258760107816, "grad_norm": 0.8981974720954895, "learning_rate": 5.881489476524554e-05, "loss": 3.1157, "step": 83700 }, { "epoch": 9.029649595687331, "grad_norm": 0.8441501259803772, "learning_rate": 5.849109552077711e-05, "loss": 3.1226, "step": 83750 }, { "epoch": 9.035040431266847, "grad_norm": 0.851274847984314, "learning_rate": 5.8167296276308683e-05, "loss": 3.1231, "step": 83800 }, { "epoch": 9.040431266846362, "grad_norm": 0.883053183555603, "learning_rate": 5.784349703184026e-05, "loss": 3.1184, "step": 83850 }, { "epoch": 9.045822102425875, "grad_norm": 0.8847874999046326, "learning_rate": 5.7519697787371825e-05, "loss": 3.1098, "step": 83900 }, { "epoch": 9.05121293800539, "grad_norm": 0.9161601662635803, "learning_rate": 5.7195898542903396e-05, "loss": 3.1108, "step": 83950 }, { "epoch": 9.056603773584905, "grad_norm": 0.9136478304862976, "learning_rate": 5.687209929843497e-05, "loss": 3.1285, "step": 84000 }, { "epoch": 9.056603773584905, "eval_accuracy": 0.3927140198406692, "eval_loss": 3.3139076232910156, "eval_runtime": 184.0412, "eval_samples_per_second": 97.864, "eval_steps_per_second": 6.118, "step": 84000 }, { "epoch": 9.06199460916442, "grad_norm": 0.858832597732544, "learning_rate": 5.654830005396654e-05, "loss": 3.1166, "step": 84050 }, { "epoch": 9.067385444743936, "grad_norm": 0.8940949440002441, "learning_rate": 5.62245008094981e-05, "loss": 3.1129, "step": 84100 }, { "epoch": 9.07277628032345, "grad_norm": 0.916739284992218, "learning_rate": 5.590070156502968e-05, "loss": 3.1193, "step": 84150 }, { "epoch": 9.078167115902964, "grad_norm": 0.8455508947372437, "learning_rate": 5.5576902320561244e-05, "loss": 3.1238, "step": 84200 }, { "epoch": 9.08355795148248, "grad_norm": 0.8741229772567749, "learning_rate": 5.525310307609282e-05, "loss": 3.1417, "step": 84250 }, { "epoch": 9.088948787061994, "grad_norm": 0.8635639548301697, "learning_rate": 5.4929303831624386e-05, "loss": 3.1083, "step": 84300 }, { "epoch": 9.09433962264151, "grad_norm": 0.8472112417221069, "learning_rate": 5.460550458715596e-05, "loss": 3.1345, "step": 84350 }, { "epoch": 9.099730458221025, "grad_norm": 0.9317988157272339, "learning_rate": 5.428170534268753e-05, "loss": 3.1339, "step": 84400 }, { "epoch": 9.10512129380054, "grad_norm": 0.8693691492080688, "learning_rate": 5.39579060982191e-05, "loss": 3.1286, "step": 84450 }, { "epoch": 9.110512129380053, "grad_norm": 0.8881381154060364, "learning_rate": 5.363410685375067e-05, "loss": 3.1367, "step": 84500 }, { "epoch": 9.115902964959568, "grad_norm": 0.8703860640525818, "learning_rate": 5.331030760928224e-05, "loss": 3.1097, "step": 84550 }, { "epoch": 9.121293800539084, "grad_norm": 0.8484287261962891, "learning_rate": 5.2986508364813804e-05, "loss": 3.1149, "step": 84600 }, { "epoch": 9.126684636118599, "grad_norm": 0.8474869132041931, "learning_rate": 5.266270912034538e-05, "loss": 3.1091, "step": 84650 }, { "epoch": 9.132075471698114, "grad_norm": 0.9253382086753845, "learning_rate": 5.233890987587695e-05, "loss": 3.129, "step": 84700 }, { "epoch": 9.137466307277627, "grad_norm": 0.9027813076972961, "learning_rate": 5.2015110631408524e-05, "loss": 3.1257, "step": 84750 }, { "epoch": 9.142857142857142, "grad_norm": 0.8579105734825134, "learning_rate": 5.1691311386940094e-05, "loss": 3.1329, "step": 84800 }, { "epoch": 9.148247978436657, "grad_norm": 0.9061856269836426, "learning_rate": 5.136751214247166e-05, "loss": 3.1346, "step": 84850 }, { "epoch": 9.153638814016173, "grad_norm": 0.8676990866661072, "learning_rate": 5.1043712898003236e-05, "loss": 3.1205, "step": 84900 }, { "epoch": 9.159029649595688, "grad_norm": 0.8959469199180603, "learning_rate": 5.07199136535348e-05, "loss": 3.1265, "step": 84950 }, { "epoch": 9.164420485175203, "grad_norm": 0.8786721229553223, "learning_rate": 5.039611440906638e-05, "loss": 3.127, "step": 85000 }, { "epoch": 9.164420485175203, "eval_accuracy": 0.392937301525128, "eval_loss": 3.313495397567749, "eval_runtime": 184.0599, "eval_samples_per_second": 97.854, "eval_steps_per_second": 6.118, "step": 85000 }, { "epoch": 9.169811320754716, "grad_norm": 0.8697172999382019, "learning_rate": 5.007231516459794e-05, "loss": 3.1357, "step": 85050 }, { "epoch": 9.175202156334231, "grad_norm": 0.9224244952201843, "learning_rate": 4.974851592012952e-05, "loss": 3.1413, "step": 85100 }, { "epoch": 9.180592991913747, "grad_norm": 0.8814460039138794, "learning_rate": 4.9424716675661084e-05, "loss": 3.1237, "step": 85150 }, { "epoch": 9.185983827493262, "grad_norm": 0.869848370552063, "learning_rate": 4.9100917431192655e-05, "loss": 3.115, "step": 85200 }, { "epoch": 9.191374663072777, "grad_norm": 0.8667035698890686, "learning_rate": 4.8777118186724226e-05, "loss": 3.1095, "step": 85250 }, { "epoch": 9.19676549865229, "grad_norm": 0.9251530766487122, "learning_rate": 4.8453318942255797e-05, "loss": 3.1439, "step": 85300 }, { "epoch": 9.202156334231805, "grad_norm": 0.8440174460411072, "learning_rate": 4.812951969778736e-05, "loss": 3.1382, "step": 85350 }, { "epoch": 9.20754716981132, "grad_norm": 0.8802871108055115, "learning_rate": 4.780572045331894e-05, "loss": 3.1245, "step": 85400 }, { "epoch": 9.212938005390836, "grad_norm": 0.9250563383102417, "learning_rate": 4.748192120885051e-05, "loss": 3.1276, "step": 85450 }, { "epoch": 9.21832884097035, "grad_norm": 0.8655513525009155, "learning_rate": 4.715812196438208e-05, "loss": 3.1143, "step": 85500 }, { "epoch": 9.223719676549866, "grad_norm": 0.9218123555183411, "learning_rate": 4.683432271991365e-05, "loss": 3.1186, "step": 85550 }, { "epoch": 9.22911051212938, "grad_norm": 0.8827371597290039, "learning_rate": 4.651052347544522e-05, "loss": 3.1105, "step": 85600 }, { "epoch": 9.234501347708894, "grad_norm": 0.8444175124168396, "learning_rate": 4.618672423097679e-05, "loss": 3.1338, "step": 85650 }, { "epoch": 9.23989218328841, "grad_norm": 0.8230475187301636, "learning_rate": 4.586292498650836e-05, "loss": 3.1287, "step": 85700 }, { "epoch": 9.245283018867925, "grad_norm": 0.8665199279785156, "learning_rate": 4.5539125742039935e-05, "loss": 3.1311, "step": 85750 }, { "epoch": 9.25067385444744, "grad_norm": 0.9222322702407837, "learning_rate": 4.52153264975715e-05, "loss": 3.1303, "step": 85800 }, { "epoch": 9.256064690026955, "grad_norm": 0.851323664188385, "learning_rate": 4.4891527253103076e-05, "loss": 3.1292, "step": 85850 }, { "epoch": 9.261455525606468, "grad_norm": 0.9509512782096863, "learning_rate": 4.456772800863464e-05, "loss": 3.1299, "step": 85900 }, { "epoch": 9.266846361185983, "grad_norm": 0.8725172877311707, "learning_rate": 4.424392876416621e-05, "loss": 3.1338, "step": 85950 }, { "epoch": 9.272237196765499, "grad_norm": 0.9174517393112183, "learning_rate": 4.392012951969778e-05, "loss": 3.109, "step": 86000 }, { "epoch": 9.272237196765499, "eval_accuracy": 0.3932686928329622, "eval_loss": 3.3107893466949463, "eval_runtime": 183.9664, "eval_samples_per_second": 97.904, "eval_steps_per_second": 6.121, "step": 86000 }, { "epoch": 9.277628032345014, "grad_norm": 0.8653820753097534, "learning_rate": 4.359633027522935e-05, "loss": 3.1292, "step": 86050 }, { "epoch": 9.283018867924529, "grad_norm": 0.9035626649856567, "learning_rate": 4.3272531030760924e-05, "loss": 3.1232, "step": 86100 }, { "epoch": 9.288409703504042, "grad_norm": 0.8988853096961975, "learning_rate": 4.2948731786292495e-05, "loss": 3.1223, "step": 86150 }, { "epoch": 9.293800539083557, "grad_norm": 0.8584290742874146, "learning_rate": 4.262493254182407e-05, "loss": 3.1422, "step": 86200 }, { "epoch": 9.299191374663073, "grad_norm": 0.926313042640686, "learning_rate": 4.230113329735564e-05, "loss": 3.1234, "step": 86250 }, { "epoch": 9.304582210242588, "grad_norm": 0.8818143606185913, "learning_rate": 4.197733405288721e-05, "loss": 3.1299, "step": 86300 }, { "epoch": 9.309973045822103, "grad_norm": 0.8349787592887878, "learning_rate": 4.165353480841878e-05, "loss": 3.1019, "step": 86350 }, { "epoch": 9.315363881401618, "grad_norm": 0.8978606462478638, "learning_rate": 4.132973556395035e-05, "loss": 3.1234, "step": 86400 }, { "epoch": 9.320754716981131, "grad_norm": 0.919739842414856, "learning_rate": 4.1005936319481913e-05, "loss": 3.1331, "step": 86450 }, { "epoch": 9.326145552560646, "grad_norm": 0.8913258910179138, "learning_rate": 4.068213707501349e-05, "loss": 3.1361, "step": 86500 }, { "epoch": 9.331536388140162, "grad_norm": 0.9055688381195068, "learning_rate": 4.0358337830545055e-05, "loss": 3.1123, "step": 86550 }, { "epoch": 9.336927223719677, "grad_norm": 0.9090213179588318, "learning_rate": 4.003453858607663e-05, "loss": 3.1128, "step": 86600 }, { "epoch": 9.342318059299192, "grad_norm": 0.8740040063858032, "learning_rate": 3.97107393416082e-05, "loss": 3.1319, "step": 86650 }, { "epoch": 9.347708894878707, "grad_norm": 0.895162045955658, "learning_rate": 3.9386940097139775e-05, "loss": 3.1115, "step": 86700 }, { "epoch": 9.35309973045822, "grad_norm": 0.8940107822418213, "learning_rate": 3.906314085267134e-05, "loss": 3.1138, "step": 86750 }, { "epoch": 9.358490566037736, "grad_norm": 0.8494407534599304, "learning_rate": 3.873934160820291e-05, "loss": 3.1352, "step": 86800 }, { "epoch": 9.36388140161725, "grad_norm": 0.9394447207450867, "learning_rate": 3.841554236373448e-05, "loss": 3.1219, "step": 86850 }, { "epoch": 9.369272237196766, "grad_norm": 0.8646783828735352, "learning_rate": 3.809174311926605e-05, "loss": 3.1223, "step": 86900 }, { "epoch": 9.374663072776281, "grad_norm": 0.8925577998161316, "learning_rate": 3.7767943874797616e-05, "loss": 3.1348, "step": 86950 }, { "epoch": 9.380053908355794, "grad_norm": 0.8575171232223511, "learning_rate": 3.744414463032919e-05, "loss": 3.1161, "step": 87000 }, { "epoch": 9.380053908355794, "eval_accuracy": 0.3934189597768096, "eval_loss": 3.309941053390503, "eval_runtime": 184.0384, "eval_samples_per_second": 97.865, "eval_steps_per_second": 6.118, "step": 87000 }, { "epoch": 9.38544474393531, "grad_norm": 0.8544990420341492, "learning_rate": 3.7120345385860764e-05, "loss": 3.1265, "step": 87050 }, { "epoch": 9.390835579514825, "grad_norm": 0.8940395712852478, "learning_rate": 3.6796546141392335e-05, "loss": 3.1151, "step": 87100 }, { "epoch": 9.39622641509434, "grad_norm": 0.8310646414756775, "learning_rate": 3.6472746896923906e-05, "loss": 3.1139, "step": 87150 }, { "epoch": 9.401617250673855, "grad_norm": 0.9821915030479431, "learning_rate": 3.614894765245547e-05, "loss": 3.1308, "step": 87200 }, { "epoch": 9.40700808625337, "grad_norm": 0.90660160779953, "learning_rate": 3.582514840798704e-05, "loss": 3.1328, "step": 87250 }, { "epoch": 9.412398921832883, "grad_norm": 0.927312970161438, "learning_rate": 3.550134916351861e-05, "loss": 3.133, "step": 87300 }, { "epoch": 9.417789757412399, "grad_norm": 0.8718499541282654, "learning_rate": 3.517754991905018e-05, "loss": 3.1129, "step": 87350 }, { "epoch": 9.423180592991914, "grad_norm": 0.8952221870422363, "learning_rate": 3.485375067458176e-05, "loss": 3.1247, "step": 87400 }, { "epoch": 9.428571428571429, "grad_norm": 0.8873728513717651, "learning_rate": 3.452995143011333e-05, "loss": 3.1104, "step": 87450 }, { "epoch": 9.433962264150944, "grad_norm": 0.9118414521217346, "learning_rate": 3.4206152185644895e-05, "loss": 3.1336, "step": 87500 }, { "epoch": 9.439353099730457, "grad_norm": 0.931516170501709, "learning_rate": 3.3882352941176466e-05, "loss": 3.1367, "step": 87550 }, { "epoch": 9.444743935309972, "grad_norm": 0.8752477765083313, "learning_rate": 3.355855369670804e-05, "loss": 3.137, "step": 87600 }, { "epoch": 9.450134770889488, "grad_norm": 0.8985586762428284, "learning_rate": 3.323475445223961e-05, "loss": 3.1092, "step": 87650 }, { "epoch": 9.455525606469003, "grad_norm": 0.877842366695404, "learning_rate": 3.291095520777118e-05, "loss": 3.1204, "step": 87700 }, { "epoch": 9.460916442048518, "grad_norm": 0.8955590128898621, "learning_rate": 3.258715596330275e-05, "loss": 3.1372, "step": 87750 }, { "epoch": 9.466307277628033, "grad_norm": 0.8843672275543213, "learning_rate": 3.226335671883432e-05, "loss": 3.1002, "step": 87800 }, { "epoch": 9.471698113207546, "grad_norm": 0.8708178400993347, "learning_rate": 3.193955747436589e-05, "loss": 3.1287, "step": 87850 }, { "epoch": 9.477088948787062, "grad_norm": 0.8979765772819519, "learning_rate": 3.161575822989746e-05, "loss": 3.1323, "step": 87900 }, { "epoch": 9.482479784366577, "grad_norm": 0.9137650728225708, "learning_rate": 3.129195898542903e-05, "loss": 3.1247, "step": 87950 }, { "epoch": 9.487870619946092, "grad_norm": 0.87474125623703, "learning_rate": 3.09681597409606e-05, "loss": 3.1228, "step": 88000 }, { "epoch": 9.487870619946092, "eval_accuracy": 0.39371069278059145, "eval_loss": 3.306328773498535, "eval_runtime": 184.1066, "eval_samples_per_second": 97.829, "eval_steps_per_second": 6.116, "step": 88000 }, { "epoch": 9.493261455525607, "grad_norm": 0.8465813398361206, "learning_rate": 3.064436049649217e-05, "loss": 3.1099, "step": 88050 }, { "epoch": 9.498652291105122, "grad_norm": 0.8977415561676025, "learning_rate": 3.032056125202374e-05, "loss": 3.126, "step": 88100 }, { "epoch": 9.504043126684635, "grad_norm": 0.9405316114425659, "learning_rate": 2.999676200755531e-05, "loss": 3.1336, "step": 88150 }, { "epoch": 9.50943396226415, "grad_norm": 0.8797356486320496, "learning_rate": 2.9672962763086884e-05, "loss": 3.1191, "step": 88200 }, { "epoch": 9.514824797843666, "grad_norm": 0.8984405398368835, "learning_rate": 2.9349163518618455e-05, "loss": 3.1202, "step": 88250 }, { "epoch": 9.520215633423181, "grad_norm": 0.874954342842102, "learning_rate": 2.9025364274150026e-05, "loss": 3.1304, "step": 88300 }, { "epoch": 9.525606469002696, "grad_norm": 0.8673447370529175, "learning_rate": 2.8701565029681597e-05, "loss": 3.1077, "step": 88350 }, { "epoch": 9.530997304582211, "grad_norm": 0.9109377264976501, "learning_rate": 2.8377765785213165e-05, "loss": 3.137, "step": 88400 }, { "epoch": 9.536388140161725, "grad_norm": 0.9213530421257019, "learning_rate": 2.8053966540744736e-05, "loss": 3.1292, "step": 88450 }, { "epoch": 9.54177897574124, "grad_norm": 0.8363325595855713, "learning_rate": 2.7730167296276306e-05, "loss": 3.14, "step": 88500 }, { "epoch": 9.547169811320755, "grad_norm": 0.9367228746414185, "learning_rate": 2.7406368051807877e-05, "loss": 3.1314, "step": 88550 }, { "epoch": 9.55256064690027, "grad_norm": 0.8717453479766846, "learning_rate": 2.7082568807339448e-05, "loss": 3.1254, "step": 88600 }, { "epoch": 9.557951482479785, "grad_norm": 0.9186379909515381, "learning_rate": 2.6758769562871016e-05, "loss": 3.1208, "step": 88650 }, { "epoch": 9.563342318059298, "grad_norm": 0.9241523146629333, "learning_rate": 2.6434970318402587e-05, "loss": 3.1223, "step": 88700 }, { "epoch": 9.568733153638814, "grad_norm": 0.9091995358467102, "learning_rate": 2.6117647058823525e-05, "loss": 3.124, "step": 88750 }, { "epoch": 9.574123989218329, "grad_norm": 0.8714845180511475, "learning_rate": 2.57938478143551e-05, "loss": 3.1406, "step": 88800 }, { "epoch": 9.579514824797844, "grad_norm": 0.8658944964408875, "learning_rate": 2.547004856988667e-05, "loss": 3.1387, "step": 88850 }, { "epoch": 9.584905660377359, "grad_norm": 0.8937883377075195, "learning_rate": 2.514624932541824e-05, "loss": 3.1142, "step": 88900 }, { "epoch": 9.590296495956874, "grad_norm": 0.8627274036407471, "learning_rate": 2.482245008094981e-05, "loss": 3.1166, "step": 88950 }, { "epoch": 9.595687331536388, "grad_norm": 0.9279044270515442, "learning_rate": 2.449865083648138e-05, "loss": 3.1352, "step": 89000 }, { "epoch": 9.595687331536388, "eval_accuracy": 0.39397634909129786, "eval_loss": 3.3040802478790283, "eval_runtime": 183.7954, "eval_samples_per_second": 97.995, "eval_steps_per_second": 6.126, "step": 89000 }, { "epoch": 9.601078167115903, "grad_norm": 0.8863611221313477, "learning_rate": 2.417485159201295e-05, "loss": 3.1325, "step": 89050 }, { "epoch": 9.606469002695418, "grad_norm": 0.9018926024436951, "learning_rate": 2.385105234754452e-05, "loss": 3.1223, "step": 89100 }, { "epoch": 9.611859838274933, "grad_norm": 0.8995761871337891, "learning_rate": 2.3527253103076092e-05, "loss": 3.1265, "step": 89150 }, { "epoch": 9.617250673854448, "grad_norm": 0.887016236782074, "learning_rate": 2.320345385860766e-05, "loss": 3.1446, "step": 89200 }, { "epoch": 9.622641509433961, "grad_norm": 0.9198271036148071, "learning_rate": 2.287965461413923e-05, "loss": 3.1424, "step": 89250 }, { "epoch": 9.628032345013477, "grad_norm": 0.965681254863739, "learning_rate": 2.25558553696708e-05, "loss": 3.1323, "step": 89300 }, { "epoch": 9.633423180592992, "grad_norm": 0.9015873074531555, "learning_rate": 2.2232056125202372e-05, "loss": 3.1177, "step": 89350 }, { "epoch": 9.638814016172507, "grad_norm": 0.8877080082893372, "learning_rate": 2.1908256880733943e-05, "loss": 3.12, "step": 89400 }, { "epoch": 9.644204851752022, "grad_norm": 0.8573229908943176, "learning_rate": 2.158445763626551e-05, "loss": 3.1291, "step": 89450 }, { "epoch": 9.649595687331537, "grad_norm": 0.878963828086853, "learning_rate": 2.126065839179708e-05, "loss": 3.126, "step": 89500 }, { "epoch": 9.65498652291105, "grad_norm": 0.9286015629768372, "learning_rate": 2.0936859147328653e-05, "loss": 3.1131, "step": 89550 }, { "epoch": 9.660377358490566, "grad_norm": 0.8522419929504395, "learning_rate": 2.0613059902860227e-05, "loss": 3.1264, "step": 89600 }, { "epoch": 9.66576819407008, "grad_norm": 0.8949482440948486, "learning_rate": 2.0289260658391798e-05, "loss": 3.1352, "step": 89650 }, { "epoch": 9.671159029649596, "grad_norm": 0.9174593091011047, "learning_rate": 1.996546141392337e-05, "loss": 3.1471, "step": 89700 }, { "epoch": 9.676549865229111, "grad_norm": 0.8965691328048706, "learning_rate": 1.9641662169454936e-05, "loss": 3.1144, "step": 89750 }, { "epoch": 9.681940700808624, "grad_norm": 0.8825041651725769, "learning_rate": 1.9317862924986507e-05, "loss": 3.1233, "step": 89800 }, { "epoch": 9.68733153638814, "grad_norm": 0.9101404547691345, "learning_rate": 1.8994063680518078e-05, "loss": 3.1192, "step": 89850 }, { "epoch": 9.692722371967655, "grad_norm": 0.8761759996414185, "learning_rate": 1.867026443604965e-05, "loss": 3.1375, "step": 89900 }, { "epoch": 9.69811320754717, "grad_norm": 0.8935624361038208, "learning_rate": 1.834646519158122e-05, "loss": 3.1185, "step": 89950 }, { "epoch": 9.703504043126685, "grad_norm": 0.9249722361564636, "learning_rate": 1.8022665947112787e-05, "loss": 3.127, "step": 90000 }, { "epoch": 9.703504043126685, "eval_accuracy": 0.3942407015673505, "eval_loss": 3.3023488521575928, "eval_runtime": 184.0912, "eval_samples_per_second": 97.837, "eval_steps_per_second": 6.117, "step": 90000 }, { "epoch": 9.7088948787062, "grad_norm": 0.905051589012146, "learning_rate": 1.7698866702644358e-05, "loss": 3.1331, "step": 90050 }, { "epoch": 9.714285714285714, "grad_norm": 0.8957482576370239, "learning_rate": 1.737506745817593e-05, "loss": 3.1176, "step": 90100 }, { "epoch": 9.719676549865229, "grad_norm": 0.9250184297561646, "learning_rate": 1.70512682137075e-05, "loss": 3.1286, "step": 90150 }, { "epoch": 9.725067385444744, "grad_norm": 0.8829726576805115, "learning_rate": 1.672746896923907e-05, "loss": 3.1508, "step": 90200 }, { "epoch": 9.730458221024259, "grad_norm": 0.9289597272872925, "learning_rate": 1.640366972477064e-05, "loss": 3.1319, "step": 90250 }, { "epoch": 9.735849056603774, "grad_norm": 0.9138729572296143, "learning_rate": 1.6079870480302213e-05, "loss": 3.1083, "step": 90300 }, { "epoch": 9.74123989218329, "grad_norm": 0.8744782209396362, "learning_rate": 1.575607123583378e-05, "loss": 3.117, "step": 90350 }, { "epoch": 9.746630727762803, "grad_norm": 0.8819894790649414, "learning_rate": 1.543227199136535e-05, "loss": 3.1321, "step": 90400 }, { "epoch": 9.752021563342318, "grad_norm": 0.8464565873146057, "learning_rate": 1.5108472746896922e-05, "loss": 3.1234, "step": 90450 }, { "epoch": 9.757412398921833, "grad_norm": 0.913282036781311, "learning_rate": 1.4784673502428493e-05, "loss": 3.1394, "step": 90500 }, { "epoch": 9.762803234501348, "grad_norm": 0.8773240447044373, "learning_rate": 1.4460874257960065e-05, "loss": 3.1411, "step": 90550 }, { "epoch": 9.768194070080863, "grad_norm": 0.9089623093605042, "learning_rate": 1.4137075013491634e-05, "loss": 3.1192, "step": 90600 }, { "epoch": 9.773584905660378, "grad_norm": 0.8616759181022644, "learning_rate": 1.3813275769023205e-05, "loss": 3.1208, "step": 90650 }, { "epoch": 9.778975741239892, "grad_norm": 0.8891574144363403, "learning_rate": 1.3489476524554775e-05, "loss": 3.1304, "step": 90700 }, { "epoch": 9.784366576819407, "grad_norm": 0.9050963521003723, "learning_rate": 1.3165677280086345e-05, "loss": 3.1315, "step": 90750 }, { "epoch": 9.789757412398922, "grad_norm": 0.9230057001113892, "learning_rate": 1.2841878035617915e-05, "loss": 3.119, "step": 90800 }, { "epoch": 9.795148247978437, "grad_norm": 0.9277952909469604, "learning_rate": 1.2518078791149486e-05, "loss": 3.1184, "step": 90850 }, { "epoch": 9.800539083557952, "grad_norm": 0.8638946413993835, "learning_rate": 1.2194279546681056e-05, "loss": 3.1233, "step": 90900 }, { "epoch": 9.805929919137466, "grad_norm": 0.8712254166603088, "learning_rate": 1.1870480302212627e-05, "loss": 3.1248, "step": 90950 }, { "epoch": 9.81132075471698, "grad_norm": 0.9255781769752502, "learning_rate": 1.1546681057744198e-05, "loss": 3.1309, "step": 91000 }, { "epoch": 9.81132075471698, "eval_accuracy": 0.3944802811849814, "eval_loss": 3.300304651260376, "eval_runtime": 183.9703, "eval_samples_per_second": 97.902, "eval_steps_per_second": 6.121, "step": 91000 }, { "epoch": 9.816711590296496, "grad_norm": 0.9070011377334595, "learning_rate": 1.1222881813275769e-05, "loss": 3.1308, "step": 91050 }, { "epoch": 9.822102425876011, "grad_norm": 0.8719533085823059, "learning_rate": 1.0899082568807338e-05, "loss": 3.1203, "step": 91100 }, { "epoch": 9.827493261455526, "grad_norm": 0.8622357845306396, "learning_rate": 1.057528332433891e-05, "loss": 3.1135, "step": 91150 }, { "epoch": 9.832884097035041, "grad_norm": 0.8596940040588379, "learning_rate": 1.0251484079870478e-05, "loss": 3.0975, "step": 91200 }, { "epoch": 9.838274932614555, "grad_norm": 0.9166052937507629, "learning_rate": 9.92768483540205e-06, "loss": 3.1198, "step": 91250 }, { "epoch": 9.84366576819407, "grad_norm": 0.8794975280761719, "learning_rate": 9.60388559093362e-06, "loss": 3.1331, "step": 91300 }, { "epoch": 9.849056603773585, "grad_norm": 0.8526261448860168, "learning_rate": 9.280086346465191e-06, "loss": 3.1378, "step": 91350 }, { "epoch": 9.8544474393531, "grad_norm": 0.8655098080635071, "learning_rate": 8.95628710199676e-06, "loss": 3.1197, "step": 91400 }, { "epoch": 9.859838274932615, "grad_norm": 0.8488094806671143, "learning_rate": 8.632487857528333e-06, "loss": 3.1101, "step": 91450 }, { "epoch": 9.865229110512129, "grad_norm": 0.8776468634605408, "learning_rate": 8.308688613059902e-06, "loss": 3.1166, "step": 91500 }, { "epoch": 9.870619946091644, "grad_norm": 0.8622964024543762, "learning_rate": 7.984889368591473e-06, "loss": 3.1334, "step": 91550 }, { "epoch": 9.876010781671159, "grad_norm": 0.9132906794548035, "learning_rate": 7.661090124123042e-06, "loss": 3.1244, "step": 91600 }, { "epoch": 9.881401617250674, "grad_norm": 0.8868639469146729, "learning_rate": 7.337290879654614e-06, "loss": 3.1337, "step": 91650 }, { "epoch": 9.88679245283019, "grad_norm": 0.8823307156562805, "learning_rate": 7.013491635186184e-06, "loss": 3.1278, "step": 91700 }, { "epoch": 9.892183288409704, "grad_norm": 0.8843882083892822, "learning_rate": 6.689692390717754e-06, "loss": 3.1326, "step": 91750 }, { "epoch": 9.897574123989218, "grad_norm": 0.8725460171699524, "learning_rate": 6.365893146249325e-06, "loss": 3.1384, "step": 91800 }, { "epoch": 9.902964959568733, "grad_norm": 0.8682441711425781, "learning_rate": 6.042093901780896e-06, "loss": 3.1279, "step": 91850 }, { "epoch": 9.908355795148248, "grad_norm": 0.8866960406303406, "learning_rate": 5.718294657312466e-06, "loss": 3.1149, "step": 91900 }, { "epoch": 9.913746630727763, "grad_norm": 0.860151469707489, "learning_rate": 5.394495412844036e-06, "loss": 3.1298, "step": 91950 }, { "epoch": 9.919137466307278, "grad_norm": 0.915698766708374, "learning_rate": 5.070696168375607e-06, "loss": 3.1271, "step": 92000 }, { "epoch": 9.919137466307278, "eval_accuracy": 0.39464065284739563, "eval_loss": 3.298301935195923, "eval_runtime": 184.3212, "eval_samples_per_second": 97.715, "eval_steps_per_second": 6.109, "step": 92000 }, { "epoch": 9.924528301886792, "grad_norm": 0.8827758431434631, "learning_rate": 4.7468969239071776e-06, "loss": 3.1396, "step": 92050 }, { "epoch": 9.929919137466307, "grad_norm": 0.8892190456390381, "learning_rate": 4.423097679438748e-06, "loss": 3.1326, "step": 92100 }, { "epoch": 9.935309973045822, "grad_norm": 0.9178445339202881, "learning_rate": 4.099298434970318e-06, "loss": 3.1344, "step": 92150 }, { "epoch": 9.940700808625337, "grad_norm": 0.862002968788147, "learning_rate": 3.7754991905018885e-06, "loss": 3.1337, "step": 92200 }, { "epoch": 9.946091644204852, "grad_norm": 0.8599331378936768, "learning_rate": 3.451699946033459e-06, "loss": 3.1044, "step": 92250 }, { "epoch": 9.951482479784367, "grad_norm": 0.9012652039527893, "learning_rate": 3.1279007015650295e-06, "loss": 3.1089, "step": 92300 }, { "epoch": 9.95687331536388, "grad_norm": 0.9277687072753906, "learning_rate": 2.8105774419859684e-06, "loss": 3.1184, "step": 92350 }, { "epoch": 9.962264150943396, "grad_norm": 0.9038698077201843, "learning_rate": 2.486778197517539e-06, "loss": 3.1224, "step": 92400 }, { "epoch": 9.967654986522911, "grad_norm": 0.859693169593811, "learning_rate": 2.1629789530491094e-06, "loss": 3.1161, "step": 92450 }, { "epoch": 9.973045822102426, "grad_norm": 0.8956581950187683, "learning_rate": 1.8391797085806798e-06, "loss": 3.1283, "step": 92500 }, { "epoch": 9.978436657681941, "grad_norm": 0.9227294325828552, "learning_rate": 1.5153804641122503e-06, "loss": 3.1291, "step": 92550 }, { "epoch": 9.983827493261456, "grad_norm": 0.8772092461585999, "learning_rate": 1.1915812196438208e-06, "loss": 3.1279, "step": 92600 }, { "epoch": 9.98921832884097, "grad_norm": 0.8762646913528442, "learning_rate": 8.677819751753911e-07, "loss": 3.1365, "step": 92650 }, { "epoch": 9.994609164420485, "grad_norm": 0.8837000131607056, "learning_rate": 5.439827307069617e-07, "loss": 3.1319, "step": 92700 }, { "epoch": 10.0, "grad_norm": 2.040294885635376, "learning_rate": 2.201834862385321e-07, "loss": 3.1364, "step": 92750 }, { "epoch": 10.005390835579515, "grad_norm": 0.9977272748947144, "learning_rate": 0.0004800828210934972, "loss": 3.1867, "step": 92800 }, { "epoch": 10.01078167115903, "grad_norm": 1.0503565073013306, "learning_rate": 0.00048001811711420253, "loss": 3.2172, "step": 92850 }, { "epoch": 10.016172506738544, "grad_norm": 1.1011548042297363, "learning_rate": 0.00047995341313490777, "loss": 3.2555, "step": 92900 }, { "epoch": 10.021563342318059, "grad_norm": 0.9567869305610657, "learning_rate": 0.000479888709155613, "loss": 3.2694, "step": 92950 }, { "epoch": 10.026954177897574, "grad_norm": 1.1113311052322388, "learning_rate": 0.00047982400517631826, "loss": 3.2875, "step": 93000 }, { "epoch": 10.026954177897574, "eval_accuracy": 0.37822396206884223, "eval_loss": 3.4554944038391113, "eval_runtime": 146.8369, "eval_samples_per_second": 122.66, "eval_steps_per_second": 7.668, "step": 93000 }, { "epoch": 10.032345013477089, "grad_norm": 1.1876881122589111, "learning_rate": 0.0004797593011970236, "loss": 3.3014, "step": 93050 }, { "epoch": 10.037735849056604, "grad_norm": 0.9571216702461243, "learning_rate": 0.00047969459721772885, "loss": 3.2934, "step": 93100 }, { "epoch": 10.04312668463612, "grad_norm": 0.9346808195114136, "learning_rate": 0.0004796298932384341, "loss": 3.3042, "step": 93150 }, { "epoch": 10.048517520215633, "grad_norm": 0.9184759259223938, "learning_rate": 0.00047956518925913944, "loss": 3.3273, "step": 93200 }, { "epoch": 10.053908355795148, "grad_norm": 0.9030526280403137, "learning_rate": 0.0004795004852798447, "loss": 3.3273, "step": 93250 }, { "epoch": 10.059299191374663, "grad_norm": 0.9374904632568359, "learning_rate": 0.0004794357813005499, "loss": 3.3431, "step": 93300 }, { "epoch": 10.064690026954178, "grad_norm": 1.0782991647720337, "learning_rate": 0.00047937107732125517, "loss": 3.3481, "step": 93350 }, { "epoch": 10.070080862533693, "grad_norm": 0.9518354535102844, "learning_rate": 0.0004793063733419605, "loss": 3.3441, "step": 93400 }, { "epoch": 10.075471698113208, "grad_norm": 0.9269945621490479, "learning_rate": 0.00047924166936266576, "loss": 3.3473, "step": 93450 }, { "epoch": 10.080862533692722, "grad_norm": 0.9213449954986572, "learning_rate": 0.000479176965383371, "loss": 3.3259, "step": 93500 }, { "epoch": 10.086253369272237, "grad_norm": 0.9624829292297363, "learning_rate": 0.0004791122614040763, "loss": 3.3459, "step": 93550 }, { "epoch": 10.091644204851752, "grad_norm": 0.9101693630218506, "learning_rate": 0.0004790475574247816, "loss": 3.3453, "step": 93600 }, { "epoch": 10.097035040431267, "grad_norm": 1.010877251625061, "learning_rate": 0.00047898285344548684, "loss": 3.3487, "step": 93650 }, { "epoch": 10.102425876010782, "grad_norm": 0.9577584862709045, "learning_rate": 0.00047891814946619213, "loss": 3.3608, "step": 93700 }, { "epoch": 10.107816711590296, "grad_norm": 0.9266205430030823, "learning_rate": 0.0004788534454868974, "loss": 3.3607, "step": 93750 }, { "epoch": 10.11320754716981, "grad_norm": 0.9234276413917542, "learning_rate": 0.00047878874150760267, "loss": 3.3616, "step": 93800 }, { "epoch": 10.118598382749326, "grad_norm": 1.0343414545059204, "learning_rate": 0.00047872403752830797, "loss": 3.3605, "step": 93850 }, { "epoch": 10.123989218328841, "grad_norm": 0.823859691619873, "learning_rate": 0.0004786593335490132, "loss": 3.3597, "step": 93900 }, { "epoch": 10.129380053908356, "grad_norm": 0.8498820066452026, "learning_rate": 0.0004785946295697185, "loss": 3.3508, "step": 93950 }, { "epoch": 10.134770889487871, "grad_norm": 0.876309871673584, "learning_rate": 0.0004785299255904238, "loss": 3.3637, "step": 94000 }, { "epoch": 10.134770889487871, "eval_accuracy": 0.3758454417016954, "eval_loss": 3.4686508178710938, "eval_runtime": 146.2986, "eval_samples_per_second": 123.111, "eval_steps_per_second": 7.697, "step": 94000 }, { "epoch": 10.140161725067385, "grad_norm": 0.8599374294281006, "learning_rate": 0.00047846522161112905, "loss": 3.3606, "step": 94050 }, { "epoch": 10.1455525606469, "grad_norm": 0.8754469156265259, "learning_rate": 0.0004784005176318343, "loss": 3.3598, "step": 94100 }, { "epoch": 10.150943396226415, "grad_norm": 0.8947704434394836, "learning_rate": 0.00047833581365253964, "loss": 3.3447, "step": 94150 }, { "epoch": 10.15633423180593, "grad_norm": 0.9169558882713318, "learning_rate": 0.0004782711096732449, "loss": 3.3796, "step": 94200 }, { "epoch": 10.161725067385445, "grad_norm": 0.8607390522956848, "learning_rate": 0.0004782064056939501, "loss": 3.3756, "step": 94250 }, { "epoch": 10.167115902964959, "grad_norm": 0.8300653696060181, "learning_rate": 0.00047814170171465536, "loss": 3.3631, "step": 94300 }, { "epoch": 10.172506738544474, "grad_norm": 0.9180701375007629, "learning_rate": 0.0004780769977353607, "loss": 3.3624, "step": 94350 }, { "epoch": 10.177897574123989, "grad_norm": 0.7899641990661621, "learning_rate": 0.00047801229375606596, "loss": 3.3785, "step": 94400 }, { "epoch": 10.183288409703504, "grad_norm": 0.854278028011322, "learning_rate": 0.0004779475897767712, "loss": 3.3768, "step": 94450 }, { "epoch": 10.18867924528302, "grad_norm": 0.8559595942497253, "learning_rate": 0.00047788288579747655, "loss": 3.357, "step": 94500 }, { "epoch": 10.194070080862534, "grad_norm": 0.9118571877479553, "learning_rate": 0.0004778181818181818, "loss": 3.369, "step": 94550 }, { "epoch": 10.199460916442048, "grad_norm": 0.7602856755256653, "learning_rate": 0.00047775347783888703, "loss": 3.3648, "step": 94600 }, { "epoch": 10.204851752021563, "grad_norm": 0.8732795715332031, "learning_rate": 0.0004776887738595923, "loss": 3.3724, "step": 94650 }, { "epoch": 10.210242587601078, "grad_norm": 0.8687312602996826, "learning_rate": 0.0004776240698802976, "loss": 3.384, "step": 94700 }, { "epoch": 10.215633423180593, "grad_norm": 0.8585965037345886, "learning_rate": 0.00047755936590100287, "loss": 3.3806, "step": 94750 }, { "epoch": 10.221024258760108, "grad_norm": 0.8676369786262512, "learning_rate": 0.0004774946619217081, "loss": 3.3732, "step": 94800 }, { "epoch": 10.226415094339623, "grad_norm": 0.8202512264251709, "learning_rate": 0.0004774299579424134, "loss": 3.3866, "step": 94850 }, { "epoch": 10.231805929919137, "grad_norm": 0.8557911515235901, "learning_rate": 0.0004773652539631187, "loss": 3.3494, "step": 94900 }, { "epoch": 10.237196765498652, "grad_norm": 0.8179985284805298, "learning_rate": 0.00047730054998382395, "loss": 3.3798, "step": 94950 }, { "epoch": 10.242587601078167, "grad_norm": 0.882681667804718, "learning_rate": 0.00047723584600452924, "loss": 3.3803, "step": 95000 }, { "epoch": 10.242587601078167, "eval_accuracy": 0.37679104778434486, "eval_loss": 3.4614784717559814, "eval_runtime": 145.9969, "eval_samples_per_second": 123.366, "eval_steps_per_second": 7.712, "step": 95000 }, { "epoch": 10.242587601078167, "step": 95000, "total_flos": 7.94262454272e+17, "train_loss": 0.07917592998303866, "train_runtime": 1615.2207, "train_samples_per_second": 9186.825, "train_steps_per_second": 287.112 } ], "logging_steps": 50, "max_steps": 463750, "num_input_tokens_seen": 0, "num_train_epochs": 50, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.94262454272e+17, "train_batch_size": 32, "trial_name": null, "trial_params": null }