{ "best_metric": 3.3018171787261963, "best_model_checkpoint": "/scratch/cl5625/exceptions/models/100M_495/checkpoint-90000", "epoch": 10.0, "eval_steps": 1000, "global_step": 92910, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005381552039608223, "grad_norm": 2.1159660816192627, "learning_rate": 0.00028199999999999997, "loss": 8.9316, "step": 50 }, { "epoch": 0.010763104079216447, "grad_norm": 1.9895623922348022, "learning_rate": 0.0005819999999999999, "loss": 6.9096, "step": 100 }, { "epoch": 0.01614465611882467, "grad_norm": 1.8703136444091797, "learning_rate": 0.0005996961534317422, "loss": 6.4463, "step": 150 }, { "epoch": 0.021526208158432893, "grad_norm": 1.2748037576675415, "learning_rate": 0.0005993729124016807, "loss": 6.2017, "step": 200 }, { "epoch": 0.026907760198041114, "grad_norm": 1.3479403257369995, "learning_rate": 0.0005990496713716194, "loss": 6.1031, "step": 250 }, { "epoch": 0.03228931223764934, "grad_norm": 1.0388602018356323, "learning_rate": 0.000598726430341558, "loss": 5.9726, "step": 300 }, { "epoch": 0.03767086427725756, "grad_norm": 1.1332666873931885, "learning_rate": 0.0005984031893114966, "loss": 5.886, "step": 350 }, { "epoch": 0.04305241631686579, "grad_norm": 1.398230791091919, "learning_rate": 0.0005980799482814351, "loss": 5.8019, "step": 400 }, { "epoch": 0.048433968356474004, "grad_norm": 1.5357556343078613, "learning_rate": 0.0005977567072513737, "loss": 5.7437, "step": 450 }, { "epoch": 0.05381552039608223, "grad_norm": 1.175200343132019, "learning_rate": 0.0005974334662213123, "loss": 5.6616, "step": 500 }, { "epoch": 0.05919707243569045, "grad_norm": 1.386881709098816, "learning_rate": 0.0005971102251912509, "loss": 5.5895, "step": 550 }, { "epoch": 0.06457862447529868, "grad_norm": 1.3578399419784546, "learning_rate": 0.0005967869841611895, "loss": 5.4912, "step": 600 }, { "epoch": 0.0699601765149069, "grad_norm": 1.2151751518249512, "learning_rate": 0.000596463743131128, "loss": 5.4275, "step": 650 }, { "epoch": 0.07534172855451512, "grad_norm": 1.3390365839004517, "learning_rate": 0.0005961405021010667, "loss": 5.3731, "step": 700 }, { "epoch": 0.08072328059412334, "grad_norm": 1.103965401649475, "learning_rate": 0.0005958172610710052, "loss": 5.3186, "step": 750 }, { "epoch": 0.08610483263373157, "grad_norm": 1.1866347789764404, "learning_rate": 0.0005954940200409439, "loss": 5.2687, "step": 800 }, { "epoch": 0.09148638467333979, "grad_norm": 1.0097291469573975, "learning_rate": 0.0005951707790108824, "loss": 5.2072, "step": 850 }, { "epoch": 0.09686793671294801, "grad_norm": 1.3843412399291992, "learning_rate": 0.0005948475379808209, "loss": 5.1549, "step": 900 }, { "epoch": 0.10224948875255624, "grad_norm": 1.1859126091003418, "learning_rate": 0.0005945242969507596, "loss": 5.0977, "step": 950 }, { "epoch": 0.10763104079216446, "grad_norm": 1.1825942993164062, "learning_rate": 0.0005942010559206981, "loss": 5.0644, "step": 1000 }, { "epoch": 0.10763104079216446, "eval_accuracy": 0.22798678911267642, "eval_loss": 5.015804290771484, "eval_runtime": 186.987, "eval_samples_per_second": 96.322, "eval_steps_per_second": 6.022, "step": 1000 }, { "epoch": 0.11301259283177269, "grad_norm": 1.036710262298584, "learning_rate": 0.0005938778148906367, "loss": 5.045, "step": 1050 }, { "epoch": 0.1183941448713809, "grad_norm": 1.1089870929718018, "learning_rate": 0.0005935545738605753, "loss": 5.0081, "step": 1100 }, { "epoch": 0.12377569691098914, "grad_norm": 0.9135867357254028, "learning_rate": 0.000593231332830514, "loss": 4.9925, "step": 1150 }, { "epoch": 0.12915724895059735, "grad_norm": 0.8327316641807556, "learning_rate": 0.0005929080918004525, "loss": 4.9437, "step": 1200 }, { "epoch": 0.13453880099020557, "grad_norm": 1.0137102603912354, "learning_rate": 0.000592584850770391, "loss": 4.9288, "step": 1250 }, { "epoch": 0.1399203530298138, "grad_norm": 1.620171070098877, "learning_rate": 0.0005922616097403296, "loss": 4.8813, "step": 1300 }, { "epoch": 0.14530190506942203, "grad_norm": 1.1357035636901855, "learning_rate": 0.0005919383687102682, "loss": 4.8713, "step": 1350 }, { "epoch": 0.15068345710903025, "grad_norm": 0.7975968718528748, "learning_rate": 0.0005916151276802069, "loss": 4.835, "step": 1400 }, { "epoch": 0.15606500914863847, "grad_norm": 0.9254979491233826, "learning_rate": 0.0005912918866501454, "loss": 4.806, "step": 1450 }, { "epoch": 0.16144656118824668, "grad_norm": 0.8283946514129639, "learning_rate": 0.000590968645620084, "loss": 4.8075, "step": 1500 }, { "epoch": 0.1668281132278549, "grad_norm": 1.1234453916549683, "learning_rate": 0.0005906454045900226, "loss": 4.7791, "step": 1550 }, { "epoch": 0.17220966526746315, "grad_norm": 0.9318310618400574, "learning_rate": 0.0005903221635599612, "loss": 4.7354, "step": 1600 }, { "epoch": 0.17759121730707136, "grad_norm": 0.8706691861152649, "learning_rate": 0.0005899989225298998, "loss": 4.7278, "step": 1650 }, { "epoch": 0.18297276934667958, "grad_norm": 1.0983690023422241, "learning_rate": 0.0005896756814998383, "loss": 4.7116, "step": 1700 }, { "epoch": 0.1883543213862878, "grad_norm": 1.1487178802490234, "learning_rate": 0.0005893524404697769, "loss": 4.7026, "step": 1750 }, { "epoch": 0.19373587342589602, "grad_norm": 0.8954223394393921, "learning_rate": 0.0005890291994397155, "loss": 4.6491, "step": 1800 }, { "epoch": 0.19911742546550426, "grad_norm": 1.1864057779312134, "learning_rate": 0.0005887059584096541, "loss": 4.6526, "step": 1850 }, { "epoch": 0.20449897750511248, "grad_norm": 0.945884108543396, "learning_rate": 0.0005883827173795926, "loss": 4.6132, "step": 1900 }, { "epoch": 0.2098805295447207, "grad_norm": 1.2601373195648193, "learning_rate": 0.0005880594763495313, "loss": 4.6044, "step": 1950 }, { "epoch": 0.2152620815843289, "grad_norm": 0.9447725415229797, "learning_rate": 0.0005877362353194698, "loss": 4.5877, "step": 2000 }, { "epoch": 0.2152620815843289, "eval_accuracy": 0.27040933128385014, "eval_loss": 4.508606910705566, "eval_runtime": 184.8398, "eval_samples_per_second": 97.441, "eval_steps_per_second": 6.092, "step": 2000 }, { "epoch": 0.22064363362393713, "grad_norm": 0.7851864695549011, "learning_rate": 0.0005874129942894084, "loss": 4.5643, "step": 2050 }, { "epoch": 0.22602518566354537, "grad_norm": 0.862130880355835, "learning_rate": 0.000587089753259347, "loss": 4.5481, "step": 2100 }, { "epoch": 0.2314067377031536, "grad_norm": 0.8470994830131531, "learning_rate": 0.0005867665122292855, "loss": 4.5219, "step": 2150 }, { "epoch": 0.2367882897427618, "grad_norm": 0.8713746070861816, "learning_rate": 0.0005864432711992242, "loss": 4.511, "step": 2200 }, { "epoch": 0.24216984178237003, "grad_norm": 0.8879591226577759, "learning_rate": 0.0005861200301691628, "loss": 4.5126, "step": 2250 }, { "epoch": 0.24755139382197827, "grad_norm": 0.9361649751663208, "learning_rate": 0.0005857967891391014, "loss": 4.4803, "step": 2300 }, { "epoch": 0.2529329458615865, "grad_norm": 0.8480615019798279, "learning_rate": 0.0005854735481090399, "loss": 4.4446, "step": 2350 }, { "epoch": 0.2583144979011947, "grad_norm": 1.0228937864303589, "learning_rate": 0.0005851503070789784, "loss": 4.4508, "step": 2400 }, { "epoch": 0.2636960499408029, "grad_norm": 0.9693055152893066, "learning_rate": 0.0005848270660489171, "loss": 4.4283, "step": 2450 }, { "epoch": 0.26907760198041114, "grad_norm": 0.7247486114501953, "learning_rate": 0.0005845038250188556, "loss": 4.4015, "step": 2500 }, { "epoch": 0.27445915402001936, "grad_norm": 0.7183290123939514, "learning_rate": 0.0005841805839887943, "loss": 4.3918, "step": 2550 }, { "epoch": 0.2798407060596276, "grad_norm": 0.900417149066925, "learning_rate": 0.0005838573429587328, "loss": 4.3701, "step": 2600 }, { "epoch": 0.2852222580992358, "grad_norm": 0.8292120099067688, "learning_rate": 0.0005835341019286715, "loss": 4.3891, "step": 2650 }, { "epoch": 0.29060381013884407, "grad_norm": 0.798037052154541, "learning_rate": 0.00058321086089861, "loss": 4.36, "step": 2700 }, { "epoch": 0.2959853621784523, "grad_norm": 0.8176091313362122, "learning_rate": 0.0005828876198685486, "loss": 4.3491, "step": 2750 }, { "epoch": 0.3013669142180605, "grad_norm": 0.8014965057373047, "learning_rate": 0.0005825643788384872, "loss": 4.3642, "step": 2800 }, { "epoch": 0.3067484662576687, "grad_norm": 1.2503150701522827, "learning_rate": 0.0005822411378084257, "loss": 4.3251, "step": 2850 }, { "epoch": 0.31213001829727693, "grad_norm": 0.7830126285552979, "learning_rate": 0.0005819178967783644, "loss": 4.3251, "step": 2900 }, { "epoch": 0.31751157033688515, "grad_norm": 0.9035398364067078, "learning_rate": 0.0005815946557483029, "loss": 4.3313, "step": 2950 }, { "epoch": 0.32289312237649337, "grad_norm": 0.6857808828353882, "learning_rate": 0.0005812714147182415, "loss": 4.2989, "step": 3000 }, { "epoch": 0.32289312237649337, "eval_accuracy": 0.2990024252411089, "eval_loss": 4.2293620109558105, "eval_runtime": 185.031, "eval_samples_per_second": 97.34, "eval_steps_per_second": 6.085, "step": 3000 }, { "epoch": 0.3282746744161016, "grad_norm": 0.7547754049301147, "learning_rate": 0.0005809481736881801, "loss": 4.296, "step": 3050 }, { "epoch": 0.3336562264557098, "grad_norm": 0.7557756304740906, "learning_rate": 0.0005806249326581187, "loss": 4.2866, "step": 3100 }, { "epoch": 0.3390377784953181, "grad_norm": 0.7948616147041321, "learning_rate": 0.0005803016916280573, "loss": 4.2736, "step": 3150 }, { "epoch": 0.3444193305349263, "grad_norm": 0.864937424659729, "learning_rate": 0.0005799784505979959, "loss": 4.2902, "step": 3200 }, { "epoch": 0.3498008825745345, "grad_norm": 0.8035250902175903, "learning_rate": 0.0005796552095679344, "loss": 4.2524, "step": 3250 }, { "epoch": 0.35518243461414273, "grad_norm": 0.7827973365783691, "learning_rate": 0.000579331968537873, "loss": 4.2573, "step": 3300 }, { "epoch": 0.36056398665375095, "grad_norm": 0.7145817875862122, "learning_rate": 0.0005790087275078116, "loss": 4.233, "step": 3350 }, { "epoch": 0.36594553869335916, "grad_norm": 0.7260717749595642, "learning_rate": 0.0005786854864777502, "loss": 4.2351, "step": 3400 }, { "epoch": 0.3713270907329674, "grad_norm": 0.846215009689331, "learning_rate": 0.0005783622454476888, "loss": 4.236, "step": 3450 }, { "epoch": 0.3767086427725756, "grad_norm": 0.6381611227989197, "learning_rate": 0.0005780390044176273, "loss": 4.2186, "step": 3500 }, { "epoch": 0.3820901948121838, "grad_norm": 0.7176727652549744, "learning_rate": 0.000577715763387566, "loss": 4.2159, "step": 3550 }, { "epoch": 0.38747174685179203, "grad_norm": 0.7199215888977051, "learning_rate": 0.0005773925223575045, "loss": 4.2187, "step": 3600 }, { "epoch": 0.3928532988914003, "grad_norm": 0.7120049595832825, "learning_rate": 0.0005770692813274432, "loss": 4.1858, "step": 3650 }, { "epoch": 0.3982348509310085, "grad_norm": 0.682905912399292, "learning_rate": 0.0005767460402973817, "loss": 4.2067, "step": 3700 }, { "epoch": 0.40361640297061674, "grad_norm": 0.8106050491333008, "learning_rate": 0.0005764227992673203, "loss": 4.1901, "step": 3750 }, { "epoch": 0.40899795501022496, "grad_norm": 0.6139677166938782, "learning_rate": 0.0005760995582372589, "loss": 4.1922, "step": 3800 }, { "epoch": 0.4143795070498332, "grad_norm": 0.6996195912361145, "learning_rate": 0.0005757763172071974, "loss": 4.1884, "step": 3850 }, { "epoch": 0.4197610590894414, "grad_norm": 0.6972913146018982, "learning_rate": 0.000575453076177136, "loss": 4.1469, "step": 3900 }, { "epoch": 0.4251426111290496, "grad_norm": 0.7805142998695374, "learning_rate": 0.0005751298351470746, "loss": 4.1681, "step": 3950 }, { "epoch": 0.4305241631686578, "grad_norm": 0.6750819683074951, "learning_rate": 0.0005748065941170133, "loss": 4.1677, "step": 4000 }, { "epoch": 0.4305241631686578, "eval_accuracy": 0.3131229545414788, "eval_loss": 4.085065841674805, "eval_runtime": 184.7939, "eval_samples_per_second": 97.465, "eval_steps_per_second": 6.093, "step": 4000 }, { "epoch": 0.43590571520826604, "grad_norm": 0.743445098400116, "learning_rate": 0.0005744833530869518, "loss": 4.1605, "step": 4050 }, { "epoch": 0.44128726724787426, "grad_norm": 0.6816270351409912, "learning_rate": 0.0005741601120568903, "loss": 4.1494, "step": 4100 }, { "epoch": 0.44666881928748253, "grad_norm": 0.6031078696250916, "learning_rate": 0.0005738368710268289, "loss": 4.1373, "step": 4150 }, { "epoch": 0.45205037132709075, "grad_norm": 0.869440495967865, "learning_rate": 0.0005735136299967675, "loss": 4.1274, "step": 4200 }, { "epoch": 0.45743192336669897, "grad_norm": 0.6965010762214661, "learning_rate": 0.0005731903889667062, "loss": 4.1363, "step": 4250 }, { "epoch": 0.4628134754063072, "grad_norm": 0.6586082577705383, "learning_rate": 0.0005728671479366447, "loss": 4.1073, "step": 4300 }, { "epoch": 0.4681950274459154, "grad_norm": 0.6008425951004028, "learning_rate": 0.0005725439069065833, "loss": 4.1131, "step": 4350 }, { "epoch": 0.4735765794855236, "grad_norm": 0.6233063340187073, "learning_rate": 0.0005722206658765219, "loss": 4.1266, "step": 4400 }, { "epoch": 0.47895813152513184, "grad_norm": 0.7806559205055237, "learning_rate": 0.0005718974248464605, "loss": 4.1052, "step": 4450 }, { "epoch": 0.48433968356474005, "grad_norm": 0.7604005336761475, "learning_rate": 0.000571574183816399, "loss": 4.0944, "step": 4500 }, { "epoch": 0.48972123560434827, "grad_norm": 0.8495274782180786, "learning_rate": 0.0005712509427863376, "loss": 4.1176, "step": 4550 }, { "epoch": 0.49510278764395654, "grad_norm": 0.6198326349258423, "learning_rate": 0.0005709277017562762, "loss": 4.1016, "step": 4600 }, { "epoch": 0.5004843396835648, "grad_norm": 0.6641437411308289, "learning_rate": 0.0005706044607262148, "loss": 4.0835, "step": 4650 }, { "epoch": 0.505865891723173, "grad_norm": 0.7691634297370911, "learning_rate": 0.0005702812196961534, "loss": 4.0815, "step": 4700 }, { "epoch": 0.5112474437627812, "grad_norm": 0.7156651020050049, "learning_rate": 0.0005699579786660919, "loss": 4.0818, "step": 4750 }, { "epoch": 0.5166289958023894, "grad_norm": 0.6472979784011841, "learning_rate": 0.0005696347376360306, "loss": 4.067, "step": 4800 }, { "epoch": 0.5220105478419976, "grad_norm": 0.724949300289154, "learning_rate": 0.0005693114966059691, "loss": 4.0615, "step": 4850 }, { "epoch": 0.5273920998816058, "grad_norm": 0.7559602856636047, "learning_rate": 0.0005689882555759077, "loss": 4.0676, "step": 4900 }, { "epoch": 0.5327736519212141, "grad_norm": 0.6394193768501282, "learning_rate": 0.0005686650145458463, "loss": 4.0461, "step": 4950 }, { "epoch": 0.5381552039608223, "grad_norm": 0.7001257538795471, "learning_rate": 0.0005683417735157848, "loss": 4.0729, "step": 5000 }, { "epoch": 0.5381552039608223, "eval_accuracy": 0.320991922635667, "eval_loss": 3.990946054458618, "eval_runtime": 184.8283, "eval_samples_per_second": 97.447, "eval_steps_per_second": 6.092, "step": 5000 }, { "epoch": 0.5435367560004305, "grad_norm": 0.7765357494354248, "learning_rate": 0.0005680185324857235, "loss": 4.0425, "step": 5050 }, { "epoch": 0.5489183080400387, "grad_norm": 0.6510383486747742, "learning_rate": 0.0005676952914556621, "loss": 4.0239, "step": 5100 }, { "epoch": 0.5542998600796469, "grad_norm": 0.6882658004760742, "learning_rate": 0.0005673720504256007, "loss": 4.0242, "step": 5150 }, { "epoch": 0.5596814121192552, "grad_norm": 0.7188264727592468, "learning_rate": 0.0005670488093955392, "loss": 4.0442, "step": 5200 }, { "epoch": 0.5650629641588634, "grad_norm": 0.6599631905555725, "learning_rate": 0.0005667255683654777, "loss": 4.0497, "step": 5250 }, { "epoch": 0.5704445161984716, "grad_norm": 0.5674473643302917, "learning_rate": 0.0005664023273354164, "loss": 4.0271, "step": 5300 }, { "epoch": 0.5758260682380799, "grad_norm": 0.6670402884483337, "learning_rate": 0.0005660790863053549, "loss": 4.0284, "step": 5350 }, { "epoch": 0.5812076202776881, "grad_norm": 0.7389664649963379, "learning_rate": 0.0005657558452752936, "loss": 4.0449, "step": 5400 }, { "epoch": 0.5865891723172963, "grad_norm": 0.6841723322868347, "learning_rate": 0.0005654326042452321, "loss": 4.051, "step": 5450 }, { "epoch": 0.5919707243569046, "grad_norm": 0.6119431853294373, "learning_rate": 0.0005651093632151708, "loss": 4.0158, "step": 5500 }, { "epoch": 0.5973522763965128, "grad_norm": 0.8850950598716736, "learning_rate": 0.0005647861221851093, "loss": 4.0183, "step": 5550 }, { "epoch": 0.602733828436121, "grad_norm": 0.6376111507415771, "learning_rate": 0.0005644628811550479, "loss": 4.0076, "step": 5600 }, { "epoch": 0.6081153804757292, "grad_norm": 0.5115376114845276, "learning_rate": 0.0005641396401249865, "loss": 4.0006, "step": 5650 }, { "epoch": 0.6134969325153374, "grad_norm": 0.6021561026573181, "learning_rate": 0.000563816399094925, "loss": 4.0056, "step": 5700 }, { "epoch": 0.6188784845549457, "grad_norm": 0.5795146226882935, "learning_rate": 0.0005634931580648637, "loss": 4.0159, "step": 5750 }, { "epoch": 0.6242600365945539, "grad_norm": 0.6918138861656189, "learning_rate": 0.0005631699170348022, "loss": 3.991, "step": 5800 }, { "epoch": 0.6296415886341621, "grad_norm": 0.6135892271995544, "learning_rate": 0.0005628466760047408, "loss": 4.0127, "step": 5850 }, { "epoch": 0.6350231406737703, "grad_norm": 0.6184191703796387, "learning_rate": 0.0005625234349746794, "loss": 3.9862, "step": 5900 }, { "epoch": 0.6404046927133785, "grad_norm": 0.6259024739265442, "learning_rate": 0.000562200193944618, "loss": 3.9922, "step": 5950 }, { "epoch": 0.6457862447529867, "grad_norm": 0.5602654814720154, "learning_rate": 0.0005618769529145566, "loss": 3.995, "step": 6000 }, { "epoch": 0.6457862447529867, "eval_accuracy": 0.32913100180896193, "eval_loss": 3.9143569469451904, "eval_runtime": 184.5106, "eval_samples_per_second": 97.615, "eval_steps_per_second": 6.103, "step": 6000 }, { "epoch": 0.651167796792595, "grad_norm": 0.7381386160850525, "learning_rate": 0.0005615537118844952, "loss": 3.9749, "step": 6050 }, { "epoch": 0.6565493488322032, "grad_norm": 0.6130694150924683, "learning_rate": 0.0005612304708544337, "loss": 3.9791, "step": 6100 }, { "epoch": 0.6619309008718114, "grad_norm": 0.6282258629798889, "learning_rate": 0.0005609072298243723, "loss": 3.9766, "step": 6150 }, { "epoch": 0.6673124529114196, "grad_norm": 0.7320833802223206, "learning_rate": 0.0005605839887943109, "loss": 3.9744, "step": 6200 }, { "epoch": 0.6726940049510278, "grad_norm": 0.7067676782608032, "learning_rate": 0.0005602607477642495, "loss": 3.9755, "step": 6250 }, { "epoch": 0.6780755569906362, "grad_norm": 0.6241488456726074, "learning_rate": 0.0005599375067341881, "loss": 3.9639, "step": 6300 }, { "epoch": 0.6834571090302444, "grad_norm": 0.6031643152236938, "learning_rate": 0.0005596142657041266, "loss": 3.9535, "step": 6350 }, { "epoch": 0.6888386610698526, "grad_norm": 0.5485600829124451, "learning_rate": 0.0005592910246740653, "loss": 3.9616, "step": 6400 }, { "epoch": 0.6942202131094608, "grad_norm": 0.6337101459503174, "learning_rate": 0.0005589677836440038, "loss": 3.9509, "step": 6450 }, { "epoch": 0.699601765149069, "grad_norm": 0.546737790107727, "learning_rate": 0.0005586445426139425, "loss": 3.9416, "step": 6500 }, { "epoch": 0.7049833171886772, "grad_norm": 0.6616541147232056, "learning_rate": 0.000558321301583881, "loss": 3.9618, "step": 6550 }, { "epoch": 0.7103648692282855, "grad_norm": 0.5739938020706177, "learning_rate": 0.0005579980605538196, "loss": 3.949, "step": 6600 }, { "epoch": 0.7157464212678937, "grad_norm": 0.5481446981430054, "learning_rate": 0.0005576748195237582, "loss": 3.961, "step": 6650 }, { "epoch": 0.7211279733075019, "grad_norm": 0.6824391484260559, "learning_rate": 0.0005573515784936967, "loss": 3.9376, "step": 6700 }, { "epoch": 0.7265095253471101, "grad_norm": 0.6663408279418945, "learning_rate": 0.0005570283374636353, "loss": 3.9525, "step": 6750 }, { "epoch": 0.7318910773867183, "grad_norm": 0.5435226559638977, "learning_rate": 0.0005567050964335739, "loss": 3.9249, "step": 6800 }, { "epoch": 0.7372726294263265, "grad_norm": 0.6284464001655579, "learning_rate": 0.0005563818554035126, "loss": 3.9307, "step": 6850 }, { "epoch": 0.7426541814659348, "grad_norm": 0.5826367139816284, "learning_rate": 0.0005560586143734511, "loss": 3.9251, "step": 6900 }, { "epoch": 0.748035733505543, "grad_norm": 0.5755831003189087, "learning_rate": 0.0005557353733433896, "loss": 3.9271, "step": 6950 }, { "epoch": 0.7534172855451512, "grad_norm": 0.6272289156913757, "learning_rate": 0.0005554121323133283, "loss": 3.9517, "step": 7000 }, { "epoch": 0.7534172855451512, "eval_accuracy": 0.33305206722441094, "eval_loss": 3.8637194633483887, "eval_runtime": 184.9315, "eval_samples_per_second": 97.393, "eval_steps_per_second": 6.089, "step": 7000 }, { "epoch": 0.7587988375847594, "grad_norm": 0.6227930188179016, "learning_rate": 0.0005550888912832668, "loss": 3.916, "step": 7050 }, { "epoch": 0.7641803896243676, "grad_norm": 0.6726216077804565, "learning_rate": 0.0005547656502532055, "loss": 3.9217, "step": 7100 }, { "epoch": 0.7695619416639758, "grad_norm": 0.5420123338699341, "learning_rate": 0.000554442409223144, "loss": 3.9238, "step": 7150 }, { "epoch": 0.7749434937035841, "grad_norm": 0.5879136919975281, "learning_rate": 0.0005541191681930826, "loss": 3.9191, "step": 7200 }, { "epoch": 0.7803250457431924, "grad_norm": 0.5138105750083923, "learning_rate": 0.0005537959271630212, "loss": 3.9145, "step": 7250 }, { "epoch": 0.7857065977828006, "grad_norm": 0.5751107931137085, "learning_rate": 0.0005534726861329598, "loss": 3.9067, "step": 7300 }, { "epoch": 0.7910881498224088, "grad_norm": 0.6342102289199829, "learning_rate": 0.0005531494451028983, "loss": 3.9038, "step": 7350 }, { "epoch": 0.796469701862017, "grad_norm": 0.5363113284111023, "learning_rate": 0.0005528262040728369, "loss": 3.9121, "step": 7400 }, { "epoch": 0.8018512539016253, "grad_norm": 0.6412177681922913, "learning_rate": 0.0005525029630427755, "loss": 3.894, "step": 7450 }, { "epoch": 0.8072328059412335, "grad_norm": 0.6429057121276855, "learning_rate": 0.0005521797220127141, "loss": 3.904, "step": 7500 }, { "epoch": 0.8126143579808417, "grad_norm": 0.6362873315811157, "learning_rate": 0.0005518564809826527, "loss": 3.9328, "step": 7550 }, { "epoch": 0.8179959100204499, "grad_norm": 0.6349565386772156, "learning_rate": 0.0005515332399525912, "loss": 3.8807, "step": 7600 }, { "epoch": 0.8233774620600581, "grad_norm": 0.6769669651985168, "learning_rate": 0.0005512099989225299, "loss": 3.8932, "step": 7650 }, { "epoch": 0.8287590140996663, "grad_norm": 0.601761519908905, "learning_rate": 0.0005508867578924685, "loss": 3.897, "step": 7700 }, { "epoch": 0.8341405661392746, "grad_norm": 0.6315547823905945, "learning_rate": 0.000550563516862407, "loss": 3.8814, "step": 7750 }, { "epoch": 0.8395221181788828, "grad_norm": 0.5754820108413696, "learning_rate": 0.0005502402758323456, "loss": 3.9101, "step": 7800 }, { "epoch": 0.844903670218491, "grad_norm": 0.627307653427124, "learning_rate": 0.0005499170348022841, "loss": 3.9035, "step": 7850 }, { "epoch": 0.8502852222580992, "grad_norm": 0.6498885154724121, "learning_rate": 0.0005495937937722228, "loss": 3.9012, "step": 7900 }, { "epoch": 0.8556667742977074, "grad_norm": 0.5503144264221191, "learning_rate": 0.0005492705527421614, "loss": 3.8942, "step": 7950 }, { "epoch": 0.8610483263373157, "grad_norm": 0.6200529336929321, "learning_rate": 0.0005489473117121, "loss": 3.8717, "step": 8000 }, { "epoch": 0.8610483263373157, "eval_accuracy": 0.3380203291726159, "eval_loss": 3.8143210411071777, "eval_runtime": 184.8318, "eval_samples_per_second": 97.445, "eval_steps_per_second": 6.092, "step": 8000 }, { "epoch": 0.8664298783769239, "grad_norm": 0.7157406210899353, "learning_rate": 0.0005486240706820385, "loss": 3.8746, "step": 8050 }, { "epoch": 0.8718114304165321, "grad_norm": 0.643976628780365, "learning_rate": 0.0005483008296519772, "loss": 3.8859, "step": 8100 }, { "epoch": 0.8771929824561403, "grad_norm": 0.5374308824539185, "learning_rate": 0.0005479775886219157, "loss": 3.8586, "step": 8150 }, { "epoch": 0.8825745344957485, "grad_norm": 0.6169512867927551, "learning_rate": 0.0005476543475918542, "loss": 3.8734, "step": 8200 }, { "epoch": 0.8879560865353568, "grad_norm": 0.6148831248283386, "learning_rate": 0.0005473311065617929, "loss": 3.8731, "step": 8250 }, { "epoch": 0.8933376385749651, "grad_norm": 0.5680239796638489, "learning_rate": 0.0005470078655317314, "loss": 3.8674, "step": 8300 }, { "epoch": 0.8987191906145733, "grad_norm": 0.5679119825363159, "learning_rate": 0.0005466846245016701, "loss": 3.8556, "step": 8350 }, { "epoch": 0.9041007426541815, "grad_norm": 0.5395457744598389, "learning_rate": 0.0005463613834716086, "loss": 3.8536, "step": 8400 }, { "epoch": 0.9094822946937897, "grad_norm": 0.5603843331336975, "learning_rate": 0.0005460381424415472, "loss": 3.8579, "step": 8450 }, { "epoch": 0.9148638467333979, "grad_norm": 0.5550517439842224, "learning_rate": 0.0005457149014114858, "loss": 3.858, "step": 8500 }, { "epoch": 0.9202453987730062, "grad_norm": 0.6112827062606812, "learning_rate": 0.0005453916603814243, "loss": 3.8563, "step": 8550 }, { "epoch": 0.9256269508126144, "grad_norm": 0.6317242383956909, "learning_rate": 0.000545068419351363, "loss": 3.854, "step": 8600 }, { "epoch": 0.9310085028522226, "grad_norm": 0.574309766292572, "learning_rate": 0.0005447451783213015, "loss": 3.8455, "step": 8650 }, { "epoch": 0.9363900548918308, "grad_norm": 0.5467745065689087, "learning_rate": 0.0005444219372912401, "loss": 3.858, "step": 8700 }, { "epoch": 0.941771606931439, "grad_norm": 0.5530813932418823, "learning_rate": 0.0005440986962611787, "loss": 3.8518, "step": 8750 }, { "epoch": 0.9471531589710472, "grad_norm": 0.7071423530578613, "learning_rate": 0.0005437754552311173, "loss": 3.8473, "step": 8800 }, { "epoch": 0.9525347110106555, "grad_norm": 0.5127449631690979, "learning_rate": 0.0005434522142010559, "loss": 3.8471, "step": 8850 }, { "epoch": 0.9579162630502637, "grad_norm": 0.5855455994606018, "learning_rate": 0.0005431289731709945, "loss": 3.8589, "step": 8900 }, { "epoch": 0.9632978150898719, "grad_norm": 0.6145156025886536, "learning_rate": 0.000542805732140933, "loss": 3.8295, "step": 8950 }, { "epoch": 0.9686793671294801, "grad_norm": 0.5550306439399719, "learning_rate": 0.0005424824911108716, "loss": 3.8378, "step": 9000 }, { "epoch": 0.9686793671294801, "eval_accuracy": 0.3415591537287335, "eval_loss": 3.7806262969970703, "eval_runtime": 184.435, "eval_samples_per_second": 97.655, "eval_steps_per_second": 6.105, "step": 9000 }, { "epoch": 0.9740609191690883, "grad_norm": 0.5801270008087158, "learning_rate": 0.0005421592500808102, "loss": 3.8502, "step": 9050 }, { "epoch": 0.9794424712086965, "grad_norm": 0.5607963800430298, "learning_rate": 0.0005418360090507488, "loss": 3.825, "step": 9100 }, { "epoch": 0.9848240232483048, "grad_norm": 0.5534636974334717, "learning_rate": 0.0005415127680206874, "loss": 3.8361, "step": 9150 }, { "epoch": 0.9902055752879131, "grad_norm": 0.692423939704895, "learning_rate": 0.0005411895269906259, "loss": 3.8318, "step": 9200 }, { "epoch": 0.9955871273275213, "grad_norm": 0.5722038745880127, "learning_rate": 0.0005408662859605646, "loss": 3.8376, "step": 9250 }, { "epoch": 1.0009686793671295, "grad_norm": 0.6226180195808411, "learning_rate": 0.0005405430449305031, "loss": 3.8294, "step": 9300 }, { "epoch": 1.0063502314067376, "grad_norm": 0.6096346974372864, "learning_rate": 0.0005402198039004416, "loss": 3.7878, "step": 9350 }, { "epoch": 1.011731783446346, "grad_norm": 0.5371617078781128, "learning_rate": 0.0005398965628703803, "loss": 3.7627, "step": 9400 }, { "epoch": 1.017113335485954, "grad_norm": 0.6130861639976501, "learning_rate": 0.0005395733218403189, "loss": 3.769, "step": 9450 }, { "epoch": 1.0224948875255624, "grad_norm": 0.5652937889099121, "learning_rate": 0.0005392500808102575, "loss": 3.7714, "step": 9500 }, { "epoch": 1.0278764395651705, "grad_norm": 0.5810158252716064, "learning_rate": 0.000538926839780196, "loss": 3.7566, "step": 9550 }, { "epoch": 1.0332579916047788, "grad_norm": 0.6366006731987, "learning_rate": 0.0005386035987501346, "loss": 3.7647, "step": 9600 }, { "epoch": 1.0386395436443872, "grad_norm": 0.5298298597335815, "learning_rate": 0.0005382803577200732, "loss": 3.7488, "step": 9650 }, { "epoch": 1.0440210956839953, "grad_norm": 1.2535181045532227, "learning_rate": 0.0005379571166900119, "loss": 3.7539, "step": 9700 }, { "epoch": 1.0494026477236036, "grad_norm": 0.601189374923706, "learning_rate": 0.0005376338756599504, "loss": 3.7654, "step": 9750 }, { "epoch": 1.0547841997632117, "grad_norm": 0.6602137684822083, "learning_rate": 0.0005373106346298889, "loss": 3.7686, "step": 9800 }, { "epoch": 1.06016575180282, "grad_norm": 0.6176168918609619, "learning_rate": 0.0005369873935998276, "loss": 3.7589, "step": 9850 }, { "epoch": 1.0655473038424281, "grad_norm": 0.6260287165641785, "learning_rate": 0.0005366641525697661, "loss": 3.7655, "step": 9900 }, { "epoch": 1.0709288558820365, "grad_norm": 0.5720372796058655, "learning_rate": 0.0005363409115397048, "loss": 3.7559, "step": 9950 }, { "epoch": 1.0763104079216446, "grad_norm": 0.5716636180877686, "learning_rate": 0.0005360176705096433, "loss": 3.7573, "step": 10000 }, { "epoch": 1.0763104079216446, "eval_accuracy": 0.3451164492757796, "eval_loss": 3.7458271980285645, "eval_runtime": 185.1344, "eval_samples_per_second": 97.286, "eval_steps_per_second": 6.082, "step": 10000 }, { "epoch": 1.081691959961253, "grad_norm": 0.6392249464988708, "learning_rate": 0.0005356944294795819, "loss": 3.746, "step": 10050 }, { "epoch": 1.087073512000861, "grad_norm": 0.5700231194496155, "learning_rate": 0.0005353711884495205, "loss": 3.7913, "step": 10100 }, { "epoch": 1.0924550640404693, "grad_norm": 0.5578026175498962, "learning_rate": 0.000535047947419459, "loss": 3.7517, "step": 10150 }, { "epoch": 1.0978366160800774, "grad_norm": 0.5823580026626587, "learning_rate": 0.0005347247063893976, "loss": 3.7495, "step": 10200 }, { "epoch": 1.1032181681196858, "grad_norm": 0.6097853779792786, "learning_rate": 0.0005344014653593362, "loss": 3.748, "step": 10250 }, { "epoch": 1.1085997201592939, "grad_norm": 0.6023664474487305, "learning_rate": 0.0005340782243292748, "loss": 3.7464, "step": 10300 }, { "epoch": 1.1139812721989022, "grad_norm": 0.6338013410568237, "learning_rate": 0.0005337549832992134, "loss": 3.7546, "step": 10350 }, { "epoch": 1.1193628242385103, "grad_norm": 0.5424166917800903, "learning_rate": 0.000533431742269152, "loss": 3.7676, "step": 10400 }, { "epoch": 1.1247443762781186, "grad_norm": 0.5553328990936279, "learning_rate": 0.0005331085012390905, "loss": 3.7516, "step": 10450 }, { "epoch": 1.1301259283177267, "grad_norm": 0.5472305417060852, "learning_rate": 0.0005327852602090292, "loss": 3.7578, "step": 10500 }, { "epoch": 1.135507480357335, "grad_norm": 0.6304797530174255, "learning_rate": 0.0005324620191789678, "loss": 3.7302, "step": 10550 }, { "epoch": 1.1408890323969434, "grad_norm": 0.5144755244255066, "learning_rate": 0.0005321387781489063, "loss": 3.7487, "step": 10600 }, { "epoch": 1.1462705844365515, "grad_norm": 0.5785931348800659, "learning_rate": 0.0005318155371188449, "loss": 3.7557, "step": 10650 }, { "epoch": 1.1516521364761596, "grad_norm": 0.5502986907958984, "learning_rate": 0.0005314922960887834, "loss": 3.7452, "step": 10700 }, { "epoch": 1.157033688515768, "grad_norm": 0.6079390048980713, "learning_rate": 0.0005311690550587221, "loss": 3.7587, "step": 10750 }, { "epoch": 1.1624152405553763, "grad_norm": 0.5040614008903503, "learning_rate": 0.0005308458140286607, "loss": 3.7611, "step": 10800 }, { "epoch": 1.1677967925949844, "grad_norm": 0.5929045677185059, "learning_rate": 0.0005305225729985993, "loss": 3.7669, "step": 10850 }, { "epoch": 1.1731783446345927, "grad_norm": 0.535692572593689, "learning_rate": 0.0005301993319685378, "loss": 3.764, "step": 10900 }, { "epoch": 1.1785598966742008, "grad_norm": 0.5377613306045532, "learning_rate": 0.0005298760909384765, "loss": 3.7536, "step": 10950 }, { "epoch": 1.1839414487138091, "grad_norm": 0.6084091663360596, "learning_rate": 0.000529552849908415, "loss": 3.7359, "step": 11000 }, { "epoch": 1.1839414487138091, "eval_accuracy": 0.3476050350182825, "eval_loss": 3.722259044647217, "eval_runtime": 184.9277, "eval_samples_per_second": 97.395, "eval_steps_per_second": 6.089, "step": 11000 }, { "epoch": 1.1893230007534172, "grad_norm": 0.5570613741874695, "learning_rate": 0.0005292296088783535, "loss": 3.7553, "step": 11050 }, { "epoch": 1.1947045527930256, "grad_norm": 0.5923260450363159, "learning_rate": 0.0005289063678482922, "loss": 3.7489, "step": 11100 }, { "epoch": 1.2000861048326337, "grad_norm": 0.6002750396728516, "learning_rate": 0.0005285831268182307, "loss": 3.742, "step": 11150 }, { "epoch": 1.205467656872242, "grad_norm": 0.6400654911994934, "learning_rate": 0.0005282598857881694, "loss": 3.7462, "step": 11200 }, { "epoch": 1.21084920891185, "grad_norm": 0.5764461755752563, "learning_rate": 0.0005279366447581079, "loss": 3.7248, "step": 11250 }, { "epoch": 1.2162307609514584, "grad_norm": 0.5291248559951782, "learning_rate": 0.0005276134037280465, "loss": 3.7395, "step": 11300 }, { "epoch": 1.2216123129910665, "grad_norm": 0.5363642573356628, "learning_rate": 0.0005272966275185863, "loss": 3.7486, "step": 11350 }, { "epoch": 1.2269938650306749, "grad_norm": 0.5614747405052185, "learning_rate": 0.0005269733864885249, "loss": 3.7352, "step": 11400 }, { "epoch": 1.232375417070283, "grad_norm": 0.568588137626648, "learning_rate": 0.0005266501454584636, "loss": 3.7354, "step": 11450 }, { "epoch": 1.2377569691098913, "grad_norm": 0.5658771991729736, "learning_rate": 0.0005263269044284021, "loss": 3.7457, "step": 11500 }, { "epoch": 1.2431385211494996, "grad_norm": 0.5699712038040161, "learning_rate": 0.0005260036633983406, "loss": 3.7583, "step": 11550 }, { "epoch": 1.2485200731891077, "grad_norm": 0.6179584264755249, "learning_rate": 0.0005256804223682792, "loss": 3.7221, "step": 11600 }, { "epoch": 1.2539016252287158, "grad_norm": 0.5554487109184265, "learning_rate": 0.0005253571813382178, "loss": 3.7318, "step": 11650 }, { "epoch": 1.2592831772683242, "grad_norm": 0.5801360011100769, "learning_rate": 0.0005250339403081564, "loss": 3.7272, "step": 11700 }, { "epoch": 1.2646647293079325, "grad_norm": 0.58757084608078, "learning_rate": 0.000524710699278095, "loss": 3.7316, "step": 11750 }, { "epoch": 1.2700462813475406, "grad_norm": 0.6320570111274719, "learning_rate": 0.0005243874582480336, "loss": 3.7323, "step": 11800 }, { "epoch": 1.275427833387149, "grad_norm": 0.5550466179847717, "learning_rate": 0.0005240642172179722, "loss": 3.7152, "step": 11850 }, { "epoch": 1.280809385426757, "grad_norm": 0.7199887633323669, "learning_rate": 0.0005237409761879107, "loss": 3.724, "step": 11900 }, { "epoch": 1.2861909374663654, "grad_norm": 0.5186243653297424, "learning_rate": 0.0005234177351578493, "loss": 3.7295, "step": 11950 }, { "epoch": 1.2915724895059735, "grad_norm": 0.5249941945075989, "learning_rate": 0.0005230944941277878, "loss": 3.7189, "step": 12000 }, { "epoch": 1.2915724895059735, "eval_accuracy": 0.34949287894405906, "eval_loss": 3.6969494819641113, "eval_runtime": 184.7681, "eval_samples_per_second": 97.479, "eval_steps_per_second": 6.094, "step": 12000 }, { "epoch": 1.2969540415455818, "grad_norm": 0.5758280158042908, "learning_rate": 0.0005227712530977265, "loss": 3.7271, "step": 12050 }, { "epoch": 1.30233559358519, "grad_norm": 0.5738679766654968, "learning_rate": 0.0005224480120676651, "loss": 3.7388, "step": 12100 }, { "epoch": 1.3077171456247982, "grad_norm": 0.6011477708816528, "learning_rate": 0.0005221247710376037, "loss": 3.7381, "step": 12150 }, { "epoch": 1.3130986976644063, "grad_norm": 0.5592882037162781, "learning_rate": 0.0005218015300075422, "loss": 3.7326, "step": 12200 }, { "epoch": 1.3184802497040147, "grad_norm": 0.5867257714271545, "learning_rate": 0.0005214782889774809, "loss": 3.7186, "step": 12250 }, { "epoch": 1.3238618017436228, "grad_norm": 0.5920559167861938, "learning_rate": 0.0005211550479474194, "loss": 3.7018, "step": 12300 }, { "epoch": 1.329243353783231, "grad_norm": 0.5726664066314697, "learning_rate": 0.000520831806917358, "loss": 3.7209, "step": 12350 }, { "epoch": 1.3346249058228392, "grad_norm": 0.5191407799720764, "learning_rate": 0.0005205085658872966, "loss": 3.7221, "step": 12400 }, { "epoch": 1.3400064578624475, "grad_norm": 0.5495452284812927, "learning_rate": 0.0005201853248572351, "loss": 3.7186, "step": 12450 }, { "epoch": 1.3453880099020559, "grad_norm": 0.536133348941803, "learning_rate": 0.0005198620838271738, "loss": 3.727, "step": 12500 }, { "epoch": 1.350769561941664, "grad_norm": 0.6578084826469421, "learning_rate": 0.0005195388427971123, "loss": 3.7175, "step": 12550 }, { "epoch": 1.356151113981272, "grad_norm": 0.697626531124115, "learning_rate": 0.000519215601767051, "loss": 3.7129, "step": 12600 }, { "epoch": 1.3615326660208804, "grad_norm": 0.5493432879447937, "learning_rate": 0.0005188923607369895, "loss": 3.7113, "step": 12650 }, { "epoch": 1.3669142180604887, "grad_norm": 0.5725867748260498, "learning_rate": 0.000518569119706928, "loss": 3.7209, "step": 12700 }, { "epoch": 1.3722957701000968, "grad_norm": 0.5644043684005737, "learning_rate": 0.0005182458786768667, "loss": 3.7148, "step": 12750 }, { "epoch": 1.3776773221397052, "grad_norm": 0.516650378704071, "learning_rate": 0.0005179226376468052, "loss": 3.7039, "step": 12800 }, { "epoch": 1.3830588741793133, "grad_norm": 0.5859173536300659, "learning_rate": 0.0005175993966167438, "loss": 3.7156, "step": 12850 }, { "epoch": 1.3884404262189216, "grad_norm": 0.4950462579727173, "learning_rate": 0.0005172761555866824, "loss": 3.7052, "step": 12900 }, { "epoch": 1.3938219782585297, "grad_norm": 0.6020005941390991, "learning_rate": 0.0005169529145566211, "loss": 3.7147, "step": 12950 }, { "epoch": 1.399203530298138, "grad_norm": 0.5504841208457947, "learning_rate": 0.0005166296735265596, "loss": 3.7161, "step": 13000 }, { "epoch": 1.399203530298138, "eval_accuracy": 0.3519344179861383, "eval_loss": 3.6728596687316895, "eval_runtime": 184.9723, "eval_samples_per_second": 97.371, "eval_steps_per_second": 6.087, "step": 13000 }, { "epoch": 1.4045850823377461, "grad_norm": 0.5823192000389099, "learning_rate": 0.0005163064324964982, "loss": 3.7071, "step": 13050 }, { "epoch": 1.4099666343773545, "grad_norm": 0.5856388807296753, "learning_rate": 0.0005159831914664367, "loss": 3.6995, "step": 13100 }, { "epoch": 1.4153481864169626, "grad_norm": 0.6325433850288391, "learning_rate": 0.0005156599504363753, "loss": 3.717, "step": 13150 }, { "epoch": 1.420729738456571, "grad_norm": 0.6164390444755554, "learning_rate": 0.000515336709406314, "loss": 3.7094, "step": 13200 }, { "epoch": 1.426111290496179, "grad_norm": 0.5653883218765259, "learning_rate": 0.0005150134683762525, "loss": 3.6838, "step": 13250 }, { "epoch": 1.4314928425357873, "grad_norm": 0.6306009888648987, "learning_rate": 0.0005146902273461911, "loss": 3.7257, "step": 13300 }, { "epoch": 1.4368743945753955, "grad_norm": 0.5829269886016846, "learning_rate": 0.0005143669863161297, "loss": 3.7158, "step": 13350 }, { "epoch": 1.4422559466150038, "grad_norm": 0.5895189642906189, "learning_rate": 0.0005140437452860683, "loss": 3.7031, "step": 13400 }, { "epoch": 1.447637498654612, "grad_norm": 0.5675577521324158, "learning_rate": 0.0005137205042560069, "loss": 3.7066, "step": 13450 }, { "epoch": 1.4530190506942202, "grad_norm": 0.5363256335258484, "learning_rate": 0.0005133972632259455, "loss": 3.6979, "step": 13500 }, { "epoch": 1.4584006027338283, "grad_norm": 0.6230180859565735, "learning_rate": 0.000513074022195884, "loss": 3.7036, "step": 13550 }, { "epoch": 1.4637821547734367, "grad_norm": 0.566116988658905, "learning_rate": 0.0005127507811658226, "loss": 3.7168, "step": 13600 }, { "epoch": 1.469163706813045, "grad_norm": 0.5180698037147522, "learning_rate": 0.0005124275401357612, "loss": 3.7075, "step": 13650 }, { "epoch": 1.474545258852653, "grad_norm": 0.743535578250885, "learning_rate": 0.0005121042991056997, "loss": 3.7021, "step": 13700 }, { "epoch": 1.4799268108922612, "grad_norm": 0.5717049241065979, "learning_rate": 0.0005117810580756384, "loss": 3.7071, "step": 13750 }, { "epoch": 1.4853083629318695, "grad_norm": 0.5536872148513794, "learning_rate": 0.0005114578170455769, "loss": 3.7012, "step": 13800 }, { "epoch": 1.4906899149714778, "grad_norm": 0.5160924196243286, "learning_rate": 0.0005111345760155156, "loss": 3.7051, "step": 13850 }, { "epoch": 1.496071467011086, "grad_norm": 0.5639503598213196, "learning_rate": 0.0005108113349854541, "loss": 3.6988, "step": 13900 }, { "epoch": 1.501453019050694, "grad_norm": 0.5460876822471619, "learning_rate": 0.0005104880939553926, "loss": 3.6936, "step": 13950 }, { "epoch": 1.5068345710903024, "grad_norm": 0.5820977091789246, "learning_rate": 0.0005101648529253313, "loss": 3.7014, "step": 14000 }, { "epoch": 1.5068345710903024, "eval_accuracy": 0.3535329192716649, "eval_loss": 3.655021905899048, "eval_runtime": 184.7328, "eval_samples_per_second": 97.498, "eval_steps_per_second": 6.095, "step": 14000 }, { "epoch": 1.5122161231299107, "grad_norm": 0.5606687664985657, "learning_rate": 0.0005098416118952699, "loss": 3.6949, "step": 14050 }, { "epoch": 1.5175976751695188, "grad_norm": 0.5651271343231201, "learning_rate": 0.0005095183708652085, "loss": 3.6798, "step": 14100 }, { "epoch": 1.5229792272091272, "grad_norm": 0.5150020122528076, "learning_rate": 0.000509195129835147, "loss": 3.68, "step": 14150 }, { "epoch": 1.5283607792487355, "grad_norm": 0.5977632999420166, "learning_rate": 0.0005088718888050856, "loss": 3.6851, "step": 14200 }, { "epoch": 1.5337423312883436, "grad_norm": 0.6188676953315735, "learning_rate": 0.0005085486477750242, "loss": 3.6703, "step": 14250 }, { "epoch": 1.5391238833279517, "grad_norm": 0.5410431027412415, "learning_rate": 0.0005082254067449629, "loss": 3.6897, "step": 14300 }, { "epoch": 1.54450543536756, "grad_norm": 0.5880251526832581, "learning_rate": 0.0005079021657149014, "loss": 3.7039, "step": 14350 }, { "epoch": 1.5498869874071683, "grad_norm": 0.5555596947669983, "learning_rate": 0.0005075789246848399, "loss": 3.6537, "step": 14400 }, { "epoch": 1.5552685394467765, "grad_norm": 0.5395658612251282, "learning_rate": 0.0005072556836547785, "loss": 3.6913, "step": 14450 }, { "epoch": 1.5606500914863846, "grad_norm": 0.5492722392082214, "learning_rate": 0.0005069324426247171, "loss": 3.6815, "step": 14500 }, { "epoch": 1.566031643525993, "grad_norm": 0.5226956605911255, "learning_rate": 0.0005066092015946557, "loss": 3.6833, "step": 14550 }, { "epoch": 1.5714131955656012, "grad_norm": 0.5579638481140137, "learning_rate": 0.0005062859605645943, "loss": 3.6818, "step": 14600 }, { "epoch": 1.5767947476052093, "grad_norm": 0.5651870965957642, "learning_rate": 0.0005059627195345329, "loss": 3.6777, "step": 14650 }, { "epoch": 1.5821762996448174, "grad_norm": 0.5880829095840454, "learning_rate": 0.0005056394785044715, "loss": 3.6853, "step": 14700 }, { "epoch": 1.5875578516844258, "grad_norm": 0.6156719326972961, "learning_rate": 0.00050531623747441, "loss": 3.6936, "step": 14750 }, { "epoch": 1.592939403724034, "grad_norm": 0.5112903118133545, "learning_rate": 0.0005049929964443486, "loss": 3.6819, "step": 14800 }, { "epoch": 1.5983209557636422, "grad_norm": 0.6731923818588257, "learning_rate": 0.0005046697554142871, "loss": 3.6723, "step": 14850 }, { "epoch": 1.6037025078032503, "grad_norm": 0.6107453107833862, "learning_rate": 0.0005043465143842258, "loss": 3.7043, "step": 14900 }, { "epoch": 1.6090840598428586, "grad_norm": 0.6314976811408997, "learning_rate": 0.0005040232733541644, "loss": 3.6723, "step": 14950 }, { "epoch": 1.614465611882467, "grad_norm": 0.5651878714561462, "learning_rate": 0.000503700032324103, "loss": 3.684, "step": 15000 }, { "epoch": 1.614465611882467, "eval_accuracy": 0.35533384088718994, "eval_loss": 3.6347434520721436, "eval_runtime": 184.7829, "eval_samples_per_second": 97.471, "eval_steps_per_second": 6.094, "step": 15000 }, { "epoch": 1.619847163922075, "grad_norm": 0.5857282280921936, "learning_rate": 0.0005033767912940415, "loss": 3.6826, "step": 15050 }, { "epoch": 1.6252287159616834, "grad_norm": 0.5991663336753845, "learning_rate": 0.0005030535502639802, "loss": 3.6961, "step": 15100 }, { "epoch": 1.6306102680012917, "grad_norm": 0.5337246060371399, "learning_rate": 0.0005027303092339187, "loss": 3.6949, "step": 15150 }, { "epoch": 1.6359918200408998, "grad_norm": 0.6633757948875427, "learning_rate": 0.0005024070682038573, "loss": 3.6859, "step": 15200 }, { "epoch": 1.641373372080508, "grad_norm": 0.5476294755935669, "learning_rate": 0.0005020838271737959, "loss": 3.6768, "step": 15250 }, { "epoch": 1.6467549241201163, "grad_norm": 0.5833688974380493, "learning_rate": 0.0005017605861437344, "loss": 3.6723, "step": 15300 }, { "epoch": 1.6521364761597246, "grad_norm": 0.5485004782676697, "learning_rate": 0.0005014373451136731, "loss": 3.6814, "step": 15350 }, { "epoch": 1.6575180281993327, "grad_norm": 0.5738468170166016, "learning_rate": 0.0005011205689042129, "loss": 3.6761, "step": 15400 }, { "epoch": 1.6628995802389408, "grad_norm": 0.5266857147216797, "learning_rate": 0.0005007973278741514, "loss": 3.683, "step": 15450 }, { "epoch": 1.6682811322785491, "grad_norm": 0.5718429088592529, "learning_rate": 0.00050047408684409, "loss": 3.6676, "step": 15500 }, { "epoch": 1.6736626843181575, "grad_norm": 0.5431683659553528, "learning_rate": 0.0005001508458140286, "loss": 3.6648, "step": 15550 }, { "epoch": 1.6790442363577656, "grad_norm": 0.6572660803794861, "learning_rate": 0.0004998276047839673, "loss": 3.6898, "step": 15600 }, { "epoch": 1.6844257883973737, "grad_norm": 0.5896181464195251, "learning_rate": 0.0004995043637539058, "loss": 3.6605, "step": 15650 }, { "epoch": 1.689807340436982, "grad_norm": 0.549990713596344, "learning_rate": 0.0004991811227238443, "loss": 3.657, "step": 15700 }, { "epoch": 1.6951888924765903, "grad_norm": 0.5397534966468811, "learning_rate": 0.0004988578816937829, "loss": 3.6706, "step": 15750 }, { "epoch": 1.7005704445161984, "grad_norm": 0.5501967668533325, "learning_rate": 0.0004985346406637215, "loss": 3.6697, "step": 15800 }, { "epoch": 1.7059519965558065, "grad_norm": 0.6153730154037476, "learning_rate": 0.0004982113996336602, "loss": 3.6671, "step": 15850 }, { "epoch": 1.7113335485954149, "grad_norm": 0.619310200214386, "learning_rate": 0.0004978881586035987, "loss": 3.6858, "step": 15900 }, { "epoch": 1.7167151006350232, "grad_norm": 0.5859758853912354, "learning_rate": 0.0004975649175735373, "loss": 3.6639, "step": 15950 }, { "epoch": 1.7220966526746313, "grad_norm": 0.5947201251983643, "learning_rate": 0.0004972416765434759, "loss": 3.6546, "step": 16000 }, { "epoch": 1.7220966526746313, "eval_accuracy": 0.3573106715280866, "eval_loss": 3.6192057132720947, "eval_runtime": 184.9464, "eval_samples_per_second": 97.385, "eval_steps_per_second": 6.088, "step": 16000 }, { "epoch": 1.7274782047142396, "grad_norm": 0.6014895439147949, "learning_rate": 0.0004969184355134145, "loss": 3.6555, "step": 16050 }, { "epoch": 1.732859756753848, "grad_norm": 0.6242139935493469, "learning_rate": 0.0004965951944833531, "loss": 3.6725, "step": 16100 }, { "epoch": 1.738241308793456, "grad_norm": 0.6019642949104309, "learning_rate": 0.0004962719534532916, "loss": 3.6383, "step": 16150 }, { "epoch": 1.7436228608330642, "grad_norm": 0.5532971620559692, "learning_rate": 0.0004959487124232302, "loss": 3.6586, "step": 16200 }, { "epoch": 1.7490044128726725, "grad_norm": 0.684308648109436, "learning_rate": 0.0004956254713931688, "loss": 3.6486, "step": 16250 }, { "epoch": 1.7543859649122808, "grad_norm": 0.5315883755683899, "learning_rate": 0.0004953022303631074, "loss": 3.6488, "step": 16300 }, { "epoch": 1.759767516951889, "grad_norm": 0.5260655879974365, "learning_rate": 0.0004949789893330459, "loss": 3.6647, "step": 16350 }, { "epoch": 1.765149068991497, "grad_norm": 0.5171688795089722, "learning_rate": 0.0004946557483029846, "loss": 3.6572, "step": 16400 }, { "epoch": 1.7705306210311054, "grad_norm": 0.5283517241477966, "learning_rate": 0.0004943325072729231, "loss": 3.6654, "step": 16450 }, { "epoch": 1.7759121730707137, "grad_norm": 0.5800235867500305, "learning_rate": 0.0004940092662428617, "loss": 3.673, "step": 16500 }, { "epoch": 1.7812937251103218, "grad_norm": 0.6345435380935669, "learning_rate": 0.0004936860252128003, "loss": 3.671, "step": 16550 }, { "epoch": 1.78667527714993, "grad_norm": 0.5385580658912659, "learning_rate": 0.0004933627841827388, "loss": 3.6765, "step": 16600 }, { "epoch": 1.7920568291895382, "grad_norm": 0.521317720413208, "learning_rate": 0.0004930395431526775, "loss": 3.6577, "step": 16650 }, { "epoch": 1.7974383812291466, "grad_norm": 0.596302330493927, "learning_rate": 0.0004927163021226161, "loss": 3.6477, "step": 16700 }, { "epoch": 1.8028199332687547, "grad_norm": 0.5540564656257629, "learning_rate": 0.0004923930610925547, "loss": 3.6626, "step": 16750 }, { "epoch": 1.8082014853083628, "grad_norm": 0.5950270295143127, "learning_rate": 0.0004920698200624932, "loss": 3.6484, "step": 16800 }, { "epoch": 1.813583037347971, "grad_norm": 0.5792022943496704, "learning_rate": 0.0004917465790324317, "loss": 3.6412, "step": 16850 }, { "epoch": 1.8189645893875794, "grad_norm": 0.5268118381500244, "learning_rate": 0.0004914233380023704, "loss": 3.6561, "step": 16900 }, { "epoch": 1.8243461414271875, "grad_norm": 0.5259914994239807, "learning_rate": 0.0004911000969723089, "loss": 3.6361, "step": 16950 }, { "epoch": 1.8297276934667959, "grad_norm": 0.5924632549285889, "learning_rate": 0.0004907768559422476, "loss": 3.6475, "step": 17000 }, { "epoch": 1.8297276934667959, "eval_accuracy": 0.3591788492811688, "eval_loss": 3.601672410964966, "eval_runtime": 184.7368, "eval_samples_per_second": 97.495, "eval_steps_per_second": 6.095, "step": 17000 }, { "epoch": 1.8351092455064042, "grad_norm": 0.5890408754348755, "learning_rate": 0.0004904536149121861, "loss": 3.6432, "step": 17050 }, { "epoch": 1.8404907975460123, "grad_norm": 0.5994479656219482, "learning_rate": 0.0004901303738821248, "loss": 3.645, "step": 17100 }, { "epoch": 1.8458723495856204, "grad_norm": 0.5794470310211182, "learning_rate": 0.0004898071328520633, "loss": 3.6375, "step": 17150 }, { "epoch": 1.8512539016252287, "grad_norm": 0.5518653392791748, "learning_rate": 0.0004894838918220019, "loss": 3.6343, "step": 17200 }, { "epoch": 1.856635453664837, "grad_norm": 0.5658548474311829, "learning_rate": 0.0004891606507919405, "loss": 3.6607, "step": 17250 }, { "epoch": 1.8620170057044452, "grad_norm": 0.5588201880455017, "learning_rate": 0.000488837409761879, "loss": 3.6551, "step": 17300 }, { "epoch": 1.8673985577440533, "grad_norm": 0.5978080034255981, "learning_rate": 0.0004885141687318177, "loss": 3.6435, "step": 17350 }, { "epoch": 1.8727801097836616, "grad_norm": 0.5438842177391052, "learning_rate": 0.00048819092770175623, "loss": 3.6491, "step": 17400 }, { "epoch": 1.87816166182327, "grad_norm": 0.5227386951446533, "learning_rate": 0.00048787415149229604, "loss": 3.6431, "step": 17450 }, { "epoch": 1.883543213862878, "grad_norm": 0.5770353078842163, "learning_rate": 0.00048755091046223464, "loss": 3.652, "step": 17500 }, { "epoch": 1.8889247659024861, "grad_norm": 0.6320988535881042, "learning_rate": 0.0004872276694321732, "loss": 3.6296, "step": 17550 }, { "epoch": 1.8943063179420945, "grad_norm": 0.5457765460014343, "learning_rate": 0.00048690442840211177, "loss": 3.6411, "step": 17600 }, { "epoch": 1.8996878699817028, "grad_norm": 0.5683311820030212, "learning_rate": 0.0004865811873720504, "loss": 3.6369, "step": 17650 }, { "epoch": 1.905069422021311, "grad_norm": 0.6170719861984253, "learning_rate": 0.00048625794634198896, "loss": 3.6392, "step": 17700 }, { "epoch": 1.910450974060919, "grad_norm": 0.5767971277236938, "learning_rate": 0.00048593470531192756, "loss": 3.6482, "step": 17750 }, { "epoch": 1.9158325261005273, "grad_norm": 0.5953500270843506, "learning_rate": 0.00048561146428186615, "loss": 3.6474, "step": 17800 }, { "epoch": 1.9212140781401357, "grad_norm": 0.5317740440368652, "learning_rate": 0.0004852882232518047, "loss": 3.6312, "step": 17850 }, { "epoch": 1.9265956301797438, "grad_norm": 0.5526648163795471, "learning_rate": 0.00048496498222174334, "loss": 3.6434, "step": 17900 }, { "epoch": 1.931977182219352, "grad_norm": 0.5383221507072449, "learning_rate": 0.00048464174119168193, "loss": 3.6373, "step": 17950 }, { "epoch": 1.9373587342589604, "grad_norm": 0.567500114440918, "learning_rate": 0.0004843185001616205, "loss": 3.6584, "step": 18000 }, { "epoch": 1.9373587342589604, "eval_accuracy": 0.3605316863873469, "eval_loss": 3.5917465686798096, "eval_runtime": 184.8222, "eval_samples_per_second": 97.45, "eval_steps_per_second": 6.092, "step": 18000 }, { "epoch": 1.9427402862985685, "grad_norm": 0.5618447065353394, "learning_rate": 0.00048399525913155907, "loss": 3.6331, "step": 18050 }, { "epoch": 1.9481218383381766, "grad_norm": 0.5482614636421204, "learning_rate": 0.0004836720181014976, "loss": 3.6441, "step": 18100 }, { "epoch": 1.953503390377785, "grad_norm": 0.5669938325881958, "learning_rate": 0.0004833487770714362, "loss": 3.6514, "step": 18150 }, { "epoch": 1.9588849424173933, "grad_norm": 0.5920496582984924, "learning_rate": 0.00048302553604137485, "loss": 3.6434, "step": 18200 }, { "epoch": 1.9642664944570014, "grad_norm": 0.5437381267547607, "learning_rate": 0.0004827022950113134, "loss": 3.664, "step": 18250 }, { "epoch": 1.9696480464966095, "grad_norm": 0.5581496357917786, "learning_rate": 0.000482379053981252, "loss": 3.6224, "step": 18300 }, { "epoch": 1.9750295985362178, "grad_norm": 0.5657931566238403, "learning_rate": 0.0004820558129511906, "loss": 3.6262, "step": 18350 }, { "epoch": 1.9804111505758262, "grad_norm": 0.5733588933944702, "learning_rate": 0.0004817325719211291, "loss": 3.6271, "step": 18400 }, { "epoch": 1.9857927026154343, "grad_norm": 0.5454577803611755, "learning_rate": 0.0004814093308910677, "loss": 3.6385, "step": 18450 }, { "epoch": 1.9911742546550424, "grad_norm": 0.5400912165641785, "learning_rate": 0.00048108608986100637, "loss": 3.6269, "step": 18500 }, { "epoch": 1.9965558066946507, "grad_norm": 0.5676894187927246, "learning_rate": 0.0004807628488309449, "loss": 3.6297, "step": 18550 }, { "epoch": 2.001937358734259, "grad_norm": 0.5944864153862, "learning_rate": 0.0004804396078008835, "loss": 3.578, "step": 18600 }, { "epoch": 2.007318910773867, "grad_norm": 0.5926184058189392, "learning_rate": 0.00048011636677082204, "loss": 3.5433, "step": 18650 }, { "epoch": 2.0127004628134753, "grad_norm": 0.507103443145752, "learning_rate": 0.00047979312574076064, "loss": 3.5372, "step": 18700 }, { "epoch": 2.018082014853084, "grad_norm": 0.5739589929580688, "learning_rate": 0.0004794698847106992, "loss": 3.5428, "step": 18750 }, { "epoch": 2.023463566892692, "grad_norm": 0.5899766087532043, "learning_rate": 0.0004791466436806378, "loss": 3.5556, "step": 18800 }, { "epoch": 2.0288451189323, "grad_norm": 0.603664755821228, "learning_rate": 0.0004788234026505764, "loss": 3.5476, "step": 18850 }, { "epoch": 2.034226670971908, "grad_norm": 0.6328676342964172, "learning_rate": 0.00047850016162051496, "loss": 3.5374, "step": 18900 }, { "epoch": 2.0396082230115167, "grad_norm": 0.5715523362159729, "learning_rate": 0.00047817692059045356, "loss": 3.5504, "step": 18950 }, { "epoch": 2.044989775051125, "grad_norm": 0.554076075553894, "learning_rate": 0.00047785367956039215, "loss": 3.5579, "step": 19000 }, { "epoch": 2.044989775051125, "eval_accuracy": 0.3616910127003274, "eval_loss": 3.57794189453125, "eval_runtime": 184.5487, "eval_samples_per_second": 97.595, "eval_steps_per_second": 6.101, "step": 19000 }, { "epoch": 2.050371327090733, "grad_norm": 0.5813206434249878, "learning_rate": 0.00047753043853033075, "loss": 3.5526, "step": 19050 }, { "epoch": 2.055752879130341, "grad_norm": 0.5949262976646423, "learning_rate": 0.00047720719750026934, "loss": 3.5472, "step": 19100 }, { "epoch": 2.0611344311699495, "grad_norm": 0.5201438069343567, "learning_rate": 0.00047688395647020793, "loss": 3.5551, "step": 19150 }, { "epoch": 2.0665159832095576, "grad_norm": 0.5755208730697632, "learning_rate": 0.0004765607154401465, "loss": 3.5279, "step": 19200 }, { "epoch": 2.0718975352491658, "grad_norm": 0.5886051654815674, "learning_rate": 0.00047623747441008507, "loss": 3.5456, "step": 19250 }, { "epoch": 2.0772790872887743, "grad_norm": 0.581571102142334, "learning_rate": 0.0004759142333800236, "loss": 3.5523, "step": 19300 }, { "epoch": 2.0826606393283824, "grad_norm": 0.5398459434509277, "learning_rate": 0.00047559099234996226, "loss": 3.5488, "step": 19350 }, { "epoch": 2.0880421913679905, "grad_norm": 0.5748805403709412, "learning_rate": 0.00047526775131990085, "loss": 3.5538, "step": 19400 }, { "epoch": 2.0934237434075986, "grad_norm": 0.5900923013687134, "learning_rate": 0.0004749445102898394, "loss": 3.5743, "step": 19450 }, { "epoch": 2.098805295447207, "grad_norm": 0.5444471836090088, "learning_rate": 0.000474621269259778, "loss": 3.5705, "step": 19500 }, { "epoch": 2.1041868474868153, "grad_norm": 0.5641753673553467, "learning_rate": 0.0004742980282297166, "loss": 3.5555, "step": 19550 }, { "epoch": 2.1095683995264234, "grad_norm": 0.5689592957496643, "learning_rate": 0.0004739812520202564, "loss": 3.5609, "step": 19600 }, { "epoch": 2.1149499515660315, "grad_norm": 0.6442001461982727, "learning_rate": 0.000473658010990195, "loss": 3.5649, "step": 19650 }, { "epoch": 2.12033150360564, "grad_norm": 0.5538535118103027, "learning_rate": 0.0004733347699601336, "loss": 3.5624, "step": 19700 }, { "epoch": 2.125713055645248, "grad_norm": 0.5702688097953796, "learning_rate": 0.0004730115289300722, "loss": 3.5425, "step": 19750 }, { "epoch": 2.1310946076848563, "grad_norm": 0.5623447895050049, "learning_rate": 0.000472694752720612, "loss": 3.5462, "step": 19800 }, { "epoch": 2.1364761597244644, "grad_norm": 0.6005106568336487, "learning_rate": 0.0004723715116905506, "loss": 3.5623, "step": 19850 }, { "epoch": 2.141857711764073, "grad_norm": 0.530021607875824, "learning_rate": 0.0004720482706604891, "loss": 3.552, "step": 19900 }, { "epoch": 2.147239263803681, "grad_norm": 0.6486440896987915, "learning_rate": 0.0004717250296304277, "loss": 3.5587, "step": 19950 }, { "epoch": 2.152620815843289, "grad_norm": 0.5921491980552673, "learning_rate": 0.0004714017886003663, "loss": 3.5664, "step": 20000 }, { "epoch": 2.152620815843289, "eval_accuracy": 0.36275624561246056, "eval_loss": 3.568859815597534, "eval_runtime": 184.9316, "eval_samples_per_second": 97.393, "eval_steps_per_second": 6.089, "step": 20000 }, { "epoch": 2.1580023678828972, "grad_norm": 0.5587154030799866, "learning_rate": 0.00047107854757030485, "loss": 3.5344, "step": 20050 }, { "epoch": 2.163383919922506, "grad_norm": 0.6131019592285156, "learning_rate": 0.0004707553065402435, "loss": 3.563, "step": 20100 }, { "epoch": 2.168765471962114, "grad_norm": 0.602177083492279, "learning_rate": 0.0004704320655101821, "loss": 3.5386, "step": 20150 }, { "epoch": 2.174147024001722, "grad_norm": 0.5748916864395142, "learning_rate": 0.00047010882448012063, "loss": 3.5412, "step": 20200 }, { "epoch": 2.1795285760413305, "grad_norm": 0.5983401536941528, "learning_rate": 0.0004697855834500592, "loss": 3.5439, "step": 20250 }, { "epoch": 2.1849101280809387, "grad_norm": 0.5558173656463623, "learning_rate": 0.00046946234241999776, "loss": 3.5476, "step": 20300 }, { "epoch": 2.1902916801205468, "grad_norm": 0.5931748151779175, "learning_rate": 0.00046913910138993636, "loss": 3.5686, "step": 20350 }, { "epoch": 2.195673232160155, "grad_norm": 0.5662775635719299, "learning_rate": 0.000468815860359875, "loss": 3.5691, "step": 20400 }, { "epoch": 2.2010547841997634, "grad_norm": 0.6415780186653137, "learning_rate": 0.00046849261932981355, "loss": 3.54, "step": 20450 }, { "epoch": 2.2064363362393715, "grad_norm": 0.5790057182312012, "learning_rate": 0.00046816937829975214, "loss": 3.5471, "step": 20500 }, { "epoch": 2.2118178882789796, "grad_norm": 0.559398353099823, "learning_rate": 0.00046784613726969074, "loss": 3.5688, "step": 20550 }, { "epoch": 2.2171994403185877, "grad_norm": 0.5323479175567627, "learning_rate": 0.0004675228962396293, "loss": 3.5536, "step": 20600 }, { "epoch": 2.2225809923581963, "grad_norm": 0.5640736818313599, "learning_rate": 0.0004671996552095679, "loss": 3.5444, "step": 20650 }, { "epoch": 2.2279625443978044, "grad_norm": 0.6212958097457886, "learning_rate": 0.0004668764141795065, "loss": 3.5705, "step": 20700 }, { "epoch": 2.2333440964374125, "grad_norm": 0.5601683855056763, "learning_rate": 0.00046655317314944506, "loss": 3.5538, "step": 20750 }, { "epoch": 2.2387256484770206, "grad_norm": 0.5923741459846497, "learning_rate": 0.00046622993211938366, "loss": 3.5587, "step": 20800 }, { "epoch": 2.244107200516629, "grad_norm": 0.5725792646408081, "learning_rate": 0.0004659066910893222, "loss": 3.5558, "step": 20850 }, { "epoch": 2.2494887525562373, "grad_norm": 0.577241063117981, "learning_rate": 0.0004655834500592608, "loss": 3.5422, "step": 20900 }, { "epoch": 2.2548703045958454, "grad_norm": 0.5885173082351685, "learning_rate": 0.00046526020902919944, "loss": 3.5613, "step": 20950 }, { "epoch": 2.2602518566354535, "grad_norm": 0.6301099061965942, "learning_rate": 0.000464936967999138, "loss": 3.5668, "step": 21000 }, { "epoch": 2.2602518566354535, "eval_accuracy": 0.3639710935511143, "eval_loss": 3.557680606842041, "eval_runtime": 184.4972, "eval_samples_per_second": 97.622, "eval_steps_per_second": 6.103, "step": 21000 }, { "epoch": 2.265633408675062, "grad_norm": 0.6086539030075073, "learning_rate": 0.0004646137269690766, "loss": 3.5513, "step": 21050 }, { "epoch": 2.27101496071467, "grad_norm": 0.5513246059417725, "learning_rate": 0.00046429048593901517, "loss": 3.5549, "step": 21100 }, { "epoch": 2.2763965127542782, "grad_norm": 0.5785967111587524, "learning_rate": 0.0004639672449089537, "loss": 3.5596, "step": 21150 }, { "epoch": 2.281778064793887, "grad_norm": 0.5956328511238098, "learning_rate": 0.0004636440038788923, "loss": 3.5427, "step": 21200 }, { "epoch": 2.287159616833495, "grad_norm": 0.6081483364105225, "learning_rate": 0.00046332076284883095, "loss": 3.5325, "step": 21250 }, { "epoch": 2.292541168873103, "grad_norm": 0.5518943071365356, "learning_rate": 0.0004629975218187695, "loss": 3.5368, "step": 21300 }, { "epoch": 2.297922720912711, "grad_norm": 0.5766976475715637, "learning_rate": 0.0004626742807887081, "loss": 3.5642, "step": 21350 }, { "epoch": 2.303304272952319, "grad_norm": 0.582249104976654, "learning_rate": 0.00046235103975864663, "loss": 3.5492, "step": 21400 }, { "epoch": 2.3086858249919278, "grad_norm": 0.6385564208030701, "learning_rate": 0.0004620277987285852, "loss": 3.5574, "step": 21450 }, { "epoch": 2.314067377031536, "grad_norm": 0.5533836483955383, "learning_rate": 0.0004617045576985239, "loss": 3.5461, "step": 21500 }, { "epoch": 2.319448929071144, "grad_norm": 0.6168516278266907, "learning_rate": 0.0004613813166684624, "loss": 3.536, "step": 21550 }, { "epoch": 2.3248304811107525, "grad_norm": 0.5422675013542175, "learning_rate": 0.000461058075638401, "loss": 3.5478, "step": 21600 }, { "epoch": 2.3302120331503606, "grad_norm": 0.673985481262207, "learning_rate": 0.00046073483460833955, "loss": 3.5648, "step": 21650 }, { "epoch": 2.3355935851899687, "grad_norm": 0.6507932543754578, "learning_rate": 0.00046041159357827814, "loss": 3.5514, "step": 21700 }, { "epoch": 2.340975137229577, "grad_norm": 0.5993643403053284, "learning_rate": 0.00046008835254821674, "loss": 3.5723, "step": 21750 }, { "epoch": 2.3463566892691854, "grad_norm": 0.629294753074646, "learning_rate": 0.0004597651115181554, "loss": 3.5385, "step": 21800 }, { "epoch": 2.3517382413087935, "grad_norm": 0.5397011041641235, "learning_rate": 0.00045944187048809393, "loss": 3.5378, "step": 21850 }, { "epoch": 2.3571197933484016, "grad_norm": 0.5139660835266113, "learning_rate": 0.0004591186294580325, "loss": 3.5584, "step": 21900 }, { "epoch": 2.3625013453880097, "grad_norm": 0.6737929582595825, "learning_rate": 0.00045879538842797106, "loss": 3.5637, "step": 21950 }, { "epoch": 2.3678828974276183, "grad_norm": 0.5924611687660217, "learning_rate": 0.00045847214739790966, "loss": 3.5642, "step": 22000 }, { "epoch": 2.3678828974276183, "eval_accuracy": 0.36480739482862234, "eval_loss": 3.550454616546631, "eval_runtime": 184.9625, "eval_samples_per_second": 97.376, "eval_steps_per_second": 6.088, "step": 22000 }, { "epoch": 2.3732644494672264, "grad_norm": 0.5367729067802429, "learning_rate": 0.0004581489063678482, "loss": 3.5564, "step": 22050 }, { "epoch": 2.3786460015068345, "grad_norm": 0.6676405072212219, "learning_rate": 0.00045782566533778685, "loss": 3.5521, "step": 22100 }, { "epoch": 2.384027553546443, "grad_norm": 0.5662420988082886, "learning_rate": 0.00045750242430772544, "loss": 3.5702, "step": 22150 }, { "epoch": 2.389409105586051, "grad_norm": 0.5759710073471069, "learning_rate": 0.000457179183277664, "loss": 3.5667, "step": 22200 }, { "epoch": 2.3947906576256592, "grad_norm": 0.5874856114387512, "learning_rate": 0.0004568559422476026, "loss": 3.5669, "step": 22250 }, { "epoch": 2.4001722096652673, "grad_norm": 0.6054186820983887, "learning_rate": 0.00045653270121754117, "loss": 3.5455, "step": 22300 }, { "epoch": 2.4055537617048754, "grad_norm": 0.5471639037132263, "learning_rate": 0.0004562094601874797, "loss": 3.5433, "step": 22350 }, { "epoch": 2.410935313744484, "grad_norm": 0.5708964467048645, "learning_rate": 0.00045588621915741836, "loss": 3.5434, "step": 22400 }, { "epoch": 2.416316865784092, "grad_norm": 0.5709865093231201, "learning_rate": 0.00045556297812735696, "loss": 3.5641, "step": 22450 }, { "epoch": 2.4216984178237, "grad_norm": 0.608943521976471, "learning_rate": 0.0004552397370972955, "loss": 3.5373, "step": 22500 }, { "epoch": 2.4270799698633088, "grad_norm": 0.5688168406486511, "learning_rate": 0.0004549164960672341, "loss": 3.5529, "step": 22550 }, { "epoch": 2.432461521902917, "grad_norm": 0.5552303791046143, "learning_rate": 0.00045459325503717263, "loss": 3.5403, "step": 22600 }, { "epoch": 2.437843073942525, "grad_norm": 0.5680282711982727, "learning_rate": 0.0004542700140071113, "loss": 3.5457, "step": 22650 }, { "epoch": 2.443224625982133, "grad_norm": 0.584951639175415, "learning_rate": 0.0004539467729770499, "loss": 3.5575, "step": 22700 }, { "epoch": 2.4486061780217416, "grad_norm": 0.5572960376739502, "learning_rate": 0.0004536235319469884, "loss": 3.5445, "step": 22750 }, { "epoch": 2.4539877300613497, "grad_norm": 0.5390388369560242, "learning_rate": 0.000453300290916927, "loss": 3.5575, "step": 22800 }, { "epoch": 2.459369282100958, "grad_norm": 0.6123374700546265, "learning_rate": 0.0004529770498868656, "loss": 3.5767, "step": 22850 }, { "epoch": 2.464750834140566, "grad_norm": 0.5894783139228821, "learning_rate": 0.00045265380885680414, "loss": 3.5542, "step": 22900 }, { "epoch": 2.4701323861801745, "grad_norm": 0.5763325095176697, "learning_rate": 0.0004523305678267428, "loss": 3.5624, "step": 22950 }, { "epoch": 2.4755139382197826, "grad_norm": 0.5496013760566711, "learning_rate": 0.0004520073267966814, "loss": 3.5787, "step": 23000 }, { "epoch": 2.4755139382197826, "eval_accuracy": 0.3661909807020519, "eval_loss": 3.5384585857391357, "eval_runtime": 184.7612, "eval_samples_per_second": 97.483, "eval_steps_per_second": 6.094, "step": 23000 }, { "epoch": 2.4808954902593907, "grad_norm": 0.6314433217048645, "learning_rate": 0.00045168408576661993, "loss": 3.538, "step": 23050 }, { "epoch": 2.4862770422989993, "grad_norm": 0.5926424860954285, "learning_rate": 0.0004513608447365585, "loss": 3.5478, "step": 23100 }, { "epoch": 2.4916585943386074, "grad_norm": 0.5751678943634033, "learning_rate": 0.00045103760370649706, "loss": 3.5523, "step": 23150 }, { "epoch": 2.4970401463782155, "grad_norm": 0.5919656753540039, "learning_rate": 0.00045071436267643566, "loss": 3.5424, "step": 23200 }, { "epoch": 2.5024216984178236, "grad_norm": 0.5607296228408813, "learning_rate": 0.0004503911216463743, "loss": 3.5337, "step": 23250 }, { "epoch": 2.5078032504574317, "grad_norm": 0.5334439873695374, "learning_rate": 0.00045006788061631285, "loss": 3.5375, "step": 23300 }, { "epoch": 2.5131848024970402, "grad_norm": 0.5940549969673157, "learning_rate": 0.00044974463958625144, "loss": 3.5429, "step": 23350 }, { "epoch": 2.5185663545366483, "grad_norm": 0.5540521740913391, "learning_rate": 0.00044942139855619004, "loss": 3.5407, "step": 23400 }, { "epoch": 2.5239479065762565, "grad_norm": 0.6090623140335083, "learning_rate": 0.0004490981575261286, "loss": 3.5343, "step": 23450 }, { "epoch": 2.529329458615865, "grad_norm": 0.5577390789985657, "learning_rate": 0.0004487749164960672, "loss": 3.5486, "step": 23500 }, { "epoch": 2.534711010655473, "grad_norm": 0.5548477172851562, "learning_rate": 0.0004484516754660058, "loss": 3.5276, "step": 23550 }, { "epoch": 2.540092562695081, "grad_norm": 0.6063026189804077, "learning_rate": 0.00044812843443594436, "loss": 3.5385, "step": 23600 }, { "epoch": 2.5454741147346893, "grad_norm": 0.5368787050247192, "learning_rate": 0.00044780519340588296, "loss": 3.5378, "step": 23650 }, { "epoch": 2.550855666774298, "grad_norm": 0.6063023805618286, "learning_rate": 0.0004474819523758215, "loss": 3.5421, "step": 23700 }, { "epoch": 2.556237218813906, "grad_norm": 0.5546424984931946, "learning_rate": 0.0004471587113457601, "loss": 3.5394, "step": 23750 }, { "epoch": 2.561618770853514, "grad_norm": 0.5690875053405762, "learning_rate": 0.00044683547031569874, "loss": 3.5454, "step": 23800 }, { "epoch": 2.567000322893122, "grad_norm": 0.5386408567428589, "learning_rate": 0.0004465186941062385, "loss": 3.5472, "step": 23850 }, { "epoch": 2.5723818749327307, "grad_norm": 0.5763099193572998, "learning_rate": 0.00044619545307617714, "loss": 3.5556, "step": 23900 }, { "epoch": 2.577763426972339, "grad_norm": 0.5599791407585144, "learning_rate": 0.0004458722120461157, "loss": 3.5531, "step": 23950 }, { "epoch": 2.583144979011947, "grad_norm": 0.5776103138923645, "learning_rate": 0.0004455489710160543, "loss": 3.5471, "step": 24000 }, { "epoch": 2.583144979011947, "eval_accuracy": 0.3669588306602369, "eval_loss": 3.5304083824157715, "eval_runtime": 185.0981, "eval_samples_per_second": 97.305, "eval_steps_per_second": 6.083, "step": 24000 }, { "epoch": 2.5885265310515555, "grad_norm": 0.5959745049476624, "learning_rate": 0.0004452257299859928, "loss": 3.5302, "step": 24050 }, { "epoch": 2.5939080830911636, "grad_norm": 0.5823758840560913, "learning_rate": 0.0004449024889559314, "loss": 3.5519, "step": 24100 }, { "epoch": 2.5992896351307717, "grad_norm": 0.5374528765678406, "learning_rate": 0.00044457924792587, "loss": 3.541, "step": 24150 }, { "epoch": 2.60467118717038, "grad_norm": 0.5732862949371338, "learning_rate": 0.0004442560068958086, "loss": 3.5239, "step": 24200 }, { "epoch": 2.610052739209988, "grad_norm": 0.5892502069473267, "learning_rate": 0.0004439327658657472, "loss": 3.5252, "step": 24250 }, { "epoch": 2.6154342912495965, "grad_norm": 0.572582483291626, "learning_rate": 0.0004436095248356858, "loss": 3.5351, "step": 24300 }, { "epoch": 2.6208158432892046, "grad_norm": 0.5686654448509216, "learning_rate": 0.00044328628380562433, "loss": 3.5611, "step": 24350 }, { "epoch": 2.6261973953288127, "grad_norm": 0.6537500619888306, "learning_rate": 0.0004429630427755629, "loss": 3.5398, "step": 24400 }, { "epoch": 2.6315789473684212, "grad_norm": 0.5685673356056213, "learning_rate": 0.0004426398017455016, "loss": 3.5416, "step": 24450 }, { "epoch": 2.6369604994080293, "grad_norm": 0.5964141488075256, "learning_rate": 0.0004423165607154401, "loss": 3.5345, "step": 24500 }, { "epoch": 2.6423420514476375, "grad_norm": 0.5742844343185425, "learning_rate": 0.0004419933196853787, "loss": 3.5365, "step": 24550 }, { "epoch": 2.6477236034872456, "grad_norm": 0.6200198531150818, "learning_rate": 0.00044167007865531725, "loss": 3.5483, "step": 24600 }, { "epoch": 2.653105155526854, "grad_norm": 0.5747848153114319, "learning_rate": 0.00044134683762525584, "loss": 3.5481, "step": 24650 }, { "epoch": 2.658486707566462, "grad_norm": 0.5651057958602905, "learning_rate": 0.00044102359659519444, "loss": 3.5375, "step": 24700 }, { "epoch": 2.6638682596060703, "grad_norm": 0.6583673357963562, "learning_rate": 0.00044070035556513303, "loss": 3.5528, "step": 24750 }, { "epoch": 2.6692498116456784, "grad_norm": 0.6136596202850342, "learning_rate": 0.00044037711453507163, "loss": 3.5511, "step": 24800 }, { "epoch": 2.674631363685287, "grad_norm": 0.534138560295105, "learning_rate": 0.0004400538735050102, "loss": 3.5538, "step": 24850 }, { "epoch": 2.680012915724895, "grad_norm": 0.5683159828186035, "learning_rate": 0.00043973063247494876, "loss": 3.5404, "step": 24900 }, { "epoch": 2.685394467764503, "grad_norm": 0.5965563058853149, "learning_rate": 0.00043940739144488736, "loss": 3.5361, "step": 24950 }, { "epoch": 2.6907760198041117, "grad_norm": 0.6091514825820923, "learning_rate": 0.0004390841504148259, "loss": 3.5407, "step": 25000 }, { "epoch": 2.6907760198041117, "eval_accuracy": 0.368015914605784, "eval_loss": 3.5202255249023438, "eval_runtime": 184.4748, "eval_samples_per_second": 97.634, "eval_steps_per_second": 6.104, "step": 25000 }, { "epoch": 2.69615757184372, "grad_norm": 0.5996729135513306, "learning_rate": 0.00043876090938476455, "loss": 3.5246, "step": 25050 }, { "epoch": 2.701539123883328, "grad_norm": 0.5679774284362793, "learning_rate": 0.00043843766835470314, "loss": 3.5314, "step": 25100 }, { "epoch": 2.706920675922936, "grad_norm": 0.6458651423454285, "learning_rate": 0.0004381144273246417, "loss": 3.5488, "step": 25150 }, { "epoch": 2.712302227962544, "grad_norm": 0.5411641597747803, "learning_rate": 0.0004377911862945803, "loss": 3.5364, "step": 25200 }, { "epoch": 2.7176837800021527, "grad_norm": 0.5480161309242249, "learning_rate": 0.00043746794526451887, "loss": 3.5359, "step": 25250 }, { "epoch": 2.723065332041761, "grad_norm": 0.5879181623458862, "learning_rate": 0.00043714470423445747, "loss": 3.5329, "step": 25300 }, { "epoch": 2.728446884081369, "grad_norm": 0.5946080088615417, "learning_rate": 0.00043682146320439606, "loss": 3.54, "step": 25350 }, { "epoch": 2.7338284361209775, "grad_norm": 0.5855832099914551, "learning_rate": 0.00043649822217433466, "loss": 3.5489, "step": 25400 }, { "epoch": 2.7392099881605856, "grad_norm": 0.5638436675071716, "learning_rate": 0.0004361749811442732, "loss": 3.5306, "step": 25450 }, { "epoch": 2.7445915402001937, "grad_norm": 0.5559232831001282, "learning_rate": 0.0004358517401142118, "loss": 3.5255, "step": 25500 }, { "epoch": 2.749973092239802, "grad_norm": 0.5973818302154541, "learning_rate": 0.00043552849908415033, "loss": 3.5437, "step": 25550 }, { "epoch": 2.7553546442794103, "grad_norm": 0.5868799090385437, "learning_rate": 0.000435205258054089, "loss": 3.5466, "step": 25600 }, { "epoch": 2.7607361963190185, "grad_norm": 0.6294220089912415, "learning_rate": 0.0004348820170240276, "loss": 3.543, "step": 25650 }, { "epoch": 2.7661177483586266, "grad_norm": 0.5231223702430725, "learning_rate": 0.0004345587759939661, "loss": 3.5294, "step": 25700 }, { "epoch": 2.7714993003982347, "grad_norm": 0.634873628616333, "learning_rate": 0.0004342355349639047, "loss": 3.537, "step": 25750 }, { "epoch": 2.776880852437843, "grad_norm": 0.5652748346328735, "learning_rate": 0.0004339122939338433, "loss": 3.5232, "step": 25800 }, { "epoch": 2.7822624044774513, "grad_norm": 0.6224278807640076, "learning_rate": 0.00043358905290378184, "loss": 3.5462, "step": 25850 }, { "epoch": 2.7876439565170594, "grad_norm": 0.6616294980049133, "learning_rate": 0.0004332658118737205, "loss": 3.5231, "step": 25900 }, { "epoch": 2.793025508556668, "grad_norm": 0.566481351852417, "learning_rate": 0.0004329425708436591, "loss": 3.5384, "step": 25950 }, { "epoch": 2.798407060596276, "grad_norm": 0.5798117518424988, "learning_rate": 0.00043261932981359763, "loss": 3.5461, "step": 26000 }, { "epoch": 2.798407060596276, "eval_accuracy": 0.3685674366643303, "eval_loss": 3.5130412578582764, "eval_runtime": 184.8813, "eval_samples_per_second": 97.419, "eval_steps_per_second": 6.09, "step": 26000 }, { "epoch": 2.803788612635884, "grad_norm": 0.5411988496780396, "learning_rate": 0.0004322960887835362, "loss": 3.5413, "step": 26050 }, { "epoch": 2.8091701646754923, "grad_norm": 0.554651141166687, "learning_rate": 0.00043197284775347476, "loss": 3.5336, "step": 26100 }, { "epoch": 2.8145517167151004, "grad_norm": 0.5864933133125305, "learning_rate": 0.00043164960672341336, "loss": 3.5489, "step": 26150 }, { "epoch": 2.819933268754709, "grad_norm": 0.549628734588623, "learning_rate": 0.000431326365693352, "loss": 3.534, "step": 26200 }, { "epoch": 2.825314820794317, "grad_norm": 0.5752499103546143, "learning_rate": 0.00043100312466329055, "loss": 3.5373, "step": 26250 }, { "epoch": 2.830696372833925, "grad_norm": 0.6056687831878662, "learning_rate": 0.00043067988363322914, "loss": 3.5467, "step": 26300 }, { "epoch": 2.8360779248735337, "grad_norm": 0.5557623505592346, "learning_rate": 0.00043035664260316774, "loss": 3.5347, "step": 26350 }, { "epoch": 2.841459476913142, "grad_norm": 0.5566017031669617, "learning_rate": 0.00043003986639370754, "loss": 3.5568, "step": 26400 }, { "epoch": 2.84684102895275, "grad_norm": 0.6013063192367554, "learning_rate": 0.0004297166253636461, "loss": 3.5329, "step": 26450 }, { "epoch": 2.852222580992358, "grad_norm": 0.560697078704834, "learning_rate": 0.0004293933843335847, "loss": 3.5274, "step": 26500 }, { "epoch": 2.857604133031966, "grad_norm": 0.5921282768249512, "learning_rate": 0.00042907014330352333, "loss": 3.538, "step": 26550 }, { "epoch": 2.8629856850715747, "grad_norm": 0.5770512223243713, "learning_rate": 0.00042874690227346187, "loss": 3.5341, "step": 26600 }, { "epoch": 2.868367237111183, "grad_norm": 0.5724422931671143, "learning_rate": 0.00042842366124340046, "loss": 3.5228, "step": 26650 }, { "epoch": 2.873748789150791, "grad_norm": 0.593283474445343, "learning_rate": 0.00042810042021333906, "loss": 3.535, "step": 26700 }, { "epoch": 2.8791303411903995, "grad_norm": 0.5896148681640625, "learning_rate": 0.0004277771791832776, "loss": 3.5122, "step": 26750 }, { "epoch": 2.8845118932300076, "grad_norm": 0.5596243143081665, "learning_rate": 0.0004274539381532162, "loss": 3.5354, "step": 26800 }, { "epoch": 2.8898934452696157, "grad_norm": 0.592008650302887, "learning_rate": 0.00042713069712315484, "loss": 3.5354, "step": 26850 }, { "epoch": 2.895274997309224, "grad_norm": 0.5777035355567932, "learning_rate": 0.0004268074560930934, "loss": 3.5294, "step": 26900 }, { "epoch": 2.9006565493488323, "grad_norm": 0.6243908405303955, "learning_rate": 0.000426484215063032, "loss": 3.5134, "step": 26950 }, { "epoch": 2.9060381013884404, "grad_norm": 0.5806100368499756, "learning_rate": 0.0004261609740329705, "loss": 3.5291, "step": 27000 }, { "epoch": 2.9060381013884404, "eval_accuracy": 0.36983183031982736, "eval_loss": 3.504441738128662, "eval_runtime": 185.054, "eval_samples_per_second": 97.328, "eval_steps_per_second": 6.085, "step": 27000 }, { "epoch": 2.9114196534280485, "grad_norm": 0.6097036600112915, "learning_rate": 0.0004258377330029091, "loss": 3.5165, "step": 27050 }, { "epoch": 2.9168012054676566, "grad_norm": 0.6032792925834656, "learning_rate": 0.00042551449197284776, "loss": 3.543, "step": 27100 }, { "epoch": 2.922182757507265, "grad_norm": 0.5959144830703735, "learning_rate": 0.0004251912509427863, "loss": 3.5216, "step": 27150 }, { "epoch": 2.9275643095468733, "grad_norm": 0.6524012684822083, "learning_rate": 0.0004248680099127249, "loss": 3.5105, "step": 27200 }, { "epoch": 2.9329458615864814, "grad_norm": 0.5885307788848877, "learning_rate": 0.0004245447688826635, "loss": 3.5274, "step": 27250 }, { "epoch": 2.93832741362609, "grad_norm": 0.6017599105834961, "learning_rate": 0.00042422152785260203, "loss": 3.5332, "step": 27300 }, { "epoch": 2.943708965665698, "grad_norm": 0.6001046895980835, "learning_rate": 0.0004238982868225406, "loss": 3.5223, "step": 27350 }, { "epoch": 2.949090517705306, "grad_norm": 0.5998550057411194, "learning_rate": 0.0004235750457924793, "loss": 3.5066, "step": 27400 }, { "epoch": 2.9544720697449143, "grad_norm": 0.6128471493721008, "learning_rate": 0.0004232518047624178, "loss": 3.5448, "step": 27450 }, { "epoch": 2.9598536217845224, "grad_norm": 0.5720791816711426, "learning_rate": 0.0004229285637323564, "loss": 3.5259, "step": 27500 }, { "epoch": 2.965235173824131, "grad_norm": 0.5852633118629456, "learning_rate": 0.00042260532270229495, "loss": 3.5241, "step": 27550 }, { "epoch": 2.970616725863739, "grad_norm": 0.6088577508926392, "learning_rate": 0.00042228208167223354, "loss": 3.518, "step": 27600 }, { "epoch": 2.975998277903347, "grad_norm": 0.5983795523643494, "learning_rate": 0.00042195884064217214, "loss": 3.5264, "step": 27650 }, { "epoch": 2.9813798299429557, "grad_norm": 0.5886030197143555, "learning_rate": 0.00042163559961211073, "loss": 3.5261, "step": 27700 }, { "epoch": 2.986761381982564, "grad_norm": 0.612271249294281, "learning_rate": 0.00042131235858204933, "loss": 3.5311, "step": 27750 }, { "epoch": 2.992142934022172, "grad_norm": 0.6254613399505615, "learning_rate": 0.0004209891175519879, "loss": 3.5097, "step": 27800 }, { "epoch": 2.9975244860617805, "grad_norm": 0.6479254364967346, "learning_rate": 0.00042066587652192646, "loss": 3.5238, "step": 27850 }, { "epoch": 3.0029060381013886, "grad_norm": 0.6095746755599976, "learning_rate": 0.00042034263549186506, "loss": 3.4804, "step": 27900 }, { "epoch": 3.0082875901409967, "grad_norm": 0.6377231478691101, "learning_rate": 0.0004200193944618036, "loss": 3.4438, "step": 27950 }, { "epoch": 3.0136691421806048, "grad_norm": 0.5965224504470825, "learning_rate": 0.00041969615343174225, "loss": 3.4265, "step": 28000 }, { "epoch": 3.0136691421806048, "eval_accuracy": 0.3701457285127234, "eval_loss": 3.5000498294830322, "eval_runtime": 184.7752, "eval_samples_per_second": 97.475, "eval_steps_per_second": 6.094, "step": 28000 }, { "epoch": 3.0190506942202133, "grad_norm": 0.5932812690734863, "learning_rate": 0.00041937291240168084, "loss": 3.4383, "step": 28050 }, { "epoch": 3.0244322462598214, "grad_norm": 0.6348116993904114, "learning_rate": 0.0004190496713716194, "loss": 3.4435, "step": 28100 }, { "epoch": 3.0298137982994295, "grad_norm": 0.623078465461731, "learning_rate": 0.000418726430341558, "loss": 3.4472, "step": 28150 }, { "epoch": 3.0351953503390376, "grad_norm": 0.5842487215995789, "learning_rate": 0.00041840318931149657, "loss": 3.4284, "step": 28200 }, { "epoch": 3.040576902378646, "grad_norm": 0.6037656664848328, "learning_rate": 0.00041807994828143517, "loss": 3.4273, "step": 28250 }, { "epoch": 3.0459584544182543, "grad_norm": 0.5624446868896484, "learning_rate": 0.00041775670725137376, "loss": 3.4341, "step": 28300 }, { "epoch": 3.0513400064578624, "grad_norm": 0.6236454248428345, "learning_rate": 0.00041743346622131236, "loss": 3.4384, "step": 28350 }, { "epoch": 3.0567215584974705, "grad_norm": 0.5729033350944519, "learning_rate": 0.0004171102251912509, "loss": 3.4509, "step": 28400 }, { "epoch": 3.062103110537079, "grad_norm": 0.570470929145813, "learning_rate": 0.0004167869841611895, "loss": 3.4454, "step": 28450 }, { "epoch": 3.067484662576687, "grad_norm": 0.5834895372390747, "learning_rate": 0.00041646374313112803, "loss": 3.4471, "step": 28500 }, { "epoch": 3.0728662146162953, "grad_norm": 0.6381838917732239, "learning_rate": 0.0004161405021010667, "loss": 3.4475, "step": 28550 }, { "epoch": 3.0782477666559034, "grad_norm": 0.611271321773529, "learning_rate": 0.00041582372589160643, "loss": 3.452, "step": 28600 }, { "epoch": 3.083629318695512, "grad_norm": 0.6530306339263916, "learning_rate": 0.0004155004848615451, "loss": 3.4317, "step": 28650 }, { "epoch": 3.08901087073512, "grad_norm": 0.6146801710128784, "learning_rate": 0.0004151772438314837, "loss": 3.4379, "step": 28700 }, { "epoch": 3.094392422774728, "grad_norm": 0.5884636640548706, "learning_rate": 0.0004148540028014222, "loss": 3.4381, "step": 28750 }, { "epoch": 3.0997739748143363, "grad_norm": 0.6561384797096252, "learning_rate": 0.0004145307617713608, "loss": 3.4375, "step": 28800 }, { "epoch": 3.105155526853945, "grad_norm": 0.5567984580993652, "learning_rate": 0.00041420752074129935, "loss": 3.4381, "step": 28850 }, { "epoch": 3.110537078893553, "grad_norm": 0.583527684211731, "learning_rate": 0.000413884279711238, "loss": 3.4351, "step": 28900 }, { "epoch": 3.115918630933161, "grad_norm": 0.580026388168335, "learning_rate": 0.0004135610386811766, "loss": 3.443, "step": 28950 }, { "epoch": 3.121300182972769, "grad_norm": 0.6263661980628967, "learning_rate": 0.00041323779765111514, "loss": 3.4472, "step": 29000 }, { "epoch": 3.121300182972769, "eval_accuracy": 0.37119868364853353, "eval_loss": 3.4976308345794678, "eval_runtime": 184.8903, "eval_samples_per_second": 97.415, "eval_steps_per_second": 6.09, "step": 29000 }, { "epoch": 3.1266817350123777, "grad_norm": 0.5766123533248901, "learning_rate": 0.00041291455662105373, "loss": 3.4438, "step": 29050 }, { "epoch": 3.132063287051986, "grad_norm": 0.5877207517623901, "learning_rate": 0.0004125913155909923, "loss": 3.4321, "step": 29100 }, { "epoch": 3.137444839091594, "grad_norm": 0.6023469567298889, "learning_rate": 0.00041226807456093087, "loss": 3.4396, "step": 29150 }, { "epoch": 3.1428263911312024, "grad_norm": 0.5752248167991638, "learning_rate": 0.0004119448335308695, "loss": 3.4291, "step": 29200 }, { "epoch": 3.1482079431708105, "grad_norm": 0.5830261707305908, "learning_rate": 0.0004116215925008081, "loss": 3.4434, "step": 29250 }, { "epoch": 3.1535894952104186, "grad_norm": 0.6252485513687134, "learning_rate": 0.00041129835147074665, "loss": 3.458, "step": 29300 }, { "epoch": 3.1589710472500268, "grad_norm": 0.6096091866493225, "learning_rate": 0.00041097511044068524, "loss": 3.4718, "step": 29350 }, { "epoch": 3.1643525992896353, "grad_norm": 0.6204565167427063, "learning_rate": 0.0004106518694106238, "loss": 3.4728, "step": 29400 }, { "epoch": 3.1697341513292434, "grad_norm": 0.6110973954200745, "learning_rate": 0.0004103286283805624, "loss": 3.4484, "step": 29450 }, { "epoch": 3.1751157033688515, "grad_norm": 0.6513490080833435, "learning_rate": 0.00041000538735050103, "loss": 3.4548, "step": 29500 }, { "epoch": 3.1804972554084596, "grad_norm": 0.6093899607658386, "learning_rate": 0.00040968214632043957, "loss": 3.4613, "step": 29550 }, { "epoch": 3.185878807448068, "grad_norm": 0.6108792424201965, "learning_rate": 0.00040935890529037816, "loss": 3.4307, "step": 29600 }, { "epoch": 3.1912603594876763, "grad_norm": 0.6220118999481201, "learning_rate": 0.00040903566426031676, "loss": 3.4599, "step": 29650 }, { "epoch": 3.1966419115272844, "grad_norm": 0.5830521583557129, "learning_rate": 0.0004087124232302553, "loss": 3.4543, "step": 29700 }, { "epoch": 3.2020234635668925, "grad_norm": 0.6141335368156433, "learning_rate": 0.0004083891822001939, "loss": 3.4441, "step": 29750 }, { "epoch": 3.207405015606501, "grad_norm": 0.6098508238792419, "learning_rate": 0.00040806594117013254, "loss": 3.4538, "step": 29800 }, { "epoch": 3.212786567646109, "grad_norm": 0.5632219314575195, "learning_rate": 0.0004077427001400711, "loss": 3.4533, "step": 29850 }, { "epoch": 3.2181681196857173, "grad_norm": 0.6396539807319641, "learning_rate": 0.0004074194591100097, "loss": 3.4618, "step": 29900 }, { "epoch": 3.2235496717253254, "grad_norm": 0.6783267855644226, "learning_rate": 0.0004070962180799482, "loss": 3.4458, "step": 29950 }, { "epoch": 3.228931223764934, "grad_norm": 0.6168944239616394, "learning_rate": 0.0004067729770498868, "loss": 3.4667, "step": 30000 }, { "epoch": 3.228931223764934, "eval_accuracy": 0.37168392744551304, "eval_loss": 3.4887137413024902, "eval_runtime": 184.9069, "eval_samples_per_second": 97.406, "eval_steps_per_second": 6.09, "step": 30000 }, { "epoch": 3.234312775804542, "grad_norm": 0.6674462556838989, "learning_rate": 0.00040644973601982546, "loss": 3.4648, "step": 30050 }, { "epoch": 3.23969432784415, "grad_norm": 0.5786466598510742, "learning_rate": 0.000406126494989764, "loss": 3.4439, "step": 30100 }, { "epoch": 3.2450758798837587, "grad_norm": 0.6295211315155029, "learning_rate": 0.0004058032539597026, "loss": 3.4525, "step": 30150 }, { "epoch": 3.250457431923367, "grad_norm": 0.659146249294281, "learning_rate": 0.00040548001292964114, "loss": 3.4615, "step": 30200 }, { "epoch": 3.255838983962975, "grad_norm": 0.5946276187896729, "learning_rate": 0.00040515677189957973, "loss": 3.4612, "step": 30250 }, { "epoch": 3.261220536002583, "grad_norm": 0.6724011898040771, "learning_rate": 0.0004048335308695183, "loss": 3.4394, "step": 30300 }, { "epoch": 3.2666020880421915, "grad_norm": 0.6301489472389221, "learning_rate": 0.0004045102898394569, "loss": 3.4501, "step": 30350 }, { "epoch": 3.2719836400817996, "grad_norm": 0.666102945804596, "learning_rate": 0.0004041870488093955, "loss": 3.4573, "step": 30400 }, { "epoch": 3.2773651921214078, "grad_norm": 0.6429094672203064, "learning_rate": 0.0004038638077793341, "loss": 3.4675, "step": 30450 }, { "epoch": 3.282746744161016, "grad_norm": 0.6529164910316467, "learning_rate": 0.00040354056674927265, "loss": 3.4523, "step": 30500 }, { "epoch": 3.2881282962006244, "grad_norm": 0.593303918838501, "learning_rate": 0.00040321732571921124, "loss": 3.4455, "step": 30550 }, { "epoch": 3.2935098482402325, "grad_norm": 0.587743878364563, "learning_rate": 0.0004029070143303523, "loss": 3.4432, "step": 30600 }, { "epoch": 3.2988914002798406, "grad_norm": 0.6212865114212036, "learning_rate": 0.00040258377330029086, "loss": 3.4705, "step": 30650 }, { "epoch": 3.304272952319449, "grad_norm": 0.5814943313598633, "learning_rate": 0.00040226053227022945, "loss": 3.4567, "step": 30700 }, { "epoch": 3.3096545043590573, "grad_norm": 0.5858567953109741, "learning_rate": 0.00040193729124016805, "loss": 3.453, "step": 30750 }, { "epoch": 3.3150360563986654, "grad_norm": 0.6212705373764038, "learning_rate": 0.00040161405021010664, "loss": 3.4501, "step": 30800 }, { "epoch": 3.3204176084382735, "grad_norm": 0.6279528141021729, "learning_rate": 0.00040129080918004524, "loss": 3.4556, "step": 30850 }, { "epoch": 3.3257991604778816, "grad_norm": 0.6483830213546753, "learning_rate": 0.00040096756814998383, "loss": 3.4473, "step": 30900 }, { "epoch": 3.33118071251749, "grad_norm": 0.6192697286605835, "learning_rate": 0.00040064432711992237, "loss": 3.47, "step": 30950 }, { "epoch": 3.3365622645570983, "grad_norm": 0.6230787038803101, "learning_rate": 0.00040032108608986097, "loss": 3.4857, "step": 31000 }, { "epoch": 3.3365622645570983, "eval_accuracy": 0.3728218491395941, "eval_loss": 3.4819231033325195, "eval_runtime": 185.0123, "eval_samples_per_second": 97.35, "eval_steps_per_second": 6.086, "step": 31000 }, { "epoch": 3.3419438165967064, "grad_norm": 0.6070190072059631, "learning_rate": 0.0003999978450597995, "loss": 3.4448, "step": 31050 }, { "epoch": 3.347325368636315, "grad_norm": 0.638231098651886, "learning_rate": 0.00039967460402973816, "loss": 3.4718, "step": 31100 }, { "epoch": 3.352706920675923, "grad_norm": 0.7029089331626892, "learning_rate": 0.00039935136299967675, "loss": 3.4692, "step": 31150 }, { "epoch": 3.358088472715531, "grad_norm": 0.565448522567749, "learning_rate": 0.0003990281219696153, "loss": 3.4601, "step": 31200 }, { "epoch": 3.3634700247551392, "grad_norm": 0.6335359811782837, "learning_rate": 0.0003987048809395539, "loss": 3.4573, "step": 31250 }, { "epoch": 3.368851576794748, "grad_norm": 0.604525625705719, "learning_rate": 0.0003983816399094925, "loss": 3.4508, "step": 31300 }, { "epoch": 3.374233128834356, "grad_norm": 0.634780764579773, "learning_rate": 0.000398058398879431, "loss": 3.4762, "step": 31350 }, { "epoch": 3.379614680873964, "grad_norm": 0.636223316192627, "learning_rate": 0.00039773515784936967, "loss": 3.4521, "step": 31400 }, { "epoch": 3.384996232913572, "grad_norm": 0.6109070181846619, "learning_rate": 0.00039741191681930826, "loss": 3.4659, "step": 31450 }, { "epoch": 3.3903777849531807, "grad_norm": 0.6315115094184875, "learning_rate": 0.0003970886757892468, "loss": 3.4598, "step": 31500 }, { "epoch": 3.3957593369927888, "grad_norm": 0.5966545343399048, "learning_rate": 0.0003967654347591854, "loss": 3.458, "step": 31550 }, { "epoch": 3.401140889032397, "grad_norm": 0.6154278516769409, "learning_rate": 0.00039644219372912394, "loss": 3.4428, "step": 31600 }, { "epoch": 3.4065224410720054, "grad_norm": 0.6580645442008972, "learning_rate": 0.0003961189526990626, "loss": 3.4463, "step": 31650 }, { "epoch": 3.4119039931116135, "grad_norm": 0.6818550825119019, "learning_rate": 0.0003957957116690012, "loss": 3.4425, "step": 31700 }, { "epoch": 3.4172855451512216, "grad_norm": 0.6167941689491272, "learning_rate": 0.0003954724706389397, "loss": 3.4583, "step": 31750 }, { "epoch": 3.4226670971908297, "grad_norm": 0.6051561236381531, "learning_rate": 0.0003951492296088783, "loss": 3.4626, "step": 31800 }, { "epoch": 3.428048649230438, "grad_norm": 0.604656457901001, "learning_rate": 0.0003948259885788169, "loss": 3.456, "step": 31850 }, { "epoch": 3.4334302012700464, "grad_norm": 0.6075143814086914, "learning_rate": 0.00039450274754875545, "loss": 3.4499, "step": 31900 }, { "epoch": 3.4388117533096545, "grad_norm": 0.6349363327026367, "learning_rate": 0.0003941795065186941, "loss": 3.4608, "step": 31950 }, { "epoch": 3.4441933053492626, "grad_norm": 0.6251042485237122, "learning_rate": 0.0003938562654886327, "loss": 3.4601, "step": 32000 }, { "epoch": 3.4441933053492626, "eval_accuracy": 0.372848251791333, "eval_loss": 3.476959466934204, "eval_runtime": 185.0651, "eval_samples_per_second": 97.323, "eval_steps_per_second": 6.084, "step": 32000 }, { "epoch": 3.449574857388871, "grad_norm": 0.6375144720077515, "learning_rate": 0.00039353302445857124, "loss": 3.4643, "step": 32050 }, { "epoch": 3.4549564094284793, "grad_norm": 0.6365067958831787, "learning_rate": 0.00039320978342850983, "loss": 3.4534, "step": 32100 }, { "epoch": 3.4603379614680874, "grad_norm": 0.6374215483665466, "learning_rate": 0.00039288654239844837, "loss": 3.4464, "step": 32150 }, { "epoch": 3.4657195135076955, "grad_norm": 0.6080241203308105, "learning_rate": 0.00039256330136838697, "loss": 3.4582, "step": 32200 }, { "epoch": 3.471101065547304, "grad_norm": 0.5984114408493042, "learning_rate": 0.0003922400603383256, "loss": 3.456, "step": 32250 }, { "epoch": 3.476482617586912, "grad_norm": 0.6118507981300354, "learning_rate": 0.00039191681930826416, "loss": 3.4786, "step": 32300 }, { "epoch": 3.4818641696265202, "grad_norm": 0.6618357300758362, "learning_rate": 0.00039159357827820275, "loss": 3.4559, "step": 32350 }, { "epoch": 3.4872457216661283, "grad_norm": 0.6183602809906006, "learning_rate": 0.00039127033724814135, "loss": 3.462, "step": 32400 }, { "epoch": 3.492627273705737, "grad_norm": 0.6030476689338684, "learning_rate": 0.0003909470962180799, "loss": 3.4661, "step": 32450 }, { "epoch": 3.498008825745345, "grad_norm": 0.6316002011299133, "learning_rate": 0.00039062385518801854, "loss": 3.4361, "step": 32500 }, { "epoch": 3.503390377784953, "grad_norm": 0.592577338218689, "learning_rate": 0.00039030061415795713, "loss": 3.438, "step": 32550 }, { "epoch": 3.5087719298245617, "grad_norm": 0.6018152236938477, "learning_rate": 0.00038997737312789567, "loss": 3.4465, "step": 32600 }, { "epoch": 3.5141534818641698, "grad_norm": 0.657094419002533, "learning_rate": 0.00038965413209783426, "loss": 3.4449, "step": 32650 }, { "epoch": 3.519535033903778, "grad_norm": 0.6502700448036194, "learning_rate": 0.0003893308910677728, "loss": 3.4711, "step": 32700 }, { "epoch": 3.524916585943386, "grad_norm": 0.6325830817222595, "learning_rate": 0.0003890076500377114, "loss": 3.4545, "step": 32750 }, { "epoch": 3.530298137982994, "grad_norm": 0.6116040349006653, "learning_rate": 0.00038868440900765005, "loss": 3.4498, "step": 32800 }, { "epoch": 3.5356796900226026, "grad_norm": 0.6353474855422974, "learning_rate": 0.0003883611679775886, "loss": 3.4359, "step": 32850 }, { "epoch": 3.5410612420622107, "grad_norm": 0.6394508481025696, "learning_rate": 0.0003880379269475272, "loss": 3.4538, "step": 32900 }, { "epoch": 3.546442794101819, "grad_norm": 0.5920386910438538, "learning_rate": 0.0003877146859174657, "loss": 3.4638, "step": 32950 }, { "epoch": 3.5518243461414274, "grad_norm": 0.6352172493934631, "learning_rate": 0.0003873914448874043, "loss": 3.4542, "step": 33000 }, { "epoch": 3.5518243461414274, "eval_accuracy": 0.3739409738840832, "eval_loss": 3.467642068862915, "eval_runtime": 184.7948, "eval_samples_per_second": 97.465, "eval_steps_per_second": 6.093, "step": 33000 }, { "epoch": 3.5572058981810355, "grad_norm": 0.5835293531417847, "learning_rate": 0.0003870682038573429, "loss": 3.4507, "step": 33050 }, { "epoch": 3.5625874502206436, "grad_norm": 0.6255052089691162, "learning_rate": 0.0003867449628272815, "loss": 3.4522, "step": 33100 }, { "epoch": 3.5679690022602517, "grad_norm": 0.6663792133331299, "learning_rate": 0.0003864217217972201, "loss": 3.4373, "step": 33150 }, { "epoch": 3.57335055429986, "grad_norm": 0.5987926125526428, "learning_rate": 0.0003860984807671587, "loss": 3.4538, "step": 33200 }, { "epoch": 3.5787321063394684, "grad_norm": 0.6144534945487976, "learning_rate": 0.00038577523973709724, "loss": 3.466, "step": 33250 }, { "epoch": 3.5841136583790765, "grad_norm": 0.6252511739730835, "learning_rate": 0.00038545199870703583, "loss": 3.4454, "step": 33300 }, { "epoch": 3.5894952104186846, "grad_norm": 0.6555866599082947, "learning_rate": 0.0003851287576769744, "loss": 3.455, "step": 33350 }, { "epoch": 3.594876762458293, "grad_norm": 0.6036868691444397, "learning_rate": 0.000384805516646913, "loss": 3.4569, "step": 33400 }, { "epoch": 3.6002583144979012, "grad_norm": 0.6310072541236877, "learning_rate": 0.0003844822756168516, "loss": 3.4375, "step": 33450 }, { "epoch": 3.6056398665375093, "grad_norm": 0.6304979920387268, "learning_rate": 0.00038415903458679016, "loss": 3.4567, "step": 33500 }, { "epoch": 3.611021418577118, "grad_norm": 0.6037408113479614, "learning_rate": 0.00038383579355672875, "loss": 3.4643, "step": 33550 }, { "epoch": 3.616402970616726, "grad_norm": 0.6693500280380249, "learning_rate": 0.00038351255252666735, "loss": 3.4298, "step": 33600 }, { "epoch": 3.621784522656334, "grad_norm": 0.6045053005218506, "learning_rate": 0.00038318931149660594, "loss": 3.4583, "step": 33650 }, { "epoch": 3.627166074695942, "grad_norm": 0.5890701413154602, "learning_rate": 0.00038286607046654454, "loss": 3.4497, "step": 33700 }, { "epoch": 3.6325476267355503, "grad_norm": 0.6815541386604309, "learning_rate": 0.00038254282943648313, "loss": 3.4522, "step": 33750 }, { "epoch": 3.637929178775159, "grad_norm": 0.6552605032920837, "learning_rate": 0.00038221958840642167, "loss": 3.466, "step": 33800 }, { "epoch": 3.643310730814767, "grad_norm": 0.6101818084716797, "learning_rate": 0.00038189634737636027, "loss": 3.4601, "step": 33850 }, { "epoch": 3.648692282854375, "grad_norm": 0.636333703994751, "learning_rate": 0.0003815731063462988, "loss": 3.4451, "step": 33900 }, { "epoch": 3.6540738348939836, "grad_norm": 0.6014490723609924, "learning_rate": 0.00038124986531623745, "loss": 3.4576, "step": 33950 }, { "epoch": 3.6594553869335917, "grad_norm": 0.6211981773376465, "learning_rate": 0.00038092662428617605, "loss": 3.4527, "step": 34000 }, { "epoch": 3.6594553869335917, "eval_accuracy": 0.374348748172051, "eval_loss": 3.4629900455474854, "eval_runtime": 184.9732, "eval_samples_per_second": 97.371, "eval_steps_per_second": 6.087, "step": 34000 }, { "epoch": 3.6648369389732, "grad_norm": 0.65215665102005, "learning_rate": 0.0003806033832561146, "loss": 3.4392, "step": 34050 }, { "epoch": 3.670218491012808, "grad_norm": 0.6055262684822083, "learning_rate": 0.0003802801422260532, "loss": 3.4235, "step": 34100 }, { "epoch": 3.675600043052416, "grad_norm": 0.6137178540229797, "learning_rate": 0.0003799569011959918, "loss": 3.4482, "step": 34150 }, { "epoch": 3.6809815950920246, "grad_norm": 0.6887802481651306, "learning_rate": 0.0003796336601659303, "loss": 3.4466, "step": 34200 }, { "epoch": 3.6863631471316327, "grad_norm": 0.634081244468689, "learning_rate": 0.00037931041913586897, "loss": 3.4466, "step": 34250 }, { "epoch": 3.691744699171241, "grad_norm": 0.622760534286499, "learning_rate": 0.00037898717810580756, "loss": 3.4431, "step": 34300 }, { "epoch": 3.6971262512108494, "grad_norm": 0.616561770439148, "learning_rate": 0.0003786639370757461, "loss": 3.454, "step": 34350 }, { "epoch": 3.7025078032504575, "grad_norm": 0.6342828273773193, "learning_rate": 0.0003783406960456847, "loss": 3.4736, "step": 34400 }, { "epoch": 3.7078893552900656, "grad_norm": 0.7093660235404968, "learning_rate": 0.00037801745501562324, "loss": 3.4731, "step": 34450 }, { "epoch": 3.713270907329674, "grad_norm": 0.6421014666557312, "learning_rate": 0.0003776942139855619, "loss": 3.4593, "step": 34500 }, { "epoch": 3.7186524593692822, "grad_norm": 0.6160980463027954, "learning_rate": 0.0003773709729555005, "loss": 3.4535, "step": 34550 }, { "epoch": 3.7240340114088903, "grad_norm": 0.6290106177330017, "learning_rate": 0.000377047731925439, "loss": 3.461, "step": 34600 }, { "epoch": 3.7294155634484984, "grad_norm": 0.6211118698120117, "learning_rate": 0.0003767244908953776, "loss": 3.4576, "step": 34650 }, { "epoch": 3.7347971154881066, "grad_norm": 0.6442049145698547, "learning_rate": 0.0003764077146859174, "loss": 3.4573, "step": 34700 }, { "epoch": 3.740178667527715, "grad_norm": 0.6050986647605896, "learning_rate": 0.000376084473655856, "loss": 3.4425, "step": 34750 }, { "epoch": 3.745560219567323, "grad_norm": 0.6353182792663574, "learning_rate": 0.00037576123262579456, "loss": 3.437, "step": 34800 }, { "epoch": 3.7509417716069313, "grad_norm": 0.6667701005935669, "learning_rate": 0.00037543799159573315, "loss": 3.4571, "step": 34850 }, { "epoch": 3.75632332364654, "grad_norm": 0.6093037128448486, "learning_rate": 0.0003751147505656718, "loss": 3.4453, "step": 34900 }, { "epoch": 3.761704875686148, "grad_norm": 0.5967413783073425, "learning_rate": 0.00037479150953561034, "loss": 3.4659, "step": 34950 }, { "epoch": 3.767086427725756, "grad_norm": 0.6182402968406677, "learning_rate": 0.00037446826850554894, "loss": 3.4603, "step": 35000 }, { "epoch": 3.767086427725756, "eval_accuracy": 0.37532618955083, "eval_loss": 3.454987049102783, "eval_runtime": 184.7797, "eval_samples_per_second": 97.473, "eval_steps_per_second": 6.094, "step": 35000 }, { "epoch": 3.772467979765364, "grad_norm": 0.6796460747718811, "learning_rate": 0.00037414502747548753, "loss": 3.4427, "step": 35050 }, { "epoch": 3.7778495318049723, "grad_norm": 0.6759485006332397, "learning_rate": 0.00037382178644542607, "loss": 3.4735, "step": 35100 }, { "epoch": 3.783231083844581, "grad_norm": 0.6431259512901306, "learning_rate": 0.00037349854541536467, "loss": 3.4669, "step": 35150 }, { "epoch": 3.788612635884189, "grad_norm": 0.6208698153495789, "learning_rate": 0.0003731753043853033, "loss": 3.4517, "step": 35200 }, { "epoch": 3.793994187923797, "grad_norm": 0.6531832218170166, "learning_rate": 0.00037285206335524186, "loss": 3.459, "step": 35250 }, { "epoch": 3.7993757399634056, "grad_norm": 0.6968291997909546, "learning_rate": 0.00037252882232518045, "loss": 3.45, "step": 35300 }, { "epoch": 3.8047572920030137, "grad_norm": 0.6449758410453796, "learning_rate": 0.000372205581295119, "loss": 3.4481, "step": 35350 }, { "epoch": 3.810138844042622, "grad_norm": 0.6235898733139038, "learning_rate": 0.0003718823402650576, "loss": 3.4659, "step": 35400 }, { "epoch": 3.8155203960822304, "grad_norm": 0.7655190825462341, "learning_rate": 0.00037155909923499624, "loss": 3.4616, "step": 35450 }, { "epoch": 3.8209019481218385, "grad_norm": 0.6297181248664856, "learning_rate": 0.0003712358582049348, "loss": 3.4774, "step": 35500 }, { "epoch": 3.8262835001614466, "grad_norm": 0.6621859073638916, "learning_rate": 0.00037091261717487337, "loss": 3.4521, "step": 35550 }, { "epoch": 3.8316650522010547, "grad_norm": 0.5900010466575623, "learning_rate": 0.00037058937614481197, "loss": 3.447, "step": 35600 }, { "epoch": 3.837046604240663, "grad_norm": 0.6074267625808716, "learning_rate": 0.0003702661351147505, "loss": 3.4478, "step": 35650 }, { "epoch": 3.8424281562802713, "grad_norm": 0.616184413433075, "learning_rate": 0.0003699428940846891, "loss": 3.431, "step": 35700 }, { "epoch": 3.8478097083198795, "grad_norm": 0.6139175891876221, "learning_rate": 0.00036961965305462775, "loss": 3.4688, "step": 35750 }, { "epoch": 3.8531912603594876, "grad_norm": 0.6691179871559143, "learning_rate": 0.0003692964120245663, "loss": 3.4461, "step": 35800 }, { "epoch": 3.858572812399096, "grad_norm": 0.6166695356369019, "learning_rate": 0.0003689731709945049, "loss": 3.4642, "step": 35850 }, { "epoch": 3.863954364438704, "grad_norm": 0.6620895266532898, "learning_rate": 0.0003686499299644434, "loss": 3.4599, "step": 35900 }, { "epoch": 3.8693359164783123, "grad_norm": 0.58867347240448, "learning_rate": 0.000368326688934382, "loss": 3.4494, "step": 35950 }, { "epoch": 3.8747174685179204, "grad_norm": 0.6114178895950317, "learning_rate": 0.0003680034479043206, "loss": 3.4545, "step": 36000 }, { "epoch": 3.8747174685179204, "eval_accuracy": 0.37561944369504136, "eval_loss": 3.451740026473999, "eval_runtime": 184.9248, "eval_samples_per_second": 97.396, "eval_steps_per_second": 6.089, "step": 36000 }, { "epoch": 3.8800990205575285, "grad_norm": 0.6434826850891113, "learning_rate": 0.0003676802068742592, "loss": 3.4546, "step": 36050 }, { "epoch": 3.885480572597137, "grad_norm": 0.6056340336799622, "learning_rate": 0.0003673569658441978, "loss": 3.4587, "step": 36100 }, { "epoch": 3.890862124636745, "grad_norm": 0.6036874055862427, "learning_rate": 0.0003670337248141364, "loss": 3.4457, "step": 36150 }, { "epoch": 3.8962436766763533, "grad_norm": 0.6100910305976868, "learning_rate": 0.00036671048378407494, "loss": 3.4508, "step": 36200 }, { "epoch": 3.901625228715962, "grad_norm": 0.6239060759544373, "learning_rate": 0.00036638724275401353, "loss": 3.4465, "step": 36250 }, { "epoch": 3.90700678075557, "grad_norm": 0.6256414651870728, "learning_rate": 0.00036607046654455334, "loss": 3.4362, "step": 36300 }, { "epoch": 3.912388332795178, "grad_norm": 0.5970678329467773, "learning_rate": 0.00036574722551449193, "loss": 3.4634, "step": 36350 }, { "epoch": 3.9177698848347866, "grad_norm": 0.6506580114364624, "learning_rate": 0.00036542398448443053, "loss": 3.4448, "step": 36400 }, { "epoch": 3.9231514368743947, "grad_norm": 0.64057856798172, "learning_rate": 0.0003651007434543691, "loss": 3.445, "step": 36450 }, { "epoch": 3.928532988914003, "grad_norm": 0.6665002703666687, "learning_rate": 0.0003647775024243077, "loss": 3.4366, "step": 36500 }, { "epoch": 3.933914540953611, "grad_norm": 0.6165843605995178, "learning_rate": 0.00036445426139424626, "loss": 3.4316, "step": 36550 }, { "epoch": 3.939296092993219, "grad_norm": 0.6074864268302917, "learning_rate": 0.00036413102036418485, "loss": 3.4501, "step": 36600 }, { "epoch": 3.9446776450328276, "grad_norm": 0.7129690647125244, "learning_rate": 0.0003638077793341234, "loss": 3.4409, "step": 36650 }, { "epoch": 3.9500591970724357, "grad_norm": 0.6536222100257874, "learning_rate": 0.00036348453830406204, "loss": 3.4472, "step": 36700 }, { "epoch": 3.955440749112044, "grad_norm": 0.5931937098503113, "learning_rate": 0.00036316129727400064, "loss": 3.4439, "step": 36750 }, { "epoch": 3.9608223011516523, "grad_norm": 0.6528921127319336, "learning_rate": 0.0003628380562439392, "loss": 3.4427, "step": 36800 }, { "epoch": 3.9662038531912605, "grad_norm": 0.5929076075553894, "learning_rate": 0.00036251481521387777, "loss": 3.4536, "step": 36850 }, { "epoch": 3.9715854052308686, "grad_norm": 0.6636005640029907, "learning_rate": 0.00036219157418381637, "loss": 3.4588, "step": 36900 }, { "epoch": 3.9769669572704767, "grad_norm": 0.5989009737968445, "learning_rate": 0.0003618683331537549, "loss": 3.4483, "step": 36950 }, { "epoch": 3.9823485093100848, "grad_norm": 0.6094232201576233, "learning_rate": 0.00036154509212369356, "loss": 3.456, "step": 37000 }, { "epoch": 3.9823485093100848, "eval_accuracy": 0.37639761567756863, "eval_loss": 3.444646120071411, "eval_runtime": 184.7393, "eval_samples_per_second": 97.494, "eval_steps_per_second": 6.095, "step": 37000 }, { "epoch": 3.9877300613496933, "grad_norm": 0.6560502648353577, "learning_rate": 0.00036122185109363215, "loss": 3.4508, "step": 37050 }, { "epoch": 3.9931116133893014, "grad_norm": 0.5955866575241089, "learning_rate": 0.0003608986100635707, "loss": 3.4503, "step": 37100 }, { "epoch": 3.9984931654289095, "grad_norm": 0.6154381036758423, "learning_rate": 0.0003605753690335093, "loss": 3.4548, "step": 37150 }, { "epoch": 4.003874717468518, "grad_norm": 0.6376234292984009, "learning_rate": 0.0003602521280034478, "loss": 3.3711, "step": 37200 }, { "epoch": 4.009256269508126, "grad_norm": 0.6253707408905029, "learning_rate": 0.0003599288869733865, "loss": 3.3483, "step": 37250 }, { "epoch": 4.014637821547734, "grad_norm": 0.6334782838821411, "learning_rate": 0.00035960564594332507, "loss": 3.3408, "step": 37300 }, { "epoch": 4.020019373587343, "grad_norm": 0.6270228624343872, "learning_rate": 0.0003592824049132636, "loss": 3.3435, "step": 37350 }, { "epoch": 4.0254009256269505, "grad_norm": 0.6254666447639465, "learning_rate": 0.0003589591638832022, "loss": 3.3457, "step": 37400 }, { "epoch": 4.030782477666559, "grad_norm": 0.7249552607536316, "learning_rate": 0.0003586359228531408, "loss": 3.3514, "step": 37450 }, { "epoch": 4.036164029706168, "grad_norm": 0.6332011222839355, "learning_rate": 0.00035831268182307934, "loss": 3.3702, "step": 37500 }, { "epoch": 4.041545581745775, "grad_norm": 0.6144351959228516, "learning_rate": 0.000357989440793018, "loss": 3.3694, "step": 37550 }, { "epoch": 4.046927133785384, "grad_norm": 0.6586161851882935, "learning_rate": 0.0003576661997629566, "loss": 3.3739, "step": 37600 }, { "epoch": 4.0523086858249915, "grad_norm": 0.6490854024887085, "learning_rate": 0.0003573429587328951, "loss": 3.3556, "step": 37650 }, { "epoch": 4.0576902378646, "grad_norm": 0.6474526524543762, "learning_rate": 0.0003570197177028337, "loss": 3.3655, "step": 37700 }, { "epoch": 4.063071789904209, "grad_norm": 0.6664512157440186, "learning_rate": 0.00035669647667277226, "loss": 3.362, "step": 37750 }, { "epoch": 4.068453341943816, "grad_norm": 0.6286465525627136, "learning_rate": 0.00035637323564271085, "loss": 3.385, "step": 37800 }, { "epoch": 4.073834893983425, "grad_norm": 0.6096540093421936, "learning_rate": 0.0003560499946126495, "loss": 3.3604, "step": 37850 }, { "epoch": 4.079216446023033, "grad_norm": 0.6467686295509338, "learning_rate": 0.00035572675358258804, "loss": 3.3638, "step": 37900 }, { "epoch": 4.084597998062641, "grad_norm": 0.6298966407775879, "learning_rate": 0.00035540351255252664, "loss": 3.3726, "step": 37950 }, { "epoch": 4.08997955010225, "grad_norm": 0.7064566612243652, "learning_rate": 0.00035508027152246523, "loss": 3.376, "step": 38000 }, { "epoch": 4.08997955010225, "eval_accuracy": 0.37656059500929034, "eval_loss": 3.446718692779541, "eval_runtime": 184.5393, "eval_samples_per_second": 97.6, "eval_steps_per_second": 6.102, "step": 38000 }, { "epoch": 4.095361102141858, "grad_norm": 0.6466246843338013, "learning_rate": 0.0003547570304924038, "loss": 3.3837, "step": 38050 }, { "epoch": 4.100742654181466, "grad_norm": 0.6541309952735901, "learning_rate": 0.0003544337894623424, "loss": 3.3768, "step": 38100 }, { "epoch": 4.106124206221074, "grad_norm": 0.6219347715377808, "learning_rate": 0.000354110548432281, "loss": 3.361, "step": 38150 }, { "epoch": 4.111505758260682, "grad_norm": 0.6315196752548218, "learning_rate": 0.00035378730740221956, "loss": 3.3712, "step": 38200 }, { "epoch": 4.1168873103002905, "grad_norm": 0.6126587390899658, "learning_rate": 0.00035346406637215815, "loss": 3.3761, "step": 38250 }, { "epoch": 4.122268862339899, "grad_norm": 0.7233909368515015, "learning_rate": 0.0003531408253420967, "loss": 3.344, "step": 38300 }, { "epoch": 4.127650414379507, "grad_norm": 0.6340131163597107, "learning_rate": 0.0003528175843120353, "loss": 3.3787, "step": 38350 }, { "epoch": 4.133031966419115, "grad_norm": 0.6910709738731384, "learning_rate": 0.00035249434328197394, "loss": 3.3778, "step": 38400 }, { "epoch": 4.138413518458724, "grad_norm": 0.6195011138916016, "learning_rate": 0.0003521711022519125, "loss": 3.3878, "step": 38450 }, { "epoch": 4.1437950704983315, "grad_norm": 0.6645599007606506, "learning_rate": 0.00035184786122185107, "loss": 3.3773, "step": 38500 }, { "epoch": 4.14917662253794, "grad_norm": 0.6299995183944702, "learning_rate": 0.0003515310850123909, "loss": 3.3798, "step": 38550 }, { "epoch": 4.154558174577549, "grad_norm": 0.6655153632164001, "learning_rate": 0.00035120784398232947, "loss": 3.3805, "step": 38600 }, { "epoch": 4.159939726617156, "grad_norm": 0.626811683177948, "learning_rate": 0.000350884602952268, "loss": 3.3592, "step": 38650 }, { "epoch": 4.165321278656765, "grad_norm": 0.647689938545227, "learning_rate": 0.0003505613619222066, "loss": 3.383, "step": 38700 }, { "epoch": 4.1707028306963725, "grad_norm": 0.6576713919639587, "learning_rate": 0.0003502381208921452, "loss": 3.3792, "step": 38750 }, { "epoch": 4.176084382735981, "grad_norm": 0.6431198716163635, "learning_rate": 0.0003499148798620838, "loss": 3.3732, "step": 38800 }, { "epoch": 4.18146593477559, "grad_norm": 0.691774845123291, "learning_rate": 0.0003495916388320224, "loss": 3.3809, "step": 38850 }, { "epoch": 4.186847486815197, "grad_norm": 0.6469290852546692, "learning_rate": 0.000349268397801961, "loss": 3.3734, "step": 38900 }, { "epoch": 4.192229038854806, "grad_norm": 0.7078692317008972, "learning_rate": 0.0003489451567718995, "loss": 3.3864, "step": 38950 }, { "epoch": 4.197610590894414, "grad_norm": 0.6617368459701538, "learning_rate": 0.0003486219157418381, "loss": 3.3961, "step": 39000 }, { "epoch": 4.197610590894414, "eval_accuracy": 0.3772209872614268, "eval_loss": 3.4436137676239014, "eval_runtime": 184.7722, "eval_samples_per_second": 97.477, "eval_steps_per_second": 6.094, "step": 39000 }, { "epoch": 4.202992142934022, "grad_norm": 0.6935890316963196, "learning_rate": 0.00034829867471177677, "loss": 3.3808, "step": 39050 }, { "epoch": 4.208373694973631, "grad_norm": 0.6806594133377075, "learning_rate": 0.0003479754336817153, "loss": 3.3771, "step": 39100 }, { "epoch": 4.213755247013238, "grad_norm": 0.6456026434898376, "learning_rate": 0.0003476521926516539, "loss": 3.4124, "step": 39150 }, { "epoch": 4.219136799052847, "grad_norm": 0.5707885026931763, "learning_rate": 0.00034732895162159245, "loss": 3.3916, "step": 39200 }, { "epoch": 4.224518351092455, "grad_norm": 0.6426082849502563, "learning_rate": 0.00034700571059153104, "loss": 3.376, "step": 39250 }, { "epoch": 4.229899903132063, "grad_norm": 0.6489585638046265, "learning_rate": 0.00034668246956146963, "loss": 3.3809, "step": 39300 }, { "epoch": 4.2352814551716715, "grad_norm": 0.6257960796356201, "learning_rate": 0.00034635922853140823, "loss": 3.3873, "step": 39350 }, { "epoch": 4.24066300721128, "grad_norm": 0.698000967502594, "learning_rate": 0.0003460359875013468, "loss": 3.3771, "step": 39400 }, { "epoch": 4.246044559250888, "grad_norm": 0.6346425414085388, "learning_rate": 0.0003457127464712854, "loss": 3.3877, "step": 39450 }, { "epoch": 4.251426111290496, "grad_norm": 0.6701180338859558, "learning_rate": 0.00034538950544122396, "loss": 3.3697, "step": 39500 }, { "epoch": 4.256807663330104, "grad_norm": 0.6371496319770813, "learning_rate": 0.00034506626441116255, "loss": 3.3843, "step": 39550 }, { "epoch": 4.2621892153697125, "grad_norm": 0.6401668190956116, "learning_rate": 0.0003447430233811011, "loss": 3.3913, "step": 39600 }, { "epoch": 4.267570767409321, "grad_norm": 0.6551806926727295, "learning_rate": 0.00034441978235103974, "loss": 3.3918, "step": 39650 }, { "epoch": 4.272952319448929, "grad_norm": 0.6413207054138184, "learning_rate": 0.00034409654132097834, "loss": 3.3873, "step": 39700 }, { "epoch": 4.278333871488537, "grad_norm": 0.6472829580307007, "learning_rate": 0.0003437733002909169, "loss": 3.3974, "step": 39750 }, { "epoch": 4.283715423528146, "grad_norm": 0.6183416247367859, "learning_rate": 0.00034345005926085547, "loss": 3.3703, "step": 39800 }, { "epoch": 4.2890969755677535, "grad_norm": 0.715303361415863, "learning_rate": 0.00034312681823079407, "loss": 3.3877, "step": 39850 }, { "epoch": 4.294478527607362, "grad_norm": 0.6685398817062378, "learning_rate": 0.00034280357720073266, "loss": 3.3878, "step": 39900 }, { "epoch": 4.299860079646971, "grad_norm": 0.6901618242263794, "learning_rate": 0.00034248033617067126, "loss": 3.3789, "step": 39950 }, { "epoch": 4.305241631686578, "grad_norm": 0.6296851634979248, "learning_rate": 0.00034215709514060985, "loss": 3.3729, "step": 40000 }, { "epoch": 4.305241631686578, "eval_accuracy": 0.3777952177735263, "eval_loss": 3.437654972076416, "eval_runtime": 184.7949, "eval_samples_per_second": 97.465, "eval_steps_per_second": 6.093, "step": 40000 }, { "epoch": 4.310623183726187, "grad_norm": 0.7259793877601624, "learning_rate": 0.0003418338541105484, "loss": 3.3616, "step": 40050 }, { "epoch": 4.3160047357657945, "grad_norm": 0.6109891533851624, "learning_rate": 0.000341510613080487, "loss": 3.3857, "step": 40100 }, { "epoch": 4.321386287805403, "grad_norm": 0.6450995802879333, "learning_rate": 0.0003411873720504255, "loss": 3.389, "step": 40150 }, { "epoch": 4.326767839845012, "grad_norm": 0.6640621423721313, "learning_rate": 0.0003408641310203642, "loss": 3.3955, "step": 40200 }, { "epoch": 4.332149391884619, "grad_norm": 0.6404122710227966, "learning_rate": 0.00034054088999030277, "loss": 3.3787, "step": 40250 }, { "epoch": 4.337530943924228, "grad_norm": 0.6834564208984375, "learning_rate": 0.0003402176489602413, "loss": 3.3903, "step": 40300 }, { "epoch": 4.342912495963836, "grad_norm": 0.7080681324005127, "learning_rate": 0.0003398944079301799, "loss": 3.3965, "step": 40350 }, { "epoch": 4.348294048003444, "grad_norm": 0.6617289185523987, "learning_rate": 0.0003395711669001185, "loss": 3.4053, "step": 40400 }, { "epoch": 4.3536756000430525, "grad_norm": 0.6993653178215027, "learning_rate": 0.00033924792587005704, "loss": 3.3868, "step": 40450 }, { "epoch": 4.359057152082661, "grad_norm": 0.6167160868644714, "learning_rate": 0.0003389246848399957, "loss": 3.3719, "step": 40500 }, { "epoch": 4.364438704122269, "grad_norm": 0.669032096862793, "learning_rate": 0.0003386014438099343, "loss": 3.3802, "step": 40550 }, { "epoch": 4.369820256161877, "grad_norm": 0.619784414768219, "learning_rate": 0.0003382782027798728, "loss": 3.3882, "step": 40600 }, { "epoch": 4.375201808201485, "grad_norm": 0.6738349199295044, "learning_rate": 0.0003379549617498114, "loss": 3.3755, "step": 40650 }, { "epoch": 4.3805833602410935, "grad_norm": 0.6499971151351929, "learning_rate": 0.00033763172071974996, "loss": 3.3892, "step": 40700 }, { "epoch": 4.385964912280702, "grad_norm": 0.6084420084953308, "learning_rate": 0.00033730847968968855, "loss": 3.3835, "step": 40750 }, { "epoch": 4.39134646432031, "grad_norm": 0.6882473230361938, "learning_rate": 0.0003369852386596272, "loss": 3.3852, "step": 40800 }, { "epoch": 4.396728016359918, "grad_norm": 0.6895729303359985, "learning_rate": 0.00033666199762956574, "loss": 3.3923, "step": 40850 }, { "epoch": 4.402109568399527, "grad_norm": 0.6523504853248596, "learning_rate": 0.00033633875659950434, "loss": 3.3847, "step": 40900 }, { "epoch": 4.4074911204391345, "grad_norm": 0.619976282119751, "learning_rate": 0.0003360155155694429, "loss": 3.3709, "step": 40950 }, { "epoch": 4.412872672478743, "grad_norm": 0.6657698750495911, "learning_rate": 0.0003356922745393815, "loss": 3.3866, "step": 41000 }, { "epoch": 4.412872672478743, "eval_accuracy": 0.3780848863724397, "eval_loss": 3.4327142238616943, "eval_runtime": 184.5831, "eval_samples_per_second": 97.577, "eval_steps_per_second": 6.1, "step": 41000 }, { "epoch": 4.418254224518351, "grad_norm": 0.6433376669883728, "learning_rate": 0.0003353690335093201, "loss": 3.3842, "step": 41050 }, { "epoch": 4.423635776557959, "grad_norm": 0.676450252532959, "learning_rate": 0.0003350457924792587, "loss": 3.4097, "step": 41100 }, { "epoch": 4.429017328597568, "grad_norm": 0.6363053321838379, "learning_rate": 0.00033472255144919726, "loss": 3.389, "step": 41150 }, { "epoch": 4.4343988806371755, "grad_norm": 0.6782971620559692, "learning_rate": 0.00033439931041913585, "loss": 3.4028, "step": 41200 }, { "epoch": 4.439780432676784, "grad_norm": 0.6248264908790588, "learning_rate": 0.00033408253420967566, "loss": 3.3992, "step": 41250 }, { "epoch": 4.445161984716393, "grad_norm": 0.617743194103241, "learning_rate": 0.00033375929317961425, "loss": 3.3682, "step": 41300 }, { "epoch": 4.450543536756, "grad_norm": 0.7119070887565613, "learning_rate": 0.0003334360521495528, "loss": 3.3864, "step": 41350 }, { "epoch": 4.455925088795609, "grad_norm": 0.6565623879432678, "learning_rate": 0.0003331128111194914, "loss": 3.3814, "step": 41400 }, { "epoch": 4.461306640835216, "grad_norm": 0.7476305365562439, "learning_rate": 0.00033278957008943004, "loss": 3.384, "step": 41450 }, { "epoch": 4.466688192874825, "grad_norm": 0.6427007913589478, "learning_rate": 0.0003324663290593686, "loss": 3.3901, "step": 41500 }, { "epoch": 4.4720697449144335, "grad_norm": 0.6926642656326294, "learning_rate": 0.00033214308802930717, "loss": 3.3919, "step": 41550 }, { "epoch": 4.477451296954041, "grad_norm": 0.687822163105011, "learning_rate": 0.0003318198469992457, "loss": 3.3926, "step": 41600 }, { "epoch": 4.48283284899365, "grad_norm": 0.695188045501709, "learning_rate": 0.0003314966059691843, "loss": 3.3966, "step": 41650 }, { "epoch": 4.488214401033258, "grad_norm": 0.6738727688789368, "learning_rate": 0.00033117336493912296, "loss": 3.4002, "step": 41700 }, { "epoch": 4.493595953072866, "grad_norm": 0.6818880438804626, "learning_rate": 0.0003308501239090615, "loss": 3.3993, "step": 41750 }, { "epoch": 4.4989775051124745, "grad_norm": 0.6498254537582397, "learning_rate": 0.0003305268828790001, "loss": 3.4005, "step": 41800 }, { "epoch": 4.504359057152083, "grad_norm": 0.6302522420883179, "learning_rate": 0.0003302036418489387, "loss": 3.3798, "step": 41850 }, { "epoch": 4.509740609191691, "grad_norm": 0.6108139157295227, "learning_rate": 0.0003298804008188772, "loss": 3.36, "step": 41900 }, { "epoch": 4.515122161231299, "grad_norm": 0.6787891983985901, "learning_rate": 0.0003295571597888158, "loss": 3.3822, "step": 41950 }, { "epoch": 4.520503713270907, "grad_norm": 0.6845345497131348, "learning_rate": 0.00032923391875875447, "loss": 3.3773, "step": 42000 }, { "epoch": 4.520503713270907, "eval_accuracy": 0.3784473524061888, "eval_loss": 3.427448272705078, "eval_runtime": 184.8487, "eval_samples_per_second": 97.436, "eval_steps_per_second": 6.091, "step": 42000 }, { "epoch": 4.5258852653105155, "grad_norm": 0.6910591125488281, "learning_rate": 0.000328910677728693, "loss": 3.3944, "step": 42050 }, { "epoch": 4.531266817350124, "grad_norm": 0.6579834222793579, "learning_rate": 0.0003285874366986316, "loss": 3.3767, "step": 42100 }, { "epoch": 4.536648369389732, "grad_norm": 0.6646836996078491, "learning_rate": 0.00032826419566857015, "loss": 3.3925, "step": 42150 }, { "epoch": 4.54202992142934, "grad_norm": 0.6295957565307617, "learning_rate": 0.00032794095463850874, "loss": 3.3939, "step": 42200 }, { "epoch": 4.547411473468949, "grad_norm": 0.6931584477424622, "learning_rate": 0.0003276177136084473, "loss": 3.3813, "step": 42250 }, { "epoch": 4.5527930255085565, "grad_norm": 0.728244423866272, "learning_rate": 0.00032729447257838593, "loss": 3.3823, "step": 42300 }, { "epoch": 4.558174577548165, "grad_norm": 0.6908766627311707, "learning_rate": 0.0003269712315483245, "loss": 3.3996, "step": 42350 }, { "epoch": 4.563556129587774, "grad_norm": 0.6593059301376343, "learning_rate": 0.00032664799051826306, "loss": 3.3678, "step": 42400 }, { "epoch": 4.568937681627381, "grad_norm": 0.7266299724578857, "learning_rate": 0.00032632474948820166, "loss": 3.3908, "step": 42450 }, { "epoch": 4.57431923366699, "grad_norm": 0.6872548460960388, "learning_rate": 0.00032600150845814025, "loss": 3.3937, "step": 42500 }, { "epoch": 4.579700785706597, "grad_norm": 0.7212083339691162, "learning_rate": 0.0003256782674280788, "loss": 3.3884, "step": 42550 }, { "epoch": 4.585082337746206, "grad_norm": 0.6755133867263794, "learning_rate": 0.00032535502639801744, "loss": 3.3797, "step": 42600 }, { "epoch": 4.5904638897858145, "grad_norm": 0.634061336517334, "learning_rate": 0.00032503178536795604, "loss": 3.3579, "step": 42650 }, { "epoch": 4.595845441825422, "grad_norm": 0.6569266319274902, "learning_rate": 0.0003247085443378946, "loss": 3.3904, "step": 42700 }, { "epoch": 4.601226993865031, "grad_norm": 0.6839199066162109, "learning_rate": 0.0003243853033078332, "loss": 3.3873, "step": 42750 }, { "epoch": 4.606608545904638, "grad_norm": 0.6381998062133789, "learning_rate": 0.0003240620622777717, "loss": 3.3958, "step": 42800 }, { "epoch": 4.611990097944247, "grad_norm": 0.6814762949943542, "learning_rate": 0.00032373882124771036, "loss": 3.3978, "step": 42850 }, { "epoch": 4.6173716499838555, "grad_norm": 0.6227290630340576, "learning_rate": 0.00032341558021764896, "loss": 3.3834, "step": 42900 }, { "epoch": 4.622753202023463, "grad_norm": 0.644250750541687, "learning_rate": 0.0003230923391875875, "loss": 3.3759, "step": 42950 }, { "epoch": 4.628134754063072, "grad_norm": 0.6935346722602844, "learning_rate": 0.0003227690981575261, "loss": 3.4146, "step": 43000 }, { "epoch": 4.628134754063072, "eval_accuracy": 0.37911926186443357, "eval_loss": 3.4232747554779053, "eval_runtime": 184.608, "eval_samples_per_second": 97.563, "eval_steps_per_second": 6.099, "step": 43000 }, { "epoch": 4.63351630610268, "grad_norm": 0.6674992442131042, "learning_rate": 0.0003224458571274647, "loss": 3.3878, "step": 43050 }, { "epoch": 4.638897858142288, "grad_norm": 0.6759820580482483, "learning_rate": 0.00032212261609740323, "loss": 3.3895, "step": 43100 }, { "epoch": 4.6442794101818965, "grad_norm": 0.687537670135498, "learning_rate": 0.0003217993750673419, "loss": 3.3907, "step": 43150 }, { "epoch": 4.649660962221505, "grad_norm": 0.6385945081710815, "learning_rate": 0.00032147613403728047, "loss": 3.3825, "step": 43200 }, { "epoch": 4.655042514261113, "grad_norm": 0.7066798210144043, "learning_rate": 0.000321152893007219, "loss": 3.3945, "step": 43250 }, { "epoch": 4.660424066300721, "grad_norm": 0.7198359966278076, "learning_rate": 0.0003208296519771576, "loss": 3.4057, "step": 43300 }, { "epoch": 4.665805618340329, "grad_norm": 0.6627112627029419, "learning_rate": 0.00032050641094709615, "loss": 3.3895, "step": 43350 }, { "epoch": 4.6711871703799375, "grad_norm": 0.6638057827949524, "learning_rate": 0.00032018316991703474, "loss": 3.3827, "step": 43400 }, { "epoch": 4.676568722419546, "grad_norm": 0.641120970249176, "learning_rate": 0.0003198599288869734, "loss": 3.3974, "step": 43450 }, { "epoch": 4.681950274459154, "grad_norm": 0.6419323682785034, "learning_rate": 0.00031953668785691193, "loss": 3.3961, "step": 43500 }, { "epoch": 4.687331826498762, "grad_norm": 0.7281374335289001, "learning_rate": 0.0003192134468268505, "loss": 3.3956, "step": 43550 }, { "epoch": 4.692713378538371, "grad_norm": 0.6420050859451294, "learning_rate": 0.0003188902057967891, "loss": 3.3683, "step": 43600 }, { "epoch": 4.6980949305779784, "grad_norm": 0.6763567328453064, "learning_rate": 0.00031856696476672766, "loss": 3.3961, "step": 43650 }, { "epoch": 4.703476482617587, "grad_norm": 0.7230156064033508, "learning_rate": 0.0003182437237366663, "loss": 3.4111, "step": 43700 }, { "epoch": 4.7088580346571955, "grad_norm": 0.7006603479385376, "learning_rate": 0.0003179204827066049, "loss": 3.4003, "step": 43750 }, { "epoch": 4.714239586696803, "grad_norm": 0.6924633383750916, "learning_rate": 0.00031759724167654344, "loss": 3.3832, "step": 43800 }, { "epoch": 4.719621138736412, "grad_norm": 0.7065801024436951, "learning_rate": 0.00031727400064648204, "loss": 3.3894, "step": 43850 }, { "epoch": 4.725002690776019, "grad_norm": 0.6304015517234802, "learning_rate": 0.0003169507596164206, "loss": 3.3659, "step": 43900 }, { "epoch": 4.730384242815628, "grad_norm": 0.6763742566108704, "learning_rate": 0.0003166275185863592, "loss": 3.3802, "step": 43950 }, { "epoch": 4.7357657948552365, "grad_norm": 0.6544877886772156, "learning_rate": 0.0003163042775562978, "loss": 3.3892, "step": 44000 }, { "epoch": 4.7357657948552365, "eval_accuracy": 0.3791140465258185, "eval_loss": 3.4197874069213867, "eval_runtime": 184.7556, "eval_samples_per_second": 97.486, "eval_steps_per_second": 6.095, "step": 44000 }, { "epoch": 4.741147346894844, "grad_norm": 0.6808891296386719, "learning_rate": 0.00031598103652623636, "loss": 3.3747, "step": 44050 }, { "epoch": 4.746528898934453, "grad_norm": 0.6510489583015442, "learning_rate": 0.00031565779549617496, "loss": 3.3851, "step": 44100 }, { "epoch": 4.751910450974061, "grad_norm": 0.659172773361206, "learning_rate": 0.00031534101928671476, "loss": 3.3789, "step": 44150 }, { "epoch": 4.757292003013669, "grad_norm": 0.7054234743118286, "learning_rate": 0.00031501777825665336, "loss": 3.3755, "step": 44200 }, { "epoch": 4.7626735550532775, "grad_norm": 0.7176251411437988, "learning_rate": 0.0003146945372265919, "loss": 3.3863, "step": 44250 }, { "epoch": 4.768055107092886, "grad_norm": 0.751799464225769, "learning_rate": 0.0003143712961965305, "loss": 3.3894, "step": 44300 }, { "epoch": 4.773436659132494, "grad_norm": 0.6770383715629578, "learning_rate": 0.0003140480551664691, "loss": 3.3808, "step": 44350 }, { "epoch": 4.778818211172102, "grad_norm": 0.6924847960472107, "learning_rate": 0.0003137248141364077, "loss": 3.3874, "step": 44400 }, { "epoch": 4.78419976321171, "grad_norm": 0.633284330368042, "learning_rate": 0.0003134015731063463, "loss": 3.3971, "step": 44450 }, { "epoch": 4.7895813152513185, "grad_norm": 0.6829878091812134, "learning_rate": 0.0003130783320762849, "loss": 3.3811, "step": 44500 }, { "epoch": 4.794962867290927, "grad_norm": 0.6898048520088196, "learning_rate": 0.0003127550910462234, "loss": 3.3897, "step": 44550 }, { "epoch": 4.800344419330535, "grad_norm": 0.6750717163085938, "learning_rate": 0.000312431850016162, "loss": 3.3971, "step": 44600 }, { "epoch": 4.805725971370143, "grad_norm": 0.6479890942573547, "learning_rate": 0.00031210860898610066, "loss": 3.406, "step": 44650 }, { "epoch": 4.811107523409751, "grad_norm": 0.7591122984886169, "learning_rate": 0.0003117853679560392, "loss": 3.3913, "step": 44700 }, { "epoch": 4.8164890754493594, "grad_norm": 0.6572939157485962, "learning_rate": 0.0003114621269259778, "loss": 3.38, "step": 44750 }, { "epoch": 4.821870627488968, "grad_norm": 0.6863782405853271, "learning_rate": 0.00031113888589591633, "loss": 3.3788, "step": 44800 }, { "epoch": 4.827252179528576, "grad_norm": 0.6886435747146606, "learning_rate": 0.00031081564486585493, "loss": 3.3875, "step": 44850 }, { "epoch": 4.832633731568184, "grad_norm": 0.7485862970352173, "learning_rate": 0.0003104924038357935, "loss": 3.3789, "step": 44900 }, { "epoch": 4.838015283607793, "grad_norm": 0.666218638420105, "learning_rate": 0.0003101691628057321, "loss": 3.3945, "step": 44950 }, { "epoch": 4.8433968356474, "grad_norm": 0.6517705917358398, "learning_rate": 0.0003098459217756707, "loss": 3.401, "step": 45000 }, { "epoch": 4.8433968356474, "eval_accuracy": 0.38019872830487045, "eval_loss": 3.4125325679779053, "eval_runtime": 184.9758, "eval_samples_per_second": 97.369, "eval_steps_per_second": 6.087, "step": 45000 }, { "epoch": 4.848778387687009, "grad_norm": 0.6660953760147095, "learning_rate": 0.0003095226807456093, "loss": 3.3751, "step": 45050 }, { "epoch": 4.8541599397266175, "grad_norm": 0.6609524488449097, "learning_rate": 0.00030919943971554785, "loss": 3.3946, "step": 45100 }, { "epoch": 4.859541491766225, "grad_norm": 0.6775374412536621, "learning_rate": 0.00030887619868548644, "loss": 3.3883, "step": 45150 }, { "epoch": 4.864923043805834, "grad_norm": 0.664930522441864, "learning_rate": 0.000308552957655425, "loss": 3.3975, "step": 45200 }, { "epoch": 4.870304595845441, "grad_norm": 0.6616421341896057, "learning_rate": 0.00030822971662536363, "loss": 3.3899, "step": 45250 }, { "epoch": 4.87568614788505, "grad_norm": 0.7150900959968567, "learning_rate": 0.0003079064755953022, "loss": 3.3785, "step": 45300 }, { "epoch": 4.8810676999246585, "grad_norm": 0.6499496698379517, "learning_rate": 0.00030758323456524077, "loss": 3.3818, "step": 45350 }, { "epoch": 4.886449251964266, "grad_norm": 0.658064603805542, "learning_rate": 0.00030725999353517936, "loss": 3.3796, "step": 45400 }, { "epoch": 4.891830804003875, "grad_norm": 0.6768323183059692, "learning_rate": 0.00030693675250511795, "loss": 3.3858, "step": 45450 }, { "epoch": 4.897212356043483, "grad_norm": 0.679030179977417, "learning_rate": 0.00030661351147505655, "loss": 3.3737, "step": 45500 }, { "epoch": 4.902593908083091, "grad_norm": 0.6560843586921692, "learning_rate": 0.00030629027044499514, "loss": 3.3828, "step": 45550 }, { "epoch": 4.9079754601226995, "grad_norm": 0.6728476881980896, "learning_rate": 0.00030596702941493374, "loss": 3.3804, "step": 45600 }, { "epoch": 4.913357012162308, "grad_norm": 0.736443817615509, "learning_rate": 0.0003056437883848723, "loss": 3.3771, "step": 45650 }, { "epoch": 4.918738564201916, "grad_norm": 0.6769527792930603, "learning_rate": 0.0003053205473548109, "loss": 3.4014, "step": 45700 }, { "epoch": 4.924120116241524, "grad_norm": 0.6532332301139832, "learning_rate": 0.0003049973063247494, "loss": 3.3797, "step": 45750 }, { "epoch": 4.929501668281132, "grad_norm": 0.6099948287010193, "learning_rate": 0.00030467406529468806, "loss": 3.3837, "step": 45800 }, { "epoch": 4.9348832203207404, "grad_norm": 0.6482587456703186, "learning_rate": 0.00030435082426462666, "loss": 3.3776, "step": 45850 }, { "epoch": 4.940264772360349, "grad_norm": 0.6524862051010132, "learning_rate": 0.0003040275832345652, "loss": 3.3893, "step": 45900 }, { "epoch": 4.945646324399957, "grad_norm": 0.6760078072547913, "learning_rate": 0.0003037043422045038, "loss": 3.386, "step": 45950 }, { "epoch": 4.951027876439565, "grad_norm": 0.7402188777923584, "learning_rate": 0.0003033811011744424, "loss": 3.3919, "step": 46000 }, { "epoch": 4.951027876439565, "eval_accuracy": 0.3805924863703101, "eval_loss": 3.4080684185028076, "eval_runtime": 184.4763, "eval_samples_per_second": 97.633, "eval_steps_per_second": 6.104, "step": 46000 }, { "epoch": 4.956409428479174, "grad_norm": 0.7083448767662048, "learning_rate": 0.00030305786014438093, "loss": 3.3806, "step": 46050 }, { "epoch": 4.961790980518781, "grad_norm": 0.7174310684204102, "learning_rate": 0.0003027346191143196, "loss": 3.37, "step": 46100 }, { "epoch": 4.96717253255839, "grad_norm": 0.6411959528923035, "learning_rate": 0.00030241137808425817, "loss": 3.3813, "step": 46150 }, { "epoch": 4.9725540845979985, "grad_norm": 0.6771125197410583, "learning_rate": 0.0003020881370541967, "loss": 3.3759, "step": 46200 }, { "epoch": 4.977935636637606, "grad_norm": 0.6481576561927795, "learning_rate": 0.0003017648960241353, "loss": 3.3786, "step": 46250 }, { "epoch": 4.983317188677215, "grad_norm": 0.6613046526908875, "learning_rate": 0.00030144165499407385, "loss": 3.4187, "step": 46300 }, { "epoch": 4.988698740716822, "grad_norm": 0.6821366548538208, "learning_rate": 0.00030111841396401244, "loss": 3.3862, "step": 46350 }, { "epoch": 4.994080292756431, "grad_norm": 0.7078256607055664, "learning_rate": 0.0003007951729339511, "loss": 3.3917, "step": 46400 }, { "epoch": 4.9994618447960395, "grad_norm": 0.6903404593467712, "learning_rate": 0.00030047193190388963, "loss": 3.3968, "step": 46450 }, { "epoch": 5.004843396835647, "grad_norm": 0.6922981142997742, "learning_rate": 0.0003001486908738282, "loss": 3.3085, "step": 46500 }, { "epoch": 5.010224948875256, "grad_norm": 0.75609290599823, "learning_rate": 0.0002998254498437668, "loss": 3.2916, "step": 46550 }, { "epoch": 5.015606500914864, "grad_norm": 0.7146428823471069, "learning_rate": 0.0002995022088137054, "loss": 3.3115, "step": 46600 }, { "epoch": 5.020988052954472, "grad_norm": 0.7553977966308594, "learning_rate": 0.00029917896778364396, "loss": 3.3028, "step": 46650 }, { "epoch": 5.0263696049940805, "grad_norm": 0.6598968505859375, "learning_rate": 0.00029885572675358255, "loss": 3.2995, "step": 46700 }, { "epoch": 5.031751157033688, "grad_norm": 0.6931416988372803, "learning_rate": 0.00029853248572352114, "loss": 3.2986, "step": 46750 }, { "epoch": 5.037132709073297, "grad_norm": 0.6862365007400513, "learning_rate": 0.00029820924469345974, "loss": 3.3076, "step": 46800 }, { "epoch": 5.042514261112905, "grad_norm": 0.674670398235321, "learning_rate": 0.0002978860036633983, "loss": 3.282, "step": 46850 }, { "epoch": 5.047895813152513, "grad_norm": 0.6725071668624878, "learning_rate": 0.00029756276263333693, "loss": 3.3175, "step": 46900 }, { "epoch": 5.0532773651921215, "grad_norm": 0.715003252029419, "learning_rate": 0.00029723952160327547, "loss": 3.3009, "step": 46950 }, { "epoch": 5.05865891723173, "grad_norm": 0.6437234878540039, "learning_rate": 0.00029691628057321406, "loss": 3.2957, "step": 47000 }, { "epoch": 5.05865891723173, "eval_accuracy": 0.3807873009981615, "eval_loss": 3.412716865539551, "eval_runtime": 184.8748, "eval_samples_per_second": 97.423, "eval_steps_per_second": 6.091, "step": 47000 }, { "epoch": 5.064040469271338, "grad_norm": 0.6628379225730896, "learning_rate": 0.00029659303954315266, "loss": 3.3054, "step": 47050 }, { "epoch": 5.069422021310946, "grad_norm": 0.6656396985054016, "learning_rate": 0.00029626979851309125, "loss": 3.3118, "step": 47100 }, { "epoch": 5.074803573350554, "grad_norm": 0.6756969094276428, "learning_rate": 0.00029594655748302985, "loss": 3.2969, "step": 47150 }, { "epoch": 5.080185125390162, "grad_norm": 0.6976021528244019, "learning_rate": 0.0002956233164529684, "loss": 3.2881, "step": 47200 }, { "epoch": 5.085566677429771, "grad_norm": 0.7710605263710022, "learning_rate": 0.000295300075422907, "loss": 3.3236, "step": 47250 }, { "epoch": 5.090948229469379, "grad_norm": 0.6808165907859802, "learning_rate": 0.0002949768343928456, "loss": 3.3053, "step": 47300 }, { "epoch": 5.096329781508987, "grad_norm": 0.6778217554092407, "learning_rate": 0.00029465359336278417, "loss": 3.3169, "step": 47350 }, { "epoch": 5.101711333548596, "grad_norm": 0.7051342725753784, "learning_rate": 0.0002943303523327227, "loss": 3.3148, "step": 47400 }, { "epoch": 5.107092885588203, "grad_norm": 0.6564488410949707, "learning_rate": 0.00029400711130266136, "loss": 3.2898, "step": 47450 }, { "epoch": 5.112474437627812, "grad_norm": 0.6671553254127502, "learning_rate": 0.0002936838702725999, "loss": 3.3092, "step": 47500 }, { "epoch": 5.1178559896674205, "grad_norm": 0.7125695943832397, "learning_rate": 0.0002933606292425385, "loss": 3.3329, "step": 47550 }, { "epoch": 5.123237541707028, "grad_norm": 0.7285773754119873, "learning_rate": 0.0002930373882124771, "loss": 3.3005, "step": 47600 }, { "epoch": 5.128619093746637, "grad_norm": 0.7182151675224304, "learning_rate": 0.0002927141471824157, "loss": 3.3026, "step": 47650 }, { "epoch": 5.134000645786244, "grad_norm": 0.6936147212982178, "learning_rate": 0.0002923909061523542, "loss": 3.3223, "step": 47700 }, { "epoch": 5.139382197825853, "grad_norm": 0.7076971530914307, "learning_rate": 0.0002920676651222928, "loss": 3.3132, "step": 47750 }, { "epoch": 5.1447637498654615, "grad_norm": 0.7523118257522583, "learning_rate": 0.0002917444240922314, "loss": 3.331, "step": 47800 }, { "epoch": 5.150145301905069, "grad_norm": 0.6461667418479919, "learning_rate": 0.00029142118306216996, "loss": 3.3323, "step": 47850 }, { "epoch": 5.155526853944678, "grad_norm": 0.6895806193351746, "learning_rate": 0.0002910979420321086, "loss": 3.3141, "step": 47900 }, { "epoch": 5.160908405984286, "grad_norm": 0.715032696723938, "learning_rate": 0.00029077470100204715, "loss": 3.3178, "step": 47950 }, { "epoch": 5.166289958023894, "grad_norm": 0.7430562376976013, "learning_rate": 0.00029045145997198574, "loss": 3.3146, "step": 48000 }, { "epoch": 5.166289958023894, "eval_accuracy": 0.38082771987242847, "eval_loss": 3.4105722904205322, "eval_runtime": 184.8541, "eval_samples_per_second": 97.434, "eval_steps_per_second": 6.091, "step": 48000 }, { "epoch": 5.1716715100635025, "grad_norm": 0.6621975302696228, "learning_rate": 0.00029012821894192433, "loss": 3.3294, "step": 48050 }, { "epoch": 5.17705306210311, "grad_norm": 0.6793010830879211, "learning_rate": 0.00028980497791186293, "loss": 3.3091, "step": 48100 }, { "epoch": 5.182434614142719, "grad_norm": 0.7010188698768616, "learning_rate": 0.00028948820170240274, "loss": 3.3182, "step": 48150 }, { "epoch": 5.187816166182327, "grad_norm": 0.6465266942977905, "learning_rate": 0.00028916496067234133, "loss": 3.3094, "step": 48200 }, { "epoch": 5.193197718221935, "grad_norm": 0.7279083132743835, "learning_rate": 0.0002888417196422799, "loss": 3.3092, "step": 48250 }, { "epoch": 5.198579270261543, "grad_norm": 0.6986241936683655, "learning_rate": 0.00028851847861221847, "loss": 3.2993, "step": 48300 }, { "epoch": 5.203960822301152, "grad_norm": 0.6948270797729492, "learning_rate": 0.00028819523758215706, "loss": 3.3012, "step": 48350 }, { "epoch": 5.20934237434076, "grad_norm": 0.6764682531356812, "learning_rate": 0.00028787199655209566, "loss": 3.3296, "step": 48400 }, { "epoch": 5.214723926380368, "grad_norm": 0.7321643233299255, "learning_rate": 0.00028754875552203425, "loss": 3.3373, "step": 48450 }, { "epoch": 5.220105478419977, "grad_norm": 0.666074275970459, "learning_rate": 0.0002872255144919728, "loss": 3.315, "step": 48500 }, { "epoch": 5.225487030459584, "grad_norm": 0.7783872485160828, "learning_rate": 0.00028690227346191144, "loss": 3.3212, "step": 48550 }, { "epoch": 5.230868582499193, "grad_norm": 0.726060152053833, "learning_rate": 0.00028657903243185, "loss": 3.3096, "step": 48600 }, { "epoch": 5.236250134538801, "grad_norm": 0.6934331655502319, "learning_rate": 0.0002862557914017886, "loss": 3.3258, "step": 48650 }, { "epoch": 5.241631686578409, "grad_norm": 0.684074878692627, "learning_rate": 0.00028593255037172717, "loss": 3.3216, "step": 48700 }, { "epoch": 5.247013238618018, "grad_norm": 0.6767908334732056, "learning_rate": 0.00028560930934166576, "loss": 3.3207, "step": 48750 }, { "epoch": 5.252394790657625, "grad_norm": 0.691230058670044, "learning_rate": 0.00028528606831160436, "loss": 3.3434, "step": 48800 }, { "epoch": 5.257776342697234, "grad_norm": 0.7426613569259644, "learning_rate": 0.0002849628272815429, "loss": 3.3263, "step": 48850 }, { "epoch": 5.2631578947368425, "grad_norm": 0.6871113777160645, "learning_rate": 0.0002846395862514815, "loss": 3.3332, "step": 48900 }, { "epoch": 5.26853944677645, "grad_norm": 0.6987447142601013, "learning_rate": 0.0002843228100420213, "loss": 3.3349, "step": 48950 }, { "epoch": 5.273920998816059, "grad_norm": 0.6851251125335693, "learning_rate": 0.0002839995690119599, "loss": 3.3387, "step": 49000 }, { "epoch": 5.273920998816059, "eval_accuracy": 0.3813237203053016, "eval_loss": 3.4056828022003174, "eval_runtime": 184.6477, "eval_samples_per_second": 97.543, "eval_steps_per_second": 6.098, "step": 49000 }, { "epoch": 5.279302550855666, "grad_norm": 0.6929810047149658, "learning_rate": 0.0002836763279818985, "loss": 3.3249, "step": 49050 }, { "epoch": 5.284684102895275, "grad_norm": 0.7261629104614258, "learning_rate": 0.0002833530869518371, "loss": 3.3331, "step": 49100 }, { "epoch": 5.2900656549348835, "grad_norm": 0.6819291114807129, "learning_rate": 0.0002830298459217756, "loss": 3.326, "step": 49150 }, { "epoch": 5.295447206974491, "grad_norm": 0.7193529009819031, "learning_rate": 0.0002827066048917142, "loss": 3.3126, "step": 49200 }, { "epoch": 5.3008287590141, "grad_norm": 0.7201562523841858, "learning_rate": 0.0002823833638616528, "loss": 3.3113, "step": 49250 }, { "epoch": 5.306210311053708, "grad_norm": 0.6952700614929199, "learning_rate": 0.0002820601228315914, "loss": 3.3216, "step": 49300 }, { "epoch": 5.311591863093316, "grad_norm": 0.6874309778213501, "learning_rate": 0.00028173688180153, "loss": 3.3205, "step": 49350 }, { "epoch": 5.316973415132924, "grad_norm": 0.6886963844299316, "learning_rate": 0.00028141364077146854, "loss": 3.3213, "step": 49400 }, { "epoch": 5.322354967172533, "grad_norm": 0.7013046741485596, "learning_rate": 0.00028109039974140714, "loss": 3.344, "step": 49450 }, { "epoch": 5.327736519212141, "grad_norm": 0.73094242811203, "learning_rate": 0.00028076715871134573, "loss": 3.3239, "step": 49500 }, { "epoch": 5.333118071251749, "grad_norm": 0.7266910076141357, "learning_rate": 0.00028044391768128433, "loss": 3.3593, "step": 49550 }, { "epoch": 5.338499623291357, "grad_norm": 0.6710968017578125, "learning_rate": 0.0002801206766512229, "loss": 3.3211, "step": 49600 }, { "epoch": 5.343881175330965, "grad_norm": 0.6896623373031616, "learning_rate": 0.0002797974356211615, "loss": 3.3291, "step": 49650 }, { "epoch": 5.349262727370574, "grad_norm": 0.7244205474853516, "learning_rate": 0.00027947419459110006, "loss": 3.3124, "step": 49700 }, { "epoch": 5.354644279410182, "grad_norm": 0.7222291827201843, "learning_rate": 0.00027915095356103865, "loss": 3.3257, "step": 49750 }, { "epoch": 5.36002583144979, "grad_norm": 0.6412253379821777, "learning_rate": 0.00027882771253097725, "loss": 3.3272, "step": 49800 }, { "epoch": 5.365407383489399, "grad_norm": 0.699478268623352, "learning_rate": 0.00027850447150091584, "loss": 3.335, "step": 49850 }, { "epoch": 5.370788935529006, "grad_norm": 0.6737022399902344, "learning_rate": 0.00027818123047085444, "loss": 3.3194, "step": 49900 }, { "epoch": 5.376170487568615, "grad_norm": 0.6805015802383423, "learning_rate": 0.000277857989440793, "loss": 3.3292, "step": 49950 }, { "epoch": 5.3815520396082235, "grad_norm": 0.7129510641098022, "learning_rate": 0.00027753474841073157, "loss": 3.3388, "step": 50000 }, { "epoch": 5.3815520396082235, "eval_accuracy": 0.3819017536684746, "eval_loss": 3.400902032852173, "eval_runtime": 184.6882, "eval_samples_per_second": 97.521, "eval_steps_per_second": 6.097, "step": 50000 }, { "epoch": 5.386933591647831, "grad_norm": 0.7065452933311462, "learning_rate": 0.00027721150738067017, "loss": 3.3545, "step": 50050 }, { "epoch": 5.39231514368744, "grad_norm": 0.714790403842926, "learning_rate": 0.00027688826635060876, "loss": 3.3329, "step": 50100 }, { "epoch": 5.397696695727047, "grad_norm": 0.7263872027397156, "learning_rate": 0.0002765650253205473, "loss": 3.3431, "step": 50150 }, { "epoch": 5.403078247766656, "grad_norm": 0.6966455578804016, "learning_rate": 0.00027624178429048595, "loss": 3.3286, "step": 50200 }, { "epoch": 5.4084597998062645, "grad_norm": 0.716153621673584, "learning_rate": 0.0002759185432604245, "loss": 3.3326, "step": 50250 }, { "epoch": 5.413841351845872, "grad_norm": 0.7463889718055725, "learning_rate": 0.0002755953022303631, "loss": 3.3212, "step": 50300 }, { "epoch": 5.419222903885481, "grad_norm": 0.6929644346237183, "learning_rate": 0.0002752720612003017, "loss": 3.3268, "step": 50350 }, { "epoch": 5.424604455925088, "grad_norm": 0.7609820365905762, "learning_rate": 0.0002749488201702403, "loss": 3.3446, "step": 50400 }, { "epoch": 5.429986007964697, "grad_norm": 0.721838116645813, "learning_rate": 0.0002746255791401788, "loss": 3.3271, "step": 50450 }, { "epoch": 5.435367560004305, "grad_norm": 0.7337953448295593, "learning_rate": 0.0002743023381101174, "loss": 3.3351, "step": 50500 }, { "epoch": 5.440749112043913, "grad_norm": 0.7078254818916321, "learning_rate": 0.000273979097080056, "loss": 3.326, "step": 50550 }, { "epoch": 5.446130664083522, "grad_norm": 0.6884511709213257, "learning_rate": 0.0002736558560499946, "loss": 3.3364, "step": 50600 }, { "epoch": 5.45151221612313, "grad_norm": 0.6717587113380432, "learning_rate": 0.0002733326150199332, "loss": 3.3209, "step": 50650 }, { "epoch": 5.456893768162738, "grad_norm": 0.7340549826622009, "learning_rate": 0.00027300937398987173, "loss": 3.3344, "step": 50700 }, { "epoch": 5.462275320202346, "grad_norm": 0.7046282291412354, "learning_rate": 0.0002726861329598104, "loss": 3.3286, "step": 50750 }, { "epoch": 5.467656872241955, "grad_norm": 0.7176598906517029, "learning_rate": 0.0002723628919297489, "loss": 3.335, "step": 50800 }, { "epoch": 5.473038424281563, "grad_norm": 0.7335805296897888, "learning_rate": 0.0002720396508996875, "loss": 3.3395, "step": 50850 }, { "epoch": 5.478419976321171, "grad_norm": 0.7153794169425964, "learning_rate": 0.0002717164098696261, "loss": 3.3335, "step": 50900 }, { "epoch": 5.483801528360779, "grad_norm": 0.7129613757133484, "learning_rate": 0.00027139316883956465, "loss": 3.34, "step": 50950 }, { "epoch": 5.489183080400387, "grad_norm": 0.701945424079895, "learning_rate": 0.00027106992780950325, "loss": 3.3306, "step": 51000 }, { "epoch": 5.489183080400387, "eval_accuracy": 0.3824516459337037, "eval_loss": 3.3981103897094727, "eval_runtime": 184.7536, "eval_samples_per_second": 97.487, "eval_steps_per_second": 6.095, "step": 51000 }, { "epoch": 5.494564632439996, "grad_norm": 0.6667730212211609, "learning_rate": 0.00027074668677944184, "loss": 3.3239, "step": 51050 }, { "epoch": 5.499946184479604, "grad_norm": 0.6833287477493286, "learning_rate": 0.00027042344574938044, "loss": 3.33, "step": 51100 }, { "epoch": 5.505327736519212, "grad_norm": 0.7144169211387634, "learning_rate": 0.000270100204719319, "loss": 3.3382, "step": 51150 }, { "epoch": 5.510709288558821, "grad_norm": 0.7473929524421692, "learning_rate": 0.0002697769636892576, "loss": 3.3129, "step": 51200 }, { "epoch": 5.516090840598428, "grad_norm": 0.7221124172210693, "learning_rate": 0.00026945372265919617, "loss": 3.3577, "step": 51250 }, { "epoch": 5.521472392638037, "grad_norm": 0.7008888721466064, "learning_rate": 0.00026913048162913476, "loss": 3.315, "step": 51300 }, { "epoch": 5.5268539446776455, "grad_norm": 0.7081114649772644, "learning_rate": 0.00026880724059907336, "loss": 3.3402, "step": 51350 }, { "epoch": 5.532235496717253, "grad_norm": 0.7221799492835999, "learning_rate": 0.00026848399956901195, "loss": 3.3182, "step": 51400 }, { "epoch": 5.537617048756862, "grad_norm": 0.6946343779563904, "learning_rate": 0.0002681607585389505, "loss": 3.3361, "step": 51450 }, { "epoch": 5.542998600796469, "grad_norm": 0.702681303024292, "learning_rate": 0.0002678375175088891, "loss": 3.3399, "step": 51500 }, { "epoch": 5.548380152836078, "grad_norm": 0.7525026798248291, "learning_rate": 0.0002675142764788277, "loss": 3.3242, "step": 51550 }, { "epoch": 5.553761704875686, "grad_norm": 0.772919774055481, "learning_rate": 0.0002671910354487663, "loss": 3.322, "step": 51600 }, { "epoch": 5.559143256915294, "grad_norm": 0.6999807953834534, "learning_rate": 0.00026686779441870487, "loss": 3.3294, "step": 51650 }, { "epoch": 5.564524808954903, "grad_norm": 0.7455286979675293, "learning_rate": 0.0002665445533886434, "loss": 3.3344, "step": 51700 }, { "epoch": 5.569906360994511, "grad_norm": 0.7175178527832031, "learning_rate": 0.00026622131235858206, "loss": 3.3322, "step": 51750 }, { "epoch": 5.575287913034119, "grad_norm": 0.7250377535820007, "learning_rate": 0.0002658980713285206, "loss": 3.3416, "step": 51800 }, { "epoch": 5.580669465073727, "grad_norm": 0.7095009088516235, "learning_rate": 0.0002655748302984592, "loss": 3.3336, "step": 51850 }, { "epoch": 5.586051017113336, "grad_norm": 0.757713258266449, "learning_rate": 0.0002652515892683978, "loss": 3.3384, "step": 51900 }, { "epoch": 5.591432569152944, "grad_norm": 0.7264611124992371, "learning_rate": 0.0002649283482383364, "loss": 3.3279, "step": 51950 }, { "epoch": 5.596814121192552, "grad_norm": 0.7121990323066711, "learning_rate": 0.0002646051072082749, "loss": 3.3177, "step": 52000 }, { "epoch": 5.596814121192552, "eval_accuracy": 0.3827728238700833, "eval_loss": 3.3929855823516846, "eval_runtime": 184.7294, "eval_samples_per_second": 97.499, "eval_steps_per_second": 6.095, "step": 52000 }, { "epoch": 5.60219567323216, "grad_norm": 0.7471087574958801, "learning_rate": 0.0002642818661782135, "loss": 3.3679, "step": 52050 }, { "epoch": 5.607577225271768, "grad_norm": 0.7511977553367615, "learning_rate": 0.0002639586251481521, "loss": 3.3404, "step": 52100 }, { "epoch": 5.612958777311377, "grad_norm": 0.6835429072380066, "learning_rate": 0.0002636353841180907, "loss": 3.3509, "step": 52150 }, { "epoch": 5.618340329350985, "grad_norm": 0.7123720645904541, "learning_rate": 0.0002633121430880293, "loss": 3.3175, "step": 52200 }, { "epoch": 5.623721881390593, "grad_norm": 0.7043327689170837, "learning_rate": 0.00026298890205796784, "loss": 3.3321, "step": 52250 }, { "epoch": 5.629103433430201, "grad_norm": 0.7442822456359863, "learning_rate": 0.00026266566102790644, "loss": 3.3466, "step": 52300 }, { "epoch": 5.634484985469809, "grad_norm": 0.6664136648178101, "learning_rate": 0.00026234241999784503, "loss": 3.3312, "step": 52350 }, { "epoch": 5.639866537509418, "grad_norm": 0.7649520039558411, "learning_rate": 0.0002620191789677836, "loss": 3.3195, "step": 52400 }, { "epoch": 5.645248089549026, "grad_norm": 0.717771053314209, "learning_rate": 0.00026169593793772217, "loss": 3.3188, "step": 52450 }, { "epoch": 5.650629641588634, "grad_norm": 0.7109053134918213, "learning_rate": 0.0002613726969076608, "loss": 3.3444, "step": 52500 }, { "epoch": 5.656011193628243, "grad_norm": 0.6929636001586914, "learning_rate": 0.00026104945587759936, "loss": 3.3419, "step": 52550 }, { "epoch": 5.66139274566785, "grad_norm": 0.7861093282699585, "learning_rate": 0.00026072621484753795, "loss": 3.3373, "step": 52600 }, { "epoch": 5.666774297707459, "grad_norm": 0.7142403721809387, "learning_rate": 0.00026040297381747655, "loss": 3.3326, "step": 52650 }, { "epoch": 5.672155849747067, "grad_norm": 0.7610068321228027, "learning_rate": 0.00026007973278741514, "loss": 3.3319, "step": 52700 }, { "epoch": 5.677537401786675, "grad_norm": 0.7404228448867798, "learning_rate": 0.00025975649175735373, "loss": 3.3378, "step": 52750 }, { "epoch": 5.682918953826284, "grad_norm": 0.6821990013122559, "learning_rate": 0.0002594332507272923, "loss": 3.3357, "step": 52800 }, { "epoch": 5.688300505865891, "grad_norm": 0.7214856743812561, "learning_rate": 0.00025911000969723087, "loss": 3.3244, "step": 52850 }, { "epoch": 5.6936820579055, "grad_norm": 0.7556411623954773, "learning_rate": 0.00025878676866716946, "loss": 3.3249, "step": 52900 }, { "epoch": 5.699063609945108, "grad_norm": 0.7074944376945496, "learning_rate": 0.00025846352763710806, "loss": 3.3231, "step": 52950 }, { "epoch": 5.704445161984716, "grad_norm": 0.7696301341056824, "learning_rate": 0.00025814675142764787, "loss": 3.3337, "step": 53000 }, { "epoch": 5.704445161984716, "eval_accuracy": 0.38316332234888856, "eval_loss": 3.3885035514831543, "eval_runtime": 184.6186, "eval_samples_per_second": 97.558, "eval_steps_per_second": 6.099, "step": 53000 }, { "epoch": 5.709826714024325, "grad_norm": 0.7107107043266296, "learning_rate": 0.00025782351039758646, "loss": 3.3093, "step": 53050 }, { "epoch": 5.715208266063933, "grad_norm": 0.7684431672096252, "learning_rate": 0.000257500269367525, "loss": 3.3285, "step": 53100 }, { "epoch": 5.720589818103541, "grad_norm": 0.7184401154518127, "learning_rate": 0.0002571770283374636, "loss": 3.3186, "step": 53150 }, { "epoch": 5.725971370143149, "grad_norm": 0.6955456733703613, "learning_rate": 0.0002568537873074022, "loss": 3.3399, "step": 53200 }, { "epoch": 5.731352922182758, "grad_norm": 0.7292144298553467, "learning_rate": 0.0002565305462773408, "loss": 3.3194, "step": 53250 }, { "epoch": 5.736734474222366, "grad_norm": 0.7374510169029236, "learning_rate": 0.0002562073052472794, "loss": 3.3199, "step": 53300 }, { "epoch": 5.742116026261974, "grad_norm": 0.6997143030166626, "learning_rate": 0.0002558840642172179, "loss": 3.3359, "step": 53350 }, { "epoch": 5.747497578301582, "grad_norm": 0.755856454372406, "learning_rate": 0.00025556082318715657, "loss": 3.3198, "step": 53400 }, { "epoch": 5.75287913034119, "grad_norm": 0.7310744524002075, "learning_rate": 0.0002552375821570951, "loss": 3.3347, "step": 53450 }, { "epoch": 5.758260682380799, "grad_norm": 0.767074704170227, "learning_rate": 0.0002549143411270337, "loss": 3.3416, "step": 53500 }, { "epoch": 5.763642234420407, "grad_norm": 0.7051040530204773, "learning_rate": 0.0002545911000969723, "loss": 3.3338, "step": 53550 }, { "epoch": 5.769023786460015, "grad_norm": 0.7256135940551758, "learning_rate": 0.0002542678590669109, "loss": 3.3486, "step": 53600 }, { "epoch": 5.774405338499624, "grad_norm": 0.7255988121032715, "learning_rate": 0.00025394461803684943, "loss": 3.3406, "step": 53650 }, { "epoch": 5.779786890539231, "grad_norm": 0.7210702300071716, "learning_rate": 0.00025362137700678803, "loss": 3.3282, "step": 53700 }, { "epoch": 5.78516844257884, "grad_norm": 0.7039930820465088, "learning_rate": 0.0002532981359767266, "loss": 3.3353, "step": 53750 }, { "epoch": 5.790549994618448, "grad_norm": 0.7262842059135437, "learning_rate": 0.0002529748949466652, "loss": 3.3549, "step": 53800 }, { "epoch": 5.795931546658056, "grad_norm": 0.718593180179596, "learning_rate": 0.0002526516539166038, "loss": 3.3341, "step": 53850 }, { "epoch": 5.801313098697665, "grad_norm": 0.7095365524291992, "learning_rate": 0.00025232841288654235, "loss": 3.3327, "step": 53900 }, { "epoch": 5.806694650737272, "grad_norm": 0.7695627212524414, "learning_rate": 0.00025200517185648095, "loss": 3.3315, "step": 53950 }, { "epoch": 5.812076202776881, "grad_norm": 0.699380099773407, "learning_rate": 0.00025168193082641954, "loss": 3.344, "step": 54000 }, { "epoch": 5.812076202776881, "eval_accuracy": 0.38343777954350794, "eval_loss": 3.3852038383483887, "eval_runtime": 184.9031, "eval_samples_per_second": 97.408, "eval_steps_per_second": 6.09, "step": 54000 }, { "epoch": 5.817457754816489, "grad_norm": 0.7718612551689148, "learning_rate": 0.00025135868979635814, "loss": 3.3475, "step": 54050 }, { "epoch": 5.822839306856097, "grad_norm": 0.7421673536300659, "learning_rate": 0.00025104191358689794, "loss": 3.3234, "step": 54100 }, { "epoch": 5.828220858895706, "grad_norm": 0.728031575679779, "learning_rate": 0.00025071867255683654, "loss": 3.3411, "step": 54150 }, { "epoch": 5.833602410935313, "grad_norm": 0.7653467655181885, "learning_rate": 0.00025039543152677513, "loss": 3.3354, "step": 54200 }, { "epoch": 5.838983962974922, "grad_norm": 0.6954116821289062, "learning_rate": 0.0002500721904967137, "loss": 3.3503, "step": 54250 }, { "epoch": 5.84436551501453, "grad_norm": 0.7203173637390137, "learning_rate": 0.00024974894946665227, "loss": 3.3408, "step": 54300 }, { "epoch": 5.849747067054138, "grad_norm": 0.7083039879798889, "learning_rate": 0.00024942570843659086, "loss": 3.3285, "step": 54350 }, { "epoch": 5.855128619093747, "grad_norm": 0.8395054936408997, "learning_rate": 0.00024910246740652946, "loss": 3.3423, "step": 54400 }, { "epoch": 5.860510171133355, "grad_norm": 0.7595872282981873, "learning_rate": 0.000248779226376468, "loss": 3.3395, "step": 54450 }, { "epoch": 5.865891723172963, "grad_norm": 0.7334976196289062, "learning_rate": 0.00024845598534640665, "loss": 3.3299, "step": 54500 }, { "epoch": 5.871273275212571, "grad_norm": 0.7502398490905762, "learning_rate": 0.0002481327443163452, "loss": 3.3213, "step": 54550 }, { "epoch": 5.87665482725218, "grad_norm": 0.8335617184638977, "learning_rate": 0.0002478095032862838, "loss": 3.3317, "step": 54600 }, { "epoch": 5.882036379291788, "grad_norm": 0.7220059037208557, "learning_rate": 0.0002474862622562224, "loss": 3.3367, "step": 54650 }, { "epoch": 5.887417931331396, "grad_norm": 0.7250782251358032, "learning_rate": 0.00024716302122616097, "loss": 3.3122, "step": 54700 }, { "epoch": 5.892799483371004, "grad_norm": 0.7305442690849304, "learning_rate": 0.0002468397801960995, "loss": 3.3405, "step": 54750 }, { "epoch": 5.898181035410612, "grad_norm": 0.6805377006530762, "learning_rate": 0.0002465165391660381, "loss": 3.3371, "step": 54800 }, { "epoch": 5.903562587450221, "grad_norm": 0.7381418347358704, "learning_rate": 0.0002461932981359767, "loss": 3.3307, "step": 54850 }, { "epoch": 5.9089441394898286, "grad_norm": 0.7628136873245239, "learning_rate": 0.0002458700571059153, "loss": 3.3325, "step": 54900 }, { "epoch": 5.914325691529437, "grad_norm": 0.6967856884002686, "learning_rate": 0.0002455468160758539, "loss": 3.3264, "step": 54950 }, { "epoch": 5.919707243569046, "grad_norm": 0.7036657333374023, "learning_rate": 0.00024522357504579243, "loss": 3.3345, "step": 55000 }, { "epoch": 5.919707243569046, "eval_accuracy": 0.3839816958999072, "eval_loss": 3.3790087699890137, "eval_runtime": 184.9878, "eval_samples_per_second": 97.363, "eval_steps_per_second": 6.087, "step": 55000 }, { "epoch": 5.925088795608653, "grad_norm": 0.7337253093719482, "learning_rate": 0.000244900334015731, "loss": 3.333, "step": 55050 }, { "epoch": 5.930470347648262, "grad_norm": 0.7193291783332825, "learning_rate": 0.0002445770929856696, "loss": 3.3383, "step": 55100 }, { "epoch": 5.93585189968787, "grad_norm": 0.6923545598983765, "learning_rate": 0.0002442538519556082, "loss": 3.3076, "step": 55150 }, { "epoch": 5.941233451727478, "grad_norm": 0.7571631073951721, "learning_rate": 0.0002439306109255468, "loss": 3.3281, "step": 55200 }, { "epoch": 5.946615003767087, "grad_norm": 0.7402598857879639, "learning_rate": 0.00024360736989548538, "loss": 3.333, "step": 55250 }, { "epoch": 5.951996555806694, "grad_norm": 0.7690083384513855, "learning_rate": 0.00024328412886542394, "loss": 3.3345, "step": 55300 }, { "epoch": 5.957378107846303, "grad_norm": 0.6994602084159851, "learning_rate": 0.00024296088783536257, "loss": 3.3188, "step": 55350 }, { "epoch": 5.962759659885911, "grad_norm": 0.7294467091560364, "learning_rate": 0.00024263764680530113, "loss": 3.3436, "step": 55400 }, { "epoch": 5.968141211925519, "grad_norm": 0.7167765498161316, "learning_rate": 0.0002423144057752397, "loss": 3.3152, "step": 55450 }, { "epoch": 5.973522763965128, "grad_norm": 0.733696699142456, "learning_rate": 0.00024199116474517832, "loss": 3.3195, "step": 55500 }, { "epoch": 5.978904316004736, "grad_norm": 0.7189485430717468, "learning_rate": 0.0002416679237151169, "loss": 3.3247, "step": 55550 }, { "epoch": 5.984285868044344, "grad_norm": 0.7302901148796082, "learning_rate": 0.00024134468268505546, "loss": 3.3269, "step": 55600 }, { "epoch": 5.989667420083952, "grad_norm": 0.7239100337028503, "learning_rate": 0.00024102144165499405, "loss": 3.3427, "step": 55650 }, { "epoch": 5.995048972123561, "grad_norm": 0.7289806604385376, "learning_rate": 0.00024069820062493265, "loss": 3.3162, "step": 55700 }, { "epoch": 6.000430524163169, "grad_norm": 0.7002869248390198, "learning_rate": 0.00024037495959487121, "loss": 3.3247, "step": 55750 }, { "epoch": 6.005812076202777, "grad_norm": 0.7519763708114624, "learning_rate": 0.0002400517185648098, "loss": 3.2449, "step": 55800 }, { "epoch": 6.011193628242385, "grad_norm": 0.7565599679946899, "learning_rate": 0.00023972847753474838, "loss": 3.2444, "step": 55850 }, { "epoch": 6.016575180281993, "grad_norm": 0.7228876352310181, "learning_rate": 0.00023940523650468697, "loss": 3.2454, "step": 55900 }, { "epoch": 6.021956732321602, "grad_norm": 0.7430625557899475, "learning_rate": 0.00023908199547462557, "loss": 3.26, "step": 55950 }, { "epoch": 6.0273382843612096, "grad_norm": 0.7767698764801025, "learning_rate": 0.00023875875444456413, "loss": 3.2586, "step": 56000 }, { "epoch": 6.0273382843612096, "eval_accuracy": 0.3844594426476275, "eval_loss": 3.383514642715454, "eval_runtime": 184.7551, "eval_samples_per_second": 97.486, "eval_steps_per_second": 6.095, "step": 56000 }, { "epoch": 6.032719836400818, "grad_norm": 0.7343913316726685, "learning_rate": 0.0002384355134145027, "loss": 3.2529, "step": 56050 }, { "epoch": 6.038101388440427, "grad_norm": 0.745909571647644, "learning_rate": 0.00023811227238444132, "loss": 3.242, "step": 56100 }, { "epoch": 6.043482940480034, "grad_norm": 0.7055720090866089, "learning_rate": 0.0002377890313543799, "loss": 3.2403, "step": 56150 }, { "epoch": 6.048864492519643, "grad_norm": 0.7497826218605042, "learning_rate": 0.00023746579032431849, "loss": 3.243, "step": 56200 }, { "epoch": 6.0542460445592505, "grad_norm": 0.7048562169075012, "learning_rate": 0.00023714254929425708, "loss": 3.242, "step": 56250 }, { "epoch": 6.059627596598859, "grad_norm": 0.7660420536994934, "learning_rate": 0.00023681930826419565, "loss": 3.2648, "step": 56300 }, { "epoch": 6.065009148638468, "grad_norm": 0.7854742407798767, "learning_rate": 0.00023649606723413424, "loss": 3.2551, "step": 56350 }, { "epoch": 6.070390700678075, "grad_norm": 0.7162970900535583, "learning_rate": 0.0002361728262040728, "loss": 3.2563, "step": 56400 }, { "epoch": 6.075772252717684, "grad_norm": 0.7376888990402222, "learning_rate": 0.00023584958517401138, "loss": 3.2507, "step": 56450 }, { "epoch": 6.081153804757292, "grad_norm": 0.7727888226509094, "learning_rate": 0.00023552634414395, "loss": 3.2685, "step": 56500 }, { "epoch": 6.0865353567969, "grad_norm": 0.7515780925750732, "learning_rate": 0.00023520310311388857, "loss": 3.2594, "step": 56550 }, { "epoch": 6.091916908836509, "grad_norm": 0.721759557723999, "learning_rate": 0.00023487986208382713, "loss": 3.2608, "step": 56600 }, { "epoch": 6.097298460876116, "grad_norm": 0.6812422871589661, "learning_rate": 0.00023455662105376576, "loss": 3.2503, "step": 56650 }, { "epoch": 6.102680012915725, "grad_norm": 0.7772680521011353, "learning_rate": 0.00023423338002370432, "loss": 3.2606, "step": 56700 }, { "epoch": 6.108061564955333, "grad_norm": 0.7657935619354248, "learning_rate": 0.0002339101389936429, "loss": 3.2543, "step": 56750 }, { "epoch": 6.113443116994941, "grad_norm": 0.7089388370513916, "learning_rate": 0.00023358689796358149, "loss": 3.2707, "step": 56800 }, { "epoch": 6.11882466903455, "grad_norm": 0.7366516590118408, "learning_rate": 0.00023326365693352008, "loss": 3.2698, "step": 56850 }, { "epoch": 6.124206221074158, "grad_norm": 0.7500792145729065, "learning_rate": 0.00023294041590345865, "loss": 3.2517, "step": 56900 }, { "epoch": 6.129587773113766, "grad_norm": 0.7379442453384399, "learning_rate": 0.00023261717487339724, "loss": 3.2377, "step": 56950 }, { "epoch": 6.134969325153374, "grad_norm": 0.7361008524894714, "learning_rate": 0.0002322939338433358, "loss": 3.2657, "step": 57000 }, { "epoch": 6.134969325153374, "eval_accuracy": 0.3848917724882413, "eval_loss": 3.3805084228515625, "eval_runtime": 184.8323, "eval_samples_per_second": 97.445, "eval_steps_per_second": 6.092, "step": 57000 }, { "epoch": 6.140350877192983, "grad_norm": 0.7512038350105286, "learning_rate": 0.0002319706928132744, "loss": 3.2544, "step": 57050 }, { "epoch": 6.1457324292325906, "grad_norm": 0.7466521859169006, "learning_rate": 0.000231647451783213, "loss": 3.2559, "step": 57100 }, { "epoch": 6.151113981272199, "grad_norm": 0.7655845880508423, "learning_rate": 0.00023132421075315157, "loss": 3.2716, "step": 57150 }, { "epoch": 6.156495533311807, "grad_norm": 0.7715071439743042, "learning_rate": 0.0002310009697230902, "loss": 3.2597, "step": 57200 }, { "epoch": 6.161877085351415, "grad_norm": 0.7462551593780518, "learning_rate": 0.00023067772869302876, "loss": 3.2619, "step": 57250 }, { "epoch": 6.167258637391024, "grad_norm": 0.6912649273872375, "learning_rate": 0.00023035448766296732, "loss": 3.254, "step": 57300 }, { "epoch": 6.1726401894306315, "grad_norm": 0.7506750822067261, "learning_rate": 0.00023003124663290592, "loss": 3.2706, "step": 57350 }, { "epoch": 6.17802174147024, "grad_norm": 0.7650833129882812, "learning_rate": 0.0002297080056028445, "loss": 3.2581, "step": 57400 }, { "epoch": 6.183403293509849, "grad_norm": 0.7462422251701355, "learning_rate": 0.00022938476457278308, "loss": 3.2885, "step": 57450 }, { "epoch": 6.188784845549456, "grad_norm": 0.7575925588607788, "learning_rate": 0.00022906152354272168, "loss": 3.2697, "step": 57500 }, { "epoch": 6.194166397589065, "grad_norm": 0.7644687294960022, "learning_rate": 0.00022873828251266024, "loss": 3.2552, "step": 57550 }, { "epoch": 6.1995479496286725, "grad_norm": 0.7716279625892639, "learning_rate": 0.00022841504148259884, "loss": 3.2557, "step": 57600 }, { "epoch": 6.204929501668281, "grad_norm": 0.7543048858642578, "learning_rate": 0.00022809180045253743, "loss": 3.2624, "step": 57650 }, { "epoch": 6.21031105370789, "grad_norm": 0.7387022376060486, "learning_rate": 0.000227768559422476, "loss": 3.2646, "step": 57700 }, { "epoch": 6.215692605747497, "grad_norm": 0.8519375324249268, "learning_rate": 0.00022744531839241457, "loss": 3.2812, "step": 57750 }, { "epoch": 6.221074157787106, "grad_norm": 0.7521229982376099, "learning_rate": 0.0002271220773623532, "loss": 3.2558, "step": 57800 }, { "epoch": 6.226455709826714, "grad_norm": 0.7615373134613037, "learning_rate": 0.00022679883633229176, "loss": 3.2676, "step": 57850 }, { "epoch": 6.231837261866322, "grad_norm": 0.7787977457046509, "learning_rate": 0.00022647559530223032, "loss": 3.2631, "step": 57900 }, { "epoch": 6.237218813905931, "grad_norm": 0.7316073179244995, "learning_rate": 0.00022615235427216895, "loss": 3.2677, "step": 57950 }, { "epoch": 6.242600365945538, "grad_norm": 0.7252430319786072, "learning_rate": 0.0002258291132421075, "loss": 3.2792, "step": 58000 }, { "epoch": 6.242600365945538, "eval_accuracy": 0.38463469975567227, "eval_loss": 3.379333019256592, "eval_runtime": 184.8261, "eval_samples_per_second": 97.448, "eval_steps_per_second": 6.092, "step": 58000 }, { "epoch": 6.247981917985147, "grad_norm": 0.7599756121635437, "learning_rate": 0.0002255058722120461, "loss": 3.2763, "step": 58050 }, { "epoch": 6.253363470024755, "grad_norm": 0.717863142490387, "learning_rate": 0.0002251890960025859, "loss": 3.2631, "step": 58100 }, { "epoch": 6.258745022064363, "grad_norm": 0.7755172848701477, "learning_rate": 0.0002248658549725245, "loss": 3.2734, "step": 58150 }, { "epoch": 6.264126574103972, "grad_norm": 0.737862229347229, "learning_rate": 0.00022454261394246308, "loss": 3.2671, "step": 58200 }, { "epoch": 6.26950812614358, "grad_norm": 0.7289722561836243, "learning_rate": 0.00022421937291240164, "loss": 3.2787, "step": 58250 }, { "epoch": 6.274889678183188, "grad_norm": 0.737337052822113, "learning_rate": 0.00022389613188234027, "loss": 3.2799, "step": 58300 }, { "epoch": 6.280271230222796, "grad_norm": 0.7309353351593018, "learning_rate": 0.00022357289085227883, "loss": 3.2639, "step": 58350 }, { "epoch": 6.285652782262405, "grad_norm": 0.7572712302207947, "learning_rate": 0.0002232496498222174, "loss": 3.2858, "step": 58400 }, { "epoch": 6.2910343343020125, "grad_norm": 0.7290258407592773, "learning_rate": 0.000222926408792156, "loss": 3.2798, "step": 58450 }, { "epoch": 6.296415886341621, "grad_norm": 0.7298064827919006, "learning_rate": 0.0002226031677620946, "loss": 3.2858, "step": 58500 }, { "epoch": 6.301797438381229, "grad_norm": 0.7554293274879456, "learning_rate": 0.00022227992673203316, "loss": 3.2744, "step": 58550 }, { "epoch": 6.307178990420837, "grad_norm": 0.7588570713996887, "learning_rate": 0.00022195668570197175, "loss": 3.2767, "step": 58600 }, { "epoch": 6.312560542460446, "grad_norm": 0.7721754312515259, "learning_rate": 0.00022163344467191032, "loss": 3.26, "step": 58650 }, { "epoch": 6.3179420945000535, "grad_norm": 0.7513089179992676, "learning_rate": 0.00022131020364184891, "loss": 3.2749, "step": 58700 }, { "epoch": 6.323323646539662, "grad_norm": 0.8285422921180725, "learning_rate": 0.0002209869626117875, "loss": 3.281, "step": 58750 }, { "epoch": 6.328705198579271, "grad_norm": 0.7624416947364807, "learning_rate": 0.00022066372158172608, "loss": 3.2768, "step": 58800 }, { "epoch": 6.334086750618878, "grad_norm": 0.7431687712669373, "learning_rate": 0.0002203469453722659, "loss": 3.2844, "step": 58850 }, { "epoch": 6.339468302658487, "grad_norm": 0.7476500272750854, "learning_rate": 0.00022002370434220448, "loss": 3.2766, "step": 58900 }, { "epoch": 6.344849854698095, "grad_norm": 0.7640836238861084, "learning_rate": 0.00021970046331214307, "loss": 3.2819, "step": 58950 }, { "epoch": 6.350231406737703, "grad_norm": 0.8359128832817078, "learning_rate": 0.00021937722228208167, "loss": 3.2714, "step": 59000 }, { "epoch": 6.350231406737703, "eval_accuracy": 0.38533931373314917, "eval_loss": 3.3757359981536865, "eval_runtime": 184.9919, "eval_samples_per_second": 97.361, "eval_steps_per_second": 6.087, "step": 59000 }, { "epoch": 6.355612958777312, "grad_norm": 0.7550880312919617, "learning_rate": 0.00021905398125202024, "loss": 3.2664, "step": 59050 }, { "epoch": 6.360994510816919, "grad_norm": 0.7995261549949646, "learning_rate": 0.00021873074022195883, "loss": 3.2856, "step": 59100 }, { "epoch": 6.366376062856528, "grad_norm": 0.7748396992683411, "learning_rate": 0.0002184074991918974, "loss": 3.2692, "step": 59150 }, { "epoch": 6.371757614896136, "grad_norm": 0.7827137112617493, "learning_rate": 0.00021808425816183597, "loss": 3.298, "step": 59200 }, { "epoch": 6.377139166935744, "grad_norm": 0.7864527702331543, "learning_rate": 0.0002177610171317746, "loss": 3.2718, "step": 59250 }, { "epoch": 6.382520718975353, "grad_norm": 0.739658772945404, "learning_rate": 0.00021743777610171315, "loss": 3.2979, "step": 59300 }, { "epoch": 6.387902271014961, "grad_norm": 0.7770798206329346, "learning_rate": 0.00021711453507165172, "loss": 3.2887, "step": 59350 }, { "epoch": 6.393283823054569, "grad_norm": 0.7680968046188354, "learning_rate": 0.00021679129404159034, "loss": 3.2854, "step": 59400 }, { "epoch": 6.398665375094177, "grad_norm": 0.7551103830337524, "learning_rate": 0.0002164680530115289, "loss": 3.2648, "step": 59450 }, { "epoch": 6.404046927133785, "grad_norm": 0.7820691466331482, "learning_rate": 0.00021614481198146748, "loss": 3.2598, "step": 59500 }, { "epoch": 6.4094284791733935, "grad_norm": 0.7943058013916016, "learning_rate": 0.00021582157095140607, "loss": 3.2879, "step": 59550 }, { "epoch": 6.414810031213002, "grad_norm": 0.7810237407684326, "learning_rate": 0.00021549832992134467, "loss": 3.2768, "step": 59600 }, { "epoch": 6.42019158325261, "grad_norm": 0.7318100333213806, "learning_rate": 0.00021517508889128324, "loss": 3.3002, "step": 59650 }, { "epoch": 6.425573135292218, "grad_norm": 0.7224478721618652, "learning_rate": 0.00021485184786122183, "loss": 3.2621, "step": 59700 }, { "epoch": 6.430954687331827, "grad_norm": 0.7693201303482056, "learning_rate": 0.0002145286068311604, "loss": 3.2572, "step": 59750 }, { "epoch": 6.4363362393714345, "grad_norm": 0.7094773650169373, "learning_rate": 0.00021420536580109902, "loss": 3.2818, "step": 59800 }, { "epoch": 6.441717791411043, "grad_norm": 0.761139988899231, "learning_rate": 0.0002138821247710376, "loss": 3.2627, "step": 59850 }, { "epoch": 6.447099343450651, "grad_norm": 0.7221077084541321, "learning_rate": 0.00021355888374097615, "loss": 3.2619, "step": 59900 }, { "epoch": 6.452480895490259, "grad_norm": 0.7425982356071472, "learning_rate": 0.00021323564271091478, "loss": 3.297, "step": 59950 }, { "epoch": 6.457862447529868, "grad_norm": 0.8136957287788391, "learning_rate": 0.00021291240168085334, "loss": 3.2827, "step": 60000 }, { "epoch": 6.457862447529868, "eval_accuracy": 0.3858433544797205, "eval_loss": 3.369577169418335, "eval_runtime": 184.5983, "eval_samples_per_second": 97.569, "eval_steps_per_second": 6.1, "step": 60000 }, { "epoch": 6.4632439995694755, "grad_norm": 0.7648407220840454, "learning_rate": 0.0002125891606507919, "loss": 3.2689, "step": 60050 }, { "epoch": 6.468625551609084, "grad_norm": 0.7560406923294067, "learning_rate": 0.0002122659196207305, "loss": 3.2828, "step": 60100 }, { "epoch": 6.474007103648693, "grad_norm": 0.7538021206855774, "learning_rate": 0.0002119426785906691, "loss": 3.2751, "step": 60150 }, { "epoch": 6.4793886556883, "grad_norm": 0.796912670135498, "learning_rate": 0.00021161943756060767, "loss": 3.2776, "step": 60200 }, { "epoch": 6.484770207727909, "grad_norm": 0.7616314888000488, "learning_rate": 0.00021129619653054626, "loss": 3.2964, "step": 60250 }, { "epoch": 6.490151759767517, "grad_norm": 0.7025566101074219, "learning_rate": 0.00021097295550048483, "loss": 3.2876, "step": 60300 }, { "epoch": 6.495533311807125, "grad_norm": 0.775004506111145, "learning_rate": 0.00021064971447042343, "loss": 3.2618, "step": 60350 }, { "epoch": 6.500914863846734, "grad_norm": 0.7972927689552307, "learning_rate": 0.00021032647344036202, "loss": 3.2837, "step": 60400 }, { "epoch": 6.506296415886341, "grad_norm": 0.7656959295272827, "learning_rate": 0.0002100032324103006, "loss": 3.2853, "step": 60450 }, { "epoch": 6.51167796792595, "grad_norm": 0.7854245901107788, "learning_rate": 0.00020967999138023916, "loss": 3.2795, "step": 60500 }, { "epoch": 6.517059519965558, "grad_norm": 0.7886183857917786, "learning_rate": 0.00020935675035017778, "loss": 3.2849, "step": 60550 }, { "epoch": 6.522441072005166, "grad_norm": 0.8237834572792053, "learning_rate": 0.00020903350932011634, "loss": 3.285, "step": 60600 }, { "epoch": 6.5278226240447745, "grad_norm": 0.7524319291114807, "learning_rate": 0.0002087102682900549, "loss": 3.2721, "step": 60650 }, { "epoch": 6.533204176084383, "grad_norm": 0.7918885946273804, "learning_rate": 0.00020838702725999353, "loss": 3.2733, "step": 60700 }, { "epoch": 6.538585728123991, "grad_norm": 0.742572546005249, "learning_rate": 0.0002080637862299321, "loss": 3.2713, "step": 60750 }, { "epoch": 6.543967280163599, "grad_norm": 0.7656800150871277, "learning_rate": 0.0002077405451998707, "loss": 3.2906, "step": 60800 }, { "epoch": 6.549348832203208, "grad_norm": 0.7438012361526489, "learning_rate": 0.00020741730416980926, "loss": 3.2745, "step": 60850 }, { "epoch": 6.5547303842428155, "grad_norm": 0.793135404586792, "learning_rate": 0.00020709406313974786, "loss": 3.2778, "step": 60900 }, { "epoch": 6.560111936282424, "grad_norm": 0.7506826519966125, "learning_rate": 0.00020677082210968645, "loss": 3.2682, "step": 60950 }, { "epoch": 6.565493488322032, "grad_norm": 0.7645418047904968, "learning_rate": 0.00020644758107962502, "loss": 3.2671, "step": 61000 }, { "epoch": 6.565493488322032, "eval_accuracy": 0.386022305785951, "eval_loss": 3.368492603302002, "eval_runtime": 185.0491, "eval_samples_per_second": 97.331, "eval_steps_per_second": 6.085, "step": 61000 }, { "epoch": 6.57087504036164, "grad_norm": 0.7325201630592346, "learning_rate": 0.0002061243400495636, "loss": 3.2796, "step": 61050 }, { "epoch": 6.576256592401249, "grad_norm": 0.7918336987495422, "learning_rate": 0.0002058010990195022, "loss": 3.2781, "step": 61100 }, { "epoch": 6.5816381444408565, "grad_norm": 0.7214813232421875, "learning_rate": 0.00020547785798944078, "loss": 3.2881, "step": 61150 }, { "epoch": 6.587019696480465, "grad_norm": 0.7659515738487244, "learning_rate": 0.00020515461695937934, "loss": 3.2731, "step": 61200 }, { "epoch": 6.592401248520073, "grad_norm": 0.7384321093559265, "learning_rate": 0.00020483137592931797, "loss": 3.2826, "step": 61250 }, { "epoch": 6.597782800559681, "grad_norm": 0.7732715010643005, "learning_rate": 0.00020450813489925653, "loss": 3.2796, "step": 61300 }, { "epoch": 6.60316435259929, "grad_norm": 0.7654476165771484, "learning_rate": 0.0002041848938691951, "loss": 3.2582, "step": 61350 }, { "epoch": 6.608545904638898, "grad_norm": 0.7931333184242249, "learning_rate": 0.0002038616528391337, "loss": 3.2684, "step": 61400 }, { "epoch": 6.613927456678506, "grad_norm": 0.7446388602256775, "learning_rate": 0.00020353841180907226, "loss": 3.2826, "step": 61450 }, { "epoch": 6.619309008718115, "grad_norm": 0.7642561197280884, "learning_rate": 0.00020321517077901086, "loss": 3.2918, "step": 61500 }, { "epoch": 6.624690560757722, "grad_norm": 0.8336501717567444, "learning_rate": 0.00020289192974894945, "loss": 3.2695, "step": 61550 }, { "epoch": 6.630072112797331, "grad_norm": 0.7792569994926453, "learning_rate": 0.00020256868871888802, "loss": 3.285, "step": 61600 }, { "epoch": 6.635453664836939, "grad_norm": 0.8366368412971497, "learning_rate": 0.00020224544768882664, "loss": 3.2801, "step": 61650 }, { "epoch": 6.640835216876547, "grad_norm": 0.7799500226974487, "learning_rate": 0.0002019222066587652, "loss": 3.2867, "step": 61700 }, { "epoch": 6.6462167689161555, "grad_norm": 0.8103165626525879, "learning_rate": 0.00020159896562870378, "loss": 3.2709, "step": 61750 }, { "epoch": 6.651598320955763, "grad_norm": 0.7403290867805481, "learning_rate": 0.00020127572459864237, "loss": 3.2901, "step": 61800 }, { "epoch": 6.656979872995372, "grad_norm": 0.7185531854629517, "learning_rate": 0.00020095248356858097, "loss": 3.2843, "step": 61850 }, { "epoch": 6.66236142503498, "grad_norm": 0.7514901161193848, "learning_rate": 0.00020062924253851953, "loss": 3.2866, "step": 61900 }, { "epoch": 6.667742977074588, "grad_norm": 0.7954921126365662, "learning_rate": 0.00020030600150845813, "loss": 3.2607, "step": 61950 }, { "epoch": 6.6731245291141965, "grad_norm": 0.7863759398460388, "learning_rate": 0.0001999827604783967, "loss": 3.3058, "step": 62000 }, { "epoch": 6.6731245291141965, "eval_accuracy": 0.3863287069295878, "eval_loss": 3.3633809089660645, "eval_runtime": 184.8165, "eval_samples_per_second": 97.453, "eval_steps_per_second": 6.093, "step": 62000 }, { "epoch": 6.678506081153805, "grad_norm": 0.7923064827919006, "learning_rate": 0.0001996595194483353, "loss": 3.3203, "step": 62050 }, { "epoch": 6.683887633193413, "grad_norm": 0.7636295557022095, "learning_rate": 0.00019933627841827389, "loss": 3.297, "step": 62100 }, { "epoch": 6.689269185233021, "grad_norm": 0.740822970867157, "learning_rate": 0.00019901303738821245, "loss": 3.2925, "step": 62150 }, { "epoch": 6.69465073727263, "grad_norm": 0.7824040651321411, "learning_rate": 0.00019868979635815102, "loss": 3.2845, "step": 62200 }, { "epoch": 6.7000322893122375, "grad_norm": 0.7348474860191345, "learning_rate": 0.00019836655532808964, "loss": 3.2755, "step": 62250 }, { "epoch": 6.705413841351846, "grad_norm": 0.8276709318161011, "learning_rate": 0.0001980433142980282, "loss": 3.302, "step": 62300 }, { "epoch": 6.710795393391454, "grad_norm": 0.7462751865386963, "learning_rate": 0.00019772007326796678, "loss": 3.2734, "step": 62350 }, { "epoch": 6.716176945431062, "grad_norm": 0.8004351258277893, "learning_rate": 0.0001973968322379054, "loss": 3.278, "step": 62400 }, { "epoch": 6.721558497470671, "grad_norm": 0.7743335366249084, "learning_rate": 0.00019707359120784397, "loss": 3.2841, "step": 62450 }, { "epoch": 6.7269400495102785, "grad_norm": 0.7511734366416931, "learning_rate": 0.00019675035017778253, "loss": 3.277, "step": 62500 }, { "epoch": 6.732321601549887, "grad_norm": 0.7555059194564819, "learning_rate": 0.00019642710914772113, "loss": 3.2914, "step": 62550 }, { "epoch": 6.737703153589496, "grad_norm": 0.7607969641685486, "learning_rate": 0.00019610386811765972, "loss": 3.2923, "step": 62600 }, { "epoch": 6.743084705629103, "grad_norm": 0.7745279669761658, "learning_rate": 0.00019578062708759832, "loss": 3.2707, "step": 62650 }, { "epoch": 6.748466257668712, "grad_norm": 0.7459644675254822, "learning_rate": 0.00019545738605753689, "loss": 3.2656, "step": 62700 }, { "epoch": 6.75384780970832, "grad_norm": 0.7880685329437256, "learning_rate": 0.00019513414502747545, "loss": 3.271, "step": 62750 }, { "epoch": 6.759229361747928, "grad_norm": 0.7434306740760803, "learning_rate": 0.00019481090399741408, "loss": 3.2812, "step": 62800 }, { "epoch": 6.7646109137875365, "grad_norm": 0.8286969065666199, "learning_rate": 0.0001945005926085551, "loss": 3.2948, "step": 62850 }, { "epoch": 6.769992465827144, "grad_norm": 0.7807492613792419, "learning_rate": 0.0001941773515784937, "loss": 3.296, "step": 62900 }, { "epoch": 6.775374017866753, "grad_norm": 0.8009621500968933, "learning_rate": 0.00019385411054843226, "loss": 3.2822, "step": 62950 }, { "epoch": 6.780755569906361, "grad_norm": 0.7488981485366821, "learning_rate": 0.00019353086951837085, "loss": 3.2775, "step": 63000 }, { "epoch": 6.780755569906361, "eval_accuracy": 0.3868306832712907, "eval_loss": 3.3595683574676514, "eval_runtime": 184.9791, "eval_samples_per_second": 97.368, "eval_steps_per_second": 6.087, "step": 63000 }, { "epoch": 6.786137121945969, "grad_norm": 0.7897948026657104, "learning_rate": 0.00019320762848830942, "loss": 3.277, "step": 63050 }, { "epoch": 6.7915186739855775, "grad_norm": 0.7589372992515564, "learning_rate": 0.000192884387458248, "loss": 3.2875, "step": 63100 }, { "epoch": 6.796900226025185, "grad_norm": 0.7196517586708069, "learning_rate": 0.0001925611464281866, "loss": 3.2844, "step": 63150 }, { "epoch": 6.802281778064794, "grad_norm": 0.7708859443664551, "learning_rate": 0.00019223790539812518, "loss": 3.2796, "step": 63200 }, { "epoch": 6.807663330104402, "grad_norm": 0.7655634880065918, "learning_rate": 0.00019191466436806374, "loss": 3.2928, "step": 63250 }, { "epoch": 6.813044882144011, "grad_norm": 0.8020054697990417, "learning_rate": 0.00019159142333800236, "loss": 3.259, "step": 63300 }, { "epoch": 6.8184264341836185, "grad_norm": 0.7741644382476807, "learning_rate": 0.00019126818230794093, "loss": 3.2918, "step": 63350 }, { "epoch": 6.823807986223227, "grad_norm": 0.7251269221305847, "learning_rate": 0.00019094494127787953, "loss": 3.2822, "step": 63400 }, { "epoch": 6.829189538262835, "grad_norm": 0.7983183860778809, "learning_rate": 0.00019062170024781812, "loss": 3.3032, "step": 63450 }, { "epoch": 6.834571090302443, "grad_norm": 0.7949510216712952, "learning_rate": 0.0001902984592177567, "loss": 3.284, "step": 63500 }, { "epoch": 6.839952642342052, "grad_norm": 0.788966715335846, "learning_rate": 0.00018997521818769528, "loss": 3.2741, "step": 63550 }, { "epoch": 6.8453341943816595, "grad_norm": 0.7868644595146179, "learning_rate": 0.00018965197715763385, "loss": 3.2781, "step": 63600 }, { "epoch": 6.850715746421268, "grad_norm": 0.7467278242111206, "learning_rate": 0.00018932873612757245, "loss": 3.2761, "step": 63650 }, { "epoch": 6.856097298460876, "grad_norm": 0.7820574641227722, "learning_rate": 0.00018900549509751104, "loss": 3.2763, "step": 63700 }, { "epoch": 6.861478850500484, "grad_norm": 0.7491395473480225, "learning_rate": 0.0001886822540674496, "loss": 3.2862, "step": 63750 }, { "epoch": 6.866860402540093, "grad_norm": 0.8236872553825378, "learning_rate": 0.00018835901303738818, "loss": 3.2784, "step": 63800 }, { "epoch": 6.8722419545797, "grad_norm": 0.7589198350906372, "learning_rate": 0.0001880357720073268, "loss": 3.2835, "step": 63850 }, { "epoch": 6.877623506619309, "grad_norm": 0.7768234014511108, "learning_rate": 0.00018771253097726537, "loss": 3.2868, "step": 63900 }, { "epoch": 6.8830050586589175, "grad_norm": 0.7730334997177124, "learning_rate": 0.00018738928994720393, "loss": 3.292, "step": 63950 }, { "epoch": 6.888386610698525, "grad_norm": 0.7769656181335449, "learning_rate": 0.00018706604891714255, "loss": 3.2919, "step": 64000 }, { "epoch": 6.888386610698525, "eval_accuracy": 0.3873528690501271, "eval_loss": 3.354077100753784, "eval_runtime": 185.2307, "eval_samples_per_second": 97.236, "eval_steps_per_second": 6.079, "step": 64000 }, { "epoch": 6.893768162738134, "grad_norm": 0.8213059902191162, "learning_rate": 0.00018674280788708112, "loss": 3.2788, "step": 64050 }, { "epoch": 6.899149714777742, "grad_norm": 0.7441644668579102, "learning_rate": 0.0001864195668570197, "loss": 3.2803, "step": 64100 }, { "epoch": 6.90453126681735, "grad_norm": 0.7904473543167114, "learning_rate": 0.00018609632582695828, "loss": 3.2874, "step": 64150 }, { "epoch": 6.9099128188569585, "grad_norm": 0.7749623656272888, "learning_rate": 0.00018577308479689685, "loss": 3.2675, "step": 64200 }, { "epoch": 6.915294370896566, "grad_norm": 0.7651856541633606, "learning_rate": 0.00018544984376683545, "loss": 3.2799, "step": 64250 }, { "epoch": 6.920675922936175, "grad_norm": 0.8032562136650085, "learning_rate": 0.00018512660273677404, "loss": 3.2993, "step": 64300 }, { "epoch": 6.926057474975783, "grad_norm": 0.7830508351325989, "learning_rate": 0.0001848033617067126, "loss": 3.2769, "step": 64350 }, { "epoch": 6.931439027015391, "grad_norm": 0.778607964515686, "learning_rate": 0.00018448012067665123, "loss": 3.2785, "step": 64400 }, { "epoch": 6.9368205790549995, "grad_norm": 0.7424008846282959, "learning_rate": 0.0001841568796465898, "loss": 3.2746, "step": 64450 }, { "epoch": 6.942202131094608, "grad_norm": 0.7957499027252197, "learning_rate": 0.00018383363861652837, "loss": 3.2933, "step": 64500 }, { "epoch": 6.947583683134216, "grad_norm": 0.8147248029708862, "learning_rate": 0.00018351039758646696, "loss": 3.273, "step": 64550 }, { "epoch": 6.952965235173824, "grad_norm": 0.7838407158851624, "learning_rate": 0.00018318715655640555, "loss": 3.2845, "step": 64600 }, { "epoch": 6.958346787213433, "grad_norm": 0.7947826385498047, "learning_rate": 0.00018286391552634412, "loss": 3.299, "step": 64650 }, { "epoch": 6.9637283392530405, "grad_norm": 0.8972591161727905, "learning_rate": 0.00018254067449628272, "loss": 3.3021, "step": 64700 }, { "epoch": 6.969109891292649, "grad_norm": 0.7764172554016113, "learning_rate": 0.00018221743346622128, "loss": 3.2696, "step": 64750 }, { "epoch": 6.974491443332257, "grad_norm": 0.8290959000587463, "learning_rate": 0.00018189419243615988, "loss": 3.2747, "step": 64800 }, { "epoch": 6.979872995371865, "grad_norm": 0.751587450504303, "learning_rate": 0.00018157095140609847, "loss": 3.2917, "step": 64850 }, { "epoch": 6.985254547411474, "grad_norm": 0.7895004153251648, "learning_rate": 0.00018124771037603704, "loss": 3.2852, "step": 64900 }, { "epoch": 6.990636099451081, "grad_norm": 0.7497310638427734, "learning_rate": 0.0001809244693459756, "loss": 3.3, "step": 64950 }, { "epoch": 6.99601765149069, "grad_norm": 0.763469398021698, "learning_rate": 0.00018060122831591423, "loss": 3.2603, "step": 65000 }, { "epoch": 6.99601765149069, "eval_accuracy": 0.38770675150573886, "eval_loss": 3.351874828338623, "eval_runtime": 184.7786, "eval_samples_per_second": 97.473, "eval_steps_per_second": 6.094, "step": 65000 }, { "epoch": 7.0013992035302985, "grad_norm": 0.7597163915634155, "learning_rate": 0.0001802779872858528, "loss": 3.2679, "step": 65050 }, { "epoch": 7.006780755569906, "grad_norm": 0.7525449395179749, "learning_rate": 0.00017995474625579137, "loss": 3.2098, "step": 65100 }, { "epoch": 7.012162307609515, "grad_norm": 0.8813150525093079, "learning_rate": 0.00017963150522573, "loss": 3.1881, "step": 65150 }, { "epoch": 7.017543859649122, "grad_norm": 0.787375271320343, "learning_rate": 0.0001793147290162698, "loss": 3.1957, "step": 65200 }, { "epoch": 7.022925411688731, "grad_norm": 0.7959186434745789, "learning_rate": 0.00017899148798620836, "loss": 3.2153, "step": 65250 }, { "epoch": 7.0283069637283395, "grad_norm": 0.7243605256080627, "learning_rate": 0.00017866824695614696, "loss": 3.1987, "step": 65300 }, { "epoch": 7.033688515767947, "grad_norm": 0.7697280049324036, "learning_rate": 0.00017834500592608555, "loss": 3.2116, "step": 65350 }, { "epoch": 7.039070067807556, "grad_norm": 0.781091034412384, "learning_rate": 0.00017802176489602412, "loss": 3.192, "step": 65400 }, { "epoch": 7.044451619847164, "grad_norm": 0.8070541620254517, "learning_rate": 0.00017769852386596269, "loss": 3.2159, "step": 65450 }, { "epoch": 7.049833171886772, "grad_norm": 0.7928311228752136, "learning_rate": 0.0001773752828359013, "loss": 3.2199, "step": 65500 }, { "epoch": 7.0552147239263805, "grad_norm": 0.8367716670036316, "learning_rate": 0.00017705204180583988, "loss": 3.1962, "step": 65550 }, { "epoch": 7.060596275965988, "grad_norm": 0.8304590582847595, "learning_rate": 0.00017672880077577844, "loss": 3.2124, "step": 65600 }, { "epoch": 7.065977828005597, "grad_norm": 0.789740800857544, "learning_rate": 0.00017640555974571704, "loss": 3.1983, "step": 65650 }, { "epoch": 7.071359380045205, "grad_norm": 0.7952477335929871, "learning_rate": 0.00017608231871565563, "loss": 3.2149, "step": 65700 }, { "epoch": 7.076740932084813, "grad_norm": 0.8178505897521973, "learning_rate": 0.0001757590776855942, "loss": 3.2005, "step": 65750 }, { "epoch": 7.0821224841244215, "grad_norm": 0.8267757892608643, "learning_rate": 0.0001754358366555328, "loss": 3.2145, "step": 65800 }, { "epoch": 7.08750403616403, "grad_norm": 0.7804736495018005, "learning_rate": 0.00017511259562547136, "loss": 3.204, "step": 65850 }, { "epoch": 7.092885588203638, "grad_norm": 0.8235890865325928, "learning_rate": 0.00017478935459540996, "loss": 3.2238, "step": 65900 }, { "epoch": 7.098267140243246, "grad_norm": 0.8137307167053223, "learning_rate": 0.00017446611356534855, "loss": 3.2068, "step": 65950 }, { "epoch": 7.103648692282855, "grad_norm": 0.7665497660636902, "learning_rate": 0.00017414287253528712, "loss": 3.1971, "step": 66000 }, { "epoch": 7.103648692282855, "eval_accuracy": 0.3873983946101214, "eval_loss": 3.3571224212646484, "eval_runtime": 185.0933, "eval_samples_per_second": 97.308, "eval_steps_per_second": 6.083, "step": 66000 }, { "epoch": 7.109030244322462, "grad_norm": 0.8523070812225342, "learning_rate": 0.00017381963150522569, "loss": 3.2132, "step": 66050 }, { "epoch": 7.114411796362071, "grad_norm": 0.8501766920089722, "learning_rate": 0.0001734963904751643, "loss": 3.2196, "step": 66100 }, { "epoch": 7.119793348401679, "grad_norm": 0.7479291558265686, "learning_rate": 0.00017317314944510288, "loss": 3.2021, "step": 66150 }, { "epoch": 7.125174900441287, "grad_norm": 0.8581606149673462, "learning_rate": 0.00017284990841504147, "loss": 3.238, "step": 66200 }, { "epoch": 7.130556452480896, "grad_norm": 0.8038374185562134, "learning_rate": 0.00017252666738498007, "loss": 3.2067, "step": 66250 }, { "epoch": 7.135938004520503, "grad_norm": 0.7955670356750488, "learning_rate": 0.00017220342635491863, "loss": 3.2062, "step": 66300 }, { "epoch": 7.141319556560112, "grad_norm": 0.8133190274238586, "learning_rate": 0.00017188018532485723, "loss": 3.2189, "step": 66350 }, { "epoch": 7.1467011085997205, "grad_norm": 0.7808161377906799, "learning_rate": 0.0001715569442947958, "loss": 3.2181, "step": 66400 }, { "epoch": 7.152082660639328, "grad_norm": 0.7812925577163696, "learning_rate": 0.0001712337032647344, "loss": 3.21, "step": 66450 }, { "epoch": 7.157464212678937, "grad_norm": 0.7803530097007751, "learning_rate": 0.00017091046223467298, "loss": 3.221, "step": 66500 }, { "epoch": 7.162845764718545, "grad_norm": 0.8198121786117554, "learning_rate": 0.00017058722120461155, "loss": 3.2165, "step": 66550 }, { "epoch": 7.168227316758153, "grad_norm": 0.777133047580719, "learning_rate": 0.00017026398017455012, "loss": 3.2101, "step": 66600 }, { "epoch": 7.1736088687977615, "grad_norm": 0.7378299236297607, "learning_rate": 0.00016994073914448874, "loss": 3.2223, "step": 66650 }, { "epoch": 7.178990420837369, "grad_norm": 0.8564599752426147, "learning_rate": 0.0001696174981144273, "loss": 3.2103, "step": 66700 }, { "epoch": 7.184371972876978, "grad_norm": 0.8352597951889038, "learning_rate": 0.00016929425708436588, "loss": 3.2395, "step": 66750 }, { "epoch": 7.189753524916586, "grad_norm": 0.7732496857643127, "learning_rate": 0.0001689710160543045, "loss": 3.209, "step": 66800 }, { "epoch": 7.195135076956194, "grad_norm": 0.7928017377853394, "learning_rate": 0.00016864777502424307, "loss": 3.208, "step": 66850 }, { "epoch": 7.2005166289958025, "grad_norm": 0.8274809122085571, "learning_rate": 0.00016832453399418163, "loss": 3.2147, "step": 66900 }, { "epoch": 7.205898181035411, "grad_norm": 0.812453031539917, "learning_rate": 0.00016800129296412023, "loss": 3.2238, "step": 66950 }, { "epoch": 7.211279733075019, "grad_norm": 0.7714659571647644, "learning_rate": 0.00016767805193405882, "loss": 3.2284, "step": 67000 }, { "epoch": 7.211279733075019, "eval_accuracy": 0.38780638620386476, "eval_loss": 3.356039047241211, "eval_runtime": 185.0026, "eval_samples_per_second": 97.355, "eval_steps_per_second": 6.086, "step": 67000 }, { "epoch": 7.216661285114627, "grad_norm": 0.7873409390449524, "learning_rate": 0.0001673548109039974, "loss": 3.2216, "step": 67050 }, { "epoch": 7.222042837154235, "grad_norm": 0.8289926052093506, "learning_rate": 0.00016703156987393598, "loss": 3.2315, "step": 67100 }, { "epoch": 7.2274243891938434, "grad_norm": 0.7620259523391724, "learning_rate": 0.00016670832884387455, "loss": 3.2078, "step": 67150 }, { "epoch": 7.232805941233452, "grad_norm": 0.8179631233215332, "learning_rate": 0.00016638508781381317, "loss": 3.2387, "step": 67200 }, { "epoch": 7.23818749327306, "grad_norm": 0.8092691898345947, "learning_rate": 0.00016606184678375174, "loss": 3.2218, "step": 67250 }, { "epoch": 7.243569045312668, "grad_norm": 0.8265263438224792, "learning_rate": 0.0001657386057536903, "loss": 3.2141, "step": 67300 }, { "epoch": 7.248950597352277, "grad_norm": 0.7724214196205139, "learning_rate": 0.00016541536472362893, "loss": 3.2062, "step": 67350 }, { "epoch": 7.254332149391884, "grad_norm": 0.8382948040962219, "learning_rate": 0.0001650921236935675, "loss": 3.2272, "step": 67400 }, { "epoch": 7.259713701431493, "grad_norm": 0.7633907794952393, "learning_rate": 0.00016476888266350607, "loss": 3.2307, "step": 67450 }, { "epoch": 7.265095253471101, "grad_norm": 0.8241746425628662, "learning_rate": 0.00016444564163344466, "loss": 3.2082, "step": 67500 }, { "epoch": 7.270476805510709, "grad_norm": 0.8180881142616272, "learning_rate": 0.00016412240060338326, "loss": 3.2256, "step": 67550 }, { "epoch": 7.275858357550318, "grad_norm": 0.8148139119148254, "learning_rate": 0.00016379915957332182, "loss": 3.2394, "step": 67600 }, { "epoch": 7.281239909589925, "grad_norm": 0.8224004507064819, "learning_rate": 0.00016348238336386163, "loss": 3.2298, "step": 67650 }, { "epoch": 7.286621461629534, "grad_norm": 0.7899991273880005, "learning_rate": 0.0001631591423338002, "loss": 3.2128, "step": 67700 }, { "epoch": 7.2920030136691425, "grad_norm": 0.8135297298431396, "learning_rate": 0.00016283590130373882, "loss": 3.2192, "step": 67750 }, { "epoch": 7.29738456570875, "grad_norm": 0.8687593340873718, "learning_rate": 0.00016251266027367739, "loss": 3.2169, "step": 67800 }, { "epoch": 7.302766117748359, "grad_norm": 0.8100031018257141, "learning_rate": 0.00016218941924361595, "loss": 3.2052, "step": 67850 }, { "epoch": 7.308147669787967, "grad_norm": 0.7895699739456177, "learning_rate": 0.00016186617821355458, "loss": 3.2334, "step": 67900 }, { "epoch": 7.313529221827575, "grad_norm": 0.7895455956459045, "learning_rate": 0.00016154293718349314, "loss": 3.2231, "step": 67950 }, { "epoch": 7.3189107738671835, "grad_norm": 0.8295309543609619, "learning_rate": 0.00016121969615343174, "loss": 3.2006, "step": 68000 }, { "epoch": 7.3189107738671835, "eval_accuracy": 0.3882477342341672, "eval_loss": 3.351390838623047, "eval_runtime": 184.5883, "eval_samples_per_second": 97.574, "eval_steps_per_second": 6.1, "step": 68000 }, { "epoch": 7.324292325906791, "grad_norm": 0.8855544328689575, "learning_rate": 0.0001608964551233703, "loss": 3.214, "step": 68050 }, { "epoch": 7.3296738779464, "grad_norm": 0.8429924249649048, "learning_rate": 0.0001605732140933089, "loss": 3.2306, "step": 68100 }, { "epoch": 7.335055429986008, "grad_norm": 0.8152256011962891, "learning_rate": 0.0001602499730632475, "loss": 3.2301, "step": 68150 }, { "epoch": 7.340436982025616, "grad_norm": 0.8925390243530273, "learning_rate": 0.00015992673203318606, "loss": 3.2047, "step": 68200 }, { "epoch": 7.3458185340652244, "grad_norm": 0.8430379033088684, "learning_rate": 0.00015960349100312463, "loss": 3.2294, "step": 68250 }, { "epoch": 7.351200086104833, "grad_norm": 0.8633614778518677, "learning_rate": 0.00015928024997306325, "loss": 3.2202, "step": 68300 }, { "epoch": 7.356581638144441, "grad_norm": 0.8025208711624146, "learning_rate": 0.00015895700894300182, "loss": 3.2208, "step": 68350 }, { "epoch": 7.361963190184049, "grad_norm": 0.828989565372467, "learning_rate": 0.00015863376791294039, "loss": 3.2162, "step": 68400 }, { "epoch": 7.367344742223658, "grad_norm": 0.8749890923500061, "learning_rate": 0.000158310526882879, "loss": 3.2363, "step": 68450 }, { "epoch": 7.372726294263265, "grad_norm": 0.8234397768974304, "learning_rate": 0.00015798728585281758, "loss": 3.2358, "step": 68500 }, { "epoch": 7.378107846302874, "grad_norm": 0.8228161931037903, "learning_rate": 0.00015766404482275614, "loss": 3.2272, "step": 68550 }, { "epoch": 7.383489398342482, "grad_norm": 0.8303384780883789, "learning_rate": 0.00015734080379269474, "loss": 3.2314, "step": 68600 }, { "epoch": 7.38887095038209, "grad_norm": 0.8005957007408142, "learning_rate": 0.00015701756276263333, "loss": 3.2291, "step": 68650 }, { "epoch": 7.394252502421699, "grad_norm": 0.8635926842689514, "learning_rate": 0.0001566943217325719, "loss": 3.2199, "step": 68700 }, { "epoch": 7.399634054461306, "grad_norm": 0.7820467352867126, "learning_rate": 0.0001563710807025105, "loss": 3.2154, "step": 68750 }, { "epoch": 7.405015606500915, "grad_norm": 0.8372554779052734, "learning_rate": 0.00015604783967244906, "loss": 3.246, "step": 68800 }, { "epoch": 7.4103971585405235, "grad_norm": 0.8614478707313538, "learning_rate": 0.00015572459864238763, "loss": 3.2373, "step": 68850 }, { "epoch": 7.415778710580131, "grad_norm": 0.7786409854888916, "learning_rate": 0.00015540135761232625, "loss": 3.2196, "step": 68900 }, { "epoch": 7.42116026261974, "grad_norm": 0.80152827501297, "learning_rate": 0.00015507811658226482, "loss": 3.2348, "step": 68950 }, { "epoch": 7.426541814659347, "grad_norm": 0.8448081612586975, "learning_rate": 0.00015475487555220344, "loss": 3.2225, "step": 69000 }, { "epoch": 7.426541814659347, "eval_accuracy": 0.38895625971560543, "eval_loss": 3.3472721576690674, "eval_runtime": 184.8693, "eval_samples_per_second": 97.426, "eval_steps_per_second": 6.091, "step": 69000 }, { "epoch": 7.431923366698956, "grad_norm": 0.7952991127967834, "learning_rate": 0.000154431634522142, "loss": 3.2336, "step": 69050 }, { "epoch": 7.4373049187385645, "grad_norm": 0.8257370591163635, "learning_rate": 0.00015410839349208058, "loss": 3.2352, "step": 69100 }, { "epoch": 7.442686470778172, "grad_norm": 0.7818418741226196, "learning_rate": 0.00015378515246201917, "loss": 3.2127, "step": 69150 }, { "epoch": 7.448068022817781, "grad_norm": 0.7481926083564758, "learning_rate": 0.00015346191143195774, "loss": 3.2302, "step": 69200 }, { "epoch": 7.453449574857389, "grad_norm": 0.7871079444885254, "learning_rate": 0.00015313867040189633, "loss": 3.2131, "step": 69250 }, { "epoch": 7.458831126896997, "grad_norm": 0.8238528370857239, "learning_rate": 0.00015281542937183493, "loss": 3.2312, "step": 69300 }, { "epoch": 7.4642126789366054, "grad_norm": 0.8295047283172607, "learning_rate": 0.0001524921883417735, "loss": 3.2295, "step": 69350 }, { "epoch": 7.469594230976213, "grad_norm": 0.8163410425186157, "learning_rate": 0.00015216894731171206, "loss": 3.2072, "step": 69400 }, { "epoch": 7.474975783015822, "grad_norm": 0.8015903830528259, "learning_rate": 0.00015184570628165068, "loss": 3.2306, "step": 69450 }, { "epoch": 7.48035733505543, "grad_norm": 0.8221493363380432, "learning_rate": 0.00015152246525158925, "loss": 3.2241, "step": 69500 }, { "epoch": 7.485738887095038, "grad_norm": 0.8415564298629761, "learning_rate": 0.00015119922422152782, "loss": 3.2263, "step": 69550 }, { "epoch": 7.491120439134646, "grad_norm": 0.8357605338096619, "learning_rate": 0.00015087598319146644, "loss": 3.2443, "step": 69600 }, { "epoch": 7.496501991174255, "grad_norm": 0.8570162057876587, "learning_rate": 0.000150552742161405, "loss": 3.2303, "step": 69650 }, { "epoch": 7.501883543213863, "grad_norm": 0.8457155227661133, "learning_rate": 0.00015022950113134358, "loss": 3.2263, "step": 69700 }, { "epoch": 7.507265095253471, "grad_norm": 0.8195136785507202, "learning_rate": 0.00014990626010128217, "loss": 3.246, "step": 69750 }, { "epoch": 7.51264664729308, "grad_norm": 0.9330834150314331, "learning_rate": 0.00014958301907122077, "loss": 3.2381, "step": 69800 }, { "epoch": 7.518028199332687, "grad_norm": 0.836077094078064, "learning_rate": 0.00014925977804115933, "loss": 3.2522, "step": 69850 }, { "epoch": 7.523409751372296, "grad_norm": 0.8031020760536194, "learning_rate": 0.00014893653701109793, "loss": 3.237, "step": 69900 }, { "epoch": 7.528791303411904, "grad_norm": 0.8233877420425415, "learning_rate": 0.0001486132959810365, "loss": 3.229, "step": 69950 }, { "epoch": 7.534172855451512, "grad_norm": 0.8217124342918396, "learning_rate": 0.0001482900549509751, "loss": 3.2258, "step": 70000 }, { "epoch": 7.534172855451512, "eval_accuracy": 0.3891678068881802, "eval_loss": 3.345587968826294, "eval_runtime": 184.9082, "eval_samples_per_second": 97.405, "eval_steps_per_second": 6.09, "step": 70000 }, { "epoch": 7.539554407491121, "grad_norm": 0.8223305344581604, "learning_rate": 0.00014796681392091368, "loss": 3.2386, "step": 70050 }, { "epoch": 7.544935959530728, "grad_norm": 0.8422094583511353, "learning_rate": 0.00014764357289085228, "loss": 3.2385, "step": 70100 }, { "epoch": 7.550317511570337, "grad_norm": 0.8534138798713684, "learning_rate": 0.00014732033186079085, "loss": 3.2214, "step": 70150 }, { "epoch": 7.5556990636099455, "grad_norm": 0.8637493252754211, "learning_rate": 0.00014699709083072944, "loss": 3.2428, "step": 70200 }, { "epoch": 7.561080615649553, "grad_norm": 0.831741452217102, "learning_rate": 0.00014667384980066804, "loss": 3.2154, "step": 70250 }, { "epoch": 7.566462167689162, "grad_norm": 0.8678784370422363, "learning_rate": 0.00014635707359120784, "loss": 3.2371, "step": 70300 }, { "epoch": 7.57184371972877, "grad_norm": 0.8292410373687744, "learning_rate": 0.0001460338325611464, "loss": 3.2373, "step": 70350 }, { "epoch": 7.577225271768378, "grad_norm": 0.7928509712219238, "learning_rate": 0.00014571059153108498, "loss": 3.2306, "step": 70400 }, { "epoch": 7.5826068238079865, "grad_norm": 0.8370128870010376, "learning_rate": 0.00014538735050102357, "loss": 3.2222, "step": 70450 }, { "epoch": 7.587988375847594, "grad_norm": 0.801749587059021, "learning_rate": 0.00014506410947096217, "loss": 3.2312, "step": 70500 }, { "epoch": 7.593369927887203, "grad_norm": 0.8322213292121887, "learning_rate": 0.00014474086844090076, "loss": 3.2293, "step": 70550 }, { "epoch": 7.598751479926811, "grad_norm": 0.8021792769432068, "learning_rate": 0.00014441762741083933, "loss": 3.2364, "step": 70600 }, { "epoch": 7.604133031966419, "grad_norm": 0.8352303504943848, "learning_rate": 0.00014409438638077792, "loss": 3.2124, "step": 70650 }, { "epoch": 7.609514584006027, "grad_norm": 0.850394070148468, "learning_rate": 0.00014377114535071652, "loss": 3.2298, "step": 70700 }, { "epoch": 7.614896136045635, "grad_norm": 0.8542826771736145, "learning_rate": 0.0001434479043206551, "loss": 3.2264, "step": 70750 }, { "epoch": 7.620277688085244, "grad_norm": 0.8750932216644287, "learning_rate": 0.00014312466329059368, "loss": 3.2408, "step": 70800 }, { "epoch": 7.625659240124852, "grad_norm": 0.8510986566543579, "learning_rate": 0.00014280142226053225, "loss": 3.21, "step": 70850 }, { "epoch": 7.63104079216446, "grad_norm": 0.8305497765541077, "learning_rate": 0.00014247818123047084, "loss": 3.2402, "step": 70900 }, { "epoch": 7.636422344204068, "grad_norm": 0.8636519908905029, "learning_rate": 0.0001421549402004094, "loss": 3.2141, "step": 70950 }, { "epoch": 7.641803896243677, "grad_norm": 0.871688723564148, "learning_rate": 0.000141831699170348, "loss": 3.2407, "step": 71000 }, { "epoch": 7.641803896243677, "eval_accuracy": 0.3893689233835248, "eval_loss": 3.3400113582611084, "eval_runtime": 185.1188, "eval_samples_per_second": 97.294, "eval_steps_per_second": 6.083, "step": 71000 }, { "epoch": 7.647185448283285, "grad_norm": 0.8559573888778687, "learning_rate": 0.0001415084581402866, "loss": 3.2194, "step": 71050 }, { "epoch": 7.652567000322893, "grad_norm": 0.8053033947944641, "learning_rate": 0.00014118521711022517, "loss": 3.2275, "step": 71100 }, { "epoch": 7.657948552362502, "grad_norm": 0.872286856174469, "learning_rate": 0.00014086197608016376, "loss": 3.2224, "step": 71150 }, { "epoch": 7.663330104402109, "grad_norm": 0.7740647792816162, "learning_rate": 0.00014053873505010236, "loss": 3.2056, "step": 71200 }, { "epoch": 7.668711656441718, "grad_norm": 0.858696699142456, "learning_rate": 0.00014021549402004092, "loss": 3.2429, "step": 71250 }, { "epoch": 7.674093208481326, "grad_norm": 0.825717031955719, "learning_rate": 0.00013989225298997952, "loss": 3.24, "step": 71300 }, { "epoch": 7.679474760520934, "grad_norm": 0.8377315402030945, "learning_rate": 0.00013956901195991811, "loss": 3.2464, "step": 71350 }, { "epoch": 7.684856312560543, "grad_norm": 0.8150701522827148, "learning_rate": 0.00013924577092985668, "loss": 3.2356, "step": 71400 }, { "epoch": 7.69023786460015, "grad_norm": 0.8369122743606567, "learning_rate": 0.00013892252989979528, "loss": 3.2431, "step": 71450 }, { "epoch": 7.695619416639759, "grad_norm": 0.8841993808746338, "learning_rate": 0.00013859928886973384, "loss": 3.2415, "step": 71500 }, { "epoch": 7.7010009686793675, "grad_norm": 0.8426228165626526, "learning_rate": 0.00013827604783967244, "loss": 3.2186, "step": 71550 }, { "epoch": 7.706382520718975, "grad_norm": 0.819513738155365, "learning_rate": 0.000137952806809611, "loss": 3.2303, "step": 71600 }, { "epoch": 7.711764072758584, "grad_norm": 0.8620902895927429, "learning_rate": 0.0001376295657795496, "loss": 3.247, "step": 71650 }, { "epoch": 7.717145624798192, "grad_norm": 0.8760289549827576, "learning_rate": 0.0001373063247494882, "loss": 3.2234, "step": 71700 }, { "epoch": 7.7225271768378, "grad_norm": 0.8032657504081726, "learning_rate": 0.00013698308371942676, "loss": 3.2231, "step": 71750 }, { "epoch": 7.727908728877408, "grad_norm": 0.8196533918380737, "learning_rate": 0.00013665984268936536, "loss": 3.2375, "step": 71800 }, { "epoch": 7.733290280917016, "grad_norm": 0.795331597328186, "learning_rate": 0.00013633660165930395, "loss": 3.2346, "step": 71850 }, { "epoch": 7.738671832956625, "grad_norm": 0.8460344076156616, "learning_rate": 0.00013601336062924255, "loss": 3.2446, "step": 71900 }, { "epoch": 7.744053384996233, "grad_norm": 0.862134575843811, "learning_rate": 0.00013569011959918111, "loss": 3.2438, "step": 71950 }, { "epoch": 7.749434937035841, "grad_norm": 0.8252710700035095, "learning_rate": 0.0001353668785691197, "loss": 3.228, "step": 72000 }, { "epoch": 7.749434937035841, "eval_accuracy": 0.3897550757468175, "eval_loss": 3.3368613719940186, "eval_runtime": 184.6906, "eval_samples_per_second": 97.52, "eval_steps_per_second": 6.097, "step": 72000 }, { "epoch": 7.754816489075449, "grad_norm": 0.8688374161720276, "learning_rate": 0.00013504363753905828, "loss": 3.232, "step": 72050 }, { "epoch": 7.760198041115058, "grad_norm": 0.8163436055183411, "learning_rate": 0.00013472039650899687, "loss": 3.2337, "step": 72100 }, { "epoch": 7.765579593154666, "grad_norm": 0.7999975085258484, "learning_rate": 0.00013439715547893544, "loss": 3.2324, "step": 72150 }, { "epoch": 7.770961145194274, "grad_norm": 0.8432734608650208, "learning_rate": 0.00013407391444887403, "loss": 3.2204, "step": 72200 }, { "epoch": 7.776342697233883, "grad_norm": 0.8608858585357666, "learning_rate": 0.0001337506734188126, "loss": 3.2419, "step": 72250 }, { "epoch": 7.78172424927349, "grad_norm": 0.8750514388084412, "learning_rate": 0.0001334274323887512, "loss": 3.2275, "step": 72300 }, { "epoch": 7.787105801313099, "grad_norm": 0.8133453726768494, "learning_rate": 0.0001331041913586898, "loss": 3.2265, "step": 72350 }, { "epoch": 7.792487353352707, "grad_norm": 0.8550586700439453, "learning_rate": 0.00013278095032862838, "loss": 3.2262, "step": 72400 }, { "epoch": 7.797868905392315, "grad_norm": 0.8691797256469727, "learning_rate": 0.00013245770929856695, "loss": 3.243, "step": 72450 }, { "epoch": 7.803250457431924, "grad_norm": 0.883783757686615, "learning_rate": 0.00013213446826850555, "loss": 3.2444, "step": 72500 }, { "epoch": 7.808632009471531, "grad_norm": 0.8034313917160034, "learning_rate": 0.00013181122723844411, "loss": 3.2254, "step": 72550 }, { "epoch": 7.81401356151114, "grad_norm": 0.830434262752533, "learning_rate": 0.0001314879862083827, "loss": 3.235, "step": 72600 }, { "epoch": 7.819395113550748, "grad_norm": 0.8352016806602478, "learning_rate": 0.00013116474517832128, "loss": 3.2293, "step": 72650 }, { "epoch": 7.824776665590356, "grad_norm": 0.8240856528282166, "learning_rate": 0.00013084150414825987, "loss": 3.2342, "step": 72700 }, { "epoch": 7.830158217629965, "grad_norm": 0.7969643473625183, "learning_rate": 0.00013051826311819844, "loss": 3.2313, "step": 72750 }, { "epoch": 7.835539769669572, "grad_norm": 0.8742566108703613, "learning_rate": 0.00013019502208813703, "loss": 3.2344, "step": 72800 }, { "epoch": 7.840921321709181, "grad_norm": 0.8692363500595093, "learning_rate": 0.00012987178105807563, "loss": 3.2332, "step": 72850 }, { "epoch": 7.846302873748789, "grad_norm": 0.8365722298622131, "learning_rate": 0.00012954854002801422, "loss": 3.2264, "step": 72900 }, { "epoch": 7.851684425788397, "grad_norm": 0.828915536403656, "learning_rate": 0.0001292252989979528, "loss": 3.2234, "step": 72950 }, { "epoch": 7.857065977828006, "grad_norm": 0.8495892882347107, "learning_rate": 0.00012890205796789139, "loss": 3.2466, "step": 73000 }, { "epoch": 7.857065977828006, "eval_accuracy": 0.3902502069565881, "eval_loss": 3.332028388977051, "eval_runtime": 184.8344, "eval_samples_per_second": 97.444, "eval_steps_per_second": 6.092, "step": 73000 }, { "epoch": 7.862447529867614, "grad_norm": 0.8604885339736938, "learning_rate": 0.00012857881693782998, "loss": 3.228, "step": 73050 }, { "epoch": 7.867829081907222, "grad_norm": 0.8913674354553223, "learning_rate": 0.00012825557590776855, "loss": 3.2537, "step": 73100 }, { "epoch": 7.87321063394683, "grad_norm": 0.8491577506065369, "learning_rate": 0.00012793233487770714, "loss": 3.222, "step": 73150 }, { "epoch": 7.878592185986438, "grad_norm": 0.9810826182365417, "learning_rate": 0.0001276090938476457, "loss": 3.2323, "step": 73200 }, { "epoch": 7.883973738026047, "grad_norm": 0.8327536582946777, "learning_rate": 0.0001272858528175843, "loss": 3.235, "step": 73250 }, { "epoch": 7.889355290065655, "grad_norm": 0.8036714196205139, "learning_rate": 0.00012696261178752287, "loss": 3.2236, "step": 73300 }, { "epoch": 7.894736842105263, "grad_norm": 0.8222293853759766, "learning_rate": 0.00012663937075746147, "loss": 3.2249, "step": 73350 }, { "epoch": 7.900118394144871, "grad_norm": 0.8565043210983276, "learning_rate": 0.00012631612972740006, "loss": 3.2353, "step": 73400 }, { "epoch": 7.90549994618448, "grad_norm": 0.8443336486816406, "learning_rate": 0.00012599288869733863, "loss": 3.2214, "step": 73450 }, { "epoch": 7.910881498224088, "grad_norm": 0.9020589590072632, "learning_rate": 0.00012566964766727722, "loss": 3.2482, "step": 73500 }, { "epoch": 7.916263050263696, "grad_norm": 0.8561992049217224, "learning_rate": 0.00012534640663721582, "loss": 3.2166, "step": 73550 }, { "epoch": 7.921644602303305, "grad_norm": 0.8745976090431213, "learning_rate": 0.00012502316560715439, "loss": 3.2524, "step": 73600 }, { "epoch": 7.927026154342912, "grad_norm": 0.8272269368171692, "learning_rate": 0.00012469992457709298, "loss": 3.2354, "step": 73650 }, { "epoch": 7.932407706382521, "grad_norm": 0.9126244783401489, "learning_rate": 0.00012437668354703158, "loss": 3.2354, "step": 73700 }, { "epoch": 7.937789258422129, "grad_norm": 0.8359974026679993, "learning_rate": 0.00012405344251697014, "loss": 3.2412, "step": 73750 }, { "epoch": 7.943170810461737, "grad_norm": 0.8640374541282654, "learning_rate": 0.00012373020148690874, "loss": 3.2426, "step": 73800 }, { "epoch": 7.948552362501346, "grad_norm": 0.878259003162384, "learning_rate": 0.0001234069604568473, "loss": 3.223, "step": 73850 }, { "epoch": 7.953933914540953, "grad_norm": 0.9081090688705444, "learning_rate": 0.0001230837194267859, "loss": 3.2481, "step": 73900 }, { "epoch": 7.959315466580562, "grad_norm": 0.8142951130867004, "learning_rate": 0.00012276047839672447, "loss": 3.2287, "step": 73950 }, { "epoch": 7.96469701862017, "grad_norm": 0.8311217427253723, "learning_rate": 0.00012243723736666306, "loss": 3.2126, "step": 74000 }, { "epoch": 7.96469701862017, "eval_accuracy": 0.39052531606853436, "eval_loss": 3.3307554721832275, "eval_runtime": 184.6893, "eval_samples_per_second": 97.521, "eval_steps_per_second": 6.097, "step": 74000 }, { "epoch": 7.970078570659778, "grad_norm": 0.8098016381263733, "learning_rate": 0.00012211399633660166, "loss": 3.2278, "step": 74050 }, { "epoch": 7.975460122699387, "grad_norm": 0.8477680683135986, "learning_rate": 0.00012179075530654022, "loss": 3.2304, "step": 74100 }, { "epoch": 7.980841674738995, "grad_norm": 0.825594961643219, "learning_rate": 0.00012146751427647882, "loss": 3.2386, "step": 74150 }, { "epoch": 7.986223226778603, "grad_norm": 0.8377440571784973, "learning_rate": 0.0001211442732464174, "loss": 3.2335, "step": 74200 }, { "epoch": 7.991604778818211, "grad_norm": 0.8370752930641174, "learning_rate": 0.00012082103221635598, "loss": 3.2272, "step": 74250 }, { "epoch": 7.996986330857819, "grad_norm": 0.8993654251098633, "learning_rate": 0.00012049779118629456, "loss": 3.2465, "step": 74300 }, { "epoch": 8.002367882897428, "grad_norm": 0.8433480858802795, "learning_rate": 0.0001201810149768344, "loss": 3.1976, "step": 74350 }, { "epoch": 8.007749434937036, "grad_norm": 0.8284233808517456, "learning_rate": 0.00011985777394677296, "loss": 3.1514, "step": 74400 }, { "epoch": 8.013130986976645, "grad_norm": 0.8333739638328552, "learning_rate": 0.00011953453291671154, "loss": 3.1506, "step": 74450 }, { "epoch": 8.018512539016251, "grad_norm": 0.8498620986938477, "learning_rate": 0.00011921129188665014, "loss": 3.1347, "step": 74500 }, { "epoch": 8.02389409105586, "grad_norm": 0.8060201406478882, "learning_rate": 0.0001188880508565887, "loss": 3.1446, "step": 74550 }, { "epoch": 8.029275643095469, "grad_norm": 0.8369296789169312, "learning_rate": 0.0001185648098265273, "loss": 3.1747, "step": 74600 }, { "epoch": 8.034657195135077, "grad_norm": 0.8029821515083313, "learning_rate": 0.0001182415687964659, "loss": 3.1674, "step": 74650 }, { "epoch": 8.040038747174686, "grad_norm": 0.8536189198493958, "learning_rate": 0.00011791832776640448, "loss": 3.1492, "step": 74700 }, { "epoch": 8.045420299214294, "grad_norm": 0.8600338101387024, "learning_rate": 0.00011759508673634306, "loss": 3.1523, "step": 74750 }, { "epoch": 8.050801851253901, "grad_norm": 0.838543713092804, "learning_rate": 0.00011727184570628164, "loss": 3.1805, "step": 74800 }, { "epoch": 8.05618340329351, "grad_norm": 0.7800434827804565, "learning_rate": 0.00011694860467622023, "loss": 3.172, "step": 74850 }, { "epoch": 8.061564955333118, "grad_norm": 0.8202197551727295, "learning_rate": 0.0001166253636461588, "loss": 3.1607, "step": 74900 }, { "epoch": 8.066946507372727, "grad_norm": 0.8486565947532654, "learning_rate": 0.0001163021226160974, "loss": 3.1748, "step": 74950 }, { "epoch": 8.072328059412335, "grad_norm": 0.9214633107185364, "learning_rate": 0.00011597888158603598, "loss": 3.1488, "step": 75000 }, { "epoch": 8.072328059412335, "eval_accuracy": 0.3903842846201511, "eval_loss": 3.3340110778808594, "eval_runtime": 185.0586, "eval_samples_per_second": 97.326, "eval_steps_per_second": 6.085, "step": 75000 }, { "epoch": 8.077709611451942, "grad_norm": 0.868916392326355, "learning_rate": 0.00011565564055597456, "loss": 3.1697, "step": 75050 }, { "epoch": 8.08309116349155, "grad_norm": 0.8408175706863403, "learning_rate": 0.00011533239952591314, "loss": 3.1719, "step": 75100 }, { "epoch": 8.088472715531159, "grad_norm": 0.8719428777694702, "learning_rate": 0.00011500915849585173, "loss": 3.182, "step": 75150 }, { "epoch": 8.093854267570768, "grad_norm": 0.7724485397338867, "learning_rate": 0.00011468591746579033, "loss": 3.1634, "step": 75200 }, { "epoch": 8.099235819610376, "grad_norm": 0.8824598789215088, "learning_rate": 0.0001143626764357289, "loss": 3.1758, "step": 75250 }, { "epoch": 8.104617371649983, "grad_norm": 0.8987236022949219, "learning_rate": 0.00011403943540566749, "loss": 3.1593, "step": 75300 }, { "epoch": 8.109998923689592, "grad_norm": 0.8450270295143127, "learning_rate": 0.00011371619437560607, "loss": 3.164, "step": 75350 }, { "epoch": 8.1153804757292, "grad_norm": 0.8858851790428162, "learning_rate": 0.00011339295334554464, "loss": 3.1648, "step": 75400 }, { "epoch": 8.120762027768809, "grad_norm": 0.83014315366745, "learning_rate": 0.00011306971231548323, "loss": 3.1557, "step": 75450 }, { "epoch": 8.126143579808417, "grad_norm": 0.8151319622993469, "learning_rate": 0.00011274647128542183, "loss": 3.1569, "step": 75500 }, { "epoch": 8.131525131848026, "grad_norm": 0.8844127655029297, "learning_rate": 0.0001124232302553604, "loss": 3.1876, "step": 75550 }, { "epoch": 8.136906683887632, "grad_norm": 0.8928391933441162, "learning_rate": 0.00011209998922529899, "loss": 3.1632, "step": 75600 }, { "epoch": 8.142288235927241, "grad_norm": 0.844195544719696, "learning_rate": 0.00011177674819523757, "loss": 3.1593, "step": 75650 }, { "epoch": 8.14766978796685, "grad_norm": 0.8816542625427246, "learning_rate": 0.00011145350716517617, "loss": 3.1543, "step": 75700 }, { "epoch": 8.153051340006458, "grad_norm": 0.917953372001648, "learning_rate": 0.00011113026613511473, "loss": 3.1802, "step": 75750 }, { "epoch": 8.158432892046067, "grad_norm": 0.8676404356956482, "learning_rate": 0.00011081348992565455, "loss": 3.1598, "step": 75800 }, { "epoch": 8.163814444085673, "grad_norm": 0.8764878511428833, "learning_rate": 0.00011049024889559314, "loss": 3.1617, "step": 75850 }, { "epoch": 8.169195996125282, "grad_norm": 0.8242963552474976, "learning_rate": 0.00011016700786553172, "loss": 3.1673, "step": 75900 }, { "epoch": 8.17457754816489, "grad_norm": 0.8988597989082336, "learning_rate": 0.00010984376683547031, "loss": 3.181, "step": 75950 }, { "epoch": 8.1799591002045, "grad_norm": 0.8427569270133972, "learning_rate": 0.00010952052580540889, "loss": 3.1626, "step": 76000 }, { "epoch": 8.1799591002045, "eval_accuracy": 0.3904491503941764, "eval_loss": 3.3352701663970947, "eval_runtime": 184.661, "eval_samples_per_second": 97.535, "eval_steps_per_second": 6.098, "step": 76000 }, { "epoch": 8.185340652244108, "grad_norm": 0.8113105297088623, "learning_rate": 0.00010919728477534747, "loss": 3.1803, "step": 76050 }, { "epoch": 8.190722204283716, "grad_norm": 0.8857055902481079, "learning_rate": 0.00010887404374528605, "loss": 3.1951, "step": 76100 }, { "epoch": 8.196103756323323, "grad_norm": 0.8819203972816467, "learning_rate": 0.00010855080271522465, "loss": 3.157, "step": 76150 }, { "epoch": 8.201485308362932, "grad_norm": 0.920045793056488, "learning_rate": 0.00010822756168516322, "loss": 3.1773, "step": 76200 }, { "epoch": 8.20686686040254, "grad_norm": 0.9266934990882874, "learning_rate": 0.00010790432065510181, "loss": 3.18, "step": 76250 }, { "epoch": 8.212248412442149, "grad_norm": 0.8361415266990662, "learning_rate": 0.0001075810796250404, "loss": 3.1652, "step": 76300 }, { "epoch": 8.217629964481757, "grad_norm": 0.9125852584838867, "learning_rate": 0.00010725783859497897, "loss": 3.1847, "step": 76350 }, { "epoch": 8.223011516521364, "grad_norm": 0.8211091160774231, "learning_rate": 0.00010693459756491757, "loss": 3.1481, "step": 76400 }, { "epoch": 8.228393068560973, "grad_norm": 0.8206071257591248, "learning_rate": 0.00010661135653485615, "loss": 3.1628, "step": 76450 }, { "epoch": 8.233774620600581, "grad_norm": 0.8912805318832397, "learning_rate": 0.00010628811550479474, "loss": 3.1706, "step": 76500 }, { "epoch": 8.23915617264019, "grad_norm": 0.8347760438919067, "learning_rate": 0.00010596487447473331, "loss": 3.1822, "step": 76550 }, { "epoch": 8.244537724679798, "grad_norm": 0.8456274271011353, "learning_rate": 0.0001056416334446719, "loss": 3.1729, "step": 76600 }, { "epoch": 8.249919276719407, "grad_norm": 0.8533112406730652, "learning_rate": 0.00010531839241461049, "loss": 3.1695, "step": 76650 }, { "epoch": 8.255300828759013, "grad_norm": 0.8624753952026367, "learning_rate": 0.00010499515138454907, "loss": 3.1648, "step": 76700 }, { "epoch": 8.260682380798622, "grad_norm": 0.8519622683525085, "learning_rate": 0.00010467191035448765, "loss": 3.1798, "step": 76750 }, { "epoch": 8.26606393283823, "grad_norm": 0.838147759437561, "learning_rate": 0.00010434866932442624, "loss": 3.1683, "step": 76800 }, { "epoch": 8.27144548487784, "grad_norm": 0.903648853302002, "learning_rate": 0.00010402542829436481, "loss": 3.1834, "step": 76850 }, { "epoch": 8.276827036917448, "grad_norm": 0.8623703718185425, "learning_rate": 0.0001037021872643034, "loss": 3.1707, "step": 76900 }, { "epoch": 8.282208588957054, "grad_norm": 0.8603115081787109, "learning_rate": 0.00010337894623424199, "loss": 3.1824, "step": 76950 }, { "epoch": 8.287590140996663, "grad_norm": 0.8683320879936218, "learning_rate": 0.00010305570520418058, "loss": 3.1706, "step": 77000 }, { "epoch": 8.287590140996663, "eval_accuracy": 0.39092265967927187, "eval_loss": 3.3312289714813232, "eval_runtime": 184.7891, "eval_samples_per_second": 97.468, "eval_steps_per_second": 6.093, "step": 77000 }, { "epoch": 8.292971693036272, "grad_norm": 0.8868846893310547, "learning_rate": 0.00010273246417411915, "loss": 3.1657, "step": 77050 }, { "epoch": 8.29835324507588, "grad_norm": 0.835442304611206, "learning_rate": 0.00010240922314405774, "loss": 3.1762, "step": 77100 }, { "epoch": 8.303734797115489, "grad_norm": 0.8506765961647034, "learning_rate": 0.00010208598211399634, "loss": 3.1836, "step": 77150 }, { "epoch": 8.309116349155097, "grad_norm": 0.9037638306617737, "learning_rate": 0.0001017627410839349, "loss": 3.1595, "step": 77200 }, { "epoch": 8.314497901194704, "grad_norm": 0.8438506722450256, "learning_rate": 0.0001014395000538735, "loss": 3.1774, "step": 77250 }, { "epoch": 8.319879453234313, "grad_norm": 0.8813534379005432, "learning_rate": 0.00010111625902381208, "loss": 3.1947, "step": 77300 }, { "epoch": 8.325261005273921, "grad_norm": 0.8885173797607422, "learning_rate": 0.00010079301799375066, "loss": 3.1824, "step": 77350 }, { "epoch": 8.33064255731353, "grad_norm": 0.8622186183929443, "learning_rate": 0.00010046977696368924, "loss": 3.1645, "step": 77400 }, { "epoch": 8.336024109353138, "grad_norm": 0.891852080821991, "learning_rate": 0.00010014653593362784, "loss": 3.1819, "step": 77450 }, { "epoch": 8.341405661392745, "grad_norm": 0.8418218493461609, "learning_rate": 9.982329490356642e-05, "loss": 3.1684, "step": 77500 }, { "epoch": 8.346787213432354, "grad_norm": 0.8554936051368713, "learning_rate": 9.9500053873505e-05, "loss": 3.1878, "step": 77550 }, { "epoch": 8.352168765471962, "grad_norm": 0.854648232460022, "learning_rate": 9.917681284344358e-05, "loss": 3.1779, "step": 77600 }, { "epoch": 8.35755031751157, "grad_norm": 0.8960241079330444, "learning_rate": 9.885357181338218e-05, "loss": 3.1708, "step": 77650 }, { "epoch": 8.36293186955118, "grad_norm": 0.9255419969558716, "learning_rate": 9.853033078332074e-05, "loss": 3.1909, "step": 77700 }, { "epoch": 8.368313421590786, "grad_norm": 0.8990132808685303, "learning_rate": 9.820708975325934e-05, "loss": 3.1713, "step": 77750 }, { "epoch": 8.373694973630395, "grad_norm": 0.85789954662323, "learning_rate": 9.788384872319793e-05, "loss": 3.1605, "step": 77800 }, { "epoch": 8.379076525670003, "grad_norm": 0.858098030090332, "learning_rate": 9.75606076931365e-05, "loss": 3.1687, "step": 77850 }, { "epoch": 8.384458077709612, "grad_norm": 0.8912584781646729, "learning_rate": 9.723736666307508e-05, "loss": 3.1633, "step": 77900 }, { "epoch": 8.38983962974922, "grad_norm": 0.8760279417037964, "learning_rate": 9.691412563301368e-05, "loss": 3.163, "step": 77950 }, { "epoch": 8.395221181788829, "grad_norm": 0.8484280705451965, "learning_rate": 9.659088460295227e-05, "loss": 3.1875, "step": 78000 }, { "epoch": 8.395221181788829, "eval_accuracy": 0.39122645315360116, "eval_loss": 3.326995372772217, "eval_runtime": 184.8092, "eval_samples_per_second": 97.457, "eval_steps_per_second": 6.093, "step": 78000 }, { "epoch": 8.400602733828435, "grad_norm": 0.8364370465278625, "learning_rate": 9.626764357289084e-05, "loss": 3.1871, "step": 78050 }, { "epoch": 8.405984285868044, "grad_norm": 0.8873591423034668, "learning_rate": 9.594440254282943e-05, "loss": 3.1791, "step": 78100 }, { "epoch": 8.411365837907653, "grad_norm": 0.8371385335922241, "learning_rate": 9.562116151276802e-05, "loss": 3.1628, "step": 78150 }, { "epoch": 8.416747389947261, "grad_norm": 0.8212653994560242, "learning_rate": 9.52979204827066e-05, "loss": 3.1585, "step": 78200 }, { "epoch": 8.42212894198687, "grad_norm": 0.8428384065628052, "learning_rate": 9.497467945264518e-05, "loss": 3.2135, "step": 78250 }, { "epoch": 8.427510494026476, "grad_norm": 0.8416067361831665, "learning_rate": 9.465143842258377e-05, "loss": 3.1811, "step": 78300 }, { "epoch": 8.432892046066085, "grad_norm": 0.8841109871864319, "learning_rate": 9.432819739252234e-05, "loss": 3.1834, "step": 78350 }, { "epoch": 8.438273598105694, "grad_norm": 0.8793404698371887, "learning_rate": 9.400495636246093e-05, "loss": 3.1823, "step": 78400 }, { "epoch": 8.443655150145302, "grad_norm": 0.8999808430671692, "learning_rate": 9.368171533239952e-05, "loss": 3.1813, "step": 78450 }, { "epoch": 8.44903670218491, "grad_norm": 0.832278847694397, "learning_rate": 9.335847430233811e-05, "loss": 3.1914, "step": 78500 }, { "epoch": 8.45441825422452, "grad_norm": 0.8846303820610046, "learning_rate": 9.303523327227668e-05, "loss": 3.1763, "step": 78550 }, { "epoch": 8.459799806264126, "grad_norm": 0.8395057320594788, "learning_rate": 9.271199224221527e-05, "loss": 3.1892, "step": 78600 }, { "epoch": 8.465181358303735, "grad_norm": 0.8631083369255066, "learning_rate": 9.238875121215387e-05, "loss": 3.1991, "step": 78650 }, { "epoch": 8.470562910343343, "grad_norm": 0.8669479489326477, "learning_rate": 9.206551018209243e-05, "loss": 3.1691, "step": 78700 }, { "epoch": 8.475944462382952, "grad_norm": 0.8604099154472351, "learning_rate": 9.174226915203103e-05, "loss": 3.1691, "step": 78750 }, { "epoch": 8.48132601442256, "grad_norm": 0.8545402884483337, "learning_rate": 9.141902812196961e-05, "loss": 3.1751, "step": 78800 }, { "epoch": 8.486707566462167, "grad_norm": 0.8492422699928284, "learning_rate": 9.109578709190818e-05, "loss": 3.1837, "step": 78850 }, { "epoch": 8.492089118501776, "grad_norm": 0.8209715485572815, "learning_rate": 9.077254606184677e-05, "loss": 3.1882, "step": 78900 }, { "epoch": 8.497470670541384, "grad_norm": 0.876168429851532, "learning_rate": 9.044930503178537e-05, "loss": 3.1854, "step": 78950 }, { "epoch": 8.502852222580993, "grad_norm": 0.9051874876022339, "learning_rate": 9.012606400172395e-05, "loss": 3.1783, "step": 79000 }, { "epoch": 8.502852222580993, "eval_accuracy": 0.3914579924575338, "eval_loss": 3.3256032466888428, "eval_runtime": 185.3855, "eval_samples_per_second": 97.154, "eval_steps_per_second": 6.074, "step": 79000 }, { "epoch": 8.508233774620601, "grad_norm": 0.9354413747787476, "learning_rate": 8.980282297166253e-05, "loss": 3.1785, "step": 79050 }, { "epoch": 8.513615326660208, "grad_norm": 0.8846203684806824, "learning_rate": 8.947958194160111e-05, "loss": 3.2052, "step": 79100 }, { "epoch": 8.518996878699816, "grad_norm": 0.8579129576683044, "learning_rate": 8.91563409115397e-05, "loss": 3.1818, "step": 79150 }, { "epoch": 8.524378430739425, "grad_norm": 0.8579406142234802, "learning_rate": 8.883309988147827e-05, "loss": 3.1827, "step": 79200 }, { "epoch": 8.529759982779034, "grad_norm": 0.8814583420753479, "learning_rate": 8.850985885141687e-05, "loss": 3.1749, "step": 79250 }, { "epoch": 8.535141534818642, "grad_norm": 0.8940314054489136, "learning_rate": 8.818661782135545e-05, "loss": 3.1611, "step": 79300 }, { "epoch": 8.54052308685825, "grad_norm": 0.9121882319450378, "learning_rate": 8.786337679129403e-05, "loss": 3.1679, "step": 79350 }, { "epoch": 8.545904638897857, "grad_norm": 0.9019620418548584, "learning_rate": 8.754013576123261e-05, "loss": 3.1741, "step": 79400 }, { "epoch": 8.551286190937466, "grad_norm": 0.9139887690544128, "learning_rate": 8.72168947311712e-05, "loss": 3.1732, "step": 79450 }, { "epoch": 8.556667742977075, "grad_norm": 0.8666207790374756, "learning_rate": 8.68936537011098e-05, "loss": 3.1821, "step": 79500 }, { "epoch": 8.562049295016683, "grad_norm": 0.8577825427055359, "learning_rate": 8.657041267104837e-05, "loss": 3.1826, "step": 79550 }, { "epoch": 8.567430847056292, "grad_norm": 0.8668467402458191, "learning_rate": 8.624717164098696e-05, "loss": 3.1725, "step": 79600 }, { "epoch": 8.572812399095898, "grad_norm": 0.8373755216598511, "learning_rate": 8.592393061092554e-05, "loss": 3.1797, "step": 79650 }, { "epoch": 8.578193951135507, "grad_norm": 0.8314027786254883, "learning_rate": 8.560068958086412e-05, "loss": 3.1849, "step": 79700 }, { "epoch": 8.583575503175116, "grad_norm": 0.9058263897895813, "learning_rate": 8.52774485508027e-05, "loss": 3.176, "step": 79750 }, { "epoch": 8.588957055214724, "grad_norm": 0.9133097529411316, "learning_rate": 8.49542075207413e-05, "loss": 3.1968, "step": 79800 }, { "epoch": 8.594338607254333, "grad_norm": 0.8769819140434265, "learning_rate": 8.46374313112811e-05, "loss": 3.1829, "step": 79850 }, { "epoch": 8.599720159293941, "grad_norm": 1.0664840936660767, "learning_rate": 8.431419028121969e-05, "loss": 3.1828, "step": 79900 }, { "epoch": 8.605101711333548, "grad_norm": 0.8446348905563354, "learning_rate": 8.399094925115828e-05, "loss": 3.1563, "step": 79950 }, { "epoch": 8.610483263373157, "grad_norm": 0.8764998912811279, "learning_rate": 8.366770822109685e-05, "loss": 3.1571, "step": 80000 }, { "epoch": 8.610483263373157, "eval_accuracy": 0.39208394174423306, "eval_loss": 3.321202039718628, "eval_runtime": 184.6735, "eval_samples_per_second": 97.529, "eval_steps_per_second": 6.097, "step": 80000 }, { "epoch": 8.615864815412765, "grad_norm": 0.8480128049850464, "learning_rate": 8.334446719103544e-05, "loss": 3.189, "step": 80050 }, { "epoch": 8.621246367452374, "grad_norm": 0.8969175219535828, "learning_rate": 8.302122616097403e-05, "loss": 3.1913, "step": 80100 }, { "epoch": 8.626627919491982, "grad_norm": 0.8639907240867615, "learning_rate": 8.269798513091261e-05, "loss": 3.1644, "step": 80150 }, { "epoch": 8.632009471531589, "grad_norm": 0.8794361352920532, "learning_rate": 8.237474410085119e-05, "loss": 3.1668, "step": 80200 }, { "epoch": 8.637391023571197, "grad_norm": 0.885933518409729, "learning_rate": 8.205150307078978e-05, "loss": 3.1819, "step": 80250 }, { "epoch": 8.642772575610806, "grad_norm": 0.8281596302986145, "learning_rate": 8.172826204072838e-05, "loss": 3.1691, "step": 80300 }, { "epoch": 8.648154127650415, "grad_norm": 0.9188771843910217, "learning_rate": 8.140502101066694e-05, "loss": 3.1883, "step": 80350 }, { "epoch": 8.653535679690023, "grad_norm": 0.9139700531959534, "learning_rate": 8.108177998060553e-05, "loss": 3.1712, "step": 80400 }, { "epoch": 8.658917231729632, "grad_norm": 0.8853664398193359, "learning_rate": 8.075853895054412e-05, "loss": 3.1819, "step": 80450 }, { "epoch": 8.664298783769238, "grad_norm": 0.8718395233154297, "learning_rate": 8.043529792048269e-05, "loss": 3.1864, "step": 80500 }, { "epoch": 8.669680335808847, "grad_norm": 0.8565669655799866, "learning_rate": 8.011205689042128e-05, "loss": 3.1692, "step": 80550 }, { "epoch": 8.675061887848456, "grad_norm": 0.8765687942504883, "learning_rate": 7.978881586035988e-05, "loss": 3.2026, "step": 80600 }, { "epoch": 8.680443439888064, "grad_norm": 0.8425050973892212, "learning_rate": 7.946557483029845e-05, "loss": 3.1826, "step": 80650 }, { "epoch": 8.685824991927673, "grad_norm": 0.8891403079032898, "learning_rate": 7.914233380023704e-05, "loss": 3.1744, "step": 80700 }, { "epoch": 8.69120654396728, "grad_norm": 0.8497437834739685, "learning_rate": 7.881909277017562e-05, "loss": 3.1793, "step": 80750 }, { "epoch": 8.696588096006888, "grad_norm": 0.8760660886764526, "learning_rate": 7.849585174011422e-05, "loss": 3.1883, "step": 80800 }, { "epoch": 8.701969648046497, "grad_norm": 0.9156056642532349, "learning_rate": 7.817261071005278e-05, "loss": 3.1886, "step": 80850 }, { "epoch": 8.707351200086105, "grad_norm": 0.9590579867362976, "learning_rate": 7.784936967999138e-05, "loss": 3.1879, "step": 80900 }, { "epoch": 8.712732752125714, "grad_norm": 0.8654018044471741, "learning_rate": 7.752612864992996e-05, "loss": 3.1743, "step": 80950 }, { "epoch": 8.718114304165322, "grad_norm": 1.0002062320709229, "learning_rate": 7.720288761986854e-05, "loss": 3.2066, "step": 81000 }, { "epoch": 8.718114304165322, "eval_accuracy": 0.39225344024922365, "eval_loss": 3.317610025405884, "eval_runtime": 184.756, "eval_samples_per_second": 97.485, "eval_steps_per_second": 6.095, "step": 81000 }, { "epoch": 8.723495856204929, "grad_norm": 0.8658928871154785, "learning_rate": 7.687964658980712e-05, "loss": 3.1837, "step": 81050 }, { "epoch": 8.728877408244538, "grad_norm": 0.8824866414070129, "learning_rate": 7.655640555974572e-05, "loss": 3.1794, "step": 81100 }, { "epoch": 8.734258960284146, "grad_norm": 0.9138555526733398, "learning_rate": 7.623316452968428e-05, "loss": 3.1763, "step": 81150 }, { "epoch": 8.739640512323755, "grad_norm": 0.8968687653541565, "learning_rate": 7.590992349962288e-05, "loss": 3.1867, "step": 81200 }, { "epoch": 8.745022064363363, "grad_norm": 0.8379364609718323, "learning_rate": 7.558668246956147e-05, "loss": 3.1788, "step": 81250 }, { "epoch": 8.75040361640297, "grad_norm": 0.8813756108283997, "learning_rate": 7.526344143950005e-05, "loss": 3.1934, "step": 81300 }, { "epoch": 8.755785168442578, "grad_norm": 0.8551928400993347, "learning_rate": 7.494020040943862e-05, "loss": 3.1963, "step": 81350 }, { "epoch": 8.761166720482187, "grad_norm": 0.9404810667037964, "learning_rate": 7.461695937937722e-05, "loss": 3.1712, "step": 81400 }, { "epoch": 8.766548272521796, "grad_norm": 0.9553894400596619, "learning_rate": 7.42937183493158e-05, "loss": 3.1956, "step": 81450 }, { "epoch": 8.771929824561404, "grad_norm": 0.8532766103744507, "learning_rate": 7.397047731925439e-05, "loss": 3.1838, "step": 81500 }, { "epoch": 8.777311376601011, "grad_norm": 0.8783697485923767, "learning_rate": 7.364723628919297e-05, "loss": 3.1803, "step": 81550 }, { "epoch": 8.78269292864062, "grad_norm": 0.8748445510864258, "learning_rate": 7.332399525913155e-05, "loss": 3.1718, "step": 81600 }, { "epoch": 8.788074480680228, "grad_norm": 0.8972441554069519, "learning_rate": 7.300075422907013e-05, "loss": 3.1777, "step": 81650 }, { "epoch": 8.793456032719837, "grad_norm": 0.8829938173294067, "learning_rate": 7.267751319900872e-05, "loss": 3.1816, "step": 81700 }, { "epoch": 8.798837584759445, "grad_norm": 0.8357025980949402, "learning_rate": 7.235427216894731e-05, "loss": 3.1809, "step": 81750 }, { "epoch": 8.804219136799054, "grad_norm": 0.8655660152435303, "learning_rate": 7.203103113888589e-05, "loss": 3.1801, "step": 81800 }, { "epoch": 8.80960068883866, "grad_norm": 0.8427842855453491, "learning_rate": 7.17142549294257e-05, "loss": 3.1963, "step": 81850 }, { "epoch": 8.814982240878269, "grad_norm": 0.9213821887969971, "learning_rate": 7.139101389936428e-05, "loss": 3.1815, "step": 81900 }, { "epoch": 8.820363792917878, "grad_norm": 0.8830085396766663, "learning_rate": 7.10742376899041e-05, "loss": 3.1861, "step": 81950 }, { "epoch": 8.825745344957486, "grad_norm": 0.8858305811882019, "learning_rate": 7.075099665984268e-05, "loss": 3.1893, "step": 82000 }, { "epoch": 8.825745344957486, "eval_accuracy": 0.3927499839465358, "eval_loss": 3.313309669494629, "eval_runtime": 184.9366, "eval_samples_per_second": 97.39, "eval_steps_per_second": 6.089, "step": 82000 }, { "epoch": 8.831126896997095, "grad_norm": 0.898823082447052, "learning_rate": 7.042775562978126e-05, "loss": 3.1939, "step": 82050 }, { "epoch": 8.836508449036701, "grad_norm": 0.8878447413444519, "learning_rate": 7.010451459971986e-05, "loss": 3.1949, "step": 82100 }, { "epoch": 8.84189000107631, "grad_norm": 0.853378415107727, "learning_rate": 6.978127356965844e-05, "loss": 3.1804, "step": 82150 }, { "epoch": 8.847271553115919, "grad_norm": 0.8771511316299438, "learning_rate": 6.945803253959702e-05, "loss": 3.1889, "step": 82200 }, { "epoch": 8.852653105155527, "grad_norm": 0.8539415597915649, "learning_rate": 6.91347915095356e-05, "loss": 3.1983, "step": 82250 }, { "epoch": 8.858034657195136, "grad_norm": 0.8989981412887573, "learning_rate": 6.881155047947418e-05, "loss": 3.1973, "step": 82300 }, { "epoch": 8.863416209234742, "grad_norm": 0.8486732840538025, "learning_rate": 6.848830944941278e-05, "loss": 3.1764, "step": 82350 }, { "epoch": 8.868797761274351, "grad_norm": 0.862515389919281, "learning_rate": 6.816506841935136e-05, "loss": 3.1813, "step": 82400 }, { "epoch": 8.87417931331396, "grad_norm": 0.8676425814628601, "learning_rate": 6.784182738928994e-05, "loss": 3.1833, "step": 82450 }, { "epoch": 8.879560865353568, "grad_norm": 0.8382507562637329, "learning_rate": 6.751858635922853e-05, "loss": 3.1977, "step": 82500 }, { "epoch": 8.884942417393177, "grad_norm": 0.9167499542236328, "learning_rate": 6.719534532916711e-05, "loss": 3.2017, "step": 82550 }, { "epoch": 8.890323969432785, "grad_norm": 0.9321362972259521, "learning_rate": 6.68721042991057e-05, "loss": 3.1856, "step": 82600 }, { "epoch": 8.895705521472392, "grad_norm": 0.8884369134902954, "learning_rate": 6.654886326904428e-05, "loss": 3.166, "step": 82650 }, { "epoch": 8.901087073512, "grad_norm": 0.8399180769920349, "learning_rate": 6.622562223898286e-05, "loss": 3.1608, "step": 82700 }, { "epoch": 8.906468625551609, "grad_norm": 0.9197646379470825, "learning_rate": 6.590238120892145e-05, "loss": 3.19, "step": 82750 }, { "epoch": 8.911850177591218, "grad_norm": 0.8913666605949402, "learning_rate": 6.557914017886003e-05, "loss": 3.1812, "step": 82800 }, { "epoch": 8.917231729630826, "grad_norm": 0.9225286841392517, "learning_rate": 6.525589914879861e-05, "loss": 3.1826, "step": 82850 }, { "epoch": 8.922613281670433, "grad_norm": 0.8732419610023499, "learning_rate": 6.49326581187372e-05, "loss": 3.1804, "step": 82900 }, { "epoch": 8.927994833710041, "grad_norm": 0.9099151492118835, "learning_rate": 6.460941708867578e-05, "loss": 3.1793, "step": 82950 }, { "epoch": 8.93337638574965, "grad_norm": 0.8944827318191528, "learning_rate": 6.428617605861437e-05, "loss": 3.197, "step": 83000 }, { "epoch": 8.93337638574965, "eval_accuracy": 0.39288395295721107, "eval_loss": 3.311122417449951, "eval_runtime": 185.1062, "eval_samples_per_second": 97.301, "eval_steps_per_second": 6.083, "step": 83000 }, { "epoch": 8.938757937789259, "grad_norm": 0.9626566171646118, "learning_rate": 6.396293502855295e-05, "loss": 3.1786, "step": 83050 }, { "epoch": 8.944139489828867, "grad_norm": 0.8564572334289551, "learning_rate": 6.363969399849153e-05, "loss": 3.1572, "step": 83100 }, { "epoch": 8.949521041868476, "grad_norm": 0.8902326822280884, "learning_rate": 6.331645296843011e-05, "loss": 3.1862, "step": 83150 }, { "epoch": 8.954902593908082, "grad_norm": 0.8501969575881958, "learning_rate": 6.29932119383687e-05, "loss": 3.1809, "step": 83200 }, { "epoch": 8.960284145947691, "grad_norm": 0.8566688299179077, "learning_rate": 6.266997090830729e-05, "loss": 3.1817, "step": 83250 }, { "epoch": 8.9656656979873, "grad_norm": 0.8682833909988403, "learning_rate": 6.234672987824587e-05, "loss": 3.1946, "step": 83300 }, { "epoch": 8.971047250026908, "grad_norm": 0.9486375451087952, "learning_rate": 6.202348884818447e-05, "loss": 3.1855, "step": 83350 }, { "epoch": 8.976428802066517, "grad_norm": 0.8489530086517334, "learning_rate": 6.170024781812305e-05, "loss": 3.181, "step": 83400 }, { "epoch": 8.981810354106123, "grad_norm": 0.8610159158706665, "learning_rate": 6.137700678806163e-05, "loss": 3.1787, "step": 83450 }, { "epoch": 8.987191906145732, "grad_norm": 0.926662802696228, "learning_rate": 6.105376575800021e-05, "loss": 3.1692, "step": 83500 }, { "epoch": 8.99257345818534, "grad_norm": 0.9315507411956787, "learning_rate": 6.073052472793879e-05, "loss": 3.1682, "step": 83550 }, { "epoch": 8.997955010224949, "grad_norm": 0.9029734134674072, "learning_rate": 6.0407283697877384e-05, "loss": 3.1876, "step": 83600 }, { "epoch": 9.003336562264558, "grad_norm": 0.9033173322677612, "learning_rate": 6.0084042667815966e-05, "loss": 3.1469, "step": 83650 }, { "epoch": 9.008718114304166, "grad_norm": 0.8773213028907776, "learning_rate": 5.976080163775455e-05, "loss": 3.1173, "step": 83700 }, { "epoch": 9.014099666343773, "grad_norm": 0.9068469405174255, "learning_rate": 5.9437560607693135e-05, "loss": 3.1375, "step": 83750 }, { "epoch": 9.019481218383381, "grad_norm": 0.8826611042022705, "learning_rate": 5.9114319577631716e-05, "loss": 3.0867, "step": 83800 }, { "epoch": 9.02486277042299, "grad_norm": 0.8884468674659729, "learning_rate": 5.8791078547570304e-05, "loss": 3.1317, "step": 83850 }, { "epoch": 9.030244322462599, "grad_norm": 0.8935588002204895, "learning_rate": 5.8467837517508885e-05, "loss": 3.1211, "step": 83900 }, { "epoch": 9.035625874502207, "grad_norm": 0.8488547205924988, "learning_rate": 5.8144596487447466e-05, "loss": 3.1253, "step": 83950 }, { "epoch": 9.041007426541814, "grad_norm": 0.9119958877563477, "learning_rate": 5.7821355457386054e-05, "loss": 3.1188, "step": 84000 }, { "epoch": 9.041007426541814, "eval_accuracy": 0.3928714578751124, "eval_loss": 3.3146016597747803, "eval_runtime": 184.5536, "eval_samples_per_second": 97.592, "eval_steps_per_second": 6.101, "step": 84000 }, { "epoch": 9.046388978581422, "grad_norm": 0.8722310662269592, "learning_rate": 5.7498114427324635e-05, "loss": 3.1142, "step": 84050 }, { "epoch": 9.051770530621031, "grad_norm": 0.8786686062812805, "learning_rate": 5.717487339726322e-05, "loss": 3.1196, "step": 84100 }, { "epoch": 9.05715208266064, "grad_norm": 0.873644232749939, "learning_rate": 5.6851632367201804e-05, "loss": 3.1298, "step": 84150 }, { "epoch": 9.062533634700248, "grad_norm": 0.8933513760566711, "learning_rate": 5.6528391337140385e-05, "loss": 3.1089, "step": 84200 }, { "epoch": 9.067915186739857, "grad_norm": 0.8709551095962524, "learning_rate": 5.620515030707897e-05, "loss": 3.106, "step": 84250 }, { "epoch": 9.073296738779463, "grad_norm": 0.8717238306999207, "learning_rate": 5.5881909277017554e-05, "loss": 3.1345, "step": 84300 }, { "epoch": 9.078678290819072, "grad_norm": 0.858155369758606, "learning_rate": 5.555866824695614e-05, "loss": 3.1032, "step": 84350 }, { "epoch": 9.08405984285868, "grad_norm": 0.8939986824989319, "learning_rate": 5.523542721689472e-05, "loss": 3.1245, "step": 84400 }, { "epoch": 9.089441394898289, "grad_norm": 0.9245389103889465, "learning_rate": 5.4912186186833304e-05, "loss": 3.1343, "step": 84450 }, { "epoch": 9.094822946937898, "grad_norm": 0.9213153719902039, "learning_rate": 5.45889451567719e-05, "loss": 3.1191, "step": 84500 }, { "epoch": 9.100204498977504, "grad_norm": 0.8970081210136414, "learning_rate": 5.426570412671048e-05, "loss": 3.1257, "step": 84550 }, { "epoch": 9.105586051017113, "grad_norm": 0.9398916959762573, "learning_rate": 5.394246309664907e-05, "loss": 3.1261, "step": 84600 }, { "epoch": 9.110967603056721, "grad_norm": 0.8600006103515625, "learning_rate": 5.361922206658765e-05, "loss": 3.1337, "step": 84650 }, { "epoch": 9.11634915509633, "grad_norm": 0.926282525062561, "learning_rate": 5.329598103652623e-05, "loss": 3.1305, "step": 84700 }, { "epoch": 9.121730707135939, "grad_norm": 0.8645695447921753, "learning_rate": 5.297274000646482e-05, "loss": 3.1233, "step": 84750 }, { "epoch": 9.127112259175545, "grad_norm": 0.8621381521224976, "learning_rate": 5.26494989764034e-05, "loss": 3.121, "step": 84800 }, { "epoch": 9.132493811215154, "grad_norm": 0.8733987212181091, "learning_rate": 5.232625794634199e-05, "loss": 3.1233, "step": 84850 }, { "epoch": 9.137875363254762, "grad_norm": 0.940946102142334, "learning_rate": 5.200301691628057e-05, "loss": 3.1294, "step": 84900 }, { "epoch": 9.143256915294371, "grad_norm": 0.8964011073112488, "learning_rate": 5.167977588621915e-05, "loss": 3.1282, "step": 84950 }, { "epoch": 9.14863846733398, "grad_norm": 0.9262023568153381, "learning_rate": 5.135653485615774e-05, "loss": 3.1232, "step": 85000 }, { "epoch": 9.14863846733398, "eval_accuracy": 0.3931999155549756, "eval_loss": 3.31473970413208, "eval_runtime": 184.9896, "eval_samples_per_second": 97.362, "eval_steps_per_second": 6.087, "step": 85000 }, { "epoch": 9.154020019373588, "grad_norm": 0.860278844833374, "learning_rate": 5.103329382609632e-05, "loss": 3.1191, "step": 85050 }, { "epoch": 9.159401571413195, "grad_norm": 0.9099411368370056, "learning_rate": 5.0710052796034906e-05, "loss": 3.1229, "step": 85100 }, { "epoch": 9.164783123452803, "grad_norm": 0.8791196942329407, "learning_rate": 5.038681176597349e-05, "loss": 3.0978, "step": 85150 }, { "epoch": 9.170164675492412, "grad_norm": 1.0391086339950562, "learning_rate": 5.006357073591207e-05, "loss": 3.1267, "step": 85200 }, { "epoch": 9.17554622753202, "grad_norm": 0.8792301416397095, "learning_rate": 4.974032970585066e-05, "loss": 3.1279, "step": 85250 }, { "epoch": 9.180927779571629, "grad_norm": 0.8521240949630737, "learning_rate": 4.9417088675789244e-05, "loss": 3.118, "step": 85300 }, { "epoch": 9.186309331611236, "grad_norm": 0.912763237953186, "learning_rate": 4.909384764572783e-05, "loss": 3.121, "step": 85350 }, { "epoch": 9.191690883650844, "grad_norm": 0.8790477514266968, "learning_rate": 4.877060661566641e-05, "loss": 3.129, "step": 85400 }, { "epoch": 9.197072435690453, "grad_norm": 0.8999446630477905, "learning_rate": 4.8447365585604994e-05, "loss": 3.1379, "step": 85450 }, { "epoch": 9.202453987730062, "grad_norm": 0.8669745326042175, "learning_rate": 4.812412455554358e-05, "loss": 3.1175, "step": 85500 }, { "epoch": 9.20783553976967, "grad_norm": 0.9326590895652771, "learning_rate": 4.780088352548216e-05, "loss": 3.1152, "step": 85550 }, { "epoch": 9.213217091809279, "grad_norm": 0.8711249828338623, "learning_rate": 4.747764249542075e-05, "loss": 3.1119, "step": 85600 }, { "epoch": 9.218598643848885, "grad_norm": 0.9170059561729431, "learning_rate": 4.715440146535933e-05, "loss": 3.1528, "step": 85650 }, { "epoch": 9.223980195888494, "grad_norm": 0.8670192360877991, "learning_rate": 4.683116043529791e-05, "loss": 3.1398, "step": 85700 }, { "epoch": 9.229361747928102, "grad_norm": 0.9062800407409668, "learning_rate": 4.65079194052365e-05, "loss": 3.1122, "step": 85750 }, { "epoch": 9.234743299967711, "grad_norm": 0.8952690362930298, "learning_rate": 4.618467837517508e-05, "loss": 3.1347, "step": 85800 }, { "epoch": 9.24012485200732, "grad_norm": 0.9159784317016602, "learning_rate": 4.586143734511367e-05, "loss": 3.1263, "step": 85850 }, { "epoch": 9.245506404046926, "grad_norm": 0.9201834797859192, "learning_rate": 4.553819631505225e-05, "loss": 3.1292, "step": 85900 }, { "epoch": 9.250887956086535, "grad_norm": 0.8721088767051697, "learning_rate": 4.521495528499083e-05, "loss": 3.126, "step": 85950 }, { "epoch": 9.256269508126143, "grad_norm": 0.9054554104804993, "learning_rate": 4.4898179075530645e-05, "loss": 3.1126, "step": 86000 }, { "epoch": 9.256269508126143, "eval_accuracy": 0.3934344971397671, "eval_loss": 3.3119852542877197, "eval_runtime": 184.8448, "eval_samples_per_second": 97.438, "eval_steps_per_second": 6.092, "step": 86000 }, { "epoch": 9.261651060165752, "grad_norm": 0.8648436665534973, "learning_rate": 4.457493804546923e-05, "loss": 3.1399, "step": 86050 }, { "epoch": 9.26703261220536, "grad_norm": 0.8899005651473999, "learning_rate": 4.4251697015407814e-05, "loss": 3.1283, "step": 86100 }, { "epoch": 9.272414164244967, "grad_norm": 0.8872169852256775, "learning_rate": 4.392845598534641e-05, "loss": 3.127, "step": 86150 }, { "epoch": 9.277795716284576, "grad_norm": 0.8724729418754578, "learning_rate": 4.360521495528499e-05, "loss": 3.126, "step": 86200 }, { "epoch": 9.283177268324184, "grad_norm": 0.8950654864311218, "learning_rate": 4.3281973925223564e-05, "loss": 3.1202, "step": 86250 }, { "epoch": 9.288558820363793, "grad_norm": 0.9006926417350769, "learning_rate": 4.295873289516216e-05, "loss": 3.1234, "step": 86300 }, { "epoch": 9.293940372403402, "grad_norm": 0.8863205313682556, "learning_rate": 4.263549186510074e-05, "loss": 3.112, "step": 86350 }, { "epoch": 9.29932192444301, "grad_norm": 0.8906213045120239, "learning_rate": 4.231225083503933e-05, "loss": 3.1287, "step": 86400 }, { "epoch": 9.304703476482617, "grad_norm": 0.8871445655822754, "learning_rate": 4.198900980497791e-05, "loss": 3.1298, "step": 86450 }, { "epoch": 9.310085028522225, "grad_norm": 0.9487957954406738, "learning_rate": 4.166576877491649e-05, "loss": 3.1195, "step": 86500 }, { "epoch": 9.315466580561834, "grad_norm": 0.8825381398200989, "learning_rate": 4.134252774485508e-05, "loss": 3.1436, "step": 86550 }, { "epoch": 9.320848132601443, "grad_norm": 0.8810291886329651, "learning_rate": 4.101928671479366e-05, "loss": 3.1323, "step": 86600 }, { "epoch": 9.326229684641051, "grad_norm": 0.9148359298706055, "learning_rate": 4.069604568473225e-05, "loss": 3.1344, "step": 86650 }, { "epoch": 9.331611236680658, "grad_norm": 0.8992313146591187, "learning_rate": 4.037280465467083e-05, "loss": 3.1286, "step": 86700 }, { "epoch": 9.336992788720266, "grad_norm": 0.9089197516441345, "learning_rate": 4.004956362460941e-05, "loss": 3.1332, "step": 86750 }, { "epoch": 9.342374340759875, "grad_norm": 0.868971586227417, "learning_rate": 3.9726322594548e-05, "loss": 3.1276, "step": 86800 }, { "epoch": 9.347755892799483, "grad_norm": 0.921055257320404, "learning_rate": 3.940308156448658e-05, "loss": 3.1217, "step": 86850 }, { "epoch": 9.353137444839092, "grad_norm": 0.959754228591919, "learning_rate": 3.9079840534425166e-05, "loss": 3.1177, "step": 86900 }, { "epoch": 9.3585189968787, "grad_norm": 0.892400324344635, "learning_rate": 3.875659950436375e-05, "loss": 3.1181, "step": 86950 }, { "epoch": 9.363900548918307, "grad_norm": 0.9151771068572998, "learning_rate": 3.843335847430233e-05, "loss": 3.1101, "step": 87000 }, { "epoch": 9.363900548918307, "eval_accuracy": 0.3936111667353534, "eval_loss": 3.3083672523498535, "eval_runtime": 185.0906, "eval_samples_per_second": 97.309, "eval_steps_per_second": 6.084, "step": 87000 }, { "epoch": 9.369282100957916, "grad_norm": 0.8925448656082153, "learning_rate": 3.811011744424092e-05, "loss": 3.1359, "step": 87050 }, { "epoch": 9.374663652997524, "grad_norm": 0.8919752240180969, "learning_rate": 3.7786876414179504e-05, "loss": 3.1191, "step": 87100 }, { "epoch": 9.380045205037133, "grad_norm": 0.9180679321289062, "learning_rate": 3.7463635384118085e-05, "loss": 3.1496, "step": 87150 }, { "epoch": 9.385426757076742, "grad_norm": 0.8926573991775513, "learning_rate": 3.714039435405667e-05, "loss": 3.1319, "step": 87200 }, { "epoch": 9.390808309116348, "grad_norm": 0.9110729098320007, "learning_rate": 3.6817153323995254e-05, "loss": 3.1211, "step": 87250 }, { "epoch": 9.396189861155957, "grad_norm": 0.9294565320014954, "learning_rate": 3.649391229393384e-05, "loss": 3.1246, "step": 87300 }, { "epoch": 9.401571413195565, "grad_norm": 0.88547283411026, "learning_rate": 3.617067126387242e-05, "loss": 3.137, "step": 87350 }, { "epoch": 9.406952965235174, "grad_norm": 0.8936682343482971, "learning_rate": 3.5847430233811004e-05, "loss": 3.1245, "step": 87400 }, { "epoch": 9.412334517274783, "grad_norm": 0.9498528838157654, "learning_rate": 3.552418920374959e-05, "loss": 3.1319, "step": 87450 }, { "epoch": 9.417716069314391, "grad_norm": 0.949592113494873, "learning_rate": 3.520094817368818e-05, "loss": 3.1445, "step": 87500 }, { "epoch": 9.423097621353998, "grad_norm": 0.8684675097465515, "learning_rate": 3.487770714362676e-05, "loss": 3.1472, "step": 87550 }, { "epoch": 9.428479173393606, "grad_norm": 0.950598418712616, "learning_rate": 3.455446611356535e-05, "loss": 3.1463, "step": 87600 }, { "epoch": 9.433860725433215, "grad_norm": 0.9288617968559265, "learning_rate": 3.423122508350393e-05, "loss": 3.1354, "step": 87650 }, { "epoch": 9.439242277472824, "grad_norm": 0.9261043071746826, "learning_rate": 3.390798405344251e-05, "loss": 3.1414, "step": 87700 }, { "epoch": 9.444623829512432, "grad_norm": 0.8761609196662903, "learning_rate": 3.35847430233811e-05, "loss": 3.1457, "step": 87750 }, { "epoch": 9.450005381552039, "grad_norm": 0.9005210995674133, "learning_rate": 3.326150199331968e-05, "loss": 3.1285, "step": 87800 }, { "epoch": 9.455386933591647, "grad_norm": 0.9317635297775269, "learning_rate": 3.293826096325827e-05, "loss": 3.1293, "step": 87850 }, { "epoch": 9.460768485631256, "grad_norm": 0.886011004447937, "learning_rate": 3.261501993319685e-05, "loss": 3.1344, "step": 87900 }, { "epoch": 9.466150037670864, "grad_norm": 0.9412506818771362, "learning_rate": 3.229177890313544e-05, "loss": 3.136, "step": 87950 }, { "epoch": 9.471531589710473, "grad_norm": 0.9001096487045288, "learning_rate": 3.197500269367525e-05, "loss": 3.1343, "step": 88000 }, { "epoch": 9.471531589710473, "eval_accuracy": 0.39403382646895174, "eval_loss": 3.306755781173706, "eval_runtime": 184.7543, "eval_samples_per_second": 97.486, "eval_steps_per_second": 6.095, "step": 88000 }, { "epoch": 9.476913141750082, "grad_norm": 0.9205313920974731, "learning_rate": 3.165176166361383e-05, "loss": 3.1289, "step": 88050 }, { "epoch": 9.482294693789688, "grad_norm": 0.9257974624633789, "learning_rate": 3.132852063355242e-05, "loss": 3.1247, "step": 88100 }, { "epoch": 9.487676245829297, "grad_norm": 0.9012815952301025, "learning_rate": 3.1005279603491e-05, "loss": 3.1413, "step": 88150 }, { "epoch": 9.493057797868905, "grad_norm": 0.9336722493171692, "learning_rate": 3.068203857342958e-05, "loss": 3.1185, "step": 88200 }, { "epoch": 9.498439349908514, "grad_norm": 0.9232977628707886, "learning_rate": 3.035879754336817e-05, "loss": 3.1124, "step": 88250 }, { "epoch": 9.503820901948123, "grad_norm": 0.8725977540016174, "learning_rate": 3.0035556513306754e-05, "loss": 3.1293, "step": 88300 }, { "epoch": 9.50920245398773, "grad_norm": 0.9048357009887695, "learning_rate": 2.971231548324534e-05, "loss": 3.1483, "step": 88350 }, { "epoch": 9.514584006027338, "grad_norm": 0.9110617637634277, "learning_rate": 2.938907445318392e-05, "loss": 3.1158, "step": 88400 }, { "epoch": 9.519965558066946, "grad_norm": 0.8577834963798523, "learning_rate": 2.9065833423122504e-05, "loss": 3.1259, "step": 88450 }, { "epoch": 9.525347110106555, "grad_norm": 0.8744284510612488, "learning_rate": 2.874259239306109e-05, "loss": 3.1197, "step": 88500 }, { "epoch": 9.530728662146164, "grad_norm": 0.8887792229652405, "learning_rate": 2.8419351362999673e-05, "loss": 3.1283, "step": 88550 }, { "epoch": 9.536110214185772, "grad_norm": 0.9559165835380554, "learning_rate": 2.809611033293826e-05, "loss": 3.1234, "step": 88600 }, { "epoch": 9.541491766225379, "grad_norm": 0.8760658502578735, "learning_rate": 2.7772869302876845e-05, "loss": 3.118, "step": 88650 }, { "epoch": 9.546873318264987, "grad_norm": 0.885046660900116, "learning_rate": 2.7449628272815427e-05, "loss": 3.1529, "step": 88700 }, { "epoch": 9.552254870304596, "grad_norm": 0.8885508179664612, "learning_rate": 2.712638724275401e-05, "loss": 3.1341, "step": 88750 }, { "epoch": 9.557636422344205, "grad_norm": 0.90824294090271, "learning_rate": 2.6803146212692596e-05, "loss": 3.1306, "step": 88800 }, { "epoch": 9.563017974383813, "grad_norm": 0.9239144921302795, "learning_rate": 2.647990518263118e-05, "loss": 3.1522, "step": 88850 }, { "epoch": 9.56839952642342, "grad_norm": 0.927269458770752, "learning_rate": 2.6156664152569765e-05, "loss": 3.1294, "step": 88900 }, { "epoch": 9.573781078463028, "grad_norm": 0.8974427580833435, "learning_rate": 2.5833423122508346e-05, "loss": 3.1244, "step": 88950 }, { "epoch": 9.579162630502637, "grad_norm": 0.844535231590271, "learning_rate": 2.5510182092446934e-05, "loss": 3.1383, "step": 89000 }, { "epoch": 9.579162630502637, "eval_accuracy": 0.39431393361373746, "eval_loss": 3.303724765777588, "eval_runtime": 184.8828, "eval_samples_per_second": 97.418, "eval_steps_per_second": 6.09, "step": 89000 }, { "epoch": 9.584544182542245, "grad_norm": 0.8630052804946899, "learning_rate": 2.5186941062385518e-05, "loss": 3.1265, "step": 89050 }, { "epoch": 9.589925734581854, "grad_norm": 0.8742738366127014, "learning_rate": 2.4863700032324103e-05, "loss": 3.144, "step": 89100 }, { "epoch": 9.59530728662146, "grad_norm": 0.9079467058181763, "learning_rate": 2.4540459002262687e-05, "loss": 3.1231, "step": 89150 }, { "epoch": 9.60068883866107, "grad_norm": 0.8864733576774597, "learning_rate": 2.4217217972201268e-05, "loss": 3.1178, "step": 89200 }, { "epoch": 9.606070390700678, "grad_norm": 0.9138379096984863, "learning_rate": 2.3893976942139853e-05, "loss": 3.1156, "step": 89250 }, { "epoch": 9.611451942740286, "grad_norm": 0.9160833954811096, "learning_rate": 2.3570735912078437e-05, "loss": 3.1396, "step": 89300 }, { "epoch": 9.616833494779895, "grad_norm": 0.9601799249649048, "learning_rate": 2.324749488201702e-05, "loss": 3.1191, "step": 89350 }, { "epoch": 9.622215046819504, "grad_norm": 0.9554757475852966, "learning_rate": 2.292425385195561e-05, "loss": 3.1543, "step": 89400 }, { "epoch": 9.62759659885911, "grad_norm": 0.9189677238464355, "learning_rate": 2.260101282189419e-05, "loss": 3.1164, "step": 89450 }, { "epoch": 9.632978150898719, "grad_norm": 0.888224184513092, "learning_rate": 2.2277771791832775e-05, "loss": 3.1328, "step": 89500 }, { "epoch": 9.638359702938327, "grad_norm": 0.8766180872917175, "learning_rate": 2.195453076177136e-05, "loss": 3.1125, "step": 89550 }, { "epoch": 9.643741254977936, "grad_norm": 0.9146680235862732, "learning_rate": 2.1631289731709944e-05, "loss": 3.121, "step": 89600 }, { "epoch": 9.649122807017545, "grad_norm": 0.9508671760559082, "learning_rate": 2.130804870164853e-05, "loss": 3.1331, "step": 89650 }, { "epoch": 9.654504359057151, "grad_norm": 0.8772718906402588, "learning_rate": 2.098480767158711e-05, "loss": 3.1178, "step": 89700 }, { "epoch": 9.65988591109676, "grad_norm": 0.9173164963722229, "learning_rate": 2.0661566641525694e-05, "loss": 3.1261, "step": 89750 }, { "epoch": 9.665267463136368, "grad_norm": 0.8904097676277161, "learning_rate": 2.0338325611464282e-05, "loss": 3.1378, "step": 89800 }, { "epoch": 9.670649015175977, "grad_norm": 0.9122865200042725, "learning_rate": 2.0015084581402867e-05, "loss": 3.1482, "step": 89850 }, { "epoch": 9.676030567215586, "grad_norm": 0.9031899571418762, "learning_rate": 1.969184355134145e-05, "loss": 3.1317, "step": 89900 }, { "epoch": 9.681412119255192, "grad_norm": 0.9516944885253906, "learning_rate": 1.9368602521280032e-05, "loss": 3.1399, "step": 89950 }, { "epoch": 9.6867936712948, "grad_norm": 0.8837878108024597, "learning_rate": 1.9045361491218617e-05, "loss": 3.1249, "step": 90000 }, { "epoch": 9.6867936712948, "eval_accuracy": 0.39447756486278607, "eval_loss": 3.3018171787261963, "eval_runtime": 185.0594, "eval_samples_per_second": 97.326, "eval_steps_per_second": 6.085, "step": 90000 }, { "epoch": 9.69217522333441, "grad_norm": 0.9284664988517761, "learning_rate": 1.872858528175843e-05, "loss": 3.1494, "step": 90050 }, { "epoch": 9.697556775374018, "grad_norm": 0.9753976464271545, "learning_rate": 1.8405344251697014e-05, "loss": 3.1438, "step": 90100 }, { "epoch": 9.702938327413626, "grad_norm": 0.9227730631828308, "learning_rate": 1.80821032216356e-05, "loss": 3.1289, "step": 90150 }, { "epoch": 9.708319879453235, "grad_norm": 0.8707805275917053, "learning_rate": 1.7758862191574183e-05, "loss": 3.1446, "step": 90200 }, { "epoch": 9.713701431492842, "grad_norm": 0.8503866791725159, "learning_rate": 1.7435621161512768e-05, "loss": 3.1292, "step": 90250 }, { "epoch": 9.71908298353245, "grad_norm": 0.8727654218673706, "learning_rate": 1.711238013145135e-05, "loss": 3.121, "step": 90300 }, { "epoch": 9.724464535572059, "grad_norm": 0.9083965420722961, "learning_rate": 1.6789139101389937e-05, "loss": 3.1278, "step": 90350 }, { "epoch": 9.729846087611667, "grad_norm": 0.865218997001648, "learning_rate": 1.6472362891929747e-05, "loss": 3.1158, "step": 90400 }, { "epoch": 9.735227639651276, "grad_norm": 0.9163516759872437, "learning_rate": 1.614912186186833e-05, "loss": 3.1345, "step": 90450 }, { "epoch": 9.740609191690883, "grad_norm": 0.8571822047233582, "learning_rate": 1.5825880831806916e-05, "loss": 3.1355, "step": 90500 }, { "epoch": 9.745990743730491, "grad_norm": 0.8619289398193359, "learning_rate": 1.55026398017455e-05, "loss": 3.113, "step": 90550 }, { "epoch": 9.7513722957701, "grad_norm": 0.8706497550010681, "learning_rate": 1.5179398771684085e-05, "loss": 3.1324, "step": 90600 }, { "epoch": 9.756753847809708, "grad_norm": 0.9034489989280701, "learning_rate": 1.485615774162267e-05, "loss": 3.1221, "step": 90650 }, { "epoch": 9.762135399849317, "grad_norm": 0.8666117191314697, "learning_rate": 1.4532916711561252e-05, "loss": 3.1205, "step": 90700 }, { "epoch": 9.767516951888926, "grad_norm": 0.9398617148399353, "learning_rate": 1.4209675681499837e-05, "loss": 3.1403, "step": 90750 }, { "epoch": 9.772898503928532, "grad_norm": 0.9415978789329529, "learning_rate": 1.3886434651438423e-05, "loss": 3.1329, "step": 90800 }, { "epoch": 9.77828005596814, "grad_norm": 0.8788616061210632, "learning_rate": 1.3563193621377006e-05, "loss": 3.1208, "step": 90850 }, { "epoch": 9.78366160800775, "grad_norm": 0.8646774888038635, "learning_rate": 1.323995259131559e-05, "loss": 3.1319, "step": 90900 }, { "epoch": 9.789043160047358, "grad_norm": 0.9247779846191406, "learning_rate": 1.2916711561254173e-05, "loss": 3.1319, "step": 90950 }, { "epoch": 9.794424712086967, "grad_norm": 0.9060809016227722, "learning_rate": 1.2593470531192759e-05, "loss": 3.1263, "step": 91000 }, { "epoch": 9.794424712086967, "eval_accuracy": 0.39467879001101847, "eval_loss": 3.299553394317627, "eval_runtime": 184.8969, "eval_samples_per_second": 97.411, "eval_steps_per_second": 6.09, "step": 91000 }, { "epoch": 9.799806264126573, "grad_norm": 0.8818645477294922, "learning_rate": 1.2270229501131344e-05, "loss": 3.1265, "step": 91050 }, { "epoch": 9.805187816166182, "grad_norm": 0.917174220085144, "learning_rate": 1.1946988471069926e-05, "loss": 3.1284, "step": 91100 }, { "epoch": 9.81056936820579, "grad_norm": 0.9325860738754272, "learning_rate": 1.162374744100851e-05, "loss": 3.1433, "step": 91150 }, { "epoch": 9.815950920245399, "grad_norm": 0.9602986574172974, "learning_rate": 1.1300506410947095e-05, "loss": 3.1445, "step": 91200 }, { "epoch": 9.821332472285007, "grad_norm": 0.9245926737785339, "learning_rate": 1.097726538088568e-05, "loss": 3.1266, "step": 91250 }, { "epoch": 9.826714024324616, "grad_norm": 0.9178900718688965, "learning_rate": 1.0654024350824264e-05, "loss": 3.1296, "step": 91300 }, { "epoch": 9.832095576364223, "grad_norm": 0.8924757838249207, "learning_rate": 1.0330783320762847e-05, "loss": 3.1316, "step": 91350 }, { "epoch": 9.837477128403831, "grad_norm": 0.8745411038398743, "learning_rate": 1.0007542290701433e-05, "loss": 3.1478, "step": 91400 }, { "epoch": 9.84285868044344, "grad_norm": 0.9549648761749268, "learning_rate": 9.684301260640016e-06, "loss": 3.1255, "step": 91450 }, { "epoch": 9.848240232483048, "grad_norm": 0.8879762291908264, "learning_rate": 9.3610602305786e-06, "loss": 3.1295, "step": 91500 }, { "epoch": 9.853621784522657, "grad_norm": 0.8747987747192383, "learning_rate": 9.037819200517185e-06, "loss": 3.1141, "step": 91550 }, { "epoch": 9.859003336562264, "grad_norm": 0.9374912977218628, "learning_rate": 8.71457817045577e-06, "loss": 3.133, "step": 91600 }, { "epoch": 9.864384888601872, "grad_norm": 0.8693592548370361, "learning_rate": 8.391337140394354e-06, "loss": 3.1176, "step": 91650 }, { "epoch": 9.869766440641481, "grad_norm": 0.9023551344871521, "learning_rate": 8.068096110332939e-06, "loss": 3.116, "step": 91700 }, { "epoch": 9.87514799268109, "grad_norm": 0.8818048238754272, "learning_rate": 7.744855080271521e-06, "loss": 3.134, "step": 91750 }, { "epoch": 9.880529544720698, "grad_norm": 0.9004733562469482, "learning_rate": 7.421614050210106e-06, "loss": 3.1342, "step": 91800 }, { "epoch": 9.885911096760307, "grad_norm": 0.9453480243682861, "learning_rate": 7.09837302014869e-06, "loss": 3.1202, "step": 91850 }, { "epoch": 9.891292648799913, "grad_norm": 0.8884382843971252, "learning_rate": 6.775131990087275e-06, "loss": 3.1268, "step": 91900 }, { "epoch": 9.896674200839522, "grad_norm": 0.9106020927429199, "learning_rate": 6.451890960025859e-06, "loss": 3.1243, "step": 91950 }, { "epoch": 9.90205575287913, "grad_norm": 0.9381679892539978, "learning_rate": 6.128649929964443e-06, "loss": 3.1164, "step": 92000 }, { "epoch": 9.90205575287913, "eval_accuracy": 0.39488544780364165, "eval_loss": 3.2985033988952637, "eval_runtime": 184.9441, "eval_samples_per_second": 97.386, "eval_steps_per_second": 6.088, "step": 92000 }, { "epoch": 9.907437304918739, "grad_norm": 0.9050652384757996, "learning_rate": 5.805408899903027e-06, "loss": 3.1365, "step": 92050 }, { "epoch": 9.912818856958348, "grad_norm": 0.924476146697998, "learning_rate": 5.482167869841611e-06, "loss": 3.1275, "step": 92100 }, { "epoch": 9.918200408997954, "grad_norm": 0.8838419914245605, "learning_rate": 5.1589268397801965e-06, "loss": 3.1385, "step": 92150 }, { "epoch": 9.923581961037563, "grad_norm": 0.8708315491676331, "learning_rate": 4.83568580971878e-06, "loss": 3.1325, "step": 92200 }, { "epoch": 9.928963513077171, "grad_norm": 0.9714722633361816, "learning_rate": 4.512444779657364e-06, "loss": 3.1291, "step": 92250 }, { "epoch": 9.93434506511678, "grad_norm": 0.8961527943611145, "learning_rate": 4.189203749595948e-06, "loss": 3.1339, "step": 92300 }, { "epoch": 9.939726617156388, "grad_norm": 0.9741976261138916, "learning_rate": 3.865962719534533e-06, "loss": 3.1428, "step": 92350 }, { "epoch": 9.945108169195997, "grad_norm": 0.8602545261383057, "learning_rate": 3.542721689473117e-06, "loss": 3.1453, "step": 92400 }, { "epoch": 9.950489721235604, "grad_norm": 0.8766176700592041, "learning_rate": 3.2194806594117013e-06, "loss": 3.1286, "step": 92450 }, { "epoch": 9.955871273275212, "grad_norm": 0.9090676307678223, "learning_rate": 2.8962396293502854e-06, "loss": 3.1159, "step": 92500 }, { "epoch": 9.961252825314821, "grad_norm": 0.8845322132110596, "learning_rate": 2.5729985992888694e-06, "loss": 3.1353, "step": 92550 }, { "epoch": 9.96663437735443, "grad_norm": 0.924277663230896, "learning_rate": 2.249757569227454e-06, "loss": 3.1229, "step": 92600 }, { "epoch": 9.972015929394038, "grad_norm": 0.8867099285125732, "learning_rate": 1.926516539166038e-06, "loss": 3.1248, "step": 92650 }, { "epoch": 9.977397481433645, "grad_norm": 0.9389845132827759, "learning_rate": 1.603275509104622e-06, "loss": 3.1412, "step": 92700 }, { "epoch": 9.982779033473253, "grad_norm": 0.9016231894493103, "learning_rate": 1.2800344790432064e-06, "loss": 3.1243, "step": 92750 }, { "epoch": 9.988160585512862, "grad_norm": 0.8891977071762085, "learning_rate": 9.567934489817906e-07, "loss": 3.112, "step": 92800 }, { "epoch": 9.99354213755247, "grad_norm": 0.896614134311676, "learning_rate": 6.335524189203748e-07, "loss": 3.126, "step": 92850 }, { "epoch": 9.998923689592079, "grad_norm": 0.9039609432220459, "learning_rate": 3.103113888589591e-07, "loss": 3.1234, "step": 92900 }, { "epoch": 10.0, "step": 92910, "total_flos": 7.76821211136e+17, "train_loss": 3.4538463850168144, "train_runtime": 79850.848, "train_samples_per_second": 37.232, "train_steps_per_second": 1.164 } ], "logging_steps": 50, "max_steps": 92910, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.76821211136e+17, "train_batch_size": 32, "trial_name": null, "trial_params": null }