100M_high_1000_634 / trainer_state.json
craa's picture
End of training
587d3b3 verified
{
"best_metric": 3.3065290451049805,
"best_model_checkpoint": "/scratch/cl5625/exceptions/models/100M_634/checkpoint-90000",
"epoch": 10.0,
"eval_steps": 1000,
"global_step": 92910,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.005381552039608223,
"grad_norm": 2.728989362716675,
"learning_rate": 0.00028199999999999997,
"loss": 8.6119,
"step": 50
},
{
"epoch": 0.010763104079216447,
"grad_norm": 1.724418044090271,
"learning_rate": 0.0005819999999999999,
"loss": 6.9596,
"step": 100
},
{
"epoch": 0.01614465611882467,
"grad_norm": 1.106858491897583,
"learning_rate": 0.0005996961534317422,
"loss": 6.5223,
"step": 150
},
{
"epoch": 0.021526208158432893,
"grad_norm": 1.3945527076721191,
"learning_rate": 0.0005993729124016807,
"loss": 6.2759,
"step": 200
},
{
"epoch": 0.026907760198041114,
"grad_norm": 1.685577630996704,
"learning_rate": 0.0005990496713716194,
"loss": 6.0955,
"step": 250
},
{
"epoch": 0.03228931223764934,
"grad_norm": 2.0357065200805664,
"learning_rate": 0.000598726430341558,
"loss": 6.0263,
"step": 300
},
{
"epoch": 0.03767086427725756,
"grad_norm": 0.9172292351722717,
"learning_rate": 0.0005984031893114966,
"loss": 5.8966,
"step": 350
},
{
"epoch": 0.04305241631686579,
"grad_norm": 2.3104074001312256,
"learning_rate": 0.0005980799482814351,
"loss": 5.8365,
"step": 400
},
{
"epoch": 0.048433968356474004,
"grad_norm": 1.7049633264541626,
"learning_rate": 0.0005977567072513737,
"loss": 5.7381,
"step": 450
},
{
"epoch": 0.05381552039608223,
"grad_norm": 1.1454957723617554,
"learning_rate": 0.0005974334662213123,
"loss": 5.6778,
"step": 500
},
{
"epoch": 0.05919707243569045,
"grad_norm": 1.5053794384002686,
"learning_rate": 0.0005971102251912509,
"loss": 5.6187,
"step": 550
},
{
"epoch": 0.06457862447529868,
"grad_norm": 0.9626969695091248,
"learning_rate": 0.0005967869841611895,
"loss": 5.5314,
"step": 600
},
{
"epoch": 0.0699601765149069,
"grad_norm": 1.2051186561584473,
"learning_rate": 0.000596463743131128,
"loss": 5.458,
"step": 650
},
{
"epoch": 0.07534172855451512,
"grad_norm": 1.3867778778076172,
"learning_rate": 0.0005961405021010667,
"loss": 5.3855,
"step": 700
},
{
"epoch": 0.08072328059412334,
"grad_norm": 1.4545854330062866,
"learning_rate": 0.0005958172610710052,
"loss": 5.3445,
"step": 750
},
{
"epoch": 0.08610483263373157,
"grad_norm": 1.1404812335968018,
"learning_rate": 0.0005954940200409439,
"loss": 5.2568,
"step": 800
},
{
"epoch": 0.09148638467333979,
"grad_norm": 1.3057512044906616,
"learning_rate": 0.0005951707790108824,
"loss": 5.2444,
"step": 850
},
{
"epoch": 0.09686793671294801,
"grad_norm": 1.1935104131698608,
"learning_rate": 0.0005948475379808209,
"loss": 5.1903,
"step": 900
},
{
"epoch": 0.10224948875255624,
"grad_norm": 0.9733734130859375,
"learning_rate": 0.0005945242969507596,
"loss": 5.1723,
"step": 950
},
{
"epoch": 0.10763104079216446,
"grad_norm": 1.0952446460723877,
"learning_rate": 0.0005942010559206981,
"loss": 5.1166,
"step": 1000
},
{
"epoch": 0.10763104079216446,
"eval_accuracy": 0.22600756810824782,
"eval_loss": 5.031754493713379,
"eval_runtime": 180.4654,
"eval_samples_per_second": 99.803,
"eval_steps_per_second": 6.239,
"step": 1000
},
{
"epoch": 0.11301259283177269,
"grad_norm": 1.521315097808838,
"learning_rate": 0.0005938778148906367,
"loss": 5.0458,
"step": 1050
},
{
"epoch": 0.1183941448713809,
"grad_norm": 1.205197811126709,
"learning_rate": 0.0005935545738605753,
"loss": 5.023,
"step": 1100
},
{
"epoch": 0.12377569691098914,
"grad_norm": 0.8693559169769287,
"learning_rate": 0.000593231332830514,
"loss": 4.9911,
"step": 1150
},
{
"epoch": 0.12915724895059735,
"grad_norm": 1.2472184896469116,
"learning_rate": 0.0005929080918004525,
"loss": 4.9623,
"step": 1200
},
{
"epoch": 0.13453880099020557,
"grad_norm": 0.9657862186431885,
"learning_rate": 0.000592584850770391,
"loss": 4.9466,
"step": 1250
},
{
"epoch": 0.1399203530298138,
"grad_norm": 0.8568148016929626,
"learning_rate": 0.0005922616097403296,
"loss": 4.9032,
"step": 1300
},
{
"epoch": 0.14530190506942203,
"grad_norm": 0.949377179145813,
"learning_rate": 0.0005919383687102682,
"loss": 4.8553,
"step": 1350
},
{
"epoch": 0.15068345710903025,
"grad_norm": 0.92209392786026,
"learning_rate": 0.0005916151276802069,
"loss": 4.8447,
"step": 1400
},
{
"epoch": 0.15606500914863847,
"grad_norm": 0.7755404710769653,
"learning_rate": 0.0005912918866501454,
"loss": 4.8361,
"step": 1450
},
{
"epoch": 0.16144656118824668,
"grad_norm": 1.0212815999984741,
"learning_rate": 0.000590968645620084,
"loss": 4.7901,
"step": 1500
},
{
"epoch": 0.1668281132278549,
"grad_norm": 0.9886590838432312,
"learning_rate": 0.0005906454045900226,
"loss": 4.7598,
"step": 1550
},
{
"epoch": 0.17220966526746315,
"grad_norm": 1.1899733543395996,
"learning_rate": 0.0005903221635599612,
"loss": 4.7607,
"step": 1600
},
{
"epoch": 0.17759121730707136,
"grad_norm": 0.948335587978363,
"learning_rate": 0.0005899989225298998,
"loss": 4.7664,
"step": 1650
},
{
"epoch": 0.18297276934667958,
"grad_norm": 1.285933256149292,
"learning_rate": 0.0005896756814998383,
"loss": 4.7184,
"step": 1700
},
{
"epoch": 0.1883543213862878,
"grad_norm": 0.788101315498352,
"learning_rate": 0.0005893524404697769,
"loss": 4.6973,
"step": 1750
},
{
"epoch": 0.19373587342589602,
"grad_norm": 0.6696946620941162,
"learning_rate": 0.0005890291994397155,
"loss": 4.6636,
"step": 1800
},
{
"epoch": 0.19911742546550426,
"grad_norm": 0.7915722727775574,
"learning_rate": 0.0005887059584096541,
"loss": 4.6548,
"step": 1850
},
{
"epoch": 0.20449897750511248,
"grad_norm": 0.7558673620223999,
"learning_rate": 0.0005883827173795926,
"loss": 4.6091,
"step": 1900
},
{
"epoch": 0.2098805295447207,
"grad_norm": 0.9301922917366028,
"learning_rate": 0.0005880594763495313,
"loss": 4.6026,
"step": 1950
},
{
"epoch": 0.2152620815843289,
"grad_norm": 1.3686238527297974,
"learning_rate": 0.0005877362353194698,
"loss": 4.5936,
"step": 2000
},
{
"epoch": 0.2152620815843289,
"eval_accuracy": 0.2692196908151694,
"eval_loss": 4.520735263824463,
"eval_runtime": 180.5703,
"eval_samples_per_second": 99.745,
"eval_steps_per_second": 6.236,
"step": 2000
},
{
"epoch": 0.22064363362393713,
"grad_norm": 0.808482825756073,
"learning_rate": 0.0005874129942894084,
"loss": 4.5665,
"step": 2050
},
{
"epoch": 0.22602518566354537,
"grad_norm": 0.9632678031921387,
"learning_rate": 0.000587089753259347,
"loss": 4.5801,
"step": 2100
},
{
"epoch": 0.2314067377031536,
"grad_norm": 0.8442896604537964,
"learning_rate": 0.0005867665122292855,
"loss": 4.5534,
"step": 2150
},
{
"epoch": 0.2367882897427618,
"grad_norm": 1.0314425230026245,
"learning_rate": 0.0005864432711992242,
"loss": 4.5134,
"step": 2200
},
{
"epoch": 0.24216984178237003,
"grad_norm": 1.247391700744629,
"learning_rate": 0.0005861200301691628,
"loss": 4.4984,
"step": 2250
},
{
"epoch": 0.24755139382197827,
"grad_norm": 0.8239965438842773,
"learning_rate": 0.0005857967891391014,
"loss": 4.4888,
"step": 2300
},
{
"epoch": 0.2529329458615865,
"grad_norm": 0.69472736120224,
"learning_rate": 0.0005854735481090399,
"loss": 4.4592,
"step": 2350
},
{
"epoch": 0.2583144979011947,
"grad_norm": 1.0152015686035156,
"learning_rate": 0.0005851503070789784,
"loss": 4.4467,
"step": 2400
},
{
"epoch": 0.2636960499408029,
"grad_norm": 0.9481901526451111,
"learning_rate": 0.0005848270660489171,
"loss": 4.4431,
"step": 2450
},
{
"epoch": 0.26907760198041114,
"grad_norm": 0.914855420589447,
"learning_rate": 0.0005845038250188556,
"loss": 4.4427,
"step": 2500
},
{
"epoch": 0.27445915402001936,
"grad_norm": 0.7849532961845398,
"learning_rate": 0.0005841805839887943,
"loss": 4.4114,
"step": 2550
},
{
"epoch": 0.2798407060596276,
"grad_norm": 0.8744374513626099,
"learning_rate": 0.0005838573429587328,
"loss": 4.3739,
"step": 2600
},
{
"epoch": 0.2852222580992358,
"grad_norm": 0.8036152124404907,
"learning_rate": 0.0005835341019286715,
"loss": 4.4008,
"step": 2650
},
{
"epoch": 0.29060381013884407,
"grad_norm": 0.7502331137657166,
"learning_rate": 0.00058321086089861,
"loss": 4.361,
"step": 2700
},
{
"epoch": 0.2959853621784523,
"grad_norm": 0.737133800983429,
"learning_rate": 0.0005828876198685486,
"loss": 4.3553,
"step": 2750
},
{
"epoch": 0.3013669142180605,
"grad_norm": 0.90378338098526,
"learning_rate": 0.0005825643788384872,
"loss": 4.3844,
"step": 2800
},
{
"epoch": 0.3067484662576687,
"grad_norm": 0.6965075135231018,
"learning_rate": 0.0005822411378084257,
"loss": 4.3413,
"step": 2850
},
{
"epoch": 0.31213001829727693,
"grad_norm": 0.8999699354171753,
"learning_rate": 0.0005819178967783644,
"loss": 4.3424,
"step": 2900
},
{
"epoch": 0.31751157033688515,
"grad_norm": 0.7911924719810486,
"learning_rate": 0.0005815946557483029,
"loss": 4.3326,
"step": 2950
},
{
"epoch": 0.32289312237649337,
"grad_norm": 1.0059629678726196,
"learning_rate": 0.0005812714147182415,
"loss": 4.3077,
"step": 3000
},
{
"epoch": 0.32289312237649337,
"eval_accuracy": 0.29792002517270105,
"eval_loss": 4.243895530700684,
"eval_runtime": 180.4537,
"eval_samples_per_second": 99.81,
"eval_steps_per_second": 6.24,
"step": 3000
},
{
"epoch": 0.3282746744161016,
"grad_norm": 0.7331672310829163,
"learning_rate": 0.0005809481736881801,
"loss": 4.3384,
"step": 3050
},
{
"epoch": 0.3336562264557098,
"grad_norm": 0.9308981895446777,
"learning_rate": 0.0005806249326581187,
"loss": 4.2823,
"step": 3100
},
{
"epoch": 0.3390377784953181,
"grad_norm": 0.7922755479812622,
"learning_rate": 0.0005803016916280573,
"loss": 4.3038,
"step": 3150
},
{
"epoch": 0.3444193305349263,
"grad_norm": 0.8320370316505432,
"learning_rate": 0.0005799784505979959,
"loss": 4.2708,
"step": 3200
},
{
"epoch": 0.3498008825745345,
"grad_norm": 0.7593258023262024,
"learning_rate": 0.0005796552095679344,
"loss": 4.2699,
"step": 3250
},
{
"epoch": 0.35518243461414273,
"grad_norm": 0.8462849259376526,
"learning_rate": 0.000579331968537873,
"loss": 4.2502,
"step": 3300
},
{
"epoch": 0.36056398665375095,
"grad_norm": 0.6994985938072205,
"learning_rate": 0.0005790087275078116,
"loss": 4.2565,
"step": 3350
},
{
"epoch": 0.36594553869335916,
"grad_norm": 0.7432887554168701,
"learning_rate": 0.0005786854864777502,
"loss": 4.2457,
"step": 3400
},
{
"epoch": 0.3713270907329674,
"grad_norm": 0.7776950597763062,
"learning_rate": 0.0005783622454476888,
"loss": 4.2194,
"step": 3450
},
{
"epoch": 0.3767086427725756,
"grad_norm": 0.769761323928833,
"learning_rate": 0.0005780390044176273,
"loss": 4.2308,
"step": 3500
},
{
"epoch": 0.3820901948121838,
"grad_norm": 0.739103376865387,
"learning_rate": 0.000577715763387566,
"loss": 4.2353,
"step": 3550
},
{
"epoch": 0.38747174685179203,
"grad_norm": 0.881566047668457,
"learning_rate": 0.0005773925223575045,
"loss": 4.1979,
"step": 3600
},
{
"epoch": 0.3928532988914003,
"grad_norm": 0.7779421210289001,
"learning_rate": 0.0005770692813274432,
"loss": 4.2024,
"step": 3650
},
{
"epoch": 0.3982348509310085,
"grad_norm": 0.8221611380577087,
"learning_rate": 0.0005767460402973817,
"loss": 4.2008,
"step": 3700
},
{
"epoch": 0.40361640297061674,
"grad_norm": 0.749286949634552,
"learning_rate": 0.0005764227992673203,
"loss": 4.1936,
"step": 3750
},
{
"epoch": 0.40899795501022496,
"grad_norm": 0.6680869460105896,
"learning_rate": 0.0005760995582372589,
"loss": 4.1999,
"step": 3800
},
{
"epoch": 0.4143795070498332,
"grad_norm": 0.650600254535675,
"learning_rate": 0.0005757763172071974,
"loss": 4.1712,
"step": 3850
},
{
"epoch": 0.4197610590894414,
"grad_norm": 0.7819272875785828,
"learning_rate": 0.000575453076177136,
"loss": 4.1907,
"step": 3900
},
{
"epoch": 0.4251426111290496,
"grad_norm": 0.8200270533561707,
"learning_rate": 0.0005751298351470746,
"loss": 4.1628,
"step": 3950
},
{
"epoch": 0.4305241631686578,
"grad_norm": 0.5963762998580933,
"learning_rate": 0.0005748065941170133,
"loss": 4.1514,
"step": 4000
},
{
"epoch": 0.4305241631686578,
"eval_accuracy": 0.31238215915235973,
"eval_loss": 4.085026741027832,
"eval_runtime": 180.1755,
"eval_samples_per_second": 99.964,
"eval_steps_per_second": 6.249,
"step": 4000
},
{
"epoch": 0.43590571520826604,
"grad_norm": 0.7472826838493347,
"learning_rate": 0.0005744833530869518,
"loss": 4.1579,
"step": 4050
},
{
"epoch": 0.44128726724787426,
"grad_norm": 0.5647183656692505,
"learning_rate": 0.0005741601120568903,
"loss": 4.153,
"step": 4100
},
{
"epoch": 0.44666881928748253,
"grad_norm": 0.6447238922119141,
"learning_rate": 0.0005738368710268289,
"loss": 4.1402,
"step": 4150
},
{
"epoch": 0.45205037132709075,
"grad_norm": 0.6352419257164001,
"learning_rate": 0.0005735136299967675,
"loss": 4.1218,
"step": 4200
},
{
"epoch": 0.45743192336669897,
"grad_norm": 0.7383701205253601,
"learning_rate": 0.0005731903889667062,
"loss": 4.1315,
"step": 4250
},
{
"epoch": 0.4628134754063072,
"grad_norm": 0.6032237410545349,
"learning_rate": 0.0005728671479366447,
"loss": 4.1318,
"step": 4300
},
{
"epoch": 0.4681950274459154,
"grad_norm": 0.7067080736160278,
"learning_rate": 0.0005725439069065833,
"loss": 4.1231,
"step": 4350
},
{
"epoch": 0.4735765794855236,
"grad_norm": 0.6998112797737122,
"learning_rate": 0.0005722206658765219,
"loss": 4.1332,
"step": 4400
},
{
"epoch": 0.47895813152513184,
"grad_norm": 0.6890502572059631,
"learning_rate": 0.0005718974248464605,
"loss": 4.0919,
"step": 4450
},
{
"epoch": 0.48433968356474005,
"grad_norm": 0.6913346648216248,
"learning_rate": 0.000571574183816399,
"loss": 4.0985,
"step": 4500
},
{
"epoch": 0.48972123560434827,
"grad_norm": 0.6454923748970032,
"learning_rate": 0.0005712509427863376,
"loss": 4.0914,
"step": 4550
},
{
"epoch": 0.49510278764395654,
"grad_norm": 0.6718611717224121,
"learning_rate": 0.0005709277017562762,
"loss": 4.0875,
"step": 4600
},
{
"epoch": 0.5004843396835648,
"grad_norm": 0.6115706562995911,
"learning_rate": 0.0005706044607262148,
"loss": 4.0789,
"step": 4650
},
{
"epoch": 0.505865891723173,
"grad_norm": 0.7557240724563599,
"learning_rate": 0.0005702812196961534,
"loss": 4.1053,
"step": 4700
},
{
"epoch": 0.5112474437627812,
"grad_norm": 0.6292441487312317,
"learning_rate": 0.0005699579786660919,
"loss": 4.0814,
"step": 4750
},
{
"epoch": 0.5166289958023894,
"grad_norm": 0.5616717338562012,
"learning_rate": 0.0005696347376360306,
"loss": 4.0577,
"step": 4800
},
{
"epoch": 0.5220105478419976,
"grad_norm": 0.5930373668670654,
"learning_rate": 0.0005693114966059691,
"loss": 4.0565,
"step": 4850
},
{
"epoch": 0.5273920998816058,
"grad_norm": 0.7059026956558228,
"learning_rate": 0.0005689882555759077,
"loss": 4.0902,
"step": 4900
},
{
"epoch": 0.5327736519212141,
"grad_norm": 0.6343038082122803,
"learning_rate": 0.0005686650145458463,
"loss": 4.0649,
"step": 4950
},
{
"epoch": 0.5381552039608223,
"grad_norm": 0.5513986945152283,
"learning_rate": 0.0005683417735157848,
"loss": 4.049,
"step": 5000
},
{
"epoch": 0.5381552039608223,
"eval_accuracy": 0.3214264255340371,
"eval_loss": 3.9884958267211914,
"eval_runtime": 180.1991,
"eval_samples_per_second": 99.951,
"eval_steps_per_second": 6.249,
"step": 5000
},
{
"epoch": 0.5435367560004305,
"grad_norm": 0.6979557871818542,
"learning_rate": 0.0005680185324857235,
"loss": 4.0494,
"step": 5050
},
{
"epoch": 0.5489183080400387,
"grad_norm": 0.7719425559043884,
"learning_rate": 0.0005676952914556621,
"loss": 4.0467,
"step": 5100
},
{
"epoch": 0.5542998600796469,
"grad_norm": 0.623746931552887,
"learning_rate": 0.0005673720504256007,
"loss": 4.0159,
"step": 5150
},
{
"epoch": 0.5596814121192552,
"grad_norm": 0.694391667842865,
"learning_rate": 0.0005670488093955392,
"loss": 4.0426,
"step": 5200
},
{
"epoch": 0.5650629641588634,
"grad_norm": 0.6758689284324646,
"learning_rate": 0.0005667255683654777,
"loss": 4.0391,
"step": 5250
},
{
"epoch": 0.5704445161984716,
"grad_norm": 0.6376744508743286,
"learning_rate": 0.0005664023273354164,
"loss": 4.0361,
"step": 5300
},
{
"epoch": 0.5758260682380799,
"grad_norm": 0.6066127419471741,
"learning_rate": 0.0005660790863053549,
"loss": 4.0327,
"step": 5350
},
{
"epoch": 0.5812076202776881,
"grad_norm": 0.5952569246292114,
"learning_rate": 0.0005657558452752936,
"loss": 4.0252,
"step": 5400
},
{
"epoch": 0.5865891723172963,
"grad_norm": 0.6313889026641846,
"learning_rate": 0.0005654326042452321,
"loss": 4.0221,
"step": 5450
},
{
"epoch": 0.5919707243569046,
"grad_norm": 0.596638023853302,
"learning_rate": 0.0005651093632151708,
"loss": 4.0425,
"step": 5500
},
{
"epoch": 0.5973522763965128,
"grad_norm": 0.6534123420715332,
"learning_rate": 0.0005647861221851093,
"loss": 4.0383,
"step": 5550
},
{
"epoch": 0.602733828436121,
"grad_norm": 0.5690531730651855,
"learning_rate": 0.0005644628811550479,
"loss": 4.0233,
"step": 5600
},
{
"epoch": 0.6081153804757292,
"grad_norm": 0.626017153263092,
"learning_rate": 0.0005641396401249865,
"loss": 3.9948,
"step": 5650
},
{
"epoch": 0.6134969325153374,
"grad_norm": 0.6084668636322021,
"learning_rate": 0.000563816399094925,
"loss": 4.0122,
"step": 5700
},
{
"epoch": 0.6188784845549457,
"grad_norm": 0.5626436471939087,
"learning_rate": 0.0005634931580648637,
"loss": 4.0019,
"step": 5750
},
{
"epoch": 0.6242600365945539,
"grad_norm": 0.5933359861373901,
"learning_rate": 0.0005631699170348022,
"loss": 3.9866,
"step": 5800
},
{
"epoch": 0.6296415886341621,
"grad_norm": 0.5561618804931641,
"learning_rate": 0.0005628466760047408,
"loss": 3.9969,
"step": 5850
},
{
"epoch": 0.6350231406737703,
"grad_norm": 0.6588941812515259,
"learning_rate": 0.0005625234349746794,
"loss": 3.9841,
"step": 5900
},
{
"epoch": 0.6404046927133785,
"grad_norm": 0.6015214920043945,
"learning_rate": 0.000562200193944618,
"loss": 3.9915,
"step": 5950
},
{
"epoch": 0.6457862447529867,
"grad_norm": 0.6395725607872009,
"learning_rate": 0.0005618769529145566,
"loss": 3.9874,
"step": 6000
},
{
"epoch": 0.6457862447529867,
"eval_accuracy": 0.32797471777684023,
"eval_loss": 3.916991710662842,
"eval_runtime": 180.0235,
"eval_samples_per_second": 100.048,
"eval_steps_per_second": 6.255,
"step": 6000
},
{
"epoch": 0.651167796792595,
"grad_norm": 0.6231560707092285,
"learning_rate": 0.0005615537118844952,
"loss": 3.9965,
"step": 6050
},
{
"epoch": 0.6565493488322032,
"grad_norm": 0.6274864077568054,
"learning_rate": 0.0005612304708544337,
"loss": 3.97,
"step": 6100
},
{
"epoch": 0.6619309008718114,
"grad_norm": 0.6003475189208984,
"learning_rate": 0.0005609072298243723,
"loss": 3.9817,
"step": 6150
},
{
"epoch": 0.6673124529114196,
"grad_norm": 0.6099907159805298,
"learning_rate": 0.0005605839887943109,
"loss": 3.9868,
"step": 6200
},
{
"epoch": 0.6726940049510278,
"grad_norm": 0.640876829624176,
"learning_rate": 0.0005602607477642495,
"loss": 3.9754,
"step": 6250
},
{
"epoch": 0.6780755569906362,
"grad_norm": 0.8232746720314026,
"learning_rate": 0.0005599375067341881,
"loss": 3.9749,
"step": 6300
},
{
"epoch": 0.6834571090302444,
"grad_norm": 0.6095024347305298,
"learning_rate": 0.0005596142657041266,
"loss": 3.9693,
"step": 6350
},
{
"epoch": 0.6888386610698526,
"grad_norm": 0.5760998725891113,
"learning_rate": 0.0005592910246740653,
"loss": 3.9533,
"step": 6400
},
{
"epoch": 0.6942202131094608,
"grad_norm": 0.6642665266990662,
"learning_rate": 0.0005589677836440038,
"loss": 3.9595,
"step": 6450
},
{
"epoch": 0.699601765149069,
"grad_norm": 0.5661361813545227,
"learning_rate": 0.0005586445426139425,
"loss": 3.9461,
"step": 6500
},
{
"epoch": 0.7049833171886772,
"grad_norm": 0.5967432856559753,
"learning_rate": 0.000558321301583881,
"loss": 3.9567,
"step": 6550
},
{
"epoch": 0.7103648692282855,
"grad_norm": 0.5816430449485779,
"learning_rate": 0.0005579980605538196,
"loss": 3.9586,
"step": 6600
},
{
"epoch": 0.7157464212678937,
"grad_norm": 0.5664747357368469,
"learning_rate": 0.0005576748195237582,
"loss": 3.9593,
"step": 6650
},
{
"epoch": 0.7211279733075019,
"grad_norm": 0.5324902534484863,
"learning_rate": 0.0005573515784936967,
"loss": 3.9369,
"step": 6700
},
{
"epoch": 0.7265095253471101,
"grad_norm": 0.8317891955375671,
"learning_rate": 0.0005570283374636353,
"loss": 3.9518,
"step": 6750
},
{
"epoch": 0.7318910773867183,
"grad_norm": 0.5899968147277832,
"learning_rate": 0.0005567050964335739,
"loss": 3.9414,
"step": 6800
},
{
"epoch": 0.7372726294263265,
"grad_norm": 0.5521121621131897,
"learning_rate": 0.0005563818554035126,
"loss": 3.9377,
"step": 6850
},
{
"epoch": 0.7426541814659348,
"grad_norm": 0.5316669940948486,
"learning_rate": 0.0005560586143734511,
"loss": 3.9372,
"step": 6900
},
{
"epoch": 0.748035733505543,
"grad_norm": 0.5491471290588379,
"learning_rate": 0.0005557353733433896,
"loss": 3.9333,
"step": 6950
},
{
"epoch": 0.7534172855451512,
"grad_norm": 0.5990865230560303,
"learning_rate": 0.0005554121323133283,
"loss": 3.937,
"step": 7000
},
{
"epoch": 0.7534172855451512,
"eval_accuracy": 0.33351840541891065,
"eval_loss": 3.861109733581543,
"eval_runtime": 180.3493,
"eval_samples_per_second": 99.867,
"eval_steps_per_second": 6.243,
"step": 7000
},
{
"epoch": 0.7587988375847594,
"grad_norm": 0.5951315760612488,
"learning_rate": 0.0005550888912832668,
"loss": 3.9214,
"step": 7050
},
{
"epoch": 0.7641803896243676,
"grad_norm": 0.6178938746452332,
"learning_rate": 0.0005547656502532055,
"loss": 3.9212,
"step": 7100
},
{
"epoch": 0.7695619416639758,
"grad_norm": 0.6568113565444946,
"learning_rate": 0.000554442409223144,
"loss": 3.917,
"step": 7150
},
{
"epoch": 0.7749434937035841,
"grad_norm": 0.6213955283164978,
"learning_rate": 0.0005541191681930826,
"loss": 3.9392,
"step": 7200
},
{
"epoch": 0.7803250457431924,
"grad_norm": 0.6062942743301392,
"learning_rate": 0.0005537959271630212,
"loss": 3.9189,
"step": 7250
},
{
"epoch": 0.7857065977828006,
"grad_norm": 0.6278512477874756,
"learning_rate": 0.0005534726861329598,
"loss": 3.908,
"step": 7300
},
{
"epoch": 0.7910881498224088,
"grad_norm": 0.5704114437103271,
"learning_rate": 0.0005531494451028983,
"loss": 3.9054,
"step": 7350
},
{
"epoch": 0.796469701862017,
"grad_norm": 0.579227089881897,
"learning_rate": 0.0005528262040728369,
"loss": 3.904,
"step": 7400
},
{
"epoch": 0.8018512539016253,
"grad_norm": 0.567130982875824,
"learning_rate": 0.0005525029630427755,
"loss": 3.8941,
"step": 7450
},
{
"epoch": 0.8072328059412335,
"grad_norm": 0.5331254005432129,
"learning_rate": 0.0005521797220127141,
"loss": 3.9149,
"step": 7500
},
{
"epoch": 0.8126143579808417,
"grad_norm": 0.5759610533714294,
"learning_rate": 0.0005518564809826527,
"loss": 3.9028,
"step": 7550
},
{
"epoch": 0.8179959100204499,
"grad_norm": 0.5558604001998901,
"learning_rate": 0.0005515332399525912,
"loss": 3.9277,
"step": 7600
},
{
"epoch": 0.8233774620600581,
"grad_norm": 0.6097841858863831,
"learning_rate": 0.0005512099989225299,
"loss": 3.8995,
"step": 7650
},
{
"epoch": 0.8287590140996663,
"grad_norm": 0.580385684967041,
"learning_rate": 0.0005508867578924685,
"loss": 3.8788,
"step": 7700
},
{
"epoch": 0.8341405661392746,
"grad_norm": 0.5963330268859863,
"learning_rate": 0.000550563516862407,
"loss": 3.87,
"step": 7750
},
{
"epoch": 0.8395221181788828,
"grad_norm": 0.6174511313438416,
"learning_rate": 0.0005502402758323456,
"loss": 3.9082,
"step": 7800
},
{
"epoch": 0.844903670218491,
"grad_norm": 0.5597873330116272,
"learning_rate": 0.0005499170348022841,
"loss": 3.8767,
"step": 7850
},
{
"epoch": 0.8502852222580992,
"grad_norm": 0.6453114151954651,
"learning_rate": 0.0005495937937722228,
"loss": 3.8749,
"step": 7900
},
{
"epoch": 0.8556667742977074,
"grad_norm": 0.6109079718589783,
"learning_rate": 0.0005492705527421614,
"loss": 3.8862,
"step": 7950
},
{
"epoch": 0.8610483263373157,
"grad_norm": 0.5893735885620117,
"learning_rate": 0.0005489473117121,
"loss": 3.8876,
"step": 8000
},
{
"epoch": 0.8610483263373157,
"eval_accuracy": 0.33768654750124977,
"eval_loss": 3.8177008628845215,
"eval_runtime": 180.2298,
"eval_samples_per_second": 99.934,
"eval_steps_per_second": 6.248,
"step": 8000
},
{
"epoch": 0.8664298783769239,
"grad_norm": 0.5334535241127014,
"learning_rate": 0.0005486240706820385,
"loss": 3.8913,
"step": 8050
},
{
"epoch": 0.8718114304165321,
"grad_norm": 0.6027869582176208,
"learning_rate": 0.0005483008296519772,
"loss": 3.8906,
"step": 8100
},
{
"epoch": 0.8771929824561403,
"grad_norm": 0.5669072270393372,
"learning_rate": 0.0005479775886219157,
"loss": 3.8675,
"step": 8150
},
{
"epoch": 0.8825745344957485,
"grad_norm": 0.5210901498794556,
"learning_rate": 0.0005476543475918542,
"loss": 3.8678,
"step": 8200
},
{
"epoch": 0.8879560865353568,
"grad_norm": 0.5795321464538574,
"learning_rate": 0.0005473311065617929,
"loss": 3.8764,
"step": 8250
},
{
"epoch": 0.8933376385749651,
"grad_norm": 0.5359807014465332,
"learning_rate": 0.0005470078655317314,
"loss": 3.87,
"step": 8300
},
{
"epoch": 0.8987191906145733,
"grad_norm": 0.5787212252616882,
"learning_rate": 0.0005466846245016701,
"loss": 3.8619,
"step": 8350
},
{
"epoch": 0.9041007426541815,
"grad_norm": 0.542587161064148,
"learning_rate": 0.0005463613834716086,
"loss": 3.8606,
"step": 8400
},
{
"epoch": 0.9094822946937897,
"grad_norm": 0.5083078742027283,
"learning_rate": 0.0005460381424415472,
"loss": 3.8736,
"step": 8450
},
{
"epoch": 0.9148638467333979,
"grad_norm": 0.5857490301132202,
"learning_rate": 0.0005457149014114858,
"loss": 3.8588,
"step": 8500
},
{
"epoch": 0.9202453987730062,
"grad_norm": 0.6556069850921631,
"learning_rate": 0.0005453916603814243,
"loss": 3.8751,
"step": 8550
},
{
"epoch": 0.9256269508126144,
"grad_norm": 0.6806515455245972,
"learning_rate": 0.000545068419351363,
"loss": 3.8549,
"step": 8600
},
{
"epoch": 0.9310085028522226,
"grad_norm": 0.5887377858161926,
"learning_rate": 0.0005447451783213015,
"loss": 3.8544,
"step": 8650
},
{
"epoch": 0.9363900548918308,
"grad_norm": 0.5550622344017029,
"learning_rate": 0.0005444219372912401,
"loss": 3.866,
"step": 8700
},
{
"epoch": 0.941771606931439,
"grad_norm": 0.5800760388374329,
"learning_rate": 0.0005440986962611787,
"loss": 3.837,
"step": 8750
},
{
"epoch": 0.9471531589710472,
"grad_norm": 0.6137863993644714,
"learning_rate": 0.0005437754552311173,
"loss": 3.8428,
"step": 8800
},
{
"epoch": 0.9525347110106555,
"grad_norm": 0.5557230710983276,
"learning_rate": 0.0005434522142010559,
"loss": 3.843,
"step": 8850
},
{
"epoch": 0.9579162630502637,
"grad_norm": 0.5875601172447205,
"learning_rate": 0.0005431289731709945,
"loss": 3.8428,
"step": 8900
},
{
"epoch": 0.9632978150898719,
"grad_norm": 0.5003922581672668,
"learning_rate": 0.000542805732140933,
"loss": 3.8504,
"step": 8950
},
{
"epoch": 0.9686793671294801,
"grad_norm": 0.6187765002250671,
"learning_rate": 0.0005424824911108716,
"loss": 3.8255,
"step": 9000
},
{
"epoch": 0.9686793671294801,
"eval_accuracy": 0.3412266758920212,
"eval_loss": 3.7763822078704834,
"eval_runtime": 180.2292,
"eval_samples_per_second": 99.934,
"eval_steps_per_second": 6.248,
"step": 9000
},
{
"epoch": 0.9740609191690883,
"grad_norm": 0.5244655609130859,
"learning_rate": 0.0005421592500808102,
"loss": 3.8325,
"step": 9050
},
{
"epoch": 0.9794424712086965,
"grad_norm": 0.6028016209602356,
"learning_rate": 0.0005418360090507488,
"loss": 3.8239,
"step": 9100
},
{
"epoch": 0.9848240232483048,
"grad_norm": 0.6184064745903015,
"learning_rate": 0.0005415127680206874,
"loss": 3.8494,
"step": 9150
},
{
"epoch": 0.9902055752879131,
"grad_norm": 0.5826838612556458,
"learning_rate": 0.0005411895269906259,
"loss": 3.8364,
"step": 9200
},
{
"epoch": 0.9955871273275213,
"grad_norm": 0.5381140112876892,
"learning_rate": 0.0005408662859605646,
"loss": 3.8343,
"step": 9250
},
{
"epoch": 1.0009686793671295,
"grad_norm": 0.7138656377792358,
"learning_rate": 0.0005405430449305031,
"loss": 3.8177,
"step": 9300
},
{
"epoch": 1.0063502314067376,
"grad_norm": 0.6079965233802795,
"learning_rate": 0.0005402198039004416,
"loss": 3.7709,
"step": 9350
},
{
"epoch": 1.011731783446346,
"grad_norm": 0.48963576555252075,
"learning_rate": 0.0005398965628703803,
"loss": 3.7634,
"step": 9400
},
{
"epoch": 1.017113335485954,
"grad_norm": 0.5836416482925415,
"learning_rate": 0.0005395733218403189,
"loss": 3.7652,
"step": 9450
},
{
"epoch": 1.0224948875255624,
"grad_norm": 0.6434711217880249,
"learning_rate": 0.0005392500808102575,
"loss": 3.777,
"step": 9500
},
{
"epoch": 1.0278764395651705,
"grad_norm": 0.5500202178955078,
"learning_rate": 0.000538926839780196,
"loss": 3.7877,
"step": 9550
},
{
"epoch": 1.0332579916047788,
"grad_norm": 0.5199800133705139,
"learning_rate": 0.0005386035987501346,
"loss": 3.774,
"step": 9600
},
{
"epoch": 1.0386395436443872,
"grad_norm": 0.5320031046867371,
"learning_rate": 0.0005382803577200732,
"loss": 3.78,
"step": 9650
},
{
"epoch": 1.0440210956839953,
"grad_norm": 0.5856112837791443,
"learning_rate": 0.0005379571166900119,
"loss": 3.7726,
"step": 9700
},
{
"epoch": 1.0494026477236036,
"grad_norm": 0.595620334148407,
"learning_rate": 0.0005376338756599504,
"loss": 3.7815,
"step": 9750
},
{
"epoch": 1.0547841997632117,
"grad_norm": 0.5389620661735535,
"learning_rate": 0.0005373106346298889,
"loss": 3.7741,
"step": 9800
},
{
"epoch": 1.06016575180282,
"grad_norm": 0.6535997986793518,
"learning_rate": 0.0005369873935998276,
"loss": 3.7702,
"step": 9850
},
{
"epoch": 1.0655473038424281,
"grad_norm": 0.6863070726394653,
"learning_rate": 0.0005366641525697661,
"loss": 3.7684,
"step": 9900
},
{
"epoch": 1.0709288558820365,
"grad_norm": 0.5795710682868958,
"learning_rate": 0.0005363409115397048,
"loss": 3.7652,
"step": 9950
},
{
"epoch": 1.0763104079216446,
"grad_norm": 0.6301584243774414,
"learning_rate": 0.0005360176705096433,
"loss": 3.7842,
"step": 10000
},
{
"epoch": 1.0763104079216446,
"eval_accuracy": 0.34461621138028176,
"eval_loss": 3.7496469020843506,
"eval_runtime": 180.3887,
"eval_samples_per_second": 99.845,
"eval_steps_per_second": 6.242,
"step": 10000
},
{
"epoch": 1.081691959961253,
"grad_norm": 0.6134398579597473,
"learning_rate": 0.0005356944294795819,
"loss": 3.7578,
"step": 10050
},
{
"epoch": 1.087073512000861,
"grad_norm": 0.5343470573425293,
"learning_rate": 0.0005353711884495205,
"loss": 3.7574,
"step": 10100
},
{
"epoch": 1.0924550640404693,
"grad_norm": 0.5602631568908691,
"learning_rate": 0.000535047947419459,
"loss": 3.7551,
"step": 10150
},
{
"epoch": 1.0978366160800774,
"grad_norm": 0.56365966796875,
"learning_rate": 0.0005347247063893976,
"loss": 3.7695,
"step": 10200
},
{
"epoch": 1.1032181681196858,
"grad_norm": 0.5774141550064087,
"learning_rate": 0.0005344014653593362,
"loss": 3.7648,
"step": 10250
},
{
"epoch": 1.1085997201592939,
"grad_norm": 0.5826075077056885,
"learning_rate": 0.0005340782243292748,
"loss": 3.758,
"step": 10300
},
{
"epoch": 1.1139812721989022,
"grad_norm": 0.5574244856834412,
"learning_rate": 0.0005337549832992134,
"loss": 3.7457,
"step": 10350
},
{
"epoch": 1.1193628242385103,
"grad_norm": 0.5108634829521179,
"learning_rate": 0.000533431742269152,
"loss": 3.7676,
"step": 10400
},
{
"epoch": 1.1247443762781186,
"grad_norm": 0.5607388019561768,
"learning_rate": 0.0005331085012390905,
"loss": 3.7618,
"step": 10450
},
{
"epoch": 1.1301259283177267,
"grad_norm": 0.5880976319313049,
"learning_rate": 0.0005327852602090292,
"loss": 3.7593,
"step": 10500
},
{
"epoch": 1.135507480357335,
"grad_norm": 0.5758063793182373,
"learning_rate": 0.0005324620191789678,
"loss": 3.7638,
"step": 10550
},
{
"epoch": 1.1408890323969434,
"grad_norm": 0.5253919959068298,
"learning_rate": 0.0005321387781489063,
"loss": 3.7666,
"step": 10600
},
{
"epoch": 1.1462705844365515,
"grad_norm": 0.5626659393310547,
"learning_rate": 0.0005318155371188449,
"loss": 3.734,
"step": 10650
},
{
"epoch": 1.1516521364761596,
"grad_norm": 0.5251621603965759,
"learning_rate": 0.0005314922960887834,
"loss": 3.7614,
"step": 10700
},
{
"epoch": 1.157033688515768,
"grad_norm": 0.6023606061935425,
"learning_rate": 0.0005311690550587221,
"loss": 3.768,
"step": 10750
},
{
"epoch": 1.1624152405553763,
"grad_norm": 0.5632306933403015,
"learning_rate": 0.0005308458140286607,
"loss": 3.7378,
"step": 10800
},
{
"epoch": 1.1677967925949844,
"grad_norm": 0.5231441855430603,
"learning_rate": 0.0005305225729985993,
"loss": 3.7404,
"step": 10850
},
{
"epoch": 1.1731783446345927,
"grad_norm": 0.567490816116333,
"learning_rate": 0.0005301993319685378,
"loss": 3.7422,
"step": 10900
},
{
"epoch": 1.1785598966742008,
"grad_norm": 0.5332131385803223,
"learning_rate": 0.0005298760909384765,
"loss": 3.7561,
"step": 10950
},
{
"epoch": 1.1839414487138091,
"grad_norm": 0.5323898792266846,
"learning_rate": 0.000529552849908415,
"loss": 3.7212,
"step": 11000
},
{
"epoch": 1.1839414487138091,
"eval_accuracy": 0.34692530255211507,
"eval_loss": 3.722130298614502,
"eval_runtime": 180.2165,
"eval_samples_per_second": 99.941,
"eval_steps_per_second": 6.248,
"step": 11000
},
{
"epoch": 1.1893230007534172,
"grad_norm": 0.5329837203025818,
"learning_rate": 0.0005292296088783535,
"loss": 3.7392,
"step": 11050
},
{
"epoch": 1.1947045527930256,
"grad_norm": 0.6097086071968079,
"learning_rate": 0.0005289063678482922,
"loss": 3.7493,
"step": 11100
},
{
"epoch": 1.2000861048326337,
"grad_norm": 0.5243571996688843,
"learning_rate": 0.0005285831268182307,
"loss": 3.7329,
"step": 11150
},
{
"epoch": 1.205467656872242,
"grad_norm": 0.5530942678451538,
"learning_rate": 0.0005282598857881694,
"loss": 3.7355,
"step": 11200
},
{
"epoch": 1.21084920891185,
"grad_norm": 0.5588470101356506,
"learning_rate": 0.0005279366447581079,
"loss": 3.7601,
"step": 11250
},
{
"epoch": 1.2162307609514584,
"grad_norm": 0.5571747422218323,
"learning_rate": 0.0005276134037280465,
"loss": 3.7339,
"step": 11300
},
{
"epoch": 1.2216123129910665,
"grad_norm": 0.5556563138961792,
"learning_rate": 0.0005272901626979851,
"loss": 3.7238,
"step": 11350
},
{
"epoch": 1.2269938650306749,
"grad_norm": 0.6499513387680054,
"learning_rate": 0.0005269669216679236,
"loss": 3.7475,
"step": 11400
},
{
"epoch": 1.232375417070283,
"grad_norm": 0.5881815552711487,
"learning_rate": 0.0005266436806378623,
"loss": 3.7419,
"step": 11450
},
{
"epoch": 1.2377569691098913,
"grad_norm": 0.5554115772247314,
"learning_rate": 0.0005263204396078008,
"loss": 3.7435,
"step": 11500
},
{
"epoch": 1.2431385211494996,
"grad_norm": 0.562907338142395,
"learning_rate": 0.0005259971985777394,
"loss": 3.7435,
"step": 11550
},
{
"epoch": 1.2485200731891077,
"grad_norm": 0.5462493896484375,
"learning_rate": 0.000525673957547678,
"loss": 3.7375,
"step": 11600
},
{
"epoch": 1.2539016252287158,
"grad_norm": 0.5715740323066711,
"learning_rate": 0.0005253507165176167,
"loss": 3.732,
"step": 11650
},
{
"epoch": 1.2592831772683242,
"grad_norm": 0.5162888169288635,
"learning_rate": 0.0005250274754875552,
"loss": 3.7298,
"step": 11700
},
{
"epoch": 1.2646647293079325,
"grad_norm": 0.5549124479293823,
"learning_rate": 0.0005247042344574938,
"loss": 3.73,
"step": 11750
},
{
"epoch": 1.2700462813475406,
"grad_norm": 0.5421045422554016,
"learning_rate": 0.0005243809934274323,
"loss": 3.7321,
"step": 11800
},
{
"epoch": 1.275427833387149,
"grad_norm": 0.5627176761627197,
"learning_rate": 0.0005240577523973709,
"loss": 3.7224,
"step": 11850
},
{
"epoch": 1.280809385426757,
"grad_norm": 0.5783089995384216,
"learning_rate": 0.0005237345113673095,
"loss": 3.7265,
"step": 11900
},
{
"epoch": 1.2861909374663654,
"grad_norm": 0.6225681900978088,
"learning_rate": 0.0005234112703372481,
"loss": 3.7381,
"step": 11950
},
{
"epoch": 1.2915724895059735,
"grad_norm": 0.5271321535110474,
"learning_rate": 0.0005230880293071867,
"loss": 3.7296,
"step": 12000
},
{
"epoch": 1.2915724895059735,
"eval_accuracy": 0.3496377132435158,
"eval_loss": 3.698834180831909,
"eval_runtime": 179.7345,
"eval_samples_per_second": 100.209,
"eval_steps_per_second": 6.265,
"step": 12000
},
{
"epoch": 1.2969540415455818,
"grad_norm": 0.5290063619613647,
"learning_rate": 0.0005227647882771253,
"loss": 3.7369,
"step": 12050
},
{
"epoch": 1.30233559358519,
"grad_norm": 0.5177344679832458,
"learning_rate": 0.0005224415472470639,
"loss": 3.7374,
"step": 12100
},
{
"epoch": 1.3077171456247982,
"grad_norm": 0.5714365839958191,
"learning_rate": 0.0005221247710376037,
"loss": 3.7489,
"step": 12150
},
{
"epoch": 1.3130986976644063,
"grad_norm": 0.5795943140983582,
"learning_rate": 0.0005218015300075422,
"loss": 3.7384,
"step": 12200
},
{
"epoch": 1.3184802497040147,
"grad_norm": 0.5433863401412964,
"learning_rate": 0.0005214782889774809,
"loss": 3.7228,
"step": 12250
},
{
"epoch": 1.3238618017436228,
"grad_norm": 0.5290769338607788,
"learning_rate": 0.0005211550479474194,
"loss": 3.7153,
"step": 12300
},
{
"epoch": 1.329243353783231,
"grad_norm": 0.6278223395347595,
"learning_rate": 0.000520831806917358,
"loss": 3.7361,
"step": 12350
},
{
"epoch": 1.3346249058228392,
"grad_norm": 0.5160257816314697,
"learning_rate": 0.0005205085658872966,
"loss": 3.6936,
"step": 12400
},
{
"epoch": 1.3400064578624475,
"grad_norm": 0.591951310634613,
"learning_rate": 0.0005201853248572351,
"loss": 3.7249,
"step": 12450
},
{
"epoch": 1.3453880099020559,
"grad_norm": 0.5568095445632935,
"learning_rate": 0.0005198620838271738,
"loss": 3.726,
"step": 12500
},
{
"epoch": 1.350769561941664,
"grad_norm": 0.5874963998794556,
"learning_rate": 0.0005195388427971123,
"loss": 3.7238,
"step": 12550
},
{
"epoch": 1.356151113981272,
"grad_norm": 0.5767421126365662,
"learning_rate": 0.000519215601767051,
"loss": 3.7266,
"step": 12600
},
{
"epoch": 1.3615326660208804,
"grad_norm": 0.5423089265823364,
"learning_rate": 0.0005188923607369895,
"loss": 3.7133,
"step": 12650
},
{
"epoch": 1.3669142180604887,
"grad_norm": 0.562261164188385,
"learning_rate": 0.000518569119706928,
"loss": 3.7204,
"step": 12700
},
{
"epoch": 1.3722957701000968,
"grad_norm": 0.5939424633979797,
"learning_rate": 0.0005182458786768667,
"loss": 3.7343,
"step": 12750
},
{
"epoch": 1.3776773221397052,
"grad_norm": 0.5418099164962769,
"learning_rate": 0.0005179226376468052,
"loss": 3.7304,
"step": 12800
},
{
"epoch": 1.3830588741793133,
"grad_norm": 0.5787643790245056,
"learning_rate": 0.0005175993966167438,
"loss": 3.7152,
"step": 12850
},
{
"epoch": 1.3884404262189216,
"grad_norm": 0.5920687913894653,
"learning_rate": 0.0005172761555866824,
"loss": 3.7025,
"step": 12900
},
{
"epoch": 1.3938219782585297,
"grad_norm": 0.5622273683547974,
"learning_rate": 0.0005169529145566211,
"loss": 3.7297,
"step": 12950
},
{
"epoch": 1.399203530298138,
"grad_norm": 0.5593236684799194,
"learning_rate": 0.0005166296735265596,
"loss": 3.6772,
"step": 13000
},
{
"epoch": 1.399203530298138,
"eval_accuracy": 0.351960929290765,
"eval_loss": 3.674530506134033,
"eval_runtime": 180.1009,
"eval_samples_per_second": 100.005,
"eval_steps_per_second": 6.252,
"step": 13000
},
{
"epoch": 1.4045850823377461,
"grad_norm": 0.551760733127594,
"learning_rate": 0.0005163064324964982,
"loss": 3.7236,
"step": 13050
},
{
"epoch": 1.4099666343773545,
"grad_norm": 0.48447367548942566,
"learning_rate": 0.0005159831914664367,
"loss": 3.7066,
"step": 13100
},
{
"epoch": 1.4153481864169626,
"grad_norm": 0.5610434412956238,
"learning_rate": 0.0005156599504363753,
"loss": 3.7186,
"step": 13150
},
{
"epoch": 1.420729738456571,
"grad_norm": 0.6366670727729797,
"learning_rate": 0.000515336709406314,
"loss": 3.7192,
"step": 13200
},
{
"epoch": 1.426111290496179,
"grad_norm": 0.6398788094520569,
"learning_rate": 0.0005150134683762525,
"loss": 3.7288,
"step": 13250
},
{
"epoch": 1.4314928425357873,
"grad_norm": 0.5352592468261719,
"learning_rate": 0.0005146902273461911,
"loss": 3.7099,
"step": 13300
},
{
"epoch": 1.4368743945753955,
"grad_norm": 0.5241870284080505,
"learning_rate": 0.0005143669863161297,
"loss": 3.7291,
"step": 13350
},
{
"epoch": 1.4422559466150038,
"grad_norm": 0.5873683094978333,
"learning_rate": 0.0005140437452860683,
"loss": 3.7056,
"step": 13400
},
{
"epoch": 1.447637498654612,
"grad_norm": 0.5123621225357056,
"learning_rate": 0.0005137205042560069,
"loss": 3.6975,
"step": 13450
},
{
"epoch": 1.4530190506942202,
"grad_norm": 0.5494991540908813,
"learning_rate": 0.0005133972632259455,
"loss": 3.7057,
"step": 13500
},
{
"epoch": 1.4584006027338283,
"grad_norm": 0.5739464163780212,
"learning_rate": 0.0005130804870164853,
"loss": 3.6991,
"step": 13550
},
{
"epoch": 1.4637821547734367,
"grad_norm": 0.5516660809516907,
"learning_rate": 0.0005127572459864238,
"loss": 3.7064,
"step": 13600
},
{
"epoch": 1.469163706813045,
"grad_norm": 0.5697253346443176,
"learning_rate": 0.0005124340049563624,
"loss": 3.6926,
"step": 13650
},
{
"epoch": 1.474545258852653,
"grad_norm": 0.5626788139343262,
"learning_rate": 0.000512110763926301,
"loss": 3.677,
"step": 13700
},
{
"epoch": 1.4799268108922612,
"grad_norm": 0.5842751860618591,
"learning_rate": 0.0005117875228962395,
"loss": 3.6764,
"step": 13750
},
{
"epoch": 1.4853083629318695,
"grad_norm": 0.5150489211082458,
"learning_rate": 0.0005114642818661782,
"loss": 3.6995,
"step": 13800
},
{
"epoch": 1.4906899149714778,
"grad_norm": 0.5614669322967529,
"learning_rate": 0.0005111410408361168,
"loss": 3.6865,
"step": 13850
},
{
"epoch": 1.496071467011086,
"grad_norm": 0.5543150901794434,
"learning_rate": 0.0005108177998060554,
"loss": 3.6665,
"step": 13900
},
{
"epoch": 1.501453019050694,
"grad_norm": 0.6134596467018127,
"learning_rate": 0.0005104945587759939,
"loss": 3.7014,
"step": 13950
},
{
"epoch": 1.5068345710903024,
"grad_norm": 0.564272403717041,
"learning_rate": 0.0005101713177459324,
"loss": 3.6926,
"step": 14000
},
{
"epoch": 1.5068345710903024,
"eval_accuracy": 0.353645918274992,
"eval_loss": 3.6572258472442627,
"eval_runtime": 179.942,
"eval_samples_per_second": 100.093,
"eval_steps_per_second": 6.258,
"step": 14000
},
{
"epoch": 1.5122161231299107,
"grad_norm": 0.5460367202758789,
"learning_rate": 0.0005098480767158711,
"loss": 3.7094,
"step": 14050
},
{
"epoch": 1.5175976751695188,
"grad_norm": 0.5200446844100952,
"learning_rate": 0.0005095248356858097,
"loss": 3.6915,
"step": 14100
},
{
"epoch": 1.5229792272091272,
"grad_norm": 0.5343534350395203,
"learning_rate": 0.0005092015946557483,
"loss": 3.6877,
"step": 14150
},
{
"epoch": 1.5283607792487355,
"grad_norm": 0.5287026166915894,
"learning_rate": 0.0005088783536256868,
"loss": 3.6859,
"step": 14200
},
{
"epoch": 1.5337423312883436,
"grad_norm": 0.5682664513587952,
"learning_rate": 0.0005085551125956255,
"loss": 3.6866,
"step": 14250
},
{
"epoch": 1.5391238833279517,
"grad_norm": 0.5513389110565186,
"learning_rate": 0.000508231871565564,
"loss": 3.6758,
"step": 14300
},
{
"epoch": 1.54450543536756,
"grad_norm": 0.5954504013061523,
"learning_rate": 0.0005079086305355026,
"loss": 3.6807,
"step": 14350
},
{
"epoch": 1.5498869874071683,
"grad_norm": 0.534307062625885,
"learning_rate": 0.0005075853895054412,
"loss": 3.7011,
"step": 14400
},
{
"epoch": 1.5552685394467765,
"grad_norm": 0.5764222145080566,
"learning_rate": 0.0005072621484753797,
"loss": 3.6999,
"step": 14450
},
{
"epoch": 1.5606500914863846,
"grad_norm": 0.5494203567504883,
"learning_rate": 0.0005069389074453184,
"loss": 3.6871,
"step": 14500
},
{
"epoch": 1.566031643525993,
"grad_norm": 0.5286200046539307,
"learning_rate": 0.0005066156664152569,
"loss": 3.6961,
"step": 14550
},
{
"epoch": 1.5714131955656012,
"grad_norm": 0.5253137350082397,
"learning_rate": 0.0005062924253851955,
"loss": 3.6849,
"step": 14600
},
{
"epoch": 1.5767947476052093,
"grad_norm": 0.5719892978668213,
"learning_rate": 0.0005059691843551341,
"loss": 3.6662,
"step": 14650
},
{
"epoch": 1.5821762996448174,
"grad_norm": 0.6287044882774353,
"learning_rate": 0.0005056459433250727,
"loss": 3.6673,
"step": 14700
},
{
"epoch": 1.5875578516844258,
"grad_norm": 0.5336293578147888,
"learning_rate": 0.0005053227022950113,
"loss": 3.6737,
"step": 14750
},
{
"epoch": 1.592939403724034,
"grad_norm": 0.5435810089111328,
"learning_rate": 0.0005049994612649499,
"loss": 3.6794,
"step": 14800
},
{
"epoch": 1.5983209557636422,
"grad_norm": 0.6014664769172668,
"learning_rate": 0.0005046762202348884,
"loss": 3.6995,
"step": 14850
},
{
"epoch": 1.6037025078032503,
"grad_norm": 0.5819555521011353,
"learning_rate": 0.000504352979204827,
"loss": 3.6859,
"step": 14900
},
{
"epoch": 1.6090840598428586,
"grad_norm": 0.541706919670105,
"learning_rate": 0.0005040297381747656,
"loss": 3.6838,
"step": 14950
},
{
"epoch": 1.614465611882467,
"grad_norm": 0.6835187077522278,
"learning_rate": 0.0005037064971447042,
"loss": 3.6872,
"step": 15000
},
{
"epoch": 1.614465611882467,
"eval_accuracy": 0.35586591407881746,
"eval_loss": 3.635637044906616,
"eval_runtime": 179.85,
"eval_samples_per_second": 100.145,
"eval_steps_per_second": 6.261,
"step": 15000
},
{
"epoch": 1.619847163922075,
"grad_norm": 0.5497548580169678,
"learning_rate": 0.0005033832561146428,
"loss": 3.6633,
"step": 15050
},
{
"epoch": 1.6252287159616834,
"grad_norm": 0.5530732274055481,
"learning_rate": 0.0005030600150845813,
"loss": 3.6751,
"step": 15100
},
{
"epoch": 1.6306102680012917,
"grad_norm": 0.5972110629081726,
"learning_rate": 0.00050273677405452,
"loss": 3.6909,
"step": 15150
},
{
"epoch": 1.6359918200408998,
"grad_norm": 0.5513889193534851,
"learning_rate": 0.0005024135330244585,
"loss": 3.6774,
"step": 15200
},
{
"epoch": 1.641373372080508,
"grad_norm": 0.5464329719543457,
"learning_rate": 0.0005020902919943972,
"loss": 3.6562,
"step": 15250
},
{
"epoch": 1.6467549241201163,
"grad_norm": 0.621970534324646,
"learning_rate": 0.0005017670509643357,
"loss": 3.6929,
"step": 15300
},
{
"epoch": 1.6521364761597246,
"grad_norm": 0.5582807064056396,
"learning_rate": 0.0005014438099342743,
"loss": 3.6783,
"step": 15350
},
{
"epoch": 1.6575180281993327,
"grad_norm": 0.5590953230857849,
"learning_rate": 0.0005011205689042129,
"loss": 3.6581,
"step": 15400
},
{
"epoch": 1.6628995802389408,
"grad_norm": 0.5335003137588501,
"learning_rate": 0.0005007973278741514,
"loss": 3.6591,
"step": 15450
},
{
"epoch": 1.6682811322785491,
"grad_norm": 0.5382925868034363,
"learning_rate": 0.00050047408684409,
"loss": 3.6654,
"step": 15500
},
{
"epoch": 1.6736626843181575,
"grad_norm": 0.5382738709449768,
"learning_rate": 0.0005001508458140286,
"loss": 3.6541,
"step": 15550
},
{
"epoch": 1.6790442363577656,
"grad_norm": 0.5367087721824646,
"learning_rate": 0.0004998276047839673,
"loss": 3.6594,
"step": 15600
},
{
"epoch": 1.6844257883973737,
"grad_norm": 0.5150970816612244,
"learning_rate": 0.0004995043637539058,
"loss": 3.6849,
"step": 15650
},
{
"epoch": 1.689807340436982,
"grad_norm": 0.5762928128242493,
"learning_rate": 0.0004991811227238443,
"loss": 3.6711,
"step": 15700
},
{
"epoch": 1.6951888924765903,
"grad_norm": 0.48318615555763245,
"learning_rate": 0.0004988578816937829,
"loss": 3.6631,
"step": 15750
},
{
"epoch": 1.7005704445161984,
"grad_norm": 0.5551878213882446,
"learning_rate": 0.0004985346406637215,
"loss": 3.6544,
"step": 15800
},
{
"epoch": 1.7059519965558065,
"grad_norm": 0.5169930458068848,
"learning_rate": 0.0004982113996336602,
"loss": 3.6696,
"step": 15850
},
{
"epoch": 1.7113335485954149,
"grad_norm": 0.592219352722168,
"learning_rate": 0.0004978881586035987,
"loss": 3.6806,
"step": 15900
},
{
"epoch": 1.7167151006350232,
"grad_norm": 0.492858350276947,
"learning_rate": 0.0004975649175735373,
"loss": 3.6541,
"step": 15950
},
{
"epoch": 1.7220966526746313,
"grad_norm": 0.5765631198883057,
"learning_rate": 0.0004972416765434759,
"loss": 3.6757,
"step": 16000
},
{
"epoch": 1.7220966526746313,
"eval_accuracy": 0.3572902447851775,
"eval_loss": 3.61974835395813,
"eval_runtime": 179.8687,
"eval_samples_per_second": 100.134,
"eval_steps_per_second": 6.26,
"step": 16000
},
{
"epoch": 1.7274782047142396,
"grad_norm": 0.6189604997634888,
"learning_rate": 0.0004969184355134145,
"loss": 3.6559,
"step": 16050
},
{
"epoch": 1.732859756753848,
"grad_norm": 0.5432648658752441,
"learning_rate": 0.0004965951944833531,
"loss": 3.6684,
"step": 16100
},
{
"epoch": 1.738241308793456,
"grad_norm": 0.6157057881355286,
"learning_rate": 0.0004962719534532916,
"loss": 3.6691,
"step": 16150
},
{
"epoch": 1.7436228608330642,
"grad_norm": 0.6409887671470642,
"learning_rate": 0.0004959487124232302,
"loss": 3.6738,
"step": 16200
},
{
"epoch": 1.7490044128726725,
"grad_norm": 0.5156955122947693,
"learning_rate": 0.0004956254713931688,
"loss": 3.673,
"step": 16250
},
{
"epoch": 1.7543859649122808,
"grad_norm": 0.5291411280632019,
"learning_rate": 0.0004953022303631074,
"loss": 3.6519,
"step": 16300
},
{
"epoch": 1.759767516951889,
"grad_norm": 0.7061535120010376,
"learning_rate": 0.0004949789893330459,
"loss": 3.656,
"step": 16350
},
{
"epoch": 1.765149068991497,
"grad_norm": 0.5498104095458984,
"learning_rate": 0.0004946557483029846,
"loss": 3.6604,
"step": 16400
},
{
"epoch": 1.7705306210311054,
"grad_norm": 0.560353696346283,
"learning_rate": 0.0004943325072729231,
"loss": 3.6492,
"step": 16450
},
{
"epoch": 1.7759121730707137,
"grad_norm": 0.5988339185714722,
"learning_rate": 0.0004940092662428617,
"loss": 3.6731,
"step": 16500
},
{
"epoch": 1.7812937251103218,
"grad_norm": 0.5460607409477234,
"learning_rate": 0.0004936860252128003,
"loss": 3.6579,
"step": 16550
},
{
"epoch": 1.78667527714993,
"grad_norm": 0.5597132444381714,
"learning_rate": 0.0004933627841827388,
"loss": 3.6591,
"step": 16600
},
{
"epoch": 1.7920568291895382,
"grad_norm": 0.589875340461731,
"learning_rate": 0.0004930395431526775,
"loss": 3.6633,
"step": 16650
},
{
"epoch": 1.7974383812291466,
"grad_norm": 0.5566060543060303,
"learning_rate": 0.0004927163021226161,
"loss": 3.648,
"step": 16700
},
{
"epoch": 1.8028199332687547,
"grad_norm": 0.5171520709991455,
"learning_rate": 0.0004923930610925547,
"loss": 3.646,
"step": 16750
},
{
"epoch": 1.8082014853083628,
"grad_norm": 0.5987152457237244,
"learning_rate": 0.0004920698200624932,
"loss": 3.6482,
"step": 16800
},
{
"epoch": 1.813583037347971,
"grad_norm": 0.5647301077842712,
"learning_rate": 0.0004917465790324317,
"loss": 3.624,
"step": 16850
},
{
"epoch": 1.8189645893875794,
"grad_norm": 0.5267931818962097,
"learning_rate": 0.0004914233380023704,
"loss": 3.6343,
"step": 16900
},
{
"epoch": 1.8243461414271875,
"grad_norm": 0.5864246487617493,
"learning_rate": 0.0004911000969723089,
"loss": 3.6659,
"step": 16950
},
{
"epoch": 1.8297276934667959,
"grad_norm": 0.5181066393852234,
"learning_rate": 0.0004907768559422476,
"loss": 3.6485,
"step": 17000
},
{
"epoch": 1.8297276934667959,
"eval_accuracy": 0.3588785326992496,
"eval_loss": 3.603902816772461,
"eval_runtime": 180.6484,
"eval_samples_per_second": 99.702,
"eval_steps_per_second": 6.233,
"step": 17000
},
{
"epoch": 1.8351092455064042,
"grad_norm": 0.5298339128494263,
"learning_rate": 0.0004904536149121861,
"loss": 3.6723,
"step": 17050
},
{
"epoch": 1.8404907975460123,
"grad_norm": 0.5916500091552734,
"learning_rate": 0.0004901303738821248,
"loss": 3.643,
"step": 17100
},
{
"epoch": 1.8458723495856204,
"grad_norm": 0.6232360601425171,
"learning_rate": 0.0004898071328520633,
"loss": 3.6451,
"step": 17150
},
{
"epoch": 1.8512539016252287,
"grad_norm": 0.5699940323829651,
"learning_rate": 0.0004894838918220019,
"loss": 3.6625,
"step": 17200
},
{
"epoch": 1.856635453664837,
"grad_norm": 0.5958313345909119,
"learning_rate": 0.0004891606507919405,
"loss": 3.6632,
"step": 17250
},
{
"epoch": 1.8620170057044452,
"grad_norm": 0.5603935718536377,
"learning_rate": 0.000488837409761879,
"loss": 3.6399,
"step": 17300
},
{
"epoch": 1.8673985577440533,
"grad_norm": 0.5688992738723755,
"learning_rate": 0.0004885141687318177,
"loss": 3.6626,
"step": 17350
},
{
"epoch": 1.8727801097836616,
"grad_norm": 0.5361830592155457,
"learning_rate": 0.00048819092770175623,
"loss": 3.6463,
"step": 17400
},
{
"epoch": 1.87816166182327,
"grad_norm": 0.5704878568649292,
"learning_rate": 0.0004878676866716948,
"loss": 3.6309,
"step": 17450
},
{
"epoch": 1.883543213862878,
"grad_norm": 0.5955484509468079,
"learning_rate": 0.00048754444564163337,
"loss": 3.6267,
"step": 17500
},
{
"epoch": 1.8889247659024861,
"grad_norm": 0.5311874151229858,
"learning_rate": 0.000487221204611572,
"loss": 3.6326,
"step": 17550
},
{
"epoch": 1.8943063179420945,
"grad_norm": 0.5513229966163635,
"learning_rate": 0.00048689796358151056,
"loss": 3.6471,
"step": 17600
},
{
"epoch": 1.8996878699817028,
"grad_norm": 0.560566782951355,
"learning_rate": 0.00048657472255144915,
"loss": 3.6473,
"step": 17650
},
{
"epoch": 1.905069422021311,
"grad_norm": 0.5173212885856628,
"learning_rate": 0.00048625794634198896,
"loss": 3.6304,
"step": 17700
},
{
"epoch": 1.910450974060919,
"grad_norm": 0.550670862197876,
"learning_rate": 0.00048593470531192756,
"loss": 3.62,
"step": 17750
},
{
"epoch": 1.9158325261005273,
"grad_norm": 0.5653765797615051,
"learning_rate": 0.00048561146428186615,
"loss": 3.6364,
"step": 17800
},
{
"epoch": 1.9212140781401357,
"grad_norm": 0.4973013401031494,
"learning_rate": 0.0004852882232518047,
"loss": 3.623,
"step": 17850
},
{
"epoch": 1.9265956301797438,
"grad_norm": 0.5724383592605591,
"learning_rate": 0.00048496498222174334,
"loss": 3.6446,
"step": 17900
},
{
"epoch": 1.931977182219352,
"grad_norm": 0.49556764960289,
"learning_rate": 0.00048464174119168193,
"loss": 3.6491,
"step": 17950
},
{
"epoch": 1.9373587342589604,
"grad_norm": 0.5692288875579834,
"learning_rate": 0.0004843185001616205,
"loss": 3.6324,
"step": 18000
},
{
"epoch": 1.9373587342589604,
"eval_accuracy": 0.3603584936841706,
"eval_loss": 3.5903141498565674,
"eval_runtime": 179.9609,
"eval_samples_per_second": 100.083,
"eval_steps_per_second": 6.257,
"step": 18000
},
{
"epoch": 1.9427402862985685,
"grad_norm": 0.5478717088699341,
"learning_rate": 0.00048399525913155907,
"loss": 3.6503,
"step": 18050
},
{
"epoch": 1.9481218383381766,
"grad_norm": 0.5264731049537659,
"learning_rate": 0.0004836720181014976,
"loss": 3.6285,
"step": 18100
},
{
"epoch": 1.953503390377785,
"grad_norm": 0.5305705070495605,
"learning_rate": 0.0004833487770714362,
"loss": 3.6198,
"step": 18150
},
{
"epoch": 1.9588849424173933,
"grad_norm": 0.5922155976295471,
"learning_rate": 0.00048302553604137485,
"loss": 3.6275,
"step": 18200
},
{
"epoch": 1.9642664944570014,
"grad_norm": 0.5468623042106628,
"learning_rate": 0.0004827022950113134,
"loss": 3.6447,
"step": 18250
},
{
"epoch": 1.9696480464966095,
"grad_norm": 0.5142163038253784,
"learning_rate": 0.000482379053981252,
"loss": 3.6241,
"step": 18300
},
{
"epoch": 1.9750295985362178,
"grad_norm": 0.5266579389572144,
"learning_rate": 0.0004820558129511906,
"loss": 3.6267,
"step": 18350
},
{
"epoch": 1.9804111505758262,
"grad_norm": 0.5678869485855103,
"learning_rate": 0.0004817325719211291,
"loss": 3.6415,
"step": 18400
},
{
"epoch": 1.9857927026154343,
"grad_norm": 0.5284473299980164,
"learning_rate": 0.0004814093308910677,
"loss": 3.624,
"step": 18450
},
{
"epoch": 1.9911742546550424,
"grad_norm": 0.534938395023346,
"learning_rate": 0.00048108608986100637,
"loss": 3.6337,
"step": 18500
},
{
"epoch": 1.9965558066946507,
"grad_norm": 0.5706211924552917,
"learning_rate": 0.0004807628488309449,
"loss": 3.6429,
"step": 18550
},
{
"epoch": 2.001937358734259,
"grad_norm": 0.5797092914581299,
"learning_rate": 0.0004804396078008835,
"loss": 3.5995,
"step": 18600
},
{
"epoch": 2.007318910773867,
"grad_norm": 0.54485023021698,
"learning_rate": 0.00048011636677082204,
"loss": 3.5261,
"step": 18650
},
{
"epoch": 2.0127004628134753,
"grad_norm": 0.5473525524139404,
"learning_rate": 0.00047979312574076064,
"loss": 3.5536,
"step": 18700
},
{
"epoch": 2.018082014853084,
"grad_norm": 0.556251049041748,
"learning_rate": 0.0004794698847106992,
"loss": 3.5245,
"step": 18750
},
{
"epoch": 2.023463566892692,
"grad_norm": 0.5699254870414734,
"learning_rate": 0.0004791466436806378,
"loss": 3.5467,
"step": 18800
},
{
"epoch": 2.0288451189323,
"grad_norm": 0.5414472222328186,
"learning_rate": 0.0004788234026505764,
"loss": 3.5511,
"step": 18850
},
{
"epoch": 2.034226670971908,
"grad_norm": 0.5470781922340393,
"learning_rate": 0.00047850016162051496,
"loss": 3.5416,
"step": 18900
},
{
"epoch": 2.0396082230115167,
"grad_norm": 0.5483036041259766,
"learning_rate": 0.00047817692059045356,
"loss": 3.5396,
"step": 18950
},
{
"epoch": 2.044989775051125,
"grad_norm": 0.5391370058059692,
"learning_rate": 0.00047785367956039215,
"loss": 3.5384,
"step": 19000
},
{
"epoch": 2.044989775051125,
"eval_accuracy": 0.36190886174039544,
"eval_loss": 3.5802509784698486,
"eval_runtime": 179.9777,
"eval_samples_per_second": 100.074,
"eval_steps_per_second": 6.256,
"step": 19000
},
{
"epoch": 2.050371327090733,
"grad_norm": 0.6000381708145142,
"learning_rate": 0.00047753043853033075,
"loss": 3.5573,
"step": 19050
},
{
"epoch": 2.055752879130341,
"grad_norm": 0.5803197026252747,
"learning_rate": 0.00047720719750026934,
"loss": 3.5533,
"step": 19100
},
{
"epoch": 2.0611344311699495,
"grad_norm": 0.6103200912475586,
"learning_rate": 0.00047688395647020793,
"loss": 3.5525,
"step": 19150
},
{
"epoch": 2.0665159832095576,
"grad_norm": 0.5805879831314087,
"learning_rate": 0.0004765607154401465,
"loss": 3.5686,
"step": 19200
},
{
"epoch": 2.0718975352491658,
"grad_norm": 0.5827572345733643,
"learning_rate": 0.00047623747441008507,
"loss": 3.5494,
"step": 19250
},
{
"epoch": 2.0772790872887743,
"grad_norm": 0.6824135184288025,
"learning_rate": 0.0004759142333800236,
"loss": 3.551,
"step": 19300
},
{
"epoch": 2.0826606393283824,
"grad_norm": 0.5492932200431824,
"learning_rate": 0.00047559099234996226,
"loss": 3.5614,
"step": 19350
},
{
"epoch": 2.0880421913679905,
"grad_norm": 0.5775820016860962,
"learning_rate": 0.00047526775131990085,
"loss": 3.5566,
"step": 19400
},
{
"epoch": 2.0934237434075986,
"grad_norm": 0.5697767734527588,
"learning_rate": 0.0004749445102898394,
"loss": 3.5567,
"step": 19450
},
{
"epoch": 2.098805295447207,
"grad_norm": 0.5625877380371094,
"learning_rate": 0.000474621269259778,
"loss": 3.5373,
"step": 19500
},
{
"epoch": 2.1041868474868153,
"grad_norm": 0.5145209431648254,
"learning_rate": 0.0004742980282297166,
"loss": 3.5329,
"step": 19550
},
{
"epoch": 2.1095683995264234,
"grad_norm": 0.5731906294822693,
"learning_rate": 0.0004739747871996551,
"loss": 3.5426,
"step": 19600
},
{
"epoch": 2.1149499515660315,
"grad_norm": 0.5850881934165955,
"learning_rate": 0.00047365154616959377,
"loss": 3.536,
"step": 19650
},
{
"epoch": 2.12033150360564,
"grad_norm": 0.5699858069419861,
"learning_rate": 0.00047332830513953237,
"loss": 3.5535,
"step": 19700
},
{
"epoch": 2.125713055645248,
"grad_norm": 0.6195251941680908,
"learning_rate": 0.0004730050641094709,
"loss": 3.5599,
"step": 19750
},
{
"epoch": 2.1310946076848563,
"grad_norm": 0.598713755607605,
"learning_rate": 0.0004726818230794095,
"loss": 3.5721,
"step": 19800
},
{
"epoch": 2.1364761597244644,
"grad_norm": 0.5637555718421936,
"learning_rate": 0.0004723650468699493,
"loss": 3.5659,
"step": 19850
},
{
"epoch": 2.141857711764073,
"grad_norm": 0.6245604753494263,
"learning_rate": 0.0004720418058398879,
"loss": 3.56,
"step": 19900
},
{
"epoch": 2.147239263803681,
"grad_norm": 0.5333353877067566,
"learning_rate": 0.00047171856480982644,
"loss": 3.5668,
"step": 19950
},
{
"epoch": 2.152620815843289,
"grad_norm": 0.5466655492782593,
"learning_rate": 0.0004713953237797651,
"loss": 3.5558,
"step": 20000
},
{
"epoch": 2.152620815843289,
"eval_accuracy": 0.3628188296758417,
"eval_loss": 3.5698418617248535,
"eval_runtime": 180.3488,
"eval_samples_per_second": 99.868,
"eval_steps_per_second": 6.243,
"step": 20000
},
{
"epoch": 2.1580023678828972,
"grad_norm": 0.5573538541793823,
"learning_rate": 0.0004710720827497037,
"loss": 3.5583,
"step": 20050
},
{
"epoch": 2.163383919922506,
"grad_norm": 0.5291585326194763,
"learning_rate": 0.00047074884171964223,
"loss": 3.5514,
"step": 20100
},
{
"epoch": 2.168765471962114,
"grad_norm": 0.5464843511581421,
"learning_rate": 0.0004704256006895808,
"loss": 3.5752,
"step": 20150
},
{
"epoch": 2.174147024001722,
"grad_norm": 0.5665469765663147,
"learning_rate": 0.00047010235965951936,
"loss": 3.5353,
"step": 20200
},
{
"epoch": 2.1795285760413305,
"grad_norm": 0.5358191728591919,
"learning_rate": 0.00046977911862945796,
"loss": 3.5741,
"step": 20250
},
{
"epoch": 2.1849101280809387,
"grad_norm": 0.6494370698928833,
"learning_rate": 0.0004694558775993966,
"loss": 3.5496,
"step": 20300
},
{
"epoch": 2.1902916801205468,
"grad_norm": 0.5425307750701904,
"learning_rate": 0.00046913263656933515,
"loss": 3.5726,
"step": 20350
},
{
"epoch": 2.195673232160155,
"grad_norm": 0.5559353828430176,
"learning_rate": 0.00046880939553927374,
"loss": 3.5448,
"step": 20400
},
{
"epoch": 2.2010547841997634,
"grad_norm": 0.5595205426216125,
"learning_rate": 0.00046848615450921234,
"loss": 3.5519,
"step": 20450
},
{
"epoch": 2.2064363362393715,
"grad_norm": 0.5760559439659119,
"learning_rate": 0.0004681629134791509,
"loss": 3.5624,
"step": 20500
},
{
"epoch": 2.2118178882789796,
"grad_norm": 0.5371535420417786,
"learning_rate": 0.00046783967244908947,
"loss": 3.5661,
"step": 20550
},
{
"epoch": 2.2171994403185877,
"grad_norm": 0.5587826371192932,
"learning_rate": 0.0004675164314190281,
"loss": 3.5571,
"step": 20600
},
{
"epoch": 2.2225809923581963,
"grad_norm": 0.6047276258468628,
"learning_rate": 0.00046719319038896666,
"loss": 3.5486,
"step": 20650
},
{
"epoch": 2.2279625443978044,
"grad_norm": 0.5768632292747498,
"learning_rate": 0.00046686994935890526,
"loss": 3.5546,
"step": 20700
},
{
"epoch": 2.2333440964374125,
"grad_norm": 0.6263005137443542,
"learning_rate": 0.0004665467083288438,
"loss": 3.554,
"step": 20750
},
{
"epoch": 2.2387256484770206,
"grad_norm": 0.6004786491394043,
"learning_rate": 0.0004662234672987824,
"loss": 3.5518,
"step": 20800
},
{
"epoch": 2.244107200516629,
"grad_norm": 0.6552075147628784,
"learning_rate": 0.00046590022626872104,
"loss": 3.549,
"step": 20850
},
{
"epoch": 2.2494887525562373,
"grad_norm": 0.5540356636047363,
"learning_rate": 0.0004655769852386596,
"loss": 3.5524,
"step": 20900
},
{
"epoch": 2.2548703045958454,
"grad_norm": 0.593418300151825,
"learning_rate": 0.0004652537442085982,
"loss": 3.5517,
"step": 20950
},
{
"epoch": 2.2602518566354535,
"grad_norm": 0.5552871823310852,
"learning_rate": 0.00046493050317853677,
"loss": 3.5791,
"step": 21000
},
{
"epoch": 2.2602518566354535,
"eval_accuracy": 0.3637226043966826,
"eval_loss": 3.56265926361084,
"eval_runtime": 179.7058,
"eval_samples_per_second": 100.225,
"eval_steps_per_second": 6.266,
"step": 21000
},
{
"epoch": 2.265633408675062,
"grad_norm": 0.5663247108459473,
"learning_rate": 0.0004646072621484753,
"loss": 3.5696,
"step": 21050
},
{
"epoch": 2.27101496071467,
"grad_norm": 0.5453855395317078,
"learning_rate": 0.0004642840211184139,
"loss": 3.5571,
"step": 21100
},
{
"epoch": 2.2763965127542782,
"grad_norm": 0.5427026748657227,
"learning_rate": 0.00046396078008835255,
"loss": 3.5705,
"step": 21150
},
{
"epoch": 2.281778064793887,
"grad_norm": 0.9001597166061401,
"learning_rate": 0.0004636375390582911,
"loss": 3.5635,
"step": 21200
},
{
"epoch": 2.287159616833495,
"grad_norm": 0.5693010687828064,
"learning_rate": 0.0004633142980282297,
"loss": 3.5494,
"step": 21250
},
{
"epoch": 2.292541168873103,
"grad_norm": 0.5871009230613708,
"learning_rate": 0.00046299105699816823,
"loss": 3.5662,
"step": 21300
},
{
"epoch": 2.297922720912711,
"grad_norm": 0.5679400563240051,
"learning_rate": 0.0004626678159681068,
"loss": 3.5698,
"step": 21350
},
{
"epoch": 2.303304272952319,
"grad_norm": 0.5684141516685486,
"learning_rate": 0.0004623445749380454,
"loss": 3.5673,
"step": 21400
},
{
"epoch": 2.3086858249919278,
"grad_norm": 0.5958677530288696,
"learning_rate": 0.000462021333907984,
"loss": 3.5621,
"step": 21450
},
{
"epoch": 2.314067377031536,
"grad_norm": 0.5784631967544556,
"learning_rate": 0.0004616980928779226,
"loss": 3.5556,
"step": 21500
},
{
"epoch": 2.319448929071144,
"grad_norm": 0.5525918006896973,
"learning_rate": 0.0004613748518478612,
"loss": 3.5443,
"step": 21550
},
{
"epoch": 2.3248304811107525,
"grad_norm": 0.561515212059021,
"learning_rate": 0.00046105161081779974,
"loss": 3.5642,
"step": 21600
},
{
"epoch": 2.3302120331503606,
"grad_norm": 0.5596020817756653,
"learning_rate": 0.00046072836978773834,
"loss": 3.5656,
"step": 21650
},
{
"epoch": 2.3355935851899687,
"grad_norm": 0.6286318898200989,
"learning_rate": 0.000460405128757677,
"loss": 3.5574,
"step": 21700
},
{
"epoch": 2.340975137229577,
"grad_norm": 0.58785080909729,
"learning_rate": 0.0004600818877276155,
"loss": 3.5552,
"step": 21750
},
{
"epoch": 2.3463566892691854,
"grad_norm": 0.5441408753395081,
"learning_rate": 0.0004597586466975541,
"loss": 3.5477,
"step": 21800
},
{
"epoch": 2.3517382413087935,
"grad_norm": 0.5354775190353394,
"learning_rate": 0.00045943540566749266,
"loss": 3.5428,
"step": 21850
},
{
"epoch": 2.3571197933484016,
"grad_norm": 0.5691161155700684,
"learning_rate": 0.00045911216463743126,
"loss": 3.5492,
"step": 21900
},
{
"epoch": 2.3625013453880097,
"grad_norm": 0.5708732008934021,
"learning_rate": 0.00045878892360736985,
"loss": 3.568,
"step": 21950
},
{
"epoch": 2.3678828974276183,
"grad_norm": 0.5911089777946472,
"learning_rate": 0.00045847214739790966,
"loss": 3.5606,
"step": 22000
},
{
"epoch": 2.3678828974276183,
"eval_accuracy": 0.3649360398477947,
"eval_loss": 3.547260284423828,
"eval_runtime": 180.1433,
"eval_samples_per_second": 99.981,
"eval_steps_per_second": 6.251,
"step": 22000
},
{
"epoch": 2.3732644494672264,
"grad_norm": 0.5441485047340393,
"learning_rate": 0.0004581489063678482,
"loss": 3.5519,
"step": 22050
},
{
"epoch": 2.3786460015068345,
"grad_norm": 0.561090350151062,
"learning_rate": 0.00045782566533778685,
"loss": 3.5399,
"step": 22100
},
{
"epoch": 2.384027553546443,
"grad_norm": 0.5627390146255493,
"learning_rate": 0.00045750242430772544,
"loss": 3.564,
"step": 22150
},
{
"epoch": 2.389409105586051,
"grad_norm": 0.5869255065917969,
"learning_rate": 0.000457179183277664,
"loss": 3.55,
"step": 22200
},
{
"epoch": 2.3947906576256592,
"grad_norm": 0.6055136322975159,
"learning_rate": 0.0004568559422476026,
"loss": 3.5617,
"step": 22250
},
{
"epoch": 2.4001722096652673,
"grad_norm": 0.598421037197113,
"learning_rate": 0.00045653270121754117,
"loss": 3.5477,
"step": 22300
},
{
"epoch": 2.4055537617048754,
"grad_norm": 0.6216960549354553,
"learning_rate": 0.0004562094601874797,
"loss": 3.5399,
"step": 22350
},
{
"epoch": 2.410935313744484,
"grad_norm": 0.5606902837753296,
"learning_rate": 0.00045588621915741836,
"loss": 3.5615,
"step": 22400
},
{
"epoch": 2.416316865784092,
"grad_norm": 0.635993242263794,
"learning_rate": 0.00045556297812735696,
"loss": 3.5505,
"step": 22450
},
{
"epoch": 2.4216984178237,
"grad_norm": 0.5704584121704102,
"learning_rate": 0.0004552397370972955,
"loss": 3.5438,
"step": 22500
},
{
"epoch": 2.4270799698633088,
"grad_norm": 0.564420759677887,
"learning_rate": 0.0004549164960672341,
"loss": 3.5614,
"step": 22550
},
{
"epoch": 2.432461521902917,
"grad_norm": 0.5742602348327637,
"learning_rate": 0.00045459325503717263,
"loss": 3.5337,
"step": 22600
},
{
"epoch": 2.437843073942525,
"grad_norm": 0.5805179476737976,
"learning_rate": 0.0004542700140071113,
"loss": 3.5488,
"step": 22650
},
{
"epoch": 2.443224625982133,
"grad_norm": 0.5458112359046936,
"learning_rate": 0.0004539467729770499,
"loss": 3.547,
"step": 22700
},
{
"epoch": 2.4486061780217416,
"grad_norm": 0.5278791189193726,
"learning_rate": 0.0004536235319469884,
"loss": 3.5493,
"step": 22750
},
{
"epoch": 2.4539877300613497,
"grad_norm": 0.5699906945228577,
"learning_rate": 0.000453300290916927,
"loss": 3.5378,
"step": 22800
},
{
"epoch": 2.459369282100958,
"grad_norm": 0.5342678427696228,
"learning_rate": 0.0004529770498868656,
"loss": 3.5485,
"step": 22850
},
{
"epoch": 2.464750834140566,
"grad_norm": 0.5861589908599854,
"learning_rate": 0.00045265380885680414,
"loss": 3.576,
"step": 22900
},
{
"epoch": 2.4701323861801745,
"grad_norm": 0.5228945016860962,
"learning_rate": 0.0004523305678267428,
"loss": 3.5571,
"step": 22950
},
{
"epoch": 2.4755139382197826,
"grad_norm": 0.5882893204689026,
"learning_rate": 0.0004520073267966814,
"loss": 3.5327,
"step": 23000
},
{
"epoch": 2.4755139382197826,
"eval_accuracy": 0.3658152590159895,
"eval_loss": 3.540963888168335,
"eval_runtime": 179.7322,
"eval_samples_per_second": 100.21,
"eval_steps_per_second": 6.265,
"step": 23000
},
{
"epoch": 2.4808954902593907,
"grad_norm": 0.6533792018890381,
"learning_rate": 0.00045168408576661993,
"loss": 3.542,
"step": 23050
},
{
"epoch": 2.4862770422989993,
"grad_norm": 0.5604386329650879,
"learning_rate": 0.0004513608447365585,
"loss": 3.5655,
"step": 23100
},
{
"epoch": 2.4916585943386074,
"grad_norm": 0.6017301678657532,
"learning_rate": 0.00045103760370649706,
"loss": 3.5436,
"step": 23150
},
{
"epoch": 2.4970401463782155,
"grad_norm": 0.5612083077430725,
"learning_rate": 0.00045071436267643566,
"loss": 3.5427,
"step": 23200
},
{
"epoch": 2.5024216984178236,
"grad_norm": 0.5622110366821289,
"learning_rate": 0.0004503911216463743,
"loss": 3.5466,
"step": 23250
},
{
"epoch": 2.5078032504574317,
"grad_norm": 0.5365000367164612,
"learning_rate": 0.00045006788061631285,
"loss": 3.5615,
"step": 23300
},
{
"epoch": 2.5131848024970402,
"grad_norm": 0.5494401454925537,
"learning_rate": 0.00044974463958625144,
"loss": 3.5428,
"step": 23350
},
{
"epoch": 2.5185663545366483,
"grad_norm": 0.5880193114280701,
"learning_rate": 0.00044942139855619004,
"loss": 3.5372,
"step": 23400
},
{
"epoch": 2.5239479065762565,
"grad_norm": 0.520449161529541,
"learning_rate": 0.0004490981575261286,
"loss": 3.5542,
"step": 23450
},
{
"epoch": 2.529329458615865,
"grad_norm": 0.5952346920967102,
"learning_rate": 0.0004487749164960672,
"loss": 3.5421,
"step": 23500
},
{
"epoch": 2.534711010655473,
"grad_norm": 0.5716413259506226,
"learning_rate": 0.0004484516754660058,
"loss": 3.5611,
"step": 23550
},
{
"epoch": 2.540092562695081,
"grad_norm": 0.5482928156852722,
"learning_rate": 0.00044812843443594436,
"loss": 3.536,
"step": 23600
},
{
"epoch": 2.5454741147346893,
"grad_norm": 0.5782982110977173,
"learning_rate": 0.00044780519340588296,
"loss": 3.5429,
"step": 23650
},
{
"epoch": 2.550855666774298,
"grad_norm": 0.607822060585022,
"learning_rate": 0.0004474819523758215,
"loss": 3.5595,
"step": 23700
},
{
"epoch": 2.556237218813906,
"grad_norm": 0.5348266363143921,
"learning_rate": 0.0004471587113457601,
"loss": 3.5514,
"step": 23750
},
{
"epoch": 2.561618770853514,
"grad_norm": 0.5660938620567322,
"learning_rate": 0.00044683547031569874,
"loss": 3.5448,
"step": 23800
},
{
"epoch": 2.567000322893122,
"grad_norm": 0.5412974953651428,
"learning_rate": 0.0004465122292856373,
"loss": 3.5323,
"step": 23850
},
{
"epoch": 2.5723818749327307,
"grad_norm": 0.5600056648254395,
"learning_rate": 0.0004461889882555759,
"loss": 3.5325,
"step": 23900
},
{
"epoch": 2.577763426972339,
"grad_norm": 0.5820670127868652,
"learning_rate": 0.00044586574722551447,
"loss": 3.5247,
"step": 23950
},
{
"epoch": 2.583144979011947,
"grad_norm": 0.5861896276473999,
"learning_rate": 0.0004455489710160543,
"loss": 3.5435,
"step": 24000
},
{
"epoch": 2.583144979011947,
"eval_accuracy": 0.3670534673255233,
"eval_loss": 3.5298831462860107,
"eval_runtime": 180.4928,
"eval_samples_per_second": 99.788,
"eval_steps_per_second": 6.238,
"step": 24000
},
{
"epoch": 2.5885265310515555,
"grad_norm": 0.5437816381454468,
"learning_rate": 0.0004452257299859928,
"loss": 3.5307,
"step": 24050
},
{
"epoch": 2.5939080830911636,
"grad_norm": 0.5888408422470093,
"learning_rate": 0.0004449024889559314,
"loss": 3.5398,
"step": 24100
},
{
"epoch": 2.5992896351307717,
"grad_norm": 0.5409483909606934,
"learning_rate": 0.00044457924792587,
"loss": 3.5417,
"step": 24150
},
{
"epoch": 2.60467118717038,
"grad_norm": 0.5518124103546143,
"learning_rate": 0.0004442560068958086,
"loss": 3.5487,
"step": 24200
},
{
"epoch": 2.610052739209988,
"grad_norm": 0.5882364511489868,
"learning_rate": 0.0004439327658657472,
"loss": 3.5487,
"step": 24250
},
{
"epoch": 2.6154342912495965,
"grad_norm": 0.6176393032073975,
"learning_rate": 0.0004436095248356858,
"loss": 3.5503,
"step": 24300
},
{
"epoch": 2.6208158432892046,
"grad_norm": 0.6049776077270508,
"learning_rate": 0.00044328628380562433,
"loss": 3.5415,
"step": 24350
},
{
"epoch": 2.6261973953288127,
"grad_norm": 0.585410475730896,
"learning_rate": 0.0004429630427755629,
"loss": 3.5487,
"step": 24400
},
{
"epoch": 2.6315789473684212,
"grad_norm": 0.5842284560203552,
"learning_rate": 0.0004426398017455016,
"loss": 3.5549,
"step": 24450
},
{
"epoch": 2.6369604994080293,
"grad_norm": 0.57343590259552,
"learning_rate": 0.0004423165607154401,
"loss": 3.5565,
"step": 24500
},
{
"epoch": 2.6423420514476375,
"grad_norm": 0.6217878460884094,
"learning_rate": 0.0004419933196853787,
"loss": 3.5196,
"step": 24550
},
{
"epoch": 2.6477236034872456,
"grad_norm": 0.5150393843650818,
"learning_rate": 0.00044167007865531725,
"loss": 3.5427,
"step": 24600
},
{
"epoch": 2.653105155526854,
"grad_norm": 0.5708527565002441,
"learning_rate": 0.00044134683762525584,
"loss": 3.5328,
"step": 24650
},
{
"epoch": 2.658486707566462,
"grad_norm": 0.5596299767494202,
"learning_rate": 0.00044102359659519444,
"loss": 3.5424,
"step": 24700
},
{
"epoch": 2.6638682596060703,
"grad_norm": 0.5434728860855103,
"learning_rate": 0.00044070035556513303,
"loss": 3.5557,
"step": 24750
},
{
"epoch": 2.6692498116456784,
"grad_norm": 0.5621779561042786,
"learning_rate": 0.00044037711453507163,
"loss": 3.5385,
"step": 24800
},
{
"epoch": 2.674631363685287,
"grad_norm": 0.5845193862915039,
"learning_rate": 0.0004400538735050102,
"loss": 3.5583,
"step": 24850
},
{
"epoch": 2.680012915724895,
"grad_norm": 0.5741780996322632,
"learning_rate": 0.00043973063247494876,
"loss": 3.5429,
"step": 24900
},
{
"epoch": 2.685394467764503,
"grad_norm": 0.5549833178520203,
"learning_rate": 0.00043940739144488736,
"loss": 3.5355,
"step": 24950
},
{
"epoch": 2.6907760198041117,
"grad_norm": 0.5826306939125061,
"learning_rate": 0.0004390841504148259,
"loss": 3.5384,
"step": 25000
},
{
"epoch": 2.6907760198041117,
"eval_accuracy": 0.3680882774290684,
"eval_loss": 3.5210931301116943,
"eval_runtime": 179.7785,
"eval_samples_per_second": 100.184,
"eval_steps_per_second": 6.263,
"step": 25000
},
{
"epoch": 2.69615757184372,
"grad_norm": 0.5777888894081116,
"learning_rate": 0.00043876090938476455,
"loss": 3.5599,
"step": 25050
},
{
"epoch": 2.701539123883328,
"grad_norm": 0.5324108600616455,
"learning_rate": 0.00043843766835470314,
"loss": 3.5358,
"step": 25100
},
{
"epoch": 2.706920675922936,
"grad_norm": 0.6265562772750854,
"learning_rate": 0.0004381144273246417,
"loss": 3.5241,
"step": 25150
},
{
"epoch": 2.712302227962544,
"grad_norm": 0.5996240377426147,
"learning_rate": 0.0004377911862945803,
"loss": 3.5503,
"step": 25200
},
{
"epoch": 2.7176837800021527,
"grad_norm": 0.5723291635513306,
"learning_rate": 0.00043746794526451887,
"loss": 3.531,
"step": 25250
},
{
"epoch": 2.723065332041761,
"grad_norm": 0.5591868758201599,
"learning_rate": 0.00043714470423445747,
"loss": 3.5112,
"step": 25300
},
{
"epoch": 2.728446884081369,
"grad_norm": 0.6413562893867493,
"learning_rate": 0.00043682146320439606,
"loss": 3.547,
"step": 25350
},
{
"epoch": 2.7338284361209775,
"grad_norm": 0.5785355567932129,
"learning_rate": 0.00043649822217433466,
"loss": 3.5309,
"step": 25400
},
{
"epoch": 2.7392099881605856,
"grad_norm": 0.5443965792655945,
"learning_rate": 0.0004361749811442732,
"loss": 3.5263,
"step": 25450
},
{
"epoch": 2.7445915402001937,
"grad_norm": 0.5387240052223206,
"learning_rate": 0.0004358517401142118,
"loss": 3.528,
"step": 25500
},
{
"epoch": 2.749973092239802,
"grad_norm": 0.5816372632980347,
"learning_rate": 0.00043552849908415033,
"loss": 3.5582,
"step": 25550
},
{
"epoch": 2.7553546442794103,
"grad_norm": 0.5390015244483948,
"learning_rate": 0.000435205258054089,
"loss": 3.5369,
"step": 25600
},
{
"epoch": 2.7607361963190185,
"grad_norm": 0.5933911204338074,
"learning_rate": 0.0004348820170240276,
"loss": 3.5282,
"step": 25650
},
{
"epoch": 2.7661177483586266,
"grad_norm": 0.5716332793235779,
"learning_rate": 0.0004345587759939661,
"loss": 3.5247,
"step": 25700
},
{
"epoch": 2.7714993003982347,
"grad_norm": 0.5888131856918335,
"learning_rate": 0.0004342355349639047,
"loss": 3.5367,
"step": 25750
},
{
"epoch": 2.776880852437843,
"grad_norm": 0.5930463671684265,
"learning_rate": 0.0004339122939338433,
"loss": 3.5348,
"step": 25800
},
{
"epoch": 2.7822624044774513,
"grad_norm": 0.5617492198944092,
"learning_rate": 0.00043358905290378184,
"loss": 3.5135,
"step": 25850
},
{
"epoch": 2.7876439565170594,
"grad_norm": 0.5550035834312439,
"learning_rate": 0.0004332658118737205,
"loss": 3.5386,
"step": 25900
},
{
"epoch": 2.793025508556668,
"grad_norm": 0.545727014541626,
"learning_rate": 0.0004329425708436591,
"loss": 3.5242,
"step": 25950
},
{
"epoch": 2.798407060596276,
"grad_norm": 0.588079035282135,
"learning_rate": 0.0004326257946341989,
"loss": 3.5324,
"step": 26000
},
{
"epoch": 2.798407060596276,
"eval_accuracy": 0.36903094988374685,
"eval_loss": 3.512333631515503,
"eval_runtime": 180.2641,
"eval_samples_per_second": 99.914,
"eval_steps_per_second": 6.246,
"step": 26000
},
{
"epoch": 2.803788612635884,
"grad_norm": 0.5898637175559998,
"learning_rate": 0.00043230255360413744,
"loss": 3.5297,
"step": 26050
},
{
"epoch": 2.8091701646754923,
"grad_norm": 0.6078748106956482,
"learning_rate": 0.00043197931257407603,
"loss": 3.5293,
"step": 26100
},
{
"epoch": 2.8145517167151004,
"grad_norm": 0.5512140989303589,
"learning_rate": 0.0004316560715440146,
"loss": 3.5403,
"step": 26150
},
{
"epoch": 2.819933268754709,
"grad_norm": 0.5708469748497009,
"learning_rate": 0.00043133283051395317,
"loss": 3.516,
"step": 26200
},
{
"epoch": 2.825314820794317,
"grad_norm": 0.5742740035057068,
"learning_rate": 0.0004310095894838918,
"loss": 3.5201,
"step": 26250
},
{
"epoch": 2.830696372833925,
"grad_norm": 0.5645484924316406,
"learning_rate": 0.0004306863484538304,
"loss": 3.5243,
"step": 26300
},
{
"epoch": 2.8360779248735337,
"grad_norm": 0.5661312341690063,
"learning_rate": 0.00043036310742376895,
"loss": 3.5189,
"step": 26350
},
{
"epoch": 2.841459476913142,
"grad_norm": 0.5699840784072876,
"learning_rate": 0.00043003986639370754,
"loss": 3.5436,
"step": 26400
},
{
"epoch": 2.84684102895275,
"grad_norm": 0.5516175031661987,
"learning_rate": 0.0004297166253636461,
"loss": 3.5463,
"step": 26450
},
{
"epoch": 2.852222580992358,
"grad_norm": 0.5519968271255493,
"learning_rate": 0.0004293933843335847,
"loss": 3.5243,
"step": 26500
},
{
"epoch": 2.857604133031966,
"grad_norm": 0.5642961859703064,
"learning_rate": 0.00042907014330352333,
"loss": 3.5221,
"step": 26550
},
{
"epoch": 2.8629856850715747,
"grad_norm": 0.623792290687561,
"learning_rate": 0.00042874690227346187,
"loss": 3.5376,
"step": 26600
},
{
"epoch": 2.868367237111183,
"grad_norm": 0.5826441049575806,
"learning_rate": 0.00042842366124340046,
"loss": 3.5282,
"step": 26650
},
{
"epoch": 2.873748789150791,
"grad_norm": 0.5754335522651672,
"learning_rate": 0.00042810042021333906,
"loss": 3.5277,
"step": 26700
},
{
"epoch": 2.8791303411903995,
"grad_norm": 0.5910266041755676,
"learning_rate": 0.0004277771791832776,
"loss": 3.5288,
"step": 26750
},
{
"epoch": 2.8845118932300076,
"grad_norm": 0.5650596022605896,
"learning_rate": 0.0004274539381532162,
"loss": 3.5197,
"step": 26800
},
{
"epoch": 2.8898934452696157,
"grad_norm": 0.6088457703590393,
"learning_rate": 0.00042713069712315484,
"loss": 3.51,
"step": 26850
},
{
"epoch": 2.895274997309224,
"grad_norm": 0.5383467674255371,
"learning_rate": 0.0004268074560930934,
"loss": 3.5174,
"step": 26900
},
{
"epoch": 2.9006565493488323,
"grad_norm": 0.5610182881355286,
"learning_rate": 0.000426484215063032,
"loss": 3.5204,
"step": 26950
},
{
"epoch": 2.9060381013884404,
"grad_norm": 0.5684370994567871,
"learning_rate": 0.0004261609740329705,
"loss": 3.5153,
"step": 27000
},
{
"epoch": 2.9060381013884404,
"eval_accuracy": 0.36983009187362237,
"eval_loss": 3.501441717147827,
"eval_runtime": 179.9533,
"eval_samples_per_second": 100.087,
"eval_steps_per_second": 6.257,
"step": 27000
},
{
"epoch": 2.9114196534280485,
"grad_norm": 0.6579453349113464,
"learning_rate": 0.0004258377330029091,
"loss": 3.5223,
"step": 27050
},
{
"epoch": 2.9168012054676566,
"grad_norm": 0.5974162220954895,
"learning_rate": 0.00042551449197284776,
"loss": 3.5281,
"step": 27100
},
{
"epoch": 2.922182757507265,
"grad_norm": 0.5756824016571045,
"learning_rate": 0.0004251912509427863,
"loss": 3.5124,
"step": 27150
},
{
"epoch": 2.9275643095468733,
"grad_norm": 0.639370858669281,
"learning_rate": 0.0004248680099127249,
"loss": 3.5413,
"step": 27200
},
{
"epoch": 2.9329458615864814,
"grad_norm": 0.597856879234314,
"learning_rate": 0.0004245447688826635,
"loss": 3.5282,
"step": 27250
},
{
"epoch": 2.93832741362609,
"grad_norm": 0.6288902163505554,
"learning_rate": 0.00042422152785260203,
"loss": 3.532,
"step": 27300
},
{
"epoch": 2.943708965665698,
"grad_norm": 0.7320722937583923,
"learning_rate": 0.0004238982868225406,
"loss": 3.5306,
"step": 27350
},
{
"epoch": 2.949090517705306,
"grad_norm": 0.5934067964553833,
"learning_rate": 0.0004235750457924793,
"loss": 3.5274,
"step": 27400
},
{
"epoch": 2.9544720697449143,
"grad_norm": 0.554151177406311,
"learning_rate": 0.0004232518047624178,
"loss": 3.5108,
"step": 27450
},
{
"epoch": 2.9598536217845224,
"grad_norm": 0.552830159664154,
"learning_rate": 0.0004229285637323564,
"loss": 3.515,
"step": 27500
},
{
"epoch": 2.965235173824131,
"grad_norm": 0.5718715190887451,
"learning_rate": 0.00042260532270229495,
"loss": 3.5394,
"step": 27550
},
{
"epoch": 2.970616725863739,
"grad_norm": 0.5800749063491821,
"learning_rate": 0.00042228208167223354,
"loss": 3.5282,
"step": 27600
},
{
"epoch": 2.975998277903347,
"grad_norm": 0.6078159809112549,
"learning_rate": 0.00042195884064217214,
"loss": 3.5321,
"step": 27650
},
{
"epoch": 2.9813798299429557,
"grad_norm": 0.5575968623161316,
"learning_rate": 0.00042163559961211073,
"loss": 3.524,
"step": 27700
},
{
"epoch": 2.986761381982564,
"grad_norm": 0.5276609659194946,
"learning_rate": 0.00042131235858204933,
"loss": 3.5273,
"step": 27750
},
{
"epoch": 2.992142934022172,
"grad_norm": 0.5870263576507568,
"learning_rate": 0.0004209891175519879,
"loss": 3.5096,
"step": 27800
},
{
"epoch": 2.9975244860617805,
"grad_norm": 0.5546600818634033,
"learning_rate": 0.00042066587652192646,
"loss": 3.5217,
"step": 27850
},
{
"epoch": 3.0029060381013886,
"grad_norm": 0.5607388019561768,
"learning_rate": 0.00042034263549186506,
"loss": 3.46,
"step": 27900
},
{
"epoch": 3.0082875901409967,
"grad_norm": 0.6114853620529175,
"learning_rate": 0.0004200193944618036,
"loss": 3.4226,
"step": 27950
},
{
"epoch": 3.0136691421806048,
"grad_norm": 0.5920767188072205,
"learning_rate": 0.00041970261825234346,
"loss": 3.4362,
"step": 28000
},
{
"epoch": 3.0136691421806048,
"eval_accuracy": 0.3704642987797955,
"eval_loss": 3.497429132461548,
"eval_runtime": 180.2708,
"eval_samples_per_second": 99.911,
"eval_steps_per_second": 6.246,
"step": 28000
},
{
"epoch": 3.0190506942202133,
"grad_norm": 0.6086764335632324,
"learning_rate": 0.00041937937722228205,
"loss": 3.4374,
"step": 28050
},
{
"epoch": 3.0244322462598214,
"grad_norm": 0.573062539100647,
"learning_rate": 0.00041905613619222065,
"loss": 3.4365,
"step": 28100
},
{
"epoch": 3.0298137982994295,
"grad_norm": 0.6111119389533997,
"learning_rate": 0.00041873289516215924,
"loss": 3.4324,
"step": 28150
},
{
"epoch": 3.0351953503390376,
"grad_norm": 0.5908578634262085,
"learning_rate": 0.0004184096541320978,
"loss": 3.4512,
"step": 28200
},
{
"epoch": 3.040576902378646,
"grad_norm": 0.8571264147758484,
"learning_rate": 0.0004180864131020364,
"loss": 3.4453,
"step": 28250
},
{
"epoch": 3.0459584544182543,
"grad_norm": 0.5717503428459167,
"learning_rate": 0.0004177631720719749,
"loss": 3.4228,
"step": 28300
},
{
"epoch": 3.0513400064578624,
"grad_norm": 0.606881320476532,
"learning_rate": 0.00041743993104191357,
"loss": 3.4489,
"step": 28350
},
{
"epoch": 3.0567215584974705,
"grad_norm": 0.579672634601593,
"learning_rate": 0.00041711669001185216,
"loss": 3.4317,
"step": 28400
},
{
"epoch": 3.062103110537079,
"grad_norm": 0.6016443371772766,
"learning_rate": 0.0004167934489817907,
"loss": 3.4341,
"step": 28450
},
{
"epoch": 3.067484662576687,
"grad_norm": 0.5978291034698486,
"learning_rate": 0.0004164702079517293,
"loss": 3.4344,
"step": 28500
},
{
"epoch": 3.0728662146162953,
"grad_norm": 0.5813153386116028,
"learning_rate": 0.0004161469669216679,
"loss": 3.4579,
"step": 28550
},
{
"epoch": 3.0782477666559034,
"grad_norm": 0.5806988477706909,
"learning_rate": 0.00041582372589160643,
"loss": 3.4485,
"step": 28600
},
{
"epoch": 3.083629318695512,
"grad_norm": 0.5982168316841125,
"learning_rate": 0.0004155004848615451,
"loss": 3.4556,
"step": 28650
},
{
"epoch": 3.08901087073512,
"grad_norm": 0.6008811593055725,
"learning_rate": 0.0004151772438314837,
"loss": 3.4323,
"step": 28700
},
{
"epoch": 3.094392422774728,
"grad_norm": 0.5557597875595093,
"learning_rate": 0.0004148540028014222,
"loss": 3.4236,
"step": 28750
},
{
"epoch": 3.0997739748143363,
"grad_norm": 0.584446370601654,
"learning_rate": 0.0004145307617713608,
"loss": 3.4504,
"step": 28800
},
{
"epoch": 3.105155526853945,
"grad_norm": 0.576360821723938,
"learning_rate": 0.00041420752074129935,
"loss": 3.4543,
"step": 28850
},
{
"epoch": 3.110537078893553,
"grad_norm": 0.5923529267311096,
"learning_rate": 0.000413884279711238,
"loss": 3.4149,
"step": 28900
},
{
"epoch": 3.115918630933161,
"grad_norm": 0.5620018243789673,
"learning_rate": 0.00041356750350177775,
"loss": 3.4458,
"step": 28950
},
{
"epoch": 3.121300182972769,
"grad_norm": 0.658657968044281,
"learning_rate": 0.0004132442624717164,
"loss": 3.448,
"step": 29000
},
{
"epoch": 3.121300182972769,
"eval_accuracy": 0.371386979103116,
"eval_loss": 3.4938414096832275,
"eval_runtime": 180.2028,
"eval_samples_per_second": 99.948,
"eval_steps_per_second": 6.249,
"step": 29000
},
{
"epoch": 3.1266817350123777,
"grad_norm": 0.5916072726249695,
"learning_rate": 0.000412921021441655,
"loss": 3.4481,
"step": 29050
},
{
"epoch": 3.132063287051986,
"grad_norm": 0.5693389177322388,
"learning_rate": 0.00041259778041159354,
"loss": 3.4506,
"step": 29100
},
{
"epoch": 3.137444839091594,
"grad_norm": 0.5854905247688293,
"learning_rate": 0.00041227453938153213,
"loss": 3.4614,
"step": 29150
},
{
"epoch": 3.1428263911312024,
"grad_norm": 0.5985913276672363,
"learning_rate": 0.00041195129835147067,
"loss": 3.4455,
"step": 29200
},
{
"epoch": 3.1482079431708105,
"grad_norm": 0.5984156727790833,
"learning_rate": 0.00041162805732140927,
"loss": 3.447,
"step": 29250
},
{
"epoch": 3.1535894952104186,
"grad_norm": 0.5744922757148743,
"learning_rate": 0.0004113048162913479,
"loss": 3.4795,
"step": 29300
},
{
"epoch": 3.1589710472500268,
"grad_norm": 0.6045938730239868,
"learning_rate": 0.00041098157526128646,
"loss": 3.4548,
"step": 29350
},
{
"epoch": 3.1643525992896353,
"grad_norm": 0.6241233348846436,
"learning_rate": 0.00041065833423122505,
"loss": 3.4531,
"step": 29400
},
{
"epoch": 3.1697341513292434,
"grad_norm": 0.5867751836776733,
"learning_rate": 0.00041033509320116365,
"loss": 3.441,
"step": 29450
},
{
"epoch": 3.1751157033688515,
"grad_norm": 0.6040583252906799,
"learning_rate": 0.0004100118521711022,
"loss": 3.461,
"step": 29500
},
{
"epoch": 3.1804972554084596,
"grad_norm": 0.6353616118431091,
"learning_rate": 0.0004096886111410408,
"loss": 3.4538,
"step": 29550
},
{
"epoch": 3.185878807448068,
"grad_norm": 0.609554648399353,
"learning_rate": 0.00040936537011097943,
"loss": 3.4705,
"step": 29600
},
{
"epoch": 3.1912603594876763,
"grad_norm": 0.5806785821914673,
"learning_rate": 0.00040904212908091797,
"loss": 3.444,
"step": 29650
},
{
"epoch": 3.1966419115272844,
"grad_norm": 0.7444307804107666,
"learning_rate": 0.00040871888805085656,
"loss": 3.4575,
"step": 29700
},
{
"epoch": 3.2020234635668925,
"grad_norm": 0.6058987379074097,
"learning_rate": 0.0004083956470207951,
"loss": 3.4707,
"step": 29750
},
{
"epoch": 3.207405015606501,
"grad_norm": 0.5647191405296326,
"learning_rate": 0.0004080724059907337,
"loss": 3.4502,
"step": 29800
},
{
"epoch": 3.212786567646109,
"grad_norm": 0.6401323080062866,
"learning_rate": 0.00040774916496067235,
"loss": 3.4423,
"step": 29850
},
{
"epoch": 3.2181681196857173,
"grad_norm": 0.6080923676490784,
"learning_rate": 0.0004074259239306109,
"loss": 3.4446,
"step": 29900
},
{
"epoch": 3.2235496717253254,
"grad_norm": 0.5872158408164978,
"learning_rate": 0.0004071026829005495,
"loss": 3.4658,
"step": 29950
},
{
"epoch": 3.228931223764934,
"grad_norm": 0.5680561065673828,
"learning_rate": 0.0004067794418704881,
"loss": 3.4482,
"step": 30000
},
{
"epoch": 3.228931223764934,
"eval_accuracy": 0.3716701285287606,
"eval_loss": 3.4923532009124756,
"eval_runtime": 180.3398,
"eval_samples_per_second": 99.873,
"eval_steps_per_second": 6.244,
"step": 30000
},
{
"epoch": 3.234312775804542,
"grad_norm": 0.6124100685119629,
"learning_rate": 0.0004064562008404266,
"loss": 3.4498,
"step": 30050
},
{
"epoch": 3.23969432784415,
"grad_norm": 0.6356387138366699,
"learning_rate": 0.0004061329598103652,
"loss": 3.4562,
"step": 30100
},
{
"epoch": 3.2450758798837587,
"grad_norm": 0.6179471611976624,
"learning_rate": 0.00040580971878030386,
"loss": 3.4407,
"step": 30150
},
{
"epoch": 3.250457431923367,
"grad_norm": 0.5960007905960083,
"learning_rate": 0.0004054864777502424,
"loss": 3.4524,
"step": 30200
},
{
"epoch": 3.255838983962975,
"grad_norm": 0.6349455118179321,
"learning_rate": 0.000405163236720181,
"loss": 3.4605,
"step": 30250
},
{
"epoch": 3.261220536002583,
"grad_norm": 0.5955315232276917,
"learning_rate": 0.00040483999569011954,
"loss": 3.4601,
"step": 30300
},
{
"epoch": 3.2666020880421915,
"grad_norm": 0.6215502023696899,
"learning_rate": 0.00040451675466005813,
"loss": 3.4626,
"step": 30350
},
{
"epoch": 3.2719836400817996,
"grad_norm": 0.5687872171401978,
"learning_rate": 0.0004041935136299967,
"loss": 3.4495,
"step": 30400
},
{
"epoch": 3.2773651921214078,
"grad_norm": 0.6149540543556213,
"learning_rate": 0.0004038702725999353,
"loss": 3.4654,
"step": 30450
},
{
"epoch": 3.282746744161016,
"grad_norm": 0.5696414709091187,
"learning_rate": 0.0004035470315698739,
"loss": 3.4482,
"step": 30500
},
{
"epoch": 3.2881282962006244,
"grad_norm": 0.5921586751937866,
"learning_rate": 0.0004032237905398125,
"loss": 3.4423,
"step": 30550
},
{
"epoch": 3.2935098482402325,
"grad_norm": 0.6276850700378418,
"learning_rate": 0.00040290054950975105,
"loss": 3.4592,
"step": 30600
},
{
"epoch": 3.2988914002798406,
"grad_norm": 0.597515344619751,
"learning_rate": 0.00040257730847968965,
"loss": 3.4615,
"step": 30650
},
{
"epoch": 3.304272952319449,
"grad_norm": 0.6534189581871033,
"learning_rate": 0.0004022540674496283,
"loss": 3.465,
"step": 30700
},
{
"epoch": 3.3096545043590573,
"grad_norm": 0.6232941746711731,
"learning_rate": 0.00040193082641956684,
"loss": 3.4645,
"step": 30750
},
{
"epoch": 3.3150360563986654,
"grad_norm": 0.6439439654350281,
"learning_rate": 0.00040160758538950543,
"loss": 3.4445,
"step": 30800
},
{
"epoch": 3.3204176084382735,
"grad_norm": 0.6507943868637085,
"learning_rate": 0.00040128434435944397,
"loss": 3.4739,
"step": 30850
},
{
"epoch": 3.3257991604778816,
"grad_norm": 0.6493215560913086,
"learning_rate": 0.00040096110332938257,
"loss": 3.4431,
"step": 30900
},
{
"epoch": 3.33118071251749,
"grad_norm": 0.6146352291107178,
"learning_rate": 0.00040063786229932116,
"loss": 3.4583,
"step": 30950
},
{
"epoch": 3.3365622645570983,
"grad_norm": 0.6426176428794861,
"learning_rate": 0.00040031462126925975,
"loss": 3.4714,
"step": 31000
},
{
"epoch": 3.3365622645570983,
"eval_accuracy": 0.3725597783741855,
"eval_loss": 3.482445240020752,
"eval_runtime": 179.9242,
"eval_samples_per_second": 100.103,
"eval_steps_per_second": 6.258,
"step": 31000
},
{
"epoch": 3.3419438165967064,
"grad_norm": 0.6456658244132996,
"learning_rate": 0.00039999138023919835,
"loss": 3.4466,
"step": 31050
},
{
"epoch": 3.347325368636315,
"grad_norm": 0.6010160446166992,
"learning_rate": 0.00039966813920913694,
"loss": 3.4282,
"step": 31100
},
{
"epoch": 3.352706920675923,
"grad_norm": 0.6064096689224243,
"learning_rate": 0.0003993448981790755,
"loss": 3.4615,
"step": 31150
},
{
"epoch": 3.358088472715531,
"grad_norm": 0.6383664011955261,
"learning_rate": 0.0003990216571490141,
"loss": 3.454,
"step": 31200
},
{
"epoch": 3.3634700247551392,
"grad_norm": 0.6114600896835327,
"learning_rate": 0.0003986984161189526,
"loss": 3.4583,
"step": 31250
},
{
"epoch": 3.368851576794748,
"grad_norm": 0.6199226975440979,
"learning_rate": 0.00039837517508889127,
"loss": 3.447,
"step": 31300
},
{
"epoch": 3.374233128834356,
"grad_norm": 0.6489754915237427,
"learning_rate": 0.00039805193405882986,
"loss": 3.4586,
"step": 31350
},
{
"epoch": 3.379614680873964,
"grad_norm": 0.57906174659729,
"learning_rate": 0.0003977286930287684,
"loss": 3.459,
"step": 31400
},
{
"epoch": 3.384996232913572,
"grad_norm": 0.5564760565757751,
"learning_rate": 0.000397405451998707,
"loss": 3.4716,
"step": 31450
},
{
"epoch": 3.3903777849531807,
"grad_norm": 0.6209973096847534,
"learning_rate": 0.00039708221096864554,
"loss": 3.471,
"step": 31500
},
{
"epoch": 3.3957593369927888,
"grad_norm": 0.5805020928382874,
"learning_rate": 0.00039675896993858413,
"loss": 3.4692,
"step": 31550
},
{
"epoch": 3.401140889032397,
"grad_norm": 0.598192572593689,
"learning_rate": 0.0003964357289085228,
"loss": 3.4498,
"step": 31600
},
{
"epoch": 3.4065224410720054,
"grad_norm": 0.6662253737449646,
"learning_rate": 0.0003961124878784613,
"loss": 3.4412,
"step": 31650
},
{
"epoch": 3.4119039931116135,
"grad_norm": 0.5580045580863953,
"learning_rate": 0.0003957892468483999,
"loss": 3.4489,
"step": 31700
},
{
"epoch": 3.4172855451512216,
"grad_norm": 0.6119815111160278,
"learning_rate": 0.0003954660058183385,
"loss": 3.4623,
"step": 31750
},
{
"epoch": 3.4226670971908297,
"grad_norm": 0.5958937406539917,
"learning_rate": 0.00039514276478827705,
"loss": 3.4371,
"step": 31800
},
{
"epoch": 3.428048649230438,
"grad_norm": 0.6033514738082886,
"learning_rate": 0.0003948195237582157,
"loss": 3.456,
"step": 31850
},
{
"epoch": 3.4334302012700464,
"grad_norm": 0.5335016846656799,
"learning_rate": 0.0003944962827281543,
"loss": 3.4487,
"step": 31900
},
{
"epoch": 3.4388117533096545,
"grad_norm": 0.6240798234939575,
"learning_rate": 0.00039417304169809284,
"loss": 3.4618,
"step": 31950
},
{
"epoch": 3.4441933053492626,
"grad_norm": 0.6087186932563782,
"learning_rate": 0.00039384980066803143,
"loss": 3.4663,
"step": 32000
},
{
"epoch": 3.4441933053492626,
"eval_accuracy": 0.3730128609163719,
"eval_loss": 3.4758286476135254,
"eval_runtime": 180.9046,
"eval_samples_per_second": 99.561,
"eval_steps_per_second": 6.224,
"step": 32000
},
{
"epoch": 3.449574857388871,
"grad_norm": 0.5830413103103638,
"learning_rate": 0.00039352655963796997,
"loss": 3.4582,
"step": 32050
},
{
"epoch": 3.4549564094284793,
"grad_norm": 0.6102946996688843,
"learning_rate": 0.00039320331860790857,
"loss": 3.4796,
"step": 32100
},
{
"epoch": 3.4603379614680874,
"grad_norm": 0.5744568705558777,
"learning_rate": 0.0003928800775778472,
"loss": 3.448,
"step": 32150
},
{
"epoch": 3.4657195135076955,
"grad_norm": 0.6026392579078674,
"learning_rate": 0.00039255683654778576,
"loss": 3.451,
"step": 32200
},
{
"epoch": 3.471101065547304,
"grad_norm": 0.6402329206466675,
"learning_rate": 0.00039223359551772435,
"loss": 3.4786,
"step": 32250
},
{
"epoch": 3.476482617586912,
"grad_norm": 0.6022635102272034,
"learning_rate": 0.00039191035448766294,
"loss": 3.461,
"step": 32300
},
{
"epoch": 3.4818641696265202,
"grad_norm": 0.6575260758399963,
"learning_rate": 0.0003915871134576015,
"loss": 3.4667,
"step": 32350
},
{
"epoch": 3.4872457216661283,
"grad_norm": 0.6149677634239197,
"learning_rate": 0.0003912638724275401,
"loss": 3.4635,
"step": 32400
},
{
"epoch": 3.492627273705737,
"grad_norm": 0.7180488705635071,
"learning_rate": 0.00039094063139747873,
"loss": 3.4736,
"step": 32450
},
{
"epoch": 3.498008825745345,
"grad_norm": 0.6193327903747559,
"learning_rate": 0.00039061739036741727,
"loss": 3.4594,
"step": 32500
},
{
"epoch": 3.503390377784953,
"grad_norm": 0.6941162943840027,
"learning_rate": 0.00039029414933735586,
"loss": 3.4656,
"step": 32550
},
{
"epoch": 3.5087719298245617,
"grad_norm": 0.5805447101593018,
"learning_rate": 0.0003899709083072944,
"loss": 3.4647,
"step": 32600
},
{
"epoch": 3.5141534818641698,
"grad_norm": 0.6239206790924072,
"learning_rate": 0.000389647667277233,
"loss": 3.4568,
"step": 32650
},
{
"epoch": 3.519535033903778,
"grad_norm": 0.6483040452003479,
"learning_rate": 0.00038932442624717165,
"loss": 3.4532,
"step": 32700
},
{
"epoch": 3.524916585943386,
"grad_norm": 0.6193462014198303,
"learning_rate": 0.0003890011852171102,
"loss": 3.4473,
"step": 32750
},
{
"epoch": 3.530298137982994,
"grad_norm": 0.6228381991386414,
"learning_rate": 0.0003886779441870488,
"loss": 3.4511,
"step": 32800
},
{
"epoch": 3.5356796900226026,
"grad_norm": 0.5783094167709351,
"learning_rate": 0.0003883547031569874,
"loss": 3.444,
"step": 32850
},
{
"epoch": 3.5410612420622107,
"grad_norm": 0.5559167861938477,
"learning_rate": 0.0003880314621269259,
"loss": 3.4532,
"step": 32900
},
{
"epoch": 3.546442794101819,
"grad_norm": 0.603102445602417,
"learning_rate": 0.0003877146859174657,
"loss": 3.4626,
"step": 32950
},
{
"epoch": 3.5518243461414274,
"grad_norm": 0.5790406465530396,
"learning_rate": 0.0003873914448874043,
"loss": 3.461,
"step": 33000
},
{
"epoch": 3.5518243461414274,
"eval_accuracy": 0.3738632870692959,
"eval_loss": 3.4701783657073975,
"eval_runtime": 180.005,
"eval_samples_per_second": 100.058,
"eval_steps_per_second": 6.255,
"step": 33000
},
{
"epoch": 3.5572058981810355,
"grad_norm": 0.6018248796463013,
"learning_rate": 0.0003870682038573429,
"loss": 3.4504,
"step": 33050
},
{
"epoch": 3.5625874502206436,
"grad_norm": 0.5604973435401917,
"learning_rate": 0.0003867449628272815,
"loss": 3.4534,
"step": 33100
},
{
"epoch": 3.5679690022602517,
"grad_norm": 0.5984012484550476,
"learning_rate": 0.0003864217217972201,
"loss": 3.4518,
"step": 33150
},
{
"epoch": 3.57335055429986,
"grad_norm": 0.6406269669532776,
"learning_rate": 0.0003860984807671587,
"loss": 3.4586,
"step": 33200
},
{
"epoch": 3.5787321063394684,
"grad_norm": 0.6193456053733826,
"learning_rate": 0.00038577523973709724,
"loss": 3.4617,
"step": 33250
},
{
"epoch": 3.5841136583790765,
"grad_norm": 0.6093300580978394,
"learning_rate": 0.00038545199870703583,
"loss": 3.4654,
"step": 33300
},
{
"epoch": 3.5894952104186846,
"grad_norm": 0.5693668723106384,
"learning_rate": 0.0003851287576769744,
"loss": 3.4363,
"step": 33350
},
{
"epoch": 3.594876762458293,
"grad_norm": 0.6020606160163879,
"learning_rate": 0.000384805516646913,
"loss": 3.4486,
"step": 33400
},
{
"epoch": 3.6002583144979012,
"grad_norm": 0.5911715030670166,
"learning_rate": 0.0003844822756168516,
"loss": 3.4621,
"step": 33450
},
{
"epoch": 3.6056398665375093,
"grad_norm": 0.6165992021560669,
"learning_rate": 0.00038415903458679016,
"loss": 3.456,
"step": 33500
},
{
"epoch": 3.611021418577118,
"grad_norm": 0.6295124292373657,
"learning_rate": 0.00038383579355672875,
"loss": 3.4562,
"step": 33550
},
{
"epoch": 3.616402970616726,
"grad_norm": 0.5734518766403198,
"learning_rate": 0.00038351255252666735,
"loss": 3.4466,
"step": 33600
},
{
"epoch": 3.621784522656334,
"grad_norm": 0.6045584082603455,
"learning_rate": 0.00038318931149660594,
"loss": 3.4463,
"step": 33650
},
{
"epoch": 3.627166074695942,
"grad_norm": 0.625198483467102,
"learning_rate": 0.00038286607046654454,
"loss": 3.4344,
"step": 33700
},
{
"epoch": 3.6325476267355503,
"grad_norm": 0.6212444305419922,
"learning_rate": 0.00038254282943648313,
"loss": 3.4403,
"step": 33750
},
{
"epoch": 3.637929178775159,
"grad_norm": 0.6339898109436035,
"learning_rate": 0.00038221958840642167,
"loss": 3.4669,
"step": 33800
},
{
"epoch": 3.643310730814767,
"grad_norm": 0.5889163017272949,
"learning_rate": 0.00038189634737636027,
"loss": 3.4513,
"step": 33850
},
{
"epoch": 3.648692282854375,
"grad_norm": 0.5536656975746155,
"learning_rate": 0.0003815731063462988,
"loss": 3.4398,
"step": 33900
},
{
"epoch": 3.6540738348939836,
"grad_norm": 0.5549130439758301,
"learning_rate": 0.00038124986531623745,
"loss": 3.4713,
"step": 33950
},
{
"epoch": 3.6594553869335917,
"grad_norm": 0.6131840944290161,
"learning_rate": 0.00038092662428617605,
"loss": 3.4507,
"step": 34000
},
{
"epoch": 3.6594553869335917,
"eval_accuracy": 0.37475293691472084,
"eval_loss": 3.4609808921813965,
"eval_runtime": 180.1311,
"eval_samples_per_second": 99.988,
"eval_steps_per_second": 6.251,
"step": 34000
},
{
"epoch": 3.6648369389732,
"grad_norm": 0.5674228072166443,
"learning_rate": 0.0003806033832561146,
"loss": 3.4384,
"step": 34050
},
{
"epoch": 3.670218491012808,
"grad_norm": 0.6528271436691284,
"learning_rate": 0.0003802801422260532,
"loss": 3.4544,
"step": 34100
},
{
"epoch": 3.675600043052416,
"grad_norm": 0.59865802526474,
"learning_rate": 0.0003799569011959918,
"loss": 3.4432,
"step": 34150
},
{
"epoch": 3.6809815950920246,
"grad_norm": 0.6160247325897217,
"learning_rate": 0.0003796336601659303,
"loss": 3.4473,
"step": 34200
},
{
"epoch": 3.6863631471316327,
"grad_norm": 0.6500553488731384,
"learning_rate": 0.00037931041913586897,
"loss": 3.4589,
"step": 34250
},
{
"epoch": 3.691744699171241,
"grad_norm": 0.6018146276473999,
"learning_rate": 0.00037898717810580756,
"loss": 3.4633,
"step": 34300
},
{
"epoch": 3.6971262512108494,
"grad_norm": 0.614810585975647,
"learning_rate": 0.0003786639370757461,
"loss": 3.4687,
"step": 34350
},
{
"epoch": 3.7025078032504575,
"grad_norm": 0.5721622109413147,
"learning_rate": 0.0003783406960456847,
"loss": 3.4664,
"step": 34400
},
{
"epoch": 3.7078893552900656,
"grad_norm": 0.626194953918457,
"learning_rate": 0.00037801745501562324,
"loss": 3.4689,
"step": 34450
},
{
"epoch": 3.713270907329674,
"grad_norm": 0.6698977947235107,
"learning_rate": 0.0003776942139855619,
"loss": 3.4617,
"step": 34500
},
{
"epoch": 3.7186524593692822,
"grad_norm": 0.5798797607421875,
"learning_rate": 0.0003773709729555005,
"loss": 3.4756,
"step": 34550
},
{
"epoch": 3.7240340114088903,
"grad_norm": 0.648026704788208,
"learning_rate": 0.000377047731925439,
"loss": 3.4632,
"step": 34600
},
{
"epoch": 3.7294155634484984,
"grad_norm": 0.6278716921806335,
"learning_rate": 0.0003767244908953776,
"loss": 3.4511,
"step": 34650
},
{
"epoch": 3.7347971154881066,
"grad_norm": 0.5706120729446411,
"learning_rate": 0.0003764012498653162,
"loss": 3.4438,
"step": 34700
},
{
"epoch": 3.740178667527715,
"grad_norm": 0.6078208088874817,
"learning_rate": 0.00037607800883525475,
"loss": 3.4435,
"step": 34750
},
{
"epoch": 3.745560219567323,
"grad_norm": 0.5775535702705383,
"learning_rate": 0.0003757547678051934,
"loss": 3.4463,
"step": 34800
},
{
"epoch": 3.7509417716069313,
"grad_norm": 0.6093581914901733,
"learning_rate": 0.000375431526775132,
"loss": 3.4498,
"step": 34850
},
{
"epoch": 3.75632332364654,
"grad_norm": 0.6318726539611816,
"learning_rate": 0.00037510828574507054,
"loss": 3.4469,
"step": 34900
},
{
"epoch": 3.761704875686148,
"grad_norm": 0.5926064252853394,
"learning_rate": 0.00037479150953561034,
"loss": 3.4331,
"step": 34950
},
{
"epoch": 3.767086427725756,
"grad_norm": 0.5718520283699036,
"learning_rate": 0.00037446826850554894,
"loss": 3.4391,
"step": 35000
},
{
"epoch": 3.767086427725756,
"eval_accuracy": 0.3750746581155395,
"eval_loss": 3.45552396774292,
"eval_runtime": 180.4933,
"eval_samples_per_second": 99.788,
"eval_steps_per_second": 6.238,
"step": 35000
},
{
"epoch": 3.772467979765364,
"grad_norm": 0.6313642859458923,
"learning_rate": 0.00037414502747548753,
"loss": 3.4635,
"step": 35050
},
{
"epoch": 3.7778495318049723,
"grad_norm": 0.6304598450660706,
"learning_rate": 0.00037382178644542607,
"loss": 3.4657,
"step": 35100
},
{
"epoch": 3.783231083844581,
"grad_norm": 0.5782265067100525,
"learning_rate": 0.00037349854541536467,
"loss": 3.4483,
"step": 35150
},
{
"epoch": 3.788612635884189,
"grad_norm": 0.6751189231872559,
"learning_rate": 0.0003731753043853033,
"loss": 3.4494,
"step": 35200
},
{
"epoch": 3.793994187923797,
"grad_norm": 0.6328338980674744,
"learning_rate": 0.00037285206335524186,
"loss": 3.4462,
"step": 35250
},
{
"epoch": 3.7993757399634056,
"grad_norm": 0.608933687210083,
"learning_rate": 0.00037252882232518045,
"loss": 3.451,
"step": 35300
},
{
"epoch": 3.8047572920030137,
"grad_norm": 0.5944675803184509,
"learning_rate": 0.000372205581295119,
"loss": 3.4391,
"step": 35350
},
{
"epoch": 3.810138844042622,
"grad_norm": 0.5831218957901001,
"learning_rate": 0.0003718823402650576,
"loss": 3.4609,
"step": 35400
},
{
"epoch": 3.8155203960822304,
"grad_norm": 0.6026729345321655,
"learning_rate": 0.00037155909923499624,
"loss": 3.43,
"step": 35450
},
{
"epoch": 3.8209019481218385,
"grad_norm": 0.5765225887298584,
"learning_rate": 0.0003712358582049348,
"loss": 3.4546,
"step": 35500
},
{
"epoch": 3.8262835001614466,
"grad_norm": 0.6434342861175537,
"learning_rate": 0.00037091261717487337,
"loss": 3.4432,
"step": 35550
},
{
"epoch": 3.8316650522010547,
"grad_norm": 0.6166746020317078,
"learning_rate": 0.00037058937614481197,
"loss": 3.4316,
"step": 35600
},
{
"epoch": 3.837046604240663,
"grad_norm": 0.5961298942565918,
"learning_rate": 0.0003702661351147505,
"loss": 3.449,
"step": 35650
},
{
"epoch": 3.8424281562802713,
"grad_norm": 0.5844510793685913,
"learning_rate": 0.0003699428940846891,
"loss": 3.4387,
"step": 35700
},
{
"epoch": 3.8478097083198795,
"grad_norm": 0.6083241105079651,
"learning_rate": 0.00036961965305462775,
"loss": 3.4513,
"step": 35750
},
{
"epoch": 3.8531912603594876,
"grad_norm": 0.5536696314811707,
"learning_rate": 0.0003692964120245663,
"loss": 3.435,
"step": 35800
},
{
"epoch": 3.858572812399096,
"grad_norm": 0.5852833390235901,
"learning_rate": 0.0003689731709945049,
"loss": 3.4571,
"step": 35850
},
{
"epoch": 3.863954364438704,
"grad_norm": 0.6565154194831848,
"learning_rate": 0.0003686499299644434,
"loss": 3.4436,
"step": 35900
},
{
"epoch": 3.8693359164783123,
"grad_norm": 0.6170244812965393,
"learning_rate": 0.000368326688934382,
"loss": 3.4601,
"step": 35950
},
{
"epoch": 3.8747174685179204,
"grad_norm": 0.6194032430648804,
"learning_rate": 0.0003680034479043206,
"loss": 3.4502,
"step": 36000
},
{
"epoch": 3.8747174685179204,
"eval_accuracy": 0.3757220120211382,
"eval_loss": 3.4532008171081543,
"eval_runtime": 180.0191,
"eval_samples_per_second": 100.05,
"eval_steps_per_second": 6.255,
"step": 36000
},
{
"epoch": 3.8800990205575285,
"grad_norm": 0.6787853837013245,
"learning_rate": 0.0003676802068742592,
"loss": 3.4645,
"step": 36050
},
{
"epoch": 3.885480572597137,
"grad_norm": 0.5699832439422607,
"learning_rate": 0.0003673569658441978,
"loss": 3.4565,
"step": 36100
},
{
"epoch": 3.890862124636745,
"grad_norm": 0.6348679661750793,
"learning_rate": 0.0003670337248141364,
"loss": 3.4306,
"step": 36150
},
{
"epoch": 3.8962436766763533,
"grad_norm": 0.611460268497467,
"learning_rate": 0.00036671048378407494,
"loss": 3.4289,
"step": 36200
},
{
"epoch": 3.901625228715962,
"grad_norm": 0.6488123536109924,
"learning_rate": 0.00036638724275401353,
"loss": 3.4341,
"step": 36250
},
{
"epoch": 3.90700678075557,
"grad_norm": 0.5952411890029907,
"learning_rate": 0.0003660640017239522,
"loss": 3.4465,
"step": 36300
},
{
"epoch": 3.912388332795178,
"grad_norm": 0.5958573222160339,
"learning_rate": 0.0003657407606938907,
"loss": 3.4567,
"step": 36350
},
{
"epoch": 3.9177698848347866,
"grad_norm": 0.6205569505691528,
"learning_rate": 0.0003654175196638293,
"loss": 3.4473,
"step": 36400
},
{
"epoch": 3.9231514368743947,
"grad_norm": 0.6399919390678406,
"learning_rate": 0.00036509427863376786,
"loss": 3.4574,
"step": 36450
},
{
"epoch": 3.928532988914003,
"grad_norm": 0.6450821161270142,
"learning_rate": 0.00036477103760370645,
"loss": 3.4467,
"step": 36500
},
{
"epoch": 3.933914540953611,
"grad_norm": 0.6391284465789795,
"learning_rate": 0.00036444779657364505,
"loss": 3.4279,
"step": 36550
},
{
"epoch": 3.939296092993219,
"grad_norm": 0.5766438245773315,
"learning_rate": 0.00036412455554358364,
"loss": 3.4366,
"step": 36600
},
{
"epoch": 3.9446776450328276,
"grad_norm": 0.6284584403038025,
"learning_rate": 0.00036380131451352224,
"loss": 3.4329,
"step": 36650
},
{
"epoch": 3.9500591970724357,
"grad_norm": 0.5596027374267578,
"learning_rate": 0.00036347807348346083,
"loss": 3.4525,
"step": 36700
},
{
"epoch": 3.955440749112044,
"grad_norm": 0.6084979772567749,
"learning_rate": 0.00036315483245339937,
"loss": 3.4452,
"step": 36750
},
{
"epoch": 3.9608223011516523,
"grad_norm": 0.6172415614128113,
"learning_rate": 0.00036283159142333797,
"loss": 3.4485,
"step": 36800
},
{
"epoch": 3.9662038531912605,
"grad_norm": 0.6101384162902832,
"learning_rate": 0.0003625083503932765,
"loss": 3.4566,
"step": 36850
},
{
"epoch": 3.9715854052308686,
"grad_norm": 0.5563173890113831,
"learning_rate": 0.00036218510936321516,
"loss": 3.4444,
"step": 36900
},
{
"epoch": 3.9769669572704767,
"grad_norm": 0.6216126084327698,
"learning_rate": 0.00036186186833315375,
"loss": 3.4486,
"step": 36950
},
{
"epoch": 3.9823485093100848,
"grad_norm": 0.6059814691543579,
"learning_rate": 0.0003615386273030923,
"loss": 3.4418,
"step": 37000
},
{
"epoch": 3.9823485093100848,
"eval_accuracy": 0.37635067763003277,
"eval_loss": 3.4460902214050293,
"eval_runtime": 180.6452,
"eval_samples_per_second": 99.704,
"eval_steps_per_second": 6.233,
"step": 37000
},
{
"epoch": 3.9877300613496933,
"grad_norm": 0.5866619944572449,
"learning_rate": 0.0003612153862730309,
"loss": 3.4566,
"step": 37050
},
{
"epoch": 3.9931116133893014,
"grad_norm": 0.6042488813400269,
"learning_rate": 0.0003608986100635707,
"loss": 3.4515,
"step": 37100
},
{
"epoch": 3.9984931654289095,
"grad_norm": 0.6132956147193909,
"learning_rate": 0.0003605753690335093,
"loss": 3.4339,
"step": 37150
},
{
"epoch": 4.003874717468518,
"grad_norm": 0.5894633531570435,
"learning_rate": 0.0003602521280034478,
"loss": 3.3714,
"step": 37200
},
{
"epoch": 4.009256269508126,
"grad_norm": 0.5937063694000244,
"learning_rate": 0.0003599288869733865,
"loss": 3.3621,
"step": 37250
},
{
"epoch": 4.014637821547734,
"grad_norm": 0.6420974135398865,
"learning_rate": 0.00035960564594332507,
"loss": 3.3542,
"step": 37300
},
{
"epoch": 4.020019373587343,
"grad_norm": 0.5973079800605774,
"learning_rate": 0.0003592824049132636,
"loss": 3.3511,
"step": 37350
},
{
"epoch": 4.0254009256269505,
"grad_norm": 0.5899896025657654,
"learning_rate": 0.0003589591638832022,
"loss": 3.3511,
"step": 37400
},
{
"epoch": 4.030782477666559,
"grad_norm": 0.6211423277854919,
"learning_rate": 0.0003586359228531408,
"loss": 3.365,
"step": 37450
},
{
"epoch": 4.036164029706168,
"grad_norm": 0.5809051394462585,
"learning_rate": 0.00035831268182307934,
"loss": 3.3682,
"step": 37500
},
{
"epoch": 4.041545581745775,
"grad_norm": 0.6257584691047668,
"learning_rate": 0.000357989440793018,
"loss": 3.3659,
"step": 37550
},
{
"epoch": 4.046927133785384,
"grad_norm": 0.6464672088623047,
"learning_rate": 0.0003576661997629566,
"loss": 3.3591,
"step": 37600
},
{
"epoch": 4.0523086858249915,
"grad_norm": 0.6366177201271057,
"learning_rate": 0.0003573429587328951,
"loss": 3.3684,
"step": 37650
},
{
"epoch": 4.0576902378646,
"grad_norm": 0.6766300797462463,
"learning_rate": 0.0003570197177028337,
"loss": 3.3451,
"step": 37700
},
{
"epoch": 4.063071789904209,
"grad_norm": 0.620211124420166,
"learning_rate": 0.00035669647667277226,
"loss": 3.3523,
"step": 37750
},
{
"epoch": 4.068453341943816,
"grad_norm": 0.6328372955322266,
"learning_rate": 0.00035637323564271085,
"loss": 3.369,
"step": 37800
},
{
"epoch": 4.073834893983425,
"grad_norm": 0.6072499752044678,
"learning_rate": 0.0003560499946126495,
"loss": 3.3667,
"step": 37850
},
{
"epoch": 4.079216446023033,
"grad_norm": 0.5885031819343567,
"learning_rate": 0.00035572675358258804,
"loss": 3.3755,
"step": 37900
},
{
"epoch": 4.084597998062641,
"grad_norm": 0.5657071471214294,
"learning_rate": 0.00035540351255252664,
"loss": 3.3716,
"step": 37950
},
{
"epoch": 4.08997955010225,
"grad_norm": 0.6271093487739563,
"learning_rate": 0.00035508027152246523,
"loss": 3.3718,
"step": 38000
},
{
"epoch": 4.08997955010225,
"eval_accuracy": 0.3768978535730665,
"eval_loss": 3.4471983909606934,
"eval_runtime": 180.3248,
"eval_samples_per_second": 99.881,
"eval_steps_per_second": 6.244,
"step": 38000
},
{
"epoch": 4.095361102141858,
"grad_norm": 0.6071150302886963,
"learning_rate": 0.0003547570304924038,
"loss": 3.3633,
"step": 38050
},
{
"epoch": 4.100742654181466,
"grad_norm": 0.5937423706054688,
"learning_rate": 0.0003544337894623424,
"loss": 3.3846,
"step": 38100
},
{
"epoch": 4.106124206221074,
"grad_norm": 0.5915195345878601,
"learning_rate": 0.000354110548432281,
"loss": 3.3781,
"step": 38150
},
{
"epoch": 4.111505758260682,
"grad_norm": 0.6321580410003662,
"learning_rate": 0.00035378730740221956,
"loss": 3.3696,
"step": 38200
},
{
"epoch": 4.1168873103002905,
"grad_norm": 0.6198523640632629,
"learning_rate": 0.00035346406637215815,
"loss": 3.3694,
"step": 38250
},
{
"epoch": 4.122268862339899,
"grad_norm": 0.6256786584854126,
"learning_rate": 0.0003531408253420967,
"loss": 3.375,
"step": 38300
},
{
"epoch": 4.127650414379507,
"grad_norm": 0.6233739256858826,
"learning_rate": 0.0003528175843120353,
"loss": 3.3891,
"step": 38350
},
{
"epoch": 4.133031966419115,
"grad_norm": 0.6164677143096924,
"learning_rate": 0.00035249434328197394,
"loss": 3.366,
"step": 38400
},
{
"epoch": 4.138413518458724,
"grad_norm": 0.6025086045265198,
"learning_rate": 0.0003521711022519125,
"loss": 3.3528,
"step": 38450
},
{
"epoch": 4.1437950704983315,
"grad_norm": 0.6623359322547913,
"learning_rate": 0.00035184786122185107,
"loss": 3.3703,
"step": 38500
},
{
"epoch": 4.14917662253794,
"grad_norm": 0.6005052924156189,
"learning_rate": 0.00035152462019178967,
"loss": 3.3746,
"step": 38550
},
{
"epoch": 4.154558174577549,
"grad_norm": 0.6520372033119202,
"learning_rate": 0.0003512013791617282,
"loss": 3.3717,
"step": 38600
},
{
"epoch": 4.159939726617156,
"grad_norm": 0.5965567827224731,
"learning_rate": 0.0003508781381316668,
"loss": 3.3652,
"step": 38650
},
{
"epoch": 4.165321278656765,
"grad_norm": 0.6479796171188354,
"learning_rate": 0.00035055489710160545,
"loss": 3.3732,
"step": 38700
},
{
"epoch": 4.1707028306963725,
"grad_norm": 0.5954575538635254,
"learning_rate": 0.000350231656071544,
"loss": 3.3845,
"step": 38750
},
{
"epoch": 4.176084382735981,
"grad_norm": 0.6750271916389465,
"learning_rate": 0.0003499084150414826,
"loss": 3.3609,
"step": 38800
},
{
"epoch": 4.18146593477559,
"grad_norm": 0.6075855493545532,
"learning_rate": 0.0003495851740114211,
"loss": 3.3825,
"step": 38850
},
{
"epoch": 4.186847486815197,
"grad_norm": 0.6466742753982544,
"learning_rate": 0.0003492619329813597,
"loss": 3.3572,
"step": 38900
},
{
"epoch": 4.192229038854806,
"grad_norm": 0.6090953350067139,
"learning_rate": 0.0003489386919512983,
"loss": 3.3637,
"step": 38950
},
{
"epoch": 4.197610590894414,
"grad_norm": 0.6194248795509338,
"learning_rate": 0.0003486154509212369,
"loss": 3.3859,
"step": 39000
},
{
"epoch": 4.197610590894414,
"eval_accuracy": 0.3771731799907884,
"eval_loss": 3.4430482387542725,
"eval_runtime": 180.581,
"eval_samples_per_second": 99.739,
"eval_steps_per_second": 6.235,
"step": 39000
},
{
"epoch": 4.202992142934022,
"grad_norm": 0.6370955109596252,
"learning_rate": 0.0003482922098911755,
"loss": 3.3882,
"step": 39050
},
{
"epoch": 4.208373694973631,
"grad_norm": 0.6059002876281738,
"learning_rate": 0.0003479689688611141,
"loss": 3.3683,
"step": 39100
},
{
"epoch": 4.213755247013238,
"grad_norm": 0.6166430711746216,
"learning_rate": 0.00034764572783105264,
"loss": 3.3749,
"step": 39150
},
{
"epoch": 4.219136799052847,
"grad_norm": 0.6343166828155518,
"learning_rate": 0.00034732248680099123,
"loss": 3.3703,
"step": 39200
},
{
"epoch": 4.224518351092455,
"grad_norm": 0.6673994064331055,
"learning_rate": 0.00034700571059153104,
"loss": 3.3923,
"step": 39250
},
{
"epoch": 4.229899903132063,
"grad_norm": 0.5913220643997192,
"learning_rate": 0.00034668246956146963,
"loss": 3.394,
"step": 39300
},
{
"epoch": 4.2352814551716715,
"grad_norm": 0.6904832720756531,
"learning_rate": 0.00034635922853140823,
"loss": 3.3711,
"step": 39350
},
{
"epoch": 4.24066300721128,
"grad_norm": 0.6442829966545105,
"learning_rate": 0.0003460359875013468,
"loss": 3.3899,
"step": 39400
},
{
"epoch": 4.246044559250888,
"grad_norm": 0.5766376852989197,
"learning_rate": 0.0003457127464712854,
"loss": 3.3693,
"step": 39450
},
{
"epoch": 4.251426111290496,
"grad_norm": 0.6282715797424316,
"learning_rate": 0.00034538950544122396,
"loss": 3.3743,
"step": 39500
},
{
"epoch": 4.256807663330104,
"grad_norm": 0.5965016484260559,
"learning_rate": 0.00034506626441116255,
"loss": 3.3786,
"step": 39550
},
{
"epoch": 4.2621892153697125,
"grad_norm": 0.6471324563026428,
"learning_rate": 0.0003447430233811011,
"loss": 3.3732,
"step": 39600
},
{
"epoch": 4.267570767409321,
"grad_norm": 0.6036005616188049,
"learning_rate": 0.00034441978235103974,
"loss": 3.3769,
"step": 39650
},
{
"epoch": 4.272952319448929,
"grad_norm": 0.6378051042556763,
"learning_rate": 0.00034409654132097834,
"loss": 3.3983,
"step": 39700
},
{
"epoch": 4.278333871488537,
"grad_norm": 0.6300690770149231,
"learning_rate": 0.0003437733002909169,
"loss": 3.3651,
"step": 39750
},
{
"epoch": 4.283715423528146,
"grad_norm": 0.5953393578529358,
"learning_rate": 0.00034345005926085547,
"loss": 3.3929,
"step": 39800
},
{
"epoch": 4.2890969755677535,
"grad_norm": 0.5986993312835693,
"learning_rate": 0.00034312681823079407,
"loss": 3.3837,
"step": 39850
},
{
"epoch": 4.294478527607362,
"grad_norm": 0.6578076481819153,
"learning_rate": 0.00034280357720073266,
"loss": 3.3791,
"step": 39900
},
{
"epoch": 4.299860079646971,
"grad_norm": 0.6022131443023682,
"learning_rate": 0.00034248033617067126,
"loss": 3.3983,
"step": 39950
},
{
"epoch": 4.305241631686578,
"grad_norm": 0.6208112835884094,
"learning_rate": 0.00034215709514060985,
"loss": 3.3809,
"step": 40000
},
{
"epoch": 4.305241631686578,
"eval_accuracy": 0.37746382646569215,
"eval_loss": 3.43764591217041,
"eval_runtime": 180.2919,
"eval_samples_per_second": 99.899,
"eval_steps_per_second": 6.245,
"step": 40000
},
{
"epoch": 4.310623183726187,
"grad_norm": 0.5868871212005615,
"learning_rate": 0.0003418338541105484,
"loss": 3.3803,
"step": 40050
},
{
"epoch": 4.3160047357657945,
"grad_norm": 0.5954568982124329,
"learning_rate": 0.000341510613080487,
"loss": 3.3836,
"step": 40100
},
{
"epoch": 4.321386287805403,
"grad_norm": 0.6300106644630432,
"learning_rate": 0.0003411873720504255,
"loss": 3.3725,
"step": 40150
},
{
"epoch": 4.326767839845012,
"grad_norm": 0.6355574131011963,
"learning_rate": 0.0003408641310203642,
"loss": 3.3852,
"step": 40200
},
{
"epoch": 4.332149391884619,
"grad_norm": 0.6317576766014099,
"learning_rate": 0.00034054088999030277,
"loss": 3.3944,
"step": 40250
},
{
"epoch": 4.337530943924228,
"grad_norm": 0.6220029592514038,
"learning_rate": 0.0003402176489602413,
"loss": 3.3843,
"step": 40300
},
{
"epoch": 4.342912495963836,
"grad_norm": 0.6280874013900757,
"learning_rate": 0.0003398944079301799,
"loss": 3.3892,
"step": 40350
},
{
"epoch": 4.348294048003444,
"grad_norm": 0.6555618047714233,
"learning_rate": 0.0003395711669001185,
"loss": 3.3741,
"step": 40400
},
{
"epoch": 4.3536756000430525,
"grad_norm": 0.6413003206253052,
"learning_rate": 0.00033924792587005704,
"loss": 3.3906,
"step": 40450
},
{
"epoch": 4.359057152082661,
"grad_norm": 0.6359059810638428,
"learning_rate": 0.0003389246848399957,
"loss": 3.3967,
"step": 40500
},
{
"epoch": 4.364438704122269,
"grad_norm": 0.6359884738922119,
"learning_rate": 0.0003386014438099343,
"loss": 3.3777,
"step": 40550
},
{
"epoch": 4.369820256161877,
"grad_norm": 0.6541460156440735,
"learning_rate": 0.0003382782027798728,
"loss": 3.3911,
"step": 40600
},
{
"epoch": 4.375201808201485,
"grad_norm": 0.6076342463493347,
"learning_rate": 0.0003379549617498114,
"loss": 3.388,
"step": 40650
},
{
"epoch": 4.3805833602410935,
"grad_norm": 0.6610375046730042,
"learning_rate": 0.00033763172071974996,
"loss": 3.3873,
"step": 40700
},
{
"epoch": 4.385964912280702,
"grad_norm": 0.6042363047599792,
"learning_rate": 0.00033730847968968855,
"loss": 3.3692,
"step": 40750
},
{
"epoch": 4.39134646432031,
"grad_norm": 0.6662680506706238,
"learning_rate": 0.0003369852386596272,
"loss": 3.3931,
"step": 40800
},
{
"epoch": 4.396728016359918,
"grad_norm": 0.6107580065727234,
"learning_rate": 0.00033666199762956574,
"loss": 3.3989,
"step": 40850
},
{
"epoch": 4.402109568399527,
"grad_norm": 0.6042559146881104,
"learning_rate": 0.00033633875659950434,
"loss": 3.3916,
"step": 40900
},
{
"epoch": 4.4074911204391345,
"grad_norm": 0.6182630062103271,
"learning_rate": 0.0003360155155694429,
"loss": 3.4034,
"step": 40950
},
{
"epoch": 4.412872672478743,
"grad_norm": 0.6265887022018433,
"learning_rate": 0.0003356922745393815,
"loss": 3.384,
"step": 41000
},
{
"epoch": 4.412872672478743,
"eval_accuracy": 0.3774822974566206,
"eval_loss": 3.438303232192993,
"eval_runtime": 180.488,
"eval_samples_per_second": 99.791,
"eval_steps_per_second": 6.239,
"step": 41000
},
{
"epoch": 4.418254224518351,
"grad_norm": 0.607819676399231,
"learning_rate": 0.0003353690335093201,
"loss": 3.3809,
"step": 41050
},
{
"epoch": 4.423635776557959,
"grad_norm": 0.6153405904769897,
"learning_rate": 0.0003350457924792587,
"loss": 3.3748,
"step": 41100
},
{
"epoch": 4.429017328597568,
"grad_norm": 0.6520572900772095,
"learning_rate": 0.00033472255144919726,
"loss": 3.3901,
"step": 41150
},
{
"epoch": 4.4343988806371755,
"grad_norm": 0.6429771780967712,
"learning_rate": 0.00033439931041913585,
"loss": 3.3964,
"step": 41200
},
{
"epoch": 4.439780432676784,
"grad_norm": 0.6534312963485718,
"learning_rate": 0.0003340760693890744,
"loss": 3.3768,
"step": 41250
},
{
"epoch": 4.445161984716393,
"grad_norm": 0.639164388179779,
"learning_rate": 0.00033375929317961425,
"loss": 3.3835,
"step": 41300
},
{
"epoch": 4.450543536756,
"grad_norm": 0.6154906153678894,
"learning_rate": 0.0003334360521495528,
"loss": 3.4067,
"step": 41350
},
{
"epoch": 4.455925088795609,
"grad_norm": 0.6143419146537781,
"learning_rate": 0.0003331128111194914,
"loss": 3.3849,
"step": 41400
},
{
"epoch": 4.461306640835216,
"grad_norm": 0.6307302117347717,
"learning_rate": 0.00033278957008943004,
"loss": 3.3831,
"step": 41450
},
{
"epoch": 4.466688192874825,
"grad_norm": 0.6415225267410278,
"learning_rate": 0.0003324663290593686,
"loss": 3.3719,
"step": 41500
},
{
"epoch": 4.4720697449144335,
"grad_norm": 0.594627320766449,
"learning_rate": 0.00033214308802930717,
"loss": 3.3902,
"step": 41550
},
{
"epoch": 4.477451296954041,
"grad_norm": 0.6894360184669495,
"learning_rate": 0.0003318198469992457,
"loss": 3.398,
"step": 41600
},
{
"epoch": 4.48283284899365,
"grad_norm": 0.5951968431472778,
"learning_rate": 0.0003314966059691843,
"loss": 3.3817,
"step": 41650
},
{
"epoch": 4.488214401033258,
"grad_norm": 0.6982078552246094,
"learning_rate": 0.00033117336493912296,
"loss": 3.3932,
"step": 41700
},
{
"epoch": 4.493595953072866,
"grad_norm": 0.6590989828109741,
"learning_rate": 0.0003308501239090615,
"loss": 3.3916,
"step": 41750
},
{
"epoch": 4.4989775051124745,
"grad_norm": 0.626447319984436,
"learning_rate": 0.0003305268828790001,
"loss": 3.3732,
"step": 41800
},
{
"epoch": 4.504359057152083,
"grad_norm": 0.624427080154419,
"learning_rate": 0.0003302036418489387,
"loss": 3.3799,
"step": 41850
},
{
"epoch": 4.509740609191691,
"grad_norm": 0.6644538640975952,
"learning_rate": 0.0003298804008188772,
"loss": 3.4,
"step": 41900
},
{
"epoch": 4.515122161231299,
"grad_norm": 0.6066876649856567,
"learning_rate": 0.0003295571597888158,
"loss": 3.3932,
"step": 41950
},
{
"epoch": 4.520503713270907,
"grad_norm": 0.5975305438041687,
"learning_rate": 0.00032923391875875447,
"loss": 3.3768,
"step": 42000
},
{
"epoch": 4.520503713270907,
"eval_accuracy": 0.37895345755762866,
"eval_loss": 3.4301979541778564,
"eval_runtime": 180.422,
"eval_samples_per_second": 99.827,
"eval_steps_per_second": 6.241,
"step": 42000
},
{
"epoch": 4.5258852653105155,
"grad_norm": 0.6146498918533325,
"learning_rate": 0.000328910677728693,
"loss": 3.3973,
"step": 42050
},
{
"epoch": 4.531266817350124,
"grad_norm": 0.6150146722793579,
"learning_rate": 0.0003285874366986316,
"loss": 3.3871,
"step": 42100
},
{
"epoch": 4.536648369389732,
"grad_norm": 0.5906317830085754,
"learning_rate": 0.00032826419566857015,
"loss": 3.3871,
"step": 42150
},
{
"epoch": 4.54202992142934,
"grad_norm": 0.6632665395736694,
"learning_rate": 0.00032794095463850874,
"loss": 3.3829,
"step": 42200
},
{
"epoch": 4.547411473468949,
"grad_norm": 0.6187365055084229,
"learning_rate": 0.0003276177136084473,
"loss": 3.3977,
"step": 42250
},
{
"epoch": 4.5527930255085565,
"grad_norm": 0.6101699471473694,
"learning_rate": 0.00032729447257838593,
"loss": 3.3902,
"step": 42300
},
{
"epoch": 4.558174577548165,
"grad_norm": 0.6219088435173035,
"learning_rate": 0.0003269712315483245,
"loss": 3.3671,
"step": 42350
},
{
"epoch": 4.563556129587774,
"grad_norm": 0.6886587142944336,
"learning_rate": 0.00032664799051826306,
"loss": 3.3877,
"step": 42400
},
{
"epoch": 4.568937681627381,
"grad_norm": 0.7076488137245178,
"learning_rate": 0.00032632474948820166,
"loss": 3.3835,
"step": 42450
},
{
"epoch": 4.57431923366699,
"grad_norm": 0.6349630355834961,
"learning_rate": 0.00032600150845814025,
"loss": 3.4027,
"step": 42500
},
{
"epoch": 4.579700785706597,
"grad_norm": 0.6757253408432007,
"learning_rate": 0.0003256782674280788,
"loss": 3.376,
"step": 42550
},
{
"epoch": 4.585082337746206,
"grad_norm": 0.6440753936767578,
"learning_rate": 0.00032535502639801744,
"loss": 3.387,
"step": 42600
},
{
"epoch": 4.5904638897858145,
"grad_norm": 0.5789088606834412,
"learning_rate": 0.00032503178536795604,
"loss": 3.3952,
"step": 42650
},
{
"epoch": 4.595845441825422,
"grad_norm": 0.6196984052658081,
"learning_rate": 0.0003247085443378946,
"loss": 3.3845,
"step": 42700
},
{
"epoch": 4.601226993865031,
"grad_norm": 0.5972916483879089,
"learning_rate": 0.0003243853033078332,
"loss": 3.3952,
"step": 42750
},
{
"epoch": 4.606608545904638,
"grad_norm": 0.6403440237045288,
"learning_rate": 0.0003240620622777717,
"loss": 3.3873,
"step": 42800
},
{
"epoch": 4.611990097944247,
"grad_norm": 0.6815773844718933,
"learning_rate": 0.00032373882124771036,
"loss": 3.415,
"step": 42850
},
{
"epoch": 4.6173716499838555,
"grad_norm": 0.6044899225234985,
"learning_rate": 0.00032341558021764896,
"loss": 3.385,
"step": 42900
},
{
"epoch": 4.622753202023463,
"grad_norm": 0.6453890800476074,
"learning_rate": 0.0003230923391875875,
"loss": 3.3874,
"step": 42950
},
{
"epoch": 4.628134754063072,
"grad_norm": 0.6140652298927307,
"learning_rate": 0.0003227690981575261,
"loss": 3.381,
"step": 43000
},
{
"epoch": 4.628134754063072,
"eval_accuracy": 0.3790061542082187,
"eval_loss": 3.4236814975738525,
"eval_runtime": 180.3345,
"eval_samples_per_second": 99.876,
"eval_steps_per_second": 6.244,
"step": 43000
},
{
"epoch": 4.63351630610268,
"grad_norm": 0.6114611029624939,
"learning_rate": 0.0003224458571274647,
"loss": 3.3835,
"step": 43050
},
{
"epoch": 4.638897858142288,
"grad_norm": 0.621130645275116,
"learning_rate": 0.00032212261609740323,
"loss": 3.3746,
"step": 43100
},
{
"epoch": 4.6442794101818965,
"grad_norm": 0.7222752571105957,
"learning_rate": 0.0003217993750673419,
"loss": 3.4018,
"step": 43150
},
{
"epoch": 4.649660962221505,
"grad_norm": 0.6595478057861328,
"learning_rate": 0.00032147613403728047,
"loss": 3.4018,
"step": 43200
},
{
"epoch": 4.655042514261113,
"grad_norm": 0.6354619860649109,
"learning_rate": 0.000321152893007219,
"loss": 3.403,
"step": 43250
},
{
"epoch": 4.660424066300721,
"grad_norm": 0.7083771228790283,
"learning_rate": 0.0003208296519771576,
"loss": 3.3735,
"step": 43300
},
{
"epoch": 4.665805618340329,
"grad_norm": 0.62490314245224,
"learning_rate": 0.0003205128757676974,
"loss": 3.3931,
"step": 43350
},
{
"epoch": 4.6711871703799375,
"grad_norm": 0.6417445540428162,
"learning_rate": 0.000320189634737636,
"loss": 3.402,
"step": 43400
},
{
"epoch": 4.676568722419546,
"grad_norm": 0.613502025604248,
"learning_rate": 0.00031986639370757455,
"loss": 3.3852,
"step": 43450
},
{
"epoch": 4.681950274459154,
"grad_norm": 0.5854348540306091,
"learning_rate": 0.0003195431526775132,
"loss": 3.3849,
"step": 43500
},
{
"epoch": 4.687331826498762,
"grad_norm": 0.6281751990318298,
"learning_rate": 0.0003192199116474518,
"loss": 3.3938,
"step": 43550
},
{
"epoch": 4.692713378538371,
"grad_norm": 0.6472344398498535,
"learning_rate": 0.00031889667061739033,
"loss": 3.3916,
"step": 43600
},
{
"epoch": 4.6980949305779784,
"grad_norm": 0.66495680809021,
"learning_rate": 0.0003185734295873289,
"loss": 3.3904,
"step": 43650
},
{
"epoch": 4.703476482617587,
"grad_norm": 0.6219391822814941,
"learning_rate": 0.00031825665337786873,
"loss": 3.3885,
"step": 43700
},
{
"epoch": 4.7088580346571955,
"grad_norm": 0.5999084115028381,
"learning_rate": 0.00031793341234780733,
"loss": 3.3887,
"step": 43750
},
{
"epoch": 4.714239586696803,
"grad_norm": 0.6449667811393738,
"learning_rate": 0.00031761017131774587,
"loss": 3.3878,
"step": 43800
},
{
"epoch": 4.719621138736412,
"grad_norm": 0.6235863566398621,
"learning_rate": 0.00031728693028768446,
"loss": 3.3608,
"step": 43850
},
{
"epoch": 4.725002690776019,
"grad_norm": 0.6473997831344604,
"learning_rate": 0.0003169636892576231,
"loss": 3.3974,
"step": 43900
},
{
"epoch": 4.730384242815628,
"grad_norm": 0.6401399970054626,
"learning_rate": 0.00031664044822756165,
"loss": 3.4097,
"step": 43950
},
{
"epoch": 4.7357657948552365,
"grad_norm": 0.6407394409179688,
"learning_rate": 0.00031631720719750025,
"loss": 3.3897,
"step": 44000
},
{
"epoch": 4.7357657948552365,
"eval_accuracy": 0.3795691934728733,
"eval_loss": 3.419826030731201,
"eval_runtime": 180.2132,
"eval_samples_per_second": 99.943,
"eval_steps_per_second": 6.248,
"step": 44000
},
{
"epoch": 4.741147346894844,
"grad_norm": 0.6270405650138855,
"learning_rate": 0.00031599396616743884,
"loss": 3.3685,
"step": 44050
},
{
"epoch": 4.746528898934453,
"grad_norm": 0.6910510063171387,
"learning_rate": 0.0003156707251373774,
"loss": 3.3945,
"step": 44100
},
{
"epoch": 4.751910450974061,
"grad_norm": 0.602331280708313,
"learning_rate": 0.000315347484107316,
"loss": 3.3932,
"step": 44150
},
{
"epoch": 4.757292003013669,
"grad_norm": 0.6058678030967712,
"learning_rate": 0.0003150242430772546,
"loss": 3.3845,
"step": 44200
},
{
"epoch": 4.7626735550532775,
"grad_norm": 0.6464012861251831,
"learning_rate": 0.00031470100204719317,
"loss": 3.3934,
"step": 44250
},
{
"epoch": 4.768055107092886,
"grad_norm": 0.6263419985771179,
"learning_rate": 0.00031437776101713176,
"loss": 3.3844,
"step": 44300
},
{
"epoch": 4.773436659132494,
"grad_norm": 0.6222793459892273,
"learning_rate": 0.0003140545199870703,
"loss": 3.385,
"step": 44350
},
{
"epoch": 4.778818211172102,
"grad_norm": 0.7262570261955261,
"learning_rate": 0.0003137312789570089,
"loss": 3.3795,
"step": 44400
},
{
"epoch": 4.78419976321171,
"grad_norm": 0.625326931476593,
"learning_rate": 0.00031340803792694754,
"loss": 3.3824,
"step": 44450
},
{
"epoch": 4.7895813152513185,
"grad_norm": 0.6770031452178955,
"learning_rate": 0.0003130847968968861,
"loss": 3.3799,
"step": 44500
},
{
"epoch": 4.794962867290927,
"grad_norm": 0.6451786160469055,
"learning_rate": 0.0003127615558668247,
"loss": 3.3916,
"step": 44550
},
{
"epoch": 4.800344419330535,
"grad_norm": 0.6783530116081238,
"learning_rate": 0.0003124383148367633,
"loss": 3.3967,
"step": 44600
},
{
"epoch": 4.805725971370143,
"grad_norm": 0.6572859287261963,
"learning_rate": 0.0003121150738067018,
"loss": 3.3786,
"step": 44650
},
{
"epoch": 4.811107523409751,
"grad_norm": 0.6286121606826782,
"learning_rate": 0.0003117918327766404,
"loss": 3.3847,
"step": 44700
},
{
"epoch": 4.8164890754493594,
"grad_norm": 0.6354442238807678,
"learning_rate": 0.00031146859174657906,
"loss": 3.3885,
"step": 44750
},
{
"epoch": 4.821870627488968,
"grad_norm": 0.6219363212585449,
"learning_rate": 0.0003111453507165176,
"loss": 3.3731,
"step": 44800
},
{
"epoch": 4.827252179528576,
"grad_norm": 0.6498527526855469,
"learning_rate": 0.0003108221096864562,
"loss": 3.3757,
"step": 44850
},
{
"epoch": 4.832633731568184,
"grad_norm": 0.6466692090034485,
"learning_rate": 0.00031049886865639473,
"loss": 3.3938,
"step": 44900
},
{
"epoch": 4.838015283607793,
"grad_norm": 0.6414273381233215,
"learning_rate": 0.00031017562762633333,
"loss": 3.3908,
"step": 44950
},
{
"epoch": 4.8433968356474,
"grad_norm": 0.6336743831634521,
"learning_rate": 0.00030985238659627187,
"loss": 3.4056,
"step": 45000
},
{
"epoch": 4.8433968356474,
"eval_accuracy": 0.37976868017490073,
"eval_loss": 3.4165797233581543,
"eval_runtime": 179.7792,
"eval_samples_per_second": 100.184,
"eval_steps_per_second": 6.263,
"step": 45000
},
{
"epoch": 4.848778387687009,
"grad_norm": 0.6573559045791626,
"learning_rate": 0.0003095291455662105,
"loss": 3.4003,
"step": 45050
},
{
"epoch": 4.8541599397266175,
"grad_norm": 0.6122683882713318,
"learning_rate": 0.0003092059045361491,
"loss": 3.4081,
"step": 45100
},
{
"epoch": 4.859541491766225,
"grad_norm": 0.6225839853286743,
"learning_rate": 0.00030888266350608765,
"loss": 3.3859,
"step": 45150
},
{
"epoch": 4.864923043805834,
"grad_norm": 0.6035463213920593,
"learning_rate": 0.00030855942247602625,
"loss": 3.3995,
"step": 45200
},
{
"epoch": 4.870304595845441,
"grad_norm": 0.620333194732666,
"learning_rate": 0.00030823618144596484,
"loss": 3.3876,
"step": 45250
},
{
"epoch": 4.87568614788505,
"grad_norm": 0.6101974844932556,
"learning_rate": 0.0003079129404159035,
"loss": 3.3992,
"step": 45300
},
{
"epoch": 4.8810676999246585,
"grad_norm": 0.6473581194877625,
"learning_rate": 0.00030758969938584203,
"loss": 3.4209,
"step": 45350
},
{
"epoch": 4.886449251964266,
"grad_norm": 0.6419587731361389,
"learning_rate": 0.0003072664583557806,
"loss": 3.3862,
"step": 45400
},
{
"epoch": 4.891830804003875,
"grad_norm": 0.6355916261672974,
"learning_rate": 0.00030694321732571917,
"loss": 3.3831,
"step": 45450
},
{
"epoch": 4.897212356043483,
"grad_norm": 0.6475318670272827,
"learning_rate": 0.00030661997629565776,
"loss": 3.3749,
"step": 45500
},
{
"epoch": 4.902593908083091,
"grad_norm": 0.6229646801948547,
"learning_rate": 0.0003062967352655963,
"loss": 3.3826,
"step": 45550
},
{
"epoch": 4.9079754601226995,
"grad_norm": 0.6335062384605408,
"learning_rate": 0.00030597349423553495,
"loss": 3.3829,
"step": 45600
},
{
"epoch": 4.913357012162308,
"grad_norm": 0.6275141835212708,
"learning_rate": 0.00030565025320547355,
"loss": 3.3979,
"step": 45650
},
{
"epoch": 4.918738564201916,
"grad_norm": 0.6986739039421082,
"learning_rate": 0.0003053270121754121,
"loss": 3.4014,
"step": 45700
},
{
"epoch": 4.924120116241524,
"grad_norm": 0.7142667174339294,
"learning_rate": 0.0003050037711453507,
"loss": 3.3784,
"step": 45750
},
{
"epoch": 4.929501668281132,
"grad_norm": 0.6080502271652222,
"learning_rate": 0.0003046805301152893,
"loss": 3.3835,
"step": 45800
},
{
"epoch": 4.9348832203207404,
"grad_norm": 0.6908614635467529,
"learning_rate": 0.0003043572890852278,
"loss": 3.3739,
"step": 45850
},
{
"epoch": 4.940264772360349,
"grad_norm": 0.653312623500824,
"learning_rate": 0.00030403404805516646,
"loss": 3.3974,
"step": 45900
},
{
"epoch": 4.945646324399957,
"grad_norm": 0.6278221607208252,
"learning_rate": 0.00030371080702510506,
"loss": 3.3654,
"step": 45950
},
{
"epoch": 4.951027876439565,
"grad_norm": 0.6530224680900574,
"learning_rate": 0.0003033875659950436,
"loss": 3.3793,
"step": 46000
},
{
"epoch": 4.951027876439565,
"eval_accuracy": 0.3805414195130373,
"eval_loss": 3.408653974533081,
"eval_runtime": 180.4298,
"eval_samples_per_second": 99.823,
"eval_steps_per_second": 6.241,
"step": 46000
},
{
"epoch": 4.956409428479174,
"grad_norm": 0.6106470227241516,
"learning_rate": 0.0003030643249649822,
"loss": 3.3896,
"step": 46050
},
{
"epoch": 4.961790980518781,
"grad_norm": 0.620707631111145,
"learning_rate": 0.00030274108393492073,
"loss": 3.3754,
"step": 46100
},
{
"epoch": 4.96717253255839,
"grad_norm": 0.6534816026687622,
"learning_rate": 0.00030241784290485933,
"loss": 3.3799,
"step": 46150
},
{
"epoch": 4.9725540845979985,
"grad_norm": 0.6243362426757812,
"learning_rate": 0.000302094601874798,
"loss": 3.3661,
"step": 46200
},
{
"epoch": 4.977935636637606,
"grad_norm": 0.6367123126983643,
"learning_rate": 0.0003017713608447365,
"loss": 3.3833,
"step": 46250
},
{
"epoch": 4.983317188677215,
"grad_norm": 0.6724542379379272,
"learning_rate": 0.0003014481198146751,
"loss": 3.3736,
"step": 46300
},
{
"epoch": 4.988698740716822,
"grad_norm": 0.6005682349205017,
"learning_rate": 0.0003011248787846137,
"loss": 3.3951,
"step": 46350
},
{
"epoch": 4.994080292756431,
"grad_norm": 0.6530224084854126,
"learning_rate": 0.00030080163775455225,
"loss": 3.3865,
"step": 46400
},
{
"epoch": 4.9994618447960395,
"grad_norm": 0.6524396538734436,
"learning_rate": 0.0003004783967244909,
"loss": 3.3978,
"step": 46450
},
{
"epoch": 5.004843396835647,
"grad_norm": 0.664138674736023,
"learning_rate": 0.0003001551556944295,
"loss": 3.2914,
"step": 46500
},
{
"epoch": 5.010224948875256,
"grad_norm": 0.6500561237335205,
"learning_rate": 0.00029983191466436803,
"loss": 3.2781,
"step": 46550
},
{
"epoch": 5.015606500914864,
"grad_norm": 0.6164722442626953,
"learning_rate": 0.0002995086736343066,
"loss": 3.2889,
"step": 46600
},
{
"epoch": 5.020988052954472,
"grad_norm": 0.6303194761276245,
"learning_rate": 0.00029918543260424517,
"loss": 3.2927,
"step": 46650
},
{
"epoch": 5.0263696049940805,
"grad_norm": 0.640316367149353,
"learning_rate": 0.0002988621915741838,
"loss": 3.3064,
"step": 46700
},
{
"epoch": 5.031751157033688,
"grad_norm": 0.6622575521469116,
"learning_rate": 0.00029853895054412236,
"loss": 3.3021,
"step": 46750
},
{
"epoch": 5.037132709073297,
"grad_norm": 0.5933485627174377,
"learning_rate": 0.00029821570951406095,
"loss": 3.294,
"step": 46800
},
{
"epoch": 5.042514261112905,
"grad_norm": 0.6456048488616943,
"learning_rate": 0.00029789246848399955,
"loss": 3.2845,
"step": 46850
},
{
"epoch": 5.047895813152513,
"grad_norm": 0.6657723188400269,
"learning_rate": 0.00029756922745393814,
"loss": 3.3037,
"step": 46900
},
{
"epoch": 5.0532773651921215,
"grad_norm": 0.6260988116264343,
"learning_rate": 0.00029724598642387674,
"loss": 3.3016,
"step": 46950
},
{
"epoch": 5.05865891723173,
"grad_norm": 0.6429811716079712,
"learning_rate": 0.0002969227453938153,
"loss": 3.294,
"step": 47000
},
{
"epoch": 5.05865891723173,
"eval_accuracy": 0.3809833108077788,
"eval_loss": 3.4148974418640137,
"eval_runtime": 180.1259,
"eval_samples_per_second": 99.991,
"eval_steps_per_second": 6.251,
"step": 47000
},
{
"epoch": 5.064040469271338,
"grad_norm": 0.6801566481590271,
"learning_rate": 0.00029659950436375387,
"loss": 3.3172,
"step": 47050
},
{
"epoch": 5.069422021310946,
"grad_norm": 0.636012852191925,
"learning_rate": 0.00029627626333369246,
"loss": 3.3021,
"step": 47100
},
{
"epoch": 5.074803573350554,
"grad_norm": 0.6584773063659668,
"learning_rate": 0.00029595302230363106,
"loss": 3.3106,
"step": 47150
},
{
"epoch": 5.080185125390162,
"grad_norm": 0.656566321849823,
"learning_rate": 0.0002956297812735696,
"loss": 3.3225,
"step": 47200
},
{
"epoch": 5.085566677429771,
"grad_norm": 0.6332939267158508,
"learning_rate": 0.00029530654024350825,
"loss": 3.3065,
"step": 47250
},
{
"epoch": 5.090948229469379,
"grad_norm": 0.6884245872497559,
"learning_rate": 0.0002949832992134468,
"loss": 3.2999,
"step": 47300
},
{
"epoch": 5.096329781508987,
"grad_norm": 0.6462180018424988,
"learning_rate": 0.0002946600581833854,
"loss": 3.3009,
"step": 47350
},
{
"epoch": 5.101711333548596,
"grad_norm": 0.6998089551925659,
"learning_rate": 0.000294336817153324,
"loss": 3.3156,
"step": 47400
},
{
"epoch": 5.107092885588203,
"grad_norm": 0.6399083733558655,
"learning_rate": 0.0002940135761232626,
"loss": 3.3166,
"step": 47450
},
{
"epoch": 5.112474437627812,
"grad_norm": 0.6904787421226501,
"learning_rate": 0.0002936903350932011,
"loss": 3.3293,
"step": 47500
},
{
"epoch": 5.1178559896674205,
"grad_norm": 0.7392804026603699,
"learning_rate": 0.0002933670940631397,
"loss": 3.3106,
"step": 47550
},
{
"epoch": 5.123237541707028,
"grad_norm": 0.6379286050796509,
"learning_rate": 0.0002930438530330783,
"loss": 3.3172,
"step": 47600
},
{
"epoch": 5.128619093746637,
"grad_norm": 0.6620005965232849,
"learning_rate": 0.0002927206120030169,
"loss": 3.3154,
"step": 47650
},
{
"epoch": 5.134000645786244,
"grad_norm": 0.6485711932182312,
"learning_rate": 0.0002923973709729555,
"loss": 3.3219,
"step": 47700
},
{
"epoch": 5.139382197825853,
"grad_norm": 0.6479270458221436,
"learning_rate": 0.00029207412994289403,
"loss": 3.3062,
"step": 47750
},
{
"epoch": 5.1447637498654615,
"grad_norm": 0.6357467770576477,
"learning_rate": 0.0002917508889128327,
"loss": 3.3221,
"step": 47800
},
{
"epoch": 5.150145301905069,
"grad_norm": 0.7255944013595581,
"learning_rate": 0.0002914276478827712,
"loss": 3.3055,
"step": 47850
},
{
"epoch": 5.155526853944678,
"grad_norm": 0.6277051568031311,
"learning_rate": 0.00029111087167331103,
"loss": 3.2959,
"step": 47900
},
{
"epoch": 5.160908405984286,
"grad_norm": 0.6564116477966309,
"learning_rate": 0.0002907876306432496,
"loss": 3.3198,
"step": 47950
},
{
"epoch": 5.166289958023894,
"grad_norm": 0.6674607396125793,
"learning_rate": 0.0002904643896131882,
"loss": 3.3209,
"step": 48000
},
{
"epoch": 5.166289958023894,
"eval_accuracy": 0.38105034963956036,
"eval_loss": 3.410404920578003,
"eval_runtime": 180.0365,
"eval_samples_per_second": 100.041,
"eval_steps_per_second": 6.254,
"step": 48000
},
{
"epoch": 5.1716715100635025,
"grad_norm": 0.6171758770942688,
"learning_rate": 0.0002901411485831268,
"loss": 3.319,
"step": 48050
},
{
"epoch": 5.17705306210311,
"grad_norm": 0.7050843238830566,
"learning_rate": 0.00028981790755306535,
"loss": 3.3137,
"step": 48100
},
{
"epoch": 5.182434614142719,
"grad_norm": 0.6586178541183472,
"learning_rate": 0.00028949466652300395,
"loss": 3.3183,
"step": 48150
},
{
"epoch": 5.187816166182327,
"grad_norm": 0.6953296661376953,
"learning_rate": 0.00028917142549294254,
"loss": 3.3212,
"step": 48200
},
{
"epoch": 5.193197718221935,
"grad_norm": 0.6340330839157104,
"learning_rate": 0.00028884818446288114,
"loss": 3.3081,
"step": 48250
},
{
"epoch": 5.198579270261543,
"grad_norm": 0.6623373627662659,
"learning_rate": 0.0002885249434328197,
"loss": 3.3145,
"step": 48300
},
{
"epoch": 5.203960822301152,
"grad_norm": 0.6238620281219482,
"learning_rate": 0.0002882017024027583,
"loss": 3.3138,
"step": 48350
},
{
"epoch": 5.20934237434076,
"grad_norm": 0.6164125204086304,
"learning_rate": 0.00028787846137269687,
"loss": 3.3321,
"step": 48400
},
{
"epoch": 5.214723926380368,
"grad_norm": 0.6289418339729309,
"learning_rate": 0.00028755522034263546,
"loss": 3.3213,
"step": 48450
},
{
"epoch": 5.220105478419977,
"grad_norm": 0.7005164623260498,
"learning_rate": 0.00028723197931257406,
"loss": 3.3121,
"step": 48500
},
{
"epoch": 5.225487030459584,
"grad_norm": 0.6364693641662598,
"learning_rate": 0.00028690873828251265,
"loss": 3.3398,
"step": 48550
},
{
"epoch": 5.230868582499193,
"grad_norm": 0.6689774394035339,
"learning_rate": 0.00028658549725245125,
"loss": 3.305,
"step": 48600
},
{
"epoch": 5.236250134538801,
"grad_norm": 0.6482794880867004,
"learning_rate": 0.0002862622562223898,
"loss": 3.3205,
"step": 48650
},
{
"epoch": 5.241631686578409,
"grad_norm": 0.6583704352378845,
"learning_rate": 0.0002859390151923284,
"loss": 3.3276,
"step": 48700
},
{
"epoch": 5.247013238618018,
"grad_norm": 0.7061691880226135,
"learning_rate": 0.000285615774162267,
"loss": 3.3136,
"step": 48750
},
{
"epoch": 5.252394790657625,
"grad_norm": 0.660162091255188,
"learning_rate": 0.00028529253313220557,
"loss": 3.3249,
"step": 48800
},
{
"epoch": 5.257776342697234,
"grad_norm": 0.7076647877693176,
"learning_rate": 0.0002849692921021441,
"loss": 3.3087,
"step": 48850
},
{
"epoch": 5.2631578947368425,
"grad_norm": 0.6797081828117371,
"learning_rate": 0.00028464605107208276,
"loss": 3.335,
"step": 48900
},
{
"epoch": 5.26853944677645,
"grad_norm": 0.6778848171234131,
"learning_rate": 0.0002843228100420213,
"loss": 3.316,
"step": 48950
},
{
"epoch": 5.273920998816059,
"grad_norm": 0.6917376518249512,
"learning_rate": 0.0002839995690119599,
"loss": 3.3007,
"step": 49000
},
{
"epoch": 5.273920998816059,
"eval_accuracy": 0.38130709641346594,
"eval_loss": 3.407010793685913,
"eval_runtime": 180.5321,
"eval_samples_per_second": 99.766,
"eval_steps_per_second": 6.237,
"step": 49000
},
{
"epoch": 5.279302550855666,
"grad_norm": 0.6813348531723022,
"learning_rate": 0.0002836763279818985,
"loss": 3.3281,
"step": 49050
},
{
"epoch": 5.284684102895275,
"grad_norm": 0.6178494095802307,
"learning_rate": 0.0002833530869518371,
"loss": 3.3281,
"step": 49100
},
{
"epoch": 5.2900656549348835,
"grad_norm": 0.7000295519828796,
"learning_rate": 0.0002830298459217756,
"loss": 3.3227,
"step": 49150
},
{
"epoch": 5.295447206974491,
"grad_norm": 0.6339635252952576,
"learning_rate": 0.0002827066048917142,
"loss": 3.3207,
"step": 49200
},
{
"epoch": 5.3008287590141,
"grad_norm": 0.6519315838813782,
"learning_rate": 0.0002823833638616528,
"loss": 3.3175,
"step": 49250
},
{
"epoch": 5.306210311053708,
"grad_norm": 0.7280599474906921,
"learning_rate": 0.0002820601228315914,
"loss": 3.3503,
"step": 49300
},
{
"epoch": 5.311591863093316,
"grad_norm": 0.6790415048599243,
"learning_rate": 0.00028173688180153,
"loss": 3.3403,
"step": 49350
},
{
"epoch": 5.316973415132924,
"grad_norm": 0.6788349151611328,
"learning_rate": 0.00028141364077146854,
"loss": 3.3234,
"step": 49400
},
{
"epoch": 5.322354967172533,
"grad_norm": 0.702439546585083,
"learning_rate": 0.00028109039974140714,
"loss": 3.335,
"step": 49450
},
{
"epoch": 5.327736519212141,
"grad_norm": 0.6609926223754883,
"learning_rate": 0.00028076715871134573,
"loss": 3.3331,
"step": 49500
},
{
"epoch": 5.333118071251749,
"grad_norm": 0.6514111757278442,
"learning_rate": 0.00028044391768128433,
"loss": 3.3408,
"step": 49550
},
{
"epoch": 5.338499623291357,
"grad_norm": 0.6543302536010742,
"learning_rate": 0.0002801206766512229,
"loss": 3.3421,
"step": 49600
},
{
"epoch": 5.343881175330965,
"grad_norm": 0.6633315086364746,
"learning_rate": 0.0002797974356211615,
"loss": 3.3449,
"step": 49650
},
{
"epoch": 5.349262727370574,
"grad_norm": 0.6488233804702759,
"learning_rate": 0.00027947419459110006,
"loss": 3.3186,
"step": 49700
},
{
"epoch": 5.354644279410182,
"grad_norm": 0.7404241561889648,
"learning_rate": 0.00027915095356103865,
"loss": 3.3246,
"step": 49750
},
{
"epoch": 5.36002583144979,
"grad_norm": 0.6592351198196411,
"learning_rate": 0.00027882771253097725,
"loss": 3.3266,
"step": 49800
},
{
"epoch": 5.365407383489399,
"grad_norm": 0.6456713676452637,
"learning_rate": 0.00027850447150091584,
"loss": 3.3199,
"step": 49850
},
{
"epoch": 5.370788935529006,
"grad_norm": 0.6827576160430908,
"learning_rate": 0.00027818769529145565,
"loss": 3.3212,
"step": 49900
},
{
"epoch": 5.376170487568615,
"grad_norm": 0.6642194390296936,
"learning_rate": 0.0002778644542613942,
"loss": 3.3367,
"step": 49950
},
{
"epoch": 5.3815520396082235,
"grad_norm": 0.6693667769432068,
"learning_rate": 0.00027754121323133284,
"loss": 3.3337,
"step": 50000
},
{
"epoch": 5.3815520396082235,
"eval_accuracy": 0.38180950736672015,
"eval_loss": 3.403062343597412,
"eval_runtime": 180.255,
"eval_samples_per_second": 99.92,
"eval_steps_per_second": 6.247,
"step": 50000
},
{
"epoch": 5.386933591647831,
"grad_norm": 0.6618473529815674,
"learning_rate": 0.0002772179722012714,
"loss": 3.3359,
"step": 50050
},
{
"epoch": 5.39231514368744,
"grad_norm": 0.6960184574127197,
"learning_rate": 0.00027689473117120997,
"loss": 3.3385,
"step": 50100
},
{
"epoch": 5.397696695727047,
"grad_norm": 0.6597319841384888,
"learning_rate": 0.00027657149014114857,
"loss": 3.3239,
"step": 50150
},
{
"epoch": 5.403078247766656,
"grad_norm": 0.6477595567703247,
"learning_rate": 0.00027624824911108716,
"loss": 3.337,
"step": 50200
},
{
"epoch": 5.4084597998062645,
"grad_norm": 0.6981415152549744,
"learning_rate": 0.0002759250080810257,
"loss": 3.3085,
"step": 50250
},
{
"epoch": 5.413841351845872,
"grad_norm": 0.6679623126983643,
"learning_rate": 0.0002756017670509643,
"loss": 3.3104,
"step": 50300
},
{
"epoch": 5.419222903885481,
"grad_norm": 0.6863991618156433,
"learning_rate": 0.0002752785260209029,
"loss": 3.3437,
"step": 50350
},
{
"epoch": 5.424604455925088,
"grad_norm": 0.6082434058189392,
"learning_rate": 0.0002749552849908415,
"loss": 3.3244,
"step": 50400
},
{
"epoch": 5.429986007964697,
"grad_norm": 0.6896102428436279,
"learning_rate": 0.0002746320439607801,
"loss": 3.3639,
"step": 50450
},
{
"epoch": 5.435367560004305,
"grad_norm": 0.688849151134491,
"learning_rate": 0.0002743088029307186,
"loss": 3.3353,
"step": 50500
},
{
"epoch": 5.440749112043913,
"grad_norm": 0.7596466541290283,
"learning_rate": 0.00027398556190065727,
"loss": 3.3247,
"step": 50550
},
{
"epoch": 5.446130664083522,
"grad_norm": 0.6528745293617249,
"learning_rate": 0.0002736623208705958,
"loss": 3.3308,
"step": 50600
},
{
"epoch": 5.45151221612313,
"grad_norm": 0.6976600885391235,
"learning_rate": 0.0002733390798405344,
"loss": 3.3428,
"step": 50650
},
{
"epoch": 5.456893768162738,
"grad_norm": 0.6412066221237183,
"learning_rate": 0.000273015838810473,
"loss": 3.337,
"step": 50700
},
{
"epoch": 5.462275320202346,
"grad_norm": 0.6509308815002441,
"learning_rate": 0.0002726925977804116,
"loss": 3.3395,
"step": 50750
},
{
"epoch": 5.467656872241955,
"grad_norm": 0.661870002746582,
"learning_rate": 0.00027236935675035013,
"loss": 3.3315,
"step": 50800
},
{
"epoch": 5.473038424281563,
"grad_norm": 0.6559614539146423,
"learning_rate": 0.00027204611572028873,
"loss": 3.3327,
"step": 50850
},
{
"epoch": 5.478419976321171,
"grad_norm": 0.7063403129577637,
"learning_rate": 0.0002717228746902273,
"loss": 3.3128,
"step": 50900
},
{
"epoch": 5.483801528360779,
"grad_norm": 0.6607415676116943,
"learning_rate": 0.0002713996336601659,
"loss": 3.3446,
"step": 50950
},
{
"epoch": 5.489183080400387,
"grad_norm": 0.6556530594825745,
"learning_rate": 0.0002710763926301045,
"loss": 3.3386,
"step": 51000
},
{
"epoch": 5.489183080400387,
"eval_accuracy": 0.3822826906931522,
"eval_loss": 3.398730516433716,
"eval_runtime": 180.1394,
"eval_samples_per_second": 99.984,
"eval_steps_per_second": 6.251,
"step": 51000
},
{
"epoch": 5.494564632439996,
"grad_norm": 0.6305342316627502,
"learning_rate": 0.00027075315160004305,
"loss": 3.3359,
"step": 51050
},
{
"epoch": 5.499946184479604,
"grad_norm": 0.7005680799484253,
"learning_rate": 0.00027042991056998165,
"loss": 3.3388,
"step": 51100
},
{
"epoch": 5.505327736519212,
"grad_norm": 0.6419593095779419,
"learning_rate": 0.00027010666953992024,
"loss": 3.3246,
"step": 51150
},
{
"epoch": 5.510709288558821,
"grad_norm": 0.6594645380973816,
"learning_rate": 0.00026978342850985884,
"loss": 3.3368,
"step": 51200
},
{
"epoch": 5.516090840598428,
"grad_norm": 0.6918158531188965,
"learning_rate": 0.0002694601874797974,
"loss": 3.3387,
"step": 51250
},
{
"epoch": 5.521472392638037,
"grad_norm": 0.6681293249130249,
"learning_rate": 0.000269136946449736,
"loss": 3.3247,
"step": 51300
},
{
"epoch": 5.5268539446776455,
"grad_norm": 0.6619294881820679,
"learning_rate": 0.00026881370541967457,
"loss": 3.3303,
"step": 51350
},
{
"epoch": 5.532235496717253,
"grad_norm": 0.6340798735618591,
"learning_rate": 0.00026849046438961316,
"loss": 3.3283,
"step": 51400
},
{
"epoch": 5.537617048756862,
"grad_norm": 0.715398371219635,
"learning_rate": 0.00026816722335955176,
"loss": 3.3455,
"step": 51450
},
{
"epoch": 5.542998600796469,
"grad_norm": 0.6638710498809814,
"learning_rate": 0.00026784398232949035,
"loss": 3.3138,
"step": 51500
},
{
"epoch": 5.548380152836078,
"grad_norm": 0.6782832741737366,
"learning_rate": 0.00026752074129942895,
"loss": 3.3473,
"step": 51550
},
{
"epoch": 5.553761704875686,
"grad_norm": 0.6959323287010193,
"learning_rate": 0.0002671975002693675,
"loss": 3.3321,
"step": 51600
},
{
"epoch": 5.559143256915294,
"grad_norm": 0.6669702529907227,
"learning_rate": 0.0002668742592393061,
"loss": 3.3203,
"step": 51650
},
{
"epoch": 5.564524808954903,
"grad_norm": 0.6613565683364868,
"learning_rate": 0.0002665510182092447,
"loss": 3.3511,
"step": 51700
},
{
"epoch": 5.569906360994511,
"grad_norm": 0.6775053143501282,
"learning_rate": 0.00026622777717918327,
"loss": 3.3312,
"step": 51750
},
{
"epoch": 5.575287913034119,
"grad_norm": 0.6518484354019165,
"learning_rate": 0.0002659045361491218,
"loss": 3.3478,
"step": 51800
},
{
"epoch": 5.580669465073727,
"grad_norm": 0.718291699886322,
"learning_rate": 0.00026558129511906046,
"loss": 3.3449,
"step": 51850
},
{
"epoch": 5.586051017113336,
"grad_norm": 0.6551558971405029,
"learning_rate": 0.000265258054088999,
"loss": 3.3506,
"step": 51900
},
{
"epoch": 5.591432569152944,
"grad_norm": 0.622401773929596,
"learning_rate": 0.0002649412778795388,
"loss": 3.3253,
"step": 51950
},
{
"epoch": 5.596814121192552,
"grad_norm": 0.6426682472229004,
"learning_rate": 0.0002646180368494774,
"loss": 3.3294,
"step": 52000
},
{
"epoch": 5.596814121192552,
"eval_accuracy": 0.3830105563886214,
"eval_loss": 3.3924288749694824,
"eval_runtime": 180.6606,
"eval_samples_per_second": 99.695,
"eval_steps_per_second": 6.233,
"step": 52000
},
{
"epoch": 5.60219567323216,
"grad_norm": 0.6933045387268066,
"learning_rate": 0.000264294795819416,
"loss": 3.3382,
"step": 52050
},
{
"epoch": 5.607577225271768,
"grad_norm": 0.6460835933685303,
"learning_rate": 0.0002639715547893546,
"loss": 3.3259,
"step": 52100
},
{
"epoch": 5.612958777311377,
"grad_norm": 0.718722403049469,
"learning_rate": 0.00026364831375929313,
"loss": 3.342,
"step": 52150
},
{
"epoch": 5.618340329350985,
"grad_norm": 0.6578804850578308,
"learning_rate": 0.0002633250727292318,
"loss": 3.3247,
"step": 52200
},
{
"epoch": 5.623721881390593,
"grad_norm": 0.7196077704429626,
"learning_rate": 0.0002630018316991703,
"loss": 3.3305,
"step": 52250
},
{
"epoch": 5.629103433430201,
"grad_norm": 0.7222257256507874,
"learning_rate": 0.0002626785906691089,
"loss": 3.3448,
"step": 52300
},
{
"epoch": 5.634484985469809,
"grad_norm": 0.6821712255477905,
"learning_rate": 0.0002623553496390475,
"loss": 3.3404,
"step": 52350
},
{
"epoch": 5.639866537509418,
"grad_norm": 0.6659912467002869,
"learning_rate": 0.0002620321086089861,
"loss": 3.3263,
"step": 52400
},
{
"epoch": 5.645248089549026,
"grad_norm": 0.6843506097793579,
"learning_rate": 0.00026170886757892464,
"loss": 3.316,
"step": 52450
},
{
"epoch": 5.650629641588634,
"grad_norm": 0.6823989748954773,
"learning_rate": 0.00026138562654886324,
"loss": 3.3422,
"step": 52500
},
{
"epoch": 5.656011193628243,
"grad_norm": 0.6962025761604309,
"learning_rate": 0.00026106238551880183,
"loss": 3.3361,
"step": 52550
},
{
"epoch": 5.66139274566785,
"grad_norm": 0.6766282916069031,
"learning_rate": 0.00026073914448874043,
"loss": 3.328,
"step": 52600
},
{
"epoch": 5.666774297707459,
"grad_norm": 0.6884021759033203,
"learning_rate": 0.000260415903458679,
"loss": 3.3356,
"step": 52650
},
{
"epoch": 5.672155849747067,
"grad_norm": 0.6552942395210266,
"learning_rate": 0.00026009266242861756,
"loss": 3.3326,
"step": 52700
},
{
"epoch": 5.677537401786675,
"grad_norm": 0.649552583694458,
"learning_rate": 0.00025976942139855616,
"loss": 3.3303,
"step": 52750
},
{
"epoch": 5.682918953826284,
"grad_norm": 0.7173223495483398,
"learning_rate": 0.00025945264518909596,
"loss": 3.3183,
"step": 52800
},
{
"epoch": 5.688300505865891,
"grad_norm": 0.7656635046005249,
"learning_rate": 0.00025912940415903456,
"loss": 3.3251,
"step": 52850
},
{
"epoch": 5.6936820579055,
"grad_norm": 0.740873396396637,
"learning_rate": 0.00025880616312897315,
"loss": 3.3179,
"step": 52900
},
{
"epoch": 5.699063609945108,
"grad_norm": 0.7117483019828796,
"learning_rate": 0.00025848292209891175,
"loss": 3.3273,
"step": 52950
},
{
"epoch": 5.704445161984716,
"grad_norm": 0.68827223777771,
"learning_rate": 0.00025815968106885034,
"loss": 3.3153,
"step": 53000
},
{
"epoch": 5.704445161984716,
"eval_accuracy": 0.38329913845865665,
"eval_loss": 3.3908591270446777,
"eval_runtime": 180.1829,
"eval_samples_per_second": 99.96,
"eval_steps_per_second": 6.249,
"step": 53000
},
{
"epoch": 5.709826714024325,
"grad_norm": 0.6453783512115479,
"learning_rate": 0.0002578364400387889,
"loss": 3.3325,
"step": 53050
},
{
"epoch": 5.715208266063933,
"grad_norm": 0.696941077709198,
"learning_rate": 0.0002575131990087275,
"loss": 3.3351,
"step": 53100
},
{
"epoch": 5.720589818103541,
"grad_norm": 0.6282665133476257,
"learning_rate": 0.0002571899579786661,
"loss": 3.3288,
"step": 53150
},
{
"epoch": 5.725971370143149,
"grad_norm": 0.6461380124092102,
"learning_rate": 0.00025686671694860467,
"loss": 3.3403,
"step": 53200
},
{
"epoch": 5.731352922182758,
"grad_norm": 0.6844844818115234,
"learning_rate": 0.0002565434759185432,
"loss": 3.3306,
"step": 53250
},
{
"epoch": 5.736734474222366,
"grad_norm": 0.6883606314659119,
"learning_rate": 0.00025622023488848186,
"loss": 3.3242,
"step": 53300
},
{
"epoch": 5.742116026261974,
"grad_norm": 0.7068212628364563,
"learning_rate": 0.0002558969938584204,
"loss": 3.3241,
"step": 53350
},
{
"epoch": 5.747497578301582,
"grad_norm": 0.7148516774177551,
"learning_rate": 0.000255573752828359,
"loss": 3.326,
"step": 53400
},
{
"epoch": 5.75287913034119,
"grad_norm": 0.6874505877494812,
"learning_rate": 0.0002552505117982976,
"loss": 3.3322,
"step": 53450
},
{
"epoch": 5.758260682380799,
"grad_norm": 0.655957818031311,
"learning_rate": 0.0002549272707682362,
"loss": 3.3385,
"step": 53500
},
{
"epoch": 5.763642234420407,
"grad_norm": 0.6629857420921326,
"learning_rate": 0.0002546040297381747,
"loss": 3.3132,
"step": 53550
},
{
"epoch": 5.769023786460015,
"grad_norm": 0.671413779258728,
"learning_rate": 0.0002542807887081133,
"loss": 3.3402,
"step": 53600
},
{
"epoch": 5.774405338499624,
"grad_norm": 0.6932768225669861,
"learning_rate": 0.0002539575476780519,
"loss": 3.3157,
"step": 53650
},
{
"epoch": 5.779786890539231,
"grad_norm": 0.6655086874961853,
"learning_rate": 0.0002536343066479905,
"loss": 3.3101,
"step": 53700
},
{
"epoch": 5.78516844257884,
"grad_norm": 0.6800814270973206,
"learning_rate": 0.0002533110656179291,
"loss": 3.3382,
"step": 53750
},
{
"epoch": 5.790549994618448,
"grad_norm": 0.665797233581543,
"learning_rate": 0.00025298782458786764,
"loss": 3.3159,
"step": 53800
},
{
"epoch": 5.795931546658056,
"grad_norm": 0.701047956943512,
"learning_rate": 0.00025266458355780624,
"loss": 3.3319,
"step": 53850
},
{
"epoch": 5.801313098697665,
"grad_norm": 0.6783666014671326,
"learning_rate": 0.00025234134252774483,
"loss": 3.3272,
"step": 53900
},
{
"epoch": 5.806694650737272,
"grad_norm": 0.6685265302658081,
"learning_rate": 0.0002520181014976834,
"loss": 3.3389,
"step": 53950
},
{
"epoch": 5.812076202776881,
"grad_norm": 0.6971936821937561,
"learning_rate": 0.000251694860467622,
"loss": 3.3324,
"step": 54000
},
{
"epoch": 5.812076202776881,
"eval_accuracy": 0.3839824564701219,
"eval_loss": 3.3846652507781982,
"eval_runtime": 180.3207,
"eval_samples_per_second": 99.883,
"eval_steps_per_second": 6.244,
"step": 54000
},
{
"epoch": 5.817457754816489,
"grad_norm": 0.6983776092529297,
"learning_rate": 0.0002513716194375606,
"loss": 3.3241,
"step": 54050
},
{
"epoch": 5.822839306856097,
"grad_norm": 0.6729536056518555,
"learning_rate": 0.00025104837840749916,
"loss": 3.3321,
"step": 54100
},
{
"epoch": 5.828220858895706,
"grad_norm": 0.6928292512893677,
"learning_rate": 0.00025072513737743775,
"loss": 3.3442,
"step": 54150
},
{
"epoch": 5.833602410935313,
"grad_norm": 0.6832127571105957,
"learning_rate": 0.00025040189634737634,
"loss": 3.33,
"step": 54200
},
{
"epoch": 5.838983962974922,
"grad_norm": 0.7145037651062012,
"learning_rate": 0.00025007865531731494,
"loss": 3.3557,
"step": 54250
},
{
"epoch": 5.84436551501453,
"grad_norm": 0.7388347387313843,
"learning_rate": 0.00024975541428725353,
"loss": 3.3417,
"step": 54300
},
{
"epoch": 5.849747067054138,
"grad_norm": 0.6939401030540466,
"learning_rate": 0.0002494321732571921,
"loss": 3.3285,
"step": 54350
},
{
"epoch": 5.855128619093747,
"grad_norm": 0.7111252546310425,
"learning_rate": 0.00024910893222713067,
"loss": 3.3386,
"step": 54400
},
{
"epoch": 5.860510171133355,
"grad_norm": 0.7177287936210632,
"learning_rate": 0.00024878569119706926,
"loss": 3.3222,
"step": 54450
},
{
"epoch": 5.865891723172963,
"grad_norm": 0.6975011229515076,
"learning_rate": 0.00024846245016700786,
"loss": 3.3301,
"step": 54500
},
{
"epoch": 5.871273275212571,
"grad_norm": 0.8165584802627563,
"learning_rate": 0.0002481392091369464,
"loss": 3.3396,
"step": 54550
},
{
"epoch": 5.87665482725218,
"grad_norm": 0.7520706653594971,
"learning_rate": 0.00024781596810688505,
"loss": 3.3314,
"step": 54600
},
{
"epoch": 5.882036379291788,
"grad_norm": 0.6771239042282104,
"learning_rate": 0.0002474927270768236,
"loss": 3.3574,
"step": 54650
},
{
"epoch": 5.887417931331396,
"grad_norm": 0.6594865918159485,
"learning_rate": 0.0002471694860467622,
"loss": 3.3361,
"step": 54700
},
{
"epoch": 5.892799483371004,
"grad_norm": 0.6948037147521973,
"learning_rate": 0.0002468462450167008,
"loss": 3.3186,
"step": 54750
},
{
"epoch": 5.898181035410612,
"grad_norm": 0.680067241191864,
"learning_rate": 0.0002465230039866393,
"loss": 3.3348,
"step": 54800
},
{
"epoch": 5.903562587450221,
"grad_norm": 0.6603071689605713,
"learning_rate": 0.0002461997629565779,
"loss": 3.3266,
"step": 54850
},
{
"epoch": 5.9089441394898286,
"grad_norm": 0.6521521210670471,
"learning_rate": 0.0002458765219265165,
"loss": 3.3288,
"step": 54900
},
{
"epoch": 5.914325691529437,
"grad_norm": 0.8012475371360779,
"learning_rate": 0.0002455532808964551,
"loss": 3.3194,
"step": 54950
},
{
"epoch": 5.919707243569046,
"grad_norm": 0.7362468242645264,
"learning_rate": 0.0002452300398663937,
"loss": 3.3298,
"step": 55000
},
{
"epoch": 5.919707243569046,
"eval_accuracy": 0.3843108054970973,
"eval_loss": 3.38219952583313,
"eval_runtime": 180.5029,
"eval_samples_per_second": 99.782,
"eval_steps_per_second": 6.238,
"step": 55000
},
{
"epoch": 5.925088795608653,
"grad_norm": 0.7196722030639648,
"learning_rate": 0.0002449067988363323,
"loss": 3.3229,
"step": 55050
},
{
"epoch": 5.930470347648262,
"grad_norm": 0.7012231945991516,
"learning_rate": 0.00024458355780627083,
"loss": 3.3314,
"step": 55100
},
{
"epoch": 5.93585189968787,
"grad_norm": 0.7560854554176331,
"learning_rate": 0.0002442603167762094,
"loss": 3.3366,
"step": 55150
},
{
"epoch": 5.941233451727478,
"grad_norm": 0.6536340713500977,
"learning_rate": 0.00024393707574614802,
"loss": 3.3336,
"step": 55200
},
{
"epoch": 5.946615003767087,
"grad_norm": 0.7389876246452332,
"learning_rate": 0.0002436138347160866,
"loss": 3.3283,
"step": 55250
},
{
"epoch": 5.951996555806694,
"grad_norm": 0.6520614624023438,
"learning_rate": 0.0002432905936860252,
"loss": 3.3385,
"step": 55300
},
{
"epoch": 5.957378107846303,
"grad_norm": 0.6752536296844482,
"learning_rate": 0.00024296735265596378,
"loss": 3.3286,
"step": 55350
},
{
"epoch": 5.962759659885911,
"grad_norm": 0.6813256740570068,
"learning_rate": 0.00024264411162590235,
"loss": 3.3388,
"step": 55400
},
{
"epoch": 5.968141211925519,
"grad_norm": 0.6881508827209473,
"learning_rate": 0.00024232087059584097,
"loss": 3.325,
"step": 55450
},
{
"epoch": 5.973522763965128,
"grad_norm": 0.7486839890480042,
"learning_rate": 0.00024199762956577953,
"loss": 3.3334,
"step": 55500
},
{
"epoch": 5.978904316004736,
"grad_norm": 0.6768556237220764,
"learning_rate": 0.0002416743885357181,
"loss": 3.331,
"step": 55550
},
{
"epoch": 5.984285868044344,
"grad_norm": 0.7191193699836731,
"learning_rate": 0.0002413511475056567,
"loss": 3.3214,
"step": 55600
},
{
"epoch": 5.989667420083952,
"grad_norm": 0.7417619228363037,
"learning_rate": 0.0002410279064755953,
"loss": 3.3383,
"step": 55650
},
{
"epoch": 5.995048972123561,
"grad_norm": 0.7486613392829895,
"learning_rate": 0.00024070466544553386,
"loss": 3.3266,
"step": 55700
},
{
"epoch": 6.000430524163169,
"grad_norm": 0.6875068545341492,
"learning_rate": 0.00024038142441547245,
"loss": 3.3292,
"step": 55750
},
{
"epoch": 6.005812076202777,
"grad_norm": 0.7089745998382568,
"learning_rate": 0.00024005818338541102,
"loss": 3.2554,
"step": 55800
},
{
"epoch": 6.011193628242385,
"grad_norm": 0.6928079724311829,
"learning_rate": 0.0002397349423553496,
"loss": 3.2292,
"step": 55850
},
{
"epoch": 6.016575180281993,
"grad_norm": 0.6740532517433167,
"learning_rate": 0.0002394117013252882,
"loss": 3.2425,
"step": 55900
},
{
"epoch": 6.021956732321602,
"grad_norm": 0.6799925565719604,
"learning_rate": 0.00023908846029522678,
"loss": 3.2304,
"step": 55950
},
{
"epoch": 6.0273382843612096,
"grad_norm": 0.7598007917404175,
"learning_rate": 0.00023876521926516537,
"loss": 3.2505,
"step": 56000
},
{
"epoch": 6.0273382843612096,
"eval_accuracy": 0.3845337612228926,
"eval_loss": 3.3847992420196533,
"eval_runtime": 180.0957,
"eval_samples_per_second": 100.008,
"eval_steps_per_second": 6.252,
"step": 56000
},
{
"epoch": 6.032719836400818,
"grad_norm": 0.7275548577308655,
"learning_rate": 0.00023844197823510397,
"loss": 3.248,
"step": 56050
},
{
"epoch": 6.038101388440427,
"grad_norm": 0.7083311676979065,
"learning_rate": 0.00023811873720504253,
"loss": 3.2203,
"step": 56100
},
{
"epoch": 6.043482940480034,
"grad_norm": 0.7197110056877136,
"learning_rate": 0.00023779549617498113,
"loss": 3.2358,
"step": 56150
},
{
"epoch": 6.048864492519643,
"grad_norm": 0.6651777029037476,
"learning_rate": 0.0002374722551449197,
"loss": 3.2521,
"step": 56200
},
{
"epoch": 6.0542460445592505,
"grad_norm": 0.7160053253173828,
"learning_rate": 0.0002371490141148583,
"loss": 3.2404,
"step": 56250
},
{
"epoch": 6.059627596598859,
"grad_norm": 0.7208572626113892,
"learning_rate": 0.00023682577308479689,
"loss": 3.244,
"step": 56300
},
{
"epoch": 6.065009148638468,
"grad_norm": 0.6583945155143738,
"learning_rate": 0.00023650253205473545,
"loss": 3.2551,
"step": 56350
},
{
"epoch": 6.070390700678075,
"grad_norm": 0.695590078830719,
"learning_rate": 0.00023617929102467402,
"loss": 3.25,
"step": 56400
},
{
"epoch": 6.075772252717684,
"grad_norm": 0.6976557374000549,
"learning_rate": 0.00023586251481521386,
"loss": 3.2608,
"step": 56450
},
{
"epoch": 6.081153804757292,
"grad_norm": 0.7817869782447815,
"learning_rate": 0.00023553927378515242,
"loss": 3.2698,
"step": 56500
},
{
"epoch": 6.0865353567969,
"grad_norm": 0.6705769300460815,
"learning_rate": 0.00023521603275509104,
"loss": 3.2602,
"step": 56550
},
{
"epoch": 6.091916908836509,
"grad_norm": 0.7016733884811401,
"learning_rate": 0.0002348927917250296,
"loss": 3.2514,
"step": 56600
},
{
"epoch": 6.097298460876116,
"grad_norm": 0.7753743529319763,
"learning_rate": 0.00023456955069496818,
"loss": 3.2367,
"step": 56650
},
{
"epoch": 6.102680012915725,
"grad_norm": 0.6739595532417297,
"learning_rate": 0.00023424630966490677,
"loss": 3.2541,
"step": 56700
},
{
"epoch": 6.108061564955333,
"grad_norm": 0.732443630695343,
"learning_rate": 0.00023392306863484537,
"loss": 3.2537,
"step": 56750
},
{
"epoch": 6.113443116994941,
"grad_norm": 0.7694481611251831,
"learning_rate": 0.00023359982760478396,
"loss": 3.2594,
"step": 56800
},
{
"epoch": 6.11882466903455,
"grad_norm": 0.7056264877319336,
"learning_rate": 0.00023327658657472253,
"loss": 3.2455,
"step": 56850
},
{
"epoch": 6.124206221074158,
"grad_norm": 0.6392900943756104,
"learning_rate": 0.0002329533455446611,
"loss": 3.2531,
"step": 56900
},
{
"epoch": 6.129587773113766,
"grad_norm": 0.6789774298667908,
"learning_rate": 0.00023263010451459972,
"loss": 3.2526,
"step": 56950
},
{
"epoch": 6.134969325153374,
"grad_norm": 0.7180214524269104,
"learning_rate": 0.0002323068634845383,
"loss": 3.2374,
"step": 57000
},
{
"epoch": 6.134969325153374,
"eval_accuracy": 0.3843350350910799,
"eval_loss": 3.387458562850952,
"eval_runtime": 179.945,
"eval_samples_per_second": 100.092,
"eval_steps_per_second": 6.257,
"step": 57000
},
{
"epoch": 6.140350877192983,
"grad_norm": 0.6634199619293213,
"learning_rate": 0.00023198362245447686,
"loss": 3.2663,
"step": 57050
},
{
"epoch": 6.1457324292325906,
"grad_norm": 0.7024934887886047,
"learning_rate": 0.00023166038142441548,
"loss": 3.2761,
"step": 57100
},
{
"epoch": 6.151113981272199,
"grad_norm": 0.7180144190788269,
"learning_rate": 0.00023133714039435404,
"loss": 3.2654,
"step": 57150
},
{
"epoch": 6.156495533311807,
"grad_norm": 0.6728076934814453,
"learning_rate": 0.0002310138993642926,
"loss": 3.2579,
"step": 57200
},
{
"epoch": 6.161877085351415,
"grad_norm": 0.7283917665481567,
"learning_rate": 0.0002306906583342312,
"loss": 3.2565,
"step": 57250
},
{
"epoch": 6.167258637391024,
"grad_norm": 0.7351763248443604,
"learning_rate": 0.00023036741730416977,
"loss": 3.2523,
"step": 57300
},
{
"epoch": 6.1726401894306315,
"grad_norm": 0.7366397380828857,
"learning_rate": 0.00023004417627410837,
"loss": 3.2665,
"step": 57350
},
{
"epoch": 6.17802174147024,
"grad_norm": 0.6817381381988525,
"learning_rate": 0.00022972093524404696,
"loss": 3.2791,
"step": 57400
},
{
"epoch": 6.183403293509849,
"grad_norm": 0.6933722496032715,
"learning_rate": 0.00022939769421398553,
"loss": 3.2675,
"step": 57450
},
{
"epoch": 6.188784845549456,
"grad_norm": 0.6675575375556946,
"learning_rate": 0.0002290744531839241,
"loss": 3.252,
"step": 57500
},
{
"epoch": 6.194166397589065,
"grad_norm": 0.7353582382202148,
"learning_rate": 0.00022875121215386272,
"loss": 3.2649,
"step": 57550
},
{
"epoch": 6.1995479496286725,
"grad_norm": 0.7077396512031555,
"learning_rate": 0.0002284279711238013,
"loss": 3.2762,
"step": 57600
},
{
"epoch": 6.204929501668281,
"grad_norm": 0.761920154094696,
"learning_rate": 0.00022810473009373986,
"loss": 3.2494,
"step": 57650
},
{
"epoch": 6.21031105370789,
"grad_norm": 0.6998116970062256,
"learning_rate": 0.00022778148906367848,
"loss": 3.2744,
"step": 57700
},
{
"epoch": 6.215692605747497,
"grad_norm": 0.6889002919197083,
"learning_rate": 0.00022745824803361705,
"loss": 3.2544,
"step": 57750
},
{
"epoch": 6.221074157787106,
"grad_norm": 0.7440342307090759,
"learning_rate": 0.00022713500700355564,
"loss": 3.2774,
"step": 57800
},
{
"epoch": 6.226455709826714,
"grad_norm": 0.6621478199958801,
"learning_rate": 0.0002268117659734942,
"loss": 3.2952,
"step": 57850
},
{
"epoch": 6.231837261866322,
"grad_norm": 0.7194118499755859,
"learning_rate": 0.0002264885249434328,
"loss": 3.2817,
"step": 57900
},
{
"epoch": 6.237218813905931,
"grad_norm": 0.7072327136993408,
"learning_rate": 0.0002261652839133714,
"loss": 3.2649,
"step": 57950
},
{
"epoch": 6.242600365945538,
"grad_norm": 0.7320728898048401,
"learning_rate": 0.00022584204288330996,
"loss": 3.2673,
"step": 58000
},
{
"epoch": 6.242600365945538,
"eval_accuracy": 0.3850747439513209,
"eval_loss": 3.380539894104004,
"eval_runtime": 179.9128,
"eval_samples_per_second": 100.11,
"eval_steps_per_second": 6.259,
"step": 58000
},
{
"epoch": 6.247981917985147,
"grad_norm": 0.7100010514259338,
"learning_rate": 0.00022551880185324853,
"loss": 3.2534,
"step": 58050
},
{
"epoch": 6.253363470024755,
"grad_norm": 0.6797668933868408,
"learning_rate": 0.00022519556082318715,
"loss": 3.2727,
"step": 58100
},
{
"epoch": 6.258745022064363,
"grad_norm": 0.7310889959335327,
"learning_rate": 0.00022487231979312572,
"loss": 3.2822,
"step": 58150
},
{
"epoch": 6.264126574103972,
"grad_norm": 0.7226874232292175,
"learning_rate": 0.0002245490787630643,
"loss": 3.2832,
"step": 58200
},
{
"epoch": 6.26950812614358,
"grad_norm": 0.7260108590126038,
"learning_rate": 0.0002242258377330029,
"loss": 3.2615,
"step": 58250
},
{
"epoch": 6.274889678183188,
"grad_norm": 0.6836520433425903,
"learning_rate": 0.00022390259670294148,
"loss": 3.2678,
"step": 58300
},
{
"epoch": 6.280271230222796,
"grad_norm": 0.7636072039604187,
"learning_rate": 0.00022357935567288005,
"loss": 3.27,
"step": 58350
},
{
"epoch": 6.285652782262405,
"grad_norm": 0.7172989845275879,
"learning_rate": 0.00022325611464281864,
"loss": 3.2646,
"step": 58400
},
{
"epoch": 6.2910343343020125,
"grad_norm": 0.7145113945007324,
"learning_rate": 0.00022293287361275723,
"loss": 3.2717,
"step": 58450
},
{
"epoch": 6.296415886341621,
"grad_norm": 0.7093191146850586,
"learning_rate": 0.0002226096325826958,
"loss": 3.2584,
"step": 58500
},
{
"epoch": 6.301797438381229,
"grad_norm": 0.6939222812652588,
"learning_rate": 0.0002222863915526344,
"loss": 3.2631,
"step": 58550
},
{
"epoch": 6.307178990420837,
"grad_norm": 0.7617230415344238,
"learning_rate": 0.00022196315052257296,
"loss": 3.2769,
"step": 58600
},
{
"epoch": 6.312560542460446,
"grad_norm": 0.694742739200592,
"learning_rate": 0.00022163990949251156,
"loss": 3.2843,
"step": 58650
},
{
"epoch": 6.3179420945000535,
"grad_norm": 0.6639745235443115,
"learning_rate": 0.00022131666846245015,
"loss": 3.2728,
"step": 58700
},
{
"epoch": 6.323323646539662,
"grad_norm": 0.7379493117332458,
"learning_rate": 0.00022099342743238872,
"loss": 3.2663,
"step": 58750
},
{
"epoch": 6.328705198579271,
"grad_norm": 0.7672418355941772,
"learning_rate": 0.00022067018640232734,
"loss": 3.2839,
"step": 58800
},
{
"epoch": 6.334086750618878,
"grad_norm": 0.7370325922966003,
"learning_rate": 0.0002203469453722659,
"loss": 3.2852,
"step": 58850
},
{
"epoch": 6.339468302658487,
"grad_norm": 0.7163979411125183,
"learning_rate": 0.00022002370434220448,
"loss": 3.2848,
"step": 58900
},
{
"epoch": 6.344849854698095,
"grad_norm": 0.7386805415153503,
"learning_rate": 0.00021970046331214307,
"loss": 3.2539,
"step": 58950
},
{
"epoch": 6.350231406737703,
"grad_norm": 0.7379843592643738,
"learning_rate": 0.00021937722228208167,
"loss": 3.2727,
"step": 59000
},
{
"epoch": 6.350231406737703,
"eval_accuracy": 0.3851716623272514,
"eval_loss": 3.3763844966888428,
"eval_runtime": 180.0816,
"eval_samples_per_second": 100.016,
"eval_steps_per_second": 6.253,
"step": 59000
},
{
"epoch": 6.355612958777312,
"grad_norm": 0.6990553736686707,
"learning_rate": 0.00021905398125202024,
"loss": 3.2635,
"step": 59050
},
{
"epoch": 6.360994510816919,
"grad_norm": 0.7014365792274475,
"learning_rate": 0.00021873074022195883,
"loss": 3.2834,
"step": 59100
},
{
"epoch": 6.366376062856528,
"grad_norm": 0.6956151127815247,
"learning_rate": 0.0002184074991918974,
"loss": 3.2644,
"step": 59150
},
{
"epoch": 6.371757614896136,
"grad_norm": 0.7179141640663147,
"learning_rate": 0.00021808425816183597,
"loss": 3.2749,
"step": 59200
},
{
"epoch": 6.377139166935744,
"grad_norm": 0.7074273824691772,
"learning_rate": 0.0002177610171317746,
"loss": 3.2607,
"step": 59250
},
{
"epoch": 6.382520718975353,
"grad_norm": 0.6928147673606873,
"learning_rate": 0.00021743777610171315,
"loss": 3.2804,
"step": 59300
},
{
"epoch": 6.387902271014961,
"grad_norm": 0.7080892324447632,
"learning_rate": 0.00021711453507165172,
"loss": 3.2897,
"step": 59350
},
{
"epoch": 6.393283823054569,
"grad_norm": 0.7480721473693848,
"learning_rate": 0.00021679129404159034,
"loss": 3.275,
"step": 59400
},
{
"epoch": 6.398665375094177,
"grad_norm": 0.7683929800987244,
"learning_rate": 0.0002164680530115289,
"loss": 3.2511,
"step": 59450
},
{
"epoch": 6.404046927133785,
"grad_norm": 0.7302149534225464,
"learning_rate": 0.00021614481198146748,
"loss": 3.2587,
"step": 59500
},
{
"epoch": 6.4094284791733935,
"grad_norm": 0.7285591959953308,
"learning_rate": 0.00021582157095140607,
"loss": 3.2727,
"step": 59550
},
{
"epoch": 6.414810031213002,
"grad_norm": 0.7607294917106628,
"learning_rate": 0.00021549832992134467,
"loss": 3.2927,
"step": 59600
},
{
"epoch": 6.42019158325261,
"grad_norm": 0.7400927543640137,
"learning_rate": 0.00021517508889128324,
"loss": 3.2775,
"step": 59650
},
{
"epoch": 6.425573135292218,
"grad_norm": 0.7432016730308533,
"learning_rate": 0.00021485184786122183,
"loss": 3.2883,
"step": 59700
},
{
"epoch": 6.430954687331827,
"grad_norm": 0.754863977432251,
"learning_rate": 0.0002145286068311604,
"loss": 3.2949,
"step": 59750
},
{
"epoch": 6.4363362393714345,
"grad_norm": 0.7521831393241882,
"learning_rate": 0.00021420536580109902,
"loss": 3.2847,
"step": 59800
},
{
"epoch": 6.441717791411043,
"grad_norm": 0.6796225309371948,
"learning_rate": 0.0002138821247710376,
"loss": 3.2842,
"step": 59850
},
{
"epoch": 6.447099343450651,
"grad_norm": 0.7475155591964722,
"learning_rate": 0.00021355888374097615,
"loss": 3.2863,
"step": 59900
},
{
"epoch": 6.452480895490259,
"grad_norm": 0.7627286314964294,
"learning_rate": 0.00021323564271091478,
"loss": 3.2928,
"step": 59950
},
{
"epoch": 6.457862447529868,
"grad_norm": 0.7810362577438354,
"learning_rate": 0.00021291240168085334,
"loss": 3.2923,
"step": 60000
},
{
"epoch": 6.457862447529868,
"eval_accuracy": 0.38576631958226004,
"eval_loss": 3.371659278869629,
"eval_runtime": 179.7476,
"eval_samples_per_second": 100.202,
"eval_steps_per_second": 6.264,
"step": 60000
},
{
"epoch": 6.4632439995694755,
"grad_norm": 0.731270968914032,
"learning_rate": 0.0002125891606507919,
"loss": 3.2814,
"step": 60050
},
{
"epoch": 6.468625551609084,
"grad_norm": 0.7113689184188843,
"learning_rate": 0.0002122659196207305,
"loss": 3.2691,
"step": 60100
},
{
"epoch": 6.474007103648693,
"grad_norm": 0.7589284777641296,
"learning_rate": 0.0002119426785906691,
"loss": 3.2758,
"step": 60150
},
{
"epoch": 6.4793886556883,
"grad_norm": 0.7309033870697021,
"learning_rate": 0.00021161943756060767,
"loss": 3.2725,
"step": 60200
},
{
"epoch": 6.484770207727909,
"grad_norm": 0.7444686889648438,
"learning_rate": 0.00021130266135114747,
"loss": 3.2851,
"step": 60250
},
{
"epoch": 6.490151759767517,
"grad_norm": 0.7868755459785461,
"learning_rate": 0.00021097942032108607,
"loss": 3.2974,
"step": 60300
},
{
"epoch": 6.495533311807125,
"grad_norm": 0.720824122428894,
"learning_rate": 0.00021065617929102466,
"loss": 3.2745,
"step": 60350
},
{
"epoch": 6.500914863846734,
"grad_norm": 0.717765748500824,
"learning_rate": 0.00021033293826096323,
"loss": 3.2889,
"step": 60400
},
{
"epoch": 6.506296415886341,
"grad_norm": 0.7933386564254761,
"learning_rate": 0.0002100096972309018,
"loss": 3.2683,
"step": 60450
},
{
"epoch": 6.51167796792595,
"grad_norm": 0.7101936936378479,
"learning_rate": 0.00020968645620084042,
"loss": 3.2728,
"step": 60500
},
{
"epoch": 6.517059519965558,
"grad_norm": 0.6985874176025391,
"learning_rate": 0.000209363215170779,
"loss": 3.2901,
"step": 60550
},
{
"epoch": 6.522441072005166,
"grad_norm": 0.7263524532318115,
"learning_rate": 0.00020903997414071758,
"loss": 3.2667,
"step": 60600
},
{
"epoch": 6.5278226240447745,
"grad_norm": 0.7178045511245728,
"learning_rate": 0.00020871673311065618,
"loss": 3.282,
"step": 60650
},
{
"epoch": 6.533204176084383,
"grad_norm": 0.7246670126914978,
"learning_rate": 0.00020839349208059475,
"loss": 3.2883,
"step": 60700
},
{
"epoch": 6.538585728123991,
"grad_norm": 0.7198054194450378,
"learning_rate": 0.00020807025105053334,
"loss": 3.2974,
"step": 60750
},
{
"epoch": 6.543967280163599,
"grad_norm": 0.7221017479896545,
"learning_rate": 0.0002077470100204719,
"loss": 3.2856,
"step": 60800
},
{
"epoch": 6.549348832203208,
"grad_norm": 0.7302899360656738,
"learning_rate": 0.00020742376899041048,
"loss": 3.2856,
"step": 60850
},
{
"epoch": 6.5547303842428155,
"grad_norm": 0.7253843545913696,
"learning_rate": 0.0002071005279603491,
"loss": 3.2752,
"step": 60900
},
{
"epoch": 6.560111936282424,
"grad_norm": 0.7374371886253357,
"learning_rate": 0.00020677728693028766,
"loss": 3.2728,
"step": 60950
},
{
"epoch": 6.565493488322032,
"grad_norm": 0.70008385181427,
"learning_rate": 0.00020645404590022623,
"loss": 3.2689,
"step": 61000
},
{
"epoch": 6.565493488322032,
"eval_accuracy": 0.3859744985153126,
"eval_loss": 3.3697760105133057,
"eval_runtime": 180.5619,
"eval_samples_per_second": 99.75,
"eval_steps_per_second": 6.236,
"step": 61000
},
{
"epoch": 6.57087504036164,
"grad_norm": 0.7110406160354614,
"learning_rate": 0.00020613080487016485,
"loss": 3.289,
"step": 61050
},
{
"epoch": 6.576256592401249,
"grad_norm": 0.8671207427978516,
"learning_rate": 0.00020580756384010342,
"loss": 3.2676,
"step": 61100
},
{
"epoch": 6.5816381444408565,
"grad_norm": 0.797752320766449,
"learning_rate": 0.000205484322810042,
"loss": 3.2827,
"step": 61150
},
{
"epoch": 6.587019696480465,
"grad_norm": 0.7081204652786255,
"learning_rate": 0.00020516108177998058,
"loss": 3.2917,
"step": 61200
},
{
"epoch": 6.592401248520073,
"grad_norm": 0.7679519057273865,
"learning_rate": 0.00020483784074991918,
"loss": 3.2761,
"step": 61250
},
{
"epoch": 6.597782800559681,
"grad_norm": 0.7073348164558411,
"learning_rate": 0.00020451459971985775,
"loss": 3.2917,
"step": 61300
},
{
"epoch": 6.60316435259929,
"grad_norm": 0.714297354221344,
"learning_rate": 0.00020419135868979634,
"loss": 3.2842,
"step": 61350
},
{
"epoch": 6.608545904638898,
"grad_norm": 0.7196522355079651,
"learning_rate": 0.0002038681176597349,
"loss": 3.2688,
"step": 61400
},
{
"epoch": 6.613927456678506,
"grad_norm": 0.7500666975975037,
"learning_rate": 0.0002035448766296735,
"loss": 3.2923,
"step": 61450
},
{
"epoch": 6.619309008718115,
"grad_norm": 0.7291411757469177,
"learning_rate": 0.0002032216355996121,
"loss": 3.2894,
"step": 61500
},
{
"epoch": 6.624690560757722,
"grad_norm": 0.7208287119865417,
"learning_rate": 0.00020289839456955067,
"loss": 3.3013,
"step": 61550
},
{
"epoch": 6.630072112797331,
"grad_norm": 0.7192732691764832,
"learning_rate": 0.0002025751535394893,
"loss": 3.2926,
"step": 61600
},
{
"epoch": 6.635453664836939,
"grad_norm": 0.769903302192688,
"learning_rate": 0.00020225191250942785,
"loss": 3.2713,
"step": 61650
},
{
"epoch": 6.640835216876547,
"grad_norm": 0.758073091506958,
"learning_rate": 0.00020192867147936642,
"loss": 3.2694,
"step": 61700
},
{
"epoch": 6.6462167689161555,
"grad_norm": 0.7105669379234314,
"learning_rate": 0.00020160543044930502,
"loss": 3.2889,
"step": 61750
},
{
"epoch": 6.651598320955763,
"grad_norm": 0.7682111263275146,
"learning_rate": 0.0002012821894192436,
"loss": 3.2897,
"step": 61800
},
{
"epoch": 6.656979872995372,
"grad_norm": 0.7516341209411621,
"learning_rate": 0.00020095894838918218,
"loss": 3.2793,
"step": 61850
},
{
"epoch": 6.66236142503498,
"grad_norm": 0.7740298509597778,
"learning_rate": 0.00020063570735912077,
"loss": 3.2811,
"step": 61900
},
{
"epoch": 6.667742977074588,
"grad_norm": 0.7486993074417114,
"learning_rate": 0.00020031246632905934,
"loss": 3.2811,
"step": 61950
},
{
"epoch": 6.6731245291141965,
"grad_norm": 0.8277780413627625,
"learning_rate": 0.00019998922529899794,
"loss": 3.2853,
"step": 62000
},
{
"epoch": 6.6731245291141965,
"eval_accuracy": 0.3862082108770016,
"eval_loss": 3.3669943809509277,
"eval_runtime": 179.7428,
"eval_samples_per_second": 100.204,
"eval_steps_per_second": 6.265,
"step": 62000
},
{
"epoch": 6.678506081153805,
"grad_norm": 0.7306733727455139,
"learning_rate": 0.00019966598426893653,
"loss": 3.2939,
"step": 62050
},
{
"epoch": 6.683887633193413,
"grad_norm": 0.7362749576568604,
"learning_rate": 0.0001993427432388751,
"loss": 3.2821,
"step": 62100
},
{
"epoch": 6.689269185233021,
"grad_norm": 0.7791106700897217,
"learning_rate": 0.00019901950220881367,
"loss": 3.2954,
"step": 62150
},
{
"epoch": 6.69465073727263,
"grad_norm": 0.7366771101951599,
"learning_rate": 0.0001986962611787523,
"loss": 3.2765,
"step": 62200
},
{
"epoch": 6.7000322893122375,
"grad_norm": 0.7491232752799988,
"learning_rate": 0.00019837302014869085,
"loss": 3.301,
"step": 62250
},
{
"epoch": 6.705413841351846,
"grad_norm": 0.742691695690155,
"learning_rate": 0.00019804977911862942,
"loss": 3.2811,
"step": 62300
},
{
"epoch": 6.710795393391454,
"grad_norm": 0.7395520210266113,
"learning_rate": 0.00019772653808856804,
"loss": 3.2865,
"step": 62350
},
{
"epoch": 6.716176945431062,
"grad_norm": 0.7324630618095398,
"learning_rate": 0.0001974032970585066,
"loss": 3.2892,
"step": 62400
},
{
"epoch": 6.721558497470671,
"grad_norm": 0.7284770011901855,
"learning_rate": 0.00019708005602844518,
"loss": 3.3008,
"step": 62450
},
{
"epoch": 6.7269400495102785,
"grad_norm": 0.7420316338539124,
"learning_rate": 0.00019675681499838377,
"loss": 3.2911,
"step": 62500
},
{
"epoch": 6.732321601549887,
"grad_norm": 0.7340070605278015,
"learning_rate": 0.00019643357396832237,
"loss": 3.2747,
"step": 62550
},
{
"epoch": 6.737703153589496,
"grad_norm": 0.7871265411376953,
"learning_rate": 0.00019611033293826096,
"loss": 3.297,
"step": 62600
},
{
"epoch": 6.743084705629103,
"grad_norm": 0.7386654615402222,
"learning_rate": 0.00019578709190819953,
"loss": 3.2891,
"step": 62650
},
{
"epoch": 6.748466257668712,
"grad_norm": 0.7021500468254089,
"learning_rate": 0.0001954638508781381,
"loss": 3.2665,
"step": 62700
},
{
"epoch": 6.75384780970832,
"grad_norm": 0.7563971281051636,
"learning_rate": 0.00019514060984807672,
"loss": 3.2821,
"step": 62750
},
{
"epoch": 6.759229361747928,
"grad_norm": 0.7392533421516418,
"learning_rate": 0.0001948173688180153,
"loss": 3.2866,
"step": 62800
},
{
"epoch": 6.7646109137875365,
"grad_norm": 0.7882697582244873,
"learning_rate": 0.00019449412778795386,
"loss": 3.3085,
"step": 62850
},
{
"epoch": 6.769992465827144,
"grad_norm": 0.751366138458252,
"learning_rate": 0.00019417088675789245,
"loss": 3.2849,
"step": 62900
},
{
"epoch": 6.775374017866753,
"grad_norm": 0.7440650463104248,
"learning_rate": 0.00019384764572783104,
"loss": 3.2891,
"step": 62950
},
{
"epoch": 6.780755569906361,
"grad_norm": 0.7576360702514648,
"learning_rate": 0.0001935244046977696,
"loss": 3.2815,
"step": 63000
},
{
"epoch": 6.780755569906361,
"eval_accuracy": 0.3868667560300451,
"eval_loss": 3.360293388366699,
"eval_runtime": 180.0715,
"eval_samples_per_second": 100.021,
"eval_steps_per_second": 6.253,
"step": 63000
},
{
"epoch": 6.786137121945969,
"grad_norm": 0.7954241633415222,
"learning_rate": 0.0001932011636677082,
"loss": 3.2861,
"step": 63050
},
{
"epoch": 6.7915186739855775,
"grad_norm": 0.7142677307128906,
"learning_rate": 0.00019287792263764677,
"loss": 3.2811,
"step": 63100
},
{
"epoch": 6.796900226025185,
"grad_norm": 0.7331296801567078,
"learning_rate": 0.00019255468160758537,
"loss": 3.2915,
"step": 63150
},
{
"epoch": 6.802281778064794,
"grad_norm": 0.7577782273292542,
"learning_rate": 0.00019223144057752396,
"loss": 3.2868,
"step": 63200
},
{
"epoch": 6.807663330104402,
"grad_norm": 0.7805274128913879,
"learning_rate": 0.00019190819954746253,
"loss": 3.2869,
"step": 63250
},
{
"epoch": 6.813044882144011,
"grad_norm": 0.7110979557037354,
"learning_rate": 0.0001915849585174011,
"loss": 3.2906,
"step": 63300
},
{
"epoch": 6.8184264341836185,
"grad_norm": 0.7497155070304871,
"learning_rate": 0.00019126171748733972,
"loss": 3.2839,
"step": 63350
},
{
"epoch": 6.823807986223227,
"grad_norm": 0.7474818825721741,
"learning_rate": 0.0001909384764572783,
"loss": 3.277,
"step": 63400
},
{
"epoch": 6.829189538262835,
"grad_norm": 0.7573341727256775,
"learning_rate": 0.00019061523542721688,
"loss": 3.2829,
"step": 63450
},
{
"epoch": 6.834571090302443,
"grad_norm": 0.6964619755744934,
"learning_rate": 0.00019029199439715548,
"loss": 3.2674,
"step": 63500
},
{
"epoch": 6.839952642342052,
"grad_norm": 0.7417646646499634,
"learning_rate": 0.00018996875336709404,
"loss": 3.3016,
"step": 63550
},
{
"epoch": 6.8453341943816595,
"grad_norm": 0.7879679203033447,
"learning_rate": 0.00018964551233703264,
"loss": 3.2684,
"step": 63600
},
{
"epoch": 6.850715746421268,
"grad_norm": 0.7311843633651733,
"learning_rate": 0.0001893222713069712,
"loss": 3.286,
"step": 63650
},
{
"epoch": 6.856097298460876,
"grad_norm": 0.7371758222579956,
"learning_rate": 0.0001889990302769098,
"loss": 3.2775,
"step": 63700
},
{
"epoch": 6.861478850500484,
"grad_norm": 0.7263456583023071,
"learning_rate": 0.0001886757892468484,
"loss": 3.2652,
"step": 63750
},
{
"epoch": 6.866860402540093,
"grad_norm": 0.7316994071006775,
"learning_rate": 0.00018835254821678696,
"loss": 3.2727,
"step": 63800
},
{
"epoch": 6.8722419545797,
"grad_norm": 0.7389889359474182,
"learning_rate": 0.00018802930718672553,
"loss": 3.3056,
"step": 63850
},
{
"epoch": 6.877623506619309,
"grad_norm": 0.7888100743293762,
"learning_rate": 0.00018770606615666415,
"loss": 3.2851,
"step": 63900
},
{
"epoch": 6.8830050586589175,
"grad_norm": 0.7792270183563232,
"learning_rate": 0.00018738282512660272,
"loss": 3.2949,
"step": 63950
},
{
"epoch": 6.888386610698525,
"grad_norm": 0.7635626196861267,
"learning_rate": 0.0001870595840965413,
"loss": 3.2878,
"step": 64000
},
{
"epoch": 6.888386610698525,
"eval_accuracy": 0.3867621232990798,
"eval_loss": 3.3606150150299072,
"eval_runtime": 180.6437,
"eval_samples_per_second": 99.705,
"eval_steps_per_second": 6.233,
"step": 64000
},
{
"epoch": 6.893768162738134,
"grad_norm": 0.7803065180778503,
"learning_rate": 0.0001867363430664799,
"loss": 3.2846,
"step": 64050
},
{
"epoch": 6.899149714777742,
"grad_norm": 0.7100071310997009,
"learning_rate": 0.00018641310203641848,
"loss": 3.3034,
"step": 64100
},
{
"epoch": 6.90453126681735,
"grad_norm": 0.7439180016517639,
"learning_rate": 0.00018608986100635705,
"loss": 3.2756,
"step": 64150
},
{
"epoch": 6.9099128188569585,
"grad_norm": 0.7899887561798096,
"learning_rate": 0.00018576661997629564,
"loss": 3.2778,
"step": 64200
},
{
"epoch": 6.915294370896566,
"grad_norm": 0.7477514147758484,
"learning_rate": 0.00018544337894623423,
"loss": 3.2992,
"step": 64250
},
{
"epoch": 6.920675922936175,
"grad_norm": 0.7305505275726318,
"learning_rate": 0.00018512660273677404,
"loss": 3.2864,
"step": 64300
},
{
"epoch": 6.926057474975783,
"grad_norm": 0.7840530872344971,
"learning_rate": 0.0001848033617067126,
"loss": 3.2804,
"step": 64350
},
{
"epoch": 6.931439027015391,
"grad_norm": 0.7968816161155701,
"learning_rate": 0.00018448012067665123,
"loss": 3.2859,
"step": 64400
},
{
"epoch": 6.9368205790549995,
"grad_norm": 0.7263173460960388,
"learning_rate": 0.0001841568796465898,
"loss": 3.2938,
"step": 64450
},
{
"epoch": 6.942202131094608,
"grad_norm": 0.74007248878479,
"learning_rate": 0.00018383363861652837,
"loss": 3.2647,
"step": 64500
},
{
"epoch": 6.947583683134216,
"grad_norm": 0.730910062789917,
"learning_rate": 0.00018351039758646696,
"loss": 3.2913,
"step": 64550
},
{
"epoch": 6.952965235173824,
"grad_norm": 0.7842426300048828,
"learning_rate": 0.00018318715655640555,
"loss": 3.2938,
"step": 64600
},
{
"epoch": 6.958346787213433,
"grad_norm": 0.7127736210823059,
"learning_rate": 0.00018286391552634412,
"loss": 3.2837,
"step": 64650
},
{
"epoch": 6.9637283392530405,
"grad_norm": 0.7509408593177795,
"learning_rate": 0.00018254067449628272,
"loss": 3.2829,
"step": 64700
},
{
"epoch": 6.969109891292649,
"grad_norm": 0.7198226451873779,
"learning_rate": 0.00018221743346622128,
"loss": 3.2862,
"step": 64750
},
{
"epoch": 6.974491443332257,
"grad_norm": 0.7937029600143433,
"learning_rate": 0.00018189419243615988,
"loss": 3.2979,
"step": 64800
},
{
"epoch": 6.979872995371865,
"grad_norm": 0.758535623550415,
"learning_rate": 0.00018157095140609847,
"loss": 3.2594,
"step": 64850
},
{
"epoch": 6.985254547411474,
"grad_norm": 0.8164926767349243,
"learning_rate": 0.00018124771037603704,
"loss": 3.2843,
"step": 64900
},
{
"epoch": 6.990636099451081,
"grad_norm": 0.7364434003829956,
"learning_rate": 0.0001809244693459756,
"loss": 3.2898,
"step": 64950
},
{
"epoch": 6.99601765149069,
"grad_norm": 0.7529191970825195,
"learning_rate": 0.00018060122831591423,
"loss": 3.2674,
"step": 65000
},
{
"epoch": 6.99601765149069,
"eval_accuracy": 0.3877488001733231,
"eval_loss": 3.354095458984375,
"eval_runtime": 180.0368,
"eval_samples_per_second": 100.041,
"eval_steps_per_second": 6.254,
"step": 65000
},
{
"epoch": 7.0013992035302985,
"grad_norm": 0.7346994876861572,
"learning_rate": 0.0001802779872858528,
"loss": 3.2675,
"step": 65050
},
{
"epoch": 7.006780755569906,
"grad_norm": 0.7564079165458679,
"learning_rate": 0.00017995474625579137,
"loss": 3.2023,
"step": 65100
},
{
"epoch": 7.012162307609515,
"grad_norm": 0.7516438364982605,
"learning_rate": 0.00017963150522573,
"loss": 3.2093,
"step": 65150
},
{
"epoch": 7.017543859649122,
"grad_norm": 0.757781445980072,
"learning_rate": 0.00017930826419566856,
"loss": 3.2116,
"step": 65200
},
{
"epoch": 7.022925411688731,
"grad_norm": 0.7500360608100891,
"learning_rate": 0.00017898502316560712,
"loss": 3.2007,
"step": 65250
},
{
"epoch": 7.0283069637283395,
"grad_norm": 0.7233052849769592,
"learning_rate": 0.00017866178213554572,
"loss": 3.2189,
"step": 65300
},
{
"epoch": 7.033688515767947,
"grad_norm": 0.8041335940361023,
"learning_rate": 0.0001783385411054843,
"loss": 3.1901,
"step": 65350
},
{
"epoch": 7.039070067807556,
"grad_norm": 0.7104873657226562,
"learning_rate": 0.0001780153000754229,
"loss": 3.1969,
"step": 65400
},
{
"epoch": 7.044451619847164,
"grad_norm": 0.7547011375427246,
"learning_rate": 0.00017769205904536147,
"loss": 3.2141,
"step": 65450
},
{
"epoch": 7.049833171886772,
"grad_norm": 0.7907810211181641,
"learning_rate": 0.00017736881801530004,
"loss": 3.2051,
"step": 65500
},
{
"epoch": 7.0552147239263805,
"grad_norm": 0.7557851076126099,
"learning_rate": 0.00017704557698523866,
"loss": 3.209,
"step": 65550
},
{
"epoch": 7.060596275965988,
"grad_norm": 0.7390865087509155,
"learning_rate": 0.00017672233595517723,
"loss": 3.2206,
"step": 65600
},
{
"epoch": 7.065977828005597,
"grad_norm": 0.7852340340614319,
"learning_rate": 0.0001763990949251158,
"loss": 3.2176,
"step": 65650
},
{
"epoch": 7.071359380045205,
"grad_norm": 0.7627012729644775,
"learning_rate": 0.00017607585389505442,
"loss": 3.2183,
"step": 65700
},
{
"epoch": 7.076740932084813,
"grad_norm": 0.7575951814651489,
"learning_rate": 0.000175752612864993,
"loss": 3.2181,
"step": 65750
},
{
"epoch": 7.0821224841244215,
"grad_norm": 0.7672650218009949,
"learning_rate": 0.00017542937183493156,
"loss": 3.2102,
"step": 65800
},
{
"epoch": 7.08750403616403,
"grad_norm": 0.7804780006408691,
"learning_rate": 0.00017510613080487015,
"loss": 3.2121,
"step": 65850
},
{
"epoch": 7.092885588203638,
"grad_norm": 0.7676131725311279,
"learning_rate": 0.00017478288977480874,
"loss": 3.2017,
"step": 65900
},
{
"epoch": 7.098267140243246,
"grad_norm": 0.7814217805862427,
"learning_rate": 0.00017446611356534855,
"loss": 3.2032,
"step": 65950
},
{
"epoch": 7.103648692282855,
"grad_norm": 0.7216848731040955,
"learning_rate": 0.00017414287253528712,
"loss": 3.2193,
"step": 66000
},
{
"epoch": 7.103648692282855,
"eval_accuracy": 0.38735189117413676,
"eval_loss": 3.357109785079956,
"eval_runtime": 179.9063,
"eval_samples_per_second": 100.113,
"eval_steps_per_second": 6.259,
"step": 66000
},
{
"epoch": 7.109030244322462,
"grad_norm": 0.754563570022583,
"learning_rate": 0.00017381963150522569,
"loss": 3.2194,
"step": 66050
},
{
"epoch": 7.114411796362071,
"grad_norm": 0.7799685597419739,
"learning_rate": 0.0001734963904751643,
"loss": 3.2178,
"step": 66100
},
{
"epoch": 7.119793348401679,
"grad_norm": 0.7888051271438599,
"learning_rate": 0.00017317314944510288,
"loss": 3.2021,
"step": 66150
},
{
"epoch": 7.125174900441287,
"grad_norm": 0.7203174829483032,
"learning_rate": 0.00017284990841504147,
"loss": 3.211,
"step": 66200
},
{
"epoch": 7.130556452480896,
"grad_norm": 0.8019355535507202,
"learning_rate": 0.00017252666738498007,
"loss": 3.203,
"step": 66250
},
{
"epoch": 7.135938004520503,
"grad_norm": 0.7878087162971497,
"learning_rate": 0.00017220342635491863,
"loss": 3.2177,
"step": 66300
},
{
"epoch": 7.141319556560112,
"grad_norm": 0.7539237141609192,
"learning_rate": 0.00017188018532485723,
"loss": 3.2079,
"step": 66350
},
{
"epoch": 7.1467011085997205,
"grad_norm": 0.7446153163909912,
"learning_rate": 0.0001715569442947958,
"loss": 3.1975,
"step": 66400
},
{
"epoch": 7.152082660639328,
"grad_norm": 0.733919084072113,
"learning_rate": 0.0001712337032647344,
"loss": 3.243,
"step": 66450
},
{
"epoch": 7.157464212678937,
"grad_norm": 0.806229829788208,
"learning_rate": 0.00017091046223467298,
"loss": 3.2213,
"step": 66500
},
{
"epoch": 7.162845764718545,
"grad_norm": 0.7871570587158203,
"learning_rate": 0.00017058722120461155,
"loss": 3.2023,
"step": 66550
},
{
"epoch": 7.168227316758153,
"grad_norm": 0.7633065581321716,
"learning_rate": 0.00017026398017455012,
"loss": 3.2327,
"step": 66600
},
{
"epoch": 7.1736088687977615,
"grad_norm": 0.7800832986831665,
"learning_rate": 0.00016994073914448874,
"loss": 3.2351,
"step": 66650
},
{
"epoch": 7.178990420837369,
"grad_norm": 0.7342001795768738,
"learning_rate": 0.0001696174981144273,
"loss": 3.2212,
"step": 66700
},
{
"epoch": 7.184371972876978,
"grad_norm": 0.7810934782028198,
"learning_rate": 0.00016929425708436588,
"loss": 3.2014,
"step": 66750
},
{
"epoch": 7.189753524916586,
"grad_norm": 0.7123428583145142,
"learning_rate": 0.0001689710160543045,
"loss": 3.2019,
"step": 66800
},
{
"epoch": 7.195135076956194,
"grad_norm": 0.8075796365737915,
"learning_rate": 0.00016864777502424307,
"loss": 3.2242,
"step": 66850
},
{
"epoch": 7.2005166289958025,
"grad_norm": 0.8010225892066956,
"learning_rate": 0.00016832453399418163,
"loss": 3.2183,
"step": 66900
},
{
"epoch": 7.205898181035411,
"grad_norm": 0.7485325336456299,
"learning_rate": 0.00016800129296412023,
"loss": 3.2022,
"step": 66950
},
{
"epoch": 7.211279733075019,
"grad_norm": 0.8330503702163696,
"learning_rate": 0.00016767805193405882,
"loss": 3.2274,
"step": 67000
},
{
"epoch": 7.211279733075019,
"eval_accuracy": 0.3875922313619824,
"eval_loss": 3.357809066772461,
"eval_runtime": 179.9928,
"eval_samples_per_second": 100.065,
"eval_steps_per_second": 6.256,
"step": 67000
},
{
"epoch": 7.216661285114627,
"grad_norm": 0.7660570740699768,
"learning_rate": 0.0001673548109039974,
"loss": 3.218,
"step": 67050
},
{
"epoch": 7.222042837154235,
"grad_norm": 0.7876999974250793,
"learning_rate": 0.00016703156987393598,
"loss": 3.2167,
"step": 67100
},
{
"epoch": 7.2274243891938434,
"grad_norm": 0.7447673082351685,
"learning_rate": 0.00016670832884387455,
"loss": 3.2343,
"step": 67150
},
{
"epoch": 7.232805941233452,
"grad_norm": 0.7839945554733276,
"learning_rate": 0.00016638508781381317,
"loss": 3.2155,
"step": 67200
},
{
"epoch": 7.23818749327306,
"grad_norm": 0.7349887490272522,
"learning_rate": 0.00016606184678375174,
"loss": 3.2318,
"step": 67250
},
{
"epoch": 7.243569045312668,
"grad_norm": 0.7882452011108398,
"learning_rate": 0.0001657386057536903,
"loss": 3.2308,
"step": 67300
},
{
"epoch": 7.248950597352277,
"grad_norm": 0.7886552810668945,
"learning_rate": 0.00016541536472362893,
"loss": 3.228,
"step": 67350
},
{
"epoch": 7.254332149391884,
"grad_norm": 0.7716850638389587,
"learning_rate": 0.0001650921236935675,
"loss": 3.2112,
"step": 67400
},
{
"epoch": 7.259713701431493,
"grad_norm": 0.7802140712738037,
"learning_rate": 0.00016476888266350607,
"loss": 3.2307,
"step": 67450
},
{
"epoch": 7.265095253471101,
"grad_norm": 0.7871243953704834,
"learning_rate": 0.00016444564163344466,
"loss": 3.2068,
"step": 67500
},
{
"epoch": 7.270476805510709,
"grad_norm": 0.8147279024124146,
"learning_rate": 0.00016412240060338326,
"loss": 3.2212,
"step": 67550
},
{
"epoch": 7.275858357550318,
"grad_norm": 0.8272632956504822,
"learning_rate": 0.00016379915957332182,
"loss": 3.2287,
"step": 67600
},
{
"epoch": 7.281239909589925,
"grad_norm": 0.7926257252693176,
"learning_rate": 0.00016347591854326042,
"loss": 3.2024,
"step": 67650
},
{
"epoch": 7.286621461629534,
"grad_norm": 0.8117291927337646,
"learning_rate": 0.00016315267751319898,
"loss": 3.2295,
"step": 67700
},
{
"epoch": 7.2920030136691425,
"grad_norm": 0.7746281623840332,
"learning_rate": 0.00016282943648313755,
"loss": 3.2416,
"step": 67750
},
{
"epoch": 7.29738456570875,
"grad_norm": 0.7859088182449341,
"learning_rate": 0.00016250619545307617,
"loss": 3.2179,
"step": 67800
},
{
"epoch": 7.302766117748359,
"grad_norm": 0.749521017074585,
"learning_rate": 0.00016218295442301474,
"loss": 3.2167,
"step": 67850
},
{
"epoch": 7.308147669787967,
"grad_norm": 0.7174946665763855,
"learning_rate": 0.0001618597133929533,
"loss": 3.2142,
"step": 67900
},
{
"epoch": 7.313529221827575,
"grad_norm": 0.7507534027099609,
"learning_rate": 0.00016153647236289193,
"loss": 3.2187,
"step": 67950
},
{
"epoch": 7.3189107738671835,
"grad_norm": 0.7594087719917297,
"learning_rate": 0.0001612132313328305,
"loss": 3.2201,
"step": 68000
},
{
"epoch": 7.3189107738671835,
"eval_accuracy": 0.38803912068956337,
"eval_loss": 3.35626482963562,
"eval_runtime": 180.0467,
"eval_samples_per_second": 100.035,
"eval_steps_per_second": 6.254,
"step": 68000
},
{
"epoch": 7.324292325906791,
"grad_norm": 0.9152792096138,
"learning_rate": 0.0001608899903027691,
"loss": 3.2376,
"step": 68050
},
{
"epoch": 7.3296738779464,
"grad_norm": 0.770829975605011,
"learning_rate": 0.00016056674927270766,
"loss": 3.2143,
"step": 68100
},
{
"epoch": 7.335055429986008,
"grad_norm": 0.7539659142494202,
"learning_rate": 0.00016024350824264626,
"loss": 3.2165,
"step": 68150
},
{
"epoch": 7.340436982025616,
"grad_norm": 0.7711495757102966,
"learning_rate": 0.00015992026721258485,
"loss": 3.2089,
"step": 68200
},
{
"epoch": 7.3458185340652244,
"grad_norm": 0.8086374402046204,
"learning_rate": 0.00015959702618252342,
"loss": 3.2291,
"step": 68250
},
{
"epoch": 7.351200086104833,
"grad_norm": 0.7436811327934265,
"learning_rate": 0.00015927378515246199,
"loss": 3.2341,
"step": 68300
},
{
"epoch": 7.356581638144441,
"grad_norm": 0.7765557169914246,
"learning_rate": 0.0001589505441224006,
"loss": 3.2137,
"step": 68350
},
{
"epoch": 7.361963190184049,
"grad_norm": 0.8380234837532043,
"learning_rate": 0.00015862730309233917,
"loss": 3.2272,
"step": 68400
},
{
"epoch": 7.367344742223658,
"grad_norm": 0.7880136966705322,
"learning_rate": 0.00015830406206227774,
"loss": 3.2346,
"step": 68450
},
{
"epoch": 7.372726294263265,
"grad_norm": 0.7508086562156677,
"learning_rate": 0.00015798082103221636,
"loss": 3.2303,
"step": 68500
},
{
"epoch": 7.378107846302874,
"grad_norm": 0.7732205390930176,
"learning_rate": 0.00015765758000215493,
"loss": 3.2266,
"step": 68550
},
{
"epoch": 7.383489398342482,
"grad_norm": 0.7558825016021729,
"learning_rate": 0.0001573343389720935,
"loss": 3.2174,
"step": 68600
},
{
"epoch": 7.38887095038209,
"grad_norm": 0.8041996359825134,
"learning_rate": 0.0001570110979420321,
"loss": 3.2201,
"step": 68650
},
{
"epoch": 7.394252502421699,
"grad_norm": 0.7656950354576111,
"learning_rate": 0.0001566878569119707,
"loss": 3.2176,
"step": 68700
},
{
"epoch": 7.399634054461306,
"grad_norm": 0.7538243532180786,
"learning_rate": 0.00015636461588190926,
"loss": 3.223,
"step": 68750
},
{
"epoch": 7.405015606500915,
"grad_norm": 0.8169330358505249,
"learning_rate": 0.00015604137485184785,
"loss": 3.2286,
"step": 68800
},
{
"epoch": 7.4103971585405235,
"grad_norm": 0.7956921458244324,
"learning_rate": 0.00015571813382178642,
"loss": 3.2345,
"step": 68850
},
{
"epoch": 7.415778710580131,
"grad_norm": 0.7846260070800781,
"learning_rate": 0.000155394892791725,
"loss": 3.2288,
"step": 68900
},
{
"epoch": 7.42116026261974,
"grad_norm": 0.7715317606925964,
"learning_rate": 0.0001550716517616636,
"loss": 3.2327,
"step": 68950
},
{
"epoch": 7.426541814659347,
"grad_norm": 0.7787097692489624,
"learning_rate": 0.00015474841073160217,
"loss": 3.2384,
"step": 69000
},
{
"epoch": 7.426541814659347,
"eval_accuracy": 0.3884835110007246,
"eval_loss": 3.3530378341674805,
"eval_runtime": 179.8752,
"eval_samples_per_second": 100.131,
"eval_steps_per_second": 6.26,
"step": 69000
},
{
"epoch": 7.431923366698956,
"grad_norm": 0.8361562490463257,
"learning_rate": 0.0001544251697015408,
"loss": 3.2302,
"step": 69050
},
{
"epoch": 7.4373049187385645,
"grad_norm": 0.7875263094902039,
"learning_rate": 0.00015410192867147936,
"loss": 3.2302,
"step": 69100
},
{
"epoch": 7.442686470778172,
"grad_norm": 0.8438063263893127,
"learning_rate": 0.00015377868764141793,
"loss": 3.2276,
"step": 69150
},
{
"epoch": 7.448068022817781,
"grad_norm": 0.8052544593811035,
"learning_rate": 0.00015345544661135653,
"loss": 3.2535,
"step": 69200
},
{
"epoch": 7.453449574857389,
"grad_norm": 0.8062599301338196,
"learning_rate": 0.00015313220558129512,
"loss": 3.2247,
"step": 69250
},
{
"epoch": 7.458831126896997,
"grad_norm": 0.7678873538970947,
"learning_rate": 0.0001528089645512337,
"loss": 3.2306,
"step": 69300
},
{
"epoch": 7.4642126789366054,
"grad_norm": 0.8218041658401489,
"learning_rate": 0.00015248572352117228,
"loss": 3.2194,
"step": 69350
},
{
"epoch": 7.469594230976213,
"grad_norm": 0.7693276405334473,
"learning_rate": 0.00015216248249111085,
"loss": 3.2326,
"step": 69400
},
{
"epoch": 7.474975783015822,
"grad_norm": 0.7829477787017822,
"learning_rate": 0.00015183924146104945,
"loss": 3.2393,
"step": 69450
},
{
"epoch": 7.48035733505543,
"grad_norm": 0.78624427318573,
"learning_rate": 0.00015151600043098804,
"loss": 3.2352,
"step": 69500
},
{
"epoch": 7.485738887095038,
"grad_norm": 0.8021953701972961,
"learning_rate": 0.0001511927594009266,
"loss": 3.23,
"step": 69550
},
{
"epoch": 7.491120439134646,
"grad_norm": 0.8379894495010376,
"learning_rate": 0.00015086951837086518,
"loss": 3.2369,
"step": 69600
},
{
"epoch": 7.496501991174255,
"grad_norm": 0.8550412654876709,
"learning_rate": 0.0001505462773408038,
"loss": 3.2208,
"step": 69650
},
{
"epoch": 7.501883543213863,
"grad_norm": 0.7694771885871887,
"learning_rate": 0.00015022303631074236,
"loss": 3.2236,
"step": 69700
},
{
"epoch": 7.507265095253471,
"grad_norm": 0.8577457666397095,
"learning_rate": 0.00014989979528068096,
"loss": 3.2487,
"step": 69750
},
{
"epoch": 7.51264664729308,
"grad_norm": 0.7951626181602478,
"learning_rate": 0.00014957655425061953,
"loss": 3.2151,
"step": 69800
},
{
"epoch": 7.518028199332687,
"grad_norm": 0.770133376121521,
"learning_rate": 0.00014925331322055812,
"loss": 3.2279,
"step": 69850
},
{
"epoch": 7.523409751372296,
"grad_norm": 0.7697238326072693,
"learning_rate": 0.0001489300721904967,
"loss": 3.2426,
"step": 69900
},
{
"epoch": 7.528791303411904,
"grad_norm": 0.8447310328483582,
"learning_rate": 0.0001486132959810365,
"loss": 3.2232,
"step": 69950
},
{
"epoch": 7.534172855451512,
"grad_norm": 0.7980256676673889,
"learning_rate": 0.0001482900549509751,
"loss": 3.2464,
"step": 70000
},
{
"epoch": 7.534172855451512,
"eval_accuracy": 0.38879534478875216,
"eval_loss": 3.348252534866333,
"eval_runtime": 180.2429,
"eval_samples_per_second": 99.926,
"eval_steps_per_second": 6.247,
"step": 70000
},
{
"epoch": 7.539554407491121,
"grad_norm": 0.7334508299827576,
"learning_rate": 0.00014796681392091368,
"loss": 3.2262,
"step": 70050
},
{
"epoch": 7.544935959530728,
"grad_norm": 0.7756717801094055,
"learning_rate": 0.00014764357289085228,
"loss": 3.236,
"step": 70100
},
{
"epoch": 7.550317511570337,
"grad_norm": 0.852515459060669,
"learning_rate": 0.00014732033186079085,
"loss": 3.2377,
"step": 70150
},
{
"epoch": 7.5556990636099455,
"grad_norm": 0.7722263932228088,
"learning_rate": 0.00014699709083072944,
"loss": 3.222,
"step": 70200
},
{
"epoch": 7.561080615649553,
"grad_norm": 0.8237218856811523,
"learning_rate": 0.00014667384980066804,
"loss": 3.2207,
"step": 70250
},
{
"epoch": 7.566462167689162,
"grad_norm": 0.7758240699768066,
"learning_rate": 0.0001463506087706066,
"loss": 3.2207,
"step": 70300
},
{
"epoch": 7.57184371972877,
"grad_norm": 0.7621212601661682,
"learning_rate": 0.0001460273677405452,
"loss": 3.2295,
"step": 70350
},
{
"epoch": 7.577225271768378,
"grad_norm": 0.7284106016159058,
"learning_rate": 0.00014570412671048377,
"loss": 3.2325,
"step": 70400
},
{
"epoch": 7.5826068238079865,
"grad_norm": 0.793984055519104,
"learning_rate": 0.00014538088568042236,
"loss": 3.2337,
"step": 70450
},
{
"epoch": 7.587988375847594,
"grad_norm": 0.7775974869728088,
"learning_rate": 0.00014505764465036093,
"loss": 3.226,
"step": 70500
},
{
"epoch": 7.593369927887203,
"grad_norm": 0.7938340306282043,
"learning_rate": 0.00014473440362029952,
"loss": 3.2377,
"step": 70550
},
{
"epoch": 7.598751479926811,
"grad_norm": 0.8512852191925049,
"learning_rate": 0.00014441116259023812,
"loss": 3.2208,
"step": 70600
},
{
"epoch": 7.604133031966419,
"grad_norm": 0.8293440341949463,
"learning_rate": 0.00014408792156017669,
"loss": 3.2424,
"step": 70650
},
{
"epoch": 7.609514584006027,
"grad_norm": 0.7990338206291199,
"learning_rate": 0.00014376468053011528,
"loss": 3.2255,
"step": 70700
},
{
"epoch": 7.614896136045635,
"grad_norm": 0.7645559906959534,
"learning_rate": 0.00014344143950005387,
"loss": 3.2468,
"step": 70750
},
{
"epoch": 7.620277688085244,
"grad_norm": 0.7902533411979675,
"learning_rate": 0.00014311819846999244,
"loss": 3.2283,
"step": 70800
},
{
"epoch": 7.625659240124852,
"grad_norm": 0.7919648885726929,
"learning_rate": 0.00014279495743993104,
"loss": 3.2399,
"step": 70850
},
{
"epoch": 7.63104079216446,
"grad_norm": 0.7276238799095154,
"learning_rate": 0.00014247171640986963,
"loss": 3.2308,
"step": 70900
},
{
"epoch": 7.636422344204068,
"grad_norm": 0.7745274901390076,
"learning_rate": 0.0001421484753798082,
"loss": 3.2317,
"step": 70950
},
{
"epoch": 7.641803896243677,
"grad_norm": 0.7933140397071838,
"learning_rate": 0.0001418252343497468,
"loss": 3.2376,
"step": 71000
},
{
"epoch": 7.641803896243677,
"eval_accuracy": 0.389115218890478,
"eval_loss": 3.343194007873535,
"eval_runtime": 179.6738,
"eval_samples_per_second": 100.243,
"eval_steps_per_second": 6.267,
"step": 71000
},
{
"epoch": 7.647185448283285,
"grad_norm": 0.7818636298179626,
"learning_rate": 0.00014150199331968536,
"loss": 3.2169,
"step": 71050
},
{
"epoch": 7.652567000322893,
"grad_norm": 0.8027185797691345,
"learning_rate": 0.00014117875228962396,
"loss": 3.219,
"step": 71100
},
{
"epoch": 7.657948552362502,
"grad_norm": 0.8079972863197327,
"learning_rate": 0.00014085551125956252,
"loss": 3.231,
"step": 71150
},
{
"epoch": 7.663330104402109,
"grad_norm": 0.7901222109794617,
"learning_rate": 0.00014053227022950112,
"loss": 3.2243,
"step": 71200
},
{
"epoch": 7.668711656441718,
"grad_norm": 0.8354732394218445,
"learning_rate": 0.0001402090291994397,
"loss": 3.2378,
"step": 71250
},
{
"epoch": 7.674093208481326,
"grad_norm": 0.8050693869590759,
"learning_rate": 0.00013988578816937828,
"loss": 3.2202,
"step": 71300
},
{
"epoch": 7.679474760520934,
"grad_norm": 0.7609599828720093,
"learning_rate": 0.00013956254713931687,
"loss": 3.2262,
"step": 71350
},
{
"epoch": 7.684856312560543,
"grad_norm": 0.8210611343383789,
"learning_rate": 0.00013923930610925547,
"loss": 3.233,
"step": 71400
},
{
"epoch": 7.69023786460015,
"grad_norm": 0.7654331922531128,
"learning_rate": 0.00013891606507919404,
"loss": 3.2487,
"step": 71450
},
{
"epoch": 7.695619416639759,
"grad_norm": 0.8369812965393066,
"learning_rate": 0.00013859282404913263,
"loss": 3.2273,
"step": 71500
},
{
"epoch": 7.7010009686793675,
"grad_norm": 0.8057788610458374,
"learning_rate": 0.0001382695830190712,
"loss": 3.2144,
"step": 71550
},
{
"epoch": 7.706382520718975,
"grad_norm": 0.7989969849586487,
"learning_rate": 0.0001379463419890098,
"loss": 3.2386,
"step": 71600
},
{
"epoch": 7.711764072758584,
"grad_norm": 0.8108153939247131,
"learning_rate": 0.00013762310095894836,
"loss": 3.2351,
"step": 71650
},
{
"epoch": 7.717145624798192,
"grad_norm": 0.7994381785392761,
"learning_rate": 0.0001373063247494882,
"loss": 3.2329,
"step": 71700
},
{
"epoch": 7.7225271768378,
"grad_norm": 0.7698352932929993,
"learning_rate": 0.00013698308371942676,
"loss": 3.2261,
"step": 71750
},
{
"epoch": 7.727908728877408,
"grad_norm": 0.7884970903396606,
"learning_rate": 0.00013665984268936536,
"loss": 3.2273,
"step": 71800
},
{
"epoch": 7.733290280917016,
"grad_norm": 0.8238911032676697,
"learning_rate": 0.00013633660165930395,
"loss": 3.2296,
"step": 71850
},
{
"epoch": 7.738671832956625,
"grad_norm": 0.804165780544281,
"learning_rate": 0.00013601336062924255,
"loss": 3.2403,
"step": 71900
},
{
"epoch": 7.744053384996233,
"grad_norm": 0.8062986731529236,
"learning_rate": 0.00013569011959918111,
"loss": 3.2474,
"step": 71950
},
{
"epoch": 7.749434937035841,
"grad_norm": 0.772881805896759,
"learning_rate": 0.0001353668785691197,
"loss": 3.2447,
"step": 72000
},
{
"epoch": 7.749434937035841,
"eval_accuracy": 0.389392827018844,
"eval_loss": 3.343226909637451,
"eval_runtime": 180.5115,
"eval_samples_per_second": 99.778,
"eval_steps_per_second": 6.238,
"step": 72000
},
{
"epoch": 7.754816489075449,
"grad_norm": 0.7907891869544983,
"learning_rate": 0.00013504363753905828,
"loss": 3.2411,
"step": 72050
},
{
"epoch": 7.760198041115058,
"grad_norm": 0.8008028864860535,
"learning_rate": 0.00013472039650899687,
"loss": 3.2389,
"step": 72100
},
{
"epoch": 7.765579593154666,
"grad_norm": 0.8094232082366943,
"learning_rate": 0.00013439715547893544,
"loss": 3.2357,
"step": 72150
},
{
"epoch": 7.770961145194274,
"grad_norm": 0.7857368588447571,
"learning_rate": 0.00013407391444887403,
"loss": 3.2359,
"step": 72200
},
{
"epoch": 7.776342697233883,
"grad_norm": 0.7756304740905762,
"learning_rate": 0.0001337506734188126,
"loss": 3.2329,
"step": 72250
},
{
"epoch": 7.78172424927349,
"grad_norm": 0.7991217374801636,
"learning_rate": 0.0001334274323887512,
"loss": 3.2244,
"step": 72300
},
{
"epoch": 7.787105801313099,
"grad_norm": 0.8120244741439819,
"learning_rate": 0.0001331041913586898,
"loss": 3.2231,
"step": 72350
},
{
"epoch": 7.792487353352707,
"grad_norm": 0.8280323147773743,
"learning_rate": 0.00013278095032862838,
"loss": 3.2446,
"step": 72400
},
{
"epoch": 7.797868905392315,
"grad_norm": 0.7918257713317871,
"learning_rate": 0.00013245770929856695,
"loss": 3.2223,
"step": 72450
},
{
"epoch": 7.803250457431924,
"grad_norm": 0.7558754086494446,
"learning_rate": 0.00013213446826850555,
"loss": 3.241,
"step": 72500
},
{
"epoch": 7.808632009471531,
"grad_norm": 0.8165222406387329,
"learning_rate": 0.00013181122723844411,
"loss": 3.2175,
"step": 72550
},
{
"epoch": 7.81401356151114,
"grad_norm": 0.8103949427604675,
"learning_rate": 0.0001314879862083827,
"loss": 3.2113,
"step": 72600
},
{
"epoch": 7.819395113550748,
"grad_norm": 0.7988500595092773,
"learning_rate": 0.00013116474517832128,
"loss": 3.2328,
"step": 72650
},
{
"epoch": 7.824776665590356,
"grad_norm": 0.838823139667511,
"learning_rate": 0.00013084150414825987,
"loss": 3.2343,
"step": 72700
},
{
"epoch": 7.830158217629965,
"grad_norm": 0.8075284957885742,
"learning_rate": 0.00013051826311819844,
"loss": 3.2446,
"step": 72750
},
{
"epoch": 7.835539769669572,
"grad_norm": 0.7953897714614868,
"learning_rate": 0.00013019502208813703,
"loss": 3.2419,
"step": 72800
},
{
"epoch": 7.840921321709181,
"grad_norm": 0.7763371467590332,
"learning_rate": 0.00012987178105807563,
"loss": 3.227,
"step": 72850
},
{
"epoch": 7.846302873748789,
"grad_norm": 0.8023640513420105,
"learning_rate": 0.00012954854002801422,
"loss": 3.2381,
"step": 72900
},
{
"epoch": 7.851684425788397,
"grad_norm": 0.7901715040206909,
"learning_rate": 0.0001292252989979528,
"loss": 3.2443,
"step": 72950
},
{
"epoch": 7.857065977828006,
"grad_norm": 0.7858378887176514,
"learning_rate": 0.00012890205796789139,
"loss": 3.2332,
"step": 73000
},
{
"epoch": 7.857065977828006,
"eval_accuracy": 0.39001747247088947,
"eval_loss": 3.336719274520874,
"eval_runtime": 179.8309,
"eval_samples_per_second": 100.155,
"eval_steps_per_second": 6.261,
"step": 73000
},
{
"epoch": 7.862447529867614,
"grad_norm": 0.795928418636322,
"learning_rate": 0.00012857881693782998,
"loss": 3.2402,
"step": 73050
},
{
"epoch": 7.867829081907222,
"grad_norm": 0.8185012340545654,
"learning_rate": 0.00012825557590776855,
"loss": 3.2294,
"step": 73100
},
{
"epoch": 7.87321063394683,
"grad_norm": 0.8013267517089844,
"learning_rate": 0.00012793233487770714,
"loss": 3.2336,
"step": 73150
},
{
"epoch": 7.878592185986438,
"grad_norm": 0.8170375227928162,
"learning_rate": 0.0001276090938476457,
"loss": 3.2552,
"step": 73200
},
{
"epoch": 7.883973738026047,
"grad_norm": 0.7894675731658936,
"learning_rate": 0.0001272858528175843,
"loss": 3.2187,
"step": 73250
},
{
"epoch": 7.889355290065655,
"grad_norm": 0.7849202752113342,
"learning_rate": 0.00012696261178752287,
"loss": 3.2207,
"step": 73300
},
{
"epoch": 7.894736842105263,
"grad_norm": 0.799229621887207,
"learning_rate": 0.00012663937075746147,
"loss": 3.2559,
"step": 73350
},
{
"epoch": 7.900118394144871,
"grad_norm": 0.7787337303161621,
"learning_rate": 0.00012631612972740006,
"loss": 3.2399,
"step": 73400
},
{
"epoch": 7.90549994618448,
"grad_norm": 0.7923638224601746,
"learning_rate": 0.00012599288869733863,
"loss": 3.2334,
"step": 73450
},
{
"epoch": 7.910881498224088,
"grad_norm": 0.7863078713417053,
"learning_rate": 0.00012566964766727722,
"loss": 3.2534,
"step": 73500
},
{
"epoch": 7.916263050263696,
"grad_norm": 0.8022647500038147,
"learning_rate": 0.00012534640663721582,
"loss": 3.2506,
"step": 73550
},
{
"epoch": 7.921644602303305,
"grad_norm": 0.8193617463111877,
"learning_rate": 0.00012502316560715439,
"loss": 3.2194,
"step": 73600
},
{
"epoch": 7.927026154342912,
"grad_norm": 0.7646448612213135,
"learning_rate": 0.00012469992457709298,
"loss": 3.2243,
"step": 73650
},
{
"epoch": 7.932407706382521,
"grad_norm": 0.7733369469642639,
"learning_rate": 0.00012437668354703158,
"loss": 3.2191,
"step": 73700
},
{
"epoch": 7.937789258422129,
"grad_norm": 0.8152474164962769,
"learning_rate": 0.00012405344251697014,
"loss": 3.2421,
"step": 73750
},
{
"epoch": 7.943170810461737,
"grad_norm": 0.786149263381958,
"learning_rate": 0.00012373020148690874,
"loss": 3.2335,
"step": 73800
},
{
"epoch": 7.948552362501346,
"grad_norm": 0.8432660102844238,
"learning_rate": 0.0001234069604568473,
"loss": 3.2521,
"step": 73850
},
{
"epoch": 7.953933914540953,
"grad_norm": 0.8004118204116821,
"learning_rate": 0.0001230837194267859,
"loss": 3.2383,
"step": 73900
},
{
"epoch": 7.959315466580562,
"grad_norm": 0.7988388538360596,
"learning_rate": 0.00012276047839672447,
"loss": 3.2287,
"step": 73950
},
{
"epoch": 7.96469701862017,
"grad_norm": 0.7863640189170837,
"learning_rate": 0.00012243723736666306,
"loss": 3.212,
"step": 74000
},
{
"epoch": 7.96469701862017,
"eval_accuracy": 0.39020229103306187,
"eval_loss": 3.332491636276245,
"eval_runtime": 180.1547,
"eval_samples_per_second": 99.975,
"eval_steps_per_second": 6.25,
"step": 74000
},
{
"epoch": 7.970078570659778,
"grad_norm": 0.7968112826347351,
"learning_rate": 0.00012211399633660166,
"loss": 3.2303,
"step": 74050
},
{
"epoch": 7.975460122699387,
"grad_norm": 0.7782838344573975,
"learning_rate": 0.00012179075530654022,
"loss": 3.2324,
"step": 74100
},
{
"epoch": 7.980841674738995,
"grad_norm": 0.8536900877952576,
"learning_rate": 0.00012146751427647882,
"loss": 3.2439,
"step": 74150
},
{
"epoch": 7.986223226778603,
"grad_norm": 0.8339260220527649,
"learning_rate": 0.0001211442732464174,
"loss": 3.2338,
"step": 74200
},
{
"epoch": 7.991604778818211,
"grad_norm": 0.8594132661819458,
"learning_rate": 0.00012082103221635598,
"loss": 3.2435,
"step": 74250
},
{
"epoch": 7.996986330857819,
"grad_norm": 0.8051967620849609,
"learning_rate": 0.00012049779118629456,
"loss": 3.2442,
"step": 74300
},
{
"epoch": 8.002367882897428,
"grad_norm": 0.8067830801010132,
"learning_rate": 0.00012017455015623316,
"loss": 3.1685,
"step": 74350
},
{
"epoch": 8.007749434937036,
"grad_norm": 0.8121262788772583,
"learning_rate": 0.00011985130912617175,
"loss": 3.1723,
"step": 74400
},
{
"epoch": 8.013130986976645,
"grad_norm": 0.8569990992546082,
"learning_rate": 0.00011952806809611032,
"loss": 3.1684,
"step": 74450
},
{
"epoch": 8.018512539016251,
"grad_norm": 0.8412384986877441,
"learning_rate": 0.00011920482706604891,
"loss": 3.1601,
"step": 74500
},
{
"epoch": 8.02389409105586,
"grad_norm": 0.8731394410133362,
"learning_rate": 0.0001188815860359875,
"loss": 3.1824,
"step": 74550
},
{
"epoch": 8.029275643095469,
"grad_norm": 0.8120042681694031,
"learning_rate": 0.00011855834500592608,
"loss": 3.1549,
"step": 74600
},
{
"epoch": 8.034657195135077,
"grad_norm": 0.8084120750427246,
"learning_rate": 0.00011823510397586466,
"loss": 3.1696,
"step": 74650
},
{
"epoch": 8.040038747174686,
"grad_norm": 0.8044456839561462,
"learning_rate": 0.00011791186294580325,
"loss": 3.1584,
"step": 74700
},
{
"epoch": 8.045420299214294,
"grad_norm": 0.8172317147254944,
"learning_rate": 0.00011758862191574182,
"loss": 3.1491,
"step": 74750
},
{
"epoch": 8.050801851253901,
"grad_norm": 0.8528711795806885,
"learning_rate": 0.00011726538088568041,
"loss": 3.1622,
"step": 74800
},
{
"epoch": 8.05618340329351,
"grad_norm": 0.7957918643951416,
"learning_rate": 0.000116942139855619,
"loss": 3.1509,
"step": 74850
},
{
"epoch": 8.061564955333118,
"grad_norm": 0.8333423137664795,
"learning_rate": 0.00011661889882555759,
"loss": 3.1685,
"step": 74900
},
{
"epoch": 8.066946507372727,
"grad_norm": 0.8132439851760864,
"learning_rate": 0.00011629565779549616,
"loss": 3.1513,
"step": 74950
},
{
"epoch": 8.072328059412335,
"grad_norm": 0.7772408127784729,
"learning_rate": 0.00011597241676543475,
"loss": 3.1428,
"step": 75000
},
{
"epoch": 8.072328059412335,
"eval_accuracy": 0.38983102411539977,
"eval_loss": 3.3393983840942383,
"eval_runtime": 179.925,
"eval_samples_per_second": 100.103,
"eval_steps_per_second": 6.258,
"step": 75000
},
{
"epoch": 8.077709611451942,
"grad_norm": 0.8047040700912476,
"learning_rate": 0.00011564917573537335,
"loss": 3.1564,
"step": 75050
},
{
"epoch": 8.08309116349155,
"grad_norm": 0.8130930662155151,
"learning_rate": 0.00011532593470531191,
"loss": 3.1592,
"step": 75100
},
{
"epoch": 8.088472715531159,
"grad_norm": 0.7982088923454285,
"learning_rate": 0.0001150026936752505,
"loss": 3.1624,
"step": 75150
},
{
"epoch": 8.093854267570768,
"grad_norm": 0.8022302389144897,
"learning_rate": 0.00011467945264518909,
"loss": 3.1626,
"step": 75200
},
{
"epoch": 8.099235819610376,
"grad_norm": 0.7814152240753174,
"learning_rate": 0.00011435621161512766,
"loss": 3.1724,
"step": 75250
},
{
"epoch": 8.104617371649983,
"grad_norm": 0.8569167852401733,
"learning_rate": 0.00011403297058506625,
"loss": 3.1842,
"step": 75300
},
{
"epoch": 8.109998923689592,
"grad_norm": 0.8122818470001221,
"learning_rate": 0.00011370972955500485,
"loss": 3.163,
"step": 75350
},
{
"epoch": 8.1153804757292,
"grad_norm": 0.7728685736656189,
"learning_rate": 0.00011338648852494343,
"loss": 3.1683,
"step": 75400
},
{
"epoch": 8.120762027768809,
"grad_norm": 0.8322961330413818,
"learning_rate": 0.00011306324749488201,
"loss": 3.1758,
"step": 75450
},
{
"epoch": 8.126143579808417,
"grad_norm": 0.8351318836212158,
"learning_rate": 0.00011274000646482059,
"loss": 3.1767,
"step": 75500
},
{
"epoch": 8.131525131848026,
"grad_norm": 0.8136656880378723,
"learning_rate": 0.00011241676543475918,
"loss": 3.1513,
"step": 75550
},
{
"epoch": 8.136906683887632,
"grad_norm": 0.8244020342826843,
"learning_rate": 0.00011209352440469775,
"loss": 3.1645,
"step": 75600
},
{
"epoch": 8.142288235927241,
"grad_norm": 0.8306747078895569,
"learning_rate": 0.00011177028337463635,
"loss": 3.161,
"step": 75650
},
{
"epoch": 8.14766978796685,
"grad_norm": 0.7907711267471313,
"learning_rate": 0.00011144704234457493,
"loss": 3.1824,
"step": 75700
},
{
"epoch": 8.153051340006458,
"grad_norm": 0.8126032948493958,
"learning_rate": 0.00011113026613511473,
"loss": 3.161,
"step": 75750
},
{
"epoch": 8.158432892046067,
"grad_norm": 0.8425970077514648,
"learning_rate": 0.00011080702510505333,
"loss": 3.1704,
"step": 75800
},
{
"epoch": 8.163814444085673,
"grad_norm": 0.8157441020011902,
"learning_rate": 0.00011048378407499191,
"loss": 3.1555,
"step": 75850
},
{
"epoch": 8.169195996125282,
"grad_norm": 0.8338661789894104,
"learning_rate": 0.00011016054304493049,
"loss": 3.1604,
"step": 75900
},
{
"epoch": 8.17457754816489,
"grad_norm": 0.797713041305542,
"learning_rate": 0.00010983730201486907,
"loss": 3.1651,
"step": 75950
},
{
"epoch": 8.1799591002045,
"grad_norm": 0.8311964869499207,
"learning_rate": 0.00010951406098480767,
"loss": 3.1588,
"step": 76000
},
{
"epoch": 8.1799591002045,
"eval_accuracy": 0.3903796125459751,
"eval_loss": 3.3365893363952637,
"eval_runtime": 180.0247,
"eval_samples_per_second": 100.047,
"eval_steps_per_second": 6.255,
"step": 76000
},
{
"epoch": 8.185340652244108,
"grad_norm": 0.8243706226348877,
"learning_rate": 0.00010919081995474623,
"loss": 3.1717,
"step": 76050
},
{
"epoch": 8.190722204283716,
"grad_norm": 0.8235000371932983,
"learning_rate": 0.00010886757892468483,
"loss": 3.1711,
"step": 76100
},
{
"epoch": 8.196103756323323,
"grad_norm": 0.8117742538452148,
"learning_rate": 0.00010854433789462342,
"loss": 3.1864,
"step": 76150
},
{
"epoch": 8.201485308362932,
"grad_norm": 0.820876955986023,
"learning_rate": 0.000108221096864562,
"loss": 3.1615,
"step": 76200
},
{
"epoch": 8.20686686040254,
"grad_norm": 0.8476978540420532,
"learning_rate": 0.00010789785583450059,
"loss": 3.1629,
"step": 76250
},
{
"epoch": 8.212248412442149,
"grad_norm": 0.8074275851249695,
"learning_rate": 0.00010757461480443917,
"loss": 3.1628,
"step": 76300
},
{
"epoch": 8.217629964481757,
"grad_norm": 0.8188003897666931,
"learning_rate": 0.00010725137377437776,
"loss": 3.1659,
"step": 76350
},
{
"epoch": 8.223011516521364,
"grad_norm": 0.8572345972061157,
"learning_rate": 0.00010692813274431633,
"loss": 3.1835,
"step": 76400
},
{
"epoch": 8.228393068560973,
"grad_norm": 0.8235215544700623,
"learning_rate": 0.00010660489171425492,
"loss": 3.1772,
"step": 76450
},
{
"epoch": 8.233774620600581,
"grad_norm": 0.767663836479187,
"learning_rate": 0.0001062816506841935,
"loss": 3.1676,
"step": 76500
},
{
"epoch": 8.23915617264019,
"grad_norm": 0.8580052852630615,
"learning_rate": 0.00010595840965413209,
"loss": 3.192,
"step": 76550
},
{
"epoch": 8.244537724679798,
"grad_norm": 0.8200854063034058,
"learning_rate": 0.00010563516862407067,
"loss": 3.1725,
"step": 76600
},
{
"epoch": 8.249919276719407,
"grad_norm": 0.8221836686134338,
"learning_rate": 0.00010531192759400926,
"loss": 3.1761,
"step": 76650
},
{
"epoch": 8.255300828759013,
"grad_norm": 0.8396674394607544,
"learning_rate": 0.00010498868656394784,
"loss": 3.1504,
"step": 76700
},
{
"epoch": 8.260682380798622,
"grad_norm": 0.8755212426185608,
"learning_rate": 0.00010466544553388642,
"loss": 3.1605,
"step": 76750
},
{
"epoch": 8.26606393283823,
"grad_norm": 0.7833921909332275,
"learning_rate": 0.000104342204503825,
"loss": 3.1879,
"step": 76800
},
{
"epoch": 8.27144548487784,
"grad_norm": 0.8280619382858276,
"learning_rate": 0.0001040189634737636,
"loss": 3.1607,
"step": 76850
},
{
"epoch": 8.276827036917448,
"grad_norm": 0.8352559804916382,
"learning_rate": 0.00010369572244370217,
"loss": 3.1848,
"step": 76900
},
{
"epoch": 8.282208588957054,
"grad_norm": 0.8176780343055725,
"learning_rate": 0.00010337248141364076,
"loss": 3.1873,
"step": 76950
},
{
"epoch": 8.287590140996663,
"grad_norm": 0.868302583694458,
"learning_rate": 0.00010304924038357936,
"loss": 3.1633,
"step": 77000
},
{
"epoch": 8.287590140996663,
"eval_accuracy": 0.39057431852093866,
"eval_loss": 3.3340141773223877,
"eval_runtime": 179.9597,
"eval_samples_per_second": 100.084,
"eval_steps_per_second": 6.257,
"step": 77000
},
{
"epoch": 8.292971693036272,
"grad_norm": 0.8199362754821777,
"learning_rate": 0.00010272599935351792,
"loss": 3.1816,
"step": 77050
},
{
"epoch": 8.29835324507588,
"grad_norm": 0.8597607612609863,
"learning_rate": 0.00010240275832345652,
"loss": 3.182,
"step": 77100
},
{
"epoch": 8.303734797115489,
"grad_norm": 0.8568953275680542,
"learning_rate": 0.0001020795172933951,
"loss": 3.1675,
"step": 77150
},
{
"epoch": 8.309116349155097,
"grad_norm": 0.8705286979675293,
"learning_rate": 0.0001017562762633337,
"loss": 3.1871,
"step": 77200
},
{
"epoch": 8.314497901194704,
"grad_norm": 0.8231798410415649,
"learning_rate": 0.00010143303523327226,
"loss": 3.1839,
"step": 77250
},
{
"epoch": 8.319879453234313,
"grad_norm": 0.8620304465293884,
"learning_rate": 0.00010110979420321086,
"loss": 3.1756,
"step": 77300
},
{
"epoch": 8.325261005273921,
"grad_norm": 0.8248611688613892,
"learning_rate": 0.00010078655317314944,
"loss": 3.1995,
"step": 77350
},
{
"epoch": 8.33064255731353,
"grad_norm": 0.8094890117645264,
"learning_rate": 0.00010046331214308802,
"loss": 3.1779,
"step": 77400
},
{
"epoch": 8.336024109353138,
"grad_norm": 0.844918966293335,
"learning_rate": 0.0001001400711130266,
"loss": 3.1811,
"step": 77450
},
{
"epoch": 8.341405661392745,
"grad_norm": 0.8428331017494202,
"learning_rate": 9.98168300829652e-05,
"loss": 3.1706,
"step": 77500
},
{
"epoch": 8.346787213432354,
"grad_norm": 0.8277503252029419,
"learning_rate": 9.949358905290376e-05,
"loss": 3.1794,
"step": 77550
},
{
"epoch": 8.352168765471962,
"grad_norm": 0.870583713054657,
"learning_rate": 9.917034802284236e-05,
"loss": 3.1925,
"step": 77600
},
{
"epoch": 8.35755031751157,
"grad_norm": 0.8874440789222717,
"learning_rate": 9.884710699278094e-05,
"loss": 3.1897,
"step": 77650
},
{
"epoch": 8.36293186955118,
"grad_norm": 0.8165091276168823,
"learning_rate": 9.852386596271953e-05,
"loss": 3.1606,
"step": 77700
},
{
"epoch": 8.368313421590786,
"grad_norm": 0.8494898080825806,
"learning_rate": 9.82006249326581e-05,
"loss": 3.189,
"step": 77750
},
{
"epoch": 8.373694973630395,
"grad_norm": 0.8409062027931213,
"learning_rate": 9.78773839025967e-05,
"loss": 3.1569,
"step": 77800
},
{
"epoch": 8.379076525670003,
"grad_norm": 0.8248782753944397,
"learning_rate": 9.755414287253529e-05,
"loss": 3.1796,
"step": 77850
},
{
"epoch": 8.384458077709612,
"grad_norm": 0.8239468932151794,
"learning_rate": 9.723736666307508e-05,
"loss": 3.1884,
"step": 77900
},
{
"epoch": 8.38983962974922,
"grad_norm": 0.8291215896606445,
"learning_rate": 9.691412563301368e-05,
"loss": 3.1874,
"step": 77950
},
{
"epoch": 8.395221181788829,
"grad_norm": 0.8552483916282654,
"learning_rate": 9.659088460295227e-05,
"loss": 3.1619,
"step": 78000
},
{
"epoch": 8.395221181788829,
"eval_accuracy": 0.3911120416627325,
"eval_loss": 3.330475330352783,
"eval_runtime": 179.7478,
"eval_samples_per_second": 100.201,
"eval_steps_per_second": 6.264,
"step": 78000
},
{
"epoch": 8.400602733828435,
"grad_norm": 0.8288816213607788,
"learning_rate": 9.626764357289084e-05,
"loss": 3.1793,
"step": 78050
},
{
"epoch": 8.405984285868044,
"grad_norm": 0.8704893589019775,
"learning_rate": 9.594440254282943e-05,
"loss": 3.1677,
"step": 78100
},
{
"epoch": 8.411365837907653,
"grad_norm": 0.8380488157272339,
"learning_rate": 9.562116151276802e-05,
"loss": 3.1952,
"step": 78150
},
{
"epoch": 8.416747389947261,
"grad_norm": 0.8037657141685486,
"learning_rate": 9.52979204827066e-05,
"loss": 3.1903,
"step": 78200
},
{
"epoch": 8.42212894198687,
"grad_norm": 0.8423017263412476,
"learning_rate": 9.497467945264518e-05,
"loss": 3.1957,
"step": 78250
},
{
"epoch": 8.427510494026476,
"grad_norm": 0.7998765707015991,
"learning_rate": 9.465143842258377e-05,
"loss": 3.2026,
"step": 78300
},
{
"epoch": 8.432892046066085,
"grad_norm": 0.8594872355461121,
"learning_rate": 9.432819739252234e-05,
"loss": 3.174,
"step": 78350
},
{
"epoch": 8.438273598105694,
"grad_norm": 0.8684812784194946,
"learning_rate": 9.400495636246093e-05,
"loss": 3.1623,
"step": 78400
},
{
"epoch": 8.443655150145302,
"grad_norm": 0.8077685236930847,
"learning_rate": 9.368171533239952e-05,
"loss": 3.1733,
"step": 78450
},
{
"epoch": 8.44903670218491,
"grad_norm": 0.847590982913971,
"learning_rate": 9.335847430233811e-05,
"loss": 3.1807,
"step": 78500
},
{
"epoch": 8.45441825422452,
"grad_norm": 0.866955041885376,
"learning_rate": 9.303523327227668e-05,
"loss": 3.1811,
"step": 78550
},
{
"epoch": 8.459799806264126,
"grad_norm": 0.8266087174415588,
"learning_rate": 9.271199224221527e-05,
"loss": 3.1786,
"step": 78600
},
{
"epoch": 8.465181358303735,
"grad_norm": 0.8498027920722961,
"learning_rate": 9.238875121215387e-05,
"loss": 3.1924,
"step": 78650
},
{
"epoch": 8.470562910343343,
"grad_norm": 0.9029237627983093,
"learning_rate": 9.206551018209243e-05,
"loss": 3.192,
"step": 78700
},
{
"epoch": 8.475944462382952,
"grad_norm": 0.8091332912445068,
"learning_rate": 9.174226915203103e-05,
"loss": 3.1902,
"step": 78750
},
{
"epoch": 8.48132601442256,
"grad_norm": 0.8806012868881226,
"learning_rate": 9.141902812196961e-05,
"loss": 3.1834,
"step": 78800
},
{
"epoch": 8.486707566462167,
"grad_norm": 0.8003334403038025,
"learning_rate": 9.109578709190818e-05,
"loss": 3.1967,
"step": 78850
},
{
"epoch": 8.492089118501776,
"grad_norm": 0.8782911896705627,
"learning_rate": 9.077254606184677e-05,
"loss": 3.1777,
"step": 78900
},
{
"epoch": 8.497470670541384,
"grad_norm": 0.834248960018158,
"learning_rate": 9.044930503178537e-05,
"loss": 3.1673,
"step": 78950
},
{
"epoch": 8.502852222580993,
"grad_norm": 0.8210733532905579,
"learning_rate": 9.012606400172395e-05,
"loss": 3.1821,
"step": 79000
},
{
"epoch": 8.502852222580993,
"eval_accuracy": 0.39133499738852784,
"eval_loss": 3.3288872241973877,
"eval_runtime": 179.8126,
"eval_samples_per_second": 100.165,
"eval_steps_per_second": 6.262,
"step": 79000
},
{
"epoch": 8.508233774620601,
"grad_norm": 0.8497635126113892,
"learning_rate": 8.980282297166253e-05,
"loss": 3.1702,
"step": 79050
},
{
"epoch": 8.513615326660208,
"grad_norm": 0.8306612372398376,
"learning_rate": 8.947958194160111e-05,
"loss": 3.1834,
"step": 79100
},
{
"epoch": 8.518996878699816,
"grad_norm": 0.8895464539527893,
"learning_rate": 8.91563409115397e-05,
"loss": 3.1958,
"step": 79150
},
{
"epoch": 8.524378430739425,
"grad_norm": 0.8079034686088562,
"learning_rate": 8.883309988147827e-05,
"loss": 3.1914,
"step": 79200
},
{
"epoch": 8.529759982779034,
"grad_norm": 0.8202755451202393,
"learning_rate": 8.850985885141687e-05,
"loss": 3.1948,
"step": 79250
},
{
"epoch": 8.535141534818642,
"grad_norm": 0.8289138674736023,
"learning_rate": 8.818661782135545e-05,
"loss": 3.1863,
"step": 79300
},
{
"epoch": 8.54052308685825,
"grad_norm": 0.8435153961181641,
"learning_rate": 8.786337679129403e-05,
"loss": 3.1997,
"step": 79350
},
{
"epoch": 8.545904638897857,
"grad_norm": 0.9007829427719116,
"learning_rate": 8.754013576123261e-05,
"loss": 3.1776,
"step": 79400
},
{
"epoch": 8.551286190937466,
"grad_norm": 0.8751116394996643,
"learning_rate": 8.72168947311712e-05,
"loss": 3.1819,
"step": 79450
},
{
"epoch": 8.556667742977075,
"grad_norm": 0.8555277585983276,
"learning_rate": 8.68936537011098e-05,
"loss": 3.1839,
"step": 79500
},
{
"epoch": 8.562049295016683,
"grad_norm": 0.8604739308357239,
"learning_rate": 8.657041267104837e-05,
"loss": 3.1696,
"step": 79550
},
{
"epoch": 8.567430847056292,
"grad_norm": 0.8286479711532593,
"learning_rate": 8.624717164098696e-05,
"loss": 3.1646,
"step": 79600
},
{
"epoch": 8.572812399095898,
"grad_norm": 0.826781690120697,
"learning_rate": 8.592393061092554e-05,
"loss": 3.1811,
"step": 79650
},
{
"epoch": 8.578193951135507,
"grad_norm": 0.8096086978912354,
"learning_rate": 8.560068958086412e-05,
"loss": 3.1804,
"step": 79700
},
{
"epoch": 8.583575503175116,
"grad_norm": 0.8319725394248962,
"learning_rate": 8.528391337140394e-05,
"loss": 3.204,
"step": 79750
},
{
"epoch": 8.588957055214724,
"grad_norm": 0.8044587969779968,
"learning_rate": 8.496067234134253e-05,
"loss": 3.1926,
"step": 79800
},
{
"epoch": 8.594338607254333,
"grad_norm": 0.7995679974555969,
"learning_rate": 8.46374313112811e-05,
"loss": 3.1776,
"step": 79850
},
{
"epoch": 8.599720159293941,
"grad_norm": 0.8578434586524963,
"learning_rate": 8.431419028121969e-05,
"loss": 3.1978,
"step": 79900
},
{
"epoch": 8.605101711333548,
"grad_norm": 0.8371725678443909,
"learning_rate": 8.399094925115828e-05,
"loss": 3.1848,
"step": 79950
},
{
"epoch": 8.610483263373157,
"grad_norm": 0.856256902217865,
"learning_rate": 8.366770822109685e-05,
"loss": 3.1881,
"step": 80000
},
{
"epoch": 8.610483263373157,
"eval_accuracy": 0.3915523031641568,
"eval_loss": 3.3257627487182617,
"eval_runtime": 180.2073,
"eval_samples_per_second": 99.946,
"eval_steps_per_second": 6.248,
"step": 80000
},
{
"epoch": 8.615864815412765,
"grad_norm": 0.8404597640037537,
"learning_rate": 8.334446719103544e-05,
"loss": 3.1712,
"step": 80050
},
{
"epoch": 8.621246367452374,
"grad_norm": 0.8627851009368896,
"learning_rate": 8.302122616097403e-05,
"loss": 3.1643,
"step": 80100
},
{
"epoch": 8.626627919491982,
"grad_norm": 0.8004584312438965,
"learning_rate": 8.269798513091261e-05,
"loss": 3.16,
"step": 80150
},
{
"epoch": 8.632009471531589,
"grad_norm": 0.8076578974723816,
"learning_rate": 8.237474410085119e-05,
"loss": 3.1925,
"step": 80200
},
{
"epoch": 8.637391023571197,
"grad_norm": 0.8029952645301819,
"learning_rate": 8.205150307078978e-05,
"loss": 3.2021,
"step": 80250
},
{
"epoch": 8.642772575610806,
"grad_norm": 0.8227396607398987,
"learning_rate": 8.172826204072838e-05,
"loss": 3.1918,
"step": 80300
},
{
"epoch": 8.648154127650415,
"grad_norm": 0.8576304912567139,
"learning_rate": 8.140502101066694e-05,
"loss": 3.1858,
"step": 80350
},
{
"epoch": 8.653535679690023,
"grad_norm": 0.802401602268219,
"learning_rate": 8.108177998060553e-05,
"loss": 3.1806,
"step": 80400
},
{
"epoch": 8.658917231729632,
"grad_norm": 0.9308738112449646,
"learning_rate": 8.075853895054412e-05,
"loss": 3.1773,
"step": 80450
},
{
"epoch": 8.664298783769238,
"grad_norm": 0.8083286881446838,
"learning_rate": 8.043529792048269e-05,
"loss": 3.1862,
"step": 80500
},
{
"epoch": 8.669680335808847,
"grad_norm": 0.8586329221725464,
"learning_rate": 8.011205689042128e-05,
"loss": 3.1954,
"step": 80550
},
{
"epoch": 8.675061887848456,
"grad_norm": 0.8661782741546631,
"learning_rate": 7.978881586035988e-05,
"loss": 3.1814,
"step": 80600
},
{
"epoch": 8.680443439888064,
"grad_norm": 0.8471584916114807,
"learning_rate": 7.946557483029845e-05,
"loss": 3.1782,
"step": 80650
},
{
"epoch": 8.685824991927673,
"grad_norm": 0.8519724011421204,
"learning_rate": 7.914233380023704e-05,
"loss": 3.1871,
"step": 80700
},
{
"epoch": 8.69120654396728,
"grad_norm": 0.883844792842865,
"learning_rate": 7.881909277017562e-05,
"loss": 3.2014,
"step": 80750
},
{
"epoch": 8.696588096006888,
"grad_norm": 0.8341808915138245,
"learning_rate": 7.849585174011422e-05,
"loss": 3.1888,
"step": 80800
},
{
"epoch": 8.701969648046497,
"grad_norm": 0.8479406833648682,
"learning_rate": 7.817261071005278e-05,
"loss": 3.1704,
"step": 80850
},
{
"epoch": 8.707351200086105,
"grad_norm": 0.8445674777030945,
"learning_rate": 7.784936967999138e-05,
"loss": 3.1814,
"step": 80900
},
{
"epoch": 8.712732752125714,
"grad_norm": 0.876797616481781,
"learning_rate": 7.752612864992996e-05,
"loss": 3.1678,
"step": 80950
},
{
"epoch": 8.718114304165322,
"grad_norm": 0.8637148141860962,
"learning_rate": 7.720288761986854e-05,
"loss": 3.1781,
"step": 81000
},
{
"epoch": 8.718114304165322,
"eval_accuracy": 0.3921533709395465,
"eval_loss": 3.320990800857544,
"eval_runtime": 179.8841,
"eval_samples_per_second": 100.126,
"eval_steps_per_second": 6.26,
"step": 81000
},
{
"epoch": 8.723495856204929,
"grad_norm": 0.8150270581245422,
"learning_rate": 7.687964658980712e-05,
"loss": 3.1736,
"step": 81050
},
{
"epoch": 8.728877408244538,
"grad_norm": 0.8342413306236267,
"learning_rate": 7.655640555974572e-05,
"loss": 3.1732,
"step": 81100
},
{
"epoch": 8.734258960284146,
"grad_norm": 0.8617803454399109,
"learning_rate": 7.623316452968428e-05,
"loss": 3.1916,
"step": 81150
},
{
"epoch": 8.739640512323755,
"grad_norm": 0.8338009119033813,
"learning_rate": 7.590992349962288e-05,
"loss": 3.1914,
"step": 81200
},
{
"epoch": 8.745022064363363,
"grad_norm": 0.8587698340415955,
"learning_rate": 7.558668246956147e-05,
"loss": 3.1983,
"step": 81250
},
{
"epoch": 8.75040361640297,
"grad_norm": 0.8290186524391174,
"learning_rate": 7.526344143950005e-05,
"loss": 3.1846,
"step": 81300
},
{
"epoch": 8.755785168442578,
"grad_norm": 0.8477612137794495,
"learning_rate": 7.494020040943862e-05,
"loss": 3.1712,
"step": 81350
},
{
"epoch": 8.761166720482187,
"grad_norm": 0.8442912101745605,
"learning_rate": 7.461695937937722e-05,
"loss": 3.1715,
"step": 81400
},
{
"epoch": 8.766548272521796,
"grad_norm": 0.8489503264427185,
"learning_rate": 7.42937183493158e-05,
"loss": 3.1831,
"step": 81450
},
{
"epoch": 8.771929824561404,
"grad_norm": 0.8153248429298401,
"learning_rate": 7.397047731925439e-05,
"loss": 3.1897,
"step": 81500
},
{
"epoch": 8.777311376601011,
"grad_norm": 0.8930414319038391,
"learning_rate": 7.364723628919297e-05,
"loss": 3.1778,
"step": 81550
},
{
"epoch": 8.78269292864062,
"grad_norm": 0.845933198928833,
"learning_rate": 7.332399525913155e-05,
"loss": 3.1975,
"step": 81600
},
{
"epoch": 8.788074480680228,
"grad_norm": 0.8592908382415771,
"learning_rate": 7.300075422907013e-05,
"loss": 3.1731,
"step": 81650
},
{
"epoch": 8.793456032719837,
"grad_norm": 0.8641341924667358,
"learning_rate": 7.267751319900872e-05,
"loss": 3.1821,
"step": 81700
},
{
"epoch": 8.798837584759445,
"grad_norm": 0.853470504283905,
"learning_rate": 7.235427216894731e-05,
"loss": 3.1721,
"step": 81750
},
{
"epoch": 8.804219136799054,
"grad_norm": 0.8515709042549133,
"learning_rate": 7.203103113888589e-05,
"loss": 3.1734,
"step": 81800
},
{
"epoch": 8.80960068883866,
"grad_norm": 0.802684485912323,
"learning_rate": 7.170779010882447e-05,
"loss": 3.1746,
"step": 81850
},
{
"epoch": 8.814982240878269,
"grad_norm": 0.8838043212890625,
"learning_rate": 7.138454907876305e-05,
"loss": 3.1732,
"step": 81900
},
{
"epoch": 8.820363792917878,
"grad_norm": 0.8429666757583618,
"learning_rate": 7.106130804870164e-05,
"loss": 3.1743,
"step": 81950
},
{
"epoch": 8.825745344957486,
"grad_norm": 0.8476669788360596,
"learning_rate": 7.073806701864023e-05,
"loss": 3.1834,
"step": 82000
},
{
"epoch": 8.825745344957486,
"eval_accuracy": 0.39253778485663415,
"eval_loss": 3.3173086643218994,
"eval_runtime": 180.2273,
"eval_samples_per_second": 99.935,
"eval_steps_per_second": 6.248,
"step": 82000
},
{
"epoch": 8.831126896997095,
"grad_norm": 0.8718939423561096,
"learning_rate": 7.041482598857881e-05,
"loss": 3.1812,
"step": 82050
},
{
"epoch": 8.836508449036701,
"grad_norm": 0.8387258648872375,
"learning_rate": 7.009158495851739e-05,
"loss": 3.1846,
"step": 82100
},
{
"epoch": 8.84189000107631,
"grad_norm": 0.8789180517196655,
"learning_rate": 6.976834392845599e-05,
"loss": 3.178,
"step": 82150
},
{
"epoch": 8.847271553115919,
"grad_norm": 0.8595734238624573,
"learning_rate": 6.944510289839457e-05,
"loss": 3.1809,
"step": 82200
},
{
"epoch": 8.852653105155527,
"grad_norm": 0.8221034407615662,
"learning_rate": 6.912186186833315e-05,
"loss": 3.1896,
"step": 82250
},
{
"epoch": 8.858034657195136,
"grad_norm": 0.8374698758125305,
"learning_rate": 6.879862083827173e-05,
"loss": 3.1773,
"step": 82300
},
{
"epoch": 8.863416209234742,
"grad_norm": 0.8906286358833313,
"learning_rate": 6.847537980821031e-05,
"loss": 3.1929,
"step": 82350
},
{
"epoch": 8.868797761274351,
"grad_norm": 0.8791503310203552,
"learning_rate": 6.81521387781489e-05,
"loss": 3.1769,
"step": 82400
},
{
"epoch": 8.87417931331396,
"grad_norm": 0.8439499735832214,
"learning_rate": 6.782889774808749e-05,
"loss": 3.1919,
"step": 82450
},
{
"epoch": 8.879560865353568,
"grad_norm": 0.8035529851913452,
"learning_rate": 6.750565671802607e-05,
"loss": 3.1862,
"step": 82500
},
{
"epoch": 8.884942417393177,
"grad_norm": 0.8311242461204529,
"learning_rate": 6.718241568796465e-05,
"loss": 3.192,
"step": 82550
},
{
"epoch": 8.890323969432785,
"grad_norm": 0.8353226780891418,
"learning_rate": 6.685917465790323e-05,
"loss": 3.1812,
"step": 82600
},
{
"epoch": 8.895705521472392,
"grad_norm": 0.8998394012451172,
"learning_rate": 6.653593362784182e-05,
"loss": 3.1869,
"step": 82650
},
{
"epoch": 8.901087073512,
"grad_norm": 0.8866690397262573,
"learning_rate": 6.62126925977804e-05,
"loss": 3.1857,
"step": 82700
},
{
"epoch": 8.906468625551609,
"grad_norm": 0.9431129097938538,
"learning_rate": 6.588945156771899e-05,
"loss": 3.1655,
"step": 82750
},
{
"epoch": 8.911850177591218,
"grad_norm": 0.8412008285522461,
"learning_rate": 6.556621053765757e-05,
"loss": 3.1803,
"step": 82800
},
{
"epoch": 8.917231729630826,
"grad_norm": 0.8523921966552734,
"learning_rate": 6.524296950759615e-05,
"loss": 3.1745,
"step": 82850
},
{
"epoch": 8.922613281670433,
"grad_norm": 0.8443444967269897,
"learning_rate": 6.491972847753474e-05,
"loss": 3.2073,
"step": 82900
},
{
"epoch": 8.927994833710041,
"grad_norm": 0.8373326063156128,
"learning_rate": 6.459648744747333e-05,
"loss": 3.1818,
"step": 82950
},
{
"epoch": 8.93337638574965,
"grad_norm": 0.8930068612098694,
"learning_rate": 6.427324641741192e-05,
"loss": 3.1769,
"step": 83000
},
{
"epoch": 8.93337638574965,
"eval_accuracy": 0.3927184746090696,
"eval_loss": 3.3146286010742188,
"eval_runtime": 180.16,
"eval_samples_per_second": 99.972,
"eval_steps_per_second": 6.25,
"step": 83000
},
{
"epoch": 8.938757937789259,
"grad_norm": 0.850643515586853,
"learning_rate": 6.39500053873505e-05,
"loss": 3.1891,
"step": 83050
},
{
"epoch": 8.944139489828867,
"grad_norm": 0.8928857445716858,
"learning_rate": 6.362676435728908e-05,
"loss": 3.1738,
"step": 83100
},
{
"epoch": 8.949521041868476,
"grad_norm": 0.8975667953491211,
"learning_rate": 6.330352332722766e-05,
"loss": 3.1725,
"step": 83150
},
{
"epoch": 8.954902593908082,
"grad_norm": 0.8075838088989258,
"learning_rate": 6.298028229716624e-05,
"loss": 3.1837,
"step": 83200
},
{
"epoch": 8.960284145947691,
"grad_norm": 0.8257818222045898,
"learning_rate": 6.265704126710484e-05,
"loss": 3.1772,
"step": 83250
},
{
"epoch": 8.9656656979873,
"grad_norm": 0.874556303024292,
"learning_rate": 6.233380023704342e-05,
"loss": 3.188,
"step": 83300
},
{
"epoch": 8.971047250026908,
"grad_norm": 0.9293429255485535,
"learning_rate": 6.2010559206982e-05,
"loss": 3.1731,
"step": 83350
},
{
"epoch": 8.976428802066517,
"grad_norm": 0.8618453741073608,
"learning_rate": 6.168731817692058e-05,
"loss": 3.1832,
"step": 83400
},
{
"epoch": 8.981810354106123,
"grad_norm": 0.846990168094635,
"learning_rate": 6.136407714685916e-05,
"loss": 3.1839,
"step": 83450
},
{
"epoch": 8.987191906145732,
"grad_norm": 0.8998017311096191,
"learning_rate": 6.104083611679776e-05,
"loss": 3.1788,
"step": 83500
},
{
"epoch": 8.99257345818534,
"grad_norm": 0.8273147940635681,
"learning_rate": 6.071759508673634e-05,
"loss": 3.1743,
"step": 83550
},
{
"epoch": 8.997955010224949,
"grad_norm": 0.8635324239730835,
"learning_rate": 6.039435405667492e-05,
"loss": 3.1996,
"step": 83600
},
{
"epoch": 9.003336562264558,
"grad_norm": 0.8398813605308533,
"learning_rate": 6.007111302661351e-05,
"loss": 3.1242,
"step": 83650
},
{
"epoch": 9.008718114304166,
"grad_norm": 0.858457088470459,
"learning_rate": 5.974787199655209e-05,
"loss": 3.1269,
"step": 83700
},
{
"epoch": 9.014099666343773,
"grad_norm": 0.8467790484428406,
"learning_rate": 5.942463096649068e-05,
"loss": 3.1305,
"step": 83750
},
{
"epoch": 9.019481218383381,
"grad_norm": 0.8237690925598145,
"learning_rate": 5.910138993642926e-05,
"loss": 3.1128,
"step": 83800
},
{
"epoch": 9.02486277042299,
"grad_norm": 0.9006997346878052,
"learning_rate": 5.878461372696907e-05,
"loss": 3.1345,
"step": 83850
},
{
"epoch": 9.030244322462599,
"grad_norm": 0.899467945098877,
"learning_rate": 5.846137269690765e-05,
"loss": 3.1302,
"step": 83900
},
{
"epoch": 9.035625874502207,
"grad_norm": 0.8287666440010071,
"learning_rate": 5.813813166684624e-05,
"loss": 3.1283,
"step": 83950
},
{
"epoch": 9.041007426541814,
"grad_norm": 0.8582010269165039,
"learning_rate": 5.781489063678482e-05,
"loss": 3.1148,
"step": 84000
},
{
"epoch": 9.041007426541814,
"eval_accuracy": 0.39264361276936544,
"eval_loss": 3.318713426589966,
"eval_runtime": 179.9421,
"eval_samples_per_second": 100.093,
"eval_steps_per_second": 6.258,
"step": 84000
},
{
"epoch": 9.046388978581422,
"grad_norm": 0.8633049726486206,
"learning_rate": 5.749164960672341e-05,
"loss": 3.1233,
"step": 84050
},
{
"epoch": 9.051770530621031,
"grad_norm": 0.8720656037330627,
"learning_rate": 5.716840857666199e-05,
"loss": 3.1223,
"step": 84100
},
{
"epoch": 9.05715208266064,
"grad_norm": 0.8847928047180176,
"learning_rate": 5.684516754660057e-05,
"loss": 3.1114,
"step": 84150
},
{
"epoch": 9.062533634700248,
"grad_norm": 0.8395546078681946,
"learning_rate": 5.652192651653916e-05,
"loss": 3.1071,
"step": 84200
},
{
"epoch": 9.067915186739857,
"grad_norm": 0.8343135118484497,
"learning_rate": 5.619868548647774e-05,
"loss": 3.1269,
"step": 84250
},
{
"epoch": 9.073296738779463,
"grad_norm": 0.8847300410270691,
"learning_rate": 5.5875444456416335e-05,
"loss": 3.1341,
"step": 84300
},
{
"epoch": 9.078678290819072,
"grad_norm": 0.857685387134552,
"learning_rate": 5.5552203426354916e-05,
"loss": 3.1307,
"step": 84350
},
{
"epoch": 9.08405984285868,
"grad_norm": 0.8166672587394714,
"learning_rate": 5.52289623962935e-05,
"loss": 3.1287,
"step": 84400
},
{
"epoch": 9.089441394898289,
"grad_norm": 0.843408465385437,
"learning_rate": 5.4905721366232085e-05,
"loss": 3.1302,
"step": 84450
},
{
"epoch": 9.094822946937898,
"grad_norm": 0.8354735970497131,
"learning_rate": 5.4582480336170666e-05,
"loss": 3.1264,
"step": 84500
},
{
"epoch": 9.100204498977504,
"grad_norm": 0.8715230226516724,
"learning_rate": 5.4259239306109254e-05,
"loss": 3.1349,
"step": 84550
},
{
"epoch": 9.105586051017113,
"grad_norm": 0.794308066368103,
"learning_rate": 5.3935998276047835e-05,
"loss": 3.1194,
"step": 84600
},
{
"epoch": 9.110967603056721,
"grad_norm": 0.8556493520736694,
"learning_rate": 5.3612757245986416e-05,
"loss": 3.1273,
"step": 84650
},
{
"epoch": 9.11634915509633,
"grad_norm": 0.8844385147094727,
"learning_rate": 5.3289516215925004e-05,
"loss": 3.111,
"step": 84700
},
{
"epoch": 9.121730707135939,
"grad_norm": 0.816949188709259,
"learning_rate": 5.2966275185863585e-05,
"loss": 3.1177,
"step": 84750
},
{
"epoch": 9.127112259175545,
"grad_norm": 0.8969637751579285,
"learning_rate": 5.264303415580217e-05,
"loss": 3.1527,
"step": 84800
},
{
"epoch": 9.132493811215154,
"grad_norm": 0.8661097884178162,
"learning_rate": 5.2319793125740754e-05,
"loss": 3.1297,
"step": 84850
},
{
"epoch": 9.137875363254762,
"grad_norm": 0.8353490233421326,
"learning_rate": 5.1996552095679336e-05,
"loss": 3.1286,
"step": 84900
},
{
"epoch": 9.143256915294371,
"grad_norm": 0.869312584400177,
"learning_rate": 5.1673311065617923e-05,
"loss": 3.1292,
"step": 84950
},
{
"epoch": 9.14863846733398,
"grad_norm": 0.8817721605300903,
"learning_rate": 5.1350070035556505e-05,
"loss": 3.101,
"step": 85000
},
{
"epoch": 9.14863846733398,
"eval_accuracy": 0.39279127204390535,
"eval_loss": 3.3183066844940186,
"eval_runtime": 180.0381,
"eval_samples_per_second": 100.04,
"eval_steps_per_second": 6.254,
"step": 85000
},
{
"epoch": 9.154020019373588,
"grad_norm": 0.8361867666244507,
"learning_rate": 5.102682900549509e-05,
"loss": 3.1389,
"step": 85050
},
{
"epoch": 9.159401571413195,
"grad_norm": 0.8956885933876038,
"learning_rate": 5.0703587975433674e-05,
"loss": 3.1193,
"step": 85100
},
{
"epoch": 9.164783123452803,
"grad_norm": 0.8367882370948792,
"learning_rate": 5.0380346945372255e-05,
"loss": 3.1208,
"step": 85150
},
{
"epoch": 9.170164675492412,
"grad_norm": 0.8613442778587341,
"learning_rate": 5.005710591531085e-05,
"loss": 3.1245,
"step": 85200
},
{
"epoch": 9.17554622753202,
"grad_norm": 0.8361165523529053,
"learning_rate": 4.973386488524943e-05,
"loss": 3.1413,
"step": 85250
},
{
"epoch": 9.180927779571629,
"grad_norm": 0.8090612292289734,
"learning_rate": 4.941062385518802e-05,
"loss": 3.1308,
"step": 85300
},
{
"epoch": 9.186309331611236,
"grad_norm": 0.8353068828582764,
"learning_rate": 4.90873828251266e-05,
"loss": 3.1282,
"step": 85350
},
{
"epoch": 9.191690883650844,
"grad_norm": 0.8467201590538025,
"learning_rate": 4.876414179506518e-05,
"loss": 3.1291,
"step": 85400
},
{
"epoch": 9.197072435690453,
"grad_norm": 0.9284381866455078,
"learning_rate": 4.844090076500377e-05,
"loss": 3.138,
"step": 85450
},
{
"epoch": 9.202453987730062,
"grad_norm": 0.8707318902015686,
"learning_rate": 4.811765973494235e-05,
"loss": 3.1146,
"step": 85500
},
{
"epoch": 9.20783553976967,
"grad_norm": 0.8575342297554016,
"learning_rate": 4.779441870488094e-05,
"loss": 3.1414,
"step": 85550
},
{
"epoch": 9.213217091809279,
"grad_norm": 0.8592131733894348,
"learning_rate": 4.747117767481952e-05,
"loss": 3.1118,
"step": 85600
},
{
"epoch": 9.218598643848885,
"grad_norm": 0.868922770023346,
"learning_rate": 4.71479366447581e-05,
"loss": 3.1254,
"step": 85650
},
{
"epoch": 9.223980195888494,
"grad_norm": 0.8733725547790527,
"learning_rate": 4.682469561469669e-05,
"loss": 3.1187,
"step": 85700
},
{
"epoch": 9.229361747928102,
"grad_norm": 0.853115439414978,
"learning_rate": 4.650145458463527e-05,
"loss": 3.1316,
"step": 85750
},
{
"epoch": 9.234743299967711,
"grad_norm": 0.8624739050865173,
"learning_rate": 4.6178213554573856e-05,
"loss": 3.1467,
"step": 85800
},
{
"epoch": 9.24012485200732,
"grad_norm": 0.830511748790741,
"learning_rate": 4.585497252451244e-05,
"loss": 3.1283,
"step": 85850
},
{
"epoch": 9.245506404046926,
"grad_norm": 0.843909740447998,
"learning_rate": 4.553173149445102e-05,
"loss": 3.1194,
"step": 85900
},
{
"epoch": 9.250887956086535,
"grad_norm": 0.8719884753227234,
"learning_rate": 4.520849046438961e-05,
"loss": 3.1206,
"step": 85950
},
{
"epoch": 9.256269508126143,
"grad_norm": 0.8853352665901184,
"learning_rate": 4.488524943432819e-05,
"loss": 3.1249,
"step": 86000
},
{
"epoch": 9.256269508126143,
"eval_accuracy": 0.3932848821132465,
"eval_loss": 3.315089702606201,
"eval_runtime": 179.9252,
"eval_samples_per_second": 100.103,
"eval_steps_per_second": 6.258,
"step": 86000
},
{
"epoch": 9.261651060165752,
"grad_norm": 0.8370910286903381,
"learning_rate": 4.456200840426678e-05,
"loss": 3.1212,
"step": 86050
},
{
"epoch": 9.26703261220536,
"grad_norm": 0.8677020072937012,
"learning_rate": 4.4245232194806596e-05,
"loss": 3.1426,
"step": 86100
},
{
"epoch": 9.272414164244967,
"grad_norm": 0.8646278977394104,
"learning_rate": 4.392199116474518e-05,
"loss": 3.1316,
"step": 86150
},
{
"epoch": 9.277795716284576,
"grad_norm": 0.8564377427101135,
"learning_rate": 4.359875013468376e-05,
"loss": 3.1319,
"step": 86200
},
{
"epoch": 9.283177268324184,
"grad_norm": 0.8211710453033447,
"learning_rate": 4.3275509104622346e-05,
"loss": 3.1192,
"step": 86250
},
{
"epoch": 9.288558820363793,
"grad_norm": 0.8410226702690125,
"learning_rate": 4.295226807456093e-05,
"loss": 3.1324,
"step": 86300
},
{
"epoch": 9.293940372403402,
"grad_norm": 0.8313528299331665,
"learning_rate": 4.2629027044499515e-05,
"loss": 3.1334,
"step": 86350
},
{
"epoch": 9.29932192444301,
"grad_norm": 0.8200216889381409,
"learning_rate": 4.2305786014438096e-05,
"loss": 3.1349,
"step": 86400
},
{
"epoch": 9.304703476482617,
"grad_norm": 0.8536425232887268,
"learning_rate": 4.198254498437668e-05,
"loss": 3.1515,
"step": 86450
},
{
"epoch": 9.310085028522225,
"grad_norm": 0.8287447690963745,
"learning_rate": 4.1659303954315265e-05,
"loss": 3.1178,
"step": 86500
},
{
"epoch": 9.315466580561834,
"grad_norm": 0.8226425051689148,
"learning_rate": 4.1336062924253846e-05,
"loss": 3.1261,
"step": 86550
},
{
"epoch": 9.320848132601443,
"grad_norm": 0.8848347663879395,
"learning_rate": 4.1012821894192434e-05,
"loss": 3.1234,
"step": 86600
},
{
"epoch": 9.326229684641051,
"grad_norm": 0.8347705006599426,
"learning_rate": 4.0689580864131015e-05,
"loss": 3.1335,
"step": 86650
},
{
"epoch": 9.331611236680658,
"grad_norm": 0.8725020885467529,
"learning_rate": 4.0366339834069596e-05,
"loss": 3.1235,
"step": 86700
},
{
"epoch": 9.336992788720266,
"grad_norm": 0.834156334400177,
"learning_rate": 4.0043098804008184e-05,
"loss": 3.1187,
"step": 86750
},
{
"epoch": 9.342374340759875,
"grad_norm": 0.8515089154243469,
"learning_rate": 3.9719857773946765e-05,
"loss": 3.1172,
"step": 86800
},
{
"epoch": 9.347755892799483,
"grad_norm": 0.8820292949676514,
"learning_rate": 3.939661674388536e-05,
"loss": 3.1238,
"step": 86850
},
{
"epoch": 9.353137444839092,
"grad_norm": 0.8537640571594238,
"learning_rate": 3.9073375713823934e-05,
"loss": 3.1401,
"step": 86900
},
{
"epoch": 9.3585189968787,
"grad_norm": 0.900911808013916,
"learning_rate": 3.8750134683762515e-05,
"loss": 3.1283,
"step": 86950
},
{
"epoch": 9.363900548918307,
"grad_norm": 0.8385894298553467,
"learning_rate": 3.842689365370111e-05,
"loss": 3.1282,
"step": 87000
},
{
"epoch": 9.363900548918307,
"eval_accuracy": 0.3934482960565195,
"eval_loss": 3.3138954639434814,
"eval_runtime": 180.0796,
"eval_samples_per_second": 100.017,
"eval_steps_per_second": 6.253,
"step": 87000
},
{
"epoch": 9.369282100957916,
"grad_norm": 0.8399181365966797,
"learning_rate": 3.810365262363969e-05,
"loss": 3.1466,
"step": 87050
},
{
"epoch": 9.374663652997524,
"grad_norm": 0.8344743251800537,
"learning_rate": 3.778041159357828e-05,
"loss": 3.1362,
"step": 87100
},
{
"epoch": 9.380045205037133,
"grad_norm": 0.8556305170059204,
"learning_rate": 3.745717056351686e-05,
"loss": 3.1363,
"step": 87150
},
{
"epoch": 9.385426757076742,
"grad_norm": 0.8326317071914673,
"learning_rate": 3.713392953345544e-05,
"loss": 3.1301,
"step": 87200
},
{
"epoch": 9.390808309116348,
"grad_norm": 0.8645597100257874,
"learning_rate": 3.681068850339403e-05,
"loss": 3.13,
"step": 87250
},
{
"epoch": 9.396189861155957,
"grad_norm": 0.8261982798576355,
"learning_rate": 3.648744747333262e-05,
"loss": 3.1417,
"step": 87300
},
{
"epoch": 9.401571413195565,
"grad_norm": 0.8522182106971741,
"learning_rate": 3.61642064432712e-05,
"loss": 3.1177,
"step": 87350
},
{
"epoch": 9.406952965235174,
"grad_norm": 0.8481505513191223,
"learning_rate": 3.584096541320978e-05,
"loss": 3.1478,
"step": 87400
},
{
"epoch": 9.412334517274783,
"grad_norm": 0.8410167098045349,
"learning_rate": 3.551772438314837e-05,
"loss": 3.1244,
"step": 87450
},
{
"epoch": 9.417716069314391,
"grad_norm": 0.8503827452659607,
"learning_rate": 3.519448335308695e-05,
"loss": 3.1273,
"step": 87500
},
{
"epoch": 9.423097621353998,
"grad_norm": 0.8531398177146912,
"learning_rate": 3.4871242323025536e-05,
"loss": 3.1317,
"step": 87550
},
{
"epoch": 9.428479173393606,
"grad_norm": 0.8031463027000427,
"learning_rate": 3.454800129296412e-05,
"loss": 3.1272,
"step": 87600
},
{
"epoch": 9.433860725433215,
"grad_norm": 0.8540205955505371,
"learning_rate": 3.42247602629027e-05,
"loss": 3.137,
"step": 87650
},
{
"epoch": 9.439242277472824,
"grad_norm": 0.8364588022232056,
"learning_rate": 3.3901519232841286e-05,
"loss": 3.1339,
"step": 87700
},
{
"epoch": 9.444623829512432,
"grad_norm": 0.8611851930618286,
"learning_rate": 3.3578278202779874e-05,
"loss": 3.1277,
"step": 87750
},
{
"epoch": 9.450005381552039,
"grad_norm": 0.8790317177772522,
"learning_rate": 3.3255037172718455e-05,
"loss": 3.1342,
"step": 87800
},
{
"epoch": 9.455386933591647,
"grad_norm": 0.858058750629425,
"learning_rate": 3.2931796142657036e-05,
"loss": 3.1399,
"step": 87850
},
{
"epoch": 9.460768485631256,
"grad_norm": 0.8198880553245544,
"learning_rate": 3.2608555112595624e-05,
"loss": 3.1449,
"step": 87900
},
{
"epoch": 9.466150037670864,
"grad_norm": 0.8721171021461487,
"learning_rate": 3.2285314082534205e-05,
"loss": 3.1154,
"step": 87950
},
{
"epoch": 9.471531589710473,
"grad_norm": 0.8698805570602417,
"learning_rate": 3.196207305247279e-05,
"loss": 3.1402,
"step": 88000
},
{
"epoch": 9.471531589710473,
"eval_accuracy": 0.39370221785534193,
"eval_loss": 3.3107540607452393,
"eval_runtime": 180.1755,
"eval_samples_per_second": 99.964,
"eval_steps_per_second": 6.249,
"step": 88000
},
{
"epoch": 9.476913141750082,
"grad_norm": 0.8367273807525635,
"learning_rate": 3.1638832022411374e-05,
"loss": 3.14,
"step": 88050
},
{
"epoch": 9.482294693789688,
"grad_norm": 0.8261746764183044,
"learning_rate": 3.1315590992349955e-05,
"loss": 3.1287,
"step": 88100
},
{
"epoch": 9.487676245829297,
"grad_norm": 0.8505935668945312,
"learning_rate": 3.099234996228854e-05,
"loss": 3.1379,
"step": 88150
},
{
"epoch": 9.493057797868905,
"grad_norm": 0.8693322539329529,
"learning_rate": 3.0675573752828356e-05,
"loss": 3.1351,
"step": 88200
},
{
"epoch": 9.498439349908514,
"grad_norm": 0.8701393604278564,
"learning_rate": 3.035233272276694e-05,
"loss": 3.116,
"step": 88250
},
{
"epoch": 9.503820901948123,
"grad_norm": 0.8780934810638428,
"learning_rate": 3.0029091692705525e-05,
"loss": 3.138,
"step": 88300
},
{
"epoch": 9.50920245398773,
"grad_norm": 0.8750477433204651,
"learning_rate": 2.970585066264411e-05,
"loss": 3.113,
"step": 88350
},
{
"epoch": 9.514584006027338,
"grad_norm": 0.8323451280593872,
"learning_rate": 2.938260963258269e-05,
"loss": 3.1459,
"step": 88400
},
{
"epoch": 9.519965558066946,
"grad_norm": 0.8494437336921692,
"learning_rate": 2.9059368602521275e-05,
"loss": 3.1316,
"step": 88450
},
{
"epoch": 9.525347110106555,
"grad_norm": 0.8687306046485901,
"learning_rate": 2.8736127572459863e-05,
"loss": 3.1182,
"step": 88500
},
{
"epoch": 9.530728662146164,
"grad_norm": 0.8358095288276672,
"learning_rate": 2.8412886542398448e-05,
"loss": 3.1418,
"step": 88550
},
{
"epoch": 9.536110214185772,
"grad_norm": 0.8712921142578125,
"learning_rate": 2.8089645512337032e-05,
"loss": 3.1184,
"step": 88600
},
{
"epoch": 9.541491766225379,
"grad_norm": 0.9299412965774536,
"learning_rate": 2.7766404482275613e-05,
"loss": 3.116,
"step": 88650
},
{
"epoch": 9.546873318264987,
"grad_norm": 0.8581258654594421,
"learning_rate": 2.7443163452214198e-05,
"loss": 3.142,
"step": 88700
},
{
"epoch": 9.552254870304596,
"grad_norm": 0.8373107314109802,
"learning_rate": 2.7119922422152782e-05,
"loss": 3.108,
"step": 88750
},
{
"epoch": 9.557636422344205,
"grad_norm": 0.83504718542099,
"learning_rate": 2.6796681392091367e-05,
"loss": 3.1487,
"step": 88800
},
{
"epoch": 9.563017974383813,
"grad_norm": 0.8694041967391968,
"learning_rate": 2.6473440362029955e-05,
"loss": 3.147,
"step": 88850
},
{
"epoch": 9.56839952642342,
"grad_norm": 0.7938987016677856,
"learning_rate": 2.6150199331968536e-05,
"loss": 3.1254,
"step": 88900
},
{
"epoch": 9.573781078463028,
"grad_norm": 0.9023017287254333,
"learning_rate": 2.582695830190712e-05,
"loss": 3.1391,
"step": 88950
},
{
"epoch": 9.579162630502637,
"grad_norm": 0.8764638304710388,
"learning_rate": 2.5503717271845705e-05,
"loss": 3.1279,
"step": 89000
},
{
"epoch": 9.579162630502637,
"eval_accuracy": 0.39404023698933277,
"eval_loss": 3.308558940887451,
"eval_runtime": 180.0618,
"eval_samples_per_second": 100.027,
"eval_steps_per_second": 6.253,
"step": 89000
},
{
"epoch": 9.584544182542245,
"grad_norm": 0.8714621663093567,
"learning_rate": 2.518047624178429e-05,
"loss": 3.1327,
"step": 89050
},
{
"epoch": 9.589925734581854,
"grad_norm": 0.872740626335144,
"learning_rate": 2.4857235211722874e-05,
"loss": 3.135,
"step": 89100
},
{
"epoch": 9.59530728662146,
"grad_norm": 0.8976700901985168,
"learning_rate": 2.4533994181661455e-05,
"loss": 3.1357,
"step": 89150
},
{
"epoch": 9.60068883866107,
"grad_norm": 0.8642706871032715,
"learning_rate": 2.421075315160004e-05,
"loss": 3.1514,
"step": 89200
},
{
"epoch": 9.606070390700678,
"grad_norm": 0.8876418471336365,
"learning_rate": 2.3887512121538624e-05,
"loss": 3.1272,
"step": 89250
},
{
"epoch": 9.611451942740286,
"grad_norm": 0.8454718589782715,
"learning_rate": 2.3564271091477212e-05,
"loss": 3.1263,
"step": 89300
},
{
"epoch": 9.616833494779895,
"grad_norm": 0.8721873760223389,
"learning_rate": 2.3241030061415796e-05,
"loss": 3.133,
"step": 89350
},
{
"epoch": 9.622215046819504,
"grad_norm": 0.8209410905838013,
"learning_rate": 2.2917789031354377e-05,
"loss": 3.1391,
"step": 89400
},
{
"epoch": 9.62759659885911,
"grad_norm": 0.8578438758850098,
"learning_rate": 2.2594548001292962e-05,
"loss": 3.1284,
"step": 89450
},
{
"epoch": 9.632978150898719,
"grad_norm": 0.8609829545021057,
"learning_rate": 2.2271306971231546e-05,
"loss": 3.1445,
"step": 89500
},
{
"epoch": 9.638359702938327,
"grad_norm": 0.8142579793930054,
"learning_rate": 2.194806594117013e-05,
"loss": 3.1321,
"step": 89550
},
{
"epoch": 9.643741254977936,
"grad_norm": 0.9409341812133789,
"learning_rate": 2.1624824911108715e-05,
"loss": 3.1258,
"step": 89600
},
{
"epoch": 9.649122807017545,
"grad_norm": 0.8790959119796753,
"learning_rate": 2.1301583881047296e-05,
"loss": 3.1316,
"step": 89650
},
{
"epoch": 9.654504359057151,
"grad_norm": 0.8953224420547485,
"learning_rate": 2.097834285098588e-05,
"loss": 3.1157,
"step": 89700
},
{
"epoch": 9.65988591109676,
"grad_norm": 0.8896967172622681,
"learning_rate": 2.065510182092447e-05,
"loss": 3.1224,
"step": 89750
},
{
"epoch": 9.665267463136368,
"grad_norm": 0.8611626625061035,
"learning_rate": 2.0331860790863053e-05,
"loss": 3.1174,
"step": 89800
},
{
"epoch": 9.670649015175977,
"grad_norm": 0.8367219567298889,
"learning_rate": 2.0008619760801638e-05,
"loss": 3.1503,
"step": 89850
},
{
"epoch": 9.676030567215586,
"grad_norm": 0.8824328184127808,
"learning_rate": 1.968537873074022e-05,
"loss": 3.1359,
"step": 89900
},
{
"epoch": 9.681412119255192,
"grad_norm": 0.8521384000778198,
"learning_rate": 1.9362137700678803e-05,
"loss": 3.1371,
"step": 89950
},
{
"epoch": 9.6867936712948,
"grad_norm": 0.8393240571022034,
"learning_rate": 1.9038896670617388e-05,
"loss": 3.159,
"step": 90000
},
{
"epoch": 9.6867936712948,
"eval_accuracy": 0.3942384198567064,
"eval_loss": 3.3065290451049805,
"eval_runtime": 180.214,
"eval_samples_per_second": 99.942,
"eval_steps_per_second": 6.248,
"step": 90000
},
{
"epoch": 9.69217522333441,
"grad_norm": 0.8758241534233093,
"learning_rate": 1.8715655640555972e-05,
"loss": 3.1504,
"step": 90050
},
{
"epoch": 9.697556775374018,
"grad_norm": 0.8368741273880005,
"learning_rate": 1.8392414610494557e-05,
"loss": 3.1238,
"step": 90100
},
{
"epoch": 9.702938327413626,
"grad_norm": 0.8934403657913208,
"learning_rate": 1.806917358043314e-05,
"loss": 3.1538,
"step": 90150
},
{
"epoch": 9.708319879453235,
"grad_norm": 0.9281708598136902,
"learning_rate": 1.7745932550371726e-05,
"loss": 3.1454,
"step": 90200
},
{
"epoch": 9.713701431492842,
"grad_norm": 0.8880597949028015,
"learning_rate": 1.742269152031031e-05,
"loss": 3.128,
"step": 90250
},
{
"epoch": 9.71908298353245,
"grad_norm": 0.8629451394081116,
"learning_rate": 1.7105915310850124e-05,
"loss": 3.1369,
"step": 90300
},
{
"epoch": 9.724464535572059,
"grad_norm": 0.8905329704284668,
"learning_rate": 1.6782674280788708e-05,
"loss": 3.1219,
"step": 90350
},
{
"epoch": 9.729846087611667,
"grad_norm": 0.9040992856025696,
"learning_rate": 1.645943325072729e-05,
"loss": 3.1318,
"step": 90400
},
{
"epoch": 9.735227639651276,
"grad_norm": 0.8780677318572998,
"learning_rate": 1.6136192220665877e-05,
"loss": 3.1296,
"step": 90450
},
{
"epoch": 9.740609191690883,
"grad_norm": 0.8315309882164001,
"learning_rate": 1.5812951190604458e-05,
"loss": 3.1468,
"step": 90500
},
{
"epoch": 9.745990743730491,
"grad_norm": 0.8430354595184326,
"learning_rate": 1.5489710160543043e-05,
"loss": 3.1279,
"step": 90550
},
{
"epoch": 9.7513722957701,
"grad_norm": 0.8890626430511475,
"learning_rate": 1.5166469130481629e-05,
"loss": 3.1419,
"step": 90600
},
{
"epoch": 9.756753847809708,
"grad_norm": 0.8835778832435608,
"learning_rate": 1.4843228100420212e-05,
"loss": 3.1615,
"step": 90650
},
{
"epoch": 9.762135399849317,
"grad_norm": 0.8142911791801453,
"learning_rate": 1.4519987070358796e-05,
"loss": 3.1359,
"step": 90700
},
{
"epoch": 9.767516951888926,
"grad_norm": 0.8582029938697815,
"learning_rate": 1.419674604029738e-05,
"loss": 3.1239,
"step": 90750
},
{
"epoch": 9.772898503928532,
"grad_norm": 0.9165640473365784,
"learning_rate": 1.3873505010235965e-05,
"loss": 3.1209,
"step": 90800
},
{
"epoch": 9.77828005596814,
"grad_norm": 0.8222701549530029,
"learning_rate": 1.355026398017455e-05,
"loss": 3.1337,
"step": 90850
},
{
"epoch": 9.78366160800775,
"grad_norm": 0.8537663221359253,
"learning_rate": 1.3227022950113132e-05,
"loss": 3.14,
"step": 90900
},
{
"epoch": 9.789043160047358,
"grad_norm": 0.8751277923583984,
"learning_rate": 1.2903781920051719e-05,
"loss": 3.1155,
"step": 90950
},
{
"epoch": 9.794424712086967,
"grad_norm": 0.8880753517150879,
"learning_rate": 1.2580540889990301e-05,
"loss": 3.1357,
"step": 91000
},
{
"epoch": 9.794424712086967,
"eval_accuracy": 0.3946130550138907,
"eval_loss": 3.3034253120422363,
"eval_runtime": 180.0837,
"eval_samples_per_second": 100.015,
"eval_steps_per_second": 6.253,
"step": 91000
},
{
"epoch": 9.799806264126573,
"grad_norm": 0.8514898419380188,
"learning_rate": 1.2257299859928886e-05,
"loss": 3.1331,
"step": 91050
},
{
"epoch": 9.805187816166182,
"grad_norm": 0.8708495497703552,
"learning_rate": 1.193405882986747e-05,
"loss": 3.1113,
"step": 91100
},
{
"epoch": 9.81056936820579,
"grad_norm": 0.8504621982574463,
"learning_rate": 1.1610817799806053e-05,
"loss": 3.1098,
"step": 91150
},
{
"epoch": 9.815950920245399,
"grad_norm": 0.8342603445053101,
"learning_rate": 1.128757676974464e-05,
"loss": 3.1181,
"step": 91200
},
{
"epoch": 9.821332472285007,
"grad_norm": 0.8990119099617004,
"learning_rate": 1.0964335739683222e-05,
"loss": 3.1207,
"step": 91250
},
{
"epoch": 9.826714024324616,
"grad_norm": 0.8407130241394043,
"learning_rate": 1.0641094709621807e-05,
"loss": 3.124,
"step": 91300
},
{
"epoch": 9.832095576364223,
"grad_norm": 0.8398468494415283,
"learning_rate": 1.0317853679560393e-05,
"loss": 3.1244,
"step": 91350
},
{
"epoch": 9.837477128403831,
"grad_norm": 0.8928409814834595,
"learning_rate": 9.994612649498976e-06,
"loss": 3.1233,
"step": 91400
},
{
"epoch": 9.84285868044344,
"grad_norm": 0.8549565672874451,
"learning_rate": 9.67137161943756e-06,
"loss": 3.1248,
"step": 91450
},
{
"epoch": 9.848240232483048,
"grad_norm": 0.8319708108901978,
"learning_rate": 9.348130589376145e-06,
"loss": 3.1021,
"step": 91500
},
{
"epoch": 9.853621784522657,
"grad_norm": 0.8330086469650269,
"learning_rate": 9.024889559314727e-06,
"loss": 3.1326,
"step": 91550
},
{
"epoch": 9.859003336562264,
"grad_norm": 0.8426081538200378,
"learning_rate": 8.701648529253312e-06,
"loss": 3.1118,
"step": 91600
},
{
"epoch": 9.864384888601872,
"grad_norm": 0.8548322916030884,
"learning_rate": 8.378407499191896e-06,
"loss": 3.1311,
"step": 91650
},
{
"epoch": 9.869766440641481,
"grad_norm": 0.8335317969322205,
"learning_rate": 8.055166469130481e-06,
"loss": 3.1387,
"step": 91700
},
{
"epoch": 9.87514799268109,
"grad_norm": 0.9084175229072571,
"learning_rate": 7.731925439069065e-06,
"loss": 3.1367,
"step": 91750
},
{
"epoch": 9.880529544720698,
"grad_norm": 0.8292434215545654,
"learning_rate": 7.408684409007649e-06,
"loss": 3.1127,
"step": 91800
},
{
"epoch": 9.885911096760307,
"grad_norm": 0.8056044578552246,
"learning_rate": 7.085443378946234e-06,
"loss": 3.1378,
"step": 91850
},
{
"epoch": 9.891292648799913,
"grad_norm": 0.8410173654556274,
"learning_rate": 6.762202348884817e-06,
"loss": 3.1161,
"step": 91900
},
{
"epoch": 9.896674200839522,
"grad_norm": 0.9032399654388428,
"learning_rate": 6.438961318823403e-06,
"loss": 3.1386,
"step": 91950
},
{
"epoch": 9.90205575287913,
"grad_norm": 0.8298187851905823,
"learning_rate": 6.115720288761986e-06,
"loss": 3.1212,
"step": 92000
},
{
"epoch": 9.90205575287913,
"eval_accuracy": 0.39463576346744395,
"eval_loss": 3.303168535232544,
"eval_runtime": 180.0813,
"eval_samples_per_second": 100.016,
"eval_steps_per_second": 6.253,
"step": 92000
},
{
"epoch": 9.907437304918739,
"grad_norm": 0.8719784021377563,
"learning_rate": 5.792479258700571e-06,
"loss": 3.1241,
"step": 92050
},
{
"epoch": 9.912818856958348,
"grad_norm": 0.8631389141082764,
"learning_rate": 5.469238228639154e-06,
"loss": 3.1118,
"step": 92100
},
{
"epoch": 9.918200408997954,
"grad_norm": 0.8735671639442444,
"learning_rate": 5.145997198577739e-06,
"loss": 3.1353,
"step": 92150
},
{
"epoch": 9.923581961037563,
"grad_norm": 0.8469117879867554,
"learning_rate": 4.822756168516323e-06,
"loss": 3.128,
"step": 92200
},
{
"epoch": 9.928963513077171,
"grad_norm": 0.8447126150131226,
"learning_rate": 4.499515138454908e-06,
"loss": 3.1262,
"step": 92250
},
{
"epoch": 9.93434506511678,
"grad_norm": 0.8380026817321777,
"learning_rate": 4.1762741083934915e-06,
"loss": 3.1138,
"step": 92300
},
{
"epoch": 9.939726617156388,
"grad_norm": 0.838094174861908,
"learning_rate": 3.853033078332076e-06,
"loss": 3.1297,
"step": 92350
},
{
"epoch": 9.945108169195997,
"grad_norm": 0.9846398830413818,
"learning_rate": 3.52979204827066e-06,
"loss": 3.1255,
"step": 92400
},
{
"epoch": 9.950489721235604,
"grad_norm": 0.8288633823394775,
"learning_rate": 3.206551018209244e-06,
"loss": 3.1255,
"step": 92450
},
{
"epoch": 9.955871273275212,
"grad_norm": 0.8590484261512756,
"learning_rate": 2.8833099881478286e-06,
"loss": 3.1037,
"step": 92500
},
{
"epoch": 9.961252825314821,
"grad_norm": 0.8735798001289368,
"learning_rate": 2.5600689580864127e-06,
"loss": 3.1176,
"step": 92550
},
{
"epoch": 9.96663437735443,
"grad_norm": 0.8541606068611145,
"learning_rate": 2.2368279280249972e-06,
"loss": 3.1224,
"step": 92600
},
{
"epoch": 9.972015929394038,
"grad_norm": 0.8478603363037109,
"learning_rate": 1.9135868979635813e-06,
"loss": 3.1271,
"step": 92650
},
{
"epoch": 9.977397481433645,
"grad_norm": 0.8639833331108093,
"learning_rate": 1.596810688503394e-06,
"loss": 3.1237,
"step": 92700
},
{
"epoch": 9.982779033473253,
"grad_norm": 0.8723976612091064,
"learning_rate": 1.2735696584419782e-06,
"loss": 3.1151,
"step": 92750
},
{
"epoch": 9.988160585512862,
"grad_norm": 0.8547508716583252,
"learning_rate": 9.503286283805623e-07,
"loss": 3.1201,
"step": 92800
},
{
"epoch": 9.99354213755247,
"grad_norm": 0.8455697894096375,
"learning_rate": 6.270875983191466e-07,
"loss": 3.1272,
"step": 92850
},
{
"epoch": 9.998923689592079,
"grad_norm": 0.8672316670417786,
"learning_rate": 3.038465682577308e-07,
"loss": 3.1172,
"step": 92900
},
{
"epoch": 10.0,
"step": 92910,
"total_flos": 7.7683166281728e+17,
"train_loss": 3.4544442354291984,
"train_runtime": 80531.2369,
"train_samples_per_second": 36.918,
"train_steps_per_second": 1.154
}
],
"logging_steps": 50,
"max_steps": 92910,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7.7683166281728e+17,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}