{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.984202967927238, "eval_steps": 500, "global_step": 650, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007659167065581618, "grad_norm": 5.921403529288989, "learning_rate": 6.153846153846155e-07, "loss": 0.9112, "step": 1 }, { "epoch": 0.015318334131163236, "grad_norm": 6.161237327775742, "learning_rate": 1.230769230769231e-06, "loss": 0.9517, "step": 2 }, { "epoch": 0.022977501196744854, "grad_norm": 6.058434736522049, "learning_rate": 1.8461538461538465e-06, "loss": 0.9295, "step": 3 }, { "epoch": 0.030636668262326472, "grad_norm": 6.016130530129469, "learning_rate": 2.461538461538462e-06, "loss": 0.9235, "step": 4 }, { "epoch": 0.03829583532790809, "grad_norm": 5.729487082326455, "learning_rate": 3.0769230769230774e-06, "loss": 0.9035, "step": 5 }, { "epoch": 0.04595500239348971, "grad_norm": 4.266411955281039, "learning_rate": 3.692307692307693e-06, "loss": 0.8544, "step": 6 }, { "epoch": 0.053614169459071326, "grad_norm": 2.4685484993069466, "learning_rate": 4.307692307692308e-06, "loss": 0.8081, "step": 7 }, { "epoch": 0.061273336524652944, "grad_norm": 2.2509637870140136, "learning_rate": 4.923076923076924e-06, "loss": 0.836, "step": 8 }, { "epoch": 0.06893250359023456, "grad_norm": 3.6503885662831137, "learning_rate": 5.538461538461539e-06, "loss": 0.7788, "step": 9 }, { "epoch": 0.07659167065581618, "grad_norm": 4.501975701703904, "learning_rate": 6.153846153846155e-06, "loss": 0.8309, "step": 10 }, { "epoch": 0.0842508377213978, "grad_norm": 4.370930073810917, "learning_rate": 6.76923076923077e-06, "loss": 0.8224, "step": 11 }, { "epoch": 0.09191000478697942, "grad_norm": 3.546183001564244, "learning_rate": 7.384615384615386e-06, "loss": 0.7941, "step": 12 }, { "epoch": 0.09956917185256103, "grad_norm": 3.2331489380759555, "learning_rate": 8.000000000000001e-06, "loss": 0.7635, "step": 13 }, { "epoch": 0.10722833891814265, "grad_norm": 2.5015396395106966, "learning_rate": 8.615384615384617e-06, "loss": 0.7327, "step": 14 }, { "epoch": 0.11488750598372427, "grad_norm": 1.8441949964024125, "learning_rate": 9.230769230769232e-06, "loss": 0.7094, "step": 15 }, { "epoch": 0.12254667304930589, "grad_norm": 1.732314176486123, "learning_rate": 9.846153846153848e-06, "loss": 0.6849, "step": 16 }, { "epoch": 0.1302058401148875, "grad_norm": 1.8608940893991852, "learning_rate": 1.0461538461538463e-05, "loss": 0.6811, "step": 17 }, { "epoch": 0.13786500718046912, "grad_norm": 1.8077531019190372, "learning_rate": 1.1076923076923079e-05, "loss": 0.6745, "step": 18 }, { "epoch": 0.14552417424605074, "grad_norm": 1.2577918457624666, "learning_rate": 1.1692307692307694e-05, "loss": 0.6578, "step": 19 }, { "epoch": 0.15318334131163236, "grad_norm": 1.289096433457239, "learning_rate": 1.230769230769231e-05, "loss": 0.6754, "step": 20 }, { "epoch": 0.16084250837721398, "grad_norm": 1.4570529134788588, "learning_rate": 1.2923076923076925e-05, "loss": 0.6655, "step": 21 }, { "epoch": 0.1685016754427956, "grad_norm": 1.1952396245104895, "learning_rate": 1.353846153846154e-05, "loss": 0.64, "step": 22 }, { "epoch": 0.1761608425083772, "grad_norm": 1.0573908695788548, "learning_rate": 1.4153846153846156e-05, "loss": 0.6226, "step": 23 }, { "epoch": 0.18382000957395883, "grad_norm": 1.0531289868486842, "learning_rate": 1.4769230769230772e-05, "loss": 0.5999, "step": 24 }, { "epoch": 0.19147917663954045, "grad_norm": 0.8767976703429683, "learning_rate": 1.5384615384615387e-05, "loss": 0.6224, "step": 25 }, { "epoch": 0.19913834370512207, "grad_norm": 1.0191023286379297, "learning_rate": 1.6000000000000003e-05, "loss": 0.6382, "step": 26 }, { "epoch": 0.20679751077070369, "grad_norm": 0.8038756547216995, "learning_rate": 1.6615384615384618e-05, "loss": 0.6216, "step": 27 }, { "epoch": 0.2144566778362853, "grad_norm": 0.958933004581782, "learning_rate": 1.7230769230769234e-05, "loss": 0.6154, "step": 28 }, { "epoch": 0.22211584490186692, "grad_norm": 0.9664195885794775, "learning_rate": 1.784615384615385e-05, "loss": 0.6128, "step": 29 }, { "epoch": 0.22977501196744854, "grad_norm": 0.9729481103689919, "learning_rate": 1.8461538461538465e-05, "loss": 0.5937, "step": 30 }, { "epoch": 0.23743417903303016, "grad_norm": 0.9246652793706185, "learning_rate": 1.907692307692308e-05, "loss": 0.6247, "step": 31 }, { "epoch": 0.24509334609861178, "grad_norm": 0.9530006352752787, "learning_rate": 1.9692307692307696e-05, "loss": 0.5816, "step": 32 }, { "epoch": 0.2527525131641934, "grad_norm": 0.8884934035836074, "learning_rate": 2.0307692307692308e-05, "loss": 0.6311, "step": 33 }, { "epoch": 0.260411680229775, "grad_norm": 0.8458697710494548, "learning_rate": 2.0923076923076927e-05, "loss": 0.6026, "step": 34 }, { "epoch": 0.26807084729535663, "grad_norm": 0.8241647035938812, "learning_rate": 2.153846153846154e-05, "loss": 0.5813, "step": 35 }, { "epoch": 0.27573001436093825, "grad_norm": 0.6637335260284863, "learning_rate": 2.2153846153846158e-05, "loss": 0.5826, "step": 36 }, { "epoch": 0.28338918142651986, "grad_norm": 0.9863751042251347, "learning_rate": 2.276923076923077e-05, "loss": 0.6061, "step": 37 }, { "epoch": 0.2910483484921015, "grad_norm": 0.7725648338774505, "learning_rate": 2.338461538461539e-05, "loss": 0.5535, "step": 38 }, { "epoch": 0.2987075155576831, "grad_norm": 0.9197682556877761, "learning_rate": 2.4e-05, "loss": 0.5891, "step": 39 }, { "epoch": 0.3063666826232647, "grad_norm": 0.9877497296915406, "learning_rate": 2.461538461538462e-05, "loss": 0.575, "step": 40 }, { "epoch": 0.31402584968884634, "grad_norm": 0.720541520642535, "learning_rate": 2.523076923076923e-05, "loss": 0.5779, "step": 41 }, { "epoch": 0.32168501675442795, "grad_norm": 0.9225926491059669, "learning_rate": 2.584615384615385e-05, "loss": 0.5528, "step": 42 }, { "epoch": 0.3293441838200096, "grad_norm": 0.8068833036305191, "learning_rate": 2.6461538461538463e-05, "loss": 0.5888, "step": 43 }, { "epoch": 0.3370033508855912, "grad_norm": 0.957481417916251, "learning_rate": 2.707692307692308e-05, "loss": 0.5559, "step": 44 }, { "epoch": 0.3446625179511728, "grad_norm": 0.78572658067857, "learning_rate": 2.7692307692307694e-05, "loss": 0.5612, "step": 45 }, { "epoch": 0.3523216850167544, "grad_norm": 0.9209329658610131, "learning_rate": 2.8307692307692312e-05, "loss": 0.584, "step": 46 }, { "epoch": 0.35998085208233604, "grad_norm": 1.044590316803605, "learning_rate": 2.8923076923076925e-05, "loss": 0.5804, "step": 47 }, { "epoch": 0.36764001914791766, "grad_norm": 1.172192022892308, "learning_rate": 2.9538461538461543e-05, "loss": 0.5518, "step": 48 }, { "epoch": 0.3752991862134993, "grad_norm": 1.0766200831878052, "learning_rate": 3.0153846153846155e-05, "loss": 0.5846, "step": 49 }, { "epoch": 0.3829583532790809, "grad_norm": 1.1041677201522513, "learning_rate": 3.0769230769230774e-05, "loss": 0.5487, "step": 50 }, { "epoch": 0.3906175203446625, "grad_norm": 1.0469042529413854, "learning_rate": 3.1384615384615386e-05, "loss": 0.5395, "step": 51 }, { "epoch": 0.39827668741024413, "grad_norm": 1.0791323635301755, "learning_rate": 3.2000000000000005e-05, "loss": 0.5657, "step": 52 }, { "epoch": 0.40593585447582575, "grad_norm": 0.8078486400788046, "learning_rate": 3.261538461538462e-05, "loss": 0.5399, "step": 53 }, { "epoch": 0.41359502154140737, "grad_norm": 1.191539346367779, "learning_rate": 3.3230769230769236e-05, "loss": 0.5377, "step": 54 }, { "epoch": 0.421254188606989, "grad_norm": 0.8698447542598498, "learning_rate": 3.384615384615385e-05, "loss": 0.5551, "step": 55 }, { "epoch": 0.4289133556725706, "grad_norm": 1.1800967382678582, "learning_rate": 3.446153846153847e-05, "loss": 0.5702, "step": 56 }, { "epoch": 0.4365725227381522, "grad_norm": 1.220687522421987, "learning_rate": 3.507692307692308e-05, "loss": 0.5552, "step": 57 }, { "epoch": 0.44423168980373384, "grad_norm": 0.9041645568135938, "learning_rate": 3.56923076923077e-05, "loss": 0.5688, "step": 58 }, { "epoch": 0.45189085686931546, "grad_norm": 1.5855877495279507, "learning_rate": 3.630769230769231e-05, "loss": 0.5641, "step": 59 }, { "epoch": 0.4595500239348971, "grad_norm": 0.8216109564331809, "learning_rate": 3.692307692307693e-05, "loss": 0.5536, "step": 60 }, { "epoch": 0.4672091910004787, "grad_norm": 1.5857353518922366, "learning_rate": 3.753846153846154e-05, "loss": 0.5497, "step": 61 }, { "epoch": 0.4748683580660603, "grad_norm": 1.0645534816694728, "learning_rate": 3.815384615384616e-05, "loss": 0.5611, "step": 62 }, { "epoch": 0.48252752513164193, "grad_norm": 1.8572686093355073, "learning_rate": 3.876923076923077e-05, "loss": 0.5703, "step": 63 }, { "epoch": 0.49018669219722355, "grad_norm": 1.6301218272572524, "learning_rate": 3.938461538461539e-05, "loss": 0.5468, "step": 64 }, { "epoch": 0.49784585926280517, "grad_norm": 1.1854602191040988, "learning_rate": 4e-05, "loss": 0.564, "step": 65 }, { "epoch": 0.5055050263283868, "grad_norm": 1.5458244211037195, "learning_rate": 3.999971160550277e-05, "loss": 0.5677, "step": 66 }, { "epoch": 0.5131641933939685, "grad_norm": 1.2355216531245246, "learning_rate": 3.999884643032821e-05, "loss": 0.5606, "step": 67 }, { "epoch": 0.52082336045955, "grad_norm": 1.0712313677482361, "learning_rate": 3.9997404499427494e-05, "loss": 0.551, "step": 68 }, { "epoch": 0.5284825275251317, "grad_norm": 1.4808309855361168, "learning_rate": 3.9995385854385124e-05, "loss": 0.542, "step": 69 }, { "epoch": 0.5361416945907133, "grad_norm": 0.8373088986931322, "learning_rate": 3.999279055341771e-05, "loss": 0.5255, "step": 70 }, { "epoch": 0.5438008616562949, "grad_norm": 1.683135657978014, "learning_rate": 3.9989618671372304e-05, "loss": 0.5578, "step": 71 }, { "epoch": 0.5514600287218765, "grad_norm": 1.1420245332567331, "learning_rate": 3.998587029972423e-05, "loss": 0.5729, "step": 72 }, { "epoch": 0.5591191957874582, "grad_norm": 1.6242190582145548, "learning_rate": 3.998154554657448e-05, "loss": 0.5621, "step": 73 }, { "epoch": 0.5667783628530397, "grad_norm": 1.4494865534227146, "learning_rate": 3.997664453664654e-05, "loss": 0.539, "step": 74 }, { "epoch": 0.5744375299186214, "grad_norm": 1.2416402505339934, "learning_rate": 3.9971167411282835e-05, "loss": 0.5451, "step": 75 }, { "epoch": 0.582096696984203, "grad_norm": 1.6591387835608409, "learning_rate": 3.996511432844067e-05, "loss": 0.5649, "step": 76 }, { "epoch": 0.5897558640497846, "grad_norm": 1.2982089198354367, "learning_rate": 3.9958485462687606e-05, "loss": 0.5518, "step": 77 }, { "epoch": 0.5974150311153662, "grad_norm": 1.3123456841661971, "learning_rate": 3.9951281005196486e-05, "loss": 0.5514, "step": 78 }, { "epoch": 0.6050741981809479, "grad_norm": 1.189744218079718, "learning_rate": 3.994350116373991e-05, "loss": 0.5408, "step": 79 }, { "epoch": 0.6127333652465294, "grad_norm": 1.4721204904424363, "learning_rate": 3.9935146162684206e-05, "loss": 0.5507, "step": 80 }, { "epoch": 0.6203925323121111, "grad_norm": 1.1623645549927197, "learning_rate": 3.9926216242983017e-05, "loss": 0.5569, "step": 81 }, { "epoch": 0.6280516993776927, "grad_norm": 1.325416617911315, "learning_rate": 3.991671166217031e-05, "loss": 0.5312, "step": 82 }, { "epoch": 0.6357108664432743, "grad_norm": 1.1877951056068266, "learning_rate": 3.990663269435298e-05, "loss": 0.5644, "step": 83 }, { "epoch": 0.6433700335088559, "grad_norm": 1.2286620885326376, "learning_rate": 3.989597963020289e-05, "loss": 0.5756, "step": 84 }, { "epoch": 0.6510292005744376, "grad_norm": 1.1060602604243368, "learning_rate": 3.9884752776948564e-05, "loss": 0.5311, "step": 85 }, { "epoch": 0.6586883676400191, "grad_norm": 1.0698164056511321, "learning_rate": 3.9872952458366267e-05, "loss": 0.5574, "step": 86 }, { "epoch": 0.6663475347056008, "grad_norm": 1.1820578926976948, "learning_rate": 3.986057901477069e-05, "loss": 0.5513, "step": 87 }, { "epoch": 0.6740067017711824, "grad_norm": 0.8169595363079514, "learning_rate": 3.984763280300514e-05, "loss": 0.5483, "step": 88 }, { "epoch": 0.681665868836764, "grad_norm": 1.379920176811677, "learning_rate": 3.983411419643125e-05, "loss": 0.5892, "step": 89 }, { "epoch": 0.6893250359023456, "grad_norm": 0.9681470926974101, "learning_rate": 3.982002358491817e-05, "loss": 0.5342, "step": 90 }, { "epoch": 0.6969842029679273, "grad_norm": 1.119318426323717, "learning_rate": 3.980536137483141e-05, "loss": 0.5552, "step": 91 }, { "epoch": 0.7046433700335089, "grad_norm": 1.4254561226777611, "learning_rate": 3.9790127989021024e-05, "loss": 0.5336, "step": 92 }, { "epoch": 0.7123025370990905, "grad_norm": 1.0062543492174385, "learning_rate": 3.9774323866809485e-05, "loss": 0.5601, "step": 93 }, { "epoch": 0.7199617041646721, "grad_norm": 1.705345700534379, "learning_rate": 3.9757949463978975e-05, "loss": 0.5659, "step": 94 }, { "epoch": 0.7276208712302538, "grad_norm": 1.1181524317220572, "learning_rate": 3.9741005252758255e-05, "loss": 0.5204, "step": 95 }, { "epoch": 0.7352800382958353, "grad_norm": 1.6689670685483973, "learning_rate": 3.9723491721809076e-05, "loss": 0.5498, "step": 96 }, { "epoch": 0.742939205361417, "grad_norm": 1.4295009254142887, "learning_rate": 3.970540937621201e-05, "loss": 0.565, "step": 97 }, { "epoch": 0.7505983724269986, "grad_norm": 1.4548186946761517, "learning_rate": 3.9686758737451955e-05, "loss": 0.55, "step": 98 }, { "epoch": 0.7582575394925802, "grad_norm": 1.213694031644175, "learning_rate": 3.966754034340308e-05, "loss": 0.5348, "step": 99 }, { "epoch": 0.7659167065581618, "grad_norm": 1.0580342997486925, "learning_rate": 3.9647754748313294e-05, "loss": 0.5242, "step": 100 }, { "epoch": 0.7735758736237435, "grad_norm": 1.0821077174340752, "learning_rate": 3.962740252278827e-05, "loss": 0.542, "step": 101 }, { "epoch": 0.781235040689325, "grad_norm": 1.2618844489347851, "learning_rate": 3.960648425377499e-05, "loss": 0.5465, "step": 102 }, { "epoch": 0.7888942077549067, "grad_norm": 0.9576457209595916, "learning_rate": 3.958500054454482e-05, "loss": 0.5206, "step": 103 }, { "epoch": 0.7965533748204883, "grad_norm": 1.6879319110317588, "learning_rate": 3.9562952014676116e-05, "loss": 0.5554, "step": 104 }, { "epoch": 0.8042125418860699, "grad_norm": 1.3708595612461836, "learning_rate": 3.954033930003634e-05, "loss": 0.5745, "step": 105 }, { "epoch": 0.8118717089516515, "grad_norm": 1.2986325728462578, "learning_rate": 3.9517163052763756e-05, "loss": 0.5413, "step": 106 }, { "epoch": 0.8195308760172332, "grad_norm": 1.1532883579115862, "learning_rate": 3.9493423941248564e-05, "loss": 0.5367, "step": 107 }, { "epoch": 0.8271900430828147, "grad_norm": 1.0800362940814041, "learning_rate": 3.946912265011368e-05, "loss": 0.5643, "step": 108 }, { "epoch": 0.8348492101483964, "grad_norm": 1.0542424383185787, "learning_rate": 3.944425988019498e-05, "loss": 0.5434, "step": 109 }, { "epoch": 0.842508377213978, "grad_norm": 0.9223839325666725, "learning_rate": 3.9418836348521045e-05, "loss": 0.5313, "step": 110 }, { "epoch": 0.8501675442795597, "grad_norm": 0.9786970098640274, "learning_rate": 3.9392852788292556e-05, "loss": 0.5257, "step": 111 }, { "epoch": 0.8578267113451412, "grad_norm": 0.8883004076978934, "learning_rate": 3.936630994886109e-05, "loss": 0.5315, "step": 112 }, { "epoch": 0.8654858784107229, "grad_norm": 0.9210396895221461, "learning_rate": 3.933920859570753e-05, "loss": 0.5332, "step": 113 }, { "epoch": 0.8731450454763044, "grad_norm": 1.0300033230225123, "learning_rate": 3.931154951041998e-05, "loss": 0.5364, "step": 114 }, { "epoch": 0.8808042125418861, "grad_norm": 1.07072986303729, "learning_rate": 3.928333349067125e-05, "loss": 0.5581, "step": 115 }, { "epoch": 0.8884633796074677, "grad_norm": 0.7856322271677167, "learning_rate": 3.925456135019582e-05, "loss": 0.5153, "step": 116 }, { "epoch": 0.8961225466730494, "grad_norm": 1.3208925243839957, "learning_rate": 3.922523391876638e-05, "loss": 0.5397, "step": 117 }, { "epoch": 0.9037817137386309, "grad_norm": 1.0321781910738392, "learning_rate": 3.9195352042169924e-05, "loss": 0.5434, "step": 118 }, { "epoch": 0.9114408808042126, "grad_norm": 1.2263516987444574, "learning_rate": 3.916491658218333e-05, "loss": 0.5412, "step": 119 }, { "epoch": 0.9191000478697942, "grad_norm": 1.0824446357123727, "learning_rate": 3.913392841654851e-05, "loss": 0.5345, "step": 120 }, { "epoch": 0.9267592149353758, "grad_norm": 0.9492587544198882, "learning_rate": 3.9102388438947104e-05, "loss": 0.5374, "step": 121 }, { "epoch": 0.9344183820009574, "grad_norm": 1.200654956363631, "learning_rate": 3.907029755897473e-05, "loss": 0.528, "step": 122 }, { "epoch": 0.9420775490665391, "grad_norm": 0.8940904112935483, "learning_rate": 3.903765670211469e-05, "loss": 0.5432, "step": 123 }, { "epoch": 0.9497367161321206, "grad_norm": 1.513206646677525, "learning_rate": 3.9004466809711343e-05, "loss": 0.55, "step": 124 }, { "epoch": 0.9573958831977023, "grad_norm": 0.9181110535142029, "learning_rate": 3.897072883894291e-05, "loss": 0.5322, "step": 125 }, { "epoch": 0.9650550502632839, "grad_norm": 1.493130538955249, "learning_rate": 3.893644376279392e-05, "loss": 0.5118, "step": 126 }, { "epoch": 0.9727142173288655, "grad_norm": 1.199837488596111, "learning_rate": 3.89016125700271e-05, "loss": 0.5279, "step": 127 }, { "epoch": 0.9803733843944471, "grad_norm": 1.7015944265286984, "learning_rate": 3.8866236265154864e-05, "loss": 0.5464, "step": 128 }, { "epoch": 0.9880325514600288, "grad_norm": 1.4588454807007443, "learning_rate": 3.88303158684104e-05, "loss": 0.5544, "step": 129 }, { "epoch": 0.9956917185256103, "grad_norm": 1.4514845707429012, "learning_rate": 3.879385241571817e-05, "loss": 0.5686, "step": 130 }, { "epoch": 1.0047869794159885, "grad_norm": 2.2360445683138384, "learning_rate": 3.875684695866409e-05, "loss": 0.8604, "step": 131 }, { "epoch": 1.0124461464815702, "grad_norm": 1.085982951452546, "learning_rate": 3.871930056446518e-05, "loss": 0.4109, "step": 132 }, { "epoch": 1.0201053135471518, "grad_norm": 0.9862931565568734, "learning_rate": 3.8681214315938786e-05, "loss": 0.475, "step": 133 }, { "epoch": 1.0277644806127333, "grad_norm": 1.33743772292278, "learning_rate": 3.864258931147136e-05, "loss": 0.4306, "step": 134 }, { "epoch": 1.035423647678315, "grad_norm": 1.0921525549828976, "learning_rate": 3.860342666498677e-05, "loss": 0.4848, "step": 135 }, { "epoch": 1.0430828147438966, "grad_norm": 1.2645353404433375, "learning_rate": 3.856372750591419e-05, "loss": 0.4485, "step": 136 }, { "epoch": 1.0507419818094783, "grad_norm": 1.1984966128303223, "learning_rate": 3.8523492979155534e-05, "loss": 0.4808, "step": 137 }, { "epoch": 1.0584011488750598, "grad_norm": 1.0500948118376512, "learning_rate": 3.84827242450524e-05, "loss": 0.4881, "step": 138 }, { "epoch": 1.0660603159406414, "grad_norm": 0.746032471414876, "learning_rate": 3.844142247935265e-05, "loss": 0.4028, "step": 139 }, { "epoch": 1.073719483006223, "grad_norm": 0.9623091350387484, "learning_rate": 3.839958887317649e-05, "loss": 0.4715, "step": 140 }, { "epoch": 1.0813786500718048, "grad_norm": 0.7201641889161347, "learning_rate": 3.835722463298208e-05, "loss": 0.4606, "step": 141 }, { "epoch": 1.0890378171373862, "grad_norm": 0.8943620531687253, "learning_rate": 3.831433098053082e-05, "loss": 0.4712, "step": 142 }, { "epoch": 1.096696984202968, "grad_norm": 0.7046882034899636, "learning_rate": 3.827090915285202e-05, "loss": 0.4458, "step": 143 }, { "epoch": 1.1043561512685496, "grad_norm": 0.6548369768890453, "learning_rate": 3.8226960402207316e-05, "loss": 0.3904, "step": 144 }, { "epoch": 1.1120153183341313, "grad_norm": 0.8530160058379846, "learning_rate": 3.818248599605448e-05, "loss": 0.546, "step": 145 }, { "epoch": 1.1196744853997127, "grad_norm": 0.8081896528154348, "learning_rate": 3.813748721701091e-05, "loss": 0.4651, "step": 146 }, { "epoch": 1.1273336524652944, "grad_norm": 0.7600777048606053, "learning_rate": 3.809196536281665e-05, "loss": 0.3961, "step": 147 }, { "epoch": 1.134992819530876, "grad_norm": 0.8170599343972957, "learning_rate": 3.80459217462969e-05, "loss": 0.4859, "step": 148 }, { "epoch": 1.1426519865964577, "grad_norm": 0.5981126204379706, "learning_rate": 3.799935769532425e-05, "loss": 0.4112, "step": 149 }, { "epoch": 1.1503111536620392, "grad_norm": 0.6918899342824092, "learning_rate": 3.795227455278029e-05, "loss": 0.445, "step": 150 }, { "epoch": 1.1579703207276208, "grad_norm": 0.7615214743583981, "learning_rate": 3.790467367651694e-05, "loss": 0.5105, "step": 151 }, { "epoch": 1.1656294877932025, "grad_norm": 0.5625993273766481, "learning_rate": 3.785655643931728e-05, "loss": 0.3587, "step": 152 }, { "epoch": 1.1732886548587842, "grad_norm": 0.9402358775980364, "learning_rate": 3.780792422885597e-05, "loss": 0.5268, "step": 153 }, { "epoch": 1.1809478219243656, "grad_norm": 1.0131842602903653, "learning_rate": 3.7758778447659184e-05, "loss": 0.4696, "step": 154 }, { "epoch": 1.1886069889899473, "grad_norm": 0.7498386211674599, "learning_rate": 3.7709120513064196e-05, "loss": 0.4471, "step": 155 }, { "epoch": 1.196266156055529, "grad_norm": 0.7182193922598371, "learning_rate": 3.7658951857178544e-05, "loss": 0.4344, "step": 156 }, { "epoch": 1.2039253231211107, "grad_norm": 0.6854353371665008, "learning_rate": 3.760827392683863e-05, "loss": 0.4195, "step": 157 }, { "epoch": 1.2115844901866921, "grad_norm": 1.1427693762223539, "learning_rate": 3.755708818356809e-05, "loss": 0.5111, "step": 158 }, { "epoch": 1.2192436572522738, "grad_norm": 0.6625293413220111, "learning_rate": 3.75053961035356e-05, "loss": 0.4288, "step": 159 }, { "epoch": 1.2269028243178555, "grad_norm": 0.9123675189917038, "learning_rate": 3.745319917751229e-05, "loss": 0.4656, "step": 160 }, { "epoch": 1.2345619913834371, "grad_norm": 0.7341125406296108, "learning_rate": 3.740049891082879e-05, "loss": 0.425, "step": 161 }, { "epoch": 1.2422211584490186, "grad_norm": 0.8631406479236612, "learning_rate": 3.734729682333179e-05, "loss": 0.4911, "step": 162 }, { "epoch": 1.2498803255146003, "grad_norm": 0.7496345502817138, "learning_rate": 3.729359444934022e-05, "loss": 0.3874, "step": 163 }, { "epoch": 1.257539492580182, "grad_norm": 0.7894969941658941, "learning_rate": 3.723939333760099e-05, "loss": 0.4578, "step": 164 }, { "epoch": 1.2651986596457636, "grad_norm": 0.9223288305001206, "learning_rate": 3.718469505124434e-05, "loss": 0.523, "step": 165 }, { "epoch": 1.272857826711345, "grad_norm": 0.7992363424810383, "learning_rate": 3.712950116773875e-05, "loss": 0.46, "step": 166 }, { "epoch": 1.2805169937769267, "grad_norm": 0.8478230607138995, "learning_rate": 3.707381327884545e-05, "loss": 0.4749, "step": 167 }, { "epoch": 1.2881761608425084, "grad_norm": 0.5454929496812748, "learning_rate": 3.70176329905725e-05, "loss": 0.4463, "step": 168 }, { "epoch": 1.2958353279080899, "grad_norm": 0.750962890324772, "learning_rate": 3.696096192312852e-05, "loss": 0.4562, "step": 169 }, { "epoch": 1.3034944949736715, "grad_norm": 0.6464014219567544, "learning_rate": 3.69038017108759e-05, "loss": 0.458, "step": 170 }, { "epoch": 1.3111536620392532, "grad_norm": 0.6483521349636168, "learning_rate": 3.6846154002283696e-05, "loss": 0.4709, "step": 171 }, { "epoch": 1.3188128291048349, "grad_norm": 0.7936957658021746, "learning_rate": 3.678802045988012e-05, "loss": 0.4581, "step": 172 }, { "epoch": 1.3264719961704166, "grad_norm": 0.9116293942129717, "learning_rate": 3.6729402760204535e-05, "loss": 0.4367, "step": 173 }, { "epoch": 1.334131163235998, "grad_norm": 0.8432492215433733, "learning_rate": 3.667030259375915e-05, "loss": 0.5148, "step": 174 }, { "epoch": 1.3417903303015797, "grad_norm": 0.7438248026517662, "learning_rate": 3.6610721664960236e-05, "loss": 0.4653, "step": 175 }, { "epoch": 1.3494494973671614, "grad_norm": 0.6889892205036577, "learning_rate": 3.6550661692089e-05, "loss": 0.4231, "step": 176 }, { "epoch": 1.3571086644327428, "grad_norm": 0.7784374117922759, "learning_rate": 3.6490124407242007e-05, "loss": 0.5052, "step": 177 }, { "epoch": 1.3647678314983245, "grad_norm": 0.8790420908019635, "learning_rate": 3.642911155628124e-05, "loss": 0.4415, "step": 178 }, { "epoch": 1.3724269985639062, "grad_norm": 0.7174987210830999, "learning_rate": 3.636762489878374e-05, "loss": 0.4421, "step": 179 }, { "epoch": 1.3800861656294878, "grad_norm": 0.8020599752892646, "learning_rate": 3.6305666207990886e-05, "loss": 0.4863, "step": 180 }, { "epoch": 1.3877453326950695, "grad_norm": 0.7802012248909737, "learning_rate": 3.624323727075723e-05, "loss": 0.4426, "step": 181 }, { "epoch": 1.395404499760651, "grad_norm": 0.671987752263627, "learning_rate": 3.6180339887498953e-05, "loss": 0.4077, "step": 182 }, { "epoch": 1.4030636668262326, "grad_norm": 0.9587427415971348, "learning_rate": 3.6116975872141984e-05, "loss": 0.5101, "step": 183 }, { "epoch": 1.4107228338918143, "grad_norm": 0.9118291287887855, "learning_rate": 3.605314705206966e-05, "loss": 0.4823, "step": 184 }, { "epoch": 1.4183820009573957, "grad_norm": 0.8693454740841087, "learning_rate": 3.598885526807003e-05, "loss": 0.4712, "step": 185 }, { "epoch": 1.4260411680229774, "grad_norm": 0.7409071346511643, "learning_rate": 3.5924102374282754e-05, "loss": 0.4422, "step": 186 }, { "epoch": 1.433700335088559, "grad_norm": 0.9303869755499337, "learning_rate": 3.5858890238145674e-05, "loss": 0.4207, "step": 187 }, { "epoch": 1.4413595021541408, "grad_norm": 0.852020974421498, "learning_rate": 3.5793220740340904e-05, "loss": 0.4927, "step": 188 }, { "epoch": 1.4490186692197224, "grad_norm": 0.7715289615754756, "learning_rate": 3.572709577474062e-05, "loss": 0.4546, "step": 189 }, { "epoch": 1.456677836285304, "grad_norm": 0.7648123067206034, "learning_rate": 3.566051724835245e-05, "loss": 0.4775, "step": 190 }, { "epoch": 1.4643370033508856, "grad_norm": 0.6233854224617866, "learning_rate": 3.559348708126445e-05, "loss": 0.4428, "step": 191 }, { "epoch": 1.4719961704164672, "grad_norm": 0.7266153094875722, "learning_rate": 3.552600720658976e-05, "loss": 0.4775, "step": 192 }, { "epoch": 1.4796553374820487, "grad_norm": 0.6034805115192634, "learning_rate": 3.545807957041084e-05, "loss": 0.4353, "step": 193 }, { "epoch": 1.4873145045476304, "grad_norm": 0.733938783383156, "learning_rate": 3.538970613172332e-05, "loss": 0.492, "step": 194 }, { "epoch": 1.494973671613212, "grad_norm": 0.6587084782665951, "learning_rate": 3.532088886237956e-05, "loss": 0.4471, "step": 195 }, { "epoch": 1.5026328386787937, "grad_norm": 0.7660206322795097, "learning_rate": 3.525162974703174e-05, "loss": 0.4278, "step": 196 }, { "epoch": 1.5102920057443754, "grad_norm": 0.6137621338422651, "learning_rate": 3.518193078307463e-05, "loss": 0.4325, "step": 197 }, { "epoch": 1.517951172809957, "grad_norm": 0.5768603184148955, "learning_rate": 3.5111793980588006e-05, "loss": 0.4529, "step": 198 }, { "epoch": 1.5256103398755385, "grad_norm": 0.7888259317826345, "learning_rate": 3.5041221362278644e-05, "loss": 0.44, "step": 199 }, { "epoch": 1.5332695069411202, "grad_norm": 0.5650383799794519, "learning_rate": 3.497021496342203e-05, "loss": 0.4585, "step": 200 }, { "epoch": 1.5409286740067016, "grad_norm": 0.705765230853562, "learning_rate": 3.489877683180362e-05, "loss": 0.4616, "step": 201 }, { "epoch": 1.5485878410722833, "grad_norm": 0.6873526187506311, "learning_rate": 3.482690902765984e-05, "loss": 0.4433, "step": 202 }, { "epoch": 1.556247008137865, "grad_norm": 0.8007803368306625, "learning_rate": 3.475461362361861e-05, "loss": 0.5263, "step": 203 }, { "epoch": 1.5639061752034467, "grad_norm": 0.6918325743424758, "learning_rate": 3.468189270463959e-05, "loss": 0.4325, "step": 204 }, { "epoch": 1.5715653422690283, "grad_norm": 0.5677928436218839, "learning_rate": 3.4608748367954064e-05, "loss": 0.4654, "step": 205 }, { "epoch": 1.57922450933461, "grad_norm": 0.7751869494065315, "learning_rate": 3.4535182723004466e-05, "loss": 0.461, "step": 206 }, { "epoch": 1.5868836764001915, "grad_norm": 0.5725206135864617, "learning_rate": 3.446119789138351e-05, "loss": 0.4952, "step": 207 }, { "epoch": 1.5945428434657731, "grad_norm": 0.7296776133168584, "learning_rate": 3.438679600677303e-05, "loss": 0.4313, "step": 208 }, { "epoch": 1.6022020105313546, "grad_norm": 0.587167588865063, "learning_rate": 3.431197921488242e-05, "loss": 0.433, "step": 209 }, { "epoch": 1.6098611775969363, "grad_norm": 0.8056341147184974, "learning_rate": 3.423674967338681e-05, "loss": 0.4514, "step": 210 }, { "epoch": 1.617520344662518, "grad_norm": 0.7051970461447443, "learning_rate": 3.416110955186477e-05, "loss": 0.5431, "step": 211 }, { "epoch": 1.6251795117280996, "grad_norm": 0.5607663565934765, "learning_rate": 3.4085061031735794e-05, "loss": 0.4408, "step": 212 }, { "epoch": 1.6328386787936813, "grad_norm": 0.8246463721021458, "learning_rate": 3.4008606306197336e-05, "loss": 0.4864, "step": 213 }, { "epoch": 1.640497845859263, "grad_norm": 0.6500034356382017, "learning_rate": 3.393174758016161e-05, "loss": 0.4215, "step": 214 }, { "epoch": 1.6481570129248444, "grad_norm": 0.7062452977875294, "learning_rate": 3.385448707019199e-05, "loss": 0.4731, "step": 215 }, { "epoch": 1.655816179990426, "grad_norm": 0.5942221349239608, "learning_rate": 3.377682700443907e-05, "loss": 0.4164, "step": 216 }, { "epoch": 1.6634753470560075, "grad_norm": 0.7905876229344655, "learning_rate": 3.3698769622576404e-05, "loss": 0.5147, "step": 217 }, { "epoch": 1.6711345141215892, "grad_norm": 0.5859170654910125, "learning_rate": 3.3620317175735945e-05, "loss": 0.4275, "step": 218 }, { "epoch": 1.6787936811871709, "grad_norm": 0.6690808146840328, "learning_rate": 3.3541471926443084e-05, "loss": 0.4789, "step": 219 }, { "epoch": 1.6864528482527525, "grad_norm": 0.6561200573249233, "learning_rate": 3.34622361485514e-05, "loss": 0.4991, "step": 220 }, { "epoch": 1.6941120153183342, "grad_norm": 0.5562256326390935, "learning_rate": 3.3382612127177166e-05, "loss": 0.4343, "step": 221 }, { "epoch": 1.701771182383916, "grad_norm": 0.7170323098042807, "learning_rate": 3.330260215863332e-05, "loss": 0.4486, "step": 222 }, { "epoch": 1.7094303494494973, "grad_norm": 0.6014886240988541, "learning_rate": 3.322220855036333e-05, "loss": 0.4957, "step": 223 }, { "epoch": 1.717089516515079, "grad_norm": 0.5278357930923132, "learning_rate": 3.314143362087462e-05, "loss": 0.442, "step": 224 }, { "epoch": 1.7247486835806605, "grad_norm": 0.6035858404605409, "learning_rate": 3.30602796996717e-05, "loss": 0.4506, "step": 225 }, { "epoch": 1.7324078506462421, "grad_norm": 0.6363716300860364, "learning_rate": 3.297874912718902e-05, "loss": 0.4663, "step": 226 }, { "epoch": 1.7400670177118238, "grad_norm": 0.47737502299597623, "learning_rate": 3.2896844254723414e-05, "loss": 0.418, "step": 227 }, { "epoch": 1.7477261847774055, "grad_norm": 0.8067471733349434, "learning_rate": 3.281456744436634e-05, "loss": 0.4957, "step": 228 }, { "epoch": 1.7553853518429872, "grad_norm": 0.6089284826032468, "learning_rate": 3.273192106893572e-05, "loss": 0.4418, "step": 229 }, { "epoch": 1.7630445189085688, "grad_norm": 0.5962202659179022, "learning_rate": 3.2648907511907544e-05, "loss": 0.4669, "step": 230 }, { "epoch": 1.7707036859741503, "grad_norm": 0.5728200467314639, "learning_rate": 3.256552916734713e-05, "loss": 0.4678, "step": 231 }, { "epoch": 1.778362853039732, "grad_norm": 0.541716837396919, "learning_rate": 3.248178843984006e-05, "loss": 0.3918, "step": 232 }, { "epoch": 1.7860220201053134, "grad_norm": 0.6534786961268279, "learning_rate": 3.239768774442281e-05, "loss": 0.4922, "step": 233 }, { "epoch": 1.793681187170895, "grad_norm": 0.8358591875149838, "learning_rate": 3.2313229506513167e-05, "loss": 0.4902, "step": 234 }, { "epoch": 1.8013403542364768, "grad_norm": 0.591595847556571, "learning_rate": 3.222841616184025e-05, "loss": 0.4138, "step": 235 }, { "epoch": 1.8089995213020584, "grad_norm": 0.6277241932218716, "learning_rate": 3.2143250156374226e-05, "loss": 0.4474, "step": 236 }, { "epoch": 1.81665868836764, "grad_norm": 0.7278025713494423, "learning_rate": 3.2057733946255844e-05, "loss": 0.4709, "step": 237 }, { "epoch": 1.8243178554332218, "grad_norm": 0.6410859956043007, "learning_rate": 3.197186999772555e-05, "loss": 0.4583, "step": 238 }, { "epoch": 1.8319770224988032, "grad_norm": 0.6757146265800846, "learning_rate": 3.188566078705235e-05, "loss": 0.4704, "step": 239 }, { "epoch": 1.839636189564385, "grad_norm": 0.7572042816661453, "learning_rate": 3.1799108800462466e-05, "loss": 0.4119, "step": 240 }, { "epoch": 1.8472953566299664, "grad_norm": 0.590906796662085, "learning_rate": 3.1712216534067536e-05, "loss": 0.47, "step": 241 }, { "epoch": 1.854954523695548, "grad_norm": 0.8241586342465295, "learning_rate": 3.1624986493792735e-05, "loss": 0.4624, "step": 242 }, { "epoch": 1.8626136907611297, "grad_norm": 0.5604470218245483, "learning_rate": 3.153742119530441e-05, "loss": 0.4831, "step": 243 }, { "epoch": 1.8702728578267114, "grad_norm": 0.8158955643638057, "learning_rate": 3.144952316393758e-05, "loss": 0.4204, "step": 244 }, { "epoch": 1.877932024892293, "grad_norm": 0.5828232072086235, "learning_rate": 3.136129493462312e-05, "loss": 0.4646, "step": 245 }, { "epoch": 1.8855911919578747, "grad_norm": 0.7063142478848118, "learning_rate": 3.1272739051814594e-05, "loss": 0.4576, "step": 246 }, { "epoch": 1.8932503590234562, "grad_norm": 0.6560131266274781, "learning_rate": 3.1183858069414936e-05, "loss": 0.4537, "step": 247 }, { "epoch": 1.9009095260890378, "grad_norm": 0.5709968540583878, "learning_rate": 3.109465455070278e-05, "loss": 0.4559, "step": 248 }, { "epoch": 1.9085686931546193, "grad_norm": 0.6268646820645886, "learning_rate": 3.1005131068258506e-05, "loss": 0.3999, "step": 249 }, { "epoch": 1.916227860220201, "grad_norm": 0.48362971083345346, "learning_rate": 3.091529020389009e-05, "loss": 0.4246, "step": 250 }, { "epoch": 1.9238870272857826, "grad_norm": 0.6630981220865457, "learning_rate": 3.082513454855863e-05, "loss": 0.5091, "step": 251 }, { "epoch": 1.9315461943513643, "grad_norm": 0.670450600935426, "learning_rate": 3.073466670230361e-05, "loss": 0.3812, "step": 252 }, { "epoch": 1.939205361416946, "grad_norm": 0.552225236659151, "learning_rate": 3.0643889274167926e-05, "loss": 0.4504, "step": 253 }, { "epoch": 1.9468645284825277, "grad_norm": 0.570779364405288, "learning_rate": 3.055280488212266e-05, "loss": 0.4444, "step": 254 }, { "epoch": 1.9545236955481091, "grad_norm": 0.5961792678109743, "learning_rate": 3.0461416152991555e-05, "loss": 0.4763, "step": 255 }, { "epoch": 1.9621828626136908, "grad_norm": 0.5968250026211483, "learning_rate": 3.0369725722375274e-05, "loss": 0.4754, "step": 256 }, { "epoch": 1.9698420296792722, "grad_norm": 0.649954877349215, "learning_rate": 3.0277736234575378e-05, "loss": 0.4345, "step": 257 }, { "epoch": 1.977501196744854, "grad_norm": 0.6643346772869196, "learning_rate": 3.0185450342518075e-05, "loss": 0.4532, "step": 258 }, { "epoch": 1.9851603638104356, "grad_norm": 0.47332581564622644, "learning_rate": 3.009287070767771e-05, "loss": 0.4881, "step": 259 }, { "epoch": 1.9928195308760173, "grad_norm": 0.5015320907864502, "learning_rate": 3.0000000000000004e-05, "loss": 0.4242, "step": 260 }, { "epoch": 2.0019147917663953, "grad_norm": 0.8833235426774388, "learning_rate": 2.990684089782507e-05, "loss": 0.6896, "step": 261 }, { "epoch": 2.009573958831977, "grad_norm": 0.8871508493864276, "learning_rate": 2.9813396087810134e-05, "loss": 0.3589, "step": 262 }, { "epoch": 2.0172331258975587, "grad_norm": 1.0765160264598035, "learning_rate": 2.971966826485212e-05, "loss": 0.3823, "step": 263 }, { "epoch": 2.0248922929631403, "grad_norm": 1.2741275889363188, "learning_rate": 2.962566013200986e-05, "loss": 0.3751, "step": 264 }, { "epoch": 2.032551460028722, "grad_norm": 0.7162387782791261, "learning_rate": 2.9531374400426158e-05, "loss": 0.3232, "step": 265 }, { "epoch": 2.0402106270943037, "grad_norm": 0.972447687697029, "learning_rate": 2.943681378924964e-05, "loss": 0.3447, "step": 266 }, { "epoch": 2.047869794159885, "grad_norm": 0.7472088308314776, "learning_rate": 2.934198102555631e-05, "loss": 0.3502, "step": 267 }, { "epoch": 2.0555289612254666, "grad_norm": 0.9904131833727945, "learning_rate": 2.924687884427087e-05, "loss": 0.3328, "step": 268 }, { "epoch": 2.0631881282910483, "grad_norm": 0.8268717893678047, "learning_rate": 2.9151509988087912e-05, "loss": 0.323, "step": 269 }, { "epoch": 2.07084729535663, "grad_norm": 1.0133251834372121, "learning_rate": 2.9055877207392752e-05, "loss": 0.3314, "step": 270 }, { "epoch": 2.0785064624222116, "grad_norm": 0.9295212140684717, "learning_rate": 2.8959983260182166e-05, "loss": 0.3433, "step": 271 }, { "epoch": 2.0861656294877933, "grad_norm": 0.7493423793907141, "learning_rate": 2.886383091198483e-05, "loss": 0.3174, "step": 272 }, { "epoch": 2.093824796553375, "grad_norm": 1.1449594495047748, "learning_rate": 2.876742293578155e-05, "loss": 0.4027, "step": 273 }, { "epoch": 2.1014839636189566, "grad_norm": 0.8495410462496846, "learning_rate": 2.8670762111925313e-05, "loss": 0.3762, "step": 274 }, { "epoch": 2.109143130684538, "grad_norm": 0.7356070365983308, "learning_rate": 2.8573851228061084e-05, "loss": 0.3116, "step": 275 }, { "epoch": 2.1168022977501195, "grad_norm": 0.7959446709248563, "learning_rate": 2.8476693079045432e-05, "loss": 0.3379, "step": 276 }, { "epoch": 2.124461464815701, "grad_norm": 0.8100378872218936, "learning_rate": 2.8379290466865906e-05, "loss": 0.3305, "step": 277 }, { "epoch": 2.132120631881283, "grad_norm": 0.8085735074400692, "learning_rate": 2.828164620056024e-05, "loss": 0.3598, "step": 278 }, { "epoch": 2.1397797989468645, "grad_norm": 0.6344360858538197, "learning_rate": 2.818376309613535e-05, "loss": 0.302, "step": 279 }, { "epoch": 2.147438966012446, "grad_norm": 0.57409154364413, "learning_rate": 2.80856439764861e-05, "loss": 0.3686, "step": 280 }, { "epoch": 2.155098133078028, "grad_norm": 0.5845504651635646, "learning_rate": 2.798729167131391e-05, "loss": 0.3526, "step": 281 }, { "epoch": 2.1627573001436096, "grad_norm": 0.4776801469885686, "learning_rate": 2.7888709017045146e-05, "loss": 0.3326, "step": 282 }, { "epoch": 2.170416467209191, "grad_norm": 0.48684077839230894, "learning_rate": 2.7789898856749297e-05, "loss": 0.3342, "step": 283 }, { "epoch": 2.1780756342747725, "grad_norm": 0.49773085707991915, "learning_rate": 2.7690864040057023e-05, "loss": 0.3051, "step": 284 }, { "epoch": 2.185734801340354, "grad_norm": 0.4525197933418408, "learning_rate": 2.7591607423077932e-05, "loss": 0.3087, "step": 285 }, { "epoch": 2.193393968405936, "grad_norm": 0.5639448536870705, "learning_rate": 2.7492131868318247e-05, "loss": 0.3503, "step": 286 }, { "epoch": 2.2010531354715175, "grad_norm": 0.5607287965131402, "learning_rate": 2.739244024459822e-05, "loss": 0.3654, "step": 287 }, { "epoch": 2.208712302537099, "grad_norm": 0.4515128468379325, "learning_rate": 2.7292535426969436e-05, "loss": 0.3149, "step": 288 }, { "epoch": 2.216371469602681, "grad_norm": 0.6011149807368861, "learning_rate": 2.7192420296631835e-05, "loss": 0.3911, "step": 289 }, { "epoch": 2.2240306366682625, "grad_norm": 0.4554658307562261, "learning_rate": 2.7092097740850712e-05, "loss": 0.3068, "step": 290 }, { "epoch": 2.231689803733844, "grad_norm": 0.5520804650288836, "learning_rate": 2.6991570652873357e-05, "loss": 0.3681, "step": 291 }, { "epoch": 2.2393489707994254, "grad_norm": 0.48661650666410344, "learning_rate": 2.6890841931845674e-05, "loss": 0.3022, "step": 292 }, { "epoch": 2.247008137865007, "grad_norm": 0.5416624941038123, "learning_rate": 2.6789914482728546e-05, "loss": 0.359, "step": 293 }, { "epoch": 2.2546673049305888, "grad_norm": 0.582685239343103, "learning_rate": 2.6688791216214064e-05, "loss": 0.3693, "step": 294 }, { "epoch": 2.2623264719961704, "grad_norm": 0.4568416200895835, "learning_rate": 2.6587475048641596e-05, "loss": 0.32, "step": 295 }, { "epoch": 2.269985639061752, "grad_norm": 0.6028222179058447, "learning_rate": 2.6485968901913658e-05, "loss": 0.3726, "step": 296 }, { "epoch": 2.2776448061273338, "grad_norm": 0.5497203840683457, "learning_rate": 2.6384275703411666e-05, "loss": 0.3584, "step": 297 }, { "epoch": 2.2853039731929154, "grad_norm": 0.4601068271433548, "learning_rate": 2.6282398385911503e-05, "loss": 0.3263, "step": 298 }, { "epoch": 2.2929631402584967, "grad_norm": 0.5843782768173683, "learning_rate": 2.618033988749895e-05, "loss": 0.3854, "step": 299 }, { "epoch": 2.3006223073240784, "grad_norm": 0.4668138387272922, "learning_rate": 2.607810315148494e-05, "loss": 0.3445, "step": 300 }, { "epoch": 2.30828147438966, "grad_norm": 0.4966561419101829, "learning_rate": 2.5975691126320678e-05, "loss": 0.3486, "step": 301 }, { "epoch": 2.3159406414552417, "grad_norm": 0.552195612826531, "learning_rate": 2.587310676551262e-05, "loss": 0.3478, "step": 302 }, { "epoch": 2.3235998085208234, "grad_norm": 0.49951171616233425, "learning_rate": 2.5770353027537276e-05, "loss": 0.3494, "step": 303 }, { "epoch": 2.331258975586405, "grad_norm": 0.4487518241170282, "learning_rate": 2.5667432875755904e-05, "loss": 0.3267, "step": 304 }, { "epoch": 2.3389181426519867, "grad_norm": 0.47306289952945507, "learning_rate": 2.5564349278329056e-05, "loss": 0.3246, "step": 305 }, { "epoch": 2.3465773097175684, "grad_norm": 0.4974553141384758, "learning_rate": 2.5461105208130953e-05, "loss": 0.2918, "step": 306 }, { "epoch": 2.35423647678315, "grad_norm": 0.4809251554335802, "learning_rate": 2.5357703642663766e-05, "loss": 0.357, "step": 307 }, { "epoch": 2.3618956438487313, "grad_norm": 0.5596662616004654, "learning_rate": 2.525414756397174e-05, "loss": 0.3287, "step": 308 }, { "epoch": 2.369554810914313, "grad_norm": 0.46194307500162646, "learning_rate": 2.5150439958555205e-05, "loss": 0.323, "step": 309 }, { "epoch": 2.3772139779798946, "grad_norm": 0.5354975145364615, "learning_rate": 2.5046583817284437e-05, "loss": 0.3641, "step": 310 }, { "epoch": 2.3848731450454763, "grad_norm": 0.5273653633413932, "learning_rate": 2.4942582135313393e-05, "loss": 0.3553, "step": 311 }, { "epoch": 2.392532312111058, "grad_norm": 0.49817920339367244, "learning_rate": 2.4838437911993355e-05, "loss": 0.3455, "step": 312 }, { "epoch": 2.4001914791766397, "grad_norm": 0.515170607443925, "learning_rate": 2.473415415078642e-05, "loss": 0.3619, "step": 313 }, { "epoch": 2.4078506462422213, "grad_norm": 0.4926042303192558, "learning_rate": 2.4629733859178867e-05, "loss": 0.3485, "step": 314 }, { "epoch": 2.4155098133078026, "grad_norm": 0.42905408480479645, "learning_rate": 2.4525180048594452e-05, "loss": 0.3057, "step": 315 }, { "epoch": 2.4231689803733842, "grad_norm": 0.4867004151742306, "learning_rate": 2.4420495734307527e-05, "loss": 0.3747, "step": 316 }, { "epoch": 2.430828147438966, "grad_norm": 0.41338427949555484, "learning_rate": 2.4315683935356127e-05, "loss": 0.3475, "step": 317 }, { "epoch": 2.4384873145045476, "grad_norm": 0.44888986646854223, "learning_rate": 2.421074767445485e-05, "loss": 0.303, "step": 318 }, { "epoch": 2.4461464815701293, "grad_norm": 0.5491552096723435, "learning_rate": 2.4105689977907722e-05, "loss": 0.3979, "step": 319 }, { "epoch": 2.453805648635711, "grad_norm": 0.4927516444394549, "learning_rate": 2.4000513875520892e-05, "loss": 0.3407, "step": 320 }, { "epoch": 2.4614648157012926, "grad_norm": 0.466811974445456, "learning_rate": 2.3895222400515282e-05, "loss": 0.3425, "step": 321 }, { "epoch": 2.4691239827668743, "grad_norm": 0.47821175171844904, "learning_rate": 2.3789818589439094e-05, "loss": 0.3098, "step": 322 }, { "epoch": 2.476783149832456, "grad_norm": 0.4912242879477096, "learning_rate": 2.3684305482080233e-05, "loss": 0.3591, "step": 323 }, { "epoch": 2.484442316898037, "grad_norm": 0.4811573112209665, "learning_rate": 2.357868612137866e-05, "loss": 0.3773, "step": 324 }, { "epoch": 2.492101483963619, "grad_norm": 0.4870566155608247, "learning_rate": 2.3472963553338614e-05, "loss": 0.3228, "step": 325 }, { "epoch": 2.4997606510292005, "grad_norm": 0.5310392825279873, "learning_rate": 2.3367140826940768e-05, "loss": 0.3546, "step": 326 }, { "epoch": 2.507419818094782, "grad_norm": 0.5483105250492419, "learning_rate": 2.326122099405435e-05, "loss": 0.3375, "step": 327 }, { "epoch": 2.515078985160364, "grad_norm": 0.5003667793105148, "learning_rate": 2.315520710934903e-05, "loss": 0.3687, "step": 328 }, { "epoch": 2.5227381522259456, "grad_norm": 0.533844789442384, "learning_rate": 2.304910223020691e-05, "loss": 0.3548, "step": 329 }, { "epoch": 2.5303973192915272, "grad_norm": 0.5149064035903343, "learning_rate": 2.2942909416634326e-05, "loss": 0.3446, "step": 330 }, { "epoch": 2.5380564863571085, "grad_norm": 0.4151169766555542, "learning_rate": 2.2836631731173577e-05, "loss": 0.3281, "step": 331 }, { "epoch": 2.54571565342269, "grad_norm": 0.5156566437281681, "learning_rate": 2.2730272238814636e-05, "loss": 0.3479, "step": 332 }, { "epoch": 2.553374820488272, "grad_norm": 0.536538863268432, "learning_rate": 2.2623834006906732e-05, "loss": 0.3468, "step": 333 }, { "epoch": 2.5610339875538535, "grad_norm": 0.41537339391659595, "learning_rate": 2.25173201050699e-05, "loss": 0.311, "step": 334 }, { "epoch": 2.568693154619435, "grad_norm": 0.5021990261148753, "learning_rate": 2.2410733605106462e-05, "loss": 0.3803, "step": 335 }, { "epoch": 2.576352321685017, "grad_norm": 0.39621064090295866, "learning_rate": 2.2304077580912423e-05, "loss": 0.2936, "step": 336 }, { "epoch": 2.5840114887505985, "grad_norm": 0.5372701642790138, "learning_rate": 2.2197355108388835e-05, "loss": 0.3184, "step": 337 }, { "epoch": 2.5916706558161797, "grad_norm": 0.41383864714124363, "learning_rate": 2.209056926535307e-05, "loss": 0.3287, "step": 338 }, { "epoch": 2.599329822881762, "grad_norm": 0.5114947611293079, "learning_rate": 2.1983723131450088e-05, "loss": 0.3683, "step": 339 }, { "epoch": 2.606988989947343, "grad_norm": 0.4173762346729875, "learning_rate": 2.1876819788063586e-05, "loss": 0.3086, "step": 340 }, { "epoch": 2.6146481570129247, "grad_norm": 0.5192212848345572, "learning_rate": 2.176986231822717e-05, "loss": 0.375, "step": 341 }, { "epoch": 2.6223073240785064, "grad_norm": 0.5669533936852463, "learning_rate": 2.166285380653541e-05, "loss": 0.3432, "step": 342 }, { "epoch": 2.629966491144088, "grad_norm": 0.4671450099921592, "learning_rate": 2.1555797339054898e-05, "loss": 0.3207, "step": 343 }, { "epoch": 2.6376256582096698, "grad_norm": 0.5035698279484611, "learning_rate": 2.1448696003235252e-05, "loss": 0.3586, "step": 344 }, { "epoch": 2.6452848252752514, "grad_norm": 0.49482528595990216, "learning_rate": 2.1341552887820048e-05, "loss": 0.3419, "step": 345 }, { "epoch": 2.652943992340833, "grad_norm": 0.5124832493661444, "learning_rate": 2.123437108275779e-05, "loss": 0.3605, "step": 346 }, { "epoch": 2.6606031594064143, "grad_norm": 0.3916326803311756, "learning_rate": 2.112715367911275e-05, "loss": 0.3102, "step": 347 }, { "epoch": 2.668262326471996, "grad_norm": 0.41173103844867237, "learning_rate": 2.1019903768975852e-05, "loss": 0.3645, "step": 348 }, { "epoch": 2.6759214935375777, "grad_norm": 0.48815468984239224, "learning_rate": 2.0912624445375483e-05, "loss": 0.3641, "step": 349 }, { "epoch": 2.6835806606031594, "grad_norm": 0.40031092411352465, "learning_rate": 2.0805318802188307e-05, "loss": 0.3233, "step": 350 }, { "epoch": 2.691239827668741, "grad_norm": 0.4801007606225588, "learning_rate": 2.0697989934050025e-05, "loss": 0.3492, "step": 351 }, { "epoch": 2.6988989947343227, "grad_norm": 0.465959760136513, "learning_rate": 2.0590640936266132e-05, "loss": 0.3492, "step": 352 }, { "epoch": 2.7065581617999044, "grad_norm": 0.432367828293715, "learning_rate": 2.0483274904722647e-05, "loss": 0.3478, "step": 353 }, { "epoch": 2.7142173288654856, "grad_norm": 0.4510586932875561, "learning_rate": 2.037589493579685e-05, "loss": 0.3553, "step": 354 }, { "epoch": 2.7218764959310677, "grad_norm": 0.467950300228232, "learning_rate": 2.0268504126267952e-05, "loss": 0.3653, "step": 355 }, { "epoch": 2.729535662996649, "grad_norm": 0.5009397300600292, "learning_rate": 2.0161105573227798e-05, "loss": 0.36, "step": 356 }, { "epoch": 2.7371948300622306, "grad_norm": 0.49408159207397456, "learning_rate": 2.005370237399157e-05, "loss": 0.3515, "step": 357 }, { "epoch": 2.7448539971278123, "grad_norm": 0.35262515813271533, "learning_rate": 1.9946297626008432e-05, "loss": 0.2689, "step": 358 }, { "epoch": 2.752513164193394, "grad_norm": 0.4827674543765931, "learning_rate": 1.9838894426772205e-05, "loss": 0.3706, "step": 359 }, { "epoch": 2.7601723312589757, "grad_norm": 0.3780483362953392, "learning_rate": 1.9731495873732055e-05, "loss": 0.3135, "step": 360 }, { "epoch": 2.7678314983245573, "grad_norm": 0.3811396663153698, "learning_rate": 1.9624105064203157e-05, "loss": 0.3576, "step": 361 }, { "epoch": 2.775490665390139, "grad_norm": 0.45864928956878004, "learning_rate": 1.951672509527736e-05, "loss": 0.3321, "step": 362 }, { "epoch": 2.7831498324557202, "grad_norm": 0.4507638328421475, "learning_rate": 1.940935906373388e-05, "loss": 0.3715, "step": 363 }, { "epoch": 2.790808999521302, "grad_norm": 0.43134511596968567, "learning_rate": 1.930201006594999e-05, "loss": 0.3412, "step": 364 }, { "epoch": 2.7984681665868836, "grad_norm": 0.52057087403904, "learning_rate": 1.9194681197811703e-05, "loss": 0.3849, "step": 365 }, { "epoch": 2.8061273336524652, "grad_norm": 0.37380617464599, "learning_rate": 1.9087375554624527e-05, "loss": 0.3072, "step": 366 }, { "epoch": 2.813786500718047, "grad_norm": 0.3665120285057917, "learning_rate": 1.898009623102415e-05, "loss": 0.3261, "step": 367 }, { "epoch": 2.8214456677836286, "grad_norm": 0.45855248957112604, "learning_rate": 1.887284632088725e-05, "loss": 0.4029, "step": 368 }, { "epoch": 2.8291048348492103, "grad_norm": 0.42598192016232783, "learning_rate": 1.8765628917242213e-05, "loss": 0.3588, "step": 369 }, { "epoch": 2.8367640019147915, "grad_norm": 0.4155845765749022, "learning_rate": 1.8658447112179952e-05, "loss": 0.3039, "step": 370 }, { "epoch": 2.8444231689803736, "grad_norm": 0.4601308365256824, "learning_rate": 1.8551303996764755e-05, "loss": 0.3836, "step": 371 }, { "epoch": 2.852082336045955, "grad_norm": 0.4914229018003968, "learning_rate": 1.8444202660945105e-05, "loss": 0.3679, "step": 372 }, { "epoch": 2.8597415031115365, "grad_norm": 0.46287822342404056, "learning_rate": 1.8337146193464595e-05, "loss": 0.288, "step": 373 }, { "epoch": 2.867400670177118, "grad_norm": 0.49183098229163497, "learning_rate": 1.8230137681772836e-05, "loss": 0.3741, "step": 374 }, { "epoch": 2.8750598372427, "grad_norm": 0.40911104249164226, "learning_rate": 1.8123180211936417e-05, "loss": 0.3425, "step": 375 }, { "epoch": 2.8827190043082815, "grad_norm": 0.4656657585294656, "learning_rate": 1.801627686854992e-05, "loss": 0.3436, "step": 376 }, { "epoch": 2.890378171373863, "grad_norm": 0.3955542489042361, "learning_rate": 1.7909430734646936e-05, "loss": 0.3142, "step": 377 }, { "epoch": 2.898037338439445, "grad_norm": 0.3919921799832172, "learning_rate": 1.780264489161117e-05, "loss": 0.327, "step": 378 }, { "epoch": 2.905696505505026, "grad_norm": 0.40565680454375597, "learning_rate": 1.769592241908758e-05, "loss": 0.3469, "step": 379 }, { "epoch": 2.913355672570608, "grad_norm": 0.4612451296187882, "learning_rate": 1.758926639489354e-05, "loss": 0.398, "step": 380 }, { "epoch": 2.9210148396361895, "grad_norm": 0.374323134300942, "learning_rate": 1.748267989493011e-05, "loss": 0.3077, "step": 381 }, { "epoch": 2.928674006701771, "grad_norm": 0.447528401086559, "learning_rate": 1.7376165993093278e-05, "loss": 0.3791, "step": 382 }, { "epoch": 2.936333173767353, "grad_norm": 0.3684289290106699, "learning_rate": 1.7269727761185374e-05, "loss": 0.3203, "step": 383 }, { "epoch": 2.9439923408329345, "grad_norm": 0.3713011538957981, "learning_rate": 1.7163368268826433e-05, "loss": 0.3247, "step": 384 }, { "epoch": 2.951651507898516, "grad_norm": 0.4288353235268095, "learning_rate": 1.7057090583365678e-05, "loss": 0.3714, "step": 385 }, { "epoch": 2.9593106749640974, "grad_norm": 0.35222146097913287, "learning_rate": 1.6950897769793093e-05, "loss": 0.2943, "step": 386 }, { "epoch": 2.9669698420296795, "grad_norm": 0.38643823068088784, "learning_rate": 1.6844792890650976e-05, "loss": 0.3443, "step": 387 }, { "epoch": 2.9746290090952607, "grad_norm": 0.38938210740713053, "learning_rate": 1.673877900594566e-05, "loss": 0.3399, "step": 388 }, { "epoch": 2.9822881761608424, "grad_norm": 0.43159107146354114, "learning_rate": 1.6632859173059232e-05, "loss": 0.344, "step": 389 }, { "epoch": 2.989947343226424, "grad_norm": 0.38084254373242143, "learning_rate": 1.6527036446661396e-05, "loss": 0.3455, "step": 390 }, { "epoch": 2.9976065102920058, "grad_norm": 0.5938797732009524, "learning_rate": 1.6421313878621344e-05, "loss": 0.4699, "step": 391 }, { "epoch": 3.006701771182384, "grad_norm": 0.6492454140958951, "learning_rate": 1.631569451791977e-05, "loss": 0.3313, "step": 392 }, { "epoch": 3.0143609382479655, "grad_norm": 0.4219633362824025, "learning_rate": 1.6210181410560912e-05, "loss": 0.2324, "step": 393 }, { "epoch": 3.022020105313547, "grad_norm": 0.5182168578367924, "learning_rate": 1.610477759948472e-05, "loss": 0.245, "step": 394 }, { "epoch": 3.029679272379129, "grad_norm": 0.6377564861463809, "learning_rate": 1.5999486124479115e-05, "loss": 0.2515, "step": 395 }, { "epoch": 3.0373384394447105, "grad_norm": 0.5098950195718823, "learning_rate": 1.5894310022092288e-05, "loss": 0.2364, "step": 396 }, { "epoch": 3.044997606510292, "grad_norm": 0.5168454095271555, "learning_rate": 1.5789252325545157e-05, "loss": 0.2334, "step": 397 }, { "epoch": 3.052656773575874, "grad_norm": 0.5240320586315429, "learning_rate": 1.568431606464388e-05, "loss": 0.2675, "step": 398 }, { "epoch": 3.060315940641455, "grad_norm": 0.44117429526871266, "learning_rate": 1.557950426569248e-05, "loss": 0.2315, "step": 399 }, { "epoch": 3.0679751077070367, "grad_norm": 0.48765308835045107, "learning_rate": 1.547481995140556e-05, "loss": 0.2437, "step": 400 }, { "epoch": 3.0756342747726184, "grad_norm": 0.4520916642343294, "learning_rate": 1.5370266140821143e-05, "loss": 0.2341, "step": 401 }, { "epoch": 3.0832934418382, "grad_norm": 0.4647186437171434, "learning_rate": 1.5265845849213588e-05, "loss": 0.2468, "step": 402 }, { "epoch": 3.0909526089037818, "grad_norm": 0.46140225100955573, "learning_rate": 1.5161562088006649e-05, "loss": 0.2345, "step": 403 }, { "epoch": 3.0986117759693634, "grad_norm": 0.4382412999538287, "learning_rate": 1.5057417864686607e-05, "loss": 0.2386, "step": 404 }, { "epoch": 3.106270943034945, "grad_norm": 0.45590994359507603, "learning_rate": 1.4953416182715566e-05, "loss": 0.2443, "step": 405 }, { "epoch": 3.113930110100527, "grad_norm": 0.44632890994369173, "learning_rate": 1.4849560041444795e-05, "loss": 0.2216, "step": 406 }, { "epoch": 3.121589277166108, "grad_norm": 0.3690202710694795, "learning_rate": 1.4745852436028262e-05, "loss": 0.2289, "step": 407 }, { "epoch": 3.1292484442316897, "grad_norm": 0.3987565576025537, "learning_rate": 1.464229635733624e-05, "loss": 0.2097, "step": 408 }, { "epoch": 3.1369076112972714, "grad_norm": 0.3922785950430538, "learning_rate": 1.4538894791869052e-05, "loss": 0.2205, "step": 409 }, { "epoch": 3.144566778362853, "grad_norm": 0.4019487118753477, "learning_rate": 1.443565072167095e-05, "loss": 0.2536, "step": 410 }, { "epoch": 3.1522259454284347, "grad_norm": 0.37368100647725233, "learning_rate": 1.43325671242441e-05, "loss": 0.2514, "step": 411 }, { "epoch": 3.1598851124940164, "grad_norm": 0.3884417771257955, "learning_rate": 1.4229646972462732e-05, "loss": 0.2456, "step": 412 }, { "epoch": 3.167544279559598, "grad_norm": 0.3767796290523846, "learning_rate": 1.412689323448739e-05, "loss": 0.2237, "step": 413 }, { "epoch": 3.1752034466251793, "grad_norm": 0.3674660595771459, "learning_rate": 1.4024308873679327e-05, "loss": 0.2591, "step": 414 }, { "epoch": 3.182862613690761, "grad_norm": 0.3572633760098233, "learning_rate": 1.3921896848515064e-05, "loss": 0.2163, "step": 415 }, { "epoch": 3.1905217807563426, "grad_norm": 0.381550127572975, "learning_rate": 1.3819660112501054e-05, "loss": 0.2442, "step": 416 }, { "epoch": 3.1981809478219243, "grad_norm": 0.37631597527744115, "learning_rate": 1.37176016140885e-05, "loss": 0.2323, "step": 417 }, { "epoch": 3.205840114887506, "grad_norm": 0.35956143405057833, "learning_rate": 1.3615724296588342e-05, "loss": 0.2044, "step": 418 }, { "epoch": 3.2134992819530876, "grad_norm": 0.38268492786329705, "learning_rate": 1.3514031098086349e-05, "loss": 0.2402, "step": 419 }, { "epoch": 3.2211584490186693, "grad_norm": 0.38994977473417125, "learning_rate": 1.341252495135841e-05, "loss": 0.2319, "step": 420 }, { "epoch": 3.228817616084251, "grad_norm": 0.35677922752266494, "learning_rate": 1.3311208783785945e-05, "loss": 0.2341, "step": 421 }, { "epoch": 3.2364767831498327, "grad_norm": 0.37528548994142574, "learning_rate": 1.3210085517271459e-05, "loss": 0.25, "step": 422 }, { "epoch": 3.244135950215414, "grad_norm": 0.382737915829752, "learning_rate": 1.3109158068154329e-05, "loss": 0.2501, "step": 423 }, { "epoch": 3.2517951172809956, "grad_norm": 0.3540547128374988, "learning_rate": 1.3008429347126641e-05, "loss": 0.2482, "step": 424 }, { "epoch": 3.2594542843465772, "grad_norm": 0.3474058116435946, "learning_rate": 1.2907902259149287e-05, "loss": 0.2359, "step": 425 }, { "epoch": 3.267113451412159, "grad_norm": 0.3903729184634132, "learning_rate": 1.2807579703368162e-05, "loss": 0.2122, "step": 426 }, { "epoch": 3.2747726184777406, "grad_norm": 0.40188660683701266, "learning_rate": 1.2707464573030572e-05, "loss": 0.2884, "step": 427 }, { "epoch": 3.2824317855433223, "grad_norm": 0.3471695792330223, "learning_rate": 1.260755975540178e-05, "loss": 0.217, "step": 428 }, { "epoch": 3.290090952608904, "grad_norm": 0.36589717504907115, "learning_rate": 1.250786813168176e-05, "loss": 0.2379, "step": 429 }, { "epoch": 3.297750119674485, "grad_norm": 0.3688472677382392, "learning_rate": 1.2408392576922075e-05, "loss": 0.2317, "step": 430 }, { "epoch": 3.305409286740067, "grad_norm": 0.3822664640855971, "learning_rate": 1.2309135959942986e-05, "loss": 0.2376, "step": 431 }, { "epoch": 3.3130684538056485, "grad_norm": 0.35469891768647493, "learning_rate": 1.2210101143250708e-05, "loss": 0.2183, "step": 432 }, { "epoch": 3.32072762087123, "grad_norm": 0.3609442874261945, "learning_rate": 1.211129098295486e-05, "loss": 0.2372, "step": 433 }, { "epoch": 3.328386787936812, "grad_norm": 0.40218147421255007, "learning_rate": 1.2012708328686093e-05, "loss": 0.2722, "step": 434 }, { "epoch": 3.3360459550023935, "grad_norm": 0.34284386522492444, "learning_rate": 1.1914356023513904e-05, "loss": 0.2249, "step": 435 }, { "epoch": 3.343705122067975, "grad_norm": 0.3398937321234149, "learning_rate": 1.1816236903864656e-05, "loss": 0.2314, "step": 436 }, { "epoch": 3.351364289133557, "grad_norm": 0.3645220719375267, "learning_rate": 1.1718353799439766e-05, "loss": 0.2415, "step": 437 }, { "epoch": 3.3590234561991386, "grad_norm": 0.35966116512097646, "learning_rate": 1.1620709533134104e-05, "loss": 0.2539, "step": 438 }, { "epoch": 3.36668262326472, "grad_norm": 0.35249135244554275, "learning_rate": 1.1523306920954571e-05, "loss": 0.207, "step": 439 }, { "epoch": 3.3743417903303015, "grad_norm": 0.350527349824925, "learning_rate": 1.1426148771938915e-05, "loss": 0.215, "step": 440 }, { "epoch": 3.382000957395883, "grad_norm": 0.3645807698031509, "learning_rate": 1.1329237888074691e-05, "loss": 0.2294, "step": 441 }, { "epoch": 3.389660124461465, "grad_norm": 0.3526979558756505, "learning_rate": 1.123257706421845e-05, "loss": 0.2357, "step": 442 }, { "epoch": 3.3973192915270465, "grad_norm": 0.3298007139758826, "learning_rate": 1.1136169088015177e-05, "loss": 0.2196, "step": 443 }, { "epoch": 3.404978458592628, "grad_norm": 0.3901583239274404, "learning_rate": 1.1040016739817836e-05, "loss": 0.2593, "step": 444 }, { "epoch": 3.41263762565821, "grad_norm": 0.3312336548111748, "learning_rate": 1.094412279260726e-05, "loss": 0.2201, "step": 445 }, { "epoch": 3.420296792723791, "grad_norm": 0.37712815006038436, "learning_rate": 1.0848490011912096e-05, "loss": 0.2353, "step": 446 }, { "epoch": 3.4279559597893727, "grad_norm": 0.33427689171156727, "learning_rate": 1.0753121155729133e-05, "loss": 0.2247, "step": 447 }, { "epoch": 3.4356151268549544, "grad_norm": 0.3403837414551559, "learning_rate": 1.0658018974443692e-05, "loss": 0.2334, "step": 448 }, { "epoch": 3.443274293920536, "grad_norm": 0.3427488817401635, "learning_rate": 1.056318621075036e-05, "loss": 0.2394, "step": 449 }, { "epoch": 3.4509334609861178, "grad_norm": 0.35721031062542025, "learning_rate": 1.0468625599573842e-05, "loss": 0.238, "step": 450 }, { "epoch": 3.4585926280516994, "grad_norm": 0.35498625884545276, "learning_rate": 1.037433986799015e-05, "loss": 0.248, "step": 451 }, { "epoch": 3.466251795117281, "grad_norm": 0.3686856880290374, "learning_rate": 1.028033173514788e-05, "loss": 0.2602, "step": 452 }, { "epoch": 3.4739109621828628, "grad_norm": 0.32569736084610973, "learning_rate": 1.0186603912189867e-05, "loss": 0.2261, "step": 453 }, { "epoch": 3.4815701292484444, "grad_norm": 0.35179935045391797, "learning_rate": 1.0093159102174938e-05, "loss": 0.228, "step": 454 }, { "epoch": 3.4892292963140257, "grad_norm": 0.37412117087411056, "learning_rate": 1.0000000000000006e-05, "loss": 0.2558, "step": 455 }, { "epoch": 3.4968884633796073, "grad_norm": 0.3196396284614333, "learning_rate": 9.907129292322298e-06, "loss": 0.2016, "step": 456 }, { "epoch": 3.504547630445189, "grad_norm": 0.367949651744729, "learning_rate": 9.814549657481935e-06, "loss": 0.2345, "step": 457 }, { "epoch": 3.5122067975107707, "grad_norm": 0.36477967052612537, "learning_rate": 9.722263765424628e-06, "loss": 0.2276, "step": 458 }, { "epoch": 3.5198659645763524, "grad_norm": 0.3332594226268887, "learning_rate": 9.630274277624729e-06, "loss": 0.23, "step": 459 }, { "epoch": 3.527525131641934, "grad_norm": 0.3477521018845692, "learning_rate": 9.538583847008452e-06, "loss": 0.2352, "step": 460 }, { "epoch": 3.5351842987075157, "grad_norm": 0.32086860442292786, "learning_rate": 9.447195117877343e-06, "loss": 0.2125, "step": 461 }, { "epoch": 3.542843465773097, "grad_norm": 0.3678326552292364, "learning_rate": 9.356110725832081e-06, "loss": 0.2576, "step": 462 }, { "epoch": 3.550502632838679, "grad_norm": 0.3656606530951392, "learning_rate": 9.265333297696395e-06, "loss": 0.2461, "step": 463 }, { "epoch": 3.5581617999042603, "grad_norm": 0.3229274366638359, "learning_rate": 9.174865451441375e-06, "loss": 0.2298, "step": 464 }, { "epoch": 3.565820966969842, "grad_norm": 0.32293232206607697, "learning_rate": 9.084709796109907e-06, "loss": 0.2166, "step": 465 }, { "epoch": 3.5734801340354236, "grad_norm": 0.3461993603233461, "learning_rate": 8.994868931741499e-06, "loss": 0.2342, "step": 466 }, { "epoch": 3.5811393011010053, "grad_norm": 0.31931497478868687, "learning_rate": 8.905345449297223e-06, "loss": 0.2245, "step": 467 }, { "epoch": 3.588798468166587, "grad_norm": 0.34882777970263285, "learning_rate": 8.816141930585067e-06, "loss": 0.2412, "step": 468 }, { "epoch": 3.5964576352321687, "grad_norm": 0.31810617065052815, "learning_rate": 8.72726094818541e-06, "loss": 0.2202, "step": 469 }, { "epoch": 3.6041168022977503, "grad_norm": 0.36687350590070866, "learning_rate": 8.638705065376887e-06, "loss": 0.225, "step": 470 }, { "epoch": 3.6117759693633316, "grad_norm": 0.35726654000876984, "learning_rate": 8.550476836062419e-06, "loss": 0.2343, "step": 471 }, { "epoch": 3.6194351364289132, "grad_norm": 0.33724954427539455, "learning_rate": 8.462578804695595e-06, "loss": 0.2328, "step": 472 }, { "epoch": 3.627094303494495, "grad_norm": 0.33977950988225286, "learning_rate": 8.375013506207275e-06, "loss": 0.2593, "step": 473 }, { "epoch": 3.6347534705600766, "grad_norm": 0.3357461336976667, "learning_rate": 8.287783465932466e-06, "loss": 0.2162, "step": 474 }, { "epoch": 3.6424126376256583, "grad_norm": 0.33938581871332074, "learning_rate": 8.200891199537549e-06, "loss": 0.2159, "step": 475 }, { "epoch": 3.65007180469124, "grad_norm": 0.30850555682034936, "learning_rate": 8.114339212947655e-06, "loss": 0.2308, "step": 476 }, { "epoch": 3.6577309717568216, "grad_norm": 0.33712030493831774, "learning_rate": 8.028130002274459e-06, "loss": 0.2617, "step": 477 }, { "epoch": 3.665390138822403, "grad_norm": 0.33952468314921025, "learning_rate": 7.942266053744155e-06, "loss": 0.2277, "step": 478 }, { "epoch": 3.673049305887985, "grad_norm": 0.32790624314094596, "learning_rate": 7.856749843625777e-06, "loss": 0.2459, "step": 479 }, { "epoch": 3.680708472953566, "grad_norm": 0.30426795240880533, "learning_rate": 7.771583838159756e-06, "loss": 0.2373, "step": 480 }, { "epoch": 3.688367640019148, "grad_norm": 0.32139830397764274, "learning_rate": 7.686770493486835e-06, "loss": 0.265, "step": 481 }, { "epoch": 3.6960268070847295, "grad_norm": 0.30384977031133725, "learning_rate": 7.602312255577193e-06, "loss": 0.2166, "step": 482 }, { "epoch": 3.703685974150311, "grad_norm": 0.33085512515068155, "learning_rate": 7.518211560159949e-06, "loss": 0.2511, "step": 483 }, { "epoch": 3.711345141215893, "grad_norm": 0.3176611144031733, "learning_rate": 7.434470832652865e-06, "loss": 0.2084, "step": 484 }, { "epoch": 3.7190043082814745, "grad_norm": 0.33574434095691075, "learning_rate": 7.3510924880924575e-06, "loss": 0.2308, "step": 485 }, { "epoch": 3.726663475347056, "grad_norm": 0.2971615945594848, "learning_rate": 7.268078931064293e-06, "loss": 0.2217, "step": 486 }, { "epoch": 3.7343226424126374, "grad_norm": 0.32961234977638115, "learning_rate": 7.185432555633672e-06, "loss": 0.2548, "step": 487 }, { "epoch": 3.741981809478219, "grad_norm": 0.32814885991409504, "learning_rate": 7.1031557452765934e-06, "loss": 0.2439, "step": 488 }, { "epoch": 3.749640976543801, "grad_norm": 0.34360104917258844, "learning_rate": 7.021250872810983e-06, "loss": 0.224, "step": 489 }, { "epoch": 3.7573001436093825, "grad_norm": 0.3505861571637595, "learning_rate": 6.939720300328303e-06, "loss": 0.2459, "step": 490 }, { "epoch": 3.764959310674964, "grad_norm": 0.33976097846578596, "learning_rate": 6.858566379125389e-06, "loss": 0.2329, "step": 491 }, { "epoch": 3.772618477740546, "grad_norm": 0.32593443522992444, "learning_rate": 6.777791449636681e-06, "loss": 0.2295, "step": 492 }, { "epoch": 3.7802776448061275, "grad_norm": 0.33560750807739526, "learning_rate": 6.697397841366686e-06, "loss": 0.2148, "step": 493 }, { "epoch": 3.7879368118717087, "grad_norm": 0.3347449057734125, "learning_rate": 6.617387872822842e-06, "loss": 0.2197, "step": 494 }, { "epoch": 3.795595978937291, "grad_norm": 0.3467626472846674, "learning_rate": 6.537763851448593e-06, "loss": 0.2391, "step": 495 }, { "epoch": 3.803255146002872, "grad_norm": 0.32356879416863116, "learning_rate": 6.458528073556925e-06, "loss": 0.2268, "step": 496 }, { "epoch": 3.8109143130684537, "grad_norm": 0.3283370951000147, "learning_rate": 6.379682824264055e-06, "loss": 0.2337, "step": 497 }, { "epoch": 3.8185734801340354, "grad_norm": 0.30770811712569635, "learning_rate": 6.301230377423595e-06, "loss": 0.2224, "step": 498 }, { "epoch": 3.826232647199617, "grad_norm": 0.3253183469854635, "learning_rate": 6.223172995560935e-06, "loss": 0.2193, "step": 499 }, { "epoch": 3.8338918142651988, "grad_norm": 0.3348399208299024, "learning_rate": 6.145512929808013e-06, "loss": 0.241, "step": 500 }, { "epoch": 3.8415509813307804, "grad_norm": 0.33895071979833463, "learning_rate": 6.068252419838399e-06, "loss": 0.2427, "step": 501 }, { "epoch": 3.849210148396362, "grad_norm": 0.3343502843045349, "learning_rate": 5.991393693802674e-06, "loss": 0.2319, "step": 502 }, { "epoch": 3.8568693154619433, "grad_norm": 0.32674599015773287, "learning_rate": 5.9149389682642165e-06, "loss": 0.2438, "step": 503 }, { "epoch": 3.864528482527525, "grad_norm": 0.324625635133392, "learning_rate": 5.838890448135228e-06, "loss": 0.2464, "step": 504 }, { "epoch": 3.8721876495931067, "grad_norm": 0.33427787078954374, "learning_rate": 5.7632503266131925e-06, "loss": 0.216, "step": 505 }, { "epoch": 3.8798468166586884, "grad_norm": 0.2975052588803942, "learning_rate": 5.688020785117581e-06, "loss": 0.2193, "step": 506 }, { "epoch": 3.88750598372427, "grad_norm": 0.35541630429900906, "learning_rate": 5.613203993226981e-06, "loss": 0.2416, "step": 507 }, { "epoch": 3.8951651507898517, "grad_norm": 0.3667627536084296, "learning_rate": 5.538802108616494e-06, "loss": 0.2339, "step": 508 }, { "epoch": 3.9028243178554334, "grad_norm": 0.32563123367734753, "learning_rate": 5.46481727699554e-06, "loss": 0.2259, "step": 509 }, { "epoch": 3.9104834849210146, "grad_norm": 0.3147349932477261, "learning_rate": 5.39125163204594e-06, "loss": 0.2166, "step": 510 }, { "epoch": 3.9181426519865967, "grad_norm": 0.3264794954593056, "learning_rate": 5.318107295360424e-06, "loss": 0.2191, "step": 511 }, { "epoch": 3.925801819052178, "grad_norm": 0.3493610273060971, "learning_rate": 5.245386376381398e-06, "loss": 0.223, "step": 512 }, { "epoch": 3.9334609861177596, "grad_norm": 0.31947742082366326, "learning_rate": 5.17309097234016e-06, "loss": 0.2286, "step": 513 }, { "epoch": 3.9411201531833413, "grad_norm": 0.3273845550020943, "learning_rate": 5.101223168196381e-06, "loss": 0.2425, "step": 514 }, { "epoch": 3.948779320248923, "grad_norm": 0.33071553976584445, "learning_rate": 5.029785036577976e-06, "loss": 0.2142, "step": 515 }, { "epoch": 3.9564384873145046, "grad_norm": 0.29920602227346177, "learning_rate": 4.958778637721364e-06, "loss": 0.2255, "step": 516 }, { "epoch": 3.9640976543800863, "grad_norm": 0.3220643902328496, "learning_rate": 4.8882060194119985e-06, "loss": 0.2316, "step": 517 }, { "epoch": 3.971756821445668, "grad_norm": 0.3348880545917222, "learning_rate": 4.8180692169253714e-06, "loss": 0.2319, "step": 518 }, { "epoch": 3.9794159885112492, "grad_norm": 0.3183864890443684, "learning_rate": 4.74837025296826e-06, "loss": 0.2398, "step": 519 }, { "epoch": 3.987075155576831, "grad_norm": 0.3170966409917256, "learning_rate": 4.679111137620442e-06, "loss": 0.2235, "step": 520 }, { "epoch": 3.9947343226424126, "grad_norm": 0.2929870337825558, "learning_rate": 4.610293868276681e-06, "loss": 0.2118, "step": 521 }, { "epoch": 4.003829583532791, "grad_norm": 0.6366855376673984, "learning_rate": 4.541920429589168e-06, "loss": 0.3375, "step": 522 }, { "epoch": 4.011488750598373, "grad_norm": 0.39332713233031513, "learning_rate": 4.47399279341024e-06, "loss": 0.1091, "step": 523 }, { "epoch": 4.019147917663954, "grad_norm": 0.46392007812335173, "learning_rate": 4.406512918735555e-06, "loss": 0.2075, "step": 524 }, { "epoch": 4.026807084729536, "grad_norm": 0.32163834207187486, "learning_rate": 4.339482751647557e-06, "loss": 0.1486, "step": 525 }, { "epoch": 4.034466251795117, "grad_norm": 0.346990014252883, "learning_rate": 4.272904225259387e-06, "loss": 0.1689, "step": 526 }, { "epoch": 4.0421254188606985, "grad_norm": 0.4470904005969031, "learning_rate": 4.206779259659102e-06, "loss": 0.1619, "step": 527 }, { "epoch": 4.049784585926281, "grad_norm": 0.5423468490222553, "learning_rate": 4.141109761854332e-06, "loss": 0.1583, "step": 528 }, { "epoch": 4.057443752991862, "grad_norm": 0.4455453744472251, "learning_rate": 4.075897625717249e-06, "loss": 0.1561, "step": 529 }, { "epoch": 4.065102920057444, "grad_norm": 0.40070748734343326, "learning_rate": 4.011144731929981e-06, "loss": 0.1711, "step": 530 }, { "epoch": 4.072762087123025, "grad_norm": 0.3857460690221848, "learning_rate": 3.9468529479303445e-06, "loss": 0.1874, "step": 531 }, { "epoch": 4.080421254188607, "grad_norm": 0.3386030886136809, "learning_rate": 3.883024127858017e-06, "loss": 0.1676, "step": 532 }, { "epoch": 4.088080421254189, "grad_norm": 0.3665495913561285, "learning_rate": 3.819660112501053e-06, "loss": 0.1605, "step": 533 }, { "epoch": 4.09573958831977, "grad_norm": 0.35417549475087723, "learning_rate": 3.756762729242773e-06, "loss": 0.1504, "step": 534 }, { "epoch": 4.103398755385352, "grad_norm": 0.3396123769965615, "learning_rate": 3.694333792009115e-06, "loss": 0.1375, "step": 535 }, { "epoch": 4.111057922450933, "grad_norm": 0.32569257575060306, "learning_rate": 3.632375101216259e-06, "loss": 0.1518, "step": 536 }, { "epoch": 4.118717089516515, "grad_norm": 0.3219593442840899, "learning_rate": 3.5708884437187673e-06, "loss": 0.168, "step": 537 }, { "epoch": 4.1263762565820965, "grad_norm": 0.31324123070534954, "learning_rate": 3.509875592757999e-06, "loss": 0.1681, "step": 538 }, { "epoch": 4.134035423647679, "grad_norm": 0.3836751303057977, "learning_rate": 3.4493383079110054e-06, "loss": 0.1898, "step": 539 }, { "epoch": 4.14169459071326, "grad_norm": 0.3500886642993717, "learning_rate": 3.3892783350397675e-06, "loss": 0.1793, "step": 540 }, { "epoch": 4.149353757778842, "grad_norm": 0.3188272991338731, "learning_rate": 3.329697406240855e-06, "loss": 0.1721, "step": 541 }, { "epoch": 4.157012924844423, "grad_norm": 0.3012590925590462, "learning_rate": 3.2705972397954655e-06, "loss": 0.1537, "step": 542 }, { "epoch": 4.164672091910004, "grad_norm": 0.3367792622931613, "learning_rate": 3.211979540119883e-06, "loss": 0.1797, "step": 543 }, { "epoch": 4.1723312589755865, "grad_norm": 0.33133930012065166, "learning_rate": 3.153845997716303e-06, "loss": 0.1925, "step": 544 }, { "epoch": 4.179990426041168, "grad_norm": 0.3094187164114957, "learning_rate": 3.0961982891241083e-06, "loss": 0.1481, "step": 545 }, { "epoch": 4.18764959310675, "grad_norm": 0.317376469580326, "learning_rate": 3.039038076871481e-06, "loss": 0.1727, "step": 546 }, { "epoch": 4.195308760172331, "grad_norm": 0.27650160786044625, "learning_rate": 2.9823670094275e-06, "loss": 0.1446, "step": 547 }, { "epoch": 4.202967927237913, "grad_norm": 0.30890083357655423, "learning_rate": 2.9261867211545603e-06, "loss": 0.1622, "step": 548 }, { "epoch": 4.2106270943034945, "grad_norm": 0.33112327548553416, "learning_rate": 2.870498832261257e-06, "loss": 0.1663, "step": 549 }, { "epoch": 4.218286261369076, "grad_norm": 0.3120284953069467, "learning_rate": 2.815304948755664e-06, "loss": 0.1617, "step": 550 }, { "epoch": 4.225945428434658, "grad_norm": 0.31255801847483555, "learning_rate": 2.7606066623990145e-06, "loss": 0.1673, "step": 551 }, { "epoch": 4.233604595500239, "grad_norm": 0.3087850852760736, "learning_rate": 2.7064055506597875e-06, "loss": 0.1608, "step": 552 }, { "epoch": 4.241263762565821, "grad_norm": 0.29550399939819244, "learning_rate": 2.6527031766682142e-06, "loss": 0.149, "step": 553 }, { "epoch": 4.248922929631402, "grad_norm": 0.28827764887901935, "learning_rate": 2.599501089171217e-06, "loss": 0.1545, "step": 554 }, { "epoch": 4.2565820966969845, "grad_norm": 0.31764263264427883, "learning_rate": 2.546800822487714e-06, "loss": 0.1581, "step": 555 }, { "epoch": 4.264241263762566, "grad_norm": 0.2896451556809014, "learning_rate": 2.494603896464405e-06, "loss": 0.1432, "step": 556 }, { "epoch": 4.271900430828148, "grad_norm": 0.2846656674626245, "learning_rate": 2.4429118164319076e-06, "loss": 0.1506, "step": 557 }, { "epoch": 4.279559597893729, "grad_norm": 0.3167572946509413, "learning_rate": 2.3917260731613733e-06, "loss": 0.1681, "step": 558 }, { "epoch": 4.28721876495931, "grad_norm": 0.2994943543117942, "learning_rate": 2.3410481428214602e-06, "loss": 0.1582, "step": 559 }, { "epoch": 4.294877932024892, "grad_norm": 0.2958413618672273, "learning_rate": 2.2908794869358044e-06, "loss": 0.1629, "step": 560 }, { "epoch": 4.302537099090474, "grad_norm": 0.2854040506798909, "learning_rate": 2.2412215523408266e-06, "loss": 0.1504, "step": 561 }, { "epoch": 4.310196266156056, "grad_norm": 0.27706245243194494, "learning_rate": 2.1920757711440354e-06, "loss": 0.1335, "step": 562 }, { "epoch": 4.317855433221637, "grad_norm": 0.3002281927963276, "learning_rate": 2.143443560682721e-06, "loss": 0.1656, "step": 563 }, { "epoch": 4.325514600287219, "grad_norm": 0.3053710633138099, "learning_rate": 2.0953263234830667e-06, "loss": 0.1793, "step": 564 }, { "epoch": 4.3331737673528, "grad_norm": 0.29373337337239025, "learning_rate": 2.0477254472197237e-06, "loss": 0.1607, "step": 565 }, { "epoch": 4.340832934418382, "grad_norm": 0.29361686795928127, "learning_rate": 2.0006423046757596e-06, "loss": 0.1417, "step": 566 }, { "epoch": 4.348492101483964, "grad_norm": 0.30321625536955094, "learning_rate": 1.9540782537031045e-06, "loss": 0.1561, "step": 567 }, { "epoch": 4.356151268549545, "grad_norm": 0.294608355859984, "learning_rate": 1.908034637183356e-06, "loss": 0.1436, "step": 568 }, { "epoch": 4.363810435615127, "grad_norm": 0.2981183902576262, "learning_rate": 1.8625127829890922e-06, "loss": 0.1829, "step": 569 }, { "epoch": 4.371469602680708, "grad_norm": 0.2662376876124655, "learning_rate": 1.817514003945524e-06, "loss": 0.1321, "step": 570 }, { "epoch": 4.37912876974629, "grad_norm": 0.2857595944219623, "learning_rate": 1.7730395977926917e-06, "loss": 0.1568, "step": 571 }, { "epoch": 4.386787936811872, "grad_norm": 0.29444711124099954, "learning_rate": 1.7290908471479805e-06, "loss": 0.1619, "step": 572 }, { "epoch": 4.394447103877454, "grad_norm": 0.299861629149527, "learning_rate": 1.6856690194691872e-06, "loss": 0.1765, "step": 573 }, { "epoch": 4.402106270943035, "grad_norm": 0.2885973088406149, "learning_rate": 1.6427753670179214e-06, "loss": 0.1599, "step": 574 }, { "epoch": 4.409765438008616, "grad_norm": 0.284486199205028, "learning_rate": 1.6004111268235156e-06, "loss": 0.1595, "step": 575 }, { "epoch": 4.417424605074198, "grad_norm": 0.2941729164644181, "learning_rate": 1.5585775206473508e-06, "loss": 0.161, "step": 576 }, { "epoch": 4.4250837721397795, "grad_norm": 0.29517993856011077, "learning_rate": 1.5172757549476024e-06, "loss": 0.1608, "step": 577 }, { "epoch": 4.432742939205362, "grad_norm": 0.3009878509805784, "learning_rate": 1.4765070208444732e-06, "loss": 0.1657, "step": 578 }, { "epoch": 4.440402106270943, "grad_norm": 0.2832295303483787, "learning_rate": 1.4362724940858109e-06, "loss": 0.1481, "step": 579 }, { "epoch": 4.448061273336525, "grad_norm": 0.30017167488484514, "learning_rate": 1.396573335013236e-06, "loss": 0.1758, "step": 580 }, { "epoch": 4.455720440402106, "grad_norm": 0.26948255003997973, "learning_rate": 1.3574106885286465e-06, "loss": 0.1506, "step": 581 }, { "epoch": 4.463379607467688, "grad_norm": 0.2989296760767762, "learning_rate": 1.3187856840612167e-06, "loss": 0.1745, "step": 582 }, { "epoch": 4.47103877453327, "grad_norm": 0.3057384139973513, "learning_rate": 1.2806994355348224e-06, "loss": 0.1721, "step": 583 }, { "epoch": 4.478697941598851, "grad_norm": 0.2636716603605, "learning_rate": 1.2431530413359138e-06, "loss": 0.1367, "step": 584 }, { "epoch": 4.486357108664433, "grad_norm": 0.29595905590077376, "learning_rate": 1.2061475842818337e-06, "loss": 0.1651, "step": 585 }, { "epoch": 4.494016275730014, "grad_norm": 0.28723857409594045, "learning_rate": 1.169684131589608e-06, "loss": 0.1686, "step": 586 }, { "epoch": 4.501675442795596, "grad_norm": 0.28232920958455504, "learning_rate": 1.1337637348451369e-06, "loss": 0.1478, "step": 587 }, { "epoch": 4.5093346098611775, "grad_norm": 0.2945977053685175, "learning_rate": 1.0983874299729092e-06, "loss": 0.1698, "step": 588 }, { "epoch": 4.516993776926759, "grad_norm": 0.29635940814476264, "learning_rate": 1.0635562372060825e-06, "loss": 0.1684, "step": 589 }, { "epoch": 4.524652943992341, "grad_norm": 0.28161502295558277, "learning_rate": 1.0292711610570904e-06, "loss": 0.1435, "step": 590 }, { "epoch": 4.532312111057922, "grad_norm": 0.29973282189790135, "learning_rate": 9.955331902886645e-07, "loss": 0.1601, "step": 591 }, { "epoch": 4.539971278123504, "grad_norm": 0.2912268259487591, "learning_rate": 9.62343297885313e-07, "loss": 0.1676, "step": 592 }, { "epoch": 4.547630445189085, "grad_norm": 0.2804062252790609, "learning_rate": 9.297024410252753e-07, "loss": 0.1572, "step": 593 }, { "epoch": 4.5552896122546676, "grad_norm": 0.2902402528058867, "learning_rate": 8.976115610528957e-07, "loss": 0.1665, "step": 594 }, { "epoch": 4.562948779320249, "grad_norm": 0.2948360443924776, "learning_rate": 8.660715834514977e-07, "loss": 0.1798, "step": 595 }, { "epoch": 4.570607946385831, "grad_norm": 0.30613400428544785, "learning_rate": 8.350834178166755e-07, "loss": 0.1708, "step": 596 }, { "epoch": 4.578267113451412, "grad_norm": 0.29386467356756857, "learning_rate": 8.046479578300803e-07, "loss": 0.1674, "step": 597 }, { "epoch": 4.585926280516993, "grad_norm": 0.26855112533508585, "learning_rate": 7.747660812336221e-07, "loss": 0.127, "step": 598 }, { "epoch": 4.5935854475825755, "grad_norm": 0.30897844150919035, "learning_rate": 7.454386498041865e-07, "loss": 0.1832, "step": 599 }, { "epoch": 4.601244614648157, "grad_norm": 0.26907534103375785, "learning_rate": 7.166665093287539e-07, "loss": 0.1514, "step": 600 }, { "epoch": 4.608903781713739, "grad_norm": 0.2779300127385892, "learning_rate": 6.884504895800237e-07, "loss": 0.1596, "step": 601 }, { "epoch": 4.61656294877932, "grad_norm": 0.283076309215598, "learning_rate": 6.607914042924756e-07, "loss": 0.1493, "step": 602 }, { "epoch": 4.624222115844902, "grad_norm": 0.30512219679760166, "learning_rate": 6.336900511389133e-07, "loss": 0.1839, "step": 603 }, { "epoch": 4.631881282910483, "grad_norm": 0.2784336379714144, "learning_rate": 6.071472117074462e-07, "loss": 0.1553, "step": 604 }, { "epoch": 4.6395404499760655, "grad_norm": 0.2908490571653058, "learning_rate": 5.811636514789598e-07, "loss": 0.1735, "step": 605 }, { "epoch": 4.647199617041647, "grad_norm": 0.2689077917871218, "learning_rate": 5.557401198050327e-07, "loss": 0.1406, "step": 606 }, { "epoch": 4.654858784107228, "grad_norm": 0.30361924900865483, "learning_rate": 5.308773498863251e-07, "loss": 0.1806, "step": 607 }, { "epoch": 4.66251795117281, "grad_norm": 0.2738298236426758, "learning_rate": 5.065760587514446e-07, "loss": 0.1555, "step": 608 }, { "epoch": 4.670177118238391, "grad_norm": 0.28066674173939915, "learning_rate": 4.828369472362493e-07, "loss": 0.1456, "step": 609 }, { "epoch": 4.677836285303973, "grad_norm": 0.28254726890674403, "learning_rate": 4.5966069996365993e-07, "loss": 0.1666, "step": 610 }, { "epoch": 4.685495452369555, "grad_norm": 0.28356791946970256, "learning_rate": 4.3704798532388624e-07, "loss": 0.1623, "step": 611 }, { "epoch": 4.693154619435137, "grad_norm": 0.27978608677096656, "learning_rate": 4.1499945545518283e-07, "loss": 0.1573, "step": 612 }, { "epoch": 4.700813786500718, "grad_norm": 0.2807595614948837, "learning_rate": 3.935157462250128e-07, "loss": 0.1645, "step": 613 }, { "epoch": 4.7084729535663, "grad_norm": 0.2903423678415902, "learning_rate": 3.7259747721173134e-07, "loss": 0.1697, "step": 614 }, { "epoch": 4.716132120631881, "grad_norm": 0.27761326426161137, "learning_rate": 3.522452516867048e-07, "loss": 0.154, "step": 615 }, { "epoch": 4.723791287697463, "grad_norm": 0.2867472262572132, "learning_rate": 3.324596565969174e-07, "loss": 0.1565, "step": 616 }, { "epoch": 4.731450454763045, "grad_norm": 0.28465179062440776, "learning_rate": 3.1324126254804524e-07, "loss": 0.1401, "step": 617 }, { "epoch": 4.739109621828626, "grad_norm": 0.2947996164689416, "learning_rate": 2.9459062378799806e-07, "loss": 0.1754, "step": 618 }, { "epoch": 4.746768788894208, "grad_norm": 0.2863382862190152, "learning_rate": 2.7650827819093005e-07, "loss": 0.1599, "step": 619 }, { "epoch": 4.754427955959789, "grad_norm": 0.2709108924581573, "learning_rate": 2.5899474724174313e-07, "loss": 0.1398, "step": 620 }, { "epoch": 4.7620871230253705, "grad_norm": 0.30022047839665467, "learning_rate": 2.4205053602103015e-07, "loss": 0.1792, "step": 621 }, { "epoch": 4.769746290090953, "grad_norm": 0.2981470620905006, "learning_rate": 2.2567613319051997e-07, "loss": 0.1449, "step": 622 }, { "epoch": 4.777405457156534, "grad_norm": 0.2875561082245844, "learning_rate": 2.0987201097897757e-07, "loss": 0.1395, "step": 623 }, { "epoch": 4.785064624222116, "grad_norm": 0.27525535566929177, "learning_rate": 1.9463862516859277e-07, "loss": 0.1436, "step": 624 }, { "epoch": 4.792723791287697, "grad_norm": 0.2940141095677646, "learning_rate": 1.799764150818306e-07, "loss": 0.1655, "step": 625 }, { "epoch": 4.800382958353279, "grad_norm": 0.3081953681996578, "learning_rate": 1.658858035687594e-07, "loss": 0.1667, "step": 626 }, { "epoch": 4.808042125418861, "grad_norm": 0.26637485304480285, "learning_rate": 1.5236719699486256e-07, "loss": 0.1309, "step": 627 }, { "epoch": 4.815701292484443, "grad_norm": 0.2705718541638546, "learning_rate": 1.3942098522931491e-07, "loss": 0.1613, "step": 628 }, { "epoch": 4.823360459550024, "grad_norm": 0.3013254867336271, "learning_rate": 1.2704754163374022e-07, "loss": 0.1871, "step": 629 }, { "epoch": 4.831019626615605, "grad_norm": 0.2781156531330604, "learning_rate": 1.1524722305144231e-07, "loss": 0.1583, "step": 630 }, { "epoch": 4.838678793681187, "grad_norm": 0.27865348896247544, "learning_rate": 1.0402036979711317e-07, "loss": 0.1541, "step": 631 }, { "epoch": 4.8463379607467685, "grad_norm": 0.3006123678070983, "learning_rate": 9.336730564702745e-08, "loss": 0.177, "step": 632 }, { "epoch": 4.853997127812351, "grad_norm": 0.2933355244436529, "learning_rate": 8.328833782969003e-08, "loss": 0.1735, "step": 633 }, { "epoch": 4.861656294877932, "grad_norm": 0.2612274384851061, "learning_rate": 7.378375701698748e-08, "loss": 0.142, "step": 634 }, { "epoch": 4.869315461943514, "grad_norm": 0.27506364744141204, "learning_rate": 6.485383731580142e-08, "loss": 0.1599, "step": 635 }, { "epoch": 4.876974629009095, "grad_norm": 0.2740746508966746, "learning_rate": 5.649883626009933e-08, "loss": 0.1515, "step": 636 }, { "epoch": 4.884633796074677, "grad_norm": 0.2923776048680673, "learning_rate": 4.871899480351605e-08, "loss": 0.1614, "step": 637 }, { "epoch": 4.8922929631402585, "grad_norm": 0.2563881864024306, "learning_rate": 4.151453731239707e-08, "loss": 0.1431, "step": 638 }, { "epoch": 4.89995213020584, "grad_norm": 0.31290090906865226, "learning_rate": 3.4885671559332645e-08, "loss": 0.1929, "step": 639 }, { "epoch": 4.907611297271422, "grad_norm": 0.2833641820727045, "learning_rate": 2.8832588717164766e-08, "loss": 0.1354, "step": 640 }, { "epoch": 4.915270464337003, "grad_norm": 0.2886325657329948, "learning_rate": 2.3355463353467168e-08, "loss": 0.1543, "step": 641 }, { "epoch": 4.922929631402585, "grad_norm": 0.27914356344167157, "learning_rate": 1.8454453425527098e-08, "loss": 0.1511, "step": 642 }, { "epoch": 4.9305887984681664, "grad_norm": 0.3094563610476224, "learning_rate": 1.4129700275771208e-08, "loss": 0.1843, "step": 643 }, { "epoch": 4.938247965533749, "grad_norm": 0.25323095165276716, "learning_rate": 1.0381328627702136e-08, "loss": 0.1128, "step": 644 }, { "epoch": 4.94590713259933, "grad_norm": 0.2790763770971488, "learning_rate": 7.209446582292501e-09, "loss": 0.1528, "step": 645 }, { "epoch": 4.953566299664912, "grad_norm": 0.294849164666762, "learning_rate": 4.614145614876275e-09, "loss": 0.1657, "step": 646 }, { "epoch": 4.961225466730493, "grad_norm": 0.30506766088706516, "learning_rate": 2.5955005725064597e-09, "loss": 0.178, "step": 647 }, { "epoch": 4.968884633796074, "grad_norm": 0.281891446902876, "learning_rate": 1.1535696717945855e-09, "loss": 0.1473, "step": 648 }, { "epoch": 4.9765438008616565, "grad_norm": 0.2745560443089173, "learning_rate": 2.8839449723205847e-10, "loss": 0.1509, "step": 649 }, { "epoch": 4.984202967927238, "grad_norm": 0.2816379591846223, "learning_rate": 0.0, "loss": 0.1471, "step": 650 }, { "epoch": 4.984202967927238, "step": 650, "total_flos": 1.3070689797343805e+18, "train_loss": 0.36027532513325033, "train_runtime": 68390.8162, "train_samples_per_second": 1.222, "train_steps_per_second": 0.01 } ], "logging_steps": 1.0, "max_steps": 650, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.3070689797343805e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }