{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 782, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0025575447570332483, "grad_norm": 21.953563303736964, "learning_rate": 9.999959651660741e-06, "loss": 1.3458, "step": 1 }, { "epoch": 0.005115089514066497, "grad_norm": 11.652464724143648, "learning_rate": 9.999838607294157e-06, "loss": 1.293, "step": 2 }, { "epoch": 0.0076726342710997444, "grad_norm": 14.373069724258055, "learning_rate": 9.999636868853824e-06, "loss": 0.9527, "step": 3 }, { "epoch": 0.010230179028132993, "grad_norm": 9.875205922264723, "learning_rate": 9.999354439595668e-06, "loss": 0.9394, "step": 4 }, { "epoch": 0.01278772378516624, "grad_norm": 15.116196776322745, "learning_rate": 9.998991324077906e-06, "loss": 0.8439, "step": 5 }, { "epoch": 0.015345268542199489, "grad_norm": 11.057152894995923, "learning_rate": 9.998547528160987e-06, "loss": 0.8405, "step": 6 }, { "epoch": 0.017902813299232736, "grad_norm": 9.365688071554938, "learning_rate": 9.998023059007477e-06, "loss": 0.8734, "step": 7 }, { "epoch": 0.020460358056265986, "grad_norm": 8.392619519229692, "learning_rate": 9.997417925081963e-06, "loss": 0.7449, "step": 8 }, { "epoch": 0.023017902813299233, "grad_norm": 9.340778471253389, "learning_rate": 9.996732136150902e-06, "loss": 0.7876, "step": 9 }, { "epoch": 0.02557544757033248, "grad_norm": 8.20102806123342, "learning_rate": 9.995965703282472e-06, "loss": 0.7954, "step": 10 }, { "epoch": 0.028132992327365727, "grad_norm": 8.76971572309029, "learning_rate": 9.995118638846394e-06, "loss": 0.8333, "step": 11 }, { "epoch": 0.030690537084398978, "grad_norm": 7.557620850007077, "learning_rate": 9.99419095651372e-06, "loss": 0.8554, "step": 12 }, { "epoch": 0.03324808184143223, "grad_norm": 7.743714182477662, "learning_rate": 9.993182671256633e-06, "loss": 0.8637, "step": 13 }, { "epoch": 0.03580562659846547, "grad_norm": 7.32084774886768, "learning_rate": 9.992093799348182e-06, "loss": 0.8621, "step": 14 }, { "epoch": 0.03836317135549872, "grad_norm": 7.6800585762630655, "learning_rate": 9.990924358362037e-06, "loss": 0.7075, "step": 15 }, { "epoch": 0.04092071611253197, "grad_norm": 7.053712401961599, "learning_rate": 9.9896743671722e-06, "loss": 0.8078, "step": 16 }, { "epoch": 0.043478260869565216, "grad_norm": 7.957894711517596, "learning_rate": 9.988343845952697e-06, "loss": 0.8534, "step": 17 }, { "epoch": 0.04603580562659847, "grad_norm": 7.77708191899406, "learning_rate": 9.986932816177258e-06, "loss": 0.7898, "step": 18 }, { "epoch": 0.04859335038363171, "grad_norm": 8.86563729131372, "learning_rate": 9.985441300618966e-06, "loss": 0.7867, "step": 19 }, { "epoch": 0.05115089514066496, "grad_norm": 7.9582904003704655, "learning_rate": 9.98386932334989e-06, "loss": 0.7112, "step": 20 }, { "epoch": 0.05370843989769821, "grad_norm": 7.46439913319433, "learning_rate": 9.982216909740703e-06, "loss": 0.7207, "step": 21 }, { "epoch": 0.056265984654731455, "grad_norm": 7.502417267747896, "learning_rate": 9.980484086460258e-06, "loss": 0.7635, "step": 22 }, { "epoch": 0.058823529411764705, "grad_norm": 8.077410576519545, "learning_rate": 9.978670881475173e-06, "loss": 0.7568, "step": 23 }, { "epoch": 0.061381074168797956, "grad_norm": 8.530645418517887, "learning_rate": 9.976777324049374e-06, "loss": 0.8006, "step": 24 }, { "epoch": 0.0639386189258312, "grad_norm": 7.884014258298183, "learning_rate": 9.974803444743617e-06, "loss": 0.741, "step": 25 }, { "epoch": 0.06649616368286446, "grad_norm": 6.917821196530823, "learning_rate": 9.972749275415005e-06, "loss": 0.6668, "step": 26 }, { "epoch": 0.06905370843989769, "grad_norm": 6.951997744847425, "learning_rate": 9.970614849216465e-06, "loss": 0.7619, "step": 27 }, { "epoch": 0.07161125319693094, "grad_norm": 6.541689929654185, "learning_rate": 9.96840020059622e-06, "loss": 0.7667, "step": 28 }, { "epoch": 0.0741687979539642, "grad_norm": 7.022380585505507, "learning_rate": 9.966105365297226e-06, "loss": 0.7176, "step": 29 }, { "epoch": 0.07672634271099744, "grad_norm": 6.912790106638039, "learning_rate": 9.963730380356599e-06, "loss": 0.7199, "step": 30 }, { "epoch": 0.0792838874680307, "grad_norm": 6.0140784283728275, "learning_rate": 9.96127528410502e-06, "loss": 0.7356, "step": 31 }, { "epoch": 0.08184143222506395, "grad_norm": 6.50099127837641, "learning_rate": 9.958740116166113e-06, "loss": 0.6741, "step": 32 }, { "epoch": 0.08439897698209718, "grad_norm": 6.196159967289502, "learning_rate": 9.9561249174558e-06, "loss": 0.6453, "step": 33 }, { "epoch": 0.08695652173913043, "grad_norm": 7.115083432855918, "learning_rate": 9.953429730181653e-06, "loss": 0.6921, "step": 34 }, { "epoch": 0.08951406649616368, "grad_norm": 8.373221612447116, "learning_rate": 9.950654597842209e-06, "loss": 0.6904, "step": 35 }, { "epoch": 0.09207161125319693, "grad_norm": 7.708793853244428, "learning_rate": 9.947799565226253e-06, "loss": 0.7295, "step": 36 }, { "epoch": 0.09462915601023018, "grad_norm": 8.81215615740888, "learning_rate": 9.944864678412118e-06, "loss": 0.6806, "step": 37 }, { "epoch": 0.09718670076726342, "grad_norm": 7.456995220385941, "learning_rate": 9.94184998476693e-06, "loss": 0.6659, "step": 38 }, { "epoch": 0.09974424552429667, "grad_norm": 7.636846126721769, "learning_rate": 9.938755532945838e-06, "loss": 0.7073, "step": 39 }, { "epoch": 0.10230179028132992, "grad_norm": 8.018319876671674, "learning_rate": 9.93558137289124e-06, "loss": 0.8063, "step": 40 }, { "epoch": 0.10485933503836317, "grad_norm": 6.429546682427428, "learning_rate": 9.932327555831972e-06, "loss": 0.5498, "step": 41 }, { "epoch": 0.10741687979539642, "grad_norm": 6.106558471191083, "learning_rate": 9.928994134282477e-06, "loss": 0.6125, "step": 42 }, { "epoch": 0.10997442455242967, "grad_norm": 7.220242425194578, "learning_rate": 9.925581162041967e-06, "loss": 0.6553, "step": 43 }, { "epoch": 0.11253196930946291, "grad_norm": 7.406518370609504, "learning_rate": 9.922088694193546e-06, "loss": 0.7235, "step": 44 }, { "epoch": 0.11508951406649616, "grad_norm": 6.232863127816167, "learning_rate": 9.918516787103322e-06, "loss": 0.6302, "step": 45 }, { "epoch": 0.11764705882352941, "grad_norm": 6.960585716404723, "learning_rate": 9.91486549841951e-06, "loss": 0.6631, "step": 46 }, { "epoch": 0.12020460358056266, "grad_norm": 7.014565196822424, "learning_rate": 9.911134887071477e-06, "loss": 0.6013, "step": 47 }, { "epoch": 0.12276214833759591, "grad_norm": 6.747724971317866, "learning_rate": 9.907325013268816e-06, "loss": 0.7311, "step": 48 }, { "epoch": 0.12531969309462915, "grad_norm": 6.343001355283702, "learning_rate": 9.903435938500356e-06, "loss": 0.5662, "step": 49 }, { "epoch": 0.1278772378516624, "grad_norm": 6.68715844293194, "learning_rate": 9.899467725533181e-06, "loss": 0.6746, "step": 50 }, { "epoch": 0.13043478260869565, "grad_norm": 6.670242029824649, "learning_rate": 9.895420438411616e-06, "loss": 0.5636, "step": 51 }, { "epoch": 0.1329923273657289, "grad_norm": 7.6565238394007, "learning_rate": 9.89129414245618e-06, "loss": 0.6959, "step": 52 }, { "epoch": 0.13554987212276215, "grad_norm": 6.322260761537018, "learning_rate": 9.887088904262557e-06, "loss": 0.6475, "step": 53 }, { "epoch": 0.13810741687979539, "grad_norm": 7.424958009645482, "learning_rate": 9.882804791700488e-06, "loss": 0.7381, "step": 54 }, { "epoch": 0.14066496163682865, "grad_norm": 6.484080185567617, "learning_rate": 9.878441873912712e-06, "loss": 0.5978, "step": 55 }, { "epoch": 0.1432225063938619, "grad_norm": 7.904610864584768, "learning_rate": 9.87400022131382e-06, "loss": 0.7156, "step": 56 }, { "epoch": 0.14578005115089515, "grad_norm": 6.091768577876145, "learning_rate": 9.869479905589136e-06, "loss": 0.5674, "step": 57 }, { "epoch": 0.1483375959079284, "grad_norm": 6.338206394229405, "learning_rate": 9.864880999693551e-06, "loss": 0.5511, "step": 58 }, { "epoch": 0.15089514066496162, "grad_norm": 7.393436860670507, "learning_rate": 9.860203577850353e-06, "loss": 0.6305, "step": 59 }, { "epoch": 0.1534526854219949, "grad_norm": 6.520098900315275, "learning_rate": 9.855447715550024e-06, "loss": 0.6191, "step": 60 }, { "epoch": 0.15601023017902813, "grad_norm": 6.415362809418058, "learning_rate": 9.850613489549018e-06, "loss": 0.6296, "step": 61 }, { "epoch": 0.1585677749360614, "grad_norm": 7.9580764103093635, "learning_rate": 9.845700977868536e-06, "loss": 0.6384, "step": 62 }, { "epoch": 0.16112531969309463, "grad_norm": 7.1082899489573315, "learning_rate": 9.840710259793251e-06, "loss": 0.6021, "step": 63 }, { "epoch": 0.1636828644501279, "grad_norm": 6.386504712691233, "learning_rate": 9.835641415870038e-06, "loss": 0.6873, "step": 64 }, { "epoch": 0.16624040920716113, "grad_norm": 7.4797007816006555, "learning_rate": 9.830494527906671e-06, "loss": 0.5648, "step": 65 }, { "epoch": 0.16879795396419436, "grad_norm": 6.748679766187366, "learning_rate": 9.825269678970502e-06, "loss": 0.566, "step": 66 }, { "epoch": 0.17135549872122763, "grad_norm": 6.56170099270124, "learning_rate": 9.819966953387122e-06, "loss": 0.6398, "step": 67 }, { "epoch": 0.17391304347826086, "grad_norm": 5.941005918686002, "learning_rate": 9.814586436738998e-06, "loss": 0.6658, "step": 68 }, { "epoch": 0.17647058823529413, "grad_norm": 5.942719440981736, "learning_rate": 9.809128215864096e-06, "loss": 0.5971, "step": 69 }, { "epoch": 0.17902813299232737, "grad_norm": 6.4968806279365925, "learning_rate": 9.803592378854476e-06, "loss": 0.6047, "step": 70 }, { "epoch": 0.1815856777493606, "grad_norm": 5.577555703482892, "learning_rate": 9.797979015054868e-06, "loss": 0.5534, "step": 71 }, { "epoch": 0.18414322250639387, "grad_norm": 6.262423254220968, "learning_rate": 9.792288215061237e-06, "loss": 0.5755, "step": 72 }, { "epoch": 0.1867007672634271, "grad_norm": 7.17658874763877, "learning_rate": 9.786520070719313e-06, "loss": 0.5511, "step": 73 }, { "epoch": 0.18925831202046037, "grad_norm": 5.532267867303149, "learning_rate": 9.780674675123113e-06, "loss": 0.4965, "step": 74 }, { "epoch": 0.1918158567774936, "grad_norm": 7.040268722404183, "learning_rate": 9.77475212261344e-06, "loss": 0.6644, "step": 75 }, { "epoch": 0.19437340153452684, "grad_norm": 6.774497918030386, "learning_rate": 9.768752508776358e-06, "loss": 0.491, "step": 76 }, { "epoch": 0.1969309462915601, "grad_norm": 5.776934238941505, "learning_rate": 9.762675930441647e-06, "loss": 0.4861, "step": 77 }, { "epoch": 0.19948849104859334, "grad_norm": 6.47046232974907, "learning_rate": 9.756522485681247e-06, "loss": 0.6273, "step": 78 }, { "epoch": 0.2020460358056266, "grad_norm": 6.712438744552012, "learning_rate": 9.750292273807666e-06, "loss": 0.7368, "step": 79 }, { "epoch": 0.20460358056265984, "grad_norm": 8.615585919768426, "learning_rate": 9.743985395372387e-06, "loss": 0.6335, "step": 80 }, { "epoch": 0.2071611253196931, "grad_norm": 8.071799145378323, "learning_rate": 9.737601952164238e-06, "loss": 0.6612, "step": 81 }, { "epoch": 0.20971867007672634, "grad_norm": 6.823667160150434, "learning_rate": 9.73114204720775e-06, "loss": 0.5456, "step": 82 }, { "epoch": 0.21227621483375958, "grad_norm": 5.888063194216176, "learning_rate": 9.724605784761501e-06, "loss": 0.5959, "step": 83 }, { "epoch": 0.21483375959079284, "grad_norm": 7.210449720464657, "learning_rate": 9.717993270316421e-06, "loss": 0.5919, "step": 84 }, { "epoch": 0.21739130434782608, "grad_norm": 7.06901541191221, "learning_rate": 9.711304610594104e-06, "loss": 0.5882, "step": 85 }, { "epoch": 0.21994884910485935, "grad_norm": 6.374122555190542, "learning_rate": 9.704539913545073e-06, "loss": 0.6133, "step": 86 }, { "epoch": 0.22250639386189258, "grad_norm": 6.630925756972144, "learning_rate": 9.697699288347043e-06, "loss": 0.5664, "step": 87 }, { "epoch": 0.22506393861892582, "grad_norm": 5.966658672241774, "learning_rate": 9.690782845403164e-06, "loss": 0.6244, "step": 88 }, { "epoch": 0.22762148337595908, "grad_norm": 5.5770246123900264, "learning_rate": 9.683790696340229e-06, "loss": 0.5334, "step": 89 }, { "epoch": 0.23017902813299232, "grad_norm": 6.522499319244262, "learning_rate": 9.676722954006878e-06, "loss": 0.6739, "step": 90 }, { "epoch": 0.23273657289002558, "grad_norm": 6.4235805218363025, "learning_rate": 9.669579732471779e-06, "loss": 0.6595, "step": 91 }, { "epoch": 0.23529411764705882, "grad_norm": 6.3431696131235284, "learning_rate": 9.66236114702178e-06, "loss": 0.6023, "step": 92 }, { "epoch": 0.23785166240409208, "grad_norm": 6.429883886786425, "learning_rate": 9.655067314160058e-06, "loss": 0.5986, "step": 93 }, { "epoch": 0.24040920716112532, "grad_norm": 6.701368521252926, "learning_rate": 9.647698351604227e-06, "loss": 0.6569, "step": 94 }, { "epoch": 0.24296675191815856, "grad_norm": 6.626338500498468, "learning_rate": 9.640254378284447e-06, "loss": 0.5552, "step": 95 }, { "epoch": 0.24552429667519182, "grad_norm": 6.872036104522696, "learning_rate": 9.632735514341508e-06, "loss": 0.5384, "step": 96 }, { "epoch": 0.24808184143222506, "grad_norm": 7.976561364452108, "learning_rate": 9.625141881124874e-06, "loss": 0.6225, "step": 97 }, { "epoch": 0.2506393861892583, "grad_norm": 6.979721775572128, "learning_rate": 9.617473601190743e-06, "loss": 0.5937, "step": 98 }, { "epoch": 0.2531969309462916, "grad_norm": 7.519173375318072, "learning_rate": 9.609730798300056e-06, "loss": 0.5673, "step": 99 }, { "epoch": 0.2557544757033248, "grad_norm": 6.501376080036557, "learning_rate": 9.601913597416513e-06, "loss": 0.6167, "step": 100 }, { "epoch": 0.25831202046035806, "grad_norm": 7.232025532728033, "learning_rate": 9.594022124704541e-06, "loss": 0.6528, "step": 101 }, { "epoch": 0.2608695652173913, "grad_norm": 5.821804513422366, "learning_rate": 9.586056507527266e-06, "loss": 0.5839, "step": 102 }, { "epoch": 0.26342710997442453, "grad_norm": 7.447246210872055, "learning_rate": 9.578016874444459e-06, "loss": 0.5425, "step": 103 }, { "epoch": 0.2659846547314578, "grad_norm": 5.023721202880345, "learning_rate": 9.569903355210457e-06, "loss": 0.4649, "step": 104 }, { "epoch": 0.26854219948849106, "grad_norm": 5.3855220410063165, "learning_rate": 9.561716080772072e-06, "loss": 0.5362, "step": 105 }, { "epoch": 0.2710997442455243, "grad_norm": 5.755012965635693, "learning_rate": 9.55345518326647e-06, "loss": 0.637, "step": 106 }, { "epoch": 0.27365728900255754, "grad_norm": 6.317278618265475, "learning_rate": 9.545120796019056e-06, "loss": 0.6073, "step": 107 }, { "epoch": 0.27621483375959077, "grad_norm": 6.556750652969424, "learning_rate": 9.5367130535413e-06, "loss": 0.6106, "step": 108 }, { "epoch": 0.27877237851662406, "grad_norm": 7.234204791297877, "learning_rate": 9.528232091528578e-06, "loss": 0.5537, "step": 109 }, { "epoch": 0.2813299232736573, "grad_norm": 6.44863975829238, "learning_rate": 9.519678046857987e-06, "loss": 0.6654, "step": 110 }, { "epoch": 0.28388746803069054, "grad_norm": 5.882964477582141, "learning_rate": 9.511051057586125e-06, "loss": 0.5723, "step": 111 }, { "epoch": 0.2864450127877238, "grad_norm": 5.682881601819309, "learning_rate": 9.502351262946865e-06, "loss": 0.5325, "step": 112 }, { "epoch": 0.289002557544757, "grad_norm": 5.907234833224297, "learning_rate": 9.493578803349117e-06, "loss": 0.6238, "step": 113 }, { "epoch": 0.2915601023017903, "grad_norm": 6.539157912120955, "learning_rate": 9.48473382037455e-06, "loss": 0.6228, "step": 114 }, { "epoch": 0.29411764705882354, "grad_norm": 6.263387218751874, "learning_rate": 9.475816456775313e-06, "loss": 0.5954, "step": 115 }, { "epoch": 0.2966751918158568, "grad_norm": 5.93110985413684, "learning_rate": 9.466826856471728e-06, "loss": 0.6008, "step": 116 }, { "epoch": 0.29923273657289, "grad_norm": 6.326093043591936, "learning_rate": 9.457765164549979e-06, "loss": 0.5834, "step": 117 }, { "epoch": 0.30179028132992325, "grad_norm": 8.333855109469555, "learning_rate": 9.448631527259749e-06, "loss": 0.7357, "step": 118 }, { "epoch": 0.30434782608695654, "grad_norm": 7.352912958641833, "learning_rate": 9.439426092011877e-06, "loss": 0.5621, "step": 119 }, { "epoch": 0.3069053708439898, "grad_norm": 6.812130929920315, "learning_rate": 9.430149007375974e-06, "loss": 0.6281, "step": 120 }, { "epoch": 0.309462915601023, "grad_norm": 6.244862089177942, "learning_rate": 9.42080042307802e-06, "loss": 0.6083, "step": 121 }, { "epoch": 0.31202046035805625, "grad_norm": 5.91824427054237, "learning_rate": 9.411380489997962e-06, "loss": 0.5141, "step": 122 }, { "epoch": 0.3145780051150895, "grad_norm": 5.9138584985365075, "learning_rate": 9.401889360167256e-06, "loss": 0.5525, "step": 123 }, { "epoch": 0.3171355498721228, "grad_norm": 6.264682424037789, "learning_rate": 9.392327186766434e-06, "loss": 0.5049, "step": 124 }, { "epoch": 0.319693094629156, "grad_norm": 5.838854049045799, "learning_rate": 9.382694124122624e-06, "loss": 0.5835, "step": 125 }, { "epoch": 0.32225063938618925, "grad_norm": 6.39551091760669, "learning_rate": 9.372990327707057e-06, "loss": 0.5132, "step": 126 }, { "epoch": 0.3248081841432225, "grad_norm": 6.716488574379217, "learning_rate": 9.36321595413256e-06, "loss": 0.5372, "step": 127 }, { "epoch": 0.3273657289002558, "grad_norm": 5.742568960781894, "learning_rate": 9.353371161151032e-06, "loss": 0.6203, "step": 128 }, { "epoch": 0.329923273657289, "grad_norm": 6.772868273583815, "learning_rate": 9.34345610765089e-06, "loss": 0.4926, "step": 129 }, { "epoch": 0.33248081841432225, "grad_norm": 7.185477633140159, "learning_rate": 9.333470953654513e-06, "loss": 0.6842, "step": 130 }, { "epoch": 0.3350383631713555, "grad_norm": 6.296716433927121, "learning_rate": 9.32341586031565e-06, "loss": 0.5163, "step": 131 }, { "epoch": 0.3375959079283887, "grad_norm": 8.081152679238958, "learning_rate": 9.31329098991683e-06, "loss": 0.7902, "step": 132 }, { "epoch": 0.340153452685422, "grad_norm": 6.861874747525335, "learning_rate": 9.303096505866734e-06, "loss": 0.5645, "step": 133 }, { "epoch": 0.34271099744245526, "grad_norm": 6.171268132782227, "learning_rate": 9.292832572697566e-06, "loss": 0.6114, "step": 134 }, { "epoch": 0.3452685421994885, "grad_norm": 6.552747924757749, "learning_rate": 9.282499356062385e-06, "loss": 0.5954, "step": 135 }, { "epoch": 0.34782608695652173, "grad_norm": 6.847724657541522, "learning_rate": 9.272097022732444e-06, "loss": 0.6485, "step": 136 }, { "epoch": 0.35038363171355497, "grad_norm": 6.507209809740027, "learning_rate": 9.261625740594494e-06, "loss": 0.6159, "step": 137 }, { "epoch": 0.35294117647058826, "grad_norm": 7.252605300424873, "learning_rate": 9.251085678648072e-06, "loss": 0.5576, "step": 138 }, { "epoch": 0.3554987212276215, "grad_norm": 7.119115054572448, "learning_rate": 9.240477007002777e-06, "loss": 0.7135, "step": 139 }, { "epoch": 0.35805626598465473, "grad_norm": 7.349847252060697, "learning_rate": 9.22979989687552e-06, "loss": 0.6444, "step": 140 }, { "epoch": 0.36061381074168797, "grad_norm": 5.765325529890724, "learning_rate": 9.219054520587766e-06, "loss": 0.4233, "step": 141 }, { "epoch": 0.3631713554987212, "grad_norm": 5.112678313504443, "learning_rate": 9.208241051562753e-06, "loss": 0.5447, "step": 142 }, { "epoch": 0.3657289002557545, "grad_norm": 7.144208807574378, "learning_rate": 9.197359664322684e-06, "loss": 0.5891, "step": 143 }, { "epoch": 0.36828644501278773, "grad_norm": 6.278104080681577, "learning_rate": 9.186410534485924e-06, "loss": 0.5701, "step": 144 }, { "epoch": 0.37084398976982097, "grad_norm": 6.887145963626663, "learning_rate": 9.175393838764153e-06, "loss": 0.5502, "step": 145 }, { "epoch": 0.3734015345268542, "grad_norm": 6.023815922626531, "learning_rate": 9.164309754959523e-06, "loss": 0.5286, "step": 146 }, { "epoch": 0.37595907928388744, "grad_norm": 5.517318940161725, "learning_rate": 9.153158461961782e-06, "loss": 0.4433, "step": 147 }, { "epoch": 0.37851662404092073, "grad_norm": 6.019318968473694, "learning_rate": 9.14194013974539e-06, "loss": 0.5065, "step": 148 }, { "epoch": 0.38107416879795397, "grad_norm": 5.6444749821227385, "learning_rate": 9.130654969366619e-06, "loss": 0.501, "step": 149 }, { "epoch": 0.3836317135549872, "grad_norm": 6.559732586020477, "learning_rate": 9.11930313296062e-06, "loss": 0.6101, "step": 150 }, { "epoch": 0.38618925831202044, "grad_norm": 6.2279396393587705, "learning_rate": 9.107884813738492e-06, "loss": 0.5938, "step": 151 }, { "epoch": 0.3887468030690537, "grad_norm": 7.319750378815782, "learning_rate": 9.096400195984322e-06, "loss": 0.4252, "step": 152 }, { "epoch": 0.391304347826087, "grad_norm": 7.088600256578154, "learning_rate": 9.08484946505221e-06, "loss": 0.5793, "step": 153 }, { "epoch": 0.3938618925831202, "grad_norm": 6.573409174775933, "learning_rate": 9.073232807363283e-06, "loss": 0.5026, "step": 154 }, { "epoch": 0.39641943734015345, "grad_norm": 7.980317521312881, "learning_rate": 9.061550410402677e-06, "loss": 0.6736, "step": 155 }, { "epoch": 0.3989769820971867, "grad_norm": 6.87810163307716, "learning_rate": 9.049802462716521e-06, "loss": 0.493, "step": 156 }, { "epoch": 0.40153452685422, "grad_norm": 7.407535434490462, "learning_rate": 9.037989153908882e-06, "loss": 0.5762, "step": 157 }, { "epoch": 0.4040920716112532, "grad_norm": 6.310545687817972, "learning_rate": 9.026110674638722e-06, "loss": 0.5802, "step": 158 }, { "epoch": 0.40664961636828645, "grad_norm": 6.538278944739297, "learning_rate": 9.0141672166168e-06, "loss": 0.4665, "step": 159 }, { "epoch": 0.4092071611253197, "grad_norm": 6.25525186329276, "learning_rate": 9.002158972602599e-06, "loss": 0.65, "step": 160 }, { "epoch": 0.4117647058823529, "grad_norm": 7.134439997933502, "learning_rate": 8.990086136401199e-06, "loss": 0.6436, "step": 161 }, { "epoch": 0.4143222506393862, "grad_norm": 6.907730334313879, "learning_rate": 8.977948902860154e-06, "loss": 0.6688, "step": 162 }, { "epoch": 0.41687979539641945, "grad_norm": 5.60949076779962, "learning_rate": 8.965747467866355e-06, "loss": 0.4263, "step": 163 }, { "epoch": 0.4194373401534527, "grad_norm": 7.153866044006984, "learning_rate": 8.953482028342853e-06, "loss": 0.707, "step": 164 }, { "epoch": 0.4219948849104859, "grad_norm": 5.749873195369449, "learning_rate": 8.9411527822457e-06, "loss": 0.5522, "step": 165 }, { "epoch": 0.42455242966751916, "grad_norm": 6.450113679037296, "learning_rate": 8.92875992856073e-06, "loss": 0.491, "step": 166 }, { "epoch": 0.42710997442455245, "grad_norm": 7.10664387732353, "learning_rate": 8.916303667300373e-06, "loss": 0.5526, "step": 167 }, { "epoch": 0.4296675191815857, "grad_norm": 5.908129226489756, "learning_rate": 8.903784199500412e-06, "loss": 0.4589, "step": 168 }, { "epoch": 0.4322250639386189, "grad_norm": 6.036965674381112, "learning_rate": 8.89120172721674e-06, "loss": 0.5393, "step": 169 }, { "epoch": 0.43478260869565216, "grad_norm": 6.557422143271899, "learning_rate": 8.8785564535221e-06, "loss": 0.4947, "step": 170 }, { "epoch": 0.4373401534526854, "grad_norm": 7.091654604179938, "learning_rate": 8.86584858250281e-06, "loss": 0.5744, "step": 171 }, { "epoch": 0.4398976982097187, "grad_norm": 6.839942726126082, "learning_rate": 8.853078319255466e-06, "loss": 0.5621, "step": 172 }, { "epoch": 0.4424552429667519, "grad_norm": 6.705305452145543, "learning_rate": 8.840245869883635e-06, "loss": 0.6277, "step": 173 }, { "epoch": 0.44501278772378516, "grad_norm": 6.410914353629197, "learning_rate": 8.827351441494525e-06, "loss": 0.5795, "step": 174 }, { "epoch": 0.4475703324808184, "grad_norm": 6.178508928681474, "learning_rate": 8.814395242195642e-06, "loss": 0.5039, "step": 175 }, { "epoch": 0.45012787723785164, "grad_norm": 5.283454576226612, "learning_rate": 8.80137748109144e-06, "loss": 0.4565, "step": 176 }, { "epoch": 0.45268542199488493, "grad_norm": 5.78397157032685, "learning_rate": 8.78829836827993e-06, "loss": 0.5435, "step": 177 }, { "epoch": 0.45524296675191817, "grad_norm": 5.014711664858047, "learning_rate": 8.77515811484931e-06, "loss": 0.4744, "step": 178 }, { "epoch": 0.4578005115089514, "grad_norm": 5.673473816390766, "learning_rate": 8.761956932874539e-06, "loss": 0.4794, "step": 179 }, { "epoch": 0.46035805626598464, "grad_norm": 6.548843702355434, "learning_rate": 8.748695035413925e-06, "loss": 0.5124, "step": 180 }, { "epoch": 0.4629156010230179, "grad_norm": 5.294743068866496, "learning_rate": 8.735372636505681e-06, "loss": 0.4964, "step": 181 }, { "epoch": 0.46547314578005117, "grad_norm": 5.307479046129796, "learning_rate": 8.72198995116448e-06, "loss": 0.4848, "step": 182 }, { "epoch": 0.4680306905370844, "grad_norm": 5.862290813329295, "learning_rate": 8.708547195377968e-06, "loss": 0.6168, "step": 183 }, { "epoch": 0.47058823529411764, "grad_norm": 6.660472406940894, "learning_rate": 8.695044586103297e-06, "loss": 0.6317, "step": 184 }, { "epoch": 0.4731457800511509, "grad_norm": 6.239173715990654, "learning_rate": 8.68148234126361e-06, "loss": 0.5712, "step": 185 }, { "epoch": 0.47570332480818417, "grad_norm": 6.855169313192307, "learning_rate": 8.667860679744529e-06, "loss": 0.4569, "step": 186 }, { "epoch": 0.4782608695652174, "grad_norm": 6.098286103768463, "learning_rate": 8.65417982139062e-06, "loss": 0.5377, "step": 187 }, { "epoch": 0.48081841432225064, "grad_norm": 6.687927077404218, "learning_rate": 8.640439987001855e-06, "loss": 0.5466, "step": 188 }, { "epoch": 0.4833759590792839, "grad_norm": 5.06986215439764, "learning_rate": 8.626641398330027e-06, "loss": 0.4597, "step": 189 }, { "epoch": 0.4859335038363171, "grad_norm": 6.615207228778788, "learning_rate": 8.612784278075195e-06, "loss": 0.6486, "step": 190 }, { "epoch": 0.4884910485933504, "grad_norm": 8.206032710195597, "learning_rate": 8.598868849882074e-06, "loss": 0.53, "step": 191 }, { "epoch": 0.49104859335038364, "grad_norm": 6.512201203410748, "learning_rate": 8.58489533833643e-06, "loss": 0.5075, "step": 192 }, { "epoch": 0.4936061381074169, "grad_norm": 6.623708653660542, "learning_rate": 8.570863968961456e-06, "loss": 0.4697, "step": 193 }, { "epoch": 0.4961636828644501, "grad_norm": 4.93093459028815, "learning_rate": 8.556774968214134e-06, "loss": 0.5169, "step": 194 }, { "epoch": 0.49872122762148335, "grad_norm": 5.787452319450779, "learning_rate": 8.542628563481577e-06, "loss": 0.5482, "step": 195 }, { "epoch": 0.5012787723785166, "grad_norm": 5.88850708880366, "learning_rate": 8.52842498307736e-06, "loss": 0.6134, "step": 196 }, { "epoch": 0.5038363171355499, "grad_norm": 6.696557687225988, "learning_rate": 8.514164456237835e-06, "loss": 0.6447, "step": 197 }, { "epoch": 0.5063938618925832, "grad_norm": 7.195386414110228, "learning_rate": 8.499847213118431e-06, "loss": 0.5117, "step": 198 }, { "epoch": 0.5089514066496164, "grad_norm": 5.825208905742397, "learning_rate": 8.485473484789944e-06, "loss": 0.5152, "step": 199 }, { "epoch": 0.5115089514066496, "grad_norm": 6.118462869888847, "learning_rate": 8.471043503234796e-06, "loss": 0.6536, "step": 200 }, { "epoch": 0.5140664961636828, "grad_norm": 7.551332041886624, "learning_rate": 8.45655750134331e-06, "loss": 0.6084, "step": 201 }, { "epoch": 0.5166240409207161, "grad_norm": 5.225118027592022, "learning_rate": 8.442015712909926e-06, "loss": 0.4555, "step": 202 }, { "epoch": 0.5191815856777494, "grad_norm": 5.3592959184851265, "learning_rate": 8.427418372629456e-06, "loss": 0.4821, "step": 203 }, { "epoch": 0.5217391304347826, "grad_norm": 6.212770160868918, "learning_rate": 8.412765716093273e-06, "loss": 0.5149, "step": 204 }, { "epoch": 0.5242966751918159, "grad_norm": 5.519259201547804, "learning_rate": 8.398057979785515e-06, "loss": 0.4876, "step": 205 }, { "epoch": 0.5268542199488491, "grad_norm": 6.847844140759948, "learning_rate": 8.383295401079284e-06, "loss": 0.5245, "step": 206 }, { "epoch": 0.5294117647058824, "grad_norm": 6.473882379231715, "learning_rate": 8.368478218232787e-06, "loss": 0.5319, "step": 207 }, { "epoch": 0.5319693094629157, "grad_norm": 4.995240865011453, "learning_rate": 8.353606670385514e-06, "loss": 0.4201, "step": 208 }, { "epoch": 0.5345268542199488, "grad_norm": 5.272197527549254, "learning_rate": 8.338680997554372e-06, "loss": 0.4832, "step": 209 }, { "epoch": 0.5370843989769821, "grad_norm": 5.585738561949535, "learning_rate": 8.3237014406298e-06, "loss": 0.4929, "step": 210 }, { "epoch": 0.5396419437340153, "grad_norm": 6.239332915949274, "learning_rate": 8.308668241371897e-06, "loss": 0.4171, "step": 211 }, { "epoch": 0.5421994884910486, "grad_norm": 5.322513595323884, "learning_rate": 8.293581642406517e-06, "loss": 0.4073, "step": 212 }, { "epoch": 0.5447570332480819, "grad_norm": 7.46575629890418, "learning_rate": 8.278441887221338e-06, "loss": 0.6626, "step": 213 }, { "epoch": 0.5473145780051151, "grad_norm": 6.052661000824651, "learning_rate": 8.263249220161957e-06, "loss": 0.5068, "step": 214 }, { "epoch": 0.5498721227621484, "grad_norm": 6.708569886061961, "learning_rate": 8.248003886427927e-06, "loss": 0.4966, "step": 215 }, { "epoch": 0.5524296675191815, "grad_norm": 5.68545952396897, "learning_rate": 8.232706132068806e-06, "loss": 0.3861, "step": 216 }, { "epoch": 0.5549872122762148, "grad_norm": 5.421452942064916, "learning_rate": 8.217356203980187e-06, "loss": 0.3885, "step": 217 }, { "epoch": 0.5575447570332481, "grad_norm": 6.028892220556533, "learning_rate": 8.201954349899712e-06, "loss": 0.5848, "step": 218 }, { "epoch": 0.5601023017902813, "grad_norm": 5.85865954700368, "learning_rate": 8.186500818403076e-06, "loss": 0.4014, "step": 219 }, { "epoch": 0.5626598465473146, "grad_norm": 7.224336653003557, "learning_rate": 8.17099585890001e-06, "loss": 0.6191, "step": 220 }, { "epoch": 0.5652173913043478, "grad_norm": 6.30772616724702, "learning_rate": 8.155439721630265e-06, "loss": 0.4756, "step": 221 }, { "epoch": 0.5677749360613811, "grad_norm": 6.236289350658551, "learning_rate": 8.139832657659557e-06, "loss": 0.4964, "step": 222 }, { "epoch": 0.5703324808184144, "grad_norm": 6.209791638515158, "learning_rate": 8.124174918875532e-06, "loss": 0.5958, "step": 223 }, { "epoch": 0.5728900255754475, "grad_norm": 6.377139101398714, "learning_rate": 8.108466757983695e-06, "loss": 0.3906, "step": 224 }, { "epoch": 0.5754475703324808, "grad_norm": 5.907378705805276, "learning_rate": 8.092708428503324e-06, "loss": 0.5376, "step": 225 }, { "epoch": 0.578005115089514, "grad_norm": 6.743429798340147, "learning_rate": 8.076900184763394e-06, "loss": 0.4802, "step": 226 }, { "epoch": 0.5805626598465473, "grad_norm": 5.376964104341389, "learning_rate": 8.061042281898453e-06, "loss": 0.4509, "step": 227 }, { "epoch": 0.5831202046035806, "grad_norm": 6.511105613676549, "learning_rate": 8.04513497584452e-06, "loss": 0.4214, "step": 228 }, { "epoch": 0.5856777493606138, "grad_norm": 6.852114152108356, "learning_rate": 8.02917852333495e-06, "loss": 0.6038, "step": 229 }, { "epoch": 0.5882352941176471, "grad_norm": 5.572255565814968, "learning_rate": 8.013173181896283e-06, "loss": 0.5224, "step": 230 }, { "epoch": 0.5907928388746803, "grad_norm": 7.0760399259778435, "learning_rate": 7.9971192098441e-06, "loss": 0.4502, "step": 231 }, { "epoch": 0.5933503836317136, "grad_norm": 5.207560218712082, "learning_rate": 7.981016866278843e-06, "loss": 0.4027, "step": 232 }, { "epoch": 0.5959079283887468, "grad_norm": 5.796635616276233, "learning_rate": 7.964866411081645e-06, "loss": 0.5675, "step": 233 }, { "epoch": 0.59846547314578, "grad_norm": 5.708164631421739, "learning_rate": 7.94866810491012e-06, "loss": 0.4437, "step": 234 }, { "epoch": 0.6010230179028133, "grad_norm": 6.436431203161013, "learning_rate": 7.93242220919417e-06, "loss": 0.5474, "step": 235 }, { "epoch": 0.6035805626598465, "grad_norm": 5.574144310350591, "learning_rate": 7.916128986131761e-06, "loss": 0.5439, "step": 236 }, { "epoch": 0.6061381074168798, "grad_norm": 4.8240456797819835, "learning_rate": 7.899788698684687e-06, "loss": 0.4686, "step": 237 }, { "epoch": 0.6086956521739131, "grad_norm": 6.5914849870729055, "learning_rate": 7.883401610574338e-06, "loss": 0.5512, "step": 238 }, { "epoch": 0.6112531969309463, "grad_norm": 4.739332602957458, "learning_rate": 7.866967986277423e-06, "loss": 0.4204, "step": 239 }, { "epoch": 0.6138107416879796, "grad_norm": 6.116120073202256, "learning_rate": 7.850488091021726e-06, "loss": 0.5596, "step": 240 }, { "epoch": 0.6163682864450127, "grad_norm": 6.299579832647148, "learning_rate": 7.833962190781809e-06, "loss": 0.5729, "step": 241 }, { "epoch": 0.618925831202046, "grad_norm": 5.77832842987742, "learning_rate": 7.817390552274721e-06, "loss": 0.4062, "step": 242 }, { "epoch": 0.6214833759590793, "grad_norm": 5.6928424134185365, "learning_rate": 7.800773442955703e-06, "loss": 0.562, "step": 243 }, { "epoch": 0.6240409207161125, "grad_norm": 5.754032663780959, "learning_rate": 7.784111131013858e-06, "loss": 0.4763, "step": 244 }, { "epoch": 0.6265984654731458, "grad_norm": 6.1254790208347, "learning_rate": 7.767403885367832e-06, "loss": 0.4931, "step": 245 }, { "epoch": 0.629156010230179, "grad_norm": 6.088136454995643, "learning_rate": 7.750651975661471e-06, "loss": 0.5366, "step": 246 }, { "epoch": 0.6317135549872123, "grad_norm": 6.525801581028963, "learning_rate": 7.733855672259472e-06, "loss": 0.5869, "step": 247 }, { "epoch": 0.6342710997442456, "grad_norm": 5.142226783902718, "learning_rate": 7.717015246243012e-06, "loss": 0.4107, "step": 248 }, { "epoch": 0.6368286445012787, "grad_norm": 5.884475685733821, "learning_rate": 7.700130969405377e-06, "loss": 0.5575, "step": 249 }, { "epoch": 0.639386189258312, "grad_norm": 5.430956350007929, "learning_rate": 7.683203114247587e-06, "loss": 0.4316, "step": 250 }, { "epoch": 0.6419437340153452, "grad_norm": 5.852470007876826, "learning_rate": 7.66623195397397e-06, "loss": 0.5228, "step": 251 }, { "epoch": 0.6445012787723785, "grad_norm": 6.316931409524609, "learning_rate": 7.649217762487786e-06, "loss": 0.6069, "step": 252 }, { "epoch": 0.6470588235294118, "grad_norm": 5.981156306158716, "learning_rate": 7.63216081438678e-06, "loss": 0.4525, "step": 253 }, { "epoch": 0.649616368286445, "grad_norm": 6.193124213697377, "learning_rate": 7.615061384958764e-06, "loss": 0.6367, "step": 254 }, { "epoch": 0.6521739130434783, "grad_norm": 6.145394633019291, "learning_rate": 7.597919750177168e-06, "loss": 0.5622, "step": 255 }, { "epoch": 0.6547314578005116, "grad_norm": 6.076895232152138, "learning_rate": 7.580736186696593e-06, "loss": 0.5016, "step": 256 }, { "epoch": 0.6572890025575447, "grad_norm": 5.562852949209647, "learning_rate": 7.563510971848339e-06, "loss": 0.5739, "step": 257 }, { "epoch": 0.659846547314578, "grad_norm": 6.358261854476947, "learning_rate": 7.546244383635929e-06, "loss": 0.5783, "step": 258 }, { "epoch": 0.6624040920716112, "grad_norm": 5.61873313563532, "learning_rate": 7.528936700730627e-06, "loss": 0.5671, "step": 259 }, { "epoch": 0.6649616368286445, "grad_norm": 5.007471092297137, "learning_rate": 7.5115882024669375e-06, "loss": 0.4238, "step": 260 }, { "epoch": 0.6675191815856778, "grad_norm": 5.02718344671977, "learning_rate": 7.494199168838099e-06, "loss": 0.431, "step": 261 }, { "epoch": 0.670076726342711, "grad_norm": 5.822025275525143, "learning_rate": 7.476769880491561e-06, "loss": 0.555, "step": 262 }, { "epoch": 0.6726342710997443, "grad_norm": 5.852672525450696, "learning_rate": 7.459300618724462e-06, "loss": 0.4537, "step": 263 }, { "epoch": 0.6751918158567775, "grad_norm": 6.36830749484907, "learning_rate": 7.44179166547908e-06, "loss": 0.5466, "step": 264 }, { "epoch": 0.6777493606138107, "grad_norm": 5.187641704740303, "learning_rate": 7.42424330333829e-06, "loss": 0.4966, "step": 265 }, { "epoch": 0.680306905370844, "grad_norm": 5.862839321803861, "learning_rate": 7.406655815520998e-06, "loss": 0.4902, "step": 266 }, { "epoch": 0.6828644501278772, "grad_norm": 6.529161660718858, "learning_rate": 7.389029485877577e-06, "loss": 0.493, "step": 267 }, { "epoch": 0.6854219948849105, "grad_norm": 5.732050686572585, "learning_rate": 7.371364598885276e-06, "loss": 0.4744, "step": 268 }, { "epoch": 0.6879795396419437, "grad_norm": 5.533174363200175, "learning_rate": 7.353661439643638e-06, "loss": 0.3833, "step": 269 }, { "epoch": 0.690537084398977, "grad_norm": 5.768168615527615, "learning_rate": 7.335920293869891e-06, "loss": 0.423, "step": 270 }, { "epoch": 0.6930946291560103, "grad_norm": 5.852266644103708, "learning_rate": 7.318141447894344e-06, "loss": 0.3371, "step": 271 }, { "epoch": 0.6956521739130435, "grad_norm": 6.038563114564619, "learning_rate": 7.300325188655762e-06, "loss": 0.4891, "step": 272 }, { "epoch": 0.6982097186700768, "grad_norm": 6.280803826327464, "learning_rate": 7.28247180369673e-06, "loss": 0.5385, "step": 273 }, { "epoch": 0.7007672634271099, "grad_norm": 7.292365659382516, "learning_rate": 7.264581581159024e-06, "loss": 0.6148, "step": 274 }, { "epoch": 0.7033248081841432, "grad_norm": 6.763240999324924, "learning_rate": 7.246654809778951e-06, "loss": 0.5272, "step": 275 }, { "epoch": 0.7058823529411765, "grad_norm": 6.444401975777849, "learning_rate": 7.2286917788826926e-06, "loss": 0.4879, "step": 276 }, { "epoch": 0.7084398976982097, "grad_norm": 5.037923525497081, "learning_rate": 7.210692778381634e-06, "loss": 0.5377, "step": 277 }, { "epoch": 0.710997442455243, "grad_norm": 6.327806611970394, "learning_rate": 7.192658098767686e-06, "loss": 0.4654, "step": 278 }, { "epoch": 0.7135549872122762, "grad_norm": 5.832786135763086, "learning_rate": 7.174588031108598e-06, "loss": 0.5921, "step": 279 }, { "epoch": 0.7161125319693095, "grad_norm": 6.032098832742715, "learning_rate": 7.1564828670432595e-06, "loss": 0.5032, "step": 280 }, { "epoch": 0.7186700767263428, "grad_norm": 7.528807903355475, "learning_rate": 7.138342898776989e-06, "loss": 0.5143, "step": 281 }, { "epoch": 0.7212276214833759, "grad_norm": 5.662236290695636, "learning_rate": 7.120168419076825e-06, "loss": 0.5752, "step": 282 }, { "epoch": 0.7237851662404092, "grad_norm": 5.922836277812778, "learning_rate": 7.101959721266798e-06, "loss": 0.5907, "step": 283 }, { "epoch": 0.7263427109974424, "grad_norm": 6.258012070363337, "learning_rate": 7.083717099223192e-06, "loss": 0.5447, "step": 284 }, { "epoch": 0.7289002557544757, "grad_norm": 5.261480296532744, "learning_rate": 7.0654408473698084e-06, "loss": 0.4521, "step": 285 }, { "epoch": 0.731457800511509, "grad_norm": 5.918110722172615, "learning_rate": 7.047131260673214e-06, "loss": 0.4637, "step": 286 }, { "epoch": 0.7340153452685422, "grad_norm": 5.741282290810403, "learning_rate": 7.0287886346379755e-06, "loss": 0.4131, "step": 287 }, { "epoch": 0.7365728900255755, "grad_norm": 5.943919434881143, "learning_rate": 7.010413265301888e-06, "loss": 0.4712, "step": 288 }, { "epoch": 0.7391304347826086, "grad_norm": 6.1059644383499885, "learning_rate": 6.9920054492312086e-06, "loss": 0.6022, "step": 289 }, { "epoch": 0.7416879795396419, "grad_norm": 6.884474367848085, "learning_rate": 6.97356548351586e-06, "loss": 0.5212, "step": 290 }, { "epoch": 0.7442455242966752, "grad_norm": 5.758493578440039, "learning_rate": 6.9550936657646386e-06, "loss": 0.507, "step": 291 }, { "epoch": 0.7468030690537084, "grad_norm": 5.5122177192122415, "learning_rate": 6.936590294100414e-06, "loss": 0.4096, "step": 292 }, { "epoch": 0.7493606138107417, "grad_norm": 6.529147733060143, "learning_rate": 6.918055667155311e-06, "loss": 0.4668, "step": 293 }, { "epoch": 0.7519181585677749, "grad_norm": 5.580434227838566, "learning_rate": 6.899490084065897e-06, "loss": 0.4825, "step": 294 }, { "epoch": 0.7544757033248082, "grad_norm": 6.141771398723171, "learning_rate": 6.8808938444683505e-06, "loss": 0.5189, "step": 295 }, { "epoch": 0.7570332480818415, "grad_norm": 6.258445455734687, "learning_rate": 6.862267248493624e-06, "loss": 0.4217, "step": 296 }, { "epoch": 0.7595907928388747, "grad_norm": 5.577447249480196, "learning_rate": 6.843610596762606e-06, "loss": 0.4574, "step": 297 }, { "epoch": 0.7621483375959079, "grad_norm": 6.508949986966596, "learning_rate": 6.824924190381257e-06, "loss": 0.4512, "step": 298 }, { "epoch": 0.7647058823529411, "grad_norm": 6.487835617711066, "learning_rate": 6.806208330935766e-06, "loss": 0.4817, "step": 299 }, { "epoch": 0.7672634271099744, "grad_norm": 6.733508290939032, "learning_rate": 6.7874633204876705e-06, "loss": 0.4648, "step": 300 }, { "epoch": 0.7698209718670077, "grad_norm": 6.6391089444926195, "learning_rate": 6.768689461568987e-06, "loss": 0.4959, "step": 301 }, { "epoch": 0.7723785166240409, "grad_norm": 5.426573235588597, "learning_rate": 6.7498870571773275e-06, "loss": 0.4101, "step": 302 }, { "epoch": 0.7749360613810742, "grad_norm": 5.27403325499086, "learning_rate": 6.731056410771008e-06, "loss": 0.4183, "step": 303 }, { "epoch": 0.7774936061381074, "grad_norm": 5.602097858442588, "learning_rate": 6.712197826264154e-06, "loss": 0.4712, "step": 304 }, { "epoch": 0.7800511508951407, "grad_norm": 5.602023911663575, "learning_rate": 6.69331160802179e-06, "loss": 0.376, "step": 305 }, { "epoch": 0.782608695652174, "grad_norm": 6.057331404811353, "learning_rate": 6.674398060854931e-06, "loss": 0.3333, "step": 306 }, { "epoch": 0.7851662404092071, "grad_norm": 5.453910778706793, "learning_rate": 6.655457490015667e-06, "loss": 0.5251, "step": 307 }, { "epoch": 0.7877237851662404, "grad_norm": 6.209463244054028, "learning_rate": 6.636490201192229e-06, "loss": 0.5256, "step": 308 }, { "epoch": 0.7902813299232737, "grad_norm": 5.6125006489249145, "learning_rate": 6.617496500504056e-06, "loss": 0.35, "step": 309 }, { "epoch": 0.7928388746803069, "grad_norm": 6.414498410153366, "learning_rate": 6.5984766944968636e-06, "loss": 0.5181, "step": 310 }, { "epoch": 0.7953964194373402, "grad_norm": 6.101747662704522, "learning_rate": 6.579431090137681e-06, "loss": 0.4106, "step": 311 }, { "epoch": 0.7979539641943734, "grad_norm": 6.462777333488606, "learning_rate": 6.560359994809916e-06, "loss": 0.6125, "step": 312 }, { "epoch": 0.8005115089514067, "grad_norm": 5.9920053305051875, "learning_rate": 6.541263716308375e-06, "loss": 0.4968, "step": 313 }, { "epoch": 0.80306905370844, "grad_norm": 6.671005371719509, "learning_rate": 6.522142562834307e-06, "loss": 0.5637, "step": 314 }, { "epoch": 0.8056265984654731, "grad_norm": 5.361336122168199, "learning_rate": 6.502996842990431e-06, "loss": 0.4208, "step": 315 }, { "epoch": 0.8081841432225064, "grad_norm": 5.670064103939166, "learning_rate": 6.483826865775941e-06, "loss": 0.5278, "step": 316 }, { "epoch": 0.8107416879795396, "grad_norm": 5.5103998057715105, "learning_rate": 6.46463294058154e-06, "loss": 0.4007, "step": 317 }, { "epoch": 0.8132992327365729, "grad_norm": 5.659095784663181, "learning_rate": 6.445415377184427e-06, "loss": 0.4742, "step": 318 }, { "epoch": 0.8158567774936062, "grad_norm": 6.30132561670194, "learning_rate": 6.426174485743309e-06, "loss": 0.4078, "step": 319 }, { "epoch": 0.8184143222506394, "grad_norm": 5.643268096385628, "learning_rate": 6.4069105767933944e-06, "loss": 0.46, "step": 320 }, { "epoch": 0.8209718670076727, "grad_norm": 7.788725418859061, "learning_rate": 6.387623961241375e-06, "loss": 0.6119, "step": 321 }, { "epoch": 0.8235294117647058, "grad_norm": 5.927896020719375, "learning_rate": 6.368314950360416e-06, "loss": 0.5225, "step": 322 }, { "epoch": 0.8260869565217391, "grad_norm": 5.296955151964955, "learning_rate": 6.348983855785122e-06, "loss": 0.3126, "step": 323 }, { "epoch": 0.8286445012787724, "grad_norm": 5.095733276738074, "learning_rate": 6.3296309895065215e-06, "loss": 0.3639, "step": 324 }, { "epoch": 0.8312020460358056, "grad_norm": 6.080988913298908, "learning_rate": 6.310256663867019e-06, "loss": 0.5063, "step": 325 }, { "epoch": 0.8337595907928389, "grad_norm": 7.036245894709906, "learning_rate": 6.290861191555359e-06, "loss": 0.4578, "step": 326 }, { "epoch": 0.8363171355498721, "grad_norm": 5.580633409599807, "learning_rate": 6.271444885601583e-06, "loss": 0.4639, "step": 327 }, { "epoch": 0.8388746803069054, "grad_norm": 6.523213064272758, "learning_rate": 6.252008059371968e-06, "loss": 0.4699, "step": 328 }, { "epoch": 0.8414322250639387, "grad_norm": 4.85798591447732, "learning_rate": 6.2325510265639785e-06, "loss": 0.2973, "step": 329 }, { "epoch": 0.8439897698209718, "grad_norm": 6.805564012992218, "learning_rate": 6.213074101201202e-06, "loss": 0.4894, "step": 330 }, { "epoch": 0.8465473145780051, "grad_norm": 6.052583253032932, "learning_rate": 6.193577597628268e-06, "loss": 0.5193, "step": 331 }, { "epoch": 0.8491048593350383, "grad_norm": 6.230894069829904, "learning_rate": 6.174061830505801e-06, "loss": 0.5028, "step": 332 }, { "epoch": 0.8516624040920716, "grad_norm": 7.201079594756455, "learning_rate": 6.154527114805312e-06, "loss": 0.618, "step": 333 }, { "epoch": 0.8542199488491049, "grad_norm": 6.600993850416883, "learning_rate": 6.1349737658041385e-06, "loss": 0.5133, "step": 334 }, { "epoch": 0.8567774936061381, "grad_norm": 5.724822321191247, "learning_rate": 6.115402099080345e-06, "loss": 0.4838, "step": 335 }, { "epoch": 0.8593350383631714, "grad_norm": 5.613108225355487, "learning_rate": 6.095812430507627e-06, "loss": 0.3442, "step": 336 }, { "epoch": 0.8618925831202046, "grad_norm": 5.965477408637899, "learning_rate": 6.076205076250227e-06, "loss": 0.5109, "step": 337 }, { "epoch": 0.8644501278772379, "grad_norm": 5.9857901067552, "learning_rate": 6.056580352757813e-06, "loss": 0.4511, "step": 338 }, { "epoch": 0.8670076726342711, "grad_norm": 6.174131332105638, "learning_rate": 6.036938576760388e-06, "loss": 0.4419, "step": 339 }, { "epoch": 0.8695652173913043, "grad_norm": 5.075847963553367, "learning_rate": 6.0172800652631706e-06, "loss": 0.3777, "step": 340 }, { "epoch": 0.8721227621483376, "grad_norm": 7.190594951132575, "learning_rate": 5.997605135541472e-06, "loss": 0.5106, "step": 341 }, { "epoch": 0.8746803069053708, "grad_norm": 6.747196680979683, "learning_rate": 5.977914105135594e-06, "loss": 0.4762, "step": 342 }, { "epoch": 0.8772378516624041, "grad_norm": 5.714370912906624, "learning_rate": 5.9582072918456805e-06, "loss": 0.362, "step": 343 }, { "epoch": 0.8797953964194374, "grad_norm": 5.236876408531075, "learning_rate": 5.938485013726612e-06, "loss": 0.3947, "step": 344 }, { "epoch": 0.8823529411764706, "grad_norm": 5.116217278468624, "learning_rate": 5.918747589082853e-06, "loss": 0.4747, "step": 345 }, { "epoch": 0.8849104859335039, "grad_norm": 4.749747087957306, "learning_rate": 5.898995336463326e-06, "loss": 0.4274, "step": 346 }, { "epoch": 0.887468030690537, "grad_norm": 5.230418676152823, "learning_rate": 5.879228574656269e-06, "loss": 0.3441, "step": 347 }, { "epoch": 0.8900255754475703, "grad_norm": 6.303242756857959, "learning_rate": 5.859447622684084e-06, "loss": 0.5131, "step": 348 }, { "epoch": 0.8925831202046036, "grad_norm": 5.4435933577635645, "learning_rate": 5.839652799798197e-06, "loss": 0.4243, "step": 349 }, { "epoch": 0.8951406649616368, "grad_norm": 6.921408500128556, "learning_rate": 5.819844425473899e-06, "loss": 0.5549, "step": 350 }, { "epoch": 0.8976982097186701, "grad_norm": 6.635487149449039, "learning_rate": 5.800022819405194e-06, "loss": 0.5061, "step": 351 }, { "epoch": 0.9002557544757033, "grad_norm": 6.3209563326259515, "learning_rate": 5.780188301499636e-06, "loss": 0.5999, "step": 352 }, { "epoch": 0.9028132992327366, "grad_norm": 6.077422261762329, "learning_rate": 5.760341191873167e-06, "loss": 0.5111, "step": 353 }, { "epoch": 0.9053708439897699, "grad_norm": 5.867259188193717, "learning_rate": 5.740481810844952e-06, "loss": 0.4771, "step": 354 }, { "epoch": 0.907928388746803, "grad_norm": 4.777893746653604, "learning_rate": 5.720610478932211e-06, "loss": 0.3242, "step": 355 }, { "epoch": 0.9104859335038363, "grad_norm": 5.764419825204445, "learning_rate": 5.700727516845038e-06, "loss": 0.3306, "step": 356 }, { "epoch": 0.9130434782608695, "grad_norm": 5.510318712985209, "learning_rate": 5.680833245481234e-06, "loss": 0.4642, "step": 357 }, { "epoch": 0.9156010230179028, "grad_norm": 6.53182819796998, "learning_rate": 5.660927985921122e-06, "loss": 0.51, "step": 358 }, { "epoch": 0.9181585677749361, "grad_norm": 6.086318246788371, "learning_rate": 5.641012059422369e-06, "loss": 0.5472, "step": 359 }, { "epoch": 0.9207161125319693, "grad_norm": 5.544550425971534, "learning_rate": 5.621085787414799e-06, "loss": 0.4603, "step": 360 }, { "epoch": 0.9232736572890026, "grad_norm": 6.1043161575819616, "learning_rate": 5.601149491495206e-06, "loss": 0.485, "step": 361 }, { "epoch": 0.9258312020460358, "grad_norm": 5.827986489165051, "learning_rate": 5.581203493422161e-06, "loss": 0.5864, "step": 362 }, { "epoch": 0.928388746803069, "grad_norm": 5.172049016763376, "learning_rate": 5.561248115110822e-06, "loss": 0.4517, "step": 363 }, { "epoch": 0.9309462915601023, "grad_norm": 6.640832021653832, "learning_rate": 5.541283678627742e-06, "loss": 0.3703, "step": 364 }, { "epoch": 0.9335038363171355, "grad_norm": 5.361432485515485, "learning_rate": 5.521310506185661e-06, "loss": 0.4262, "step": 365 }, { "epoch": 0.9360613810741688, "grad_norm": 5.808037599792696, "learning_rate": 5.501328920138314e-06, "loss": 0.544, "step": 366 }, { "epoch": 0.9386189258312021, "grad_norm": 6.0541662716251095, "learning_rate": 5.481339242975227e-06, "loss": 0.4024, "step": 367 }, { "epoch": 0.9411764705882353, "grad_norm": 6.232047697753629, "learning_rate": 5.46134179731651e-06, "loss": 0.4862, "step": 368 }, { "epoch": 0.9437340153452686, "grad_norm": 6.000932213910604, "learning_rate": 5.441336905907653e-06, "loss": 0.4635, "step": 369 }, { "epoch": 0.9462915601023018, "grad_norm": 7.611546241496742, "learning_rate": 5.421324891614312e-06, "loss": 0.4135, "step": 370 }, { "epoch": 0.948849104859335, "grad_norm": 5.362330490202002, "learning_rate": 5.4013060774171055e-06, "loss": 0.4506, "step": 371 }, { "epoch": 0.9514066496163683, "grad_norm": 5.346120483245015, "learning_rate": 5.3812807864063946e-06, "loss": 0.4576, "step": 372 }, { "epoch": 0.9539641943734015, "grad_norm": 6.100965821026688, "learning_rate": 5.361249341777075e-06, "loss": 0.5165, "step": 373 }, { "epoch": 0.9565217391304348, "grad_norm": 4.676613084116823, "learning_rate": 5.341212066823356e-06, "loss": 0.4383, "step": 374 }, { "epoch": 0.959079283887468, "grad_norm": 6.564693244827939, "learning_rate": 5.321169284933543e-06, "loss": 0.5044, "step": 375 }, { "epoch": 0.9616368286445013, "grad_norm": 4.752286756978166, "learning_rate": 5.3011213195848245e-06, "loss": 0.5422, "step": 376 }, { "epoch": 0.9641943734015346, "grad_norm": 6.298216680263071, "learning_rate": 5.281068494338039e-06, "loss": 0.3751, "step": 377 }, { "epoch": 0.9667519181585678, "grad_norm": 5.041453892793781, "learning_rate": 5.26101113283247e-06, "loss": 0.3732, "step": 378 }, { "epoch": 0.969309462915601, "grad_norm": 6.072922237394072, "learning_rate": 5.240949558780605e-06, "loss": 0.4873, "step": 379 }, { "epoch": 0.9718670076726342, "grad_norm": 5.712375997105542, "learning_rate": 5.220884095962924e-06, "loss": 0.4877, "step": 380 }, { "epoch": 0.9744245524296675, "grad_norm": 5.5897510763688585, "learning_rate": 5.200815068222666e-06, "loss": 0.386, "step": 381 }, { "epoch": 0.9769820971867008, "grad_norm": 6.160592176666778, "learning_rate": 5.1807427994606065e-06, "loss": 0.369, "step": 382 }, { "epoch": 0.979539641943734, "grad_norm": 5.482848722330923, "learning_rate": 5.1606676136298305e-06, "loss": 0.4618, "step": 383 }, { "epoch": 0.9820971867007673, "grad_norm": 5.96411837712334, "learning_rate": 5.140589834730503e-06, "loss": 0.4286, "step": 384 }, { "epoch": 0.9846547314578005, "grad_norm": 6.021924443213883, "learning_rate": 5.120509786804635e-06, "loss": 0.4545, "step": 385 }, { "epoch": 0.9872122762148338, "grad_norm": 4.860619712046072, "learning_rate": 5.100427793930862e-06, "loss": 0.4847, "step": 386 }, { "epoch": 0.989769820971867, "grad_norm": 6.16249371180202, "learning_rate": 5.08034418021921e-06, "loss": 0.4119, "step": 387 }, { "epoch": 0.9923273657289002, "grad_norm": 5.190813395638479, "learning_rate": 5.06025926980586e-06, "loss": 0.3609, "step": 388 }, { "epoch": 0.9948849104859335, "grad_norm": 5.3092155597782025, "learning_rate": 5.040173386847926e-06, "loss": 0.3655, "step": 389 }, { "epoch": 0.9974424552429667, "grad_norm": 5.7321767766000935, "learning_rate": 5.0200868555182155e-06, "loss": 0.5465, "step": 390 }, { "epoch": 1.0, "grad_norm": 5.719257228559791, "learning_rate": 5e-06, "loss": 0.4347, "step": 391 }, { "epoch": 1.0025575447570332, "grad_norm": 4.2391486408050145, "learning_rate": 4.979913144481785e-06, "loss": 0.2179, "step": 392 }, { "epoch": 1.0051150895140666, "grad_norm": 4.604106412801713, "learning_rate": 4.959826613152074e-06, "loss": 0.2675, "step": 393 }, { "epoch": 1.0076726342710998, "grad_norm": 4.113700302207306, "learning_rate": 4.939740730194141e-06, "loss": 0.2586, "step": 394 }, { "epoch": 1.010230179028133, "grad_norm": 4.894518417900452, "learning_rate": 4.919655819780792e-06, "loss": 0.278, "step": 395 }, { "epoch": 1.0127877237851663, "grad_norm": 4.3903147870940815, "learning_rate": 4.899572206069138e-06, "loss": 0.2175, "step": 396 }, { "epoch": 1.0153452685421995, "grad_norm": 4.395292367085451, "learning_rate": 4.879490213195366e-06, "loss": 0.2597, "step": 397 }, { "epoch": 1.0179028132992327, "grad_norm": 4.399892144315872, "learning_rate": 4.8594101652694996e-06, "loss": 0.2806, "step": 398 }, { "epoch": 1.020460358056266, "grad_norm": 5.345621237626395, "learning_rate": 4.839332386370171e-06, "loss": 0.2571, "step": 399 }, { "epoch": 1.0230179028132993, "grad_norm": 4.533667138931473, "learning_rate": 4.819257200539394e-06, "loss": 0.2646, "step": 400 }, { "epoch": 1.0255754475703325, "grad_norm": 5.037019368108468, "learning_rate": 4.799184931777337e-06, "loss": 0.1862, "step": 401 }, { "epoch": 1.0281329923273657, "grad_norm": 5.628048303078724, "learning_rate": 4.779115904037079e-06, "loss": 0.239, "step": 402 }, { "epoch": 1.030690537084399, "grad_norm": 6.657437275552853, "learning_rate": 4.759050441219395e-06, "loss": 0.2032, "step": 403 }, { "epoch": 1.0332480818414322, "grad_norm": 7.066364647253674, "learning_rate": 4.738988867167531e-06, "loss": 0.2686, "step": 404 }, { "epoch": 1.0358056265984654, "grad_norm": 6.242504066160711, "learning_rate": 4.718931505661961e-06, "loss": 0.2039, "step": 405 }, { "epoch": 1.0383631713554988, "grad_norm": 7.053347520596888, "learning_rate": 4.698878680415176e-06, "loss": 0.2677, "step": 406 }, { "epoch": 1.040920716112532, "grad_norm": 6.980982091790442, "learning_rate": 4.678830715066458e-06, "loss": 0.2986, "step": 407 }, { "epoch": 1.0434782608695652, "grad_norm": 7.1587957747958715, "learning_rate": 4.6587879331766465e-06, "loss": 0.2742, "step": 408 }, { "epoch": 1.0460358056265984, "grad_norm": 6.506842088570043, "learning_rate": 4.638750658222927e-06, "loss": 0.327, "step": 409 }, { "epoch": 1.0485933503836318, "grad_norm": 5.343374587345889, "learning_rate": 4.618719213593605e-06, "loss": 0.2399, "step": 410 }, { "epoch": 1.051150895140665, "grad_norm": 5.952604432226821, "learning_rate": 4.598693922582896e-06, "loss": 0.3362, "step": 411 }, { "epoch": 1.0537084398976981, "grad_norm": 5.292017124731555, "learning_rate": 4.5786751083856895e-06, "loss": 0.2229, "step": 412 }, { "epoch": 1.0562659846547315, "grad_norm": 5.7674633387114, "learning_rate": 4.558663094092348e-06, "loss": 0.2716, "step": 413 }, { "epoch": 1.0588235294117647, "grad_norm": 4.654675739900228, "learning_rate": 4.53865820268349e-06, "loss": 0.2439, "step": 414 }, { "epoch": 1.061381074168798, "grad_norm": 6.404992169914481, "learning_rate": 4.518660757024774e-06, "loss": 0.2713, "step": 415 }, { "epoch": 1.0639386189258313, "grad_norm": 4.940754639892674, "learning_rate": 4.498671079861686e-06, "loss": 0.2225, "step": 416 }, { "epoch": 1.0664961636828645, "grad_norm": 4.76891987704093, "learning_rate": 4.478689493814341e-06, "loss": 0.2676, "step": 417 }, { "epoch": 1.0690537084398977, "grad_norm": 5.023886841031966, "learning_rate": 4.4587163213722595e-06, "loss": 0.2594, "step": 418 }, { "epoch": 1.0716112531969308, "grad_norm": 4.184649476108494, "learning_rate": 4.438751884889179e-06, "loss": 0.1791, "step": 419 }, { "epoch": 1.0741687979539642, "grad_norm": 5.029459043202698, "learning_rate": 4.41879650657784e-06, "loss": 0.2588, "step": 420 }, { "epoch": 1.0767263427109974, "grad_norm": 4.795405784853634, "learning_rate": 4.398850508504795e-06, "loss": 0.2273, "step": 421 }, { "epoch": 1.0792838874680306, "grad_norm": 4.137020254719259, "learning_rate": 4.3789142125852015e-06, "loss": 0.2682, "step": 422 }, { "epoch": 1.081841432225064, "grad_norm": 5.222698810349911, "learning_rate": 4.358987940577631e-06, "loss": 0.2186, "step": 423 }, { "epoch": 1.0843989769820972, "grad_norm": 5.743293194926402, "learning_rate": 4.339072014078879e-06, "loss": 0.1999, "step": 424 }, { "epoch": 1.0869565217391304, "grad_norm": 5.052413496789277, "learning_rate": 4.319166754518768e-06, "loss": 0.2047, "step": 425 }, { "epoch": 1.0895140664961638, "grad_norm": 6.345187477594701, "learning_rate": 4.299272483154963e-06, "loss": 0.314, "step": 426 }, { "epoch": 1.092071611253197, "grad_norm": 5.440791188116355, "learning_rate": 4.27938952106779e-06, "loss": 0.1874, "step": 427 }, { "epoch": 1.0946291560102301, "grad_norm": 4.59572793131487, "learning_rate": 4.259518189155049e-06, "loss": 0.2259, "step": 428 }, { "epoch": 1.0971867007672633, "grad_norm": 7.1276467094843925, "learning_rate": 4.2396588081268355e-06, "loss": 0.2556, "step": 429 }, { "epoch": 1.0997442455242967, "grad_norm": 6.0356348861027564, "learning_rate": 4.219811698500365e-06, "loss": 0.3379, "step": 430 }, { "epoch": 1.10230179028133, "grad_norm": 5.364328953584096, "learning_rate": 4.199977180594807e-06, "loss": 0.1789, "step": 431 }, { "epoch": 1.104859335038363, "grad_norm": 6.395865991308588, "learning_rate": 4.1801555745261025e-06, "loss": 0.3364, "step": 432 }, { "epoch": 1.1074168797953965, "grad_norm": 5.307623822753343, "learning_rate": 4.160347200201804e-06, "loss": 0.3123, "step": 433 }, { "epoch": 1.1099744245524297, "grad_norm": 5.8960456968886055, "learning_rate": 4.140552377315918e-06, "loss": 0.3115, "step": 434 }, { "epoch": 1.1125319693094629, "grad_norm": 5.5625443910431915, "learning_rate": 4.120771425343733e-06, "loss": 0.2276, "step": 435 }, { "epoch": 1.1150895140664963, "grad_norm": 5.922218273593377, "learning_rate": 4.101004663536675e-06, "loss": 0.2569, "step": 436 }, { "epoch": 1.1176470588235294, "grad_norm": 5.263570887474491, "learning_rate": 4.081252410917148e-06, "loss": 0.2444, "step": 437 }, { "epoch": 1.1202046035805626, "grad_norm": 6.181574590354267, "learning_rate": 4.061514986273391e-06, "loss": 0.251, "step": 438 }, { "epoch": 1.1227621483375958, "grad_norm": 6.1580046071843935, "learning_rate": 4.041792708154321e-06, "loss": 0.2596, "step": 439 }, { "epoch": 1.1253196930946292, "grad_norm": 5.458696870991704, "learning_rate": 4.022085894864408e-06, "loss": 0.2237, "step": 440 }, { "epoch": 1.1278772378516624, "grad_norm": 4.70048293838936, "learning_rate": 4.0023948644585294e-06, "loss": 0.2476, "step": 441 }, { "epoch": 1.1304347826086956, "grad_norm": 4.466773072234314, "learning_rate": 3.982719934736832e-06, "loss": 0.2, "step": 442 }, { "epoch": 1.132992327365729, "grad_norm": 4.27013775097264, "learning_rate": 3.963061423239612e-06, "loss": 0.1978, "step": 443 }, { "epoch": 1.1355498721227621, "grad_norm": 5.689677983167152, "learning_rate": 3.943419647242189e-06, "loss": 0.3034, "step": 444 }, { "epoch": 1.1381074168797953, "grad_norm": 4.25975811296343, "learning_rate": 3.923794923749775e-06, "loss": 0.2187, "step": 445 }, { "epoch": 1.1406649616368287, "grad_norm": 4.876551747060212, "learning_rate": 3.904187569492373e-06, "loss": 0.257, "step": 446 }, { "epoch": 1.143222506393862, "grad_norm": 7.4635507464539765, "learning_rate": 3.884597900919656e-06, "loss": 0.2295, "step": 447 }, { "epoch": 1.145780051150895, "grad_norm": 5.715365590596455, "learning_rate": 3.865026234195863e-06, "loss": 0.2771, "step": 448 }, { "epoch": 1.1483375959079285, "grad_norm": 5.275576356708176, "learning_rate": 3.8454728851946885e-06, "loss": 0.1754, "step": 449 }, { "epoch": 1.1508951406649617, "grad_norm": 5.062656319020689, "learning_rate": 3.8259381694942e-06, "loss": 0.2382, "step": 450 }, { "epoch": 1.1534526854219949, "grad_norm": 5.798164920877668, "learning_rate": 3.806422402371733e-06, "loss": 0.288, "step": 451 }, { "epoch": 1.156010230179028, "grad_norm": 5.272021205396943, "learning_rate": 3.786925898798801e-06, "loss": 0.2654, "step": 452 }, { "epoch": 1.1585677749360614, "grad_norm": 4.739102069258541, "learning_rate": 3.767448973436021e-06, "loss": 0.1996, "step": 453 }, { "epoch": 1.1611253196930946, "grad_norm": 5.933924764643274, "learning_rate": 3.7479919406280334e-06, "loss": 0.3014, "step": 454 }, { "epoch": 1.1636828644501278, "grad_norm": 5.230108631129741, "learning_rate": 3.728555114398419e-06, "loss": 0.2306, "step": 455 }, { "epoch": 1.1662404092071612, "grad_norm": 5.1311748950507905, "learning_rate": 3.709138808444641e-06, "loss": 0.2426, "step": 456 }, { "epoch": 1.1687979539641944, "grad_norm": 4.8518342992156365, "learning_rate": 3.689743336132982e-06, "loss": 0.148, "step": 457 }, { "epoch": 1.1713554987212276, "grad_norm": 4.141242492696908, "learning_rate": 3.6703690104934806e-06, "loss": 0.1969, "step": 458 }, { "epoch": 1.1739130434782608, "grad_norm": 5.575504964227216, "learning_rate": 3.6510161442148783e-06, "loss": 0.2847, "step": 459 }, { "epoch": 1.1764705882352942, "grad_norm": 5.205832366429269, "learning_rate": 3.6316850496395863e-06, "loss": 0.3003, "step": 460 }, { "epoch": 1.1790281329923273, "grad_norm": 5.640735065432005, "learning_rate": 3.6123760387586265e-06, "loss": 0.2454, "step": 461 }, { "epoch": 1.1815856777493605, "grad_norm": 5.129282050106555, "learning_rate": 3.5930894232066072e-06, "loss": 0.1335, "step": 462 }, { "epoch": 1.184143222506394, "grad_norm": 5.640092025988574, "learning_rate": 3.5738255142566912e-06, "loss": 0.2181, "step": 463 }, { "epoch": 1.186700767263427, "grad_norm": 5.386833673551597, "learning_rate": 3.5545846228155743e-06, "loss": 0.2176, "step": 464 }, { "epoch": 1.1892583120204603, "grad_norm": 5.704105112383357, "learning_rate": 3.5353670594184623e-06, "loss": 0.2497, "step": 465 }, { "epoch": 1.1918158567774937, "grad_norm": 5.93251011789455, "learning_rate": 3.516173134224059e-06, "loss": 0.3096, "step": 466 }, { "epoch": 1.1943734015345269, "grad_norm": 5.135070353719214, "learning_rate": 3.4970031570095707e-06, "loss": 0.2201, "step": 467 }, { "epoch": 1.19693094629156, "grad_norm": 5.865866790423223, "learning_rate": 3.477857437165694e-06, "loss": 0.2694, "step": 468 }, { "epoch": 1.1994884910485935, "grad_norm": 6.268703077398429, "learning_rate": 3.458736283691626e-06, "loss": 0.2555, "step": 469 }, { "epoch": 1.2020460358056266, "grad_norm": 6.588443786412173, "learning_rate": 3.4396400051900846e-06, "loss": 0.2546, "step": 470 }, { "epoch": 1.2046035805626598, "grad_norm": 5.327149282563148, "learning_rate": 3.4205689098623195e-06, "loss": 0.1783, "step": 471 }, { "epoch": 1.207161125319693, "grad_norm": 4.958795590477734, "learning_rate": 3.401523305503139e-06, "loss": 0.1755, "step": 472 }, { "epoch": 1.2097186700767264, "grad_norm": 5.8814512868036735, "learning_rate": 3.3825034994959445e-06, "loss": 0.2751, "step": 473 }, { "epoch": 1.2122762148337596, "grad_norm": 6.346180998183863, "learning_rate": 3.3635097988077724e-06, "loss": 0.2825, "step": 474 }, { "epoch": 1.2148337595907928, "grad_norm": 4.824141112150008, "learning_rate": 3.3445425099843343e-06, "loss": 0.193, "step": 475 }, { "epoch": 1.2173913043478262, "grad_norm": 5.212763506921242, "learning_rate": 3.3256019391450696e-06, "loss": 0.2539, "step": 476 }, { "epoch": 1.2199488491048593, "grad_norm": 5.731843763443905, "learning_rate": 3.3066883919782116e-06, "loss": 0.2647, "step": 477 }, { "epoch": 1.2225063938618925, "grad_norm": 5.069916401820976, "learning_rate": 3.287802173735848e-06, "loss": 0.2466, "step": 478 }, { "epoch": 1.2250639386189257, "grad_norm": 5.351892499599417, "learning_rate": 3.268943589228992e-06, "loss": 0.2369, "step": 479 }, { "epoch": 1.227621483375959, "grad_norm": 5.226527301587174, "learning_rate": 3.250112942822673e-06, "loss": 0.2063, "step": 480 }, { "epoch": 1.2301790281329923, "grad_norm": 5.599367983952576, "learning_rate": 3.231310538431015e-06, "loss": 0.3108, "step": 481 }, { "epoch": 1.2327365728900257, "grad_norm": 4.949232012109662, "learning_rate": 3.212536679512332e-06, "loss": 0.2778, "step": 482 }, { "epoch": 1.2352941176470589, "grad_norm": 4.692407945943108, "learning_rate": 3.1937916690642356e-06, "loss": 0.2972, "step": 483 }, { "epoch": 1.237851662404092, "grad_norm": 4.946087662063809, "learning_rate": 3.1750758096187446e-06, "loss": 0.2155, "step": 484 }, { "epoch": 1.2404092071611252, "grad_norm": 4.8219004232287, "learning_rate": 3.1563894032373977e-06, "loss": 0.22, "step": 485 }, { "epoch": 1.2429667519181586, "grad_norm": 5.621400027096903, "learning_rate": 3.137732751506376e-06, "loss": 0.2286, "step": 486 }, { "epoch": 1.2455242966751918, "grad_norm": 5.479705903908969, "learning_rate": 3.1191061555316503e-06, "loss": 0.2534, "step": 487 }, { "epoch": 1.248081841432225, "grad_norm": 5.794553272942662, "learning_rate": 3.1005099159341044e-06, "loss": 0.2618, "step": 488 }, { "epoch": 1.2506393861892584, "grad_norm": 6.017425638082386, "learning_rate": 3.08194433284469e-06, "loss": 0.2843, "step": 489 }, { "epoch": 1.2531969309462916, "grad_norm": 4.5326844807976805, "learning_rate": 3.0634097058995877e-06, "loss": 0.1928, "step": 490 }, { "epoch": 1.2557544757033248, "grad_norm": 6.4463104104564115, "learning_rate": 3.0449063342353635e-06, "loss": 0.2478, "step": 491 }, { "epoch": 1.258312020460358, "grad_norm": 4.5887061422293565, "learning_rate": 3.0264345164841426e-06, "loss": 0.204, "step": 492 }, { "epoch": 1.2608695652173914, "grad_norm": 5.508160093203401, "learning_rate": 3.007994550768793e-06, "loss": 0.2159, "step": 493 }, { "epoch": 1.2634271099744245, "grad_norm": 4.737111003577314, "learning_rate": 2.989586734698113e-06, "loss": 0.2093, "step": 494 }, { "epoch": 1.265984654731458, "grad_norm": 4.947584431847608, "learning_rate": 2.971211365362028e-06, "loss": 0.1984, "step": 495 }, { "epoch": 1.2685421994884911, "grad_norm": 4.800447464091687, "learning_rate": 2.9528687393267865e-06, "loss": 0.2396, "step": 496 }, { "epoch": 1.2710997442455243, "grad_norm": 5.424287981614177, "learning_rate": 2.934559152630192e-06, "loss": 0.2752, "step": 497 }, { "epoch": 1.2736572890025575, "grad_norm": 4.885472933623559, "learning_rate": 2.9162829007768103e-06, "loss": 0.2778, "step": 498 }, { "epoch": 1.2762148337595907, "grad_norm": 5.051334757075423, "learning_rate": 2.898040278733203e-06, "loss": 0.2728, "step": 499 }, { "epoch": 1.278772378516624, "grad_norm": 5.103121044913984, "learning_rate": 2.879831580923176e-06, "loss": 0.2276, "step": 500 }, { "epoch": 1.278772378516624, "eval_loss": 0.4383206367492676, "eval_runtime": 0.9493, "eval_samples_per_second": 33.71, "eval_steps_per_second": 8.428, "step": 500 }, { "epoch": 1.2813299232736572, "grad_norm": 6.302964407184529, "learning_rate": 2.8616571012230134e-06, "loss": 0.324, "step": 501 }, { "epoch": 1.2838874680306906, "grad_norm": 4.570682847412226, "learning_rate": 2.843517132956742e-06, "loss": 0.2638, "step": 502 }, { "epoch": 1.2864450127877238, "grad_norm": 5.1557886110649065, "learning_rate": 2.8254119688914017e-06, "loss": 0.2901, "step": 503 }, { "epoch": 1.289002557544757, "grad_norm": 4.59205694943921, "learning_rate": 2.8073419012323154e-06, "loss": 0.246, "step": 504 }, { "epoch": 1.2915601023017902, "grad_norm": 5.220862830802117, "learning_rate": 2.789307221618369e-06, "loss": 0.2665, "step": 505 }, { "epoch": 1.2941176470588236, "grad_norm": 5.6445669568957895, "learning_rate": 2.771308221117309e-06, "loss": 0.2565, "step": 506 }, { "epoch": 1.2966751918158568, "grad_norm": 5.578642462958587, "learning_rate": 2.7533451902210512e-06, "loss": 0.2805, "step": 507 }, { "epoch": 1.29923273657289, "grad_norm": 5.9960066898200814, "learning_rate": 2.7354184188409773e-06, "loss": 0.3228, "step": 508 }, { "epoch": 1.3017902813299234, "grad_norm": 4.257760920803555, "learning_rate": 2.71752819630327e-06, "loss": 0.1833, "step": 509 }, { "epoch": 1.3043478260869565, "grad_norm": 6.1683068408323605, "learning_rate": 2.6996748113442397e-06, "loss": 0.185, "step": 510 }, { "epoch": 1.3069053708439897, "grad_norm": 5.042423224754131, "learning_rate": 2.6818585521056573e-06, "loss": 0.241, "step": 511 }, { "epoch": 1.309462915601023, "grad_norm": 4.727233078424547, "learning_rate": 2.66407970613011e-06, "loss": 0.1559, "step": 512 }, { "epoch": 1.3120204603580563, "grad_norm": 5.0241427635079505, "learning_rate": 2.646338560356363e-06, "loss": 0.1877, "step": 513 }, { "epoch": 1.3145780051150895, "grad_norm": 6.126061319891909, "learning_rate": 2.6286354011147252e-06, "loss": 0.2004, "step": 514 }, { "epoch": 1.317135549872123, "grad_norm": 5.24912663925928, "learning_rate": 2.6109705141224255e-06, "loss": 0.2364, "step": 515 }, { "epoch": 1.319693094629156, "grad_norm": 6.034124083029447, "learning_rate": 2.593344184479003e-06, "loss": 0.2788, "step": 516 }, { "epoch": 1.3222506393861893, "grad_norm": 6.608163697606231, "learning_rate": 2.575756696661713e-06, "loss": 0.2416, "step": 517 }, { "epoch": 1.3248081841432224, "grad_norm": 4.758894846786754, "learning_rate": 2.5582083345209217e-06, "loss": 0.2124, "step": 518 }, { "epoch": 1.3273657289002558, "grad_norm": 5.719504905312417, "learning_rate": 2.540699381275539e-06, "loss": 0.297, "step": 519 }, { "epoch": 1.329923273657289, "grad_norm": 5.86724061218003, "learning_rate": 2.5232301195084395e-06, "loss": 0.3234, "step": 520 }, { "epoch": 1.3324808184143222, "grad_norm": 5.7644650842566945, "learning_rate": 2.5058008311619035e-06, "loss": 0.2615, "step": 521 }, { "epoch": 1.3350383631713556, "grad_norm": 5.745064713338971, "learning_rate": 2.488411797533064e-06, "loss": 0.2118, "step": 522 }, { "epoch": 1.3375959079283888, "grad_norm": 4.576869292927031, "learning_rate": 2.4710632992693737e-06, "loss": 0.2278, "step": 523 }, { "epoch": 1.340153452685422, "grad_norm": 5.037272698300459, "learning_rate": 2.4537556163640726e-06, "loss": 0.236, "step": 524 }, { "epoch": 1.3427109974424551, "grad_norm": 5.421366472711992, "learning_rate": 2.4364890281516633e-06, "loss": 0.2577, "step": 525 }, { "epoch": 1.3452685421994885, "grad_norm": 4.866078798724115, "learning_rate": 2.4192638133034074e-06, "loss": 0.2318, "step": 526 }, { "epoch": 1.3478260869565217, "grad_norm": 4.7964708165794265, "learning_rate": 2.4020802498228333e-06, "loss": 0.2054, "step": 527 }, { "epoch": 1.350383631713555, "grad_norm": 4.928572310449122, "learning_rate": 2.384938615041238e-06, "loss": 0.2681, "step": 528 }, { "epoch": 1.3529411764705883, "grad_norm": 4.737745386088544, "learning_rate": 2.3678391856132203e-06, "loss": 0.23, "step": 529 }, { "epoch": 1.3554987212276215, "grad_norm": 4.546792180996154, "learning_rate": 2.350782237512215e-06, "loss": 0.2158, "step": 530 }, { "epoch": 1.3580562659846547, "grad_norm": 5.14316611376246, "learning_rate": 2.3337680460260314e-06, "loss": 0.196, "step": 531 }, { "epoch": 1.3606138107416879, "grad_norm": 5.834646238531157, "learning_rate": 2.316796885752415e-06, "loss": 0.2587, "step": 532 }, { "epoch": 1.3631713554987213, "grad_norm": 5.780876751105411, "learning_rate": 2.299869030594622e-06, "loss": 0.1942, "step": 533 }, { "epoch": 1.3657289002557544, "grad_norm": 4.548997507254711, "learning_rate": 2.2829847537569904e-06, "loss": 0.176, "step": 534 }, { "epoch": 1.3682864450127878, "grad_norm": 5.731302369481431, "learning_rate": 2.266144327740531e-06, "loss": 0.3019, "step": 535 }, { "epoch": 1.370843989769821, "grad_norm": 5.073679478403773, "learning_rate": 2.2493480243385298e-06, "loss": 0.2065, "step": 536 }, { "epoch": 1.3734015345268542, "grad_norm": 5.850299158046723, "learning_rate": 2.2325961146321683e-06, "loss": 0.2714, "step": 537 }, { "epoch": 1.3759590792838874, "grad_norm": 5.7360121373777035, "learning_rate": 2.2158888689861434e-06, "loss": 0.2249, "step": 538 }, { "epoch": 1.3785166240409208, "grad_norm": 4.828788116522103, "learning_rate": 2.1992265570442974e-06, "loss": 0.2304, "step": 539 }, { "epoch": 1.381074168797954, "grad_norm": 5.160394779683274, "learning_rate": 2.182609447725279e-06, "loss": 0.2259, "step": 540 }, { "epoch": 1.3836317135549872, "grad_norm": 5.290912122436337, "learning_rate": 2.1660378092181935e-06, "loss": 0.2104, "step": 541 }, { "epoch": 1.3861892583120206, "grad_norm": 5.983793408233051, "learning_rate": 2.149511908978275e-06, "loss": 0.2372, "step": 542 }, { "epoch": 1.3887468030690537, "grad_norm": 5.146327357469625, "learning_rate": 2.1330320137225773e-06, "loss": 0.2594, "step": 543 }, { "epoch": 1.391304347826087, "grad_norm": 6.1200616170554465, "learning_rate": 2.1165983894256647e-06, "loss": 0.2858, "step": 544 }, { "epoch": 1.39386189258312, "grad_norm": 5.18913029029246, "learning_rate": 2.100211301315315e-06, "loss": 0.1924, "step": 545 }, { "epoch": 1.3964194373401535, "grad_norm": 5.568160243169466, "learning_rate": 2.0838710138682412e-06, "loss": 0.193, "step": 546 }, { "epoch": 1.3989769820971867, "grad_norm": 4.998024067501082, "learning_rate": 2.0675777908058307e-06, "loss": 0.2753, "step": 547 }, { "epoch": 1.40153452685422, "grad_norm": 5.644184048080268, "learning_rate": 2.051331895089882e-06, "loss": 0.2865, "step": 548 }, { "epoch": 1.4040920716112533, "grad_norm": 5.40264585527916, "learning_rate": 2.035133588918356e-06, "loss": 0.183, "step": 549 }, { "epoch": 1.4066496163682864, "grad_norm": 4.477185903868405, "learning_rate": 2.0189831337211573e-06, "loss": 0.1957, "step": 550 }, { "epoch": 1.4092071611253196, "grad_norm": 5.199329102721851, "learning_rate": 2.0028807901559027e-06, "loss": 0.2067, "step": 551 }, { "epoch": 1.4117647058823528, "grad_norm": 4.040859440248337, "learning_rate": 1.9868268181037186e-06, "loss": 0.188, "step": 552 }, { "epoch": 1.4143222506393862, "grad_norm": 4.554963712348274, "learning_rate": 1.970821476665051e-06, "loss": 0.1872, "step": 553 }, { "epoch": 1.4168797953964194, "grad_norm": 5.430125296226688, "learning_rate": 1.9548650241554812e-06, "loss": 0.2154, "step": 554 }, { "epoch": 1.4194373401534528, "grad_norm": 5.328931412586536, "learning_rate": 1.9389577181015496e-06, "loss": 0.1755, "step": 555 }, { "epoch": 1.421994884910486, "grad_norm": 4.249854605239616, "learning_rate": 1.923099815236608e-06, "loss": 0.1593, "step": 556 }, { "epoch": 1.4245524296675192, "grad_norm": 3.907795738528054, "learning_rate": 1.9072915714966761e-06, "loss": 0.1251, "step": 557 }, { "epoch": 1.4271099744245523, "grad_norm": 5.529382370073762, "learning_rate": 1.8915332420163074e-06, "loss": 0.1817, "step": 558 }, { "epoch": 1.4296675191815857, "grad_norm": 5.035709320601147, "learning_rate": 1.8758250811244682e-06, "loss": 0.2079, "step": 559 }, { "epoch": 1.432225063938619, "grad_norm": 4.581843899498873, "learning_rate": 1.8601673423404449e-06, "loss": 0.2037, "step": 560 }, { "epoch": 1.434782608695652, "grad_norm": 6.5976247360019045, "learning_rate": 1.8445602783697375e-06, "loss": 0.2877, "step": 561 }, { "epoch": 1.4373401534526855, "grad_norm": 4.959100730308061, "learning_rate": 1.8290041410999893e-06, "loss": 0.1957, "step": 562 }, { "epoch": 1.4398976982097187, "grad_norm": 5.169599476739708, "learning_rate": 1.8134991815969238e-06, "loss": 0.2358, "step": 563 }, { "epoch": 1.4424552429667519, "grad_norm": 5.894324327311631, "learning_rate": 1.798045650100289e-06, "loss": 0.224, "step": 564 }, { "epoch": 1.445012787723785, "grad_norm": 5.802424880479901, "learning_rate": 1.782643796019814e-06, "loss": 0.2429, "step": 565 }, { "epoch": 1.4475703324808185, "grad_norm": 4.922016237101965, "learning_rate": 1.7672938679311957e-06, "loss": 0.2266, "step": 566 }, { "epoch": 1.4501278772378516, "grad_norm": 5.755816715889569, "learning_rate": 1.7519961135720737e-06, "loss": 0.2376, "step": 567 }, { "epoch": 1.452685421994885, "grad_norm": 5.220912348487513, "learning_rate": 1.736750779838044e-06, "loss": 0.2216, "step": 568 }, { "epoch": 1.4552429667519182, "grad_norm": 5.210521733834325, "learning_rate": 1.7215581127786624e-06, "loss": 0.2615, "step": 569 }, { "epoch": 1.4578005115089514, "grad_norm": 5.082754952247903, "learning_rate": 1.7064183575934856e-06, "loss": 0.2395, "step": 570 }, { "epoch": 1.4603580562659846, "grad_norm": 4.668118713570096, "learning_rate": 1.6913317586281048e-06, "loss": 0.1761, "step": 571 }, { "epoch": 1.4629156010230178, "grad_norm": 5.582283863667842, "learning_rate": 1.676298559370202e-06, "loss": 0.2342, "step": 572 }, { "epoch": 1.4654731457800512, "grad_norm": 5.88971284160264, "learning_rate": 1.6613190024456293e-06, "loss": 0.3086, "step": 573 }, { "epoch": 1.4680306905370843, "grad_norm": 5.053648614417428, "learning_rate": 1.6463933296144863e-06, "loss": 0.2169, "step": 574 }, { "epoch": 1.4705882352941178, "grad_norm": 5.3893999941494055, "learning_rate": 1.6315217817672142e-06, "loss": 0.2483, "step": 575 }, { "epoch": 1.473145780051151, "grad_norm": 6.003001211927369, "learning_rate": 1.6167045989207185e-06, "loss": 0.2488, "step": 576 }, { "epoch": 1.4757033248081841, "grad_norm": 5.8557134317975965, "learning_rate": 1.6019420202144853e-06, "loss": 0.2349, "step": 577 }, { "epoch": 1.4782608695652173, "grad_norm": 5.263808302802599, "learning_rate": 1.5872342839067305e-06, "loss": 0.1857, "step": 578 }, { "epoch": 1.4808184143222507, "grad_norm": 4.433242818261617, "learning_rate": 1.5725816273705453e-06, "loss": 0.1888, "step": 579 }, { "epoch": 1.4833759590792839, "grad_norm": 4.993905414062928, "learning_rate": 1.5579842870900746e-06, "loss": 0.2154, "step": 580 }, { "epoch": 1.485933503836317, "grad_norm": 4.381712629412114, "learning_rate": 1.5434424986566938e-06, "loss": 0.222, "step": 581 }, { "epoch": 1.4884910485933505, "grad_norm": 4.174610676856918, "learning_rate": 1.5289564967652033e-06, "loss": 0.1991, "step": 582 }, { "epoch": 1.4910485933503836, "grad_norm": 4.819672292661032, "learning_rate": 1.5145265152100574e-06, "loss": 0.2425, "step": 583 }, { "epoch": 1.4936061381074168, "grad_norm": 5.297609781727035, "learning_rate": 1.5001527868815702e-06, "loss": 0.3006, "step": 584 }, { "epoch": 1.49616368286445, "grad_norm": 5.79687615069299, "learning_rate": 1.4858355437621663e-06, "loss": 0.27, "step": 585 }, { "epoch": 1.4987212276214834, "grad_norm": 4.762605653912011, "learning_rate": 1.4715750169226417e-06, "loss": 0.2548, "step": 586 }, { "epoch": 1.5012787723785166, "grad_norm": 5.7198048630752725, "learning_rate": 1.457371436518424e-06, "loss": 0.2594, "step": 587 }, { "epoch": 1.50383631713555, "grad_norm": 4.802403477803473, "learning_rate": 1.4432250317858675e-06, "loss": 0.19, "step": 588 }, { "epoch": 1.5063938618925832, "grad_norm": 6.635420914435053, "learning_rate": 1.4291360310385455e-06, "loss": 0.2405, "step": 589 }, { "epoch": 1.5089514066496164, "grad_norm": 4.539889740139929, "learning_rate": 1.4151046616635727e-06, "loss": 0.1389, "step": 590 }, { "epoch": 1.5115089514066495, "grad_norm": 5.38811895384659, "learning_rate": 1.4011311501179287e-06, "loss": 0.2662, "step": 591 }, { "epoch": 1.5140664961636827, "grad_norm": 5.109717895825474, "learning_rate": 1.3872157219248045e-06, "loss": 0.2043, "step": 592 }, { "epoch": 1.5166240409207161, "grad_norm": 5.0959884186276625, "learning_rate": 1.373358601669973e-06, "loss": 0.1719, "step": 593 }, { "epoch": 1.5191815856777495, "grad_norm": 5.469942867350689, "learning_rate": 1.3595600129981469e-06, "loss": 0.1644, "step": 594 }, { "epoch": 1.5217391304347827, "grad_norm": 3.8395720981267885, "learning_rate": 1.3458201786093795e-06, "loss": 0.1706, "step": 595 }, { "epoch": 1.5242966751918159, "grad_norm": 5.518424328264906, "learning_rate": 1.3321393202554739e-06, "loss": 0.23, "step": 596 }, { "epoch": 1.526854219948849, "grad_norm": 6.325750802451315, "learning_rate": 1.3185176587363919e-06, "loss": 0.3093, "step": 597 }, { "epoch": 1.5294117647058822, "grad_norm": 5.306860678214338, "learning_rate": 1.3049554138967052e-06, "loss": 0.1669, "step": 598 }, { "epoch": 1.5319693094629157, "grad_norm": 5.531828559483773, "learning_rate": 1.2914528046220332e-06, "loss": 0.2521, "step": 599 }, { "epoch": 1.5345268542199488, "grad_norm": 4.928627869404515, "learning_rate": 1.278010048835523e-06, "loss": 0.1606, "step": 600 }, { "epoch": 1.5370843989769822, "grad_norm": 4.736176509824429, "learning_rate": 1.2646273634943195e-06, "loss": 0.2382, "step": 601 }, { "epoch": 1.5396419437340154, "grad_norm": 5.273813096512506, "learning_rate": 1.2513049645860759e-06, "loss": 0.1809, "step": 602 }, { "epoch": 1.5421994884910486, "grad_norm": 5.824236440729651, "learning_rate": 1.2380430671254618e-06, "loss": 0.2194, "step": 603 }, { "epoch": 1.5447570332480818, "grad_norm": 5.0892118395960395, "learning_rate": 1.224841885150691e-06, "loss": 0.2326, "step": 604 }, { "epoch": 1.547314578005115, "grad_norm": 4.776070837603081, "learning_rate": 1.2117016317200702e-06, "loss": 0.202, "step": 605 }, { "epoch": 1.5498721227621484, "grad_norm": 4.827682184305226, "learning_rate": 1.1986225189085627e-06, "loss": 0.2047, "step": 606 }, { "epoch": 1.5524296675191815, "grad_norm": 6.354947289868109, "learning_rate": 1.185604757804359e-06, "loss": 0.2489, "step": 607 }, { "epoch": 1.554987212276215, "grad_norm": 6.389932140184679, "learning_rate": 1.172648558505477e-06, "loss": 0.2039, "step": 608 }, { "epoch": 1.5575447570332481, "grad_norm": 5.4707664838796815, "learning_rate": 1.1597541301163655e-06, "loss": 0.2242, "step": 609 }, { "epoch": 1.5601023017902813, "grad_norm": 5.852672190013367, "learning_rate": 1.1469216807445348e-06, "loss": 0.1804, "step": 610 }, { "epoch": 1.5626598465473145, "grad_norm": 6.351885487364482, "learning_rate": 1.1341514174971907e-06, "loss": 0.2128, "step": 611 }, { "epoch": 1.5652173913043477, "grad_norm": 5.1770067311650605, "learning_rate": 1.1214435464779006e-06, "loss": 0.2993, "step": 612 }, { "epoch": 1.567774936061381, "grad_norm": 5.447984746676171, "learning_rate": 1.1087982727832613e-06, "loss": 0.307, "step": 613 }, { "epoch": 1.5703324808184145, "grad_norm": 5.182727825447981, "learning_rate": 1.0962158004995893e-06, "loss": 0.2687, "step": 614 }, { "epoch": 1.5728900255754477, "grad_norm": 4.64809564738005, "learning_rate": 1.083696332699628e-06, "loss": 0.1572, "step": 615 }, { "epoch": 1.5754475703324808, "grad_norm": 4.7325713746238876, "learning_rate": 1.0712400714392723e-06, "loss": 0.18, "step": 616 }, { "epoch": 1.578005115089514, "grad_norm": 5.269623806320352, "learning_rate": 1.058847217754303e-06, "loss": 0.2037, "step": 617 }, { "epoch": 1.5805626598465472, "grad_norm": 5.304757901803767, "learning_rate": 1.0465179716571467e-06, "loss": 0.1971, "step": 618 }, { "epoch": 1.5831202046035806, "grad_norm": 6.728413029180757, "learning_rate": 1.034252532133646e-06, "loss": 0.2494, "step": 619 }, { "epoch": 1.5856777493606138, "grad_norm": 3.4223539307887454, "learning_rate": 1.0220510971398473e-06, "loss": 0.1613, "step": 620 }, { "epoch": 1.5882352941176472, "grad_norm": 5.305411837383943, "learning_rate": 1.0099138635988026e-06, "loss": 0.2349, "step": 621 }, { "epoch": 1.5907928388746804, "grad_norm": 4.856914484373969, "learning_rate": 9.978410273974015e-07, "loss": 0.1723, "step": 622 }, { "epoch": 1.5933503836317136, "grad_norm": 5.358684756870008, "learning_rate": 9.858327833832004e-07, "loss": 0.2593, "step": 623 }, { "epoch": 1.5959079283887467, "grad_norm": 4.542731858430077, "learning_rate": 9.738893253612808e-07, "loss": 0.1939, "step": 624 }, { "epoch": 1.59846547314578, "grad_norm": 4.5148886714925345, "learning_rate": 9.620108460911181e-07, "loss": 0.1901, "step": 625 }, { "epoch": 1.6010230179028133, "grad_norm": 5.107745783747687, "learning_rate": 9.50197537283481e-07, "loss": 0.2062, "step": 626 }, { "epoch": 1.6035805626598465, "grad_norm": 6.591147405424452, "learning_rate": 9.384495895973227e-07, "loss": 0.2293, "step": 627 }, { "epoch": 1.60613810741688, "grad_norm": 5.607830445145212, "learning_rate": 9.267671926367166e-07, "loss": 0.2449, "step": 628 }, { "epoch": 1.608695652173913, "grad_norm": 5.457362074950791, "learning_rate": 9.151505349477901e-07, "loss": 0.2638, "step": 629 }, { "epoch": 1.6112531969309463, "grad_norm": 4.821051497667833, "learning_rate": 9.035998040156801e-07, "loss": 0.1902, "step": 630 }, { "epoch": 1.6138107416879794, "grad_norm": 5.171122943753494, "learning_rate": 8.921151862615091e-07, "loss": 0.1422, "step": 631 }, { "epoch": 1.6163682864450126, "grad_norm": 6.259146518225325, "learning_rate": 8.806968670393801e-07, "loss": 0.315, "step": 632 }, { "epoch": 1.618925831202046, "grad_norm": 5.397528592846895, "learning_rate": 8.693450306333818e-07, "loss": 0.21, "step": 633 }, { "epoch": 1.6214833759590794, "grad_norm": 5.180249332098257, "learning_rate": 8.580598602546109e-07, "loss": 0.2556, "step": 634 }, { "epoch": 1.6240409207161126, "grad_norm": 4.562634677645485, "learning_rate": 8.4684153803822e-07, "loss": 0.2216, "step": 635 }, { "epoch": 1.6265984654731458, "grad_norm": 5.390722528674729, "learning_rate": 8.356902450404792e-07, "loss": 0.233, "step": 636 }, { "epoch": 1.629156010230179, "grad_norm": 4.775191692612564, "learning_rate": 8.246061612358475e-07, "loss": 0.2287, "step": 637 }, { "epoch": 1.6317135549872122, "grad_norm": 4.852828279434425, "learning_rate": 8.135894655140758e-07, "loss": 0.2191, "step": 638 }, { "epoch": 1.6342710997442456, "grad_norm": 5.704310246599364, "learning_rate": 8.026403356773161e-07, "loss": 0.2047, "step": 639 }, { "epoch": 1.6368286445012787, "grad_norm": 4.793807479675851, "learning_rate": 7.91758948437249e-07, "loss": 0.1618, "step": 640 }, { "epoch": 1.6393861892583121, "grad_norm": 4.754502329271982, "learning_rate": 7.809454794122346e-07, "loss": 0.2781, "step": 641 }, { "epoch": 1.6419437340153453, "grad_norm": 7.018986504016533, "learning_rate": 7.702001031244816e-07, "loss": 0.2729, "step": 642 }, { "epoch": 1.6445012787723785, "grad_norm": 4.41916704046766, "learning_rate": 7.595229929972253e-07, "loss": 0.2092, "step": 643 }, { "epoch": 1.6470588235294117, "grad_norm": 5.336932050022912, "learning_rate": 7.489143213519301e-07, "loss": 0.2267, "step": 644 }, { "epoch": 1.6496163682864449, "grad_norm": 4.654218849496965, "learning_rate": 7.383742594055077e-07, "loss": 0.2136, "step": 645 }, { "epoch": 1.6521739130434783, "grad_norm": 5.1623224835718515, "learning_rate": 7.279029772675572e-07, "loss": 0.221, "step": 646 }, { "epoch": 1.6547314578005117, "grad_norm": 5.686897560893601, "learning_rate": 7.17500643937617e-07, "loss": 0.243, "step": 647 }, { "epoch": 1.6572890025575449, "grad_norm": 4.738802336926236, "learning_rate": 7.071674273024353e-07, "loss": 0.2059, "step": 648 }, { "epoch": 1.659846547314578, "grad_norm": 6.305944976355385, "learning_rate": 6.969034941332664e-07, "loss": 0.2147, "step": 649 }, { "epoch": 1.6624040920716112, "grad_norm": 5.190380929928794, "learning_rate": 6.86709010083172e-07, "loss": 0.1909, "step": 650 }, { "epoch": 1.6649616368286444, "grad_norm": 5.659990773064172, "learning_rate": 6.765841396843514e-07, "loss": 0.1913, "step": 651 }, { "epoch": 1.6675191815856778, "grad_norm": 4.841853616494702, "learning_rate": 6.665290463454882e-07, "loss": 0.201, "step": 652 }, { "epoch": 1.670076726342711, "grad_norm": 6.0589020041416575, "learning_rate": 6.565438923491102e-07, "loss": 0.2908, "step": 653 }, { "epoch": 1.6726342710997444, "grad_norm": 4.716079545501844, "learning_rate": 6.466288388489689e-07, "loss": 0.2169, "step": 654 }, { "epoch": 1.6751918158567776, "grad_norm": 4.268196527173102, "learning_rate": 6.367840458674401e-07, "loss": 0.202, "step": 655 }, { "epoch": 1.6777493606138107, "grad_norm": 5.8944147568821625, "learning_rate": 6.270096722929442e-07, "loss": 0.2586, "step": 656 }, { "epoch": 1.680306905370844, "grad_norm": 5.564876452450793, "learning_rate": 6.173058758773775e-07, "loss": 0.2391, "step": 657 }, { "epoch": 1.682864450127877, "grad_norm": 5.003498689191179, "learning_rate": 6.076728132335669e-07, "loss": 0.1756, "step": 658 }, { "epoch": 1.6854219948849105, "grad_norm": 4.993122576064924, "learning_rate": 5.981106398327463e-07, "loss": 0.1984, "step": 659 }, { "epoch": 1.6879795396419437, "grad_norm": 5.118160492207756, "learning_rate": 5.886195100020408e-07, "loss": 0.2879, "step": 660 }, { "epoch": 1.690537084398977, "grad_norm": 5.396969385626712, "learning_rate": 5.7919957692198e-07, "loss": 0.1767, "step": 661 }, { "epoch": 1.6930946291560103, "grad_norm": 4.709841482974692, "learning_rate": 5.698509926240275e-07, "loss": 0.2265, "step": 662 }, { "epoch": 1.6956521739130435, "grad_norm": 5.74378925311425, "learning_rate": 5.60573907988124e-07, "loss": 0.2585, "step": 663 }, { "epoch": 1.6982097186700766, "grad_norm": 5.340843111211026, "learning_rate": 5.513684727402529e-07, "loss": 0.1713, "step": 664 }, { "epoch": 1.7007672634271098, "grad_norm": 5.095292121289202, "learning_rate": 5.422348354500217e-07, "loss": 0.1969, "step": 665 }, { "epoch": 1.7033248081841432, "grad_norm": 6.38491199555891, "learning_rate": 5.331731435282705e-07, "loss": 0.1764, "step": 666 }, { "epoch": 1.7058823529411766, "grad_norm": 6.33773326776292, "learning_rate": 5.241835432246888e-07, "loss": 0.2176, "step": 667 }, { "epoch": 1.7084398976982098, "grad_norm": 4.747290198021267, "learning_rate": 5.152661796254505e-07, "loss": 0.2194, "step": 668 }, { "epoch": 1.710997442455243, "grad_norm": 5.180305167036278, "learning_rate": 5.064211966508837e-07, "loss": 0.1838, "step": 669 }, { "epoch": 1.7135549872122762, "grad_norm": 5.158569886057061, "learning_rate": 4.976487370531352e-07, "loss": 0.1714, "step": 670 }, { "epoch": 1.7161125319693094, "grad_norm": 6.7699503709640405, "learning_rate": 4.88948942413876e-07, "loss": 0.2794, "step": 671 }, { "epoch": 1.7186700767263428, "grad_norm": 4.330269645663795, "learning_rate": 4.803219531420128e-07, "loss": 0.1567, "step": 672 }, { "epoch": 1.721227621483376, "grad_norm": 6.071163454166977, "learning_rate": 4.717679084714222e-07, "loss": 0.2268, "step": 673 }, { "epoch": 1.7237851662404093, "grad_norm": 4.455492690019572, "learning_rate": 4.6328694645870254e-07, "loss": 0.2092, "step": 674 }, { "epoch": 1.7263427109974425, "grad_norm": 5.5297942207932165, "learning_rate": 4.5487920398094465e-07, "loss": 0.2037, "step": 675 }, { "epoch": 1.7289002557544757, "grad_norm": 4.8818412705137115, "learning_rate": 4.46544816733529e-07, "loss": 0.2475, "step": 676 }, { "epoch": 1.7314578005115089, "grad_norm": 5.627569483357689, "learning_rate": 4.382839192279303e-07, "loss": 0.2585, "step": 677 }, { "epoch": 1.734015345268542, "grad_norm": 5.1192641037653965, "learning_rate": 4.3009664478954384e-07, "loss": 0.2449, "step": 678 }, { "epoch": 1.7365728900255755, "grad_norm": 5.286870696275029, "learning_rate": 4.219831255555423e-07, "loss": 0.2004, "step": 679 }, { "epoch": 1.7391304347826086, "grad_norm": 5.5327277136832675, "learning_rate": 4.139434924727359e-07, "loss": 0.1921, "step": 680 }, { "epoch": 1.741687979539642, "grad_norm": 4.552608688952828, "learning_rate": 4.059778752954607e-07, "loss": 0.1432, "step": 681 }, { "epoch": 1.7442455242966752, "grad_norm": 5.216152442361206, "learning_rate": 3.9808640258348686e-07, "loss": 0.1754, "step": 682 }, { "epoch": 1.7468030690537084, "grad_norm": 5.756818223369158, "learning_rate": 3.9026920169994374e-07, "loss": 0.21, "step": 683 }, { "epoch": 1.7493606138107416, "grad_norm": 5.29491810051506, "learning_rate": 3.825263988092587e-07, "loss": 0.2228, "step": 684 }, { "epoch": 1.7519181585677748, "grad_norm": 5.164127220528273, "learning_rate": 3.7485811887512714e-07, "loss": 0.2144, "step": 685 }, { "epoch": 1.7544757033248082, "grad_norm": 5.314203439099788, "learning_rate": 3.672644856584928e-07, "loss": 0.2911, "step": 686 }, { "epoch": 1.7570332480818416, "grad_norm": 4.972659541709094, "learning_rate": 3.597456217155526e-07, "loss": 0.1519, "step": 687 }, { "epoch": 1.7595907928388748, "grad_norm": 5.199722512215739, "learning_rate": 3.523016483957742e-07, "loss": 0.2739, "step": 688 }, { "epoch": 1.762148337595908, "grad_norm": 5.21999031799005, "learning_rate": 3.4493268583994434e-07, "loss": 0.201, "step": 689 }, { "epoch": 1.7647058823529411, "grad_norm": 4.364821712930802, "learning_rate": 3.3763885297822153e-07, "loss": 0.1881, "step": 690 }, { "epoch": 1.7672634271099743, "grad_norm": 5.202401016558692, "learning_rate": 3.3042026752822254e-07, "loss": 0.2125, "step": 691 }, { "epoch": 1.7698209718670077, "grad_norm": 5.684342025889591, "learning_rate": 3.2327704599312283e-07, "loss": 0.2636, "step": 692 }, { "epoch": 1.772378516624041, "grad_norm": 5.59338748722469, "learning_rate": 3.16209303659773e-07, "loss": 0.2489, "step": 693 }, { "epoch": 1.7749360613810743, "grad_norm": 4.23116707170702, "learning_rate": 3.0921715459683753e-07, "loss": 0.153, "step": 694 }, { "epoch": 1.7774936061381075, "grad_norm": 5.558061912478462, "learning_rate": 3.0230071165295804e-07, "loss": 0.2291, "step": 695 }, { "epoch": 1.7800511508951407, "grad_norm": 5.197951332859319, "learning_rate": 2.95460086454929e-07, "loss": 0.2267, "step": 696 }, { "epoch": 1.7826086956521738, "grad_norm": 5.668569655261014, "learning_rate": 2.88695389405898e-07, "loss": 0.2191, "step": 697 }, { "epoch": 1.785166240409207, "grad_norm": 4.00218617616878, "learning_rate": 2.820067296835799e-07, "loss": 0.1415, "step": 698 }, { "epoch": 1.7877237851662404, "grad_norm": 5.580094964915473, "learning_rate": 2.753942152385014e-07, "loss": 0.2282, "step": 699 }, { "epoch": 1.7902813299232738, "grad_norm": 5.157143860228644, "learning_rate": 2.688579527922514e-07, "loss": 0.228, "step": 700 }, { "epoch": 1.792838874680307, "grad_norm": 6.375062738221611, "learning_rate": 2.6239804783576294e-07, "loss": 0.227, "step": 701 }, { "epoch": 1.7953964194373402, "grad_norm": 5.7437796904243115, "learning_rate": 2.560146046276135e-07, "loss": 0.2823, "step": 702 }, { "epoch": 1.7979539641943734, "grad_norm": 5.624690347371044, "learning_rate": 2.4970772619233475e-07, "loss": 0.2042, "step": 703 }, { "epoch": 1.8005115089514065, "grad_norm": 6.498262506148002, "learning_rate": 2.4347751431875453e-07, "loss": 0.2598, "step": 704 }, { "epoch": 1.80306905370844, "grad_norm": 6.186981848018963, "learning_rate": 2.373240695583534e-07, "loss": 0.2505, "step": 705 }, { "epoch": 1.8056265984654731, "grad_norm": 4.425992548953008, "learning_rate": 2.3124749122364286e-07, "loss": 0.2145, "step": 706 }, { "epoch": 1.8081841432225065, "grad_norm": 4.690718811893992, "learning_rate": 2.2524787738656073e-07, "loss": 0.2106, "step": 707 }, { "epoch": 1.8107416879795397, "grad_norm": 6.156573679807675, "learning_rate": 2.1932532487688784e-07, "loss": 0.222, "step": 708 }, { "epoch": 1.813299232736573, "grad_norm": 5.586951769770138, "learning_rate": 2.1347992928068884e-07, "loss": 0.2469, "step": 709 }, { "epoch": 1.815856777493606, "grad_norm": 5.817151417919622, "learning_rate": 2.0771178493876387e-07, "loss": 0.2514, "step": 710 }, { "epoch": 1.8184143222506393, "grad_norm": 5.920190459702125, "learning_rate": 2.0202098494513157e-07, "loss": 0.2034, "step": 711 }, { "epoch": 1.8209718670076727, "grad_norm": 5.8230410453933015, "learning_rate": 1.964076211455246e-07, "loss": 0.2364, "step": 712 }, { "epoch": 1.8235294117647058, "grad_norm": 5.7392724174161405, "learning_rate": 1.908717841359048e-07, "loss": 0.2135, "step": 713 }, { "epoch": 1.8260869565217392, "grad_norm": 5.775722433349359, "learning_rate": 1.8541356326100436e-07, "loss": 0.2462, "step": 714 }, { "epoch": 1.8286445012787724, "grad_norm": 5.591670023546011, "learning_rate": 1.800330466128808e-07, "loss": 0.228, "step": 715 }, { "epoch": 1.8312020460358056, "grad_norm": 4.711360255327199, "learning_rate": 1.7473032102949983e-07, "loss": 0.2075, "step": 716 }, { "epoch": 1.8337595907928388, "grad_norm": 5.929377308741077, "learning_rate": 1.695054720933309e-07, "loss": 0.257, "step": 717 }, { "epoch": 1.836317135549872, "grad_norm": 5.279542039345601, "learning_rate": 1.6435858412996275e-07, "loss": 0.1529, "step": 718 }, { "epoch": 1.8388746803069054, "grad_norm": 5.1397676431266595, "learning_rate": 1.5928974020674947e-07, "loss": 0.1491, "step": 719 }, { "epoch": 1.8414322250639388, "grad_norm": 5.512326623165319, "learning_rate": 1.542990221314644e-07, "loss": 0.1774, "step": 720 }, { "epoch": 1.843989769820972, "grad_norm": 5.66627970312488, "learning_rate": 1.4938651045098174e-07, "loss": 0.2278, "step": 721 }, { "epoch": 1.8465473145780051, "grad_norm": 4.424031839884274, "learning_rate": 1.445522844499775e-07, "loss": 0.1713, "step": 722 }, { "epoch": 1.8491048593350383, "grad_norm": 6.625716850892438, "learning_rate": 1.3979642214964728e-07, "loss": 0.254, "step": 723 }, { "epoch": 1.8516624040920715, "grad_norm": 5.181446281414914, "learning_rate": 1.3511900030644954e-07, "loss": 0.1753, "step": 724 }, { "epoch": 1.854219948849105, "grad_norm": 6.694348870428919, "learning_rate": 1.3052009441086533e-07, "loss": 0.2535, "step": 725 }, { "epoch": 1.856777493606138, "grad_norm": 6.09347774518519, "learning_rate": 1.2599977868618052e-07, "loss": 0.2617, "step": 726 }, { "epoch": 1.8593350383631715, "grad_norm": 6.126918278593983, "learning_rate": 1.215581260872889e-07, "loss": 0.1734, "step": 727 }, { "epoch": 1.8618925831202047, "grad_norm": 4.9394061667025175, "learning_rate": 1.1719520829951203e-07, "loss": 0.1727, "step": 728 }, { "epoch": 1.8644501278772379, "grad_norm": 5.679604325861919, "learning_rate": 1.1291109573744574e-07, "loss": 0.2195, "step": 729 }, { "epoch": 1.867007672634271, "grad_norm": 5.5914554870990525, "learning_rate": 1.087058575438199e-07, "loss": 0.2528, "step": 730 }, { "epoch": 1.8695652173913042, "grad_norm": 5.877663743949046, "learning_rate": 1.0457956158838545e-07, "loss": 0.2491, "step": 731 }, { "epoch": 1.8721227621483376, "grad_norm": 3.8851223721617663, "learning_rate": 1.0053227446681912e-07, "loss": 0.1317, "step": 732 }, { "epoch": 1.8746803069053708, "grad_norm": 5.412598503950858, "learning_rate": 9.656406149964548e-08, "loss": 0.2314, "step": 733 }, { "epoch": 1.8772378516624042, "grad_norm": 4.937142526302332, "learning_rate": 9.267498673118547e-08, "loss": 0.1964, "step": 734 }, { "epoch": 1.8797953964194374, "grad_norm": 5.6093925489434735, "learning_rate": 8.886511292852395e-08, "loss": 0.2169, "step": 735 }, { "epoch": 1.8823529411764706, "grad_norm": 5.544589584646933, "learning_rate": 8.513450158049109e-08, "loss": 0.2262, "step": 736 }, { "epoch": 1.8849104859335037, "grad_norm": 4.522441136690872, "learning_rate": 8.148321289667749e-08, "loss": 0.1836, "step": 737 }, { "epoch": 1.887468030690537, "grad_norm": 4.735747110712485, "learning_rate": 7.791130580645623e-08, "loss": 0.2037, "step": 738 }, { "epoch": 1.8900255754475703, "grad_norm": 5.424105506041224, "learning_rate": 7.441883795803462e-08, "loss": 0.181, "step": 739 }, { "epoch": 1.8925831202046037, "grad_norm": 5.57291250707421, "learning_rate": 7.100586571752444e-08, "loss": 0.2009, "step": 740 }, { "epoch": 1.895140664961637, "grad_norm": 4.723852414529119, "learning_rate": 6.767244416802988e-08, "loss": 0.2135, "step": 741 }, { "epoch": 1.89769820971867, "grad_norm": 4.625168893146648, "learning_rate": 6.441862710876102e-08, "loss": 0.182, "step": 742 }, { "epoch": 1.9002557544757033, "grad_norm": 5.742248230826967, "learning_rate": 6.124446705416343e-08, "loss": 0.2256, "step": 743 }, { "epoch": 1.9028132992327365, "grad_norm": 5.446320051746138, "learning_rate": 5.815001523307162e-08, "loss": 0.1844, "step": 744 }, { "epoch": 1.9053708439897699, "grad_norm": 5.776856594363444, "learning_rate": 5.513532158788193e-08, "loss": 0.2793, "step": 745 }, { "epoch": 1.907928388746803, "grad_norm": 6.11898381729892, "learning_rate": 5.220043477374759e-08, "loss": 0.228, "step": 746 }, { "epoch": 1.9104859335038364, "grad_norm": 5.485214100078597, "learning_rate": 4.934540215779271e-08, "loss": 0.2315, "step": 747 }, { "epoch": 1.9130434782608696, "grad_norm": 5.1671494114407155, "learning_rate": 4.657026981834623e-08, "loss": 0.2191, "step": 748 }, { "epoch": 1.9156010230179028, "grad_norm": 5.544282249484376, "learning_rate": 4.3875082544201364e-08, "loss": 0.2446, "step": 749 }, { "epoch": 1.918158567774936, "grad_norm": 4.828585643362734, "learning_rate": 4.125988383388957e-08, "loss": 0.2293, "step": 750 }, { "epoch": 1.9207161125319692, "grad_norm": 6.32383213055493, "learning_rate": 3.87247158949805e-08, "loss": 0.2332, "step": 751 }, { "epoch": 1.9232736572890026, "grad_norm": 5.560416321895529, "learning_rate": 3.626961964340203e-08, "loss": 0.2383, "step": 752 }, { "epoch": 1.9258312020460358, "grad_norm": 5.865352141528642, "learning_rate": 3.389463470277576e-08, "loss": 0.2104, "step": 753 }, { "epoch": 1.9283887468030692, "grad_norm": 5.341153765939635, "learning_rate": 3.159979940378088e-08, "loss": 0.2128, "step": 754 }, { "epoch": 1.9309462915601023, "grad_norm": 5.291758252142984, "learning_rate": 2.938515078353521e-08, "loss": 0.1886, "step": 755 }, { "epoch": 1.9335038363171355, "grad_norm": 5.163447364959534, "learning_rate": 2.725072458499567e-08, "loss": 0.2996, "step": 756 }, { "epoch": 1.9360613810741687, "grad_norm": 6.676816256652347, "learning_rate": 2.519655525638376e-08, "loss": 0.2902, "step": 757 }, { "epoch": 1.938618925831202, "grad_norm": 6.29139443707462, "learning_rate": 2.3222675950627106e-08, "loss": 0.1572, "step": 758 }, { "epoch": 1.9411764705882353, "grad_norm": 4.037989400305309, "learning_rate": 2.1329118524827662e-08, "loss": 0.1417, "step": 759 }, { "epoch": 1.9437340153452687, "grad_norm": 4.5423783674836695, "learning_rate": 1.9515913539743247e-08, "loss": 0.233, "step": 760 }, { "epoch": 1.9462915601023019, "grad_norm": 3.859294316006792, "learning_rate": 1.7783090259297918e-08, "loss": 0.1918, "step": 761 }, { "epoch": 1.948849104859335, "grad_norm": 5.265098117553532, "learning_rate": 1.613067665010959e-08, "loss": 0.1965, "step": 762 }, { "epoch": 1.9514066496163682, "grad_norm": 4.559779542352908, "learning_rate": 1.4558699381034825e-08, "loss": 0.1888, "step": 763 }, { "epoch": 1.9539641943734014, "grad_norm": 6.603926656306259, "learning_rate": 1.3067183822742525e-08, "loss": 0.2729, "step": 764 }, { "epoch": 1.9565217391304348, "grad_norm": 5.966454788112439, "learning_rate": 1.1656154047303691e-08, "loss": 0.1994, "step": 765 }, { "epoch": 1.959079283887468, "grad_norm": 5.135081772282031, "learning_rate": 1.0325632827801745e-08, "loss": 0.1875, "step": 766 }, { "epoch": 1.9616368286445014, "grad_norm": 4.6463930984383275, "learning_rate": 9.075641637964483e-09, "loss": 0.213, "step": 767 }, { "epoch": 1.9641943734015346, "grad_norm": 5.19696586284704, "learning_rate": 7.906200651819907e-09, "loss": 0.188, "step": 768 }, { "epoch": 1.9667519181585678, "grad_norm": 4.1586996620802585, "learning_rate": 6.817328743368712e-09, "loss": 0.1314, "step": 769 }, { "epoch": 1.969309462915601, "grad_norm": 6.428796482081383, "learning_rate": 5.809043486279531e-09, "loss": 0.265, "step": 770 }, { "epoch": 1.9718670076726341, "grad_norm": 6.071118019652648, "learning_rate": 4.881361153606934e-09, "loss": 0.1938, "step": 771 }, { "epoch": 1.9744245524296675, "grad_norm": 6.004256284137117, "learning_rate": 4.034296717527752e-09, "loss": 0.258, "step": 772 }, { "epoch": 1.976982097186701, "grad_norm": 4.424170682035935, "learning_rate": 3.2678638490996064e-09, "loss": 0.1944, "step": 773 }, { "epoch": 1.979539641943734, "grad_norm": 4.5443816733282025, "learning_rate": 2.5820749180388573e-09, "loss": 0.225, "step": 774 }, { "epoch": 1.9820971867007673, "grad_norm": 6.813889752181503, "learning_rate": 1.976940992523546e-09, "loss": 0.242, "step": 775 }, { "epoch": 1.9846547314578005, "grad_norm": 6.943362317683093, "learning_rate": 1.4524718390140913e-09, "loss": 0.3044, "step": 776 }, { "epoch": 1.9872122762148337, "grad_norm": 5.440758095117894, "learning_rate": 1.0086759220934162e-09, "loss": 0.2411, "step": 777 }, { "epoch": 1.989769820971867, "grad_norm": 3.8628110410030914, "learning_rate": 6.455604043331676e-10, "loss": 0.1659, "step": 778 }, { "epoch": 1.9923273657289002, "grad_norm": 5.170928482570995, "learning_rate": 3.631311461765874e-10, "loss": 0.1393, "step": 779 }, { "epoch": 1.9948849104859336, "grad_norm": 5.769015016748028, "learning_rate": 1.6139270584358823e-10, "loss": 0.178, "step": 780 }, { "epoch": 1.9974424552429668, "grad_norm": 5.070011448256778, "learning_rate": 4.034833925969928e-11, "loss": 0.2705, "step": 781 }, { "epoch": 2.0, "grad_norm": 6.222117398419603, "learning_rate": 0.0, "loss": 0.1879, "step": 782 }, { "epoch": 2.0, "step": 782, "total_flos": 850857689088.0, "train_loss": 0.39088617330012115, "train_runtime": 416.9902, "train_samples_per_second": 15.003, "train_steps_per_second": 1.875 } ], "logging_steps": 1, "max_steps": 782, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 70000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 850857689088.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }