{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9907877816893985, "eval_steps": 500, "global_step": 23500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 4.2161182199548876e-05, "grad_norm": 13.292810440063477, "learning_rate": 0.0, "loss": 2.3944, "step": 1 }, { "epoch": 8.432236439909775e-05, "grad_norm": 9.827836990356445, "learning_rate": 7.026419336706015e-09, "loss": 2.3063, "step": 2 }, { "epoch": 0.00012648354659864662, "grad_norm": 8.07567310333252, "learning_rate": 1.405283867341203e-08, "loss": 2.1541, "step": 3 }, { "epoch": 0.0001686447287981955, "grad_norm": 15.019342422485352, "learning_rate": 2.1079258010118043e-08, "loss": 2.3662, "step": 4 }, { "epoch": 0.0002108059109977444, "grad_norm": 12.74185562133789, "learning_rate": 2.810567734682406e-08, "loss": 2.3179, "step": 5 }, { "epoch": 0.00025296709319729325, "grad_norm": 5.6531081199646, "learning_rate": 3.5132096683530076e-08, "loss": 1.4374, "step": 6 }, { "epoch": 0.00029512827539684213, "grad_norm": 11.508322715759277, "learning_rate": 4.2158516020236085e-08, "loss": 2.2826, "step": 7 }, { "epoch": 0.000337289457596391, "grad_norm": 13.205543518066406, "learning_rate": 4.918493535694211e-08, "loss": 2.5462, "step": 8 }, { "epoch": 0.0003794506397959399, "grad_norm": 14.691112518310547, "learning_rate": 5.621135469364812e-08, "loss": 2.725, "step": 9 }, { "epoch": 0.0004216118219954888, "grad_norm": 9.999246597290039, "learning_rate": 6.323777403035413e-08, "loss": 2.1269, "step": 10 }, { "epoch": 0.0004637730041950376, "grad_norm": 10.9578857421875, "learning_rate": 7.026419336706015e-08, "loss": 2.1552, "step": 11 }, { "epoch": 0.0005059341863945865, "grad_norm": 8.872416496276855, "learning_rate": 7.729061270376616e-08, "loss": 2.0987, "step": 12 }, { "epoch": 0.0005480953685941354, "grad_norm": 11.674811363220215, "learning_rate": 8.431703204047217e-08, "loss": 2.1093, "step": 13 }, { "epoch": 0.0005902565507936843, "grad_norm": 14.053629875183105, "learning_rate": 9.13434513771782e-08, "loss": 2.5702, "step": 14 }, { "epoch": 0.0006324177329932331, "grad_norm": 12.205204010009766, "learning_rate": 9.836987071388422e-08, "loss": 2.2232, "step": 15 }, { "epoch": 0.000674578915192782, "grad_norm": 12.968454360961914, "learning_rate": 1.0539629005059021e-07, "loss": 2.607, "step": 16 }, { "epoch": 0.0007167400973923309, "grad_norm": 13.988900184631348, "learning_rate": 1.1242270938729624e-07, "loss": 2.4074, "step": 17 }, { "epoch": 0.0007589012795918798, "grad_norm": 17.978530883789062, "learning_rate": 1.1944912872400227e-07, "loss": 2.7069, "step": 18 }, { "epoch": 0.0008010624617914286, "grad_norm": 8.27976131439209, "learning_rate": 1.2647554806070826e-07, "loss": 1.7074, "step": 19 }, { "epoch": 0.0008432236439909776, "grad_norm": 11.803068161010742, "learning_rate": 1.335019673974143e-07, "loss": 2.059, "step": 20 }, { "epoch": 0.0008853848261905264, "grad_norm": 6.375721454620361, "learning_rate": 1.405283867341203e-07, "loss": 1.4733, "step": 21 }, { "epoch": 0.0009275460083900752, "grad_norm": 15.85426139831543, "learning_rate": 1.475548060708263e-07, "loss": 3.0087, "step": 22 }, { "epoch": 0.0009697071905896242, "grad_norm": 9.469091415405273, "learning_rate": 1.5458122540753232e-07, "loss": 2.2263, "step": 23 }, { "epoch": 0.001011868372789173, "grad_norm": 7.771282196044922, "learning_rate": 1.6160764474423836e-07, "loss": 1.5916, "step": 24 }, { "epoch": 0.0010540295549887218, "grad_norm": 11.653564453125, "learning_rate": 1.6863406408094434e-07, "loss": 2.3494, "step": 25 }, { "epoch": 0.0010961907371882709, "grad_norm": 7.823153972625732, "learning_rate": 1.7566048341765038e-07, "loss": 1.8818, "step": 26 }, { "epoch": 0.0011383519193878197, "grad_norm": 7.253396987915039, "learning_rate": 1.826869027543564e-07, "loss": 1.6429, "step": 27 }, { "epoch": 0.0011805131015873685, "grad_norm": 11.600074768066406, "learning_rate": 1.897133220910624e-07, "loss": 2.2976, "step": 28 }, { "epoch": 0.0012226742837869173, "grad_norm": 13.965734481811523, "learning_rate": 1.9673974142776843e-07, "loss": 2.9063, "step": 29 }, { "epoch": 0.0012648354659864662, "grad_norm": 13.23375129699707, "learning_rate": 2.0376616076447442e-07, "loss": 2.3106, "step": 30 }, { "epoch": 0.0013069966481860152, "grad_norm": 12.43603801727295, "learning_rate": 2.1079258010118043e-07, "loss": 2.4484, "step": 31 }, { "epoch": 0.001349157830385564, "grad_norm": 13.698935508728027, "learning_rate": 2.1781899943788646e-07, "loss": 2.4134, "step": 32 }, { "epoch": 0.0013913190125851129, "grad_norm": 5.343920707702637, "learning_rate": 2.2484541877459247e-07, "loss": 1.1156, "step": 33 }, { "epoch": 0.0014334801947846617, "grad_norm": 10.105884552001953, "learning_rate": 2.318718381112985e-07, "loss": 2.0759, "step": 34 }, { "epoch": 0.0014756413769842105, "grad_norm": 12.164115905761719, "learning_rate": 2.3889825744800455e-07, "loss": 2.2586, "step": 35 }, { "epoch": 0.0015178025591837596, "grad_norm": 10.596022605895996, "learning_rate": 2.4592467678471053e-07, "loss": 2.2996, "step": 36 }, { "epoch": 0.0015599637413833084, "grad_norm": 11.034714698791504, "learning_rate": 2.529510961214165e-07, "loss": 2.1207, "step": 37 }, { "epoch": 0.0016021249235828572, "grad_norm": 9.249011993408203, "learning_rate": 2.5997751545812255e-07, "loss": 2.0278, "step": 38 }, { "epoch": 0.001644286105782406, "grad_norm": 12.26615047454834, "learning_rate": 2.670039347948286e-07, "loss": 2.3267, "step": 39 }, { "epoch": 0.0016864472879819551, "grad_norm": 16.289743423461914, "learning_rate": 2.7403035413153457e-07, "loss": 2.8577, "step": 40 }, { "epoch": 0.001728608470181504, "grad_norm": 9.91475772857666, "learning_rate": 2.810567734682406e-07, "loss": 2.1258, "step": 41 }, { "epoch": 0.0017707696523810528, "grad_norm": 7.819747447967529, "learning_rate": 2.8808319280494664e-07, "loss": 2.117, "step": 42 }, { "epoch": 0.0018129308345806016, "grad_norm": 10.145683288574219, "learning_rate": 2.951096121416526e-07, "loss": 2.2898, "step": 43 }, { "epoch": 0.0018550920167801504, "grad_norm": 14.144338607788086, "learning_rate": 3.021360314783586e-07, "loss": 2.4333, "step": 44 }, { "epoch": 0.0018972531989796995, "grad_norm": 9.046385765075684, "learning_rate": 3.0916245081506464e-07, "loss": 2.3465, "step": 45 }, { "epoch": 0.0019394143811792483, "grad_norm": 16.218324661254883, "learning_rate": 3.161888701517707e-07, "loss": 2.8051, "step": 46 }, { "epoch": 0.0019815755633787973, "grad_norm": 6.503572463989258, "learning_rate": 3.232152894884767e-07, "loss": 1.8521, "step": 47 }, { "epoch": 0.002023736745578346, "grad_norm": 6.277205467224121, "learning_rate": 3.302417088251827e-07, "loss": 1.8123, "step": 48 }, { "epoch": 0.002065897927777895, "grad_norm": 10.172727584838867, "learning_rate": 3.372681281618887e-07, "loss": 2.2624, "step": 49 }, { "epoch": 0.0021080591099774436, "grad_norm": 16.198806762695312, "learning_rate": 3.442945474985947e-07, "loss": 2.3858, "step": 50 }, { "epoch": 0.0021502202921769927, "grad_norm": 11.533013343811035, "learning_rate": 3.5132096683530076e-07, "loss": 2.3208, "step": 51 }, { "epoch": 0.0021923814743765417, "grad_norm": 6.535120487213135, "learning_rate": 3.5834738617200674e-07, "loss": 1.8716, "step": 52 }, { "epoch": 0.0022345426565760903, "grad_norm": 11.608048439025879, "learning_rate": 3.653738055087128e-07, "loss": 2.2832, "step": 53 }, { "epoch": 0.0022767038387756394, "grad_norm": 9.380041122436523, "learning_rate": 3.724002248454188e-07, "loss": 1.9012, "step": 54 }, { "epoch": 0.002318865020975188, "grad_norm": 10.371058464050293, "learning_rate": 3.794266441821248e-07, "loss": 2.0618, "step": 55 }, { "epoch": 0.002361026203174737, "grad_norm": 8.633589744567871, "learning_rate": 3.8645306351883083e-07, "loss": 2.1203, "step": 56 }, { "epoch": 0.002403187385374286, "grad_norm": 12.603894233703613, "learning_rate": 3.9347948285553687e-07, "loss": 2.616, "step": 57 }, { "epoch": 0.0024453485675738347, "grad_norm": 6.148855686187744, "learning_rate": 4.0050590219224285e-07, "loss": 1.9321, "step": 58 }, { "epoch": 0.0024875097497733837, "grad_norm": 7.670295715332031, "learning_rate": 4.0753232152894883e-07, "loss": 2.1575, "step": 59 }, { "epoch": 0.0025296709319729323, "grad_norm": 6.001925468444824, "learning_rate": 4.1455874086565487e-07, "loss": 1.8939, "step": 60 }, { "epoch": 0.0025718321141724814, "grad_norm": 10.126086235046387, "learning_rate": 4.2158516020236085e-07, "loss": 2.2297, "step": 61 }, { "epoch": 0.0026139932963720304, "grad_norm": 9.288052558898926, "learning_rate": 4.2861157953906694e-07, "loss": 2.2611, "step": 62 }, { "epoch": 0.002656154478571579, "grad_norm": 7.51863956451416, "learning_rate": 4.356379988757729e-07, "loss": 1.9889, "step": 63 }, { "epoch": 0.002698315660771128, "grad_norm": 7.739619731903076, "learning_rate": 4.4266441821247896e-07, "loss": 1.7813, "step": 64 }, { "epoch": 0.0027404768429706767, "grad_norm": 8.466773986816406, "learning_rate": 4.4969083754918495e-07, "loss": 2.1567, "step": 65 }, { "epoch": 0.0027826380251702258, "grad_norm": 8.682568550109863, "learning_rate": 4.5671725688589093e-07, "loss": 1.963, "step": 66 }, { "epoch": 0.002824799207369775, "grad_norm": 5.652549743652344, "learning_rate": 4.63743676222597e-07, "loss": 2.2341, "step": 67 }, { "epoch": 0.0028669603895693234, "grad_norm": 8.726757049560547, "learning_rate": 4.70770095559303e-07, "loss": 2.2343, "step": 68 }, { "epoch": 0.0029091215717688725, "grad_norm": 8.267630577087402, "learning_rate": 4.777965148960091e-07, "loss": 2.0228, "step": 69 }, { "epoch": 0.002951282753968421, "grad_norm": 7.755598068237305, "learning_rate": 4.848229342327151e-07, "loss": 2.2916, "step": 70 }, { "epoch": 0.00299344393616797, "grad_norm": 7.865350723266602, "learning_rate": 4.918493535694211e-07, "loss": 2.2163, "step": 71 }, { "epoch": 0.003035605118367519, "grad_norm": 9.473390579223633, "learning_rate": 4.98875772906127e-07, "loss": 2.1383, "step": 72 }, { "epoch": 0.0030777663005670678, "grad_norm": 4.517635345458984, "learning_rate": 5.05902192242833e-07, "loss": 1.6915, "step": 73 }, { "epoch": 0.003119927482766617, "grad_norm": 6.74133825302124, "learning_rate": 5.129286115795391e-07, "loss": 2.0919, "step": 74 }, { "epoch": 0.003162088664966166, "grad_norm": 3.8873074054718018, "learning_rate": 5.199550309162451e-07, "loss": 1.7293, "step": 75 }, { "epoch": 0.0032042498471657145, "grad_norm": 5.331579685211182, "learning_rate": 5.269814502529512e-07, "loss": 1.9583, "step": 76 }, { "epoch": 0.0032464110293652635, "grad_norm": 8.006376266479492, "learning_rate": 5.340078695896572e-07, "loss": 2.3437, "step": 77 }, { "epoch": 0.003288572211564812, "grad_norm": 7.240278720855713, "learning_rate": 5.410342889263632e-07, "loss": 2.0833, "step": 78 }, { "epoch": 0.003330733393764361, "grad_norm": 10.352540969848633, "learning_rate": 5.480607082630691e-07, "loss": 2.373, "step": 79 }, { "epoch": 0.0033728945759639102, "grad_norm": 4.930374622344971, "learning_rate": 5.550871275997751e-07, "loss": 2.0172, "step": 80 }, { "epoch": 0.003415055758163459, "grad_norm": 4.248158931732178, "learning_rate": 5.621135469364812e-07, "loss": 1.9551, "step": 81 }, { "epoch": 0.003457216940363008, "grad_norm": 6.0731201171875, "learning_rate": 5.691399662731872e-07, "loss": 1.8628, "step": 82 }, { "epoch": 0.0034993781225625565, "grad_norm": 4.317824363708496, "learning_rate": 5.761663856098933e-07, "loss": 1.9419, "step": 83 }, { "epoch": 0.0035415393047621055, "grad_norm": 3.199214220046997, "learning_rate": 5.831928049465993e-07, "loss": 1.4856, "step": 84 }, { "epoch": 0.0035837004869616546, "grad_norm": 6.204014301300049, "learning_rate": 5.902192242833052e-07, "loss": 2.0476, "step": 85 }, { "epoch": 0.003625861669161203, "grad_norm": 6.120738506317139, "learning_rate": 5.972456436200113e-07, "loss": 2.0301, "step": 86 }, { "epoch": 0.0036680228513607522, "grad_norm": 2.778083086013794, "learning_rate": 6.042720629567172e-07, "loss": 1.715, "step": 87 }, { "epoch": 0.003710184033560301, "grad_norm": 1.8544280529022217, "learning_rate": 6.112984822934233e-07, "loss": 1.0743, "step": 88 }, { "epoch": 0.00375234521575985, "grad_norm": 5.810383319854736, "learning_rate": 6.183249016301293e-07, "loss": 1.8338, "step": 89 }, { "epoch": 0.003794506397959399, "grad_norm": 6.121753692626953, "learning_rate": 6.253513209668353e-07, "loss": 2.2375, "step": 90 }, { "epoch": 0.0038366675801589476, "grad_norm": 4.129782199859619, "learning_rate": 6.323777403035414e-07, "loss": 1.86, "step": 91 }, { "epoch": 0.0038788287623584966, "grad_norm": 5.186487197875977, "learning_rate": 6.394041596402473e-07, "loss": 2.201, "step": 92 }, { "epoch": 0.003920989944558045, "grad_norm": 2.258085250854492, "learning_rate": 6.464305789769534e-07, "loss": 1.6976, "step": 93 }, { "epoch": 0.003963151126757595, "grad_norm": 5.353981971740723, "learning_rate": 6.534569983136594e-07, "loss": 1.9612, "step": 94 }, { "epoch": 0.004005312308957143, "grad_norm": 3.2475574016571045, "learning_rate": 6.604834176503654e-07, "loss": 1.8627, "step": 95 }, { "epoch": 0.004047473491156692, "grad_norm": 1.7484928369522095, "learning_rate": 6.675098369870714e-07, "loss": 1.1422, "step": 96 }, { "epoch": 0.0040896346733562405, "grad_norm": 4.084005832672119, "learning_rate": 6.745362563237774e-07, "loss": 1.9461, "step": 97 }, { "epoch": 0.00413179585555579, "grad_norm": 3.4996533393859863, "learning_rate": 6.815626756604835e-07, "loss": 2.0259, "step": 98 }, { "epoch": 0.004173957037755339, "grad_norm": 4.819375038146973, "learning_rate": 6.885890949971894e-07, "loss": 1.9668, "step": 99 }, { "epoch": 0.004216118219954887, "grad_norm": 2.417015790939331, "learning_rate": 6.956155143338955e-07, "loss": 1.7845, "step": 100 }, { "epoch": 0.004258279402154437, "grad_norm": 7.097517967224121, "learning_rate": 7.026419336706015e-07, "loss": 1.6876, "step": 101 }, { "epoch": 0.004300440584353985, "grad_norm": 2.837721824645996, "learning_rate": 7.096683530073076e-07, "loss": 1.9373, "step": 102 }, { "epoch": 0.004342601766553534, "grad_norm": 7.416475772857666, "learning_rate": 7.166947723440135e-07, "loss": 2.0264, "step": 103 }, { "epoch": 0.004384762948753083, "grad_norm": 4.350272178649902, "learning_rate": 7.237211916807195e-07, "loss": 1.9517, "step": 104 }, { "epoch": 0.004426924130952632, "grad_norm": 4.746272563934326, "learning_rate": 7.307476110174256e-07, "loss": 2.0398, "step": 105 }, { "epoch": 0.004469085313152181, "grad_norm": 2.8965442180633545, "learning_rate": 7.377740303541315e-07, "loss": 2.0257, "step": 106 }, { "epoch": 0.004511246495351729, "grad_norm": 2.649604082107544, "learning_rate": 7.448004496908376e-07, "loss": 2.0183, "step": 107 }, { "epoch": 0.004553407677551279, "grad_norm": 3.4210212230682373, "learning_rate": 7.518268690275436e-07, "loss": 1.789, "step": 108 }, { "epoch": 0.004595568859750827, "grad_norm": 14.049697875976562, "learning_rate": 7.588532883642496e-07, "loss": 1.6985, "step": 109 }, { "epoch": 0.004637730041950376, "grad_norm": 2.357064723968506, "learning_rate": 7.658797077009557e-07, "loss": 1.9052, "step": 110 }, { "epoch": 0.0046798912241499254, "grad_norm": 2.7837681770324707, "learning_rate": 7.729061270376617e-07, "loss": 1.9577, "step": 111 }, { "epoch": 0.004722052406349474, "grad_norm": 2.9170267581939697, "learning_rate": 7.799325463743676e-07, "loss": 1.8045, "step": 112 }, { "epoch": 0.004764213588549023, "grad_norm": 2.298275947570801, "learning_rate": 7.869589657110737e-07, "loss": 1.883, "step": 113 }, { "epoch": 0.004806374770748572, "grad_norm": 3.022411346435547, "learning_rate": 7.939853850477796e-07, "loss": 1.8221, "step": 114 }, { "epoch": 0.004848535952948121, "grad_norm": 1.752841591835022, "learning_rate": 8.010118043844857e-07, "loss": 1.3699, "step": 115 }, { "epoch": 0.004890697135147669, "grad_norm": 3.090153694152832, "learning_rate": 8.080382237211918e-07, "loss": 1.6901, "step": 116 }, { "epoch": 0.004932858317347218, "grad_norm": 1.9334534406661987, "learning_rate": 8.150646430578977e-07, "loss": 1.7497, "step": 117 }, { "epoch": 0.0049750194995467675, "grad_norm": 1.5166704654693604, "learning_rate": 8.220910623946038e-07, "loss": 1.3795, "step": 118 }, { "epoch": 0.005017180681746316, "grad_norm": 2.0587782859802246, "learning_rate": 8.291174817313097e-07, "loss": 1.5798, "step": 119 }, { "epoch": 0.005059341863945865, "grad_norm": 2.4106192588806152, "learning_rate": 8.361439010680158e-07, "loss": 1.6824, "step": 120 }, { "epoch": 0.005101503046145414, "grad_norm": 3.339653491973877, "learning_rate": 8.431703204047217e-07, "loss": 1.7853, "step": 121 }, { "epoch": 0.005143664228344963, "grad_norm": 1.9808452129364014, "learning_rate": 8.501967397414278e-07, "loss": 1.9483, "step": 122 }, { "epoch": 0.005185825410544511, "grad_norm": 1.6235350370407104, "learning_rate": 8.572231590781339e-07, "loss": 1.6397, "step": 123 }, { "epoch": 0.005227986592744061, "grad_norm": 5.085844993591309, "learning_rate": 8.642495784148398e-07, "loss": 2.013, "step": 124 }, { "epoch": 0.0052701477749436095, "grad_norm": 1.5156265497207642, "learning_rate": 8.712759977515459e-07, "loss": 1.6219, "step": 125 }, { "epoch": 0.005312308957143158, "grad_norm": 3.389305591583252, "learning_rate": 8.783024170882519e-07, "loss": 2.1132, "step": 126 }, { "epoch": 0.005354470139342708, "grad_norm": 1.4653531312942505, "learning_rate": 8.853288364249579e-07, "loss": 1.5619, "step": 127 }, { "epoch": 0.005396631321542256, "grad_norm": 2.920271396636963, "learning_rate": 8.923552557616639e-07, "loss": 1.8821, "step": 128 }, { "epoch": 0.005438792503741805, "grad_norm": 2.7970175743103027, "learning_rate": 8.993816750983699e-07, "loss": 1.79, "step": 129 }, { "epoch": 0.005480953685941353, "grad_norm": 1.6761364936828613, "learning_rate": 9.06408094435076e-07, "loss": 1.7708, "step": 130 }, { "epoch": 0.005523114868140903, "grad_norm": 1.9545729160308838, "learning_rate": 9.134345137717819e-07, "loss": 1.7088, "step": 131 }, { "epoch": 0.0055652760503404515, "grad_norm": 1.4782054424285889, "learning_rate": 9.20460933108488e-07, "loss": 1.6972, "step": 132 }, { "epoch": 0.00560743723254, "grad_norm": 1.6577332019805908, "learning_rate": 9.27487352445194e-07, "loss": 1.6011, "step": 133 }, { "epoch": 0.00564959841473955, "grad_norm": 1.1006001234054565, "learning_rate": 9.345137717818999e-07, "loss": 1.0764, "step": 134 }, { "epoch": 0.005691759596939098, "grad_norm": 3.8846168518066406, "learning_rate": 9.41540191118606e-07, "loss": 1.8036, "step": 135 }, { "epoch": 0.005733920779138647, "grad_norm": 2.0692474842071533, "learning_rate": 9.48566610455312e-07, "loss": 1.7789, "step": 136 }, { "epoch": 0.005776081961338196, "grad_norm": 0.9167259335517883, "learning_rate": 9.555930297920182e-07, "loss": 1.2063, "step": 137 }, { "epoch": 0.005818243143537745, "grad_norm": 3.4215033054351807, "learning_rate": 9.62619449128724e-07, "loss": 2.4444, "step": 138 }, { "epoch": 0.0058604043257372935, "grad_norm": 2.8543312549591064, "learning_rate": 9.696458684654301e-07, "loss": 1.5733, "step": 139 }, { "epoch": 0.005902565507936842, "grad_norm": 1.3607869148254395, "learning_rate": 9.76672287802136e-07, "loss": 1.7665, "step": 140 }, { "epoch": 0.005944726690136392, "grad_norm": 2.046088933944702, "learning_rate": 9.836987071388421e-07, "loss": 1.6488, "step": 141 }, { "epoch": 0.00598688787233594, "grad_norm": 1.676144003868103, "learning_rate": 9.90725126475548e-07, "loss": 1.5555, "step": 142 }, { "epoch": 0.006029049054535489, "grad_norm": 1.2840907573699951, "learning_rate": 9.97751545812254e-07, "loss": 1.7213, "step": 143 }, { "epoch": 0.006071210236735038, "grad_norm": 1.6809825897216797, "learning_rate": 1.0047779651489602e-06, "loss": 2.0289, "step": 144 }, { "epoch": 0.006113371418934587, "grad_norm": 1.5389162302017212, "learning_rate": 1.011804384485666e-06, "loss": 1.6327, "step": 145 }, { "epoch": 0.0061555326011341355, "grad_norm": 2.5830929279327393, "learning_rate": 1.0188308038223721e-06, "loss": 1.5591, "step": 146 }, { "epoch": 0.006197693783333685, "grad_norm": 2.7121338844299316, "learning_rate": 1.0258572231590782e-06, "loss": 2.1349, "step": 147 }, { "epoch": 0.006239854965533234, "grad_norm": 1.050451397895813, "learning_rate": 1.032883642495784e-06, "loss": 0.9861, "step": 148 }, { "epoch": 0.006282016147732782, "grad_norm": 1.2004363536834717, "learning_rate": 1.0399100618324902e-06, "loss": 1.5741, "step": 149 }, { "epoch": 0.006324177329932332, "grad_norm": 1.1234078407287598, "learning_rate": 1.0469364811691963e-06, "loss": 1.608, "step": 150 }, { "epoch": 0.00636633851213188, "grad_norm": 1.3859996795654297, "learning_rate": 1.0539629005059024e-06, "loss": 1.7803, "step": 151 }, { "epoch": 0.006408499694331429, "grad_norm": 1.4821362495422363, "learning_rate": 1.0609893198426083e-06, "loss": 1.7405, "step": 152 }, { "epoch": 0.0064506608765309776, "grad_norm": 1.199167013168335, "learning_rate": 1.0680157391793143e-06, "loss": 1.263, "step": 153 }, { "epoch": 0.006492822058730527, "grad_norm": 3.8304531574249268, "learning_rate": 1.0750421585160204e-06, "loss": 1.9458, "step": 154 }, { "epoch": 0.006534983240930076, "grad_norm": 1.5649083852767944, "learning_rate": 1.0820685778527263e-06, "loss": 1.6794, "step": 155 }, { "epoch": 0.006577144423129624, "grad_norm": 1.2089709043502808, "learning_rate": 1.0890949971894324e-06, "loss": 1.7291, "step": 156 }, { "epoch": 0.006619305605329174, "grad_norm": 1.3002452850341797, "learning_rate": 1.0961214165261383e-06, "loss": 1.6186, "step": 157 }, { "epoch": 0.006661466787528722, "grad_norm": 1.1235030889511108, "learning_rate": 1.1031478358628444e-06, "loss": 1.4576, "step": 158 }, { "epoch": 0.006703627969728271, "grad_norm": 1.3695344924926758, "learning_rate": 1.1101742551995502e-06, "loss": 1.821, "step": 159 }, { "epoch": 0.0067457891519278204, "grad_norm": 1.318076491355896, "learning_rate": 1.1172006745362563e-06, "loss": 1.748, "step": 160 }, { "epoch": 0.006787950334127369, "grad_norm": 1.5532612800598145, "learning_rate": 1.1242270938729624e-06, "loss": 1.8706, "step": 161 }, { "epoch": 0.006830111516326918, "grad_norm": 1.6164528131484985, "learning_rate": 1.1312535132096683e-06, "loss": 1.1193, "step": 162 }, { "epoch": 0.006872272698526466, "grad_norm": 1.4164823293685913, "learning_rate": 1.1382799325463744e-06, "loss": 1.8419, "step": 163 }, { "epoch": 0.006914433880726016, "grad_norm": 1.1721580028533936, "learning_rate": 1.1453063518830805e-06, "loss": 1.468, "step": 164 }, { "epoch": 0.006956595062925564, "grad_norm": 1.184140682220459, "learning_rate": 1.1523327712197866e-06, "loss": 1.4852, "step": 165 }, { "epoch": 0.006998756245125113, "grad_norm": 1.06318998336792, "learning_rate": 1.1593591905564924e-06, "loss": 1.6974, "step": 166 }, { "epoch": 0.0070409174273246625, "grad_norm": 1.2340329885482788, "learning_rate": 1.1663856098931985e-06, "loss": 1.8071, "step": 167 }, { "epoch": 0.007083078609524211, "grad_norm": 1.049524188041687, "learning_rate": 1.1734120292299046e-06, "loss": 1.2945, "step": 168 }, { "epoch": 0.00712523979172376, "grad_norm": 1.5732600688934326, "learning_rate": 1.1804384485666105e-06, "loss": 1.5151, "step": 169 }, { "epoch": 0.007167400973923309, "grad_norm": 1.1310746669769287, "learning_rate": 1.1874648679033166e-06, "loss": 1.6779, "step": 170 }, { "epoch": 0.007209562156122858, "grad_norm": 1.0383481979370117, "learning_rate": 1.1944912872400227e-06, "loss": 1.6468, "step": 171 }, { "epoch": 0.007251723338322406, "grad_norm": 1.614333987236023, "learning_rate": 1.2015177065767286e-06, "loss": 1.8469, "step": 172 }, { "epoch": 0.007293884520521955, "grad_norm": 1.3124310970306396, "learning_rate": 1.2085441259134344e-06, "loss": 1.4393, "step": 173 }, { "epoch": 0.0073360457027215045, "grad_norm": 1.3593467473983765, "learning_rate": 1.2155705452501405e-06, "loss": 1.5475, "step": 174 }, { "epoch": 0.007378206884921053, "grad_norm": 1.5464469194412231, "learning_rate": 1.2225969645868466e-06, "loss": 1.4594, "step": 175 }, { "epoch": 0.007420368067120602, "grad_norm": 1.3222824335098267, "learning_rate": 1.2296233839235525e-06, "loss": 1.5965, "step": 176 }, { "epoch": 0.007462529249320151, "grad_norm": 1.0900923013687134, "learning_rate": 1.2366498032602586e-06, "loss": 1.408, "step": 177 }, { "epoch": 0.0075046904315197, "grad_norm": 1.3595432043075562, "learning_rate": 1.2436762225969647e-06, "loss": 1.4918, "step": 178 }, { "epoch": 0.007546851613719248, "grad_norm": 1.345959186553955, "learning_rate": 1.2507026419336705e-06, "loss": 1.6414, "step": 179 }, { "epoch": 0.007589012795918798, "grad_norm": 1.5702348947525024, "learning_rate": 1.2577290612703766e-06, "loss": 1.7961, "step": 180 }, { "epoch": 0.0076311739781183465, "grad_norm": 0.9913703799247742, "learning_rate": 1.2647554806070827e-06, "loss": 0.9045, "step": 181 }, { "epoch": 0.007673335160317895, "grad_norm": 2.876697063446045, "learning_rate": 1.2717818999437888e-06, "loss": 2.4151, "step": 182 }, { "epoch": 0.007715496342517445, "grad_norm": 1.1710377931594849, "learning_rate": 1.2788083192804947e-06, "loss": 1.3918, "step": 183 }, { "epoch": 0.007757657524716993, "grad_norm": 1.0853137969970703, "learning_rate": 1.2858347386172008e-06, "loss": 1.6026, "step": 184 }, { "epoch": 0.007799818706916542, "grad_norm": 1.0273147821426392, "learning_rate": 1.2928611579539069e-06, "loss": 1.392, "step": 185 }, { "epoch": 0.00784197988911609, "grad_norm": 0.9112412929534912, "learning_rate": 1.2998875772906127e-06, "loss": 1.3965, "step": 186 }, { "epoch": 0.00788414107131564, "grad_norm": 0.6272501945495605, "learning_rate": 1.3069139966273188e-06, "loss": 1.0521, "step": 187 }, { "epoch": 0.00792630225351519, "grad_norm": 2.5938806533813477, "learning_rate": 1.313940415964025e-06, "loss": 2.1425, "step": 188 }, { "epoch": 0.007968463435714737, "grad_norm": 0.9216694235801697, "learning_rate": 1.3209668353007308e-06, "loss": 1.4915, "step": 189 }, { "epoch": 0.008010624617914287, "grad_norm": 0.9356284141540527, "learning_rate": 1.3279932546374367e-06, "loss": 1.3343, "step": 190 }, { "epoch": 0.008052785800113834, "grad_norm": 0.6138125061988831, "learning_rate": 1.3350196739741428e-06, "loss": 1.0317, "step": 191 }, { "epoch": 0.008094946982313384, "grad_norm": 1.474182367324829, "learning_rate": 1.3420460933108489e-06, "loss": 1.5404, "step": 192 }, { "epoch": 0.008137108164512933, "grad_norm": 1.0405019521713257, "learning_rate": 1.3490725126475547e-06, "loss": 1.5514, "step": 193 }, { "epoch": 0.008179269346712481, "grad_norm": 1.0377756357192993, "learning_rate": 1.3560989319842608e-06, "loss": 1.6142, "step": 194 }, { "epoch": 0.00822143052891203, "grad_norm": 1.1223738193511963, "learning_rate": 1.363125351320967e-06, "loss": 1.3493, "step": 195 }, { "epoch": 0.00826359171111158, "grad_norm": 4.027480602264404, "learning_rate": 1.370151770657673e-06, "loss": 2.6943, "step": 196 }, { "epoch": 0.008305752893311128, "grad_norm": 0.7396878600120544, "learning_rate": 1.3771781899943789e-06, "loss": 1.1936, "step": 197 }, { "epoch": 0.008347914075510677, "grad_norm": 0.9531145691871643, "learning_rate": 1.384204609331085e-06, "loss": 1.2762, "step": 198 }, { "epoch": 0.008390075257710227, "grad_norm": 1.9607980251312256, "learning_rate": 1.391231028667791e-06, "loss": 1.1246, "step": 199 }, { "epoch": 0.008432236439909774, "grad_norm": 0.9082164764404297, "learning_rate": 1.398257448004497e-06, "loss": 1.4696, "step": 200 }, { "epoch": 0.008474397622109324, "grad_norm": 0.9004429578781128, "learning_rate": 1.405283867341203e-06, "loss": 1.3563, "step": 201 }, { "epoch": 0.008516558804308873, "grad_norm": 1.266679048538208, "learning_rate": 1.4123102866779091e-06, "loss": 1.4856, "step": 202 }, { "epoch": 0.008558719986508421, "grad_norm": 0.9642159938812256, "learning_rate": 1.4193367060146152e-06, "loss": 1.357, "step": 203 }, { "epoch": 0.00860088116870797, "grad_norm": 1.1820989847183228, "learning_rate": 1.426363125351321e-06, "loss": 1.5099, "step": 204 }, { "epoch": 0.00864304235090752, "grad_norm": 1.449539303779602, "learning_rate": 1.433389544688027e-06, "loss": 1.4839, "step": 205 }, { "epoch": 0.008685203533107068, "grad_norm": 1.1227446794509888, "learning_rate": 1.440415964024733e-06, "loss": 1.531, "step": 206 }, { "epoch": 0.008727364715306617, "grad_norm": 1.199463963508606, "learning_rate": 1.447442383361439e-06, "loss": 1.504, "step": 207 }, { "epoch": 0.008769525897506167, "grad_norm": 1.2566994428634644, "learning_rate": 1.454468802698145e-06, "loss": 1.4437, "step": 208 }, { "epoch": 0.008811687079705715, "grad_norm": 0.7420704960823059, "learning_rate": 1.461495222034851e-06, "loss": 0.8894, "step": 209 }, { "epoch": 0.008853848261905264, "grad_norm": 1.1516697406768799, "learning_rate": 1.468521641371557e-06, "loss": 1.3341, "step": 210 }, { "epoch": 0.008896009444104814, "grad_norm": 1.6382427215576172, "learning_rate": 1.475548060708263e-06, "loss": 1.3102, "step": 211 }, { "epoch": 0.008938170626304361, "grad_norm": 1.289696455001831, "learning_rate": 1.4825744800449692e-06, "loss": 1.3536, "step": 212 }, { "epoch": 0.00898033180850391, "grad_norm": 0.8224149346351624, "learning_rate": 1.4896008993816752e-06, "loss": 1.0534, "step": 213 }, { "epoch": 0.009022492990703459, "grad_norm": 1.3075286149978638, "learning_rate": 1.4966273187183811e-06, "loss": 1.3976, "step": 214 }, { "epoch": 0.009064654172903008, "grad_norm": 2.0153791904449463, "learning_rate": 1.5036537380550872e-06, "loss": 1.303, "step": 215 }, { "epoch": 0.009106815355102557, "grad_norm": 1.1419061422348022, "learning_rate": 1.5106801573917933e-06, "loss": 1.2655, "step": 216 }, { "epoch": 0.009148976537302105, "grad_norm": 2.223149299621582, "learning_rate": 1.5177065767284992e-06, "loss": 1.6959, "step": 217 }, { "epoch": 0.009191137719501655, "grad_norm": 1.226813554763794, "learning_rate": 1.5247329960652053e-06, "loss": 1.2449, "step": 218 }, { "epoch": 0.009233298901701204, "grad_norm": 1.077543020248413, "learning_rate": 1.5317594154019114e-06, "loss": 1.4889, "step": 219 }, { "epoch": 0.009275460083900752, "grad_norm": 1.1297895908355713, "learning_rate": 1.5387858347386174e-06, "loss": 1.4151, "step": 220 }, { "epoch": 0.009317621266100301, "grad_norm": 0.8562735915184021, "learning_rate": 1.5458122540753233e-06, "loss": 1.2464, "step": 221 }, { "epoch": 0.009359782448299851, "grad_norm": 0.9254936575889587, "learning_rate": 1.5528386734120292e-06, "loss": 1.234, "step": 222 }, { "epoch": 0.009401943630499399, "grad_norm": 0.9382602572441101, "learning_rate": 1.5598650927487353e-06, "loss": 1.5004, "step": 223 }, { "epoch": 0.009444104812698948, "grad_norm": 1.3466649055480957, "learning_rate": 1.5668915120854412e-06, "loss": 1.2234, "step": 224 }, { "epoch": 0.009486265994898498, "grad_norm": 1.270646572113037, "learning_rate": 1.5739179314221475e-06, "loss": 1.3093, "step": 225 }, { "epoch": 0.009528427177098045, "grad_norm": 1.1392769813537598, "learning_rate": 1.5809443507588533e-06, "loss": 1.215, "step": 226 }, { "epoch": 0.009570588359297595, "grad_norm": 2.5960748195648193, "learning_rate": 1.5879707700955592e-06, "loss": 2.1944, "step": 227 }, { "epoch": 0.009612749541497144, "grad_norm": 2.3791592121124268, "learning_rate": 1.5949971894322655e-06, "loss": 2.0418, "step": 228 }, { "epoch": 0.009654910723696692, "grad_norm": 1.5944149494171143, "learning_rate": 1.6020236087689714e-06, "loss": 1.9429, "step": 229 }, { "epoch": 0.009697071905896242, "grad_norm": 1.7780572175979614, "learning_rate": 1.6090500281056773e-06, "loss": 1.2155, "step": 230 }, { "epoch": 0.009739233088095791, "grad_norm": 1.7258695363998413, "learning_rate": 1.6160764474423836e-06, "loss": 1.3852, "step": 231 }, { "epoch": 0.009781394270295339, "grad_norm": 1.02475905418396, "learning_rate": 1.6231028667790895e-06, "loss": 1.341, "step": 232 }, { "epoch": 0.009823555452494888, "grad_norm": 1.1996458768844604, "learning_rate": 1.6301292861157953e-06, "loss": 1.3582, "step": 233 }, { "epoch": 0.009865716634694436, "grad_norm": 1.3359684944152832, "learning_rate": 1.6371557054525016e-06, "loss": 1.3378, "step": 234 }, { "epoch": 0.009907877816893985, "grad_norm": 1.1263518333435059, "learning_rate": 1.6441821247892075e-06, "loss": 1.3785, "step": 235 }, { "epoch": 0.009950038999093535, "grad_norm": 1.1446256637573242, "learning_rate": 1.6512085441259134e-06, "loss": 1.1653, "step": 236 }, { "epoch": 0.009992200181293083, "grad_norm": 1.2582635879516602, "learning_rate": 1.6582349634626195e-06, "loss": 1.4483, "step": 237 }, { "epoch": 0.010034361363492632, "grad_norm": 1.39429771900177, "learning_rate": 1.6652613827993256e-06, "loss": 1.513, "step": 238 }, { "epoch": 0.010076522545692182, "grad_norm": 1.4948402643203735, "learning_rate": 1.6722878021360317e-06, "loss": 1.163, "step": 239 }, { "epoch": 0.01011868372789173, "grad_norm": 1.1873283386230469, "learning_rate": 1.6793142214727375e-06, "loss": 1.2395, "step": 240 }, { "epoch": 0.010160844910091279, "grad_norm": 1.5302131175994873, "learning_rate": 1.6863406408094434e-06, "loss": 1.5014, "step": 241 }, { "epoch": 0.010203006092290828, "grad_norm": 1.1987978219985962, "learning_rate": 1.6933670601461497e-06, "loss": 1.432, "step": 242 }, { "epoch": 0.010245167274490376, "grad_norm": 1.1788595914840698, "learning_rate": 1.7003934794828556e-06, "loss": 1.4981, "step": 243 }, { "epoch": 0.010287328456689926, "grad_norm": 1.1288137435913086, "learning_rate": 1.7074198988195615e-06, "loss": 1.3413, "step": 244 }, { "epoch": 0.010329489638889475, "grad_norm": 1.1237248182296753, "learning_rate": 1.7144463181562678e-06, "loss": 1.0391, "step": 245 }, { "epoch": 0.010371650821089023, "grad_norm": 1.332138180732727, "learning_rate": 1.7214727374929736e-06, "loss": 1.1721, "step": 246 }, { "epoch": 0.010413812003288572, "grad_norm": 1.6286427974700928, "learning_rate": 1.7284991568296795e-06, "loss": 1.2817, "step": 247 }, { "epoch": 0.010455973185488122, "grad_norm": 1.56279718875885, "learning_rate": 1.7355255761663858e-06, "loss": 1.6498, "step": 248 }, { "epoch": 0.01049813436768767, "grad_norm": 2.341465711593628, "learning_rate": 1.7425519955030917e-06, "loss": 1.5656, "step": 249 }, { "epoch": 0.010540295549887219, "grad_norm": 1.518951177597046, "learning_rate": 1.7495784148397976e-06, "loss": 1.2004, "step": 250 }, { "epoch": 0.010582456732086768, "grad_norm": 0.9176362156867981, "learning_rate": 1.7566048341765039e-06, "loss": 0.6291, "step": 251 }, { "epoch": 0.010624617914286316, "grad_norm": 1.4625276327133179, "learning_rate": 1.7636312535132098e-06, "loss": 1.149, "step": 252 }, { "epoch": 0.010666779096485866, "grad_norm": 1.3040893077850342, "learning_rate": 1.7706576728499159e-06, "loss": 1.0583, "step": 253 }, { "epoch": 0.010708940278685415, "grad_norm": 1.2280478477478027, "learning_rate": 1.7776840921866217e-06, "loss": 1.5161, "step": 254 }, { "epoch": 0.010751101460884963, "grad_norm": 1.3811821937561035, "learning_rate": 1.7847105115233278e-06, "loss": 1.246, "step": 255 }, { "epoch": 0.010793262643084512, "grad_norm": 0.7167531251907349, "learning_rate": 1.791736930860034e-06, "loss": 0.871, "step": 256 }, { "epoch": 0.01083542382528406, "grad_norm": 1.2268483638763428, "learning_rate": 1.7987633501967398e-06, "loss": 1.3092, "step": 257 }, { "epoch": 0.01087758500748361, "grad_norm": 1.5290592908859253, "learning_rate": 1.8057897695334457e-06, "loss": 1.0724, "step": 258 }, { "epoch": 0.010919746189683159, "grad_norm": 1.0740593671798706, "learning_rate": 1.812816188870152e-06, "loss": 1.3327, "step": 259 }, { "epoch": 0.010961907371882707, "grad_norm": 1.6105196475982666, "learning_rate": 1.8198426082068578e-06, "loss": 1.443, "step": 260 }, { "epoch": 0.011004068554082256, "grad_norm": 2.846116542816162, "learning_rate": 1.8268690275435637e-06, "loss": 1.0243, "step": 261 }, { "epoch": 0.011046229736281806, "grad_norm": 1.262549877166748, "learning_rate": 1.83389544688027e-06, "loss": 0.9862, "step": 262 }, { "epoch": 0.011088390918481354, "grad_norm": 1.6165436506271362, "learning_rate": 1.840921866216976e-06, "loss": 1.1139, "step": 263 }, { "epoch": 0.011130552100680903, "grad_norm": 1.7001943588256836, "learning_rate": 1.8479482855536818e-06, "loss": 1.2429, "step": 264 }, { "epoch": 0.011172713282880452, "grad_norm": 1.12836492061615, "learning_rate": 1.854974704890388e-06, "loss": 1.1927, "step": 265 }, { "epoch": 0.01121487446508, "grad_norm": 1.2059670686721802, "learning_rate": 1.862001124227094e-06, "loss": 1.1419, "step": 266 }, { "epoch": 0.01125703564727955, "grad_norm": 1.1556992530822754, "learning_rate": 1.8690275435637998e-06, "loss": 1.1532, "step": 267 }, { "epoch": 0.0112991968294791, "grad_norm": 1.048303484916687, "learning_rate": 1.8760539629005061e-06, "loss": 0.991, "step": 268 }, { "epoch": 0.011341358011678647, "grad_norm": 1.427024006843567, "learning_rate": 1.883080382237212e-06, "loss": 1.3049, "step": 269 }, { "epoch": 0.011383519193878196, "grad_norm": 1.167110562324524, "learning_rate": 1.890106801573918e-06, "loss": 1.4692, "step": 270 }, { "epoch": 0.011425680376077746, "grad_norm": 1.1814876794815063, "learning_rate": 1.897133220910624e-06, "loss": 1.494, "step": 271 }, { "epoch": 0.011467841558277294, "grad_norm": 1.2540966272354126, "learning_rate": 1.9041596402473299e-06, "loss": 1.0489, "step": 272 }, { "epoch": 0.011510002740476843, "grad_norm": 1.8231898546218872, "learning_rate": 1.9111860595840364e-06, "loss": 0.8261, "step": 273 }, { "epoch": 0.011552163922676393, "grad_norm": 1.6342029571533203, "learning_rate": 1.918212478920742e-06, "loss": 1.1629, "step": 274 }, { "epoch": 0.01159432510487594, "grad_norm": 1.4157445430755615, "learning_rate": 1.925238898257448e-06, "loss": 1.4811, "step": 275 }, { "epoch": 0.01163648628707549, "grad_norm": 1.3290406465530396, "learning_rate": 1.932265317594154e-06, "loss": 1.5863, "step": 276 }, { "epoch": 0.01167864746927504, "grad_norm": 1.979476809501648, "learning_rate": 1.9392917369308603e-06, "loss": 1.4557, "step": 277 }, { "epoch": 0.011720808651474587, "grad_norm": 1.2370938062667847, "learning_rate": 1.946318156267566e-06, "loss": 1.3237, "step": 278 }, { "epoch": 0.011762969833674137, "grad_norm": 1.7224557399749756, "learning_rate": 1.953344575604272e-06, "loss": 0.985, "step": 279 }, { "epoch": 0.011805131015873684, "grad_norm": 1.4851806163787842, "learning_rate": 1.960370994940978e-06, "loss": 1.2455, "step": 280 }, { "epoch": 0.011847292198073234, "grad_norm": 1.271363615989685, "learning_rate": 1.9673974142776842e-06, "loss": 0.9592, "step": 281 }, { "epoch": 0.011889453380272783, "grad_norm": 1.0950145721435547, "learning_rate": 1.9744238336143903e-06, "loss": 1.13, "step": 282 }, { "epoch": 0.011931614562472331, "grad_norm": 1.2840460538864136, "learning_rate": 1.981450252951096e-06, "loss": 0.9954, "step": 283 }, { "epoch": 0.01197377574467188, "grad_norm": 1.1775761842727661, "learning_rate": 1.9884766722878025e-06, "loss": 1.2646, "step": 284 }, { "epoch": 0.01201593692687143, "grad_norm": 1.6483384370803833, "learning_rate": 1.995503091624508e-06, "loss": 1.214, "step": 285 }, { "epoch": 0.012058098109070978, "grad_norm": 1.2853058576583862, "learning_rate": 2.0025295109612143e-06, "loss": 1.0837, "step": 286 }, { "epoch": 0.012100259291270527, "grad_norm": 2.1699957847595215, "learning_rate": 2.0095559302979203e-06, "loss": 1.5705, "step": 287 }, { "epoch": 0.012142420473470077, "grad_norm": 2.092994451522827, "learning_rate": 2.0165823496346264e-06, "loss": 0.9993, "step": 288 }, { "epoch": 0.012184581655669624, "grad_norm": 2.7220849990844727, "learning_rate": 2.023608768971332e-06, "loss": 0.9613, "step": 289 }, { "epoch": 0.012226742837869174, "grad_norm": 1.7108561992645264, "learning_rate": 2.0306351883080386e-06, "loss": 0.9645, "step": 290 }, { "epoch": 0.012268904020068723, "grad_norm": 1.6314144134521484, "learning_rate": 2.0376616076447443e-06, "loss": 0.7821, "step": 291 }, { "epoch": 0.012311065202268271, "grad_norm": 2.017411470413208, "learning_rate": 2.0446880269814504e-06, "loss": 1.3691, "step": 292 }, { "epoch": 0.01235322638446782, "grad_norm": 1.0389409065246582, "learning_rate": 2.0517144463181565e-06, "loss": 0.8533, "step": 293 }, { "epoch": 0.01239538756666737, "grad_norm": 1.3327276706695557, "learning_rate": 2.0587408656548625e-06, "loss": 0.9145, "step": 294 }, { "epoch": 0.012437548748866918, "grad_norm": 1.4158109426498413, "learning_rate": 2.065767284991568e-06, "loss": 1.4209, "step": 295 }, { "epoch": 0.012479709931066467, "grad_norm": 1.198509693145752, "learning_rate": 2.0727937043282743e-06, "loss": 0.8306, "step": 296 }, { "epoch": 0.012521871113266017, "grad_norm": 2.444201946258545, "learning_rate": 2.0798201236649804e-06, "loss": 1.883, "step": 297 }, { "epoch": 0.012564032295465564, "grad_norm": 1.3035672903060913, "learning_rate": 2.0868465430016865e-06, "loss": 1.5461, "step": 298 }, { "epoch": 0.012606193477665114, "grad_norm": 1.5803570747375488, "learning_rate": 2.0938729623383926e-06, "loss": 0.9891, "step": 299 }, { "epoch": 0.012648354659864663, "grad_norm": 1.4919393062591553, "learning_rate": 2.1008993816750982e-06, "loss": 0.7796, "step": 300 }, { "epoch": 0.012690515842064211, "grad_norm": 1.4624124765396118, "learning_rate": 2.1079258010118047e-06, "loss": 1.0035, "step": 301 }, { "epoch": 0.01273267702426376, "grad_norm": 1.1518349647521973, "learning_rate": 2.1149522203485104e-06, "loss": 0.9002, "step": 302 }, { "epoch": 0.012774838206463308, "grad_norm": 0.8686354160308838, "learning_rate": 2.1219786396852165e-06, "loss": 0.8526, "step": 303 }, { "epoch": 0.012816999388662858, "grad_norm": 1.1832365989685059, "learning_rate": 2.1290050590219226e-06, "loss": 1.3146, "step": 304 }, { "epoch": 0.012859160570862407, "grad_norm": 1.095974087715149, "learning_rate": 2.1360314783586287e-06, "loss": 1.0456, "step": 305 }, { "epoch": 0.012901321753061955, "grad_norm": 1.3608756065368652, "learning_rate": 2.1430578976953343e-06, "loss": 0.7621, "step": 306 }, { "epoch": 0.012943482935261505, "grad_norm": 1.3847835063934326, "learning_rate": 2.150084317032041e-06, "loss": 0.8537, "step": 307 }, { "epoch": 0.012985644117461054, "grad_norm": 1.2197686433792114, "learning_rate": 2.1571107363687465e-06, "loss": 1.0166, "step": 308 }, { "epoch": 0.013027805299660602, "grad_norm": 1.208760142326355, "learning_rate": 2.1641371557054526e-06, "loss": 0.7636, "step": 309 }, { "epoch": 0.013069966481860151, "grad_norm": 1.2490131855010986, "learning_rate": 2.1711635750421587e-06, "loss": 0.7619, "step": 310 }, { "epoch": 0.0131121276640597, "grad_norm": 2.906015396118164, "learning_rate": 2.178189994378865e-06, "loss": 1.8767, "step": 311 }, { "epoch": 0.013154288846259249, "grad_norm": 1.479222059249878, "learning_rate": 2.1852164137155705e-06, "loss": 1.5735, "step": 312 }, { "epoch": 0.013196450028458798, "grad_norm": 1.437944769859314, "learning_rate": 2.1922428330522765e-06, "loss": 1.1252, "step": 313 }, { "epoch": 0.013238611210658347, "grad_norm": 1.1903713941574097, "learning_rate": 2.1992692523889826e-06, "loss": 1.1956, "step": 314 }, { "epoch": 0.013280772392857895, "grad_norm": 1.2647218704223633, "learning_rate": 2.2062956717256887e-06, "loss": 0.8725, "step": 315 }, { "epoch": 0.013322933575057445, "grad_norm": 1.2363508939743042, "learning_rate": 2.213322091062395e-06, "loss": 1.3429, "step": 316 }, { "epoch": 0.013365094757256994, "grad_norm": 1.3618125915527344, "learning_rate": 2.2203485103991005e-06, "loss": 0.6183, "step": 317 }, { "epoch": 0.013407255939456542, "grad_norm": 1.5565606355667114, "learning_rate": 2.227374929735807e-06, "loss": 1.7063, "step": 318 }, { "epoch": 0.013449417121656091, "grad_norm": 1.0590002536773682, "learning_rate": 2.2344013490725127e-06, "loss": 1.0657, "step": 319 }, { "epoch": 0.013491578303855641, "grad_norm": 1.1470359563827515, "learning_rate": 2.2414277684092187e-06, "loss": 1.1793, "step": 320 }, { "epoch": 0.013533739486055189, "grad_norm": 1.5479295253753662, "learning_rate": 2.248454187745925e-06, "loss": 0.7241, "step": 321 }, { "epoch": 0.013575900668254738, "grad_norm": 2.200061559677124, "learning_rate": 2.255480607082631e-06, "loss": 0.8698, "step": 322 }, { "epoch": 0.013618061850454288, "grad_norm": 1.1960633993148804, "learning_rate": 2.2625070264193366e-06, "loss": 1.2207, "step": 323 }, { "epoch": 0.013660223032653835, "grad_norm": 1.0216068029403687, "learning_rate": 2.269533445756043e-06, "loss": 0.7692, "step": 324 }, { "epoch": 0.013702384214853385, "grad_norm": 1.4180759191513062, "learning_rate": 2.2765598650927488e-06, "loss": 1.6178, "step": 325 }, { "epoch": 0.013744545397052933, "grad_norm": 1.412074327468872, "learning_rate": 2.283586284429455e-06, "loss": 0.791, "step": 326 }, { "epoch": 0.013786706579252482, "grad_norm": 1.4691176414489746, "learning_rate": 2.290612703766161e-06, "loss": 0.8971, "step": 327 }, { "epoch": 0.013828867761452032, "grad_norm": 1.0744913816452026, "learning_rate": 2.297639123102867e-06, "loss": 1.0815, "step": 328 }, { "epoch": 0.01387102894365158, "grad_norm": 1.1239975690841675, "learning_rate": 2.304665542439573e-06, "loss": 0.9754, "step": 329 }, { "epoch": 0.013913190125851129, "grad_norm": 1.1991620063781738, "learning_rate": 2.311691961776279e-06, "loss": 1.0014, "step": 330 }, { "epoch": 0.013955351308050678, "grad_norm": 1.4413331747055054, "learning_rate": 2.318718381112985e-06, "loss": 1.0532, "step": 331 }, { "epoch": 0.013997512490250226, "grad_norm": 1.3226820230484009, "learning_rate": 2.325744800449691e-06, "loss": 0.7775, "step": 332 }, { "epoch": 0.014039673672449775, "grad_norm": 1.4561959505081177, "learning_rate": 2.332771219786397e-06, "loss": 0.924, "step": 333 }, { "epoch": 0.014081834854649325, "grad_norm": 1.0878831148147583, "learning_rate": 2.3397976391231027e-06, "loss": 0.9755, "step": 334 }, { "epoch": 0.014123996036848873, "grad_norm": 1.2494256496429443, "learning_rate": 2.3468240584598092e-06, "loss": 0.9208, "step": 335 }, { "epoch": 0.014166157219048422, "grad_norm": 2.7417263984680176, "learning_rate": 2.353850477796515e-06, "loss": 1.6341, "step": 336 }, { "epoch": 0.014208318401247972, "grad_norm": 0.9767966866493225, "learning_rate": 2.360876897133221e-06, "loss": 1.098, "step": 337 }, { "epoch": 0.01425047958344752, "grad_norm": 0.9176629781723022, "learning_rate": 2.367903316469927e-06, "loss": 0.7657, "step": 338 }, { "epoch": 0.014292640765647069, "grad_norm": 1.6359131336212158, "learning_rate": 2.374929735806633e-06, "loss": 0.9563, "step": 339 }, { "epoch": 0.014334801947846618, "grad_norm": 1.3233318328857422, "learning_rate": 2.381956155143339e-06, "loss": 0.7735, "step": 340 }, { "epoch": 0.014376963130046166, "grad_norm": 2.7205467224121094, "learning_rate": 2.3889825744800453e-06, "loss": 1.5156, "step": 341 }, { "epoch": 0.014419124312245716, "grad_norm": 1.635703444480896, "learning_rate": 2.396008993816751e-06, "loss": 0.6245, "step": 342 }, { "epoch": 0.014461285494445265, "grad_norm": 1.3880913257598877, "learning_rate": 2.403035413153457e-06, "loss": 0.8228, "step": 343 }, { "epoch": 0.014503446676644813, "grad_norm": 1.1999727487564087, "learning_rate": 2.410061832490163e-06, "loss": 0.7379, "step": 344 }, { "epoch": 0.014545607858844362, "grad_norm": 1.8649147748947144, "learning_rate": 2.417088251826869e-06, "loss": 0.7357, "step": 345 }, { "epoch": 0.01458776904104391, "grad_norm": 1.237351894378662, "learning_rate": 2.4241146711635754e-06, "loss": 0.85, "step": 346 }, { "epoch": 0.01462993022324346, "grad_norm": 1.5110304355621338, "learning_rate": 2.431141090500281e-06, "loss": 0.4658, "step": 347 }, { "epoch": 0.014672091405443009, "grad_norm": 1.8209739923477173, "learning_rate": 2.438167509836987e-06, "loss": 0.8802, "step": 348 }, { "epoch": 0.014714252587642557, "grad_norm": 1.2533769607543945, "learning_rate": 2.4451939291736932e-06, "loss": 0.7023, "step": 349 }, { "epoch": 0.014756413769842106, "grad_norm": 1.0561774969100952, "learning_rate": 2.4522203485103993e-06, "loss": 1.0015, "step": 350 }, { "epoch": 0.014798574952041656, "grad_norm": 1.5839861631393433, "learning_rate": 2.459246767847105e-06, "loss": 1.0145, "step": 351 }, { "epoch": 0.014840736134241203, "grad_norm": 2.4012317657470703, "learning_rate": 2.4662731871838115e-06, "loss": 1.2074, "step": 352 }, { "epoch": 0.014882897316440753, "grad_norm": 1.28380286693573, "learning_rate": 2.473299606520517e-06, "loss": 0.7227, "step": 353 }, { "epoch": 0.014925058498640302, "grad_norm": 2.2158093452453613, "learning_rate": 2.4803260258572232e-06, "loss": 1.5881, "step": 354 }, { "epoch": 0.01496721968083985, "grad_norm": 1.092411994934082, "learning_rate": 2.4873524451939293e-06, "loss": 1.0193, "step": 355 }, { "epoch": 0.0150093808630394, "grad_norm": 1.2933763265609741, "learning_rate": 2.4943788645306354e-06, "loss": 0.9311, "step": 356 }, { "epoch": 0.015051542045238949, "grad_norm": 1.1762230396270752, "learning_rate": 2.501405283867341e-06, "loss": 0.5591, "step": 357 }, { "epoch": 0.015093703227438497, "grad_norm": 0.9711911678314209, "learning_rate": 2.5084317032040476e-06, "loss": 1.4196, "step": 358 }, { "epoch": 0.015135864409638046, "grad_norm": 1.622582197189331, "learning_rate": 2.5154581225407533e-06, "loss": 0.9002, "step": 359 }, { "epoch": 0.015178025591837596, "grad_norm": 1.1095366477966309, "learning_rate": 2.5224845418774594e-06, "loss": 0.7383, "step": 360 }, { "epoch": 0.015220186774037144, "grad_norm": 1.4771978855133057, "learning_rate": 2.5295109612141654e-06, "loss": 0.7943, "step": 361 }, { "epoch": 0.015262347956236693, "grad_norm": 1.5038524866104126, "learning_rate": 2.536537380550871e-06, "loss": 0.8969, "step": 362 }, { "epoch": 0.015304509138436242, "grad_norm": 1.3949494361877441, "learning_rate": 2.5435637998875776e-06, "loss": 0.9001, "step": 363 }, { "epoch": 0.01534667032063579, "grad_norm": 1.3076879978179932, "learning_rate": 2.5505902192242833e-06, "loss": 0.8108, "step": 364 }, { "epoch": 0.01538883150283534, "grad_norm": 1.3277374505996704, "learning_rate": 2.5576166385609894e-06, "loss": 0.9624, "step": 365 }, { "epoch": 0.01543099268503489, "grad_norm": 1.0665467977523804, "learning_rate": 2.5646430578976955e-06, "loss": 0.9491, "step": 366 }, { "epoch": 0.015473153867234437, "grad_norm": 1.7915523052215576, "learning_rate": 2.5716694772344016e-06, "loss": 0.685, "step": 367 }, { "epoch": 0.015515315049433986, "grad_norm": 1.0303057432174683, "learning_rate": 2.5786958965711072e-06, "loss": 0.5708, "step": 368 }, { "epoch": 0.015557476231633534, "grad_norm": 1.7112817764282227, "learning_rate": 2.5857223159078137e-06, "loss": 1.355, "step": 369 }, { "epoch": 0.015599637413833084, "grad_norm": 1.5220093727111816, "learning_rate": 2.5927487352445194e-06, "loss": 0.7216, "step": 370 }, { "epoch": 0.01564179859603263, "grad_norm": 1.377610683441162, "learning_rate": 2.5997751545812255e-06, "loss": 0.6682, "step": 371 }, { "epoch": 0.01568395977823218, "grad_norm": 1.0823999643325806, "learning_rate": 2.6068015739179316e-06, "loss": 0.4137, "step": 372 }, { "epoch": 0.01572612096043173, "grad_norm": 1.4850409030914307, "learning_rate": 2.6138279932546377e-06, "loss": 0.3781, "step": 373 }, { "epoch": 0.01576828214263128, "grad_norm": 1.1681004762649536, "learning_rate": 2.6208544125913438e-06, "loss": 0.7265, "step": 374 }, { "epoch": 0.01581044332483083, "grad_norm": 0.7816203832626343, "learning_rate": 2.62788083192805e-06, "loss": 0.9181, "step": 375 }, { "epoch": 0.01585260450703038, "grad_norm": 1.4013453722000122, "learning_rate": 2.6349072512647555e-06, "loss": 0.8592, "step": 376 }, { "epoch": 0.015894765689229925, "grad_norm": 1.5999376773834229, "learning_rate": 2.6419336706014616e-06, "loss": 1.0355, "step": 377 }, { "epoch": 0.015936926871429474, "grad_norm": 1.5257543325424194, "learning_rate": 2.6489600899381677e-06, "loss": 0.8274, "step": 378 }, { "epoch": 0.015979088053629024, "grad_norm": 0.9036352634429932, "learning_rate": 2.6559865092748734e-06, "loss": 0.4826, "step": 379 }, { "epoch": 0.016021249235828573, "grad_norm": 0.9361503720283508, "learning_rate": 2.66301292861158e-06, "loss": 0.7258, "step": 380 }, { "epoch": 0.016063410418028123, "grad_norm": 1.4573709964752197, "learning_rate": 2.6700393479482855e-06, "loss": 0.8237, "step": 381 }, { "epoch": 0.01610557160022767, "grad_norm": 1.1076159477233887, "learning_rate": 2.6770657672849916e-06, "loss": 1.0509, "step": 382 }, { "epoch": 0.016147732782427218, "grad_norm": 1.0787880420684814, "learning_rate": 2.6840921866216977e-06, "loss": 0.7058, "step": 383 }, { "epoch": 0.016189893964626768, "grad_norm": 3.6805310249328613, "learning_rate": 2.691118605958404e-06, "loss": 1.6302, "step": 384 }, { "epoch": 0.016232055146826317, "grad_norm": 2.124643325805664, "learning_rate": 2.6981450252951095e-06, "loss": 0.5132, "step": 385 }, { "epoch": 0.016274216329025867, "grad_norm": 1.6828583478927612, "learning_rate": 2.705171444631816e-06, "loss": 0.5824, "step": 386 }, { "epoch": 0.016316377511225416, "grad_norm": 3.444493293762207, "learning_rate": 2.7121978639685216e-06, "loss": 1.394, "step": 387 }, { "epoch": 0.016358538693424962, "grad_norm": 1.1182111501693726, "learning_rate": 2.7192242833052277e-06, "loss": 0.9659, "step": 388 }, { "epoch": 0.01640069987562451, "grad_norm": 1.3943028450012207, "learning_rate": 2.726250702641934e-06, "loss": 0.9635, "step": 389 }, { "epoch": 0.01644286105782406, "grad_norm": 1.2187474966049194, "learning_rate": 2.73327712197864e-06, "loss": 0.6402, "step": 390 }, { "epoch": 0.01648502224002361, "grad_norm": 1.1339359283447266, "learning_rate": 2.740303541315346e-06, "loss": 0.7379, "step": 391 }, { "epoch": 0.01652718342222316, "grad_norm": 1.0967634916305542, "learning_rate": 2.747329960652052e-06, "loss": 0.8738, "step": 392 }, { "epoch": 0.01656934460442271, "grad_norm": 5.805525302886963, "learning_rate": 2.7543563799887578e-06, "loss": 1.7817, "step": 393 }, { "epoch": 0.016611505786622256, "grad_norm": 1.523934245109558, "learning_rate": 2.761382799325464e-06, "loss": 0.7824, "step": 394 }, { "epoch": 0.016653666968821805, "grad_norm": 1.9693888425827026, "learning_rate": 2.76840921866217e-06, "loss": 0.75, "step": 395 }, { "epoch": 0.016695828151021355, "grad_norm": 2.2070672512054443, "learning_rate": 2.7754356379988756e-06, "loss": 0.7688, "step": 396 }, { "epoch": 0.016737989333220904, "grad_norm": 0.9869478940963745, "learning_rate": 2.782462057335582e-06, "loss": 0.7143, "step": 397 }, { "epoch": 0.016780150515420453, "grad_norm": 1.1068414449691772, "learning_rate": 2.7894884766722878e-06, "loss": 0.746, "step": 398 }, { "epoch": 0.016822311697620003, "grad_norm": 0.9384260177612305, "learning_rate": 2.796514896008994e-06, "loss": 0.7144, "step": 399 }, { "epoch": 0.01686447287981955, "grad_norm": 0.8328140377998352, "learning_rate": 2.8035413153457e-06, "loss": 0.7496, "step": 400 }, { "epoch": 0.0169066340620191, "grad_norm": 1.7657179832458496, "learning_rate": 2.810567734682406e-06, "loss": 1.038, "step": 401 }, { "epoch": 0.016948795244218648, "grad_norm": 1.3386183977127075, "learning_rate": 2.8175941540191117e-06, "loss": 0.7413, "step": 402 }, { "epoch": 0.016990956426418197, "grad_norm": 1.4442965984344482, "learning_rate": 2.8246205733558182e-06, "loss": 1.3432, "step": 403 }, { "epoch": 0.017033117608617747, "grad_norm": 1.1995635032653809, "learning_rate": 2.831646992692524e-06, "loss": 0.4421, "step": 404 }, { "epoch": 0.017075278790817293, "grad_norm": 1.553778052330017, "learning_rate": 2.8386734120292304e-06, "loss": 0.9864, "step": 405 }, { "epoch": 0.017117439973016842, "grad_norm": 1.1403580904006958, "learning_rate": 2.845699831365936e-06, "loss": 1.0, "step": 406 }, { "epoch": 0.017159601155216392, "grad_norm": 0.6638301610946655, "learning_rate": 2.852726250702642e-06, "loss": 0.3518, "step": 407 }, { "epoch": 0.01720176233741594, "grad_norm": 1.2993419170379639, "learning_rate": 2.8597526700393482e-06, "loss": 0.9828, "step": 408 }, { "epoch": 0.01724392351961549, "grad_norm": 1.4930028915405273, "learning_rate": 2.866779089376054e-06, "loss": 0.8078, "step": 409 }, { "epoch": 0.01728608470181504, "grad_norm": 1.1494414806365967, "learning_rate": 2.87380550871276e-06, "loss": 0.602, "step": 410 }, { "epoch": 0.017328245884014586, "grad_norm": 3.714118242263794, "learning_rate": 2.880831928049466e-06, "loss": 1.5593, "step": 411 }, { "epoch": 0.017370407066214136, "grad_norm": 1.1589826345443726, "learning_rate": 2.887858347386172e-06, "loss": 0.5917, "step": 412 }, { "epoch": 0.017412568248413685, "grad_norm": 1.1813617944717407, "learning_rate": 2.894884766722878e-06, "loss": 0.8764, "step": 413 }, { "epoch": 0.017454729430613235, "grad_norm": 1.771054983139038, "learning_rate": 2.9019111860595844e-06, "loss": 0.8222, "step": 414 }, { "epoch": 0.017496890612812784, "grad_norm": 0.8509282469749451, "learning_rate": 2.90893760539629e-06, "loss": 0.6918, "step": 415 }, { "epoch": 0.017539051795012334, "grad_norm": 1.3561125993728638, "learning_rate": 2.915964024732996e-06, "loss": 0.9417, "step": 416 }, { "epoch": 0.01758121297721188, "grad_norm": 1.1123842000961304, "learning_rate": 2.922990444069702e-06, "loss": 0.6375, "step": 417 }, { "epoch": 0.01762337415941143, "grad_norm": 1.3375004529953003, "learning_rate": 2.9300168634064083e-06, "loss": 0.5833, "step": 418 }, { "epoch": 0.01766553534161098, "grad_norm": 1.9937835931777954, "learning_rate": 2.937043282743114e-06, "loss": 1.193, "step": 419 }, { "epoch": 0.017707696523810528, "grad_norm": 1.361673355102539, "learning_rate": 2.9440697020798205e-06, "loss": 0.665, "step": 420 }, { "epoch": 0.017749857706010078, "grad_norm": 1.1907340288162231, "learning_rate": 2.951096121416526e-06, "loss": 0.4808, "step": 421 }, { "epoch": 0.017792018888209627, "grad_norm": 1.9677084684371948, "learning_rate": 2.9581225407532326e-06, "loss": 1.5686, "step": 422 }, { "epoch": 0.017834180070409173, "grad_norm": 1.4330767393112183, "learning_rate": 2.9651489600899383e-06, "loss": 0.6407, "step": 423 }, { "epoch": 0.017876341252608723, "grad_norm": 1.0757044553756714, "learning_rate": 2.9721753794266444e-06, "loss": 0.6189, "step": 424 }, { "epoch": 0.017918502434808272, "grad_norm": 1.1422349214553833, "learning_rate": 2.9792017987633505e-06, "loss": 0.5476, "step": 425 }, { "epoch": 0.01796066361700782, "grad_norm": 2.0129024982452393, "learning_rate": 2.986228218100056e-06, "loss": 0.749, "step": 426 }, { "epoch": 0.01800282479920737, "grad_norm": 1.2586308717727661, "learning_rate": 2.9932546374367622e-06, "loss": 0.7196, "step": 427 }, { "epoch": 0.018044985981406917, "grad_norm": 1.3948662281036377, "learning_rate": 3.0002810567734683e-06, "loss": 0.6335, "step": 428 }, { "epoch": 0.018087147163606467, "grad_norm": 1.136612057685852, "learning_rate": 3.0073074761101744e-06, "loss": 0.4553, "step": 429 }, { "epoch": 0.018129308345806016, "grad_norm": 1.8347485065460205, "learning_rate": 3.01433389544688e-06, "loss": 0.8833, "step": 430 }, { "epoch": 0.018171469528005565, "grad_norm": 1.322221040725708, "learning_rate": 3.0213603147835866e-06, "loss": 1.4585, "step": 431 }, { "epoch": 0.018213630710205115, "grad_norm": 1.490116000175476, "learning_rate": 3.0283867341202923e-06, "loss": 0.5038, "step": 432 }, { "epoch": 0.018255791892404664, "grad_norm": 1.238089680671692, "learning_rate": 3.0354131534569984e-06, "loss": 1.5344, "step": 433 }, { "epoch": 0.01829795307460421, "grad_norm": 1.5193126201629639, "learning_rate": 3.0424395727937044e-06, "loss": 0.4887, "step": 434 }, { "epoch": 0.01834011425680376, "grad_norm": 1.3485597372055054, "learning_rate": 3.0494659921304105e-06, "loss": 0.5793, "step": 435 }, { "epoch": 0.01838227543900331, "grad_norm": 1.3789498805999756, "learning_rate": 3.0564924114671166e-06, "loss": 0.9235, "step": 436 }, { "epoch": 0.01842443662120286, "grad_norm": 1.5377343893051147, "learning_rate": 3.0635188308038227e-06, "loss": 0.5732, "step": 437 }, { "epoch": 0.01846659780340241, "grad_norm": 1.0283747911453247, "learning_rate": 3.0705452501405284e-06, "loss": 0.4105, "step": 438 }, { "epoch": 0.018508758985601958, "grad_norm": 1.0047868490219116, "learning_rate": 3.077571669477235e-06, "loss": 0.4447, "step": 439 }, { "epoch": 0.018550920167801504, "grad_norm": 1.6522831916809082, "learning_rate": 3.0845980888139406e-06, "loss": 0.7001, "step": 440 }, { "epoch": 0.018593081350001053, "grad_norm": 1.2638013362884521, "learning_rate": 3.0916245081506466e-06, "loss": 0.7661, "step": 441 }, { "epoch": 0.018635242532200603, "grad_norm": 1.0334484577178955, "learning_rate": 3.0986509274873527e-06, "loss": 1.2686, "step": 442 }, { "epoch": 0.018677403714400152, "grad_norm": 1.303143858909607, "learning_rate": 3.1056773468240584e-06, "loss": 1.0084, "step": 443 }, { "epoch": 0.018719564896599702, "grad_norm": 1.289279818534851, "learning_rate": 3.1127037661607645e-06, "loss": 0.9002, "step": 444 }, { "epoch": 0.01876172607879925, "grad_norm": 1.1880784034729004, "learning_rate": 3.1197301854974706e-06, "loss": 0.4082, "step": 445 }, { "epoch": 0.018803887260998797, "grad_norm": 1.906272053718567, "learning_rate": 3.1267566048341767e-06, "loss": 0.6278, "step": 446 }, { "epoch": 0.018846048443198347, "grad_norm": 1.5854572057724, "learning_rate": 3.1337830241708823e-06, "loss": 0.4943, "step": 447 }, { "epoch": 0.018888209625397896, "grad_norm": 1.414692997932434, "learning_rate": 3.1408094435075884e-06, "loss": 0.4576, "step": 448 }, { "epoch": 0.018930370807597446, "grad_norm": 1.167584776878357, "learning_rate": 3.147835862844295e-06, "loss": 0.9143, "step": 449 }, { "epoch": 0.018972531989796995, "grad_norm": 1.2257081270217896, "learning_rate": 3.154862282181001e-06, "loss": 0.6279, "step": 450 }, { "epoch": 0.01901469317199654, "grad_norm": 1.4712598323822021, "learning_rate": 3.1618887015177067e-06, "loss": 0.4519, "step": 451 }, { "epoch": 0.01905685435419609, "grad_norm": 1.1191139221191406, "learning_rate": 3.1689151208544128e-06, "loss": 1.4539, "step": 452 }, { "epoch": 0.01909901553639564, "grad_norm": 1.1994318962097168, "learning_rate": 3.1759415401911184e-06, "loss": 0.4938, "step": 453 }, { "epoch": 0.01914117671859519, "grad_norm": 1.7503246068954468, "learning_rate": 3.1829679595278245e-06, "loss": 0.7281, "step": 454 }, { "epoch": 0.01918333790079474, "grad_norm": 2.021331548690796, "learning_rate": 3.189994378864531e-06, "loss": 1.3088, "step": 455 }, { "epoch": 0.01922549908299429, "grad_norm": 1.0639694929122925, "learning_rate": 3.1970207982012367e-06, "loss": 0.4223, "step": 456 }, { "epoch": 0.019267660265193835, "grad_norm": 1.2248399257659912, "learning_rate": 3.204047217537943e-06, "loss": 1.475, "step": 457 }, { "epoch": 0.019309821447393384, "grad_norm": 1.7997770309448242, "learning_rate": 3.211073636874649e-06, "loss": 0.5841, "step": 458 }, { "epoch": 0.019351982629592934, "grad_norm": 1.872942566871643, "learning_rate": 3.2181000562113546e-06, "loss": 0.6062, "step": 459 }, { "epoch": 0.019394143811792483, "grad_norm": 1.8351565599441528, "learning_rate": 3.225126475548061e-06, "loss": 0.5554, "step": 460 }, { "epoch": 0.019436304993992033, "grad_norm": 1.4064991474151611, "learning_rate": 3.232152894884767e-06, "loss": 0.4806, "step": 461 }, { "epoch": 0.019478466176191582, "grad_norm": 1.2027555704116821, "learning_rate": 3.239179314221473e-06, "loss": 0.7349, "step": 462 }, { "epoch": 0.019520627358391128, "grad_norm": 1.1064081192016602, "learning_rate": 3.246205733558179e-06, "loss": 0.8672, "step": 463 }, { "epoch": 0.019562788540590677, "grad_norm": 0.9629604816436768, "learning_rate": 3.2532321528948846e-06, "loss": 0.8886, "step": 464 }, { "epoch": 0.019604949722790227, "grad_norm": 2.2624120712280273, "learning_rate": 3.2602585722315907e-06, "loss": 0.4739, "step": 465 }, { "epoch": 0.019647110904989776, "grad_norm": 1.2926243543624878, "learning_rate": 3.267284991568297e-06, "loss": 0.6745, "step": 466 }, { "epoch": 0.019689272087189326, "grad_norm": 1.5190688371658325, "learning_rate": 3.2743114109050033e-06, "loss": 0.7159, "step": 467 }, { "epoch": 0.019731433269388872, "grad_norm": 1.3555351495742798, "learning_rate": 3.281337830241709e-06, "loss": 1.2444, "step": 468 }, { "epoch": 0.01977359445158842, "grad_norm": 3.215022087097168, "learning_rate": 3.288364249578415e-06, "loss": 0.9382, "step": 469 }, { "epoch": 0.01981575563378797, "grad_norm": 1.2831155061721802, "learning_rate": 3.2953906689151207e-06, "loss": 0.7962, "step": 470 }, { "epoch": 0.01985791681598752, "grad_norm": 0.9507866501808167, "learning_rate": 3.3024170882518268e-06, "loss": 0.4473, "step": 471 }, { "epoch": 0.01990007799818707, "grad_norm": 1.2401431798934937, "learning_rate": 3.3094435075885333e-06, "loss": 0.8937, "step": 472 }, { "epoch": 0.01994223918038662, "grad_norm": 1.5420396327972412, "learning_rate": 3.316469926925239e-06, "loss": 0.4474, "step": 473 }, { "epoch": 0.019984400362586165, "grad_norm": 1.3603413105010986, "learning_rate": 3.323496346261945e-06, "loss": 1.3777, "step": 474 }, { "epoch": 0.020026561544785715, "grad_norm": 1.4202021360397339, "learning_rate": 3.330522765598651e-06, "loss": 0.5702, "step": 475 }, { "epoch": 0.020068722726985264, "grad_norm": 1.0666612386703491, "learning_rate": 3.337549184935357e-06, "loss": 0.3721, "step": 476 }, { "epoch": 0.020110883909184814, "grad_norm": 1.3715894222259521, "learning_rate": 3.3445756042720633e-06, "loss": 0.4346, "step": 477 }, { "epoch": 0.020153045091384363, "grad_norm": 1.2678554058074951, "learning_rate": 3.3516020236087694e-06, "loss": 1.428, "step": 478 }, { "epoch": 0.020195206273583913, "grad_norm": 1.1936759948730469, "learning_rate": 3.358628442945475e-06, "loss": 0.3322, "step": 479 }, { "epoch": 0.02023736745578346, "grad_norm": 1.2128477096557617, "learning_rate": 3.365654862282181e-06, "loss": 0.6642, "step": 480 }, { "epoch": 0.020279528637983008, "grad_norm": 1.8041155338287354, "learning_rate": 3.372681281618887e-06, "loss": 0.5863, "step": 481 }, { "epoch": 0.020321689820182558, "grad_norm": 1.4484094381332397, "learning_rate": 3.379707700955593e-06, "loss": 0.9547, "step": 482 }, { "epoch": 0.020363851002382107, "grad_norm": 1.0908854007720947, "learning_rate": 3.3867341202922994e-06, "loss": 1.0588, "step": 483 }, { "epoch": 0.020406012184581657, "grad_norm": 3.542285203933716, "learning_rate": 3.3937605396290055e-06, "loss": 1.4847, "step": 484 }, { "epoch": 0.020448173366781206, "grad_norm": 1.1464065313339233, "learning_rate": 3.400786958965711e-06, "loss": 0.4529, "step": 485 }, { "epoch": 0.020490334548980752, "grad_norm": 1.285136342048645, "learning_rate": 3.4078133783024173e-06, "loss": 0.4578, "step": 486 }, { "epoch": 0.0205324957311803, "grad_norm": 1.3954039812088013, "learning_rate": 3.414839797639123e-06, "loss": 0.3642, "step": 487 }, { "epoch": 0.02057465691337985, "grad_norm": 4.998807907104492, "learning_rate": 3.4218662169758295e-06, "loss": 1.2869, "step": 488 }, { "epoch": 0.0206168180955794, "grad_norm": 1.3401496410369873, "learning_rate": 3.4288926363125355e-06, "loss": 0.3272, "step": 489 }, { "epoch": 0.02065897927777895, "grad_norm": 1.358222246170044, "learning_rate": 3.435919055649241e-06, "loss": 0.3858, "step": 490 }, { "epoch": 0.020701140459978496, "grad_norm": 2.051882028579712, "learning_rate": 3.4429454749859473e-06, "loss": 1.2064, "step": 491 }, { "epoch": 0.020743301642178046, "grad_norm": 1.5642008781433105, "learning_rate": 3.4499718943226534e-06, "loss": 1.1857, "step": 492 }, { "epoch": 0.020785462824377595, "grad_norm": 1.9305403232574463, "learning_rate": 3.456998313659359e-06, "loss": 0.446, "step": 493 }, { "epoch": 0.020827624006577145, "grad_norm": 1.789806842803955, "learning_rate": 3.4640247329960656e-06, "loss": 0.8889, "step": 494 }, { "epoch": 0.020869785188776694, "grad_norm": 1.5596835613250732, "learning_rate": 3.4710511523327717e-06, "loss": 1.4669, "step": 495 }, { "epoch": 0.020911946370976243, "grad_norm": 1.4839788675308228, "learning_rate": 3.4780775716694773e-06, "loss": 1.3254, "step": 496 }, { "epoch": 0.02095410755317579, "grad_norm": 1.255504846572876, "learning_rate": 3.4851039910061834e-06, "loss": 0.5054, "step": 497 }, { "epoch": 0.02099626873537534, "grad_norm": 1.3478630781173706, "learning_rate": 3.492130410342889e-06, "loss": 0.3766, "step": 498 }, { "epoch": 0.02103842991757489, "grad_norm": 1.7024043798446655, "learning_rate": 3.499156829679595e-06, "loss": 0.53, "step": 499 }, { "epoch": 0.021080591099774438, "grad_norm": 2.1321537494659424, "learning_rate": 3.5061832490163017e-06, "loss": 1.4679, "step": 500 }, { "epoch": 0.021122752281973987, "grad_norm": 1.2200685739517212, "learning_rate": 3.5132096683530078e-06, "loss": 0.2983, "step": 501 }, { "epoch": 0.021164913464173537, "grad_norm": 1.0821027755737305, "learning_rate": 3.5202360876897134e-06, "loss": 0.3423, "step": 502 }, { "epoch": 0.021207074646373083, "grad_norm": 1.2871267795562744, "learning_rate": 3.5272625070264195e-06, "loss": 0.4245, "step": 503 }, { "epoch": 0.021249235828572632, "grad_norm": 1.0613694190979004, "learning_rate": 3.534288926363125e-06, "loss": 0.935, "step": 504 }, { "epoch": 0.021291397010772182, "grad_norm": 2.915148973464966, "learning_rate": 3.5413153456998317e-06, "loss": 1.0236, "step": 505 }, { "epoch": 0.02133355819297173, "grad_norm": 1.6012650728225708, "learning_rate": 3.548341765036538e-06, "loss": 0.3493, "step": 506 }, { "epoch": 0.02137571937517128, "grad_norm": 1.495587944984436, "learning_rate": 3.5553681843732435e-06, "loss": 0.3448, "step": 507 }, { "epoch": 0.02141788055737083, "grad_norm": 1.930396556854248, "learning_rate": 3.5623946037099495e-06, "loss": 0.4783, "step": 508 }, { "epoch": 0.021460041739570376, "grad_norm": 1.7716583013534546, "learning_rate": 3.5694210230466556e-06, "loss": 0.6779, "step": 509 }, { "epoch": 0.021502202921769926, "grad_norm": 1.7069264650344849, "learning_rate": 3.5764474423833613e-06, "loss": 0.604, "step": 510 }, { "epoch": 0.021544364103969475, "grad_norm": 1.5648213624954224, "learning_rate": 3.583473861720068e-06, "loss": 0.553, "step": 511 }, { "epoch": 0.021586525286169025, "grad_norm": 1.413510799407959, "learning_rate": 3.590500281056774e-06, "loss": 0.3315, "step": 512 }, { "epoch": 0.021628686468368574, "grad_norm": 1.0293033123016357, "learning_rate": 3.5975267003934796e-06, "loss": 0.3985, "step": 513 }, { "epoch": 0.02167084765056812, "grad_norm": 1.3213144540786743, "learning_rate": 3.6045531197301857e-06, "loss": 0.5447, "step": 514 }, { "epoch": 0.02171300883276767, "grad_norm": 1.1708276271820068, "learning_rate": 3.6115795390668913e-06, "loss": 0.7151, "step": 515 }, { "epoch": 0.02175517001496722, "grad_norm": 4.622429370880127, "learning_rate": 3.6186059584035974e-06, "loss": 1.5565, "step": 516 }, { "epoch": 0.02179733119716677, "grad_norm": 1.1917428970336914, "learning_rate": 3.625632377740304e-06, "loss": 0.5417, "step": 517 }, { "epoch": 0.021839492379366318, "grad_norm": 2.073526382446289, "learning_rate": 3.63265879707701e-06, "loss": 0.8585, "step": 518 }, { "epoch": 0.021881653561565868, "grad_norm": 1.4803080558776855, "learning_rate": 3.6396852164137157e-06, "loss": 1.3608, "step": 519 }, { "epoch": 0.021923814743765414, "grad_norm": 0.9618265628814697, "learning_rate": 3.6467116357504218e-06, "loss": 0.4213, "step": 520 }, { "epoch": 0.021965975925964963, "grad_norm": 1.2747958898544312, "learning_rate": 3.6537380550871274e-06, "loss": 0.7048, "step": 521 }, { "epoch": 0.022008137108164513, "grad_norm": 1.289934754371643, "learning_rate": 3.660764474423834e-06, "loss": 0.28, "step": 522 }, { "epoch": 0.022050298290364062, "grad_norm": 1.7034169435501099, "learning_rate": 3.66779089376054e-06, "loss": 0.677, "step": 523 }, { "epoch": 0.02209245947256361, "grad_norm": 1.3161852359771729, "learning_rate": 3.6748173130972457e-06, "loss": 0.4657, "step": 524 }, { "epoch": 0.02213462065476316, "grad_norm": 1.4165502786636353, "learning_rate": 3.681843732433952e-06, "loss": 1.4247, "step": 525 }, { "epoch": 0.022176781836962707, "grad_norm": 1.4379388093948364, "learning_rate": 3.688870151770658e-06, "loss": 0.3762, "step": 526 }, { "epoch": 0.022218943019162257, "grad_norm": 1.0862252712249756, "learning_rate": 3.6958965711073635e-06, "loss": 0.9073, "step": 527 }, { "epoch": 0.022261104201361806, "grad_norm": 1.3146408796310425, "learning_rate": 3.70292299044407e-06, "loss": 0.4608, "step": 528 }, { "epoch": 0.022303265383561355, "grad_norm": 1.3607866764068604, "learning_rate": 3.709949409780776e-06, "loss": 0.5028, "step": 529 }, { "epoch": 0.022345426565760905, "grad_norm": 0.9588279128074646, "learning_rate": 3.716975829117482e-06, "loss": 0.4883, "step": 530 }, { "epoch": 0.022387587747960454, "grad_norm": 2.086728811264038, "learning_rate": 3.724002248454188e-06, "loss": 0.7962, "step": 531 }, { "epoch": 0.02242974893016, "grad_norm": 2.060075521469116, "learning_rate": 3.7310286677908936e-06, "loss": 0.4586, "step": 532 }, { "epoch": 0.02247191011235955, "grad_norm": 1.4732993841171265, "learning_rate": 3.7380550871275997e-06, "loss": 1.4861, "step": 533 }, { "epoch": 0.0225140712945591, "grad_norm": 1.6340360641479492, "learning_rate": 3.745081506464306e-06, "loss": 0.5489, "step": 534 }, { "epoch": 0.02255623247675865, "grad_norm": 1.5460063219070435, "learning_rate": 3.7521079258010123e-06, "loss": 0.4668, "step": 535 }, { "epoch": 0.0225983936589582, "grad_norm": 1.6337792873382568, "learning_rate": 3.759134345137718e-06, "loss": 1.3713, "step": 536 }, { "epoch": 0.022640554841157744, "grad_norm": 1.1203831434249878, "learning_rate": 3.766160764474424e-06, "loss": 0.5089, "step": 537 }, { "epoch": 0.022682716023357294, "grad_norm": 1.7591861486434937, "learning_rate": 3.7731871838111297e-06, "loss": 0.4976, "step": 538 }, { "epoch": 0.022724877205556843, "grad_norm": 1.0981132984161377, "learning_rate": 3.780213603147836e-06, "loss": 0.8647, "step": 539 }, { "epoch": 0.022767038387756393, "grad_norm": 1.66310715675354, "learning_rate": 3.7872400224845423e-06, "loss": 0.5178, "step": 540 }, { "epoch": 0.022809199569955942, "grad_norm": 1.3163211345672607, "learning_rate": 3.794266441821248e-06, "loss": 0.7659, "step": 541 }, { "epoch": 0.022851360752155492, "grad_norm": 2.4486124515533447, "learning_rate": 3.801292861157954e-06, "loss": 0.9911, "step": 542 }, { "epoch": 0.022893521934355038, "grad_norm": 1.7964993715286255, "learning_rate": 3.8083192804946597e-06, "loss": 1.4343, "step": 543 }, { "epoch": 0.022935683116554587, "grad_norm": 1.3804799318313599, "learning_rate": 3.815345699831366e-06, "loss": 0.9982, "step": 544 }, { "epoch": 0.022977844298754137, "grad_norm": 1.7669315338134766, "learning_rate": 3.822372119168073e-06, "loss": 1.4211, "step": 545 }, { "epoch": 0.023020005480953686, "grad_norm": 1.5056713819503784, "learning_rate": 3.829398538504778e-06, "loss": 0.371, "step": 546 }, { "epoch": 0.023062166663153236, "grad_norm": 0.6709935069084167, "learning_rate": 3.836424957841484e-06, "loss": 0.2379, "step": 547 }, { "epoch": 0.023104327845352785, "grad_norm": 1.2807109355926514, "learning_rate": 3.84345137717819e-06, "loss": 0.7564, "step": 548 }, { "epoch": 0.02314648902755233, "grad_norm": 2.6952168941497803, "learning_rate": 3.850477796514896e-06, "loss": 1.2746, "step": 549 }, { "epoch": 0.02318865020975188, "grad_norm": 1.1244404315948486, "learning_rate": 3.857504215851602e-06, "loss": 0.9998, "step": 550 }, { "epoch": 0.02323081139195143, "grad_norm": 1.5839787721633911, "learning_rate": 3.864530635188308e-06, "loss": 0.9707, "step": 551 }, { "epoch": 0.02327297257415098, "grad_norm": 1.429267406463623, "learning_rate": 3.8715570545250145e-06, "loss": 0.9871, "step": 552 }, { "epoch": 0.02331513375635053, "grad_norm": 1.033122181892395, "learning_rate": 3.878583473861721e-06, "loss": 1.0721, "step": 553 }, { "epoch": 0.02335729493855008, "grad_norm": 1.4751434326171875, "learning_rate": 3.885609893198426e-06, "loss": 0.5088, "step": 554 }, { "epoch": 0.023399456120749625, "grad_norm": 1.0616769790649414, "learning_rate": 3.892636312535132e-06, "loss": 0.398, "step": 555 }, { "epoch": 0.023441617302949174, "grad_norm": 1.1346887350082397, "learning_rate": 3.899662731871839e-06, "loss": 0.4117, "step": 556 }, { "epoch": 0.023483778485148724, "grad_norm": 1.4016139507293701, "learning_rate": 3.906689151208544e-06, "loss": 0.834, "step": 557 }, { "epoch": 0.023525939667348273, "grad_norm": 1.3796902894973755, "learning_rate": 3.91371557054525e-06, "loss": 0.6508, "step": 558 }, { "epoch": 0.023568100849547823, "grad_norm": 1.9771815538406372, "learning_rate": 3.920741989881956e-06, "loss": 0.358, "step": 559 }, { "epoch": 0.02361026203174737, "grad_norm": 1.1195144653320312, "learning_rate": 3.927768409218662e-06, "loss": 0.7229, "step": 560 }, { "epoch": 0.023652423213946918, "grad_norm": 1.2300177812576294, "learning_rate": 3.9347948285553685e-06, "loss": 0.45, "step": 561 }, { "epoch": 0.023694584396146467, "grad_norm": 1.1353102922439575, "learning_rate": 3.9418212478920746e-06, "loss": 0.3332, "step": 562 }, { "epoch": 0.023736745578346017, "grad_norm": 2.1426470279693604, "learning_rate": 3.948847667228781e-06, "loss": 0.5338, "step": 563 }, { "epoch": 0.023778906760545566, "grad_norm": 1.338463306427002, "learning_rate": 3.955874086565487e-06, "loss": 0.7465, "step": 564 }, { "epoch": 0.023821067942745116, "grad_norm": 1.393051028251648, "learning_rate": 3.962900505902192e-06, "loss": 0.4946, "step": 565 }, { "epoch": 0.023863229124944662, "grad_norm": 0.9457556009292603, "learning_rate": 3.969926925238898e-06, "loss": 1.3141, "step": 566 }, { "epoch": 0.02390539030714421, "grad_norm": 1.5232703685760498, "learning_rate": 3.976953344575605e-06, "loss": 0.3322, "step": 567 }, { "epoch": 0.02394755148934376, "grad_norm": 1.094777226448059, "learning_rate": 3.983979763912311e-06, "loss": 0.8575, "step": 568 }, { "epoch": 0.02398971267154331, "grad_norm": 1.0862489938735962, "learning_rate": 3.991006183249016e-06, "loss": 0.4167, "step": 569 }, { "epoch": 0.02403187385374286, "grad_norm": 0.9702075123786926, "learning_rate": 3.998032602585722e-06, "loss": 0.34, "step": 570 }, { "epoch": 0.02407403503594241, "grad_norm": 0.890038251876831, "learning_rate": 4.0050590219224285e-06, "loss": 0.6693, "step": 571 }, { "epoch": 0.024116196218141955, "grad_norm": 2.202669382095337, "learning_rate": 4.012085441259135e-06, "loss": 0.6149, "step": 572 }, { "epoch": 0.024158357400341505, "grad_norm": 0.7001133561134338, "learning_rate": 4.019111860595841e-06, "loss": 0.6249, "step": 573 }, { "epoch": 0.024200518582541054, "grad_norm": 1.3851453065872192, "learning_rate": 4.026138279932547e-06, "loss": 0.9184, "step": 574 }, { "epoch": 0.024242679764740604, "grad_norm": 3.481912851333618, "learning_rate": 4.033164699269253e-06, "loss": 0.972, "step": 575 }, { "epoch": 0.024284840946940153, "grad_norm": 1.2147611379623413, "learning_rate": 4.040191118605958e-06, "loss": 0.6226, "step": 576 }, { "epoch": 0.024327002129139703, "grad_norm": 1.1452105045318604, "learning_rate": 4.047217537942664e-06, "loss": 0.393, "step": 577 }, { "epoch": 0.02436916331133925, "grad_norm": 2.662726640701294, "learning_rate": 4.05424395727937e-06, "loss": 0.3113, "step": 578 }, { "epoch": 0.024411324493538798, "grad_norm": 1.3274255990982056, "learning_rate": 4.061270376616077e-06, "loss": 0.7432, "step": 579 }, { "epoch": 0.024453485675738348, "grad_norm": 1.5222967863082886, "learning_rate": 4.0682967959527825e-06, "loss": 0.5166, "step": 580 }, { "epoch": 0.024495646857937897, "grad_norm": 2.5178518295288086, "learning_rate": 4.0753232152894886e-06, "loss": 0.8487, "step": 581 }, { "epoch": 0.024537808040137447, "grad_norm": 1.5287705659866333, "learning_rate": 4.082349634626195e-06, "loss": 0.5937, "step": 582 }, { "epoch": 0.024579969222336993, "grad_norm": 1.357215166091919, "learning_rate": 4.089376053962901e-06, "loss": 0.486, "step": 583 }, { "epoch": 0.024622130404536542, "grad_norm": 1.8128976821899414, "learning_rate": 4.096402473299607e-06, "loss": 0.5169, "step": 584 }, { "epoch": 0.02466429158673609, "grad_norm": 1.4323080778121948, "learning_rate": 4.103428892636313e-06, "loss": 0.3418, "step": 585 }, { "epoch": 0.02470645276893564, "grad_norm": 1.7128921747207642, "learning_rate": 4.110455311973019e-06, "loss": 0.7136, "step": 586 }, { "epoch": 0.02474861395113519, "grad_norm": 1.2227338552474976, "learning_rate": 4.117481731309725e-06, "loss": 1.4138, "step": 587 }, { "epoch": 0.02479077513333474, "grad_norm": 1.236534833908081, "learning_rate": 4.12450815064643e-06, "loss": 0.665, "step": 588 }, { "epoch": 0.024832936315534286, "grad_norm": 0.8387237787246704, "learning_rate": 4.131534569983136e-06, "loss": 0.6775, "step": 589 }, { "epoch": 0.024875097497733836, "grad_norm": 1.8365809917449951, "learning_rate": 4.138560989319843e-06, "loss": 0.4779, "step": 590 }, { "epoch": 0.024917258679933385, "grad_norm": 1.2188957929611206, "learning_rate": 4.145587408656549e-06, "loss": 0.7825, "step": 591 }, { "epoch": 0.024959419862132935, "grad_norm": 1.9012184143066406, "learning_rate": 4.152613827993255e-06, "loss": 0.8468, "step": 592 }, { "epoch": 0.025001581044332484, "grad_norm": 1.4180734157562256, "learning_rate": 4.159640247329961e-06, "loss": 0.503, "step": 593 }, { "epoch": 0.025043742226532033, "grad_norm": 1.1028966903686523, "learning_rate": 4.166666666666667e-06, "loss": 0.3512, "step": 594 }, { "epoch": 0.02508590340873158, "grad_norm": 1.2560821771621704, "learning_rate": 4.173693086003373e-06, "loss": 0.3795, "step": 595 }, { "epoch": 0.02512806459093113, "grad_norm": 1.0627657175064087, "learning_rate": 4.180719505340079e-06, "loss": 0.6699, "step": 596 }, { "epoch": 0.02517022577313068, "grad_norm": 1.7449195384979248, "learning_rate": 4.187745924676785e-06, "loss": 1.4211, "step": 597 }, { "epoch": 0.025212386955330228, "grad_norm": 1.3074584007263184, "learning_rate": 4.194772344013491e-06, "loss": 0.43, "step": 598 }, { "epoch": 0.025254548137529777, "grad_norm": 1.5814141035079956, "learning_rate": 4.2017987633501965e-06, "loss": 0.5865, "step": 599 }, { "epoch": 0.025296709319729327, "grad_norm": 2.108703136444092, "learning_rate": 4.2088251826869026e-06, "loss": 0.6084, "step": 600 }, { "epoch": 0.025338870501928873, "grad_norm": 1.3351589441299438, "learning_rate": 4.2158516020236095e-06, "loss": 0.5879, "step": 601 }, { "epoch": 0.025381031684128422, "grad_norm": 1.1589059829711914, "learning_rate": 4.222878021360315e-06, "loss": 0.3826, "step": 602 }, { "epoch": 0.025423192866327972, "grad_norm": 1.3212857246398926, "learning_rate": 4.229904440697021e-06, "loss": 0.4598, "step": 603 }, { "epoch": 0.02546535404852752, "grad_norm": 3.404123544692993, "learning_rate": 4.236930860033727e-06, "loss": 0.5111, "step": 604 }, { "epoch": 0.02550751523072707, "grad_norm": 1.33980393409729, "learning_rate": 4.243957279370433e-06, "loss": 0.4131, "step": 605 }, { "epoch": 0.025549676412926617, "grad_norm": 1.4270391464233398, "learning_rate": 4.250983698707139e-06, "loss": 0.4536, "step": 606 }, { "epoch": 0.025591837595126166, "grad_norm": 1.0157338380813599, "learning_rate": 4.258010118043845e-06, "loss": 0.2965, "step": 607 }, { "epoch": 0.025633998777325716, "grad_norm": 3.3252408504486084, "learning_rate": 4.265036537380551e-06, "loss": 0.9352, "step": 608 }, { "epoch": 0.025676159959525265, "grad_norm": 1.8139435052871704, "learning_rate": 4.272062956717257e-06, "loss": 0.4086, "step": 609 }, { "epoch": 0.025718321141724815, "grad_norm": 1.5767617225646973, "learning_rate": 4.279089376053963e-06, "loss": 0.7402, "step": 610 }, { "epoch": 0.025760482323924364, "grad_norm": 1.1882880926132202, "learning_rate": 4.286115795390669e-06, "loss": 0.4698, "step": 611 }, { "epoch": 0.02580264350612391, "grad_norm": 1.1259174346923828, "learning_rate": 4.293142214727376e-06, "loss": 0.6439, "step": 612 }, { "epoch": 0.02584480468832346, "grad_norm": 1.1298385858535767, "learning_rate": 4.300168634064082e-06, "loss": 0.8824, "step": 613 }, { "epoch": 0.02588696587052301, "grad_norm": 1.1287230253219604, "learning_rate": 4.307195053400787e-06, "loss": 0.3032, "step": 614 }, { "epoch": 0.02592912705272256, "grad_norm": 2.622828245162964, "learning_rate": 4.314221472737493e-06, "loss": 0.7711, "step": 615 }, { "epoch": 0.025971288234922108, "grad_norm": 1.5737247467041016, "learning_rate": 4.321247892074199e-06, "loss": 0.5668, "step": 616 }, { "epoch": 0.026013449417121658, "grad_norm": 1.328611969947815, "learning_rate": 4.328274311410905e-06, "loss": 0.3216, "step": 617 }, { "epoch": 0.026055610599321204, "grad_norm": 2.313944101333618, "learning_rate": 4.335300730747611e-06, "loss": 1.5233, "step": 618 }, { "epoch": 0.026097771781520753, "grad_norm": 1.0444631576538086, "learning_rate": 4.342327150084317e-06, "loss": 0.8106, "step": 619 }, { "epoch": 0.026139932963720303, "grad_norm": 1.4128941297531128, "learning_rate": 4.3493535694210235e-06, "loss": 0.8791, "step": 620 }, { "epoch": 0.026182094145919852, "grad_norm": 2.738297700881958, "learning_rate": 4.35637998875773e-06, "loss": 1.0886, "step": 621 }, { "epoch": 0.0262242553281194, "grad_norm": 1.2269871234893799, "learning_rate": 4.363406408094435e-06, "loss": 0.4836, "step": 622 }, { "epoch": 0.02626641651031895, "grad_norm": 1.6952067613601685, "learning_rate": 4.370432827431141e-06, "loss": 0.3689, "step": 623 }, { "epoch": 0.026308577692518497, "grad_norm": 1.3282369375228882, "learning_rate": 4.377459246767848e-06, "loss": 0.9354, "step": 624 }, { "epoch": 0.026350738874718047, "grad_norm": 1.820481777191162, "learning_rate": 4.384485666104553e-06, "loss": 0.6738, "step": 625 }, { "epoch": 0.026392900056917596, "grad_norm": 0.7900989055633545, "learning_rate": 4.391512085441259e-06, "loss": 0.3752, "step": 626 }, { "epoch": 0.026435061239117146, "grad_norm": 3.6905100345611572, "learning_rate": 4.398538504777965e-06, "loss": 1.0026, "step": 627 }, { "epoch": 0.026477222421316695, "grad_norm": 2.640894651412964, "learning_rate": 4.405564924114671e-06, "loss": 0.9301, "step": 628 }, { "epoch": 0.02651938360351624, "grad_norm": 2.09220814704895, "learning_rate": 4.4125913434513774e-06, "loss": 0.4612, "step": 629 }, { "epoch": 0.02656154478571579, "grad_norm": 1.524882197380066, "learning_rate": 4.4196177627880835e-06, "loss": 1.2803, "step": 630 }, { "epoch": 0.02660370596791534, "grad_norm": 1.4292328357696533, "learning_rate": 4.42664418212479e-06, "loss": 0.2652, "step": 631 }, { "epoch": 0.02664586715011489, "grad_norm": 1.4660543203353882, "learning_rate": 4.433670601461496e-06, "loss": 0.7089, "step": 632 }, { "epoch": 0.02668802833231444, "grad_norm": 1.1556602716445923, "learning_rate": 4.440697020798201e-06, "loss": 1.0914, "step": 633 }, { "epoch": 0.02673018951451399, "grad_norm": 1.2739765644073486, "learning_rate": 4.447723440134907e-06, "loss": 0.5308, "step": 634 }, { "epoch": 0.026772350696713534, "grad_norm": 1.2705059051513672, "learning_rate": 4.454749859471614e-06, "loss": 0.9082, "step": 635 }, { "epoch": 0.026814511878913084, "grad_norm": 1.1524523496627808, "learning_rate": 4.461776278808319e-06, "loss": 0.7234, "step": 636 }, { "epoch": 0.026856673061112633, "grad_norm": 1.4900529384613037, "learning_rate": 4.468802698145025e-06, "loss": 0.3807, "step": 637 }, { "epoch": 0.026898834243312183, "grad_norm": 1.2081925868988037, "learning_rate": 4.475829117481731e-06, "loss": 0.4485, "step": 638 }, { "epoch": 0.026940995425511732, "grad_norm": 1.4500597715377808, "learning_rate": 4.4828555368184375e-06, "loss": 0.892, "step": 639 }, { "epoch": 0.026983156607711282, "grad_norm": 1.1480845212936401, "learning_rate": 4.489881956155144e-06, "loss": 0.8546, "step": 640 }, { "epoch": 0.027025317789910828, "grad_norm": 1.421226978302002, "learning_rate": 4.49690837549185e-06, "loss": 0.6783, "step": 641 }, { "epoch": 0.027067478972110377, "grad_norm": 1.1610575914382935, "learning_rate": 4.503934794828556e-06, "loss": 0.3021, "step": 642 }, { "epoch": 0.027109640154309927, "grad_norm": 1.9941579103469849, "learning_rate": 4.510961214165262e-06, "loss": 1.3912, "step": 643 }, { "epoch": 0.027151801336509476, "grad_norm": 1.3929412364959717, "learning_rate": 4.517987633501967e-06, "loss": 0.3823, "step": 644 }, { "epoch": 0.027193962518709026, "grad_norm": 2.0383141040802, "learning_rate": 4.525014052838673e-06, "loss": 0.9835, "step": 645 }, { "epoch": 0.027236123700908575, "grad_norm": 1.4103384017944336, "learning_rate": 4.53204047217538e-06, "loss": 1.3805, "step": 646 }, { "epoch": 0.02727828488310812, "grad_norm": 2.106969118118286, "learning_rate": 4.539066891512086e-06, "loss": 0.7454, "step": 647 }, { "epoch": 0.02732044606530767, "grad_norm": 1.4829233884811401, "learning_rate": 4.5460933108487914e-06, "loss": 0.5482, "step": 648 }, { "epoch": 0.02736260724750722, "grad_norm": 1.2409590482711792, "learning_rate": 4.5531197301854975e-06, "loss": 0.6766, "step": 649 }, { "epoch": 0.02740476842970677, "grad_norm": 1.8973960876464844, "learning_rate": 4.560146149522204e-06, "loss": 0.2849, "step": 650 }, { "epoch": 0.02744692961190632, "grad_norm": 1.5250637531280518, "learning_rate": 4.56717256885891e-06, "loss": 0.3572, "step": 651 }, { "epoch": 0.027489090794105865, "grad_norm": 2.392425060272217, "learning_rate": 4.574198988195616e-06, "loss": 0.3991, "step": 652 }, { "epoch": 0.027531251976305415, "grad_norm": 1.1243300437927246, "learning_rate": 4.581225407532322e-06, "loss": 1.4519, "step": 653 }, { "epoch": 0.027573413158504964, "grad_norm": 2.6889522075653076, "learning_rate": 4.588251826869028e-06, "loss": 0.6813, "step": 654 }, { "epoch": 0.027615574340704514, "grad_norm": 1.5349897146224976, "learning_rate": 4.595278246205734e-06, "loss": 1.1946, "step": 655 }, { "epoch": 0.027657735522904063, "grad_norm": 2.069342613220215, "learning_rate": 4.602304665542439e-06, "loss": 0.3669, "step": 656 }, { "epoch": 0.027699896705103613, "grad_norm": 1.5284658670425415, "learning_rate": 4.609331084879146e-06, "loss": 0.5392, "step": 657 }, { "epoch": 0.02774205788730316, "grad_norm": 1.0176739692687988, "learning_rate": 4.616357504215852e-06, "loss": 0.7716, "step": 658 }, { "epoch": 0.027784219069502708, "grad_norm": 1.7101421356201172, "learning_rate": 4.623383923552558e-06, "loss": 1.0902, "step": 659 }, { "epoch": 0.027826380251702258, "grad_norm": 1.029341459274292, "learning_rate": 4.630410342889264e-06, "loss": 0.3507, "step": 660 }, { "epoch": 0.027868541433901807, "grad_norm": 1.3473013639450073, "learning_rate": 4.63743676222597e-06, "loss": 0.7936, "step": 661 }, { "epoch": 0.027910702616101356, "grad_norm": 1.1900042295455933, "learning_rate": 4.644463181562676e-06, "loss": 0.2288, "step": 662 }, { "epoch": 0.027952863798300906, "grad_norm": 1.5453113317489624, "learning_rate": 4.651489600899382e-06, "loss": 0.4201, "step": 663 }, { "epoch": 0.027995024980500452, "grad_norm": 1.4766831398010254, "learning_rate": 4.658516020236088e-06, "loss": 0.5378, "step": 664 }, { "epoch": 0.0280371861627, "grad_norm": 1.130664587020874, "learning_rate": 4.665542439572794e-06, "loss": 0.703, "step": 665 }, { "epoch": 0.02807934734489955, "grad_norm": 1.1406097412109375, "learning_rate": 4.6725688589095e-06, "loss": 1.4361, "step": 666 }, { "epoch": 0.0281215085270991, "grad_norm": 1.8192108869552612, "learning_rate": 4.6795952782462055e-06, "loss": 0.6574, "step": 667 }, { "epoch": 0.02816366970929865, "grad_norm": 1.5360537767410278, "learning_rate": 4.6866216975829115e-06, "loss": 0.476, "step": 668 }, { "epoch": 0.028205830891498196, "grad_norm": 1.592335820198059, "learning_rate": 4.6936481169196185e-06, "loss": 0.4319, "step": 669 }, { "epoch": 0.028247992073697745, "grad_norm": 2.2944371700286865, "learning_rate": 4.700674536256324e-06, "loss": 0.8055, "step": 670 }, { "epoch": 0.028290153255897295, "grad_norm": 1.3635362386703491, "learning_rate": 4.70770095559303e-06, "loss": 0.3301, "step": 671 }, { "epoch": 0.028332314438096844, "grad_norm": 1.3555938005447388, "learning_rate": 4.714727374929736e-06, "loss": 0.4595, "step": 672 }, { "epoch": 0.028374475620296394, "grad_norm": 1.3994652032852173, "learning_rate": 4.721753794266442e-06, "loss": 0.4038, "step": 673 }, { "epoch": 0.028416636802495943, "grad_norm": 1.4577797651290894, "learning_rate": 4.728780213603148e-06, "loss": 0.297, "step": 674 }, { "epoch": 0.02845879798469549, "grad_norm": 1.4295732975006104, "learning_rate": 4.735806632939854e-06, "loss": 0.9429, "step": 675 }, { "epoch": 0.02850095916689504, "grad_norm": 1.4650098085403442, "learning_rate": 4.74283305227656e-06, "loss": 0.4724, "step": 676 }, { "epoch": 0.028543120349094588, "grad_norm": 1.5009890794754028, "learning_rate": 4.749859471613266e-06, "loss": 0.8242, "step": 677 }, { "epoch": 0.028585281531294138, "grad_norm": 1.1672989130020142, "learning_rate": 4.756885890949972e-06, "loss": 0.2967, "step": 678 }, { "epoch": 0.028627442713493687, "grad_norm": 1.3087189197540283, "learning_rate": 4.763912310286678e-06, "loss": 0.3967, "step": 679 }, { "epoch": 0.028669603895693237, "grad_norm": 1.6527498960494995, "learning_rate": 4.770938729623385e-06, "loss": 0.5536, "step": 680 }, { "epoch": 0.028711765077892783, "grad_norm": 1.2708995342254639, "learning_rate": 4.777965148960091e-06, "loss": 0.3577, "step": 681 }, { "epoch": 0.028753926260092332, "grad_norm": 1.7054883241653442, "learning_rate": 4.784991568296796e-06, "loss": 0.4475, "step": 682 }, { "epoch": 0.02879608744229188, "grad_norm": 1.660606026649475, "learning_rate": 4.792017987633502e-06, "loss": 0.6149, "step": 683 }, { "epoch": 0.02883824862449143, "grad_norm": 1.4119787216186523, "learning_rate": 4.799044406970208e-06, "loss": 0.2452, "step": 684 }, { "epoch": 0.02888040980669098, "grad_norm": 1.5333820581436157, "learning_rate": 4.806070826306914e-06, "loss": 0.543, "step": 685 }, { "epoch": 0.02892257098889053, "grad_norm": 3.344421148300171, "learning_rate": 4.81309724564362e-06, "loss": 1.1875, "step": 686 }, { "epoch": 0.028964732171090076, "grad_norm": 1.166094422340393, "learning_rate": 4.820123664980326e-06, "loss": 0.8665, "step": 687 }, { "epoch": 0.029006893353289626, "grad_norm": 0.7620944380760193, "learning_rate": 4.8271500843170325e-06, "loss": 0.2487, "step": 688 }, { "epoch": 0.029049054535489175, "grad_norm": 1.312497854232788, "learning_rate": 4.834176503653738e-06, "loss": 0.5743, "step": 689 }, { "epoch": 0.029091215717688725, "grad_norm": 0.9545327425003052, "learning_rate": 4.841202922990444e-06, "loss": 0.2713, "step": 690 }, { "epoch": 0.029133376899888274, "grad_norm": 1.7937405109405518, "learning_rate": 4.848229342327151e-06, "loss": 0.5211, "step": 691 }, { "epoch": 0.02917553808208782, "grad_norm": 1.4521936178207397, "learning_rate": 4.855255761663857e-06, "loss": 0.6465, "step": 692 }, { "epoch": 0.02921769926428737, "grad_norm": 2.2277729511260986, "learning_rate": 4.862282181000562e-06, "loss": 0.6951, "step": 693 }, { "epoch": 0.02925986044648692, "grad_norm": 0.933417797088623, "learning_rate": 4.869308600337268e-06, "loss": 0.4385, "step": 694 }, { "epoch": 0.02930202162868647, "grad_norm": 1.1421314477920532, "learning_rate": 4.876335019673974e-06, "loss": 0.3044, "step": 695 }, { "epoch": 0.029344182810886018, "grad_norm": 1.7097927331924438, "learning_rate": 4.88336143901068e-06, "loss": 0.7613, "step": 696 }, { "epoch": 0.029386343993085567, "grad_norm": 1.5051414966583252, "learning_rate": 4.8903878583473864e-06, "loss": 0.6849, "step": 697 }, { "epoch": 0.029428505175285113, "grad_norm": 1.4448155164718628, "learning_rate": 4.8974142776840925e-06, "loss": 0.3285, "step": 698 }, { "epoch": 0.029470666357484663, "grad_norm": 1.3994874954223633, "learning_rate": 4.904440697020799e-06, "loss": 0.2411, "step": 699 }, { "epoch": 0.029512827539684212, "grad_norm": 1.0161045789718628, "learning_rate": 4.911467116357505e-06, "loss": 0.5459, "step": 700 }, { "epoch": 0.029554988721883762, "grad_norm": 1.0129061937332153, "learning_rate": 4.91849353569421e-06, "loss": 0.2588, "step": 701 }, { "epoch": 0.02959714990408331, "grad_norm": 1.1822490692138672, "learning_rate": 4.925519955030917e-06, "loss": 0.8826, "step": 702 }, { "epoch": 0.02963931108628286, "grad_norm": 1.8852972984313965, "learning_rate": 4.932546374367623e-06, "loss": 0.5506, "step": 703 }, { "epoch": 0.029681472268482407, "grad_norm": 0.8445301651954651, "learning_rate": 4.939572793704328e-06, "loss": 0.6949, "step": 704 }, { "epoch": 0.029723633450681956, "grad_norm": 1.3621866703033447, "learning_rate": 4.946599213041034e-06, "loss": 0.8774, "step": 705 }, { "epoch": 0.029765794632881506, "grad_norm": 2.048412322998047, "learning_rate": 4.95362563237774e-06, "loss": 0.7488, "step": 706 }, { "epoch": 0.029807955815081055, "grad_norm": 1.5289115905761719, "learning_rate": 4.9606520517144465e-06, "loss": 0.5077, "step": 707 }, { "epoch": 0.029850116997280605, "grad_norm": 1.3090804815292358, "learning_rate": 4.9676784710511526e-06, "loss": 0.4248, "step": 708 }, { "epoch": 0.029892278179480154, "grad_norm": 1.1718096733093262, "learning_rate": 4.974704890387859e-06, "loss": 0.5857, "step": 709 }, { "epoch": 0.0299344393616797, "grad_norm": 1.4924635887145996, "learning_rate": 4.981731309724565e-06, "loss": 0.6452, "step": 710 }, { "epoch": 0.02997660054387925, "grad_norm": 1.3171330690383911, "learning_rate": 4.988757729061271e-06, "loss": 0.288, "step": 711 }, { "epoch": 0.0300187617260788, "grad_norm": 1.3025236129760742, "learning_rate": 4.995784148397976e-06, "loss": 0.2303, "step": 712 }, { "epoch": 0.03006092290827835, "grad_norm": 2.472097635269165, "learning_rate": 5.002810567734682e-06, "loss": 1.4134, "step": 713 }, { "epoch": 0.030103084090477898, "grad_norm": 1.1957499980926514, "learning_rate": 5.009836987071389e-06, "loss": 0.4209, "step": 714 }, { "epoch": 0.030145245272677444, "grad_norm": 1.5497026443481445, "learning_rate": 5.016863406408095e-06, "loss": 0.7961, "step": 715 }, { "epoch": 0.030187406454876994, "grad_norm": 1.943673014640808, "learning_rate": 5.0238898257448004e-06, "loss": 0.7679, "step": 716 }, { "epoch": 0.030229567637076543, "grad_norm": 1.6687383651733398, "learning_rate": 5.0309162450815065e-06, "loss": 0.4306, "step": 717 }, { "epoch": 0.030271728819276093, "grad_norm": 0.9398708343505859, "learning_rate": 5.037942664418213e-06, "loss": 0.2254, "step": 718 }, { "epoch": 0.030313890001475642, "grad_norm": 1.1356337070465088, "learning_rate": 5.044969083754919e-06, "loss": 0.9159, "step": 719 }, { "epoch": 0.03035605118367519, "grad_norm": 1.1809285879135132, "learning_rate": 5.051995503091625e-06, "loss": 0.4346, "step": 720 }, { "epoch": 0.030398212365874738, "grad_norm": 1.4534380435943604, "learning_rate": 5.059021922428331e-06, "loss": 0.4896, "step": 721 }, { "epoch": 0.030440373548074287, "grad_norm": 0.8467216491699219, "learning_rate": 5.066048341765037e-06, "loss": 0.2552, "step": 722 }, { "epoch": 0.030482534730273837, "grad_norm": 1.3421591520309448, "learning_rate": 5.073074761101742e-06, "loss": 0.6306, "step": 723 }, { "epoch": 0.030524695912473386, "grad_norm": 1.3328514099121094, "learning_rate": 5.080101180438448e-06, "loss": 0.6352, "step": 724 }, { "epoch": 0.030566857094672936, "grad_norm": 0.7330204844474792, "learning_rate": 5.087127599775155e-06, "loss": 0.1529, "step": 725 }, { "epoch": 0.030609018276872485, "grad_norm": 1.4849166870117188, "learning_rate": 5.094154019111861e-06, "loss": 0.4989, "step": 726 }, { "epoch": 0.03065117945907203, "grad_norm": 1.3607146739959717, "learning_rate": 5.1011804384485666e-06, "loss": 0.3071, "step": 727 }, { "epoch": 0.03069334064127158, "grad_norm": 4.1666460037231445, "learning_rate": 5.108206857785273e-06, "loss": 0.7891, "step": 728 }, { "epoch": 0.03073550182347113, "grad_norm": 1.2478276491165161, "learning_rate": 5.115233277121979e-06, "loss": 0.4305, "step": 729 }, { "epoch": 0.03077766300567068, "grad_norm": 1.2152276039123535, "learning_rate": 5.122259696458685e-06, "loss": 0.8322, "step": 730 }, { "epoch": 0.03081982418787023, "grad_norm": 1.1073685884475708, "learning_rate": 5.129286115795391e-06, "loss": 0.6098, "step": 731 }, { "epoch": 0.03086198537006978, "grad_norm": 1.6668719053268433, "learning_rate": 5.136312535132097e-06, "loss": 0.7312, "step": 732 }, { "epoch": 0.030904146552269324, "grad_norm": 2.287731647491455, "learning_rate": 5.143338954468803e-06, "loss": 0.956, "step": 733 }, { "epoch": 0.030946307734468874, "grad_norm": 1.7079722881317139, "learning_rate": 5.150365373805509e-06, "loss": 0.2691, "step": 734 }, { "epoch": 0.030988468916668423, "grad_norm": 2.104652166366577, "learning_rate": 5.1573917931422144e-06, "loss": 0.4532, "step": 735 }, { "epoch": 0.031030630098867973, "grad_norm": 3.9226648807525635, "learning_rate": 5.164418212478921e-06, "loss": 0.906, "step": 736 }, { "epoch": 0.031072791281067522, "grad_norm": 1.8192204236984253, "learning_rate": 5.1714446318156275e-06, "loss": 0.4515, "step": 737 }, { "epoch": 0.03111495246326707, "grad_norm": 1.7427585124969482, "learning_rate": 5.178471051152333e-06, "loss": 0.3363, "step": 738 }, { "epoch": 0.031157113645466618, "grad_norm": 1.155247449874878, "learning_rate": 5.185497470489039e-06, "loss": 0.8963, "step": 739 }, { "epoch": 0.031199274827666167, "grad_norm": 1.7176406383514404, "learning_rate": 5.192523889825745e-06, "loss": 1.1651, "step": 740 }, { "epoch": 0.031241436009865717, "grad_norm": 3.4727580547332764, "learning_rate": 5.199550309162451e-06, "loss": 0.8664, "step": 741 }, { "epoch": 0.03128359719206526, "grad_norm": 1.045898675918579, "learning_rate": 5.206576728499157e-06, "loss": 0.4066, "step": 742 }, { "epoch": 0.031325758374264816, "grad_norm": 1.6439234018325806, "learning_rate": 5.213603147835863e-06, "loss": 0.7582, "step": 743 }, { "epoch": 0.03136791955646436, "grad_norm": 5.022782325744629, "learning_rate": 5.220629567172569e-06, "loss": 1.162, "step": 744 }, { "epoch": 0.031410080738663915, "grad_norm": 1.7178332805633545, "learning_rate": 5.227655986509275e-06, "loss": 0.3136, "step": 745 }, { "epoch": 0.03145224192086346, "grad_norm": 1.2951769828796387, "learning_rate": 5.2346824058459806e-06, "loss": 0.6513, "step": 746 }, { "epoch": 0.03149440310306301, "grad_norm": 1.0553522109985352, "learning_rate": 5.2417088251826875e-06, "loss": 0.4411, "step": 747 }, { "epoch": 0.03153656428526256, "grad_norm": 1.151970624923706, "learning_rate": 5.248735244519394e-06, "loss": 0.7257, "step": 748 }, { "epoch": 0.031578725467462106, "grad_norm": 1.3030213117599487, "learning_rate": 5.2557616638561e-06, "loss": 0.8287, "step": 749 }, { "epoch": 0.03162088664966166, "grad_norm": 1.051535725593567, "learning_rate": 5.262788083192805e-06, "loss": 0.4847, "step": 750 }, { "epoch": 0.031663047831861205, "grad_norm": 1.9529950618743896, "learning_rate": 5.269814502529511e-06, "loss": 0.5516, "step": 751 }, { "epoch": 0.03170520901406076, "grad_norm": 1.8829514980316162, "learning_rate": 5.276840921866217e-06, "loss": 0.9472, "step": 752 }, { "epoch": 0.031747370196260304, "grad_norm": 1.5808563232421875, "learning_rate": 5.283867341202923e-06, "loss": 0.8848, "step": 753 }, { "epoch": 0.03178953137845985, "grad_norm": 1.620773434638977, "learning_rate": 5.290893760539629e-06, "loss": 0.7208, "step": 754 }, { "epoch": 0.0318316925606594, "grad_norm": 2.1548869609832764, "learning_rate": 5.297920179876335e-06, "loss": 0.9232, "step": 755 }, { "epoch": 0.03187385374285895, "grad_norm": 1.4608380794525146, "learning_rate": 5.3049465992130415e-06, "loss": 0.3039, "step": 756 }, { "epoch": 0.0319160149250585, "grad_norm": 2.0819129943847656, "learning_rate": 5.311973018549747e-06, "loss": 0.4799, "step": 757 }, { "epoch": 0.03195817610725805, "grad_norm": 2.208596706390381, "learning_rate": 5.318999437886453e-06, "loss": 0.5289, "step": 758 }, { "epoch": 0.032000337289457594, "grad_norm": 4.547977924346924, "learning_rate": 5.32602585722316e-06, "loss": 1.3114, "step": 759 }, { "epoch": 0.032042498471657146, "grad_norm": 1.2995729446411133, "learning_rate": 5.333052276559866e-06, "loss": 0.563, "step": 760 }, { "epoch": 0.03208465965385669, "grad_norm": 0.9667707085609436, "learning_rate": 5.340078695896571e-06, "loss": 0.3196, "step": 761 }, { "epoch": 0.032126820836056245, "grad_norm": 1.8068325519561768, "learning_rate": 5.347105115233277e-06, "loss": 1.1975, "step": 762 }, { "epoch": 0.03216898201825579, "grad_norm": 2.106139659881592, "learning_rate": 5.354131534569983e-06, "loss": 0.4, "step": 763 }, { "epoch": 0.03221114320045534, "grad_norm": 1.4956883192062378, "learning_rate": 5.361157953906689e-06, "loss": 0.2871, "step": 764 }, { "epoch": 0.03225330438265489, "grad_norm": 1.191948413848877, "learning_rate": 5.368184373243395e-06, "loss": 0.2558, "step": 765 }, { "epoch": 0.032295465564854436, "grad_norm": 1.9758968353271484, "learning_rate": 5.3752107925801015e-06, "loss": 1.1954, "step": 766 }, { "epoch": 0.03233762674705399, "grad_norm": 1.3643250465393066, "learning_rate": 5.382237211916808e-06, "loss": 0.5176, "step": 767 }, { "epoch": 0.032379787929253535, "grad_norm": 1.4761117696762085, "learning_rate": 5.389263631253514e-06, "loss": 0.4775, "step": 768 }, { "epoch": 0.03242194911145309, "grad_norm": 1.010472059249878, "learning_rate": 5.396290050590219e-06, "loss": 0.4766, "step": 769 }, { "epoch": 0.032464110293652634, "grad_norm": 1.1926066875457764, "learning_rate": 5.403316469926926e-06, "loss": 0.3575, "step": 770 }, { "epoch": 0.03250627147585218, "grad_norm": 1.903672456741333, "learning_rate": 5.410342889263632e-06, "loss": 1.3304, "step": 771 }, { "epoch": 0.03254843265805173, "grad_norm": 1.2123774290084839, "learning_rate": 5.417369308600337e-06, "loss": 0.6428, "step": 772 }, { "epoch": 0.03259059384025128, "grad_norm": 1.61211359500885, "learning_rate": 5.424395727937043e-06, "loss": 0.4332, "step": 773 }, { "epoch": 0.03263275502245083, "grad_norm": 1.4781628847122192, "learning_rate": 5.431422147273749e-06, "loss": 1.3128, "step": 774 }, { "epoch": 0.03267491620465038, "grad_norm": 2.2108066082000732, "learning_rate": 5.4384485666104555e-06, "loss": 0.4026, "step": 775 }, { "epoch": 0.032717077386849924, "grad_norm": 2.1329030990600586, "learning_rate": 5.4454749859471616e-06, "loss": 0.3948, "step": 776 }, { "epoch": 0.03275923856904948, "grad_norm": 1.160630226135254, "learning_rate": 5.452501405283868e-06, "loss": 0.367, "step": 777 }, { "epoch": 0.03280139975124902, "grad_norm": 1.699188470840454, "learning_rate": 5.459527824620574e-06, "loss": 0.2817, "step": 778 }, { "epoch": 0.032843560933448576, "grad_norm": 1.8819931745529175, "learning_rate": 5.46655424395728e-06, "loss": 0.3937, "step": 779 }, { "epoch": 0.03288572211564812, "grad_norm": 1.5977668762207031, "learning_rate": 5.473580663293985e-06, "loss": 1.1933, "step": 780 }, { "epoch": 0.03292788329784767, "grad_norm": 1.3904931545257568, "learning_rate": 5.480607082630692e-06, "loss": 1.3501, "step": 781 }, { "epoch": 0.03297004448004722, "grad_norm": 1.438415288925171, "learning_rate": 5.487633501967398e-06, "loss": 0.4084, "step": 782 }, { "epoch": 0.03301220566224677, "grad_norm": 1.3242145776748657, "learning_rate": 5.494659921304104e-06, "loss": 0.4409, "step": 783 }, { "epoch": 0.03305436684444632, "grad_norm": 0.8390516042709351, "learning_rate": 5.501686340640809e-06, "loss": 0.269, "step": 784 }, { "epoch": 0.033096528026645866, "grad_norm": 1.9372568130493164, "learning_rate": 5.5087127599775155e-06, "loss": 0.3803, "step": 785 }, { "epoch": 0.03313868920884542, "grad_norm": 2.334665536880493, "learning_rate": 5.515739179314222e-06, "loss": 0.8274, "step": 786 }, { "epoch": 0.033180850391044965, "grad_norm": 0.9892289042472839, "learning_rate": 5.522765598650928e-06, "loss": 0.3998, "step": 787 }, { "epoch": 0.03322301157324451, "grad_norm": 1.7748768329620361, "learning_rate": 5.529792017987634e-06, "loss": 1.2796, "step": 788 }, { "epoch": 0.033265172755444064, "grad_norm": 1.6590083837509155, "learning_rate": 5.53681843732434e-06, "loss": 1.2732, "step": 789 }, { "epoch": 0.03330733393764361, "grad_norm": 1.6982626914978027, "learning_rate": 5.543844856661046e-06, "loss": 1.3357, "step": 790 }, { "epoch": 0.03334949511984316, "grad_norm": 1.3305805921554565, "learning_rate": 5.550871275997751e-06, "loss": 0.7172, "step": 791 }, { "epoch": 0.03339165630204271, "grad_norm": 1.0644007921218872, "learning_rate": 5.557897695334458e-06, "loss": 0.3385, "step": 792 }, { "epoch": 0.033433817484242255, "grad_norm": 1.6416895389556885, "learning_rate": 5.564924114671164e-06, "loss": 0.3803, "step": 793 }, { "epoch": 0.03347597866644181, "grad_norm": 1.4739817380905151, "learning_rate": 5.57195053400787e-06, "loss": 0.4768, "step": 794 }, { "epoch": 0.033518139848641354, "grad_norm": 1.216623306274414, "learning_rate": 5.5789769533445756e-06, "loss": 0.3097, "step": 795 }, { "epoch": 0.03356030103084091, "grad_norm": 1.3396941423416138, "learning_rate": 5.586003372681282e-06, "loss": 0.3508, "step": 796 }, { "epoch": 0.03360246221304045, "grad_norm": 1.011696219444275, "learning_rate": 5.593029792017988e-06, "loss": 0.1772, "step": 797 }, { "epoch": 0.033644623395240006, "grad_norm": 1.596599817276001, "learning_rate": 5.600056211354694e-06, "loss": 0.8389, "step": 798 }, { "epoch": 0.03368678457743955, "grad_norm": 1.5029172897338867, "learning_rate": 5.6070826306914e-06, "loss": 0.5206, "step": 799 }, { "epoch": 0.0337289457596391, "grad_norm": 1.843036413192749, "learning_rate": 5.614109050028106e-06, "loss": 0.3065, "step": 800 }, { "epoch": 0.03377110694183865, "grad_norm": 1.414650559425354, "learning_rate": 5.621135469364812e-06, "loss": 0.2045, "step": 801 }, { "epoch": 0.0338132681240382, "grad_norm": 4.964562892913818, "learning_rate": 5.628161888701518e-06, "loss": 1.2935, "step": 802 }, { "epoch": 0.03385542930623775, "grad_norm": 1.2844018936157227, "learning_rate": 5.635188308038223e-06, "loss": 0.5342, "step": 803 }, { "epoch": 0.033897590488437296, "grad_norm": 2.2775027751922607, "learning_rate": 5.64221472737493e-06, "loss": 0.7833, "step": 804 }, { "epoch": 0.03393975167063684, "grad_norm": 1.9576585292816162, "learning_rate": 5.6492411467116364e-06, "loss": 1.3138, "step": 805 }, { "epoch": 0.033981912852836395, "grad_norm": 1.3421295881271362, "learning_rate": 5.656267566048342e-06, "loss": 0.9275, "step": 806 }, { "epoch": 0.03402407403503594, "grad_norm": 1.6284939050674438, "learning_rate": 5.663293985385048e-06, "loss": 0.252, "step": 807 }, { "epoch": 0.034066235217235494, "grad_norm": 1.127935767173767, "learning_rate": 5.670320404721754e-06, "loss": 0.345, "step": 808 }, { "epoch": 0.03410839639943504, "grad_norm": 1.2212494611740112, "learning_rate": 5.677346824058461e-06, "loss": 0.3347, "step": 809 }, { "epoch": 0.034150557581634586, "grad_norm": 1.5453197956085205, "learning_rate": 5.684373243395166e-06, "loss": 1.1632, "step": 810 }, { "epoch": 0.03419271876383414, "grad_norm": 1.3125333786010742, "learning_rate": 5.691399662731872e-06, "loss": 0.2806, "step": 811 }, { "epoch": 0.034234879946033685, "grad_norm": 1.3289680480957031, "learning_rate": 5.698426082068578e-06, "loss": 0.3384, "step": 812 }, { "epoch": 0.03427704112823324, "grad_norm": 1.2932226657867432, "learning_rate": 5.705452501405284e-06, "loss": 0.4118, "step": 813 }, { "epoch": 0.034319202310432784, "grad_norm": 3.2671117782592773, "learning_rate": 5.7124789207419896e-06, "loss": 0.3124, "step": 814 }, { "epoch": 0.03436136349263234, "grad_norm": 0.9932941794395447, "learning_rate": 5.7195053400786965e-06, "loss": 0.1922, "step": 815 }, { "epoch": 0.03440352467483188, "grad_norm": 1.581778645515442, "learning_rate": 5.726531759415403e-06, "loss": 0.4042, "step": 816 }, { "epoch": 0.03444568585703143, "grad_norm": 1.190858006477356, "learning_rate": 5.733558178752108e-06, "loss": 0.3085, "step": 817 }, { "epoch": 0.03448784703923098, "grad_norm": 1.8388848304748535, "learning_rate": 5.740584598088814e-06, "loss": 0.5452, "step": 818 }, { "epoch": 0.03453000822143053, "grad_norm": 1.6239627599716187, "learning_rate": 5.74761101742552e-06, "loss": 0.5466, "step": 819 }, { "epoch": 0.03457216940363008, "grad_norm": 1.6850695610046387, "learning_rate": 5.754637436762226e-06, "loss": 0.4815, "step": 820 }, { "epoch": 0.03461433058582963, "grad_norm": 3.3327596187591553, "learning_rate": 5.761663856098932e-06, "loss": 0.8438, "step": 821 }, { "epoch": 0.03465649176802917, "grad_norm": 2.205491065979004, "learning_rate": 5.768690275435638e-06, "loss": 0.4672, "step": 822 }, { "epoch": 0.034698652950228726, "grad_norm": 1.374959111213684, "learning_rate": 5.775716694772344e-06, "loss": 0.7139, "step": 823 }, { "epoch": 0.03474081413242827, "grad_norm": 2.1067094802856445, "learning_rate": 5.7827431141090504e-06, "loss": 0.6706, "step": 824 }, { "epoch": 0.034782975314627824, "grad_norm": 1.4119555950164795, "learning_rate": 5.789769533445756e-06, "loss": 0.7248, "step": 825 }, { "epoch": 0.03482513649682737, "grad_norm": 1.707403302192688, "learning_rate": 5.796795952782463e-06, "loss": 1.2078, "step": 826 }, { "epoch": 0.034867297679026917, "grad_norm": 0.9917463064193726, "learning_rate": 5.803822372119169e-06, "loss": 0.269, "step": 827 }, { "epoch": 0.03490945886122647, "grad_norm": 1.743405818939209, "learning_rate": 5.810848791455875e-06, "loss": 1.1062, "step": 828 }, { "epoch": 0.034951620043426015, "grad_norm": 1.7177977561950684, "learning_rate": 5.81787521079258e-06, "loss": 0.3599, "step": 829 }, { "epoch": 0.03499378122562557, "grad_norm": 2.21469783782959, "learning_rate": 5.824901630129286e-06, "loss": 1.4137, "step": 830 }, { "epoch": 0.035035942407825114, "grad_norm": 2.724259376525879, "learning_rate": 5.831928049465992e-06, "loss": 1.5634, "step": 831 }, { "epoch": 0.03507810359002467, "grad_norm": 0.7098842859268188, "learning_rate": 5.838954468802698e-06, "loss": 0.2314, "step": 832 }, { "epoch": 0.03512026477222421, "grad_norm": 1.3362654447555542, "learning_rate": 5.845980888139404e-06, "loss": 0.758, "step": 833 }, { "epoch": 0.03516242595442376, "grad_norm": 1.7936257123947144, "learning_rate": 5.8530073074761105e-06, "loss": 0.7219, "step": 834 }, { "epoch": 0.03520458713662331, "grad_norm": 1.2100876569747925, "learning_rate": 5.860033726812817e-06, "loss": 0.5356, "step": 835 }, { "epoch": 0.03524674831882286, "grad_norm": 1.9358171224594116, "learning_rate": 5.867060146149523e-06, "loss": 0.3488, "step": 836 }, { "epoch": 0.03528890950102241, "grad_norm": 0.801558256149292, "learning_rate": 5.874086565486228e-06, "loss": 0.2168, "step": 837 }, { "epoch": 0.03533107068322196, "grad_norm": 1.2969213724136353, "learning_rate": 5.881112984822935e-06, "loss": 0.3705, "step": 838 }, { "epoch": 0.0353732318654215, "grad_norm": 1.131314754486084, "learning_rate": 5.888139404159641e-06, "loss": 0.5874, "step": 839 }, { "epoch": 0.035415393047621056, "grad_norm": 1.0355558395385742, "learning_rate": 5.895165823496346e-06, "loss": 0.2205, "step": 840 }, { "epoch": 0.0354575542298206, "grad_norm": 1.74932062625885, "learning_rate": 5.902192242833052e-06, "loss": 0.5416, "step": 841 }, { "epoch": 0.035499715412020155, "grad_norm": 1.0953749418258667, "learning_rate": 5.909218662169758e-06, "loss": 0.7763, "step": 842 }, { "epoch": 0.0355418765942197, "grad_norm": 3.0940427780151367, "learning_rate": 5.916245081506465e-06, "loss": 0.8796, "step": 843 }, { "epoch": 0.035584037776419254, "grad_norm": 2.7786433696746826, "learning_rate": 5.9232715008431705e-06, "loss": 0.763, "step": 844 }, { "epoch": 0.0356261989586188, "grad_norm": 1.105903148651123, "learning_rate": 5.930297920179877e-06, "loss": 0.2256, "step": 845 }, { "epoch": 0.035668360140818346, "grad_norm": 1.785275936126709, "learning_rate": 5.937324339516583e-06, "loss": 0.5968, "step": 846 }, { "epoch": 0.0357105213230179, "grad_norm": 1.0452378988265991, "learning_rate": 5.944350758853289e-06, "loss": 0.3004, "step": 847 }, { "epoch": 0.035752682505217445, "grad_norm": 2.1335973739624023, "learning_rate": 5.951377178189994e-06, "loss": 0.3817, "step": 848 }, { "epoch": 0.035794843687417, "grad_norm": 1.5548580884933472, "learning_rate": 5.958403597526701e-06, "loss": 0.9773, "step": 849 }, { "epoch": 0.035837004869616544, "grad_norm": 1.5015443563461304, "learning_rate": 5.965430016863407e-06, "loss": 0.2815, "step": 850 }, { "epoch": 0.03587916605181609, "grad_norm": 1.0823595523834229, "learning_rate": 5.972456436200112e-06, "loss": 0.1817, "step": 851 }, { "epoch": 0.03592132723401564, "grad_norm": 1.006870985031128, "learning_rate": 5.979482855536818e-06, "loss": 0.6433, "step": 852 }, { "epoch": 0.03596348841621519, "grad_norm": 1.2464817762374878, "learning_rate": 5.9865092748735245e-06, "loss": 0.5884, "step": 853 }, { "epoch": 0.03600564959841474, "grad_norm": 0.9834656715393066, "learning_rate": 5.9935356942102314e-06, "loss": 0.2933, "step": 854 }, { "epoch": 0.03604781078061429, "grad_norm": 1.2459535598754883, "learning_rate": 6.000562113546937e-06, "loss": 0.3622, "step": 855 }, { "epoch": 0.036089971962813834, "grad_norm": 2.0507357120513916, "learning_rate": 6.007588532883643e-06, "loss": 1.2522, "step": 856 }, { "epoch": 0.03613213314501339, "grad_norm": 1.2972501516342163, "learning_rate": 6.014614952220349e-06, "loss": 0.2993, "step": 857 }, { "epoch": 0.03617429432721293, "grad_norm": 1.5866713523864746, "learning_rate": 6.021641371557055e-06, "loss": 0.4591, "step": 858 }, { "epoch": 0.036216455509412486, "grad_norm": 1.3285431861877441, "learning_rate": 6.02866779089376e-06, "loss": 0.969, "step": 859 }, { "epoch": 0.03625861669161203, "grad_norm": 1.942370057106018, "learning_rate": 6.035694210230467e-06, "loss": 1.5011, "step": 860 }, { "epoch": 0.036300777873811585, "grad_norm": 1.5665981769561768, "learning_rate": 6.042720629567173e-06, "loss": 0.3592, "step": 861 }, { "epoch": 0.03634293905601113, "grad_norm": 1.3138203620910645, "learning_rate": 6.049747048903879e-06, "loss": 0.2803, "step": 862 }, { "epoch": 0.03638510023821068, "grad_norm": 2.323747158050537, "learning_rate": 6.0567734682405845e-06, "loss": 0.7553, "step": 863 }, { "epoch": 0.03642726142041023, "grad_norm": 1.2220163345336914, "learning_rate": 6.063799887577291e-06, "loss": 0.2651, "step": 864 }, { "epoch": 0.036469422602609776, "grad_norm": 1.538148045539856, "learning_rate": 6.070826306913997e-06, "loss": 0.5644, "step": 865 }, { "epoch": 0.03651158378480933, "grad_norm": 2.418821334838867, "learning_rate": 6.077852726250703e-06, "loss": 0.259, "step": 866 }, { "epoch": 0.036553744967008875, "grad_norm": 2.9334447383880615, "learning_rate": 6.084879145587409e-06, "loss": 1.3906, "step": 867 }, { "epoch": 0.03659590614920842, "grad_norm": 1.7132469415664673, "learning_rate": 6.091905564924115e-06, "loss": 0.3704, "step": 868 }, { "epoch": 0.036638067331407974, "grad_norm": 1.0885459184646606, "learning_rate": 6.098931984260821e-06, "loss": 0.2213, "step": 869 }, { "epoch": 0.03668022851360752, "grad_norm": 1.8941434621810913, "learning_rate": 6.105958403597527e-06, "loss": 0.383, "step": 870 }, { "epoch": 0.03672238969580707, "grad_norm": 1.5712002515792847, "learning_rate": 6.112984822934233e-06, "loss": 0.5494, "step": 871 }, { "epoch": 0.03676455087800662, "grad_norm": 3.7983059883117676, "learning_rate": 6.120011242270939e-06, "loss": 1.398, "step": 872 }, { "epoch": 0.036806712060206165, "grad_norm": 1.1845130920410156, "learning_rate": 6.1270376616076454e-06, "loss": 0.233, "step": 873 }, { "epoch": 0.03684887324240572, "grad_norm": 2.008737802505493, "learning_rate": 6.134064080944351e-06, "loss": 0.6019, "step": 874 }, { "epoch": 0.036891034424605264, "grad_norm": 1.868309497833252, "learning_rate": 6.141090500281057e-06, "loss": 0.3079, "step": 875 }, { "epoch": 0.03693319560680482, "grad_norm": 1.0725858211517334, "learning_rate": 6.148116919617763e-06, "loss": 0.501, "step": 876 }, { "epoch": 0.03697535678900436, "grad_norm": 1.2656912803649902, "learning_rate": 6.15514333895447e-06, "loss": 0.3363, "step": 877 }, { "epoch": 0.037017517971203916, "grad_norm": 1.5942836999893188, "learning_rate": 6.162169758291175e-06, "loss": 0.6011, "step": 878 }, { "epoch": 0.03705967915340346, "grad_norm": 1.613444209098816, "learning_rate": 6.169196177627881e-06, "loss": 0.5482, "step": 879 }, { "epoch": 0.03710184033560301, "grad_norm": 1.0983059406280518, "learning_rate": 6.176222596964587e-06, "loss": 0.5575, "step": 880 }, { "epoch": 0.03714400151780256, "grad_norm": 1.2426254749298096, "learning_rate": 6.183249016301293e-06, "loss": 0.853, "step": 881 }, { "epoch": 0.03718616270000211, "grad_norm": 2.000128984451294, "learning_rate": 6.1902754356379985e-06, "loss": 0.6325, "step": 882 }, { "epoch": 0.03722832388220166, "grad_norm": 4.171704292297363, "learning_rate": 6.1973018549747055e-06, "loss": 1.47, "step": 883 }, { "epoch": 0.037270485064401206, "grad_norm": 1.1666878461837769, "learning_rate": 6.2043282743114116e-06, "loss": 0.2236, "step": 884 }, { "epoch": 0.03731264624660075, "grad_norm": 1.5172467231750488, "learning_rate": 6.211354693648117e-06, "loss": 0.3985, "step": 885 }, { "epoch": 0.037354807428800305, "grad_norm": 5.148082256317139, "learning_rate": 6.218381112984823e-06, "loss": 1.0229, "step": 886 }, { "epoch": 0.03739696861099985, "grad_norm": 2.869997024536133, "learning_rate": 6.225407532321529e-06, "loss": 0.7571, "step": 887 }, { "epoch": 0.037439129793199404, "grad_norm": 2.083317518234253, "learning_rate": 6.232433951658236e-06, "loss": 0.5431, "step": 888 }, { "epoch": 0.03748129097539895, "grad_norm": 2.1301429271698, "learning_rate": 6.239460370994941e-06, "loss": 1.3732, "step": 889 }, { "epoch": 0.0375234521575985, "grad_norm": 1.530592918395996, "learning_rate": 6.246486790331647e-06, "loss": 0.4518, "step": 890 }, { "epoch": 0.03756561333979805, "grad_norm": 1.7092742919921875, "learning_rate": 6.253513209668353e-06, "loss": 0.7042, "step": 891 }, { "epoch": 0.037607774521997595, "grad_norm": 1.9030189514160156, "learning_rate": 6.2605396290050594e-06, "loss": 0.3759, "step": 892 }, { "epoch": 0.03764993570419715, "grad_norm": 1.1517022848129272, "learning_rate": 6.267566048341765e-06, "loss": 0.2945, "step": 893 }, { "epoch": 0.037692096886396693, "grad_norm": 1.7743033170700073, "learning_rate": 6.274592467678471e-06, "loss": 0.1994, "step": 894 }, { "epoch": 0.037734258068596246, "grad_norm": 1.0855284929275513, "learning_rate": 6.281618887015177e-06, "loss": 0.44, "step": 895 }, { "epoch": 0.03777641925079579, "grad_norm": 3.4053752422332764, "learning_rate": 6.288645306351883e-06, "loss": 1.0355, "step": 896 }, { "epoch": 0.03781858043299534, "grad_norm": 1.0806801319122314, "learning_rate": 6.29567172568859e-06, "loss": 0.7625, "step": 897 }, { "epoch": 0.03786074161519489, "grad_norm": 1.416821002960205, "learning_rate": 6.302698145025296e-06, "loss": 0.3505, "step": 898 }, { "epoch": 0.03790290279739444, "grad_norm": 1.4637300968170166, "learning_rate": 6.309724564362002e-06, "loss": 0.4529, "step": 899 }, { "epoch": 0.03794506397959399, "grad_norm": 1.344240665435791, "learning_rate": 6.316750983698707e-06, "loss": 0.5298, "step": 900 }, { "epoch": 0.037987225161793536, "grad_norm": 1.219529390335083, "learning_rate": 6.323777403035413e-06, "loss": 0.42, "step": 901 }, { "epoch": 0.03802938634399308, "grad_norm": 1.3811469078063965, "learning_rate": 6.3308038223721195e-06, "loss": 0.2869, "step": 902 }, { "epoch": 0.038071547526192635, "grad_norm": 1.41340172290802, "learning_rate": 6.3378302417088256e-06, "loss": 0.4874, "step": 903 }, { "epoch": 0.03811370870839218, "grad_norm": 1.5951117277145386, "learning_rate": 6.344856661045531e-06, "loss": 0.3836, "step": 904 }, { "epoch": 0.038155869890591734, "grad_norm": 1.2155559062957764, "learning_rate": 6.351883080382237e-06, "loss": 0.3391, "step": 905 }, { "epoch": 0.03819803107279128, "grad_norm": 0.96473228931427, "learning_rate": 6.358909499718943e-06, "loss": 0.431, "step": 906 }, { "epoch": 0.03824019225499083, "grad_norm": 0.9362431168556213, "learning_rate": 6.365935919055649e-06, "loss": 0.2279, "step": 907 }, { "epoch": 0.03828235343719038, "grad_norm": 1.3570897579193115, "learning_rate": 6.372962338392356e-06, "loss": 0.4458, "step": 908 }, { "epoch": 0.038324514619389925, "grad_norm": 1.4160271883010864, "learning_rate": 6.379988757729062e-06, "loss": 0.6961, "step": 909 }, { "epoch": 0.03836667580158948, "grad_norm": 1.1530215740203857, "learning_rate": 6.387015177065768e-06, "loss": 0.4141, "step": 910 }, { "epoch": 0.038408836983789024, "grad_norm": 2.7842044830322266, "learning_rate": 6.3940415964024734e-06, "loss": 0.6486, "step": 911 }, { "epoch": 0.03845099816598858, "grad_norm": 2.3755416870117188, "learning_rate": 6.4010680157391795e-06, "loss": 0.9409, "step": 912 }, { "epoch": 0.03849315934818812, "grad_norm": 1.7784771919250488, "learning_rate": 6.408094435075886e-06, "loss": 0.6505, "step": 913 }, { "epoch": 0.03853532053038767, "grad_norm": 1.3027615547180176, "learning_rate": 6.415120854412592e-06, "loss": 0.3378, "step": 914 }, { "epoch": 0.03857748171258722, "grad_norm": 1.3348805904388428, "learning_rate": 6.422147273749298e-06, "loss": 0.3049, "step": 915 }, { "epoch": 0.03861964289478677, "grad_norm": 1.3548800945281982, "learning_rate": 6.429173693086003e-06, "loss": 0.3535, "step": 916 }, { "epoch": 0.03866180407698632, "grad_norm": 2.3435866832733154, "learning_rate": 6.436200112422709e-06, "loss": 1.2378, "step": 917 }, { "epoch": 0.03870396525918587, "grad_norm": 1.1227998733520508, "learning_rate": 6.443226531759415e-06, "loss": 0.3011, "step": 918 }, { "epoch": 0.03874612644138541, "grad_norm": 3.3043136596679688, "learning_rate": 6.450252951096122e-06, "loss": 0.4922, "step": 919 }, { "epoch": 0.038788287623584966, "grad_norm": 1.3203891515731812, "learning_rate": 6.457279370432828e-06, "loss": 0.9533, "step": 920 }, { "epoch": 0.03883044880578451, "grad_norm": 1.3922054767608643, "learning_rate": 6.464305789769534e-06, "loss": 0.9316, "step": 921 }, { "epoch": 0.038872609987984065, "grad_norm": 1.241379976272583, "learning_rate": 6.47133220910624e-06, "loss": 0.3126, "step": 922 }, { "epoch": 0.03891477117018361, "grad_norm": 1.7146387100219727, "learning_rate": 6.478358628442946e-06, "loss": 0.5267, "step": 923 }, { "epoch": 0.038956932352383164, "grad_norm": 1.9210695028305054, "learning_rate": 6.485385047779652e-06, "loss": 0.2496, "step": 924 }, { "epoch": 0.03899909353458271, "grad_norm": 1.2400544881820679, "learning_rate": 6.492411467116358e-06, "loss": 0.3358, "step": 925 }, { "epoch": 0.039041254716782256, "grad_norm": 1.725960373878479, "learning_rate": 6.499437886453064e-06, "loss": 0.7102, "step": 926 }, { "epoch": 0.03908341589898181, "grad_norm": 1.5507935285568237, "learning_rate": 6.506464305789769e-06, "loss": 0.5068, "step": 927 }, { "epoch": 0.039125577081181355, "grad_norm": 1.6655462980270386, "learning_rate": 6.513490725126475e-06, "loss": 0.1873, "step": 928 }, { "epoch": 0.03916773826338091, "grad_norm": 1.5826166868209839, "learning_rate": 6.520517144463181e-06, "loss": 0.3321, "step": 929 }, { "epoch": 0.039209899445580454, "grad_norm": 1.2874232530593872, "learning_rate": 6.527543563799888e-06, "loss": 0.2518, "step": 930 }, { "epoch": 0.03925206062778, "grad_norm": 2.3786568641662598, "learning_rate": 6.534569983136594e-06, "loss": 1.382, "step": 931 }, { "epoch": 0.03929422180997955, "grad_norm": 1.7865623235702515, "learning_rate": 6.5415964024733005e-06, "loss": 0.4249, "step": 932 }, { "epoch": 0.0393363829921791, "grad_norm": 1.7334256172180176, "learning_rate": 6.5486228218100066e-06, "loss": 1.1784, "step": 933 }, { "epoch": 0.03937854417437865, "grad_norm": 1.8816149234771729, "learning_rate": 6.555649241146712e-06, "loss": 0.3129, "step": 934 }, { "epoch": 0.0394207053565782, "grad_norm": 1.236637830734253, "learning_rate": 6.562675660483418e-06, "loss": 0.3388, "step": 935 }, { "epoch": 0.039462866538777744, "grad_norm": 1.454307198524475, "learning_rate": 6.569702079820124e-06, "loss": 0.8897, "step": 936 }, { "epoch": 0.0395050277209773, "grad_norm": 1.103830099105835, "learning_rate": 6.57672849915683e-06, "loss": 0.1906, "step": 937 }, { "epoch": 0.03954718890317684, "grad_norm": 1.8840138912200928, "learning_rate": 6.583754918493535e-06, "loss": 0.5095, "step": 938 }, { "epoch": 0.039589350085376396, "grad_norm": 1.7100461721420288, "learning_rate": 6.590781337830241e-06, "loss": 1.2952, "step": 939 }, { "epoch": 0.03963151126757594, "grad_norm": 2.966034412384033, "learning_rate": 6.5978077571669475e-06, "loss": 1.3401, "step": 940 }, { "epoch": 0.039673672449775495, "grad_norm": 1.2899831533432007, "learning_rate": 6.6048341765036536e-06, "loss": 0.728, "step": 941 }, { "epoch": 0.03971583363197504, "grad_norm": 1.5556449890136719, "learning_rate": 6.6118605958403605e-06, "loss": 0.6855, "step": 942 }, { "epoch": 0.03975799481417459, "grad_norm": 1.8210214376449585, "learning_rate": 6.618887015177067e-06, "loss": 0.4839, "step": 943 }, { "epoch": 0.03980015599637414, "grad_norm": 1.679740309715271, "learning_rate": 6.625913434513773e-06, "loss": 0.2522, "step": 944 }, { "epoch": 0.039842317178573686, "grad_norm": 1.04052734375, "learning_rate": 6.632939853850478e-06, "loss": 0.6236, "step": 945 }, { "epoch": 0.03988447836077324, "grad_norm": 1.7853015661239624, "learning_rate": 6.639966273187184e-06, "loss": 0.3493, "step": 946 }, { "epoch": 0.039926639542972785, "grad_norm": 1.010725736618042, "learning_rate": 6.64699269252389e-06, "loss": 0.3235, "step": 947 }, { "epoch": 0.03996880072517233, "grad_norm": 0.9533209204673767, "learning_rate": 6.654019111860596e-06, "loss": 0.4554, "step": 948 }, { "epoch": 0.040010961907371884, "grad_norm": 1.273255705833435, "learning_rate": 6.661045531197302e-06, "loss": 0.3518, "step": 949 }, { "epoch": 0.04005312308957143, "grad_norm": 1.6803784370422363, "learning_rate": 6.6680719505340075e-06, "loss": 0.4874, "step": 950 }, { "epoch": 0.04009528427177098, "grad_norm": 1.7515426874160767, "learning_rate": 6.675098369870714e-06, "loss": 0.4588, "step": 951 }, { "epoch": 0.04013744545397053, "grad_norm": 1.27505624294281, "learning_rate": 6.68212478920742e-06, "loss": 0.3361, "step": 952 }, { "epoch": 0.04017960663617008, "grad_norm": 1.9163579940795898, "learning_rate": 6.689151208544127e-06, "loss": 0.9432, "step": 953 }, { "epoch": 0.04022176781836963, "grad_norm": 1.269455075263977, "learning_rate": 6.696177627880833e-06, "loss": 0.5871, "step": 954 }, { "epoch": 0.040263929000569174, "grad_norm": 1.2105664014816284, "learning_rate": 6.703204047217539e-06, "loss": 0.687, "step": 955 }, { "epoch": 0.040306090182768727, "grad_norm": 1.3454768657684326, "learning_rate": 6.710230466554245e-06, "loss": 0.7573, "step": 956 }, { "epoch": 0.04034825136496827, "grad_norm": 1.5538028478622437, "learning_rate": 6.71725688589095e-06, "loss": 0.7929, "step": 957 }, { "epoch": 0.040390412547167825, "grad_norm": 1.9180506467819214, "learning_rate": 6.724283305227656e-06, "loss": 0.3794, "step": 958 }, { "epoch": 0.04043257372936737, "grad_norm": 1.2038573026657104, "learning_rate": 6.731309724564362e-06, "loss": 0.5013, "step": 959 }, { "epoch": 0.04047473491156692, "grad_norm": 1.8348650932312012, "learning_rate": 6.738336143901068e-06, "loss": 0.3524, "step": 960 }, { "epoch": 0.04051689609376647, "grad_norm": 1.6746485233306885, "learning_rate": 6.745362563237774e-06, "loss": 0.8912, "step": 961 }, { "epoch": 0.040559057275966016, "grad_norm": 3.064556121826172, "learning_rate": 6.75238898257448e-06, "loss": 0.4339, "step": 962 }, { "epoch": 0.04060121845816557, "grad_norm": 1.9949438571929932, "learning_rate": 6.759415401911186e-06, "loss": 0.482, "step": 963 }, { "epoch": 0.040643379640365115, "grad_norm": 1.2086482048034668, "learning_rate": 6.766441821247893e-06, "loss": 0.3272, "step": 964 }, { "epoch": 0.04068554082256466, "grad_norm": 1.076113224029541, "learning_rate": 6.773468240584599e-06, "loss": 0.4664, "step": 965 }, { "epoch": 0.040727702004764214, "grad_norm": 1.3651829957962036, "learning_rate": 6.780494659921305e-06, "loss": 0.681, "step": 966 }, { "epoch": 0.04076986318696376, "grad_norm": 1.8692117929458618, "learning_rate": 6.787521079258011e-06, "loss": 0.5017, "step": 967 }, { "epoch": 0.04081202436916331, "grad_norm": 1.3985631465911865, "learning_rate": 6.794547498594716e-06, "loss": 0.3366, "step": 968 }, { "epoch": 0.04085418555136286, "grad_norm": 0.9384720921516418, "learning_rate": 6.801573917931422e-06, "loss": 0.2775, "step": 969 }, { "epoch": 0.04089634673356241, "grad_norm": 1.2334085702896118, "learning_rate": 6.8086003372681285e-06, "loss": 0.4064, "step": 970 }, { "epoch": 0.04093850791576196, "grad_norm": 1.2833694219589233, "learning_rate": 6.8156267566048346e-06, "loss": 0.4948, "step": 971 }, { "epoch": 0.040980669097961504, "grad_norm": 1.009611964225769, "learning_rate": 6.82265317594154e-06, "loss": 0.2465, "step": 972 }, { "epoch": 0.04102283028016106, "grad_norm": 4.219306945800781, "learning_rate": 6.829679595278246e-06, "loss": 0.814, "step": 973 }, { "epoch": 0.0410649914623606, "grad_norm": 1.742983102798462, "learning_rate": 6.836706014614952e-06, "loss": 0.5178, "step": 974 }, { "epoch": 0.041107152644560156, "grad_norm": 1.9818884134292603, "learning_rate": 6.843732433951659e-06, "loss": 0.3489, "step": 975 }, { "epoch": 0.0411493138267597, "grad_norm": 1.2811986207962036, "learning_rate": 6.850758853288365e-06, "loss": 0.7414, "step": 976 }, { "epoch": 0.04119147500895925, "grad_norm": 2.060426712036133, "learning_rate": 6.857785272625071e-06, "loss": 0.2938, "step": 977 }, { "epoch": 0.0412336361911588, "grad_norm": 1.2986096143722534, "learning_rate": 6.864811691961777e-06, "loss": 1.3719, "step": 978 }, { "epoch": 0.04127579737335835, "grad_norm": 1.105978012084961, "learning_rate": 6.871838111298482e-06, "loss": 0.3621, "step": 979 }, { "epoch": 0.0413179585555579, "grad_norm": 4.408673286437988, "learning_rate": 6.8788645306351885e-06, "loss": 1.1491, "step": 980 }, { "epoch": 0.041360119737757446, "grad_norm": 1.366074800491333, "learning_rate": 6.885890949971895e-06, "loss": 0.5905, "step": 981 }, { "epoch": 0.04140228091995699, "grad_norm": 2.0720815658569336, "learning_rate": 6.892917369308601e-06, "loss": 1.1996, "step": 982 }, { "epoch": 0.041444442102156545, "grad_norm": 1.097666621208191, "learning_rate": 6.899943788645307e-06, "loss": 0.7081, "step": 983 }, { "epoch": 0.04148660328435609, "grad_norm": 2.454022169113159, "learning_rate": 6.906970207982012e-06, "loss": 1.3543, "step": 984 }, { "epoch": 0.041528764466555644, "grad_norm": 1.7964640855789185, "learning_rate": 6.913996627318718e-06, "loss": 1.3708, "step": 985 }, { "epoch": 0.04157092564875519, "grad_norm": 1.2597754001617432, "learning_rate": 6.921023046655424e-06, "loss": 0.7413, "step": 986 }, { "epoch": 0.04161308683095474, "grad_norm": 1.6903427839279175, "learning_rate": 6.928049465992131e-06, "loss": 0.3725, "step": 987 }, { "epoch": 0.04165524801315429, "grad_norm": 1.7425850629806519, "learning_rate": 6.935075885328837e-06, "loss": 0.4246, "step": 988 }, { "epoch": 0.041697409195353835, "grad_norm": 1.4564940929412842, "learning_rate": 6.942102304665543e-06, "loss": 0.644, "step": 989 }, { "epoch": 0.04173957037755339, "grad_norm": 1.815222144126892, "learning_rate": 6.949128724002249e-06, "loss": 0.64, "step": 990 }, { "epoch": 0.041781731559752934, "grad_norm": 2.282236099243164, "learning_rate": 6.956155143338955e-06, "loss": 0.5965, "step": 991 }, { "epoch": 0.04182389274195249, "grad_norm": 3.524174928665161, "learning_rate": 6.963181562675661e-06, "loss": 0.9599, "step": 992 }, { "epoch": 0.04186605392415203, "grad_norm": 1.768968105316162, "learning_rate": 6.970207982012367e-06, "loss": 0.2557, "step": 993 }, { "epoch": 0.04190821510635158, "grad_norm": 2.0018999576568604, "learning_rate": 6.977234401349073e-06, "loss": 1.2374, "step": 994 }, { "epoch": 0.04195037628855113, "grad_norm": 1.5031346082687378, "learning_rate": 6.984260820685778e-06, "loss": 0.4926, "step": 995 }, { "epoch": 0.04199253747075068, "grad_norm": 1.0482735633850098, "learning_rate": 6.991287240022484e-06, "loss": 0.2635, "step": 996 }, { "epoch": 0.04203469865295023, "grad_norm": 1.7747528553009033, "learning_rate": 6.99831365935919e-06, "loss": 0.4517, "step": 997 }, { "epoch": 0.04207685983514978, "grad_norm": 1.267944574356079, "learning_rate": 7.005340078695897e-06, "loss": 0.1899, "step": 998 }, { "epoch": 0.04211902101734933, "grad_norm": 1.8609451055526733, "learning_rate": 7.012366498032603e-06, "loss": 0.6656, "step": 999 }, { "epoch": 0.042161182199548876, "grad_norm": 1.906689167022705, "learning_rate": 7.0193929173693094e-06, "loss": 0.4812, "step": 1000 }, { "epoch": 0.04220334338174842, "grad_norm": 1.1601983308792114, "learning_rate": 7.0264193367060155e-06, "loss": 0.2327, "step": 1001 }, { "epoch": 0.042245504563947975, "grad_norm": 1.5886507034301758, "learning_rate": 7.033445756042721e-06, "loss": 0.449, "step": 1002 }, { "epoch": 0.04228766574614752, "grad_norm": 1.249035358428955, "learning_rate": 7.040472175379427e-06, "loss": 0.8946, "step": 1003 }, { "epoch": 0.042329826928347074, "grad_norm": 1.4004759788513184, "learning_rate": 7.047498594716133e-06, "loss": 0.3647, "step": 1004 }, { "epoch": 0.04237198811054662, "grad_norm": 1.3023096323013306, "learning_rate": 7.054525014052839e-06, "loss": 0.4126, "step": 1005 }, { "epoch": 0.042414149292746166, "grad_norm": 1.237644910812378, "learning_rate": 7.061551433389544e-06, "loss": 0.4293, "step": 1006 }, { "epoch": 0.04245631047494572, "grad_norm": 1.783080816268921, "learning_rate": 7.06857785272625e-06, "loss": 0.312, "step": 1007 }, { "epoch": 0.042498471657145265, "grad_norm": 1.7641983032226562, "learning_rate": 7.0756042720629565e-06, "loss": 0.5286, "step": 1008 }, { "epoch": 0.04254063283934482, "grad_norm": 5.957270622253418, "learning_rate": 7.082630691399663e-06, "loss": 0.178, "step": 1009 }, { "epoch": 0.042582794021544364, "grad_norm": 1.6377474069595337, "learning_rate": 7.0896571107363695e-06, "loss": 0.7491, "step": 1010 }, { "epoch": 0.04262495520374391, "grad_norm": 4.895644664764404, "learning_rate": 7.096683530073076e-06, "loss": 1.0788, "step": 1011 }, { "epoch": 0.04266711638594346, "grad_norm": 2.469578266143799, "learning_rate": 7.103709949409782e-06, "loss": 0.5232, "step": 1012 }, { "epoch": 0.04270927756814301, "grad_norm": 1.8156929016113281, "learning_rate": 7.110736368746487e-06, "loss": 1.2222, "step": 1013 }, { "epoch": 0.04275143875034256, "grad_norm": 2.060593366622925, "learning_rate": 7.117762788083193e-06, "loss": 0.6379, "step": 1014 }, { "epoch": 0.04279359993254211, "grad_norm": 4.535948753356934, "learning_rate": 7.124789207419899e-06, "loss": 1.4308, "step": 1015 }, { "epoch": 0.04283576111474166, "grad_norm": 1.7831075191497803, "learning_rate": 7.131815626756605e-06, "loss": 0.3618, "step": 1016 }, { "epoch": 0.04287792229694121, "grad_norm": 1.8455501794815063, "learning_rate": 7.138842046093311e-06, "loss": 0.7737, "step": 1017 }, { "epoch": 0.04292008347914075, "grad_norm": 1.4251656532287598, "learning_rate": 7.1458684654300165e-06, "loss": 0.8718, "step": 1018 }, { "epoch": 0.042962244661340306, "grad_norm": 1.2022640705108643, "learning_rate": 7.152894884766723e-06, "loss": 0.1908, "step": 1019 }, { "epoch": 0.04300440584353985, "grad_norm": 1.172989010810852, "learning_rate": 7.1599213041034295e-06, "loss": 0.7636, "step": 1020 }, { "epoch": 0.043046567025739405, "grad_norm": 1.8079180717468262, "learning_rate": 7.166947723440136e-06, "loss": 0.6192, "step": 1021 }, { "epoch": 0.04308872820793895, "grad_norm": 1.3286118507385254, "learning_rate": 7.173974142776842e-06, "loss": 0.5204, "step": 1022 }, { "epoch": 0.0431308893901385, "grad_norm": 1.6683731079101562, "learning_rate": 7.181000562113548e-06, "loss": 1.2619, "step": 1023 }, { "epoch": 0.04317305057233805, "grad_norm": 0.9985883235931396, "learning_rate": 7.188026981450254e-06, "loss": 0.2659, "step": 1024 }, { "epoch": 0.043215211754537596, "grad_norm": 1.2359994649887085, "learning_rate": 7.195053400786959e-06, "loss": 0.9165, "step": 1025 }, { "epoch": 0.04325737293673715, "grad_norm": 4.6300835609436035, "learning_rate": 7.202079820123665e-06, "loss": 1.2746, "step": 1026 }, { "epoch": 0.043299534118936694, "grad_norm": 1.2137240171432495, "learning_rate": 7.209106239460371e-06, "loss": 0.3147, "step": 1027 }, { "epoch": 0.04334169530113624, "grad_norm": 1.3816229104995728, "learning_rate": 7.216132658797077e-06, "loss": 0.4198, "step": 1028 }, { "epoch": 0.04338385648333579, "grad_norm": 1.7320383787155151, "learning_rate": 7.223159078133783e-06, "loss": 0.5808, "step": 1029 }, { "epoch": 0.04342601766553534, "grad_norm": 1.4037796258926392, "learning_rate": 7.230185497470489e-06, "loss": 0.8036, "step": 1030 }, { "epoch": 0.04346817884773489, "grad_norm": 1.138725757598877, "learning_rate": 7.237211916807195e-06, "loss": 0.268, "step": 1031 }, { "epoch": 0.04351034002993444, "grad_norm": 1.2624071836471558, "learning_rate": 7.244238336143902e-06, "loss": 0.176, "step": 1032 }, { "epoch": 0.04355250121213399, "grad_norm": 1.2927238941192627, "learning_rate": 7.251264755480608e-06, "loss": 0.5862, "step": 1033 }, { "epoch": 0.04359466239433354, "grad_norm": 1.1908035278320312, "learning_rate": 7.258291174817314e-06, "loss": 0.8683, "step": 1034 }, { "epoch": 0.04363682357653308, "grad_norm": 2.3194687366485596, "learning_rate": 7.26531759415402e-06, "loss": 0.6062, "step": 1035 }, { "epoch": 0.043678984758732636, "grad_norm": 1.7395912408828735, "learning_rate": 7.272344013490725e-06, "loss": 0.2541, "step": 1036 }, { "epoch": 0.04372114594093218, "grad_norm": 1.2964140176773071, "learning_rate": 7.279370432827431e-06, "loss": 0.3394, "step": 1037 }, { "epoch": 0.043763307123131735, "grad_norm": 2.4319725036621094, "learning_rate": 7.2863968521641374e-06, "loss": 0.33, "step": 1038 }, { "epoch": 0.04380546830533128, "grad_norm": 1.166049838066101, "learning_rate": 7.2934232715008435e-06, "loss": 0.7258, "step": 1039 }, { "epoch": 0.04384762948753083, "grad_norm": 1.4341591596603394, "learning_rate": 7.300449690837549e-06, "loss": 0.3964, "step": 1040 }, { "epoch": 0.04388979066973038, "grad_norm": 1.6908578872680664, "learning_rate": 7.307476110174255e-06, "loss": 0.2903, "step": 1041 }, { "epoch": 0.043931951851929926, "grad_norm": 1.0962231159210205, "learning_rate": 7.314502529510961e-06, "loss": 0.546, "step": 1042 }, { "epoch": 0.04397411303412948, "grad_norm": 2.087338447570801, "learning_rate": 7.321528948847668e-06, "loss": 0.3078, "step": 1043 }, { "epoch": 0.044016274216329025, "grad_norm": 1.0675804615020752, "learning_rate": 7.328555368184374e-06, "loss": 0.3004, "step": 1044 }, { "epoch": 0.04405843539852858, "grad_norm": 1.4058401584625244, "learning_rate": 7.33558178752108e-06, "loss": 0.7482, "step": 1045 }, { "epoch": 0.044100596580728124, "grad_norm": 1.0940971374511719, "learning_rate": 7.342608206857786e-06, "loss": 0.2892, "step": 1046 }, { "epoch": 0.04414275776292767, "grad_norm": 1.1179842948913574, "learning_rate": 7.349634626194491e-06, "loss": 0.7621, "step": 1047 }, { "epoch": 0.04418491894512722, "grad_norm": 0.9562937617301941, "learning_rate": 7.3566610455311975e-06, "loss": 0.2017, "step": 1048 }, { "epoch": 0.04422708012732677, "grad_norm": 2.065741777420044, "learning_rate": 7.363687464867904e-06, "loss": 1.2284, "step": 1049 }, { "epoch": 0.04426924130952632, "grad_norm": 1.3736014366149902, "learning_rate": 7.37071388420461e-06, "loss": 0.7004, "step": 1050 }, { "epoch": 0.04431140249172587, "grad_norm": 1.4621387720108032, "learning_rate": 7.377740303541316e-06, "loss": 0.3885, "step": 1051 }, { "epoch": 0.044353563673925414, "grad_norm": 1.571630597114563, "learning_rate": 7.384766722878021e-06, "loss": 0.2868, "step": 1052 }, { "epoch": 0.04439572485612497, "grad_norm": 1.279402494430542, "learning_rate": 7.391793142214727e-06, "loss": 0.3517, "step": 1053 }, { "epoch": 0.04443788603832451, "grad_norm": 1.6892627477645874, "learning_rate": 7.398819561551434e-06, "loss": 0.4048, "step": 1054 }, { "epoch": 0.044480047220524066, "grad_norm": 1.238236904144287, "learning_rate": 7.40584598088814e-06, "loss": 0.1931, "step": 1055 }, { "epoch": 0.04452220840272361, "grad_norm": 1.935970664024353, "learning_rate": 7.412872400224846e-06, "loss": 1.2461, "step": 1056 }, { "epoch": 0.04456436958492316, "grad_norm": 1.397702932357788, "learning_rate": 7.419898819561552e-06, "loss": 0.2495, "step": 1057 }, { "epoch": 0.04460653076712271, "grad_norm": 1.4698858261108398, "learning_rate": 7.426925238898258e-06, "loss": 0.2915, "step": 1058 }, { "epoch": 0.04464869194932226, "grad_norm": 1.9516552686691284, "learning_rate": 7.433951658234964e-06, "loss": 0.501, "step": 1059 }, { "epoch": 0.04469085313152181, "grad_norm": 1.9591681957244873, "learning_rate": 7.44097807757167e-06, "loss": 0.6035, "step": 1060 }, { "epoch": 0.044733014313721356, "grad_norm": 1.6081504821777344, "learning_rate": 7.448004496908376e-06, "loss": 1.4149, "step": 1061 }, { "epoch": 0.04477517549592091, "grad_norm": 1.5127315521240234, "learning_rate": 7.455030916245082e-06, "loss": 0.6901, "step": 1062 }, { "epoch": 0.044817336678120455, "grad_norm": 2.5906527042388916, "learning_rate": 7.462057335581787e-06, "loss": 0.359, "step": 1063 }, { "epoch": 0.04485949786032, "grad_norm": 1.228767991065979, "learning_rate": 7.469083754918493e-06, "loss": 0.3602, "step": 1064 }, { "epoch": 0.044901659042519554, "grad_norm": 1.1951590776443481, "learning_rate": 7.476110174255199e-06, "loss": 0.1966, "step": 1065 }, { "epoch": 0.0449438202247191, "grad_norm": 1.4105260372161865, "learning_rate": 7.483136593591906e-06, "loss": 0.221, "step": 1066 }, { "epoch": 0.04498598140691865, "grad_norm": 2.864956855773926, "learning_rate": 7.490163012928612e-06, "loss": 0.8749, "step": 1067 }, { "epoch": 0.0450281425891182, "grad_norm": 1.331316590309143, "learning_rate": 7.4971894322653184e-06, "loss": 0.2999, "step": 1068 }, { "epoch": 0.045070303771317745, "grad_norm": 1.3778735399246216, "learning_rate": 7.5042158516020245e-06, "loss": 0.6749, "step": 1069 }, { "epoch": 0.0451124649535173, "grad_norm": 1.6131253242492676, "learning_rate": 7.51124227093873e-06, "loss": 0.3587, "step": 1070 }, { "epoch": 0.045154626135716844, "grad_norm": 1.380683422088623, "learning_rate": 7.518268690275436e-06, "loss": 0.3766, "step": 1071 }, { "epoch": 0.0451967873179164, "grad_norm": 2.1984286308288574, "learning_rate": 7.525295109612142e-06, "loss": 1.0505, "step": 1072 }, { "epoch": 0.04523894850011594, "grad_norm": 1.6607657670974731, "learning_rate": 7.532321528948848e-06, "loss": 0.2352, "step": 1073 }, { "epoch": 0.04528110968231549, "grad_norm": 3.89013409614563, "learning_rate": 7.539347948285553e-06, "loss": 1.4597, "step": 1074 }, { "epoch": 0.04532327086451504, "grad_norm": 1.8222886323928833, "learning_rate": 7.546374367622259e-06, "loss": 0.6555, "step": 1075 }, { "epoch": 0.04536543204671459, "grad_norm": 1.4334282875061035, "learning_rate": 7.5534007869589655e-06, "loss": 0.8226, "step": 1076 }, { "epoch": 0.04540759322891414, "grad_norm": 3.9131507873535156, "learning_rate": 7.560427206295672e-06, "loss": 0.9164, "step": 1077 }, { "epoch": 0.04544975441111369, "grad_norm": 1.1299448013305664, "learning_rate": 7.5674536256323785e-06, "loss": 0.4271, "step": 1078 }, { "epoch": 0.04549191559331324, "grad_norm": 2.446009874343872, "learning_rate": 7.5744800449690846e-06, "loss": 1.1032, "step": 1079 }, { "epoch": 0.045534076775512786, "grad_norm": 1.2657387256622314, "learning_rate": 7.581506464305791e-06, "loss": 0.2248, "step": 1080 }, { "epoch": 0.04557623795771233, "grad_norm": 1.857369065284729, "learning_rate": 7.588532883642496e-06, "loss": 1.1945, "step": 1081 }, { "epoch": 0.045618399139911885, "grad_norm": 2.748706579208374, "learning_rate": 7.595559302979202e-06, "loss": 0.8063, "step": 1082 }, { "epoch": 0.04566056032211143, "grad_norm": 1.1701796054840088, "learning_rate": 7.602585722315908e-06, "loss": 0.2121, "step": 1083 }, { "epoch": 0.045702721504310984, "grad_norm": 1.4875092506408691, "learning_rate": 7.609612141652614e-06, "loss": 0.3887, "step": 1084 }, { "epoch": 0.04574488268651053, "grad_norm": 1.3819332122802734, "learning_rate": 7.616638560989319e-06, "loss": 0.2733, "step": 1085 }, { "epoch": 0.045787043868710076, "grad_norm": 1.5295535326004028, "learning_rate": 7.6236649803260255e-06, "loss": 0.2101, "step": 1086 }, { "epoch": 0.04582920505090963, "grad_norm": 1.5508413314819336, "learning_rate": 7.630691399662732e-06, "loss": 0.9269, "step": 1087 }, { "epoch": 0.045871366233109175, "grad_norm": 1.8231744766235352, "learning_rate": 7.63771781899944e-06, "loss": 1.3186, "step": 1088 }, { "epoch": 0.04591352741530873, "grad_norm": 2.130164623260498, "learning_rate": 7.644744238336145e-06, "loss": 0.314, "step": 1089 }, { "epoch": 0.045955688597508274, "grad_norm": 1.8341094255447388, "learning_rate": 7.65177065767285e-06, "loss": 0.3002, "step": 1090 }, { "epoch": 0.045997849779707826, "grad_norm": 3.0572831630706787, "learning_rate": 7.658797077009556e-06, "loss": 0.6176, "step": 1091 }, { "epoch": 0.04604001096190737, "grad_norm": 5.696110248565674, "learning_rate": 7.665823496346262e-06, "loss": 1.6057, "step": 1092 }, { "epoch": 0.04608217214410692, "grad_norm": 1.1245245933532715, "learning_rate": 7.672849915682968e-06, "loss": 0.558, "step": 1093 }, { "epoch": 0.04612433332630647, "grad_norm": 1.7663133144378662, "learning_rate": 7.679876335019674e-06, "loss": 0.4784, "step": 1094 }, { "epoch": 0.04616649450850602, "grad_norm": 1.1142815351486206, "learning_rate": 7.68690275435638e-06, "loss": 0.2302, "step": 1095 }, { "epoch": 0.04620865569070557, "grad_norm": 1.4743399620056152, "learning_rate": 7.693929173693086e-06, "loss": 0.3508, "step": 1096 }, { "epoch": 0.046250816872905116, "grad_norm": 3.885040521621704, "learning_rate": 7.700955593029792e-06, "loss": 1.0149, "step": 1097 }, { "epoch": 0.04629297805510466, "grad_norm": 2.0303492546081543, "learning_rate": 7.707982012366499e-06, "loss": 0.3022, "step": 1098 }, { "epoch": 0.046335139237304215, "grad_norm": 2.0907652378082275, "learning_rate": 7.715008431703205e-06, "loss": 0.4109, "step": 1099 }, { "epoch": 0.04637730041950376, "grad_norm": 1.9842629432678223, "learning_rate": 7.72203485103991e-06, "loss": 1.2897, "step": 1100 }, { "epoch": 0.046419461601703314, "grad_norm": 1.137969970703125, "learning_rate": 7.729061270376617e-06, "loss": 0.1511, "step": 1101 }, { "epoch": 0.04646162278390286, "grad_norm": 1.8940831422805786, "learning_rate": 7.736087689713323e-06, "loss": 0.4729, "step": 1102 }, { "epoch": 0.046503783966102406, "grad_norm": 1.3407858610153198, "learning_rate": 7.743114109050029e-06, "loss": 0.2127, "step": 1103 }, { "epoch": 0.04654594514830196, "grad_norm": 1.755263328552246, "learning_rate": 7.750140528386735e-06, "loss": 0.2068, "step": 1104 }, { "epoch": 0.046588106330501505, "grad_norm": 1.2864375114440918, "learning_rate": 7.757166947723441e-06, "loss": 0.6623, "step": 1105 }, { "epoch": 0.04663026751270106, "grad_norm": 1.6317973136901855, "learning_rate": 7.764193367060146e-06, "loss": 0.5535, "step": 1106 }, { "epoch": 0.046672428694900604, "grad_norm": 2.5037214756011963, "learning_rate": 7.771219786396852e-06, "loss": 0.4634, "step": 1107 }, { "epoch": 0.04671458987710016, "grad_norm": 0.9817124009132385, "learning_rate": 7.778246205733558e-06, "loss": 0.2161, "step": 1108 }, { "epoch": 0.0467567510592997, "grad_norm": 2.8957407474517822, "learning_rate": 7.785272625070264e-06, "loss": 0.6024, "step": 1109 }, { "epoch": 0.04679891224149925, "grad_norm": 1.2450400590896606, "learning_rate": 7.79229904440697e-06, "loss": 0.2906, "step": 1110 }, { "epoch": 0.0468410734236988, "grad_norm": 1.184675931930542, "learning_rate": 7.799325463743678e-06, "loss": 0.5449, "step": 1111 }, { "epoch": 0.04688323460589835, "grad_norm": 2.5737011432647705, "learning_rate": 7.806351883080384e-06, "loss": 0.5021, "step": 1112 }, { "epoch": 0.0469253957880979, "grad_norm": 1.095366358757019, "learning_rate": 7.813378302417088e-06, "loss": 0.5868, "step": 1113 }, { "epoch": 0.04696755697029745, "grad_norm": 1.2145016193389893, "learning_rate": 7.820404721753794e-06, "loss": 0.3142, "step": 1114 }, { "epoch": 0.04700971815249699, "grad_norm": 1.9990302324295044, "learning_rate": 7.8274311410905e-06, "loss": 0.2339, "step": 1115 }, { "epoch": 0.047051879334696546, "grad_norm": 2.1226861476898193, "learning_rate": 7.834457560427206e-06, "loss": 0.3446, "step": 1116 }, { "epoch": 0.04709404051689609, "grad_norm": 1.2972615957260132, "learning_rate": 7.841483979763913e-06, "loss": 0.7, "step": 1117 }, { "epoch": 0.047136201699095645, "grad_norm": 1.9866552352905273, "learning_rate": 7.848510399100619e-06, "loss": 0.375, "step": 1118 }, { "epoch": 0.04717836288129519, "grad_norm": 1.4595929384231567, "learning_rate": 7.855536818437325e-06, "loss": 0.5501, "step": 1119 }, { "epoch": 0.04722052406349474, "grad_norm": 1.2458776235580444, "learning_rate": 7.86256323777403e-06, "loss": 0.3361, "step": 1120 }, { "epoch": 0.04726268524569429, "grad_norm": 1.3252484798431396, "learning_rate": 7.869589657110737e-06, "loss": 0.3628, "step": 1121 }, { "epoch": 0.047304846427893836, "grad_norm": 1.2845587730407715, "learning_rate": 7.876616076447443e-06, "loss": 0.8405, "step": 1122 }, { "epoch": 0.04734700761009339, "grad_norm": 1.9510364532470703, "learning_rate": 7.883642495784149e-06, "loss": 0.458, "step": 1123 }, { "epoch": 0.047389168792292935, "grad_norm": 1.4263027906417847, "learning_rate": 7.890668915120855e-06, "loss": 0.2422, "step": 1124 }, { "epoch": 0.04743132997449249, "grad_norm": 0.9342586398124695, "learning_rate": 7.897695334457561e-06, "loss": 0.6748, "step": 1125 }, { "epoch": 0.047473491156692034, "grad_norm": 0.8664495348930359, "learning_rate": 7.904721753794267e-06, "loss": 0.3081, "step": 1126 }, { "epoch": 0.04751565233889158, "grad_norm": 1.7938369512557983, "learning_rate": 7.911748173130973e-06, "loss": 0.4326, "step": 1127 }, { "epoch": 0.04755781352109113, "grad_norm": 1.1822329759597778, "learning_rate": 7.91877459246768e-06, "loss": 0.5207, "step": 1128 }, { "epoch": 0.04759997470329068, "grad_norm": 1.8143631219863892, "learning_rate": 7.925801011804384e-06, "loss": 0.5561, "step": 1129 }, { "epoch": 0.04764213588549023, "grad_norm": 1.5523428916931152, "learning_rate": 7.93282743114109e-06, "loss": 0.4273, "step": 1130 }, { "epoch": 0.04768429706768978, "grad_norm": 1.771988868713379, "learning_rate": 7.939853850477796e-06, "loss": 1.4911, "step": 1131 }, { "epoch": 0.047726458249889324, "grad_norm": 3.0724310874938965, "learning_rate": 7.946880269814502e-06, "loss": 0.2076, "step": 1132 }, { "epoch": 0.04776861943208888, "grad_norm": 1.2098034620285034, "learning_rate": 7.95390668915121e-06, "loss": 0.6387, "step": 1133 }, { "epoch": 0.04781078061428842, "grad_norm": 1.3503488302230835, "learning_rate": 7.960933108487916e-06, "loss": 0.4724, "step": 1134 }, { "epoch": 0.047852941796487976, "grad_norm": 1.2094571590423584, "learning_rate": 7.967959527824622e-06, "loss": 0.7259, "step": 1135 }, { "epoch": 0.04789510297868752, "grad_norm": 2.953150749206543, "learning_rate": 7.974985947161327e-06, "loss": 1.2061, "step": 1136 }, { "epoch": 0.04793726416088707, "grad_norm": 1.5169070959091187, "learning_rate": 7.982012366498033e-06, "loss": 1.2422, "step": 1137 }, { "epoch": 0.04797942534308662, "grad_norm": 1.407329797744751, "learning_rate": 7.989038785834739e-06, "loss": 0.4741, "step": 1138 }, { "epoch": 0.04802158652528617, "grad_norm": 1.4308336973190308, "learning_rate": 7.996065205171445e-06, "loss": 0.6188, "step": 1139 }, { "epoch": 0.04806374770748572, "grad_norm": 1.1423250436782837, "learning_rate": 8.003091624508151e-06, "loss": 0.1985, "step": 1140 }, { "epoch": 0.048105908889685266, "grad_norm": 1.1360450983047485, "learning_rate": 8.010118043844857e-06, "loss": 0.1328, "step": 1141 }, { "epoch": 0.04814807007188482, "grad_norm": 1.0810821056365967, "learning_rate": 8.017144463181563e-06, "loss": 0.2027, "step": 1142 }, { "epoch": 0.048190231254084365, "grad_norm": 1.2416305541992188, "learning_rate": 8.02417088251827e-06, "loss": 0.2768, "step": 1143 }, { "epoch": 0.04823239243628391, "grad_norm": 1.1537754535675049, "learning_rate": 8.031197301854975e-06, "loss": 0.6114, "step": 1144 }, { "epoch": 0.048274553618483464, "grad_norm": 2.183936834335327, "learning_rate": 8.038223721191681e-06, "loss": 0.3129, "step": 1145 }, { "epoch": 0.04831671480068301, "grad_norm": 2.04573655128479, "learning_rate": 8.045250140528387e-06, "loss": 0.5518, "step": 1146 }, { "epoch": 0.04835887598288256, "grad_norm": 1.6284170150756836, "learning_rate": 8.052276559865094e-06, "loss": 0.2848, "step": 1147 }, { "epoch": 0.04840103716508211, "grad_norm": 1.0759027004241943, "learning_rate": 8.0593029792018e-06, "loss": 0.2769, "step": 1148 }, { "epoch": 0.048443198347281655, "grad_norm": 1.6106829643249512, "learning_rate": 8.066329398538506e-06, "loss": 0.369, "step": 1149 }, { "epoch": 0.04848535952948121, "grad_norm": 1.56550931930542, "learning_rate": 8.073355817875212e-06, "loss": 0.3211, "step": 1150 }, { "epoch": 0.048527520711680754, "grad_norm": 1.513877034187317, "learning_rate": 8.080382237211916e-06, "loss": 0.6985, "step": 1151 }, { "epoch": 0.04856968189388031, "grad_norm": 1.047068476676941, "learning_rate": 8.087408656548622e-06, "loss": 0.2093, "step": 1152 }, { "epoch": 0.04861184307607985, "grad_norm": 1.2494375705718994, "learning_rate": 8.094435075885328e-06, "loss": 0.5728, "step": 1153 }, { "epoch": 0.048654004258279406, "grad_norm": 1.2324718236923218, "learning_rate": 8.101461495222034e-06, "loss": 1.2229, "step": 1154 }, { "epoch": 0.04869616544047895, "grad_norm": 1.148698329925537, "learning_rate": 8.10848791455874e-06, "loss": 0.3459, "step": 1155 }, { "epoch": 0.0487383266226785, "grad_norm": 2.0867085456848145, "learning_rate": 8.115514333895448e-06, "loss": 1.2728, "step": 1156 }, { "epoch": 0.04878048780487805, "grad_norm": 1.8715708255767822, "learning_rate": 8.122540753232154e-06, "loss": 0.6024, "step": 1157 }, { "epoch": 0.048822648987077596, "grad_norm": 1.4458502531051636, "learning_rate": 8.129567172568859e-06, "loss": 0.731, "step": 1158 }, { "epoch": 0.04886481016927715, "grad_norm": 1.1426278352737427, "learning_rate": 8.136593591905565e-06, "loss": 0.2182, "step": 1159 }, { "epoch": 0.048906971351476695, "grad_norm": 1.6031323671340942, "learning_rate": 8.143620011242271e-06, "loss": 1.1872, "step": 1160 }, { "epoch": 0.04894913253367624, "grad_norm": 2.2044923305511475, "learning_rate": 8.150646430578977e-06, "loss": 0.6878, "step": 1161 }, { "epoch": 0.048991293715875794, "grad_norm": 1.3961584568023682, "learning_rate": 8.157672849915683e-06, "loss": 0.4542, "step": 1162 }, { "epoch": 0.04903345489807534, "grad_norm": 2.7100794315338135, "learning_rate": 8.16469926925239e-06, "loss": 1.0775, "step": 1163 }, { "epoch": 0.04907561608027489, "grad_norm": 1.5736992359161377, "learning_rate": 8.171725688589095e-06, "loss": 0.9394, "step": 1164 }, { "epoch": 0.04911777726247444, "grad_norm": 1.3036327362060547, "learning_rate": 8.178752107925801e-06, "loss": 0.8762, "step": 1165 }, { "epoch": 0.049159938444673985, "grad_norm": 1.6747138500213623, "learning_rate": 8.185778527262508e-06, "loss": 0.3347, "step": 1166 }, { "epoch": 0.04920209962687354, "grad_norm": 1.7027616500854492, "learning_rate": 8.192804946599214e-06, "loss": 0.5619, "step": 1167 }, { "epoch": 0.049244260809073084, "grad_norm": 1.5633643865585327, "learning_rate": 8.19983136593592e-06, "loss": 0.2904, "step": 1168 }, { "epoch": 0.04928642199127264, "grad_norm": 1.5155401229858398, "learning_rate": 8.206857785272626e-06, "loss": 1.3028, "step": 1169 }, { "epoch": 0.04932858317347218, "grad_norm": 1.8796577453613281, "learning_rate": 8.213884204609332e-06, "loss": 0.695, "step": 1170 }, { "epoch": 0.049370744355671736, "grad_norm": 1.3417179584503174, "learning_rate": 8.220910623946038e-06, "loss": 0.8055, "step": 1171 }, { "epoch": 0.04941290553787128, "grad_norm": 1.5632892847061157, "learning_rate": 8.227937043282744e-06, "loss": 0.2136, "step": 1172 }, { "epoch": 0.04945506672007083, "grad_norm": 1.4465909004211426, "learning_rate": 8.23496346261945e-06, "loss": 0.6205, "step": 1173 }, { "epoch": 0.04949722790227038, "grad_norm": 1.1098965406417847, "learning_rate": 8.241989881956155e-06, "loss": 0.297, "step": 1174 }, { "epoch": 0.04953938908446993, "grad_norm": 1.8335508108139038, "learning_rate": 8.24901630129286e-06, "loss": 0.2999, "step": 1175 }, { "epoch": 0.04958155026666948, "grad_norm": 1.3932517766952515, "learning_rate": 8.256042720629567e-06, "loss": 0.2447, "step": 1176 }, { "epoch": 0.049623711448869026, "grad_norm": 1.0906912088394165, "learning_rate": 8.263069139966273e-06, "loss": 0.7263, "step": 1177 }, { "epoch": 0.04966587263106857, "grad_norm": 1.7731130123138428, "learning_rate": 8.27009555930298e-06, "loss": 0.2795, "step": 1178 }, { "epoch": 0.049708033813268125, "grad_norm": 1.231270432472229, "learning_rate": 8.277121978639687e-06, "loss": 0.2479, "step": 1179 }, { "epoch": 0.04975019499546767, "grad_norm": 1.3190851211547852, "learning_rate": 8.284148397976393e-06, "loss": 0.3102, "step": 1180 }, { "epoch": 0.049792356177667224, "grad_norm": 0.8402856588363647, "learning_rate": 8.291174817313097e-06, "loss": 0.2158, "step": 1181 }, { "epoch": 0.04983451735986677, "grad_norm": 1.6577967405319214, "learning_rate": 8.298201236649803e-06, "loss": 0.4213, "step": 1182 }, { "epoch": 0.049876678542066316, "grad_norm": 0.9961962103843689, "learning_rate": 8.30522765598651e-06, "loss": 0.8045, "step": 1183 }, { "epoch": 0.04991883972426587, "grad_norm": 1.2622770071029663, "learning_rate": 8.312254075323215e-06, "loss": 0.7774, "step": 1184 }, { "epoch": 0.049961000906465415, "grad_norm": 2.057685375213623, "learning_rate": 8.319280494659922e-06, "loss": 1.1671, "step": 1185 }, { "epoch": 0.05000316208866497, "grad_norm": 0.9492290616035461, "learning_rate": 8.326306913996628e-06, "loss": 0.1765, "step": 1186 }, { "epoch": 0.050045323270864514, "grad_norm": 1.2844526767730713, "learning_rate": 8.333333333333334e-06, "loss": 0.5084, "step": 1187 }, { "epoch": 0.05008748445306407, "grad_norm": 3.9633233547210693, "learning_rate": 8.34035975267004e-06, "loss": 0.7602, "step": 1188 }, { "epoch": 0.05012964563526361, "grad_norm": 3.590808629989624, "learning_rate": 8.347386172006746e-06, "loss": 1.5937, "step": 1189 }, { "epoch": 0.05017180681746316, "grad_norm": 1.5765467882156372, "learning_rate": 8.354412591343452e-06, "loss": 0.3857, "step": 1190 }, { "epoch": 0.05021396799966271, "grad_norm": 1.3965893983840942, "learning_rate": 8.361439010680158e-06, "loss": 1.0274, "step": 1191 }, { "epoch": 0.05025612918186226, "grad_norm": 0.9623571038246155, "learning_rate": 8.368465430016864e-06, "loss": 0.3719, "step": 1192 }, { "epoch": 0.05029829036406181, "grad_norm": 1.206365942955017, "learning_rate": 8.37549184935357e-06, "loss": 0.7382, "step": 1193 }, { "epoch": 0.05034045154626136, "grad_norm": 1.0646415948867798, "learning_rate": 8.382518268690276e-06, "loss": 0.1784, "step": 1194 }, { "epoch": 0.0503826127284609, "grad_norm": 1.4773790836334229, "learning_rate": 8.389544688026982e-06, "loss": 0.2506, "step": 1195 }, { "epoch": 0.050424773910660456, "grad_norm": 2.0948750972747803, "learning_rate": 8.396571107363687e-06, "loss": 1.1941, "step": 1196 }, { "epoch": 0.05046693509286, "grad_norm": 1.1195640563964844, "learning_rate": 8.403597526700393e-06, "loss": 0.3038, "step": 1197 }, { "epoch": 0.050509096275059555, "grad_norm": 1.3096790313720703, "learning_rate": 8.410623946037099e-06, "loss": 0.2428, "step": 1198 }, { "epoch": 0.0505512574572591, "grad_norm": 1.6866222620010376, "learning_rate": 8.417650365373805e-06, "loss": 0.5552, "step": 1199 }, { "epoch": 0.050593418639458654, "grad_norm": 1.8585898876190186, "learning_rate": 8.424676784710511e-06, "loss": 0.2407, "step": 1200 }, { "epoch": 0.0506355798216582, "grad_norm": 0.8452637791633606, "learning_rate": 8.431703204047219e-06, "loss": 0.2654, "step": 1201 }, { "epoch": 0.050677741003857746, "grad_norm": 1.44948148727417, "learning_rate": 8.438729623383925e-06, "loss": 0.2558, "step": 1202 }, { "epoch": 0.0507199021860573, "grad_norm": 3.0557608604431152, "learning_rate": 8.44575604272063e-06, "loss": 0.6011, "step": 1203 }, { "epoch": 0.050762063368256845, "grad_norm": 1.611918568611145, "learning_rate": 8.452782462057336e-06, "loss": 0.5587, "step": 1204 }, { "epoch": 0.0508042245504564, "grad_norm": 0.9033512473106384, "learning_rate": 8.459808881394042e-06, "loss": 0.2113, "step": 1205 }, { "epoch": 0.050846385732655944, "grad_norm": 1.3137130737304688, "learning_rate": 8.466835300730748e-06, "loss": 0.3355, "step": 1206 }, { "epoch": 0.05088854691485549, "grad_norm": 0.9260578751564026, "learning_rate": 8.473861720067454e-06, "loss": 0.3044, "step": 1207 }, { "epoch": 0.05093070809705504, "grad_norm": 2.7034707069396973, "learning_rate": 8.48088813940416e-06, "loss": 1.5234, "step": 1208 }, { "epoch": 0.05097286927925459, "grad_norm": 1.318774700164795, "learning_rate": 8.487914558740866e-06, "loss": 0.564, "step": 1209 }, { "epoch": 0.05101503046145414, "grad_norm": 2.911646604537964, "learning_rate": 8.494940978077572e-06, "loss": 0.9091, "step": 1210 }, { "epoch": 0.05105719164365369, "grad_norm": 2.320826768875122, "learning_rate": 8.501967397414278e-06, "loss": 0.4202, "step": 1211 }, { "epoch": 0.051099352825853234, "grad_norm": 1.2716996669769287, "learning_rate": 8.508993816750984e-06, "loss": 0.3322, "step": 1212 }, { "epoch": 0.05114151400805279, "grad_norm": 1.842079997062683, "learning_rate": 8.51602023608769e-06, "loss": 0.2974, "step": 1213 }, { "epoch": 0.05118367519025233, "grad_norm": 1.3111495971679688, "learning_rate": 8.523046655424396e-06, "loss": 0.637, "step": 1214 }, { "epoch": 0.051225836372451886, "grad_norm": 2.3382818698883057, "learning_rate": 8.530073074761103e-06, "loss": 1.8752, "step": 1215 }, { "epoch": 0.05126799755465143, "grad_norm": 1.928278923034668, "learning_rate": 8.537099494097809e-06, "loss": 0.4151, "step": 1216 }, { "epoch": 0.051310158736850985, "grad_norm": 1.3854990005493164, "learning_rate": 8.544125913434515e-06, "loss": 0.2126, "step": 1217 }, { "epoch": 0.05135231991905053, "grad_norm": 1.8257838487625122, "learning_rate": 8.55115233277122e-06, "loss": 0.2777, "step": 1218 }, { "epoch": 0.05139448110125008, "grad_norm": 1.365844964981079, "learning_rate": 8.558178752107925e-06, "loss": 0.7257, "step": 1219 }, { "epoch": 0.05143664228344963, "grad_norm": 1.0709753036499023, "learning_rate": 8.565205171444631e-06, "loss": 0.2932, "step": 1220 }, { "epoch": 0.051478803465649176, "grad_norm": 1.280192494392395, "learning_rate": 8.572231590781337e-06, "loss": 0.5849, "step": 1221 }, { "epoch": 0.05152096464784873, "grad_norm": 1.2889926433563232, "learning_rate": 8.579258010118043e-06, "loss": 0.3067, "step": 1222 }, { "epoch": 0.051563125830048274, "grad_norm": 1.1904828548431396, "learning_rate": 8.586284429454751e-06, "loss": 0.2859, "step": 1223 }, { "epoch": 0.05160528701224782, "grad_norm": 1.3540124893188477, "learning_rate": 8.593310848791457e-06, "loss": 0.3364, "step": 1224 }, { "epoch": 0.05164744819444737, "grad_norm": 0.9349501729011536, "learning_rate": 8.600337268128163e-06, "loss": 0.3854, "step": 1225 }, { "epoch": 0.05168960937664692, "grad_norm": 1.5837514400482178, "learning_rate": 8.607363687464868e-06, "loss": 0.3445, "step": 1226 }, { "epoch": 0.05173177055884647, "grad_norm": 1.1360251903533936, "learning_rate": 8.614390106801574e-06, "loss": 0.6524, "step": 1227 }, { "epoch": 0.05177393174104602, "grad_norm": 1.4043136835098267, "learning_rate": 8.62141652613828e-06, "loss": 0.7942, "step": 1228 }, { "epoch": 0.051816092923245564, "grad_norm": 1.2737352848052979, "learning_rate": 8.628442945474986e-06, "loss": 0.29, "step": 1229 }, { "epoch": 0.05185825410544512, "grad_norm": 1.2851632833480835, "learning_rate": 8.635469364811692e-06, "loss": 0.3116, "step": 1230 }, { "epoch": 0.05190041528764466, "grad_norm": 1.5908334255218506, "learning_rate": 8.642495784148398e-06, "loss": 0.3571, "step": 1231 }, { "epoch": 0.051942576469844216, "grad_norm": 1.829674243927002, "learning_rate": 8.649522203485104e-06, "loss": 0.769, "step": 1232 }, { "epoch": 0.05198473765204376, "grad_norm": 1.3920788764953613, "learning_rate": 8.65654862282181e-06, "loss": 0.3305, "step": 1233 }, { "epoch": 0.052026898834243315, "grad_norm": 1.611855387687683, "learning_rate": 8.663575042158517e-06, "loss": 0.226, "step": 1234 }, { "epoch": 0.05206906001644286, "grad_norm": 1.5700756311416626, "learning_rate": 8.670601461495223e-06, "loss": 0.8243, "step": 1235 }, { "epoch": 0.05211122119864241, "grad_norm": 0.9904049634933472, "learning_rate": 8.677627880831929e-06, "loss": 0.1802, "step": 1236 }, { "epoch": 0.05215338238084196, "grad_norm": 1.0270366668701172, "learning_rate": 8.684654300168635e-06, "loss": 0.1663, "step": 1237 }, { "epoch": 0.052195543563041506, "grad_norm": 1.8603193759918213, "learning_rate": 8.691680719505341e-06, "loss": 0.5342, "step": 1238 }, { "epoch": 0.05223770474524106, "grad_norm": 1.3358033895492554, "learning_rate": 8.698707138842047e-06, "loss": 0.2319, "step": 1239 }, { "epoch": 0.052279865927440605, "grad_norm": 0.9203743934631348, "learning_rate": 8.705733558178753e-06, "loss": 0.1847, "step": 1240 }, { "epoch": 0.05232202710964015, "grad_norm": 1.5330768823623657, "learning_rate": 8.71275997751546e-06, "loss": 1.2127, "step": 1241 }, { "epoch": 0.052364188291839704, "grad_norm": 1.287917137145996, "learning_rate": 8.719786396852164e-06, "loss": 1.8481, "step": 1242 }, { "epoch": 0.05240634947403925, "grad_norm": 3.0774929523468018, "learning_rate": 8.72681281618887e-06, "loss": 0.6429, "step": 1243 }, { "epoch": 0.0524485106562388, "grad_norm": 1.8059226274490356, "learning_rate": 8.733839235525576e-06, "loss": 0.5034, "step": 1244 }, { "epoch": 0.05249067183843835, "grad_norm": 1.6489454507827759, "learning_rate": 8.740865654862282e-06, "loss": 0.4244, "step": 1245 }, { "epoch": 0.0525328330206379, "grad_norm": 1.5253233909606934, "learning_rate": 8.74789207419899e-06, "loss": 0.276, "step": 1246 }, { "epoch": 0.05257499420283745, "grad_norm": 1.3276424407958984, "learning_rate": 8.754918493535696e-06, "loss": 0.2683, "step": 1247 }, { "epoch": 0.052617155385036994, "grad_norm": 1.3836637735366821, "learning_rate": 8.761944912872402e-06, "loss": 0.2854, "step": 1248 }, { "epoch": 0.05265931656723655, "grad_norm": 2.4531328678131104, "learning_rate": 8.768971332209106e-06, "loss": 0.2748, "step": 1249 }, { "epoch": 0.05270147774943609, "grad_norm": 1.565891981124878, "learning_rate": 8.775997751545812e-06, "loss": 0.5192, "step": 1250 }, { "epoch": 0.052743638931635646, "grad_norm": 1.6064774990081787, "learning_rate": 8.783024170882518e-06, "loss": 0.3828, "step": 1251 }, { "epoch": 0.05278580011383519, "grad_norm": 1.1992932558059692, "learning_rate": 8.790050590219224e-06, "loss": 0.5332, "step": 1252 }, { "epoch": 0.05282796129603474, "grad_norm": 1.3768059015274048, "learning_rate": 8.79707700955593e-06, "loss": 0.3506, "step": 1253 }, { "epoch": 0.05287012247823429, "grad_norm": 1.893083930015564, "learning_rate": 8.804103428892637e-06, "loss": 1.2731, "step": 1254 }, { "epoch": 0.05291228366043384, "grad_norm": 1.285947322845459, "learning_rate": 8.811129848229343e-06, "loss": 0.6544, "step": 1255 }, { "epoch": 0.05295444484263339, "grad_norm": 1.8031820058822632, "learning_rate": 8.818156267566049e-06, "loss": 0.5, "step": 1256 }, { "epoch": 0.052996606024832936, "grad_norm": 2.96545672416687, "learning_rate": 8.825182686902755e-06, "loss": 0.4943, "step": 1257 }, { "epoch": 0.05303876720703248, "grad_norm": 1.786203145980835, "learning_rate": 8.832209106239461e-06, "loss": 0.2103, "step": 1258 }, { "epoch": 0.053080928389232035, "grad_norm": 1.9152168035507202, "learning_rate": 8.839235525576167e-06, "loss": 0.3058, "step": 1259 }, { "epoch": 0.05312308957143158, "grad_norm": 1.322634220123291, "learning_rate": 8.846261944912873e-06, "loss": 0.6931, "step": 1260 }, { "epoch": 0.053165250753631134, "grad_norm": 1.2104766368865967, "learning_rate": 8.85328836424958e-06, "loss": 0.2536, "step": 1261 }, { "epoch": 0.05320741193583068, "grad_norm": 1.2877295017242432, "learning_rate": 8.860314783586285e-06, "loss": 0.25, "step": 1262 }, { "epoch": 0.05324957311803023, "grad_norm": 2.621222734451294, "learning_rate": 8.867341202922991e-06, "loss": 0.6739, "step": 1263 }, { "epoch": 0.05329173430022978, "grad_norm": 1.193718433380127, "learning_rate": 8.874367622259696e-06, "loss": 0.4209, "step": 1264 }, { "epoch": 0.053333895482429325, "grad_norm": 1.4825630187988281, "learning_rate": 8.881394041596402e-06, "loss": 0.4617, "step": 1265 }, { "epoch": 0.05337605666462888, "grad_norm": 0.7940189838409424, "learning_rate": 8.888420460933108e-06, "loss": 0.1595, "step": 1266 }, { "epoch": 0.053418217846828424, "grad_norm": 1.376886248588562, "learning_rate": 8.895446880269814e-06, "loss": 0.7387, "step": 1267 }, { "epoch": 0.05346037902902798, "grad_norm": 1.9813263416290283, "learning_rate": 8.902473299606522e-06, "loss": 1.3514, "step": 1268 }, { "epoch": 0.05350254021122752, "grad_norm": 1.5005571842193604, "learning_rate": 8.909499718943228e-06, "loss": 0.236, "step": 1269 }, { "epoch": 0.05354470139342707, "grad_norm": 1.146235704421997, "learning_rate": 8.916526138279934e-06, "loss": 0.303, "step": 1270 }, { "epoch": 0.05358686257562662, "grad_norm": 1.554982304573059, "learning_rate": 8.923552557616638e-06, "loss": 0.3564, "step": 1271 }, { "epoch": 0.05362902375782617, "grad_norm": 1.4240577220916748, "learning_rate": 8.930578976953345e-06, "loss": 0.326, "step": 1272 }, { "epoch": 0.05367118494002572, "grad_norm": 1.308176040649414, "learning_rate": 8.93760539629005e-06, "loss": 0.4225, "step": 1273 }, { "epoch": 0.05371334612222527, "grad_norm": 0.9478961825370789, "learning_rate": 8.944631815626757e-06, "loss": 0.6097, "step": 1274 }, { "epoch": 0.05375550730442481, "grad_norm": 1.0787806510925293, "learning_rate": 8.951658234963463e-06, "loss": 0.1852, "step": 1275 }, { "epoch": 0.053797668486624366, "grad_norm": 1.186705231666565, "learning_rate": 8.958684654300169e-06, "loss": 0.5379, "step": 1276 }, { "epoch": 0.05383982966882391, "grad_norm": 1.1022108793258667, "learning_rate": 8.965711073636875e-06, "loss": 0.688, "step": 1277 }, { "epoch": 0.053881990851023465, "grad_norm": 1.5096981525421143, "learning_rate": 8.972737492973581e-06, "loss": 1.325, "step": 1278 }, { "epoch": 0.05392415203322301, "grad_norm": 1.0293947458267212, "learning_rate": 8.979763912310287e-06, "loss": 0.2004, "step": 1279 }, { "epoch": 0.053966313215422564, "grad_norm": 2.25895357131958, "learning_rate": 8.986790331646993e-06, "loss": 1.4963, "step": 1280 }, { "epoch": 0.05400847439762211, "grad_norm": 1.3131515979766846, "learning_rate": 8.9938167509837e-06, "loss": 0.7349, "step": 1281 }, { "epoch": 0.054050635579821656, "grad_norm": 1.318566918373108, "learning_rate": 9.000843170320405e-06, "loss": 0.9422, "step": 1282 }, { "epoch": 0.05409279676202121, "grad_norm": 1.9134503602981567, "learning_rate": 9.007869589657112e-06, "loss": 0.3196, "step": 1283 }, { "epoch": 0.054134957944220755, "grad_norm": 2.1740708351135254, "learning_rate": 9.014896008993818e-06, "loss": 0.3766, "step": 1284 }, { "epoch": 0.05417711912642031, "grad_norm": 1.2782965898513794, "learning_rate": 9.021922428330524e-06, "loss": 0.3987, "step": 1285 }, { "epoch": 0.054219280308619854, "grad_norm": 1.6459155082702637, "learning_rate": 9.02894884766723e-06, "loss": 0.3164, "step": 1286 }, { "epoch": 0.0542614414908194, "grad_norm": 1.333305835723877, "learning_rate": 9.035975267003934e-06, "loss": 0.2372, "step": 1287 }, { "epoch": 0.05430360267301895, "grad_norm": 1.073891282081604, "learning_rate": 9.04300168634064e-06, "loss": 0.7998, "step": 1288 }, { "epoch": 0.0543457638552185, "grad_norm": 1.485111951828003, "learning_rate": 9.050028105677346e-06, "loss": 1.2915, "step": 1289 }, { "epoch": 0.05438792503741805, "grad_norm": 1.0499136447906494, "learning_rate": 9.057054525014052e-06, "loss": 0.1414, "step": 1290 }, { "epoch": 0.0544300862196176, "grad_norm": 1.1510971784591675, "learning_rate": 9.06408094435076e-06, "loss": 0.2299, "step": 1291 }, { "epoch": 0.05447224740181715, "grad_norm": 1.073493242263794, "learning_rate": 9.071107363687466e-06, "loss": 0.5359, "step": 1292 }, { "epoch": 0.054514408584016696, "grad_norm": 1.169134259223938, "learning_rate": 9.078133783024172e-06, "loss": 0.3264, "step": 1293 }, { "epoch": 0.05455656976621624, "grad_norm": 3.3078911304473877, "learning_rate": 9.085160202360877e-06, "loss": 1.211, "step": 1294 }, { "epoch": 0.054598730948415795, "grad_norm": 1.7747799158096313, "learning_rate": 9.092186621697583e-06, "loss": 0.4795, "step": 1295 }, { "epoch": 0.05464089213061534, "grad_norm": 2.194331169128418, "learning_rate": 9.099213041034289e-06, "loss": 1.1625, "step": 1296 }, { "epoch": 0.054683053312814894, "grad_norm": 1.1791743040084839, "learning_rate": 9.106239460370995e-06, "loss": 0.6663, "step": 1297 }, { "epoch": 0.05472521449501444, "grad_norm": 1.2490041255950928, "learning_rate": 9.113265879707701e-06, "loss": 0.1878, "step": 1298 }, { "epoch": 0.054767375677213986, "grad_norm": 1.617965817451477, "learning_rate": 9.120292299044407e-06, "loss": 0.4491, "step": 1299 }, { "epoch": 0.05480953685941354, "grad_norm": 1.5959750413894653, "learning_rate": 9.127318718381113e-06, "loss": 0.2959, "step": 1300 }, { "epoch": 0.054851698041613085, "grad_norm": 2.311511754989624, "learning_rate": 9.13434513771782e-06, "loss": 0.482, "step": 1301 }, { "epoch": 0.05489385922381264, "grad_norm": 1.9167619943618774, "learning_rate": 9.141371557054526e-06, "loss": 1.2508, "step": 1302 }, { "epoch": 0.054936020406012184, "grad_norm": 1.9263558387756348, "learning_rate": 9.148397976391232e-06, "loss": 1.0374, "step": 1303 }, { "epoch": 0.05497818158821173, "grad_norm": 1.355507731437683, "learning_rate": 9.155424395727938e-06, "loss": 0.2118, "step": 1304 }, { "epoch": 0.05502034277041128, "grad_norm": 1.7365087270736694, "learning_rate": 9.162450815064644e-06, "loss": 1.1101, "step": 1305 }, { "epoch": 0.05506250395261083, "grad_norm": 0.8089714646339417, "learning_rate": 9.16947723440135e-06, "loss": 0.2301, "step": 1306 }, { "epoch": 0.05510466513481038, "grad_norm": 1.0284323692321777, "learning_rate": 9.176503653738056e-06, "loss": 0.3703, "step": 1307 }, { "epoch": 0.05514682631700993, "grad_norm": 0.9314206838607788, "learning_rate": 9.183530073074762e-06, "loss": 0.2584, "step": 1308 }, { "epoch": 0.05518898749920948, "grad_norm": 1.7700293064117432, "learning_rate": 9.190556492411468e-06, "loss": 0.2926, "step": 1309 }, { "epoch": 0.05523114868140903, "grad_norm": 1.2107380628585815, "learning_rate": 9.197582911748173e-06, "loss": 0.4291, "step": 1310 }, { "epoch": 0.05527330986360857, "grad_norm": 1.2112022638320923, "learning_rate": 9.204609331084879e-06, "loss": 0.1988, "step": 1311 }, { "epoch": 0.055315471045808126, "grad_norm": 1.172565221786499, "learning_rate": 9.211635750421585e-06, "loss": 0.345, "step": 1312 }, { "epoch": 0.05535763222800767, "grad_norm": 1.9092353582382202, "learning_rate": 9.218662169758293e-06, "loss": 0.5156, "step": 1313 }, { "epoch": 0.055399793410207225, "grad_norm": 3.814584493637085, "learning_rate": 9.225688589094999e-06, "loss": 0.3711, "step": 1314 }, { "epoch": 0.05544195459240677, "grad_norm": 1.253511667251587, "learning_rate": 9.232715008431705e-06, "loss": 0.5092, "step": 1315 }, { "epoch": 0.05548411577460632, "grad_norm": 1.4369354248046875, "learning_rate": 9.23974142776841e-06, "loss": 0.4014, "step": 1316 }, { "epoch": 0.05552627695680587, "grad_norm": 4.226372241973877, "learning_rate": 9.246767847105115e-06, "loss": 1.3461, "step": 1317 }, { "epoch": 0.055568438139005416, "grad_norm": 1.7871670722961426, "learning_rate": 9.253794266441821e-06, "loss": 0.4873, "step": 1318 }, { "epoch": 0.05561059932120497, "grad_norm": 4.344130516052246, "learning_rate": 9.260820685778527e-06, "loss": 0.7244, "step": 1319 }, { "epoch": 0.055652760503404515, "grad_norm": 1.5691595077514648, "learning_rate": 9.267847105115233e-06, "loss": 1.2447, "step": 1320 }, { "epoch": 0.05569492168560406, "grad_norm": 1.5300785303115845, "learning_rate": 9.27487352445194e-06, "loss": 0.3581, "step": 1321 }, { "epoch": 0.055737082867803614, "grad_norm": 1.1548781394958496, "learning_rate": 9.281899943788646e-06, "loss": 0.1925, "step": 1322 }, { "epoch": 0.05577924405000316, "grad_norm": 1.587262749671936, "learning_rate": 9.288926363125352e-06, "loss": 0.6461, "step": 1323 }, { "epoch": 0.05582140523220271, "grad_norm": 1.3476506471633911, "learning_rate": 9.295952782462058e-06, "loss": 0.5209, "step": 1324 }, { "epoch": 0.05586356641440226, "grad_norm": 1.2796893119812012, "learning_rate": 9.302979201798764e-06, "loss": 0.2674, "step": 1325 }, { "epoch": 0.05590572759660181, "grad_norm": 1.2291227579116821, "learning_rate": 9.31000562113547e-06, "loss": 0.3049, "step": 1326 }, { "epoch": 0.05594788877880136, "grad_norm": 1.222464919090271, "learning_rate": 9.317032040472176e-06, "loss": 0.7518, "step": 1327 }, { "epoch": 0.055990049961000904, "grad_norm": 1.2564846277236938, "learning_rate": 9.324058459808882e-06, "loss": 0.2022, "step": 1328 }, { "epoch": 0.05603221114320046, "grad_norm": 1.1090724468231201, "learning_rate": 9.331084879145588e-06, "loss": 0.4349, "step": 1329 }, { "epoch": 0.0560743723254, "grad_norm": 1.974184274673462, "learning_rate": 9.338111298482294e-06, "loss": 0.263, "step": 1330 }, { "epoch": 0.056116533507599556, "grad_norm": 1.270318627357483, "learning_rate": 9.345137717819e-06, "loss": 0.1689, "step": 1331 }, { "epoch": 0.0561586946897991, "grad_norm": 2.916696548461914, "learning_rate": 9.352164137155705e-06, "loss": 1.2935, "step": 1332 }, { "epoch": 0.05620085587199865, "grad_norm": 1.027160406112671, "learning_rate": 9.359190556492411e-06, "loss": 0.6687, "step": 1333 }, { "epoch": 0.0562430170541982, "grad_norm": 1.1377447843551636, "learning_rate": 9.366216975829117e-06, "loss": 0.5859, "step": 1334 }, { "epoch": 0.05628517823639775, "grad_norm": 0.9758487343788147, "learning_rate": 9.373243395165823e-06, "loss": 0.2002, "step": 1335 }, { "epoch": 0.0563273394185973, "grad_norm": 1.1942638158798218, "learning_rate": 9.380269814502531e-06, "loss": 0.3531, "step": 1336 }, { "epoch": 0.056369500600796846, "grad_norm": 1.3675893545150757, "learning_rate": 9.387296233839237e-06, "loss": 0.5208, "step": 1337 }, { "epoch": 0.05641166178299639, "grad_norm": 0.871523380279541, "learning_rate": 9.394322653175943e-06, "loss": 0.207, "step": 1338 }, { "epoch": 0.056453822965195945, "grad_norm": 1.1577624082565308, "learning_rate": 9.401349072512647e-06, "loss": 0.2554, "step": 1339 }, { "epoch": 0.05649598414739549, "grad_norm": 1.8336256742477417, "learning_rate": 9.408375491849354e-06, "loss": 0.5646, "step": 1340 }, { "epoch": 0.056538145329595044, "grad_norm": 1.633349061012268, "learning_rate": 9.41540191118606e-06, "loss": 0.2712, "step": 1341 }, { "epoch": 0.05658030651179459, "grad_norm": 2.645556688308716, "learning_rate": 9.422428330522766e-06, "loss": 1.0804, "step": 1342 }, { "epoch": 0.05662246769399414, "grad_norm": 1.9881744384765625, "learning_rate": 9.429454749859472e-06, "loss": 0.5019, "step": 1343 }, { "epoch": 0.05666462887619369, "grad_norm": 4.7081170082092285, "learning_rate": 9.436481169196178e-06, "loss": 0.9075, "step": 1344 }, { "epoch": 0.056706790058393235, "grad_norm": 1.4289876222610474, "learning_rate": 9.443507588532884e-06, "loss": 0.3566, "step": 1345 }, { "epoch": 0.05674895124059279, "grad_norm": 1.7952483892440796, "learning_rate": 9.45053400786959e-06, "loss": 0.2059, "step": 1346 }, { "epoch": 0.056791112422792334, "grad_norm": 1.443226933479309, "learning_rate": 9.457560427206296e-06, "loss": 0.5789, "step": 1347 }, { "epoch": 0.05683327360499189, "grad_norm": 2.7480032444000244, "learning_rate": 9.464586846543002e-06, "loss": 1.2388, "step": 1348 }, { "epoch": 0.05687543478719143, "grad_norm": 2.1489064693450928, "learning_rate": 9.471613265879708e-06, "loss": 0.5108, "step": 1349 }, { "epoch": 0.05691759596939098, "grad_norm": 1.4481550455093384, "learning_rate": 9.478639685216414e-06, "loss": 0.436, "step": 1350 }, { "epoch": 0.05695975715159053, "grad_norm": 1.8367429971694946, "learning_rate": 9.48566610455312e-06, "loss": 0.4603, "step": 1351 }, { "epoch": 0.05700191833379008, "grad_norm": 1.1813764572143555, "learning_rate": 9.492692523889827e-06, "loss": 0.6863, "step": 1352 }, { "epoch": 0.05704407951598963, "grad_norm": 1.6922006607055664, "learning_rate": 9.499718943226533e-06, "loss": 0.3901, "step": 1353 }, { "epoch": 0.057086240698189177, "grad_norm": 1.2398368120193481, "learning_rate": 9.506745362563239e-06, "loss": 0.2648, "step": 1354 }, { "epoch": 0.05712840188038873, "grad_norm": 1.2059574127197266, "learning_rate": 9.513771781899943e-06, "loss": 0.2359, "step": 1355 }, { "epoch": 0.057170563062588275, "grad_norm": 0.9033559560775757, "learning_rate": 9.52079820123665e-06, "loss": 0.1735, "step": 1356 }, { "epoch": 0.05721272424478782, "grad_norm": 1.6294171810150146, "learning_rate": 9.527824620573355e-06, "loss": 0.4017, "step": 1357 }, { "epoch": 0.057254885426987374, "grad_norm": 1.1049506664276123, "learning_rate": 9.534851039910063e-06, "loss": 0.1972, "step": 1358 }, { "epoch": 0.05729704660918692, "grad_norm": 1.0948749780654907, "learning_rate": 9.54187745924677e-06, "loss": 0.2725, "step": 1359 }, { "epoch": 0.05733920779138647, "grad_norm": 1.209409475326538, "learning_rate": 9.548903878583475e-06, "loss": 0.2053, "step": 1360 }, { "epoch": 0.05738136897358602, "grad_norm": 1.7226285934448242, "learning_rate": 9.555930297920181e-06, "loss": 1.3625, "step": 1361 }, { "epoch": 0.057423530155785565, "grad_norm": 0.8964847326278687, "learning_rate": 9.562956717256886e-06, "loss": 0.1883, "step": 1362 }, { "epoch": 0.05746569133798512, "grad_norm": 1.4092998504638672, "learning_rate": 9.569983136593592e-06, "loss": 0.3584, "step": 1363 }, { "epoch": 0.057507852520184664, "grad_norm": 1.0498043298721313, "learning_rate": 9.577009555930298e-06, "loss": 0.2757, "step": 1364 }, { "epoch": 0.05755001370238422, "grad_norm": 1.1560651063919067, "learning_rate": 9.584035975267004e-06, "loss": 0.2982, "step": 1365 }, { "epoch": 0.05759217488458376, "grad_norm": 1.172688603401184, "learning_rate": 9.59106239460371e-06, "loss": 0.2433, "step": 1366 }, { "epoch": 0.05763433606678331, "grad_norm": 1.7810211181640625, "learning_rate": 9.598088813940416e-06, "loss": 0.53, "step": 1367 }, { "epoch": 0.05767649724898286, "grad_norm": 1.2978967428207397, "learning_rate": 9.605115233277122e-06, "loss": 0.7846, "step": 1368 }, { "epoch": 0.05771865843118241, "grad_norm": 1.022594928741455, "learning_rate": 9.612141652613828e-06, "loss": 0.2807, "step": 1369 }, { "epoch": 0.05776081961338196, "grad_norm": 1.5567643642425537, "learning_rate": 9.619168071950535e-06, "loss": 0.7707, "step": 1370 }, { "epoch": 0.05780298079558151, "grad_norm": 1.4474247694015503, "learning_rate": 9.62619449128724e-06, "loss": 0.359, "step": 1371 }, { "epoch": 0.05784514197778106, "grad_norm": 1.1458693742752075, "learning_rate": 9.633220910623947e-06, "loss": 0.2113, "step": 1372 }, { "epoch": 0.057887303159980606, "grad_norm": 1.5147417783737183, "learning_rate": 9.640247329960653e-06, "loss": 0.6451, "step": 1373 }, { "epoch": 0.05792946434218015, "grad_norm": 1.2210522890090942, "learning_rate": 9.647273749297359e-06, "loss": 0.6122, "step": 1374 }, { "epoch": 0.057971625524379705, "grad_norm": 1.2474802732467651, "learning_rate": 9.654300168634065e-06, "loss": 0.6994, "step": 1375 }, { "epoch": 0.05801378670657925, "grad_norm": 1.2499315738677979, "learning_rate": 9.661326587970771e-06, "loss": 0.931, "step": 1376 }, { "epoch": 0.058055947888778804, "grad_norm": 2.3771121501922607, "learning_rate": 9.668353007307475e-06, "loss": 0.2374, "step": 1377 }, { "epoch": 0.05809810907097835, "grad_norm": 1.5726052522659302, "learning_rate": 9.675379426644182e-06, "loss": 0.1786, "step": 1378 }, { "epoch": 0.058140270253177896, "grad_norm": 2.371479034423828, "learning_rate": 9.682405845980888e-06, "loss": 0.4683, "step": 1379 }, { "epoch": 0.05818243143537745, "grad_norm": 1.1438789367675781, "learning_rate": 9.689432265317594e-06, "loss": 0.5505, "step": 1380 }, { "epoch": 0.058224592617576995, "grad_norm": 1.5275663137435913, "learning_rate": 9.696458684654301e-06, "loss": 0.4179, "step": 1381 }, { "epoch": 0.05826675379977655, "grad_norm": 1.0355230569839478, "learning_rate": 9.703485103991008e-06, "loss": 0.5577, "step": 1382 }, { "epoch": 0.058308914981976094, "grad_norm": 1.3457167148590088, "learning_rate": 9.710511523327714e-06, "loss": 0.9829, "step": 1383 }, { "epoch": 0.05835107616417564, "grad_norm": 1.7313858270645142, "learning_rate": 9.71753794266442e-06, "loss": 0.2839, "step": 1384 }, { "epoch": 0.05839323734637519, "grad_norm": 1.3631694316864014, "learning_rate": 9.724564362001124e-06, "loss": 0.7901, "step": 1385 }, { "epoch": 0.05843539852857474, "grad_norm": 1.175742268562317, "learning_rate": 9.73159078133783e-06, "loss": 0.1896, "step": 1386 }, { "epoch": 0.05847755971077429, "grad_norm": 1.0787748098373413, "learning_rate": 9.738617200674536e-06, "loss": 0.4276, "step": 1387 }, { "epoch": 0.05851972089297384, "grad_norm": 1.073104739189148, "learning_rate": 9.745643620011242e-06, "loss": 0.7985, "step": 1388 }, { "epoch": 0.05856188207517339, "grad_norm": 1.8927150964736938, "learning_rate": 9.752670039347949e-06, "loss": 0.2478, "step": 1389 }, { "epoch": 0.05860404325737294, "grad_norm": 1.2396366596221924, "learning_rate": 9.759696458684655e-06, "loss": 0.1885, "step": 1390 }, { "epoch": 0.05864620443957248, "grad_norm": 1.248690128326416, "learning_rate": 9.76672287802136e-06, "loss": 0.7602, "step": 1391 }, { "epoch": 0.058688365621772036, "grad_norm": 1.9421744346618652, "learning_rate": 9.773749297358067e-06, "loss": 0.7583, "step": 1392 }, { "epoch": 0.05873052680397158, "grad_norm": 1.4533225297927856, "learning_rate": 9.780775716694773e-06, "loss": 0.3109, "step": 1393 }, { "epoch": 0.058772687986171135, "grad_norm": 1.4615795612335205, "learning_rate": 9.787802136031479e-06, "loss": 0.826, "step": 1394 }, { "epoch": 0.05881484916837068, "grad_norm": 1.4008944034576416, "learning_rate": 9.794828555368185e-06, "loss": 0.3564, "step": 1395 }, { "epoch": 0.05885701035057023, "grad_norm": 1.087762713432312, "learning_rate": 9.801854974704891e-06, "loss": 0.153, "step": 1396 }, { "epoch": 0.05889917153276978, "grad_norm": 2.382561206817627, "learning_rate": 9.808881394041597e-06, "loss": 0.7083, "step": 1397 }, { "epoch": 0.058941332714969326, "grad_norm": 1.3437849283218384, "learning_rate": 9.815907813378303e-06, "loss": 0.9293, "step": 1398 }, { "epoch": 0.05898349389716888, "grad_norm": 1.0945082902908325, "learning_rate": 9.82293423271501e-06, "loss": 0.3001, "step": 1399 }, { "epoch": 0.059025655079368425, "grad_norm": 1.0034037828445435, "learning_rate": 9.829960652051714e-06, "loss": 0.3183, "step": 1400 }, { "epoch": 0.05906781626156798, "grad_norm": 1.0575206279754639, "learning_rate": 9.83698707138842e-06, "loss": 0.165, "step": 1401 }, { "epoch": 0.059109977443767524, "grad_norm": 1.0866156816482544, "learning_rate": 9.844013490725126e-06, "loss": 0.1848, "step": 1402 }, { "epoch": 0.05915213862596707, "grad_norm": 2.1066386699676514, "learning_rate": 9.851039910061834e-06, "loss": 1.3199, "step": 1403 }, { "epoch": 0.05919429980816662, "grad_norm": 1.0519124269485474, "learning_rate": 9.85806632939854e-06, "loss": 0.6556, "step": 1404 }, { "epoch": 0.05923646099036617, "grad_norm": 1.073811411857605, "learning_rate": 9.865092748735246e-06, "loss": 0.2397, "step": 1405 }, { "epoch": 0.05927862217256572, "grad_norm": 0.9705523252487183, "learning_rate": 9.872119168071952e-06, "loss": 0.1894, "step": 1406 }, { "epoch": 0.05932078335476527, "grad_norm": 1.4762380123138428, "learning_rate": 9.879145587408656e-06, "loss": 0.2506, "step": 1407 }, { "epoch": 0.059362944536964814, "grad_norm": 1.1909141540527344, "learning_rate": 9.886172006745363e-06, "loss": 0.2379, "step": 1408 }, { "epoch": 0.05940510571916437, "grad_norm": 1.5683788061141968, "learning_rate": 9.893198426082069e-06, "loss": 0.5768, "step": 1409 }, { "epoch": 0.05944726690136391, "grad_norm": 1.3705250024795532, "learning_rate": 9.900224845418775e-06, "loss": 0.2705, "step": 1410 }, { "epoch": 0.059489428083563466, "grad_norm": 1.2445309162139893, "learning_rate": 9.90725126475548e-06, "loss": 0.2081, "step": 1411 }, { "epoch": 0.05953158926576301, "grad_norm": 1.7141690254211426, "learning_rate": 9.914277684092187e-06, "loss": 0.7375, "step": 1412 }, { "epoch": 0.05957375044796256, "grad_norm": 1.7567377090454102, "learning_rate": 9.921304103428893e-06, "loss": 0.5781, "step": 1413 }, { "epoch": 0.05961591163016211, "grad_norm": 1.3351843357086182, "learning_rate": 9.928330522765599e-06, "loss": 0.8167, "step": 1414 }, { "epoch": 0.05965807281236166, "grad_norm": 1.289315104484558, "learning_rate": 9.935356942102305e-06, "loss": 0.8421, "step": 1415 }, { "epoch": 0.05970023399456121, "grad_norm": 1.3480315208435059, "learning_rate": 9.942383361439011e-06, "loss": 0.2115, "step": 1416 }, { "epoch": 0.059742395176760756, "grad_norm": 1.6186758279800415, "learning_rate": 9.949409780775717e-06, "loss": 1.3558, "step": 1417 }, { "epoch": 0.05978455635896031, "grad_norm": 1.1556967496871948, "learning_rate": 9.956436200112423e-06, "loss": 0.6238, "step": 1418 }, { "epoch": 0.059826717541159855, "grad_norm": 1.4033899307250977, "learning_rate": 9.96346261944913e-06, "loss": 0.472, "step": 1419 }, { "epoch": 0.0598688787233594, "grad_norm": 1.1183867454528809, "learning_rate": 9.970489038785836e-06, "loss": 0.6222, "step": 1420 }, { "epoch": 0.059911039905558953, "grad_norm": 0.8255959153175354, "learning_rate": 9.977515458122542e-06, "loss": 0.2415, "step": 1421 }, { "epoch": 0.0599532010877585, "grad_norm": 1.4539659023284912, "learning_rate": 9.984541877459248e-06, "loss": 0.8862, "step": 1422 }, { "epoch": 0.05999536226995805, "grad_norm": 1.4011214971542358, "learning_rate": 9.991568296795952e-06, "loss": 0.3289, "step": 1423 }, { "epoch": 0.0600375234521576, "grad_norm": 1.9306048154830933, "learning_rate": 9.998594716132658e-06, "loss": 0.3257, "step": 1424 }, { "epoch": 0.060079684634357144, "grad_norm": 1.205496907234192, "learning_rate": 1.0005621135469364e-05, "loss": 0.2568, "step": 1425 }, { "epoch": 0.0601218458165567, "grad_norm": 1.3860353231430054, "learning_rate": 1.0012647554806072e-05, "loss": 0.5465, "step": 1426 }, { "epoch": 0.06016400699875624, "grad_norm": 1.6860933303833008, "learning_rate": 1.0019673974142778e-05, "loss": 0.2743, "step": 1427 }, { "epoch": 0.060206168180955796, "grad_norm": 1.9053183794021606, "learning_rate": 1.0026700393479484e-05, "loss": 1.3556, "step": 1428 }, { "epoch": 0.06024832936315534, "grad_norm": 1.4707179069519043, "learning_rate": 1.003372681281619e-05, "loss": 0.4126, "step": 1429 }, { "epoch": 0.06029049054535489, "grad_norm": 1.5473463535308838, "learning_rate": 1.0040753232152895e-05, "loss": 0.4417, "step": 1430 }, { "epoch": 0.06033265172755444, "grad_norm": 1.1698969602584839, "learning_rate": 1.0047779651489601e-05, "loss": 0.1918, "step": 1431 }, { "epoch": 0.06037481290975399, "grad_norm": 1.227696418762207, "learning_rate": 1.0054806070826307e-05, "loss": 0.3432, "step": 1432 }, { "epoch": 0.06041697409195354, "grad_norm": 0.8750402927398682, "learning_rate": 1.0061832490163013e-05, "loss": 0.1747, "step": 1433 }, { "epoch": 0.060459135274153086, "grad_norm": 1.707050085067749, "learning_rate": 1.0068858909499719e-05, "loss": 0.3794, "step": 1434 }, { "epoch": 0.06050129645635264, "grad_norm": 1.1733381748199463, "learning_rate": 1.0075885328836425e-05, "loss": 0.5346, "step": 1435 }, { "epoch": 0.060543457638552185, "grad_norm": 1.0923545360565186, "learning_rate": 1.0082911748173131e-05, "loss": 0.2959, "step": 1436 }, { "epoch": 0.06058561882075173, "grad_norm": 2.067121982574463, "learning_rate": 1.0089938167509837e-05, "loss": 1.2913, "step": 1437 }, { "epoch": 0.060627780002951284, "grad_norm": 8.009230613708496, "learning_rate": 1.0096964586846543e-05, "loss": 1.4619, "step": 1438 }, { "epoch": 0.06066994118515083, "grad_norm": 2.221025228500366, "learning_rate": 1.010399100618325e-05, "loss": 0.4138, "step": 1439 }, { "epoch": 0.06071210236735038, "grad_norm": 1.153619647026062, "learning_rate": 1.0111017425519956e-05, "loss": 0.3125, "step": 1440 }, { "epoch": 0.06075426354954993, "grad_norm": 1.4961402416229248, "learning_rate": 1.0118043844856662e-05, "loss": 0.3723, "step": 1441 }, { "epoch": 0.060796424731749475, "grad_norm": 1.7841932773590088, "learning_rate": 1.0125070264193368e-05, "loss": 1.1344, "step": 1442 }, { "epoch": 0.06083858591394903, "grad_norm": 1.65251886844635, "learning_rate": 1.0132096683530074e-05, "loss": 0.4408, "step": 1443 }, { "epoch": 0.060880747096148574, "grad_norm": 1.1687577962875366, "learning_rate": 1.013912310286678e-05, "loss": 0.6848, "step": 1444 }, { "epoch": 0.06092290827834813, "grad_norm": 1.6161577701568604, "learning_rate": 1.0146149522203484e-05, "loss": 0.5245, "step": 1445 }, { "epoch": 0.06096506946054767, "grad_norm": 1.0752640962600708, "learning_rate": 1.015317594154019e-05, "loss": 0.3958, "step": 1446 }, { "epoch": 0.061007230642747226, "grad_norm": 1.2910407781600952, "learning_rate": 1.0160202360876897e-05, "loss": 0.5711, "step": 1447 }, { "epoch": 0.06104939182494677, "grad_norm": 1.486136794090271, "learning_rate": 1.0167228780213604e-05, "loss": 0.4655, "step": 1448 }, { "epoch": 0.06109155300714632, "grad_norm": 1.153999924659729, "learning_rate": 1.017425519955031e-05, "loss": 0.9009, "step": 1449 }, { "epoch": 0.06113371418934587, "grad_norm": 1.6990150213241577, "learning_rate": 1.0181281618887017e-05, "loss": 1.022, "step": 1450 }, { "epoch": 0.06117587537154542, "grad_norm": 1.7483028173446655, "learning_rate": 1.0188308038223723e-05, "loss": 0.3146, "step": 1451 }, { "epoch": 0.06121803655374497, "grad_norm": 1.2785097360610962, "learning_rate": 1.0195334457560427e-05, "loss": 0.2829, "step": 1452 }, { "epoch": 0.061260197735944516, "grad_norm": 2.27341890335083, "learning_rate": 1.0202360876897133e-05, "loss": 0.4821, "step": 1453 }, { "epoch": 0.06130235891814406, "grad_norm": 8.779407501220703, "learning_rate": 1.020938729623384e-05, "loss": 0.9024, "step": 1454 }, { "epoch": 0.061344520100343615, "grad_norm": 1.3682435750961304, "learning_rate": 1.0216413715570545e-05, "loss": 0.2533, "step": 1455 }, { "epoch": 0.06138668128254316, "grad_norm": 1.4017326831817627, "learning_rate": 1.0223440134907251e-05, "loss": 0.2775, "step": 1456 }, { "epoch": 0.061428842464742714, "grad_norm": 1.2764971256256104, "learning_rate": 1.0230466554243957e-05, "loss": 0.1451, "step": 1457 }, { "epoch": 0.06147100364694226, "grad_norm": 1.8614312410354614, "learning_rate": 1.0237492973580664e-05, "loss": 1.2439, "step": 1458 }, { "epoch": 0.061513164829141806, "grad_norm": 1.5381152629852295, "learning_rate": 1.024451939291737e-05, "loss": 0.3028, "step": 1459 }, { "epoch": 0.06155532601134136, "grad_norm": 1.494185447692871, "learning_rate": 1.0251545812254076e-05, "loss": 0.3651, "step": 1460 }, { "epoch": 0.061597487193540905, "grad_norm": 1.9612345695495605, "learning_rate": 1.0258572231590782e-05, "loss": 0.2501, "step": 1461 }, { "epoch": 0.06163964837574046, "grad_norm": 0.9894653558731079, "learning_rate": 1.0265598650927488e-05, "loss": 0.2088, "step": 1462 }, { "epoch": 0.061681809557940004, "grad_norm": 1.188314437866211, "learning_rate": 1.0272625070264194e-05, "loss": 0.6446, "step": 1463 }, { "epoch": 0.06172397074013956, "grad_norm": 1.7020102739334106, "learning_rate": 1.02796514896009e-05, "loss": 0.2642, "step": 1464 }, { "epoch": 0.0617661319223391, "grad_norm": 1.5551507472991943, "learning_rate": 1.0286677908937606e-05, "loss": 0.1732, "step": 1465 }, { "epoch": 0.06180829310453865, "grad_norm": 1.1805700063705444, "learning_rate": 1.0293704328274312e-05, "loss": 0.5931, "step": 1466 }, { "epoch": 0.0618504542867382, "grad_norm": 1.211138129234314, "learning_rate": 1.0300730747611018e-05, "loss": 0.7291, "step": 1467 }, { "epoch": 0.06189261546893775, "grad_norm": 1.2622820138931274, "learning_rate": 1.0307757166947723e-05, "loss": 0.2855, "step": 1468 }, { "epoch": 0.0619347766511373, "grad_norm": 1.3433748483657837, "learning_rate": 1.0314783586284429e-05, "loss": 0.7792, "step": 1469 }, { "epoch": 0.06197693783333685, "grad_norm": 1.0281000137329102, "learning_rate": 1.0321810005621135e-05, "loss": 0.2376, "step": 1470 }, { "epoch": 0.06201909901553639, "grad_norm": 1.2717220783233643, "learning_rate": 1.0328836424957843e-05, "loss": 0.7664, "step": 1471 }, { "epoch": 0.062061260197735946, "grad_norm": 4.79361629486084, "learning_rate": 1.0335862844294549e-05, "loss": 1.2705, "step": 1472 }, { "epoch": 0.06210342137993549, "grad_norm": 2.2982044219970703, "learning_rate": 1.0342889263631255e-05, "loss": 1.1753, "step": 1473 }, { "epoch": 0.062145582562135045, "grad_norm": 1.7493696212768555, "learning_rate": 1.0349915682967961e-05, "loss": 0.2077, "step": 1474 }, { "epoch": 0.06218774374433459, "grad_norm": 1.808449387550354, "learning_rate": 1.0356942102304665e-05, "loss": 0.6641, "step": 1475 }, { "epoch": 0.06222990492653414, "grad_norm": 1.8265372514724731, "learning_rate": 1.0363968521641371e-05, "loss": 1.0571, "step": 1476 }, { "epoch": 0.06227206610873369, "grad_norm": 1.3455318212509155, "learning_rate": 1.0370994940978078e-05, "loss": 0.335, "step": 1477 }, { "epoch": 0.062314227290933236, "grad_norm": 1.6104298830032349, "learning_rate": 1.0378021360314784e-05, "loss": 0.2622, "step": 1478 }, { "epoch": 0.06235638847313279, "grad_norm": 3.249091863632202, "learning_rate": 1.038504777965149e-05, "loss": 0.7464, "step": 1479 }, { "epoch": 0.062398549655332335, "grad_norm": 1.3511147499084473, "learning_rate": 1.0392074198988196e-05, "loss": 0.6878, "step": 1480 }, { "epoch": 0.06244071083753189, "grad_norm": 1.4036122560501099, "learning_rate": 1.0399100618324902e-05, "loss": 0.4473, "step": 1481 }, { "epoch": 0.062482872019731434, "grad_norm": 1.7298213243484497, "learning_rate": 1.0406127037661608e-05, "loss": 0.2276, "step": 1482 }, { "epoch": 0.06252503320193098, "grad_norm": 1.7111176252365112, "learning_rate": 1.0413153456998314e-05, "loss": 0.4533, "step": 1483 }, { "epoch": 0.06256719438413053, "grad_norm": 4.905400276184082, "learning_rate": 1.042017987633502e-05, "loss": 1.0822, "step": 1484 }, { "epoch": 0.06260935556633009, "grad_norm": 1.4035499095916748, "learning_rate": 1.0427206295671726e-05, "loss": 0.1998, "step": 1485 }, { "epoch": 0.06265151674852963, "grad_norm": 1.4817776679992676, "learning_rate": 1.0434232715008432e-05, "loss": 0.3041, "step": 1486 }, { "epoch": 0.06269367793072918, "grad_norm": 1.3986297845840454, "learning_rate": 1.0441259134345138e-05, "loss": 0.4856, "step": 1487 }, { "epoch": 0.06273583911292872, "grad_norm": 1.2490465641021729, "learning_rate": 1.0448285553681845e-05, "loss": 0.7635, "step": 1488 }, { "epoch": 0.06277800029512827, "grad_norm": 0.9116742610931396, "learning_rate": 1.045531197301855e-05, "loss": 0.2217, "step": 1489 }, { "epoch": 0.06282016147732783, "grad_norm": 0.9162681698799133, "learning_rate": 1.0462338392355257e-05, "loss": 0.1984, "step": 1490 }, { "epoch": 0.06286232265952738, "grad_norm": 1.0141868591308594, "learning_rate": 1.0469364811691961e-05, "loss": 0.5992, "step": 1491 }, { "epoch": 0.06290448384172692, "grad_norm": 1.6040453910827637, "learning_rate": 1.0476391231028667e-05, "loss": 0.4606, "step": 1492 }, { "epoch": 0.06294664502392647, "grad_norm": 1.3693901300430298, "learning_rate": 1.0483417650365375e-05, "loss": 0.7854, "step": 1493 }, { "epoch": 0.06298880620612601, "grad_norm": 1.5630213022232056, "learning_rate": 1.0490444069702081e-05, "loss": 0.3514, "step": 1494 }, { "epoch": 0.06303096738832557, "grad_norm": 1.434631586074829, "learning_rate": 1.0497470489038787e-05, "loss": 0.6446, "step": 1495 }, { "epoch": 0.06307312857052512, "grad_norm": 0.9819357395172119, "learning_rate": 1.0504496908375493e-05, "loss": 0.6148, "step": 1496 }, { "epoch": 0.06311528975272467, "grad_norm": 1.3419209718704224, "learning_rate": 1.05115233277122e-05, "loss": 0.3196, "step": 1497 }, { "epoch": 0.06315745093492421, "grad_norm": 1.1596966981887817, "learning_rate": 1.0518549747048904e-05, "loss": 0.2054, "step": 1498 }, { "epoch": 0.06319961211712376, "grad_norm": 2.246758460998535, "learning_rate": 1.052557616638561e-05, "loss": 1.2643, "step": 1499 }, { "epoch": 0.06324177329932332, "grad_norm": 1.9094634056091309, "learning_rate": 1.0532602585722316e-05, "loss": 0.6596, "step": 1500 }, { "epoch": 0.06328393448152286, "grad_norm": 1.230527400970459, "learning_rate": 1.0539629005059022e-05, "loss": 0.2048, "step": 1501 }, { "epoch": 0.06332609566372241, "grad_norm": 1.4739593267440796, "learning_rate": 1.0546655424395728e-05, "loss": 0.1841, "step": 1502 }, { "epoch": 0.06336825684592196, "grad_norm": 1.795667052268982, "learning_rate": 1.0553681843732434e-05, "loss": 0.5578, "step": 1503 }, { "epoch": 0.06341041802812152, "grad_norm": 1.8344093561172485, "learning_rate": 1.056070826306914e-05, "loss": 0.4607, "step": 1504 }, { "epoch": 0.06345257921032106, "grad_norm": 1.4047662019729614, "learning_rate": 1.0567734682405846e-05, "loss": 0.2146, "step": 1505 }, { "epoch": 0.06349474039252061, "grad_norm": 1.79444420337677, "learning_rate": 1.0574761101742552e-05, "loss": 0.3607, "step": 1506 }, { "epoch": 0.06353690157472015, "grad_norm": 1.4809244871139526, "learning_rate": 1.0581787521079259e-05, "loss": 0.4063, "step": 1507 }, { "epoch": 0.0635790627569197, "grad_norm": 1.9847618341445923, "learning_rate": 1.0588813940415965e-05, "loss": 0.2491, "step": 1508 }, { "epoch": 0.06362122393911926, "grad_norm": 1.680014729499817, "learning_rate": 1.059584035975267e-05, "loss": 0.6753, "step": 1509 }, { "epoch": 0.0636633851213188, "grad_norm": 2.286989212036133, "learning_rate": 1.0602866779089377e-05, "loss": 0.5874, "step": 1510 }, { "epoch": 0.06370554630351835, "grad_norm": 5.180171966552734, "learning_rate": 1.0609893198426083e-05, "loss": 0.8519, "step": 1511 }, { "epoch": 0.0637477074857179, "grad_norm": 1.2233259677886963, "learning_rate": 1.0616919617762789e-05, "loss": 0.5375, "step": 1512 }, { "epoch": 0.06378986866791744, "grad_norm": 1.0547716617584229, "learning_rate": 1.0623946037099493e-05, "loss": 0.4917, "step": 1513 }, { "epoch": 0.063832029850117, "grad_norm": 1.0767533779144287, "learning_rate": 1.06309724564362e-05, "loss": 0.7031, "step": 1514 }, { "epoch": 0.06387419103231655, "grad_norm": 1.3136532306671143, "learning_rate": 1.0637998875772906e-05, "loss": 0.7901, "step": 1515 }, { "epoch": 0.0639163522145161, "grad_norm": 1.408179521560669, "learning_rate": 1.0645025295109613e-05, "loss": 0.3752, "step": 1516 }, { "epoch": 0.06395851339671564, "grad_norm": 2.710160493850708, "learning_rate": 1.065205171444632e-05, "loss": 0.8091, "step": 1517 }, { "epoch": 0.06400067457891519, "grad_norm": 1.2084609270095825, "learning_rate": 1.0659078133783026e-05, "loss": 0.2313, "step": 1518 }, { "epoch": 0.06404283576111475, "grad_norm": 1.135506272315979, "learning_rate": 1.0666104553119732e-05, "loss": 0.3059, "step": 1519 }, { "epoch": 0.06408499694331429, "grad_norm": 1.173912763595581, "learning_rate": 1.0673130972456436e-05, "loss": 0.5866, "step": 1520 }, { "epoch": 0.06412715812551384, "grad_norm": 1.2886641025543213, "learning_rate": 1.0680157391793142e-05, "loss": 0.3378, "step": 1521 }, { "epoch": 0.06416931930771338, "grad_norm": 1.9204789400100708, "learning_rate": 1.0687183811129848e-05, "loss": 0.2117, "step": 1522 }, { "epoch": 0.06421148048991293, "grad_norm": 1.4352269172668457, "learning_rate": 1.0694210230466554e-05, "loss": 0.5709, "step": 1523 }, { "epoch": 0.06425364167211249, "grad_norm": 1.217194676399231, "learning_rate": 1.070123664980326e-05, "loss": 0.49, "step": 1524 }, { "epoch": 0.06429580285431204, "grad_norm": 1.7930516004562378, "learning_rate": 1.0708263069139966e-05, "loss": 1.3797, "step": 1525 }, { "epoch": 0.06433796403651158, "grad_norm": 1.3107243776321411, "learning_rate": 1.0715289488476673e-05, "loss": 0.3048, "step": 1526 }, { "epoch": 0.06438012521871113, "grad_norm": 1.1610642671585083, "learning_rate": 1.0722315907813379e-05, "loss": 0.5225, "step": 1527 }, { "epoch": 0.06442228640091067, "grad_norm": 1.6879764795303345, "learning_rate": 1.0729342327150085e-05, "loss": 1.1613, "step": 1528 }, { "epoch": 0.06446444758311023, "grad_norm": 3.26578426361084, "learning_rate": 1.073636874648679e-05, "loss": 1.0537, "step": 1529 }, { "epoch": 0.06450660876530978, "grad_norm": 1.3370931148529053, "learning_rate": 1.0743395165823497e-05, "loss": 0.2834, "step": 1530 }, { "epoch": 0.06454876994750933, "grad_norm": 1.7943699359893799, "learning_rate": 1.0750421585160203e-05, "loss": 1.1304, "step": 1531 }, { "epoch": 0.06459093112970887, "grad_norm": 1.330148458480835, "learning_rate": 1.0757448004496909e-05, "loss": 0.6502, "step": 1532 }, { "epoch": 0.06463309231190843, "grad_norm": 0.8661600351333618, "learning_rate": 1.0764474423833615e-05, "loss": 0.8622, "step": 1533 }, { "epoch": 0.06467525349410798, "grad_norm": 1.4716988801956177, "learning_rate": 1.0771500843170321e-05, "loss": 0.4405, "step": 1534 }, { "epoch": 0.06471741467630752, "grad_norm": 1.7810665369033813, "learning_rate": 1.0778527262507027e-05, "loss": 0.4944, "step": 1535 }, { "epoch": 0.06475957585850707, "grad_norm": 1.3858407735824585, "learning_rate": 1.0785553681843732e-05, "loss": 0.3543, "step": 1536 }, { "epoch": 0.06480173704070662, "grad_norm": 1.3549268245697021, "learning_rate": 1.0792580101180438e-05, "loss": 0.3014, "step": 1537 }, { "epoch": 0.06484389822290618, "grad_norm": 1.1119199991226196, "learning_rate": 1.0799606520517146e-05, "loss": 0.2741, "step": 1538 }, { "epoch": 0.06488605940510572, "grad_norm": 1.4847760200500488, "learning_rate": 1.0806632939853852e-05, "loss": 0.2563, "step": 1539 }, { "epoch": 0.06492822058730527, "grad_norm": 0.9686020612716675, "learning_rate": 1.0813659359190558e-05, "loss": 0.2065, "step": 1540 }, { "epoch": 0.06497038176950481, "grad_norm": 2.6256284713745117, "learning_rate": 1.0820685778527264e-05, "loss": 1.2888, "step": 1541 }, { "epoch": 0.06501254295170436, "grad_norm": 1.3497343063354492, "learning_rate": 1.082771219786397e-05, "loss": 0.3183, "step": 1542 }, { "epoch": 0.06505470413390392, "grad_norm": 0.8125966191291809, "learning_rate": 1.0834738617200674e-05, "loss": 0.2176, "step": 1543 }, { "epoch": 0.06509686531610347, "grad_norm": 1.1437333822250366, "learning_rate": 1.084176503653738e-05, "loss": 0.3085, "step": 1544 }, { "epoch": 0.06513902649830301, "grad_norm": 1.334371566772461, "learning_rate": 1.0848791455874087e-05, "loss": 0.9039, "step": 1545 }, { "epoch": 0.06518118768050256, "grad_norm": 1.5547131299972534, "learning_rate": 1.0855817875210793e-05, "loss": 0.2056, "step": 1546 }, { "epoch": 0.0652233488627021, "grad_norm": 1.396592378616333, "learning_rate": 1.0862844294547499e-05, "loss": 0.9375, "step": 1547 }, { "epoch": 0.06526551004490166, "grad_norm": 1.1541625261306763, "learning_rate": 1.0869870713884205e-05, "loss": 0.5208, "step": 1548 }, { "epoch": 0.06530767122710121, "grad_norm": 2.0975751876831055, "learning_rate": 1.0876897133220911e-05, "loss": 0.9604, "step": 1549 }, { "epoch": 0.06534983240930076, "grad_norm": 0.8914410471916199, "learning_rate": 1.0883923552557617e-05, "loss": 0.255, "step": 1550 }, { "epoch": 0.0653919935915003, "grad_norm": 0.8643156886100769, "learning_rate": 1.0890949971894323e-05, "loss": 0.2099, "step": 1551 }, { "epoch": 0.06543415477369985, "grad_norm": 0.9693585634231567, "learning_rate": 1.089797639123103e-05, "loss": 0.5427, "step": 1552 }, { "epoch": 0.06547631595589941, "grad_norm": 2.146669864654541, "learning_rate": 1.0905002810567735e-05, "loss": 0.4727, "step": 1553 }, { "epoch": 0.06551847713809895, "grad_norm": 2.0976572036743164, "learning_rate": 1.0912029229904441e-05, "loss": 0.4863, "step": 1554 }, { "epoch": 0.0655606383202985, "grad_norm": 1.1166956424713135, "learning_rate": 1.0919055649241147e-05, "loss": 0.4647, "step": 1555 }, { "epoch": 0.06560279950249805, "grad_norm": 1.8876404762268066, "learning_rate": 1.0926082068577854e-05, "loss": 1.1596, "step": 1556 }, { "epoch": 0.06564496068469759, "grad_norm": 1.0948317050933838, "learning_rate": 1.093310848791456e-05, "loss": 0.2636, "step": 1557 }, { "epoch": 0.06568712186689715, "grad_norm": 1.122372031211853, "learning_rate": 1.0940134907251266e-05, "loss": 0.2195, "step": 1558 }, { "epoch": 0.0657292830490967, "grad_norm": 1.4354122877120972, "learning_rate": 1.094716132658797e-05, "loss": 0.306, "step": 1559 }, { "epoch": 0.06577144423129624, "grad_norm": 1.4385533332824707, "learning_rate": 1.0954187745924676e-05, "loss": 0.671, "step": 1560 }, { "epoch": 0.06581360541349579, "grad_norm": 1.427394151687622, "learning_rate": 1.0961214165261384e-05, "loss": 0.3343, "step": 1561 }, { "epoch": 0.06585576659569534, "grad_norm": 1.0922075510025024, "learning_rate": 1.096824058459809e-05, "loss": 0.6852, "step": 1562 }, { "epoch": 0.0658979277778949, "grad_norm": 1.3677270412445068, "learning_rate": 1.0975267003934796e-05, "loss": 0.2143, "step": 1563 }, { "epoch": 0.06594008896009444, "grad_norm": 1.383069634437561, "learning_rate": 1.0982293423271502e-05, "loss": 0.3271, "step": 1564 }, { "epoch": 0.06598225014229399, "grad_norm": 0.7595987915992737, "learning_rate": 1.0989319842608208e-05, "loss": 0.1257, "step": 1565 }, { "epoch": 0.06602441132449353, "grad_norm": 1.0944857597351074, "learning_rate": 1.0996346261944913e-05, "loss": 0.1651, "step": 1566 }, { "epoch": 0.0660665725066931, "grad_norm": 1.6839845180511475, "learning_rate": 1.1003372681281619e-05, "loss": 0.3285, "step": 1567 }, { "epoch": 0.06610873368889264, "grad_norm": 1.7454127073287964, "learning_rate": 1.1010399100618325e-05, "loss": 1.3006, "step": 1568 }, { "epoch": 0.06615089487109219, "grad_norm": 0.8722062706947327, "learning_rate": 1.1017425519955031e-05, "loss": 0.1941, "step": 1569 }, { "epoch": 0.06619305605329173, "grad_norm": 2.4285147190093994, "learning_rate": 1.1024451939291737e-05, "loss": 1.3044, "step": 1570 }, { "epoch": 0.06623521723549128, "grad_norm": 1.8088754415512085, "learning_rate": 1.1031478358628443e-05, "loss": 1.3191, "step": 1571 }, { "epoch": 0.06627737841769084, "grad_norm": 2.4225776195526123, "learning_rate": 1.1038504777965151e-05, "loss": 0.5891, "step": 1572 }, { "epoch": 0.06631953959989038, "grad_norm": 1.2909201383590698, "learning_rate": 1.1045531197301855e-05, "loss": 0.29, "step": 1573 }, { "epoch": 0.06636170078208993, "grad_norm": 1.1591103076934814, "learning_rate": 1.1052557616638561e-05, "loss": 0.2422, "step": 1574 }, { "epoch": 0.06640386196428948, "grad_norm": 1.429079532623291, "learning_rate": 1.1059584035975268e-05, "loss": 0.7961, "step": 1575 }, { "epoch": 0.06644602314648902, "grad_norm": 0.7375262975692749, "learning_rate": 1.1066610455311974e-05, "loss": 0.1948, "step": 1576 }, { "epoch": 0.06648818432868858, "grad_norm": 1.051797866821289, "learning_rate": 1.107363687464868e-05, "loss": 0.2787, "step": 1577 }, { "epoch": 0.06653034551088813, "grad_norm": 1.4860717058181763, "learning_rate": 1.1080663293985386e-05, "loss": 0.6685, "step": 1578 }, { "epoch": 0.06657250669308767, "grad_norm": 1.3156912326812744, "learning_rate": 1.1087689713322092e-05, "loss": 0.2125, "step": 1579 }, { "epoch": 0.06661466787528722, "grad_norm": 1.7454779148101807, "learning_rate": 1.1094716132658798e-05, "loss": 1.2501, "step": 1580 }, { "epoch": 0.06665682905748677, "grad_norm": 0.7230253219604492, "learning_rate": 1.1101742551995502e-05, "loss": 0.1734, "step": 1581 }, { "epoch": 0.06669899023968633, "grad_norm": 2.0993967056274414, "learning_rate": 1.1108768971332208e-05, "loss": 1.2563, "step": 1582 }, { "epoch": 0.06674115142188587, "grad_norm": 1.3295912742614746, "learning_rate": 1.1115795390668916e-05, "loss": 0.648, "step": 1583 }, { "epoch": 0.06678331260408542, "grad_norm": 1.287355661392212, "learning_rate": 1.1122821810005622e-05, "loss": 0.2495, "step": 1584 }, { "epoch": 0.06682547378628496, "grad_norm": 1.2764739990234375, "learning_rate": 1.1129848229342328e-05, "loss": 0.4514, "step": 1585 }, { "epoch": 0.06686763496848451, "grad_norm": 0.9429674744606018, "learning_rate": 1.1136874648679035e-05, "loss": 0.3224, "step": 1586 }, { "epoch": 0.06690979615068407, "grad_norm": 1.089879035949707, "learning_rate": 1.114390106801574e-05, "loss": 0.6165, "step": 1587 }, { "epoch": 0.06695195733288362, "grad_norm": 1.2138869762420654, "learning_rate": 1.1150927487352445e-05, "loss": 0.6639, "step": 1588 }, { "epoch": 0.06699411851508316, "grad_norm": 2.0005691051483154, "learning_rate": 1.1157953906689151e-05, "loss": 0.5443, "step": 1589 }, { "epoch": 0.06703627969728271, "grad_norm": 1.3874558210372925, "learning_rate": 1.1164980326025857e-05, "loss": 0.4565, "step": 1590 }, { "epoch": 0.06707844087948225, "grad_norm": 1.375388503074646, "learning_rate": 1.1172006745362563e-05, "loss": 0.5647, "step": 1591 }, { "epoch": 0.06712060206168181, "grad_norm": 1.9436942338943481, "learning_rate": 1.117903316469927e-05, "loss": 1.0595, "step": 1592 }, { "epoch": 0.06716276324388136, "grad_norm": 1.2318906784057617, "learning_rate": 1.1186059584035975e-05, "loss": 0.2214, "step": 1593 }, { "epoch": 0.0672049244260809, "grad_norm": 4.086311340332031, "learning_rate": 1.1193086003372682e-05, "loss": 1.1262, "step": 1594 }, { "epoch": 0.06724708560828045, "grad_norm": 1.2877278327941895, "learning_rate": 1.1200112422709388e-05, "loss": 0.3402, "step": 1595 }, { "epoch": 0.06728924679048001, "grad_norm": 1.1125104427337646, "learning_rate": 1.1207138842046094e-05, "loss": 0.4885, "step": 1596 }, { "epoch": 0.06733140797267956, "grad_norm": 1.1461193561553955, "learning_rate": 1.12141652613828e-05, "loss": 0.2495, "step": 1597 }, { "epoch": 0.0673735691548791, "grad_norm": 1.1505095958709717, "learning_rate": 1.1221191680719506e-05, "loss": 0.7995, "step": 1598 }, { "epoch": 0.06741573033707865, "grad_norm": 1.481842041015625, "learning_rate": 1.1228218100056212e-05, "loss": 0.2708, "step": 1599 }, { "epoch": 0.0674578915192782, "grad_norm": 1.291895866394043, "learning_rate": 1.1235244519392918e-05, "loss": 0.2232, "step": 1600 }, { "epoch": 0.06750005270147776, "grad_norm": 1.6065996885299683, "learning_rate": 1.1242270938729624e-05, "loss": 0.5382, "step": 1601 }, { "epoch": 0.0675422138836773, "grad_norm": 1.031853437423706, "learning_rate": 1.124929735806633e-05, "loss": 0.1642, "step": 1602 }, { "epoch": 0.06758437506587685, "grad_norm": 0.8386584520339966, "learning_rate": 1.1256323777403036e-05, "loss": 0.1482, "step": 1603 }, { "epoch": 0.0676265362480764, "grad_norm": 0.9435812830924988, "learning_rate": 1.126335019673974e-05, "loss": 0.5126, "step": 1604 }, { "epoch": 0.06766869743027594, "grad_norm": 1.1831287145614624, "learning_rate": 1.1270376616076447e-05, "loss": 0.7093, "step": 1605 }, { "epoch": 0.0677108586124755, "grad_norm": 1.777194857597351, "learning_rate": 1.1277403035413155e-05, "loss": 0.5377, "step": 1606 }, { "epoch": 0.06775301979467505, "grad_norm": 1.7349321842193604, "learning_rate": 1.128442945474986e-05, "loss": 0.1828, "step": 1607 }, { "epoch": 0.06779518097687459, "grad_norm": 0.9402649402618408, "learning_rate": 1.1291455874086567e-05, "loss": 0.2939, "step": 1608 }, { "epoch": 0.06783734215907414, "grad_norm": 1.5485703945159912, "learning_rate": 1.1298482293423273e-05, "loss": 0.3721, "step": 1609 }, { "epoch": 0.06787950334127368, "grad_norm": 1.1994740962982178, "learning_rate": 1.1305508712759979e-05, "loss": 0.1456, "step": 1610 }, { "epoch": 0.06792166452347324, "grad_norm": 1.359230637550354, "learning_rate": 1.1312535132096683e-05, "loss": 0.1978, "step": 1611 }, { "epoch": 0.06796382570567279, "grad_norm": 1.5837509632110596, "learning_rate": 1.131956155143339e-05, "loss": 0.3259, "step": 1612 }, { "epoch": 0.06800598688787234, "grad_norm": 6.289361953735352, "learning_rate": 1.1326587970770096e-05, "loss": 0.9108, "step": 1613 }, { "epoch": 0.06804814807007188, "grad_norm": 1.1751134395599365, "learning_rate": 1.1333614390106802e-05, "loss": 0.6205, "step": 1614 }, { "epoch": 0.06809030925227143, "grad_norm": 1.053402066230774, "learning_rate": 1.1340640809443508e-05, "loss": 0.2839, "step": 1615 }, { "epoch": 0.06813247043447099, "grad_norm": 1.826067566871643, "learning_rate": 1.1347667228780214e-05, "loss": 0.9402, "step": 1616 }, { "epoch": 0.06817463161667053, "grad_norm": 1.2135528326034546, "learning_rate": 1.1354693648116922e-05, "loss": 0.3616, "step": 1617 }, { "epoch": 0.06821679279887008, "grad_norm": 1.4278461933135986, "learning_rate": 1.1361720067453626e-05, "loss": 0.4041, "step": 1618 }, { "epoch": 0.06825895398106963, "grad_norm": 1.8255501985549927, "learning_rate": 1.1368746486790332e-05, "loss": 0.6269, "step": 1619 }, { "epoch": 0.06830111516326917, "grad_norm": 3.170072317123413, "learning_rate": 1.1375772906127038e-05, "loss": 1.0153, "step": 1620 }, { "epoch": 0.06834327634546873, "grad_norm": 1.2721643447875977, "learning_rate": 1.1382799325463744e-05, "loss": 0.3892, "step": 1621 }, { "epoch": 0.06838543752766828, "grad_norm": 1.0475698709487915, "learning_rate": 1.138982574480045e-05, "loss": 0.3119, "step": 1622 }, { "epoch": 0.06842759870986782, "grad_norm": 1.7641953229904175, "learning_rate": 1.1396852164137156e-05, "loss": 0.2323, "step": 1623 }, { "epoch": 0.06846975989206737, "grad_norm": 1.2979716062545776, "learning_rate": 1.1403878583473863e-05, "loss": 0.5368, "step": 1624 }, { "epoch": 0.06851192107426693, "grad_norm": 1.486194372177124, "learning_rate": 1.1410905002810569e-05, "loss": 0.1347, "step": 1625 }, { "epoch": 0.06855408225646648, "grad_norm": 0.9987601041793823, "learning_rate": 1.1417931422147273e-05, "loss": 0.2085, "step": 1626 }, { "epoch": 0.06859624343866602, "grad_norm": 1.4398937225341797, "learning_rate": 1.1424957841483979e-05, "loss": 0.5669, "step": 1627 }, { "epoch": 0.06863840462086557, "grad_norm": 1.0490971803665161, "learning_rate": 1.1431984260820687e-05, "loss": 0.1943, "step": 1628 }, { "epoch": 0.06868056580306511, "grad_norm": 1.062172532081604, "learning_rate": 1.1439010680157393e-05, "loss": 0.4233, "step": 1629 }, { "epoch": 0.06872272698526467, "grad_norm": 2.1207470893859863, "learning_rate": 1.1446037099494099e-05, "loss": 1.0599, "step": 1630 }, { "epoch": 0.06876488816746422, "grad_norm": 2.9607443809509277, "learning_rate": 1.1453063518830805e-05, "loss": 1.1746, "step": 1631 }, { "epoch": 0.06880704934966377, "grad_norm": 0.8366866707801819, "learning_rate": 1.1460089938167511e-05, "loss": 0.1923, "step": 1632 }, { "epoch": 0.06884921053186331, "grad_norm": 1.6873162984848022, "learning_rate": 1.1467116357504216e-05, "loss": 0.3086, "step": 1633 }, { "epoch": 0.06889137171406286, "grad_norm": 1.0169156789779663, "learning_rate": 1.1474142776840922e-05, "loss": 0.2035, "step": 1634 }, { "epoch": 0.06893353289626242, "grad_norm": 1.6022756099700928, "learning_rate": 1.1481169196177628e-05, "loss": 0.3488, "step": 1635 }, { "epoch": 0.06897569407846196, "grad_norm": 1.4865529537200928, "learning_rate": 1.1488195615514334e-05, "loss": 0.8664, "step": 1636 }, { "epoch": 0.06901785526066151, "grad_norm": 1.0216110944747925, "learning_rate": 1.149522203485104e-05, "loss": 0.5829, "step": 1637 }, { "epoch": 0.06906001644286106, "grad_norm": 0.968539834022522, "learning_rate": 1.1502248454187746e-05, "loss": 0.4001, "step": 1638 }, { "epoch": 0.0691021776250606, "grad_norm": 1.6950291395187378, "learning_rate": 1.1509274873524452e-05, "loss": 1.1439, "step": 1639 }, { "epoch": 0.06914433880726016, "grad_norm": 1.2385958433151245, "learning_rate": 1.1516301292861158e-05, "loss": 0.1447, "step": 1640 }, { "epoch": 0.06918649998945971, "grad_norm": 1.2622110843658447, "learning_rate": 1.1523327712197864e-05, "loss": 0.2077, "step": 1641 }, { "epoch": 0.06922866117165925, "grad_norm": 1.4020189046859741, "learning_rate": 1.153035413153457e-05, "loss": 1.0574, "step": 1642 }, { "epoch": 0.0692708223538588, "grad_norm": 1.1326895952224731, "learning_rate": 1.1537380550871277e-05, "loss": 0.4898, "step": 1643 }, { "epoch": 0.06931298353605835, "grad_norm": 1.9310038089752197, "learning_rate": 1.1544406970207983e-05, "loss": 0.8632, "step": 1644 }, { "epoch": 0.0693551447182579, "grad_norm": 1.1396918296813965, "learning_rate": 1.1551433389544689e-05, "loss": 0.6926, "step": 1645 }, { "epoch": 0.06939730590045745, "grad_norm": 1.6484512090682983, "learning_rate": 1.1558459808881395e-05, "loss": 0.2307, "step": 1646 }, { "epoch": 0.069439467082657, "grad_norm": 2.2454416751861572, "learning_rate": 1.1565486228218101e-05, "loss": 0.2888, "step": 1647 }, { "epoch": 0.06948162826485654, "grad_norm": 1.7475181818008423, "learning_rate": 1.1572512647554807e-05, "loss": 1.2015, "step": 1648 }, { "epoch": 0.06952378944705609, "grad_norm": 2.115004539489746, "learning_rate": 1.1579539066891511e-05, "loss": 0.4155, "step": 1649 }, { "epoch": 0.06956595062925565, "grad_norm": 1.0879888534545898, "learning_rate": 1.1586565486228217e-05, "loss": 0.1697, "step": 1650 }, { "epoch": 0.0696081118114552, "grad_norm": 1.146032691001892, "learning_rate": 1.1593591905564925e-05, "loss": 0.8549, "step": 1651 }, { "epoch": 0.06965027299365474, "grad_norm": 2.342744827270508, "learning_rate": 1.1600618324901631e-05, "loss": 1.3112, "step": 1652 }, { "epoch": 0.06969243417585429, "grad_norm": 1.333519458770752, "learning_rate": 1.1607644744238337e-05, "loss": 1.2714, "step": 1653 }, { "epoch": 0.06973459535805383, "grad_norm": 1.2375476360321045, "learning_rate": 1.1614671163575044e-05, "loss": 0.7726, "step": 1654 }, { "epoch": 0.06977675654025339, "grad_norm": 1.3301458358764648, "learning_rate": 1.162169758291175e-05, "loss": 0.7644, "step": 1655 }, { "epoch": 0.06981891772245294, "grad_norm": 0.9691226482391357, "learning_rate": 1.1628724002248454e-05, "loss": 0.227, "step": 1656 }, { "epoch": 0.06986107890465248, "grad_norm": 0.9037390351295471, "learning_rate": 1.163575042158516e-05, "loss": 0.1854, "step": 1657 }, { "epoch": 0.06990324008685203, "grad_norm": 1.3180558681488037, "learning_rate": 1.1642776840921866e-05, "loss": 0.338, "step": 1658 }, { "epoch": 0.06994540126905159, "grad_norm": 1.2237848043441772, "learning_rate": 1.1649803260258572e-05, "loss": 0.2165, "step": 1659 }, { "epoch": 0.06998756245125114, "grad_norm": 0.855137288570404, "learning_rate": 1.1656829679595278e-05, "loss": 0.4867, "step": 1660 }, { "epoch": 0.07002972363345068, "grad_norm": 0.9680227637290955, "learning_rate": 1.1663856098931984e-05, "loss": 0.4043, "step": 1661 }, { "epoch": 0.07007188481565023, "grad_norm": 1.351072907447815, "learning_rate": 1.1670882518268692e-05, "loss": 0.321, "step": 1662 }, { "epoch": 0.07011404599784977, "grad_norm": 1.8027880191802979, "learning_rate": 1.1677908937605397e-05, "loss": 1.1784, "step": 1663 }, { "epoch": 0.07015620718004933, "grad_norm": 1.0331264734268188, "learning_rate": 1.1684935356942103e-05, "loss": 0.2486, "step": 1664 }, { "epoch": 0.07019836836224888, "grad_norm": 1.2980626821517944, "learning_rate": 1.1691961776278809e-05, "loss": 0.2514, "step": 1665 }, { "epoch": 0.07024052954444843, "grad_norm": 1.1854840517044067, "learning_rate": 1.1698988195615515e-05, "loss": 0.2592, "step": 1666 }, { "epoch": 0.07028269072664797, "grad_norm": 1.163012146949768, "learning_rate": 1.1706014614952221e-05, "loss": 0.6407, "step": 1667 }, { "epoch": 0.07032485190884752, "grad_norm": 1.8870790004730225, "learning_rate": 1.1713041034288927e-05, "loss": 1.2687, "step": 1668 }, { "epoch": 0.07036701309104708, "grad_norm": 0.986548900604248, "learning_rate": 1.1720067453625633e-05, "loss": 0.2351, "step": 1669 }, { "epoch": 0.07040917427324662, "grad_norm": 0.9221737384796143, "learning_rate": 1.172709387296234e-05, "loss": 0.2648, "step": 1670 }, { "epoch": 0.07045133545544617, "grad_norm": 2.165529251098633, "learning_rate": 1.1734120292299045e-05, "loss": 1.1162, "step": 1671 }, { "epoch": 0.07049349663764572, "grad_norm": 1.295311689376831, "learning_rate": 1.174114671163575e-05, "loss": 0.5687, "step": 1672 }, { "epoch": 0.07053565781984526, "grad_norm": 1.3063442707061768, "learning_rate": 1.1748173130972456e-05, "loss": 0.2453, "step": 1673 }, { "epoch": 0.07057781900204482, "grad_norm": 1.3919248580932617, "learning_rate": 1.1755199550309164e-05, "loss": 0.5805, "step": 1674 }, { "epoch": 0.07061998018424437, "grad_norm": 1.1651173830032349, "learning_rate": 1.176222596964587e-05, "loss": 0.8645, "step": 1675 }, { "epoch": 0.07066214136644391, "grad_norm": 1.151070475578308, "learning_rate": 1.1769252388982576e-05, "loss": 0.7032, "step": 1676 }, { "epoch": 0.07070430254864346, "grad_norm": 1.0305547714233398, "learning_rate": 1.1776278808319282e-05, "loss": 0.3457, "step": 1677 }, { "epoch": 0.070746463730843, "grad_norm": 0.8141192197799683, "learning_rate": 1.1783305227655988e-05, "loss": 0.1626, "step": 1678 }, { "epoch": 0.07078862491304257, "grad_norm": 1.5212124586105347, "learning_rate": 1.1790331646992692e-05, "loss": 0.656, "step": 1679 }, { "epoch": 0.07083078609524211, "grad_norm": 1.432627558708191, "learning_rate": 1.1797358066329398e-05, "loss": 0.6759, "step": 1680 }, { "epoch": 0.07087294727744166, "grad_norm": 1.0842461585998535, "learning_rate": 1.1804384485666105e-05, "loss": 0.7763, "step": 1681 }, { "epoch": 0.0709151084596412, "grad_norm": 1.8081040382385254, "learning_rate": 1.181141090500281e-05, "loss": 0.1918, "step": 1682 }, { "epoch": 0.07095726964184075, "grad_norm": 1.3080852031707764, "learning_rate": 1.1818437324339517e-05, "loss": 1.2255, "step": 1683 }, { "epoch": 0.07099943082404031, "grad_norm": 0.8824946284294128, "learning_rate": 1.1825463743676223e-05, "loss": 0.3096, "step": 1684 }, { "epoch": 0.07104159200623986, "grad_norm": 1.1844924688339233, "learning_rate": 1.183249016301293e-05, "loss": 0.326, "step": 1685 }, { "epoch": 0.0710837531884394, "grad_norm": 2.895507574081421, "learning_rate": 1.1839516582349635e-05, "loss": 0.3131, "step": 1686 }, { "epoch": 0.07112591437063895, "grad_norm": 1.2858359813690186, "learning_rate": 1.1846543001686341e-05, "loss": 0.2285, "step": 1687 }, { "epoch": 0.07116807555283851, "grad_norm": 3.116950750350952, "learning_rate": 1.1853569421023047e-05, "loss": 1.2325, "step": 1688 }, { "epoch": 0.07121023673503805, "grad_norm": 1.0263723134994507, "learning_rate": 1.1860595840359753e-05, "loss": 0.1522, "step": 1689 }, { "epoch": 0.0712523979172376, "grad_norm": 1.1800376176834106, "learning_rate": 1.186762225969646e-05, "loss": 0.2771, "step": 1690 }, { "epoch": 0.07129455909943715, "grad_norm": 0.9522534012794495, "learning_rate": 1.1874648679033165e-05, "loss": 0.171, "step": 1691 }, { "epoch": 0.07133672028163669, "grad_norm": 0.8425891995429993, "learning_rate": 1.1881675098369872e-05, "loss": 0.1531, "step": 1692 }, { "epoch": 0.07137888146383625, "grad_norm": 2.347036123275757, "learning_rate": 1.1888701517706578e-05, "loss": 0.394, "step": 1693 }, { "epoch": 0.0714210426460358, "grad_norm": 1.2311586141586304, "learning_rate": 1.1895727937043282e-05, "loss": 0.7856, "step": 1694 }, { "epoch": 0.07146320382823534, "grad_norm": 1.2288106679916382, "learning_rate": 1.1902754356379988e-05, "loss": 0.3983, "step": 1695 }, { "epoch": 0.07150536501043489, "grad_norm": 2.1572840213775635, "learning_rate": 1.1909780775716696e-05, "loss": 0.3845, "step": 1696 }, { "epoch": 0.07154752619263444, "grad_norm": 0.9323369860649109, "learning_rate": 1.1916807195053402e-05, "loss": 0.2886, "step": 1697 }, { "epoch": 0.071589687374834, "grad_norm": 1.6015433073043823, "learning_rate": 1.1923833614390108e-05, "loss": 1.1356, "step": 1698 }, { "epoch": 0.07163184855703354, "grad_norm": 1.2193889617919922, "learning_rate": 1.1930860033726814e-05, "loss": 0.2838, "step": 1699 }, { "epoch": 0.07167400973923309, "grad_norm": 1.0566426515579224, "learning_rate": 1.193788645306352e-05, "loss": 0.4284, "step": 1700 }, { "epoch": 0.07171617092143263, "grad_norm": 1.1757360696792603, "learning_rate": 1.1944912872400225e-05, "loss": 0.7961, "step": 1701 }, { "epoch": 0.07175833210363218, "grad_norm": 1.3012064695358276, "learning_rate": 1.195193929173693e-05, "loss": 0.3104, "step": 1702 }, { "epoch": 0.07180049328583174, "grad_norm": 1.1169227361679077, "learning_rate": 1.1958965711073637e-05, "loss": 0.307, "step": 1703 }, { "epoch": 0.07184265446803129, "grad_norm": 1.3164385557174683, "learning_rate": 1.1965992130410343e-05, "loss": 0.3388, "step": 1704 }, { "epoch": 0.07188481565023083, "grad_norm": 1.9858678579330444, "learning_rate": 1.1973018549747049e-05, "loss": 0.3129, "step": 1705 }, { "epoch": 0.07192697683243038, "grad_norm": 1.2599104642868042, "learning_rate": 1.1980044969083755e-05, "loss": 0.1343, "step": 1706 }, { "epoch": 0.07196913801462992, "grad_norm": 1.096877098083496, "learning_rate": 1.1987071388420463e-05, "loss": 0.5902, "step": 1707 }, { "epoch": 0.07201129919682948, "grad_norm": 1.4007869958877563, "learning_rate": 1.1994097807757167e-05, "loss": 0.3578, "step": 1708 }, { "epoch": 0.07205346037902903, "grad_norm": 1.650864839553833, "learning_rate": 1.2001124227093873e-05, "loss": 1.3108, "step": 1709 }, { "epoch": 0.07209562156122858, "grad_norm": 1.5932680368423462, "learning_rate": 1.200815064643058e-05, "loss": 0.253, "step": 1710 }, { "epoch": 0.07213778274342812, "grad_norm": 1.3771414756774902, "learning_rate": 1.2015177065767286e-05, "loss": 0.231, "step": 1711 }, { "epoch": 0.07217994392562767, "grad_norm": 1.422321081161499, "learning_rate": 1.2022203485103992e-05, "loss": 0.4056, "step": 1712 }, { "epoch": 0.07222210510782723, "grad_norm": 1.5894345045089722, "learning_rate": 1.2029229904440698e-05, "loss": 1.2818, "step": 1713 }, { "epoch": 0.07226426629002677, "grad_norm": 5.0574049949646, "learning_rate": 1.2036256323777404e-05, "loss": 0.7847, "step": 1714 }, { "epoch": 0.07230642747222632, "grad_norm": 1.901742696762085, "learning_rate": 1.204328274311411e-05, "loss": 0.4694, "step": 1715 }, { "epoch": 0.07234858865442587, "grad_norm": 1.7519307136535645, "learning_rate": 1.2050309162450816e-05, "loss": 0.6766, "step": 1716 }, { "epoch": 0.07239074983662541, "grad_norm": 1.193814992904663, "learning_rate": 1.205733558178752e-05, "loss": 0.4519, "step": 1717 }, { "epoch": 0.07243291101882497, "grad_norm": 1.6163188219070435, "learning_rate": 1.2064362001124226e-05, "loss": 0.5696, "step": 1718 }, { "epoch": 0.07247507220102452, "grad_norm": 2.9988818168640137, "learning_rate": 1.2071388420460934e-05, "loss": 1.1737, "step": 1719 }, { "epoch": 0.07251723338322406, "grad_norm": 2.0087592601776123, "learning_rate": 1.207841483979764e-05, "loss": 0.2559, "step": 1720 }, { "epoch": 0.07255939456542361, "grad_norm": 0.9207276105880737, "learning_rate": 1.2085441259134346e-05, "loss": 0.2011, "step": 1721 }, { "epoch": 0.07260155574762317, "grad_norm": 2.763491153717041, "learning_rate": 1.2092467678471053e-05, "loss": 1.0102, "step": 1722 }, { "epoch": 0.07264371692982272, "grad_norm": 1.4780703783035278, "learning_rate": 1.2099494097807759e-05, "loss": 0.2087, "step": 1723 }, { "epoch": 0.07268587811202226, "grad_norm": 5.222771167755127, "learning_rate": 1.2106520517144463e-05, "loss": 0.7579, "step": 1724 }, { "epoch": 0.07272803929422181, "grad_norm": 1.1128129959106445, "learning_rate": 1.2113546936481169e-05, "loss": 0.6296, "step": 1725 }, { "epoch": 0.07277020047642135, "grad_norm": 3.29953932762146, "learning_rate": 1.2120573355817875e-05, "loss": 0.8512, "step": 1726 }, { "epoch": 0.07281236165862091, "grad_norm": 0.8998052477836609, "learning_rate": 1.2127599775154581e-05, "loss": 0.1636, "step": 1727 }, { "epoch": 0.07285452284082046, "grad_norm": 0.9238392114639282, "learning_rate": 1.2134626194491287e-05, "loss": 0.2525, "step": 1728 }, { "epoch": 0.07289668402302, "grad_norm": 1.2436487674713135, "learning_rate": 1.2141652613827993e-05, "loss": 0.3945, "step": 1729 }, { "epoch": 0.07293884520521955, "grad_norm": 1.0524195432662964, "learning_rate": 1.2148679033164701e-05, "loss": 0.3256, "step": 1730 }, { "epoch": 0.0729810063874191, "grad_norm": 1.2487611770629883, "learning_rate": 1.2155705452501406e-05, "loss": 0.5189, "step": 1731 }, { "epoch": 0.07302316756961866, "grad_norm": 1.6214015483856201, "learning_rate": 1.2162731871838112e-05, "loss": 0.6816, "step": 1732 }, { "epoch": 0.0730653287518182, "grad_norm": 2.985703468322754, "learning_rate": 1.2169758291174818e-05, "loss": 0.5952, "step": 1733 }, { "epoch": 0.07310748993401775, "grad_norm": 0.8368833065032959, "learning_rate": 1.2176784710511524e-05, "loss": 0.1873, "step": 1734 }, { "epoch": 0.0731496511162173, "grad_norm": 1.1143851280212402, "learning_rate": 1.218381112984823e-05, "loss": 0.7655, "step": 1735 }, { "epoch": 0.07319181229841684, "grad_norm": 0.915005087852478, "learning_rate": 1.2190837549184936e-05, "loss": 0.6067, "step": 1736 }, { "epoch": 0.0732339734806164, "grad_norm": 1.3638432025909424, "learning_rate": 1.2197863968521642e-05, "loss": 0.5624, "step": 1737 }, { "epoch": 0.07327613466281595, "grad_norm": 1.5020257234573364, "learning_rate": 1.2204890387858348e-05, "loss": 0.2765, "step": 1738 }, { "epoch": 0.0733182958450155, "grad_norm": 1.5870850086212158, "learning_rate": 1.2211916807195054e-05, "loss": 0.3025, "step": 1739 }, { "epoch": 0.07336045702721504, "grad_norm": 1.1108744144439697, "learning_rate": 1.2218943226531759e-05, "loss": 0.5562, "step": 1740 }, { "epoch": 0.07340261820941459, "grad_norm": 1.6598539352416992, "learning_rate": 1.2225969645868467e-05, "loss": 0.2725, "step": 1741 }, { "epoch": 0.07344477939161415, "grad_norm": 1.3312420845031738, "learning_rate": 1.2232996065205173e-05, "loss": 0.3975, "step": 1742 }, { "epoch": 0.07348694057381369, "grad_norm": 1.0821735858917236, "learning_rate": 1.2240022484541879e-05, "loss": 0.8426, "step": 1743 }, { "epoch": 0.07352910175601324, "grad_norm": 1.2528570890426636, "learning_rate": 1.2247048903878585e-05, "loss": 0.3317, "step": 1744 }, { "epoch": 0.07357126293821278, "grad_norm": 1.4866958856582642, "learning_rate": 1.2254075323215291e-05, "loss": 0.4994, "step": 1745 }, { "epoch": 0.07361342412041233, "grad_norm": 1.256890058517456, "learning_rate": 1.2261101742551997e-05, "loss": 0.4149, "step": 1746 }, { "epoch": 0.07365558530261189, "grad_norm": 0.8218117952346802, "learning_rate": 1.2268128161888701e-05, "loss": 0.1739, "step": 1747 }, { "epoch": 0.07369774648481144, "grad_norm": 1.253204107284546, "learning_rate": 1.2275154581225407e-05, "loss": 0.3363, "step": 1748 }, { "epoch": 0.07373990766701098, "grad_norm": 1.099855661392212, "learning_rate": 1.2282181000562114e-05, "loss": 0.6071, "step": 1749 }, { "epoch": 0.07378206884921053, "grad_norm": 2.5004160404205322, "learning_rate": 1.228920741989882e-05, "loss": 0.7044, "step": 1750 }, { "epoch": 0.07382423003141009, "grad_norm": 1.0889348983764648, "learning_rate": 1.2296233839235526e-05, "loss": 0.1467, "step": 1751 }, { "epoch": 0.07386639121360963, "grad_norm": 1.4374996423721313, "learning_rate": 1.2303260258572233e-05, "loss": 0.4619, "step": 1752 }, { "epoch": 0.07390855239580918, "grad_norm": 1.8478505611419678, "learning_rate": 1.231028667790894e-05, "loss": 1.2012, "step": 1753 }, { "epoch": 0.07395071357800873, "grad_norm": 1.257399559020996, "learning_rate": 1.2317313097245644e-05, "loss": 0.6353, "step": 1754 }, { "epoch": 0.07399287476020827, "grad_norm": 1.1237415075302124, "learning_rate": 1.232433951658235e-05, "loss": 0.3196, "step": 1755 }, { "epoch": 0.07403503594240783, "grad_norm": 0.9754087328910828, "learning_rate": 1.2331365935919056e-05, "loss": 0.6963, "step": 1756 }, { "epoch": 0.07407719712460738, "grad_norm": 1.4302005767822266, "learning_rate": 1.2338392355255762e-05, "loss": 0.4203, "step": 1757 }, { "epoch": 0.07411935830680692, "grad_norm": 1.7765754461288452, "learning_rate": 1.2345418774592468e-05, "loss": 0.6737, "step": 1758 }, { "epoch": 0.07416151948900647, "grad_norm": 1.4586628675460815, "learning_rate": 1.2352445193929174e-05, "loss": 0.3647, "step": 1759 }, { "epoch": 0.07420368067120602, "grad_norm": 3.344291925430298, "learning_rate": 1.235947161326588e-05, "loss": 0.567, "step": 1760 }, { "epoch": 0.07424584185340558, "grad_norm": 1.4862778186798096, "learning_rate": 1.2366498032602587e-05, "loss": 1.2787, "step": 1761 }, { "epoch": 0.07428800303560512, "grad_norm": 0.8357715606689453, "learning_rate": 1.2373524451939291e-05, "loss": 0.2631, "step": 1762 }, { "epoch": 0.07433016421780467, "grad_norm": 1.2396225929260254, "learning_rate": 1.2380550871275997e-05, "loss": 0.2603, "step": 1763 }, { "epoch": 0.07437232540000421, "grad_norm": 1.2478702068328857, "learning_rate": 1.2387577290612705e-05, "loss": 0.3079, "step": 1764 }, { "epoch": 0.07441448658220376, "grad_norm": 1.6615021228790283, "learning_rate": 1.2394603709949411e-05, "loss": 0.9066, "step": 1765 }, { "epoch": 0.07445664776440332, "grad_norm": 1.6534345149993896, "learning_rate": 1.2401630129286117e-05, "loss": 0.5551, "step": 1766 }, { "epoch": 0.07449880894660287, "grad_norm": 1.1617016792297363, "learning_rate": 1.2408656548622823e-05, "loss": 0.2371, "step": 1767 }, { "epoch": 0.07454097012880241, "grad_norm": 1.4910341501235962, "learning_rate": 1.241568296795953e-05, "loss": 0.7428, "step": 1768 }, { "epoch": 0.07458313131100196, "grad_norm": 1.3487643003463745, "learning_rate": 1.2422709387296234e-05, "loss": 0.2905, "step": 1769 }, { "epoch": 0.0746252924932015, "grad_norm": 3.374401092529297, "learning_rate": 1.242973580663294e-05, "loss": 1.0796, "step": 1770 }, { "epoch": 0.07466745367540106, "grad_norm": 1.403910756111145, "learning_rate": 1.2436762225969646e-05, "loss": 0.2315, "step": 1771 }, { "epoch": 0.07470961485760061, "grad_norm": 1.4569628238677979, "learning_rate": 1.2443788645306352e-05, "loss": 0.4445, "step": 1772 }, { "epoch": 0.07475177603980016, "grad_norm": 2.036498785018921, "learning_rate": 1.2450815064643058e-05, "loss": 0.6718, "step": 1773 }, { "epoch": 0.0747939372219997, "grad_norm": 2.2484166622161865, "learning_rate": 1.2457841483979764e-05, "loss": 0.9249, "step": 1774 }, { "epoch": 0.07483609840419925, "grad_norm": 0.9758821129798889, "learning_rate": 1.2464867903316472e-05, "loss": 0.1907, "step": 1775 }, { "epoch": 0.07487825958639881, "grad_norm": 1.1275018453598022, "learning_rate": 1.2471894322653176e-05, "loss": 0.3231, "step": 1776 }, { "epoch": 0.07492042076859835, "grad_norm": 1.6134731769561768, "learning_rate": 1.2478920741989882e-05, "loss": 0.6351, "step": 1777 }, { "epoch": 0.0749625819507979, "grad_norm": 1.01411771774292, "learning_rate": 1.2485947161326588e-05, "loss": 0.2771, "step": 1778 }, { "epoch": 0.07500474313299745, "grad_norm": 1.2648595571517944, "learning_rate": 1.2492973580663295e-05, "loss": 0.5654, "step": 1779 }, { "epoch": 0.075046904315197, "grad_norm": 1.0891114473342896, "learning_rate": 1.25e-05, "loss": 0.3204, "step": 1780 }, { "epoch": 0.07508906549739655, "grad_norm": 1.2382352352142334, "learning_rate": 1.2507026419336707e-05, "loss": 0.4019, "step": 1781 }, { "epoch": 0.0751312266795961, "grad_norm": 4.84035062789917, "learning_rate": 1.2514052838673413e-05, "loss": 0.9456, "step": 1782 }, { "epoch": 0.07517338786179564, "grad_norm": 1.8589112758636475, "learning_rate": 1.2521079258010119e-05, "loss": 1.1132, "step": 1783 }, { "epoch": 0.07521554904399519, "grad_norm": 1.2867213487625122, "learning_rate": 1.2528105677346825e-05, "loss": 0.2449, "step": 1784 }, { "epoch": 0.07525771022619475, "grad_norm": 0.907678484916687, "learning_rate": 1.253513209668353e-05, "loss": 0.2764, "step": 1785 }, { "epoch": 0.0752998714083943, "grad_norm": 1.332050085067749, "learning_rate": 1.2542158516020235e-05, "loss": 0.4229, "step": 1786 }, { "epoch": 0.07534203259059384, "grad_norm": 2.049990653991699, "learning_rate": 1.2549184935356942e-05, "loss": 0.956, "step": 1787 }, { "epoch": 0.07538419377279339, "grad_norm": 3.6090316772460938, "learning_rate": 1.2556211354693648e-05, "loss": 0.6261, "step": 1788 }, { "epoch": 0.07542635495499293, "grad_norm": 1.3667415380477905, "learning_rate": 1.2563237774030354e-05, "loss": 0.4754, "step": 1789 }, { "epoch": 0.07546851613719249, "grad_norm": 2.8401126861572266, "learning_rate": 1.257026419336706e-05, "loss": 0.5132, "step": 1790 }, { "epoch": 0.07551067731939204, "grad_norm": 0.9466784596443176, "learning_rate": 1.2577290612703766e-05, "loss": 0.2787, "step": 1791 }, { "epoch": 0.07555283850159158, "grad_norm": 0.9167592525482178, "learning_rate": 1.2584317032040474e-05, "loss": 0.2121, "step": 1792 }, { "epoch": 0.07559499968379113, "grad_norm": 1.1485406160354614, "learning_rate": 1.259134345137718e-05, "loss": 0.1864, "step": 1793 }, { "epoch": 0.07563716086599068, "grad_norm": 1.7851614952087402, "learning_rate": 1.2598369870713886e-05, "loss": 1.169, "step": 1794 }, { "epoch": 0.07567932204819024, "grad_norm": 2.1610474586486816, "learning_rate": 1.2605396290050592e-05, "loss": 0.2846, "step": 1795 }, { "epoch": 0.07572148323038978, "grad_norm": 4.7764434814453125, "learning_rate": 1.2612422709387298e-05, "loss": 0.2409, "step": 1796 }, { "epoch": 0.07576364441258933, "grad_norm": 2.354644298553467, "learning_rate": 1.2619449128724004e-05, "loss": 0.9196, "step": 1797 }, { "epoch": 0.07580580559478887, "grad_norm": 0.9537601470947266, "learning_rate": 1.262647554806071e-05, "loss": 0.1905, "step": 1798 }, { "epoch": 0.07584796677698842, "grad_norm": 1.8570191860198975, "learning_rate": 1.2633501967397415e-05, "loss": 1.3474, "step": 1799 }, { "epoch": 0.07589012795918798, "grad_norm": 1.5875866413116455, "learning_rate": 1.264052838673412e-05, "loss": 0.6033, "step": 1800 }, { "epoch": 0.07593228914138753, "grad_norm": 1.7635935544967651, "learning_rate": 1.2647554806070827e-05, "loss": 0.2848, "step": 1801 }, { "epoch": 0.07597445032358707, "grad_norm": 1.517852544784546, "learning_rate": 1.2654581225407533e-05, "loss": 0.1837, "step": 1802 }, { "epoch": 0.07601661150578662, "grad_norm": 1.1565732955932617, "learning_rate": 1.2661607644744239e-05, "loss": 0.1439, "step": 1803 }, { "epoch": 0.07605877268798616, "grad_norm": 1.3748160600662231, "learning_rate": 1.2668634064080945e-05, "loss": 1.1807, "step": 1804 }, { "epoch": 0.07610093387018572, "grad_norm": 1.4957550764083862, "learning_rate": 1.2675660483417651e-05, "loss": 0.8064, "step": 1805 }, { "epoch": 0.07614309505238527, "grad_norm": 0.9634281992912292, "learning_rate": 1.2682686902754357e-05, "loss": 0.3368, "step": 1806 }, { "epoch": 0.07618525623458482, "grad_norm": 1.2870081663131714, "learning_rate": 1.2689713322091062e-05, "loss": 0.2317, "step": 1807 }, { "epoch": 0.07622741741678436, "grad_norm": 1.0458406209945679, "learning_rate": 1.2696739741427768e-05, "loss": 0.3006, "step": 1808 }, { "epoch": 0.07626957859898391, "grad_norm": 1.2285569906234741, "learning_rate": 1.2703766160764474e-05, "loss": 0.3601, "step": 1809 }, { "epoch": 0.07631173978118347, "grad_norm": 1.2361598014831543, "learning_rate": 1.271079258010118e-05, "loss": 0.5282, "step": 1810 }, { "epoch": 0.07635390096338301, "grad_norm": 0.8521959781646729, "learning_rate": 1.2717818999437886e-05, "loss": 0.4195, "step": 1811 }, { "epoch": 0.07639606214558256, "grad_norm": 0.8436295986175537, "learning_rate": 1.2724845418774592e-05, "loss": 0.1872, "step": 1812 }, { "epoch": 0.0764382233277821, "grad_norm": 0.7525056600570679, "learning_rate": 1.2731871838111298e-05, "loss": 0.1982, "step": 1813 }, { "epoch": 0.07648038450998167, "grad_norm": 1.7795183658599854, "learning_rate": 1.2738898257448006e-05, "loss": 0.4151, "step": 1814 }, { "epoch": 0.07652254569218121, "grad_norm": 1.6641020774841309, "learning_rate": 1.2745924676784712e-05, "loss": 1.245, "step": 1815 }, { "epoch": 0.07656470687438076, "grad_norm": 0.7499412894248962, "learning_rate": 1.2752951096121418e-05, "loss": 0.2667, "step": 1816 }, { "epoch": 0.0766068680565803, "grad_norm": 0.98284512758255, "learning_rate": 1.2759977515458124e-05, "loss": 0.4269, "step": 1817 }, { "epoch": 0.07664902923877985, "grad_norm": 1.0170466899871826, "learning_rate": 1.276700393479483e-05, "loss": 0.4438, "step": 1818 }, { "epoch": 0.07669119042097941, "grad_norm": 0.9783209562301636, "learning_rate": 1.2774030354131536e-05, "loss": 0.2891, "step": 1819 }, { "epoch": 0.07673335160317896, "grad_norm": 0.9748091101646423, "learning_rate": 1.2781056773468242e-05, "loss": 0.181, "step": 1820 }, { "epoch": 0.0767755127853785, "grad_norm": 1.197115182876587, "learning_rate": 1.2788083192804947e-05, "loss": 0.2819, "step": 1821 }, { "epoch": 0.07681767396757805, "grad_norm": 1.3627209663391113, "learning_rate": 1.2795109612141653e-05, "loss": 0.1699, "step": 1822 }, { "epoch": 0.0768598351497776, "grad_norm": 1.1070533990859985, "learning_rate": 1.2802136031478359e-05, "loss": 0.2205, "step": 1823 }, { "epoch": 0.07690199633197715, "grad_norm": 0.7767760753631592, "learning_rate": 1.2809162450815065e-05, "loss": 0.1652, "step": 1824 }, { "epoch": 0.0769441575141767, "grad_norm": 1.0658442974090576, "learning_rate": 1.2816188870151771e-05, "loss": 0.6451, "step": 1825 }, { "epoch": 0.07698631869637625, "grad_norm": 1.5734785795211792, "learning_rate": 1.2823215289488477e-05, "loss": 0.3387, "step": 1826 }, { "epoch": 0.07702847987857579, "grad_norm": 0.9739288091659546, "learning_rate": 1.2830241708825183e-05, "loss": 0.4815, "step": 1827 }, { "epoch": 0.07707064106077534, "grad_norm": 1.1785039901733398, "learning_rate": 1.283726812816189e-05, "loss": 0.4026, "step": 1828 }, { "epoch": 0.0771128022429749, "grad_norm": 1.050552248954773, "learning_rate": 1.2844294547498596e-05, "loss": 0.3565, "step": 1829 }, { "epoch": 0.07715496342517444, "grad_norm": 1.8672223091125488, "learning_rate": 1.28513209668353e-05, "loss": 0.4658, "step": 1830 }, { "epoch": 0.07719712460737399, "grad_norm": 1.5881975889205933, "learning_rate": 1.2858347386172006e-05, "loss": 0.195, "step": 1831 }, { "epoch": 0.07723928578957354, "grad_norm": 1.057969331741333, "learning_rate": 1.2865373805508712e-05, "loss": 0.4809, "step": 1832 }, { "epoch": 0.07728144697177308, "grad_norm": 1.1075375080108643, "learning_rate": 1.2872400224845418e-05, "loss": 0.5925, "step": 1833 }, { "epoch": 0.07732360815397264, "grad_norm": 1.820286750793457, "learning_rate": 1.2879426644182124e-05, "loss": 0.9586, "step": 1834 }, { "epoch": 0.07736576933617219, "grad_norm": 1.4162778854370117, "learning_rate": 1.288645306351883e-05, "loss": 0.621, "step": 1835 }, { "epoch": 0.07740793051837173, "grad_norm": 0.7786826491355896, "learning_rate": 1.2893479482855537e-05, "loss": 0.1897, "step": 1836 }, { "epoch": 0.07745009170057128, "grad_norm": 1.493818759918213, "learning_rate": 1.2900505902192244e-05, "loss": 0.3247, "step": 1837 }, { "epoch": 0.07749225288277083, "grad_norm": 1.5463671684265137, "learning_rate": 1.290753232152895e-05, "loss": 0.6478, "step": 1838 }, { "epoch": 0.07753441406497039, "grad_norm": 0.8495244383811951, "learning_rate": 1.2914558740865656e-05, "loss": 0.2481, "step": 1839 }, { "epoch": 0.07757657524716993, "grad_norm": 3.785245656967163, "learning_rate": 1.2921585160202363e-05, "loss": 0.9947, "step": 1840 }, { "epoch": 0.07761873642936948, "grad_norm": 1.0129694938659668, "learning_rate": 1.2928611579539069e-05, "loss": 0.4466, "step": 1841 }, { "epoch": 0.07766089761156902, "grad_norm": 1.467731237411499, "learning_rate": 1.2935637998875775e-05, "loss": 0.314, "step": 1842 }, { "epoch": 0.07770305879376858, "grad_norm": 1.5821683406829834, "learning_rate": 1.294266441821248e-05, "loss": 0.6304, "step": 1843 }, { "epoch": 0.07774521997596813, "grad_norm": 1.0961233377456665, "learning_rate": 1.2949690837549185e-05, "loss": 0.6476, "step": 1844 }, { "epoch": 0.07778738115816768, "grad_norm": 1.2907387018203735, "learning_rate": 1.2956717256885891e-05, "loss": 0.3302, "step": 1845 }, { "epoch": 0.07782954234036722, "grad_norm": 1.0800646543502808, "learning_rate": 1.2963743676222597e-05, "loss": 0.69, "step": 1846 }, { "epoch": 0.07787170352256677, "grad_norm": 1.2510703802108765, "learning_rate": 1.2970770095559303e-05, "loss": 0.6251, "step": 1847 }, { "epoch": 0.07791386470476633, "grad_norm": 1.1476949453353882, "learning_rate": 1.297779651489601e-05, "loss": 0.4108, "step": 1848 }, { "epoch": 0.07795602588696587, "grad_norm": 1.1962419748306274, "learning_rate": 1.2984822934232716e-05, "loss": 0.7805, "step": 1849 }, { "epoch": 0.07799818706916542, "grad_norm": 1.851266860961914, "learning_rate": 1.2991849353569422e-05, "loss": 0.7502, "step": 1850 }, { "epoch": 0.07804034825136497, "grad_norm": 1.352166771888733, "learning_rate": 1.2998875772906128e-05, "loss": 0.4604, "step": 1851 }, { "epoch": 0.07808250943356451, "grad_norm": 1.8456850051879883, "learning_rate": 1.3005902192242834e-05, "loss": 0.5501, "step": 1852 }, { "epoch": 0.07812467061576407, "grad_norm": 1.9343726634979248, "learning_rate": 1.3012928611579538e-05, "loss": 0.5963, "step": 1853 }, { "epoch": 0.07816683179796362, "grad_norm": 2.4756879806518555, "learning_rate": 1.3019955030916244e-05, "loss": 0.4772, "step": 1854 }, { "epoch": 0.07820899298016316, "grad_norm": 1.3872653245925903, "learning_rate": 1.302698145025295e-05, "loss": 0.1977, "step": 1855 }, { "epoch": 0.07825115416236271, "grad_norm": 1.354936122894287, "learning_rate": 1.3034007869589657e-05, "loss": 0.3245, "step": 1856 }, { "epoch": 0.07829331534456226, "grad_norm": 1.3862073421478271, "learning_rate": 1.3041034288926363e-05, "loss": 0.5069, "step": 1857 }, { "epoch": 0.07833547652676182, "grad_norm": 1.1290524005889893, "learning_rate": 1.3048060708263069e-05, "loss": 0.2135, "step": 1858 }, { "epoch": 0.07837763770896136, "grad_norm": 1.0373313426971436, "learning_rate": 1.3055087127599777e-05, "loss": 0.7711, "step": 1859 }, { "epoch": 0.07841979889116091, "grad_norm": 0.9753422737121582, "learning_rate": 1.3062113546936483e-05, "loss": 0.4499, "step": 1860 }, { "epoch": 0.07846196007336045, "grad_norm": 0.8022780418395996, "learning_rate": 1.3069139966273189e-05, "loss": 0.14, "step": 1861 }, { "epoch": 0.07850412125556, "grad_norm": 1.0713173151016235, "learning_rate": 1.3076166385609895e-05, "loss": 0.197, "step": 1862 }, { "epoch": 0.07854628243775956, "grad_norm": 0.9650975465774536, "learning_rate": 1.3083192804946601e-05, "loss": 0.5843, "step": 1863 }, { "epoch": 0.0785884436199591, "grad_norm": 1.145531177520752, "learning_rate": 1.3090219224283307e-05, "loss": 0.781, "step": 1864 }, { "epoch": 0.07863060480215865, "grad_norm": 1.441093921661377, "learning_rate": 1.3097245643620013e-05, "loss": 0.4911, "step": 1865 }, { "epoch": 0.0786727659843582, "grad_norm": 1.0742123126983643, "learning_rate": 1.310427206295672e-05, "loss": 0.9037, "step": 1866 }, { "epoch": 0.07871492716655774, "grad_norm": 1.6010463237762451, "learning_rate": 1.3111298482293424e-05, "loss": 0.182, "step": 1867 }, { "epoch": 0.0787570883487573, "grad_norm": 1.4817732572555542, "learning_rate": 1.311832490163013e-05, "loss": 0.2286, "step": 1868 }, { "epoch": 0.07879924953095685, "grad_norm": 1.6451010704040527, "learning_rate": 1.3125351320966836e-05, "loss": 0.3352, "step": 1869 }, { "epoch": 0.0788414107131564, "grad_norm": 0.9548959136009216, "learning_rate": 1.3132377740303542e-05, "loss": 0.1708, "step": 1870 }, { "epoch": 0.07888357189535594, "grad_norm": 1.2657971382141113, "learning_rate": 1.3139404159640248e-05, "loss": 0.501, "step": 1871 }, { "epoch": 0.07892573307755549, "grad_norm": 1.2933050394058228, "learning_rate": 1.3146430578976954e-05, "loss": 0.2404, "step": 1872 }, { "epoch": 0.07896789425975505, "grad_norm": 3.633690357208252, "learning_rate": 1.315345699831366e-05, "loss": 1.2232, "step": 1873 }, { "epoch": 0.0790100554419546, "grad_norm": 1.065700888633728, "learning_rate": 1.3160483417650366e-05, "loss": 0.4676, "step": 1874 }, { "epoch": 0.07905221662415414, "grad_norm": 0.7873468995094299, "learning_rate": 1.316750983698707e-05, "loss": 0.1014, "step": 1875 }, { "epoch": 0.07909437780635369, "grad_norm": 0.9652047157287598, "learning_rate": 1.3174536256323777e-05, "loss": 0.301, "step": 1876 }, { "epoch": 0.07913653898855325, "grad_norm": 1.0933969020843506, "learning_rate": 1.3181562675660483e-05, "loss": 0.8269, "step": 1877 }, { "epoch": 0.07917870017075279, "grad_norm": 1.2341898679733276, "learning_rate": 1.3188589094997189e-05, "loss": 0.32, "step": 1878 }, { "epoch": 0.07922086135295234, "grad_norm": 1.6306465864181519, "learning_rate": 1.3195615514333895e-05, "loss": 0.3492, "step": 1879 }, { "epoch": 0.07926302253515188, "grad_norm": 1.4216798543930054, "learning_rate": 1.3202641933670601e-05, "loss": 0.2158, "step": 1880 }, { "epoch": 0.07930518371735143, "grad_norm": 0.8093234896659851, "learning_rate": 1.3209668353007307e-05, "loss": 0.2423, "step": 1881 }, { "epoch": 0.07934734489955099, "grad_norm": 1.1519405841827393, "learning_rate": 1.3216694772344015e-05, "loss": 0.3074, "step": 1882 }, { "epoch": 0.07938950608175054, "grad_norm": 0.7440341711044312, "learning_rate": 1.3223721191680721e-05, "loss": 0.1796, "step": 1883 }, { "epoch": 0.07943166726395008, "grad_norm": 0.9862023591995239, "learning_rate": 1.3230747611017427e-05, "loss": 0.6173, "step": 1884 }, { "epoch": 0.07947382844614963, "grad_norm": 1.17363703250885, "learning_rate": 1.3237774030354133e-05, "loss": 0.6711, "step": 1885 }, { "epoch": 0.07951598962834917, "grad_norm": 1.0723882913589478, "learning_rate": 1.324480044969084e-05, "loss": 0.2162, "step": 1886 }, { "epoch": 0.07955815081054873, "grad_norm": 0.6879196166992188, "learning_rate": 1.3251826869027545e-05, "loss": 0.1626, "step": 1887 }, { "epoch": 0.07960031199274828, "grad_norm": 1.210741639137268, "learning_rate": 1.3258853288364251e-05, "loss": 0.2084, "step": 1888 }, { "epoch": 0.07964247317494783, "grad_norm": 0.9421078562736511, "learning_rate": 1.3265879707700956e-05, "loss": 0.3358, "step": 1889 }, { "epoch": 0.07968463435714737, "grad_norm": 1.9219120740890503, "learning_rate": 1.3272906127037662e-05, "loss": 0.9862, "step": 1890 }, { "epoch": 0.07972679553934692, "grad_norm": 2.7171289920806885, "learning_rate": 1.3279932546374368e-05, "loss": 1.5596, "step": 1891 }, { "epoch": 0.07976895672154648, "grad_norm": 1.8948395252227783, "learning_rate": 1.3286958965711074e-05, "loss": 1.0257, "step": 1892 }, { "epoch": 0.07981111790374602, "grad_norm": 0.9696780443191528, "learning_rate": 1.329398538504778e-05, "loss": 0.6782, "step": 1893 }, { "epoch": 0.07985327908594557, "grad_norm": 1.3132100105285645, "learning_rate": 1.3301011804384486e-05, "loss": 0.2868, "step": 1894 }, { "epoch": 0.07989544026814512, "grad_norm": 1.5684733390808105, "learning_rate": 1.3308038223721192e-05, "loss": 0.4037, "step": 1895 }, { "epoch": 0.07993760145034466, "grad_norm": 1.405311942100525, "learning_rate": 1.3315064643057898e-05, "loss": 0.6796, "step": 1896 }, { "epoch": 0.07997976263254422, "grad_norm": 1.3751789331436157, "learning_rate": 1.3322091062394605e-05, "loss": 0.4204, "step": 1897 }, { "epoch": 0.08002192381474377, "grad_norm": 1.4307725429534912, "learning_rate": 1.3329117481731309e-05, "loss": 0.3518, "step": 1898 }, { "epoch": 0.08006408499694331, "grad_norm": 1.714074969291687, "learning_rate": 1.3336143901068015e-05, "loss": 1.1637, "step": 1899 }, { "epoch": 0.08010624617914286, "grad_norm": 1.5303789377212524, "learning_rate": 1.3343170320404721e-05, "loss": 0.2992, "step": 1900 }, { "epoch": 0.0801484073613424, "grad_norm": 0.8955725431442261, "learning_rate": 1.3350196739741427e-05, "loss": 0.2911, "step": 1901 }, { "epoch": 0.08019056854354197, "grad_norm": 1.1362520456314087, "learning_rate": 1.3357223159078133e-05, "loss": 0.7576, "step": 1902 }, { "epoch": 0.08023272972574151, "grad_norm": 1.068878412246704, "learning_rate": 1.336424957841484e-05, "loss": 0.3807, "step": 1903 }, { "epoch": 0.08027489090794106, "grad_norm": 1.1257987022399902, "learning_rate": 1.3371275997751547e-05, "loss": 0.1427, "step": 1904 }, { "epoch": 0.0803170520901406, "grad_norm": 2.409278392791748, "learning_rate": 1.3378302417088253e-05, "loss": 0.5836, "step": 1905 }, { "epoch": 0.08035921327234016, "grad_norm": 1.3176885843276978, "learning_rate": 1.338532883642496e-05, "loss": 0.6113, "step": 1906 }, { "epoch": 0.08040137445453971, "grad_norm": 1.1374763250350952, "learning_rate": 1.3392355255761665e-05, "loss": 0.4888, "step": 1907 }, { "epoch": 0.08044353563673926, "grad_norm": 1.1390682458877563, "learning_rate": 1.3399381675098372e-05, "loss": 0.4337, "step": 1908 }, { "epoch": 0.0804856968189388, "grad_norm": 0.8255512714385986, "learning_rate": 1.3406408094435078e-05, "loss": 0.3768, "step": 1909 }, { "epoch": 0.08052785800113835, "grad_norm": 3.073127269744873, "learning_rate": 1.3413434513771784e-05, "loss": 0.8225, "step": 1910 }, { "epoch": 0.08057001918333791, "grad_norm": 1.3266936540603638, "learning_rate": 1.342046093310849e-05, "loss": 0.3347, "step": 1911 }, { "epoch": 0.08061218036553745, "grad_norm": 0.7605717778205872, "learning_rate": 1.3427487352445194e-05, "loss": 0.1845, "step": 1912 }, { "epoch": 0.080654341547737, "grad_norm": 1.674100637435913, "learning_rate": 1.34345137717819e-05, "loss": 0.5501, "step": 1913 }, { "epoch": 0.08069650272993655, "grad_norm": 1.1325522661209106, "learning_rate": 1.3441540191118606e-05, "loss": 0.3005, "step": 1914 }, { "epoch": 0.08073866391213609, "grad_norm": 1.279091238975525, "learning_rate": 1.3448566610455312e-05, "loss": 0.4475, "step": 1915 }, { "epoch": 0.08078082509433565, "grad_norm": 1.1532442569732666, "learning_rate": 1.3455593029792019e-05, "loss": 0.6514, "step": 1916 }, { "epoch": 0.0808229862765352, "grad_norm": 0.918988049030304, "learning_rate": 1.3462619449128725e-05, "loss": 0.2931, "step": 1917 }, { "epoch": 0.08086514745873474, "grad_norm": 0.8649528622627258, "learning_rate": 1.346964586846543e-05, "loss": 0.2149, "step": 1918 }, { "epoch": 0.08090730864093429, "grad_norm": 1.1567426919937134, "learning_rate": 1.3476672287802137e-05, "loss": 0.3861, "step": 1919 }, { "epoch": 0.08094946982313383, "grad_norm": 1.1616439819335938, "learning_rate": 1.3483698707138843e-05, "loss": 0.4012, "step": 1920 }, { "epoch": 0.0809916310053334, "grad_norm": 2.849546432495117, "learning_rate": 1.3490725126475547e-05, "loss": 0.7563, "step": 1921 }, { "epoch": 0.08103379218753294, "grad_norm": 1.6661607027053833, "learning_rate": 1.3497751545812253e-05, "loss": 0.2393, "step": 1922 }, { "epoch": 0.08107595336973249, "grad_norm": 1.1386582851409912, "learning_rate": 1.350477796514896e-05, "loss": 0.1918, "step": 1923 }, { "epoch": 0.08111811455193203, "grad_norm": 1.3499857187271118, "learning_rate": 1.3511804384485666e-05, "loss": 0.7123, "step": 1924 }, { "epoch": 0.08116027573413158, "grad_norm": 1.4565643072128296, "learning_rate": 1.3518830803822372e-05, "loss": 0.2948, "step": 1925 }, { "epoch": 0.08120243691633114, "grad_norm": 1.424516201019287, "learning_rate": 1.3525857223159078e-05, "loss": 0.3824, "step": 1926 }, { "epoch": 0.08124459809853068, "grad_norm": 1.4528895616531372, "learning_rate": 1.3532883642495786e-05, "loss": 0.5688, "step": 1927 }, { "epoch": 0.08128675928073023, "grad_norm": 0.8568162322044373, "learning_rate": 1.3539910061832492e-05, "loss": 0.2567, "step": 1928 }, { "epoch": 0.08132892046292978, "grad_norm": 1.2253319025039673, "learning_rate": 1.3546936481169198e-05, "loss": 0.6001, "step": 1929 }, { "epoch": 0.08137108164512932, "grad_norm": 0.9355097413063049, "learning_rate": 1.3553962900505904e-05, "loss": 0.171, "step": 1930 }, { "epoch": 0.08141324282732888, "grad_norm": 2.35823130607605, "learning_rate": 1.356098931984261e-05, "loss": 0.2704, "step": 1931 }, { "epoch": 0.08145540400952843, "grad_norm": 0.8829721808433533, "learning_rate": 1.3568015739179316e-05, "loss": 0.4022, "step": 1932 }, { "epoch": 0.08149756519172797, "grad_norm": 0.9730018377304077, "learning_rate": 1.3575042158516022e-05, "loss": 0.1611, "step": 1933 }, { "epoch": 0.08153972637392752, "grad_norm": 1.440317988395691, "learning_rate": 1.3582068577852728e-05, "loss": 0.5839, "step": 1934 }, { "epoch": 0.08158188755612708, "grad_norm": 2.3842039108276367, "learning_rate": 1.3589094997189433e-05, "loss": 0.4601, "step": 1935 }, { "epoch": 0.08162404873832663, "grad_norm": 0.8921994566917419, "learning_rate": 1.3596121416526139e-05, "loss": 0.2169, "step": 1936 }, { "epoch": 0.08166620992052617, "grad_norm": 0.5762099623680115, "learning_rate": 1.3603147835862845e-05, "loss": 0.1435, "step": 1937 }, { "epoch": 0.08170837110272572, "grad_norm": 1.2286874055862427, "learning_rate": 1.361017425519955e-05, "loss": 0.3657, "step": 1938 }, { "epoch": 0.08175053228492526, "grad_norm": 0.8092181086540222, "learning_rate": 1.3617200674536257e-05, "loss": 0.315, "step": 1939 }, { "epoch": 0.08179269346712482, "grad_norm": 1.4818620681762695, "learning_rate": 1.3624227093872963e-05, "loss": 0.3683, "step": 1940 }, { "epoch": 0.08183485464932437, "grad_norm": 2.128532648086548, "learning_rate": 1.3631253513209669e-05, "loss": 0.7185, "step": 1941 }, { "epoch": 0.08187701583152392, "grad_norm": 1.3293224573135376, "learning_rate": 1.3638279932546375e-05, "loss": 0.6396, "step": 1942 }, { "epoch": 0.08191917701372346, "grad_norm": 1.151833415031433, "learning_rate": 1.364530635188308e-05, "loss": 0.2578, "step": 1943 }, { "epoch": 0.08196133819592301, "grad_norm": 1.1179039478302002, "learning_rate": 1.3652332771219786e-05, "loss": 0.7191, "step": 1944 }, { "epoch": 0.08200349937812257, "grad_norm": 5.392355442047119, "learning_rate": 1.3659359190556492e-05, "loss": 0.7367, "step": 1945 }, { "epoch": 0.08204566056032211, "grad_norm": 1.1811708211898804, "learning_rate": 1.3666385609893198e-05, "loss": 0.3459, "step": 1946 }, { "epoch": 0.08208782174252166, "grad_norm": 1.396986961364746, "learning_rate": 1.3673412029229904e-05, "loss": 0.6727, "step": 1947 }, { "epoch": 0.0821299829247212, "grad_norm": 2.6647586822509766, "learning_rate": 1.368043844856661e-05, "loss": 0.8037, "step": 1948 }, { "epoch": 0.08217214410692075, "grad_norm": 1.8513829708099365, "learning_rate": 1.3687464867903318e-05, "loss": 0.1883, "step": 1949 }, { "epoch": 0.08221430528912031, "grad_norm": 0.8601763844490051, "learning_rate": 1.3694491287240024e-05, "loss": 0.2822, "step": 1950 }, { "epoch": 0.08225646647131986, "grad_norm": 0.9267552495002747, "learning_rate": 1.370151770657673e-05, "loss": 0.3068, "step": 1951 }, { "epoch": 0.0822986276535194, "grad_norm": 1.0662130117416382, "learning_rate": 1.3708544125913436e-05, "loss": 0.331, "step": 1952 }, { "epoch": 0.08234078883571895, "grad_norm": 1.714898943901062, "learning_rate": 1.3715570545250142e-05, "loss": 0.3398, "step": 1953 }, { "epoch": 0.0823829500179185, "grad_norm": 0.8465660810470581, "learning_rate": 1.3722596964586848e-05, "loss": 0.5435, "step": 1954 }, { "epoch": 0.08242511120011806, "grad_norm": 1.2211757898330688, "learning_rate": 1.3729623383923554e-05, "loss": 0.7267, "step": 1955 }, { "epoch": 0.0824672723823176, "grad_norm": 1.7721339464187622, "learning_rate": 1.373664980326026e-05, "loss": 1.1061, "step": 1956 }, { "epoch": 0.08250943356451715, "grad_norm": 1.6217381954193115, "learning_rate": 1.3743676222596965e-05, "loss": 0.2533, "step": 1957 }, { "epoch": 0.0825515947467167, "grad_norm": 2.1737353801727295, "learning_rate": 1.3750702641933671e-05, "loss": 0.5781, "step": 1958 }, { "epoch": 0.08259375592891624, "grad_norm": 1.2643487453460693, "learning_rate": 1.3757729061270377e-05, "loss": 0.2389, "step": 1959 }, { "epoch": 0.0826359171111158, "grad_norm": 1.239367127418518, "learning_rate": 1.3764755480607083e-05, "loss": 0.6089, "step": 1960 }, { "epoch": 0.08267807829331535, "grad_norm": 1.2133945226669312, "learning_rate": 1.377178189994379e-05, "loss": 0.7702, "step": 1961 }, { "epoch": 0.08272023947551489, "grad_norm": 1.217620611190796, "learning_rate": 1.3778808319280495e-05, "loss": 0.6358, "step": 1962 }, { "epoch": 0.08276240065771444, "grad_norm": 0.971266508102417, "learning_rate": 1.3785834738617201e-05, "loss": 0.3545, "step": 1963 }, { "epoch": 0.08280456183991398, "grad_norm": 1.0625343322753906, "learning_rate": 1.3792861157953907e-05, "loss": 0.5674, "step": 1964 }, { "epoch": 0.08284672302211354, "grad_norm": 1.165047526359558, "learning_rate": 1.3799887577290614e-05, "loss": 0.5059, "step": 1965 }, { "epoch": 0.08288888420431309, "grad_norm": 0.9820457100868225, "learning_rate": 1.3806913996627318e-05, "loss": 0.6291, "step": 1966 }, { "epoch": 0.08293104538651264, "grad_norm": 2.622889280319214, "learning_rate": 1.3813940415964024e-05, "loss": 0.8754, "step": 1967 }, { "epoch": 0.08297320656871218, "grad_norm": 1.2098208665847778, "learning_rate": 1.382096683530073e-05, "loss": 1.1575, "step": 1968 }, { "epoch": 0.08301536775091174, "grad_norm": 1.3175551891326904, "learning_rate": 1.3827993254637436e-05, "loss": 0.2315, "step": 1969 }, { "epoch": 0.08305752893311129, "grad_norm": 1.1317572593688965, "learning_rate": 1.3835019673974142e-05, "loss": 0.4333, "step": 1970 }, { "epoch": 0.08309969011531083, "grad_norm": 0.8845009803771973, "learning_rate": 1.3842046093310848e-05, "loss": 0.2666, "step": 1971 }, { "epoch": 0.08314185129751038, "grad_norm": 1.4397294521331787, "learning_rate": 1.3849072512647556e-05, "loss": 0.4027, "step": 1972 }, { "epoch": 0.08318401247970993, "grad_norm": 1.1258960962295532, "learning_rate": 1.3856098931984262e-05, "loss": 0.1899, "step": 1973 }, { "epoch": 0.08322617366190949, "grad_norm": 1.0973466634750366, "learning_rate": 1.3863125351320968e-05, "loss": 0.5059, "step": 1974 }, { "epoch": 0.08326833484410903, "grad_norm": 1.019799828529358, "learning_rate": 1.3870151770657674e-05, "loss": 0.3712, "step": 1975 }, { "epoch": 0.08331049602630858, "grad_norm": 1.2393711805343628, "learning_rate": 1.387717818999438e-05, "loss": 0.7091, "step": 1976 }, { "epoch": 0.08335265720850812, "grad_norm": 1.4884675741195679, "learning_rate": 1.3884204609331087e-05, "loss": 0.4972, "step": 1977 }, { "epoch": 0.08339481839070767, "grad_norm": 1.5196855068206787, "learning_rate": 1.3891231028667793e-05, "loss": 0.3769, "step": 1978 }, { "epoch": 0.08343697957290723, "grad_norm": 3.5579335689544678, "learning_rate": 1.3898257448004499e-05, "loss": 0.953, "step": 1979 }, { "epoch": 0.08347914075510678, "grad_norm": 1.5936665534973145, "learning_rate": 1.3905283867341203e-05, "loss": 0.397, "step": 1980 }, { "epoch": 0.08352130193730632, "grad_norm": 0.9408168792724609, "learning_rate": 1.391231028667791e-05, "loss": 0.42, "step": 1981 }, { "epoch": 0.08356346311950587, "grad_norm": 1.1253178119659424, "learning_rate": 1.3919336706014615e-05, "loss": 0.2325, "step": 1982 }, { "epoch": 0.08360562430170541, "grad_norm": 1.6975622177124023, "learning_rate": 1.3926363125351321e-05, "loss": 0.4084, "step": 1983 }, { "epoch": 0.08364778548390497, "grad_norm": 0.9082750082015991, "learning_rate": 1.3933389544688028e-05, "loss": 0.2325, "step": 1984 }, { "epoch": 0.08368994666610452, "grad_norm": 1.063167929649353, "learning_rate": 1.3940415964024734e-05, "loss": 0.3389, "step": 1985 }, { "epoch": 0.08373210784830407, "grad_norm": 1.8716634511947632, "learning_rate": 1.394744238336144e-05, "loss": 0.4548, "step": 1986 }, { "epoch": 0.08377426903050361, "grad_norm": 1.0164196491241455, "learning_rate": 1.3954468802698146e-05, "loss": 0.216, "step": 1987 }, { "epoch": 0.08381643021270316, "grad_norm": 1.9995391368865967, "learning_rate": 1.396149522203485e-05, "loss": 0.209, "step": 1988 }, { "epoch": 0.08385859139490272, "grad_norm": 1.4720618724822998, "learning_rate": 1.3968521641371556e-05, "loss": 0.742, "step": 1989 }, { "epoch": 0.08390075257710226, "grad_norm": 1.3696796894073486, "learning_rate": 1.3975548060708262e-05, "loss": 0.305, "step": 1990 }, { "epoch": 0.08394291375930181, "grad_norm": 1.117018222808838, "learning_rate": 1.3982574480044968e-05, "loss": 0.4575, "step": 1991 }, { "epoch": 0.08398507494150136, "grad_norm": 0.9774719476699829, "learning_rate": 1.3989600899381675e-05, "loss": 0.426, "step": 1992 }, { "epoch": 0.0840272361237009, "grad_norm": 1.2450343370437622, "learning_rate": 1.399662731871838e-05, "loss": 0.5036, "step": 1993 }, { "epoch": 0.08406939730590046, "grad_norm": 1.3892149925231934, "learning_rate": 1.4003653738055088e-05, "loss": 0.2492, "step": 1994 }, { "epoch": 0.08411155848810001, "grad_norm": 1.2795698642730713, "learning_rate": 1.4010680157391795e-05, "loss": 0.7161, "step": 1995 }, { "epoch": 0.08415371967029955, "grad_norm": 1.977491855621338, "learning_rate": 1.40177065767285e-05, "loss": 0.6148, "step": 1996 }, { "epoch": 0.0841958808524991, "grad_norm": 1.348466157913208, "learning_rate": 1.4024732996065207e-05, "loss": 0.3143, "step": 1997 }, { "epoch": 0.08423804203469866, "grad_norm": 1.016780138015747, "learning_rate": 1.4031759415401913e-05, "loss": 0.675, "step": 1998 }, { "epoch": 0.0842802032168982, "grad_norm": 3.566441297531128, "learning_rate": 1.4038785834738619e-05, "loss": 0.8212, "step": 1999 }, { "epoch": 0.08432236439909775, "grad_norm": 0.8791245818138123, "learning_rate": 1.4045812254075325e-05, "loss": 0.1957, "step": 2000 }, { "epoch": 0.0843645255812973, "grad_norm": 1.3400262594223022, "learning_rate": 1.4052838673412031e-05, "loss": 0.7243, "step": 2001 }, { "epoch": 0.08440668676349684, "grad_norm": 0.7839016318321228, "learning_rate": 1.4059865092748735e-05, "loss": 0.1518, "step": 2002 }, { "epoch": 0.0844488479456964, "grad_norm": 1.081099033355713, "learning_rate": 1.4066891512085442e-05, "loss": 0.7855, "step": 2003 }, { "epoch": 0.08449100912789595, "grad_norm": 0.7848309278488159, "learning_rate": 1.4073917931422148e-05, "loss": 0.1626, "step": 2004 }, { "epoch": 0.0845331703100955, "grad_norm": 1.3335773944854736, "learning_rate": 1.4080944350758854e-05, "loss": 0.226, "step": 2005 }, { "epoch": 0.08457533149229504, "grad_norm": 0.7125983238220215, "learning_rate": 1.408797077009556e-05, "loss": 0.116, "step": 2006 }, { "epoch": 0.08461749267449459, "grad_norm": 1.0149753093719482, "learning_rate": 1.4094997189432266e-05, "loss": 0.1346, "step": 2007 }, { "epoch": 0.08465965385669415, "grad_norm": 1.0148731470108032, "learning_rate": 1.4102023608768972e-05, "loss": 0.1791, "step": 2008 }, { "epoch": 0.0847018150388937, "grad_norm": 3.6284923553466797, "learning_rate": 1.4109050028105678e-05, "loss": 0.8272, "step": 2009 }, { "epoch": 0.08474397622109324, "grad_norm": 1.3397746086120605, "learning_rate": 1.4116076447442384e-05, "loss": 0.2697, "step": 2010 }, { "epoch": 0.08478613740329279, "grad_norm": 1.798467755317688, "learning_rate": 1.4123102866779089e-05, "loss": 0.3055, "step": 2011 }, { "epoch": 0.08482829858549233, "grad_norm": 1.2528377771377563, "learning_rate": 1.4130129286115795e-05, "loss": 0.2556, "step": 2012 }, { "epoch": 0.08487045976769189, "grad_norm": 0.8432155847549438, "learning_rate": 1.41371557054525e-05, "loss": 0.2222, "step": 2013 }, { "epoch": 0.08491262094989144, "grad_norm": 1.4720441102981567, "learning_rate": 1.4144182124789207e-05, "loss": 0.3417, "step": 2014 }, { "epoch": 0.08495478213209098, "grad_norm": 1.1776291131973267, "learning_rate": 1.4151208544125913e-05, "loss": 0.2879, "step": 2015 }, { "epoch": 0.08499694331429053, "grad_norm": 1.3053536415100098, "learning_rate": 1.4158234963462619e-05, "loss": 0.6835, "step": 2016 }, { "epoch": 0.08503910449649008, "grad_norm": 1.1134089231491089, "learning_rate": 1.4165261382799327e-05, "loss": 0.8206, "step": 2017 }, { "epoch": 0.08508126567868964, "grad_norm": 1.279104232788086, "learning_rate": 1.4172287802136033e-05, "loss": 0.3515, "step": 2018 }, { "epoch": 0.08512342686088918, "grad_norm": 1.6846407651901245, "learning_rate": 1.4179314221472739e-05, "loss": 0.5618, "step": 2019 }, { "epoch": 0.08516558804308873, "grad_norm": 0.7392069101333618, "learning_rate": 1.4186340640809445e-05, "loss": 0.1368, "step": 2020 }, { "epoch": 0.08520774922528827, "grad_norm": 1.1871479749679565, "learning_rate": 1.4193367060146151e-05, "loss": 0.6361, "step": 2021 }, { "epoch": 0.08524991040748782, "grad_norm": 3.0875747203826904, "learning_rate": 1.4200393479482857e-05, "loss": 0.7507, "step": 2022 }, { "epoch": 0.08529207158968738, "grad_norm": 1.520727276802063, "learning_rate": 1.4207419898819563e-05, "loss": 0.5003, "step": 2023 }, { "epoch": 0.08533423277188693, "grad_norm": 0.8151185512542725, "learning_rate": 1.421444631815627e-05, "loss": 0.1815, "step": 2024 }, { "epoch": 0.08537639395408647, "grad_norm": 1.0503007173538208, "learning_rate": 1.4221472737492974e-05, "loss": 0.3218, "step": 2025 }, { "epoch": 0.08541855513628602, "grad_norm": 0.8176600337028503, "learning_rate": 1.422849915682968e-05, "loss": 0.2895, "step": 2026 }, { "epoch": 0.08546071631848558, "grad_norm": 1.3479872941970825, "learning_rate": 1.4235525576166386e-05, "loss": 0.3282, "step": 2027 }, { "epoch": 0.08550287750068512, "grad_norm": 0.8554492592811584, "learning_rate": 1.4242551995503092e-05, "loss": 0.304, "step": 2028 }, { "epoch": 0.08554503868288467, "grad_norm": 0.9264929294586182, "learning_rate": 1.4249578414839798e-05, "loss": 0.3557, "step": 2029 }, { "epoch": 0.08558719986508422, "grad_norm": 1.2667937278747559, "learning_rate": 1.4256604834176504e-05, "loss": 0.7573, "step": 2030 }, { "epoch": 0.08562936104728376, "grad_norm": 1.020324945449829, "learning_rate": 1.426363125351321e-05, "loss": 0.6049, "step": 2031 }, { "epoch": 0.08567152222948332, "grad_norm": 0.9540380835533142, "learning_rate": 1.4270657672849916e-05, "loss": 0.3513, "step": 2032 }, { "epoch": 0.08571368341168287, "grad_norm": 1.2212456464767456, "learning_rate": 1.4277684092186623e-05, "loss": 0.4992, "step": 2033 }, { "epoch": 0.08575584459388241, "grad_norm": 0.7057341933250427, "learning_rate": 1.4284710511523327e-05, "loss": 0.1379, "step": 2034 }, { "epoch": 0.08579800577608196, "grad_norm": 1.1191989183425903, "learning_rate": 1.4291736930860033e-05, "loss": 0.1764, "step": 2035 }, { "epoch": 0.0858401669582815, "grad_norm": 1.4517184495925903, "learning_rate": 1.4298763350196739e-05, "loss": 0.264, "step": 2036 }, { "epoch": 0.08588232814048107, "grad_norm": 0.9868198037147522, "learning_rate": 1.4305789769533445e-05, "loss": 0.1701, "step": 2037 }, { "epoch": 0.08592448932268061, "grad_norm": 1.1268380880355835, "learning_rate": 1.4312816188870151e-05, "loss": 0.1429, "step": 2038 }, { "epoch": 0.08596665050488016, "grad_norm": 3.208641290664673, "learning_rate": 1.4319842608206859e-05, "loss": 0.6691, "step": 2039 }, { "epoch": 0.0860088116870797, "grad_norm": 0.9441951513290405, "learning_rate": 1.4326869027543565e-05, "loss": 0.5945, "step": 2040 }, { "epoch": 0.08605097286927925, "grad_norm": 1.221930742263794, "learning_rate": 1.4333895446880271e-05, "loss": 0.1663, "step": 2041 }, { "epoch": 0.08609313405147881, "grad_norm": 0.8844040036201477, "learning_rate": 1.4340921866216977e-05, "loss": 0.3095, "step": 2042 }, { "epoch": 0.08613529523367836, "grad_norm": 1.1666145324707031, "learning_rate": 1.4347948285553683e-05, "loss": 0.1978, "step": 2043 }, { "epoch": 0.0861774564158779, "grad_norm": 1.4034875631332397, "learning_rate": 1.435497470489039e-05, "loss": 0.5924, "step": 2044 }, { "epoch": 0.08621961759807745, "grad_norm": 0.9659296274185181, "learning_rate": 1.4362001124227096e-05, "loss": 0.5755, "step": 2045 }, { "epoch": 0.086261778780277, "grad_norm": 0.9499979615211487, "learning_rate": 1.4369027543563802e-05, "loss": 0.5331, "step": 2046 }, { "epoch": 0.08630393996247655, "grad_norm": 1.1836076974868774, "learning_rate": 1.4376053962900508e-05, "loss": 0.6059, "step": 2047 }, { "epoch": 0.0863461011446761, "grad_norm": 3.703467607498169, "learning_rate": 1.4383080382237212e-05, "loss": 1.393, "step": 2048 }, { "epoch": 0.08638826232687564, "grad_norm": 2.6587419509887695, "learning_rate": 1.4390106801573918e-05, "loss": 0.587, "step": 2049 }, { "epoch": 0.08643042350907519, "grad_norm": 0.7906568050384521, "learning_rate": 1.4397133220910624e-05, "loss": 0.1719, "step": 2050 }, { "epoch": 0.08647258469127474, "grad_norm": 0.7207953333854675, "learning_rate": 1.440415964024733e-05, "loss": 0.2584, "step": 2051 }, { "epoch": 0.0865147458734743, "grad_norm": 3.692176342010498, "learning_rate": 1.4411186059584037e-05, "loss": 0.4174, "step": 2052 }, { "epoch": 0.08655690705567384, "grad_norm": 0.6608405709266663, "learning_rate": 1.4418212478920743e-05, "loss": 0.1383, "step": 2053 }, { "epoch": 0.08659906823787339, "grad_norm": 1.1196304559707642, "learning_rate": 1.4425238898257449e-05, "loss": 0.1417, "step": 2054 }, { "epoch": 0.08664122942007293, "grad_norm": 1.1012355089187622, "learning_rate": 1.4432265317594155e-05, "loss": 0.342, "step": 2055 }, { "epoch": 0.08668339060227248, "grad_norm": 1.149761438369751, "learning_rate": 1.443929173693086e-05, "loss": 0.344, "step": 2056 }, { "epoch": 0.08672555178447204, "grad_norm": 0.7606672048568726, "learning_rate": 1.4446318156267565e-05, "loss": 0.3654, "step": 2057 }, { "epoch": 0.08676771296667159, "grad_norm": 0.546147882938385, "learning_rate": 1.4453344575604271e-05, "loss": 0.1349, "step": 2058 }, { "epoch": 0.08680987414887113, "grad_norm": 1.2965216636657715, "learning_rate": 1.4460370994940977e-05, "loss": 0.5767, "step": 2059 }, { "epoch": 0.08685203533107068, "grad_norm": 2.0065317153930664, "learning_rate": 1.4467397414277684e-05, "loss": 0.5519, "step": 2060 }, { "epoch": 0.08689419651327024, "grad_norm": 1.642323613166809, "learning_rate": 1.447442383361439e-05, "loss": 1.2707, "step": 2061 }, { "epoch": 0.08693635769546978, "grad_norm": 1.1296923160552979, "learning_rate": 1.4481450252951097e-05, "loss": 0.3958, "step": 2062 }, { "epoch": 0.08697851887766933, "grad_norm": 1.593084692955017, "learning_rate": 1.4488476672287804e-05, "loss": 0.5548, "step": 2063 }, { "epoch": 0.08702068005986888, "grad_norm": 1.6420636177062988, "learning_rate": 1.449550309162451e-05, "loss": 0.6844, "step": 2064 }, { "epoch": 0.08706284124206842, "grad_norm": 1.0238844156265259, "learning_rate": 1.4502529510961216e-05, "loss": 0.9585, "step": 2065 }, { "epoch": 0.08710500242426798, "grad_norm": 0.7426645755767822, "learning_rate": 1.4509555930297922e-05, "loss": 0.2364, "step": 2066 }, { "epoch": 0.08714716360646753, "grad_norm": 0.9241233468055725, "learning_rate": 1.4516582349634628e-05, "loss": 0.256, "step": 2067 }, { "epoch": 0.08718932478866707, "grad_norm": 1.0876562595367432, "learning_rate": 1.4523608768971334e-05, "loss": 0.2807, "step": 2068 }, { "epoch": 0.08723148597086662, "grad_norm": 1.2860088348388672, "learning_rate": 1.453063518830804e-05, "loss": 0.526, "step": 2069 }, { "epoch": 0.08727364715306617, "grad_norm": 0.9647375345230103, "learning_rate": 1.4537661607644744e-05, "loss": 0.3457, "step": 2070 }, { "epoch": 0.08731580833526573, "grad_norm": 1.4629980325698853, "learning_rate": 1.454468802698145e-05, "loss": 1.2209, "step": 2071 }, { "epoch": 0.08735796951746527, "grad_norm": 1.0473922491073608, "learning_rate": 1.4551714446318157e-05, "loss": 0.2556, "step": 2072 }, { "epoch": 0.08740013069966482, "grad_norm": 0.8081511855125427, "learning_rate": 1.4558740865654863e-05, "loss": 0.2061, "step": 2073 }, { "epoch": 0.08744229188186436, "grad_norm": 1.6276088953018188, "learning_rate": 1.4565767284991569e-05, "loss": 0.4721, "step": 2074 }, { "epoch": 0.08748445306406391, "grad_norm": 1.7170393466949463, "learning_rate": 1.4572793704328275e-05, "loss": 0.5444, "step": 2075 }, { "epoch": 0.08752661424626347, "grad_norm": 1.195360779762268, "learning_rate": 1.4579820123664981e-05, "loss": 0.3917, "step": 2076 }, { "epoch": 0.08756877542846302, "grad_norm": 1.7091344594955444, "learning_rate": 1.4586846543001687e-05, "loss": 1.4089, "step": 2077 }, { "epoch": 0.08761093661066256, "grad_norm": 0.615898847579956, "learning_rate": 1.4593872962338393e-05, "loss": 0.152, "step": 2078 }, { "epoch": 0.08765309779286211, "grad_norm": 0.9881889820098877, "learning_rate": 1.4600899381675098e-05, "loss": 0.5342, "step": 2079 }, { "epoch": 0.08769525897506165, "grad_norm": 1.2096545696258545, "learning_rate": 1.4607925801011804e-05, "loss": 0.6054, "step": 2080 }, { "epoch": 0.08773742015726121, "grad_norm": 0.863950252532959, "learning_rate": 1.461495222034851e-05, "loss": 0.1496, "step": 2081 }, { "epoch": 0.08777958133946076, "grad_norm": 0.985031008720398, "learning_rate": 1.4621978639685216e-05, "loss": 0.1898, "step": 2082 }, { "epoch": 0.0878217425216603, "grad_norm": 1.2726352214813232, "learning_rate": 1.4629005059021922e-05, "loss": 0.2955, "step": 2083 }, { "epoch": 0.08786390370385985, "grad_norm": 2.950773000717163, "learning_rate": 1.4636031478358628e-05, "loss": 1.0756, "step": 2084 }, { "epoch": 0.0879060648860594, "grad_norm": 0.8121505975723267, "learning_rate": 1.4643057897695336e-05, "loss": 0.1859, "step": 2085 }, { "epoch": 0.08794822606825896, "grad_norm": 1.3069409132003784, "learning_rate": 1.4650084317032042e-05, "loss": 0.616, "step": 2086 }, { "epoch": 0.0879903872504585, "grad_norm": 1.5061531066894531, "learning_rate": 1.4657110736368748e-05, "loss": 0.2911, "step": 2087 }, { "epoch": 0.08803254843265805, "grad_norm": 0.8762162327766418, "learning_rate": 1.4664137155705454e-05, "loss": 0.3896, "step": 2088 }, { "epoch": 0.0880747096148576, "grad_norm": 0.6562484502792358, "learning_rate": 1.467116357504216e-05, "loss": 0.1617, "step": 2089 }, { "epoch": 0.08811687079705716, "grad_norm": 0.7882863283157349, "learning_rate": 1.4678189994378866e-05, "loss": 0.213, "step": 2090 }, { "epoch": 0.0881590319792567, "grad_norm": 1.2291067838668823, "learning_rate": 1.4685216413715572e-05, "loss": 1.2186, "step": 2091 }, { "epoch": 0.08820119316145625, "grad_norm": 0.9324796199798584, "learning_rate": 1.4692242833052278e-05, "loss": 0.5814, "step": 2092 }, { "epoch": 0.0882433543436558, "grad_norm": 0.8012099862098694, "learning_rate": 1.4699269252388983e-05, "loss": 0.4747, "step": 2093 }, { "epoch": 0.08828551552585534, "grad_norm": 1.050950288772583, "learning_rate": 1.4706295671725689e-05, "loss": 0.5599, "step": 2094 }, { "epoch": 0.0883276767080549, "grad_norm": 0.9756861329078674, "learning_rate": 1.4713322091062395e-05, "loss": 0.3121, "step": 2095 }, { "epoch": 0.08836983789025445, "grad_norm": 0.593105673789978, "learning_rate": 1.4720348510399101e-05, "loss": 0.1517, "step": 2096 }, { "epoch": 0.08841199907245399, "grad_norm": 1.795337200164795, "learning_rate": 1.4727374929735807e-05, "loss": 1.1874, "step": 2097 }, { "epoch": 0.08845416025465354, "grad_norm": 0.7988589406013489, "learning_rate": 1.4734401349072513e-05, "loss": 0.2199, "step": 2098 }, { "epoch": 0.08849632143685308, "grad_norm": 1.1378766298294067, "learning_rate": 1.474142776840922e-05, "loss": 0.2851, "step": 2099 }, { "epoch": 0.08853848261905264, "grad_norm": 0.6205458641052246, "learning_rate": 1.4748454187745925e-05, "loss": 0.2705, "step": 2100 }, { "epoch": 0.08858064380125219, "grad_norm": 1.501665711402893, "learning_rate": 1.4755480607082632e-05, "loss": 1.2584, "step": 2101 }, { "epoch": 0.08862280498345174, "grad_norm": 0.7726740837097168, "learning_rate": 1.4762507026419336e-05, "loss": 0.2199, "step": 2102 }, { "epoch": 0.08866496616565128, "grad_norm": 1.12339448928833, "learning_rate": 1.4769533445756042e-05, "loss": 0.1449, "step": 2103 }, { "epoch": 0.08870712734785083, "grad_norm": 1.1011031866073608, "learning_rate": 1.4776559865092748e-05, "loss": 0.5383, "step": 2104 }, { "epoch": 0.08874928853005039, "grad_norm": 0.9232157468795776, "learning_rate": 1.4783586284429454e-05, "loss": 0.7762, "step": 2105 }, { "epoch": 0.08879144971224993, "grad_norm": 0.7764474153518677, "learning_rate": 1.479061270376616e-05, "loss": 0.1979, "step": 2106 }, { "epoch": 0.08883361089444948, "grad_norm": 1.2538504600524902, "learning_rate": 1.4797639123102868e-05, "loss": 0.4244, "step": 2107 }, { "epoch": 0.08887577207664903, "grad_norm": 1.3526684045791626, "learning_rate": 1.4804665542439574e-05, "loss": 0.3768, "step": 2108 }, { "epoch": 0.08891793325884857, "grad_norm": 0.9626387357711792, "learning_rate": 1.481169196177628e-05, "loss": 0.2804, "step": 2109 }, { "epoch": 0.08896009444104813, "grad_norm": 1.2525426149368286, "learning_rate": 1.4818718381112986e-05, "loss": 0.5464, "step": 2110 }, { "epoch": 0.08900225562324768, "grad_norm": 1.0189876556396484, "learning_rate": 1.4825744800449692e-05, "loss": 0.8386, "step": 2111 }, { "epoch": 0.08904441680544722, "grad_norm": 1.7038021087646484, "learning_rate": 1.4832771219786399e-05, "loss": 1.2848, "step": 2112 }, { "epoch": 0.08908657798764677, "grad_norm": 1.1564542055130005, "learning_rate": 1.4839797639123105e-05, "loss": 0.1535, "step": 2113 }, { "epoch": 0.08912873916984632, "grad_norm": 0.9906467199325562, "learning_rate": 1.484682405845981e-05, "loss": 0.2452, "step": 2114 }, { "epoch": 0.08917090035204588, "grad_norm": 1.1275194883346558, "learning_rate": 1.4853850477796517e-05, "loss": 0.164, "step": 2115 }, { "epoch": 0.08921306153424542, "grad_norm": 1.320141315460205, "learning_rate": 1.4860876897133221e-05, "loss": 0.851, "step": 2116 }, { "epoch": 0.08925522271644497, "grad_norm": 1.211073398590088, "learning_rate": 1.4867903316469927e-05, "loss": 0.7599, "step": 2117 }, { "epoch": 0.08929738389864451, "grad_norm": 1.1434625387191772, "learning_rate": 1.4874929735806633e-05, "loss": 0.2649, "step": 2118 }, { "epoch": 0.08933954508084406, "grad_norm": 2.1739916801452637, "learning_rate": 1.488195615514334e-05, "loss": 1.1768, "step": 2119 }, { "epoch": 0.08938170626304362, "grad_norm": 0.8726137280464172, "learning_rate": 1.4888982574480046e-05, "loss": 0.2101, "step": 2120 }, { "epoch": 0.08942386744524317, "grad_norm": 0.9801529049873352, "learning_rate": 1.4896008993816752e-05, "loss": 0.3755, "step": 2121 }, { "epoch": 0.08946602862744271, "grad_norm": 1.2450945377349854, "learning_rate": 1.4903035413153458e-05, "loss": 0.6208, "step": 2122 }, { "epoch": 0.08950818980964226, "grad_norm": 0.791803240776062, "learning_rate": 1.4910061832490164e-05, "loss": 0.1561, "step": 2123 }, { "epoch": 0.08955035099184182, "grad_norm": 0.9398097395896912, "learning_rate": 1.4917088251826868e-05, "loss": 0.2165, "step": 2124 }, { "epoch": 0.08959251217404136, "grad_norm": 1.418456792831421, "learning_rate": 1.4924114671163574e-05, "loss": 0.5272, "step": 2125 }, { "epoch": 0.08963467335624091, "grad_norm": 1.0449092388153076, "learning_rate": 1.493114109050028e-05, "loss": 0.5061, "step": 2126 }, { "epoch": 0.08967683453844046, "grad_norm": 1.2241297960281372, "learning_rate": 1.4938167509836986e-05, "loss": 0.2859, "step": 2127 }, { "epoch": 0.08971899572064, "grad_norm": 1.6931023597717285, "learning_rate": 1.4945193929173693e-05, "loss": 0.2505, "step": 2128 }, { "epoch": 0.08976115690283956, "grad_norm": 0.786237359046936, "learning_rate": 1.4952220348510399e-05, "loss": 0.2061, "step": 2129 }, { "epoch": 0.08980331808503911, "grad_norm": 1.0159528255462646, "learning_rate": 1.4959246767847106e-05, "loss": 0.581, "step": 2130 }, { "epoch": 0.08984547926723865, "grad_norm": 1.4921537637710571, "learning_rate": 1.4966273187183813e-05, "loss": 0.6478, "step": 2131 }, { "epoch": 0.0898876404494382, "grad_norm": 0.9723564386367798, "learning_rate": 1.4973299606520519e-05, "loss": 0.2418, "step": 2132 }, { "epoch": 0.08992980163163775, "grad_norm": 1.8536837100982666, "learning_rate": 1.4980326025857225e-05, "loss": 0.3597, "step": 2133 }, { "epoch": 0.0899719628138373, "grad_norm": 0.8628232479095459, "learning_rate": 1.498735244519393e-05, "loss": 0.1628, "step": 2134 }, { "epoch": 0.09001412399603685, "grad_norm": 0.9089725017547607, "learning_rate": 1.4994378864530637e-05, "loss": 0.8064, "step": 2135 }, { "epoch": 0.0900562851782364, "grad_norm": 1.4398620128631592, "learning_rate": 1.5001405283867343e-05, "loss": 0.8464, "step": 2136 }, { "epoch": 0.09009844636043594, "grad_norm": 2.3472025394439697, "learning_rate": 1.5008431703204049e-05, "loss": 1.089, "step": 2137 }, { "epoch": 0.09014060754263549, "grad_norm": 0.9450011253356934, "learning_rate": 1.5015458122540753e-05, "loss": 0.8615, "step": 2138 }, { "epoch": 0.09018276872483505, "grad_norm": 1.293870449066162, "learning_rate": 1.502248454187746e-05, "loss": 0.3209, "step": 2139 }, { "epoch": 0.0902249299070346, "grad_norm": 1.4864803552627563, "learning_rate": 1.5029510961214166e-05, "loss": 0.3274, "step": 2140 }, { "epoch": 0.09026709108923414, "grad_norm": 0.7684285640716553, "learning_rate": 1.5036537380550872e-05, "loss": 0.1903, "step": 2141 }, { "epoch": 0.09030925227143369, "grad_norm": 1.2274832725524902, "learning_rate": 1.5043563799887578e-05, "loss": 0.393, "step": 2142 }, { "epoch": 0.09035141345363323, "grad_norm": 0.9339274168014526, "learning_rate": 1.5050590219224284e-05, "loss": 0.4465, "step": 2143 }, { "epoch": 0.0903935746358328, "grad_norm": 1.4532850980758667, "learning_rate": 1.505761663856099e-05, "loss": 1.3593, "step": 2144 }, { "epoch": 0.09043573581803234, "grad_norm": 1.101244330406189, "learning_rate": 1.5064643057897696e-05, "loss": 0.4817, "step": 2145 }, { "epoch": 0.09047789700023189, "grad_norm": 1.154280662536621, "learning_rate": 1.5071669477234402e-05, "loss": 0.2306, "step": 2146 }, { "epoch": 0.09052005818243143, "grad_norm": 2.507031202316284, "learning_rate": 1.5078695896571107e-05, "loss": 1.3289, "step": 2147 }, { "epoch": 0.09056221936463098, "grad_norm": 0.822487473487854, "learning_rate": 1.5085722315907813e-05, "loss": 0.2892, "step": 2148 }, { "epoch": 0.09060438054683054, "grad_norm": 1.0157827138900757, "learning_rate": 1.5092748735244519e-05, "loss": 0.794, "step": 2149 }, { "epoch": 0.09064654172903008, "grad_norm": 1.1976861953735352, "learning_rate": 1.5099775154581225e-05, "loss": 0.8693, "step": 2150 }, { "epoch": 0.09068870291122963, "grad_norm": 2.6262478828430176, "learning_rate": 1.5106801573917931e-05, "loss": 0.5043, "step": 2151 }, { "epoch": 0.09073086409342918, "grad_norm": 0.5086235404014587, "learning_rate": 1.5113827993254639e-05, "loss": 0.1555, "step": 2152 }, { "epoch": 0.09077302527562874, "grad_norm": 2.271280288696289, "learning_rate": 1.5120854412591345e-05, "loss": 0.6593, "step": 2153 }, { "epoch": 0.09081518645782828, "grad_norm": 0.9887911081314087, "learning_rate": 1.5127880831928051e-05, "loss": 0.5998, "step": 2154 }, { "epoch": 0.09085734764002783, "grad_norm": 1.0143063068389893, "learning_rate": 1.5134907251264757e-05, "loss": 0.2683, "step": 2155 }, { "epoch": 0.09089950882222737, "grad_norm": 1.0227504968643188, "learning_rate": 1.5141933670601463e-05, "loss": 0.6091, "step": 2156 }, { "epoch": 0.09094167000442692, "grad_norm": 0.9085489511489868, "learning_rate": 1.5148960089938169e-05, "loss": 0.2503, "step": 2157 }, { "epoch": 0.09098383118662648, "grad_norm": 1.5190693140029907, "learning_rate": 1.5155986509274875e-05, "loss": 0.4818, "step": 2158 }, { "epoch": 0.09102599236882603, "grad_norm": 1.5608609914779663, "learning_rate": 1.5163012928611581e-05, "loss": 1.2417, "step": 2159 }, { "epoch": 0.09106815355102557, "grad_norm": 0.9936251640319824, "learning_rate": 1.5170039347948287e-05, "loss": 0.7738, "step": 2160 }, { "epoch": 0.09111031473322512, "grad_norm": 2.8873934745788574, "learning_rate": 1.5177065767284992e-05, "loss": 0.9124, "step": 2161 }, { "epoch": 0.09115247591542466, "grad_norm": 1.4168751239776611, "learning_rate": 1.5184092186621698e-05, "loss": 0.9766, "step": 2162 }, { "epoch": 0.09119463709762422, "grad_norm": 0.9800207614898682, "learning_rate": 1.5191118605958404e-05, "loss": 0.2518, "step": 2163 }, { "epoch": 0.09123679827982377, "grad_norm": 1.572056770324707, "learning_rate": 1.519814502529511e-05, "loss": 0.7972, "step": 2164 }, { "epoch": 0.09127895946202332, "grad_norm": 2.565763235092163, "learning_rate": 1.5205171444631816e-05, "loss": 0.9799, "step": 2165 }, { "epoch": 0.09132112064422286, "grad_norm": 1.1324660778045654, "learning_rate": 1.5212197863968522e-05, "loss": 0.3817, "step": 2166 }, { "epoch": 0.09136328182642241, "grad_norm": 1.7757829427719116, "learning_rate": 1.5219224283305228e-05, "loss": 1.3004, "step": 2167 }, { "epoch": 0.09140544300862197, "grad_norm": 0.8402407169342041, "learning_rate": 1.5226250702641934e-05, "loss": 0.226, "step": 2168 }, { "epoch": 0.09144760419082151, "grad_norm": 1.0693761110305786, "learning_rate": 1.5233277121978639e-05, "loss": 0.208, "step": 2169 }, { "epoch": 0.09148976537302106, "grad_norm": 0.7110028862953186, "learning_rate": 1.5240303541315345e-05, "loss": 0.1471, "step": 2170 }, { "epoch": 0.0915319265552206, "grad_norm": 3.343263626098633, "learning_rate": 1.5247329960652051e-05, "loss": 0.8603, "step": 2171 }, { "epoch": 0.09157408773742015, "grad_norm": 2.2027530670166016, "learning_rate": 1.5254356379988757e-05, "loss": 1.2436, "step": 2172 }, { "epoch": 0.09161624891961971, "grad_norm": 1.0174627304077148, "learning_rate": 1.5261382799325463e-05, "loss": 0.651, "step": 2173 }, { "epoch": 0.09165841010181926, "grad_norm": 1.0210506916046143, "learning_rate": 1.5268409218662168e-05, "loss": 0.3565, "step": 2174 }, { "epoch": 0.0917005712840188, "grad_norm": 1.17862069606781, "learning_rate": 1.527543563799888e-05, "loss": 0.8171, "step": 2175 }, { "epoch": 0.09174273246621835, "grad_norm": 1.1637362241744995, "learning_rate": 1.5282462057335583e-05, "loss": 0.205, "step": 2176 }, { "epoch": 0.0917848936484179, "grad_norm": 0.8912628889083862, "learning_rate": 1.528948847667229e-05, "loss": 0.188, "step": 2177 }, { "epoch": 0.09182705483061745, "grad_norm": 1.0424163341522217, "learning_rate": 1.5296514896008995e-05, "loss": 0.6446, "step": 2178 }, { "epoch": 0.091869216012817, "grad_norm": 1.006919503211975, "learning_rate": 1.53035413153457e-05, "loss": 0.4357, "step": 2179 }, { "epoch": 0.09191137719501655, "grad_norm": 1.034541368484497, "learning_rate": 1.5310567734682407e-05, "loss": 0.3508, "step": 2180 }, { "epoch": 0.09195353837721609, "grad_norm": 1.454710602760315, "learning_rate": 1.5317594154019112e-05, "loss": 0.3308, "step": 2181 }, { "epoch": 0.09199569955941565, "grad_norm": 0.967458188533783, "learning_rate": 1.532462057335582e-05, "loss": 0.1205, "step": 2182 }, { "epoch": 0.0920378607416152, "grad_norm": 0.9648364186286926, "learning_rate": 1.5331646992692524e-05, "loss": 0.1124, "step": 2183 }, { "epoch": 0.09208002192381474, "grad_norm": 1.0218024253845215, "learning_rate": 1.5338673412029232e-05, "loss": 0.6556, "step": 2184 }, { "epoch": 0.09212218310601429, "grad_norm": 1.4335932731628418, "learning_rate": 1.5345699831365936e-05, "loss": 0.3061, "step": 2185 }, { "epoch": 0.09216434428821384, "grad_norm": 0.7920841574668884, "learning_rate": 1.5352726250702644e-05, "loss": 0.2668, "step": 2186 }, { "epoch": 0.0922065054704134, "grad_norm": 1.4980155229568481, "learning_rate": 1.535975267003935e-05, "loss": 1.1842, "step": 2187 }, { "epoch": 0.09224866665261294, "grad_norm": 1.4668428897857666, "learning_rate": 1.5366779089376053e-05, "loss": 0.7808, "step": 2188 }, { "epoch": 0.09229082783481249, "grad_norm": 1.115371584892273, "learning_rate": 1.537380550871276e-05, "loss": 0.489, "step": 2189 }, { "epoch": 0.09233298901701203, "grad_norm": 1.0928951501846313, "learning_rate": 1.5380831928049465e-05, "loss": 0.2261, "step": 2190 }, { "epoch": 0.09237515019921158, "grad_norm": 1.5118526220321655, "learning_rate": 1.5387858347386173e-05, "loss": 1.1381, "step": 2191 }, { "epoch": 0.09241731138141114, "grad_norm": 1.8281351327896118, "learning_rate": 1.5394884766722877e-05, "loss": 0.3261, "step": 2192 }, { "epoch": 0.09245947256361069, "grad_norm": 1.1418670415878296, "learning_rate": 1.5401911186059585e-05, "loss": 0.8172, "step": 2193 }, { "epoch": 0.09250163374581023, "grad_norm": 1.2595611810684204, "learning_rate": 1.540893760539629e-05, "loss": 0.6625, "step": 2194 }, { "epoch": 0.09254379492800978, "grad_norm": 1.0345373153686523, "learning_rate": 1.5415964024732997e-05, "loss": 0.208, "step": 2195 }, { "epoch": 0.09258595611020932, "grad_norm": 0.628513753414154, "learning_rate": 1.54229904440697e-05, "loss": 0.1358, "step": 2196 }, { "epoch": 0.09262811729240888, "grad_norm": 1.5018644332885742, "learning_rate": 1.543001686340641e-05, "loss": 0.3659, "step": 2197 }, { "epoch": 0.09267027847460843, "grad_norm": 1.1094322204589844, "learning_rate": 1.5437043282743117e-05, "loss": 0.5413, "step": 2198 }, { "epoch": 0.09271243965680798, "grad_norm": 1.0229114294052124, "learning_rate": 1.544406970207982e-05, "loss": 0.2181, "step": 2199 }, { "epoch": 0.09275460083900752, "grad_norm": 1.0002880096435547, "learning_rate": 1.545109612141653e-05, "loss": 0.4283, "step": 2200 }, { "epoch": 0.09279676202120707, "grad_norm": 0.7018036246299744, "learning_rate": 1.5458122540753234e-05, "loss": 0.33, "step": 2201 }, { "epoch": 0.09283892320340663, "grad_norm": 0.5952367782592773, "learning_rate": 1.5465148960089938e-05, "loss": 0.1444, "step": 2202 }, { "epoch": 0.09288108438560617, "grad_norm": 0.8093176484107971, "learning_rate": 1.5472175379426646e-05, "loss": 0.1617, "step": 2203 }, { "epoch": 0.09292324556780572, "grad_norm": 1.2860811948776245, "learning_rate": 1.547920179876335e-05, "loss": 0.5477, "step": 2204 }, { "epoch": 0.09296540675000527, "grad_norm": 0.8687154650688171, "learning_rate": 1.5486228218100058e-05, "loss": 0.4077, "step": 2205 }, { "epoch": 0.09300756793220481, "grad_norm": 1.5752092599868774, "learning_rate": 1.5493254637436762e-05, "loss": 0.3944, "step": 2206 }, { "epoch": 0.09304972911440437, "grad_norm": 0.8923194408416748, "learning_rate": 1.550028105677347e-05, "loss": 0.5097, "step": 2207 }, { "epoch": 0.09309189029660392, "grad_norm": 1.0775636434555054, "learning_rate": 1.5507307476110175e-05, "loss": 0.7657, "step": 2208 }, { "epoch": 0.09313405147880346, "grad_norm": 0.7104566097259521, "learning_rate": 1.5514333895446882e-05, "loss": 0.2528, "step": 2209 }, { "epoch": 0.09317621266100301, "grad_norm": 0.8066586256027222, "learning_rate": 1.5521360314783587e-05, "loss": 0.1706, "step": 2210 }, { "epoch": 0.09321837384320256, "grad_norm": 0.7990266680717468, "learning_rate": 1.552838673412029e-05, "loss": 0.2095, "step": 2211 }, { "epoch": 0.09326053502540212, "grad_norm": 0.8450993895530701, "learning_rate": 1.5535413153457e-05, "loss": 0.4714, "step": 2212 }, { "epoch": 0.09330269620760166, "grad_norm": 1.7894949913024902, "learning_rate": 1.5542439572793703e-05, "loss": 1.128, "step": 2213 }, { "epoch": 0.09334485738980121, "grad_norm": 1.098630666732788, "learning_rate": 1.554946599213041e-05, "loss": 0.2019, "step": 2214 }, { "epoch": 0.09338701857200075, "grad_norm": 1.8261358737945557, "learning_rate": 1.5556492411467116e-05, "loss": 1.1948, "step": 2215 }, { "epoch": 0.09342917975420031, "grad_norm": 1.0753649473190308, "learning_rate": 1.5563518830803823e-05, "loss": 0.711, "step": 2216 }, { "epoch": 0.09347134093639986, "grad_norm": 1.7092870473861694, "learning_rate": 1.5570545250140528e-05, "loss": 0.7523, "step": 2217 }, { "epoch": 0.0935135021185994, "grad_norm": 0.7670871019363403, "learning_rate": 1.5577571669477235e-05, "loss": 0.2045, "step": 2218 }, { "epoch": 0.09355566330079895, "grad_norm": 1.2907787561416626, "learning_rate": 1.558459808881394e-05, "loss": 0.3195, "step": 2219 }, { "epoch": 0.0935978244829985, "grad_norm": 0.8480671048164368, "learning_rate": 1.5591624508150648e-05, "loss": 0.1464, "step": 2220 }, { "epoch": 0.09363998566519806, "grad_norm": 0.6255781054496765, "learning_rate": 1.5598650927487355e-05, "loss": 0.1498, "step": 2221 }, { "epoch": 0.0936821468473976, "grad_norm": 1.4679584503173828, "learning_rate": 1.560567734682406e-05, "loss": 0.2431, "step": 2222 }, { "epoch": 0.09372430802959715, "grad_norm": 0.7838168740272522, "learning_rate": 1.5612703766160768e-05, "loss": 0.2045, "step": 2223 }, { "epoch": 0.0937664692117967, "grad_norm": 0.8566824793815613, "learning_rate": 1.5619730185497472e-05, "loss": 0.1836, "step": 2224 }, { "epoch": 0.09380863039399624, "grad_norm": 0.9185778498649597, "learning_rate": 1.5626756604834176e-05, "loss": 0.6505, "step": 2225 }, { "epoch": 0.0938507915761958, "grad_norm": 0.9523715376853943, "learning_rate": 1.5633783024170884e-05, "loss": 0.7046, "step": 2226 }, { "epoch": 0.09389295275839535, "grad_norm": 1.374873161315918, "learning_rate": 1.564080944350759e-05, "loss": 0.4983, "step": 2227 }, { "epoch": 0.0939351139405949, "grad_norm": 0.7132128477096558, "learning_rate": 1.5647835862844296e-05, "loss": 0.1603, "step": 2228 }, { "epoch": 0.09397727512279444, "grad_norm": 1.213283658027649, "learning_rate": 1.5654862282181e-05, "loss": 0.4915, "step": 2229 }, { "epoch": 0.09401943630499399, "grad_norm": 1.3570339679718018, "learning_rate": 1.566188870151771e-05, "loss": 0.3102, "step": 2230 }, { "epoch": 0.09406159748719355, "grad_norm": 0.8098737597465515, "learning_rate": 1.5668915120854413e-05, "loss": 0.157, "step": 2231 }, { "epoch": 0.09410375866939309, "grad_norm": 1.074904441833496, "learning_rate": 1.567594154019112e-05, "loss": 0.164, "step": 2232 }, { "epoch": 0.09414591985159264, "grad_norm": 1.5544140338897705, "learning_rate": 1.5682967959527825e-05, "loss": 1.1666, "step": 2233 }, { "epoch": 0.09418808103379218, "grad_norm": 1.7211421728134155, "learning_rate": 1.568999437886453e-05, "loss": 1.1255, "step": 2234 }, { "epoch": 0.09423024221599173, "grad_norm": 0.7899603247642517, "learning_rate": 1.5697020798201237e-05, "loss": 0.1641, "step": 2235 }, { "epoch": 0.09427240339819129, "grad_norm": 0.7843157052993774, "learning_rate": 1.5704047217537942e-05, "loss": 0.2065, "step": 2236 }, { "epoch": 0.09431456458039084, "grad_norm": 0.7628613710403442, "learning_rate": 1.571107363687465e-05, "loss": 0.1932, "step": 2237 }, { "epoch": 0.09435672576259038, "grad_norm": 0.7953601479530334, "learning_rate": 1.5718100056211354e-05, "loss": 0.2208, "step": 2238 }, { "epoch": 0.09439888694478993, "grad_norm": 0.933014452457428, "learning_rate": 1.572512647554806e-05, "loss": 0.3726, "step": 2239 }, { "epoch": 0.09444104812698947, "grad_norm": 1.127395749092102, "learning_rate": 1.5732152894884766e-05, "loss": 0.2308, "step": 2240 }, { "epoch": 0.09448320930918903, "grad_norm": 0.7642242908477783, "learning_rate": 1.5739179314221474e-05, "loss": 0.1577, "step": 2241 }, { "epoch": 0.09452537049138858, "grad_norm": 1.8798890113830566, "learning_rate": 1.574620573355818e-05, "loss": 1.1133, "step": 2242 }, { "epoch": 0.09456753167358813, "grad_norm": 0.9287672638893127, "learning_rate": 1.5753232152894886e-05, "loss": 0.2255, "step": 2243 }, { "epoch": 0.09460969285578767, "grad_norm": 0.8773738145828247, "learning_rate": 1.5760258572231594e-05, "loss": 0.6526, "step": 2244 }, { "epoch": 0.09465185403798723, "grad_norm": 0.7242274284362793, "learning_rate": 1.5767284991568298e-05, "loss": 0.3396, "step": 2245 }, { "epoch": 0.09469401522018678, "grad_norm": 1.071089506149292, "learning_rate": 1.5774311410905006e-05, "loss": 0.158, "step": 2246 }, { "epoch": 0.09473617640238632, "grad_norm": 2.5466644763946533, "learning_rate": 1.578133783024171e-05, "loss": 0.6569, "step": 2247 }, { "epoch": 0.09477833758458587, "grad_norm": 0.7597561478614807, "learning_rate": 1.5788364249578415e-05, "loss": 0.2811, "step": 2248 }, { "epoch": 0.09482049876678542, "grad_norm": 0.694885790348053, "learning_rate": 1.5795390668915123e-05, "loss": 0.3245, "step": 2249 }, { "epoch": 0.09486265994898498, "grad_norm": 0.9576011896133423, "learning_rate": 1.5802417088251827e-05, "loss": 0.4501, "step": 2250 }, { "epoch": 0.09490482113118452, "grad_norm": 1.0112630128860474, "learning_rate": 1.5809443507588535e-05, "loss": 0.5292, "step": 2251 }, { "epoch": 0.09494698231338407, "grad_norm": 1.3806802034378052, "learning_rate": 1.581646992692524e-05, "loss": 0.2781, "step": 2252 }, { "epoch": 0.09498914349558361, "grad_norm": 1.5270287990570068, "learning_rate": 1.5823496346261947e-05, "loss": 0.4516, "step": 2253 }, { "epoch": 0.09503130467778316, "grad_norm": 0.8847482204437256, "learning_rate": 1.583052276559865e-05, "loss": 0.4175, "step": 2254 }, { "epoch": 0.09507346585998272, "grad_norm": 2.6321258544921875, "learning_rate": 1.583754918493536e-05, "loss": 0.6498, "step": 2255 }, { "epoch": 0.09511562704218227, "grad_norm": 1.233877420425415, "learning_rate": 1.5844575604272063e-05, "loss": 0.3503, "step": 2256 }, { "epoch": 0.09515778822438181, "grad_norm": 0.788101315498352, "learning_rate": 1.5851602023608768e-05, "loss": 0.2362, "step": 2257 }, { "epoch": 0.09519994940658136, "grad_norm": 3.6097750663757324, "learning_rate": 1.5858628442945476e-05, "loss": 0.7184, "step": 2258 }, { "epoch": 0.0952421105887809, "grad_norm": 1.1111245155334473, "learning_rate": 1.586565486228218e-05, "loss": 0.8391, "step": 2259 }, { "epoch": 0.09528427177098046, "grad_norm": 0.9418976902961731, "learning_rate": 1.5872681281618888e-05, "loss": 0.5359, "step": 2260 }, { "epoch": 0.09532643295318001, "grad_norm": 1.2920968532562256, "learning_rate": 1.5879707700955592e-05, "loss": 0.61, "step": 2261 }, { "epoch": 0.09536859413537956, "grad_norm": 1.1793266534805298, "learning_rate": 1.58867341202923e-05, "loss": 0.7325, "step": 2262 }, { "epoch": 0.0954107553175791, "grad_norm": 0.788335382938385, "learning_rate": 1.5893760539629004e-05, "loss": 0.2971, "step": 2263 }, { "epoch": 0.09545291649977865, "grad_norm": 1.4112133979797363, "learning_rate": 1.590078695896571e-05, "loss": 0.5308, "step": 2264 }, { "epoch": 0.09549507768197821, "grad_norm": 2.6410739421844482, "learning_rate": 1.590781337830242e-05, "loss": 1.1792, "step": 2265 }, { "epoch": 0.09553723886417775, "grad_norm": 1.0759073495864868, "learning_rate": 1.5914839797639124e-05, "loss": 0.3205, "step": 2266 }, { "epoch": 0.0955794000463773, "grad_norm": 0.8093762993812561, "learning_rate": 1.5921866216975832e-05, "loss": 0.1774, "step": 2267 }, { "epoch": 0.09562156122857685, "grad_norm": 1.2013808488845825, "learning_rate": 1.5928892636312537e-05, "loss": 0.3158, "step": 2268 }, { "epoch": 0.09566372241077639, "grad_norm": 1.101377010345459, "learning_rate": 1.5935919055649244e-05, "loss": 0.4403, "step": 2269 }, { "epoch": 0.09570588359297595, "grad_norm": 0.746199369430542, "learning_rate": 1.594294547498595e-05, "loss": 0.2491, "step": 2270 }, { "epoch": 0.0957480447751755, "grad_norm": 0.5922813415527344, "learning_rate": 1.5949971894322653e-05, "loss": 0.1395, "step": 2271 }, { "epoch": 0.09579020595737504, "grad_norm": 0.9332484006881714, "learning_rate": 1.595699831365936e-05, "loss": 0.5252, "step": 2272 }, { "epoch": 0.09583236713957459, "grad_norm": 0.6101363897323608, "learning_rate": 1.5964024732996065e-05, "loss": 0.1195, "step": 2273 }, { "epoch": 0.09587452832177414, "grad_norm": 0.8702371716499329, "learning_rate": 1.5971051152332773e-05, "loss": 0.4778, "step": 2274 }, { "epoch": 0.0959166895039737, "grad_norm": 1.9946422576904297, "learning_rate": 1.5978077571669477e-05, "loss": 1.0853, "step": 2275 }, { "epoch": 0.09595885068617324, "grad_norm": 0.9892051815986633, "learning_rate": 1.5985103991006185e-05, "loss": 0.5546, "step": 2276 }, { "epoch": 0.09600101186837279, "grad_norm": 0.8852131366729736, "learning_rate": 1.599213041034289e-05, "loss": 0.6773, "step": 2277 }, { "epoch": 0.09604317305057233, "grad_norm": 0.9530655741691589, "learning_rate": 1.5999156829679594e-05, "loss": 0.2136, "step": 2278 }, { "epoch": 0.0960853342327719, "grad_norm": 0.7774327397346497, "learning_rate": 1.6006183249016302e-05, "loss": 0.5657, "step": 2279 }, { "epoch": 0.09612749541497144, "grad_norm": 0.8427597880363464, "learning_rate": 1.6013209668353006e-05, "loss": 0.3112, "step": 2280 }, { "epoch": 0.09616965659717099, "grad_norm": 1.0168373584747314, "learning_rate": 1.6020236087689714e-05, "loss": 0.485, "step": 2281 }, { "epoch": 0.09621181777937053, "grad_norm": 0.7157989740371704, "learning_rate": 1.602726250702642e-05, "loss": 0.2615, "step": 2282 }, { "epoch": 0.09625397896157008, "grad_norm": 1.000778317451477, "learning_rate": 1.6034288926363126e-05, "loss": 0.5292, "step": 2283 }, { "epoch": 0.09629614014376964, "grad_norm": 1.2416882514953613, "learning_rate": 1.604131534569983e-05, "loss": 0.3459, "step": 2284 }, { "epoch": 0.09633830132596918, "grad_norm": 0.903559148311615, "learning_rate": 1.604834176503654e-05, "loss": 0.1785, "step": 2285 }, { "epoch": 0.09638046250816873, "grad_norm": 1.0113519430160522, "learning_rate": 1.6055368184373243e-05, "loss": 0.5308, "step": 2286 }, { "epoch": 0.09642262369036828, "grad_norm": 0.9734504818916321, "learning_rate": 1.606239460370995e-05, "loss": 0.7386, "step": 2287 }, { "epoch": 0.09646478487256782, "grad_norm": 0.7894779443740845, "learning_rate": 1.606942102304666e-05, "loss": 0.488, "step": 2288 }, { "epoch": 0.09650694605476738, "grad_norm": 1.0935769081115723, "learning_rate": 1.6076447442383363e-05, "loss": 1.1441, "step": 2289 }, { "epoch": 0.09654910723696693, "grad_norm": 0.9705507755279541, "learning_rate": 1.608347386172007e-05, "loss": 0.4017, "step": 2290 }, { "epoch": 0.09659126841916647, "grad_norm": 0.9236145615577698, "learning_rate": 1.6090500281056775e-05, "loss": 0.3329, "step": 2291 }, { "epoch": 0.09663342960136602, "grad_norm": 1.2983064651489258, "learning_rate": 1.609752670039348e-05, "loss": 0.2748, "step": 2292 }, { "epoch": 0.09667559078356557, "grad_norm": 1.0374400615692139, "learning_rate": 1.6104553119730187e-05, "loss": 0.563, "step": 2293 }, { "epoch": 0.09671775196576513, "grad_norm": 0.6260776519775391, "learning_rate": 1.611157953906689e-05, "loss": 0.1253, "step": 2294 }, { "epoch": 0.09675991314796467, "grad_norm": 0.8911309838294983, "learning_rate": 1.61186059584036e-05, "loss": 0.1623, "step": 2295 }, { "epoch": 0.09680207433016422, "grad_norm": 4.1384477615356445, "learning_rate": 1.6125632377740304e-05, "loss": 0.3391, "step": 2296 }, { "epoch": 0.09684423551236376, "grad_norm": 0.46717846393585205, "learning_rate": 1.613265879707701e-05, "loss": 0.1384, "step": 2297 }, { "epoch": 0.09688639669456331, "grad_norm": 1.4587949514389038, "learning_rate": 1.6139685216413716e-05, "loss": 0.8309, "step": 2298 }, { "epoch": 0.09692855787676287, "grad_norm": 1.0776708126068115, "learning_rate": 1.6146711635750424e-05, "loss": 0.3605, "step": 2299 }, { "epoch": 0.09697071905896242, "grad_norm": 1.1438133716583252, "learning_rate": 1.6153738055087128e-05, "loss": 0.4852, "step": 2300 }, { "epoch": 0.09701288024116196, "grad_norm": 1.678198218345642, "learning_rate": 1.6160764474423832e-05, "loss": 1.1926, "step": 2301 }, { "epoch": 0.09705504142336151, "grad_norm": 0.7521281242370605, "learning_rate": 1.616779089376054e-05, "loss": 0.2759, "step": 2302 }, { "epoch": 0.09709720260556105, "grad_norm": 0.7586576342582703, "learning_rate": 1.6174817313097245e-05, "loss": 0.208, "step": 2303 }, { "epoch": 0.09713936378776061, "grad_norm": 1.554213047027588, "learning_rate": 1.6181843732433952e-05, "loss": 0.4966, "step": 2304 }, { "epoch": 0.09718152496996016, "grad_norm": 0.8868125677108765, "learning_rate": 1.6188870151770657e-05, "loss": 0.1647, "step": 2305 }, { "epoch": 0.0972236861521597, "grad_norm": 0.7174361348152161, "learning_rate": 1.6195896571107365e-05, "loss": 0.1767, "step": 2306 }, { "epoch": 0.09726584733435925, "grad_norm": 0.9666168689727783, "learning_rate": 1.620292299044407e-05, "loss": 0.2687, "step": 2307 }, { "epoch": 0.09730800851655881, "grad_norm": 0.8797942996025085, "learning_rate": 1.6209949409780777e-05, "loss": 0.5241, "step": 2308 }, { "epoch": 0.09735016969875836, "grad_norm": 2.706228017807007, "learning_rate": 1.621697582911748e-05, "loss": 0.514, "step": 2309 }, { "epoch": 0.0973923308809579, "grad_norm": 0.7274206876754761, "learning_rate": 1.622400224845419e-05, "loss": 0.2722, "step": 2310 }, { "epoch": 0.09743449206315745, "grad_norm": 0.8582733273506165, "learning_rate": 1.6231028667790897e-05, "loss": 0.7378, "step": 2311 }, { "epoch": 0.097476653245357, "grad_norm": 1.5124377012252808, "learning_rate": 1.62380550871276e-05, "loss": 1.3136, "step": 2312 }, { "epoch": 0.09751881442755655, "grad_norm": 1.3839657306671143, "learning_rate": 1.624508150646431e-05, "loss": 1.1389, "step": 2313 }, { "epoch": 0.0975609756097561, "grad_norm": 0.7597332000732422, "learning_rate": 1.6252107925801013e-05, "loss": 0.2115, "step": 2314 }, { "epoch": 0.09760313679195565, "grad_norm": 0.7382745742797852, "learning_rate": 1.6259134345137718e-05, "loss": 0.2478, "step": 2315 }, { "epoch": 0.09764529797415519, "grad_norm": 1.0437933206558228, "learning_rate": 1.6266160764474425e-05, "loss": 0.6669, "step": 2316 }, { "epoch": 0.09768745915635474, "grad_norm": 1.1873760223388672, "learning_rate": 1.627318718381113e-05, "loss": 0.6715, "step": 2317 }, { "epoch": 0.0977296203385543, "grad_norm": 3.2446062564849854, "learning_rate": 1.6280213603147838e-05, "loss": 0.8606, "step": 2318 }, { "epoch": 0.09777178152075384, "grad_norm": 0.8518437743186951, "learning_rate": 1.6287240022484542e-05, "loss": 0.1304, "step": 2319 }, { "epoch": 0.09781394270295339, "grad_norm": 2.782433032989502, "learning_rate": 1.629426644182125e-05, "loss": 0.9175, "step": 2320 }, { "epoch": 0.09785610388515294, "grad_norm": 1.1155447959899902, "learning_rate": 1.6301292861157954e-05, "loss": 0.2995, "step": 2321 }, { "epoch": 0.09789826506735248, "grad_norm": 0.6818965673446655, "learning_rate": 1.6308319280494662e-05, "loss": 0.1572, "step": 2322 }, { "epoch": 0.09794042624955204, "grad_norm": 0.7890714406967163, "learning_rate": 1.6315345699831366e-05, "loss": 0.4797, "step": 2323 }, { "epoch": 0.09798258743175159, "grad_norm": 0.6247866749763489, "learning_rate": 1.632237211916807e-05, "loss": 0.1676, "step": 2324 }, { "epoch": 0.09802474861395113, "grad_norm": 1.132295846939087, "learning_rate": 1.632939853850478e-05, "loss": 0.1928, "step": 2325 }, { "epoch": 0.09806690979615068, "grad_norm": 1.2677146196365356, "learning_rate": 1.6336424957841483e-05, "loss": 1.1968, "step": 2326 }, { "epoch": 0.09810907097835023, "grad_norm": 1.5040820837020874, "learning_rate": 1.634345137717819e-05, "loss": 0.5627, "step": 2327 }, { "epoch": 0.09815123216054979, "grad_norm": 1.4078667163848877, "learning_rate": 1.6350477796514895e-05, "loss": 0.3681, "step": 2328 }, { "epoch": 0.09819339334274933, "grad_norm": 0.9147065281867981, "learning_rate": 1.6357504215851603e-05, "loss": 0.6591, "step": 2329 }, { "epoch": 0.09823555452494888, "grad_norm": 1.5793790817260742, "learning_rate": 1.6364530635188307e-05, "loss": 0.4755, "step": 2330 }, { "epoch": 0.09827771570714842, "grad_norm": 1.041709303855896, "learning_rate": 1.6371557054525015e-05, "loss": 0.666, "step": 2331 }, { "epoch": 0.09831987688934797, "grad_norm": 1.2551618814468384, "learning_rate": 1.6378583473861723e-05, "loss": 0.2915, "step": 2332 }, { "epoch": 0.09836203807154753, "grad_norm": 1.0351295471191406, "learning_rate": 1.6385609893198427e-05, "loss": 0.5097, "step": 2333 }, { "epoch": 0.09840419925374708, "grad_norm": 1.5957425832748413, "learning_rate": 1.6392636312535135e-05, "loss": 1.1484, "step": 2334 }, { "epoch": 0.09844636043594662, "grad_norm": 0.8396458029747009, "learning_rate": 1.639966273187184e-05, "loss": 0.365, "step": 2335 }, { "epoch": 0.09848852161814617, "grad_norm": 0.9524526596069336, "learning_rate": 1.6406689151208547e-05, "loss": 0.2174, "step": 2336 }, { "epoch": 0.09853068280034573, "grad_norm": 1.239571452140808, "learning_rate": 1.641371557054525e-05, "loss": 0.2287, "step": 2337 }, { "epoch": 0.09857284398254527, "grad_norm": 0.972112774848938, "learning_rate": 1.6420741989881956e-05, "loss": 0.6946, "step": 2338 }, { "epoch": 0.09861500516474482, "grad_norm": 1.3255845308303833, "learning_rate": 1.6427768409218664e-05, "loss": 0.8975, "step": 2339 }, { "epoch": 0.09865716634694437, "grad_norm": 0.8466047048568726, "learning_rate": 1.6434794828555368e-05, "loss": 0.4059, "step": 2340 }, { "epoch": 0.09869932752914391, "grad_norm": 1.4935544729232788, "learning_rate": 1.6441821247892076e-05, "loss": 1.1533, "step": 2341 }, { "epoch": 0.09874148871134347, "grad_norm": 0.6906057000160217, "learning_rate": 1.644884766722878e-05, "loss": 0.1302, "step": 2342 }, { "epoch": 0.09878364989354302, "grad_norm": 1.03156316280365, "learning_rate": 1.6455874086565488e-05, "loss": 0.2519, "step": 2343 }, { "epoch": 0.09882581107574256, "grad_norm": 1.20457124710083, "learning_rate": 1.6462900505902193e-05, "loss": 0.5452, "step": 2344 }, { "epoch": 0.09886797225794211, "grad_norm": 0.770566463470459, "learning_rate": 1.64699269252389e-05, "loss": 0.2336, "step": 2345 }, { "epoch": 0.09891013344014166, "grad_norm": 0.7879003882408142, "learning_rate": 1.6476953344575605e-05, "loss": 0.4357, "step": 2346 }, { "epoch": 0.09895229462234122, "grad_norm": 0.9802407026290894, "learning_rate": 1.648397976391231e-05, "loss": 0.2746, "step": 2347 }, { "epoch": 0.09899445580454076, "grad_norm": 1.2103296518325806, "learning_rate": 1.6491006183249017e-05, "loss": 0.302, "step": 2348 }, { "epoch": 0.09903661698674031, "grad_norm": 1.8574424982070923, "learning_rate": 1.649803260258572e-05, "loss": 0.6963, "step": 2349 }, { "epoch": 0.09907877816893985, "grad_norm": 1.1074202060699463, "learning_rate": 1.650505902192243e-05, "loss": 0.3091, "step": 2350 }, { "epoch": 0.0991209393511394, "grad_norm": 0.8708275556564331, "learning_rate": 1.6512085441259134e-05, "loss": 0.1717, "step": 2351 }, { "epoch": 0.09916310053333896, "grad_norm": 0.9181727170944214, "learning_rate": 1.651911186059584e-05, "loss": 0.2005, "step": 2352 }, { "epoch": 0.0992052617155385, "grad_norm": 0.9011229872703552, "learning_rate": 1.6526138279932546e-05, "loss": 0.238, "step": 2353 }, { "epoch": 0.09924742289773805, "grad_norm": 0.8504952788352966, "learning_rate": 1.6533164699269253e-05, "loss": 0.1713, "step": 2354 }, { "epoch": 0.0992895840799376, "grad_norm": 1.6677613258361816, "learning_rate": 1.654019111860596e-05, "loss": 1.2962, "step": 2355 }, { "epoch": 0.09933174526213714, "grad_norm": 0.9452931880950928, "learning_rate": 1.6547217537942666e-05, "loss": 0.8284, "step": 2356 }, { "epoch": 0.0993739064443367, "grad_norm": 2.1451468467712402, "learning_rate": 1.6554243957279373e-05, "loss": 1.2129, "step": 2357 }, { "epoch": 0.09941606762653625, "grad_norm": 0.8043136596679688, "learning_rate": 1.6561270376616078e-05, "loss": 0.5457, "step": 2358 }, { "epoch": 0.0994582288087358, "grad_norm": 0.9407879114151001, "learning_rate": 1.6568296795952786e-05, "loss": 0.4779, "step": 2359 }, { "epoch": 0.09950038999093534, "grad_norm": 1.0967912673950195, "learning_rate": 1.657532321528949e-05, "loss": 0.3377, "step": 2360 }, { "epoch": 0.09954255117313489, "grad_norm": 0.6030110120773315, "learning_rate": 1.6582349634626194e-05, "loss": 0.1666, "step": 2361 }, { "epoch": 0.09958471235533445, "grad_norm": 1.610534429550171, "learning_rate": 1.6589376053962902e-05, "loss": 1.251, "step": 2362 }, { "epoch": 0.099626873537534, "grad_norm": 0.9912439584732056, "learning_rate": 1.6596402473299607e-05, "loss": 0.3057, "step": 2363 }, { "epoch": 0.09966903471973354, "grad_norm": 1.124466061592102, "learning_rate": 1.6603428892636314e-05, "loss": 0.2327, "step": 2364 }, { "epoch": 0.09971119590193309, "grad_norm": 1.175252914428711, "learning_rate": 1.661045531197302e-05, "loss": 0.2741, "step": 2365 }, { "epoch": 0.09975335708413263, "grad_norm": 0.9805692434310913, "learning_rate": 1.6617481731309727e-05, "loss": 0.2479, "step": 2366 }, { "epoch": 0.09979551826633219, "grad_norm": 0.7462818026542664, "learning_rate": 1.662450815064643e-05, "loss": 0.4242, "step": 2367 }, { "epoch": 0.09983767944853174, "grad_norm": 0.8694794178009033, "learning_rate": 1.663153456998314e-05, "loss": 0.3566, "step": 2368 }, { "epoch": 0.09987984063073128, "grad_norm": 0.7238613367080688, "learning_rate": 1.6638560989319843e-05, "loss": 0.1698, "step": 2369 }, { "epoch": 0.09992200181293083, "grad_norm": 1.2775282859802246, "learning_rate": 1.6645587408656548e-05, "loss": 0.372, "step": 2370 }, { "epoch": 0.09996416299513039, "grad_norm": 1.1941092014312744, "learning_rate": 1.6652613827993255e-05, "loss": 0.5233, "step": 2371 }, { "epoch": 0.10000632417732994, "grad_norm": 0.816774845123291, "learning_rate": 1.665964024732996e-05, "loss": 0.2546, "step": 2372 }, { "epoch": 0.10004848535952948, "grad_norm": 1.1932986974716187, "learning_rate": 1.6666666666666667e-05, "loss": 0.5914, "step": 2373 }, { "epoch": 0.10009064654172903, "grad_norm": 0.772672712802887, "learning_rate": 1.6673693086003372e-05, "loss": 0.2552, "step": 2374 }, { "epoch": 0.10013280772392857, "grad_norm": 0.7539653182029724, "learning_rate": 1.668071950534008e-05, "loss": 0.3057, "step": 2375 }, { "epoch": 0.10017496890612813, "grad_norm": 2.0476973056793213, "learning_rate": 1.6687745924676784e-05, "loss": 1.2311, "step": 2376 }, { "epoch": 0.10021713008832768, "grad_norm": 0.8994516134262085, "learning_rate": 1.6694772344013492e-05, "loss": 0.1424, "step": 2377 }, { "epoch": 0.10025929127052723, "grad_norm": 0.8794811367988586, "learning_rate": 1.67017987633502e-05, "loss": 0.3433, "step": 2378 }, { "epoch": 0.10030145245272677, "grad_norm": 1.2254992723464966, "learning_rate": 1.6708825182686904e-05, "loss": 0.5628, "step": 2379 }, { "epoch": 0.10034361363492632, "grad_norm": 0.7176948189735413, "learning_rate": 1.6715851602023612e-05, "loss": 0.2302, "step": 2380 }, { "epoch": 0.10038577481712588, "grad_norm": 0.8328309059143066, "learning_rate": 1.6722878021360316e-05, "loss": 0.5557, "step": 2381 }, { "epoch": 0.10042793599932542, "grad_norm": 0.6629101037979126, "learning_rate": 1.6729904440697024e-05, "loss": 0.2249, "step": 2382 }, { "epoch": 0.10047009718152497, "grad_norm": 0.8014935851097107, "learning_rate": 1.673693086003373e-05, "loss": 0.2248, "step": 2383 }, { "epoch": 0.10051225836372452, "grad_norm": 2.413076877593994, "learning_rate": 1.6743957279370433e-05, "loss": 0.9876, "step": 2384 }, { "epoch": 0.10055441954592406, "grad_norm": 1.1660312414169312, "learning_rate": 1.675098369870714e-05, "loss": 0.6157, "step": 2385 }, { "epoch": 0.10059658072812362, "grad_norm": 0.9243293404579163, "learning_rate": 1.6758010118043845e-05, "loss": 0.3401, "step": 2386 }, { "epoch": 0.10063874191032317, "grad_norm": 1.8829481601715088, "learning_rate": 1.6765036537380553e-05, "loss": 0.9858, "step": 2387 }, { "epoch": 0.10068090309252271, "grad_norm": 1.3799492120742798, "learning_rate": 1.6772062956717257e-05, "loss": 1.026, "step": 2388 }, { "epoch": 0.10072306427472226, "grad_norm": 1.225683331489563, "learning_rate": 1.6779089376053965e-05, "loss": 0.2955, "step": 2389 }, { "epoch": 0.1007652254569218, "grad_norm": 0.5077125430107117, "learning_rate": 1.678611579539067e-05, "loss": 0.175, "step": 2390 }, { "epoch": 0.10080738663912137, "grad_norm": 0.7893262505531311, "learning_rate": 1.6793142214727374e-05, "loss": 0.3107, "step": 2391 }, { "epoch": 0.10084954782132091, "grad_norm": 0.9030351042747498, "learning_rate": 1.680016863406408e-05, "loss": 0.7974, "step": 2392 }, { "epoch": 0.10089170900352046, "grad_norm": 0.512031078338623, "learning_rate": 1.6807195053400786e-05, "loss": 0.2209, "step": 2393 }, { "epoch": 0.10093387018572, "grad_norm": 1.4153512716293335, "learning_rate": 1.6814221472737494e-05, "loss": 1.1839, "step": 2394 }, { "epoch": 0.10097603136791955, "grad_norm": 0.966995120048523, "learning_rate": 1.6821247892074198e-05, "loss": 0.5411, "step": 2395 }, { "epoch": 0.10101819255011911, "grad_norm": 1.5002490282058716, "learning_rate": 1.6828274311410906e-05, "loss": 0.2514, "step": 2396 }, { "epoch": 0.10106035373231866, "grad_norm": 0.7280505895614624, "learning_rate": 1.683530073074761e-05, "loss": 0.3565, "step": 2397 }, { "epoch": 0.1011025149145182, "grad_norm": 1.6467584371566772, "learning_rate": 1.6842327150084318e-05, "loss": 0.3458, "step": 2398 }, { "epoch": 0.10114467609671775, "grad_norm": 0.7520173192024231, "learning_rate": 1.6849353569421022e-05, "loss": 0.2085, "step": 2399 }, { "epoch": 0.10118683727891731, "grad_norm": 0.8783631920814514, "learning_rate": 1.685637998875773e-05, "loss": 0.4719, "step": 2400 }, { "epoch": 0.10122899846111685, "grad_norm": 0.9863819479942322, "learning_rate": 1.6863406408094438e-05, "loss": 0.7175, "step": 2401 }, { "epoch": 0.1012711596433164, "grad_norm": 2.5648269653320312, "learning_rate": 1.6870432827431142e-05, "loss": 1.2239, "step": 2402 }, { "epoch": 0.10131332082551595, "grad_norm": 1.0116239786148071, "learning_rate": 1.687745924676785e-05, "loss": 0.1915, "step": 2403 }, { "epoch": 0.10135548200771549, "grad_norm": 1.7149256467819214, "learning_rate": 1.6884485666104555e-05, "loss": 1.184, "step": 2404 }, { "epoch": 0.10139764318991505, "grad_norm": 1.1326700448989868, "learning_rate": 1.689151208544126e-05, "loss": 0.5172, "step": 2405 }, { "epoch": 0.1014398043721146, "grad_norm": 0.9430103898048401, "learning_rate": 1.6898538504777967e-05, "loss": 0.1889, "step": 2406 }, { "epoch": 0.10148196555431414, "grad_norm": 0.9900254011154175, "learning_rate": 1.690556492411467e-05, "loss": 0.2816, "step": 2407 }, { "epoch": 0.10152412673651369, "grad_norm": 2.6284239292144775, "learning_rate": 1.691259134345138e-05, "loss": 0.4586, "step": 2408 }, { "epoch": 0.10156628791871324, "grad_norm": 1.2324142456054688, "learning_rate": 1.6919617762788083e-05, "loss": 0.3132, "step": 2409 }, { "epoch": 0.1016084491009128, "grad_norm": 1.0147572755813599, "learning_rate": 1.692664418212479e-05, "loss": 0.4768, "step": 2410 }, { "epoch": 0.10165061028311234, "grad_norm": 0.9247913360595703, "learning_rate": 1.6933670601461495e-05, "loss": 0.3698, "step": 2411 }, { "epoch": 0.10169277146531189, "grad_norm": 0.7860547304153442, "learning_rate": 1.6940697020798203e-05, "loss": 0.3855, "step": 2412 }, { "epoch": 0.10173493264751143, "grad_norm": 0.8830838799476624, "learning_rate": 1.6947723440134908e-05, "loss": 0.3116, "step": 2413 }, { "epoch": 0.10177709382971098, "grad_norm": 1.033447027206421, "learning_rate": 1.6954749859471612e-05, "loss": 0.2126, "step": 2414 }, { "epoch": 0.10181925501191054, "grad_norm": 0.6135001182556152, "learning_rate": 1.696177627880832e-05, "loss": 0.2003, "step": 2415 }, { "epoch": 0.10186141619411009, "grad_norm": 0.9255470633506775, "learning_rate": 1.6968802698145024e-05, "loss": 0.2343, "step": 2416 }, { "epoch": 0.10190357737630963, "grad_norm": 1.3368135690689087, "learning_rate": 1.6975829117481732e-05, "loss": 0.4921, "step": 2417 }, { "epoch": 0.10194573855850918, "grad_norm": 1.1323732137680054, "learning_rate": 1.6982855536818436e-05, "loss": 0.5072, "step": 2418 }, { "epoch": 0.10198789974070872, "grad_norm": 0.9282884001731873, "learning_rate": 1.6989881956155144e-05, "loss": 0.5694, "step": 2419 }, { "epoch": 0.10203006092290828, "grad_norm": 0.8508337140083313, "learning_rate": 1.699690837549185e-05, "loss": 0.218, "step": 2420 }, { "epoch": 0.10207222210510783, "grad_norm": 0.7769506573677063, "learning_rate": 1.7003934794828556e-05, "loss": 0.2851, "step": 2421 }, { "epoch": 0.10211438328730738, "grad_norm": 0.981890857219696, "learning_rate": 1.7010961214165264e-05, "loss": 0.7054, "step": 2422 }, { "epoch": 0.10215654446950692, "grad_norm": 0.7530070543289185, "learning_rate": 1.701798763350197e-05, "loss": 0.1377, "step": 2423 }, { "epoch": 0.10219870565170647, "grad_norm": 0.9538373947143555, "learning_rate": 1.7025014052838676e-05, "loss": 0.5316, "step": 2424 }, { "epoch": 0.10224086683390603, "grad_norm": 0.953637421131134, "learning_rate": 1.703204047217538e-05, "loss": 0.6829, "step": 2425 }, { "epoch": 0.10228302801610557, "grad_norm": 1.0755482912063599, "learning_rate": 1.703906689151209e-05, "loss": 0.3463, "step": 2426 }, { "epoch": 0.10232518919830512, "grad_norm": 0.8545163869857788, "learning_rate": 1.7046093310848793e-05, "loss": 0.7216, "step": 2427 }, { "epoch": 0.10236735038050467, "grad_norm": 0.8097062110900879, "learning_rate": 1.7053119730185497e-05, "loss": 0.2069, "step": 2428 }, { "epoch": 0.10240951156270421, "grad_norm": 0.994231104850769, "learning_rate": 1.7060146149522205e-05, "loss": 0.601, "step": 2429 }, { "epoch": 0.10245167274490377, "grad_norm": 0.7747762799263, "learning_rate": 1.706717256885891e-05, "loss": 0.3111, "step": 2430 }, { "epoch": 0.10249383392710332, "grad_norm": 1.5214967727661133, "learning_rate": 1.7074198988195617e-05, "loss": 0.2141, "step": 2431 }, { "epoch": 0.10253599510930286, "grad_norm": 0.8469321727752686, "learning_rate": 1.708122540753232e-05, "loss": 0.3452, "step": 2432 }, { "epoch": 0.10257815629150241, "grad_norm": 0.7013143301010132, "learning_rate": 1.708825182686903e-05, "loss": 0.1975, "step": 2433 }, { "epoch": 0.10262031747370197, "grad_norm": 0.8527201414108276, "learning_rate": 1.7095278246205734e-05, "loss": 0.3286, "step": 2434 }, { "epoch": 0.10266247865590152, "grad_norm": 0.9223656058311462, "learning_rate": 1.710230466554244e-05, "loss": 0.721, "step": 2435 }, { "epoch": 0.10270463983810106, "grad_norm": 1.0109885931015015, "learning_rate": 1.7109331084879146e-05, "loss": 0.2477, "step": 2436 }, { "epoch": 0.10274680102030061, "grad_norm": 1.054618239402771, "learning_rate": 1.711635750421585e-05, "loss": 0.3201, "step": 2437 }, { "epoch": 0.10278896220250015, "grad_norm": 2.4278500080108643, "learning_rate": 1.7123383923552558e-05, "loss": 0.4343, "step": 2438 }, { "epoch": 0.10283112338469971, "grad_norm": 0.9218294024467468, "learning_rate": 1.7130410342889263e-05, "loss": 0.2124, "step": 2439 }, { "epoch": 0.10287328456689926, "grad_norm": 0.7735694050788879, "learning_rate": 1.713743676222597e-05, "loss": 0.3092, "step": 2440 }, { "epoch": 0.1029154457490988, "grad_norm": 1.2836356163024902, "learning_rate": 1.7144463181562675e-05, "loss": 1.2283, "step": 2441 }, { "epoch": 0.10295760693129835, "grad_norm": 0.8546882271766663, "learning_rate": 1.7151489600899383e-05, "loss": 0.1488, "step": 2442 }, { "epoch": 0.1029997681134979, "grad_norm": 0.7361400723457336, "learning_rate": 1.7158516020236087e-05, "loss": 0.1387, "step": 2443 }, { "epoch": 0.10304192929569746, "grad_norm": 1.0505437850952148, "learning_rate": 1.7165542439572795e-05, "loss": 0.3581, "step": 2444 }, { "epoch": 0.103084090477897, "grad_norm": 0.8092806339263916, "learning_rate": 1.7172568858909503e-05, "loss": 0.246, "step": 2445 }, { "epoch": 0.10312625166009655, "grad_norm": 1.4263575077056885, "learning_rate": 1.7179595278246207e-05, "loss": 1.0201, "step": 2446 }, { "epoch": 0.1031684128422961, "grad_norm": 1.1299189329147339, "learning_rate": 1.7186621697582915e-05, "loss": 0.7727, "step": 2447 }, { "epoch": 0.10321057402449564, "grad_norm": 1.1862003803253174, "learning_rate": 1.719364811691962e-05, "loss": 0.4701, "step": 2448 }, { "epoch": 0.1032527352066952, "grad_norm": 0.820438802242279, "learning_rate": 1.7200674536256327e-05, "loss": 0.7795, "step": 2449 }, { "epoch": 0.10329489638889475, "grad_norm": 0.8482489585876465, "learning_rate": 1.720770095559303e-05, "loss": 0.5779, "step": 2450 }, { "epoch": 0.10333705757109429, "grad_norm": 0.9473509788513184, "learning_rate": 1.7214727374929736e-05, "loss": 0.1805, "step": 2451 }, { "epoch": 0.10337921875329384, "grad_norm": 1.3817484378814697, "learning_rate": 1.7221753794266443e-05, "loss": 0.469, "step": 2452 }, { "epoch": 0.10342137993549338, "grad_norm": 1.5863399505615234, "learning_rate": 1.7228780213603148e-05, "loss": 1.2516, "step": 2453 }, { "epoch": 0.10346354111769294, "grad_norm": 1.3637020587921143, "learning_rate": 1.7235806632939856e-05, "loss": 1.1022, "step": 2454 }, { "epoch": 0.10350570229989249, "grad_norm": 0.9470034837722778, "learning_rate": 1.724283305227656e-05, "loss": 0.1681, "step": 2455 }, { "epoch": 0.10354786348209204, "grad_norm": 1.157082438468933, "learning_rate": 1.7249859471613268e-05, "loss": 1.2297, "step": 2456 }, { "epoch": 0.10359002466429158, "grad_norm": 1.4927504062652588, "learning_rate": 1.7256885890949972e-05, "loss": 1.2392, "step": 2457 }, { "epoch": 0.10363218584649113, "grad_norm": 0.9827793836593628, "learning_rate": 1.726391231028668e-05, "loss": 0.3403, "step": 2458 }, { "epoch": 0.10367434702869069, "grad_norm": 0.6441689729690552, "learning_rate": 1.7270938729623384e-05, "loss": 0.1068, "step": 2459 }, { "epoch": 0.10371650821089023, "grad_norm": 0.6090232133865356, "learning_rate": 1.727796514896009e-05, "loss": 0.1706, "step": 2460 }, { "epoch": 0.10375866939308978, "grad_norm": 1.6544067859649658, "learning_rate": 1.7284991568296797e-05, "loss": 1.2911, "step": 2461 }, { "epoch": 0.10380083057528933, "grad_norm": 1.5696126222610474, "learning_rate": 1.72920179876335e-05, "loss": 1.0798, "step": 2462 }, { "epoch": 0.10384299175748889, "grad_norm": 1.0160222053527832, "learning_rate": 1.729904440697021e-05, "loss": 0.231, "step": 2463 }, { "epoch": 0.10388515293968843, "grad_norm": 0.6313055157661438, "learning_rate": 1.7306070826306913e-05, "loss": 0.1502, "step": 2464 }, { "epoch": 0.10392731412188798, "grad_norm": 1.4705520868301392, "learning_rate": 1.731309724564362e-05, "loss": 1.2635, "step": 2465 }, { "epoch": 0.10396947530408752, "grad_norm": 0.9923672676086426, "learning_rate": 1.7320123664980325e-05, "loss": 0.2153, "step": 2466 }, { "epoch": 0.10401163648628707, "grad_norm": 2.096285104751587, "learning_rate": 1.7327150084317033e-05, "loss": 0.6934, "step": 2467 }, { "epoch": 0.10405379766848663, "grad_norm": 1.6444653272628784, "learning_rate": 1.733417650365374e-05, "loss": 0.3189, "step": 2468 }, { "epoch": 0.10409595885068618, "grad_norm": 0.8765588998794556, "learning_rate": 1.7341202922990445e-05, "loss": 0.6324, "step": 2469 }, { "epoch": 0.10413812003288572, "grad_norm": 0.8265400528907776, "learning_rate": 1.7348229342327153e-05, "loss": 0.0913, "step": 2470 }, { "epoch": 0.10418028121508527, "grad_norm": 1.248733639717102, "learning_rate": 1.7355255761663857e-05, "loss": 0.33, "step": 2471 }, { "epoch": 0.10422244239728481, "grad_norm": 0.6604008674621582, "learning_rate": 1.7362282181000565e-05, "loss": 0.2694, "step": 2472 }, { "epoch": 0.10426460357948437, "grad_norm": 2.5607872009277344, "learning_rate": 1.736930860033727e-05, "loss": 1.0602, "step": 2473 }, { "epoch": 0.10430676476168392, "grad_norm": 0.7341168522834778, "learning_rate": 1.7376335019673974e-05, "loss": 0.1949, "step": 2474 }, { "epoch": 0.10434892594388347, "grad_norm": 1.4427802562713623, "learning_rate": 1.7383361439010682e-05, "loss": 0.3341, "step": 2475 }, { "epoch": 0.10439108712608301, "grad_norm": 1.4933524131774902, "learning_rate": 1.7390387858347386e-05, "loss": 1.1716, "step": 2476 }, { "epoch": 0.10443324830828256, "grad_norm": 1.7591105699539185, "learning_rate": 1.7397414277684094e-05, "loss": 0.3211, "step": 2477 }, { "epoch": 0.10447540949048212, "grad_norm": 0.9064940214157104, "learning_rate": 1.74044406970208e-05, "loss": 0.1743, "step": 2478 }, { "epoch": 0.10451757067268166, "grad_norm": 0.8642042875289917, "learning_rate": 1.7411467116357506e-05, "loss": 0.2169, "step": 2479 }, { "epoch": 0.10455973185488121, "grad_norm": 1.3095282316207886, "learning_rate": 1.741849353569421e-05, "loss": 0.3817, "step": 2480 }, { "epoch": 0.10460189303708076, "grad_norm": 0.9330442547798157, "learning_rate": 1.742551995503092e-05, "loss": 0.233, "step": 2481 }, { "epoch": 0.1046440542192803, "grad_norm": 0.7832114100456238, "learning_rate": 1.7432546374367623e-05, "loss": 0.2511, "step": 2482 }, { "epoch": 0.10468621540147986, "grad_norm": 0.8707437515258789, "learning_rate": 1.7439572793704327e-05, "loss": 0.1956, "step": 2483 }, { "epoch": 0.10472837658367941, "grad_norm": 1.1907941102981567, "learning_rate": 1.7446599213041035e-05, "loss": 0.2512, "step": 2484 }, { "epoch": 0.10477053776587895, "grad_norm": 1.1277318000793457, "learning_rate": 1.745362563237774e-05, "loss": 0.5694, "step": 2485 }, { "epoch": 0.1048126989480785, "grad_norm": 0.7374443411827087, "learning_rate": 1.7460652051714447e-05, "loss": 0.1629, "step": 2486 }, { "epoch": 0.10485486013027805, "grad_norm": 0.5585341453552246, "learning_rate": 1.746767847105115e-05, "loss": 0.1864, "step": 2487 }, { "epoch": 0.1048970213124776, "grad_norm": 0.5925265550613403, "learning_rate": 1.747470489038786e-05, "loss": 0.1765, "step": 2488 }, { "epoch": 0.10493918249467715, "grad_norm": 2.812065362930298, "learning_rate": 1.7481731309724564e-05, "loss": 0.5861, "step": 2489 }, { "epoch": 0.1049813436768767, "grad_norm": 0.8433333039283752, "learning_rate": 1.748875772906127e-05, "loss": 0.1401, "step": 2490 }, { "epoch": 0.10502350485907624, "grad_norm": 1.1911072731018066, "learning_rate": 1.749578414839798e-05, "loss": 0.2283, "step": 2491 }, { "epoch": 0.1050656660412758, "grad_norm": 1.064063310623169, "learning_rate": 1.7502810567734684e-05, "loss": 0.3295, "step": 2492 }, { "epoch": 0.10510782722347535, "grad_norm": 0.8578768968582153, "learning_rate": 1.750983698707139e-05, "loss": 0.3133, "step": 2493 }, { "epoch": 0.1051499884056749, "grad_norm": 1.2523365020751953, "learning_rate": 1.7516863406408096e-05, "loss": 0.6669, "step": 2494 }, { "epoch": 0.10519214958787444, "grad_norm": 1.2814781665802002, "learning_rate": 1.7523889825744804e-05, "loss": 0.6888, "step": 2495 }, { "epoch": 0.10523431077007399, "grad_norm": 1.5257867574691772, "learning_rate": 1.7530916245081508e-05, "loss": 0.3436, "step": 2496 }, { "epoch": 0.10527647195227355, "grad_norm": 1.154279351234436, "learning_rate": 1.7537942664418212e-05, "loss": 0.2364, "step": 2497 }, { "epoch": 0.1053186331344731, "grad_norm": 2.2375130653381348, "learning_rate": 1.754496908375492e-05, "loss": 0.6474, "step": 2498 }, { "epoch": 0.10536079431667264, "grad_norm": 1.432780146598816, "learning_rate": 1.7551995503091625e-05, "loss": 0.7024, "step": 2499 }, { "epoch": 0.10540295549887219, "grad_norm": 1.012740135192871, "learning_rate": 1.7559021922428332e-05, "loss": 0.2421, "step": 2500 }, { "epoch": 0.10544511668107173, "grad_norm": 1.2477179765701294, "learning_rate": 1.7566048341765037e-05, "loss": 0.1542, "step": 2501 }, { "epoch": 0.10548727786327129, "grad_norm": 1.2294970750808716, "learning_rate": 1.7573074761101745e-05, "loss": 0.2527, "step": 2502 }, { "epoch": 0.10552943904547084, "grad_norm": 4.059110164642334, "learning_rate": 1.758010118043845e-05, "loss": 0.9563, "step": 2503 }, { "epoch": 0.10557160022767038, "grad_norm": 0.7264147996902466, "learning_rate": 1.7587127599775157e-05, "loss": 0.1448, "step": 2504 }, { "epoch": 0.10561376140986993, "grad_norm": 0.7198948860168457, "learning_rate": 1.759415401911186e-05, "loss": 0.1664, "step": 2505 }, { "epoch": 0.10565592259206948, "grad_norm": 1.451196312904358, "learning_rate": 1.7601180438448565e-05, "loss": 0.4458, "step": 2506 }, { "epoch": 0.10569808377426904, "grad_norm": 2.7396721839904785, "learning_rate": 1.7608206857785273e-05, "loss": 0.7228, "step": 2507 }, { "epoch": 0.10574024495646858, "grad_norm": 1.4336538314819336, "learning_rate": 1.7615233277121978e-05, "loss": 0.3985, "step": 2508 }, { "epoch": 0.10578240613866813, "grad_norm": 0.7141585946083069, "learning_rate": 1.7622259696458685e-05, "loss": 0.2077, "step": 2509 }, { "epoch": 0.10582456732086767, "grad_norm": 0.9648429751396179, "learning_rate": 1.762928611579539e-05, "loss": 0.192, "step": 2510 }, { "epoch": 0.10586672850306722, "grad_norm": 1.4635510444641113, "learning_rate": 1.7636312535132098e-05, "loss": 1.0962, "step": 2511 }, { "epoch": 0.10590888968526678, "grad_norm": 1.4602805376052856, "learning_rate": 1.7643338954468805e-05, "loss": 1.2259, "step": 2512 }, { "epoch": 0.10595105086746633, "grad_norm": 0.9179686903953552, "learning_rate": 1.765036537380551e-05, "loss": 0.2448, "step": 2513 }, { "epoch": 0.10599321204966587, "grad_norm": 1.3996691703796387, "learning_rate": 1.7657391793142218e-05, "loss": 0.9017, "step": 2514 }, { "epoch": 0.10603537323186542, "grad_norm": 1.1351149082183838, "learning_rate": 1.7664418212478922e-05, "loss": 0.4046, "step": 2515 }, { "epoch": 0.10607753441406496, "grad_norm": 1.135591983795166, "learning_rate": 1.767144463181563e-05, "loss": 0.2159, "step": 2516 }, { "epoch": 0.10611969559626452, "grad_norm": 1.2728115320205688, "learning_rate": 1.7678471051152334e-05, "loss": 0.3408, "step": 2517 }, { "epoch": 0.10616185677846407, "grad_norm": 0.9766765832901001, "learning_rate": 1.7685497470489042e-05, "loss": 0.342, "step": 2518 }, { "epoch": 0.10620401796066362, "grad_norm": 1.3713276386260986, "learning_rate": 1.7692523889825746e-05, "loss": 0.2843, "step": 2519 }, { "epoch": 0.10624617914286316, "grad_norm": 1.107482671737671, "learning_rate": 1.769955030916245e-05, "loss": 0.2771, "step": 2520 }, { "epoch": 0.10628834032506271, "grad_norm": 0.738960325717926, "learning_rate": 1.770657672849916e-05, "loss": 0.1627, "step": 2521 }, { "epoch": 0.10633050150726227, "grad_norm": 0.9950055480003357, "learning_rate": 1.7713603147835863e-05, "loss": 0.5847, "step": 2522 }, { "epoch": 0.10637266268946181, "grad_norm": 0.8460886478424072, "learning_rate": 1.772062956717257e-05, "loss": 0.2281, "step": 2523 }, { "epoch": 0.10641482387166136, "grad_norm": 1.1689327955245972, "learning_rate": 1.7727655986509275e-05, "loss": 0.8895, "step": 2524 }, { "epoch": 0.1064569850538609, "grad_norm": 1.4158886671066284, "learning_rate": 1.7734682405845983e-05, "loss": 0.3259, "step": 2525 }, { "epoch": 0.10649914623606047, "grad_norm": 1.0649921894073486, "learning_rate": 1.7741708825182687e-05, "loss": 0.1856, "step": 2526 }, { "epoch": 0.10654130741826001, "grad_norm": 2.5687644481658936, "learning_rate": 1.774873524451939e-05, "loss": 1.4219, "step": 2527 }, { "epoch": 0.10658346860045956, "grad_norm": 1.2791301012039185, "learning_rate": 1.77557616638561e-05, "loss": 0.4393, "step": 2528 }, { "epoch": 0.1066256297826591, "grad_norm": 0.7865915298461914, "learning_rate": 1.7762788083192804e-05, "loss": 0.165, "step": 2529 }, { "epoch": 0.10666779096485865, "grad_norm": 1.4292093515396118, "learning_rate": 1.776981450252951e-05, "loss": 1.2491, "step": 2530 }, { "epoch": 0.10670995214705821, "grad_norm": 0.862443208694458, "learning_rate": 1.7776840921866216e-05, "loss": 0.1539, "step": 2531 }, { "epoch": 0.10675211332925776, "grad_norm": 1.3795462846755981, "learning_rate": 1.7783867341202924e-05, "loss": 0.4437, "step": 2532 }, { "epoch": 0.1067942745114573, "grad_norm": 1.5705478191375732, "learning_rate": 1.7790893760539628e-05, "loss": 0.2418, "step": 2533 }, { "epoch": 0.10683643569365685, "grad_norm": 0.84236079454422, "learning_rate": 1.7797920179876336e-05, "loss": 0.3756, "step": 2534 }, { "epoch": 0.1068785968758564, "grad_norm": 0.7653487324714661, "learning_rate": 1.7804946599213044e-05, "loss": 0.2305, "step": 2535 }, { "epoch": 0.10692075805805595, "grad_norm": 1.109588384628296, "learning_rate": 1.7811973018549748e-05, "loss": 0.3038, "step": 2536 }, { "epoch": 0.1069629192402555, "grad_norm": 0.9823553562164307, "learning_rate": 1.7818999437886456e-05, "loss": 0.7957, "step": 2537 }, { "epoch": 0.10700508042245505, "grad_norm": 0.7647210359573364, "learning_rate": 1.782602585722316e-05, "loss": 0.5791, "step": 2538 }, { "epoch": 0.10704724160465459, "grad_norm": 1.584206223487854, "learning_rate": 1.7833052276559868e-05, "loss": 1.0746, "step": 2539 }, { "epoch": 0.10708940278685414, "grad_norm": 0.9900733232498169, "learning_rate": 1.7840078695896573e-05, "loss": 0.7232, "step": 2540 }, { "epoch": 0.1071315639690537, "grad_norm": 1.128732681274414, "learning_rate": 1.7847105115233277e-05, "loss": 0.7159, "step": 2541 }, { "epoch": 0.10717372515125324, "grad_norm": 1.0008058547973633, "learning_rate": 1.7854131534569985e-05, "loss": 0.2647, "step": 2542 }, { "epoch": 0.10721588633345279, "grad_norm": 1.338810920715332, "learning_rate": 1.786115795390669e-05, "loss": 0.5595, "step": 2543 }, { "epoch": 0.10725804751565234, "grad_norm": 0.7490121126174927, "learning_rate": 1.7868184373243397e-05, "loss": 0.3766, "step": 2544 }, { "epoch": 0.10730020869785188, "grad_norm": 2.0477044582366943, "learning_rate": 1.78752107925801e-05, "loss": 0.1504, "step": 2545 }, { "epoch": 0.10734236988005144, "grad_norm": 0.6299880743026733, "learning_rate": 1.788223721191681e-05, "loss": 0.247, "step": 2546 }, { "epoch": 0.10738453106225099, "grad_norm": 1.2438054084777832, "learning_rate": 1.7889263631253513e-05, "loss": 0.6004, "step": 2547 }, { "epoch": 0.10742669224445053, "grad_norm": 0.5492790937423706, "learning_rate": 1.789629005059022e-05, "loss": 0.1488, "step": 2548 }, { "epoch": 0.10746885342665008, "grad_norm": 0.6840423941612244, "learning_rate": 1.7903316469926926e-05, "loss": 0.1944, "step": 2549 }, { "epoch": 0.10751101460884963, "grad_norm": 0.5851161479949951, "learning_rate": 1.791034288926363e-05, "loss": 0.1469, "step": 2550 }, { "epoch": 0.10755317579104919, "grad_norm": 1.058065414428711, "learning_rate": 1.7917369308600338e-05, "loss": 0.6684, "step": 2551 }, { "epoch": 0.10759533697324873, "grad_norm": 1.0738718509674072, "learning_rate": 1.7924395727937042e-05, "loss": 0.2698, "step": 2552 }, { "epoch": 0.10763749815544828, "grad_norm": 0.7968273758888245, "learning_rate": 1.793142214727375e-05, "loss": 0.3036, "step": 2553 }, { "epoch": 0.10767965933764782, "grad_norm": 0.5332055687904358, "learning_rate": 1.7938448566610454e-05, "loss": 0.1747, "step": 2554 }, { "epoch": 0.10772182051984738, "grad_norm": 1.006462812423706, "learning_rate": 1.7945474985947162e-05, "loss": 0.3072, "step": 2555 }, { "epoch": 0.10776398170204693, "grad_norm": 0.7711588144302368, "learning_rate": 1.7952501405283867e-05, "loss": 0.3497, "step": 2556 }, { "epoch": 0.10780614288424648, "grad_norm": 0.6522648930549622, "learning_rate": 1.7959527824620574e-05, "loss": 0.2541, "step": 2557 }, { "epoch": 0.10784830406644602, "grad_norm": 0.8570578098297119, "learning_rate": 1.7966554243957282e-05, "loss": 0.7033, "step": 2558 }, { "epoch": 0.10789046524864557, "grad_norm": 0.6947705745697021, "learning_rate": 1.7973580663293987e-05, "loss": 0.1687, "step": 2559 }, { "epoch": 0.10793262643084513, "grad_norm": 1.0415416955947876, "learning_rate": 1.7980607082630694e-05, "loss": 0.3201, "step": 2560 }, { "epoch": 0.10797478761304467, "grad_norm": 0.5365840792655945, "learning_rate": 1.79876335019674e-05, "loss": 0.1574, "step": 2561 }, { "epoch": 0.10801694879524422, "grad_norm": 1.0354233980178833, "learning_rate": 1.7994659921304106e-05, "loss": 0.7646, "step": 2562 }, { "epoch": 0.10805910997744377, "grad_norm": 1.1811258792877197, "learning_rate": 1.800168634064081e-05, "loss": 0.5504, "step": 2563 }, { "epoch": 0.10810127115964331, "grad_norm": 1.0977275371551514, "learning_rate": 1.8008712759977515e-05, "loss": 0.5112, "step": 2564 }, { "epoch": 0.10814343234184287, "grad_norm": 1.0051047801971436, "learning_rate": 1.8015739179314223e-05, "loss": 0.2812, "step": 2565 }, { "epoch": 0.10818559352404242, "grad_norm": 0.7551231980323792, "learning_rate": 1.8022765598650927e-05, "loss": 0.3531, "step": 2566 }, { "epoch": 0.10822775470624196, "grad_norm": 0.7775444984436035, "learning_rate": 1.8029792017987635e-05, "loss": 0.198, "step": 2567 }, { "epoch": 0.10826991588844151, "grad_norm": 1.3468149900436401, "learning_rate": 1.803681843732434e-05, "loss": 1.0916, "step": 2568 }, { "epoch": 0.10831207707064106, "grad_norm": 0.9474807381629944, "learning_rate": 1.8043844856661047e-05, "loss": 0.895, "step": 2569 }, { "epoch": 0.10835423825284062, "grad_norm": 2.1095430850982666, "learning_rate": 1.8050871275997752e-05, "loss": 0.181, "step": 2570 }, { "epoch": 0.10839639943504016, "grad_norm": 2.2787933349609375, "learning_rate": 1.805789769533446e-05, "loss": 0.6691, "step": 2571 }, { "epoch": 0.10843856061723971, "grad_norm": 2.109574317932129, "learning_rate": 1.8064924114671164e-05, "loss": 1.2463, "step": 2572 }, { "epoch": 0.10848072179943925, "grad_norm": 0.9995289444923401, "learning_rate": 1.807195053400787e-05, "loss": 0.3151, "step": 2573 }, { "epoch": 0.1085228829816388, "grad_norm": 1.1209481954574585, "learning_rate": 1.8078976953344576e-05, "loss": 0.2789, "step": 2574 }, { "epoch": 0.10856504416383836, "grad_norm": 0.9524023532867432, "learning_rate": 1.808600337268128e-05, "loss": 0.3772, "step": 2575 }, { "epoch": 0.1086072053460379, "grad_norm": 0.8690701723098755, "learning_rate": 1.809302979201799e-05, "loss": 0.7555, "step": 2576 }, { "epoch": 0.10864936652823745, "grad_norm": 0.9481076598167419, "learning_rate": 1.8100056211354693e-05, "loss": 0.1528, "step": 2577 }, { "epoch": 0.108691527710437, "grad_norm": 0.6951153874397278, "learning_rate": 1.81070826306914e-05, "loss": 0.3093, "step": 2578 }, { "epoch": 0.10873368889263654, "grad_norm": 1.1604387760162354, "learning_rate": 1.8114109050028105e-05, "loss": 0.3419, "step": 2579 }, { "epoch": 0.1087758500748361, "grad_norm": 0.9064311981201172, "learning_rate": 1.8121135469364813e-05, "loss": 0.7672, "step": 2580 }, { "epoch": 0.10881801125703565, "grad_norm": 0.9588311910629272, "learning_rate": 1.812816188870152e-05, "loss": 0.6326, "step": 2581 }, { "epoch": 0.1088601724392352, "grad_norm": 0.7354428768157959, "learning_rate": 1.8135188308038225e-05, "loss": 0.3248, "step": 2582 }, { "epoch": 0.10890233362143474, "grad_norm": 0.9855530261993408, "learning_rate": 1.8142214727374933e-05, "loss": 0.2848, "step": 2583 }, { "epoch": 0.1089444948036343, "grad_norm": 0.8451952338218689, "learning_rate": 1.8149241146711637e-05, "loss": 0.1828, "step": 2584 }, { "epoch": 0.10898665598583385, "grad_norm": 1.0160866975784302, "learning_rate": 1.8156267566048345e-05, "loss": 0.6456, "step": 2585 }, { "epoch": 0.10902881716803339, "grad_norm": 0.7452335953712463, "learning_rate": 1.816329398538505e-05, "loss": 0.1821, "step": 2586 }, { "epoch": 0.10907097835023294, "grad_norm": 1.093971610069275, "learning_rate": 1.8170320404721754e-05, "loss": 0.3213, "step": 2587 }, { "epoch": 0.10911313953243248, "grad_norm": 1.2797037363052368, "learning_rate": 1.817734682405846e-05, "loss": 0.7513, "step": 2588 }, { "epoch": 0.10915530071463204, "grad_norm": 0.7587445378303528, "learning_rate": 1.8184373243395166e-05, "loss": 0.1912, "step": 2589 }, { "epoch": 0.10919746189683159, "grad_norm": 1.2084980010986328, "learning_rate": 1.8191399662731874e-05, "loss": 0.6496, "step": 2590 }, { "epoch": 0.10923962307903114, "grad_norm": 0.7616434097290039, "learning_rate": 1.8198426082068578e-05, "loss": 0.2323, "step": 2591 }, { "epoch": 0.10928178426123068, "grad_norm": 0.6373199224472046, "learning_rate": 1.8205452501405286e-05, "loss": 0.1695, "step": 2592 }, { "epoch": 0.10932394544343023, "grad_norm": 0.7103708386421204, "learning_rate": 1.821247892074199e-05, "loss": 0.3515, "step": 2593 }, { "epoch": 0.10936610662562979, "grad_norm": 1.634270429611206, "learning_rate": 1.8219505340078698e-05, "loss": 0.9962, "step": 2594 }, { "epoch": 0.10940826780782933, "grad_norm": 0.9365391731262207, "learning_rate": 1.8226531759415402e-05, "loss": 0.3295, "step": 2595 }, { "epoch": 0.10945042899002888, "grad_norm": 2.1446127891540527, "learning_rate": 1.8233558178752107e-05, "loss": 0.1256, "step": 2596 }, { "epoch": 0.10949259017222843, "grad_norm": 0.6591174006462097, "learning_rate": 1.8240584598088815e-05, "loss": 0.137, "step": 2597 }, { "epoch": 0.10953475135442797, "grad_norm": 2.9394919872283936, "learning_rate": 1.824761101742552e-05, "loss": 1.3014, "step": 2598 }, { "epoch": 0.10957691253662753, "grad_norm": 0.6058825254440308, "learning_rate": 1.8254637436762227e-05, "loss": 0.1079, "step": 2599 }, { "epoch": 0.10961907371882708, "grad_norm": 1.1440714597702026, "learning_rate": 1.826166385609893e-05, "loss": 0.4342, "step": 2600 }, { "epoch": 0.10966123490102662, "grad_norm": 1.1213678121566772, "learning_rate": 1.826869027543564e-05, "loss": 0.5294, "step": 2601 }, { "epoch": 0.10970339608322617, "grad_norm": 1.1105539798736572, "learning_rate": 1.8275716694772347e-05, "loss": 0.6444, "step": 2602 }, { "epoch": 0.10974555726542572, "grad_norm": 1.0976812839508057, "learning_rate": 1.828274311410905e-05, "loss": 0.7437, "step": 2603 }, { "epoch": 0.10978771844762528, "grad_norm": 0.77121901512146, "learning_rate": 1.828976953344576e-05, "loss": 0.5409, "step": 2604 }, { "epoch": 0.10982987962982482, "grad_norm": 1.2420011758804321, "learning_rate": 1.8296795952782463e-05, "loss": 0.3864, "step": 2605 }, { "epoch": 0.10987204081202437, "grad_norm": 0.9973605871200562, "learning_rate": 1.830382237211917e-05, "loss": 0.8524, "step": 2606 }, { "epoch": 0.10991420199422391, "grad_norm": 0.7933306694030762, "learning_rate": 1.8310848791455875e-05, "loss": 0.3499, "step": 2607 }, { "epoch": 0.10995636317642346, "grad_norm": 1.2275267839431763, "learning_rate": 1.8317875210792583e-05, "loss": 0.4793, "step": 2608 }, { "epoch": 0.10999852435862302, "grad_norm": 0.7563764452934265, "learning_rate": 1.8324901630129288e-05, "loss": 0.137, "step": 2609 }, { "epoch": 0.11004068554082257, "grad_norm": 1.191224455833435, "learning_rate": 1.8331928049465992e-05, "loss": 0.3602, "step": 2610 }, { "epoch": 0.11008284672302211, "grad_norm": 1.0267767906188965, "learning_rate": 1.83389544688027e-05, "loss": 0.8653, "step": 2611 }, { "epoch": 0.11012500790522166, "grad_norm": 1.0612763166427612, "learning_rate": 1.8345980888139404e-05, "loss": 0.5846, "step": 2612 }, { "epoch": 0.1101671690874212, "grad_norm": 0.6991497278213501, "learning_rate": 1.8353007307476112e-05, "loss": 0.1689, "step": 2613 }, { "epoch": 0.11020933026962076, "grad_norm": 0.7719089984893799, "learning_rate": 1.8360033726812816e-05, "loss": 0.1512, "step": 2614 }, { "epoch": 0.11025149145182031, "grad_norm": 0.8459004163742065, "learning_rate": 1.8367060146149524e-05, "loss": 0.314, "step": 2615 }, { "epoch": 0.11029365263401986, "grad_norm": 2.9382152557373047, "learning_rate": 1.837408656548623e-05, "loss": 0.4013, "step": 2616 }, { "epoch": 0.1103358138162194, "grad_norm": 0.49864187836647034, "learning_rate": 1.8381112984822936e-05, "loss": 0.2416, "step": 2617 }, { "epoch": 0.11037797499841896, "grad_norm": 1.0883930921554565, "learning_rate": 1.838813940415964e-05, "loss": 0.4373, "step": 2618 }, { "epoch": 0.11042013618061851, "grad_norm": 0.8240532875061035, "learning_rate": 1.8395165823496345e-05, "loss": 0.661, "step": 2619 }, { "epoch": 0.11046229736281805, "grad_norm": 1.159570336341858, "learning_rate": 1.8402192242833053e-05, "loss": 0.3675, "step": 2620 }, { "epoch": 0.1105044585450176, "grad_norm": 1.2543586492538452, "learning_rate": 1.8409218662169757e-05, "loss": 0.323, "step": 2621 }, { "epoch": 0.11054661972721715, "grad_norm": 0.8568741679191589, "learning_rate": 1.8416245081506465e-05, "loss": 0.2358, "step": 2622 }, { "epoch": 0.1105887809094167, "grad_norm": 0.7009987235069275, "learning_rate": 1.842327150084317e-05, "loss": 0.2254, "step": 2623 }, { "epoch": 0.11063094209161625, "grad_norm": 0.9569449424743652, "learning_rate": 1.8430297920179877e-05, "loss": 0.7121, "step": 2624 }, { "epoch": 0.1106731032738158, "grad_norm": 0.590858519077301, "learning_rate": 1.8437324339516585e-05, "loss": 0.1635, "step": 2625 }, { "epoch": 0.11071526445601534, "grad_norm": 0.6093842387199402, "learning_rate": 1.844435075885329e-05, "loss": 0.1821, "step": 2626 }, { "epoch": 0.11075742563821489, "grad_norm": 0.7224240899085999, "learning_rate": 1.8451377178189997e-05, "loss": 0.5163, "step": 2627 }, { "epoch": 0.11079958682041445, "grad_norm": 0.7770045399665833, "learning_rate": 1.84584035975267e-05, "loss": 0.3508, "step": 2628 }, { "epoch": 0.110841748002614, "grad_norm": 3.277998208999634, "learning_rate": 1.846543001686341e-05, "loss": 0.5936, "step": 2629 }, { "epoch": 0.11088390918481354, "grad_norm": 1.0021333694458008, "learning_rate": 1.8472456436200114e-05, "loss": 0.6203, "step": 2630 }, { "epoch": 0.11092607036701309, "grad_norm": 1.1785651445388794, "learning_rate": 1.847948285553682e-05, "loss": 0.2861, "step": 2631 }, { "epoch": 0.11096823154921263, "grad_norm": 1.096173644065857, "learning_rate": 1.8486509274873526e-05, "loss": 0.2206, "step": 2632 }, { "epoch": 0.1110103927314122, "grad_norm": 1.2152799367904663, "learning_rate": 1.849353569421023e-05, "loss": 0.312, "step": 2633 }, { "epoch": 0.11105255391361174, "grad_norm": 0.9438897967338562, "learning_rate": 1.8500562113546938e-05, "loss": 0.2169, "step": 2634 }, { "epoch": 0.11109471509581129, "grad_norm": 1.0920368432998657, "learning_rate": 1.8507588532883643e-05, "loss": 0.5925, "step": 2635 }, { "epoch": 0.11113687627801083, "grad_norm": 0.5903985500335693, "learning_rate": 1.851461495222035e-05, "loss": 0.2057, "step": 2636 }, { "epoch": 0.11117903746021038, "grad_norm": 0.6884585618972778, "learning_rate": 1.8521641371557055e-05, "loss": 0.3069, "step": 2637 }, { "epoch": 0.11122119864240994, "grad_norm": 0.9754734039306641, "learning_rate": 1.8528667790893762e-05, "loss": 0.9354, "step": 2638 }, { "epoch": 0.11126335982460948, "grad_norm": 0.7426806092262268, "learning_rate": 1.8535694210230467e-05, "loss": 0.4745, "step": 2639 }, { "epoch": 0.11130552100680903, "grad_norm": 0.6525565385818481, "learning_rate": 1.854272062956717e-05, "loss": 0.1066, "step": 2640 }, { "epoch": 0.11134768218900858, "grad_norm": 0.8829623460769653, "learning_rate": 1.854974704890388e-05, "loss": 0.1815, "step": 2641 }, { "epoch": 0.11138984337120812, "grad_norm": 0.7645938992500305, "learning_rate": 1.8556773468240583e-05, "loss": 0.1557, "step": 2642 }, { "epoch": 0.11143200455340768, "grad_norm": 1.1097608804702759, "learning_rate": 1.856379988757729e-05, "loss": 0.3891, "step": 2643 }, { "epoch": 0.11147416573560723, "grad_norm": 1.0978198051452637, "learning_rate": 1.8570826306913996e-05, "loss": 0.6528, "step": 2644 }, { "epoch": 0.11151632691780677, "grad_norm": 0.8170788288116455, "learning_rate": 1.8577852726250703e-05, "loss": 0.5603, "step": 2645 }, { "epoch": 0.11155848810000632, "grad_norm": 0.5007385015487671, "learning_rate": 1.8584879145587408e-05, "loss": 0.1653, "step": 2646 }, { "epoch": 0.11160064928220588, "grad_norm": 0.6196160912513733, "learning_rate": 1.8591905564924116e-05, "loss": 0.2069, "step": 2647 }, { "epoch": 0.11164281046440543, "grad_norm": 1.406990647315979, "learning_rate": 1.8598931984260823e-05, "loss": 1.3222, "step": 2648 }, { "epoch": 0.11168497164660497, "grad_norm": 1.1214677095413208, "learning_rate": 1.8605958403597528e-05, "loss": 0.2413, "step": 2649 }, { "epoch": 0.11172713282880452, "grad_norm": 1.0301134586334229, "learning_rate": 1.8612984822934236e-05, "loss": 0.9271, "step": 2650 }, { "epoch": 0.11176929401100406, "grad_norm": 1.0248744487762451, "learning_rate": 1.862001124227094e-05, "loss": 0.2885, "step": 2651 }, { "epoch": 0.11181145519320362, "grad_norm": 0.8110528588294983, "learning_rate": 1.8627037661607648e-05, "loss": 0.2039, "step": 2652 }, { "epoch": 0.11185361637540317, "grad_norm": 0.7765677571296692, "learning_rate": 1.8634064080944352e-05, "loss": 0.4889, "step": 2653 }, { "epoch": 0.11189577755760272, "grad_norm": 0.5443686246871948, "learning_rate": 1.8641090500281057e-05, "loss": 0.183, "step": 2654 }, { "epoch": 0.11193793873980226, "grad_norm": 1.0717588663101196, "learning_rate": 1.8648116919617764e-05, "loss": 0.3166, "step": 2655 }, { "epoch": 0.11198009992200181, "grad_norm": 2.1639456748962402, "learning_rate": 1.865514333895447e-05, "loss": 0.5954, "step": 2656 }, { "epoch": 0.11202226110420137, "grad_norm": 1.5671130418777466, "learning_rate": 1.8662169758291176e-05, "loss": 1.3616, "step": 2657 }, { "epoch": 0.11206442228640091, "grad_norm": 0.9292743802070618, "learning_rate": 1.866919617762788e-05, "loss": 0.1801, "step": 2658 }, { "epoch": 0.11210658346860046, "grad_norm": 1.0754090547561646, "learning_rate": 1.867622259696459e-05, "loss": 0.3907, "step": 2659 }, { "epoch": 0.1121487446508, "grad_norm": 0.7659105658531189, "learning_rate": 1.8683249016301293e-05, "loss": 0.6648, "step": 2660 }, { "epoch": 0.11219090583299955, "grad_norm": 1.9424735307693481, "learning_rate": 1.8690275435638e-05, "loss": 1.329, "step": 2661 }, { "epoch": 0.11223306701519911, "grad_norm": 0.9386926293373108, "learning_rate": 1.8697301854974705e-05, "loss": 0.2756, "step": 2662 }, { "epoch": 0.11227522819739866, "grad_norm": 0.6570949554443359, "learning_rate": 1.870432827431141e-05, "loss": 0.18, "step": 2663 }, { "epoch": 0.1123173893795982, "grad_norm": 0.6409682631492615, "learning_rate": 1.8711354693648117e-05, "loss": 0.1179, "step": 2664 }, { "epoch": 0.11235955056179775, "grad_norm": 2.3741886615753174, "learning_rate": 1.8718381112984822e-05, "loss": 0.6801, "step": 2665 }, { "epoch": 0.1124017117439973, "grad_norm": 1.3216506242752075, "learning_rate": 1.872540753232153e-05, "loss": 1.1625, "step": 2666 }, { "epoch": 0.11244387292619686, "grad_norm": 0.90913987159729, "learning_rate": 1.8732433951658234e-05, "loss": 0.3174, "step": 2667 }, { "epoch": 0.1124860341083964, "grad_norm": 0.9970049262046814, "learning_rate": 1.8739460370994942e-05, "loss": 0.5687, "step": 2668 }, { "epoch": 0.11252819529059595, "grad_norm": 1.6183234453201294, "learning_rate": 1.8746486790331646e-05, "loss": 0.449, "step": 2669 }, { "epoch": 0.1125703564727955, "grad_norm": 0.8521263003349304, "learning_rate": 1.8753513209668354e-05, "loss": 0.121, "step": 2670 }, { "epoch": 0.11261251765499504, "grad_norm": 0.8255015015602112, "learning_rate": 1.8760539629005062e-05, "loss": 0.1347, "step": 2671 }, { "epoch": 0.1126546788371946, "grad_norm": 1.2258565425872803, "learning_rate": 1.8767566048341766e-05, "loss": 0.3005, "step": 2672 }, { "epoch": 0.11269684001939415, "grad_norm": 1.8638845682144165, "learning_rate": 1.8774592467678474e-05, "loss": 1.3221, "step": 2673 }, { "epoch": 0.11273900120159369, "grad_norm": 0.7446231842041016, "learning_rate": 1.8781618887015178e-05, "loss": 0.2215, "step": 2674 }, { "epoch": 0.11278116238379324, "grad_norm": 0.7435727715492249, "learning_rate": 1.8788645306351886e-05, "loss": 0.2527, "step": 2675 }, { "epoch": 0.11282332356599278, "grad_norm": 1.0297092199325562, "learning_rate": 1.879567172568859e-05, "loss": 0.3268, "step": 2676 }, { "epoch": 0.11286548474819234, "grad_norm": 1.2791430950164795, "learning_rate": 1.8802698145025295e-05, "loss": 0.5068, "step": 2677 }, { "epoch": 0.11290764593039189, "grad_norm": 0.9408848881721497, "learning_rate": 1.8809724564362003e-05, "loss": 0.3493, "step": 2678 }, { "epoch": 0.11294980711259144, "grad_norm": 2.817441463470459, "learning_rate": 1.8816750983698707e-05, "loss": 0.9114, "step": 2679 }, { "epoch": 0.11299196829479098, "grad_norm": 0.7594043612480164, "learning_rate": 1.8823777403035415e-05, "loss": 0.2152, "step": 2680 }, { "epoch": 0.11303412947699054, "grad_norm": 0.49422383308410645, "learning_rate": 1.883080382237212e-05, "loss": 0.1464, "step": 2681 }, { "epoch": 0.11307629065919009, "grad_norm": 1.0872026681900024, "learning_rate": 1.8837830241708827e-05, "loss": 0.6542, "step": 2682 }, { "epoch": 0.11311845184138963, "grad_norm": 0.9533447623252869, "learning_rate": 1.884485666104553e-05, "loss": 0.2826, "step": 2683 }, { "epoch": 0.11316061302358918, "grad_norm": 1.1512867212295532, "learning_rate": 1.885188308038224e-05, "loss": 0.8044, "step": 2684 }, { "epoch": 0.11320277420578873, "grad_norm": 1.047504186630249, "learning_rate": 1.8858909499718944e-05, "loss": 0.7292, "step": 2685 }, { "epoch": 0.11324493538798829, "grad_norm": 0.7179593443870544, "learning_rate": 1.8865935919055648e-05, "loss": 0.508, "step": 2686 }, { "epoch": 0.11328709657018783, "grad_norm": 1.0909161567687988, "learning_rate": 1.8872962338392356e-05, "loss": 0.3966, "step": 2687 }, { "epoch": 0.11332925775238738, "grad_norm": 0.49254027009010315, "learning_rate": 1.887998875772906e-05, "loss": 0.1294, "step": 2688 }, { "epoch": 0.11337141893458692, "grad_norm": 3.5739564895629883, "learning_rate": 1.8887015177065768e-05, "loss": 1.1475, "step": 2689 }, { "epoch": 0.11341358011678647, "grad_norm": 0.9234926700592041, "learning_rate": 1.8894041596402472e-05, "loss": 0.6919, "step": 2690 }, { "epoch": 0.11345574129898603, "grad_norm": 0.9666410684585571, "learning_rate": 1.890106801573918e-05, "loss": 0.6784, "step": 2691 }, { "epoch": 0.11349790248118558, "grad_norm": 0.9357845783233643, "learning_rate": 1.8908094435075885e-05, "loss": 0.1938, "step": 2692 }, { "epoch": 0.11354006366338512, "grad_norm": 1.0519676208496094, "learning_rate": 1.8915120854412592e-05, "loss": 0.636, "step": 2693 }, { "epoch": 0.11358222484558467, "grad_norm": 0.9230110049247742, "learning_rate": 1.89221472737493e-05, "loss": 0.2518, "step": 2694 }, { "epoch": 0.11362438602778421, "grad_norm": 1.1683381795883179, "learning_rate": 1.8929173693086004e-05, "loss": 0.2637, "step": 2695 }, { "epoch": 0.11366654720998377, "grad_norm": 1.5596530437469482, "learning_rate": 1.8936200112422712e-05, "loss": 0.7844, "step": 2696 }, { "epoch": 0.11370870839218332, "grad_norm": 1.4283294677734375, "learning_rate": 1.8943226531759417e-05, "loss": 0.2944, "step": 2697 }, { "epoch": 0.11375086957438287, "grad_norm": 0.9584363102912903, "learning_rate": 1.8950252951096124e-05, "loss": 0.2405, "step": 2698 }, { "epoch": 0.11379303075658241, "grad_norm": 1.4105775356292725, "learning_rate": 1.895727937043283e-05, "loss": 0.2598, "step": 2699 }, { "epoch": 0.11383519193878196, "grad_norm": 2.0098774433135986, "learning_rate": 1.8964305789769533e-05, "loss": 1.3929, "step": 2700 }, { "epoch": 0.11387735312098152, "grad_norm": 0.787997305393219, "learning_rate": 1.897133220910624e-05, "loss": 0.3933, "step": 2701 }, { "epoch": 0.11391951430318106, "grad_norm": 0.8640748858451843, "learning_rate": 1.8978358628442945e-05, "loss": 0.1873, "step": 2702 }, { "epoch": 0.11396167548538061, "grad_norm": 1.1126971244812012, "learning_rate": 1.8985385047779653e-05, "loss": 0.2672, "step": 2703 }, { "epoch": 0.11400383666758016, "grad_norm": 1.1755001544952393, "learning_rate": 1.8992411467116358e-05, "loss": 0.4306, "step": 2704 }, { "epoch": 0.1140459978497797, "grad_norm": 0.9443737864494324, "learning_rate": 1.8999437886453065e-05, "loss": 0.7367, "step": 2705 }, { "epoch": 0.11408815903197926, "grad_norm": 0.8472340106964111, "learning_rate": 1.900646430578977e-05, "loss": 0.3377, "step": 2706 }, { "epoch": 0.11413032021417881, "grad_norm": 1.4440563917160034, "learning_rate": 1.9013490725126478e-05, "loss": 1.0267, "step": 2707 }, { "epoch": 0.11417248139637835, "grad_norm": 1.2736170291900635, "learning_rate": 1.9020517144463182e-05, "loss": 0.1399, "step": 2708 }, { "epoch": 0.1142146425785779, "grad_norm": 0.6360132694244385, "learning_rate": 1.9027543563799886e-05, "loss": 0.1243, "step": 2709 }, { "epoch": 0.11425680376077746, "grad_norm": 0.8875479102134705, "learning_rate": 1.9034569983136594e-05, "loss": 0.2928, "step": 2710 }, { "epoch": 0.114298964942977, "grad_norm": 0.6437318921089172, "learning_rate": 1.90415964024733e-05, "loss": 0.1387, "step": 2711 }, { "epoch": 0.11434112612517655, "grad_norm": 0.9926124215126038, "learning_rate": 1.9048622821810006e-05, "loss": 0.8138, "step": 2712 }, { "epoch": 0.1143832873073761, "grad_norm": 1.1007211208343506, "learning_rate": 1.905564924114671e-05, "loss": 0.307, "step": 2713 }, { "epoch": 0.11442544848957564, "grad_norm": 0.5275053977966309, "learning_rate": 1.906267566048342e-05, "loss": 0.1393, "step": 2714 }, { "epoch": 0.1144676096717752, "grad_norm": 0.8494741320610046, "learning_rate": 1.9069702079820126e-05, "loss": 0.2443, "step": 2715 }, { "epoch": 0.11450977085397475, "grad_norm": 0.9329468011856079, "learning_rate": 1.907672849915683e-05, "loss": 0.6532, "step": 2716 }, { "epoch": 0.1145519320361743, "grad_norm": 0.6717954874038696, "learning_rate": 1.908375491849354e-05, "loss": 0.187, "step": 2717 }, { "epoch": 0.11459409321837384, "grad_norm": 0.9369524121284485, "learning_rate": 1.9090781337830243e-05, "loss": 0.6618, "step": 2718 }, { "epoch": 0.11463625440057339, "grad_norm": 1.2136633396148682, "learning_rate": 1.909780775716695e-05, "loss": 0.7143, "step": 2719 }, { "epoch": 0.11467841558277295, "grad_norm": 0.778799295425415, "learning_rate": 1.9104834176503655e-05, "loss": 0.3917, "step": 2720 }, { "epoch": 0.11472057676497249, "grad_norm": 0.9076274633407593, "learning_rate": 1.9111860595840363e-05, "loss": 0.1768, "step": 2721 }, { "epoch": 0.11476273794717204, "grad_norm": 2.0783770084381104, "learning_rate": 1.9118887015177067e-05, "loss": 0.9857, "step": 2722 }, { "epoch": 0.11480489912937158, "grad_norm": 1.0661051273345947, "learning_rate": 1.912591343451377e-05, "loss": 1.2325, "step": 2723 }, { "epoch": 0.11484706031157113, "grad_norm": 0.8522204756736755, "learning_rate": 1.913293985385048e-05, "loss": 0.594, "step": 2724 }, { "epoch": 0.11488922149377069, "grad_norm": 0.6965012550354004, "learning_rate": 1.9139966273187184e-05, "loss": 0.1616, "step": 2725 }, { "epoch": 0.11493138267597024, "grad_norm": 0.8100733757019043, "learning_rate": 1.914699269252389e-05, "loss": 0.3851, "step": 2726 }, { "epoch": 0.11497354385816978, "grad_norm": 1.220278024673462, "learning_rate": 1.9154019111860596e-05, "loss": 0.4148, "step": 2727 }, { "epoch": 0.11501570504036933, "grad_norm": 0.8483852744102478, "learning_rate": 1.9161045531197304e-05, "loss": 0.2007, "step": 2728 }, { "epoch": 0.11505786622256887, "grad_norm": 0.7975591421127319, "learning_rate": 1.9168071950534008e-05, "loss": 0.4557, "step": 2729 }, { "epoch": 0.11510002740476843, "grad_norm": 2.7153611183166504, "learning_rate": 1.9175098369870716e-05, "loss": 1.0421, "step": 2730 }, { "epoch": 0.11514218858696798, "grad_norm": 1.8057773113250732, "learning_rate": 1.918212478920742e-05, "loss": 1.152, "step": 2731 }, { "epoch": 0.11518434976916753, "grad_norm": 0.8736106753349304, "learning_rate": 1.9189151208544125e-05, "loss": 0.6478, "step": 2732 }, { "epoch": 0.11522651095136707, "grad_norm": 0.7556118965148926, "learning_rate": 1.9196177627880832e-05, "loss": 0.6008, "step": 2733 }, { "epoch": 0.11526867213356662, "grad_norm": 1.1302391290664673, "learning_rate": 1.9203204047217537e-05, "loss": 0.2803, "step": 2734 }, { "epoch": 0.11531083331576618, "grad_norm": 0.7426232695579529, "learning_rate": 1.9210230466554245e-05, "loss": 0.2347, "step": 2735 }, { "epoch": 0.11535299449796572, "grad_norm": 0.9913283586502075, "learning_rate": 1.921725688589095e-05, "loss": 0.601, "step": 2736 }, { "epoch": 0.11539515568016527, "grad_norm": 0.8270602822303772, "learning_rate": 1.9224283305227657e-05, "loss": 0.7183, "step": 2737 }, { "epoch": 0.11543731686236482, "grad_norm": 1.3667492866516113, "learning_rate": 1.9231309724564365e-05, "loss": 0.319, "step": 2738 }, { "epoch": 0.11547947804456438, "grad_norm": 0.5363757610321045, "learning_rate": 1.923833614390107e-05, "loss": 0.1354, "step": 2739 }, { "epoch": 0.11552163922676392, "grad_norm": 1.8516663312911987, "learning_rate": 1.9245362563237777e-05, "loss": 1.1402, "step": 2740 }, { "epoch": 0.11556380040896347, "grad_norm": 0.9704086184501648, "learning_rate": 1.925238898257448e-05, "loss": 0.9772, "step": 2741 }, { "epoch": 0.11560596159116301, "grad_norm": 0.6837202310562134, "learning_rate": 1.925941540191119e-05, "loss": 0.1266, "step": 2742 }, { "epoch": 0.11564812277336256, "grad_norm": 0.8393951058387756, "learning_rate": 1.9266441821247893e-05, "loss": 0.7857, "step": 2743 }, { "epoch": 0.11569028395556212, "grad_norm": 0.7787613868713379, "learning_rate": 1.92734682405846e-05, "loss": 0.227, "step": 2744 }, { "epoch": 0.11573244513776167, "grad_norm": 0.9019625186920166, "learning_rate": 1.9280494659921306e-05, "loss": 0.191, "step": 2745 }, { "epoch": 0.11577460631996121, "grad_norm": 1.428175926208496, "learning_rate": 1.928752107925801e-05, "loss": 0.9343, "step": 2746 }, { "epoch": 0.11581676750216076, "grad_norm": 0.5606420636177063, "learning_rate": 1.9294547498594718e-05, "loss": 0.13, "step": 2747 }, { "epoch": 0.1158589286843603, "grad_norm": 0.7260568141937256, "learning_rate": 1.9301573917931422e-05, "loss": 0.5743, "step": 2748 }, { "epoch": 0.11590108986655986, "grad_norm": 0.7903695106506348, "learning_rate": 1.930860033726813e-05, "loss": 0.3693, "step": 2749 }, { "epoch": 0.11594325104875941, "grad_norm": 0.8763325810432434, "learning_rate": 1.9315626756604834e-05, "loss": 0.3013, "step": 2750 }, { "epoch": 0.11598541223095896, "grad_norm": 1.33938729763031, "learning_rate": 1.9322653175941542e-05, "loss": 1.2887, "step": 2751 }, { "epoch": 0.1160275734131585, "grad_norm": 0.7565689086914062, "learning_rate": 1.9329679595278246e-05, "loss": 0.1316, "step": 2752 }, { "epoch": 0.11606973459535805, "grad_norm": 1.0783562660217285, "learning_rate": 1.933670601461495e-05, "loss": 0.803, "step": 2753 }, { "epoch": 0.11611189577755761, "grad_norm": 1.4052261114120483, "learning_rate": 1.934373243395166e-05, "loss": 1.0789, "step": 2754 }, { "epoch": 0.11615405695975715, "grad_norm": 0.7908285856246948, "learning_rate": 1.9350758853288363e-05, "loss": 0.2533, "step": 2755 }, { "epoch": 0.1161962181419567, "grad_norm": 0.5161212682723999, "learning_rate": 1.935778527262507e-05, "loss": 0.1776, "step": 2756 }, { "epoch": 0.11623837932415625, "grad_norm": 2.1267831325531006, "learning_rate": 1.9364811691961775e-05, "loss": 1.3591, "step": 2757 }, { "epoch": 0.11628054050635579, "grad_norm": 1.1003061532974243, "learning_rate": 1.9371838111298483e-05, "loss": 0.5863, "step": 2758 }, { "epoch": 0.11632270168855535, "grad_norm": 0.7070544362068176, "learning_rate": 1.9378864530635187e-05, "loss": 0.1085, "step": 2759 }, { "epoch": 0.1163648628707549, "grad_norm": 0.9138835668563843, "learning_rate": 1.9385890949971895e-05, "loss": 0.1931, "step": 2760 }, { "epoch": 0.11640702405295444, "grad_norm": 0.9658528566360474, "learning_rate": 1.9392917369308603e-05, "loss": 0.703, "step": 2761 }, { "epoch": 0.11644918523515399, "grad_norm": 0.7592604756355286, "learning_rate": 1.9399943788645307e-05, "loss": 0.3134, "step": 2762 }, { "epoch": 0.11649134641735354, "grad_norm": 0.5927830338478088, "learning_rate": 1.9406970207982015e-05, "loss": 0.1186, "step": 2763 }, { "epoch": 0.1165335075995531, "grad_norm": 1.0182220935821533, "learning_rate": 1.941399662731872e-05, "loss": 0.822, "step": 2764 }, { "epoch": 0.11657566878175264, "grad_norm": 0.8474565148353577, "learning_rate": 1.9421023046655427e-05, "loss": 0.1524, "step": 2765 }, { "epoch": 0.11661782996395219, "grad_norm": 0.7461113929748535, "learning_rate": 1.9428049465992132e-05, "loss": 0.1978, "step": 2766 }, { "epoch": 0.11665999114615173, "grad_norm": 1.699808955192566, "learning_rate": 1.943507588532884e-05, "loss": 1.2169, "step": 2767 }, { "epoch": 0.11670215232835128, "grad_norm": 0.6206570863723755, "learning_rate": 1.9442102304665544e-05, "loss": 0.2811, "step": 2768 }, { "epoch": 0.11674431351055084, "grad_norm": 1.0914556980133057, "learning_rate": 1.944912872400225e-05, "loss": 0.1935, "step": 2769 }, { "epoch": 0.11678647469275039, "grad_norm": 0.5720531344413757, "learning_rate": 1.9456155143338956e-05, "loss": 0.1901, "step": 2770 }, { "epoch": 0.11682863587494993, "grad_norm": 0.6400003433227539, "learning_rate": 1.946318156267566e-05, "loss": 0.1729, "step": 2771 }, { "epoch": 0.11687079705714948, "grad_norm": 0.8416888117790222, "learning_rate": 1.9470207982012368e-05, "loss": 0.5777, "step": 2772 }, { "epoch": 0.11691295823934904, "grad_norm": 0.7771966457366943, "learning_rate": 1.9477234401349073e-05, "loss": 0.2523, "step": 2773 }, { "epoch": 0.11695511942154858, "grad_norm": 0.9584172368049622, "learning_rate": 1.948426082068578e-05, "loss": 0.4461, "step": 2774 }, { "epoch": 0.11699728060374813, "grad_norm": 0.7762075662612915, "learning_rate": 1.9491287240022485e-05, "loss": 0.1844, "step": 2775 }, { "epoch": 0.11703944178594768, "grad_norm": 0.5942303538322449, "learning_rate": 1.949831365935919e-05, "loss": 0.1305, "step": 2776 }, { "epoch": 0.11708160296814722, "grad_norm": 0.9299319386482239, "learning_rate": 1.9505340078695897e-05, "loss": 0.7483, "step": 2777 }, { "epoch": 0.11712376415034678, "grad_norm": 1.1149269342422485, "learning_rate": 1.95123664980326e-05, "loss": 0.8416, "step": 2778 }, { "epoch": 0.11716592533254633, "grad_norm": 0.9497433304786682, "learning_rate": 1.951939291736931e-05, "loss": 0.2372, "step": 2779 }, { "epoch": 0.11720808651474587, "grad_norm": 0.8132931590080261, "learning_rate": 1.9526419336706014e-05, "loss": 0.3161, "step": 2780 }, { "epoch": 0.11725024769694542, "grad_norm": 0.9634218811988831, "learning_rate": 1.953344575604272e-05, "loss": 0.6496, "step": 2781 }, { "epoch": 0.11729240887914497, "grad_norm": 0.7553243041038513, "learning_rate": 1.9540472175379426e-05, "loss": 0.6945, "step": 2782 }, { "epoch": 0.11733457006134453, "grad_norm": 3.5675528049468994, "learning_rate": 1.9547498594716134e-05, "loss": 0.8702, "step": 2783 }, { "epoch": 0.11737673124354407, "grad_norm": 1.1314598321914673, "learning_rate": 1.955452501405284e-05, "loss": 0.2106, "step": 2784 }, { "epoch": 0.11741889242574362, "grad_norm": 0.6109849810600281, "learning_rate": 1.9561551433389546e-05, "loss": 0.1498, "step": 2785 }, { "epoch": 0.11746105360794316, "grad_norm": 2.053882598876953, "learning_rate": 1.9568577852726254e-05, "loss": 0.7916, "step": 2786 }, { "epoch": 0.11750321479014271, "grad_norm": 0.8610380291938782, "learning_rate": 1.9575604272062958e-05, "loss": 0.7623, "step": 2787 }, { "epoch": 0.11754537597234227, "grad_norm": 0.5354581475257874, "learning_rate": 1.9582630691399666e-05, "loss": 0.2629, "step": 2788 }, { "epoch": 0.11758753715454182, "grad_norm": 1.1945244073867798, "learning_rate": 1.958965711073637e-05, "loss": 0.2657, "step": 2789 }, { "epoch": 0.11762969833674136, "grad_norm": 0.8333754539489746, "learning_rate": 1.9596683530073074e-05, "loss": 0.3487, "step": 2790 }, { "epoch": 0.11767185951894091, "grad_norm": 1.0668658018112183, "learning_rate": 1.9603709949409782e-05, "loss": 0.7197, "step": 2791 }, { "epoch": 0.11771402070114045, "grad_norm": 1.0169256925582886, "learning_rate": 1.9610736368746487e-05, "loss": 0.2553, "step": 2792 }, { "epoch": 0.11775618188334001, "grad_norm": 1.3814523220062256, "learning_rate": 1.9617762788083194e-05, "loss": 0.3006, "step": 2793 }, { "epoch": 0.11779834306553956, "grad_norm": 0.886273205280304, "learning_rate": 1.96247892074199e-05, "loss": 0.308, "step": 2794 }, { "epoch": 0.1178405042477391, "grad_norm": 0.6823963522911072, "learning_rate": 1.9631815626756607e-05, "loss": 0.2252, "step": 2795 }, { "epoch": 0.11788266542993865, "grad_norm": 0.7680464386940002, "learning_rate": 1.963884204609331e-05, "loss": 0.3212, "step": 2796 }, { "epoch": 0.1179248266121382, "grad_norm": 0.6217151880264282, "learning_rate": 1.964586846543002e-05, "loss": 0.1874, "step": 2797 }, { "epoch": 0.11796698779433776, "grad_norm": 1.2142707109451294, "learning_rate": 1.9652894884766723e-05, "loss": 0.5145, "step": 2798 }, { "epoch": 0.1180091489765373, "grad_norm": 1.059664249420166, "learning_rate": 1.9659921304103428e-05, "loss": 0.3003, "step": 2799 }, { "epoch": 0.11805131015873685, "grad_norm": 0.8905671834945679, "learning_rate": 1.9666947723440135e-05, "loss": 0.7518, "step": 2800 }, { "epoch": 0.1180934713409364, "grad_norm": 1.3795182704925537, "learning_rate": 1.967397414277684e-05, "loss": 1.181, "step": 2801 }, { "epoch": 0.11813563252313596, "grad_norm": 0.6350196003913879, "learning_rate": 1.9681000562113548e-05, "loss": 0.2182, "step": 2802 }, { "epoch": 0.1181777937053355, "grad_norm": 0.8169394135475159, "learning_rate": 1.9688026981450252e-05, "loss": 0.4867, "step": 2803 }, { "epoch": 0.11821995488753505, "grad_norm": 1.0930812358856201, "learning_rate": 1.969505340078696e-05, "loss": 0.3845, "step": 2804 }, { "epoch": 0.1182621160697346, "grad_norm": 0.8860512375831604, "learning_rate": 1.9702079820123668e-05, "loss": 0.751, "step": 2805 }, { "epoch": 0.11830427725193414, "grad_norm": 0.8311176300048828, "learning_rate": 1.9709106239460372e-05, "loss": 0.299, "step": 2806 }, { "epoch": 0.1183464384341337, "grad_norm": 1.0873150825500488, "learning_rate": 1.971613265879708e-05, "loss": 0.1789, "step": 2807 }, { "epoch": 0.11838859961633325, "grad_norm": 1.131394624710083, "learning_rate": 1.9723159078133784e-05, "loss": 0.9632, "step": 2808 }, { "epoch": 0.11843076079853279, "grad_norm": 0.9252196550369263, "learning_rate": 1.9730185497470492e-05, "loss": 0.2808, "step": 2809 }, { "epoch": 0.11847292198073234, "grad_norm": 0.752273678779602, "learning_rate": 1.9737211916807196e-05, "loss": 0.2354, "step": 2810 }, { "epoch": 0.11851508316293188, "grad_norm": 0.7726693153381348, "learning_rate": 1.9744238336143904e-05, "loss": 0.6788, "step": 2811 }, { "epoch": 0.11855724434513144, "grad_norm": 0.5444349050521851, "learning_rate": 1.975126475548061e-05, "loss": 0.4608, "step": 2812 }, { "epoch": 0.11859940552733099, "grad_norm": 0.9661257266998291, "learning_rate": 1.9758291174817313e-05, "loss": 0.3255, "step": 2813 }, { "epoch": 0.11864156670953054, "grad_norm": 1.287818193435669, "learning_rate": 1.976531759415402e-05, "loss": 0.6451, "step": 2814 }, { "epoch": 0.11868372789173008, "grad_norm": 0.6648878455162048, "learning_rate": 1.9772344013490725e-05, "loss": 0.1674, "step": 2815 }, { "epoch": 0.11872588907392963, "grad_norm": 0.8096155524253845, "learning_rate": 1.9779370432827433e-05, "loss": 0.3527, "step": 2816 }, { "epoch": 0.11876805025612919, "grad_norm": 0.6451486945152283, "learning_rate": 1.9786396852164137e-05, "loss": 0.094, "step": 2817 }, { "epoch": 0.11881021143832873, "grad_norm": 0.7265570759773254, "learning_rate": 1.9793423271500845e-05, "loss": 0.4512, "step": 2818 }, { "epoch": 0.11885237262052828, "grad_norm": 0.8171469569206238, "learning_rate": 1.980044969083755e-05, "loss": 0.5946, "step": 2819 }, { "epoch": 0.11889453380272783, "grad_norm": 0.545862078666687, "learning_rate": 1.9807476110174257e-05, "loss": 0.1951, "step": 2820 }, { "epoch": 0.11893669498492737, "grad_norm": 0.8621190786361694, "learning_rate": 1.981450252951096e-05, "loss": 0.1751, "step": 2821 }, { "epoch": 0.11897885616712693, "grad_norm": 1.0706636905670166, "learning_rate": 1.9821528948847666e-05, "loss": 0.3094, "step": 2822 }, { "epoch": 0.11902101734932648, "grad_norm": 0.9734317660331726, "learning_rate": 1.9828555368184374e-05, "loss": 0.2056, "step": 2823 }, { "epoch": 0.11906317853152602, "grad_norm": 0.8837046027183533, "learning_rate": 1.9835581787521078e-05, "loss": 0.1998, "step": 2824 }, { "epoch": 0.11910533971372557, "grad_norm": 0.9370532035827637, "learning_rate": 1.9842608206857786e-05, "loss": 0.1803, "step": 2825 }, { "epoch": 0.11914750089592512, "grad_norm": 1.2411952018737793, "learning_rate": 1.984963462619449e-05, "loss": 1.0498, "step": 2826 }, { "epoch": 0.11918966207812468, "grad_norm": 0.6646832227706909, "learning_rate": 1.9856661045531198e-05, "loss": 0.1097, "step": 2827 }, { "epoch": 0.11923182326032422, "grad_norm": 0.7344616055488586, "learning_rate": 1.9863687464867906e-05, "loss": 0.1281, "step": 2828 }, { "epoch": 0.11927398444252377, "grad_norm": 0.7452588677406311, "learning_rate": 1.987071388420461e-05, "loss": 0.234, "step": 2829 }, { "epoch": 0.11931614562472331, "grad_norm": 0.7722044587135315, "learning_rate": 1.9877740303541318e-05, "loss": 0.2135, "step": 2830 }, { "epoch": 0.11935830680692286, "grad_norm": 1.7133612632751465, "learning_rate": 1.9884766722878022e-05, "loss": 1.1641, "step": 2831 }, { "epoch": 0.11940046798912242, "grad_norm": 0.792268693447113, "learning_rate": 1.989179314221473e-05, "loss": 0.1581, "step": 2832 }, { "epoch": 0.11944262917132197, "grad_norm": 0.8498839735984802, "learning_rate": 1.9898819561551435e-05, "loss": 0.3563, "step": 2833 }, { "epoch": 0.11948479035352151, "grad_norm": 0.7302964329719543, "learning_rate": 1.9905845980888142e-05, "loss": 0.1923, "step": 2834 }, { "epoch": 0.11952695153572106, "grad_norm": 1.0259724855422974, "learning_rate": 1.9912872400224847e-05, "loss": 0.6097, "step": 2835 }, { "epoch": 0.11956911271792062, "grad_norm": 0.7428796291351318, "learning_rate": 1.991989881956155e-05, "loss": 0.3963, "step": 2836 }, { "epoch": 0.11961127390012016, "grad_norm": 0.9690293669700623, "learning_rate": 1.992692523889826e-05, "loss": 0.3011, "step": 2837 }, { "epoch": 0.11965343508231971, "grad_norm": 0.8418883681297302, "learning_rate": 1.9933951658234963e-05, "loss": 0.3634, "step": 2838 }, { "epoch": 0.11969559626451926, "grad_norm": 0.7344012260437012, "learning_rate": 1.994097807757167e-05, "loss": 0.3354, "step": 2839 }, { "epoch": 0.1197377574467188, "grad_norm": 0.923847496509552, "learning_rate": 1.9948004496908376e-05, "loss": 0.285, "step": 2840 }, { "epoch": 0.11977991862891836, "grad_norm": 0.8866056203842163, "learning_rate": 1.9955030916245083e-05, "loss": 0.2164, "step": 2841 }, { "epoch": 0.11982207981111791, "grad_norm": 0.9207534790039062, "learning_rate": 1.9962057335581788e-05, "loss": 0.6624, "step": 2842 }, { "epoch": 0.11986424099331745, "grad_norm": 0.9731162786483765, "learning_rate": 1.9969083754918496e-05, "loss": 0.6327, "step": 2843 }, { "epoch": 0.119906402175517, "grad_norm": 0.996025800704956, "learning_rate": 1.99761101742552e-05, "loss": 0.1694, "step": 2844 }, { "epoch": 0.11994856335771654, "grad_norm": 0.556600034236908, "learning_rate": 1.9983136593591904e-05, "loss": 0.2383, "step": 2845 }, { "epoch": 0.1199907245399161, "grad_norm": 0.7633523344993591, "learning_rate": 1.9990163012928612e-05, "loss": 0.3439, "step": 2846 }, { "epoch": 0.12003288572211565, "grad_norm": 0.6015572547912598, "learning_rate": 1.9997189432265316e-05, "loss": 0.3873, "step": 2847 }, { "epoch": 0.1200750469043152, "grad_norm": 0.9471933841705322, "learning_rate": 2.0004215851602024e-05, "loss": 0.2217, "step": 2848 }, { "epoch": 0.12011720808651474, "grad_norm": 3.1296563148498535, "learning_rate": 2.001124227093873e-05, "loss": 0.9891, "step": 2849 }, { "epoch": 0.12015936926871429, "grad_norm": 0.9852733016014099, "learning_rate": 2.0018268690275436e-05, "loss": 0.2395, "step": 2850 }, { "epoch": 0.12020153045091385, "grad_norm": 0.7894324660301208, "learning_rate": 2.0025295109612144e-05, "loss": 0.3301, "step": 2851 }, { "epoch": 0.1202436916331134, "grad_norm": 0.795404851436615, "learning_rate": 2.003232152894885e-05, "loss": 0.2258, "step": 2852 }, { "epoch": 0.12028585281531294, "grad_norm": 1.2521368265151978, "learning_rate": 2.0039347948285556e-05, "loss": 0.4246, "step": 2853 }, { "epoch": 0.12032801399751249, "grad_norm": 0.9017361998558044, "learning_rate": 2.004637436762226e-05, "loss": 0.3105, "step": 2854 }, { "epoch": 0.12037017517971203, "grad_norm": 0.5002608895301819, "learning_rate": 2.005340078695897e-05, "loss": 0.1274, "step": 2855 }, { "epoch": 0.12041233636191159, "grad_norm": 1.148764729499817, "learning_rate": 2.0060427206295673e-05, "loss": 0.3085, "step": 2856 }, { "epoch": 0.12045449754411114, "grad_norm": 0.831092894077301, "learning_rate": 2.006745362563238e-05, "loss": 0.2059, "step": 2857 }, { "epoch": 0.12049665872631068, "grad_norm": 1.065421462059021, "learning_rate": 2.0074480044969085e-05, "loss": 0.2107, "step": 2858 }, { "epoch": 0.12053881990851023, "grad_norm": 0.7934461236000061, "learning_rate": 2.008150646430579e-05, "loss": 0.1573, "step": 2859 }, { "epoch": 0.12058098109070978, "grad_norm": 0.7798044085502625, "learning_rate": 2.0088532883642497e-05, "loss": 0.1931, "step": 2860 }, { "epoch": 0.12062314227290934, "grad_norm": 0.6448240876197815, "learning_rate": 2.0095559302979202e-05, "loss": 0.9613, "step": 2861 }, { "epoch": 0.12066530345510888, "grad_norm": 0.8193891644477844, "learning_rate": 2.010258572231591e-05, "loss": 0.2608, "step": 2862 }, { "epoch": 0.12070746463730843, "grad_norm": 0.8304193615913391, "learning_rate": 2.0109612141652614e-05, "loss": 0.7231, "step": 2863 }, { "epoch": 0.12074962581950797, "grad_norm": 0.8545004725456238, "learning_rate": 2.0116638560989322e-05, "loss": 0.2152, "step": 2864 }, { "epoch": 0.12079178700170753, "grad_norm": 0.5017878413200378, "learning_rate": 2.0123664980326026e-05, "loss": 0.1369, "step": 2865 }, { "epoch": 0.12083394818390708, "grad_norm": 0.8182042241096497, "learning_rate": 2.0130691399662734e-05, "loss": 0.2372, "step": 2866 }, { "epoch": 0.12087610936610663, "grad_norm": 1.0311384201049805, "learning_rate": 2.0137717818999438e-05, "loss": 0.3303, "step": 2867 }, { "epoch": 0.12091827054830617, "grad_norm": 0.5946455597877502, "learning_rate": 2.0144744238336143e-05, "loss": 0.4726, "step": 2868 }, { "epoch": 0.12096043173050572, "grad_norm": 0.48708340525627136, "learning_rate": 2.015177065767285e-05, "loss": 0.1331, "step": 2869 }, { "epoch": 0.12100259291270528, "grad_norm": 0.7721068263053894, "learning_rate": 2.0158797077009555e-05, "loss": 0.2575, "step": 2870 }, { "epoch": 0.12104475409490482, "grad_norm": 0.491361528635025, "learning_rate": 2.0165823496346263e-05, "loss": 0.1857, "step": 2871 }, { "epoch": 0.12108691527710437, "grad_norm": 0.7505192756652832, "learning_rate": 2.0172849915682967e-05, "loss": 0.5166, "step": 2872 }, { "epoch": 0.12112907645930392, "grad_norm": 1.4088462591171265, "learning_rate": 2.0179876335019675e-05, "loss": 0.6561, "step": 2873 }, { "epoch": 0.12117123764150346, "grad_norm": 0.8664233088493347, "learning_rate": 2.0186902754356383e-05, "loss": 0.6826, "step": 2874 }, { "epoch": 0.12121339882370302, "grad_norm": 0.6953309774398804, "learning_rate": 2.0193929173693087e-05, "loss": 0.5555, "step": 2875 }, { "epoch": 0.12125556000590257, "grad_norm": 0.971642255783081, "learning_rate": 2.0200955593029795e-05, "loss": 0.306, "step": 2876 }, { "epoch": 0.12129772118810211, "grad_norm": 0.8857735395431519, "learning_rate": 2.02079820123665e-05, "loss": 0.5577, "step": 2877 }, { "epoch": 0.12133988237030166, "grad_norm": 0.7248367667198181, "learning_rate": 2.0215008431703207e-05, "loss": 0.1918, "step": 2878 }, { "epoch": 0.1213820435525012, "grad_norm": 1.1500202417373657, "learning_rate": 2.022203485103991e-05, "loss": 0.2309, "step": 2879 }, { "epoch": 0.12142420473470077, "grad_norm": 0.782890796661377, "learning_rate": 2.022906127037662e-05, "loss": 0.1933, "step": 2880 }, { "epoch": 0.12146636591690031, "grad_norm": 0.7586026191711426, "learning_rate": 2.0236087689713324e-05, "loss": 0.6324, "step": 2881 }, { "epoch": 0.12150852709909986, "grad_norm": 1.2038811445236206, "learning_rate": 2.0243114109050028e-05, "loss": 0.5439, "step": 2882 }, { "epoch": 0.1215506882812994, "grad_norm": 1.1798737049102783, "learning_rate": 2.0250140528386736e-05, "loss": 1.2094, "step": 2883 }, { "epoch": 0.12159284946349895, "grad_norm": 0.7360527515411377, "learning_rate": 2.025716694772344e-05, "loss": 0.3495, "step": 2884 }, { "epoch": 0.12163501064569851, "grad_norm": 0.7932382225990295, "learning_rate": 2.0264193367060148e-05, "loss": 0.4343, "step": 2885 }, { "epoch": 0.12167717182789806, "grad_norm": 0.727786660194397, "learning_rate": 2.0271219786396852e-05, "loss": 0.2262, "step": 2886 }, { "epoch": 0.1217193330100976, "grad_norm": 0.9940581917762756, "learning_rate": 2.027824620573356e-05, "loss": 0.6262, "step": 2887 }, { "epoch": 0.12176149419229715, "grad_norm": 1.0199846029281616, "learning_rate": 2.0285272625070264e-05, "loss": 0.2071, "step": 2888 }, { "epoch": 0.1218036553744967, "grad_norm": 0.6089363694190979, "learning_rate": 2.029229904440697e-05, "loss": 0.1507, "step": 2889 }, { "epoch": 0.12184581655669625, "grad_norm": 1.3307355642318726, "learning_rate": 2.0299325463743677e-05, "loss": 0.3978, "step": 2890 }, { "epoch": 0.1218879777388958, "grad_norm": 0.8188278079032898, "learning_rate": 2.030635188308038e-05, "loss": 0.4996, "step": 2891 }, { "epoch": 0.12193013892109535, "grad_norm": 0.8818971514701843, "learning_rate": 2.031337830241709e-05, "loss": 0.2016, "step": 2892 }, { "epoch": 0.12197230010329489, "grad_norm": 0.46538275480270386, "learning_rate": 2.0320404721753793e-05, "loss": 0.1682, "step": 2893 }, { "epoch": 0.12201446128549445, "grad_norm": 0.6662185192108154, "learning_rate": 2.03274311410905e-05, "loss": 0.1531, "step": 2894 }, { "epoch": 0.122056622467694, "grad_norm": 1.4190315008163452, "learning_rate": 2.033445756042721e-05, "loss": 0.2829, "step": 2895 }, { "epoch": 0.12209878364989354, "grad_norm": 0.9596356749534607, "learning_rate": 2.0341483979763913e-05, "loss": 0.2552, "step": 2896 }, { "epoch": 0.12214094483209309, "grad_norm": 1.3019802570343018, "learning_rate": 2.034851039910062e-05, "loss": 1.2832, "step": 2897 }, { "epoch": 0.12218310601429264, "grad_norm": 0.5265406966209412, "learning_rate": 2.0355536818437325e-05, "loss": 0.1538, "step": 2898 }, { "epoch": 0.1222252671964922, "grad_norm": 0.8048961162567139, "learning_rate": 2.0362563237774033e-05, "loss": 0.3877, "step": 2899 }, { "epoch": 0.12226742837869174, "grad_norm": 0.8595924377441406, "learning_rate": 2.0369589657110738e-05, "loss": 0.5983, "step": 2900 }, { "epoch": 0.12230958956089129, "grad_norm": 0.5683339238166809, "learning_rate": 2.0376616076447445e-05, "loss": 0.1634, "step": 2901 }, { "epoch": 0.12235175074309083, "grad_norm": 0.46578484773635864, "learning_rate": 2.038364249578415e-05, "loss": 0.1335, "step": 2902 }, { "epoch": 0.12239391192529038, "grad_norm": 0.8353554606437683, "learning_rate": 2.0390668915120854e-05, "loss": 0.6473, "step": 2903 }, { "epoch": 0.12243607310748994, "grad_norm": 1.190774917602539, "learning_rate": 2.0397695334457562e-05, "loss": 0.6688, "step": 2904 }, { "epoch": 0.12247823428968949, "grad_norm": 0.902430534362793, "learning_rate": 2.0404721753794266e-05, "loss": 0.3282, "step": 2905 }, { "epoch": 0.12252039547188903, "grad_norm": 2.518561363220215, "learning_rate": 2.0411748173130974e-05, "loss": 0.8448, "step": 2906 }, { "epoch": 0.12256255665408858, "grad_norm": 0.6554327011108398, "learning_rate": 2.041877459246768e-05, "loss": 0.2847, "step": 2907 }, { "epoch": 0.12260471783628812, "grad_norm": 1.093084692955017, "learning_rate": 2.0425801011804386e-05, "loss": 0.5039, "step": 2908 }, { "epoch": 0.12264687901848768, "grad_norm": 0.8506736755371094, "learning_rate": 2.043282743114109e-05, "loss": 0.2748, "step": 2909 }, { "epoch": 0.12268904020068723, "grad_norm": 2.34397554397583, "learning_rate": 2.04398538504778e-05, "loss": 1.0897, "step": 2910 }, { "epoch": 0.12273120138288678, "grad_norm": 0.9442713856697083, "learning_rate": 2.0446880269814503e-05, "loss": 0.2221, "step": 2911 }, { "epoch": 0.12277336256508632, "grad_norm": 1.3873348236083984, "learning_rate": 2.0453906689151207e-05, "loss": 1.2307, "step": 2912 }, { "epoch": 0.12281552374728587, "grad_norm": 0.743024468421936, "learning_rate": 2.0460933108487915e-05, "loss": 0.4966, "step": 2913 }, { "epoch": 0.12285768492948543, "grad_norm": 0.608452558517456, "learning_rate": 2.046795952782462e-05, "loss": 0.329, "step": 2914 }, { "epoch": 0.12289984611168497, "grad_norm": 0.5766218900680542, "learning_rate": 2.0474985947161327e-05, "loss": 0.1155, "step": 2915 }, { "epoch": 0.12294200729388452, "grad_norm": 1.278683066368103, "learning_rate": 2.048201236649803e-05, "loss": 0.3154, "step": 2916 }, { "epoch": 0.12298416847608407, "grad_norm": 0.7015822529792786, "learning_rate": 2.048903878583474e-05, "loss": 0.2145, "step": 2917 }, { "epoch": 0.12302632965828361, "grad_norm": 0.8843845129013062, "learning_rate": 2.0496065205171447e-05, "loss": 0.1331, "step": 2918 }, { "epoch": 0.12306849084048317, "grad_norm": 1.405419945716858, "learning_rate": 2.050309162450815e-05, "loss": 0.5016, "step": 2919 }, { "epoch": 0.12311065202268272, "grad_norm": 0.7037725448608398, "learning_rate": 2.051011804384486e-05, "loss": 0.4074, "step": 2920 }, { "epoch": 0.12315281320488226, "grad_norm": 1.7377985715866089, "learning_rate": 2.0517144463181564e-05, "loss": 0.398, "step": 2921 }, { "epoch": 0.12319497438708181, "grad_norm": 0.8320789337158203, "learning_rate": 2.052417088251827e-05, "loss": 0.1694, "step": 2922 }, { "epoch": 0.12323713556928136, "grad_norm": 0.8838518261909485, "learning_rate": 2.0531197301854976e-05, "loss": 0.2153, "step": 2923 }, { "epoch": 0.12327929675148092, "grad_norm": 0.7002869844436646, "learning_rate": 2.0538223721191684e-05, "loss": 0.5924, "step": 2924 }, { "epoch": 0.12332145793368046, "grad_norm": 0.9970077276229858, "learning_rate": 2.0545250140528388e-05, "loss": 0.3329, "step": 2925 }, { "epoch": 0.12336361911588001, "grad_norm": 2.6427788734436035, "learning_rate": 2.0552276559865092e-05, "loss": 0.748, "step": 2926 }, { "epoch": 0.12340578029807955, "grad_norm": 0.4527434706687927, "learning_rate": 2.05593029792018e-05, "loss": 0.1008, "step": 2927 }, { "epoch": 0.12344794148027911, "grad_norm": 0.7384372353553772, "learning_rate": 2.0566329398538505e-05, "loss": 0.6678, "step": 2928 }, { "epoch": 0.12349010266247866, "grad_norm": 0.7912419438362122, "learning_rate": 2.0573355817875212e-05, "loss": 0.1354, "step": 2929 }, { "epoch": 0.1235322638446782, "grad_norm": 0.7869753837585449, "learning_rate": 2.0580382237211917e-05, "loss": 0.5289, "step": 2930 }, { "epoch": 0.12357442502687775, "grad_norm": 0.6022210717201233, "learning_rate": 2.0587408656548625e-05, "loss": 0.1874, "step": 2931 }, { "epoch": 0.1236165862090773, "grad_norm": 0.7908697128295898, "learning_rate": 2.059443507588533e-05, "loss": 0.1493, "step": 2932 }, { "epoch": 0.12365874739127686, "grad_norm": 0.6208046674728394, "learning_rate": 2.0601461495222037e-05, "loss": 0.1081, "step": 2933 }, { "epoch": 0.1237009085734764, "grad_norm": 0.5488293170928955, "learning_rate": 2.060848791455874e-05, "loss": 0.1126, "step": 2934 }, { "epoch": 0.12374306975567595, "grad_norm": 0.8425131440162659, "learning_rate": 2.0615514333895446e-05, "loss": 0.5239, "step": 2935 }, { "epoch": 0.1237852309378755, "grad_norm": 0.7449964880943298, "learning_rate": 2.0622540753232153e-05, "loss": 0.2224, "step": 2936 }, { "epoch": 0.12382739212007504, "grad_norm": 0.9126968383789062, "learning_rate": 2.0629567172568858e-05, "loss": 0.1715, "step": 2937 }, { "epoch": 0.1238695533022746, "grad_norm": 1.732524037361145, "learning_rate": 2.0636593591905566e-05, "loss": 0.4476, "step": 2938 }, { "epoch": 0.12391171448447415, "grad_norm": 0.9415704011917114, "learning_rate": 2.064362001124227e-05, "loss": 0.407, "step": 2939 }, { "epoch": 0.1239538756666737, "grad_norm": 0.781982958316803, "learning_rate": 2.0650646430578978e-05, "loss": 0.3284, "step": 2940 }, { "epoch": 0.12399603684887324, "grad_norm": 0.6244618892669678, "learning_rate": 2.0657672849915685e-05, "loss": 0.2055, "step": 2941 }, { "epoch": 0.12403819803107279, "grad_norm": 0.708379328250885, "learning_rate": 2.066469926925239e-05, "loss": 0.2809, "step": 2942 }, { "epoch": 0.12408035921327235, "grad_norm": 0.7437301874160767, "learning_rate": 2.0671725688589098e-05, "loss": 0.1807, "step": 2943 }, { "epoch": 0.12412252039547189, "grad_norm": 0.8834581971168518, "learning_rate": 2.0678752107925802e-05, "loss": 0.3354, "step": 2944 }, { "epoch": 0.12416468157767144, "grad_norm": 0.8079770803451538, "learning_rate": 2.068577852726251e-05, "loss": 0.7486, "step": 2945 }, { "epoch": 0.12420684275987098, "grad_norm": 1.2861250638961792, "learning_rate": 2.0692804946599214e-05, "loss": 0.945, "step": 2946 }, { "epoch": 0.12424900394207053, "grad_norm": 1.4776761531829834, "learning_rate": 2.0699831365935922e-05, "loss": 0.8149, "step": 2947 }, { "epoch": 0.12429116512427009, "grad_norm": 0.6437259316444397, "learning_rate": 2.0706857785272626e-05, "loss": 0.1802, "step": 2948 }, { "epoch": 0.12433332630646964, "grad_norm": 0.6203144788742065, "learning_rate": 2.071388420460933e-05, "loss": 0.2033, "step": 2949 }, { "epoch": 0.12437548748866918, "grad_norm": 1.092566967010498, "learning_rate": 2.072091062394604e-05, "loss": 0.3291, "step": 2950 }, { "epoch": 0.12441764867086873, "grad_norm": 1.1383377313613892, "learning_rate": 2.0727937043282743e-05, "loss": 0.2514, "step": 2951 }, { "epoch": 0.12445980985306827, "grad_norm": 1.1761877536773682, "learning_rate": 2.073496346261945e-05, "loss": 1.1079, "step": 2952 }, { "epoch": 0.12450197103526783, "grad_norm": 1.078972339630127, "learning_rate": 2.0741989881956155e-05, "loss": 0.293, "step": 2953 }, { "epoch": 0.12454413221746738, "grad_norm": 1.408292293548584, "learning_rate": 2.0749016301292863e-05, "loss": 1.1671, "step": 2954 }, { "epoch": 0.12458629339966693, "grad_norm": 0.6663954854011536, "learning_rate": 2.0756042720629567e-05, "loss": 0.6109, "step": 2955 }, { "epoch": 0.12462845458186647, "grad_norm": 0.8013573288917542, "learning_rate": 2.0763069139966275e-05, "loss": 0.2794, "step": 2956 }, { "epoch": 0.12467061576406603, "grad_norm": 0.7314831018447876, "learning_rate": 2.077009555930298e-05, "loss": 0.449, "step": 2957 }, { "epoch": 0.12471277694626558, "grad_norm": 1.4949796199798584, "learning_rate": 2.0777121978639684e-05, "loss": 0.4588, "step": 2958 }, { "epoch": 0.12475493812846512, "grad_norm": 0.555273175239563, "learning_rate": 2.0784148397976392e-05, "loss": 0.1803, "step": 2959 }, { "epoch": 0.12479709931066467, "grad_norm": 0.9785864353179932, "learning_rate": 2.0791174817313096e-05, "loss": 0.5108, "step": 2960 }, { "epoch": 0.12483926049286422, "grad_norm": 1.2575652599334717, "learning_rate": 2.0798201236649804e-05, "loss": 0.4551, "step": 2961 }, { "epoch": 0.12488142167506378, "grad_norm": 0.6018635034561157, "learning_rate": 2.0805227655986508e-05, "loss": 0.1605, "step": 2962 }, { "epoch": 0.12492358285726332, "grad_norm": 0.5405349135398865, "learning_rate": 2.0812254075323216e-05, "loss": 0.1865, "step": 2963 }, { "epoch": 0.12496574403946287, "grad_norm": 0.7903945446014404, "learning_rate": 2.0819280494659924e-05, "loss": 0.6668, "step": 2964 }, { "epoch": 0.1250079052216624, "grad_norm": 1.1867454051971436, "learning_rate": 2.0826306913996628e-05, "loss": 0.2071, "step": 2965 }, { "epoch": 0.12505006640386196, "grad_norm": 1.0664188861846924, "learning_rate": 2.0833333333333336e-05, "loss": 0.3426, "step": 2966 }, { "epoch": 0.1250922275860615, "grad_norm": 0.9918521642684937, "learning_rate": 2.084035975267004e-05, "loss": 0.1477, "step": 2967 }, { "epoch": 0.12513438876826105, "grad_norm": 0.7957016229629517, "learning_rate": 2.0847386172006748e-05, "loss": 0.2247, "step": 2968 }, { "epoch": 0.1251765499504606, "grad_norm": 0.9968437552452087, "learning_rate": 2.0854412591343453e-05, "loss": 0.5832, "step": 2969 }, { "epoch": 0.12521871113266017, "grad_norm": 0.7790495753288269, "learning_rate": 2.086143901068016e-05, "loss": 0.3975, "step": 2970 }, { "epoch": 0.12526087231485972, "grad_norm": 0.9041679501533508, "learning_rate": 2.0868465430016865e-05, "loss": 0.4134, "step": 2971 }, { "epoch": 0.12530303349705926, "grad_norm": 1.8614683151245117, "learning_rate": 2.087549184935357e-05, "loss": 0.8197, "step": 2972 }, { "epoch": 0.1253451946792588, "grad_norm": 0.9511240720748901, "learning_rate": 2.0882518268690277e-05, "loss": 0.4861, "step": 2973 }, { "epoch": 0.12538735586145835, "grad_norm": 0.6882857084274292, "learning_rate": 2.088954468802698e-05, "loss": 0.2216, "step": 2974 }, { "epoch": 0.1254295170436579, "grad_norm": 0.7729109525680542, "learning_rate": 2.089657110736369e-05, "loss": 0.2819, "step": 2975 }, { "epoch": 0.12547167822585745, "grad_norm": 0.8817088603973389, "learning_rate": 2.0903597526700394e-05, "loss": 0.2537, "step": 2976 }, { "epoch": 0.125513839408057, "grad_norm": 0.6773881912231445, "learning_rate": 2.09106239460371e-05, "loss": 0.2155, "step": 2977 }, { "epoch": 0.12555600059025654, "grad_norm": 0.7351981997489929, "learning_rate": 2.0917650365373806e-05, "loss": 0.185, "step": 2978 }, { "epoch": 0.1255981617724561, "grad_norm": 1.439676284790039, "learning_rate": 2.0924676784710513e-05, "loss": 1.1496, "step": 2979 }, { "epoch": 0.12564032295465566, "grad_norm": 1.4263797998428345, "learning_rate": 2.0931703204047218e-05, "loss": 1.0753, "step": 2980 }, { "epoch": 0.1256824841368552, "grad_norm": 1.1563804149627686, "learning_rate": 2.0938729623383922e-05, "loss": 1.394, "step": 2981 }, { "epoch": 0.12572464531905475, "grad_norm": 0.9454512596130371, "learning_rate": 2.094575604272063e-05, "loss": 0.4347, "step": 2982 }, { "epoch": 0.1257668065012543, "grad_norm": 0.5233676433563232, "learning_rate": 2.0952782462057334e-05, "loss": 0.1378, "step": 2983 }, { "epoch": 0.12580896768345384, "grad_norm": 0.8158366084098816, "learning_rate": 2.0959808881394042e-05, "loss": 0.6911, "step": 2984 }, { "epoch": 0.1258511288656534, "grad_norm": 0.8908228278160095, "learning_rate": 2.096683530073075e-05, "loss": 0.7664, "step": 2985 }, { "epoch": 0.12589329004785293, "grad_norm": 0.8728945255279541, "learning_rate": 2.0973861720067454e-05, "loss": 0.5833, "step": 2986 }, { "epoch": 0.12593545123005248, "grad_norm": 0.9714994430541992, "learning_rate": 2.0980888139404162e-05, "loss": 0.4648, "step": 2987 }, { "epoch": 0.12597761241225203, "grad_norm": 1.4088813066482544, "learning_rate": 2.0987914558740867e-05, "loss": 1.1826, "step": 2988 }, { "epoch": 0.1260197735944516, "grad_norm": 0.7338193655014038, "learning_rate": 2.0994940978077574e-05, "loss": 0.5028, "step": 2989 }, { "epoch": 0.12606193477665115, "grad_norm": 0.8926715850830078, "learning_rate": 2.100196739741428e-05, "loss": 0.3267, "step": 2990 }, { "epoch": 0.1261040959588507, "grad_norm": 2.0857794284820557, "learning_rate": 2.1008993816750987e-05, "loss": 0.6549, "step": 2991 }, { "epoch": 0.12614625714105024, "grad_norm": 0.707119882106781, "learning_rate": 2.101602023608769e-05, "loss": 0.2631, "step": 2992 }, { "epoch": 0.12618841832324978, "grad_norm": 0.4948939383029938, "learning_rate": 2.10230466554244e-05, "loss": 0.1585, "step": 2993 }, { "epoch": 0.12623057950544933, "grad_norm": 2.854161024093628, "learning_rate": 2.1030073074761103e-05, "loss": 1.1791, "step": 2994 }, { "epoch": 0.12627274068764888, "grad_norm": 0.48194271326065063, "learning_rate": 2.1037099494097808e-05, "loss": 0.1573, "step": 2995 }, { "epoch": 0.12631490186984842, "grad_norm": 0.862860381603241, "learning_rate": 2.1044125913434515e-05, "loss": 0.185, "step": 2996 }, { "epoch": 0.12635706305204797, "grad_norm": 4.035521984100342, "learning_rate": 2.105115233277122e-05, "loss": 1.0892, "step": 2997 }, { "epoch": 0.12639922423424751, "grad_norm": 0.7548896670341492, "learning_rate": 2.1058178752107927e-05, "loss": 0.4074, "step": 2998 }, { "epoch": 0.1264413854164471, "grad_norm": 0.8819296360015869, "learning_rate": 2.1065205171444632e-05, "loss": 0.7061, "step": 2999 }, { "epoch": 0.12648354659864663, "grad_norm": 0.5577319860458374, "learning_rate": 2.107223159078134e-05, "loss": 0.1457, "step": 3000 }, { "epoch": 0.12652570778084618, "grad_norm": 0.5421215295791626, "learning_rate": 2.1079258010118044e-05, "loss": 0.1517, "step": 3001 }, { "epoch": 0.12656786896304573, "grad_norm": 0.996231734752655, "learning_rate": 2.108628442945475e-05, "loss": 0.7496, "step": 3002 }, { "epoch": 0.12661003014524527, "grad_norm": 0.7632014751434326, "learning_rate": 2.1093310848791456e-05, "loss": 0.4457, "step": 3003 }, { "epoch": 0.12665219132744482, "grad_norm": 0.8326290845870972, "learning_rate": 2.110033726812816e-05, "loss": 0.81, "step": 3004 }, { "epoch": 0.12669435250964436, "grad_norm": 0.8249830603599548, "learning_rate": 2.110736368746487e-05, "loss": 0.2592, "step": 3005 }, { "epoch": 0.1267365136918439, "grad_norm": 1.5289881229400635, "learning_rate": 2.1114390106801573e-05, "loss": 0.9978, "step": 3006 }, { "epoch": 0.12677867487404346, "grad_norm": 1.143620491027832, "learning_rate": 2.112141652613828e-05, "loss": 0.2579, "step": 3007 }, { "epoch": 0.12682083605624303, "grad_norm": 0.8610961437225342, "learning_rate": 2.112844294547499e-05, "loss": 0.552, "step": 3008 }, { "epoch": 0.12686299723844258, "grad_norm": 1.3730206489562988, "learning_rate": 2.1135469364811693e-05, "loss": 1.205, "step": 3009 }, { "epoch": 0.12690515842064212, "grad_norm": 0.6854246854782104, "learning_rate": 2.11424957841484e-05, "loss": 0.5793, "step": 3010 }, { "epoch": 0.12694731960284167, "grad_norm": 1.2559278011322021, "learning_rate": 2.1149522203485105e-05, "loss": 0.2302, "step": 3011 }, { "epoch": 0.12698948078504121, "grad_norm": 1.9263070821762085, "learning_rate": 2.1156548622821813e-05, "loss": 0.9997, "step": 3012 }, { "epoch": 0.12703164196724076, "grad_norm": 1.49770188331604, "learning_rate": 2.1163575042158517e-05, "loss": 0.5472, "step": 3013 }, { "epoch": 0.1270738031494403, "grad_norm": 0.4181506037712097, "learning_rate": 2.1170601461495225e-05, "loss": 0.1506, "step": 3014 }, { "epoch": 0.12711596433163985, "grad_norm": 0.5683713555335999, "learning_rate": 2.117762788083193e-05, "loss": 0.3552, "step": 3015 }, { "epoch": 0.1271581255138394, "grad_norm": 0.587113618850708, "learning_rate": 2.1184654300168634e-05, "loss": 0.1058, "step": 3016 }, { "epoch": 0.12720028669603894, "grad_norm": 0.6922287940979004, "learning_rate": 2.119168071950534e-05, "loss": 0.293, "step": 3017 }, { "epoch": 0.12724244787823852, "grad_norm": 0.6125219464302063, "learning_rate": 2.1198707138842046e-05, "loss": 0.1314, "step": 3018 }, { "epoch": 0.12728460906043806, "grad_norm": 0.6660592555999756, "learning_rate": 2.1205733558178754e-05, "loss": 0.1248, "step": 3019 }, { "epoch": 0.1273267702426376, "grad_norm": 1.0772544145584106, "learning_rate": 2.1212759977515458e-05, "loss": 0.8838, "step": 3020 }, { "epoch": 0.12736893142483716, "grad_norm": 0.6289644241333008, "learning_rate": 2.1219786396852166e-05, "loss": 0.2258, "step": 3021 }, { "epoch": 0.1274110926070367, "grad_norm": 0.7405238747596741, "learning_rate": 2.122681281618887e-05, "loss": 0.5194, "step": 3022 }, { "epoch": 0.12745325378923625, "grad_norm": 0.6516577005386353, "learning_rate": 2.1233839235525578e-05, "loss": 0.6739, "step": 3023 }, { "epoch": 0.1274954149714358, "grad_norm": 1.2194700241088867, "learning_rate": 2.1240865654862282e-05, "loss": 0.3531, "step": 3024 }, { "epoch": 0.12753757615363534, "grad_norm": 0.7917352318763733, "learning_rate": 2.1247892074198987e-05, "loss": 0.4685, "step": 3025 }, { "epoch": 0.1275797373358349, "grad_norm": 0.8472031950950623, "learning_rate": 2.1254918493535695e-05, "loss": 0.6758, "step": 3026 }, { "epoch": 0.12762189851803443, "grad_norm": 1.163489818572998, "learning_rate": 2.12619449128724e-05, "loss": 1.0855, "step": 3027 }, { "epoch": 0.127664059700234, "grad_norm": 0.5582273006439209, "learning_rate": 2.1268971332209107e-05, "loss": 0.1565, "step": 3028 }, { "epoch": 0.12770622088243355, "grad_norm": 0.6661915183067322, "learning_rate": 2.127599775154581e-05, "loss": 0.668, "step": 3029 }, { "epoch": 0.1277483820646331, "grad_norm": 0.5912653207778931, "learning_rate": 2.128302417088252e-05, "loss": 0.1862, "step": 3030 }, { "epoch": 0.12779054324683264, "grad_norm": 1.0522613525390625, "learning_rate": 2.1290050590219227e-05, "loss": 0.2897, "step": 3031 }, { "epoch": 0.1278327044290322, "grad_norm": 1.948025107383728, "learning_rate": 2.129707700955593e-05, "loss": 0.8047, "step": 3032 }, { "epoch": 0.12787486561123174, "grad_norm": 1.3193504810333252, "learning_rate": 2.130410342889264e-05, "loss": 0.4035, "step": 3033 }, { "epoch": 0.12791702679343128, "grad_norm": 1.085487723350525, "learning_rate": 2.1311129848229343e-05, "loss": 0.2262, "step": 3034 }, { "epoch": 0.12795918797563083, "grad_norm": 0.8356585502624512, "learning_rate": 2.131815626756605e-05, "loss": 0.2695, "step": 3035 }, { "epoch": 0.12800134915783037, "grad_norm": 0.8050744533538818, "learning_rate": 2.1325182686902755e-05, "loss": 0.1751, "step": 3036 }, { "epoch": 0.12804351034002995, "grad_norm": 0.695530354976654, "learning_rate": 2.1332209106239463e-05, "loss": 0.2094, "step": 3037 }, { "epoch": 0.1280856715222295, "grad_norm": 0.9784587025642395, "learning_rate": 2.1339235525576168e-05, "loss": 0.6062, "step": 3038 }, { "epoch": 0.12812783270442904, "grad_norm": 0.5208792090415955, "learning_rate": 2.1346261944912872e-05, "loss": 0.1377, "step": 3039 }, { "epoch": 0.12816999388662859, "grad_norm": 0.41588687896728516, "learning_rate": 2.135328836424958e-05, "loss": 0.1204, "step": 3040 }, { "epoch": 0.12821215506882813, "grad_norm": 0.7031230926513672, "learning_rate": 2.1360314783586284e-05, "loss": 0.16, "step": 3041 }, { "epoch": 0.12825431625102768, "grad_norm": 1.1357964277267456, "learning_rate": 2.1367341202922992e-05, "loss": 0.4072, "step": 3042 }, { "epoch": 0.12829647743322722, "grad_norm": 0.7601960301399231, "learning_rate": 2.1374367622259696e-05, "loss": 0.2617, "step": 3043 }, { "epoch": 0.12833863861542677, "grad_norm": 2.3407843112945557, "learning_rate": 2.1381394041596404e-05, "loss": 0.7966, "step": 3044 }, { "epoch": 0.12838079979762632, "grad_norm": 0.7015318274497986, "learning_rate": 2.138842046093311e-05, "loss": 0.2099, "step": 3045 }, { "epoch": 0.12842296097982586, "grad_norm": 1.1577609777450562, "learning_rate": 2.1395446880269816e-05, "loss": 0.1897, "step": 3046 }, { "epoch": 0.12846512216202544, "grad_norm": 0.9410524964332581, "learning_rate": 2.140247329960652e-05, "loss": 0.2585, "step": 3047 }, { "epoch": 0.12850728334422498, "grad_norm": 0.9118687510490417, "learning_rate": 2.1409499718943225e-05, "loss": 0.519, "step": 3048 }, { "epoch": 0.12854944452642453, "grad_norm": 0.8941906094551086, "learning_rate": 2.1416526138279933e-05, "loss": 0.3091, "step": 3049 }, { "epoch": 0.12859160570862407, "grad_norm": 0.8048597574234009, "learning_rate": 2.1423552557616637e-05, "loss": 0.2549, "step": 3050 }, { "epoch": 0.12863376689082362, "grad_norm": 1.0068713426589966, "learning_rate": 2.1430578976953345e-05, "loss": 0.2281, "step": 3051 }, { "epoch": 0.12867592807302317, "grad_norm": 0.7421866059303284, "learning_rate": 2.143760539629005e-05, "loss": 0.1664, "step": 3052 }, { "epoch": 0.1287180892552227, "grad_norm": 0.9702873229980469, "learning_rate": 2.1444631815626757e-05, "loss": 0.4637, "step": 3053 }, { "epoch": 0.12876025043742226, "grad_norm": 1.653759479522705, "learning_rate": 2.1451658234963465e-05, "loss": 0.9038, "step": 3054 }, { "epoch": 0.1288024116196218, "grad_norm": 0.71996009349823, "learning_rate": 2.145868465430017e-05, "loss": 0.417, "step": 3055 }, { "epoch": 0.12884457280182135, "grad_norm": 0.932466983795166, "learning_rate": 2.1465711073636877e-05, "loss": 0.4217, "step": 3056 }, { "epoch": 0.12888673398402092, "grad_norm": 0.6773529052734375, "learning_rate": 2.147273749297358e-05, "loss": 0.1181, "step": 3057 }, { "epoch": 0.12892889516622047, "grad_norm": 0.5415958166122437, "learning_rate": 2.147976391231029e-05, "loss": 0.1116, "step": 3058 }, { "epoch": 0.12897105634842002, "grad_norm": 0.8252147436141968, "learning_rate": 2.1486790331646994e-05, "loss": 0.4156, "step": 3059 }, { "epoch": 0.12901321753061956, "grad_norm": 0.8614987730979919, "learning_rate": 2.14938167509837e-05, "loss": 0.5913, "step": 3060 }, { "epoch": 0.1290553787128191, "grad_norm": 0.5663326382637024, "learning_rate": 2.1500843170320406e-05, "loss": 0.2975, "step": 3061 }, { "epoch": 0.12909753989501865, "grad_norm": 1.153889536857605, "learning_rate": 2.150786958965711e-05, "loss": 0.4941, "step": 3062 }, { "epoch": 0.1291397010772182, "grad_norm": 0.6387035846710205, "learning_rate": 2.1514896008993818e-05, "loss": 0.2452, "step": 3063 }, { "epoch": 0.12918186225941775, "grad_norm": 0.7821643948554993, "learning_rate": 2.1521922428330523e-05, "loss": 0.7539, "step": 3064 }, { "epoch": 0.1292240234416173, "grad_norm": 0.5406206846237183, "learning_rate": 2.152894884766723e-05, "loss": 0.151, "step": 3065 }, { "epoch": 0.12926618462381687, "grad_norm": 0.9373726844787598, "learning_rate": 2.1535975267003935e-05, "loss": 0.2117, "step": 3066 }, { "epoch": 0.1293083458060164, "grad_norm": 0.9727888107299805, "learning_rate": 2.1543001686340643e-05, "loss": 0.3971, "step": 3067 }, { "epoch": 0.12935050698821596, "grad_norm": 0.5707845091819763, "learning_rate": 2.1550028105677347e-05, "loss": 0.1657, "step": 3068 }, { "epoch": 0.1293926681704155, "grad_norm": 0.7589766383171082, "learning_rate": 2.1557054525014055e-05, "loss": 0.6139, "step": 3069 }, { "epoch": 0.12943482935261505, "grad_norm": 0.5283305644989014, "learning_rate": 2.156408094435076e-05, "loss": 0.157, "step": 3070 }, { "epoch": 0.1294769905348146, "grad_norm": 0.7385545372962952, "learning_rate": 2.1571107363687464e-05, "loss": 0.3029, "step": 3071 }, { "epoch": 0.12951915171701414, "grad_norm": 2.1031653881073, "learning_rate": 2.157813378302417e-05, "loss": 0.7134, "step": 3072 }, { "epoch": 0.1295613128992137, "grad_norm": 1.0725138187408447, "learning_rate": 2.1585160202360876e-05, "loss": 0.5671, "step": 3073 }, { "epoch": 0.12960347408141323, "grad_norm": 0.5061240792274475, "learning_rate": 2.1592186621697583e-05, "loss": 0.1922, "step": 3074 }, { "epoch": 0.12964563526361278, "grad_norm": 1.153135061264038, "learning_rate": 2.159921304103429e-05, "loss": 0.4973, "step": 3075 }, { "epoch": 0.12968779644581235, "grad_norm": 0.7053271532058716, "learning_rate": 2.1606239460370996e-05, "loss": 0.339, "step": 3076 }, { "epoch": 0.1297299576280119, "grad_norm": 0.8344533443450928, "learning_rate": 2.1613265879707703e-05, "loss": 0.2144, "step": 3077 }, { "epoch": 0.12977211881021145, "grad_norm": 0.9614937901496887, "learning_rate": 2.1620292299044408e-05, "loss": 0.5084, "step": 3078 }, { "epoch": 0.129814279992411, "grad_norm": 3.1122472286224365, "learning_rate": 2.1627318718381116e-05, "loss": 0.4968, "step": 3079 }, { "epoch": 0.12985644117461054, "grad_norm": 1.1264705657958984, "learning_rate": 2.163434513771782e-05, "loss": 0.5023, "step": 3080 }, { "epoch": 0.12989860235681008, "grad_norm": 0.8235370516777039, "learning_rate": 2.1641371557054528e-05, "loss": 0.2674, "step": 3081 }, { "epoch": 0.12994076353900963, "grad_norm": 0.9406670928001404, "learning_rate": 2.1648397976391232e-05, "loss": 0.2613, "step": 3082 }, { "epoch": 0.12998292472120918, "grad_norm": 0.6024072766304016, "learning_rate": 2.165542439572794e-05, "loss": 0.4397, "step": 3083 }, { "epoch": 0.13002508590340872, "grad_norm": 1.1091481447219849, "learning_rate": 2.1662450815064644e-05, "loss": 0.3047, "step": 3084 }, { "epoch": 0.13006724708560827, "grad_norm": 1.0398632287979126, "learning_rate": 2.166947723440135e-05, "loss": 0.7737, "step": 3085 }, { "epoch": 0.13010940826780784, "grad_norm": 0.6634765267372131, "learning_rate": 2.1676503653738057e-05, "loss": 0.3287, "step": 3086 }, { "epoch": 0.1301515694500074, "grad_norm": 0.8768291473388672, "learning_rate": 2.168353007307476e-05, "loss": 0.6788, "step": 3087 }, { "epoch": 0.13019373063220693, "grad_norm": 1.1576783657073975, "learning_rate": 2.169055649241147e-05, "loss": 0.6809, "step": 3088 }, { "epoch": 0.13023589181440648, "grad_norm": 1.1530929803848267, "learning_rate": 2.1697582911748173e-05, "loss": 0.6475, "step": 3089 }, { "epoch": 0.13027805299660603, "grad_norm": 0.4908033013343811, "learning_rate": 2.170460933108488e-05, "loss": 0.2226, "step": 3090 }, { "epoch": 0.13032021417880557, "grad_norm": 0.6062421202659607, "learning_rate": 2.1711635750421585e-05, "loss": 0.3509, "step": 3091 }, { "epoch": 0.13036237536100512, "grad_norm": 0.9222788214683533, "learning_rate": 2.1718662169758293e-05, "loss": 0.3797, "step": 3092 }, { "epoch": 0.13040453654320466, "grad_norm": 0.9052474498748779, "learning_rate": 2.1725688589094997e-05, "loss": 0.3179, "step": 3093 }, { "epoch": 0.1304466977254042, "grad_norm": 1.047889232635498, "learning_rate": 2.1732715008431702e-05, "loss": 0.233, "step": 3094 }, { "epoch": 0.13048885890760378, "grad_norm": 1.902291178703308, "learning_rate": 2.173974142776841e-05, "loss": 0.5153, "step": 3095 }, { "epoch": 0.13053102008980333, "grad_norm": 0.9896543025970459, "learning_rate": 2.1746767847105114e-05, "loss": 0.6596, "step": 3096 }, { "epoch": 0.13057318127200288, "grad_norm": 1.2218831777572632, "learning_rate": 2.1753794266441822e-05, "loss": 0.2823, "step": 3097 }, { "epoch": 0.13061534245420242, "grad_norm": 0.8042402863502502, "learning_rate": 2.176082068577853e-05, "loss": 0.7225, "step": 3098 }, { "epoch": 0.13065750363640197, "grad_norm": 0.8912374377250671, "learning_rate": 2.1767847105115234e-05, "loss": 0.2172, "step": 3099 }, { "epoch": 0.1306996648186015, "grad_norm": 0.671735405921936, "learning_rate": 2.1774873524451942e-05, "loss": 0.1775, "step": 3100 }, { "epoch": 0.13074182600080106, "grad_norm": 0.8769001960754395, "learning_rate": 2.1781899943788646e-05, "loss": 0.1758, "step": 3101 }, { "epoch": 0.1307839871830006, "grad_norm": 2.4098074436187744, "learning_rate": 2.1788926363125354e-05, "loss": 0.3439, "step": 3102 }, { "epoch": 0.13082614836520015, "grad_norm": 0.8451292514801025, "learning_rate": 2.179595278246206e-05, "loss": 0.248, "step": 3103 }, { "epoch": 0.1308683095473997, "grad_norm": 1.0502599477767944, "learning_rate": 2.1802979201798766e-05, "loss": 0.6307, "step": 3104 }, { "epoch": 0.13091047072959927, "grad_norm": 0.47033244371414185, "learning_rate": 2.181000562113547e-05, "loss": 0.2618, "step": 3105 }, { "epoch": 0.13095263191179882, "grad_norm": 1.1488367319107056, "learning_rate": 2.181703204047218e-05, "loss": 0.2535, "step": 3106 }, { "epoch": 0.13099479309399836, "grad_norm": 0.6509972810745239, "learning_rate": 2.1824058459808883e-05, "loss": 0.3606, "step": 3107 }, { "epoch": 0.1310369542761979, "grad_norm": 0.6853072047233582, "learning_rate": 2.1831084879145587e-05, "loss": 0.195, "step": 3108 }, { "epoch": 0.13107911545839745, "grad_norm": 0.5000185370445251, "learning_rate": 2.1838111298482295e-05, "loss": 0.1508, "step": 3109 }, { "epoch": 0.131121276640597, "grad_norm": 1.0251256227493286, "learning_rate": 2.1845137717819e-05, "loss": 0.5383, "step": 3110 }, { "epoch": 0.13116343782279655, "grad_norm": 0.8754714727401733, "learning_rate": 2.1852164137155707e-05, "loss": 0.3001, "step": 3111 }, { "epoch": 0.1312055990049961, "grad_norm": 0.6859191060066223, "learning_rate": 2.185919055649241e-05, "loss": 0.5387, "step": 3112 }, { "epoch": 0.13124776018719564, "grad_norm": 1.1568658351898193, "learning_rate": 2.186621697582912e-05, "loss": 0.2917, "step": 3113 }, { "epoch": 0.13128992136939518, "grad_norm": 0.5042809247970581, "learning_rate": 2.1873243395165824e-05, "loss": 0.1661, "step": 3114 }, { "epoch": 0.13133208255159476, "grad_norm": 0.8357123732566833, "learning_rate": 2.188026981450253e-05, "loss": 0.5184, "step": 3115 }, { "epoch": 0.1313742437337943, "grad_norm": 0.658749520778656, "learning_rate": 2.1887296233839236e-05, "loss": 0.1666, "step": 3116 }, { "epoch": 0.13141640491599385, "grad_norm": 0.4882674217224121, "learning_rate": 2.189432265317594e-05, "loss": 0.1121, "step": 3117 }, { "epoch": 0.1314585660981934, "grad_norm": 0.8305597901344299, "learning_rate": 2.1901349072512648e-05, "loss": 0.2224, "step": 3118 }, { "epoch": 0.13150072728039294, "grad_norm": 0.6591949462890625, "learning_rate": 2.1908375491849352e-05, "loss": 0.1513, "step": 3119 }, { "epoch": 0.1315428884625925, "grad_norm": 0.8742942214012146, "learning_rate": 2.1915401911186064e-05, "loss": 0.3799, "step": 3120 }, { "epoch": 0.13158504964479203, "grad_norm": 0.6357071399688721, "learning_rate": 2.1922428330522768e-05, "loss": 0.2019, "step": 3121 }, { "epoch": 0.13162721082699158, "grad_norm": 1.1266251802444458, "learning_rate": 2.1929454749859472e-05, "loss": 0.2999, "step": 3122 }, { "epoch": 0.13166937200919113, "grad_norm": 1.4860841035842896, "learning_rate": 2.193648116919618e-05, "loss": 1.1758, "step": 3123 }, { "epoch": 0.13171153319139067, "grad_norm": 0.9590775966644287, "learning_rate": 2.1943507588532885e-05, "loss": 0.289, "step": 3124 }, { "epoch": 0.13175369437359025, "grad_norm": 0.7424389719963074, "learning_rate": 2.1950534007869592e-05, "loss": 0.3273, "step": 3125 }, { "epoch": 0.1317958555557898, "grad_norm": 0.6862459182739258, "learning_rate": 2.1957560427206297e-05, "loss": 0.4609, "step": 3126 }, { "epoch": 0.13183801673798934, "grad_norm": 1.738623857498169, "learning_rate": 2.1964586846543005e-05, "loss": 0.1956, "step": 3127 }, { "epoch": 0.13188017792018888, "grad_norm": 1.391160249710083, "learning_rate": 2.197161326587971e-05, "loss": 1.2443, "step": 3128 }, { "epoch": 0.13192233910238843, "grad_norm": 0.8648914098739624, "learning_rate": 2.1978639685216417e-05, "loss": 0.348, "step": 3129 }, { "epoch": 0.13196450028458798, "grad_norm": 1.0873527526855469, "learning_rate": 2.198566610455312e-05, "loss": 0.2376, "step": 3130 }, { "epoch": 0.13200666146678752, "grad_norm": 1.0272351503372192, "learning_rate": 2.1992692523889826e-05, "loss": 0.2857, "step": 3131 }, { "epoch": 0.13204882264898707, "grad_norm": 0.7908236384391785, "learning_rate": 2.1999718943226533e-05, "loss": 0.6244, "step": 3132 }, { "epoch": 0.13209098383118661, "grad_norm": 0.841439962387085, "learning_rate": 2.2006745362563238e-05, "loss": 0.4466, "step": 3133 }, { "epoch": 0.1321331450133862, "grad_norm": 0.812468409538269, "learning_rate": 2.2013771781899945e-05, "loss": 0.4104, "step": 3134 }, { "epoch": 0.13217530619558573, "grad_norm": 0.3997057378292084, "learning_rate": 2.202079820123665e-05, "loss": 0.1481, "step": 3135 }, { "epoch": 0.13221746737778528, "grad_norm": 3.4037883281707764, "learning_rate": 2.2027824620573358e-05, "loss": 0.6929, "step": 3136 }, { "epoch": 0.13225962855998483, "grad_norm": 0.8383671045303345, "learning_rate": 2.2034851039910062e-05, "loss": 0.2386, "step": 3137 }, { "epoch": 0.13230178974218437, "grad_norm": 1.1392772197723389, "learning_rate": 2.2041877459246766e-05, "loss": 0.1328, "step": 3138 }, { "epoch": 0.13234395092438392, "grad_norm": 1.1988835334777832, "learning_rate": 2.2048903878583474e-05, "loss": 0.1402, "step": 3139 }, { "epoch": 0.13238611210658346, "grad_norm": 0.7728553414344788, "learning_rate": 2.205593029792018e-05, "loss": 0.2016, "step": 3140 }, { "epoch": 0.132428273288783, "grad_norm": 0.718893826007843, "learning_rate": 2.2062956717256886e-05, "loss": 0.3879, "step": 3141 }, { "epoch": 0.13247043447098256, "grad_norm": 0.7376164793968201, "learning_rate": 2.206998313659359e-05, "loss": 0.2561, "step": 3142 }, { "epoch": 0.1325125956531821, "grad_norm": 0.5444796085357666, "learning_rate": 2.2077009555930302e-05, "loss": 0.1878, "step": 3143 }, { "epoch": 0.13255475683538168, "grad_norm": 0.8929121494293213, "learning_rate": 2.2084035975267006e-05, "loss": 0.2322, "step": 3144 }, { "epoch": 0.13259691801758122, "grad_norm": 0.8591914772987366, "learning_rate": 2.209106239460371e-05, "loss": 0.6851, "step": 3145 }, { "epoch": 0.13263907919978077, "grad_norm": 1.2459845542907715, "learning_rate": 2.209808881394042e-05, "loss": 0.2579, "step": 3146 }, { "epoch": 0.13268124038198031, "grad_norm": 0.6150601506233215, "learning_rate": 2.2105115233277123e-05, "loss": 0.1869, "step": 3147 }, { "epoch": 0.13272340156417986, "grad_norm": 1.1217936277389526, "learning_rate": 2.211214165261383e-05, "loss": 0.24, "step": 3148 }, { "epoch": 0.1327655627463794, "grad_norm": 0.8694982528686523, "learning_rate": 2.2119168071950535e-05, "loss": 0.4654, "step": 3149 }, { "epoch": 0.13280772392857895, "grad_norm": 1.0624655485153198, "learning_rate": 2.2126194491287243e-05, "loss": 0.632, "step": 3150 }, { "epoch": 0.1328498851107785, "grad_norm": 2.7466697692871094, "learning_rate": 2.2133220910623947e-05, "loss": 0.6037, "step": 3151 }, { "epoch": 0.13289204629297804, "grad_norm": 0.7879027724266052, "learning_rate": 2.214024732996065e-05, "loss": 0.3968, "step": 3152 }, { "epoch": 0.1329342074751776, "grad_norm": 0.6098812222480774, "learning_rate": 2.214727374929736e-05, "loss": 0.1717, "step": 3153 }, { "epoch": 0.13297636865737716, "grad_norm": 0.880962073802948, "learning_rate": 2.2154300168634064e-05, "loss": 0.8394, "step": 3154 }, { "epoch": 0.1330185298395767, "grad_norm": 1.2672971487045288, "learning_rate": 2.216132658797077e-05, "loss": 1.0354, "step": 3155 }, { "epoch": 0.13306069102177626, "grad_norm": 0.7798324227333069, "learning_rate": 2.2168353007307476e-05, "loss": 0.2817, "step": 3156 }, { "epoch": 0.1331028522039758, "grad_norm": 1.0448247194290161, "learning_rate": 2.2175379426644184e-05, "loss": 0.3381, "step": 3157 }, { "epoch": 0.13314501338617535, "grad_norm": 1.2161600589752197, "learning_rate": 2.2182405845980888e-05, "loss": 0.5738, "step": 3158 }, { "epoch": 0.1331871745683749, "grad_norm": 0.7903310060501099, "learning_rate": 2.2189432265317596e-05, "loss": 0.6342, "step": 3159 }, { "epoch": 0.13322933575057444, "grad_norm": 1.0768953561782837, "learning_rate": 2.21964586846543e-05, "loss": 0.2879, "step": 3160 }, { "epoch": 0.133271496932774, "grad_norm": 0.8164755702018738, "learning_rate": 2.2203485103991005e-05, "loss": 0.5809, "step": 3161 }, { "epoch": 0.13331365811497353, "grad_norm": 0.928809404373169, "learning_rate": 2.2210511523327713e-05, "loss": 0.4484, "step": 3162 }, { "epoch": 0.1333558192971731, "grad_norm": 0.4993686079978943, "learning_rate": 2.2217537942664417e-05, "loss": 0.1233, "step": 3163 }, { "epoch": 0.13339798047937265, "grad_norm": 0.6673113703727722, "learning_rate": 2.2224564362001125e-05, "loss": 0.5385, "step": 3164 }, { "epoch": 0.1334401416615722, "grad_norm": 1.0445938110351562, "learning_rate": 2.2231590781337833e-05, "loss": 0.1999, "step": 3165 }, { "epoch": 0.13348230284377174, "grad_norm": 0.7007851600646973, "learning_rate": 2.2238617200674537e-05, "loss": 0.2289, "step": 3166 }, { "epoch": 0.1335244640259713, "grad_norm": 0.6306040287017822, "learning_rate": 2.2245643620011245e-05, "loss": 0.1856, "step": 3167 }, { "epoch": 0.13356662520817084, "grad_norm": 1.207157015800476, "learning_rate": 2.225267003934795e-05, "loss": 0.4982, "step": 3168 }, { "epoch": 0.13360878639037038, "grad_norm": 1.2139054536819458, "learning_rate": 2.2259696458684657e-05, "loss": 0.2896, "step": 3169 }, { "epoch": 0.13365094757256993, "grad_norm": 0.9359455108642578, "learning_rate": 2.226672287802136e-05, "loss": 0.2012, "step": 3170 }, { "epoch": 0.13369310875476947, "grad_norm": 1.0528563261032104, "learning_rate": 2.227374929735807e-05, "loss": 0.1708, "step": 3171 }, { "epoch": 0.13373526993696902, "grad_norm": 0.8659073114395142, "learning_rate": 2.2280775716694773e-05, "loss": 0.1648, "step": 3172 }, { "epoch": 0.1337774311191686, "grad_norm": 0.7911582589149475, "learning_rate": 2.228780213603148e-05, "loss": 0.766, "step": 3173 }, { "epoch": 0.13381959230136814, "grad_norm": 0.9988560676574707, "learning_rate": 2.2294828555368186e-05, "loss": 0.3112, "step": 3174 }, { "epoch": 0.13386175348356769, "grad_norm": 0.8067048192024231, "learning_rate": 2.230185497470489e-05, "loss": 0.2997, "step": 3175 }, { "epoch": 0.13390391466576723, "grad_norm": 0.5778231620788574, "learning_rate": 2.2308881394041598e-05, "loss": 0.2145, "step": 3176 }, { "epoch": 0.13394607584796678, "grad_norm": 2.134230375289917, "learning_rate": 2.2315907813378302e-05, "loss": 0.9435, "step": 3177 }, { "epoch": 0.13398823703016632, "grad_norm": 0.7148623466491699, "learning_rate": 2.232293423271501e-05, "loss": 0.2674, "step": 3178 }, { "epoch": 0.13403039821236587, "grad_norm": 1.4070594310760498, "learning_rate": 2.2329960652051714e-05, "loss": 0.3559, "step": 3179 }, { "epoch": 0.13407255939456542, "grad_norm": 0.9701237678527832, "learning_rate": 2.2336987071388422e-05, "loss": 0.5368, "step": 3180 }, { "epoch": 0.13411472057676496, "grad_norm": 0.5790089964866638, "learning_rate": 2.2344013490725127e-05, "loss": 0.0891, "step": 3181 }, { "epoch": 0.1341568817589645, "grad_norm": 1.5062719583511353, "learning_rate": 2.2351039910061834e-05, "loss": 1.1878, "step": 3182 }, { "epoch": 0.13419904294116408, "grad_norm": 0.9258183836936951, "learning_rate": 2.235806632939854e-05, "loss": 0.1737, "step": 3183 }, { "epoch": 0.13424120412336363, "grad_norm": 1.194156289100647, "learning_rate": 2.2365092748735243e-05, "loss": 0.9168, "step": 3184 }, { "epoch": 0.13428336530556317, "grad_norm": 0.958690345287323, "learning_rate": 2.237211916807195e-05, "loss": 0.489, "step": 3185 }, { "epoch": 0.13432552648776272, "grad_norm": 0.8525042533874512, "learning_rate": 2.2379145587408655e-05, "loss": 0.2587, "step": 3186 }, { "epoch": 0.13436768766996227, "grad_norm": 0.6434085965156555, "learning_rate": 2.2386172006745363e-05, "loss": 0.3306, "step": 3187 }, { "epoch": 0.1344098488521618, "grad_norm": 0.5308146476745605, "learning_rate": 2.239319842608207e-05, "loss": 0.2895, "step": 3188 }, { "epoch": 0.13445201003436136, "grad_norm": 0.8079679012298584, "learning_rate": 2.2400224845418775e-05, "loss": 0.6869, "step": 3189 }, { "epoch": 0.1344941712165609, "grad_norm": 1.1054980754852295, "learning_rate": 2.2407251264755483e-05, "loss": 0.8112, "step": 3190 }, { "epoch": 0.13453633239876045, "grad_norm": 1.3575745820999146, "learning_rate": 2.2414277684092187e-05, "loss": 0.7177, "step": 3191 }, { "epoch": 0.13457849358096002, "grad_norm": 0.7245213985443115, "learning_rate": 2.2421304103428895e-05, "loss": 0.246, "step": 3192 }, { "epoch": 0.13462065476315957, "grad_norm": 0.9092416763305664, "learning_rate": 2.24283305227656e-05, "loss": 0.2437, "step": 3193 }, { "epoch": 0.13466281594535912, "grad_norm": 0.6779907941818237, "learning_rate": 2.2435356942102307e-05, "loss": 0.4972, "step": 3194 }, { "epoch": 0.13470497712755866, "grad_norm": 0.9210502505302429, "learning_rate": 2.2442383361439012e-05, "loss": 0.2474, "step": 3195 }, { "epoch": 0.1347471383097582, "grad_norm": 1.3908965587615967, "learning_rate": 2.244940978077572e-05, "loss": 1.1893, "step": 3196 }, { "epoch": 0.13478929949195775, "grad_norm": 0.798539936542511, "learning_rate": 2.2456436200112424e-05, "loss": 0.4075, "step": 3197 }, { "epoch": 0.1348314606741573, "grad_norm": 1.148655652999878, "learning_rate": 2.246346261944913e-05, "loss": 0.1597, "step": 3198 }, { "epoch": 0.13487362185635685, "grad_norm": 0.7803923487663269, "learning_rate": 2.2470489038785836e-05, "loss": 0.9256, "step": 3199 }, { "epoch": 0.1349157830385564, "grad_norm": 0.8226580619812012, "learning_rate": 2.247751545812254e-05, "loss": 0.4571, "step": 3200 }, { "epoch": 0.13495794422075594, "grad_norm": 0.5814884901046753, "learning_rate": 2.248454187745925e-05, "loss": 0.1794, "step": 3201 }, { "epoch": 0.1350001054029555, "grad_norm": 1.8805553913116455, "learning_rate": 2.2491568296795953e-05, "loss": 0.9665, "step": 3202 }, { "epoch": 0.13504226658515506, "grad_norm": 0.6834903359413147, "learning_rate": 2.249859471613266e-05, "loss": 0.1977, "step": 3203 }, { "epoch": 0.1350844277673546, "grad_norm": 0.8539807796478271, "learning_rate": 2.2505621135469365e-05, "loss": 0.2391, "step": 3204 }, { "epoch": 0.13512658894955415, "grad_norm": 0.5788166522979736, "learning_rate": 2.2512647554806073e-05, "loss": 0.256, "step": 3205 }, { "epoch": 0.1351687501317537, "grad_norm": 0.7742444276809692, "learning_rate": 2.2519673974142777e-05, "loss": 0.4291, "step": 3206 }, { "epoch": 0.13521091131395324, "grad_norm": 0.9629856944084167, "learning_rate": 2.252670039347948e-05, "loss": 0.1261, "step": 3207 }, { "epoch": 0.1352530724961528, "grad_norm": 0.8214014172554016, "learning_rate": 2.253372681281619e-05, "loss": 0.0954, "step": 3208 }, { "epoch": 0.13529523367835233, "grad_norm": 0.6739509105682373, "learning_rate": 2.2540753232152894e-05, "loss": 0.665, "step": 3209 }, { "epoch": 0.13533739486055188, "grad_norm": 0.42667892575263977, "learning_rate": 2.2547779651489605e-05, "loss": 0.1122, "step": 3210 }, { "epoch": 0.13537955604275143, "grad_norm": 0.8779627680778503, "learning_rate": 2.255480607082631e-05, "loss": 0.249, "step": 3211 }, { "epoch": 0.135421717224951, "grad_norm": 0.7596721053123474, "learning_rate": 2.2561832490163014e-05, "loss": 0.226, "step": 3212 }, { "epoch": 0.13546387840715055, "grad_norm": 1.1197324991226196, "learning_rate": 2.256885890949972e-05, "loss": 1.1227, "step": 3213 }, { "epoch": 0.1355060395893501, "grad_norm": 0.9928833246231079, "learning_rate": 2.2575885328836426e-05, "loss": 0.8334, "step": 3214 }, { "epoch": 0.13554820077154964, "grad_norm": 0.8240185976028442, "learning_rate": 2.2582911748173134e-05, "loss": 0.2369, "step": 3215 }, { "epoch": 0.13559036195374918, "grad_norm": 0.689079999923706, "learning_rate": 2.2589938167509838e-05, "loss": 0.3285, "step": 3216 }, { "epoch": 0.13563252313594873, "grad_norm": 0.7191972732543945, "learning_rate": 2.2596964586846546e-05, "loss": 0.6438, "step": 3217 }, { "epoch": 0.13567468431814828, "grad_norm": 1.094480276107788, "learning_rate": 2.260399100618325e-05, "loss": 0.2991, "step": 3218 }, { "epoch": 0.13571684550034782, "grad_norm": 0.8801300525665283, "learning_rate": 2.2611017425519958e-05, "loss": 0.2596, "step": 3219 }, { "epoch": 0.13575900668254737, "grad_norm": 0.7770798802375793, "learning_rate": 2.2618043844856662e-05, "loss": 0.713, "step": 3220 }, { "epoch": 0.13580116786474694, "grad_norm": 0.9996645450592041, "learning_rate": 2.2625070264193367e-05, "loss": 0.4231, "step": 3221 }, { "epoch": 0.1358433290469465, "grad_norm": 0.9991057515144348, "learning_rate": 2.2632096683530075e-05, "loss": 0.2413, "step": 3222 }, { "epoch": 0.13588549022914603, "grad_norm": 0.6298640966415405, "learning_rate": 2.263912310286678e-05, "loss": 0.2713, "step": 3223 }, { "epoch": 0.13592765141134558, "grad_norm": 1.015650749206543, "learning_rate": 2.2646149522203487e-05, "loss": 0.4696, "step": 3224 }, { "epoch": 0.13596981259354513, "grad_norm": 0.8680262565612793, "learning_rate": 2.265317594154019e-05, "loss": 0.6703, "step": 3225 }, { "epoch": 0.13601197377574467, "grad_norm": 0.5606177449226379, "learning_rate": 2.26602023608769e-05, "loss": 0.113, "step": 3226 }, { "epoch": 0.13605413495794422, "grad_norm": 0.7808599472045898, "learning_rate": 2.2667228780213603e-05, "loss": 0.6278, "step": 3227 }, { "epoch": 0.13609629614014376, "grad_norm": 1.9560942649841309, "learning_rate": 2.267425519955031e-05, "loss": 0.5447, "step": 3228 }, { "epoch": 0.1361384573223433, "grad_norm": 1.196925163269043, "learning_rate": 2.2681281618887015e-05, "loss": 1.368, "step": 3229 }, { "epoch": 0.13618061850454286, "grad_norm": 0.7557364106178284, "learning_rate": 2.268830803822372e-05, "loss": 0.4332, "step": 3230 }, { "epoch": 0.13622277968674243, "grad_norm": 0.6922465562820435, "learning_rate": 2.2695334457560428e-05, "loss": 0.1019, "step": 3231 }, { "epoch": 0.13626494086894197, "grad_norm": 0.5660123825073242, "learning_rate": 2.2702360876897132e-05, "loss": 0.1966, "step": 3232 }, { "epoch": 0.13630710205114152, "grad_norm": 1.0844255685806274, "learning_rate": 2.2709387296233843e-05, "loss": 0.2382, "step": 3233 }, { "epoch": 0.13634926323334107, "grad_norm": 0.6239710450172424, "learning_rate": 2.2716413715570548e-05, "loss": 0.2267, "step": 3234 }, { "epoch": 0.1363914244155406, "grad_norm": 0.8466508984565735, "learning_rate": 2.2723440134907252e-05, "loss": 0.1772, "step": 3235 }, { "epoch": 0.13643358559774016, "grad_norm": 1.257082462310791, "learning_rate": 2.273046655424396e-05, "loss": 0.2244, "step": 3236 }, { "epoch": 0.1364757467799397, "grad_norm": 1.0144799947738647, "learning_rate": 2.2737492973580664e-05, "loss": 0.3706, "step": 3237 }, { "epoch": 0.13651790796213925, "grad_norm": 0.6510720252990723, "learning_rate": 2.2744519392917372e-05, "loss": 0.2884, "step": 3238 }, { "epoch": 0.1365600691443388, "grad_norm": 0.5750522017478943, "learning_rate": 2.2751545812254076e-05, "loss": 0.2173, "step": 3239 }, { "epoch": 0.13660223032653834, "grad_norm": 0.56597900390625, "learning_rate": 2.2758572231590784e-05, "loss": 0.1574, "step": 3240 }, { "epoch": 0.13664439150873792, "grad_norm": 0.7316819429397583, "learning_rate": 2.276559865092749e-05, "loss": 0.3254, "step": 3241 }, { "epoch": 0.13668655269093746, "grad_norm": 0.9612445831298828, "learning_rate": 2.2772625070264196e-05, "loss": 0.4683, "step": 3242 }, { "epoch": 0.136728713873137, "grad_norm": 0.6293829679489136, "learning_rate": 2.27796514896009e-05, "loss": 0.1912, "step": 3243 }, { "epoch": 0.13677087505533655, "grad_norm": 0.6752811670303345, "learning_rate": 2.2786677908937605e-05, "loss": 0.499, "step": 3244 }, { "epoch": 0.1368130362375361, "grad_norm": 1.3175036907196045, "learning_rate": 2.2793704328274313e-05, "loss": 0.5078, "step": 3245 }, { "epoch": 0.13685519741973565, "grad_norm": 1.7534301280975342, "learning_rate": 2.2800730747611017e-05, "loss": 0.4209, "step": 3246 }, { "epoch": 0.1368973586019352, "grad_norm": 0.7578664422035217, "learning_rate": 2.2807757166947725e-05, "loss": 0.2454, "step": 3247 }, { "epoch": 0.13693951978413474, "grad_norm": 0.5526453256607056, "learning_rate": 2.281478358628443e-05, "loss": 0.1237, "step": 3248 }, { "epoch": 0.13698168096633428, "grad_norm": 0.5352286100387573, "learning_rate": 2.2821810005621137e-05, "loss": 0.1204, "step": 3249 }, { "epoch": 0.13702384214853386, "grad_norm": 0.5409436225891113, "learning_rate": 2.282883642495784e-05, "loss": 0.1607, "step": 3250 }, { "epoch": 0.1370660033307334, "grad_norm": 0.6680830717086792, "learning_rate": 2.2835862844294546e-05, "loss": 0.3018, "step": 3251 }, { "epoch": 0.13710816451293295, "grad_norm": 0.5874895453453064, "learning_rate": 2.2842889263631254e-05, "loss": 0.1599, "step": 3252 }, { "epoch": 0.1371503256951325, "grad_norm": 0.878000795841217, "learning_rate": 2.2849915682967958e-05, "loss": 0.319, "step": 3253 }, { "epoch": 0.13719248687733204, "grad_norm": 0.9035719633102417, "learning_rate": 2.2856942102304666e-05, "loss": 0.2818, "step": 3254 }, { "epoch": 0.1372346480595316, "grad_norm": 0.8865783214569092, "learning_rate": 2.2863968521641374e-05, "loss": 0.1522, "step": 3255 }, { "epoch": 0.13727680924173113, "grad_norm": 0.5241250395774841, "learning_rate": 2.287099494097808e-05, "loss": 0.1001, "step": 3256 }, { "epoch": 0.13731897042393068, "grad_norm": 0.9335011839866638, "learning_rate": 2.2878021360314786e-05, "loss": 0.4859, "step": 3257 }, { "epoch": 0.13736113160613023, "grad_norm": 0.7219457626342773, "learning_rate": 2.288504777965149e-05, "loss": 0.277, "step": 3258 }, { "epoch": 0.13740329278832977, "grad_norm": 0.6266930103302002, "learning_rate": 2.2892074198988198e-05, "loss": 0.3069, "step": 3259 }, { "epoch": 0.13744545397052935, "grad_norm": 1.253381609916687, "learning_rate": 2.2899100618324903e-05, "loss": 0.6348, "step": 3260 }, { "epoch": 0.1374876151527289, "grad_norm": 1.2006731033325195, "learning_rate": 2.290612703766161e-05, "loss": 0.403, "step": 3261 }, { "epoch": 0.13752977633492844, "grad_norm": 0.4367143213748932, "learning_rate": 2.2913153456998315e-05, "loss": 0.1146, "step": 3262 }, { "epoch": 0.13757193751712798, "grad_norm": 0.7914383411407471, "learning_rate": 2.2920179876335023e-05, "loss": 0.214, "step": 3263 }, { "epoch": 0.13761409869932753, "grad_norm": 0.4843639135360718, "learning_rate": 2.2927206295671727e-05, "loss": 0.1907, "step": 3264 }, { "epoch": 0.13765625988152708, "grad_norm": 0.8119702339172363, "learning_rate": 2.293423271500843e-05, "loss": 0.8007, "step": 3265 }, { "epoch": 0.13769842106372662, "grad_norm": 0.7903265953063965, "learning_rate": 2.294125913434514e-05, "loss": 0.3608, "step": 3266 }, { "epoch": 0.13774058224592617, "grad_norm": 0.5896468758583069, "learning_rate": 2.2948285553681843e-05, "loss": 0.199, "step": 3267 }, { "epoch": 0.13778274342812571, "grad_norm": 0.8602905869483948, "learning_rate": 2.295531197301855e-05, "loss": 0.7663, "step": 3268 }, { "epoch": 0.13782490461032526, "grad_norm": 1.6573363542556763, "learning_rate": 2.2962338392355256e-05, "loss": 0.4075, "step": 3269 }, { "epoch": 0.13786706579252483, "grad_norm": 0.8361984491348267, "learning_rate": 2.2969364811691963e-05, "loss": 0.6111, "step": 3270 }, { "epoch": 0.13790922697472438, "grad_norm": 0.7652818560600281, "learning_rate": 2.2976391231028668e-05, "loss": 0.7117, "step": 3271 }, { "epoch": 0.13795138815692393, "grad_norm": 1.228374719619751, "learning_rate": 2.2983417650365376e-05, "loss": 1.1052, "step": 3272 }, { "epoch": 0.13799354933912347, "grad_norm": 0.922357439994812, "learning_rate": 2.299044406970208e-05, "loss": 0.4252, "step": 3273 }, { "epoch": 0.13803571052132302, "grad_norm": 0.949074387550354, "learning_rate": 2.2997470489038784e-05, "loss": 0.6271, "step": 3274 }, { "epoch": 0.13807787170352256, "grad_norm": 1.1130198240280151, "learning_rate": 2.3004496908375492e-05, "loss": 1.2029, "step": 3275 }, { "epoch": 0.1381200328857221, "grad_norm": 1.022278904914856, "learning_rate": 2.3011523327712197e-05, "loss": 0.2608, "step": 3276 }, { "epoch": 0.13816219406792166, "grad_norm": 0.6493013501167297, "learning_rate": 2.3018549747048904e-05, "loss": 0.2555, "step": 3277 }, { "epoch": 0.1382043552501212, "grad_norm": 0.5372380018234253, "learning_rate": 2.3025576166385612e-05, "loss": 0.1378, "step": 3278 }, { "epoch": 0.13824651643232075, "grad_norm": 0.7414752244949341, "learning_rate": 2.3032602585722317e-05, "loss": 0.2968, "step": 3279 }, { "epoch": 0.13828867761452032, "grad_norm": 1.012679100036621, "learning_rate": 2.3039629005059024e-05, "loss": 0.207, "step": 3280 }, { "epoch": 0.13833083879671987, "grad_norm": 1.352759838104248, "learning_rate": 2.304665542439573e-05, "loss": 1.2846, "step": 3281 }, { "epoch": 0.13837299997891941, "grad_norm": 1.9371721744537354, "learning_rate": 2.3053681843732437e-05, "loss": 0.6318, "step": 3282 }, { "epoch": 0.13841516116111896, "grad_norm": 0.7825421690940857, "learning_rate": 2.306070826306914e-05, "loss": 0.6585, "step": 3283 }, { "epoch": 0.1384573223433185, "grad_norm": 0.9071826934814453, "learning_rate": 2.306773468240585e-05, "loss": 0.2805, "step": 3284 }, { "epoch": 0.13849948352551805, "grad_norm": 1.1585158109664917, "learning_rate": 2.3074761101742553e-05, "loss": 0.6756, "step": 3285 }, { "epoch": 0.1385416447077176, "grad_norm": 0.4432256817817688, "learning_rate": 2.308178752107926e-05, "loss": 0.2161, "step": 3286 }, { "epoch": 0.13858380588991714, "grad_norm": 1.0140451192855835, "learning_rate": 2.3088813940415965e-05, "loss": 0.8736, "step": 3287 }, { "epoch": 0.1386259670721167, "grad_norm": 0.8620061874389648, "learning_rate": 2.309584035975267e-05, "loss": 0.1475, "step": 3288 }, { "epoch": 0.13866812825431626, "grad_norm": 0.8112715482711792, "learning_rate": 2.3102866779089377e-05, "loss": 0.4482, "step": 3289 }, { "epoch": 0.1387102894365158, "grad_norm": 0.5906801223754883, "learning_rate": 2.3109893198426082e-05, "loss": 0.3031, "step": 3290 }, { "epoch": 0.13875245061871536, "grad_norm": 0.7740439176559448, "learning_rate": 2.311691961776279e-05, "loss": 0.2797, "step": 3291 }, { "epoch": 0.1387946118009149, "grad_norm": 0.6047173738479614, "learning_rate": 2.3123946037099494e-05, "loss": 0.3294, "step": 3292 }, { "epoch": 0.13883677298311445, "grad_norm": 0.5845008492469788, "learning_rate": 2.3130972456436202e-05, "loss": 0.2057, "step": 3293 }, { "epoch": 0.138878934165314, "grad_norm": 1.1142144203186035, "learning_rate": 2.3137998875772906e-05, "loss": 0.5521, "step": 3294 }, { "epoch": 0.13892109534751354, "grad_norm": 2.933502674102783, "learning_rate": 2.3145025295109614e-05, "loss": 0.4136, "step": 3295 }, { "epoch": 0.1389632565297131, "grad_norm": 0.726539134979248, "learning_rate": 2.315205171444632e-05, "loss": 0.2431, "step": 3296 }, { "epoch": 0.13900541771191263, "grad_norm": 0.7547749280929565, "learning_rate": 2.3159078133783023e-05, "loss": 0.2349, "step": 3297 }, { "epoch": 0.13904757889411218, "grad_norm": 0.7169299721717834, "learning_rate": 2.316610455311973e-05, "loss": 0.6699, "step": 3298 }, { "epoch": 0.13908974007631175, "grad_norm": 0.8434359431266785, "learning_rate": 2.3173130972456435e-05, "loss": 0.1955, "step": 3299 }, { "epoch": 0.1391319012585113, "grad_norm": 0.9060443043708801, "learning_rate": 2.3180157391793143e-05, "loss": 0.2514, "step": 3300 }, { "epoch": 0.13917406244071084, "grad_norm": 0.5102148056030273, "learning_rate": 2.318718381112985e-05, "loss": 0.1427, "step": 3301 }, { "epoch": 0.1392162236229104, "grad_norm": 0.6983977556228638, "learning_rate": 2.3194210230466555e-05, "loss": 0.386, "step": 3302 }, { "epoch": 0.13925838480510994, "grad_norm": 0.9544433355331421, "learning_rate": 2.3201236649803263e-05, "loss": 0.7199, "step": 3303 }, { "epoch": 0.13930054598730948, "grad_norm": 1.8988369703292847, "learning_rate": 2.3208263069139967e-05, "loss": 0.9707, "step": 3304 }, { "epoch": 0.13934270716950903, "grad_norm": 0.4988919794559479, "learning_rate": 2.3215289488476675e-05, "loss": 0.234, "step": 3305 }, { "epoch": 0.13938486835170857, "grad_norm": 0.6638209223747253, "learning_rate": 2.322231590781338e-05, "loss": 0.2572, "step": 3306 }, { "epoch": 0.13942702953390812, "grad_norm": 1.0141276121139526, "learning_rate": 2.3229342327150087e-05, "loss": 0.3861, "step": 3307 }, { "epoch": 0.13946919071610767, "grad_norm": 0.6770129203796387, "learning_rate": 2.323636874648679e-05, "loss": 0.1952, "step": 3308 }, { "epoch": 0.13951135189830724, "grad_norm": 0.4871225357055664, "learning_rate": 2.32433951658235e-05, "loss": 0.1835, "step": 3309 }, { "epoch": 0.13955351308050679, "grad_norm": 0.5932340621948242, "learning_rate": 2.3250421585160204e-05, "loss": 0.161, "step": 3310 }, { "epoch": 0.13959567426270633, "grad_norm": 0.5334405899047852, "learning_rate": 2.3257448004496908e-05, "loss": 0.191, "step": 3311 }, { "epoch": 0.13963783544490588, "grad_norm": 0.6511231064796448, "learning_rate": 2.3264474423833616e-05, "loss": 0.3413, "step": 3312 }, { "epoch": 0.13967999662710542, "grad_norm": 0.8359463214874268, "learning_rate": 2.327150084317032e-05, "loss": 0.6876, "step": 3313 }, { "epoch": 0.13972215780930497, "grad_norm": 0.8410218358039856, "learning_rate": 2.3278527262507028e-05, "loss": 0.7599, "step": 3314 }, { "epoch": 0.13976431899150452, "grad_norm": 0.5969219207763672, "learning_rate": 2.3285553681843732e-05, "loss": 0.1461, "step": 3315 }, { "epoch": 0.13980648017370406, "grad_norm": 0.5553050637245178, "learning_rate": 2.329258010118044e-05, "loss": 0.1693, "step": 3316 }, { "epoch": 0.1398486413559036, "grad_norm": 0.528958261013031, "learning_rate": 2.3299606520517145e-05, "loss": 0.1172, "step": 3317 }, { "epoch": 0.13989080253810318, "grad_norm": 0.875069260597229, "learning_rate": 2.3306632939853852e-05, "loss": 0.275, "step": 3318 }, { "epoch": 0.13993296372030273, "grad_norm": 1.1061468124389648, "learning_rate": 2.3313659359190557e-05, "loss": 0.199, "step": 3319 }, { "epoch": 0.13997512490250227, "grad_norm": 0.9695863723754883, "learning_rate": 2.332068577852726e-05, "loss": 0.7342, "step": 3320 }, { "epoch": 0.14001728608470182, "grad_norm": 1.0436270236968994, "learning_rate": 2.332771219786397e-05, "loss": 0.2633, "step": 3321 }, { "epoch": 0.14005944726690137, "grad_norm": 0.5456156730651855, "learning_rate": 2.3334738617200673e-05, "loss": 0.1023, "step": 3322 }, { "epoch": 0.1401016084491009, "grad_norm": 0.5407366156578064, "learning_rate": 2.3341765036537384e-05, "loss": 0.1521, "step": 3323 }, { "epoch": 0.14014376963130046, "grad_norm": 0.6011607050895691, "learning_rate": 2.334879145587409e-05, "loss": 0.5861, "step": 3324 }, { "epoch": 0.1401859308135, "grad_norm": 0.9578735828399658, "learning_rate": 2.3355817875210793e-05, "loss": 0.6715, "step": 3325 }, { "epoch": 0.14022809199569955, "grad_norm": 1.2128311395645142, "learning_rate": 2.33628442945475e-05, "loss": 0.4606, "step": 3326 }, { "epoch": 0.1402702531778991, "grad_norm": 0.5857946872711182, "learning_rate": 2.3369870713884205e-05, "loss": 0.3716, "step": 3327 }, { "epoch": 0.14031241436009867, "grad_norm": 1.2555298805236816, "learning_rate": 2.3376897133220913e-05, "loss": 0.2981, "step": 3328 }, { "epoch": 0.14035457554229822, "grad_norm": 1.8394795656204224, "learning_rate": 2.3383923552557618e-05, "loss": 0.87, "step": 3329 }, { "epoch": 0.14039673672449776, "grad_norm": 0.6671781539916992, "learning_rate": 2.3390949971894325e-05, "loss": 0.136, "step": 3330 }, { "epoch": 0.1404388979066973, "grad_norm": 0.4977385103702545, "learning_rate": 2.339797639123103e-05, "loss": 0.1327, "step": 3331 }, { "epoch": 0.14048105908889685, "grad_norm": 0.9710026383399963, "learning_rate": 2.3405002810567738e-05, "loss": 1.1385, "step": 3332 }, { "epoch": 0.1405232202710964, "grad_norm": 1.0420395135879517, "learning_rate": 2.3412029229904442e-05, "loss": 0.2644, "step": 3333 }, { "epoch": 0.14056538145329595, "grad_norm": 1.212821125984192, "learning_rate": 2.3419055649241146e-05, "loss": 0.3595, "step": 3334 }, { "epoch": 0.1406075426354955, "grad_norm": 0.6664418578147888, "learning_rate": 2.3426082068577854e-05, "loss": 0.2759, "step": 3335 }, { "epoch": 0.14064970381769504, "grad_norm": 0.7677310705184937, "learning_rate": 2.343310848791456e-05, "loss": 0.755, "step": 3336 }, { "epoch": 0.14069186499989458, "grad_norm": 1.1742640733718872, "learning_rate": 2.3440134907251266e-05, "loss": 0.3886, "step": 3337 }, { "epoch": 0.14073402618209416, "grad_norm": 0.916571855545044, "learning_rate": 2.344716132658797e-05, "loss": 0.2567, "step": 3338 }, { "epoch": 0.1407761873642937, "grad_norm": 1.0332642793655396, "learning_rate": 2.345418774592468e-05, "loss": 0.2438, "step": 3339 }, { "epoch": 0.14081834854649325, "grad_norm": 2.039203643798828, "learning_rate": 2.3461214165261383e-05, "loss": 0.3264, "step": 3340 }, { "epoch": 0.1408605097286928, "grad_norm": 0.6272586584091187, "learning_rate": 2.346824058459809e-05, "loss": 0.3814, "step": 3341 }, { "epoch": 0.14090267091089234, "grad_norm": 0.7862882018089294, "learning_rate": 2.3475267003934795e-05, "loss": 0.439, "step": 3342 }, { "epoch": 0.1409448320930919, "grad_norm": 0.9211741089820862, "learning_rate": 2.34822934232715e-05, "loss": 0.4156, "step": 3343 }, { "epoch": 0.14098699327529143, "grad_norm": 1.299148678779602, "learning_rate": 2.3489319842608207e-05, "loss": 1.2342, "step": 3344 }, { "epoch": 0.14102915445749098, "grad_norm": 0.5953789353370667, "learning_rate": 2.349634626194491e-05, "loss": 0.2714, "step": 3345 }, { "epoch": 0.14107131563969053, "grad_norm": 2.485434055328369, "learning_rate": 2.3503372681281623e-05, "loss": 0.5766, "step": 3346 }, { "epoch": 0.1411134768218901, "grad_norm": 0.9570678472518921, "learning_rate": 2.3510399100618327e-05, "loss": 0.6708, "step": 3347 }, { "epoch": 0.14115563800408965, "grad_norm": 0.4225929379463196, "learning_rate": 2.351742551995503e-05, "loss": 0.0921, "step": 3348 }, { "epoch": 0.1411977991862892, "grad_norm": 0.49640920758247375, "learning_rate": 2.352445193929174e-05, "loss": 0.2399, "step": 3349 }, { "epoch": 0.14123996036848874, "grad_norm": 1.3156778812408447, "learning_rate": 2.3531478358628444e-05, "loss": 0.4873, "step": 3350 }, { "epoch": 0.14128212155068828, "grad_norm": 0.6224313378334045, "learning_rate": 2.353850477796515e-05, "loss": 0.2222, "step": 3351 }, { "epoch": 0.14132428273288783, "grad_norm": 1.37745201587677, "learning_rate": 2.3545531197301856e-05, "loss": 1.0509, "step": 3352 }, { "epoch": 0.14136644391508738, "grad_norm": 0.4442446529865265, "learning_rate": 2.3552557616638564e-05, "loss": 0.1376, "step": 3353 }, { "epoch": 0.14140860509728692, "grad_norm": 0.6762083768844604, "learning_rate": 2.3559584035975268e-05, "loss": 0.6265, "step": 3354 }, { "epoch": 0.14145076627948647, "grad_norm": 0.5105918645858765, "learning_rate": 2.3566610455311976e-05, "loss": 0.1218, "step": 3355 }, { "epoch": 0.141492927461686, "grad_norm": 0.6679373383522034, "learning_rate": 2.357363687464868e-05, "loss": 0.2968, "step": 3356 }, { "epoch": 0.1415350886438856, "grad_norm": 0.4227466285228729, "learning_rate": 2.3580663293985385e-05, "loss": 0.0881, "step": 3357 }, { "epoch": 0.14157724982608513, "grad_norm": 0.9294803738594055, "learning_rate": 2.3587689713322093e-05, "loss": 0.2894, "step": 3358 }, { "epoch": 0.14161941100828468, "grad_norm": 0.8376772403717041, "learning_rate": 2.3594716132658797e-05, "loss": 0.4505, "step": 3359 }, { "epoch": 0.14166157219048423, "grad_norm": 1.142263412475586, "learning_rate": 2.3601742551995505e-05, "loss": 1.2656, "step": 3360 }, { "epoch": 0.14170373337268377, "grad_norm": 0.8750165700912476, "learning_rate": 2.360876897133221e-05, "loss": 0.2544, "step": 3361 }, { "epoch": 0.14174589455488332, "grad_norm": 0.7718451023101807, "learning_rate": 2.3615795390668917e-05, "loss": 0.6412, "step": 3362 }, { "epoch": 0.14178805573708286, "grad_norm": 1.0630894899368286, "learning_rate": 2.362282181000562e-05, "loss": 0.4795, "step": 3363 }, { "epoch": 0.1418302169192824, "grad_norm": 0.8979605436325073, "learning_rate": 2.3629848229342326e-05, "loss": 0.2568, "step": 3364 }, { "epoch": 0.14187237810148196, "grad_norm": 0.946306586265564, "learning_rate": 2.3636874648679033e-05, "loss": 0.2959, "step": 3365 }, { "epoch": 0.1419145392836815, "grad_norm": 0.7920874357223511, "learning_rate": 2.3643901068015738e-05, "loss": 0.5965, "step": 3366 }, { "epoch": 0.14195670046588107, "grad_norm": 0.9475340247154236, "learning_rate": 2.3650927487352446e-05, "loss": 0.3324, "step": 3367 }, { "epoch": 0.14199886164808062, "grad_norm": 0.5404137969017029, "learning_rate": 2.3657953906689153e-05, "loss": 0.1115, "step": 3368 }, { "epoch": 0.14204102283028017, "grad_norm": 1.3975087404251099, "learning_rate": 2.366498032602586e-05, "loss": 1.2497, "step": 3369 }, { "epoch": 0.1420831840124797, "grad_norm": 2.502897024154663, "learning_rate": 2.3672006745362566e-05, "loss": 0.7556, "step": 3370 }, { "epoch": 0.14212534519467926, "grad_norm": 0.9690002799034119, "learning_rate": 2.367903316469927e-05, "loss": 0.1894, "step": 3371 }, { "epoch": 0.1421675063768788, "grad_norm": 0.8945736289024353, "learning_rate": 2.3686059584035978e-05, "loss": 0.3158, "step": 3372 }, { "epoch": 0.14220966755907835, "grad_norm": 0.9047762155532837, "learning_rate": 2.3693086003372682e-05, "loss": 0.2756, "step": 3373 }, { "epoch": 0.1422518287412779, "grad_norm": 1.0926927328109741, "learning_rate": 2.370011242270939e-05, "loss": 0.4884, "step": 3374 }, { "epoch": 0.14229398992347744, "grad_norm": 0.8016920685768127, "learning_rate": 2.3707138842046094e-05, "loss": 0.467, "step": 3375 }, { "epoch": 0.14233615110567702, "grad_norm": 1.3035967350006104, "learning_rate": 2.3714165261382802e-05, "loss": 1.1486, "step": 3376 }, { "epoch": 0.14237831228787656, "grad_norm": 0.6199367046356201, "learning_rate": 2.3721191680719507e-05, "loss": 0.2005, "step": 3377 }, { "epoch": 0.1424204734700761, "grad_norm": 0.8283079862594604, "learning_rate": 2.372821810005621e-05, "loss": 0.3651, "step": 3378 }, { "epoch": 0.14246263465227565, "grad_norm": 0.7193543910980225, "learning_rate": 2.373524451939292e-05, "loss": 0.2665, "step": 3379 }, { "epoch": 0.1425047958344752, "grad_norm": 0.7054423689842224, "learning_rate": 2.3742270938729623e-05, "loss": 0.2541, "step": 3380 }, { "epoch": 0.14254695701667475, "grad_norm": 0.8916118741035461, "learning_rate": 2.374929735806633e-05, "loss": 0.2688, "step": 3381 }, { "epoch": 0.1425891181988743, "grad_norm": 0.7422826886177063, "learning_rate": 2.3756323777403035e-05, "loss": 0.2265, "step": 3382 }, { "epoch": 0.14263127938107384, "grad_norm": 0.7802794575691223, "learning_rate": 2.3763350196739743e-05, "loss": 0.2022, "step": 3383 }, { "epoch": 0.14267344056327338, "grad_norm": 0.6277726888656616, "learning_rate": 2.3770376616076447e-05, "loss": 0.19, "step": 3384 }, { "epoch": 0.14271560174547293, "grad_norm": 2.121464490890503, "learning_rate": 2.3777403035413155e-05, "loss": 0.4403, "step": 3385 }, { "epoch": 0.1427577629276725, "grad_norm": 0.6666181087493896, "learning_rate": 2.378442945474986e-05, "loss": 0.2201, "step": 3386 }, { "epoch": 0.14279992410987205, "grad_norm": 0.4130385220050812, "learning_rate": 2.3791455874086564e-05, "loss": 0.1455, "step": 3387 }, { "epoch": 0.1428420852920716, "grad_norm": 0.7059040069580078, "learning_rate": 2.3798482293423272e-05, "loss": 0.3885, "step": 3388 }, { "epoch": 0.14288424647427114, "grad_norm": 0.6349446177482605, "learning_rate": 2.3805508712759976e-05, "loss": 0.1819, "step": 3389 }, { "epoch": 0.1429264076564707, "grad_norm": 0.68504798412323, "learning_rate": 2.3812535132096684e-05, "loss": 0.7922, "step": 3390 }, { "epoch": 0.14296856883867023, "grad_norm": 0.6840084791183472, "learning_rate": 2.3819561551433392e-05, "loss": 0.2427, "step": 3391 }, { "epoch": 0.14301073002086978, "grad_norm": 0.887122631072998, "learning_rate": 2.38265879707701e-05, "loss": 0.9336, "step": 3392 }, { "epoch": 0.14305289120306933, "grad_norm": 0.6184951663017273, "learning_rate": 2.3833614390106804e-05, "loss": 0.3194, "step": 3393 }, { "epoch": 0.14309505238526887, "grad_norm": 0.6465080380439758, "learning_rate": 2.384064080944351e-05, "loss": 0.1951, "step": 3394 }, { "epoch": 0.14313721356746842, "grad_norm": 0.5745826363563538, "learning_rate": 2.3847667228780216e-05, "loss": 0.2319, "step": 3395 }, { "epoch": 0.143179374749668, "grad_norm": 0.5170155763626099, "learning_rate": 2.385469364811692e-05, "loss": 0.1184, "step": 3396 }, { "epoch": 0.14322153593186754, "grad_norm": 1.447007179260254, "learning_rate": 2.3861720067453628e-05, "loss": 1.248, "step": 3397 }, { "epoch": 0.14326369711406708, "grad_norm": 0.8865573406219482, "learning_rate": 2.3868746486790333e-05, "loss": 0.2413, "step": 3398 }, { "epoch": 0.14330585829626663, "grad_norm": 0.497877836227417, "learning_rate": 2.387577290612704e-05, "loss": 0.1595, "step": 3399 }, { "epoch": 0.14334801947846618, "grad_norm": 0.6821576952934265, "learning_rate": 2.3882799325463745e-05, "loss": 0.7073, "step": 3400 }, { "epoch": 0.14339018066066572, "grad_norm": 0.7762704491615295, "learning_rate": 2.388982574480045e-05, "loss": 0.6823, "step": 3401 }, { "epoch": 0.14343234184286527, "grad_norm": 0.7543767094612122, "learning_rate": 2.3896852164137157e-05, "loss": 0.1227, "step": 3402 }, { "epoch": 0.14347450302506481, "grad_norm": 0.47513818740844727, "learning_rate": 2.390387858347386e-05, "loss": 0.1084, "step": 3403 }, { "epoch": 0.14351666420726436, "grad_norm": 1.454903244972229, "learning_rate": 2.391090500281057e-05, "loss": 0.4067, "step": 3404 }, { "epoch": 0.14355882538946393, "grad_norm": 0.5553354024887085, "learning_rate": 2.3917931422147274e-05, "loss": 0.3696, "step": 3405 }, { "epoch": 0.14360098657166348, "grad_norm": 0.9415899515151978, "learning_rate": 2.392495784148398e-05, "loss": 0.6561, "step": 3406 }, { "epoch": 0.14364314775386303, "grad_norm": 0.866699755191803, "learning_rate": 2.3931984260820686e-05, "loss": 0.6185, "step": 3407 }, { "epoch": 0.14368530893606257, "grad_norm": 1.2236826419830322, "learning_rate": 2.3939010680157394e-05, "loss": 0.997, "step": 3408 }, { "epoch": 0.14372747011826212, "grad_norm": 2.3034868240356445, "learning_rate": 2.3946037099494098e-05, "loss": 0.5173, "step": 3409 }, { "epoch": 0.14376963130046166, "grad_norm": 1.0856653451919556, "learning_rate": 2.3953063518830802e-05, "loss": 0.4184, "step": 3410 }, { "epoch": 0.1438117924826612, "grad_norm": 0.8395697474479675, "learning_rate": 2.396008993816751e-05, "loss": 0.7574, "step": 3411 }, { "epoch": 0.14385395366486076, "grad_norm": 0.6356130242347717, "learning_rate": 2.3967116357504215e-05, "loss": 0.211, "step": 3412 }, { "epoch": 0.1438961148470603, "grad_norm": 1.5552985668182373, "learning_rate": 2.3974142776840926e-05, "loss": 0.9304, "step": 3413 }, { "epoch": 0.14393827602925985, "grad_norm": 0.6362619996070862, "learning_rate": 2.398116919617763e-05, "loss": 0.2232, "step": 3414 }, { "epoch": 0.14398043721145942, "grad_norm": 0.45479318499565125, "learning_rate": 2.3988195615514335e-05, "loss": 0.1626, "step": 3415 }, { "epoch": 0.14402259839365897, "grad_norm": 0.7441087961196899, "learning_rate": 2.3995222034851042e-05, "loss": 0.3647, "step": 3416 }, { "epoch": 0.14406475957585851, "grad_norm": 0.7774801850318909, "learning_rate": 2.4002248454187747e-05, "loss": 0.2894, "step": 3417 }, { "epoch": 0.14410692075805806, "grad_norm": 0.5344486236572266, "learning_rate": 2.4009274873524454e-05, "loss": 0.1562, "step": 3418 }, { "epoch": 0.1441490819402576, "grad_norm": 0.659711480140686, "learning_rate": 2.401630129286116e-05, "loss": 0.2022, "step": 3419 }, { "epoch": 0.14419124312245715, "grad_norm": 0.614648163318634, "learning_rate": 2.4023327712197867e-05, "loss": 0.3848, "step": 3420 }, { "epoch": 0.1442334043046567, "grad_norm": 1.7609487771987915, "learning_rate": 2.403035413153457e-05, "loss": 0.7076, "step": 3421 }, { "epoch": 0.14427556548685624, "grad_norm": 0.8623009920120239, "learning_rate": 2.403738055087128e-05, "loss": 0.5475, "step": 3422 }, { "epoch": 0.1443177266690558, "grad_norm": 0.8566411137580872, "learning_rate": 2.4044406970207983e-05, "loss": 0.7405, "step": 3423 }, { "epoch": 0.14435988785125534, "grad_norm": 0.9757964015007019, "learning_rate": 2.4051433389544688e-05, "loss": 0.2979, "step": 3424 }, { "epoch": 0.1444020490334549, "grad_norm": 1.2141828536987305, "learning_rate": 2.4058459808881395e-05, "loss": 0.3145, "step": 3425 }, { "epoch": 0.14444421021565446, "grad_norm": 0.7416127324104309, "learning_rate": 2.40654862282181e-05, "loss": 0.1254, "step": 3426 }, { "epoch": 0.144486371397854, "grad_norm": 0.9173240065574646, "learning_rate": 2.4072512647554808e-05, "loss": 0.7485, "step": 3427 }, { "epoch": 0.14452853258005355, "grad_norm": 0.837141752243042, "learning_rate": 2.4079539066891512e-05, "loss": 0.1425, "step": 3428 }, { "epoch": 0.1445706937622531, "grad_norm": 0.5605728030204773, "learning_rate": 2.408656548622822e-05, "loss": 0.1832, "step": 3429 }, { "epoch": 0.14461285494445264, "grad_norm": 0.5750010013580322, "learning_rate": 2.4093591905564924e-05, "loss": 0.1205, "step": 3430 }, { "epoch": 0.14465501612665219, "grad_norm": 0.9044188857078552, "learning_rate": 2.4100618324901632e-05, "loss": 0.2135, "step": 3431 }, { "epoch": 0.14469717730885173, "grad_norm": 1.1571139097213745, "learning_rate": 2.4107644744238336e-05, "loss": 0.6077, "step": 3432 }, { "epoch": 0.14473933849105128, "grad_norm": 0.5232909321784973, "learning_rate": 2.411467116357504e-05, "loss": 0.2617, "step": 3433 }, { "epoch": 0.14478149967325082, "grad_norm": 1.0785095691680908, "learning_rate": 2.412169758291175e-05, "loss": 1.1553, "step": 3434 }, { "epoch": 0.1448236608554504, "grad_norm": 0.8861425518989563, "learning_rate": 2.4128724002248453e-05, "loss": 0.8312, "step": 3435 }, { "epoch": 0.14486582203764994, "grad_norm": 1.2578760385513306, "learning_rate": 2.4135750421585164e-05, "loss": 1.0511, "step": 3436 }, { "epoch": 0.1449079832198495, "grad_norm": 1.395623803138733, "learning_rate": 2.414277684092187e-05, "loss": 1.2882, "step": 3437 }, { "epoch": 0.14495014440204904, "grad_norm": 1.355567455291748, "learning_rate": 2.4149803260258573e-05, "loss": 0.9582, "step": 3438 }, { "epoch": 0.14499230558424858, "grad_norm": 0.5588914752006531, "learning_rate": 2.415682967959528e-05, "loss": 0.2515, "step": 3439 }, { "epoch": 0.14503446676644813, "grad_norm": 0.7220130562782288, "learning_rate": 2.4163856098931985e-05, "loss": 0.3811, "step": 3440 }, { "epoch": 0.14507662794864767, "grad_norm": 1.16886305809021, "learning_rate": 2.4170882518268693e-05, "loss": 0.9441, "step": 3441 }, { "epoch": 0.14511878913084722, "grad_norm": 0.6836094260215759, "learning_rate": 2.4177908937605397e-05, "loss": 0.2156, "step": 3442 }, { "epoch": 0.14516095031304677, "grad_norm": 0.8352932333946228, "learning_rate": 2.4184935356942105e-05, "loss": 0.261, "step": 3443 }, { "epoch": 0.14520311149524634, "grad_norm": 0.693513810634613, "learning_rate": 2.419196177627881e-05, "loss": 0.2553, "step": 3444 }, { "epoch": 0.14524527267744589, "grad_norm": 1.4492888450622559, "learning_rate": 2.4198988195615517e-05, "loss": 0.3151, "step": 3445 }, { "epoch": 0.14528743385964543, "grad_norm": 1.0737831592559814, "learning_rate": 2.420601461495222e-05, "loss": 0.4468, "step": 3446 }, { "epoch": 0.14532959504184498, "grad_norm": 0.7426244020462036, "learning_rate": 2.4213041034288926e-05, "loss": 0.5099, "step": 3447 }, { "epoch": 0.14537175622404452, "grad_norm": 0.4692436754703522, "learning_rate": 2.4220067453625634e-05, "loss": 0.1588, "step": 3448 }, { "epoch": 0.14541391740624407, "grad_norm": 0.5742311477661133, "learning_rate": 2.4227093872962338e-05, "loss": 0.4463, "step": 3449 }, { "epoch": 0.14545607858844362, "grad_norm": 0.5871426463127136, "learning_rate": 2.4234120292299046e-05, "loss": 0.2342, "step": 3450 }, { "epoch": 0.14549823977064316, "grad_norm": 0.9870034456253052, "learning_rate": 2.424114671163575e-05, "loss": 0.1995, "step": 3451 }, { "epoch": 0.1455404009528427, "grad_norm": 0.7645836472511292, "learning_rate": 2.4248173130972458e-05, "loss": 0.6281, "step": 3452 }, { "epoch": 0.14558256213504225, "grad_norm": 2.2400002479553223, "learning_rate": 2.4255199550309163e-05, "loss": 1.0507, "step": 3453 }, { "epoch": 0.14562472331724183, "grad_norm": 0.6425358653068542, "learning_rate": 2.426222596964587e-05, "loss": 0.2209, "step": 3454 }, { "epoch": 0.14566688449944137, "grad_norm": 0.8436351418495178, "learning_rate": 2.4269252388982575e-05, "loss": 0.6549, "step": 3455 }, { "epoch": 0.14570904568164092, "grad_norm": 0.9112862944602966, "learning_rate": 2.427627880831928e-05, "loss": 0.1569, "step": 3456 }, { "epoch": 0.14575120686384047, "grad_norm": 0.8792335987091064, "learning_rate": 2.4283305227655987e-05, "loss": 0.2302, "step": 3457 }, { "epoch": 0.14579336804604, "grad_norm": 1.2324631214141846, "learning_rate": 2.4290331646992695e-05, "loss": 0.5487, "step": 3458 }, { "epoch": 0.14583552922823956, "grad_norm": 0.6005709171295166, "learning_rate": 2.4297358066329402e-05, "loss": 0.1316, "step": 3459 }, { "epoch": 0.1458776904104391, "grad_norm": 0.6781803965568542, "learning_rate": 2.4304384485666107e-05, "loss": 0.1435, "step": 3460 }, { "epoch": 0.14591985159263865, "grad_norm": 0.7392992377281189, "learning_rate": 2.431141090500281e-05, "loss": 0.5363, "step": 3461 }, { "epoch": 0.1459620127748382, "grad_norm": 0.8253593444824219, "learning_rate": 2.431843732433952e-05, "loss": 0.2099, "step": 3462 }, { "epoch": 0.14600417395703774, "grad_norm": 1.2836412191390991, "learning_rate": 2.4325463743676223e-05, "loss": 1.0673, "step": 3463 }, { "epoch": 0.14604633513923732, "grad_norm": 0.6814320087432861, "learning_rate": 2.433249016301293e-05, "loss": 0.165, "step": 3464 }, { "epoch": 0.14608849632143686, "grad_norm": 0.9175514578819275, "learning_rate": 2.4339516582349636e-05, "loss": 0.5822, "step": 3465 }, { "epoch": 0.1461306575036364, "grad_norm": 1.0157177448272705, "learning_rate": 2.4346543001686343e-05, "loss": 0.5655, "step": 3466 }, { "epoch": 0.14617281868583595, "grad_norm": 0.8193447589874268, "learning_rate": 2.4353569421023048e-05, "loss": 0.8379, "step": 3467 }, { "epoch": 0.1462149798680355, "grad_norm": 0.7164490222930908, "learning_rate": 2.4360595840359756e-05, "loss": 0.5461, "step": 3468 }, { "epoch": 0.14625714105023505, "grad_norm": 0.954801619052887, "learning_rate": 2.436762225969646e-05, "loss": 0.4924, "step": 3469 }, { "epoch": 0.1462993022324346, "grad_norm": 0.907965898513794, "learning_rate": 2.4374648679033164e-05, "loss": 0.5478, "step": 3470 }, { "epoch": 0.14634146341463414, "grad_norm": 0.8908265829086304, "learning_rate": 2.4381675098369872e-05, "loss": 0.3067, "step": 3471 }, { "epoch": 0.14638362459683368, "grad_norm": 0.6836931705474854, "learning_rate": 2.4388701517706577e-05, "loss": 0.627, "step": 3472 }, { "epoch": 0.14642578577903326, "grad_norm": 0.4478677213191986, "learning_rate": 2.4395727937043284e-05, "loss": 0.6431, "step": 3473 }, { "epoch": 0.1464679469612328, "grad_norm": 0.5836754441261292, "learning_rate": 2.440275435637999e-05, "loss": 0.2265, "step": 3474 }, { "epoch": 0.14651010814343235, "grad_norm": 0.5513136386871338, "learning_rate": 2.4409780775716696e-05, "loss": 0.3918, "step": 3475 }, { "epoch": 0.1465522693256319, "grad_norm": 0.6091645359992981, "learning_rate": 2.44168071950534e-05, "loss": 0.1637, "step": 3476 }, { "epoch": 0.14659443050783144, "grad_norm": 0.5673713088035583, "learning_rate": 2.442383361439011e-05, "loss": 0.1534, "step": 3477 }, { "epoch": 0.146636591690031, "grad_norm": 0.5675061941146851, "learning_rate": 2.4430860033726813e-05, "loss": 0.1687, "step": 3478 }, { "epoch": 0.14667875287223053, "grad_norm": 0.8883116245269775, "learning_rate": 2.4437886453063517e-05, "loss": 0.9122, "step": 3479 }, { "epoch": 0.14672091405443008, "grad_norm": 1.933870792388916, "learning_rate": 2.4444912872400225e-05, "loss": 0.7627, "step": 3480 }, { "epoch": 0.14676307523662963, "grad_norm": 1.199702501296997, "learning_rate": 2.4451939291736933e-05, "loss": 1.2538, "step": 3481 }, { "epoch": 0.14680523641882917, "grad_norm": 0.9172251224517822, "learning_rate": 2.445896571107364e-05, "loss": 0.4864, "step": 3482 }, { "epoch": 0.14684739760102875, "grad_norm": 0.6828277111053467, "learning_rate": 2.4465992130410345e-05, "loss": 0.2652, "step": 3483 }, { "epoch": 0.1468895587832283, "grad_norm": 0.6931822896003723, "learning_rate": 2.447301854974705e-05, "loss": 0.3306, "step": 3484 }, { "epoch": 0.14693171996542784, "grad_norm": 0.8368985652923584, "learning_rate": 2.4480044969083757e-05, "loss": 0.2987, "step": 3485 }, { "epoch": 0.14697388114762738, "grad_norm": 0.7481760382652283, "learning_rate": 2.4487071388420462e-05, "loss": 0.3956, "step": 3486 }, { "epoch": 0.14701604232982693, "grad_norm": 0.7364850640296936, "learning_rate": 2.449409780775717e-05, "loss": 0.2176, "step": 3487 }, { "epoch": 0.14705820351202648, "grad_norm": 0.7214388251304626, "learning_rate": 2.4501124227093874e-05, "loss": 0.7626, "step": 3488 }, { "epoch": 0.14710036469422602, "grad_norm": 1.101598858833313, "learning_rate": 2.4508150646430582e-05, "loss": 0.2507, "step": 3489 }, { "epoch": 0.14714252587642557, "grad_norm": 0.7411937117576599, "learning_rate": 2.4515177065767286e-05, "loss": 0.3073, "step": 3490 }, { "epoch": 0.1471846870586251, "grad_norm": 0.8953520655632019, "learning_rate": 2.4522203485103994e-05, "loss": 0.6262, "step": 3491 }, { "epoch": 0.14722684824082466, "grad_norm": 0.7388269305229187, "learning_rate": 2.4529229904440698e-05, "loss": 0.2834, "step": 3492 }, { "epoch": 0.14726900942302423, "grad_norm": 0.5771730542182922, "learning_rate": 2.4536256323777403e-05, "loss": 0.5445, "step": 3493 }, { "epoch": 0.14731117060522378, "grad_norm": 0.6692408919334412, "learning_rate": 2.454328274311411e-05, "loss": 0.3767, "step": 3494 }, { "epoch": 0.14735333178742333, "grad_norm": 0.70201176404953, "learning_rate": 2.4550309162450815e-05, "loss": 0.5378, "step": 3495 }, { "epoch": 0.14739549296962287, "grad_norm": 0.557283341884613, "learning_rate": 2.4557335581787523e-05, "loss": 0.3654, "step": 3496 }, { "epoch": 0.14743765415182242, "grad_norm": 0.7883546352386475, "learning_rate": 2.4564362001124227e-05, "loss": 0.2927, "step": 3497 }, { "epoch": 0.14747981533402196, "grad_norm": 1.0320582389831543, "learning_rate": 2.4571388420460935e-05, "loss": 0.186, "step": 3498 }, { "epoch": 0.1475219765162215, "grad_norm": 0.6430939435958862, "learning_rate": 2.457841483979764e-05, "loss": 0.3134, "step": 3499 }, { "epoch": 0.14756413769842106, "grad_norm": 0.4076160192489624, "learning_rate": 2.4585441259134344e-05, "loss": 0.1275, "step": 3500 }, { "epoch": 0.1476062988806206, "grad_norm": 0.72706538438797, "learning_rate": 2.459246767847105e-05, "loss": 0.2617, "step": 3501 }, { "epoch": 0.14764846006282017, "grad_norm": 1.0621836185455322, "learning_rate": 2.4599494097807756e-05, "loss": 1.1805, "step": 3502 }, { "epoch": 0.14769062124501972, "grad_norm": 0.5554672479629517, "learning_rate": 2.4606520517144467e-05, "loss": 0.1684, "step": 3503 }, { "epoch": 0.14773278242721927, "grad_norm": 0.5920020937919617, "learning_rate": 2.461354693648117e-05, "loss": 0.2712, "step": 3504 }, { "epoch": 0.1477749436094188, "grad_norm": 0.6226962804794312, "learning_rate": 2.462057335581788e-05, "loss": 0.561, "step": 3505 }, { "epoch": 0.14781710479161836, "grad_norm": 0.5544708967208862, "learning_rate": 2.4627599775154584e-05, "loss": 0.27, "step": 3506 }, { "epoch": 0.1478592659738179, "grad_norm": 0.9583310484886169, "learning_rate": 2.4634626194491288e-05, "loss": 0.2146, "step": 3507 }, { "epoch": 0.14790142715601745, "grad_norm": 1.1342601776123047, "learning_rate": 2.4641652613827996e-05, "loss": 0.5042, "step": 3508 }, { "epoch": 0.147943588338217, "grad_norm": 0.8234052658081055, "learning_rate": 2.46486790331647e-05, "loss": 0.3241, "step": 3509 }, { "epoch": 0.14798574952041654, "grad_norm": 0.8945664763450623, "learning_rate": 2.4655705452501408e-05, "loss": 0.237, "step": 3510 }, { "epoch": 0.1480279107026161, "grad_norm": 3.5313210487365723, "learning_rate": 2.4662731871838112e-05, "loss": 1.0757, "step": 3511 }, { "epoch": 0.14807007188481566, "grad_norm": 0.5698738098144531, "learning_rate": 2.466975829117482e-05, "loss": 0.1497, "step": 3512 }, { "epoch": 0.1481122330670152, "grad_norm": 0.996466875076294, "learning_rate": 2.4676784710511524e-05, "loss": 0.2099, "step": 3513 }, { "epoch": 0.14815439424921475, "grad_norm": 1.1587128639221191, "learning_rate": 2.468381112984823e-05, "loss": 0.4407, "step": 3514 }, { "epoch": 0.1481965554314143, "grad_norm": 0.8887526988983154, "learning_rate": 2.4690837549184937e-05, "loss": 0.4729, "step": 3515 }, { "epoch": 0.14823871661361385, "grad_norm": 2.8844032287597656, "learning_rate": 2.469786396852164e-05, "loss": 0.8079, "step": 3516 }, { "epoch": 0.1482808777958134, "grad_norm": 1.0250738859176636, "learning_rate": 2.470489038785835e-05, "loss": 0.4825, "step": 3517 }, { "epoch": 0.14832303897801294, "grad_norm": 1.925748348236084, "learning_rate": 2.4711916807195053e-05, "loss": 0.493, "step": 3518 }, { "epoch": 0.14836520016021248, "grad_norm": 0.9066198468208313, "learning_rate": 2.471894322653176e-05, "loss": 0.3184, "step": 3519 }, { "epoch": 0.14840736134241203, "grad_norm": 1.6690788269042969, "learning_rate": 2.4725969645868465e-05, "loss": 1.1818, "step": 3520 }, { "epoch": 0.14844952252461158, "grad_norm": 1.111474871635437, "learning_rate": 2.4732996065205173e-05, "loss": 0.3756, "step": 3521 }, { "epoch": 0.14849168370681115, "grad_norm": 0.8286212086677551, "learning_rate": 2.4740022484541878e-05, "loss": 0.526, "step": 3522 }, { "epoch": 0.1485338448890107, "grad_norm": 1.7058124542236328, "learning_rate": 2.4747048903878582e-05, "loss": 1.0494, "step": 3523 }, { "epoch": 0.14857600607121024, "grad_norm": 1.1204912662506104, "learning_rate": 2.475407532321529e-05, "loss": 0.2291, "step": 3524 }, { "epoch": 0.1486181672534098, "grad_norm": 0.9520102739334106, "learning_rate": 2.4761101742551994e-05, "loss": 0.306, "step": 3525 }, { "epoch": 0.14866032843560933, "grad_norm": 0.7572147846221924, "learning_rate": 2.4768128161888705e-05, "loss": 0.7382, "step": 3526 }, { "epoch": 0.14870248961780888, "grad_norm": 0.6912574172019958, "learning_rate": 2.477515458122541e-05, "loss": 0.3134, "step": 3527 }, { "epoch": 0.14874465080000843, "grad_norm": 1.67494535446167, "learning_rate": 2.4782181000562114e-05, "loss": 0.9454, "step": 3528 }, { "epoch": 0.14878681198220797, "grad_norm": 0.773746907711029, "learning_rate": 2.4789207419898822e-05, "loss": 0.6486, "step": 3529 }, { "epoch": 0.14882897316440752, "grad_norm": 0.6497495174407959, "learning_rate": 2.4796233839235526e-05, "loss": 0.1583, "step": 3530 }, { "epoch": 0.1488711343466071, "grad_norm": 1.114556908607483, "learning_rate": 2.4803260258572234e-05, "loss": 0.4068, "step": 3531 }, { "epoch": 0.14891329552880664, "grad_norm": 0.9612045884132385, "learning_rate": 2.481028667790894e-05, "loss": 0.3988, "step": 3532 }, { "epoch": 0.14895545671100618, "grad_norm": 0.6432231664657593, "learning_rate": 2.4817313097245646e-05, "loss": 0.4917, "step": 3533 }, { "epoch": 0.14899761789320573, "grad_norm": 0.8051885366439819, "learning_rate": 2.482433951658235e-05, "loss": 0.2731, "step": 3534 }, { "epoch": 0.14903977907540528, "grad_norm": 2.3328914642333984, "learning_rate": 2.483136593591906e-05, "loss": 0.7886, "step": 3535 }, { "epoch": 0.14908194025760482, "grad_norm": 1.2402628660202026, "learning_rate": 2.4838392355255763e-05, "loss": 0.4422, "step": 3536 }, { "epoch": 0.14912410143980437, "grad_norm": 0.5697095990180969, "learning_rate": 2.4845418774592467e-05, "loss": 0.2438, "step": 3537 }, { "epoch": 0.14916626262200391, "grad_norm": 0.7527847290039062, "learning_rate": 2.4852445193929175e-05, "loss": 0.4516, "step": 3538 }, { "epoch": 0.14920842380420346, "grad_norm": 2.2049129009246826, "learning_rate": 2.485947161326588e-05, "loss": 0.3968, "step": 3539 }, { "epoch": 0.149250584986403, "grad_norm": 0.6073634028434753, "learning_rate": 2.4866498032602587e-05, "loss": 0.2304, "step": 3540 }, { "epoch": 0.14929274616860258, "grad_norm": 0.9910298585891724, "learning_rate": 2.487352445193929e-05, "loss": 0.5236, "step": 3541 }, { "epoch": 0.14933490735080213, "grad_norm": 1.188902497291565, "learning_rate": 2.4880550871276e-05, "loss": 0.6898, "step": 3542 }, { "epoch": 0.14937706853300167, "grad_norm": 0.7932275533676147, "learning_rate": 2.4887577290612704e-05, "loss": 0.1279, "step": 3543 }, { "epoch": 0.14941922971520122, "grad_norm": 0.8571106791496277, "learning_rate": 2.489460370994941e-05, "loss": 0.5396, "step": 3544 }, { "epoch": 0.14946139089740076, "grad_norm": 0.8471348881721497, "learning_rate": 2.4901630129286116e-05, "loss": 0.5359, "step": 3545 }, { "epoch": 0.1495035520796003, "grad_norm": 0.7429538369178772, "learning_rate": 2.490865654862282e-05, "loss": 0.3006, "step": 3546 }, { "epoch": 0.14954571326179986, "grad_norm": 0.8221273422241211, "learning_rate": 2.4915682967959528e-05, "loss": 0.2998, "step": 3547 }, { "epoch": 0.1495878744439994, "grad_norm": 0.9608278870582581, "learning_rate": 2.4922709387296236e-05, "loss": 0.2132, "step": 3548 }, { "epoch": 0.14963003562619895, "grad_norm": 0.7324603796005249, "learning_rate": 2.4929735806632944e-05, "loss": 0.1948, "step": 3549 }, { "epoch": 0.1496721968083985, "grad_norm": 0.8355990648269653, "learning_rate": 2.4936762225969648e-05, "loss": 0.3085, "step": 3550 }, { "epoch": 0.14971435799059807, "grad_norm": 0.7127228379249573, "learning_rate": 2.4943788645306352e-05, "loss": 0.5007, "step": 3551 }, { "epoch": 0.14975651917279761, "grad_norm": 0.8452028632164001, "learning_rate": 2.495081506464306e-05, "loss": 0.5763, "step": 3552 }, { "epoch": 0.14979868035499716, "grad_norm": 1.5302743911743164, "learning_rate": 2.4957841483979765e-05, "loss": 1.1923, "step": 3553 }, { "epoch": 0.1498408415371967, "grad_norm": 0.7529818415641785, "learning_rate": 2.4964867903316472e-05, "loss": 0.1889, "step": 3554 }, { "epoch": 0.14988300271939625, "grad_norm": 0.7073065042495728, "learning_rate": 2.4971894322653177e-05, "loss": 0.2619, "step": 3555 }, { "epoch": 0.1499251639015958, "grad_norm": 0.7551589012145996, "learning_rate": 2.4978920741989885e-05, "loss": 0.2327, "step": 3556 }, { "epoch": 0.14996732508379534, "grad_norm": 0.6063206195831299, "learning_rate": 2.498594716132659e-05, "loss": 0.2775, "step": 3557 }, { "epoch": 0.1500094862659949, "grad_norm": 0.7969081997871399, "learning_rate": 2.4992973580663297e-05, "loss": 0.6183, "step": 3558 }, { "epoch": 0.15005164744819444, "grad_norm": 0.9289552569389343, "learning_rate": 2.5e-05, "loss": 0.2589, "step": 3559 }, { "epoch": 0.150093808630394, "grad_norm": 0.7549233436584473, "learning_rate": 2.500702641933671e-05, "loss": 0.1665, "step": 3560 }, { "epoch": 0.15013596981259356, "grad_norm": 0.6727360486984253, "learning_rate": 2.5014052838673413e-05, "loss": 0.1734, "step": 3561 }, { "epoch": 0.1501781309947931, "grad_norm": 1.6968610286712646, "learning_rate": 2.502107925801012e-05, "loss": 0.2863, "step": 3562 }, { "epoch": 0.15022029217699265, "grad_norm": 1.359161138534546, "learning_rate": 2.5028105677346826e-05, "loss": 0.7294, "step": 3563 }, { "epoch": 0.1502624533591922, "grad_norm": 0.5627806782722473, "learning_rate": 2.5035132096683533e-05, "loss": 0.3409, "step": 3564 }, { "epoch": 0.15030461454139174, "grad_norm": 0.8769851922988892, "learning_rate": 2.5042158516020238e-05, "loss": 0.2127, "step": 3565 }, { "epoch": 0.15034677572359129, "grad_norm": 0.6608659029006958, "learning_rate": 2.5049184935356946e-05, "loss": 0.4174, "step": 3566 }, { "epoch": 0.15038893690579083, "grad_norm": 0.6778805255889893, "learning_rate": 2.505621135469365e-05, "loss": 0.1252, "step": 3567 }, { "epoch": 0.15043109808799038, "grad_norm": 0.47436273097991943, "learning_rate": 2.5063237774030358e-05, "loss": 0.1657, "step": 3568 }, { "epoch": 0.15047325927018992, "grad_norm": 1.1448566913604736, "learning_rate": 2.507026419336706e-05, "loss": 1.0708, "step": 3569 }, { "epoch": 0.1505154204523895, "grad_norm": 0.8740794062614441, "learning_rate": 2.507729061270377e-05, "loss": 0.7019, "step": 3570 }, { "epoch": 0.15055758163458904, "grad_norm": 0.5011260509490967, "learning_rate": 2.508431703204047e-05, "loss": 0.1065, "step": 3571 }, { "epoch": 0.1505997428167886, "grad_norm": 0.8393807411193848, "learning_rate": 2.5091343451377182e-05, "loss": 0.7042, "step": 3572 }, { "epoch": 0.15064190399898814, "grad_norm": 0.7617668509483337, "learning_rate": 2.5098369870713883e-05, "loss": 0.6774, "step": 3573 }, { "epoch": 0.15068406518118768, "grad_norm": 0.810819149017334, "learning_rate": 2.510539629005059e-05, "loss": 0.3101, "step": 3574 }, { "epoch": 0.15072622636338723, "grad_norm": 0.8605024814605713, "learning_rate": 2.5112422709387295e-05, "loss": 0.2299, "step": 3575 }, { "epoch": 0.15076838754558677, "grad_norm": 0.6064035892486572, "learning_rate": 2.5119449128724003e-05, "loss": 0.3184, "step": 3576 }, { "epoch": 0.15081054872778632, "grad_norm": 0.925980806350708, "learning_rate": 2.5126475548060707e-05, "loss": 0.5388, "step": 3577 }, { "epoch": 0.15085270990998587, "grad_norm": 0.7266318202018738, "learning_rate": 2.5133501967397415e-05, "loss": 0.9239, "step": 3578 }, { "epoch": 0.1508948710921854, "grad_norm": 0.6311729550361633, "learning_rate": 2.514052838673412e-05, "loss": 0.4264, "step": 3579 }, { "epoch": 0.15093703227438499, "grad_norm": 0.7627532482147217, "learning_rate": 2.5147554806070827e-05, "loss": 0.6628, "step": 3580 }, { "epoch": 0.15097919345658453, "grad_norm": 0.6243658065795898, "learning_rate": 2.5154581225407532e-05, "loss": 0.1544, "step": 3581 }, { "epoch": 0.15102135463878408, "grad_norm": 1.0934333801269531, "learning_rate": 2.516160764474424e-05, "loss": 0.9943, "step": 3582 }, { "epoch": 0.15106351582098362, "grad_norm": 0.9819531440734863, "learning_rate": 2.5168634064080947e-05, "loss": 0.246, "step": 3583 }, { "epoch": 0.15110567700318317, "grad_norm": 0.4846247136592865, "learning_rate": 2.5175660483417652e-05, "loss": 0.2058, "step": 3584 }, { "epoch": 0.15114783818538272, "grad_norm": 1.8617192506790161, "learning_rate": 2.518268690275436e-05, "loss": 0.7677, "step": 3585 }, { "epoch": 0.15118999936758226, "grad_norm": 0.6383134722709656, "learning_rate": 2.5189713322091064e-05, "loss": 0.203, "step": 3586 }, { "epoch": 0.1512321605497818, "grad_norm": 0.4650281071662903, "learning_rate": 2.5196739741427772e-05, "loss": 0.1831, "step": 3587 }, { "epoch": 0.15127432173198135, "grad_norm": 0.7978781461715698, "learning_rate": 2.5203766160764476e-05, "loss": 0.3176, "step": 3588 }, { "epoch": 0.1513164829141809, "grad_norm": 0.6325398087501526, "learning_rate": 2.5210792580101184e-05, "loss": 0.2547, "step": 3589 }, { "epoch": 0.15135864409638047, "grad_norm": 0.5034360289573669, "learning_rate": 2.5217818999437888e-05, "loss": 0.1955, "step": 3590 }, { "epoch": 0.15140080527858002, "grad_norm": 0.717110276222229, "learning_rate": 2.5224845418774596e-05, "loss": 0.3383, "step": 3591 }, { "epoch": 0.15144296646077957, "grad_norm": 1.3339207172393799, "learning_rate": 2.5231871838111297e-05, "loss": 1.1764, "step": 3592 }, { "epoch": 0.1514851276429791, "grad_norm": 1.0635524988174438, "learning_rate": 2.5238898257448008e-05, "loss": 1.1383, "step": 3593 }, { "epoch": 0.15152728882517866, "grad_norm": 1.0811219215393066, "learning_rate": 2.524592467678471e-05, "loss": 0.2935, "step": 3594 }, { "epoch": 0.1515694500073782, "grad_norm": 1.6868542432785034, "learning_rate": 2.525295109612142e-05, "loss": 0.6321, "step": 3595 }, { "epoch": 0.15161161118957775, "grad_norm": 0.8158791065216064, "learning_rate": 2.525997751545812e-05, "loss": 0.7254, "step": 3596 }, { "epoch": 0.1516537723717773, "grad_norm": 1.1260313987731934, "learning_rate": 2.526700393479483e-05, "loss": 0.3636, "step": 3597 }, { "epoch": 0.15169593355397684, "grad_norm": 0.676685094833374, "learning_rate": 2.5274030354131534e-05, "loss": 0.3148, "step": 3598 }, { "epoch": 0.15173809473617642, "grad_norm": 1.0311113595962524, "learning_rate": 2.528105677346824e-05, "loss": 0.6218, "step": 3599 }, { "epoch": 0.15178025591837596, "grad_norm": 0.7163869738578796, "learning_rate": 2.5288083192804946e-05, "loss": 0.3591, "step": 3600 }, { "epoch": 0.1518224171005755, "grad_norm": 1.2405966520309448, "learning_rate": 2.5295109612141654e-05, "loss": 0.2514, "step": 3601 }, { "epoch": 0.15186457828277505, "grad_norm": 0.7713985443115234, "learning_rate": 2.5302136031478358e-05, "loss": 0.2191, "step": 3602 }, { "epoch": 0.1519067394649746, "grad_norm": 0.6263036727905273, "learning_rate": 2.5309162450815066e-05, "loss": 0.2011, "step": 3603 }, { "epoch": 0.15194890064717415, "grad_norm": 0.7817750573158264, "learning_rate": 2.531618887015177e-05, "loss": 0.4461, "step": 3604 }, { "epoch": 0.1519910618293737, "grad_norm": 0.9285719394683838, "learning_rate": 2.5323215289488478e-05, "loss": 0.7043, "step": 3605 }, { "epoch": 0.15203322301157324, "grad_norm": 0.7953794598579407, "learning_rate": 2.5330241708825186e-05, "loss": 0.2973, "step": 3606 }, { "epoch": 0.15207538419377278, "grad_norm": 0.5782212615013123, "learning_rate": 2.533726812816189e-05, "loss": 0.2792, "step": 3607 }, { "epoch": 0.15211754537597233, "grad_norm": 0.6919851303100586, "learning_rate": 2.5344294547498598e-05, "loss": 0.1596, "step": 3608 }, { "epoch": 0.1521597065581719, "grad_norm": 1.4076125621795654, "learning_rate": 2.5351320966835302e-05, "loss": 1.2316, "step": 3609 }, { "epoch": 0.15220186774037145, "grad_norm": 0.5040709972381592, "learning_rate": 2.535834738617201e-05, "loss": 0.2056, "step": 3610 }, { "epoch": 0.152244028922571, "grad_norm": 1.5084927082061768, "learning_rate": 2.5365373805508714e-05, "loss": 0.9296, "step": 3611 }, { "epoch": 0.15228619010477054, "grad_norm": 1.203715443611145, "learning_rate": 2.5372400224845422e-05, "loss": 0.9519, "step": 3612 }, { "epoch": 0.1523283512869701, "grad_norm": 0.6355174779891968, "learning_rate": 2.5379426644182123e-05, "loss": 0.344, "step": 3613 }, { "epoch": 0.15237051246916963, "grad_norm": 0.929827868938446, "learning_rate": 2.5386453063518834e-05, "loss": 0.5933, "step": 3614 }, { "epoch": 0.15241267365136918, "grad_norm": 0.5844637155532837, "learning_rate": 2.5393479482855535e-05, "loss": 0.1706, "step": 3615 }, { "epoch": 0.15245483483356873, "grad_norm": 0.7071875929832458, "learning_rate": 2.5400505902192247e-05, "loss": 0.1615, "step": 3616 }, { "epoch": 0.15249699601576827, "grad_norm": 0.9495715498924255, "learning_rate": 2.5407532321528948e-05, "loss": 0.205, "step": 3617 }, { "epoch": 0.15253915719796782, "grad_norm": 0.688834547996521, "learning_rate": 2.541455874086566e-05, "loss": 0.2172, "step": 3618 }, { "epoch": 0.1525813183801674, "grad_norm": 0.6302003264427185, "learning_rate": 2.542158516020236e-05, "loss": 0.1066, "step": 3619 }, { "epoch": 0.15262347956236694, "grad_norm": 0.6755776405334473, "learning_rate": 2.5428611579539068e-05, "loss": 0.1424, "step": 3620 }, { "epoch": 0.15266564074456648, "grad_norm": 0.49681776762008667, "learning_rate": 2.5435637998875772e-05, "loss": 0.1303, "step": 3621 }, { "epoch": 0.15270780192676603, "grad_norm": 0.7044344544410706, "learning_rate": 2.544266441821248e-05, "loss": 0.2309, "step": 3622 }, { "epoch": 0.15274996310896558, "grad_norm": 0.8747916221618652, "learning_rate": 2.5449690837549184e-05, "loss": 0.6971, "step": 3623 }, { "epoch": 0.15279212429116512, "grad_norm": 0.7187075614929199, "learning_rate": 2.5456717256885892e-05, "loss": 0.1649, "step": 3624 }, { "epoch": 0.15283428547336467, "grad_norm": 0.7357186675071716, "learning_rate": 2.5463743676222596e-05, "loss": 0.2346, "step": 3625 }, { "epoch": 0.1528764466555642, "grad_norm": 0.8582876920700073, "learning_rate": 2.5470770095559304e-05, "loss": 0.6631, "step": 3626 }, { "epoch": 0.15291860783776376, "grad_norm": 0.5350353717803955, "learning_rate": 2.5477796514896012e-05, "loss": 0.2081, "step": 3627 }, { "epoch": 0.15296076901996333, "grad_norm": 0.8935354351997375, "learning_rate": 2.5484822934232716e-05, "loss": 0.4892, "step": 3628 }, { "epoch": 0.15300293020216288, "grad_norm": 0.5166090726852417, "learning_rate": 2.5491849353569424e-05, "loss": 0.4139, "step": 3629 }, { "epoch": 0.15304509138436242, "grad_norm": 0.8335846066474915, "learning_rate": 2.549887577290613e-05, "loss": 0.1346, "step": 3630 }, { "epoch": 0.15308725256656197, "grad_norm": 0.7371444702148438, "learning_rate": 2.5505902192242836e-05, "loss": 0.4927, "step": 3631 }, { "epoch": 0.15312941374876152, "grad_norm": 0.9824094176292419, "learning_rate": 2.551292861157954e-05, "loss": 0.5656, "step": 3632 }, { "epoch": 0.15317157493096106, "grad_norm": 0.5055113434791565, "learning_rate": 2.551995503091625e-05, "loss": 0.2203, "step": 3633 }, { "epoch": 0.1532137361131606, "grad_norm": 0.4334336519241333, "learning_rate": 2.5526981450252953e-05, "loss": 0.0746, "step": 3634 }, { "epoch": 0.15325589729536016, "grad_norm": 1.6481691598892212, "learning_rate": 2.553400786958966e-05, "loss": 0.5725, "step": 3635 }, { "epoch": 0.1532980584775597, "grad_norm": 0.7258505821228027, "learning_rate": 2.554103428892636e-05, "loss": 0.2374, "step": 3636 }, { "epoch": 0.15334021965975925, "grad_norm": 0.6732569336891174, "learning_rate": 2.5548060708263073e-05, "loss": 0.2115, "step": 3637 }, { "epoch": 0.15338238084195882, "grad_norm": 0.6867421269416809, "learning_rate": 2.5555087127599774e-05, "loss": 0.5091, "step": 3638 }, { "epoch": 0.15342454202415837, "grad_norm": 0.8899641036987305, "learning_rate": 2.5562113546936485e-05, "loss": 0.1474, "step": 3639 }, { "epoch": 0.1534667032063579, "grad_norm": 0.9376880526542664, "learning_rate": 2.5569139966273186e-05, "loss": 0.2112, "step": 3640 }, { "epoch": 0.15350886438855746, "grad_norm": 0.5447781682014465, "learning_rate": 2.5576166385609894e-05, "loss": 0.3846, "step": 3641 }, { "epoch": 0.153551025570757, "grad_norm": 2.527557611465454, "learning_rate": 2.5583192804946598e-05, "loss": 0.9255, "step": 3642 }, { "epoch": 0.15359318675295655, "grad_norm": 0.8319060802459717, "learning_rate": 2.5590219224283306e-05, "loss": 0.7843, "step": 3643 }, { "epoch": 0.1536353479351561, "grad_norm": 0.8677849173545837, "learning_rate": 2.559724564362001e-05, "loss": 0.2393, "step": 3644 }, { "epoch": 0.15367750911735564, "grad_norm": 0.6463651061058044, "learning_rate": 2.5604272062956718e-05, "loss": 0.1089, "step": 3645 }, { "epoch": 0.1537196702995552, "grad_norm": 0.7069364190101624, "learning_rate": 2.5611298482293422e-05, "loss": 0.278, "step": 3646 }, { "epoch": 0.15376183148175473, "grad_norm": 0.5623118281364441, "learning_rate": 2.561832490163013e-05, "loss": 0.4353, "step": 3647 }, { "epoch": 0.1538039926639543, "grad_norm": 0.9863419532775879, "learning_rate": 2.5625351320966835e-05, "loss": 1.4408, "step": 3648 }, { "epoch": 0.15384615384615385, "grad_norm": 0.7019908428192139, "learning_rate": 2.5632377740303542e-05, "loss": 0.1723, "step": 3649 }, { "epoch": 0.1538883150283534, "grad_norm": 1.1976927518844604, "learning_rate": 2.563940415964025e-05, "loss": 0.4265, "step": 3650 }, { "epoch": 0.15393047621055295, "grad_norm": 0.8688945770263672, "learning_rate": 2.5646430578976955e-05, "loss": 0.3367, "step": 3651 }, { "epoch": 0.1539726373927525, "grad_norm": 0.5405535697937012, "learning_rate": 2.5653456998313662e-05, "loss": 0.3004, "step": 3652 }, { "epoch": 0.15401479857495204, "grad_norm": 1.4231854677200317, "learning_rate": 2.5660483417650367e-05, "loss": 0.1656, "step": 3653 }, { "epoch": 0.15405695975715158, "grad_norm": 1.3553314208984375, "learning_rate": 2.5667509836987075e-05, "loss": 0.9928, "step": 3654 }, { "epoch": 0.15409912093935113, "grad_norm": 2.2189884185791016, "learning_rate": 2.567453625632378e-05, "loss": 0.7314, "step": 3655 }, { "epoch": 0.15414128212155068, "grad_norm": 1.121989130973816, "learning_rate": 2.5681562675660487e-05, "loss": 0.7808, "step": 3656 }, { "epoch": 0.15418344330375025, "grad_norm": 0.4274730980396271, "learning_rate": 2.568858909499719e-05, "loss": 0.0974, "step": 3657 }, { "epoch": 0.1542256044859498, "grad_norm": 0.63933926820755, "learning_rate": 2.56956155143339e-05, "loss": 0.2642, "step": 3658 }, { "epoch": 0.15426776566814934, "grad_norm": 0.9784722328186035, "learning_rate": 2.57026419336706e-05, "loss": 0.2144, "step": 3659 }, { "epoch": 0.1543099268503489, "grad_norm": 0.7229417562484741, "learning_rate": 2.570966835300731e-05, "loss": 0.1801, "step": 3660 }, { "epoch": 0.15435208803254843, "grad_norm": 1.1037648916244507, "learning_rate": 2.5716694772344012e-05, "loss": 0.2347, "step": 3661 }, { "epoch": 0.15439424921474798, "grad_norm": 1.0447384119033813, "learning_rate": 2.5723721191680723e-05, "loss": 0.3635, "step": 3662 }, { "epoch": 0.15443641039694753, "grad_norm": 3.7003557682037354, "learning_rate": 2.5730747611017424e-05, "loss": 0.7667, "step": 3663 }, { "epoch": 0.15447857157914707, "grad_norm": 1.055816411972046, "learning_rate": 2.5737774030354132e-05, "loss": 0.7085, "step": 3664 }, { "epoch": 0.15452073276134662, "grad_norm": 0.7546124458312988, "learning_rate": 2.5744800449690836e-05, "loss": 0.5903, "step": 3665 }, { "epoch": 0.15456289394354616, "grad_norm": 0.718043327331543, "learning_rate": 2.5751826869027544e-05, "loss": 0.686, "step": 3666 }, { "epoch": 0.15460505512574574, "grad_norm": 1.124688982963562, "learning_rate": 2.575885328836425e-05, "loss": 0.4033, "step": 3667 }, { "epoch": 0.15464721630794528, "grad_norm": 0.8100258708000183, "learning_rate": 2.5765879707700956e-05, "loss": 0.349, "step": 3668 }, { "epoch": 0.15468937749014483, "grad_norm": 0.6417360901832581, "learning_rate": 2.577290612703766e-05, "loss": 0.1229, "step": 3669 }, { "epoch": 0.15473153867234438, "grad_norm": 1.4896624088287354, "learning_rate": 2.577993254637437e-05, "loss": 0.3328, "step": 3670 }, { "epoch": 0.15477369985454392, "grad_norm": 0.6379799246788025, "learning_rate": 2.5786958965711073e-05, "loss": 0.2609, "step": 3671 }, { "epoch": 0.15481586103674347, "grad_norm": 1.012719750404358, "learning_rate": 2.579398538504778e-05, "loss": 0.7011, "step": 3672 }, { "epoch": 0.15485802221894301, "grad_norm": 1.4070185422897339, "learning_rate": 2.580101180438449e-05, "loss": 1.2451, "step": 3673 }, { "epoch": 0.15490018340114256, "grad_norm": 0.6561080813407898, "learning_rate": 2.5808038223721193e-05, "loss": 0.5393, "step": 3674 }, { "epoch": 0.1549423445833421, "grad_norm": 0.9677590131759644, "learning_rate": 2.58150646430579e-05, "loss": 0.5641, "step": 3675 }, { "epoch": 0.15498450576554165, "grad_norm": 1.2750403881072998, "learning_rate": 2.5822091062394605e-05, "loss": 0.3907, "step": 3676 }, { "epoch": 0.15502666694774123, "grad_norm": 0.7860377430915833, "learning_rate": 2.5829117481731313e-05, "loss": 0.6675, "step": 3677 }, { "epoch": 0.15506882812994077, "grad_norm": 0.8771556615829468, "learning_rate": 2.5836143901068017e-05, "loss": 0.2032, "step": 3678 }, { "epoch": 0.15511098931214032, "grad_norm": 3.005418539047241, "learning_rate": 2.5843170320404725e-05, "loss": 0.8369, "step": 3679 }, { "epoch": 0.15515315049433986, "grad_norm": 0.7506887912750244, "learning_rate": 2.585019673974143e-05, "loss": 0.2859, "step": 3680 }, { "epoch": 0.1551953116765394, "grad_norm": 0.6372743248939514, "learning_rate": 2.5857223159078137e-05, "loss": 0.1978, "step": 3681 }, { "epoch": 0.15523747285873896, "grad_norm": 0.6624013185501099, "learning_rate": 2.586424957841484e-05, "loss": 0.1405, "step": 3682 }, { "epoch": 0.1552796340409385, "grad_norm": 0.8503322601318359, "learning_rate": 2.587127599775155e-05, "loss": 0.3616, "step": 3683 }, { "epoch": 0.15532179522313805, "grad_norm": 0.8946192264556885, "learning_rate": 2.587830241708825e-05, "loss": 0.799, "step": 3684 }, { "epoch": 0.1553639564053376, "grad_norm": 0.5705595016479492, "learning_rate": 2.588532883642496e-05, "loss": 0.5688, "step": 3685 }, { "epoch": 0.15540611758753717, "grad_norm": 2.011314630508423, "learning_rate": 2.5892355255761663e-05, "loss": 0.6621, "step": 3686 }, { "epoch": 0.15544827876973671, "grad_norm": 0.9682185053825378, "learning_rate": 2.589938167509837e-05, "loss": 0.3593, "step": 3687 }, { "epoch": 0.15549043995193626, "grad_norm": 0.6577078700065613, "learning_rate": 2.5906408094435075e-05, "loss": 0.1742, "step": 3688 }, { "epoch": 0.1555326011341358, "grad_norm": 1.7379767894744873, "learning_rate": 2.5913434513771783e-05, "loss": 0.9235, "step": 3689 }, { "epoch": 0.15557476231633535, "grad_norm": 0.8594356775283813, "learning_rate": 2.5920460933108487e-05, "loss": 0.2255, "step": 3690 }, { "epoch": 0.1556169234985349, "grad_norm": 0.7917270064353943, "learning_rate": 2.5927487352445195e-05, "loss": 0.6108, "step": 3691 }, { "epoch": 0.15565908468073444, "grad_norm": 1.0669643878936768, "learning_rate": 2.59345137717819e-05, "loss": 0.2645, "step": 3692 }, { "epoch": 0.155701245862934, "grad_norm": 0.9493152499198914, "learning_rate": 2.5941540191118607e-05, "loss": 0.4479, "step": 3693 }, { "epoch": 0.15574340704513354, "grad_norm": 0.8476349711418152, "learning_rate": 2.594856661045531e-05, "loss": 0.7951, "step": 3694 }, { "epoch": 0.15578556822733308, "grad_norm": 0.5141727924346924, "learning_rate": 2.595559302979202e-05, "loss": 0.1102, "step": 3695 }, { "epoch": 0.15582772940953266, "grad_norm": 0.5643251538276672, "learning_rate": 2.5962619449128727e-05, "loss": 0.0988, "step": 3696 }, { "epoch": 0.1558698905917322, "grad_norm": 0.7770784497261047, "learning_rate": 2.596964586846543e-05, "loss": 0.119, "step": 3697 }, { "epoch": 0.15591205177393175, "grad_norm": 0.6730115413665771, "learning_rate": 2.597667228780214e-05, "loss": 0.1784, "step": 3698 }, { "epoch": 0.1559542129561313, "grad_norm": 0.7629953622817993, "learning_rate": 2.5983698707138844e-05, "loss": 0.6, "step": 3699 }, { "epoch": 0.15599637413833084, "grad_norm": 0.8060770034790039, "learning_rate": 2.599072512647555e-05, "loss": 0.098, "step": 3700 }, { "epoch": 0.15603853532053039, "grad_norm": 0.5949432849884033, "learning_rate": 2.5997751545812256e-05, "loss": 0.1233, "step": 3701 }, { "epoch": 0.15608069650272993, "grad_norm": 0.9081848859786987, "learning_rate": 2.6004777965148963e-05, "loss": 0.2547, "step": 3702 }, { "epoch": 0.15612285768492948, "grad_norm": 0.8222902417182922, "learning_rate": 2.6011804384485668e-05, "loss": 0.2588, "step": 3703 }, { "epoch": 0.15616501886712902, "grad_norm": 0.8749850392341614, "learning_rate": 2.6018830803822376e-05, "loss": 0.6137, "step": 3704 }, { "epoch": 0.15620718004932857, "grad_norm": 0.6513724327087402, "learning_rate": 2.6025857223159077e-05, "loss": 0.5512, "step": 3705 }, { "epoch": 0.15624934123152814, "grad_norm": 0.6798939108848572, "learning_rate": 2.6032883642495788e-05, "loss": 0.8413, "step": 3706 }, { "epoch": 0.1562915024137277, "grad_norm": 1.171555757522583, "learning_rate": 2.603991006183249e-05, "loss": 0.7544, "step": 3707 }, { "epoch": 0.15633366359592724, "grad_norm": 1.0296396017074585, "learning_rate": 2.60469364811692e-05, "loss": 0.4694, "step": 3708 }, { "epoch": 0.15637582477812678, "grad_norm": 0.8007457256317139, "learning_rate": 2.60539629005059e-05, "loss": 0.2787, "step": 3709 }, { "epoch": 0.15641798596032633, "grad_norm": 0.7364675998687744, "learning_rate": 2.606098931984261e-05, "loss": 0.7356, "step": 3710 }, { "epoch": 0.15646014714252587, "grad_norm": 0.5280604958534241, "learning_rate": 2.6068015739179313e-05, "loss": 0.1254, "step": 3711 }, { "epoch": 0.15650230832472542, "grad_norm": 0.7975168228149414, "learning_rate": 2.607504215851602e-05, "loss": 0.3204, "step": 3712 }, { "epoch": 0.15654446950692497, "grad_norm": 1.4238240718841553, "learning_rate": 2.6082068577852725e-05, "loss": 0.282, "step": 3713 }, { "epoch": 0.1565866306891245, "grad_norm": 0.7501258850097656, "learning_rate": 2.6089094997189433e-05, "loss": 0.1589, "step": 3714 }, { "epoch": 0.15662879187132409, "grad_norm": 0.562982976436615, "learning_rate": 2.6096121416526138e-05, "loss": 0.1425, "step": 3715 }, { "epoch": 0.15667095305352363, "grad_norm": 1.0336235761642456, "learning_rate": 2.6103147835862845e-05, "loss": 0.5609, "step": 3716 }, { "epoch": 0.15671311423572318, "grad_norm": 2.30423641204834, "learning_rate": 2.6110174255199553e-05, "loss": 0.7275, "step": 3717 }, { "epoch": 0.15675527541792272, "grad_norm": 45057.01953125, "learning_rate": 2.6117200674536258e-05, "loss": 0.7703, "step": 3718 }, { "epoch": 0.15679743660012227, "grad_norm": 0.8222125172615051, "learning_rate": 2.6124227093872965e-05, "loss": 0.6052, "step": 3719 }, { "epoch": 0.15683959778232182, "grad_norm": 0.689032256603241, "learning_rate": 2.613125351320967e-05, "loss": 0.4072, "step": 3720 }, { "epoch": 0.15688175896452136, "grad_norm": 0.6992154121398926, "learning_rate": 2.6138279932546377e-05, "loss": 0.1654, "step": 3721 }, { "epoch": 0.1569239201467209, "grad_norm": 0.717452883720398, "learning_rate": 2.6145306351883082e-05, "loss": 0.2001, "step": 3722 }, { "epoch": 0.15696608132892045, "grad_norm": 1.009034276008606, "learning_rate": 2.615233277121979e-05, "loss": 0.2047, "step": 3723 }, { "epoch": 0.15700824251112, "grad_norm": 2.13893723487854, "learning_rate": 2.6159359190556494e-05, "loss": 0.7977, "step": 3724 }, { "epoch": 0.15705040369331957, "grad_norm": 1.2483547925949097, "learning_rate": 2.6166385609893202e-05, "loss": 0.3114, "step": 3725 }, { "epoch": 0.15709256487551912, "grad_norm": 0.7942487597465515, "learning_rate": 2.6173412029229903e-05, "loss": 0.2007, "step": 3726 }, { "epoch": 0.15713472605771867, "grad_norm": 0.797238826751709, "learning_rate": 2.6180438448566614e-05, "loss": 0.7083, "step": 3727 }, { "epoch": 0.1571768872399182, "grad_norm": 1.3966213464736938, "learning_rate": 2.6187464867903315e-05, "loss": 1.0222, "step": 3728 }, { "epoch": 0.15721904842211776, "grad_norm": 1.027103304862976, "learning_rate": 2.6194491287240026e-05, "loss": 0.4893, "step": 3729 }, { "epoch": 0.1572612096043173, "grad_norm": 0.5244775414466858, "learning_rate": 2.6201517706576727e-05, "loss": 0.2493, "step": 3730 }, { "epoch": 0.15730337078651685, "grad_norm": 0.5867892503738403, "learning_rate": 2.620854412591344e-05, "loss": 0.1359, "step": 3731 }, { "epoch": 0.1573455319687164, "grad_norm": 0.8022965788841248, "learning_rate": 2.621557054525014e-05, "loss": 0.2859, "step": 3732 }, { "epoch": 0.15738769315091594, "grad_norm": 0.672375500202179, "learning_rate": 2.6222596964586847e-05, "loss": 0.1346, "step": 3733 }, { "epoch": 0.1574298543331155, "grad_norm": 2.021160125732422, "learning_rate": 2.622962338392355e-05, "loss": 1.3805, "step": 3734 }, { "epoch": 0.15747201551531506, "grad_norm": 0.8465659022331238, "learning_rate": 2.623664980326026e-05, "loss": 0.4834, "step": 3735 }, { "epoch": 0.1575141766975146, "grad_norm": 1.1654325723648071, "learning_rate": 2.6243676222596964e-05, "loss": 1.2322, "step": 3736 }, { "epoch": 0.15755633787971415, "grad_norm": 0.7082625031471252, "learning_rate": 2.625070264193367e-05, "loss": 0.6198, "step": 3737 }, { "epoch": 0.1575984990619137, "grad_norm": 0.6504243612289429, "learning_rate": 2.6257729061270376e-05, "loss": 0.2141, "step": 3738 }, { "epoch": 0.15764066024411325, "grad_norm": 1.3662605285644531, "learning_rate": 2.6264755480607084e-05, "loss": 1.0163, "step": 3739 }, { "epoch": 0.1576828214263128, "grad_norm": 0.7992802262306213, "learning_rate": 2.627178189994379e-05, "loss": 0.6964, "step": 3740 }, { "epoch": 0.15772498260851234, "grad_norm": 1.7704062461853027, "learning_rate": 2.6278808319280496e-05, "loss": 0.5717, "step": 3741 }, { "epoch": 0.15776714379071188, "grad_norm": 0.8655713200569153, "learning_rate": 2.6285834738617204e-05, "loss": 0.8417, "step": 3742 }, { "epoch": 0.15780930497291143, "grad_norm": 0.9145855903625488, "learning_rate": 2.6292861157953908e-05, "loss": 0.5871, "step": 3743 }, { "epoch": 0.15785146615511098, "grad_norm": 0.5726626515388489, "learning_rate": 2.6299887577290616e-05, "loss": 0.164, "step": 3744 }, { "epoch": 0.15789362733731055, "grad_norm": 0.8124604821205139, "learning_rate": 2.630691399662732e-05, "loss": 0.4223, "step": 3745 }, { "epoch": 0.1579357885195101, "grad_norm": 0.6204549074172974, "learning_rate": 2.6313940415964028e-05, "loss": 0.3623, "step": 3746 }, { "epoch": 0.15797794970170964, "grad_norm": 0.7305410504341125, "learning_rate": 2.6320966835300732e-05, "loss": 0.1202, "step": 3747 }, { "epoch": 0.1580201108839092, "grad_norm": 1.302307367324829, "learning_rate": 2.632799325463744e-05, "loss": 1.0685, "step": 3748 }, { "epoch": 0.15806227206610873, "grad_norm": 0.6343984007835388, "learning_rate": 2.633501967397414e-05, "loss": 0.3571, "step": 3749 }, { "epoch": 0.15810443324830828, "grad_norm": 0.737615704536438, "learning_rate": 2.6342046093310852e-05, "loss": 0.2469, "step": 3750 }, { "epoch": 0.15814659443050783, "grad_norm": 0.5751727819442749, "learning_rate": 2.6349072512647553e-05, "loss": 0.2497, "step": 3751 }, { "epoch": 0.15818875561270737, "grad_norm": 0.8505719304084778, "learning_rate": 2.6356098931984265e-05, "loss": 0.1811, "step": 3752 }, { "epoch": 0.15823091679490692, "grad_norm": 0.9545885920524597, "learning_rate": 2.6363125351320966e-05, "loss": 0.4786, "step": 3753 }, { "epoch": 0.1582730779771065, "grad_norm": 0.8652141690254211, "learning_rate": 2.6370151770657677e-05, "loss": 0.3499, "step": 3754 }, { "epoch": 0.15831523915930604, "grad_norm": 0.6200642585754395, "learning_rate": 2.6377178189994378e-05, "loss": 0.2552, "step": 3755 }, { "epoch": 0.15835740034150558, "grad_norm": 0.4548117220401764, "learning_rate": 2.6384204609331086e-05, "loss": 0.2349, "step": 3756 }, { "epoch": 0.15839956152370513, "grad_norm": 0.7935814261436462, "learning_rate": 2.639123102866779e-05, "loss": 0.3451, "step": 3757 }, { "epoch": 0.15844172270590468, "grad_norm": 1.0930887460708618, "learning_rate": 2.6398257448004498e-05, "loss": 0.4261, "step": 3758 }, { "epoch": 0.15848388388810422, "grad_norm": 0.8433278203010559, "learning_rate": 2.6405283867341202e-05, "loss": 0.1781, "step": 3759 }, { "epoch": 0.15852604507030377, "grad_norm": 0.7033737897872925, "learning_rate": 2.641231028667791e-05, "loss": 0.6155, "step": 3760 }, { "epoch": 0.1585682062525033, "grad_norm": 0.6368169188499451, "learning_rate": 2.6419336706014614e-05, "loss": 0.2923, "step": 3761 }, { "epoch": 0.15861036743470286, "grad_norm": 0.649983286857605, "learning_rate": 2.6426363125351322e-05, "loss": 0.181, "step": 3762 }, { "epoch": 0.1586525286169024, "grad_norm": 2.063356876373291, "learning_rate": 2.643338954468803e-05, "loss": 0.789, "step": 3763 }, { "epoch": 0.15869468979910198, "grad_norm": 0.9011859893798828, "learning_rate": 2.6440415964024734e-05, "loss": 0.2509, "step": 3764 }, { "epoch": 0.15873685098130152, "grad_norm": 1.265116810798645, "learning_rate": 2.6447442383361442e-05, "loss": 0.4295, "step": 3765 }, { "epoch": 0.15877901216350107, "grad_norm": 0.9299634695053101, "learning_rate": 2.6454468802698146e-05, "loss": 0.2574, "step": 3766 }, { "epoch": 0.15882117334570062, "grad_norm": 1.0914582014083862, "learning_rate": 2.6461495222034854e-05, "loss": 0.9104, "step": 3767 }, { "epoch": 0.15886333452790016, "grad_norm": 0.9169183969497681, "learning_rate": 2.646852164137156e-05, "loss": 0.4335, "step": 3768 }, { "epoch": 0.1589054957100997, "grad_norm": 0.513595700263977, "learning_rate": 2.6475548060708266e-05, "loss": 0.1314, "step": 3769 }, { "epoch": 0.15894765689229925, "grad_norm": 0.5440307855606079, "learning_rate": 2.648257448004497e-05, "loss": 0.2512, "step": 3770 }, { "epoch": 0.1589898180744988, "grad_norm": 10.38740062713623, "learning_rate": 2.648960089938168e-05, "loss": 0.3037, "step": 3771 }, { "epoch": 0.15903197925669835, "grad_norm": 1.231209635734558, "learning_rate": 2.649662731871838e-05, "loss": 1.1316, "step": 3772 }, { "epoch": 0.1590741404388979, "grad_norm": 0.8620584607124329, "learning_rate": 2.650365373805509e-05, "loss": 0.8488, "step": 3773 }, { "epoch": 0.15911630162109747, "grad_norm": 0.5258167386054993, "learning_rate": 2.6510680157391792e-05, "loss": 0.3347, "step": 3774 }, { "epoch": 0.159158462803297, "grad_norm": 2.3190646171569824, "learning_rate": 2.6517706576728503e-05, "loss": 0.4591, "step": 3775 }, { "epoch": 0.15920062398549656, "grad_norm": 0.46200862526893616, "learning_rate": 2.6524732996065204e-05, "loss": 0.2112, "step": 3776 }, { "epoch": 0.1592427851676961, "grad_norm": 0.48167091608047485, "learning_rate": 2.6531759415401912e-05, "loss": 0.1984, "step": 3777 }, { "epoch": 0.15928494634989565, "grad_norm": 0.555270254611969, "learning_rate": 2.6538785834738616e-05, "loss": 0.2261, "step": 3778 }, { "epoch": 0.1593271075320952, "grad_norm": 0.7024890780448914, "learning_rate": 2.6545812254075324e-05, "loss": 0.8306, "step": 3779 }, { "epoch": 0.15936926871429474, "grad_norm": 0.8430922627449036, "learning_rate": 2.6552838673412028e-05, "loss": 0.3098, "step": 3780 }, { "epoch": 0.1594114298964943, "grad_norm": 1.1685593128204346, "learning_rate": 2.6559865092748736e-05, "loss": 0.2308, "step": 3781 }, { "epoch": 0.15945359107869383, "grad_norm": 0.38478389382362366, "learning_rate": 2.656689151208544e-05, "loss": 0.1375, "step": 3782 }, { "epoch": 0.1594957522608934, "grad_norm": 0.8257564306259155, "learning_rate": 2.6573917931422148e-05, "loss": 0.2823, "step": 3783 }, { "epoch": 0.15953791344309295, "grad_norm": 0.5248709321022034, "learning_rate": 2.6580944350758853e-05, "loss": 0.2272, "step": 3784 }, { "epoch": 0.1595800746252925, "grad_norm": 0.6250510215759277, "learning_rate": 2.658797077009556e-05, "loss": 0.348, "step": 3785 }, { "epoch": 0.15962223580749205, "grad_norm": 0.4440321624279022, "learning_rate": 2.6594997189432268e-05, "loss": 0.1851, "step": 3786 }, { "epoch": 0.1596643969896916, "grad_norm": 0.6367289423942566, "learning_rate": 2.6602023608768973e-05, "loss": 0.159, "step": 3787 }, { "epoch": 0.15970655817189114, "grad_norm": 1.4652268886566162, "learning_rate": 2.660905002810568e-05, "loss": 1.1543, "step": 3788 }, { "epoch": 0.15974871935409068, "grad_norm": 0.34838229417800903, "learning_rate": 2.6616076447442385e-05, "loss": 0.1087, "step": 3789 }, { "epoch": 0.15979088053629023, "grad_norm": 1.042372703552246, "learning_rate": 2.6623102866779093e-05, "loss": 0.7325, "step": 3790 }, { "epoch": 0.15983304171848978, "grad_norm": 1.2787641286849976, "learning_rate": 2.6630129286115797e-05, "loss": 1.0286, "step": 3791 }, { "epoch": 0.15987520290068932, "grad_norm": 1.171874761581421, "learning_rate": 2.6637155705452505e-05, "loss": 1.0667, "step": 3792 }, { "epoch": 0.1599173640828889, "grad_norm": 0.659543514251709, "learning_rate": 2.664418212478921e-05, "loss": 0.4477, "step": 3793 }, { "epoch": 0.15995952526508844, "grad_norm": 0.7853938341140747, "learning_rate": 2.6651208544125917e-05, "loss": 0.6514, "step": 3794 }, { "epoch": 0.160001686447288, "grad_norm": 0.5809754133224487, "learning_rate": 2.6658234963462618e-05, "loss": 0.4321, "step": 3795 }, { "epoch": 0.16004384762948753, "grad_norm": 0.5032126307487488, "learning_rate": 2.666526138279933e-05, "loss": 0.2335, "step": 3796 }, { "epoch": 0.16008600881168708, "grad_norm": 0.6834444999694824, "learning_rate": 2.667228780213603e-05, "loss": 0.2282, "step": 3797 }, { "epoch": 0.16012816999388663, "grad_norm": 0.9061486124992371, "learning_rate": 2.667931422147274e-05, "loss": 0.3257, "step": 3798 }, { "epoch": 0.16017033117608617, "grad_norm": 0.7908409237861633, "learning_rate": 2.6686340640809442e-05, "loss": 0.2971, "step": 3799 }, { "epoch": 0.16021249235828572, "grad_norm": 0.8157692551612854, "learning_rate": 2.669336706014615e-05, "loss": 0.2466, "step": 3800 }, { "epoch": 0.16025465354048526, "grad_norm": 0.778475821018219, "learning_rate": 2.6700393479482854e-05, "loss": 0.5733, "step": 3801 }, { "epoch": 0.1602968147226848, "grad_norm": 1.4830368757247925, "learning_rate": 2.6707419898819562e-05, "loss": 1.097, "step": 3802 }, { "epoch": 0.16033897590488438, "grad_norm": 0.6297268867492676, "learning_rate": 2.6714446318156267e-05, "loss": 0.2695, "step": 3803 }, { "epoch": 0.16038113708708393, "grad_norm": 0.7945795059204102, "learning_rate": 2.6721472737492974e-05, "loss": 0.5295, "step": 3804 }, { "epoch": 0.16042329826928348, "grad_norm": 2.2978506088256836, "learning_rate": 2.672849915682968e-05, "loss": 1.0818, "step": 3805 }, { "epoch": 0.16046545945148302, "grad_norm": 0.5008324980735779, "learning_rate": 2.6735525576166387e-05, "loss": 0.4131, "step": 3806 }, { "epoch": 0.16050762063368257, "grad_norm": 0.5249876379966736, "learning_rate": 2.6742551995503094e-05, "loss": 0.1639, "step": 3807 }, { "epoch": 0.16054978181588211, "grad_norm": 1.3647557497024536, "learning_rate": 2.67495784148398e-05, "loss": 0.3681, "step": 3808 }, { "epoch": 0.16059194299808166, "grad_norm": 0.4543229043483734, "learning_rate": 2.6756604834176507e-05, "loss": 0.1454, "step": 3809 }, { "epoch": 0.1606341041802812, "grad_norm": 1.3641008138656616, "learning_rate": 2.676363125351321e-05, "loss": 1.1732, "step": 3810 }, { "epoch": 0.16067626536248075, "grad_norm": 0.6366008520126343, "learning_rate": 2.677065767284992e-05, "loss": 0.2024, "step": 3811 }, { "epoch": 0.16071842654468033, "grad_norm": 1.037899374961853, "learning_rate": 2.6777684092186623e-05, "loss": 0.3038, "step": 3812 }, { "epoch": 0.16076058772687987, "grad_norm": 0.9174931049346924, "learning_rate": 2.678471051152333e-05, "loss": 0.3132, "step": 3813 }, { "epoch": 0.16080274890907942, "grad_norm": 0.6646636724472046, "learning_rate": 2.6791736930860035e-05, "loss": 0.7169, "step": 3814 }, { "epoch": 0.16084491009127896, "grad_norm": 0.595757246017456, "learning_rate": 2.6798763350196743e-05, "loss": 0.1364, "step": 3815 }, { "epoch": 0.1608870712734785, "grad_norm": 0.6872008442878723, "learning_rate": 2.6805789769533447e-05, "loss": 0.1458, "step": 3816 }, { "epoch": 0.16092923245567806, "grad_norm": 0.8356547355651855, "learning_rate": 2.6812816188870155e-05, "loss": 0.3077, "step": 3817 }, { "epoch": 0.1609713936378776, "grad_norm": 0.8739213943481445, "learning_rate": 2.6819842608206856e-05, "loss": 0.483, "step": 3818 }, { "epoch": 0.16101355482007715, "grad_norm": 0.4206174910068512, "learning_rate": 2.6826869027543567e-05, "loss": 0.1151, "step": 3819 }, { "epoch": 0.1610557160022767, "grad_norm": 1.4616472721099854, "learning_rate": 2.683389544688027e-05, "loss": 1.3323, "step": 3820 }, { "epoch": 0.16109787718447624, "grad_norm": 0.7541486620903015, "learning_rate": 2.684092186621698e-05, "loss": 0.6509, "step": 3821 }, { "epoch": 0.16114003836667581, "grad_norm": 0.9511774182319641, "learning_rate": 2.684794828555368e-05, "loss": 0.3774, "step": 3822 }, { "epoch": 0.16118219954887536, "grad_norm": 0.7505674958229065, "learning_rate": 2.685497470489039e-05, "loss": 0.7054, "step": 3823 }, { "epoch": 0.1612243607310749, "grad_norm": 1.9856988191604614, "learning_rate": 2.6862001124227093e-05, "loss": 1.1235, "step": 3824 }, { "epoch": 0.16126652191327445, "grad_norm": 0.5975627899169922, "learning_rate": 2.68690275435638e-05, "loss": 0.5674, "step": 3825 }, { "epoch": 0.161308683095474, "grad_norm": 0.7494075298309326, "learning_rate": 2.6876053962900505e-05, "loss": 0.7508, "step": 3826 }, { "epoch": 0.16135084427767354, "grad_norm": 1.051649570465088, "learning_rate": 2.6883080382237213e-05, "loss": 1.1227, "step": 3827 }, { "epoch": 0.1613930054598731, "grad_norm": 0.8467483520507812, "learning_rate": 2.6890106801573917e-05, "loss": 0.2706, "step": 3828 }, { "epoch": 0.16143516664207264, "grad_norm": 0.7918856143951416, "learning_rate": 2.6897133220910625e-05, "loss": 0.38, "step": 3829 }, { "epoch": 0.16147732782427218, "grad_norm": 0.677745521068573, "learning_rate": 2.6904159640247333e-05, "loss": 0.6352, "step": 3830 }, { "epoch": 0.16151948900647173, "grad_norm": 1.02802574634552, "learning_rate": 2.6911186059584037e-05, "loss": 0.3859, "step": 3831 }, { "epoch": 0.1615616501886713, "grad_norm": 0.794178307056427, "learning_rate": 2.6918212478920745e-05, "loss": 0.7649, "step": 3832 }, { "epoch": 0.16160381137087085, "grad_norm": 0.7700927257537842, "learning_rate": 2.692523889825745e-05, "loss": 0.7334, "step": 3833 }, { "epoch": 0.1616459725530704, "grad_norm": 0.9307934045791626, "learning_rate": 2.6932265317594157e-05, "loss": 1.0231, "step": 3834 }, { "epoch": 0.16168813373526994, "grad_norm": 0.9139590263366699, "learning_rate": 2.693929173693086e-05, "loss": 0.2309, "step": 3835 }, { "epoch": 0.16173029491746949, "grad_norm": 0.866767168045044, "learning_rate": 2.694631815626757e-05, "loss": 0.3441, "step": 3836 }, { "epoch": 0.16177245609966903, "grad_norm": 1.1152112483978271, "learning_rate": 2.6953344575604274e-05, "loss": 1.187, "step": 3837 }, { "epoch": 0.16181461728186858, "grad_norm": 0.5584548711776733, "learning_rate": 2.696037099494098e-05, "loss": 0.3031, "step": 3838 }, { "epoch": 0.16185677846406812, "grad_norm": 1.0026744604110718, "learning_rate": 2.6967397414277686e-05, "loss": 0.292, "step": 3839 }, { "epoch": 0.16189893964626767, "grad_norm": 0.5410341620445251, "learning_rate": 2.6974423833614394e-05, "loss": 0.1089, "step": 3840 }, { "epoch": 0.16194110082846724, "grad_norm": 0.9832969903945923, "learning_rate": 2.6981450252951095e-05, "loss": 0.2311, "step": 3841 }, { "epoch": 0.1619832620106668, "grad_norm": 0.9441969394683838, "learning_rate": 2.6988476672287806e-05, "loss": 0.2531, "step": 3842 }, { "epoch": 0.16202542319286634, "grad_norm": 0.4858447015285492, "learning_rate": 2.6995503091624507e-05, "loss": 0.2087, "step": 3843 }, { "epoch": 0.16206758437506588, "grad_norm": 0.9879778623580933, "learning_rate": 2.7002529510961218e-05, "loss": 0.3903, "step": 3844 }, { "epoch": 0.16210974555726543, "grad_norm": 1.7065036296844482, "learning_rate": 2.700955593029792e-05, "loss": 0.5063, "step": 3845 }, { "epoch": 0.16215190673946497, "grad_norm": 0.5463408827781677, "learning_rate": 2.7016582349634627e-05, "loss": 0.1485, "step": 3846 }, { "epoch": 0.16219406792166452, "grad_norm": 2.2135393619537354, "learning_rate": 2.702360876897133e-05, "loss": 0.6687, "step": 3847 }, { "epoch": 0.16223622910386407, "grad_norm": 0.730636715888977, "learning_rate": 2.703063518830804e-05, "loss": 0.4474, "step": 3848 }, { "epoch": 0.1622783902860636, "grad_norm": 0.6437188386917114, "learning_rate": 2.7037661607644743e-05, "loss": 0.1392, "step": 3849 }, { "epoch": 0.16232055146826316, "grad_norm": 1.3939459323883057, "learning_rate": 2.704468802698145e-05, "loss": 1.005, "step": 3850 }, { "epoch": 0.16236271265046273, "grad_norm": 1.3158010244369507, "learning_rate": 2.7051714446318156e-05, "loss": 0.9389, "step": 3851 }, { "epoch": 0.16240487383266228, "grad_norm": 0.5808680653572083, "learning_rate": 2.7058740865654863e-05, "loss": 0.2285, "step": 3852 }, { "epoch": 0.16244703501486182, "grad_norm": 0.7883356213569641, "learning_rate": 2.706576728499157e-05, "loss": 0.1964, "step": 3853 }, { "epoch": 0.16248919619706137, "grad_norm": 1.0266095399856567, "learning_rate": 2.7072793704328275e-05, "loss": 0.3355, "step": 3854 }, { "epoch": 0.16253135737926092, "grad_norm": 0.8377487063407898, "learning_rate": 2.7079820123664983e-05, "loss": 0.3431, "step": 3855 }, { "epoch": 0.16257351856146046, "grad_norm": 0.38047972321510315, "learning_rate": 2.7086846543001688e-05, "loss": 0.0804, "step": 3856 }, { "epoch": 0.16261567974366, "grad_norm": 0.8642258644104004, "learning_rate": 2.7093872962338395e-05, "loss": 0.5793, "step": 3857 }, { "epoch": 0.16265784092585955, "grad_norm": 0.8451614379882812, "learning_rate": 2.71008993816751e-05, "loss": 0.2049, "step": 3858 }, { "epoch": 0.1627000021080591, "grad_norm": 1.0041329860687256, "learning_rate": 2.7107925801011808e-05, "loss": 0.2487, "step": 3859 }, { "epoch": 0.16274216329025865, "grad_norm": 1.1684221029281616, "learning_rate": 2.7114952220348512e-05, "loss": 0.8526, "step": 3860 }, { "epoch": 0.16278432447245822, "grad_norm": 0.8012048602104187, "learning_rate": 2.712197863968522e-05, "loss": 0.1756, "step": 3861 }, { "epoch": 0.16282648565465777, "grad_norm": 0.6676225066184998, "learning_rate": 2.712900505902192e-05, "loss": 0.1927, "step": 3862 }, { "epoch": 0.1628686468368573, "grad_norm": 1.254379153251648, "learning_rate": 2.7136031478358632e-05, "loss": 1.3811, "step": 3863 }, { "epoch": 0.16291080801905686, "grad_norm": 0.6105718612670898, "learning_rate": 2.7143057897695333e-05, "loss": 0.2069, "step": 3864 }, { "epoch": 0.1629529692012564, "grad_norm": 0.7803722620010376, "learning_rate": 2.7150084317032044e-05, "loss": 0.3183, "step": 3865 }, { "epoch": 0.16299513038345595, "grad_norm": 0.788148045539856, "learning_rate": 2.7157110736368745e-05, "loss": 0.3884, "step": 3866 }, { "epoch": 0.1630372915656555, "grad_norm": 0.4467121958732605, "learning_rate": 2.7164137155705456e-05, "loss": 0.1065, "step": 3867 }, { "epoch": 0.16307945274785504, "grad_norm": 0.5710030198097229, "learning_rate": 2.7171163575042157e-05, "loss": 0.12, "step": 3868 }, { "epoch": 0.1631216139300546, "grad_norm": 0.8182225227355957, "learning_rate": 2.7178189994378865e-05, "loss": 0.2324, "step": 3869 }, { "epoch": 0.16316377511225416, "grad_norm": 0.7153573036193848, "learning_rate": 2.718521641371557e-05, "loss": 0.1682, "step": 3870 }, { "epoch": 0.1632059362944537, "grad_norm": 0.7669311165809631, "learning_rate": 2.7192242833052277e-05, "loss": 0.6021, "step": 3871 }, { "epoch": 0.16324809747665325, "grad_norm": 0.6988969445228577, "learning_rate": 2.7199269252388982e-05, "loss": 0.7014, "step": 3872 }, { "epoch": 0.1632902586588528, "grad_norm": 1.5915590524673462, "learning_rate": 2.720629567172569e-05, "loss": 0.4898, "step": 3873 }, { "epoch": 0.16333241984105235, "grad_norm": 0.6231810450553894, "learning_rate": 2.7213322091062394e-05, "loss": 0.2492, "step": 3874 }, { "epoch": 0.1633745810232519, "grad_norm": 0.8914927244186401, "learning_rate": 2.72203485103991e-05, "loss": 0.224, "step": 3875 }, { "epoch": 0.16341674220545144, "grad_norm": 0.7028272747993469, "learning_rate": 2.722737492973581e-05, "loss": 0.1736, "step": 3876 }, { "epoch": 0.16345890338765098, "grad_norm": 0.6956274509429932, "learning_rate": 2.7234401349072514e-05, "loss": 0.2941, "step": 3877 }, { "epoch": 0.16350106456985053, "grad_norm": 0.701896607875824, "learning_rate": 2.724142776840922e-05, "loss": 0.1696, "step": 3878 }, { "epoch": 0.16354322575205008, "grad_norm": 0.7573440670967102, "learning_rate": 2.7248454187745926e-05, "loss": 0.2892, "step": 3879 }, { "epoch": 0.16358538693424965, "grad_norm": 0.3457406461238861, "learning_rate": 2.7255480607082634e-05, "loss": 0.0974, "step": 3880 }, { "epoch": 0.1636275481164492, "grad_norm": 0.6545485258102417, "learning_rate": 2.7262507026419338e-05, "loss": 0.1556, "step": 3881 }, { "epoch": 0.16366970929864874, "grad_norm": 0.4533800482749939, "learning_rate": 2.7269533445756046e-05, "loss": 0.0973, "step": 3882 }, { "epoch": 0.1637118704808483, "grad_norm": 0.8211630582809448, "learning_rate": 2.727655986509275e-05, "loss": 0.2601, "step": 3883 }, { "epoch": 0.16375403166304783, "grad_norm": 0.7538045644760132, "learning_rate": 2.7283586284429458e-05, "loss": 0.4278, "step": 3884 }, { "epoch": 0.16379619284524738, "grad_norm": 1.178289771080017, "learning_rate": 2.729061270376616e-05, "loss": 1.1289, "step": 3885 }, { "epoch": 0.16383835402744693, "grad_norm": 1.7947605848312378, "learning_rate": 2.729763912310287e-05, "loss": 0.8345, "step": 3886 }, { "epoch": 0.16388051520964647, "grad_norm": 0.41206982731819153, "learning_rate": 2.730466554243957e-05, "loss": 0.1418, "step": 3887 }, { "epoch": 0.16392267639184602, "grad_norm": 1.0262569189071655, "learning_rate": 2.7311691961776283e-05, "loss": 0.8861, "step": 3888 }, { "epoch": 0.16396483757404556, "grad_norm": 0.9203798770904541, "learning_rate": 2.7318718381112984e-05, "loss": 0.3525, "step": 3889 }, { "epoch": 0.16400699875624514, "grad_norm": 0.6176479458808899, "learning_rate": 2.732574480044969e-05, "loss": 0.3013, "step": 3890 }, { "epoch": 0.16404915993844468, "grad_norm": 3.0303871631622314, "learning_rate": 2.7332771219786396e-05, "loss": 0.6157, "step": 3891 }, { "epoch": 0.16409132112064423, "grad_norm": 0.5584753751754761, "learning_rate": 2.7339797639123103e-05, "loss": 0.2523, "step": 3892 }, { "epoch": 0.16413348230284378, "grad_norm": 0.630477249622345, "learning_rate": 2.7346824058459808e-05, "loss": 0.2486, "step": 3893 }, { "epoch": 0.16417564348504332, "grad_norm": 0.5503000020980835, "learning_rate": 2.7353850477796516e-05, "loss": 0.0897, "step": 3894 }, { "epoch": 0.16421780466724287, "grad_norm": 0.5513634085655212, "learning_rate": 2.736087689713322e-05, "loss": 0.1184, "step": 3895 }, { "epoch": 0.1642599658494424, "grad_norm": 0.518926739692688, "learning_rate": 2.7367903316469928e-05, "loss": 0.2998, "step": 3896 }, { "epoch": 0.16430212703164196, "grad_norm": 0.49446046352386475, "learning_rate": 2.7374929735806636e-05, "loss": 0.1659, "step": 3897 }, { "epoch": 0.1643442882138415, "grad_norm": 0.6814749836921692, "learning_rate": 2.738195615514334e-05, "loss": 0.1212, "step": 3898 }, { "epoch": 0.16438644939604108, "grad_norm": 2.7418980598449707, "learning_rate": 2.7388982574480048e-05, "loss": 0.9184, "step": 3899 }, { "epoch": 0.16442861057824062, "grad_norm": 0.8355576395988464, "learning_rate": 2.7396008993816752e-05, "loss": 0.4004, "step": 3900 }, { "epoch": 0.16447077176044017, "grad_norm": 0.9432451725006104, "learning_rate": 2.740303541315346e-05, "loss": 0.5123, "step": 3901 }, { "epoch": 0.16451293294263972, "grad_norm": 0.6235682964324951, "learning_rate": 2.7410061832490164e-05, "loss": 0.3591, "step": 3902 }, { "epoch": 0.16455509412483926, "grad_norm": 0.5524899363517761, "learning_rate": 2.7417088251826872e-05, "loss": 0.3122, "step": 3903 }, { "epoch": 0.1645972553070388, "grad_norm": 0.7125772833824158, "learning_rate": 2.7424114671163577e-05, "loss": 0.2037, "step": 3904 }, { "epoch": 0.16463941648923835, "grad_norm": 0.6017993092536926, "learning_rate": 2.7431141090500284e-05, "loss": 0.1216, "step": 3905 }, { "epoch": 0.1646815776714379, "grad_norm": 2.6717116832733154, "learning_rate": 2.743816750983699e-05, "loss": 0.3734, "step": 3906 }, { "epoch": 0.16472373885363745, "grad_norm": 0.5296234488487244, "learning_rate": 2.7445193929173697e-05, "loss": 0.1696, "step": 3907 }, { "epoch": 0.164765900035837, "grad_norm": 0.4168719947338104, "learning_rate": 2.7452220348510398e-05, "loss": 0.0975, "step": 3908 }, { "epoch": 0.16480806121803657, "grad_norm": 0.9839299917221069, "learning_rate": 2.745924676784711e-05, "loss": 0.1422, "step": 3909 }, { "epoch": 0.1648502224002361, "grad_norm": 1.1450426578521729, "learning_rate": 2.746627318718381e-05, "loss": 1.066, "step": 3910 }, { "epoch": 0.16489238358243566, "grad_norm": 0.9430334568023682, "learning_rate": 2.747329960652052e-05, "loss": 0.3807, "step": 3911 }, { "epoch": 0.1649345447646352, "grad_norm": 0.6725366711616516, "learning_rate": 2.7480326025857222e-05, "loss": 0.6638, "step": 3912 }, { "epoch": 0.16497670594683475, "grad_norm": 0.7775063514709473, "learning_rate": 2.748735244519393e-05, "loss": 0.7016, "step": 3913 }, { "epoch": 0.1650188671290343, "grad_norm": 0.5163934826850891, "learning_rate": 2.7494378864530634e-05, "loss": 0.0839, "step": 3914 }, { "epoch": 0.16506102831123384, "grad_norm": 0.5898528695106506, "learning_rate": 2.7501405283867342e-05, "loss": 0.1574, "step": 3915 }, { "epoch": 0.1651031894934334, "grad_norm": 0.9050436019897461, "learning_rate": 2.7508431703204046e-05, "loss": 0.1549, "step": 3916 }, { "epoch": 0.16514535067563293, "grad_norm": 0.6743985414505005, "learning_rate": 2.7515458122540754e-05, "loss": 0.3407, "step": 3917 }, { "epoch": 0.16518751185783248, "grad_norm": 0.7317877411842346, "learning_rate": 2.752248454187746e-05, "loss": 0.6209, "step": 3918 }, { "epoch": 0.16522967304003205, "grad_norm": 0.9668940305709839, "learning_rate": 2.7529510961214166e-05, "loss": 0.7045, "step": 3919 }, { "epoch": 0.1652718342222316, "grad_norm": 0.6069551706314087, "learning_rate": 2.7536537380550874e-05, "loss": 0.2505, "step": 3920 }, { "epoch": 0.16531399540443115, "grad_norm": 0.8535555601119995, "learning_rate": 2.754356379988758e-05, "loss": 0.2858, "step": 3921 }, { "epoch": 0.1653561565866307, "grad_norm": 0.8481611013412476, "learning_rate": 2.7550590219224286e-05, "loss": 0.645, "step": 3922 }, { "epoch": 0.16539831776883024, "grad_norm": 0.5964815020561218, "learning_rate": 2.755761663856099e-05, "loss": 0.3164, "step": 3923 }, { "epoch": 0.16544047895102978, "grad_norm": 0.7428137063980103, "learning_rate": 2.75646430578977e-05, "loss": 0.3193, "step": 3924 }, { "epoch": 0.16548264013322933, "grad_norm": 0.7856422066688538, "learning_rate": 2.7571669477234403e-05, "loss": 0.771, "step": 3925 }, { "epoch": 0.16552480131542888, "grad_norm": 1.2647649049758911, "learning_rate": 2.757869589657111e-05, "loss": 1.1774, "step": 3926 }, { "epoch": 0.16556696249762842, "grad_norm": 0.556181788444519, "learning_rate": 2.7585722315907815e-05, "loss": 0.2437, "step": 3927 }, { "epoch": 0.16560912367982797, "grad_norm": 0.679233968257904, "learning_rate": 2.7592748735244523e-05, "loss": 0.6752, "step": 3928 }, { "epoch": 0.16565128486202754, "grad_norm": 0.4895484447479248, "learning_rate": 2.7599775154581227e-05, "loss": 0.1802, "step": 3929 }, { "epoch": 0.1656934460442271, "grad_norm": 0.9912347197532654, "learning_rate": 2.7606801573917935e-05, "loss": 0.4537, "step": 3930 }, { "epoch": 0.16573560722642663, "grad_norm": 0.5690967440605164, "learning_rate": 2.7613827993254636e-05, "loss": 0.1736, "step": 3931 }, { "epoch": 0.16577776840862618, "grad_norm": 0.5673036575317383, "learning_rate": 2.7620854412591347e-05, "loss": 0.3632, "step": 3932 }, { "epoch": 0.16581992959082573, "grad_norm": 0.45212671160697937, "learning_rate": 2.7627880831928048e-05, "loss": 0.1567, "step": 3933 }, { "epoch": 0.16586209077302527, "grad_norm": 0.5840848684310913, "learning_rate": 2.763490725126476e-05, "loss": 0.1716, "step": 3934 }, { "epoch": 0.16590425195522482, "grad_norm": 0.4392696022987366, "learning_rate": 2.764193367060146e-05, "loss": 0.151, "step": 3935 }, { "epoch": 0.16594641313742436, "grad_norm": 1.6074631214141846, "learning_rate": 2.7648960089938168e-05, "loss": 0.6384, "step": 3936 }, { "epoch": 0.1659885743196239, "grad_norm": 0.9742208123207092, "learning_rate": 2.7655986509274872e-05, "loss": 0.3051, "step": 3937 }, { "epoch": 0.16603073550182348, "grad_norm": 1.551376461982727, "learning_rate": 2.766301292861158e-05, "loss": 0.308, "step": 3938 }, { "epoch": 0.16607289668402303, "grad_norm": 1.2074178457260132, "learning_rate": 2.7670039347948285e-05, "loss": 1.0569, "step": 3939 }, { "epoch": 0.16611505786622258, "grad_norm": 0.49384254217147827, "learning_rate": 2.7677065767284992e-05, "loss": 0.1026, "step": 3940 }, { "epoch": 0.16615721904842212, "grad_norm": 0.43238121271133423, "learning_rate": 2.7684092186621697e-05, "loss": 0.1282, "step": 3941 }, { "epoch": 0.16619938023062167, "grad_norm": 0.835235595703125, "learning_rate": 2.7691118605958405e-05, "loss": 0.2722, "step": 3942 }, { "epoch": 0.16624154141282121, "grad_norm": 0.44797492027282715, "learning_rate": 2.7698145025295112e-05, "loss": 0.1542, "step": 3943 }, { "epoch": 0.16628370259502076, "grad_norm": 0.6415092945098877, "learning_rate": 2.7705171444631817e-05, "loss": 0.2842, "step": 3944 }, { "epoch": 0.1663258637772203, "grad_norm": 0.6674643754959106, "learning_rate": 2.7712197863968525e-05, "loss": 0.7304, "step": 3945 }, { "epoch": 0.16636802495941985, "grad_norm": 0.5942822098731995, "learning_rate": 2.771922428330523e-05, "loss": 0.6028, "step": 3946 }, { "epoch": 0.1664101861416194, "grad_norm": 2.1412220001220703, "learning_rate": 2.7726250702641937e-05, "loss": 0.3753, "step": 3947 }, { "epoch": 0.16645234732381897, "grad_norm": 0.8126086592674255, "learning_rate": 2.773327712197864e-05, "loss": 0.2228, "step": 3948 }, { "epoch": 0.16649450850601852, "grad_norm": 0.7564843893051147, "learning_rate": 2.774030354131535e-05, "loss": 0.6626, "step": 3949 }, { "epoch": 0.16653666968821806, "grad_norm": 0.4731402099132538, "learning_rate": 2.7747329960652053e-05, "loss": 0.299, "step": 3950 }, { "epoch": 0.1665788308704176, "grad_norm": 1.1424976587295532, "learning_rate": 2.775435637998876e-05, "loss": 0.3753, "step": 3951 }, { "epoch": 0.16662099205261716, "grad_norm": 0.4423867464065552, "learning_rate": 2.7761382799325465e-05, "loss": 0.1613, "step": 3952 }, { "epoch": 0.1666631532348167, "grad_norm": 1.3428072929382324, "learning_rate": 2.7768409218662173e-05, "loss": 0.2007, "step": 3953 }, { "epoch": 0.16670531441701625, "grad_norm": 0.578816294670105, "learning_rate": 2.7775435637998874e-05, "loss": 0.2022, "step": 3954 }, { "epoch": 0.1667474755992158, "grad_norm": 0.7694022059440613, "learning_rate": 2.7782462057335585e-05, "loss": 0.7793, "step": 3955 }, { "epoch": 0.16678963678141534, "grad_norm": 0.6679232716560364, "learning_rate": 2.7789488476672286e-05, "loss": 0.3306, "step": 3956 }, { "epoch": 0.1668317979636149, "grad_norm": 0.8093929290771484, "learning_rate": 2.7796514896008998e-05, "loss": 0.1905, "step": 3957 }, { "epoch": 0.16687395914581446, "grad_norm": 0.44472482800483704, "learning_rate": 2.78035413153457e-05, "loss": 0.162, "step": 3958 }, { "epoch": 0.166916120328014, "grad_norm": 2.0450634956359863, "learning_rate": 2.7810567734682406e-05, "loss": 1.1718, "step": 3959 }, { "epoch": 0.16695828151021355, "grad_norm": 0.5115108489990234, "learning_rate": 2.781759415401911e-05, "loss": 0.1511, "step": 3960 }, { "epoch": 0.1670004426924131, "grad_norm": 0.7690759301185608, "learning_rate": 2.782462057335582e-05, "loss": 0.6339, "step": 3961 }, { "epoch": 0.16704260387461264, "grad_norm": 1.0630266666412354, "learning_rate": 2.7831646992692523e-05, "loss": 0.7512, "step": 3962 }, { "epoch": 0.1670847650568122, "grad_norm": 0.6212154030799866, "learning_rate": 2.783867341202923e-05, "loss": 0.2152, "step": 3963 }, { "epoch": 0.16712692623901174, "grad_norm": 0.5749490261077881, "learning_rate": 2.7845699831365935e-05, "loss": 0.1506, "step": 3964 }, { "epoch": 0.16716908742121128, "grad_norm": 1.2208529710769653, "learning_rate": 2.7852726250702643e-05, "loss": 0.3175, "step": 3965 }, { "epoch": 0.16721124860341083, "grad_norm": 0.47986945509910583, "learning_rate": 2.785975267003935e-05, "loss": 0.141, "step": 3966 }, { "epoch": 0.1672534097856104, "grad_norm": 0.4429464340209961, "learning_rate": 2.7866779089376055e-05, "loss": 0.1007, "step": 3967 }, { "epoch": 0.16729557096780995, "grad_norm": 0.6589441299438477, "learning_rate": 2.7873805508712763e-05, "loss": 0.1478, "step": 3968 }, { "epoch": 0.1673377321500095, "grad_norm": 0.6478138566017151, "learning_rate": 2.7880831928049467e-05, "loss": 0.5358, "step": 3969 }, { "epoch": 0.16737989333220904, "grad_norm": 0.4238664209842682, "learning_rate": 2.7887858347386175e-05, "loss": 0.1206, "step": 3970 }, { "epoch": 0.16742205451440859, "grad_norm": 0.363503634929657, "learning_rate": 2.789488476672288e-05, "loss": 0.1445, "step": 3971 }, { "epoch": 0.16746421569660813, "grad_norm": 0.8706055879592896, "learning_rate": 2.7901911186059587e-05, "loss": 0.5556, "step": 3972 }, { "epoch": 0.16750637687880768, "grad_norm": 0.6675216555595398, "learning_rate": 2.790893760539629e-05, "loss": 0.7499, "step": 3973 }, { "epoch": 0.16754853806100722, "grad_norm": 0.6213781833648682, "learning_rate": 2.7915964024733e-05, "loss": 0.3688, "step": 3974 }, { "epoch": 0.16759069924320677, "grad_norm": 0.5371435880661011, "learning_rate": 2.79229904440697e-05, "loss": 0.513, "step": 3975 }, { "epoch": 0.16763286042540632, "grad_norm": 0.3890495300292969, "learning_rate": 2.793001686340641e-05, "loss": 0.1631, "step": 3976 }, { "epoch": 0.1676750216076059, "grad_norm": 0.5338608026504517, "learning_rate": 2.7937043282743113e-05, "loss": 0.4481, "step": 3977 }, { "epoch": 0.16771718278980544, "grad_norm": 0.41447317600250244, "learning_rate": 2.7944069702079824e-05, "loss": 0.1793, "step": 3978 }, { "epoch": 0.16775934397200498, "grad_norm": 1.0587263107299805, "learning_rate": 2.7951096121416525e-05, "loss": 0.7312, "step": 3979 }, { "epoch": 0.16780150515420453, "grad_norm": 1.4694873094558716, "learning_rate": 2.7958122540753236e-05, "loss": 1.0858, "step": 3980 }, { "epoch": 0.16784366633640407, "grad_norm": 0.6077433228492737, "learning_rate": 2.7965148960089937e-05, "loss": 0.257, "step": 3981 }, { "epoch": 0.16788582751860362, "grad_norm": 1.0211597681045532, "learning_rate": 2.7972175379426645e-05, "loss": 0.907, "step": 3982 }, { "epoch": 0.16792798870080317, "grad_norm": 1.4937618970870972, "learning_rate": 2.797920179876335e-05, "loss": 0.5734, "step": 3983 }, { "epoch": 0.1679701498830027, "grad_norm": 0.6423366069793701, "learning_rate": 2.7986228218100057e-05, "loss": 0.7116, "step": 3984 }, { "epoch": 0.16801231106520226, "grad_norm": 0.5683768391609192, "learning_rate": 2.799325463743676e-05, "loss": 0.2129, "step": 3985 }, { "epoch": 0.1680544722474018, "grad_norm": 0.8045722246170044, "learning_rate": 2.800028105677347e-05, "loss": 0.5693, "step": 3986 }, { "epoch": 0.16809663342960138, "grad_norm": 0.44512924551963806, "learning_rate": 2.8007307476110177e-05, "loss": 0.1391, "step": 3987 }, { "epoch": 0.16813879461180092, "grad_norm": 1.111182689666748, "learning_rate": 2.801433389544688e-05, "loss": 1.0556, "step": 3988 }, { "epoch": 0.16818095579400047, "grad_norm": 0.5460323095321655, "learning_rate": 2.802136031478359e-05, "loss": 0.3023, "step": 3989 }, { "epoch": 0.16822311697620002, "grad_norm": 0.6659230589866638, "learning_rate": 2.8028386734120293e-05, "loss": 0.2702, "step": 3990 }, { "epoch": 0.16826527815839956, "grad_norm": 0.6432501673698425, "learning_rate": 2.8035413153457e-05, "loss": 0.3818, "step": 3991 }, { "epoch": 0.1683074393405991, "grad_norm": 0.8594971895217896, "learning_rate": 2.8042439572793706e-05, "loss": 0.6221, "step": 3992 }, { "epoch": 0.16834960052279865, "grad_norm": 0.7463109493255615, "learning_rate": 2.8049465992130413e-05, "loss": 0.4439, "step": 3993 }, { "epoch": 0.1683917617049982, "grad_norm": 0.5527853965759277, "learning_rate": 2.8056492411467118e-05, "loss": 0.1797, "step": 3994 }, { "epoch": 0.16843392288719775, "grad_norm": 0.6451491117477417, "learning_rate": 2.8063518830803826e-05, "loss": 0.6502, "step": 3995 }, { "epoch": 0.16847608406939732, "grad_norm": 0.4127362370491028, "learning_rate": 2.807054525014053e-05, "loss": 0.1487, "step": 3996 }, { "epoch": 0.16851824525159687, "grad_norm": 0.5753454566001892, "learning_rate": 2.8077571669477238e-05, "loss": 0.1268, "step": 3997 }, { "epoch": 0.1685604064337964, "grad_norm": 0.6151889562606812, "learning_rate": 2.808459808881394e-05, "loss": 0.1218, "step": 3998 }, { "epoch": 0.16860256761599596, "grad_norm": 0.5767848491668701, "learning_rate": 2.809162450815065e-05, "loss": 0.2178, "step": 3999 }, { "epoch": 0.1686447287981955, "grad_norm": 0.6553568840026855, "learning_rate": 2.809865092748735e-05, "loss": 0.4327, "step": 4000 }, { "epoch": 0.16868688998039505, "grad_norm": 1.0736579895019531, "learning_rate": 2.8105677346824062e-05, "loss": 0.4485, "step": 4001 }, { "epoch": 0.1687290511625946, "grad_norm": 0.8044307827949524, "learning_rate": 2.8112703766160763e-05, "loss": 0.2271, "step": 4002 }, { "epoch": 0.16877121234479414, "grad_norm": 0.49725911021232605, "learning_rate": 2.811973018549747e-05, "loss": 0.1228, "step": 4003 }, { "epoch": 0.1688133735269937, "grad_norm": 0.6703592538833618, "learning_rate": 2.8126756604834175e-05, "loss": 0.2069, "step": 4004 }, { "epoch": 0.16885553470919323, "grad_norm": 1.2599425315856934, "learning_rate": 2.8133783024170883e-05, "loss": 1.0417, "step": 4005 }, { "epoch": 0.1688976958913928, "grad_norm": 0.4910648763179779, "learning_rate": 2.8140809443507588e-05, "loss": 0.2272, "step": 4006 }, { "epoch": 0.16893985707359235, "grad_norm": 0.5931493639945984, "learning_rate": 2.8147835862844295e-05, "loss": 0.2369, "step": 4007 }, { "epoch": 0.1689820182557919, "grad_norm": 0.8996337056159973, "learning_rate": 2.8154862282181e-05, "loss": 0.3235, "step": 4008 }, { "epoch": 0.16902417943799145, "grad_norm": 1.1651943922042847, "learning_rate": 2.8161888701517707e-05, "loss": 0.3725, "step": 4009 }, { "epoch": 0.169066340620191, "grad_norm": 0.7734106183052063, "learning_rate": 2.8168915120854415e-05, "loss": 0.1715, "step": 4010 }, { "epoch": 0.16910850180239054, "grad_norm": 3.741276741027832, "learning_rate": 2.817594154019112e-05, "loss": 1.188, "step": 4011 }, { "epoch": 0.16915066298459008, "grad_norm": 0.4599243700504303, "learning_rate": 2.8182967959527827e-05, "loss": 0.1083, "step": 4012 }, { "epoch": 0.16919282416678963, "grad_norm": 1.0164374113082886, "learning_rate": 2.8189994378864532e-05, "loss": 1.1708, "step": 4013 }, { "epoch": 0.16923498534898918, "grad_norm": 0.7766268849372864, "learning_rate": 2.819702079820124e-05, "loss": 0.2772, "step": 4014 }, { "epoch": 0.16927714653118872, "grad_norm": 0.7096912860870361, "learning_rate": 2.8204047217537944e-05, "loss": 0.3053, "step": 4015 }, { "epoch": 0.1693193077133883, "grad_norm": 0.9768279790878296, "learning_rate": 2.8211073636874652e-05, "loss": 0.2605, "step": 4016 }, { "epoch": 0.16936146889558784, "grad_norm": 0.6084161400794983, "learning_rate": 2.8218100056211356e-05, "loss": 0.5931, "step": 4017 }, { "epoch": 0.1694036300777874, "grad_norm": 0.5572599172592163, "learning_rate": 2.8225126475548064e-05, "loss": 0.2351, "step": 4018 }, { "epoch": 0.16944579125998693, "grad_norm": 0.6384140253067017, "learning_rate": 2.823215289488477e-05, "loss": 0.2678, "step": 4019 }, { "epoch": 0.16948795244218648, "grad_norm": 0.4261181354522705, "learning_rate": 2.8239179314221476e-05, "loss": 0.2089, "step": 4020 }, { "epoch": 0.16953011362438603, "grad_norm": 2.225851535797119, "learning_rate": 2.8246205733558177e-05, "loss": 0.6479, "step": 4021 }, { "epoch": 0.16957227480658557, "grad_norm": 0.8672577142715454, "learning_rate": 2.825323215289489e-05, "loss": 0.291, "step": 4022 }, { "epoch": 0.16961443598878512, "grad_norm": 1.2798271179199219, "learning_rate": 2.826025857223159e-05, "loss": 0.8666, "step": 4023 }, { "epoch": 0.16965659717098466, "grad_norm": 0.5420345067977905, "learning_rate": 2.82672849915683e-05, "loss": 0.1741, "step": 4024 }, { "epoch": 0.16969875835318424, "grad_norm": 1.228212833404541, "learning_rate": 2.8274311410905e-05, "loss": 0.4888, "step": 4025 }, { "epoch": 0.16974091953538378, "grad_norm": 0.539869487285614, "learning_rate": 2.828133783024171e-05, "loss": 0.1818, "step": 4026 }, { "epoch": 0.16978308071758333, "grad_norm": 0.7555884718894958, "learning_rate": 2.8288364249578414e-05, "loss": 0.2059, "step": 4027 }, { "epoch": 0.16982524189978287, "grad_norm": 0.45678800344467163, "learning_rate": 2.829539066891512e-05, "loss": 0.1338, "step": 4028 }, { "epoch": 0.16986740308198242, "grad_norm": 0.47792983055114746, "learning_rate": 2.8302417088251826e-05, "loss": 0.1444, "step": 4029 }, { "epoch": 0.16990956426418197, "grad_norm": 0.863934338092804, "learning_rate": 2.8309443507588534e-05, "loss": 0.2064, "step": 4030 }, { "epoch": 0.1699517254463815, "grad_norm": 0.8243088722229004, "learning_rate": 2.8316469926925238e-05, "loss": 0.4804, "step": 4031 }, { "epoch": 0.16999388662858106, "grad_norm": 0.8050089478492737, "learning_rate": 2.8323496346261946e-05, "loss": 0.3845, "step": 4032 }, { "epoch": 0.1700360478107806, "grad_norm": 0.6011075377464294, "learning_rate": 2.8330522765598654e-05, "loss": 0.2875, "step": 4033 }, { "epoch": 0.17007820899298015, "grad_norm": 2.05527400970459, "learning_rate": 2.8337549184935358e-05, "loss": 1.4879, "step": 4034 }, { "epoch": 0.17012037017517972, "grad_norm": 0.6393519043922424, "learning_rate": 2.8344575604272066e-05, "loss": 0.404, "step": 4035 }, { "epoch": 0.17016253135737927, "grad_norm": 0.5498968958854675, "learning_rate": 2.835160202360877e-05, "loss": 0.2899, "step": 4036 }, { "epoch": 0.17020469253957882, "grad_norm": 0.6326021552085876, "learning_rate": 2.8358628442945478e-05, "loss": 0.2776, "step": 4037 }, { "epoch": 0.17024685372177836, "grad_norm": 0.7367445826530457, "learning_rate": 2.8365654862282182e-05, "loss": 0.6209, "step": 4038 }, { "epoch": 0.1702890149039779, "grad_norm": 1.2509117126464844, "learning_rate": 2.837268128161889e-05, "loss": 1.1495, "step": 4039 }, { "epoch": 0.17033117608617745, "grad_norm": 0.8470770716667175, "learning_rate": 2.8379707700955595e-05, "loss": 0.2712, "step": 4040 }, { "epoch": 0.170373337268377, "grad_norm": 1.8990068435668945, "learning_rate": 2.8386734120292302e-05, "loss": 0.3296, "step": 4041 }, { "epoch": 0.17041549845057655, "grad_norm": 0.5885207653045654, "learning_rate": 2.8393760539629007e-05, "loss": 0.6106, "step": 4042 }, { "epoch": 0.1704576596327761, "grad_norm": 0.8604275584220886, "learning_rate": 2.8400786958965715e-05, "loss": 0.5523, "step": 4043 }, { "epoch": 0.17049982081497564, "grad_norm": 0.7783792614936829, "learning_rate": 2.8407813378302416e-05, "loss": 0.2751, "step": 4044 }, { "epoch": 0.1705419819971752, "grad_norm": 0.5688410997390747, "learning_rate": 2.8414839797639127e-05, "loss": 0.3721, "step": 4045 }, { "epoch": 0.17058414317937476, "grad_norm": 0.689686119556427, "learning_rate": 2.8421866216975828e-05, "loss": 0.2277, "step": 4046 }, { "epoch": 0.1706263043615743, "grad_norm": 0.6047942042350769, "learning_rate": 2.842889263631254e-05, "loss": 0.235, "step": 4047 }, { "epoch": 0.17066846554377385, "grad_norm": 0.5624539852142334, "learning_rate": 2.843591905564924e-05, "loss": 0.1819, "step": 4048 }, { "epoch": 0.1707106267259734, "grad_norm": 0.3764422535896301, "learning_rate": 2.8442945474985948e-05, "loss": 0.1082, "step": 4049 }, { "epoch": 0.17075278790817294, "grad_norm": 0.48337414860725403, "learning_rate": 2.8449971894322652e-05, "loss": 0.1712, "step": 4050 }, { "epoch": 0.1707949490903725, "grad_norm": 1.4324710369110107, "learning_rate": 2.845699831365936e-05, "loss": 1.1085, "step": 4051 }, { "epoch": 0.17083711027257203, "grad_norm": 0.7644332051277161, "learning_rate": 2.8464024732996064e-05, "loss": 0.2908, "step": 4052 }, { "epoch": 0.17087927145477158, "grad_norm": 0.5411726236343384, "learning_rate": 2.8471051152332772e-05, "loss": 0.4306, "step": 4053 }, { "epoch": 0.17092143263697115, "grad_norm": 0.575570285320282, "learning_rate": 2.8478077571669476e-05, "loss": 0.3033, "step": 4054 }, { "epoch": 0.1709635938191707, "grad_norm": 0.8464106917381287, "learning_rate": 2.8485103991006184e-05, "loss": 0.2884, "step": 4055 }, { "epoch": 0.17100575500137025, "grad_norm": 0.5602708458900452, "learning_rate": 2.8492130410342892e-05, "loss": 0.3305, "step": 4056 }, { "epoch": 0.1710479161835698, "grad_norm": 0.40098798274993896, "learning_rate": 2.8499156829679596e-05, "loss": 0.1479, "step": 4057 }, { "epoch": 0.17109007736576934, "grad_norm": 0.5023825764656067, "learning_rate": 2.8506183249016304e-05, "loss": 0.223, "step": 4058 }, { "epoch": 0.17113223854796888, "grad_norm": 0.7636096477508545, "learning_rate": 2.851320966835301e-05, "loss": 0.6864, "step": 4059 }, { "epoch": 0.17117439973016843, "grad_norm": 1.023689866065979, "learning_rate": 2.8520236087689716e-05, "loss": 1.2157, "step": 4060 }, { "epoch": 0.17121656091236798, "grad_norm": 0.9133042097091675, "learning_rate": 2.852726250702642e-05, "loss": 0.2665, "step": 4061 }, { "epoch": 0.17125872209456752, "grad_norm": 0.4372924566268921, "learning_rate": 2.853428892636313e-05, "loss": 0.137, "step": 4062 }, { "epoch": 0.17130088327676707, "grad_norm": 2.7334554195404053, "learning_rate": 2.8541315345699833e-05, "loss": 0.9219, "step": 4063 }, { "epoch": 0.17134304445896664, "grad_norm": 0.6469718813896179, "learning_rate": 2.854834176503654e-05, "loss": 0.1847, "step": 4064 }, { "epoch": 0.1713852056411662, "grad_norm": 0.7832385897636414, "learning_rate": 2.8555368184373245e-05, "loss": 0.5424, "step": 4065 }, { "epoch": 0.17142736682336573, "grad_norm": 0.8161367774009705, "learning_rate": 2.8562394603709953e-05, "loss": 0.3059, "step": 4066 }, { "epoch": 0.17146952800556528, "grad_norm": 0.9488900899887085, "learning_rate": 2.8569421023046654e-05, "loss": 0.5148, "step": 4067 }, { "epoch": 0.17151168918776483, "grad_norm": 0.7428566813468933, "learning_rate": 2.8576447442383365e-05, "loss": 0.2444, "step": 4068 }, { "epoch": 0.17155385036996437, "grad_norm": 0.7325941324234009, "learning_rate": 2.8583473861720066e-05, "loss": 0.7129, "step": 4069 }, { "epoch": 0.17159601155216392, "grad_norm": 0.6855400204658508, "learning_rate": 2.8590500281056777e-05, "loss": 0.7353, "step": 4070 }, { "epoch": 0.17163817273436346, "grad_norm": 1.281363606452942, "learning_rate": 2.8597526700393478e-05, "loss": 1.0522, "step": 4071 }, { "epoch": 0.171680333916563, "grad_norm": 0.7663115859031677, "learning_rate": 2.8604553119730186e-05, "loss": 0.1938, "step": 4072 }, { "epoch": 0.17172249509876256, "grad_norm": 0.5022402405738831, "learning_rate": 2.861157953906689e-05, "loss": 0.1329, "step": 4073 }, { "epoch": 0.17176465628096213, "grad_norm": 0.6715866923332214, "learning_rate": 2.8618605958403598e-05, "loss": 0.1045, "step": 4074 }, { "epoch": 0.17180681746316168, "grad_norm": 1.0842047929763794, "learning_rate": 2.8625632377740303e-05, "loss": 1.0192, "step": 4075 }, { "epoch": 0.17184897864536122, "grad_norm": 1.1497137546539307, "learning_rate": 2.863265879707701e-05, "loss": 1.1174, "step": 4076 }, { "epoch": 0.17189113982756077, "grad_norm": 0.56153804063797, "learning_rate": 2.8639685216413718e-05, "loss": 0.0935, "step": 4077 }, { "epoch": 0.17193330100976031, "grad_norm": 0.8126263618469238, "learning_rate": 2.8646711635750423e-05, "loss": 0.562, "step": 4078 }, { "epoch": 0.17197546219195986, "grad_norm": 1.2021865844726562, "learning_rate": 2.865373805508713e-05, "loss": 0.2977, "step": 4079 }, { "epoch": 0.1720176233741594, "grad_norm": 0.5455647110939026, "learning_rate": 2.8660764474423835e-05, "loss": 0.2018, "step": 4080 }, { "epoch": 0.17205978455635895, "grad_norm": 1.9788988828659058, "learning_rate": 2.8667790893760543e-05, "loss": 0.7622, "step": 4081 }, { "epoch": 0.1721019457385585, "grad_norm": 0.5785194039344788, "learning_rate": 2.8674817313097247e-05, "loss": 0.1106, "step": 4082 }, { "epoch": 0.17214410692075804, "grad_norm": 0.8550437092781067, "learning_rate": 2.8681843732433955e-05, "loss": 0.6241, "step": 4083 }, { "epoch": 0.17218626810295762, "grad_norm": 0.57952481508255, "learning_rate": 2.868887015177066e-05, "loss": 0.2448, "step": 4084 }, { "epoch": 0.17222842928515716, "grad_norm": 1.1737867593765259, "learning_rate": 2.8695896571107367e-05, "loss": 0.5038, "step": 4085 }, { "epoch": 0.1722705904673567, "grad_norm": 0.41884884238243103, "learning_rate": 2.870292299044407e-05, "loss": 0.1929, "step": 4086 }, { "epoch": 0.17231275164955626, "grad_norm": 0.7262113690376282, "learning_rate": 2.870994940978078e-05, "loss": 0.2272, "step": 4087 }, { "epoch": 0.1723549128317558, "grad_norm": 2.1549155712127686, "learning_rate": 2.8716975829117483e-05, "loss": 0.7783, "step": 4088 }, { "epoch": 0.17239707401395535, "grad_norm": 0.5408034324645996, "learning_rate": 2.872400224845419e-05, "loss": 0.1122, "step": 4089 }, { "epoch": 0.1724392351961549, "grad_norm": 0.6388062238693237, "learning_rate": 2.8731028667790892e-05, "loss": 0.3693, "step": 4090 }, { "epoch": 0.17248139637835444, "grad_norm": 0.8161526322364807, "learning_rate": 2.8738055087127603e-05, "loss": 0.2623, "step": 4091 }, { "epoch": 0.172523557560554, "grad_norm": 1.84982168674469, "learning_rate": 2.8745081506464304e-05, "loss": 0.3855, "step": 4092 }, { "epoch": 0.17256571874275356, "grad_norm": 1.6874964237213135, "learning_rate": 2.8752107925801016e-05, "loss": 1.1012, "step": 4093 }, { "epoch": 0.1726078799249531, "grad_norm": 0.888473629951477, "learning_rate": 2.8759134345137717e-05, "loss": 0.1132, "step": 4094 }, { "epoch": 0.17265004110715265, "grad_norm": 0.8035871982574463, "learning_rate": 2.8766160764474424e-05, "loss": 0.2507, "step": 4095 }, { "epoch": 0.1726922022893522, "grad_norm": 0.8095989227294922, "learning_rate": 2.877318718381113e-05, "loss": 0.1398, "step": 4096 }, { "epoch": 0.17273436347155174, "grad_norm": 0.40959659218788147, "learning_rate": 2.8780213603147837e-05, "loss": 0.1352, "step": 4097 }, { "epoch": 0.1727765246537513, "grad_norm": 0.7755389213562012, "learning_rate": 2.878724002248454e-05, "loss": 0.1506, "step": 4098 }, { "epoch": 0.17281868583595084, "grad_norm": 0.6482080221176147, "learning_rate": 2.879426644182125e-05, "loss": 0.2699, "step": 4099 }, { "epoch": 0.17286084701815038, "grad_norm": 0.7819927334785461, "learning_rate": 2.8801292861157957e-05, "loss": 0.4768, "step": 4100 }, { "epoch": 0.17290300820034993, "grad_norm": 0.6833063960075378, "learning_rate": 2.880831928049466e-05, "loss": 0.1962, "step": 4101 }, { "epoch": 0.17294516938254947, "grad_norm": 2.678602695465088, "learning_rate": 2.881534569983137e-05, "loss": 0.5431, "step": 4102 }, { "epoch": 0.17298733056474905, "grad_norm": 0.7649931311607361, "learning_rate": 2.8822372119168073e-05, "loss": 0.6524, "step": 4103 }, { "epoch": 0.1730294917469486, "grad_norm": 1.7074581384658813, "learning_rate": 2.882939853850478e-05, "loss": 0.5641, "step": 4104 }, { "epoch": 0.17307165292914814, "grad_norm": 0.5052173137664795, "learning_rate": 2.8836424957841485e-05, "loss": 0.1605, "step": 4105 }, { "epoch": 0.17311381411134769, "grad_norm": 1.0365735292434692, "learning_rate": 2.8843451377178193e-05, "loss": 1.2803, "step": 4106 }, { "epoch": 0.17315597529354723, "grad_norm": 1.0681010484695435, "learning_rate": 2.8850477796514897e-05, "loss": 1.2712, "step": 4107 }, { "epoch": 0.17319813647574678, "grad_norm": 0.5564591288566589, "learning_rate": 2.8857504215851605e-05, "loss": 0.1578, "step": 4108 }, { "epoch": 0.17324029765794632, "grad_norm": 2.3976261615753174, "learning_rate": 2.886453063518831e-05, "loss": 0.8019, "step": 4109 }, { "epoch": 0.17328245884014587, "grad_norm": 0.9699140191078186, "learning_rate": 2.8871557054525017e-05, "loss": 0.144, "step": 4110 }, { "epoch": 0.17332462002234542, "grad_norm": 0.808758556842804, "learning_rate": 2.887858347386172e-05, "loss": 0.1847, "step": 4111 }, { "epoch": 0.17336678120454496, "grad_norm": 0.4874790906906128, "learning_rate": 2.888560989319843e-05, "loss": 0.1161, "step": 4112 }, { "epoch": 0.17340894238674454, "grad_norm": 0.5691295862197876, "learning_rate": 2.889263631253513e-05, "loss": 0.1644, "step": 4113 }, { "epoch": 0.17345110356894408, "grad_norm": 0.49334174394607544, "learning_rate": 2.8899662731871842e-05, "loss": 0.2016, "step": 4114 }, { "epoch": 0.17349326475114363, "grad_norm": 0.48918789625167847, "learning_rate": 2.8906689151208543e-05, "loss": 0.1875, "step": 4115 }, { "epoch": 0.17353542593334317, "grad_norm": 0.7348389625549316, "learning_rate": 2.8913715570545254e-05, "loss": 0.6692, "step": 4116 }, { "epoch": 0.17357758711554272, "grad_norm": 0.7051131725311279, "learning_rate": 2.8920741989881955e-05, "loss": 0.2602, "step": 4117 }, { "epoch": 0.17361974829774227, "grad_norm": 0.4654143154621124, "learning_rate": 2.8927768409218663e-05, "loss": 0.1248, "step": 4118 }, { "epoch": 0.1736619094799418, "grad_norm": 0.7856143116950989, "learning_rate": 2.8934794828555367e-05, "loss": 0.2626, "step": 4119 }, { "epoch": 0.17370407066214136, "grad_norm": 1.0544694662094116, "learning_rate": 2.8941821247892075e-05, "loss": 1.1843, "step": 4120 }, { "epoch": 0.1737462318443409, "grad_norm": 0.8095389604568481, "learning_rate": 2.894884766722878e-05, "loss": 0.5891, "step": 4121 }, { "epoch": 0.17378839302654048, "grad_norm": 0.7720205783843994, "learning_rate": 2.8955874086565487e-05, "loss": 0.2126, "step": 4122 }, { "epoch": 0.17383055420874002, "grad_norm": 0.6894108653068542, "learning_rate": 2.8962900505902195e-05, "loss": 0.4921, "step": 4123 }, { "epoch": 0.17387271539093957, "grad_norm": 0.5577440857887268, "learning_rate": 2.89699269252389e-05, "loss": 0.2074, "step": 4124 }, { "epoch": 0.17391487657313912, "grad_norm": 0.7552490234375, "learning_rate": 2.8976953344575607e-05, "loss": 0.6717, "step": 4125 }, { "epoch": 0.17395703775533866, "grad_norm": 0.9184363484382629, "learning_rate": 2.898397976391231e-05, "loss": 0.3175, "step": 4126 }, { "epoch": 0.1739991989375382, "grad_norm": 0.6505680680274963, "learning_rate": 2.899100618324902e-05, "loss": 0.1645, "step": 4127 }, { "epoch": 0.17404136011973775, "grad_norm": 0.88557368516922, "learning_rate": 2.8998032602585724e-05, "loss": 0.5038, "step": 4128 }, { "epoch": 0.1740835213019373, "grad_norm": 0.8815656900405884, "learning_rate": 2.900505902192243e-05, "loss": 0.4351, "step": 4129 }, { "epoch": 0.17412568248413685, "grad_norm": 0.6026009917259216, "learning_rate": 2.9012085441259136e-05, "loss": 0.2778, "step": 4130 }, { "epoch": 0.1741678436663364, "grad_norm": 1.4372035264968872, "learning_rate": 2.9019111860595844e-05, "loss": 1.2317, "step": 4131 }, { "epoch": 0.17421000484853597, "grad_norm": 0.4920685589313507, "learning_rate": 2.9026138279932548e-05, "loss": 0.1298, "step": 4132 }, { "epoch": 0.1742521660307355, "grad_norm": 1.0769490003585815, "learning_rate": 2.9033164699269256e-05, "loss": 0.7925, "step": 4133 }, { "epoch": 0.17429432721293506, "grad_norm": 0.9524666666984558, "learning_rate": 2.9040191118605957e-05, "loss": 0.6976, "step": 4134 }, { "epoch": 0.1743364883951346, "grad_norm": 1.0132604837417603, "learning_rate": 2.9047217537942668e-05, "loss": 0.3943, "step": 4135 }, { "epoch": 0.17437864957733415, "grad_norm": 1.4381208419799805, "learning_rate": 2.905424395727937e-05, "loss": 1.1941, "step": 4136 }, { "epoch": 0.1744208107595337, "grad_norm": 0.8317660689353943, "learning_rate": 2.906127037661608e-05, "loss": 0.6344, "step": 4137 }, { "epoch": 0.17446297194173324, "grad_norm": 0.41795921325683594, "learning_rate": 2.906829679595278e-05, "loss": 0.1636, "step": 4138 }, { "epoch": 0.1745051331239328, "grad_norm": 0.6953966021537781, "learning_rate": 2.907532321528949e-05, "loss": 0.3744, "step": 4139 }, { "epoch": 0.17454729430613233, "grad_norm": 1.0854862928390503, "learning_rate": 2.9082349634626193e-05, "loss": 0.6701, "step": 4140 }, { "epoch": 0.17458945548833188, "grad_norm": 0.8885095715522766, "learning_rate": 2.90893760539629e-05, "loss": 0.4591, "step": 4141 }, { "epoch": 0.17463161667053145, "grad_norm": 0.5738387703895569, "learning_rate": 2.9096402473299605e-05, "loss": 0.2138, "step": 4142 }, { "epoch": 0.174673777852731, "grad_norm": 0.7931738495826721, "learning_rate": 2.9103428892636313e-05, "loss": 0.2355, "step": 4143 }, { "epoch": 0.17471593903493055, "grad_norm": 0.8032939434051514, "learning_rate": 2.9110455311973018e-05, "loss": 0.1979, "step": 4144 }, { "epoch": 0.1747581002171301, "grad_norm": 0.7300700545310974, "learning_rate": 2.9117481731309725e-05, "loss": 0.4807, "step": 4145 }, { "epoch": 0.17480026139932964, "grad_norm": 0.4704887270927429, "learning_rate": 2.9124508150646433e-05, "loss": 0.1165, "step": 4146 }, { "epoch": 0.17484242258152918, "grad_norm": 0.6910942196846008, "learning_rate": 2.9131534569983138e-05, "loss": 0.2307, "step": 4147 }, { "epoch": 0.17488458376372873, "grad_norm": 0.8158984184265137, "learning_rate": 2.9138560989319845e-05, "loss": 0.6562, "step": 4148 }, { "epoch": 0.17492674494592828, "grad_norm": 0.6862719058990479, "learning_rate": 2.914558740865655e-05, "loss": 0.1955, "step": 4149 }, { "epoch": 0.17496890612812782, "grad_norm": 0.7759395241737366, "learning_rate": 2.9152613827993258e-05, "loss": 0.2572, "step": 4150 }, { "epoch": 0.1750110673103274, "grad_norm": 1.5154614448547363, "learning_rate": 2.9159640247329962e-05, "loss": 1.0893, "step": 4151 }, { "epoch": 0.17505322849252694, "grad_norm": 0.8735610842704773, "learning_rate": 2.916666666666667e-05, "loss": 0.2298, "step": 4152 }, { "epoch": 0.1750953896747265, "grad_norm": 0.8784050941467285, "learning_rate": 2.9173693086003374e-05, "loss": 0.5067, "step": 4153 }, { "epoch": 0.17513755085692603, "grad_norm": 0.6357749104499817, "learning_rate": 2.9180719505340082e-05, "loss": 0.4242, "step": 4154 }, { "epoch": 0.17517971203912558, "grad_norm": 0.7185232639312744, "learning_rate": 2.9187745924676786e-05, "loss": 0.2977, "step": 4155 }, { "epoch": 0.17522187322132513, "grad_norm": 0.904625415802002, "learning_rate": 2.9194772344013494e-05, "loss": 0.4682, "step": 4156 }, { "epoch": 0.17526403440352467, "grad_norm": 0.7990413308143616, "learning_rate": 2.9201798763350195e-05, "loss": 0.2938, "step": 4157 }, { "epoch": 0.17530619558572422, "grad_norm": 0.9261242151260376, "learning_rate": 2.9208825182686906e-05, "loss": 1.0495, "step": 4158 }, { "epoch": 0.17534835676792376, "grad_norm": 0.4750436842441559, "learning_rate": 2.9215851602023607e-05, "loss": 0.1902, "step": 4159 }, { "epoch": 0.1753905179501233, "grad_norm": 2.4836206436157227, "learning_rate": 2.922287802136032e-05, "loss": 0.9756, "step": 4160 }, { "epoch": 0.17543267913232288, "grad_norm": 0.75897216796875, "learning_rate": 2.922990444069702e-05, "loss": 0.2939, "step": 4161 }, { "epoch": 0.17547484031452243, "grad_norm": 0.4173595607280731, "learning_rate": 2.9236930860033727e-05, "loss": 0.2361, "step": 4162 }, { "epoch": 0.17551700149672197, "grad_norm": 0.7313351035118103, "learning_rate": 2.924395727937043e-05, "loss": 0.6926, "step": 4163 }, { "epoch": 0.17555916267892152, "grad_norm": 1.0849703550338745, "learning_rate": 2.925098369870714e-05, "loss": 1.1896, "step": 4164 }, { "epoch": 0.17560132386112107, "grad_norm": 0.6674766540527344, "learning_rate": 2.9258010118043844e-05, "loss": 0.4333, "step": 4165 }, { "epoch": 0.1756434850433206, "grad_norm": 0.974551260471344, "learning_rate": 2.926503653738055e-05, "loss": 1.1527, "step": 4166 }, { "epoch": 0.17568564622552016, "grad_norm": 0.6565431356430054, "learning_rate": 2.9272062956717256e-05, "loss": 0.6828, "step": 4167 }, { "epoch": 0.1757278074077197, "grad_norm": 0.5942827463150024, "learning_rate": 2.9279089376053964e-05, "loss": 0.3683, "step": 4168 }, { "epoch": 0.17576996858991925, "grad_norm": 0.9093635678291321, "learning_rate": 2.928611579539067e-05, "loss": 0.4003, "step": 4169 }, { "epoch": 0.1758121297721188, "grad_norm": 0.38491490483283997, "learning_rate": 2.9293142214727376e-05, "loss": 0.0967, "step": 4170 }, { "epoch": 0.17585429095431837, "grad_norm": 1.2682476043701172, "learning_rate": 2.9300168634064084e-05, "loss": 0.5861, "step": 4171 }, { "epoch": 0.17589645213651792, "grad_norm": 0.9362366795539856, "learning_rate": 2.9307195053400788e-05, "loss": 0.5768, "step": 4172 }, { "epoch": 0.17593861331871746, "grad_norm": 0.8496823906898499, "learning_rate": 2.9314221472737496e-05, "loss": 0.4832, "step": 4173 }, { "epoch": 0.175980774500917, "grad_norm": 0.7221364378929138, "learning_rate": 2.93212478920742e-05, "loss": 0.6189, "step": 4174 }, { "epoch": 0.17602293568311655, "grad_norm": 0.5159069895744324, "learning_rate": 2.9328274311410908e-05, "loss": 0.1514, "step": 4175 }, { "epoch": 0.1760650968653161, "grad_norm": 0.36960235238075256, "learning_rate": 2.9335300730747613e-05, "loss": 0.0966, "step": 4176 }, { "epoch": 0.17610725804751565, "grad_norm": 0.5530624389648438, "learning_rate": 2.934232715008432e-05, "loss": 0.2376, "step": 4177 }, { "epoch": 0.1761494192297152, "grad_norm": 0.5064297318458557, "learning_rate": 2.9349353569421025e-05, "loss": 0.193, "step": 4178 }, { "epoch": 0.17619158041191474, "grad_norm": 0.7206596732139587, "learning_rate": 2.9356379988757732e-05, "loss": 0.1907, "step": 4179 }, { "epoch": 0.1762337415941143, "grad_norm": 0.5849565267562866, "learning_rate": 2.9363406408094433e-05, "loss": 0.5342, "step": 4180 }, { "epoch": 0.17627590277631386, "grad_norm": 0.6895796656608582, "learning_rate": 2.9370432827431145e-05, "loss": 0.1733, "step": 4181 }, { "epoch": 0.1763180639585134, "grad_norm": 0.732458770275116, "learning_rate": 2.9377459246767846e-05, "loss": 0.4972, "step": 4182 }, { "epoch": 0.17636022514071295, "grad_norm": 0.4778677225112915, "learning_rate": 2.9384485666104557e-05, "loss": 0.1787, "step": 4183 }, { "epoch": 0.1764023863229125, "grad_norm": 0.6634284257888794, "learning_rate": 2.9391512085441258e-05, "loss": 0.6498, "step": 4184 }, { "epoch": 0.17644454750511204, "grad_norm": 2.152163028717041, "learning_rate": 2.9398538504777966e-05, "loss": 0.5133, "step": 4185 }, { "epoch": 0.1764867086873116, "grad_norm": 0.4338158071041107, "learning_rate": 2.940556492411467e-05, "loss": 0.097, "step": 4186 }, { "epoch": 0.17652886986951113, "grad_norm": 1.1516433954238892, "learning_rate": 2.9412591343451378e-05, "loss": 1.065, "step": 4187 }, { "epoch": 0.17657103105171068, "grad_norm": 0.6990818977355957, "learning_rate": 2.9419617762788082e-05, "loss": 0.3817, "step": 4188 }, { "epoch": 0.17661319223391023, "grad_norm": 0.6243388056755066, "learning_rate": 2.942664418212479e-05, "loss": 0.2019, "step": 4189 }, { "epoch": 0.1766553534161098, "grad_norm": 0.7198311686515808, "learning_rate": 2.9433670601461498e-05, "loss": 0.4661, "step": 4190 }, { "epoch": 0.17669751459830935, "grad_norm": 0.4532662332057953, "learning_rate": 2.9440697020798202e-05, "loss": 0.1847, "step": 4191 }, { "epoch": 0.1767396757805089, "grad_norm": 0.5080030560493469, "learning_rate": 2.944772344013491e-05, "loss": 0.1372, "step": 4192 }, { "epoch": 0.17678183696270844, "grad_norm": 0.7101351022720337, "learning_rate": 2.9454749859471614e-05, "loss": 0.2575, "step": 4193 }, { "epoch": 0.17682399814490798, "grad_norm": 0.5944372415542603, "learning_rate": 2.9461776278808322e-05, "loss": 0.3125, "step": 4194 }, { "epoch": 0.17686615932710753, "grad_norm": 0.6355330348014832, "learning_rate": 2.9468802698145027e-05, "loss": 0.5983, "step": 4195 }, { "epoch": 0.17690832050930708, "grad_norm": 0.5824501514434814, "learning_rate": 2.9475829117481734e-05, "loss": 0.1359, "step": 4196 }, { "epoch": 0.17695048169150662, "grad_norm": 1.312024474143982, "learning_rate": 2.948285553681844e-05, "loss": 0.4853, "step": 4197 }, { "epoch": 0.17699264287370617, "grad_norm": 1.2608336210250854, "learning_rate": 2.9489881956155146e-05, "loss": 0.3674, "step": 4198 }, { "epoch": 0.17703480405590571, "grad_norm": 0.7727351188659668, "learning_rate": 2.949690837549185e-05, "loss": 0.2383, "step": 4199 }, { "epoch": 0.1770769652381053, "grad_norm": 1.2871737480163574, "learning_rate": 2.950393479482856e-05, "loss": 0.4305, "step": 4200 }, { "epoch": 0.17711912642030483, "grad_norm": 0.5377796292304993, "learning_rate": 2.9510961214165263e-05, "loss": 0.2012, "step": 4201 }, { "epoch": 0.17716128760250438, "grad_norm": 0.6190678477287292, "learning_rate": 2.951798763350197e-05, "loss": 0.4701, "step": 4202 }, { "epoch": 0.17720344878470393, "grad_norm": 0.42777910828590393, "learning_rate": 2.9525014052838672e-05, "loss": 0.0899, "step": 4203 }, { "epoch": 0.17724560996690347, "grad_norm": 0.9443922638893127, "learning_rate": 2.9532040472175383e-05, "loss": 0.428, "step": 4204 }, { "epoch": 0.17728777114910302, "grad_norm": 0.9197924733161926, "learning_rate": 2.9539066891512084e-05, "loss": 0.3608, "step": 4205 }, { "epoch": 0.17732993233130256, "grad_norm": 0.431942880153656, "learning_rate": 2.9546093310848795e-05, "loss": 0.3042, "step": 4206 }, { "epoch": 0.1773720935135021, "grad_norm": 2.430588960647583, "learning_rate": 2.9553119730185496e-05, "loss": 0.5996, "step": 4207 }, { "epoch": 0.17741425469570166, "grad_norm": 0.5766056180000305, "learning_rate": 2.9560146149522204e-05, "loss": 0.108, "step": 4208 }, { "epoch": 0.17745641587790123, "grad_norm": 0.5085228085517883, "learning_rate": 2.956717256885891e-05, "loss": 0.1042, "step": 4209 }, { "epoch": 0.17749857706010078, "grad_norm": 0.7097167372703552, "learning_rate": 2.9574198988195616e-05, "loss": 0.5327, "step": 4210 }, { "epoch": 0.17754073824230032, "grad_norm": 0.7941386103630066, "learning_rate": 2.958122540753232e-05, "loss": 0.2858, "step": 4211 }, { "epoch": 0.17758289942449987, "grad_norm": 1.9070614576339722, "learning_rate": 2.958825182686903e-05, "loss": 0.5721, "step": 4212 }, { "epoch": 0.17762506060669941, "grad_norm": 1.7979551553726196, "learning_rate": 2.9595278246205736e-05, "loss": 0.6815, "step": 4213 }, { "epoch": 0.17766722178889896, "grad_norm": 0.46854865550994873, "learning_rate": 2.960230466554244e-05, "loss": 0.0897, "step": 4214 }, { "epoch": 0.1777093829710985, "grad_norm": 0.48512202501296997, "learning_rate": 2.9609331084879148e-05, "loss": 0.1354, "step": 4215 }, { "epoch": 0.17775154415329805, "grad_norm": 0.9108487963676453, "learning_rate": 2.9616357504215853e-05, "loss": 0.5914, "step": 4216 }, { "epoch": 0.1777937053354976, "grad_norm": 0.55014568567276, "learning_rate": 2.962338392355256e-05, "loss": 0.2611, "step": 4217 }, { "epoch": 0.17783586651769714, "grad_norm": 0.6729553937911987, "learning_rate": 2.9630410342889265e-05, "loss": 0.3274, "step": 4218 }, { "epoch": 0.17787802769989672, "grad_norm": 0.4930807650089264, "learning_rate": 2.9637436762225973e-05, "loss": 0.1755, "step": 4219 }, { "epoch": 0.17792018888209626, "grad_norm": 0.7341058254241943, "learning_rate": 2.9644463181562677e-05, "loss": 0.531, "step": 4220 }, { "epoch": 0.1779623500642958, "grad_norm": 0.913987934589386, "learning_rate": 2.9651489600899385e-05, "loss": 0.1405, "step": 4221 }, { "epoch": 0.17800451124649536, "grad_norm": 1.113763689994812, "learning_rate": 2.965851602023609e-05, "loss": 1.0323, "step": 4222 }, { "epoch": 0.1780466724286949, "grad_norm": 1.108303427696228, "learning_rate": 2.9665542439572797e-05, "loss": 0.3085, "step": 4223 }, { "epoch": 0.17808883361089445, "grad_norm": 0.6930459141731262, "learning_rate": 2.9672568858909498e-05, "loss": 0.208, "step": 4224 }, { "epoch": 0.178130994793094, "grad_norm": 0.6625240445137024, "learning_rate": 2.967959527824621e-05, "loss": 0.2006, "step": 4225 }, { "epoch": 0.17817315597529354, "grad_norm": 0.5851211547851562, "learning_rate": 2.968662169758291e-05, "loss": 0.624, "step": 4226 }, { "epoch": 0.17821531715749309, "grad_norm": 0.5904361009597778, "learning_rate": 2.969364811691962e-05, "loss": 0.2185, "step": 4227 }, { "epoch": 0.17825747833969263, "grad_norm": 1.1632722616195679, "learning_rate": 2.9700674536256322e-05, "loss": 1.086, "step": 4228 }, { "epoch": 0.1782996395218922, "grad_norm": 0.872548520565033, "learning_rate": 2.9707700955593034e-05, "loss": 0.3807, "step": 4229 }, { "epoch": 0.17834180070409175, "grad_norm": 0.8474454283714294, "learning_rate": 2.9714727374929735e-05, "loss": 0.6398, "step": 4230 }, { "epoch": 0.1783839618862913, "grad_norm": 1.908605933189392, "learning_rate": 2.9721753794266442e-05, "loss": 1.0552, "step": 4231 }, { "epoch": 0.17842612306849084, "grad_norm": 0.5517642498016357, "learning_rate": 2.9728780213603147e-05, "loss": 0.4022, "step": 4232 }, { "epoch": 0.1784682842506904, "grad_norm": 2.591669797897339, "learning_rate": 2.9735806632939855e-05, "loss": 0.7897, "step": 4233 }, { "epoch": 0.17851044543288994, "grad_norm": 0.8042507171630859, "learning_rate": 2.974283305227656e-05, "loss": 0.7661, "step": 4234 }, { "epoch": 0.17855260661508948, "grad_norm": 1.336078405380249, "learning_rate": 2.9749859471613267e-05, "loss": 1.2828, "step": 4235 }, { "epoch": 0.17859476779728903, "grad_norm": 0.5861207246780396, "learning_rate": 2.9756885890949974e-05, "loss": 0.1565, "step": 4236 }, { "epoch": 0.17863692897948857, "grad_norm": 0.7107581496238708, "learning_rate": 2.976391231028668e-05, "loss": 0.4191, "step": 4237 }, { "epoch": 0.17867909016168812, "grad_norm": 0.6532698273658752, "learning_rate": 2.9770938729623387e-05, "loss": 0.1362, "step": 4238 }, { "epoch": 0.1787212513438877, "grad_norm": 0.4938681423664093, "learning_rate": 2.977796514896009e-05, "loss": 0.3974, "step": 4239 }, { "epoch": 0.17876341252608724, "grad_norm": 0.5319130420684814, "learning_rate": 2.97849915682968e-05, "loss": 0.2482, "step": 4240 }, { "epoch": 0.17880557370828679, "grad_norm": 0.5361146330833435, "learning_rate": 2.9792017987633503e-05, "loss": 0.1169, "step": 4241 }, { "epoch": 0.17884773489048633, "grad_norm": 0.4815126657485962, "learning_rate": 2.979904440697021e-05, "loss": 0.1365, "step": 4242 }, { "epoch": 0.17888989607268588, "grad_norm": 0.7767727375030518, "learning_rate": 2.9806070826306915e-05, "loss": 0.2342, "step": 4243 }, { "epoch": 0.17893205725488542, "grad_norm": 0.6677666306495667, "learning_rate": 2.9813097245643623e-05, "loss": 0.6274, "step": 4244 }, { "epoch": 0.17897421843708497, "grad_norm": 1.0148863792419434, "learning_rate": 2.9820123664980328e-05, "loss": 0.4629, "step": 4245 }, { "epoch": 0.17901637961928452, "grad_norm": 0.9373753666877747, "learning_rate": 2.9827150084317035e-05, "loss": 0.7036, "step": 4246 }, { "epoch": 0.17905854080148406, "grad_norm": 0.8033308386802673, "learning_rate": 2.9834176503653736e-05, "loss": 0.1776, "step": 4247 }, { "epoch": 0.17910070198368364, "grad_norm": 0.7267220616340637, "learning_rate": 2.9841202922990448e-05, "loss": 0.3134, "step": 4248 }, { "epoch": 0.17914286316588318, "grad_norm": 0.44349566102027893, "learning_rate": 2.984822934232715e-05, "loss": 0.1676, "step": 4249 }, { "epoch": 0.17918502434808273, "grad_norm": 1.4759817123413086, "learning_rate": 2.985525576166386e-05, "loss": 1.2184, "step": 4250 }, { "epoch": 0.17922718553028227, "grad_norm": 1.610332727432251, "learning_rate": 2.986228218100056e-05, "loss": 0.7333, "step": 4251 }, { "epoch": 0.17926934671248182, "grad_norm": 0.9913310408592224, "learning_rate": 2.986930860033727e-05, "loss": 0.2682, "step": 4252 }, { "epoch": 0.17931150789468137, "grad_norm": 0.6439542174339294, "learning_rate": 2.9876335019673973e-05, "loss": 0.5755, "step": 4253 }, { "epoch": 0.1793536690768809, "grad_norm": 1.7990479469299316, "learning_rate": 2.988336143901068e-05, "loss": 1.1961, "step": 4254 }, { "epoch": 0.17939583025908046, "grad_norm": 0.7043447494506836, "learning_rate": 2.9890387858347385e-05, "loss": 0.428, "step": 4255 }, { "epoch": 0.17943799144128, "grad_norm": 0.7729419469833374, "learning_rate": 2.9897414277684093e-05, "loss": 0.1462, "step": 4256 }, { "epoch": 0.17948015262347955, "grad_norm": 0.7879393696784973, "learning_rate": 2.9904440697020797e-05, "loss": 0.1687, "step": 4257 }, { "epoch": 0.17952231380567912, "grad_norm": 0.7146335244178772, "learning_rate": 2.9911467116357505e-05, "loss": 0.2182, "step": 4258 }, { "epoch": 0.17956447498787867, "grad_norm": 0.6177911162376404, "learning_rate": 2.9918493535694213e-05, "loss": 0.4246, "step": 4259 }, { "epoch": 0.17960663617007822, "grad_norm": 0.6575135588645935, "learning_rate": 2.9925519955030917e-05, "loss": 0.1427, "step": 4260 }, { "epoch": 0.17964879735227776, "grad_norm": 0.7110143303871155, "learning_rate": 2.9932546374367625e-05, "loss": 0.234, "step": 4261 }, { "epoch": 0.1796909585344773, "grad_norm": 0.4951889216899872, "learning_rate": 2.993957279370433e-05, "loss": 0.1122, "step": 4262 }, { "epoch": 0.17973311971667685, "grad_norm": 0.7901542782783508, "learning_rate": 2.9946599213041037e-05, "loss": 0.3336, "step": 4263 }, { "epoch": 0.1797752808988764, "grad_norm": 0.7068585157394409, "learning_rate": 2.995362563237774e-05, "loss": 0.6834, "step": 4264 }, { "epoch": 0.17981744208107595, "grad_norm": 0.8283299803733826, "learning_rate": 2.996065205171445e-05, "loss": 0.3835, "step": 4265 }, { "epoch": 0.1798596032632755, "grad_norm": 0.5025455355644226, "learning_rate": 2.9967678471051154e-05, "loss": 0.1233, "step": 4266 }, { "epoch": 0.17990176444547504, "grad_norm": 2.4362435340881348, "learning_rate": 2.997470489038786e-05, "loss": 0.6982, "step": 4267 }, { "epoch": 0.1799439256276746, "grad_norm": 0.892548680305481, "learning_rate": 2.9981731309724566e-05, "loss": 0.2086, "step": 4268 }, { "epoch": 0.17998608680987416, "grad_norm": 0.7705574631690979, "learning_rate": 2.9988757729061274e-05, "loss": 0.1983, "step": 4269 }, { "epoch": 0.1800282479920737, "grad_norm": 0.5815052390098572, "learning_rate": 2.9995784148397975e-05, "loss": 0.1791, "step": 4270 }, { "epoch": 0.18007040917427325, "grad_norm": 0.6459251046180725, "learning_rate": 3.0002810567734686e-05, "loss": 0.1708, "step": 4271 }, { "epoch": 0.1801125703564728, "grad_norm": 1.258037805557251, "learning_rate": 3.0009836987071387e-05, "loss": 1.1002, "step": 4272 }, { "epoch": 0.18015473153867234, "grad_norm": 0.5303651094436646, "learning_rate": 3.0016863406408098e-05, "loss": 0.2064, "step": 4273 }, { "epoch": 0.1801968927208719, "grad_norm": 2.356437921524048, "learning_rate": 3.00238898257448e-05, "loss": 1.1111, "step": 4274 }, { "epoch": 0.18023905390307143, "grad_norm": 0.6882047057151794, "learning_rate": 3.0030916245081507e-05, "loss": 0.2038, "step": 4275 }, { "epoch": 0.18028121508527098, "grad_norm": 0.7177526950836182, "learning_rate": 3.003794266441821e-05, "loss": 0.524, "step": 4276 }, { "epoch": 0.18032337626747055, "grad_norm": 0.6538091897964478, "learning_rate": 3.004496908375492e-05, "loss": 0.6631, "step": 4277 }, { "epoch": 0.1803655374496701, "grad_norm": 0.4716065526008606, "learning_rate": 3.0051995503091623e-05, "loss": 0.1557, "step": 4278 }, { "epoch": 0.18040769863186965, "grad_norm": 0.5972790718078613, "learning_rate": 3.005902192242833e-05, "loss": 0.1994, "step": 4279 }, { "epoch": 0.1804498598140692, "grad_norm": 0.724936842918396, "learning_rate": 3.006604834176504e-05, "loss": 0.5531, "step": 4280 }, { "epoch": 0.18049202099626874, "grad_norm": 0.8081554174423218, "learning_rate": 3.0073074761101743e-05, "loss": 1.1943, "step": 4281 }, { "epoch": 0.18053418217846828, "grad_norm": 0.5658915042877197, "learning_rate": 3.008010118043845e-05, "loss": 0.1492, "step": 4282 }, { "epoch": 0.18057634336066783, "grad_norm": 1.314383864402771, "learning_rate": 3.0087127599775156e-05, "loss": 1.0133, "step": 4283 }, { "epoch": 0.18061850454286738, "grad_norm": 0.8476178050041199, "learning_rate": 3.0094154019111863e-05, "loss": 0.5753, "step": 4284 }, { "epoch": 0.18066066572506692, "grad_norm": 0.5795540809631348, "learning_rate": 3.0101180438448568e-05, "loss": 0.2059, "step": 4285 }, { "epoch": 0.18070282690726647, "grad_norm": 0.9497121572494507, "learning_rate": 3.0108206857785276e-05, "loss": 0.5091, "step": 4286 }, { "epoch": 0.18074498808946604, "grad_norm": 1.0191516876220703, "learning_rate": 3.011523327712198e-05, "loss": 0.5096, "step": 4287 }, { "epoch": 0.1807871492716656, "grad_norm": 0.5529730319976807, "learning_rate": 3.0122259696458688e-05, "loss": 0.1972, "step": 4288 }, { "epoch": 0.18082931045386513, "grad_norm": 0.6864902377128601, "learning_rate": 3.0129286115795392e-05, "loss": 0.1881, "step": 4289 }, { "epoch": 0.18087147163606468, "grad_norm": 0.7189717888832092, "learning_rate": 3.01363125351321e-05, "loss": 0.6912, "step": 4290 }, { "epoch": 0.18091363281826423, "grad_norm": 0.554392397403717, "learning_rate": 3.0143338954468804e-05, "loss": 0.444, "step": 4291 }, { "epoch": 0.18095579400046377, "grad_norm": 0.571491539478302, "learning_rate": 3.0150365373805512e-05, "loss": 0.2464, "step": 4292 }, { "epoch": 0.18099795518266332, "grad_norm": 0.6059425473213196, "learning_rate": 3.0157391793142213e-05, "loss": 0.3565, "step": 4293 }, { "epoch": 0.18104011636486286, "grad_norm": 0.8924025893211365, "learning_rate": 3.0164418212478924e-05, "loss": 0.7323, "step": 4294 }, { "epoch": 0.1810822775470624, "grad_norm": 0.6118712425231934, "learning_rate": 3.0171444631815625e-05, "loss": 0.2538, "step": 4295 }, { "epoch": 0.18112443872926196, "grad_norm": 1.0115785598754883, "learning_rate": 3.0178471051152336e-05, "loss": 0.316, "step": 4296 }, { "epoch": 0.18116659991146153, "grad_norm": 1.3033074140548706, "learning_rate": 3.0185497470489037e-05, "loss": 0.9994, "step": 4297 }, { "epoch": 0.18120876109366107, "grad_norm": 0.6357413530349731, "learning_rate": 3.0192523889825745e-05, "loss": 0.3059, "step": 4298 }, { "epoch": 0.18125092227586062, "grad_norm": 1.1265712976455688, "learning_rate": 3.019955030916245e-05, "loss": 1.1091, "step": 4299 }, { "epoch": 0.18129308345806017, "grad_norm": 0.8984273076057434, "learning_rate": 3.0206576728499157e-05, "loss": 0.4292, "step": 4300 }, { "epoch": 0.1813352446402597, "grad_norm": 0.7463333010673523, "learning_rate": 3.0213603147835862e-05, "loss": 0.2768, "step": 4301 }, { "epoch": 0.18137740582245926, "grad_norm": 0.4402792751789093, "learning_rate": 3.022062956717257e-05, "loss": 0.1155, "step": 4302 }, { "epoch": 0.1814195670046588, "grad_norm": 1.398216962814331, "learning_rate": 3.0227655986509277e-05, "loss": 0.9983, "step": 4303 }, { "epoch": 0.18146172818685835, "grad_norm": 1.020882487297058, "learning_rate": 3.0234682405845982e-05, "loss": 1.1432, "step": 4304 }, { "epoch": 0.1815038893690579, "grad_norm": 0.3518645465373993, "learning_rate": 3.024170882518269e-05, "loss": 0.1039, "step": 4305 }, { "epoch": 0.18154605055125747, "grad_norm": 0.47583094239234924, "learning_rate": 3.0248735244519394e-05, "loss": 0.1545, "step": 4306 }, { "epoch": 0.18158821173345702, "grad_norm": 0.48045915365219116, "learning_rate": 3.0255761663856102e-05, "loss": 0.1332, "step": 4307 }, { "epoch": 0.18163037291565656, "grad_norm": 0.6313443183898926, "learning_rate": 3.0262788083192806e-05, "loss": 0.1857, "step": 4308 }, { "epoch": 0.1816725340978561, "grad_norm": 0.4942306578159332, "learning_rate": 3.0269814502529514e-05, "loss": 0.4679, "step": 4309 }, { "epoch": 0.18171469528005565, "grad_norm": 0.9379483461380005, "learning_rate": 3.0276840921866218e-05, "loss": 0.2078, "step": 4310 }, { "epoch": 0.1817568564622552, "grad_norm": 0.7284715175628662, "learning_rate": 3.0283867341202926e-05, "loss": 0.5816, "step": 4311 }, { "epoch": 0.18179901764445475, "grad_norm": 0.751708447933197, "learning_rate": 3.029089376053963e-05, "loss": 0.1917, "step": 4312 }, { "epoch": 0.1818411788266543, "grad_norm": 0.49072301387786865, "learning_rate": 3.0297920179876338e-05, "loss": 0.1305, "step": 4313 }, { "epoch": 0.18188334000885384, "grad_norm": 0.3937581181526184, "learning_rate": 3.0304946599213043e-05, "loss": 0.1589, "step": 4314 }, { "epoch": 0.18192550119105338, "grad_norm": 1.1434314250946045, "learning_rate": 3.031197301854975e-05, "loss": 1.2065, "step": 4315 }, { "epoch": 0.18196766237325296, "grad_norm": 1.5285346508026123, "learning_rate": 3.031899943788645e-05, "loss": 1.1047, "step": 4316 }, { "epoch": 0.1820098235554525, "grad_norm": 0.9548388719558716, "learning_rate": 3.0326025857223163e-05, "loss": 0.6179, "step": 4317 }, { "epoch": 0.18205198473765205, "grad_norm": 0.758074939250946, "learning_rate": 3.0333052276559864e-05, "loss": 0.5158, "step": 4318 }, { "epoch": 0.1820941459198516, "grad_norm": 0.6927825212478638, "learning_rate": 3.0340078695896575e-05, "loss": 0.5915, "step": 4319 }, { "epoch": 0.18213630710205114, "grad_norm": 0.5057649612426758, "learning_rate": 3.0347105115233276e-05, "loss": 0.1151, "step": 4320 }, { "epoch": 0.1821784682842507, "grad_norm": 0.8042638301849365, "learning_rate": 3.0354131534569984e-05, "loss": 0.5148, "step": 4321 }, { "epoch": 0.18222062946645023, "grad_norm": 0.45600399374961853, "learning_rate": 3.0361157953906688e-05, "loss": 0.2155, "step": 4322 }, { "epoch": 0.18226279064864978, "grad_norm": 0.7778016328811646, "learning_rate": 3.0368184373243396e-05, "loss": 0.6028, "step": 4323 }, { "epoch": 0.18230495183084933, "grad_norm": 0.5719259977340698, "learning_rate": 3.03752107925801e-05, "loss": 0.1751, "step": 4324 }, { "epoch": 0.18234711301304887, "grad_norm": 0.866866409778595, "learning_rate": 3.0382237211916808e-05, "loss": 0.6712, "step": 4325 }, { "epoch": 0.18238927419524845, "grad_norm": 0.648175835609436, "learning_rate": 3.0389263631253516e-05, "loss": 0.3077, "step": 4326 }, { "epoch": 0.182431435377448, "grad_norm": 0.6175505518913269, "learning_rate": 3.039629005059022e-05, "loss": 0.1865, "step": 4327 }, { "epoch": 0.18247359655964754, "grad_norm": 0.9924213290214539, "learning_rate": 3.0403316469926928e-05, "loss": 0.1949, "step": 4328 }, { "epoch": 0.18251575774184708, "grad_norm": 0.7617015838623047, "learning_rate": 3.0410342889263632e-05, "loss": 0.7243, "step": 4329 }, { "epoch": 0.18255791892404663, "grad_norm": 0.7904865741729736, "learning_rate": 3.041736930860034e-05, "loss": 0.5819, "step": 4330 }, { "epoch": 0.18260008010624618, "grad_norm": 0.48372817039489746, "learning_rate": 3.0424395727937044e-05, "loss": 0.1336, "step": 4331 }, { "epoch": 0.18264224128844572, "grad_norm": 0.4207894206047058, "learning_rate": 3.0431422147273752e-05, "loss": 0.1225, "step": 4332 }, { "epoch": 0.18268440247064527, "grad_norm": 0.46137112379074097, "learning_rate": 3.0438448566610457e-05, "loss": 0.281, "step": 4333 }, { "epoch": 0.18272656365284481, "grad_norm": 0.5244501829147339, "learning_rate": 3.0445474985947164e-05, "loss": 0.1342, "step": 4334 }, { "epoch": 0.1827687248350444, "grad_norm": 0.603782594203949, "learning_rate": 3.045250140528387e-05, "loss": 0.1973, "step": 4335 }, { "epoch": 0.18281088601724393, "grad_norm": 0.7015841007232666, "learning_rate": 3.0459527824620577e-05, "loss": 0.325, "step": 4336 }, { "epoch": 0.18285304719944348, "grad_norm": 0.74788898229599, "learning_rate": 3.0466554243957278e-05, "loss": 0.4904, "step": 4337 }, { "epoch": 0.18289520838164303, "grad_norm": 0.5308661460876465, "learning_rate": 3.047358066329399e-05, "loss": 0.4549, "step": 4338 }, { "epoch": 0.18293736956384257, "grad_norm": 0.27418461441993713, "learning_rate": 3.048060708263069e-05, "loss": 0.1044, "step": 4339 }, { "epoch": 0.18297953074604212, "grad_norm": 0.5924140810966492, "learning_rate": 3.04876335019674e-05, "loss": 0.2406, "step": 4340 }, { "epoch": 0.18302169192824166, "grad_norm": 1.7748453617095947, "learning_rate": 3.0494659921304102e-05, "loss": 0.657, "step": 4341 }, { "epoch": 0.1830638531104412, "grad_norm": 0.4046837389469147, "learning_rate": 3.0501686340640813e-05, "loss": 0.1155, "step": 4342 }, { "epoch": 0.18310601429264076, "grad_norm": 0.5851184129714966, "learning_rate": 3.0508712759977514e-05, "loss": 0.5976, "step": 4343 }, { "epoch": 0.1831481754748403, "grad_norm": 0.7916445136070251, "learning_rate": 3.0515739179314222e-05, "loss": 0.5975, "step": 4344 }, { "epoch": 0.18319033665703988, "grad_norm": 0.6309211254119873, "learning_rate": 3.0522765598650926e-05, "loss": 0.1877, "step": 4345 }, { "epoch": 0.18323249783923942, "grad_norm": 0.6347910761833191, "learning_rate": 3.052979201798764e-05, "loss": 0.2667, "step": 4346 }, { "epoch": 0.18327465902143897, "grad_norm": 0.7619715332984924, "learning_rate": 3.0536818437324335e-05, "loss": 0.3178, "step": 4347 }, { "epoch": 0.18331682020363851, "grad_norm": 0.6262138485908508, "learning_rate": 3.0543844856661046e-05, "loss": 0.2537, "step": 4348 }, { "epoch": 0.18335898138583806, "grad_norm": 0.7109354734420776, "learning_rate": 3.055087127599776e-05, "loss": 0.3243, "step": 4349 }, { "epoch": 0.1834011425680376, "grad_norm": 0.5094766616821289, "learning_rate": 3.055789769533446e-05, "loss": 0.1268, "step": 4350 }, { "epoch": 0.18344330375023715, "grad_norm": 0.6921883225440979, "learning_rate": 3.0564924114671166e-05, "loss": 0.2748, "step": 4351 }, { "epoch": 0.1834854649324367, "grad_norm": 0.44410696625709534, "learning_rate": 3.057195053400787e-05, "loss": 0.161, "step": 4352 }, { "epoch": 0.18352762611463624, "grad_norm": 0.6579286456108093, "learning_rate": 3.057897695334458e-05, "loss": 0.1716, "step": 4353 }, { "epoch": 0.1835697872968358, "grad_norm": 0.8491834402084351, "learning_rate": 3.058600337268128e-05, "loss": 0.5514, "step": 4354 }, { "epoch": 0.18361194847903536, "grad_norm": 0.5626097917556763, "learning_rate": 3.059302979201799e-05, "loss": 0.2081, "step": 4355 }, { "epoch": 0.1836541096612349, "grad_norm": 0.8502147793769836, "learning_rate": 3.0600056211354695e-05, "loss": 0.291, "step": 4356 }, { "epoch": 0.18369627084343446, "grad_norm": 0.7573407888412476, "learning_rate": 3.06070826306914e-05, "loss": 0.2115, "step": 4357 }, { "epoch": 0.183738432025634, "grad_norm": 0.3519619405269623, "learning_rate": 3.0614109050028104e-05, "loss": 0.1206, "step": 4358 }, { "epoch": 0.18378059320783355, "grad_norm": 0.6757869720458984, "learning_rate": 3.0621135469364815e-05, "loss": 0.6732, "step": 4359 }, { "epoch": 0.1838227543900331, "grad_norm": 0.7321222424507141, "learning_rate": 3.062816188870152e-05, "loss": 0.3632, "step": 4360 }, { "epoch": 0.18386491557223264, "grad_norm": 0.7034833431243896, "learning_rate": 3.0635188308038224e-05, "loss": 0.5587, "step": 4361 }, { "epoch": 0.18390707675443219, "grad_norm": 0.6677839756011963, "learning_rate": 3.064221472737493e-05, "loss": 0.1499, "step": 4362 }, { "epoch": 0.18394923793663173, "grad_norm": 0.7957937717437744, "learning_rate": 3.064924114671164e-05, "loss": 0.397, "step": 4363 }, { "epoch": 0.1839913991188313, "grad_norm": 1.0341689586639404, "learning_rate": 3.0656267566048344e-05, "loss": 1.1492, "step": 4364 }, { "epoch": 0.18403356030103085, "grad_norm": 0.48813506960868835, "learning_rate": 3.066329398538505e-05, "loss": 0.2141, "step": 4365 }, { "epoch": 0.1840757214832304, "grad_norm": 0.4698444604873657, "learning_rate": 3.067032040472175e-05, "loss": 0.2787, "step": 4366 }, { "epoch": 0.18411788266542994, "grad_norm": 0.5141905546188354, "learning_rate": 3.0677346824058464e-05, "loss": 0.2703, "step": 4367 }, { "epoch": 0.1841600438476295, "grad_norm": 0.4550652503967285, "learning_rate": 3.068437324339517e-05, "loss": 0.1487, "step": 4368 }, { "epoch": 0.18420220502982904, "grad_norm": 0.6257363557815552, "learning_rate": 3.069139966273187e-05, "loss": 0.2811, "step": 4369 }, { "epoch": 0.18424436621202858, "grad_norm": 0.987160325050354, "learning_rate": 3.0698426082068584e-05, "loss": 1.1101, "step": 4370 }, { "epoch": 0.18428652739422813, "grad_norm": 0.7820318341255188, "learning_rate": 3.070545250140529e-05, "loss": 1.1296, "step": 4371 }, { "epoch": 0.18432868857642767, "grad_norm": 0.5808510780334473, "learning_rate": 3.071247892074199e-05, "loss": 0.3299, "step": 4372 }, { "epoch": 0.18437084975862722, "grad_norm": 0.9811652302742004, "learning_rate": 3.07195053400787e-05, "loss": 1.2338, "step": 4373 }, { "epoch": 0.1844130109408268, "grad_norm": 0.7212488651275635, "learning_rate": 3.072653175941541e-05, "loss": 0.2297, "step": 4374 }, { "epoch": 0.18445517212302634, "grad_norm": 0.8952778577804565, "learning_rate": 3.0733558178752106e-05, "loss": 0.3772, "step": 4375 }, { "epoch": 0.18449733330522589, "grad_norm": 0.6095819473266602, "learning_rate": 3.074058459808882e-05, "loss": 0.1316, "step": 4376 }, { "epoch": 0.18453949448742543, "grad_norm": 1.507539987564087, "learning_rate": 3.074761101742552e-05, "loss": 1.1015, "step": 4377 }, { "epoch": 0.18458165566962498, "grad_norm": 0.872505247592926, "learning_rate": 3.075463743676223e-05, "loss": 0.7849, "step": 4378 }, { "epoch": 0.18462381685182452, "grad_norm": 0.35418832302093506, "learning_rate": 3.076166385609893e-05, "loss": 0.1161, "step": 4379 }, { "epoch": 0.18466597803402407, "grad_norm": 1.0237040519714355, "learning_rate": 3.076869027543564e-05, "loss": 0.3444, "step": 4380 }, { "epoch": 0.18470813921622362, "grad_norm": 0.8713937997817993, "learning_rate": 3.0775716694772346e-05, "loss": 0.5316, "step": 4381 }, { "epoch": 0.18475030039842316, "grad_norm": 0.5254080891609192, "learning_rate": 3.078274311410905e-05, "loss": 0.2189, "step": 4382 }, { "epoch": 0.1847924615806227, "grad_norm": 0.5107836723327637, "learning_rate": 3.0789769533445754e-05, "loss": 0.2027, "step": 4383 }, { "epoch": 0.18483462276282228, "grad_norm": 0.5594856142997742, "learning_rate": 3.0796795952782466e-05, "loss": 0.3609, "step": 4384 }, { "epoch": 0.18487678394502183, "grad_norm": 1.1217135190963745, "learning_rate": 3.080382237211917e-05, "loss": 0.3321, "step": 4385 }, { "epoch": 0.18491894512722137, "grad_norm": 0.3460027575492859, "learning_rate": 3.0810848791455874e-05, "loss": 0.116, "step": 4386 }, { "epoch": 0.18496110630942092, "grad_norm": 0.6946963667869568, "learning_rate": 3.081787521079258e-05, "loss": 0.2742, "step": 4387 }, { "epoch": 0.18500326749162047, "grad_norm": 0.9089804887771606, "learning_rate": 3.082490163012929e-05, "loss": 0.4033, "step": 4388 }, { "epoch": 0.18504542867382, "grad_norm": 0.40305253863334656, "learning_rate": 3.0831928049465994e-05, "loss": 0.1147, "step": 4389 }, { "epoch": 0.18508758985601956, "grad_norm": 0.47303658723831177, "learning_rate": 3.08389544688027e-05, "loss": 0.1686, "step": 4390 }, { "epoch": 0.1851297510382191, "grad_norm": 0.9222393035888672, "learning_rate": 3.08459808881394e-05, "loss": 0.8232, "step": 4391 }, { "epoch": 0.18517191222041865, "grad_norm": 0.5974724888801575, "learning_rate": 3.0853007307476114e-05, "loss": 0.2755, "step": 4392 }, { "epoch": 0.1852140734026182, "grad_norm": 0.7585684061050415, "learning_rate": 3.086003372681282e-05, "loss": 0.576, "step": 4393 }, { "epoch": 0.18525623458481777, "grad_norm": 0.3108837306499481, "learning_rate": 3.086706014614952e-05, "loss": 0.1189, "step": 4394 }, { "epoch": 0.18529839576701732, "grad_norm": 0.7772205471992493, "learning_rate": 3.0874086565486234e-05, "loss": 0.5958, "step": 4395 }, { "epoch": 0.18534055694921686, "grad_norm": 0.5857058763504028, "learning_rate": 3.088111298482294e-05, "loss": 0.2103, "step": 4396 }, { "epoch": 0.1853827181314164, "grad_norm": 0.616783857345581, "learning_rate": 3.088813940415964e-05, "loss": 0.337, "step": 4397 }, { "epoch": 0.18542487931361595, "grad_norm": 0.8595161437988281, "learning_rate": 3.089516582349635e-05, "loss": 0.6784, "step": 4398 }, { "epoch": 0.1854670404958155, "grad_norm": 0.3628993034362793, "learning_rate": 3.090219224283306e-05, "loss": 0.1302, "step": 4399 }, { "epoch": 0.18550920167801505, "grad_norm": 2.548837661743164, "learning_rate": 3.0909218662169756e-05, "loss": 1.0687, "step": 4400 }, { "epoch": 0.1855513628602146, "grad_norm": 0.654892086982727, "learning_rate": 3.091624508150647e-05, "loss": 0.5663, "step": 4401 }, { "epoch": 0.18559352404241414, "grad_norm": 1.2665475606918335, "learning_rate": 3.092327150084317e-05, "loss": 0.4715, "step": 4402 }, { "epoch": 0.1856356852246137, "grad_norm": 0.830374538898468, "learning_rate": 3.0930297920179876e-05, "loss": 0.2342, "step": 4403 }, { "epoch": 0.18567784640681326, "grad_norm": 0.40945568680763245, "learning_rate": 3.093732433951658e-05, "loss": 0.1623, "step": 4404 }, { "epoch": 0.1857200075890128, "grad_norm": 1.4825373888015747, "learning_rate": 3.094435075885329e-05, "loss": 1.0375, "step": 4405 }, { "epoch": 0.18576216877121235, "grad_norm": 1.420012354850769, "learning_rate": 3.0951377178189996e-05, "loss": 1.2418, "step": 4406 }, { "epoch": 0.1858043299534119, "grad_norm": 0.5511519908905029, "learning_rate": 3.09584035975267e-05, "loss": 0.1649, "step": 4407 }, { "epoch": 0.18584649113561144, "grad_norm": 0.4103066623210907, "learning_rate": 3.0965430016863405e-05, "loss": 0.2332, "step": 4408 }, { "epoch": 0.185888652317811, "grad_norm": 0.7274563312530518, "learning_rate": 3.0972456436200116e-05, "loss": 0.496, "step": 4409 }, { "epoch": 0.18593081350001053, "grad_norm": 0.4054442048072815, "learning_rate": 3.097948285553682e-05, "loss": 0.1643, "step": 4410 }, { "epoch": 0.18597297468221008, "grad_norm": 0.6888635158538818, "learning_rate": 3.0986509274873525e-05, "loss": 0.5366, "step": 4411 }, { "epoch": 0.18601513586440963, "grad_norm": 0.5513761043548584, "learning_rate": 3.099353569421023e-05, "loss": 0.193, "step": 4412 }, { "epoch": 0.1860572970466092, "grad_norm": 2.3274080753326416, "learning_rate": 3.100056211354694e-05, "loss": 0.6822, "step": 4413 }, { "epoch": 0.18609945822880875, "grad_norm": 0.7442461848258972, "learning_rate": 3.100758853288364e-05, "loss": 0.6564, "step": 4414 }, { "epoch": 0.1861416194110083, "grad_norm": 0.6186177134513855, "learning_rate": 3.101461495222035e-05, "loss": 0.4984, "step": 4415 }, { "epoch": 0.18618378059320784, "grad_norm": 0.6909957528114319, "learning_rate": 3.102164137155706e-05, "loss": 0.3895, "step": 4416 }, { "epoch": 0.18622594177540738, "grad_norm": 1.1425124406814575, "learning_rate": 3.1028667790893765e-05, "loss": 0.4831, "step": 4417 }, { "epoch": 0.18626810295760693, "grad_norm": 0.6577855348587036, "learning_rate": 3.103569421023047e-05, "loss": 0.1267, "step": 4418 }, { "epoch": 0.18631026413980648, "grad_norm": 0.34668269753456116, "learning_rate": 3.1042720629567174e-05, "loss": 0.1004, "step": 4419 }, { "epoch": 0.18635242532200602, "grad_norm": 0.631098210811615, "learning_rate": 3.1049747048903885e-05, "loss": 0.1698, "step": 4420 }, { "epoch": 0.18639458650420557, "grad_norm": 0.6116508841514587, "learning_rate": 3.105677346824058e-05, "loss": 0.1479, "step": 4421 }, { "epoch": 0.1864367476864051, "grad_norm": 0.8011828064918518, "learning_rate": 3.1063799887577294e-05, "loss": 1.1611, "step": 4422 }, { "epoch": 0.1864789088686047, "grad_norm": 0.6778179407119751, "learning_rate": 3.1070826306914e-05, "loss": 0.3322, "step": 4423 }, { "epoch": 0.18652107005080423, "grad_norm": 0.4215127229690552, "learning_rate": 3.107785272625071e-05, "loss": 0.1381, "step": 4424 }, { "epoch": 0.18656323123300378, "grad_norm": 0.605778694152832, "learning_rate": 3.108487914558741e-05, "loss": 0.4428, "step": 4425 }, { "epoch": 0.18660539241520332, "grad_norm": 0.5698724985122681, "learning_rate": 3.109190556492412e-05, "loss": 0.1442, "step": 4426 }, { "epoch": 0.18664755359740287, "grad_norm": 0.9551437497138977, "learning_rate": 3.109893198426082e-05, "loss": 0.8156, "step": 4427 }, { "epoch": 0.18668971477960242, "grad_norm": 0.5887273550033569, "learning_rate": 3.110595840359753e-05, "loss": 0.211, "step": 4428 }, { "epoch": 0.18673187596180196, "grad_norm": 0.6534777283668518, "learning_rate": 3.111298482293423e-05, "loss": 0.2831, "step": 4429 }, { "epoch": 0.1867740371440015, "grad_norm": 1.031112551689148, "learning_rate": 3.112001124227094e-05, "loss": 0.3502, "step": 4430 }, { "epoch": 0.18681619832620106, "grad_norm": 0.519027590751648, "learning_rate": 3.112703766160765e-05, "loss": 0.1173, "step": 4431 }, { "epoch": 0.18685835950840063, "grad_norm": 0.7133823037147522, "learning_rate": 3.113406408094435e-05, "loss": 0.717, "step": 4432 }, { "epoch": 0.18690052069060017, "grad_norm": 0.8338465690612793, "learning_rate": 3.1141090500281055e-05, "loss": 0.6156, "step": 4433 }, { "epoch": 0.18694268187279972, "grad_norm": 0.6253776550292969, "learning_rate": 3.1148116919617767e-05, "loss": 0.4481, "step": 4434 }, { "epoch": 0.18698484305499927, "grad_norm": 0.694473385810852, "learning_rate": 3.115514333895447e-05, "loss": 0.5238, "step": 4435 }, { "epoch": 0.1870270042371988, "grad_norm": 0.49085795879364014, "learning_rate": 3.1162169758291175e-05, "loss": 0.2686, "step": 4436 }, { "epoch": 0.18706916541939836, "grad_norm": 0.8082049489021301, "learning_rate": 3.116919617762788e-05, "loss": 0.5093, "step": 4437 }, { "epoch": 0.1871113266015979, "grad_norm": 0.5988754630088806, "learning_rate": 3.117622259696459e-05, "loss": 0.3625, "step": 4438 }, { "epoch": 0.18715348778379745, "grad_norm": 0.5434418320655823, "learning_rate": 3.1183249016301295e-05, "loss": 0.1809, "step": 4439 }, { "epoch": 0.187195648965997, "grad_norm": 0.6262903809547424, "learning_rate": 3.1190275435638e-05, "loss": 0.6233, "step": 4440 }, { "epoch": 0.18723781014819654, "grad_norm": 0.5144827365875244, "learning_rate": 3.119730185497471e-05, "loss": 0.2518, "step": 4441 }, { "epoch": 0.18727997133039612, "grad_norm": 0.9069221615791321, "learning_rate": 3.120432827431141e-05, "loss": 0.3698, "step": 4442 }, { "epoch": 0.18732213251259566, "grad_norm": 0.597357988357544, "learning_rate": 3.121135469364812e-05, "loss": 0.4656, "step": 4443 }, { "epoch": 0.1873642936947952, "grad_norm": 1.178353190422058, "learning_rate": 3.1218381112984824e-05, "loss": 0.4765, "step": 4444 }, { "epoch": 0.18740645487699475, "grad_norm": 0.39415884017944336, "learning_rate": 3.1225407532321535e-05, "loss": 0.1557, "step": 4445 }, { "epoch": 0.1874486160591943, "grad_norm": 1.1600940227508545, "learning_rate": 3.123243395165823e-05, "loss": 1.0865, "step": 4446 }, { "epoch": 0.18749077724139385, "grad_norm": 0.6786896586418152, "learning_rate": 3.1239460370994944e-05, "loss": 0.143, "step": 4447 }, { "epoch": 0.1875329384235934, "grad_norm": 1.8304293155670166, "learning_rate": 3.124648679033165e-05, "loss": 0.6141, "step": 4448 }, { "epoch": 0.18757509960579294, "grad_norm": 0.8561210632324219, "learning_rate": 3.125351320966835e-05, "loss": 0.5318, "step": 4449 }, { "epoch": 0.18761726078799248, "grad_norm": 0.7397297620773315, "learning_rate": 3.126053962900506e-05, "loss": 0.3708, "step": 4450 }, { "epoch": 0.18765942197019203, "grad_norm": 0.36949270963668823, "learning_rate": 3.126756604834177e-05, "loss": 0.1049, "step": 4451 }, { "epoch": 0.1877015831523916, "grad_norm": 0.4584369659423828, "learning_rate": 3.127459246767847e-05, "loss": 0.1547, "step": 4452 }, { "epoch": 0.18774374433459115, "grad_norm": 0.6404062509536743, "learning_rate": 3.128161888701518e-05, "loss": 0.4687, "step": 4453 }, { "epoch": 0.1877859055167907, "grad_norm": 0.41123223304748535, "learning_rate": 3.128864530635188e-05, "loss": 0.2198, "step": 4454 }, { "epoch": 0.18782806669899024, "grad_norm": 0.457210510969162, "learning_rate": 3.129567172568859e-05, "loss": 0.2552, "step": 4455 }, { "epoch": 0.1878702278811898, "grad_norm": 0.5072449445724487, "learning_rate": 3.13026981450253e-05, "loss": 0.163, "step": 4456 }, { "epoch": 0.18791238906338933, "grad_norm": 0.37531375885009766, "learning_rate": 3.1309724564362e-05, "loss": 0.2202, "step": 4457 }, { "epoch": 0.18795455024558888, "grad_norm": 0.6151559352874756, "learning_rate": 3.1316750983698706e-05, "loss": 0.3507, "step": 4458 }, { "epoch": 0.18799671142778843, "grad_norm": 0.7976768016815186, "learning_rate": 3.132377740303542e-05, "loss": 0.426, "step": 4459 }, { "epoch": 0.18803887260998797, "grad_norm": 0.9057868719100952, "learning_rate": 3.133080382237212e-05, "loss": 0.4449, "step": 4460 }, { "epoch": 0.18808103379218755, "grad_norm": 0.649997353553772, "learning_rate": 3.1337830241708826e-05, "loss": 0.7282, "step": 4461 }, { "epoch": 0.1881231949743871, "grad_norm": 0.8667358160018921, "learning_rate": 3.134485666104554e-05, "loss": 0.506, "step": 4462 }, { "epoch": 0.18816535615658664, "grad_norm": 0.6305202841758728, "learning_rate": 3.135188308038224e-05, "loss": 0.5751, "step": 4463 }, { "epoch": 0.18820751733878618, "grad_norm": 0.6496320962905884, "learning_rate": 3.1358909499718946e-05, "loss": 0.4701, "step": 4464 }, { "epoch": 0.18824967852098573, "grad_norm": 0.7209200859069824, "learning_rate": 3.136593591905565e-05, "loss": 0.5545, "step": 4465 }, { "epoch": 0.18829183970318528, "grad_norm": 0.6670485734939575, "learning_rate": 3.137296233839236e-05, "loss": 0.4687, "step": 4466 }, { "epoch": 0.18833400088538482, "grad_norm": 0.5658506155014038, "learning_rate": 3.137998875772906e-05, "loss": 0.2143, "step": 4467 }, { "epoch": 0.18837616206758437, "grad_norm": 0.69085294008255, "learning_rate": 3.138701517706577e-05, "loss": 0.2936, "step": 4468 }, { "epoch": 0.18841832324978391, "grad_norm": 0.9518995881080627, "learning_rate": 3.1394041596402475e-05, "loss": 0.235, "step": 4469 }, { "epoch": 0.18846048443198346, "grad_norm": 1.120180368423462, "learning_rate": 3.140106801573918e-05, "loss": 0.9527, "step": 4470 }, { "epoch": 0.18850264561418303, "grad_norm": 1.2390774488449097, "learning_rate": 3.1408094435075883e-05, "loss": 1.1102, "step": 4471 }, { "epoch": 0.18854480679638258, "grad_norm": 0.7636764645576477, "learning_rate": 3.1415120854412595e-05, "loss": 0.2373, "step": 4472 }, { "epoch": 0.18858696797858213, "grad_norm": 0.46688276529312134, "learning_rate": 3.14221472737493e-05, "loss": 0.1282, "step": 4473 }, { "epoch": 0.18862912916078167, "grad_norm": 0.38243868947029114, "learning_rate": 3.1429173693086e-05, "loss": 0.2069, "step": 4474 }, { "epoch": 0.18867129034298122, "grad_norm": 1.9656282663345337, "learning_rate": 3.143620011242271e-05, "loss": 0.5305, "step": 4475 }, { "epoch": 0.18871345152518076, "grad_norm": 0.4868062734603882, "learning_rate": 3.144322653175942e-05, "loss": 0.2005, "step": 4476 }, { "epoch": 0.1887556127073803, "grad_norm": 0.37551525235176086, "learning_rate": 3.145025295109612e-05, "loss": 0.1868, "step": 4477 }, { "epoch": 0.18879777388957986, "grad_norm": 0.6253876686096191, "learning_rate": 3.145727937043283e-05, "loss": 0.1925, "step": 4478 }, { "epoch": 0.1888399350717794, "grad_norm": 0.4530864953994751, "learning_rate": 3.146430578976953e-05, "loss": 0.1212, "step": 4479 }, { "epoch": 0.18888209625397895, "grad_norm": 0.7246374487876892, "learning_rate": 3.147133220910624e-05, "loss": 0.6891, "step": 4480 }, { "epoch": 0.18892425743617852, "grad_norm": 0.8159753084182739, "learning_rate": 3.147835862844295e-05, "loss": 0.7153, "step": 4481 }, { "epoch": 0.18896641861837807, "grad_norm": 0.9104269742965698, "learning_rate": 3.148538504777965e-05, "loss": 1.1703, "step": 4482 }, { "epoch": 0.18900857980057761, "grad_norm": 0.7664056420326233, "learning_rate": 3.149241146711636e-05, "loss": 0.425, "step": 4483 }, { "epoch": 0.18905074098277716, "grad_norm": 0.6349123120307922, "learning_rate": 3.149943788645307e-05, "loss": 0.2759, "step": 4484 }, { "epoch": 0.1890929021649767, "grad_norm": 0.6400662064552307, "learning_rate": 3.150646430578977e-05, "loss": 0.1819, "step": 4485 }, { "epoch": 0.18913506334717625, "grad_norm": 1.8192631006240845, "learning_rate": 3.1513490725126476e-05, "loss": 0.8513, "step": 4486 }, { "epoch": 0.1891772245293758, "grad_norm": 0.386846661567688, "learning_rate": 3.152051714446319e-05, "loss": 0.1462, "step": 4487 }, { "epoch": 0.18921938571157534, "grad_norm": 0.7542777061462402, "learning_rate": 3.1527543563799885e-05, "loss": 0.6741, "step": 4488 }, { "epoch": 0.1892615468937749, "grad_norm": 0.372501939535141, "learning_rate": 3.1534569983136596e-05, "loss": 0.1496, "step": 4489 }, { "epoch": 0.18930370807597446, "grad_norm": 0.4890865683555603, "learning_rate": 3.15415964024733e-05, "loss": 0.1569, "step": 4490 }, { "epoch": 0.189345869258174, "grad_norm": 0.5065553784370422, "learning_rate": 3.154862282181001e-05, "loss": 0.2021, "step": 4491 }, { "epoch": 0.18938803044037356, "grad_norm": 1.2868835926055908, "learning_rate": 3.155564924114671e-05, "loss": 0.5866, "step": 4492 }, { "epoch": 0.1894301916225731, "grad_norm": 0.5868709087371826, "learning_rate": 3.156267566048342e-05, "loss": 0.1749, "step": 4493 }, { "epoch": 0.18947235280477265, "grad_norm": 0.8032714128494263, "learning_rate": 3.1569702079820125e-05, "loss": 0.178, "step": 4494 }, { "epoch": 0.1895145139869722, "grad_norm": 0.9807980060577393, "learning_rate": 3.157672849915683e-05, "loss": 0.8429, "step": 4495 }, { "epoch": 0.18955667516917174, "grad_norm": 0.49626004695892334, "learning_rate": 3.1583754918493534e-05, "loss": 0.2451, "step": 4496 }, { "epoch": 0.18959883635137129, "grad_norm": 0.7162432074546814, "learning_rate": 3.1590781337830245e-05, "loss": 0.5886, "step": 4497 }, { "epoch": 0.18964099753357083, "grad_norm": 0.755374550819397, "learning_rate": 3.159780775716695e-05, "loss": 0.1776, "step": 4498 }, { "epoch": 0.18968315871577038, "grad_norm": 0.7199851274490356, "learning_rate": 3.1604834176503654e-05, "loss": 0.1146, "step": 4499 }, { "epoch": 0.18972531989796995, "grad_norm": 0.7693804502487183, "learning_rate": 3.161186059584036e-05, "loss": 0.8088, "step": 4500 }, { "epoch": 0.1897674810801695, "grad_norm": 0.7538275718688965, "learning_rate": 3.161888701517707e-05, "loss": 0.6341, "step": 4501 }, { "epoch": 0.18980964226236904, "grad_norm": 0.6933778524398804, "learning_rate": 3.1625913434513774e-05, "loss": 0.1471, "step": 4502 }, { "epoch": 0.1898518034445686, "grad_norm": 1.3331998586654663, "learning_rate": 3.163293985385048e-05, "loss": 0.4152, "step": 4503 }, { "epoch": 0.18989396462676814, "grad_norm": 0.8492370247840881, "learning_rate": 3.163996627318718e-05, "loss": 0.9461, "step": 4504 }, { "epoch": 0.18993612580896768, "grad_norm": 0.47327208518981934, "learning_rate": 3.1646992692523894e-05, "loss": 0.1794, "step": 4505 }, { "epoch": 0.18997828699116723, "grad_norm": 0.6504278779029846, "learning_rate": 3.16540191118606e-05, "loss": 0.6163, "step": 4506 }, { "epoch": 0.19002044817336677, "grad_norm": 0.7608025074005127, "learning_rate": 3.16610455311973e-05, "loss": 0.311, "step": 4507 }, { "epoch": 0.19006260935556632, "grad_norm": 0.45931240916252136, "learning_rate": 3.1668071950534014e-05, "loss": 0.2373, "step": 4508 }, { "epoch": 0.19010477053776587, "grad_norm": 1.8683125972747803, "learning_rate": 3.167509836987072e-05, "loss": 0.9542, "step": 4509 }, { "epoch": 0.19014693171996544, "grad_norm": 0.6629027128219604, "learning_rate": 3.168212478920742e-05, "loss": 0.5719, "step": 4510 }, { "epoch": 0.19018909290216499, "grad_norm": 0.5276183485984802, "learning_rate": 3.168915120854413e-05, "loss": 0.1567, "step": 4511 }, { "epoch": 0.19023125408436453, "grad_norm": 0.5371227860450745, "learning_rate": 3.169617762788084e-05, "loss": 0.2603, "step": 4512 }, { "epoch": 0.19027341526656408, "grad_norm": 0.6106306314468384, "learning_rate": 3.1703204047217536e-05, "loss": 0.1968, "step": 4513 }, { "epoch": 0.19031557644876362, "grad_norm": 0.46188753843307495, "learning_rate": 3.171023046655425e-05, "loss": 0.2103, "step": 4514 }, { "epoch": 0.19035773763096317, "grad_norm": 1.0132746696472168, "learning_rate": 3.171725688589095e-05, "loss": 0.4048, "step": 4515 }, { "epoch": 0.19039989881316272, "grad_norm": 0.5665220022201538, "learning_rate": 3.1724283305227656e-05, "loss": 0.3704, "step": 4516 }, { "epoch": 0.19044205999536226, "grad_norm": 0.4391170144081116, "learning_rate": 3.173130972456436e-05, "loss": 0.2514, "step": 4517 }, { "epoch": 0.1904842211775618, "grad_norm": 0.6543424129486084, "learning_rate": 3.173833614390107e-05, "loss": 0.1647, "step": 4518 }, { "epoch": 0.19052638235976138, "grad_norm": 0.7030213475227356, "learning_rate": 3.1745362563237776e-05, "loss": 0.7709, "step": 4519 }, { "epoch": 0.19056854354196093, "grad_norm": 0.43788230419158936, "learning_rate": 3.175238898257448e-05, "loss": 0.1075, "step": 4520 }, { "epoch": 0.19061070472416047, "grad_norm": 0.4253781735897064, "learning_rate": 3.1759415401911184e-05, "loss": 0.1061, "step": 4521 }, { "epoch": 0.19065286590636002, "grad_norm": 1.98537278175354, "learning_rate": 3.1766441821247896e-05, "loss": 1.0416, "step": 4522 }, { "epoch": 0.19069502708855957, "grad_norm": 0.8260393142700195, "learning_rate": 3.17734682405846e-05, "loss": 0.4339, "step": 4523 }, { "epoch": 0.1907371882707591, "grad_norm": 0.9440877437591553, "learning_rate": 3.1780494659921304e-05, "loss": 0.4421, "step": 4524 }, { "epoch": 0.19077934945295866, "grad_norm": 0.9904062151908875, "learning_rate": 3.178752107925801e-05, "loss": 0.7137, "step": 4525 }, { "epoch": 0.1908215106351582, "grad_norm": 0.5251874327659607, "learning_rate": 3.179454749859472e-05, "loss": 0.1259, "step": 4526 }, { "epoch": 0.19086367181735775, "grad_norm": 0.48224177956581116, "learning_rate": 3.180157391793142e-05, "loss": 0.1406, "step": 4527 }, { "epoch": 0.1909058329995573, "grad_norm": 0.548183798789978, "learning_rate": 3.180860033726813e-05, "loss": 0.1595, "step": 4528 }, { "epoch": 0.19094799418175687, "grad_norm": 0.6592427492141724, "learning_rate": 3.181562675660484e-05, "loss": 0.1477, "step": 4529 }, { "epoch": 0.19099015536395642, "grad_norm": 0.4730174243450165, "learning_rate": 3.1822653175941544e-05, "loss": 0.1226, "step": 4530 }, { "epoch": 0.19103231654615596, "grad_norm": 0.6707512140274048, "learning_rate": 3.182967959527825e-05, "loss": 0.1763, "step": 4531 }, { "epoch": 0.1910744777283555, "grad_norm": 0.5684375762939453, "learning_rate": 3.183670601461495e-05, "loss": 0.3147, "step": 4532 }, { "epoch": 0.19111663891055505, "grad_norm": 0.8903200626373291, "learning_rate": 3.1843732433951664e-05, "loss": 1.2876, "step": 4533 }, { "epoch": 0.1911588000927546, "grad_norm": 0.6913869976997375, "learning_rate": 3.185075885328836e-05, "loss": 0.321, "step": 4534 }, { "epoch": 0.19120096127495415, "grad_norm": 0.890289843082428, "learning_rate": 3.185778527262507e-05, "loss": 0.198, "step": 4535 }, { "epoch": 0.1912431224571537, "grad_norm": 0.451748251914978, "learning_rate": 3.186481169196178e-05, "loss": 0.1476, "step": 4536 }, { "epoch": 0.19128528363935324, "grad_norm": 0.6343519687652588, "learning_rate": 3.187183811129849e-05, "loss": 0.1683, "step": 4537 }, { "epoch": 0.19132744482155278, "grad_norm": 0.45997199416160583, "learning_rate": 3.1878864530635186e-05, "loss": 0.1867, "step": 4538 }, { "epoch": 0.19136960600375236, "grad_norm": 0.8395001888275146, "learning_rate": 3.18858909499719e-05, "loss": 0.4293, "step": 4539 }, { "epoch": 0.1914117671859519, "grad_norm": 1.255417823791504, "learning_rate": 3.18929173693086e-05, "loss": 0.3521, "step": 4540 }, { "epoch": 0.19145392836815145, "grad_norm": 1.2005637884140015, "learning_rate": 3.1899943788645306e-05, "loss": 1.1613, "step": 4541 }, { "epoch": 0.191496089550351, "grad_norm": 0.6747391819953918, "learning_rate": 3.190697020798201e-05, "loss": 0.1658, "step": 4542 }, { "epoch": 0.19153825073255054, "grad_norm": 0.6673397421836853, "learning_rate": 3.191399662731872e-05, "loss": 0.5957, "step": 4543 }, { "epoch": 0.1915804119147501, "grad_norm": 0.7793207764625549, "learning_rate": 3.1921023046655426e-05, "loss": 0.5643, "step": 4544 }, { "epoch": 0.19162257309694963, "grad_norm": 2.076409339904785, "learning_rate": 3.192804946599213e-05, "loss": 0.8091, "step": 4545 }, { "epoch": 0.19166473427914918, "grad_norm": 0.7739682197570801, "learning_rate": 3.1935075885328835e-05, "loss": 0.804, "step": 4546 }, { "epoch": 0.19170689546134873, "grad_norm": 0.445280522108078, "learning_rate": 3.1942102304665546e-05, "loss": 0.0966, "step": 4547 }, { "epoch": 0.19174905664354827, "grad_norm": 0.8489185571670532, "learning_rate": 3.194912872400225e-05, "loss": 0.7042, "step": 4548 }, { "epoch": 0.19179121782574785, "grad_norm": 0.711320161819458, "learning_rate": 3.1956155143338955e-05, "loss": 0.2264, "step": 4549 }, { "epoch": 0.1918333790079474, "grad_norm": 0.9394399523735046, "learning_rate": 3.1963181562675666e-05, "loss": 0.1459, "step": 4550 }, { "epoch": 0.19187554019014694, "grad_norm": 0.3372669816017151, "learning_rate": 3.197020798201237e-05, "loss": 0.1314, "step": 4551 }, { "epoch": 0.19191770137234648, "grad_norm": 0.7373605966567993, "learning_rate": 3.1977234401349075e-05, "loss": 0.3156, "step": 4552 }, { "epoch": 0.19195986255454603, "grad_norm": 0.6721097826957703, "learning_rate": 3.198426082068578e-05, "loss": 0.2345, "step": 4553 }, { "epoch": 0.19200202373674558, "grad_norm": 0.6664932370185852, "learning_rate": 3.199128724002249e-05, "loss": 0.6761, "step": 4554 }, { "epoch": 0.19204418491894512, "grad_norm": 0.7290798425674438, "learning_rate": 3.199831365935919e-05, "loss": 0.5342, "step": 4555 }, { "epoch": 0.19208634610114467, "grad_norm": 0.5744290351867676, "learning_rate": 3.20053400786959e-05, "loss": 0.3624, "step": 4556 }, { "epoch": 0.1921285072833442, "grad_norm": 0.7500872015953064, "learning_rate": 3.2012366498032604e-05, "loss": 0.2407, "step": 4557 }, { "epoch": 0.1921706684655438, "grad_norm": 0.4345775544643402, "learning_rate": 3.2019392917369315e-05, "loss": 0.174, "step": 4558 }, { "epoch": 0.19221282964774333, "grad_norm": 0.803360104560852, "learning_rate": 3.202641933670601e-05, "loss": 0.195, "step": 4559 }, { "epoch": 0.19225499082994288, "grad_norm": 2.475316286087036, "learning_rate": 3.2033445756042724e-05, "loss": 0.7595, "step": 4560 }, { "epoch": 0.19229715201214242, "grad_norm": 0.6665798425674438, "learning_rate": 3.204047217537943e-05, "loss": 0.2184, "step": 4561 }, { "epoch": 0.19233931319434197, "grad_norm": 1.0213947296142578, "learning_rate": 3.204749859471613e-05, "loss": 0.2505, "step": 4562 }, { "epoch": 0.19238147437654152, "grad_norm": 0.6905348300933838, "learning_rate": 3.205452501405284e-05, "loss": 0.2626, "step": 4563 }, { "epoch": 0.19242363555874106, "grad_norm": 0.6166660189628601, "learning_rate": 3.206155143338955e-05, "loss": 0.1945, "step": 4564 }, { "epoch": 0.1924657967409406, "grad_norm": 1.25010085105896, "learning_rate": 3.206857785272625e-05, "loss": 1.3219, "step": 4565 }, { "epoch": 0.19250795792314015, "grad_norm": 1.1047301292419434, "learning_rate": 3.207560427206296e-05, "loss": 0.3634, "step": 4566 }, { "epoch": 0.1925501191053397, "grad_norm": 0.6404728293418884, "learning_rate": 3.208263069139966e-05, "loss": 0.4301, "step": 4567 }, { "epoch": 0.19259228028753927, "grad_norm": 0.5309736132621765, "learning_rate": 3.208965711073637e-05, "loss": 0.3881, "step": 4568 }, { "epoch": 0.19263444146973882, "grad_norm": 0.7863937616348267, "learning_rate": 3.209668353007308e-05, "loss": 0.4808, "step": 4569 }, { "epoch": 0.19267660265193837, "grad_norm": 0.8859933018684387, "learning_rate": 3.210370994940978e-05, "loss": 0.3285, "step": 4570 }, { "epoch": 0.1927187638341379, "grad_norm": 0.7352495193481445, "learning_rate": 3.2110736368746486e-05, "loss": 0.3243, "step": 4571 }, { "epoch": 0.19276092501633746, "grad_norm": 0.5143483281135559, "learning_rate": 3.21177627880832e-05, "loss": 0.2279, "step": 4572 }, { "epoch": 0.192803086198537, "grad_norm": 0.8830301761627197, "learning_rate": 3.21247892074199e-05, "loss": 0.6766, "step": 4573 }, { "epoch": 0.19284524738073655, "grad_norm": 0.7258024215698242, "learning_rate": 3.2131815626756606e-05, "loss": 0.3111, "step": 4574 }, { "epoch": 0.1928874085629361, "grad_norm": 1.010063886642456, "learning_rate": 3.213884204609332e-05, "loss": 0.292, "step": 4575 }, { "epoch": 0.19292956974513564, "grad_norm": 0.9375436305999756, "learning_rate": 3.214586846543002e-05, "loss": 0.3663, "step": 4576 }, { "epoch": 0.1929717309273352, "grad_norm": 1.170809268951416, "learning_rate": 3.2152894884766725e-05, "loss": 0.473, "step": 4577 }, { "epoch": 0.19301389210953476, "grad_norm": 0.610974133014679, "learning_rate": 3.215992130410343e-05, "loss": 0.161, "step": 4578 }, { "epoch": 0.1930560532917343, "grad_norm": 0.6792086958885193, "learning_rate": 3.216694772344014e-05, "loss": 0.2224, "step": 4579 }, { "epoch": 0.19309821447393385, "grad_norm": 0.6967638731002808, "learning_rate": 3.217397414277684e-05, "loss": 0.4467, "step": 4580 }, { "epoch": 0.1931403756561334, "grad_norm": 0.44037240743637085, "learning_rate": 3.218100056211355e-05, "loss": 0.231, "step": 4581 }, { "epoch": 0.19318253683833295, "grad_norm": 0.7837972044944763, "learning_rate": 3.2188026981450254e-05, "loss": 0.2228, "step": 4582 }, { "epoch": 0.1932246980205325, "grad_norm": 0.5307722687721252, "learning_rate": 3.219505340078696e-05, "loss": 0.3182, "step": 4583 }, { "epoch": 0.19326685920273204, "grad_norm": 1.35572350025177, "learning_rate": 3.220207982012366e-05, "loss": 1.1482, "step": 4584 }, { "epoch": 0.19330902038493158, "grad_norm": 0.7331713438034058, "learning_rate": 3.2209106239460374e-05, "loss": 0.5249, "step": 4585 }, { "epoch": 0.19335118156713113, "grad_norm": 0.8515654802322388, "learning_rate": 3.221613265879708e-05, "loss": 0.3203, "step": 4586 }, { "epoch": 0.1933933427493307, "grad_norm": 0.6947283148765564, "learning_rate": 3.222315907813378e-05, "loss": 0.1939, "step": 4587 }, { "epoch": 0.19343550393153025, "grad_norm": 0.8239100575447083, "learning_rate": 3.223018549747049e-05, "loss": 0.2898, "step": 4588 }, { "epoch": 0.1934776651137298, "grad_norm": 0.4453625977039337, "learning_rate": 3.22372119168072e-05, "loss": 0.1562, "step": 4589 }, { "epoch": 0.19351982629592934, "grad_norm": 0.6531926989555359, "learning_rate": 3.22442383361439e-05, "loss": 0.2833, "step": 4590 }, { "epoch": 0.1935619874781289, "grad_norm": 0.6026893854141235, "learning_rate": 3.225126475548061e-05, "loss": 0.4683, "step": 4591 }, { "epoch": 0.19360414866032843, "grad_norm": 0.42997506260871887, "learning_rate": 3.225829117481731e-05, "loss": 0.1004, "step": 4592 }, { "epoch": 0.19364630984252798, "grad_norm": 0.8295729756355286, "learning_rate": 3.226531759415402e-05, "loss": 0.7004, "step": 4593 }, { "epoch": 0.19368847102472753, "grad_norm": 0.7829402089118958, "learning_rate": 3.227234401349073e-05, "loss": 0.3392, "step": 4594 }, { "epoch": 0.19373063220692707, "grad_norm": 1.2008484601974487, "learning_rate": 3.227937043282743e-05, "loss": 0.2494, "step": 4595 }, { "epoch": 0.19377279338912662, "grad_norm": 0.33091992139816284, "learning_rate": 3.228639685216414e-05, "loss": 0.1732, "step": 4596 }, { "epoch": 0.1938149545713262, "grad_norm": 0.9878289103507996, "learning_rate": 3.229342327150085e-05, "loss": 1.188, "step": 4597 }, { "epoch": 0.19385711575352574, "grad_norm": 0.6408246755599976, "learning_rate": 3.230044969083755e-05, "loss": 0.6701, "step": 4598 }, { "epoch": 0.19389927693572528, "grad_norm": 0.5301576852798462, "learning_rate": 3.2307476110174256e-05, "loss": 0.1501, "step": 4599 }, { "epoch": 0.19394143811792483, "grad_norm": 0.5001927018165588, "learning_rate": 3.231450252951097e-05, "loss": 0.1713, "step": 4600 }, { "epoch": 0.19398359930012438, "grad_norm": 0.469438374042511, "learning_rate": 3.2321528948847665e-05, "loss": 0.1252, "step": 4601 }, { "epoch": 0.19402576048232392, "grad_norm": 0.6646157503128052, "learning_rate": 3.2328555368184376e-05, "loss": 0.4141, "step": 4602 }, { "epoch": 0.19406792166452347, "grad_norm": 0.6824544072151184, "learning_rate": 3.233558178752108e-05, "loss": 0.1778, "step": 4603 }, { "epoch": 0.19411008284672301, "grad_norm": 2.2700982093811035, "learning_rate": 3.234260820685779e-05, "loss": 0.6078, "step": 4604 }, { "epoch": 0.19415224402892256, "grad_norm": 0.7231025099754333, "learning_rate": 3.234963462619449e-05, "loss": 0.4344, "step": 4605 }, { "epoch": 0.1941944052111221, "grad_norm": 0.4500519335269928, "learning_rate": 3.23566610455312e-05, "loss": 0.3092, "step": 4606 }, { "epoch": 0.19423656639332168, "grad_norm": 0.7439253330230713, "learning_rate": 3.2363687464867905e-05, "loss": 0.1994, "step": 4607 }, { "epoch": 0.19427872757552123, "grad_norm": 0.7831011414527893, "learning_rate": 3.237071388420461e-05, "loss": 0.3142, "step": 4608 }, { "epoch": 0.19432088875772077, "grad_norm": 0.7904857397079468, "learning_rate": 3.2377740303541314e-05, "loss": 0.7184, "step": 4609 }, { "epoch": 0.19436304993992032, "grad_norm": 0.7031316757202148, "learning_rate": 3.2384766722878025e-05, "loss": 0.6399, "step": 4610 }, { "epoch": 0.19440521112211986, "grad_norm": 0.4061189889907837, "learning_rate": 3.239179314221473e-05, "loss": 0.1131, "step": 4611 }, { "epoch": 0.1944473723043194, "grad_norm": 0.5582958459854126, "learning_rate": 3.2398819561551434e-05, "loss": 0.3023, "step": 4612 }, { "epoch": 0.19448953348651896, "grad_norm": 0.9558930993080139, "learning_rate": 3.240584598088814e-05, "loss": 0.3206, "step": 4613 }, { "epoch": 0.1945316946687185, "grad_norm": 1.0618537664413452, "learning_rate": 3.241287240022485e-05, "loss": 0.9426, "step": 4614 }, { "epoch": 0.19457385585091805, "grad_norm": 0.6095786690711975, "learning_rate": 3.2419898819561553e-05, "loss": 0.173, "step": 4615 }, { "epoch": 0.19461601703311762, "grad_norm": 0.8648508787155151, "learning_rate": 3.242692523889826e-05, "loss": 0.6823, "step": 4616 }, { "epoch": 0.19465817821531717, "grad_norm": 1.3351714611053467, "learning_rate": 3.243395165823496e-05, "loss": 1.1802, "step": 4617 }, { "epoch": 0.19470033939751671, "grad_norm": 0.6401974558830261, "learning_rate": 3.2440978077571673e-05, "loss": 0.5816, "step": 4618 }, { "epoch": 0.19474250057971626, "grad_norm": 0.5848845839500427, "learning_rate": 3.244800449690838e-05, "loss": 0.5513, "step": 4619 }, { "epoch": 0.1947846617619158, "grad_norm": 0.7905059456825256, "learning_rate": 3.245503091624508e-05, "loss": 0.6177, "step": 4620 }, { "epoch": 0.19482682294411535, "grad_norm": 0.670996904373169, "learning_rate": 3.2462057335581793e-05, "loss": 0.5373, "step": 4621 }, { "epoch": 0.1948689841263149, "grad_norm": 0.8138392567634583, "learning_rate": 3.24690837549185e-05, "loss": 0.3421, "step": 4622 }, { "epoch": 0.19491114530851444, "grad_norm": 0.6846956610679626, "learning_rate": 3.24761101742552e-05, "loss": 0.2097, "step": 4623 }, { "epoch": 0.194953306490714, "grad_norm": 1.7154771089553833, "learning_rate": 3.2483136593591907e-05, "loss": 0.8727, "step": 4624 }, { "epoch": 0.19499546767291354, "grad_norm": 0.6666421294212341, "learning_rate": 3.249016301292862e-05, "loss": 0.6529, "step": 4625 }, { "epoch": 0.1950376288551131, "grad_norm": 0.7348392009735107, "learning_rate": 3.2497189432265315e-05, "loss": 0.1485, "step": 4626 }, { "epoch": 0.19507979003731266, "grad_norm": 0.6726338863372803, "learning_rate": 3.2504215851602027e-05, "loss": 0.2002, "step": 4627 }, { "epoch": 0.1951219512195122, "grad_norm": 0.707637369632721, "learning_rate": 3.251124227093873e-05, "loss": 0.1805, "step": 4628 }, { "epoch": 0.19516411240171175, "grad_norm": 0.7866166830062866, "learning_rate": 3.2518268690275435e-05, "loss": 0.3719, "step": 4629 }, { "epoch": 0.1952062735839113, "grad_norm": 0.6568343639373779, "learning_rate": 3.252529510961214e-05, "loss": 0.7222, "step": 4630 }, { "epoch": 0.19524843476611084, "grad_norm": 0.6375394463539124, "learning_rate": 3.253232152894885e-05, "loss": 0.2851, "step": 4631 }, { "epoch": 0.19529059594831039, "grad_norm": 0.9062455892562866, "learning_rate": 3.2539347948285555e-05, "loss": 0.4556, "step": 4632 }, { "epoch": 0.19533275713050993, "grad_norm": 0.5557794570922852, "learning_rate": 3.254637436762226e-05, "loss": 0.2375, "step": 4633 }, { "epoch": 0.19537491831270948, "grad_norm": 0.5492863059043884, "learning_rate": 3.2553400786958964e-05, "loss": 0.1755, "step": 4634 }, { "epoch": 0.19541707949490902, "grad_norm": 0.44956615567207336, "learning_rate": 3.2560427206295675e-05, "loss": 0.147, "step": 4635 }, { "epoch": 0.1954592406771086, "grad_norm": 0.8211267590522766, "learning_rate": 3.256745362563238e-05, "loss": 0.4244, "step": 4636 }, { "epoch": 0.19550140185930814, "grad_norm": 0.6760432124137878, "learning_rate": 3.2574480044969084e-05, "loss": 0.6412, "step": 4637 }, { "epoch": 0.1955435630415077, "grad_norm": 0.6903245449066162, "learning_rate": 3.258150646430579e-05, "loss": 0.4589, "step": 4638 }, { "epoch": 0.19558572422370724, "grad_norm": 0.46673181653022766, "learning_rate": 3.25885328836425e-05, "loss": 0.2829, "step": 4639 }, { "epoch": 0.19562788540590678, "grad_norm": 0.6954813599586487, "learning_rate": 3.2595559302979204e-05, "loss": 0.6097, "step": 4640 }, { "epoch": 0.19567004658810633, "grad_norm": 0.6522089242935181, "learning_rate": 3.260258572231591e-05, "loss": 0.574, "step": 4641 }, { "epoch": 0.19571220777030587, "grad_norm": 0.8895670175552368, "learning_rate": 3.260961214165262e-05, "loss": 0.2981, "step": 4642 }, { "epoch": 0.19575436895250542, "grad_norm": 1.5246843099594116, "learning_rate": 3.2616638560989324e-05, "loss": 1.1922, "step": 4643 }, { "epoch": 0.19579653013470497, "grad_norm": 1.0507959127426147, "learning_rate": 3.262366498032603e-05, "loss": 0.4814, "step": 4644 }, { "epoch": 0.19583869131690454, "grad_norm": 0.7431339025497437, "learning_rate": 3.263069139966273e-05, "loss": 0.3006, "step": 4645 }, { "epoch": 0.19588085249910409, "grad_norm": 0.9154266715049744, "learning_rate": 3.2637717818999444e-05, "loss": 0.5557, "step": 4646 }, { "epoch": 0.19592301368130363, "grad_norm": 0.4440683126449585, "learning_rate": 3.264474423833614e-05, "loss": 0.1966, "step": 4647 }, { "epoch": 0.19596517486350318, "grad_norm": 0.7002137303352356, "learning_rate": 3.265177065767285e-05, "loss": 0.1333, "step": 4648 }, { "epoch": 0.19600733604570272, "grad_norm": 0.7605305910110474, "learning_rate": 3.265879707700956e-05, "loss": 0.2812, "step": 4649 }, { "epoch": 0.19604949722790227, "grad_norm": 0.4591108560562134, "learning_rate": 3.266582349634627e-05, "loss": 0.2624, "step": 4650 }, { "epoch": 0.19609165841010182, "grad_norm": 0.5320200324058533, "learning_rate": 3.2672849915682966e-05, "loss": 0.1193, "step": 4651 }, { "epoch": 0.19613381959230136, "grad_norm": 0.5605942010879517, "learning_rate": 3.267987633501968e-05, "loss": 0.3347, "step": 4652 }, { "epoch": 0.1961759807745009, "grad_norm": 0.4958604574203491, "learning_rate": 3.268690275435638e-05, "loss": 0.2011, "step": 4653 }, { "epoch": 0.19621814195670045, "grad_norm": 0.5899877548217773, "learning_rate": 3.2693929173693086e-05, "loss": 0.6077, "step": 4654 }, { "epoch": 0.19626030313890003, "grad_norm": 1.1416642665863037, "learning_rate": 3.270095559302979e-05, "loss": 0.9723, "step": 4655 }, { "epoch": 0.19630246432109957, "grad_norm": 0.753693163394928, "learning_rate": 3.27079820123665e-05, "loss": 0.2464, "step": 4656 }, { "epoch": 0.19634462550329912, "grad_norm": 0.5668436288833618, "learning_rate": 3.2715008431703206e-05, "loss": 0.5366, "step": 4657 }, { "epoch": 0.19638678668549867, "grad_norm": 0.5078002214431763, "learning_rate": 3.272203485103991e-05, "loss": 0.2042, "step": 4658 }, { "epoch": 0.1964289478676982, "grad_norm": 0.6252536773681641, "learning_rate": 3.2729061270376615e-05, "loss": 0.1722, "step": 4659 }, { "epoch": 0.19647110904989776, "grad_norm": 0.7398678064346313, "learning_rate": 3.2736087689713326e-05, "loss": 0.6573, "step": 4660 }, { "epoch": 0.1965132702320973, "grad_norm": 0.4889282286167145, "learning_rate": 3.274311410905003e-05, "loss": 0.1726, "step": 4661 }, { "epoch": 0.19655543141429685, "grad_norm": 0.616923987865448, "learning_rate": 3.2750140528386735e-05, "loss": 0.179, "step": 4662 }, { "epoch": 0.1965975925964964, "grad_norm": 0.2615462839603424, "learning_rate": 3.2757166947723446e-05, "loss": 0.1081, "step": 4663 }, { "epoch": 0.19663975377869594, "grad_norm": 0.464040070772171, "learning_rate": 3.276419336706015e-05, "loss": 0.1603, "step": 4664 }, { "epoch": 0.19668191496089552, "grad_norm": 0.889462947845459, "learning_rate": 3.2771219786396855e-05, "loss": 0.3555, "step": 4665 }, { "epoch": 0.19672407614309506, "grad_norm": 0.9834058284759521, "learning_rate": 3.277824620573356e-05, "loss": 0.9863, "step": 4666 }, { "epoch": 0.1967662373252946, "grad_norm": 0.9463643431663513, "learning_rate": 3.278527262507027e-05, "loss": 0.4705, "step": 4667 }, { "epoch": 0.19680839850749415, "grad_norm": 0.5305932760238647, "learning_rate": 3.279229904440697e-05, "loss": 0.1834, "step": 4668 }, { "epoch": 0.1968505596896937, "grad_norm": 0.5704746842384338, "learning_rate": 3.279932546374368e-05, "loss": 0.1728, "step": 4669 }, { "epoch": 0.19689272087189325, "grad_norm": 0.4449191391468048, "learning_rate": 3.280635188308038e-05, "loss": 0.1347, "step": 4670 }, { "epoch": 0.1969348820540928, "grad_norm": 0.5637027621269226, "learning_rate": 3.2813378302417094e-05, "loss": 0.1295, "step": 4671 }, { "epoch": 0.19697704323629234, "grad_norm": 0.7419998049736023, "learning_rate": 3.282040472175379e-05, "loss": 0.4419, "step": 4672 }, { "epoch": 0.19701920441849188, "grad_norm": 0.6053442358970642, "learning_rate": 3.28274311410905e-05, "loss": 0.253, "step": 4673 }, { "epoch": 0.19706136560069146, "grad_norm": 0.526188850402832, "learning_rate": 3.283445756042721e-05, "loss": 0.1318, "step": 4674 }, { "epoch": 0.197103526782891, "grad_norm": 0.7905117869377136, "learning_rate": 3.284148397976391e-05, "loss": 0.6613, "step": 4675 }, { "epoch": 0.19714568796509055, "grad_norm": 0.526832640171051, "learning_rate": 3.2848510399100616e-05, "loss": 0.2291, "step": 4676 }, { "epoch": 0.1971878491472901, "grad_norm": 0.5181370973587036, "learning_rate": 3.285553681843733e-05, "loss": 0.2082, "step": 4677 }, { "epoch": 0.19723001032948964, "grad_norm": 0.6020716428756714, "learning_rate": 3.286256323777403e-05, "loss": 0.1791, "step": 4678 }, { "epoch": 0.1972721715116892, "grad_norm": 0.7857298851013184, "learning_rate": 3.2869589657110736e-05, "loss": 0.2707, "step": 4679 }, { "epoch": 0.19731433269388873, "grad_norm": 0.7712735533714294, "learning_rate": 3.287661607644744e-05, "loss": 0.5587, "step": 4680 }, { "epoch": 0.19735649387608828, "grad_norm": 0.43296054005622864, "learning_rate": 3.288364249578415e-05, "loss": 0.1054, "step": 4681 }, { "epoch": 0.19739865505828783, "grad_norm": 1.711440086364746, "learning_rate": 3.2890668915120856e-05, "loss": 0.3944, "step": 4682 }, { "epoch": 0.19744081624048737, "grad_norm": 0.7871200442314148, "learning_rate": 3.289769533445756e-05, "loss": 0.3246, "step": 4683 }, { "epoch": 0.19748297742268694, "grad_norm": 1.8213626146316528, "learning_rate": 3.2904721753794265e-05, "loss": 1.0069, "step": 4684 }, { "epoch": 0.1975251386048865, "grad_norm": 0.4468825161457062, "learning_rate": 3.2911748173130976e-05, "loss": 0.1401, "step": 4685 }, { "epoch": 0.19756729978708604, "grad_norm": 0.3996031880378723, "learning_rate": 3.291877459246768e-05, "loss": 0.1057, "step": 4686 }, { "epoch": 0.19760946096928558, "grad_norm": 0.7152394652366638, "learning_rate": 3.2925801011804385e-05, "loss": 0.1956, "step": 4687 }, { "epoch": 0.19765162215148513, "grad_norm": 1.2789504528045654, "learning_rate": 3.2932827431141096e-05, "loss": 0.8578, "step": 4688 }, { "epoch": 0.19769378333368468, "grad_norm": 0.7365677952766418, "learning_rate": 3.29398538504778e-05, "loss": 0.5365, "step": 4689 }, { "epoch": 0.19773594451588422, "grad_norm": 0.5846062302589417, "learning_rate": 3.2946880269814505e-05, "loss": 0.2342, "step": 4690 }, { "epoch": 0.19777810569808377, "grad_norm": 0.5786671042442322, "learning_rate": 3.295390668915121e-05, "loss": 0.653, "step": 4691 }, { "epoch": 0.1978202668802833, "grad_norm": 0.5107934474945068, "learning_rate": 3.296093310848792e-05, "loss": 0.1705, "step": 4692 }, { "epoch": 0.19786242806248286, "grad_norm": 0.7290273904800415, "learning_rate": 3.296795952782462e-05, "loss": 0.1563, "step": 4693 }, { "epoch": 0.19790458924468243, "grad_norm": 0.7676194906234741, "learning_rate": 3.297498594716133e-05, "loss": 0.2719, "step": 4694 }, { "epoch": 0.19794675042688198, "grad_norm": 0.6748887896537781, "learning_rate": 3.2982012366498034e-05, "loss": 0.1892, "step": 4695 }, { "epoch": 0.19798891160908152, "grad_norm": 0.5344743132591248, "learning_rate": 3.298903878583474e-05, "loss": 0.1758, "step": 4696 }, { "epoch": 0.19803107279128107, "grad_norm": 0.4457390010356903, "learning_rate": 3.299606520517144e-05, "loss": 0.1506, "step": 4697 }, { "epoch": 0.19807323397348062, "grad_norm": 0.5996521711349487, "learning_rate": 3.3003091624508154e-05, "loss": 0.2567, "step": 4698 }, { "epoch": 0.19811539515568016, "grad_norm": 0.7242110371589661, "learning_rate": 3.301011804384486e-05, "loss": 0.268, "step": 4699 }, { "epoch": 0.1981575563378797, "grad_norm": 0.644127368927002, "learning_rate": 3.301714446318156e-05, "loss": 0.3425, "step": 4700 }, { "epoch": 0.19819971752007925, "grad_norm": 0.928059458732605, "learning_rate": 3.302417088251827e-05, "loss": 1.2, "step": 4701 }, { "epoch": 0.1982418787022788, "grad_norm": 1.2542656660079956, "learning_rate": 3.303119730185498e-05, "loss": 1.095, "step": 4702 }, { "epoch": 0.19828403988447835, "grad_norm": 0.4683016836643219, "learning_rate": 3.303822372119168e-05, "loss": 0.1453, "step": 4703 }, { "epoch": 0.19832620106667792, "grad_norm": 0.8301027417182922, "learning_rate": 3.304525014052839e-05, "loss": 0.3236, "step": 4704 }, { "epoch": 0.19836836224887747, "grad_norm": 0.9263986349105835, "learning_rate": 3.305227655986509e-05, "loss": 0.5432, "step": 4705 }, { "epoch": 0.198410523431077, "grad_norm": 0.5364413857460022, "learning_rate": 3.30593029792018e-05, "loss": 0.2525, "step": 4706 }, { "epoch": 0.19845268461327656, "grad_norm": 0.4585818946361542, "learning_rate": 3.306632939853851e-05, "loss": 0.1989, "step": 4707 }, { "epoch": 0.1984948457954761, "grad_norm": 0.401481568813324, "learning_rate": 3.307335581787521e-05, "loss": 0.1237, "step": 4708 }, { "epoch": 0.19853700697767565, "grad_norm": 0.4152624309062958, "learning_rate": 3.308038223721192e-05, "loss": 0.0654, "step": 4709 }, { "epoch": 0.1985791681598752, "grad_norm": 0.8654587864875793, "learning_rate": 3.308740865654863e-05, "loss": 0.3024, "step": 4710 }, { "epoch": 0.19862132934207474, "grad_norm": 0.644884467124939, "learning_rate": 3.309443507588533e-05, "loss": 0.2536, "step": 4711 }, { "epoch": 0.1986634905242743, "grad_norm": 0.6164283156394958, "learning_rate": 3.3101461495222036e-05, "loss": 0.147, "step": 4712 }, { "epoch": 0.19870565170647386, "grad_norm": 0.47879794239997864, "learning_rate": 3.310848791455875e-05, "loss": 0.0981, "step": 4713 }, { "epoch": 0.1987478128886734, "grad_norm": 0.6657012701034546, "learning_rate": 3.3115514333895444e-05, "loss": 0.6526, "step": 4714 }, { "epoch": 0.19878997407087295, "grad_norm": 1.144178867340088, "learning_rate": 3.3122540753232156e-05, "loss": 0.5887, "step": 4715 }, { "epoch": 0.1988321352530725, "grad_norm": 0.3539597988128662, "learning_rate": 3.312956717256886e-05, "loss": 0.1013, "step": 4716 }, { "epoch": 0.19887429643527205, "grad_norm": 1.098400354385376, "learning_rate": 3.313659359190557e-05, "loss": 1.1601, "step": 4717 }, { "epoch": 0.1989164576174716, "grad_norm": 0.9728578329086304, "learning_rate": 3.314362001124227e-05, "loss": 0.3771, "step": 4718 }, { "epoch": 0.19895861879967114, "grad_norm": 0.9064515233039856, "learning_rate": 3.315064643057898e-05, "loss": 0.3158, "step": 4719 }, { "epoch": 0.19900077998187068, "grad_norm": 0.758743405342102, "learning_rate": 3.3157672849915684e-05, "loss": 0.195, "step": 4720 }, { "epoch": 0.19904294116407023, "grad_norm": 0.6547462940216064, "learning_rate": 3.316469926925239e-05, "loss": 0.6668, "step": 4721 }, { "epoch": 0.19908510234626978, "grad_norm": 0.6942564845085144, "learning_rate": 3.317172568858909e-05, "loss": 0.1816, "step": 4722 }, { "epoch": 0.19912726352846935, "grad_norm": 0.9965881705284119, "learning_rate": 3.3178752107925804e-05, "loss": 0.9842, "step": 4723 }, { "epoch": 0.1991694247106689, "grad_norm": 0.4180527329444885, "learning_rate": 3.318577852726251e-05, "loss": 0.1227, "step": 4724 }, { "epoch": 0.19921158589286844, "grad_norm": 0.5113368630409241, "learning_rate": 3.319280494659921e-05, "loss": 0.232, "step": 4725 }, { "epoch": 0.199253747075068, "grad_norm": 0.5013854503631592, "learning_rate": 3.319983136593592e-05, "loss": 0.2129, "step": 4726 }, { "epoch": 0.19929590825726753, "grad_norm": 0.5596112608909607, "learning_rate": 3.320685778527263e-05, "loss": 0.3854, "step": 4727 }, { "epoch": 0.19933806943946708, "grad_norm": 0.576474130153656, "learning_rate": 3.321388420460933e-05, "loss": 0.5031, "step": 4728 }, { "epoch": 0.19938023062166663, "grad_norm": 0.3605864942073822, "learning_rate": 3.322091062394604e-05, "loss": 0.147, "step": 4729 }, { "epoch": 0.19942239180386617, "grad_norm": 0.5241453647613525, "learning_rate": 3.322793704328275e-05, "loss": 0.164, "step": 4730 }, { "epoch": 0.19946455298606572, "grad_norm": 0.691959798336029, "learning_rate": 3.323496346261945e-05, "loss": 0.1675, "step": 4731 }, { "epoch": 0.19950671416826526, "grad_norm": 0.8255854249000549, "learning_rate": 3.324198988195616e-05, "loss": 0.24, "step": 4732 }, { "epoch": 0.19954887535046484, "grad_norm": 0.43590420484542847, "learning_rate": 3.324901630129286e-05, "loss": 0.1254, "step": 4733 }, { "epoch": 0.19959103653266438, "grad_norm": 0.8679398894309998, "learning_rate": 3.325604272062957e-05, "loss": 0.6649, "step": 4734 }, { "epoch": 0.19963319771486393, "grad_norm": 0.5272557735443115, "learning_rate": 3.326306913996628e-05, "loss": 0.4518, "step": 4735 }, { "epoch": 0.19967535889706348, "grad_norm": 0.876850426197052, "learning_rate": 3.327009555930298e-05, "loss": 0.6688, "step": 4736 }, { "epoch": 0.19971752007926302, "grad_norm": 0.6881556510925293, "learning_rate": 3.3277121978639686e-05, "loss": 0.4642, "step": 4737 }, { "epoch": 0.19975968126146257, "grad_norm": 1.577385663986206, "learning_rate": 3.32841483979764e-05, "loss": 0.5093, "step": 4738 }, { "epoch": 0.19980184244366211, "grad_norm": 0.8414198756217957, "learning_rate": 3.3291174817313095e-05, "loss": 0.7573, "step": 4739 }, { "epoch": 0.19984400362586166, "grad_norm": 0.6715496182441711, "learning_rate": 3.3298201236649806e-05, "loss": 0.5746, "step": 4740 }, { "epoch": 0.1998861648080612, "grad_norm": 0.7092502117156982, "learning_rate": 3.330522765598651e-05, "loss": 0.6785, "step": 4741 }, { "epoch": 0.19992832599026078, "grad_norm": 0.6074931025505066, "learning_rate": 3.3312254075323215e-05, "loss": 0.5537, "step": 4742 }, { "epoch": 0.19997048717246033, "grad_norm": 0.6830938458442688, "learning_rate": 3.331928049465992e-05, "loss": 0.3255, "step": 4743 }, { "epoch": 0.20001264835465987, "grad_norm": 0.9004062414169312, "learning_rate": 3.332630691399663e-05, "loss": 0.2114, "step": 4744 }, { "epoch": 0.20005480953685942, "grad_norm": 0.7893109321594238, "learning_rate": 3.3333333333333335e-05, "loss": 0.3884, "step": 4745 }, { "epoch": 0.20009697071905896, "grad_norm": 0.9105848073959351, "learning_rate": 3.334035975267004e-05, "loss": 1.2622, "step": 4746 }, { "epoch": 0.2001391319012585, "grad_norm": 1.1486996412277222, "learning_rate": 3.3347386172006744e-05, "loss": 0.2386, "step": 4747 }, { "epoch": 0.20018129308345806, "grad_norm": 0.4477214813232422, "learning_rate": 3.3354412591343455e-05, "loss": 0.1922, "step": 4748 }, { "epoch": 0.2002234542656576, "grad_norm": 0.7368886470794678, "learning_rate": 3.336143901068016e-05, "loss": 0.5644, "step": 4749 }, { "epoch": 0.20026561544785715, "grad_norm": 0.7137476205825806, "learning_rate": 3.3368465430016864e-05, "loss": 0.1982, "step": 4750 }, { "epoch": 0.2003077766300567, "grad_norm": 0.9280608892440796, "learning_rate": 3.337549184935357e-05, "loss": 0.262, "step": 4751 }, { "epoch": 0.20034993781225627, "grad_norm": 0.6526362299919128, "learning_rate": 3.338251826869028e-05, "loss": 0.1819, "step": 4752 }, { "epoch": 0.20039209899445581, "grad_norm": 0.41079357266426086, "learning_rate": 3.3389544688026984e-05, "loss": 0.1664, "step": 4753 }, { "epoch": 0.20043426017665536, "grad_norm": 0.5524877905845642, "learning_rate": 3.339657110736369e-05, "loss": 0.189, "step": 4754 }, { "epoch": 0.2004764213588549, "grad_norm": 0.5294029116630554, "learning_rate": 3.34035975267004e-05, "loss": 0.3759, "step": 4755 }, { "epoch": 0.20051858254105445, "grad_norm": 0.858825147151947, "learning_rate": 3.3410623946037104e-05, "loss": 0.4997, "step": 4756 }, { "epoch": 0.200560743723254, "grad_norm": 1.3428796529769897, "learning_rate": 3.341765036537381e-05, "loss": 0.4729, "step": 4757 }, { "epoch": 0.20060290490545354, "grad_norm": 0.5662616491317749, "learning_rate": 3.342467678471051e-05, "loss": 0.1569, "step": 4758 }, { "epoch": 0.2006450660876531, "grad_norm": 1.4118826389312744, "learning_rate": 3.3431703204047224e-05, "loss": 0.9151, "step": 4759 }, { "epoch": 0.20068722726985264, "grad_norm": 0.8984037041664124, "learning_rate": 3.343872962338392e-05, "loss": 1.1312, "step": 4760 }, { "epoch": 0.20072938845205218, "grad_norm": 0.5596324801445007, "learning_rate": 3.344575604272063e-05, "loss": 0.2745, "step": 4761 }, { "epoch": 0.20077154963425176, "grad_norm": 1.0857855081558228, "learning_rate": 3.345278246205734e-05, "loss": 1.1593, "step": 4762 }, { "epoch": 0.2008137108164513, "grad_norm": 1.8272799253463745, "learning_rate": 3.345980888139405e-05, "loss": 0.4641, "step": 4763 }, { "epoch": 0.20085587199865085, "grad_norm": 0.8309710025787354, "learning_rate": 3.3466835300730746e-05, "loss": 0.4948, "step": 4764 }, { "epoch": 0.2008980331808504, "grad_norm": 0.8132370710372925, "learning_rate": 3.347386172006746e-05, "loss": 0.5644, "step": 4765 }, { "epoch": 0.20094019436304994, "grad_norm": 1.4177706241607666, "learning_rate": 3.348088813940416e-05, "loss": 0.3285, "step": 4766 }, { "epoch": 0.20098235554524949, "grad_norm": 0.6777905821800232, "learning_rate": 3.3487914558740866e-05, "loss": 0.2252, "step": 4767 }, { "epoch": 0.20102451672744903, "grad_norm": 0.5485192537307739, "learning_rate": 3.349494097807757e-05, "loss": 0.2197, "step": 4768 }, { "epoch": 0.20106667790964858, "grad_norm": 0.574141800403595, "learning_rate": 3.350196739741428e-05, "loss": 0.1973, "step": 4769 }, { "epoch": 0.20110883909184812, "grad_norm": 0.822396457195282, "learning_rate": 3.3508993816750985e-05, "loss": 0.5705, "step": 4770 }, { "epoch": 0.2011510002740477, "grad_norm": 0.5343501567840576, "learning_rate": 3.351602023608769e-05, "loss": 0.0767, "step": 4771 }, { "epoch": 0.20119316145624724, "grad_norm": 0.4589487612247467, "learning_rate": 3.3523046655424394e-05, "loss": 0.2294, "step": 4772 }, { "epoch": 0.2012353226384468, "grad_norm": 1.141484022140503, "learning_rate": 3.3530073074761105e-05, "loss": 1.1996, "step": 4773 }, { "epoch": 0.20127748382064634, "grad_norm": 1.6038373708724976, "learning_rate": 3.353709949409781e-05, "loss": 1.3027, "step": 4774 }, { "epoch": 0.20131964500284588, "grad_norm": 0.6614691019058228, "learning_rate": 3.3544125913434514e-05, "loss": 0.7211, "step": 4775 }, { "epoch": 0.20136180618504543, "grad_norm": 0.4853542447090149, "learning_rate": 3.3551152332771225e-05, "loss": 0.1291, "step": 4776 }, { "epoch": 0.20140396736724497, "grad_norm": 0.7577099204063416, "learning_rate": 3.355817875210793e-05, "loss": 0.4815, "step": 4777 }, { "epoch": 0.20144612854944452, "grad_norm": 0.7336820363998413, "learning_rate": 3.3565205171444634e-05, "loss": 0.6091, "step": 4778 }, { "epoch": 0.20148828973164407, "grad_norm": 1.9859989881515503, "learning_rate": 3.357223159078134e-05, "loss": 0.8127, "step": 4779 }, { "epoch": 0.2015304509138436, "grad_norm": 0.8700710535049438, "learning_rate": 3.357925801011805e-05, "loss": 0.43, "step": 4780 }, { "epoch": 0.20157261209604319, "grad_norm": 0.5032647848129272, "learning_rate": 3.358628442945475e-05, "loss": 0.081, "step": 4781 }, { "epoch": 0.20161477327824273, "grad_norm": 0.5590001940727234, "learning_rate": 3.359331084879146e-05, "loss": 0.2, "step": 4782 }, { "epoch": 0.20165693446044228, "grad_norm": 0.8111636638641357, "learning_rate": 3.360033726812816e-05, "loss": 0.3508, "step": 4783 }, { "epoch": 0.20169909564264182, "grad_norm": 0.6394756436347961, "learning_rate": 3.3607363687464874e-05, "loss": 0.1296, "step": 4784 }, { "epoch": 0.20174125682484137, "grad_norm": 0.6638363599777222, "learning_rate": 3.361439010680157e-05, "loss": 0.1937, "step": 4785 }, { "epoch": 0.20178341800704092, "grad_norm": 0.8216410279273987, "learning_rate": 3.362141652613828e-05, "loss": 0.4077, "step": 4786 }, { "epoch": 0.20182557918924046, "grad_norm": 0.9762406945228577, "learning_rate": 3.362844294547499e-05, "loss": 0.5708, "step": 4787 }, { "epoch": 0.20186774037144, "grad_norm": 0.6007128953933716, "learning_rate": 3.363546936481169e-05, "loss": 0.2723, "step": 4788 }, { "epoch": 0.20190990155363955, "grad_norm": 0.8382511138916016, "learning_rate": 3.3642495784148396e-05, "loss": 0.7623, "step": 4789 }, { "epoch": 0.2019520627358391, "grad_norm": 0.5348063707351685, "learning_rate": 3.364952220348511e-05, "loss": 0.1409, "step": 4790 }, { "epoch": 0.20199422391803867, "grad_norm": 0.515358567237854, "learning_rate": 3.365654862282181e-05, "loss": 0.2614, "step": 4791 }, { "epoch": 0.20203638510023822, "grad_norm": 0.7966358661651611, "learning_rate": 3.3663575042158516e-05, "loss": 0.6818, "step": 4792 }, { "epoch": 0.20207854628243777, "grad_norm": 0.5465405583381653, "learning_rate": 3.367060146149522e-05, "loss": 0.3373, "step": 4793 }, { "epoch": 0.2021207074646373, "grad_norm": 0.8746983408927917, "learning_rate": 3.367762788083193e-05, "loss": 0.0972, "step": 4794 }, { "epoch": 0.20216286864683686, "grad_norm": 0.7267801761627197, "learning_rate": 3.3684654300168636e-05, "loss": 0.4134, "step": 4795 }, { "epoch": 0.2022050298290364, "grad_norm": 0.7022308111190796, "learning_rate": 3.369168071950534e-05, "loss": 0.297, "step": 4796 }, { "epoch": 0.20224719101123595, "grad_norm": 0.6809776425361633, "learning_rate": 3.3698707138842045e-05, "loss": 0.6842, "step": 4797 }, { "epoch": 0.2022893521934355, "grad_norm": 0.6457954049110413, "learning_rate": 3.3705733558178756e-05, "loss": 0.2991, "step": 4798 }, { "epoch": 0.20233151337563504, "grad_norm": 0.49482595920562744, "learning_rate": 3.371275997751546e-05, "loss": 0.1373, "step": 4799 }, { "epoch": 0.20237367455783462, "grad_norm": 0.4275059401988983, "learning_rate": 3.3719786396852165e-05, "loss": 0.1878, "step": 4800 }, { "epoch": 0.20241583574003416, "grad_norm": 0.5254528522491455, "learning_rate": 3.3726812816188876e-05, "loss": 0.4382, "step": 4801 }, { "epoch": 0.2024579969222337, "grad_norm": 0.4484771192073822, "learning_rate": 3.373383923552558e-05, "loss": 0.0847, "step": 4802 }, { "epoch": 0.20250015810443325, "grad_norm": 0.6699575781822205, "learning_rate": 3.3740865654862285e-05, "loss": 0.2077, "step": 4803 }, { "epoch": 0.2025423192866328, "grad_norm": 0.44688400626182556, "learning_rate": 3.374789207419899e-05, "loss": 0.1047, "step": 4804 }, { "epoch": 0.20258448046883235, "grad_norm": 1.551352858543396, "learning_rate": 3.37549184935357e-05, "loss": 0.7032, "step": 4805 }, { "epoch": 0.2026266416510319, "grad_norm": 0.696964681148529, "learning_rate": 3.37619449128724e-05, "loss": 0.5273, "step": 4806 }, { "epoch": 0.20266880283323144, "grad_norm": 0.6522461771965027, "learning_rate": 3.376897133220911e-05, "loss": 0.4526, "step": 4807 }, { "epoch": 0.20271096401543098, "grad_norm": 0.49101322889328003, "learning_rate": 3.3775997751545813e-05, "loss": 0.1599, "step": 4808 }, { "epoch": 0.20275312519763053, "grad_norm": 0.6484551429748535, "learning_rate": 3.378302417088252e-05, "loss": 0.7768, "step": 4809 }, { "epoch": 0.2027952863798301, "grad_norm": 0.6089683771133423, "learning_rate": 3.379005059021922e-05, "loss": 0.2821, "step": 4810 }, { "epoch": 0.20283744756202965, "grad_norm": 0.9306252598762512, "learning_rate": 3.3797077009555933e-05, "loss": 0.2695, "step": 4811 }, { "epoch": 0.2028796087442292, "grad_norm": 0.8401558995246887, "learning_rate": 3.380410342889264e-05, "loss": 0.2465, "step": 4812 }, { "epoch": 0.20292176992642874, "grad_norm": 0.6653099060058594, "learning_rate": 3.381112984822934e-05, "loss": 0.2908, "step": 4813 }, { "epoch": 0.2029639311086283, "grad_norm": 0.3360315263271332, "learning_rate": 3.381815626756605e-05, "loss": 0.0958, "step": 4814 }, { "epoch": 0.20300609229082783, "grad_norm": 0.9428203701972961, "learning_rate": 3.382518268690276e-05, "loss": 0.4061, "step": 4815 }, { "epoch": 0.20304825347302738, "grad_norm": 1.149843454360962, "learning_rate": 3.383220910623946e-05, "loss": 0.5101, "step": 4816 }, { "epoch": 0.20309041465522693, "grad_norm": 0.49931058287620544, "learning_rate": 3.3839235525576167e-05, "loss": 0.4126, "step": 4817 }, { "epoch": 0.20313257583742647, "grad_norm": 0.7840538024902344, "learning_rate": 3.384626194491287e-05, "loss": 0.6244, "step": 4818 }, { "epoch": 0.20317473701962602, "grad_norm": 0.8607788681983948, "learning_rate": 3.385328836424958e-05, "loss": 0.6487, "step": 4819 }, { "epoch": 0.2032168982018256, "grad_norm": 0.7974826693534851, "learning_rate": 3.3860314783586287e-05, "loss": 0.8338, "step": 4820 }, { "epoch": 0.20325905938402514, "grad_norm": 9.01984977722168, "learning_rate": 3.386734120292299e-05, "loss": 2.4241, "step": 4821 }, { "epoch": 0.20330122056622468, "grad_norm": 1.1508663892745972, "learning_rate": 3.38743676222597e-05, "loss": 1.0157, "step": 4822 }, { "epoch": 0.20334338174842423, "grad_norm": 0.43623560667037964, "learning_rate": 3.3881394041596407e-05, "loss": 0.1332, "step": 4823 }, { "epoch": 0.20338554293062377, "grad_norm": 0.9230874180793762, "learning_rate": 3.388842046093311e-05, "loss": 0.3035, "step": 4824 }, { "epoch": 0.20342770411282332, "grad_norm": 1.563681960105896, "learning_rate": 3.3895446880269815e-05, "loss": 1.055, "step": 4825 }, { "epoch": 0.20346986529502287, "grad_norm": 0.9757941365242004, "learning_rate": 3.3902473299606526e-05, "loss": 0.3131, "step": 4826 }, { "epoch": 0.2035120264772224, "grad_norm": 0.8615579605102539, "learning_rate": 3.3909499718943224e-05, "loss": 0.2056, "step": 4827 }, { "epoch": 0.20355418765942196, "grad_norm": 0.4239373505115509, "learning_rate": 3.3916526138279935e-05, "loss": 0.109, "step": 4828 }, { "epoch": 0.20359634884162153, "grad_norm": 0.5347983837127686, "learning_rate": 3.392355255761664e-05, "loss": 0.2223, "step": 4829 }, { "epoch": 0.20363851002382108, "grad_norm": 0.7498281598091125, "learning_rate": 3.393057897695335e-05, "loss": 0.2535, "step": 4830 }, { "epoch": 0.20368067120602062, "grad_norm": 0.6326190829277039, "learning_rate": 3.393760539629005e-05, "loss": 0.1085, "step": 4831 }, { "epoch": 0.20372283238822017, "grad_norm": 0.4371468722820282, "learning_rate": 3.394463181562676e-05, "loss": 0.1639, "step": 4832 }, { "epoch": 0.20376499357041972, "grad_norm": 0.9058423638343811, "learning_rate": 3.3951658234963464e-05, "loss": 0.2674, "step": 4833 }, { "epoch": 0.20380715475261926, "grad_norm": 0.7669930458068848, "learning_rate": 3.395868465430017e-05, "loss": 0.2238, "step": 4834 }, { "epoch": 0.2038493159348188, "grad_norm": 0.8134977221488953, "learning_rate": 3.396571107363687e-05, "loss": 0.2845, "step": 4835 }, { "epoch": 0.20389147711701835, "grad_norm": 0.6470930576324463, "learning_rate": 3.3972737492973584e-05, "loss": 0.1765, "step": 4836 }, { "epoch": 0.2039336382992179, "grad_norm": 0.46346884965896606, "learning_rate": 3.397976391231029e-05, "loss": 0.1517, "step": 4837 }, { "epoch": 0.20397579948141745, "grad_norm": 0.5205867886543274, "learning_rate": 3.398679033164699e-05, "loss": 0.2763, "step": 4838 }, { "epoch": 0.20401796066361702, "grad_norm": 0.6410700082778931, "learning_rate": 3.39938167509837e-05, "loss": 0.1635, "step": 4839 }, { "epoch": 0.20406012184581657, "grad_norm": 0.4975684583187103, "learning_rate": 3.400084317032041e-05, "loss": 0.1552, "step": 4840 }, { "epoch": 0.2041022830280161, "grad_norm": 2.250950813293457, "learning_rate": 3.400786958965711e-05, "loss": 0.5727, "step": 4841 }, { "epoch": 0.20414444421021566, "grad_norm": 0.558815598487854, "learning_rate": 3.401489600899382e-05, "loss": 0.2646, "step": 4842 }, { "epoch": 0.2041866053924152, "grad_norm": 0.6331632137298584, "learning_rate": 3.402192242833053e-05, "loss": 0.3017, "step": 4843 }, { "epoch": 0.20422876657461475, "grad_norm": 0.6927050948143005, "learning_rate": 3.402894884766723e-05, "loss": 0.3808, "step": 4844 }, { "epoch": 0.2042709277568143, "grad_norm": 0.7654138207435608, "learning_rate": 3.403597526700394e-05, "loss": 0.2389, "step": 4845 }, { "epoch": 0.20431308893901384, "grad_norm": 0.5742013454437256, "learning_rate": 3.404300168634064e-05, "loss": 0.5467, "step": 4846 }, { "epoch": 0.2043552501212134, "grad_norm": 0.4270583391189575, "learning_rate": 3.405002810567735e-05, "loss": 0.1936, "step": 4847 }, { "epoch": 0.20439741130341293, "grad_norm": 0.8115733861923218, "learning_rate": 3.405705452501406e-05, "loss": 0.6332, "step": 4848 }, { "epoch": 0.2044395724856125, "grad_norm": 0.7678930759429932, "learning_rate": 3.406408094435076e-05, "loss": 0.2688, "step": 4849 }, { "epoch": 0.20448173366781205, "grad_norm": 0.5094708204269409, "learning_rate": 3.4071107363687466e-05, "loss": 0.2446, "step": 4850 }, { "epoch": 0.2045238948500116, "grad_norm": 0.4785088300704956, "learning_rate": 3.407813378302418e-05, "loss": 0.1232, "step": 4851 }, { "epoch": 0.20456605603221115, "grad_norm": 0.7555882930755615, "learning_rate": 3.4085160202360875e-05, "loss": 0.3342, "step": 4852 }, { "epoch": 0.2046082172144107, "grad_norm": 0.4974837899208069, "learning_rate": 3.4092186621697586e-05, "loss": 0.2941, "step": 4853 }, { "epoch": 0.20465037839661024, "grad_norm": 0.7575719356536865, "learning_rate": 3.409921304103429e-05, "loss": 0.2841, "step": 4854 }, { "epoch": 0.20469253957880978, "grad_norm": 0.5590739846229553, "learning_rate": 3.4106239460370995e-05, "loss": 0.5897, "step": 4855 }, { "epoch": 0.20473470076100933, "grad_norm": 0.46005338430404663, "learning_rate": 3.41132658797077e-05, "loss": 0.1562, "step": 4856 }, { "epoch": 0.20477686194320888, "grad_norm": 1.5776292085647583, "learning_rate": 3.412029229904441e-05, "loss": 0.701, "step": 4857 }, { "epoch": 0.20481902312540842, "grad_norm": 0.6399824619293213, "learning_rate": 3.4127318718381115e-05, "loss": 0.2018, "step": 4858 }, { "epoch": 0.204861184307608, "grad_norm": 0.46672356128692627, "learning_rate": 3.413434513771782e-05, "loss": 0.1518, "step": 4859 }, { "epoch": 0.20490334548980754, "grad_norm": 1.6371666193008423, "learning_rate": 3.414137155705452e-05, "loss": 0.9756, "step": 4860 }, { "epoch": 0.2049455066720071, "grad_norm": 0.40706682205200195, "learning_rate": 3.4148397976391235e-05, "loss": 0.1759, "step": 4861 }, { "epoch": 0.20498766785420663, "grad_norm": 0.560975193977356, "learning_rate": 3.415542439572794e-05, "loss": 0.2228, "step": 4862 }, { "epoch": 0.20502982903640618, "grad_norm": 0.5450981259346008, "learning_rate": 3.416245081506464e-05, "loss": 0.2914, "step": 4863 }, { "epoch": 0.20507199021860573, "grad_norm": 0.6642386317253113, "learning_rate": 3.416947723440135e-05, "loss": 0.1833, "step": 4864 }, { "epoch": 0.20511415140080527, "grad_norm": 0.5424088835716248, "learning_rate": 3.417650365373806e-05, "loss": 0.461, "step": 4865 }, { "epoch": 0.20515631258300482, "grad_norm": 0.7963706851005554, "learning_rate": 3.418353007307476e-05, "loss": 0.344, "step": 4866 }, { "epoch": 0.20519847376520436, "grad_norm": 0.6085197329521179, "learning_rate": 3.419055649241147e-05, "loss": 0.1648, "step": 4867 }, { "epoch": 0.20524063494740394, "grad_norm": 0.6285789608955383, "learning_rate": 3.419758291174818e-05, "loss": 0.4054, "step": 4868 }, { "epoch": 0.20528279612960348, "grad_norm": 0.9873071908950806, "learning_rate": 3.420460933108488e-05, "loss": 0.9529, "step": 4869 }, { "epoch": 0.20532495731180303, "grad_norm": 0.4626077711582184, "learning_rate": 3.421163575042159e-05, "loss": 0.1683, "step": 4870 }, { "epoch": 0.20536711849400258, "grad_norm": 1.448872447013855, "learning_rate": 3.421866216975829e-05, "loss": 1.153, "step": 4871 }, { "epoch": 0.20540927967620212, "grad_norm": 0.4959252178668976, "learning_rate": 3.4225688589095e-05, "loss": 0.143, "step": 4872 }, { "epoch": 0.20545144085840167, "grad_norm": 0.6179549098014832, "learning_rate": 3.42327150084317e-05, "loss": 0.6137, "step": 4873 }, { "epoch": 0.20549360204060121, "grad_norm": 1.0853939056396484, "learning_rate": 3.423974142776841e-05, "loss": 1.2655, "step": 4874 }, { "epoch": 0.20553576322280076, "grad_norm": 0.6658042073249817, "learning_rate": 3.4246767847105116e-05, "loss": 0.3132, "step": 4875 }, { "epoch": 0.2055779244050003, "grad_norm": 0.5903781056404114, "learning_rate": 3.425379426644183e-05, "loss": 0.331, "step": 4876 }, { "epoch": 0.20562008558719985, "grad_norm": 0.4450789988040924, "learning_rate": 3.4260820685778525e-05, "loss": 0.2262, "step": 4877 }, { "epoch": 0.20566224676939943, "grad_norm": 0.5642912983894348, "learning_rate": 3.4267847105115236e-05, "loss": 0.3289, "step": 4878 }, { "epoch": 0.20570440795159897, "grad_norm": 0.9292883276939392, "learning_rate": 3.427487352445194e-05, "loss": 0.5726, "step": 4879 }, { "epoch": 0.20574656913379852, "grad_norm": 0.5074453353881836, "learning_rate": 3.4281899943788645e-05, "loss": 0.2391, "step": 4880 }, { "epoch": 0.20578873031599806, "grad_norm": 0.46949470043182373, "learning_rate": 3.428892636312535e-05, "loss": 0.4669, "step": 4881 }, { "epoch": 0.2058308914981976, "grad_norm": 0.5658913254737854, "learning_rate": 3.429595278246206e-05, "loss": 0.5503, "step": 4882 }, { "epoch": 0.20587305268039716, "grad_norm": 0.8613523840904236, "learning_rate": 3.4302979201798765e-05, "loss": 0.3867, "step": 4883 }, { "epoch": 0.2059152138625967, "grad_norm": 1.0312769412994385, "learning_rate": 3.431000562113547e-05, "loss": 1.2508, "step": 4884 }, { "epoch": 0.20595737504479625, "grad_norm": 0.3369012176990509, "learning_rate": 3.4317032040472174e-05, "loss": 0.1092, "step": 4885 }, { "epoch": 0.2059995362269958, "grad_norm": 0.8348138928413391, "learning_rate": 3.4324058459808885e-05, "loss": 0.6209, "step": 4886 }, { "epoch": 0.20604169740919534, "grad_norm": 0.6892932653427124, "learning_rate": 3.433108487914559e-05, "loss": 0.3578, "step": 4887 }, { "epoch": 0.20608385859139491, "grad_norm": 1.1572295427322388, "learning_rate": 3.4338111298482294e-05, "loss": 1.3256, "step": 4888 }, { "epoch": 0.20612601977359446, "grad_norm": 0.5604777336120605, "learning_rate": 3.4345137717819005e-05, "loss": 0.1715, "step": 4889 }, { "epoch": 0.206168180955794, "grad_norm": 0.6716474890708923, "learning_rate": 3.435216413715571e-05, "loss": 0.7046, "step": 4890 }, { "epoch": 0.20621034213799355, "grad_norm": 0.5828546285629272, "learning_rate": 3.4359190556492414e-05, "loss": 0.2129, "step": 4891 }, { "epoch": 0.2062525033201931, "grad_norm": 0.4849225878715515, "learning_rate": 3.436621697582912e-05, "loss": 0.2474, "step": 4892 }, { "epoch": 0.20629466450239264, "grad_norm": 0.7493168711662292, "learning_rate": 3.437324339516583e-05, "loss": 0.4205, "step": 4893 }, { "epoch": 0.2063368256845922, "grad_norm": 1.324270248413086, "learning_rate": 3.4380269814502534e-05, "loss": 1.2229, "step": 4894 }, { "epoch": 0.20637898686679174, "grad_norm": 1.385018229484558, "learning_rate": 3.438729623383924e-05, "loss": 0.9088, "step": 4895 }, { "epoch": 0.20642114804899128, "grad_norm": 0.6932553648948669, "learning_rate": 3.439432265317594e-05, "loss": 0.5445, "step": 4896 }, { "epoch": 0.20646330923119086, "grad_norm": 0.4510035216808319, "learning_rate": 3.4401349072512654e-05, "loss": 0.1364, "step": 4897 }, { "epoch": 0.2065054704133904, "grad_norm": 0.7557103037834167, "learning_rate": 3.440837549184935e-05, "loss": 0.4573, "step": 4898 }, { "epoch": 0.20654763159558995, "grad_norm": 0.8820162415504456, "learning_rate": 3.441540191118606e-05, "loss": 0.2036, "step": 4899 }, { "epoch": 0.2065897927777895, "grad_norm": 0.43025052547454834, "learning_rate": 3.442242833052277e-05, "loss": 0.2361, "step": 4900 }, { "epoch": 0.20663195395998904, "grad_norm": 0.34428951144218445, "learning_rate": 3.442945474985947e-05, "loss": 0.1252, "step": 4901 }, { "epoch": 0.20667411514218859, "grad_norm": 0.3637509346008301, "learning_rate": 3.4436481169196176e-05, "loss": 0.1264, "step": 4902 }, { "epoch": 0.20671627632438813, "grad_norm": 0.3970806300640106, "learning_rate": 3.444350758853289e-05, "loss": 0.1207, "step": 4903 }, { "epoch": 0.20675843750658768, "grad_norm": 1.0258936882019043, "learning_rate": 3.445053400786959e-05, "loss": 0.4408, "step": 4904 }, { "epoch": 0.20680059868878722, "grad_norm": 0.8902714252471924, "learning_rate": 3.4457560427206296e-05, "loss": 0.6942, "step": 4905 }, { "epoch": 0.20684275987098677, "grad_norm": 0.5731600522994995, "learning_rate": 3.4464586846543e-05, "loss": 0.1909, "step": 4906 }, { "epoch": 0.20688492105318634, "grad_norm": 0.6650870442390442, "learning_rate": 3.447161326587971e-05, "loss": 0.2138, "step": 4907 }, { "epoch": 0.2069270822353859, "grad_norm": 0.7581061124801636, "learning_rate": 3.4478639685216416e-05, "loss": 0.1502, "step": 4908 }, { "epoch": 0.20696924341758544, "grad_norm": 0.2707098126411438, "learning_rate": 3.448566610455312e-05, "loss": 0.1523, "step": 4909 }, { "epoch": 0.20701140459978498, "grad_norm": 0.6936352252960205, "learning_rate": 3.4492692523889824e-05, "loss": 0.3258, "step": 4910 }, { "epoch": 0.20705356578198453, "grad_norm": 0.5430520176887512, "learning_rate": 3.4499718943226536e-05, "loss": 0.2649, "step": 4911 }, { "epoch": 0.20709572696418407, "grad_norm": 0.6026330590248108, "learning_rate": 3.450674536256324e-05, "loss": 0.4632, "step": 4912 }, { "epoch": 0.20713788814638362, "grad_norm": 0.7385250926017761, "learning_rate": 3.4513771781899944e-05, "loss": 0.1914, "step": 4913 }, { "epoch": 0.20718004932858317, "grad_norm": 0.6854294538497925, "learning_rate": 3.4520798201236656e-05, "loss": 0.3262, "step": 4914 }, { "epoch": 0.2072222105107827, "grad_norm": 0.8175390958786011, "learning_rate": 3.452782462057336e-05, "loss": 0.2886, "step": 4915 }, { "epoch": 0.20726437169298226, "grad_norm": 1.0523823499679565, "learning_rate": 3.4534851039910064e-05, "loss": 0.3891, "step": 4916 }, { "epoch": 0.20730653287518183, "grad_norm": 0.496997594833374, "learning_rate": 3.454187745924677e-05, "loss": 0.1461, "step": 4917 }, { "epoch": 0.20734869405738138, "grad_norm": 1.0734307765960693, "learning_rate": 3.454890387858348e-05, "loss": 1.1193, "step": 4918 }, { "epoch": 0.20739085523958092, "grad_norm": 1.1792224645614624, "learning_rate": 3.455593029792018e-05, "loss": 0.9968, "step": 4919 }, { "epoch": 0.20743301642178047, "grad_norm": 0.7908415794372559, "learning_rate": 3.456295671725689e-05, "loss": 0.3499, "step": 4920 }, { "epoch": 0.20747517760398002, "grad_norm": 0.7145437002182007, "learning_rate": 3.456998313659359e-05, "loss": 0.4728, "step": 4921 }, { "epoch": 0.20751733878617956, "grad_norm": 0.6241229772567749, "learning_rate": 3.4577009555930304e-05, "loss": 0.158, "step": 4922 }, { "epoch": 0.2075594999683791, "grad_norm": 0.5533643364906311, "learning_rate": 3.4584035975267e-05, "loss": 0.5952, "step": 4923 }, { "epoch": 0.20760166115057865, "grad_norm": 1.1006239652633667, "learning_rate": 3.459106239460371e-05, "loss": 1.2618, "step": 4924 }, { "epoch": 0.2076438223327782, "grad_norm": 0.7220472693443298, "learning_rate": 3.459808881394042e-05, "loss": 0.4832, "step": 4925 }, { "epoch": 0.20768598351497777, "grad_norm": 0.37808674573898315, "learning_rate": 3.460511523327712e-05, "loss": 0.1644, "step": 4926 }, { "epoch": 0.20772814469717732, "grad_norm": 0.3368806540966034, "learning_rate": 3.4612141652613826e-05, "loss": 0.1232, "step": 4927 }, { "epoch": 0.20777030587937687, "grad_norm": 0.6481050252914429, "learning_rate": 3.461916807195054e-05, "loss": 0.1736, "step": 4928 }, { "epoch": 0.2078124670615764, "grad_norm": 0.7360457181930542, "learning_rate": 3.462619449128724e-05, "loss": 0.5036, "step": 4929 }, { "epoch": 0.20785462824377596, "grad_norm": 0.5255635380744934, "learning_rate": 3.4633220910623946e-05, "loss": 0.1899, "step": 4930 }, { "epoch": 0.2078967894259755, "grad_norm": 0.9896652102470398, "learning_rate": 3.464024732996065e-05, "loss": 0.2474, "step": 4931 }, { "epoch": 0.20793895060817505, "grad_norm": 0.426979660987854, "learning_rate": 3.464727374929736e-05, "loss": 0.1792, "step": 4932 }, { "epoch": 0.2079811117903746, "grad_norm": 0.5033552646636963, "learning_rate": 3.4654300168634066e-05, "loss": 0.1479, "step": 4933 }, { "epoch": 0.20802327297257414, "grad_norm": 0.6692194938659668, "learning_rate": 3.466132658797077e-05, "loss": 0.2732, "step": 4934 }, { "epoch": 0.2080654341547737, "grad_norm": 0.7177478075027466, "learning_rate": 3.466835300730748e-05, "loss": 0.858, "step": 4935 }, { "epoch": 0.20810759533697326, "grad_norm": 0.6094041466712952, "learning_rate": 3.4675379426644186e-05, "loss": 0.1518, "step": 4936 }, { "epoch": 0.2081497565191728, "grad_norm": 1.3481334447860718, "learning_rate": 3.468240584598089e-05, "loss": 0.3799, "step": 4937 }, { "epoch": 0.20819191770137235, "grad_norm": 1.904909372329712, "learning_rate": 3.4689432265317595e-05, "loss": 0.9844, "step": 4938 }, { "epoch": 0.2082340788835719, "grad_norm": 0.5051758885383606, "learning_rate": 3.4696458684654306e-05, "loss": 0.3931, "step": 4939 }, { "epoch": 0.20827624006577145, "grad_norm": 0.7002243995666504, "learning_rate": 3.4703485103991004e-05, "loss": 0.1219, "step": 4940 }, { "epoch": 0.208318401247971, "grad_norm": 0.7373753190040588, "learning_rate": 3.4710511523327715e-05, "loss": 0.1823, "step": 4941 }, { "epoch": 0.20836056243017054, "grad_norm": 1.357643723487854, "learning_rate": 3.471753794266442e-05, "loss": 0.4589, "step": 4942 }, { "epoch": 0.20840272361237008, "grad_norm": 0.7273041009902954, "learning_rate": 3.472456436200113e-05, "loss": 0.2328, "step": 4943 }, { "epoch": 0.20844488479456963, "grad_norm": 0.4844585359096527, "learning_rate": 3.473159078133783e-05, "loss": 0.0848, "step": 4944 }, { "epoch": 0.20848704597676918, "grad_norm": 0.44161391258239746, "learning_rate": 3.473861720067454e-05, "loss": 0.1044, "step": 4945 }, { "epoch": 0.20852920715896875, "grad_norm": 0.703851580619812, "learning_rate": 3.4745643620011244e-05, "loss": 0.6022, "step": 4946 }, { "epoch": 0.2085713683411683, "grad_norm": 2.5362493991851807, "learning_rate": 3.475267003934795e-05, "loss": 0.3866, "step": 4947 }, { "epoch": 0.20861352952336784, "grad_norm": 1.7527294158935547, "learning_rate": 3.475969645868465e-05, "loss": 0.4292, "step": 4948 }, { "epoch": 0.2086556907055674, "grad_norm": 0.6582508683204651, "learning_rate": 3.4766722878021364e-05, "loss": 0.1684, "step": 4949 }, { "epoch": 0.20869785188776693, "grad_norm": 1.1609749794006348, "learning_rate": 3.477374929735807e-05, "loss": 1.0372, "step": 4950 }, { "epoch": 0.20874001306996648, "grad_norm": 0.5549782514572144, "learning_rate": 3.478077571669477e-05, "loss": 0.1331, "step": 4951 }, { "epoch": 0.20878217425216603, "grad_norm": 0.723400890827179, "learning_rate": 3.478780213603148e-05, "loss": 0.2356, "step": 4952 }, { "epoch": 0.20882433543436557, "grad_norm": 0.8941120505332947, "learning_rate": 3.479482855536819e-05, "loss": 0.2702, "step": 4953 }, { "epoch": 0.20886649661656512, "grad_norm": 0.4423155188560486, "learning_rate": 3.480185497470489e-05, "loss": 0.1256, "step": 4954 }, { "epoch": 0.2089086577987647, "grad_norm": 0.7947689890861511, "learning_rate": 3.48088813940416e-05, "loss": 0.267, "step": 4955 }, { "epoch": 0.20895081898096424, "grad_norm": 0.5342812538146973, "learning_rate": 3.481590781337831e-05, "loss": 0.5985, "step": 4956 }, { "epoch": 0.20899298016316378, "grad_norm": 0.7001768350601196, "learning_rate": 3.482293423271501e-05, "loss": 0.2506, "step": 4957 }, { "epoch": 0.20903514134536333, "grad_norm": 0.7231866121292114, "learning_rate": 3.482996065205172e-05, "loss": 0.3945, "step": 4958 }, { "epoch": 0.20907730252756287, "grad_norm": 1.609319806098938, "learning_rate": 3.483698707138842e-05, "loss": 0.7643, "step": 4959 }, { "epoch": 0.20911946370976242, "grad_norm": 0.6426835656166077, "learning_rate": 3.484401349072513e-05, "loss": 0.6323, "step": 4960 }, { "epoch": 0.20916162489196197, "grad_norm": 0.6830517649650574, "learning_rate": 3.485103991006184e-05, "loss": 0.2248, "step": 4961 }, { "epoch": 0.2092037860741615, "grad_norm": 0.6918619275093079, "learning_rate": 3.485806632939854e-05, "loss": 0.6814, "step": 4962 }, { "epoch": 0.20924594725636106, "grad_norm": 0.9566590189933777, "learning_rate": 3.4865092748735245e-05, "loss": 1.0793, "step": 4963 }, { "epoch": 0.2092881084385606, "grad_norm": 0.5856083631515503, "learning_rate": 3.487211916807196e-05, "loss": 0.1388, "step": 4964 }, { "epoch": 0.20933026962076018, "grad_norm": 1.8122271299362183, "learning_rate": 3.4879145587408654e-05, "loss": 0.7336, "step": 4965 }, { "epoch": 0.20937243080295972, "grad_norm": 0.494584321975708, "learning_rate": 3.4886172006745365e-05, "loss": 0.2529, "step": 4966 }, { "epoch": 0.20941459198515927, "grad_norm": 0.8395365476608276, "learning_rate": 3.489319842608207e-05, "loss": 0.2202, "step": 4967 }, { "epoch": 0.20945675316735882, "grad_norm": 0.39991995692253113, "learning_rate": 3.4900224845418774e-05, "loss": 0.1885, "step": 4968 }, { "epoch": 0.20949891434955836, "grad_norm": 0.5722638964653015, "learning_rate": 3.490725126475548e-05, "loss": 0.5537, "step": 4969 }, { "epoch": 0.2095410755317579, "grad_norm": 0.8357682228088379, "learning_rate": 3.491427768409219e-05, "loss": 0.4056, "step": 4970 }, { "epoch": 0.20958323671395745, "grad_norm": 0.9372342824935913, "learning_rate": 3.4921304103428894e-05, "loss": 0.9886, "step": 4971 }, { "epoch": 0.209625397896157, "grad_norm": 0.7516882419586182, "learning_rate": 3.49283305227656e-05, "loss": 0.1264, "step": 4972 }, { "epoch": 0.20966755907835655, "grad_norm": 0.6032217144966125, "learning_rate": 3.49353569421023e-05, "loss": 0.2012, "step": 4973 }, { "epoch": 0.2097097202605561, "grad_norm": 0.6140092611312866, "learning_rate": 3.4942383361439014e-05, "loss": 0.6753, "step": 4974 }, { "epoch": 0.20975188144275567, "grad_norm": 0.8007373809814453, "learning_rate": 3.494940978077572e-05, "loss": 0.2804, "step": 4975 }, { "epoch": 0.2097940426249552, "grad_norm": 0.6630076766014099, "learning_rate": 3.495643620011242e-05, "loss": 0.7072, "step": 4976 }, { "epoch": 0.20983620380715476, "grad_norm": 1.765562891960144, "learning_rate": 3.496346261944913e-05, "loss": 0.592, "step": 4977 }, { "epoch": 0.2098783649893543, "grad_norm": 1.1702960729599, "learning_rate": 3.497048903878584e-05, "loss": 1.1266, "step": 4978 }, { "epoch": 0.20992052617155385, "grad_norm": 0.592700719833374, "learning_rate": 3.497751545812254e-05, "loss": 0.7061, "step": 4979 }, { "epoch": 0.2099626873537534, "grad_norm": 1.0938974618911743, "learning_rate": 3.498454187745925e-05, "loss": 1.1267, "step": 4980 }, { "epoch": 0.21000484853595294, "grad_norm": 0.4964422881603241, "learning_rate": 3.499156829679596e-05, "loss": 0.3548, "step": 4981 }, { "epoch": 0.2100470097181525, "grad_norm": 0.5705270767211914, "learning_rate": 3.499859471613266e-05, "loss": 0.1146, "step": 4982 }, { "epoch": 0.21008917090035203, "grad_norm": 0.7771574854850769, "learning_rate": 3.500562113546937e-05, "loss": 0.2918, "step": 4983 }, { "epoch": 0.2101313320825516, "grad_norm": 0.6309119462966919, "learning_rate": 3.501264755480607e-05, "loss": 0.6398, "step": 4984 }, { "epoch": 0.21017349326475115, "grad_norm": 0.7879749536514282, "learning_rate": 3.501967397414278e-05, "loss": 0.4496, "step": 4985 }, { "epoch": 0.2102156544469507, "grad_norm": 0.4050169885158539, "learning_rate": 3.502670039347948e-05, "loss": 0.1558, "step": 4986 }, { "epoch": 0.21025781562915025, "grad_norm": 0.5654656887054443, "learning_rate": 3.503372681281619e-05, "loss": 0.1454, "step": 4987 }, { "epoch": 0.2102999768113498, "grad_norm": 1.2121710777282715, "learning_rate": 3.5040753232152896e-05, "loss": 0.486, "step": 4988 }, { "epoch": 0.21034213799354934, "grad_norm": 0.7162886261940002, "learning_rate": 3.504777965148961e-05, "loss": 0.2369, "step": 4989 }, { "epoch": 0.21038429917574888, "grad_norm": 0.3783651292324066, "learning_rate": 3.5054806070826305e-05, "loss": 0.1189, "step": 4990 }, { "epoch": 0.21042646035794843, "grad_norm": 0.7430817484855652, "learning_rate": 3.5061832490163016e-05, "loss": 0.585, "step": 4991 }, { "epoch": 0.21046862154014798, "grad_norm": 0.5973770618438721, "learning_rate": 3.506885890949972e-05, "loss": 0.8211, "step": 4992 }, { "epoch": 0.21051078272234752, "grad_norm": 2.1586709022521973, "learning_rate": 3.5075885328836425e-05, "loss": 0.8865, "step": 4993 }, { "epoch": 0.2105529439045471, "grad_norm": 0.5909693837165833, "learning_rate": 3.508291174817313e-05, "loss": 0.2547, "step": 4994 }, { "epoch": 0.21059510508674664, "grad_norm": 0.6534302234649658, "learning_rate": 3.508993816750984e-05, "loss": 0.8193, "step": 4995 }, { "epoch": 0.2106372662689462, "grad_norm": 0.7573290467262268, "learning_rate": 3.5096964586846545e-05, "loss": 0.412, "step": 4996 }, { "epoch": 0.21067942745114573, "grad_norm": 0.7950007319450378, "learning_rate": 3.510399100618325e-05, "loss": 0.214, "step": 4997 }, { "epoch": 0.21072158863334528, "grad_norm": 0.4176011085510254, "learning_rate": 3.5111017425519953e-05, "loss": 0.1126, "step": 4998 }, { "epoch": 0.21076374981554483, "grad_norm": 0.8385640978813171, "learning_rate": 3.5118043844856665e-05, "loss": 0.41, "step": 4999 }, { "epoch": 0.21080591099774437, "grad_norm": 0.7352833151817322, "learning_rate": 3.512507026419337e-05, "loss": 0.4819, "step": 5000 }, { "epoch": 0.21084807217994392, "grad_norm": 0.8170105814933777, "learning_rate": 3.5132096683530073e-05, "loss": 0.3154, "step": 5001 }, { "epoch": 0.21089023336214346, "grad_norm": 0.47626376152038574, "learning_rate": 3.5139123102866785e-05, "loss": 0.1533, "step": 5002 }, { "epoch": 0.210932394544343, "grad_norm": 0.8738710284233093, "learning_rate": 3.514614952220349e-05, "loss": 0.9062, "step": 5003 }, { "epoch": 0.21097455572654258, "grad_norm": 0.9326266050338745, "learning_rate": 3.5153175941540193e-05, "loss": 0.5064, "step": 5004 }, { "epoch": 0.21101671690874213, "grad_norm": 0.4567812979221344, "learning_rate": 3.51602023608769e-05, "loss": 0.2178, "step": 5005 }, { "epoch": 0.21105887809094168, "grad_norm": 0.6499655246734619, "learning_rate": 3.516722878021361e-05, "loss": 0.2705, "step": 5006 }, { "epoch": 0.21110103927314122, "grad_norm": 0.6615639328956604, "learning_rate": 3.517425519955031e-05, "loss": 0.1712, "step": 5007 }, { "epoch": 0.21114320045534077, "grad_norm": 0.7790016531944275, "learning_rate": 3.518128161888702e-05, "loss": 0.2464, "step": 5008 }, { "epoch": 0.21118536163754031, "grad_norm": 0.5602735877037048, "learning_rate": 3.518830803822372e-05, "loss": 0.4666, "step": 5009 }, { "epoch": 0.21122752281973986, "grad_norm": 0.6340038180351257, "learning_rate": 3.519533445756043e-05, "loss": 0.4925, "step": 5010 }, { "epoch": 0.2112696840019394, "grad_norm": 0.7330309152603149, "learning_rate": 3.520236087689713e-05, "loss": 0.1467, "step": 5011 }, { "epoch": 0.21131184518413895, "grad_norm": 0.7457754611968994, "learning_rate": 3.520938729623384e-05, "loss": 0.1655, "step": 5012 }, { "epoch": 0.21135400636633853, "grad_norm": 1.0076394081115723, "learning_rate": 3.5216413715570547e-05, "loss": 0.6226, "step": 5013 }, { "epoch": 0.21139616754853807, "grad_norm": 1.4200869798660278, "learning_rate": 3.522344013490725e-05, "loss": 0.3498, "step": 5014 }, { "epoch": 0.21143832873073762, "grad_norm": 0.6886611580848694, "learning_rate": 3.5230466554243955e-05, "loss": 0.1304, "step": 5015 }, { "epoch": 0.21148048991293716, "grad_norm": 0.6961111426353455, "learning_rate": 3.5237492973580666e-05, "loss": 0.5518, "step": 5016 }, { "epoch": 0.2115226510951367, "grad_norm": 0.5180030465126038, "learning_rate": 3.524451939291737e-05, "loss": 0.1629, "step": 5017 }, { "epoch": 0.21156481227733626, "grad_norm": 1.1589542627334595, "learning_rate": 3.5251545812254075e-05, "loss": 1.1555, "step": 5018 }, { "epoch": 0.2116069734595358, "grad_norm": 0.4438186287879944, "learning_rate": 3.525857223159078e-05, "loss": 0.1351, "step": 5019 }, { "epoch": 0.21164913464173535, "grad_norm": 1.0454070568084717, "learning_rate": 3.526559865092749e-05, "loss": 0.3788, "step": 5020 }, { "epoch": 0.2116912958239349, "grad_norm": 0.6441635489463806, "learning_rate": 3.5272625070264195e-05, "loss": 0.4448, "step": 5021 }, { "epoch": 0.21173345700613444, "grad_norm": 0.8593316674232483, "learning_rate": 3.52796514896009e-05, "loss": 0.462, "step": 5022 }, { "epoch": 0.211775618188334, "grad_norm": 0.799714982509613, "learning_rate": 3.528667790893761e-05, "loss": 0.2302, "step": 5023 }, { "epoch": 0.21181777937053356, "grad_norm": 0.6210994720458984, "learning_rate": 3.5293704328274315e-05, "loss": 0.6096, "step": 5024 }, { "epoch": 0.2118599405527331, "grad_norm": 1.710293173789978, "learning_rate": 3.530073074761102e-05, "loss": 1.4706, "step": 5025 }, { "epoch": 0.21190210173493265, "grad_norm": 0.5098560452461243, "learning_rate": 3.5307757166947724e-05, "loss": 0.4304, "step": 5026 }, { "epoch": 0.2119442629171322, "grad_norm": 0.5049261450767517, "learning_rate": 3.5314783586284435e-05, "loss": 0.1679, "step": 5027 }, { "epoch": 0.21198642409933174, "grad_norm": 0.5757893323898315, "learning_rate": 3.532181000562114e-05, "loss": 0.1593, "step": 5028 }, { "epoch": 0.2120285852815313, "grad_norm": 0.659594714641571, "learning_rate": 3.5328836424957844e-05, "loss": 0.2728, "step": 5029 }, { "epoch": 0.21207074646373084, "grad_norm": 0.9770070314407349, "learning_rate": 3.533586284429455e-05, "loss": 1.2165, "step": 5030 }, { "epoch": 0.21211290764593038, "grad_norm": 1.3036779165267944, "learning_rate": 3.534288926363126e-05, "loss": 1.1907, "step": 5031 }, { "epoch": 0.21215506882812993, "grad_norm": 0.8424158692359924, "learning_rate": 3.534991568296796e-05, "loss": 0.369, "step": 5032 }, { "epoch": 0.2121972300103295, "grad_norm": 0.6985147595405579, "learning_rate": 3.535694210230467e-05, "loss": 0.2429, "step": 5033 }, { "epoch": 0.21223939119252905, "grad_norm": 0.9192194938659668, "learning_rate": 3.536396852164137e-05, "loss": 0.9609, "step": 5034 }, { "epoch": 0.2122815523747286, "grad_norm": 0.6659377813339233, "learning_rate": 3.5370994940978084e-05, "loss": 0.3218, "step": 5035 }, { "epoch": 0.21232371355692814, "grad_norm": 0.8199304938316345, "learning_rate": 3.537802136031478e-05, "loss": 0.8451, "step": 5036 }, { "epoch": 0.21236587473912769, "grad_norm": 0.6622576713562012, "learning_rate": 3.538504777965149e-05, "loss": 0.2391, "step": 5037 }, { "epoch": 0.21240803592132723, "grad_norm": 0.6962406635284424, "learning_rate": 3.53920741989882e-05, "loss": 0.2888, "step": 5038 }, { "epoch": 0.21245019710352678, "grad_norm": 1.0586682558059692, "learning_rate": 3.53991006183249e-05, "loss": 0.9249, "step": 5039 }, { "epoch": 0.21249235828572632, "grad_norm": 0.8315641283988953, "learning_rate": 3.5406127037661606e-05, "loss": 0.4774, "step": 5040 }, { "epoch": 0.21253451946792587, "grad_norm": 0.8194881081581116, "learning_rate": 3.541315345699832e-05, "loss": 0.3037, "step": 5041 }, { "epoch": 0.21257668065012542, "grad_norm": 3.3931667804718018, "learning_rate": 3.542017987633502e-05, "loss": 1.1865, "step": 5042 }, { "epoch": 0.212618841832325, "grad_norm": 0.4435248076915741, "learning_rate": 3.5427206295671726e-05, "loss": 0.1476, "step": 5043 }, { "epoch": 0.21266100301452454, "grad_norm": 0.668286144733429, "learning_rate": 3.543423271500843e-05, "loss": 0.6464, "step": 5044 }, { "epoch": 0.21270316419672408, "grad_norm": 1.0052024126052856, "learning_rate": 3.544125913434514e-05, "loss": 1.1022, "step": 5045 }, { "epoch": 0.21274532537892363, "grad_norm": 0.45930901169776917, "learning_rate": 3.5448285553681846e-05, "loss": 0.1669, "step": 5046 }, { "epoch": 0.21278748656112317, "grad_norm": 1.0898561477661133, "learning_rate": 3.545531197301855e-05, "loss": 0.4333, "step": 5047 }, { "epoch": 0.21282964774332272, "grad_norm": 0.6993823051452637, "learning_rate": 3.546233839235526e-05, "loss": 0.1692, "step": 5048 }, { "epoch": 0.21287180892552227, "grad_norm": 0.9467359781265259, "learning_rate": 3.5469364811691966e-05, "loss": 0.5009, "step": 5049 }, { "epoch": 0.2129139701077218, "grad_norm": 0.4867406487464905, "learning_rate": 3.547639123102867e-05, "loss": 0.3673, "step": 5050 }, { "epoch": 0.21295613128992136, "grad_norm": 0.6006426811218262, "learning_rate": 3.5483417650365375e-05, "loss": 0.2695, "step": 5051 }, { "epoch": 0.21299829247212093, "grad_norm": 1.3288829326629639, "learning_rate": 3.5490444069702086e-05, "loss": 0.864, "step": 5052 }, { "epoch": 0.21304045365432048, "grad_norm": 0.6425029635429382, "learning_rate": 3.549747048903878e-05, "loss": 0.5263, "step": 5053 }, { "epoch": 0.21308261483652002, "grad_norm": 0.5576300621032715, "learning_rate": 3.5504496908375494e-05, "loss": 0.109, "step": 5054 }, { "epoch": 0.21312477601871957, "grad_norm": 0.5954257249832153, "learning_rate": 3.55115233277122e-05, "loss": 0.1974, "step": 5055 }, { "epoch": 0.21316693720091912, "grad_norm": 0.7504906058311462, "learning_rate": 3.551854974704891e-05, "loss": 0.6701, "step": 5056 }, { "epoch": 0.21320909838311866, "grad_norm": 0.4991443157196045, "learning_rate": 3.552557616638561e-05, "loss": 0.267, "step": 5057 }, { "epoch": 0.2132512595653182, "grad_norm": 1.0627943277359009, "learning_rate": 3.553260258572232e-05, "loss": 1.1072, "step": 5058 }, { "epoch": 0.21329342074751775, "grad_norm": 0.5644214153289795, "learning_rate": 3.553962900505902e-05, "loss": 0.2341, "step": 5059 }, { "epoch": 0.2133355819297173, "grad_norm": 0.3703562915325165, "learning_rate": 3.554665542439573e-05, "loss": 0.0919, "step": 5060 }, { "epoch": 0.21337774311191685, "grad_norm": 0.9615923166275024, "learning_rate": 3.555368184373243e-05, "loss": 0.3319, "step": 5061 }, { "epoch": 0.21341990429411642, "grad_norm": 0.6810343861579895, "learning_rate": 3.556070826306914e-05, "loss": 0.1847, "step": 5062 }, { "epoch": 0.21346206547631597, "grad_norm": 0.6318798661231995, "learning_rate": 3.556773468240585e-05, "loss": 0.5575, "step": 5063 }, { "epoch": 0.2135042266585155, "grad_norm": 0.4301462769508362, "learning_rate": 3.557476110174255e-05, "loss": 0.1352, "step": 5064 }, { "epoch": 0.21354638784071506, "grad_norm": 1.9681838750839233, "learning_rate": 3.5581787521079256e-05, "loss": 0.8173, "step": 5065 }, { "epoch": 0.2135885490229146, "grad_norm": 0.45620450377464294, "learning_rate": 3.558881394041597e-05, "loss": 0.0914, "step": 5066 }, { "epoch": 0.21363071020511415, "grad_norm": 0.7005986571311951, "learning_rate": 3.559584035975267e-05, "loss": 0.2318, "step": 5067 }, { "epoch": 0.2136728713873137, "grad_norm": 2.163680076599121, "learning_rate": 3.5602866779089376e-05, "loss": 0.794, "step": 5068 }, { "epoch": 0.21371503256951324, "grad_norm": 0.654066801071167, "learning_rate": 3.560989319842609e-05, "loss": 0.5615, "step": 5069 }, { "epoch": 0.2137571937517128, "grad_norm": 1.651964783668518, "learning_rate": 3.561691961776279e-05, "loss": 0.4388, "step": 5070 }, { "epoch": 0.21379935493391233, "grad_norm": 0.4875963628292084, "learning_rate": 3.5623946037099496e-05, "loss": 0.1792, "step": 5071 }, { "epoch": 0.2138415161161119, "grad_norm": 0.6284731030464172, "learning_rate": 3.56309724564362e-05, "loss": 0.1327, "step": 5072 }, { "epoch": 0.21388367729831145, "grad_norm": 1.7648005485534668, "learning_rate": 3.563799887577291e-05, "loss": 0.994, "step": 5073 }, { "epoch": 0.213925838480511, "grad_norm": 0.5711862444877625, "learning_rate": 3.5645025295109616e-05, "loss": 0.3904, "step": 5074 }, { "epoch": 0.21396799966271055, "grad_norm": 0.5102395415306091, "learning_rate": 3.565205171444632e-05, "loss": 0.288, "step": 5075 }, { "epoch": 0.2140101608449101, "grad_norm": 1.6300811767578125, "learning_rate": 3.5659078133783025e-05, "loss": 0.5734, "step": 5076 }, { "epoch": 0.21405232202710964, "grad_norm": 0.5116466879844666, "learning_rate": 3.5666104553119736e-05, "loss": 0.128, "step": 5077 }, { "epoch": 0.21409448320930918, "grad_norm": 0.5157278776168823, "learning_rate": 3.5673130972456434e-05, "loss": 0.1507, "step": 5078 }, { "epoch": 0.21413664439150873, "grad_norm": 0.6083346605300903, "learning_rate": 3.5680157391793145e-05, "loss": 0.5202, "step": 5079 }, { "epoch": 0.21417880557370828, "grad_norm": 0.4391235411167145, "learning_rate": 3.568718381112985e-05, "loss": 0.2895, "step": 5080 }, { "epoch": 0.21422096675590785, "grad_norm": 0.4875311553478241, "learning_rate": 3.5694210230466554e-05, "loss": 0.1175, "step": 5081 }, { "epoch": 0.2142631279381074, "grad_norm": 0.9903221726417542, "learning_rate": 3.570123664980326e-05, "loss": 1.1167, "step": 5082 }, { "epoch": 0.21430528912030694, "grad_norm": 0.7965100407600403, "learning_rate": 3.570826306913997e-05, "loss": 0.3687, "step": 5083 }, { "epoch": 0.2143474503025065, "grad_norm": 0.7923727035522461, "learning_rate": 3.5715289488476674e-05, "loss": 0.3958, "step": 5084 }, { "epoch": 0.21438961148470603, "grad_norm": 1.5369970798492432, "learning_rate": 3.572231590781338e-05, "loss": 0.8856, "step": 5085 }, { "epoch": 0.21443177266690558, "grad_norm": 0.7569512128829956, "learning_rate": 3.572934232715008e-05, "loss": 0.7617, "step": 5086 }, { "epoch": 0.21447393384910513, "grad_norm": 0.5754441022872925, "learning_rate": 3.5736368746486794e-05, "loss": 0.1647, "step": 5087 }, { "epoch": 0.21451609503130467, "grad_norm": 0.3510855436325073, "learning_rate": 3.57433951658235e-05, "loss": 0.1147, "step": 5088 }, { "epoch": 0.21455825621350422, "grad_norm": 0.6774204969406128, "learning_rate": 3.57504215851602e-05, "loss": 0.1738, "step": 5089 }, { "epoch": 0.21460041739570376, "grad_norm": 1.7173575162887573, "learning_rate": 3.575744800449691e-05, "loss": 0.998, "step": 5090 }, { "epoch": 0.21464257857790334, "grad_norm": 0.682462751865387, "learning_rate": 3.576447442383362e-05, "loss": 0.6125, "step": 5091 }, { "epoch": 0.21468473976010288, "grad_norm": 0.5908157229423523, "learning_rate": 3.577150084317032e-05, "loss": 0.1254, "step": 5092 }, { "epoch": 0.21472690094230243, "grad_norm": 0.5780213475227356, "learning_rate": 3.577852726250703e-05, "loss": 0.4013, "step": 5093 }, { "epoch": 0.21476906212450197, "grad_norm": 0.7282572388648987, "learning_rate": 3.578555368184374e-05, "loss": 0.1818, "step": 5094 }, { "epoch": 0.21481122330670152, "grad_norm": 0.792317271232605, "learning_rate": 3.579258010118044e-05, "loss": 0.2943, "step": 5095 }, { "epoch": 0.21485338448890107, "grad_norm": 0.3433517813682556, "learning_rate": 3.579960652051715e-05, "loss": 0.1101, "step": 5096 }, { "epoch": 0.2148955456711006, "grad_norm": 0.5468925833702087, "learning_rate": 3.580663293985385e-05, "loss": 0.1867, "step": 5097 }, { "epoch": 0.21493770685330016, "grad_norm": 1.0651414394378662, "learning_rate": 3.581365935919056e-05, "loss": 1.2125, "step": 5098 }, { "epoch": 0.2149798680354997, "grad_norm": 0.3961642384529114, "learning_rate": 3.582068577852726e-05, "loss": 0.093, "step": 5099 }, { "epoch": 0.21502202921769925, "grad_norm": 0.6925989985466003, "learning_rate": 3.582771219786397e-05, "loss": 0.7048, "step": 5100 }, { "epoch": 0.21506419039989882, "grad_norm": 1.0096070766448975, "learning_rate": 3.5834738617200676e-05, "loss": 0.7191, "step": 5101 }, { "epoch": 0.21510635158209837, "grad_norm": 0.4988728165626526, "learning_rate": 3.584176503653739e-05, "loss": 0.1925, "step": 5102 }, { "epoch": 0.21514851276429792, "grad_norm": 1.1442376375198364, "learning_rate": 3.5848791455874084e-05, "loss": 1.1801, "step": 5103 }, { "epoch": 0.21519067394649746, "grad_norm": 0.9170005321502686, "learning_rate": 3.5855817875210796e-05, "loss": 0.7887, "step": 5104 }, { "epoch": 0.215232835128697, "grad_norm": 0.6283953189849854, "learning_rate": 3.58628442945475e-05, "loss": 0.3076, "step": 5105 }, { "epoch": 0.21527499631089655, "grad_norm": 0.38652196526527405, "learning_rate": 3.5869870713884204e-05, "loss": 0.1246, "step": 5106 }, { "epoch": 0.2153171574930961, "grad_norm": 0.4944947063922882, "learning_rate": 3.587689713322091e-05, "loss": 0.1212, "step": 5107 }, { "epoch": 0.21535931867529565, "grad_norm": 0.535452127456665, "learning_rate": 3.588392355255762e-05, "loss": 0.3299, "step": 5108 }, { "epoch": 0.2154014798574952, "grad_norm": 0.381745845079422, "learning_rate": 3.5890949971894324e-05, "loss": 0.0743, "step": 5109 }, { "epoch": 0.21544364103969477, "grad_norm": 0.9632337689399719, "learning_rate": 3.589797639123103e-05, "loss": 0.2982, "step": 5110 }, { "epoch": 0.2154858022218943, "grad_norm": 0.35973235964775085, "learning_rate": 3.590500281056773e-05, "loss": 0.1446, "step": 5111 }, { "epoch": 0.21552796340409386, "grad_norm": 0.5975242853164673, "learning_rate": 3.5912029229904444e-05, "loss": 0.3253, "step": 5112 }, { "epoch": 0.2155701245862934, "grad_norm": 0.6028047800064087, "learning_rate": 3.591905564924115e-05, "loss": 0.1542, "step": 5113 }, { "epoch": 0.21561228576849295, "grad_norm": 0.7177472710609436, "learning_rate": 3.592608206857785e-05, "loss": 0.6089, "step": 5114 }, { "epoch": 0.2156544469506925, "grad_norm": 0.6652945280075073, "learning_rate": 3.5933108487914564e-05, "loss": 0.5372, "step": 5115 }, { "epoch": 0.21569660813289204, "grad_norm": 0.6384981870651245, "learning_rate": 3.594013490725127e-05, "loss": 0.5635, "step": 5116 }, { "epoch": 0.2157387693150916, "grad_norm": 1.2504006624221802, "learning_rate": 3.594716132658797e-05, "loss": 0.4136, "step": 5117 }, { "epoch": 0.21578093049729113, "grad_norm": 0.7137653231620789, "learning_rate": 3.595418774592468e-05, "loss": 0.7005, "step": 5118 }, { "epoch": 0.21582309167949068, "grad_norm": 0.6927432417869568, "learning_rate": 3.596121416526139e-05, "loss": 1.6719, "step": 5119 }, { "epoch": 0.21586525286169025, "grad_norm": 0.6848428845405579, "learning_rate": 3.596824058459809e-05, "loss": 0.4803, "step": 5120 }, { "epoch": 0.2159074140438898, "grad_norm": 0.8200965523719788, "learning_rate": 3.59752670039348e-05, "loss": 0.7043, "step": 5121 }, { "epoch": 0.21594957522608935, "grad_norm": 1.1928292512893677, "learning_rate": 3.59822934232715e-05, "loss": 0.6308, "step": 5122 }, { "epoch": 0.2159917364082889, "grad_norm": 0.330515593290329, "learning_rate": 3.598931984260821e-05, "loss": 0.0934, "step": 5123 }, { "epoch": 0.21603389759048844, "grad_norm": 1.9550437927246094, "learning_rate": 3.599634626194491e-05, "loss": 0.9324, "step": 5124 }, { "epoch": 0.21607605877268798, "grad_norm": 0.41232073307037354, "learning_rate": 3.600337268128162e-05, "loss": 0.2424, "step": 5125 }, { "epoch": 0.21611821995488753, "grad_norm": 0.5405533909797668, "learning_rate": 3.6010399100618326e-05, "loss": 0.22, "step": 5126 }, { "epoch": 0.21616038113708708, "grad_norm": 0.49259886145591736, "learning_rate": 3.601742551995503e-05, "loss": 0.2412, "step": 5127 }, { "epoch": 0.21620254231928662, "grad_norm": 0.47107356786727905, "learning_rate": 3.6024451939291735e-05, "loss": 0.1012, "step": 5128 }, { "epoch": 0.21624470350148617, "grad_norm": 0.5754517912864685, "learning_rate": 3.6031478358628446e-05, "loss": 0.1118, "step": 5129 }, { "epoch": 0.21628686468368574, "grad_norm": 0.42240890860557556, "learning_rate": 3.603850477796515e-05, "loss": 0.1605, "step": 5130 }, { "epoch": 0.2163290258658853, "grad_norm": 0.6432012319564819, "learning_rate": 3.6045531197301855e-05, "loss": 0.1835, "step": 5131 }, { "epoch": 0.21637118704808483, "grad_norm": 0.5447766780853271, "learning_rate": 3.605255761663856e-05, "loss": 0.1626, "step": 5132 }, { "epoch": 0.21641334823028438, "grad_norm": 1.24013352394104, "learning_rate": 3.605958403597527e-05, "loss": 1.0136, "step": 5133 }, { "epoch": 0.21645550941248393, "grad_norm": 1.6153309345245361, "learning_rate": 3.6066610455311975e-05, "loss": 1.3647, "step": 5134 }, { "epoch": 0.21649767059468347, "grad_norm": 0.6735021471977234, "learning_rate": 3.607363687464868e-05, "loss": 0.6551, "step": 5135 }, { "epoch": 0.21653983177688302, "grad_norm": 0.5447114706039429, "learning_rate": 3.608066329398539e-05, "loss": 0.1261, "step": 5136 }, { "epoch": 0.21658199295908256, "grad_norm": 0.6872590184211731, "learning_rate": 3.6087689713322095e-05, "loss": 0.4604, "step": 5137 }, { "epoch": 0.2166241541412821, "grad_norm": 0.4816111624240875, "learning_rate": 3.60947161326588e-05, "loss": 0.2177, "step": 5138 }, { "epoch": 0.21666631532348168, "grad_norm": 0.6744861006736755, "learning_rate": 3.6101742551995504e-05, "loss": 0.2009, "step": 5139 }, { "epoch": 0.21670847650568123, "grad_norm": 0.762620210647583, "learning_rate": 3.6108768971332215e-05, "loss": 0.3984, "step": 5140 }, { "epoch": 0.21675063768788078, "grad_norm": 0.5901317000389099, "learning_rate": 3.611579539066892e-05, "loss": 0.2506, "step": 5141 }, { "epoch": 0.21679279887008032, "grad_norm": 0.45445674657821655, "learning_rate": 3.6122821810005624e-05, "loss": 0.1851, "step": 5142 }, { "epoch": 0.21683496005227987, "grad_norm": 0.6116588115692139, "learning_rate": 3.612984822934233e-05, "loss": 0.2222, "step": 5143 }, { "epoch": 0.21687712123447941, "grad_norm": 0.6091851592063904, "learning_rate": 3.613687464867904e-05, "loss": 0.4966, "step": 5144 }, { "epoch": 0.21691928241667896, "grad_norm": 0.8446219563484192, "learning_rate": 3.614390106801574e-05, "loss": 0.6965, "step": 5145 }, { "epoch": 0.2169614435988785, "grad_norm": 0.4400395154953003, "learning_rate": 3.615092748735245e-05, "loss": 0.1176, "step": 5146 }, { "epoch": 0.21700360478107805, "grad_norm": 2.472370147705078, "learning_rate": 3.615795390668915e-05, "loss": 0.5693, "step": 5147 }, { "epoch": 0.2170457659632776, "grad_norm": 0.6600733399391174, "learning_rate": 3.6164980326025863e-05, "loss": 0.3054, "step": 5148 }, { "epoch": 0.21708792714547717, "grad_norm": 0.5959175825119019, "learning_rate": 3.617200674536256e-05, "loss": 0.2563, "step": 5149 }, { "epoch": 0.21713008832767672, "grad_norm": 0.5680339336395264, "learning_rate": 3.617903316469927e-05, "loss": 0.3032, "step": 5150 }, { "epoch": 0.21717224950987626, "grad_norm": 0.6942978501319885, "learning_rate": 3.618605958403598e-05, "loss": 0.9192, "step": 5151 }, { "epoch": 0.2172144106920758, "grad_norm": 0.5169601440429688, "learning_rate": 3.619308600337268e-05, "loss": 0.1126, "step": 5152 }, { "epoch": 0.21725657187427536, "grad_norm": 0.8397008776664734, "learning_rate": 3.6200112422709385e-05, "loss": 0.1646, "step": 5153 }, { "epoch": 0.2172987330564749, "grad_norm": 0.5140244960784912, "learning_rate": 3.62071388420461e-05, "loss": 0.1022, "step": 5154 }, { "epoch": 0.21734089423867445, "grad_norm": 1.0480263233184814, "learning_rate": 3.62141652613828e-05, "loss": 1.0029, "step": 5155 }, { "epoch": 0.217383055420874, "grad_norm": 0.7085488438606262, "learning_rate": 3.6221191680719505e-05, "loss": 0.2065, "step": 5156 }, { "epoch": 0.21742521660307354, "grad_norm": 0.5864880681037903, "learning_rate": 3.622821810005621e-05, "loss": 0.1831, "step": 5157 }, { "epoch": 0.21746737778527309, "grad_norm": 0.6968810558319092, "learning_rate": 3.623524451939292e-05, "loss": 0.563, "step": 5158 }, { "epoch": 0.21750953896747266, "grad_norm": 0.5970064401626587, "learning_rate": 3.6242270938729625e-05, "loss": 0.2043, "step": 5159 }, { "epoch": 0.2175517001496722, "grad_norm": 0.7715983986854553, "learning_rate": 3.624929735806633e-05, "loss": 0.3598, "step": 5160 }, { "epoch": 0.21759386133187175, "grad_norm": 0.5699469447135925, "learning_rate": 3.625632377740304e-05, "loss": 0.3075, "step": 5161 }, { "epoch": 0.2176360225140713, "grad_norm": 1.0306330919265747, "learning_rate": 3.6263350196739745e-05, "loss": 1.1288, "step": 5162 }, { "epoch": 0.21767818369627084, "grad_norm": 0.8415075540542603, "learning_rate": 3.627037661607645e-05, "loss": 0.4447, "step": 5163 }, { "epoch": 0.2177203448784704, "grad_norm": 0.7718161344528198, "learning_rate": 3.6277403035413154e-05, "loss": 0.1729, "step": 5164 }, { "epoch": 0.21776250606066994, "grad_norm": 0.962016224861145, "learning_rate": 3.6284429454749865e-05, "loss": 0.2055, "step": 5165 }, { "epoch": 0.21780466724286948, "grad_norm": 0.9551565647125244, "learning_rate": 3.629145587408656e-05, "loss": 0.2431, "step": 5166 }, { "epoch": 0.21784682842506903, "grad_norm": 0.5280858874320984, "learning_rate": 3.6298482293423274e-05, "loss": 0.1998, "step": 5167 }, { "epoch": 0.2178889896072686, "grad_norm": 0.9214928150177002, "learning_rate": 3.630550871275998e-05, "loss": 0.2394, "step": 5168 }, { "epoch": 0.21793115078946815, "grad_norm": 0.41809654235839844, "learning_rate": 3.631253513209669e-05, "loss": 0.1117, "step": 5169 }, { "epoch": 0.2179733119716677, "grad_norm": 0.5355979800224304, "learning_rate": 3.631956155143339e-05, "loss": 0.1176, "step": 5170 }, { "epoch": 0.21801547315386724, "grad_norm": 0.492719441652298, "learning_rate": 3.63265879707701e-05, "loss": 0.1218, "step": 5171 }, { "epoch": 0.21805763433606679, "grad_norm": 0.6107136607170105, "learning_rate": 3.63336143901068e-05, "loss": 0.2491, "step": 5172 }, { "epoch": 0.21809979551826633, "grad_norm": 0.4668187201023102, "learning_rate": 3.634064080944351e-05, "loss": 0.1305, "step": 5173 }, { "epoch": 0.21814195670046588, "grad_norm": 0.42455825209617615, "learning_rate": 3.634766722878021e-05, "loss": 0.1252, "step": 5174 }, { "epoch": 0.21818411788266542, "grad_norm": 0.6396365761756897, "learning_rate": 3.635469364811692e-05, "loss": 0.2169, "step": 5175 }, { "epoch": 0.21822627906486497, "grad_norm": 0.695950448513031, "learning_rate": 3.636172006745363e-05, "loss": 0.1687, "step": 5176 }, { "epoch": 0.21826844024706452, "grad_norm": 0.589792788028717, "learning_rate": 3.636874648679033e-05, "loss": 0.5701, "step": 5177 }, { "epoch": 0.2183106014292641, "grad_norm": 0.764434278011322, "learning_rate": 3.6375772906127036e-05, "loss": 1.022, "step": 5178 }, { "epoch": 0.21835276261146364, "grad_norm": 0.6096768379211426, "learning_rate": 3.638279932546375e-05, "loss": 0.1805, "step": 5179 }, { "epoch": 0.21839492379366318, "grad_norm": 0.9999146461486816, "learning_rate": 3.638982574480045e-05, "loss": 1.0202, "step": 5180 }, { "epoch": 0.21843708497586273, "grad_norm": 1.2757891416549683, "learning_rate": 3.6396852164137156e-05, "loss": 1.0936, "step": 5181 }, { "epoch": 0.21847924615806227, "grad_norm": 0.755852997303009, "learning_rate": 3.640387858347387e-05, "loss": 0.7145, "step": 5182 }, { "epoch": 0.21852140734026182, "grad_norm": 0.5405344367027283, "learning_rate": 3.641090500281057e-05, "loss": 0.4697, "step": 5183 }, { "epoch": 0.21856356852246137, "grad_norm": 0.8302650451660156, "learning_rate": 3.6417931422147276e-05, "loss": 0.3846, "step": 5184 }, { "epoch": 0.2186057297046609, "grad_norm": 0.5824549794197083, "learning_rate": 3.642495784148398e-05, "loss": 0.1247, "step": 5185 }, { "epoch": 0.21864789088686046, "grad_norm": 0.8250751495361328, "learning_rate": 3.643198426082069e-05, "loss": 0.6319, "step": 5186 }, { "epoch": 0.21869005206906, "grad_norm": 0.6274031400680542, "learning_rate": 3.6439010680157396e-05, "loss": 0.1605, "step": 5187 }, { "epoch": 0.21873221325125958, "grad_norm": 0.4924379289150238, "learning_rate": 3.64460370994941e-05, "loss": 0.0903, "step": 5188 }, { "epoch": 0.21877437443345912, "grad_norm": 0.4668780565261841, "learning_rate": 3.6453063518830805e-05, "loss": 0.2182, "step": 5189 }, { "epoch": 0.21881653561565867, "grad_norm": 0.4361439645290375, "learning_rate": 3.6460089938167516e-05, "loss": 0.2274, "step": 5190 }, { "epoch": 0.21885869679785822, "grad_norm": 0.7374376654624939, "learning_rate": 3.6467116357504213e-05, "loss": 0.3403, "step": 5191 }, { "epoch": 0.21890085798005776, "grad_norm": 0.6451749801635742, "learning_rate": 3.6474142776840925e-05, "loss": 0.2608, "step": 5192 }, { "epoch": 0.2189430191622573, "grad_norm": 0.9552857279777527, "learning_rate": 3.648116919617763e-05, "loss": 0.6913, "step": 5193 }, { "epoch": 0.21898518034445685, "grad_norm": 2.0786986351013184, "learning_rate": 3.6488195615514333e-05, "loss": 0.5341, "step": 5194 }, { "epoch": 0.2190273415266564, "grad_norm": 0.5975157618522644, "learning_rate": 3.649522203485104e-05, "loss": 0.5879, "step": 5195 }, { "epoch": 0.21906950270885595, "grad_norm": 0.44939059019088745, "learning_rate": 3.650224845418775e-05, "loss": 0.1626, "step": 5196 }, { "epoch": 0.2191116638910555, "grad_norm": 0.48220768570899963, "learning_rate": 3.650927487352445e-05, "loss": 0.2059, "step": 5197 }, { "epoch": 0.21915382507325507, "grad_norm": 0.6901293992996216, "learning_rate": 3.651630129286116e-05, "loss": 0.1711, "step": 5198 }, { "epoch": 0.2191959862554546, "grad_norm": 0.3948568105697632, "learning_rate": 3.652332771219786e-05, "loss": 0.1179, "step": 5199 }, { "epoch": 0.21923814743765416, "grad_norm": 0.567825198173523, "learning_rate": 3.653035413153457e-05, "loss": 0.5122, "step": 5200 }, { "epoch": 0.2192803086198537, "grad_norm": 0.568016767501831, "learning_rate": 3.653738055087128e-05, "loss": 0.2232, "step": 5201 }, { "epoch": 0.21932246980205325, "grad_norm": 0.5909216403961182, "learning_rate": 3.654440697020798e-05, "loss": 0.2441, "step": 5202 }, { "epoch": 0.2193646309842528, "grad_norm": 0.9474493265151978, "learning_rate": 3.655143338954469e-05, "loss": 0.1767, "step": 5203 }, { "epoch": 0.21940679216645234, "grad_norm": 0.9199528098106384, "learning_rate": 3.65584598088814e-05, "loss": 0.1829, "step": 5204 }, { "epoch": 0.2194489533486519, "grad_norm": 0.9003267884254456, "learning_rate": 3.65654862282181e-05, "loss": 0.515, "step": 5205 }, { "epoch": 0.21949111453085143, "grad_norm": 0.7067112922668457, "learning_rate": 3.6572512647554806e-05, "loss": 0.4835, "step": 5206 }, { "epoch": 0.219533275713051, "grad_norm": 0.7751061320304871, "learning_rate": 3.657953906689152e-05, "loss": 0.183, "step": 5207 }, { "epoch": 0.21957543689525055, "grad_norm": 1.0703661441802979, "learning_rate": 3.658656548622822e-05, "loss": 1.0269, "step": 5208 }, { "epoch": 0.2196175980774501, "grad_norm": 0.731570839881897, "learning_rate": 3.6593591905564926e-05, "loss": 1.4514, "step": 5209 }, { "epoch": 0.21965975925964965, "grad_norm": 0.35844162106513977, "learning_rate": 3.660061832490163e-05, "loss": 0.1548, "step": 5210 }, { "epoch": 0.2197019204418492, "grad_norm": 0.4169897735118866, "learning_rate": 3.660764474423834e-05, "loss": 0.1812, "step": 5211 }, { "epoch": 0.21974408162404874, "grad_norm": 0.9607595801353455, "learning_rate": 3.661467116357504e-05, "loss": 1.2902, "step": 5212 }, { "epoch": 0.21978624280624828, "grad_norm": 0.6769492626190186, "learning_rate": 3.662169758291175e-05, "loss": 0.1741, "step": 5213 }, { "epoch": 0.21982840398844783, "grad_norm": 1.2385839223861694, "learning_rate": 3.6628724002248455e-05, "loss": 0.482, "step": 5214 }, { "epoch": 0.21987056517064738, "grad_norm": 0.585206151008606, "learning_rate": 3.6635750421585166e-05, "loss": 0.2269, "step": 5215 }, { "epoch": 0.21991272635284692, "grad_norm": 0.7006313800811768, "learning_rate": 3.6642776840921864e-05, "loss": 0.1446, "step": 5216 }, { "epoch": 0.2199548875350465, "grad_norm": 0.41866904497146606, "learning_rate": 3.6649803260258575e-05, "loss": 0.1877, "step": 5217 }, { "epoch": 0.21999704871724604, "grad_norm": 0.5537666082382202, "learning_rate": 3.665682967959528e-05, "loss": 0.1853, "step": 5218 }, { "epoch": 0.2200392098994456, "grad_norm": 0.5243241190910339, "learning_rate": 3.6663856098931984e-05, "loss": 0.1362, "step": 5219 }, { "epoch": 0.22008137108164513, "grad_norm": 1.6578484773635864, "learning_rate": 3.667088251826869e-05, "loss": 0.5084, "step": 5220 }, { "epoch": 0.22012353226384468, "grad_norm": 0.8743612170219421, "learning_rate": 3.66779089376054e-05, "loss": 0.5666, "step": 5221 }, { "epoch": 0.22016569344604422, "grad_norm": 0.49401912093162537, "learning_rate": 3.6684935356942104e-05, "loss": 0.0962, "step": 5222 }, { "epoch": 0.22020785462824377, "grad_norm": 0.744086503982544, "learning_rate": 3.669196177627881e-05, "loss": 0.3329, "step": 5223 }, { "epoch": 0.22025001581044332, "grad_norm": 0.48749759793281555, "learning_rate": 3.669898819561551e-05, "loss": 0.1455, "step": 5224 }, { "epoch": 0.22029217699264286, "grad_norm": 0.44557085633277893, "learning_rate": 3.6706014614952224e-05, "loss": 0.121, "step": 5225 }, { "epoch": 0.2203343381748424, "grad_norm": 0.7328044772148132, "learning_rate": 3.671304103428893e-05, "loss": 0.2438, "step": 5226 }, { "epoch": 0.22037649935704198, "grad_norm": 0.5647602677345276, "learning_rate": 3.672006745362563e-05, "loss": 0.2344, "step": 5227 }, { "epoch": 0.22041866053924153, "grad_norm": 0.7803656458854675, "learning_rate": 3.6727093872962344e-05, "loss": 0.321, "step": 5228 }, { "epoch": 0.22046082172144107, "grad_norm": 0.7182205319404602, "learning_rate": 3.673412029229905e-05, "loss": 0.6057, "step": 5229 }, { "epoch": 0.22050298290364062, "grad_norm": 0.5764703154563904, "learning_rate": 3.674114671163575e-05, "loss": 0.1752, "step": 5230 }, { "epoch": 0.22054514408584017, "grad_norm": 0.3534638285636902, "learning_rate": 3.674817313097246e-05, "loss": 0.0703, "step": 5231 }, { "epoch": 0.2205873052680397, "grad_norm": 0.5690711140632629, "learning_rate": 3.675519955030917e-05, "loss": 0.3013, "step": 5232 }, { "epoch": 0.22062946645023926, "grad_norm": 0.9650870561599731, "learning_rate": 3.676222596964587e-05, "loss": 0.4287, "step": 5233 }, { "epoch": 0.2206716276324388, "grad_norm": 0.42965659499168396, "learning_rate": 3.676925238898258e-05, "loss": 0.1378, "step": 5234 }, { "epoch": 0.22071378881463835, "grad_norm": 0.24469614028930664, "learning_rate": 3.677627880831928e-05, "loss": 0.0739, "step": 5235 }, { "epoch": 0.22075594999683792, "grad_norm": 0.6750406622886658, "learning_rate": 3.678330522765599e-05, "loss": 0.197, "step": 5236 }, { "epoch": 0.22079811117903747, "grad_norm": 0.39008498191833496, "learning_rate": 3.679033164699269e-05, "loss": 0.1274, "step": 5237 }, { "epoch": 0.22084027236123702, "grad_norm": 0.8135324120521545, "learning_rate": 3.67973580663294e-05, "loss": 0.3803, "step": 5238 }, { "epoch": 0.22088243354343656, "grad_norm": 0.6656026840209961, "learning_rate": 3.6804384485666106e-05, "loss": 0.5296, "step": 5239 }, { "epoch": 0.2209245947256361, "grad_norm": 0.6048420071601868, "learning_rate": 3.681141090500281e-05, "loss": 0.2132, "step": 5240 }, { "epoch": 0.22096675590783565, "grad_norm": 0.6321074366569519, "learning_rate": 3.6818437324339515e-05, "loss": 0.2793, "step": 5241 }, { "epoch": 0.2210089170900352, "grad_norm": 2.4955880641937256, "learning_rate": 3.6825463743676226e-05, "loss": 0.8736, "step": 5242 }, { "epoch": 0.22105107827223475, "grad_norm": 1.1999119520187378, "learning_rate": 3.683249016301293e-05, "loss": 0.3675, "step": 5243 }, { "epoch": 0.2210932394544343, "grad_norm": 0.7592840194702148, "learning_rate": 3.6839516582349634e-05, "loss": 0.2409, "step": 5244 }, { "epoch": 0.22113540063663384, "grad_norm": 0.5643768310546875, "learning_rate": 3.684654300168634e-05, "loss": 0.1617, "step": 5245 }, { "epoch": 0.2211775618188334, "grad_norm": 0.38674280047416687, "learning_rate": 3.685356942102305e-05, "loss": 0.1491, "step": 5246 }, { "epoch": 0.22121972300103296, "grad_norm": 0.6925111413002014, "learning_rate": 3.6860595840359754e-05, "loss": 0.1151, "step": 5247 }, { "epoch": 0.2212618841832325, "grad_norm": 0.7141271233558655, "learning_rate": 3.686762225969646e-05, "loss": 0.2268, "step": 5248 }, { "epoch": 0.22130404536543205, "grad_norm": 0.6087226271629333, "learning_rate": 3.687464867903317e-05, "loss": 0.1102, "step": 5249 }, { "epoch": 0.2213462065476316, "grad_norm": 0.4925669729709625, "learning_rate": 3.6881675098369874e-05, "loss": 0.1123, "step": 5250 }, { "epoch": 0.22138836772983114, "grad_norm": 0.5472337007522583, "learning_rate": 3.688870151770658e-05, "loss": 0.1653, "step": 5251 }, { "epoch": 0.2214305289120307, "grad_norm": 0.46495702862739563, "learning_rate": 3.689572793704328e-05, "loss": 0.232, "step": 5252 }, { "epoch": 0.22147269009423023, "grad_norm": 0.531650185585022, "learning_rate": 3.6902754356379994e-05, "loss": 0.1704, "step": 5253 }, { "epoch": 0.22151485127642978, "grad_norm": 0.5756504535675049, "learning_rate": 3.69097807757167e-05, "loss": 0.3689, "step": 5254 }, { "epoch": 0.22155701245862933, "grad_norm": 0.6250790357589722, "learning_rate": 3.69168071950534e-05, "loss": 0.2482, "step": 5255 }, { "epoch": 0.2215991736408289, "grad_norm": 0.4530881941318512, "learning_rate": 3.692383361439011e-05, "loss": 0.1329, "step": 5256 }, { "epoch": 0.22164133482302845, "grad_norm": 0.5314713716506958, "learning_rate": 3.693086003372682e-05, "loss": 0.3877, "step": 5257 }, { "epoch": 0.221683496005228, "grad_norm": 0.33016204833984375, "learning_rate": 3.6937886453063516e-05, "loss": 0.1511, "step": 5258 }, { "epoch": 0.22172565718742754, "grad_norm": 0.6931766271591187, "learning_rate": 3.694491287240023e-05, "loss": 0.7055, "step": 5259 }, { "epoch": 0.22176781836962708, "grad_norm": 0.6290098428726196, "learning_rate": 3.695193929173693e-05, "loss": 0.1891, "step": 5260 }, { "epoch": 0.22180997955182663, "grad_norm": 0.44426295161247253, "learning_rate": 3.695896571107364e-05, "loss": 0.1741, "step": 5261 }, { "epoch": 0.22185214073402618, "grad_norm": 0.35842254757881165, "learning_rate": 3.696599213041034e-05, "loss": 0.2264, "step": 5262 }, { "epoch": 0.22189430191622572, "grad_norm": 0.5241872072219849, "learning_rate": 3.697301854974705e-05, "loss": 0.2126, "step": 5263 }, { "epoch": 0.22193646309842527, "grad_norm": 0.3904970586299896, "learning_rate": 3.6980044969083756e-05, "loss": 0.2018, "step": 5264 }, { "epoch": 0.22197862428062484, "grad_norm": 1.1567493677139282, "learning_rate": 3.698707138842046e-05, "loss": 0.5903, "step": 5265 }, { "epoch": 0.2220207854628244, "grad_norm": 0.5548440217971802, "learning_rate": 3.6994097807757165e-05, "loss": 0.5461, "step": 5266 }, { "epoch": 0.22206294664502393, "grad_norm": 0.38186484575271606, "learning_rate": 3.7001124227093876e-05, "loss": 0.2094, "step": 5267 }, { "epoch": 0.22210510782722348, "grad_norm": 0.713882565498352, "learning_rate": 3.700815064643058e-05, "loss": 0.2755, "step": 5268 }, { "epoch": 0.22214726900942303, "grad_norm": 0.7963846325874329, "learning_rate": 3.7015177065767285e-05, "loss": 0.357, "step": 5269 }, { "epoch": 0.22218943019162257, "grad_norm": 0.5260047912597656, "learning_rate": 3.702220348510399e-05, "loss": 0.2537, "step": 5270 }, { "epoch": 0.22223159137382212, "grad_norm": 0.5091021060943604, "learning_rate": 3.70292299044407e-05, "loss": 0.1547, "step": 5271 }, { "epoch": 0.22227375255602166, "grad_norm": 0.6383312940597534, "learning_rate": 3.7036256323777405e-05, "loss": 0.6859, "step": 5272 }, { "epoch": 0.2223159137382212, "grad_norm": 1.3290270566940308, "learning_rate": 3.704328274311411e-05, "loss": 1.3097, "step": 5273 }, { "epoch": 0.22235807492042076, "grad_norm": 0.318589985370636, "learning_rate": 3.705030916245082e-05, "loss": 0.1087, "step": 5274 }, { "epoch": 0.22240023610262033, "grad_norm": 0.5659578442573547, "learning_rate": 3.7057335581787525e-05, "loss": 0.1612, "step": 5275 }, { "epoch": 0.22244239728481988, "grad_norm": 0.5607427358627319, "learning_rate": 3.706436200112423e-05, "loss": 0.6972, "step": 5276 }, { "epoch": 0.22248455846701942, "grad_norm": 0.6762577295303345, "learning_rate": 3.7071388420460934e-05, "loss": 0.6614, "step": 5277 }, { "epoch": 0.22252671964921897, "grad_norm": 0.5870858430862427, "learning_rate": 3.7078414839797645e-05, "loss": 0.2192, "step": 5278 }, { "epoch": 0.22256888083141851, "grad_norm": 0.48755523562431335, "learning_rate": 3.708544125913434e-05, "loss": 0.171, "step": 5279 }, { "epoch": 0.22261104201361806, "grad_norm": 0.7740013599395752, "learning_rate": 3.7092467678471054e-05, "loss": 0.727, "step": 5280 }, { "epoch": 0.2226532031958176, "grad_norm": 0.7004456520080566, "learning_rate": 3.709949409780776e-05, "loss": 0.6654, "step": 5281 }, { "epoch": 0.22269536437801715, "grad_norm": 1.3209996223449707, "learning_rate": 3.710652051714447e-05, "loss": 1.1304, "step": 5282 }, { "epoch": 0.2227375255602167, "grad_norm": 0.6819763779640198, "learning_rate": 3.711354693648117e-05, "loss": 0.7392, "step": 5283 }, { "epoch": 0.22277968674241624, "grad_norm": 0.6317685842514038, "learning_rate": 3.712057335581788e-05, "loss": 0.3423, "step": 5284 }, { "epoch": 0.22282184792461582, "grad_norm": 0.794204592704773, "learning_rate": 3.712759977515458e-05, "loss": 0.2727, "step": 5285 }, { "epoch": 0.22286400910681536, "grad_norm": 1.6220457553863525, "learning_rate": 3.713462619449129e-05, "loss": 0.4515, "step": 5286 }, { "epoch": 0.2229061702890149, "grad_norm": 0.4484397768974304, "learning_rate": 3.714165261382799e-05, "loss": 0.1619, "step": 5287 }, { "epoch": 0.22294833147121446, "grad_norm": 0.5538145899772644, "learning_rate": 3.71486790331647e-05, "loss": 0.5923, "step": 5288 }, { "epoch": 0.222990492653414, "grad_norm": 0.9479805827140808, "learning_rate": 3.715570545250141e-05, "loss": 0.9068, "step": 5289 }, { "epoch": 0.22303265383561355, "grad_norm": 0.399913489818573, "learning_rate": 3.716273187183811e-05, "loss": 0.1304, "step": 5290 }, { "epoch": 0.2230748150178131, "grad_norm": 0.6767340302467346, "learning_rate": 3.7169758291174816e-05, "loss": 0.2528, "step": 5291 }, { "epoch": 0.22311697620001264, "grad_norm": 0.5455828905105591, "learning_rate": 3.717678471051153e-05, "loss": 0.6841, "step": 5292 }, { "epoch": 0.22315913738221219, "grad_norm": 0.6424224376678467, "learning_rate": 3.718381112984823e-05, "loss": 0.324, "step": 5293 }, { "epoch": 0.22320129856441176, "grad_norm": 0.525037944316864, "learning_rate": 3.7190837549184936e-05, "loss": 0.1663, "step": 5294 }, { "epoch": 0.2232434597466113, "grad_norm": 0.7064283490180969, "learning_rate": 3.719786396852165e-05, "loss": 0.1823, "step": 5295 }, { "epoch": 0.22328562092881085, "grad_norm": 0.5338941216468811, "learning_rate": 3.720489038785835e-05, "loss": 0.2669, "step": 5296 }, { "epoch": 0.2233277821110104, "grad_norm": 0.6336111426353455, "learning_rate": 3.7211916807195056e-05, "loss": 0.1747, "step": 5297 }, { "epoch": 0.22336994329320994, "grad_norm": 0.5339506268501282, "learning_rate": 3.721894322653176e-05, "loss": 0.1037, "step": 5298 }, { "epoch": 0.2234121044754095, "grad_norm": 0.5835179090499878, "learning_rate": 3.722596964586847e-05, "loss": 0.5532, "step": 5299 }, { "epoch": 0.22345426565760904, "grad_norm": 0.7036991715431213, "learning_rate": 3.7232996065205175e-05, "loss": 0.5165, "step": 5300 }, { "epoch": 0.22349642683980858, "grad_norm": 1.02829909324646, "learning_rate": 3.724002248454188e-05, "loss": 1.1183, "step": 5301 }, { "epoch": 0.22353858802200813, "grad_norm": 0.4868949055671692, "learning_rate": 3.7247048903878584e-05, "loss": 0.1312, "step": 5302 }, { "epoch": 0.22358074920420767, "grad_norm": 0.6277164816856384, "learning_rate": 3.7254075323215295e-05, "loss": 0.2455, "step": 5303 }, { "epoch": 0.22362291038640725, "grad_norm": 0.41436317563056946, "learning_rate": 3.726110174255199e-05, "loss": 0.2563, "step": 5304 }, { "epoch": 0.2236650715686068, "grad_norm": 0.5003829002380371, "learning_rate": 3.7268128161888704e-05, "loss": 0.1567, "step": 5305 }, { "epoch": 0.22370723275080634, "grad_norm": 0.3885522186756134, "learning_rate": 3.727515458122541e-05, "loss": 0.1359, "step": 5306 }, { "epoch": 0.22374939393300589, "grad_norm": 0.7356507778167725, "learning_rate": 3.728218100056211e-05, "loss": 0.3961, "step": 5307 }, { "epoch": 0.22379155511520543, "grad_norm": 0.6132249236106873, "learning_rate": 3.728920741989882e-05, "loss": 0.3371, "step": 5308 }, { "epoch": 0.22383371629740498, "grad_norm": 0.36056333780288696, "learning_rate": 3.729623383923553e-05, "loss": 0.0956, "step": 5309 }, { "epoch": 0.22387587747960452, "grad_norm": 0.7246581315994263, "learning_rate": 3.730326025857223e-05, "loss": 0.5785, "step": 5310 }, { "epoch": 0.22391803866180407, "grad_norm": 0.8582298755645752, "learning_rate": 3.731028667790894e-05, "loss": 0.4735, "step": 5311 }, { "epoch": 0.22396019984400362, "grad_norm": 0.5140970945358276, "learning_rate": 3.731731309724564e-05, "loss": 0.2809, "step": 5312 }, { "epoch": 0.22400236102620316, "grad_norm": 0.8156713247299194, "learning_rate": 3.732433951658235e-05, "loss": 0.7497, "step": 5313 }, { "epoch": 0.22404452220840274, "grad_norm": 0.7045878767967224, "learning_rate": 3.733136593591906e-05, "loss": 0.339, "step": 5314 }, { "epoch": 0.22408668339060228, "grad_norm": 0.6648823618888855, "learning_rate": 3.733839235525576e-05, "loss": 0.6057, "step": 5315 }, { "epoch": 0.22412884457280183, "grad_norm": 0.3872586786746979, "learning_rate": 3.734541877459247e-05, "loss": 0.1997, "step": 5316 }, { "epoch": 0.22417100575500137, "grad_norm": 0.31159526109695435, "learning_rate": 3.735244519392918e-05, "loss": 0.1254, "step": 5317 }, { "epoch": 0.22421316693720092, "grad_norm": 0.45180678367614746, "learning_rate": 3.735947161326588e-05, "loss": 0.2364, "step": 5318 }, { "epoch": 0.22425532811940047, "grad_norm": 2.112959384918213, "learning_rate": 3.7366498032602586e-05, "loss": 0.5156, "step": 5319 }, { "epoch": 0.2242974893016, "grad_norm": 0.5542256236076355, "learning_rate": 3.73735244519393e-05, "loss": 0.3886, "step": 5320 }, { "epoch": 0.22433965048379956, "grad_norm": 1.3591269254684448, "learning_rate": 3.7380550871276e-05, "loss": 0.7165, "step": 5321 }, { "epoch": 0.2243818116659991, "grad_norm": 0.6855183839797974, "learning_rate": 3.7387577290612706e-05, "loss": 0.2106, "step": 5322 }, { "epoch": 0.22442397284819868, "grad_norm": 0.7793599963188171, "learning_rate": 3.739460370994941e-05, "loss": 0.4654, "step": 5323 }, { "epoch": 0.22446613403039822, "grad_norm": 0.6559529900550842, "learning_rate": 3.740163012928612e-05, "loss": 0.4732, "step": 5324 }, { "epoch": 0.22450829521259777, "grad_norm": 0.41201919317245483, "learning_rate": 3.740865654862282e-05, "loss": 0.0852, "step": 5325 }, { "epoch": 0.22455045639479732, "grad_norm": 0.642423152923584, "learning_rate": 3.741568296795953e-05, "loss": 0.6642, "step": 5326 }, { "epoch": 0.22459261757699686, "grad_norm": 0.47724902629852295, "learning_rate": 3.7422709387296235e-05, "loss": 0.1521, "step": 5327 }, { "epoch": 0.2246347787591964, "grad_norm": 0.6486107707023621, "learning_rate": 3.7429735806632946e-05, "loss": 0.7747, "step": 5328 }, { "epoch": 0.22467693994139595, "grad_norm": 0.4717431962490082, "learning_rate": 3.7436762225969644e-05, "loss": 0.1578, "step": 5329 }, { "epoch": 0.2247191011235955, "grad_norm": 0.7170950770378113, "learning_rate": 3.7443788645306355e-05, "loss": 0.827, "step": 5330 }, { "epoch": 0.22476126230579505, "grad_norm": 0.5994043946266174, "learning_rate": 3.745081506464306e-05, "loss": 0.1749, "step": 5331 }, { "epoch": 0.2248034234879946, "grad_norm": 0.48614174127578735, "learning_rate": 3.7457841483979764e-05, "loss": 0.1876, "step": 5332 }, { "epoch": 0.22484558467019417, "grad_norm": 0.5253828763961792, "learning_rate": 3.746486790331647e-05, "loss": 0.199, "step": 5333 }, { "epoch": 0.2248877458523937, "grad_norm": 0.5652293562889099, "learning_rate": 3.747189432265318e-05, "loss": 0.1974, "step": 5334 }, { "epoch": 0.22492990703459326, "grad_norm": 0.8416962027549744, "learning_rate": 3.7478920741989884e-05, "loss": 1.1763, "step": 5335 }, { "epoch": 0.2249720682167928, "grad_norm": 0.9287106394767761, "learning_rate": 3.748594716132659e-05, "loss": 0.816, "step": 5336 }, { "epoch": 0.22501422939899235, "grad_norm": 1.7027779817581177, "learning_rate": 3.749297358066329e-05, "loss": 0.939, "step": 5337 }, { "epoch": 0.2250563905811919, "grad_norm": 0.6774755120277405, "learning_rate": 3.7500000000000003e-05, "loss": 0.3146, "step": 5338 }, { "epoch": 0.22509855176339144, "grad_norm": 0.6768981218338013, "learning_rate": 3.750702641933671e-05, "loss": 0.3401, "step": 5339 }, { "epoch": 0.225140712945591, "grad_norm": 0.5331408381462097, "learning_rate": 3.751405283867341e-05, "loss": 0.1623, "step": 5340 }, { "epoch": 0.22518287412779053, "grad_norm": 0.4480690658092499, "learning_rate": 3.7521079258010123e-05, "loss": 0.1991, "step": 5341 }, { "epoch": 0.22522503530999008, "grad_norm": 0.7139310836791992, "learning_rate": 3.752810567734683e-05, "loss": 0.3302, "step": 5342 }, { "epoch": 0.22526719649218965, "grad_norm": 0.8783630728721619, "learning_rate": 3.753513209668353e-05, "loss": 0.4001, "step": 5343 }, { "epoch": 0.2253093576743892, "grad_norm": 0.8478385806083679, "learning_rate": 3.754215851602024e-05, "loss": 0.2142, "step": 5344 }, { "epoch": 0.22535151885658875, "grad_norm": 0.9222989678382874, "learning_rate": 3.754918493535695e-05, "loss": 0.9929, "step": 5345 }, { "epoch": 0.2253936800387883, "grad_norm": 0.4032059907913208, "learning_rate": 3.755621135469365e-05, "loss": 0.1695, "step": 5346 }, { "epoch": 0.22543584122098784, "grad_norm": 0.6388689279556274, "learning_rate": 3.7563237774030357e-05, "loss": 0.1525, "step": 5347 }, { "epoch": 0.22547800240318738, "grad_norm": 0.44287678599357605, "learning_rate": 3.757026419336706e-05, "loss": 0.1397, "step": 5348 }, { "epoch": 0.22552016358538693, "grad_norm": 0.8555375933647156, "learning_rate": 3.757729061270377e-05, "loss": 1.1437, "step": 5349 }, { "epoch": 0.22556232476758648, "grad_norm": 0.3653925955295563, "learning_rate": 3.758431703204047e-05, "loss": 0.1565, "step": 5350 }, { "epoch": 0.22560448594978602, "grad_norm": 0.6835444569587708, "learning_rate": 3.759134345137718e-05, "loss": 0.5818, "step": 5351 }, { "epoch": 0.22564664713198557, "grad_norm": 0.5000010132789612, "learning_rate": 3.7598369870713885e-05, "loss": 0.1407, "step": 5352 }, { "epoch": 0.22568880831418514, "grad_norm": 0.4886903464794159, "learning_rate": 3.760539629005059e-05, "loss": 0.1704, "step": 5353 }, { "epoch": 0.2257309694963847, "grad_norm": 0.4736570715904236, "learning_rate": 3.7612422709387294e-05, "loss": 0.1602, "step": 5354 }, { "epoch": 0.22577313067858423, "grad_norm": 1.8745254278182983, "learning_rate": 3.7619449128724005e-05, "loss": 0.4642, "step": 5355 }, { "epoch": 0.22581529186078378, "grad_norm": 0.552089273929596, "learning_rate": 3.762647554806071e-05, "loss": 0.2782, "step": 5356 }, { "epoch": 0.22585745304298332, "grad_norm": 0.5784171223640442, "learning_rate": 3.7633501967397414e-05, "loss": 0.4949, "step": 5357 }, { "epoch": 0.22589961422518287, "grad_norm": 0.6151466965675354, "learning_rate": 3.764052838673412e-05, "loss": 0.2181, "step": 5358 }, { "epoch": 0.22594177540738242, "grad_norm": 0.37867534160614014, "learning_rate": 3.764755480607083e-05, "loss": 0.1145, "step": 5359 }, { "epoch": 0.22598393658958196, "grad_norm": 0.5514224171638489, "learning_rate": 3.7654581225407534e-05, "loss": 0.2365, "step": 5360 }, { "epoch": 0.2260260977717815, "grad_norm": 0.5561191439628601, "learning_rate": 3.766160764474424e-05, "loss": 0.1252, "step": 5361 }, { "epoch": 0.22606825895398108, "grad_norm": 0.4334237277507782, "learning_rate": 3.766863406408095e-05, "loss": 0.1725, "step": 5362 }, { "epoch": 0.22611042013618063, "grad_norm": 0.7127219438552856, "learning_rate": 3.7675660483417654e-05, "loss": 0.1519, "step": 5363 }, { "epoch": 0.22615258131838017, "grad_norm": 1.7756654024124146, "learning_rate": 3.768268690275436e-05, "loss": 0.6513, "step": 5364 }, { "epoch": 0.22619474250057972, "grad_norm": 0.5812070965766907, "learning_rate": 3.768971332209106e-05, "loss": 0.2329, "step": 5365 }, { "epoch": 0.22623690368277927, "grad_norm": 0.5930289030075073, "learning_rate": 3.7696739741427774e-05, "loss": 0.2126, "step": 5366 }, { "epoch": 0.2262790648649788, "grad_norm": 0.9277421236038208, "learning_rate": 3.770376616076448e-05, "loss": 1.0629, "step": 5367 }, { "epoch": 0.22632122604717836, "grad_norm": 0.6116136312484741, "learning_rate": 3.771079258010118e-05, "loss": 0.5723, "step": 5368 }, { "epoch": 0.2263633872293779, "grad_norm": 0.48773348331451416, "learning_rate": 3.771781899943789e-05, "loss": 0.1517, "step": 5369 }, { "epoch": 0.22640554841157745, "grad_norm": 0.3334721624851227, "learning_rate": 3.77248454187746e-05, "loss": 0.1137, "step": 5370 }, { "epoch": 0.226447709593777, "grad_norm": 0.7132471799850464, "learning_rate": 3.7731871838111296e-05, "loss": 0.2062, "step": 5371 }, { "epoch": 0.22648987077597657, "grad_norm": 0.397170752286911, "learning_rate": 3.773889825744801e-05, "loss": 0.1953, "step": 5372 }, { "epoch": 0.22653203195817612, "grad_norm": 0.8623019456863403, "learning_rate": 3.774592467678471e-05, "loss": 0.2666, "step": 5373 }, { "epoch": 0.22657419314037566, "grad_norm": 0.4866829812526703, "learning_rate": 3.775295109612142e-05, "loss": 0.2295, "step": 5374 }, { "epoch": 0.2266163543225752, "grad_norm": 0.4535565674304962, "learning_rate": 3.775997751545812e-05, "loss": 0.2019, "step": 5375 }, { "epoch": 0.22665851550477475, "grad_norm": 0.6701686382293701, "learning_rate": 3.776700393479483e-05, "loss": 0.4399, "step": 5376 }, { "epoch": 0.2267006766869743, "grad_norm": 0.6119763255119324, "learning_rate": 3.7774030354131536e-05, "loss": 0.1889, "step": 5377 }, { "epoch": 0.22674283786917385, "grad_norm": 0.6507869362831116, "learning_rate": 3.778105677346824e-05, "loss": 0.1834, "step": 5378 }, { "epoch": 0.2267849990513734, "grad_norm": 0.6350282430648804, "learning_rate": 3.7788083192804945e-05, "loss": 0.5613, "step": 5379 }, { "epoch": 0.22682716023357294, "grad_norm": 0.9336341023445129, "learning_rate": 3.7795109612141656e-05, "loss": 1.1535, "step": 5380 }, { "epoch": 0.22686932141577248, "grad_norm": 0.6430904865264893, "learning_rate": 3.780213603147836e-05, "loss": 0.5201, "step": 5381 }, { "epoch": 0.22691148259797206, "grad_norm": 0.7000746130943298, "learning_rate": 3.7809162450815065e-05, "loss": 0.2945, "step": 5382 }, { "epoch": 0.2269536437801716, "grad_norm": 0.6186385154724121, "learning_rate": 3.781618887015177e-05, "loss": 0.1971, "step": 5383 }, { "epoch": 0.22699580496237115, "grad_norm": 0.6653255820274353, "learning_rate": 3.782321528948848e-05, "loss": 0.5152, "step": 5384 }, { "epoch": 0.2270379661445707, "grad_norm": 0.8560287356376648, "learning_rate": 3.7830241708825185e-05, "loss": 0.6073, "step": 5385 }, { "epoch": 0.22708012732677024, "grad_norm": 1.3502302169799805, "learning_rate": 3.783726812816189e-05, "loss": 0.9854, "step": 5386 }, { "epoch": 0.2271222885089698, "grad_norm": 0.5898434519767761, "learning_rate": 3.78442945474986e-05, "loss": 0.2139, "step": 5387 }, { "epoch": 0.22716444969116933, "grad_norm": 0.6739141941070557, "learning_rate": 3.7851320966835305e-05, "loss": 0.4179, "step": 5388 }, { "epoch": 0.22720661087336888, "grad_norm": 0.684119462966919, "learning_rate": 3.785834738617201e-05, "loss": 0.7262, "step": 5389 }, { "epoch": 0.22724877205556843, "grad_norm": 0.3374174237251282, "learning_rate": 3.786537380550871e-05, "loss": 0.115, "step": 5390 }, { "epoch": 0.227290933237768, "grad_norm": 0.7210511565208435, "learning_rate": 3.7872400224845425e-05, "loss": 0.1604, "step": 5391 }, { "epoch": 0.22733309441996755, "grad_norm": 0.7276901006698608, "learning_rate": 3.787942664418212e-05, "loss": 0.594, "step": 5392 }, { "epoch": 0.2273752556021671, "grad_norm": 1.3699702024459839, "learning_rate": 3.788645306351883e-05, "loss": 1.0401, "step": 5393 }, { "epoch": 0.22741741678436664, "grad_norm": 0.40715834498405457, "learning_rate": 3.789347948285554e-05, "loss": 0.1825, "step": 5394 }, { "epoch": 0.22745957796656618, "grad_norm": 0.29377275705337524, "learning_rate": 3.790050590219225e-05, "loss": 0.1175, "step": 5395 }, { "epoch": 0.22750173914876573, "grad_norm": 0.36359843611717224, "learning_rate": 3.7907532321528947e-05, "loss": 0.1306, "step": 5396 }, { "epoch": 0.22754390033096528, "grad_norm": 0.5007559061050415, "learning_rate": 3.791455874086566e-05, "loss": 0.155, "step": 5397 }, { "epoch": 0.22758606151316482, "grad_norm": 0.40349695086479187, "learning_rate": 3.792158516020236e-05, "loss": 0.1491, "step": 5398 }, { "epoch": 0.22762822269536437, "grad_norm": 1.135973572731018, "learning_rate": 3.7928611579539066e-05, "loss": 1.1352, "step": 5399 }, { "epoch": 0.22767038387756391, "grad_norm": 0.8551744818687439, "learning_rate": 3.793563799887577e-05, "loss": 0.238, "step": 5400 }, { "epoch": 0.2277125450597635, "grad_norm": 1.653762936592102, "learning_rate": 3.794266441821248e-05, "loss": 1.3712, "step": 5401 }, { "epoch": 0.22775470624196303, "grad_norm": 0.3480006158351898, "learning_rate": 3.7949690837549186e-05, "loss": 0.0823, "step": 5402 }, { "epoch": 0.22779686742416258, "grad_norm": 0.733532190322876, "learning_rate": 3.795671725688589e-05, "loss": 0.2015, "step": 5403 }, { "epoch": 0.22783902860636213, "grad_norm": 1.2521023750305176, "learning_rate": 3.7963743676222595e-05, "loss": 1.026, "step": 5404 }, { "epoch": 0.22788118978856167, "grad_norm": 0.7698987126350403, "learning_rate": 3.7970770095559306e-05, "loss": 0.2656, "step": 5405 }, { "epoch": 0.22792335097076122, "grad_norm": 0.5109981298446655, "learning_rate": 3.797779651489601e-05, "loss": 0.5226, "step": 5406 }, { "epoch": 0.22796551215296076, "grad_norm": 0.531485378742218, "learning_rate": 3.7984822934232715e-05, "loss": 0.3054, "step": 5407 }, { "epoch": 0.2280076733351603, "grad_norm": 0.5663291215896606, "learning_rate": 3.7991849353569426e-05, "loss": 0.6289, "step": 5408 }, { "epoch": 0.22804983451735986, "grad_norm": 0.4880531132221222, "learning_rate": 3.799887577290613e-05, "loss": 0.224, "step": 5409 }, { "epoch": 0.2280919956995594, "grad_norm": 0.8621692657470703, "learning_rate": 3.8005902192242835e-05, "loss": 0.4165, "step": 5410 }, { "epoch": 0.22813415688175898, "grad_norm": 0.6040752530097961, "learning_rate": 3.801292861157954e-05, "loss": 0.2401, "step": 5411 }, { "epoch": 0.22817631806395852, "grad_norm": 0.33744779229164124, "learning_rate": 3.801995503091625e-05, "loss": 0.1235, "step": 5412 }, { "epoch": 0.22821847924615807, "grad_norm": 0.7339324355125427, "learning_rate": 3.8026981450252955e-05, "loss": 0.275, "step": 5413 }, { "epoch": 0.22826064042835761, "grad_norm": 0.6293808221817017, "learning_rate": 3.803400786958966e-05, "loss": 0.1245, "step": 5414 }, { "epoch": 0.22830280161055716, "grad_norm": 0.7109682559967041, "learning_rate": 3.8041034288926364e-05, "loss": 0.1979, "step": 5415 }, { "epoch": 0.2283449627927567, "grad_norm": 0.7777795791625977, "learning_rate": 3.8048060708263075e-05, "loss": 0.2998, "step": 5416 }, { "epoch": 0.22838712397495625, "grad_norm": 0.7120689749717712, "learning_rate": 3.805508712759977e-05, "loss": 0.6851, "step": 5417 }, { "epoch": 0.2284292851571558, "grad_norm": 2.202706813812256, "learning_rate": 3.8062113546936484e-05, "loss": 0.8038, "step": 5418 }, { "epoch": 0.22847144633935534, "grad_norm": 0.6307893395423889, "learning_rate": 3.806913996627319e-05, "loss": 0.6293, "step": 5419 }, { "epoch": 0.22851360752155492, "grad_norm": 0.6360272765159607, "learning_rate": 3.807616638560989e-05, "loss": 0.2868, "step": 5420 }, { "epoch": 0.22855576870375446, "grad_norm": 0.9024614691734314, "learning_rate": 3.80831928049466e-05, "loss": 0.4587, "step": 5421 }, { "epoch": 0.228597929885954, "grad_norm": 0.8990575075149536, "learning_rate": 3.809021922428331e-05, "loss": 0.6779, "step": 5422 }, { "epoch": 0.22864009106815356, "grad_norm": 0.8696261048316956, "learning_rate": 3.809724564362001e-05, "loss": 0.2488, "step": 5423 }, { "epoch": 0.2286822522503531, "grad_norm": 1.4247032403945923, "learning_rate": 3.810427206295672e-05, "loss": 0.5587, "step": 5424 }, { "epoch": 0.22872441343255265, "grad_norm": 0.45671093463897705, "learning_rate": 3.811129848229342e-05, "loss": 0.2002, "step": 5425 }, { "epoch": 0.2287665746147522, "grad_norm": 0.3170555830001831, "learning_rate": 3.811832490163013e-05, "loss": 0.1246, "step": 5426 }, { "epoch": 0.22880873579695174, "grad_norm": 0.3701683580875397, "learning_rate": 3.812535132096684e-05, "loss": 0.0904, "step": 5427 }, { "epoch": 0.22885089697915129, "grad_norm": 0.7648754715919495, "learning_rate": 3.813237774030354e-05, "loss": 0.6293, "step": 5428 }, { "epoch": 0.22889305816135083, "grad_norm": 1.3812857866287231, "learning_rate": 3.813940415964025e-05, "loss": 0.9337, "step": 5429 }, { "epoch": 0.2289352193435504, "grad_norm": 0.7547027468681335, "learning_rate": 3.814643057897696e-05, "loss": 0.2942, "step": 5430 }, { "epoch": 0.22897738052574995, "grad_norm": 0.6032198071479797, "learning_rate": 3.815345699831366e-05, "loss": 0.2278, "step": 5431 }, { "epoch": 0.2290195417079495, "grad_norm": 0.5371900200843811, "learning_rate": 3.8160483417650366e-05, "loss": 0.1492, "step": 5432 }, { "epoch": 0.22906170289014904, "grad_norm": 0.46235883235931396, "learning_rate": 3.816750983698708e-05, "loss": 0.2075, "step": 5433 }, { "epoch": 0.2291038640723486, "grad_norm": 0.3518291413784027, "learning_rate": 3.817453625632378e-05, "loss": 0.152, "step": 5434 }, { "epoch": 0.22914602525454814, "grad_norm": 0.8946235775947571, "learning_rate": 3.8181562675660486e-05, "loss": 1.0366, "step": 5435 }, { "epoch": 0.22918818643674768, "grad_norm": 0.5750179290771484, "learning_rate": 3.818858909499719e-05, "loss": 0.1978, "step": 5436 }, { "epoch": 0.22923034761894723, "grad_norm": 0.6500300765037537, "learning_rate": 3.81956155143339e-05, "loss": 0.6679, "step": 5437 }, { "epoch": 0.22927250880114677, "grad_norm": 0.7570860385894775, "learning_rate": 3.82026419336706e-05, "loss": 0.3373, "step": 5438 }, { "epoch": 0.22931466998334632, "grad_norm": 0.4830571413040161, "learning_rate": 3.820966835300731e-05, "loss": 0.2251, "step": 5439 }, { "epoch": 0.2293568311655459, "grad_norm": 0.7418341040611267, "learning_rate": 3.8216694772344014e-05, "loss": 0.3997, "step": 5440 }, { "epoch": 0.22939899234774544, "grad_norm": 0.5266831517219543, "learning_rate": 3.8223721191680726e-05, "loss": 0.2651, "step": 5441 }, { "epoch": 0.22944115352994499, "grad_norm": 0.8691344261169434, "learning_rate": 3.823074761101742e-05, "loss": 0.1985, "step": 5442 }, { "epoch": 0.22948331471214453, "grad_norm": 0.8917213082313538, "learning_rate": 3.8237774030354134e-05, "loss": 0.9512, "step": 5443 }, { "epoch": 0.22952547589434408, "grad_norm": 0.7234621047973633, "learning_rate": 3.824480044969084e-05, "loss": 0.3357, "step": 5444 }, { "epoch": 0.22956763707654362, "grad_norm": 0.829099178314209, "learning_rate": 3.825182686902754e-05, "loss": 0.4513, "step": 5445 }, { "epoch": 0.22960979825874317, "grad_norm": 0.5974171757698059, "learning_rate": 3.825885328836425e-05, "loss": 0.1948, "step": 5446 }, { "epoch": 0.22965195944094272, "grad_norm": 0.6280362606048584, "learning_rate": 3.826587970770096e-05, "loss": 0.1755, "step": 5447 }, { "epoch": 0.22969412062314226, "grad_norm": 1.1523330211639404, "learning_rate": 3.827290612703766e-05, "loss": 0.7373, "step": 5448 }, { "epoch": 0.22973628180534184, "grad_norm": 0.827926516532898, "learning_rate": 3.827993254637437e-05, "loss": 1.0973, "step": 5449 }, { "epoch": 0.22977844298754138, "grad_norm": 0.446770042181015, "learning_rate": 3.828695896571107e-05, "loss": 0.2305, "step": 5450 }, { "epoch": 0.22982060416974093, "grad_norm": 0.6130131483078003, "learning_rate": 3.829398538504778e-05, "loss": 0.1027, "step": 5451 }, { "epoch": 0.22986276535194047, "grad_norm": 0.5243578553199768, "learning_rate": 3.830101180438449e-05, "loss": 0.2915, "step": 5452 }, { "epoch": 0.22990492653414002, "grad_norm": 0.558263897895813, "learning_rate": 3.830803822372119e-05, "loss": 0.4705, "step": 5453 }, { "epoch": 0.22994708771633957, "grad_norm": 0.6811521053314209, "learning_rate": 3.83150646430579e-05, "loss": 0.5996, "step": 5454 }, { "epoch": 0.2299892488985391, "grad_norm": 0.42687666416168213, "learning_rate": 3.832209106239461e-05, "loss": 0.1313, "step": 5455 }, { "epoch": 0.23003141008073866, "grad_norm": 0.6357313990592957, "learning_rate": 3.832911748173131e-05, "loss": 0.4807, "step": 5456 }, { "epoch": 0.2300735712629382, "grad_norm": 0.6284554600715637, "learning_rate": 3.8336143901068016e-05, "loss": 0.1942, "step": 5457 }, { "epoch": 0.23011573244513775, "grad_norm": 0.8732650279998779, "learning_rate": 3.834317032040473e-05, "loss": 0.9177, "step": 5458 }, { "epoch": 0.23015789362733732, "grad_norm": 0.3564365804195404, "learning_rate": 3.835019673974143e-05, "loss": 0.1393, "step": 5459 }, { "epoch": 0.23020005480953687, "grad_norm": 0.9707518219947815, "learning_rate": 3.8357223159078136e-05, "loss": 0.8384, "step": 5460 }, { "epoch": 0.23024221599173642, "grad_norm": 0.6190702319145203, "learning_rate": 3.836424957841484e-05, "loss": 0.1791, "step": 5461 }, { "epoch": 0.23028437717393596, "grad_norm": 0.9517836570739746, "learning_rate": 3.837127599775155e-05, "loss": 0.303, "step": 5462 }, { "epoch": 0.2303265383561355, "grad_norm": 0.4447805881500244, "learning_rate": 3.837830241708825e-05, "loss": 0.1154, "step": 5463 }, { "epoch": 0.23036869953833505, "grad_norm": 0.35049599409103394, "learning_rate": 3.838532883642496e-05, "loss": 0.1137, "step": 5464 }, { "epoch": 0.2304108607205346, "grad_norm": 0.6833480000495911, "learning_rate": 3.8392355255761665e-05, "loss": 0.2101, "step": 5465 }, { "epoch": 0.23045302190273415, "grad_norm": 0.5625246167182922, "learning_rate": 3.839938167509837e-05, "loss": 0.1457, "step": 5466 }, { "epoch": 0.2304951830849337, "grad_norm": 0.6046240329742432, "learning_rate": 3.8406408094435074e-05, "loss": 0.5829, "step": 5467 }, { "epoch": 0.23053734426713324, "grad_norm": 0.41983285546302795, "learning_rate": 3.8413434513771785e-05, "loss": 0.1701, "step": 5468 }, { "epoch": 0.2305795054493328, "grad_norm": 0.41205650568008423, "learning_rate": 3.842046093310849e-05, "loss": 0.2049, "step": 5469 }, { "epoch": 0.23062166663153236, "grad_norm": 1.5485934019088745, "learning_rate": 3.8427487352445194e-05, "loss": 1.1218, "step": 5470 }, { "epoch": 0.2306638278137319, "grad_norm": 0.38185033202171326, "learning_rate": 3.84345137717819e-05, "loss": 0.1152, "step": 5471 }, { "epoch": 0.23070598899593145, "grad_norm": 0.67266446352005, "learning_rate": 3.844154019111861e-05, "loss": 0.3643, "step": 5472 }, { "epoch": 0.230748150178131, "grad_norm": 0.5517199039459229, "learning_rate": 3.8448566610455314e-05, "loss": 0.2846, "step": 5473 }, { "epoch": 0.23079031136033054, "grad_norm": 1.5345046520233154, "learning_rate": 3.845559302979202e-05, "loss": 0.6034, "step": 5474 }, { "epoch": 0.2308324725425301, "grad_norm": 1.4232187271118164, "learning_rate": 3.846261944912873e-05, "loss": 1.1709, "step": 5475 }, { "epoch": 0.23087463372472963, "grad_norm": 0.5978676080703735, "learning_rate": 3.8469645868465434e-05, "loss": 0.2433, "step": 5476 }, { "epoch": 0.23091679490692918, "grad_norm": 0.6515754461288452, "learning_rate": 3.847667228780214e-05, "loss": 0.3846, "step": 5477 }, { "epoch": 0.23095895608912875, "grad_norm": 1.2113720178604126, "learning_rate": 3.848369870713884e-05, "loss": 0.3873, "step": 5478 }, { "epoch": 0.2310011172713283, "grad_norm": 1.2427958250045776, "learning_rate": 3.8490725126475554e-05, "loss": 1.0382, "step": 5479 }, { "epoch": 0.23104327845352784, "grad_norm": 0.586348831653595, "learning_rate": 3.849775154581226e-05, "loss": 0.1233, "step": 5480 }, { "epoch": 0.2310854396357274, "grad_norm": 0.9925923943519592, "learning_rate": 3.850477796514896e-05, "loss": 0.7479, "step": 5481 }, { "epoch": 0.23112760081792694, "grad_norm": 1.005582332611084, "learning_rate": 3.851180438448567e-05, "loss": 1.1607, "step": 5482 }, { "epoch": 0.23116976200012648, "grad_norm": 0.9464192390441895, "learning_rate": 3.851883080382238e-05, "loss": 0.3043, "step": 5483 }, { "epoch": 0.23121192318232603, "grad_norm": 0.516129732131958, "learning_rate": 3.8525857223159076e-05, "loss": 0.256, "step": 5484 }, { "epoch": 0.23125408436452558, "grad_norm": 1.2307682037353516, "learning_rate": 3.853288364249579e-05, "loss": 1.0816, "step": 5485 }, { "epoch": 0.23129624554672512, "grad_norm": 0.3337603807449341, "learning_rate": 3.853991006183249e-05, "loss": 0.1164, "step": 5486 }, { "epoch": 0.23133840672892467, "grad_norm": 0.6570212244987488, "learning_rate": 3.85469364811692e-05, "loss": 0.1945, "step": 5487 }, { "epoch": 0.23138056791112424, "grad_norm": 0.503570020198822, "learning_rate": 3.85539629005059e-05, "loss": 0.1725, "step": 5488 }, { "epoch": 0.2314227290933238, "grad_norm": 0.6105921864509583, "learning_rate": 3.856098931984261e-05, "loss": 0.3048, "step": 5489 }, { "epoch": 0.23146489027552333, "grad_norm": 0.5227676033973694, "learning_rate": 3.8568015739179316e-05, "loss": 0.2346, "step": 5490 }, { "epoch": 0.23150705145772288, "grad_norm": 0.8800788521766663, "learning_rate": 3.857504215851602e-05, "loss": 0.2117, "step": 5491 }, { "epoch": 0.23154921263992242, "grad_norm": 0.8460599780082703, "learning_rate": 3.8582068577852724e-05, "loss": 1.1082, "step": 5492 }, { "epoch": 0.23159137382212197, "grad_norm": 0.6879788637161255, "learning_rate": 3.8589094997189435e-05, "loss": 0.5717, "step": 5493 }, { "epoch": 0.23163353500432152, "grad_norm": 0.8234785199165344, "learning_rate": 3.859612141652614e-05, "loss": 0.1857, "step": 5494 }, { "epoch": 0.23167569618652106, "grad_norm": 0.7330378293991089, "learning_rate": 3.8603147835862844e-05, "loss": 0.1964, "step": 5495 }, { "epoch": 0.2317178573687206, "grad_norm": 0.6198638081550598, "learning_rate": 3.8610174255199555e-05, "loss": 0.2384, "step": 5496 }, { "epoch": 0.23176001855092015, "grad_norm": 0.5270299911499023, "learning_rate": 3.861720067453626e-05, "loss": 0.4145, "step": 5497 }, { "epoch": 0.23180217973311973, "grad_norm": 1.2256258726119995, "learning_rate": 3.8624227093872964e-05, "loss": 0.8924, "step": 5498 }, { "epoch": 0.23184434091531927, "grad_norm": 0.5659022331237793, "learning_rate": 3.863125351320967e-05, "loss": 0.331, "step": 5499 }, { "epoch": 0.23188650209751882, "grad_norm": 0.600576639175415, "learning_rate": 3.863827993254638e-05, "loss": 0.5912, "step": 5500 }, { "epoch": 0.23192866327971837, "grad_norm": 0.45926642417907715, "learning_rate": 3.8645306351883084e-05, "loss": 0.4987, "step": 5501 }, { "epoch": 0.2319708244619179, "grad_norm": 0.693645179271698, "learning_rate": 3.865233277121979e-05, "loss": 0.5652, "step": 5502 }, { "epoch": 0.23201298564411746, "grad_norm": 1.1028468608856201, "learning_rate": 3.865935919055649e-05, "loss": 1.1129, "step": 5503 }, { "epoch": 0.232055146826317, "grad_norm": 0.42857834696769714, "learning_rate": 3.8666385609893204e-05, "loss": 0.1177, "step": 5504 }, { "epoch": 0.23209730800851655, "grad_norm": 0.7083531618118286, "learning_rate": 3.86734120292299e-05, "loss": 0.7024, "step": 5505 }, { "epoch": 0.2321394691907161, "grad_norm": 0.7323026657104492, "learning_rate": 3.868043844856661e-05, "loss": 0.3374, "step": 5506 }, { "epoch": 0.23218163037291564, "grad_norm": 0.5745407938957214, "learning_rate": 3.868746486790332e-05, "loss": 0.2418, "step": 5507 }, { "epoch": 0.23222379155511522, "grad_norm": 0.4021243751049042, "learning_rate": 3.869449128724003e-05, "loss": 0.0976, "step": 5508 }, { "epoch": 0.23226595273731476, "grad_norm": 0.5823472738265991, "learning_rate": 3.8701517706576726e-05, "loss": 0.1389, "step": 5509 }, { "epoch": 0.2323081139195143, "grad_norm": 0.44686922430992126, "learning_rate": 3.870854412591344e-05, "loss": 0.1204, "step": 5510 }, { "epoch": 0.23235027510171385, "grad_norm": 0.640900731086731, "learning_rate": 3.871557054525014e-05, "loss": 0.1285, "step": 5511 }, { "epoch": 0.2323924362839134, "grad_norm": 1.6980284452438354, "learning_rate": 3.8722596964586846e-05, "loss": 0.8605, "step": 5512 }, { "epoch": 0.23243459746611295, "grad_norm": 0.4438343048095703, "learning_rate": 3.872962338392355e-05, "loss": 0.1302, "step": 5513 }, { "epoch": 0.2324767586483125, "grad_norm": 0.6148391962051392, "learning_rate": 3.873664980326026e-05, "loss": 0.2139, "step": 5514 }, { "epoch": 0.23251891983051204, "grad_norm": 1.044372320175171, "learning_rate": 3.8743676222596966e-05, "loss": 1.0512, "step": 5515 }, { "epoch": 0.23256108101271158, "grad_norm": 1.2904541492462158, "learning_rate": 3.875070264193367e-05, "loss": 0.4021, "step": 5516 }, { "epoch": 0.23260324219491116, "grad_norm": 0.6824262142181396, "learning_rate": 3.8757729061270375e-05, "loss": 0.1956, "step": 5517 }, { "epoch": 0.2326454033771107, "grad_norm": 1.1343233585357666, "learning_rate": 3.8764755480607086e-05, "loss": 0.4363, "step": 5518 }, { "epoch": 0.23268756455931025, "grad_norm": 0.8380451798439026, "learning_rate": 3.877178189994379e-05, "loss": 0.4135, "step": 5519 }, { "epoch": 0.2327297257415098, "grad_norm": 0.5222038626670837, "learning_rate": 3.8778808319280495e-05, "loss": 0.4752, "step": 5520 }, { "epoch": 0.23277188692370934, "grad_norm": 0.6745630502700806, "learning_rate": 3.8785834738617206e-05, "loss": 0.5156, "step": 5521 }, { "epoch": 0.2328140481059089, "grad_norm": 0.5916473865509033, "learning_rate": 3.879286115795391e-05, "loss": 0.1682, "step": 5522 }, { "epoch": 0.23285620928810843, "grad_norm": 0.8310933709144592, "learning_rate": 3.8799887577290615e-05, "loss": 0.269, "step": 5523 }, { "epoch": 0.23289837047030798, "grad_norm": 0.5950385332107544, "learning_rate": 3.880691399662732e-05, "loss": 0.6432, "step": 5524 }, { "epoch": 0.23294053165250753, "grad_norm": 0.44669345021247864, "learning_rate": 3.881394041596403e-05, "loss": 0.1253, "step": 5525 }, { "epoch": 0.23298269283470707, "grad_norm": 0.86366206407547, "learning_rate": 3.8820966835300735e-05, "loss": 0.8669, "step": 5526 }, { "epoch": 0.23302485401690665, "grad_norm": 0.6545134782791138, "learning_rate": 3.882799325463744e-05, "loss": 0.6146, "step": 5527 }, { "epoch": 0.2330670151991062, "grad_norm": 0.8004190325737, "learning_rate": 3.8835019673974144e-05, "loss": 0.636, "step": 5528 }, { "epoch": 0.23310917638130574, "grad_norm": 0.41786596179008484, "learning_rate": 3.8842046093310855e-05, "loss": 0.174, "step": 5529 }, { "epoch": 0.23315133756350528, "grad_norm": 1.0261484384536743, "learning_rate": 3.884907251264755e-05, "loss": 0.2026, "step": 5530 }, { "epoch": 0.23319349874570483, "grad_norm": 0.6286110281944275, "learning_rate": 3.8856098931984263e-05, "loss": 0.177, "step": 5531 }, { "epoch": 0.23323565992790438, "grad_norm": 0.7536410093307495, "learning_rate": 3.886312535132097e-05, "loss": 0.4831, "step": 5532 }, { "epoch": 0.23327782111010392, "grad_norm": 1.0148335695266724, "learning_rate": 3.887015177065768e-05, "loss": 0.4357, "step": 5533 }, { "epoch": 0.23331998229230347, "grad_norm": 0.7934555411338806, "learning_rate": 3.887717818999438e-05, "loss": 0.3152, "step": 5534 }, { "epoch": 0.23336214347450301, "grad_norm": 1.1654850244522095, "learning_rate": 3.888420460933109e-05, "loss": 0.4855, "step": 5535 }, { "epoch": 0.23340430465670256, "grad_norm": 1.2027024030685425, "learning_rate": 3.889123102866779e-05, "loss": 0.9722, "step": 5536 }, { "epoch": 0.23344646583890213, "grad_norm": 0.757015585899353, "learning_rate": 3.88982574480045e-05, "loss": 0.2337, "step": 5537 }, { "epoch": 0.23348862702110168, "grad_norm": 0.539321780204773, "learning_rate": 3.89052838673412e-05, "loss": 0.203, "step": 5538 }, { "epoch": 0.23353078820330123, "grad_norm": 0.8733016848564148, "learning_rate": 3.891231028667791e-05, "loss": 0.6504, "step": 5539 }, { "epoch": 0.23357294938550077, "grad_norm": 0.686288058757782, "learning_rate": 3.8919336706014617e-05, "loss": 0.3904, "step": 5540 }, { "epoch": 0.23361511056770032, "grad_norm": 0.8377251029014587, "learning_rate": 3.892636312535132e-05, "loss": 0.618, "step": 5541 }, { "epoch": 0.23365727174989986, "grad_norm": 0.5880982875823975, "learning_rate": 3.893338954468803e-05, "loss": 0.1501, "step": 5542 }, { "epoch": 0.2336994329320994, "grad_norm": 1.0219794511795044, "learning_rate": 3.8940415964024737e-05, "loss": 0.3143, "step": 5543 }, { "epoch": 0.23374159411429896, "grad_norm": 0.5797579884529114, "learning_rate": 3.894744238336144e-05, "loss": 0.4921, "step": 5544 }, { "epoch": 0.2337837552964985, "grad_norm": 0.5537455677986145, "learning_rate": 3.8954468802698145e-05, "loss": 0.2929, "step": 5545 }, { "epoch": 0.23382591647869808, "grad_norm": 0.5696252584457397, "learning_rate": 3.8961495222034856e-05, "loss": 0.2025, "step": 5546 }, { "epoch": 0.23386807766089762, "grad_norm": 0.7956192493438721, "learning_rate": 3.896852164137156e-05, "loss": 0.3998, "step": 5547 }, { "epoch": 0.23391023884309717, "grad_norm": 0.3618776500225067, "learning_rate": 3.8975548060708265e-05, "loss": 0.1259, "step": 5548 }, { "epoch": 0.23395240002529671, "grad_norm": 0.5657559633255005, "learning_rate": 3.898257448004497e-05, "loss": 0.1584, "step": 5549 }, { "epoch": 0.23399456120749626, "grad_norm": 0.665860652923584, "learning_rate": 3.898960089938168e-05, "loss": 0.4786, "step": 5550 }, { "epoch": 0.2340367223896958, "grad_norm": 0.6551241278648376, "learning_rate": 3.899662731871838e-05, "loss": 0.6063, "step": 5551 }, { "epoch": 0.23407888357189535, "grad_norm": 0.6713167428970337, "learning_rate": 3.900365373805509e-05, "loss": 0.5066, "step": 5552 }, { "epoch": 0.2341210447540949, "grad_norm": 0.4723493754863739, "learning_rate": 3.9010680157391794e-05, "loss": 0.1799, "step": 5553 }, { "epoch": 0.23416320593629444, "grad_norm": 1.1616804599761963, "learning_rate": 3.9017706576728505e-05, "loss": 1.2228, "step": 5554 }, { "epoch": 0.234205367118494, "grad_norm": 0.48867282271385193, "learning_rate": 3.90247329960652e-05, "loss": 0.3219, "step": 5555 }, { "epoch": 0.23424752830069356, "grad_norm": 0.576662003993988, "learning_rate": 3.9031759415401914e-05, "loss": 0.2896, "step": 5556 }, { "epoch": 0.2342896894828931, "grad_norm": 0.7467201948165894, "learning_rate": 3.903878583473862e-05, "loss": 0.412, "step": 5557 }, { "epoch": 0.23433185066509266, "grad_norm": 1.2516899108886719, "learning_rate": 3.904581225407532e-05, "loss": 0.2958, "step": 5558 }, { "epoch": 0.2343740118472922, "grad_norm": 2.6091887950897217, "learning_rate": 3.905283867341203e-05, "loss": 0.995, "step": 5559 }, { "epoch": 0.23441617302949175, "grad_norm": 0.5175874829292297, "learning_rate": 3.905986509274874e-05, "loss": 0.1875, "step": 5560 }, { "epoch": 0.2344583342116913, "grad_norm": 2.039403200149536, "learning_rate": 3.906689151208544e-05, "loss": 0.4266, "step": 5561 }, { "epoch": 0.23450049539389084, "grad_norm": 0.6515030264854431, "learning_rate": 3.907391793142215e-05, "loss": 0.0947, "step": 5562 }, { "epoch": 0.23454265657609039, "grad_norm": 0.740134596824646, "learning_rate": 3.908094435075885e-05, "loss": 0.5881, "step": 5563 }, { "epoch": 0.23458481775828993, "grad_norm": 0.9690190553665161, "learning_rate": 3.908797077009556e-05, "loss": 0.8951, "step": 5564 }, { "epoch": 0.23462697894048948, "grad_norm": 0.7091088891029358, "learning_rate": 3.909499718943227e-05, "loss": 0.6459, "step": 5565 }, { "epoch": 0.23466914012268905, "grad_norm": 1.2551053762435913, "learning_rate": 3.910202360876897e-05, "loss": 0.2954, "step": 5566 }, { "epoch": 0.2347113013048886, "grad_norm": 0.4546331763267517, "learning_rate": 3.910905002810568e-05, "loss": 0.1268, "step": 5567 }, { "epoch": 0.23475346248708814, "grad_norm": 0.7451272010803223, "learning_rate": 3.911607644744239e-05, "loss": 0.2675, "step": 5568 }, { "epoch": 0.2347956236692877, "grad_norm": 0.5424447655677795, "learning_rate": 3.912310286677909e-05, "loss": 0.1042, "step": 5569 }, { "epoch": 0.23483778485148724, "grad_norm": 0.7082661986351013, "learning_rate": 3.9130129286115796e-05, "loss": 0.2447, "step": 5570 }, { "epoch": 0.23487994603368678, "grad_norm": 0.5371587872505188, "learning_rate": 3.913715570545251e-05, "loss": 0.1842, "step": 5571 }, { "epoch": 0.23492210721588633, "grad_norm": 0.5918092727661133, "learning_rate": 3.914418212478921e-05, "loss": 0.4772, "step": 5572 }, { "epoch": 0.23496426839808587, "grad_norm": 0.38992172479629517, "learning_rate": 3.9151208544125916e-05, "loss": 0.1253, "step": 5573 }, { "epoch": 0.23500642958028542, "grad_norm": 0.4931984841823578, "learning_rate": 3.915823496346262e-05, "loss": 0.2261, "step": 5574 }, { "epoch": 0.235048590762485, "grad_norm": 0.6614482402801514, "learning_rate": 3.916526138279933e-05, "loss": 0.5264, "step": 5575 }, { "epoch": 0.23509075194468454, "grad_norm": 0.5877997279167175, "learning_rate": 3.917228780213603e-05, "loss": 0.4052, "step": 5576 }, { "epoch": 0.23513291312688409, "grad_norm": 0.8053550124168396, "learning_rate": 3.917931422147274e-05, "loss": 0.2156, "step": 5577 }, { "epoch": 0.23517507430908363, "grad_norm": 0.6406261920928955, "learning_rate": 3.9186340640809445e-05, "loss": 0.717, "step": 5578 }, { "epoch": 0.23521723549128318, "grad_norm": 0.5372385382652283, "learning_rate": 3.919336706014615e-05, "loss": 0.2664, "step": 5579 }, { "epoch": 0.23525939667348272, "grad_norm": 0.954778790473938, "learning_rate": 3.920039347948285e-05, "loss": 1.1077, "step": 5580 }, { "epoch": 0.23530155785568227, "grad_norm": 0.6150511503219604, "learning_rate": 3.9207419898819565e-05, "loss": 0.2997, "step": 5581 }, { "epoch": 0.23534371903788182, "grad_norm": 0.7445586323738098, "learning_rate": 3.921444631815627e-05, "loss": 0.2203, "step": 5582 }, { "epoch": 0.23538588022008136, "grad_norm": 0.6116443276405334, "learning_rate": 3.922147273749297e-05, "loss": 0.1682, "step": 5583 }, { "epoch": 0.2354280414022809, "grad_norm": 0.6853642463684082, "learning_rate": 3.922849915682968e-05, "loss": 0.29, "step": 5584 }, { "epoch": 0.23547020258448048, "grad_norm": 0.5473601222038269, "learning_rate": 3.923552557616639e-05, "loss": 0.3603, "step": 5585 }, { "epoch": 0.23551236376668003, "grad_norm": 0.3472830355167389, "learning_rate": 3.924255199550309e-05, "loss": 0.1191, "step": 5586 }, { "epoch": 0.23555452494887957, "grad_norm": 0.6488789916038513, "learning_rate": 3.92495784148398e-05, "loss": 0.5227, "step": 5587 }, { "epoch": 0.23559668613107912, "grad_norm": 0.6838273406028748, "learning_rate": 3.925660483417651e-05, "loss": 0.5623, "step": 5588 }, { "epoch": 0.23563884731327867, "grad_norm": 1.092755913734436, "learning_rate": 3.926363125351321e-05, "loss": 0.6293, "step": 5589 }, { "epoch": 0.2356810084954782, "grad_norm": 0.8779308199882507, "learning_rate": 3.927065767284992e-05, "loss": 0.3162, "step": 5590 }, { "epoch": 0.23572316967767776, "grad_norm": 0.6937261819839478, "learning_rate": 3.927768409218662e-05, "loss": 0.4217, "step": 5591 }, { "epoch": 0.2357653308598773, "grad_norm": 0.5482853055000305, "learning_rate": 3.928471051152333e-05, "loss": 0.5612, "step": 5592 }, { "epoch": 0.23580749204207685, "grad_norm": 0.8766998052597046, "learning_rate": 3.929173693086004e-05, "loss": 0.6594, "step": 5593 }, { "epoch": 0.2358496532242764, "grad_norm": 0.8564848303794861, "learning_rate": 3.929876335019674e-05, "loss": 0.9653, "step": 5594 }, { "epoch": 0.23589181440647597, "grad_norm": 0.6661531925201416, "learning_rate": 3.9305789769533446e-05, "loss": 0.5574, "step": 5595 }, { "epoch": 0.23593397558867552, "grad_norm": 0.6553901433944702, "learning_rate": 3.931281618887016e-05, "loss": 0.7578, "step": 5596 }, { "epoch": 0.23597613677087506, "grad_norm": 1.0377106666564941, "learning_rate": 3.9319842608206855e-05, "loss": 1.0898, "step": 5597 }, { "epoch": 0.2360182979530746, "grad_norm": 0.6221846342086792, "learning_rate": 3.9326869027543566e-05, "loss": 0.2272, "step": 5598 }, { "epoch": 0.23606045913527415, "grad_norm": 0.6091955900192261, "learning_rate": 3.933389544688027e-05, "loss": 0.2014, "step": 5599 }, { "epoch": 0.2361026203174737, "grad_norm": 0.5454517602920532, "learning_rate": 3.934092186621698e-05, "loss": 0.2236, "step": 5600 }, { "epoch": 0.23614478149967325, "grad_norm": 0.7551701068878174, "learning_rate": 3.934794828555368e-05, "loss": 0.2149, "step": 5601 }, { "epoch": 0.2361869426818728, "grad_norm": 0.8289546370506287, "learning_rate": 3.935497470489039e-05, "loss": 0.3402, "step": 5602 }, { "epoch": 0.23622910386407234, "grad_norm": 0.764427900314331, "learning_rate": 3.9362001124227095e-05, "loss": 0.3029, "step": 5603 }, { "epoch": 0.2362712650462719, "grad_norm": 0.5870422720909119, "learning_rate": 3.93690275435638e-05, "loss": 0.1519, "step": 5604 }, { "epoch": 0.23631342622847146, "grad_norm": 0.6174147129058838, "learning_rate": 3.9376053962900504e-05, "loss": 0.678, "step": 5605 }, { "epoch": 0.236355587410671, "grad_norm": 0.9730868935585022, "learning_rate": 3.9383080382237215e-05, "loss": 0.1922, "step": 5606 }, { "epoch": 0.23639774859287055, "grad_norm": 1.1570827960968018, "learning_rate": 3.939010680157392e-05, "loss": 1.0996, "step": 5607 }, { "epoch": 0.2364399097750701, "grad_norm": 0.459773987531662, "learning_rate": 3.9397133220910624e-05, "loss": 0.0908, "step": 5608 }, { "epoch": 0.23648207095726964, "grad_norm": 0.7430912256240845, "learning_rate": 3.9404159640247335e-05, "loss": 0.5349, "step": 5609 }, { "epoch": 0.2365242321394692, "grad_norm": 0.5978959798812866, "learning_rate": 3.941118605958404e-05, "loss": 0.2559, "step": 5610 }, { "epoch": 0.23656639332166873, "grad_norm": 0.3348560035228729, "learning_rate": 3.9418212478920744e-05, "loss": 0.0985, "step": 5611 }, { "epoch": 0.23660855450386828, "grad_norm": 0.6153833270072937, "learning_rate": 3.942523889825745e-05, "loss": 0.6725, "step": 5612 }, { "epoch": 0.23665071568606783, "grad_norm": 1.1186468601226807, "learning_rate": 3.943226531759416e-05, "loss": 0.6139, "step": 5613 }, { "epoch": 0.2366928768682674, "grad_norm": 0.7173687219619751, "learning_rate": 3.9439291736930864e-05, "loss": 0.6503, "step": 5614 }, { "epoch": 0.23673503805046694, "grad_norm": 0.7119860053062439, "learning_rate": 3.944631815626757e-05, "loss": 0.6654, "step": 5615 }, { "epoch": 0.2367771992326665, "grad_norm": 0.7381113767623901, "learning_rate": 3.945334457560427e-05, "loss": 0.2139, "step": 5616 }, { "epoch": 0.23681936041486604, "grad_norm": 0.5621922016143799, "learning_rate": 3.9460370994940984e-05, "loss": 0.2297, "step": 5617 }, { "epoch": 0.23686152159706558, "grad_norm": 0.5446730852127075, "learning_rate": 3.946739741427769e-05, "loss": 0.4381, "step": 5618 }, { "epoch": 0.23690368277926513, "grad_norm": 0.8563411831855774, "learning_rate": 3.947442383361439e-05, "loss": 1.031, "step": 5619 }, { "epoch": 0.23694584396146467, "grad_norm": 0.43891340494155884, "learning_rate": 3.94814502529511e-05, "loss": 0.1055, "step": 5620 }, { "epoch": 0.23698800514366422, "grad_norm": 1.9116781949996948, "learning_rate": 3.948847667228781e-05, "loss": 0.5293, "step": 5621 }, { "epoch": 0.23703016632586377, "grad_norm": 0.5194368362426758, "learning_rate": 3.9495503091624506e-05, "loss": 0.1066, "step": 5622 }, { "epoch": 0.2370723275080633, "grad_norm": 0.6602439880371094, "learning_rate": 3.950252951096122e-05, "loss": 0.7397, "step": 5623 }, { "epoch": 0.2371144886902629, "grad_norm": 0.43376195430755615, "learning_rate": 3.950955593029792e-05, "loss": 0.1524, "step": 5624 }, { "epoch": 0.23715664987246243, "grad_norm": 0.8209614157676697, "learning_rate": 3.9516582349634626e-05, "loss": 0.1968, "step": 5625 }, { "epoch": 0.23719881105466198, "grad_norm": 0.8068884015083313, "learning_rate": 3.952360876897133e-05, "loss": 0.1336, "step": 5626 }, { "epoch": 0.23724097223686152, "grad_norm": 0.7995896339416504, "learning_rate": 3.953063518830804e-05, "loss": 0.1854, "step": 5627 }, { "epoch": 0.23728313341906107, "grad_norm": 0.7660163640975952, "learning_rate": 3.9537661607644746e-05, "loss": 0.7287, "step": 5628 }, { "epoch": 0.23732529460126062, "grad_norm": 0.7475618124008179, "learning_rate": 3.954468802698145e-05, "loss": 0.588, "step": 5629 }, { "epoch": 0.23736745578346016, "grad_norm": 0.6159539818763733, "learning_rate": 3.9551714446318154e-05, "loss": 0.2095, "step": 5630 }, { "epoch": 0.2374096169656597, "grad_norm": 0.627856433391571, "learning_rate": 3.9558740865654866e-05, "loss": 0.3027, "step": 5631 }, { "epoch": 0.23745177814785925, "grad_norm": 0.6432468891143799, "learning_rate": 3.956576728499157e-05, "loss": 0.5955, "step": 5632 }, { "epoch": 0.23749393933005883, "grad_norm": 0.8263204097747803, "learning_rate": 3.9572793704328274e-05, "loss": 0.2143, "step": 5633 }, { "epoch": 0.23753610051225837, "grad_norm": 0.5633732676506042, "learning_rate": 3.9579820123664986e-05, "loss": 0.1794, "step": 5634 }, { "epoch": 0.23757826169445792, "grad_norm": 0.5049425363540649, "learning_rate": 3.958684654300169e-05, "loss": 0.2361, "step": 5635 }, { "epoch": 0.23762042287665747, "grad_norm": 1.1349543333053589, "learning_rate": 3.9593872962338394e-05, "loss": 1.0242, "step": 5636 }, { "epoch": 0.237662584058857, "grad_norm": 0.4849970042705536, "learning_rate": 3.96008993816751e-05, "loss": 0.101, "step": 5637 }, { "epoch": 0.23770474524105656, "grad_norm": 0.5958666801452637, "learning_rate": 3.960792580101181e-05, "loss": 0.7039, "step": 5638 }, { "epoch": 0.2377469064232561, "grad_norm": 0.5115727186203003, "learning_rate": 3.9614952220348514e-05, "loss": 0.2645, "step": 5639 }, { "epoch": 0.23778906760545565, "grad_norm": 0.6001365184783936, "learning_rate": 3.962197863968522e-05, "loss": 0.5853, "step": 5640 }, { "epoch": 0.2378312287876552, "grad_norm": 1.0845156908035278, "learning_rate": 3.962900505902192e-05, "loss": 0.5654, "step": 5641 }, { "epoch": 0.23787338996985474, "grad_norm": 0.7082664966583252, "learning_rate": 3.9636031478358634e-05, "loss": 0.5396, "step": 5642 }, { "epoch": 0.23791555115205432, "grad_norm": 0.4352554380893707, "learning_rate": 3.964305789769533e-05, "loss": 0.1316, "step": 5643 }, { "epoch": 0.23795771233425386, "grad_norm": 0.7054523825645447, "learning_rate": 3.965008431703204e-05, "loss": 0.4823, "step": 5644 }, { "epoch": 0.2379998735164534, "grad_norm": 0.3858247399330139, "learning_rate": 3.965711073636875e-05, "loss": 0.1592, "step": 5645 }, { "epoch": 0.23804203469865295, "grad_norm": 0.5421833395957947, "learning_rate": 3.966413715570546e-05, "loss": 0.1951, "step": 5646 }, { "epoch": 0.2380841958808525, "grad_norm": 1.592197299003601, "learning_rate": 3.9671163575042156e-05, "loss": 1.1096, "step": 5647 }, { "epoch": 0.23812635706305205, "grad_norm": 0.4818774163722992, "learning_rate": 3.967818999437887e-05, "loss": 0.2821, "step": 5648 }, { "epoch": 0.2381685182452516, "grad_norm": 0.9679142236709595, "learning_rate": 3.968521641371557e-05, "loss": 0.3853, "step": 5649 }, { "epoch": 0.23821067942745114, "grad_norm": 0.6212756037712097, "learning_rate": 3.9692242833052276e-05, "loss": 0.1421, "step": 5650 }, { "epoch": 0.23825284060965068, "grad_norm": 0.5529770851135254, "learning_rate": 3.969926925238898e-05, "loss": 0.2582, "step": 5651 }, { "epoch": 0.23829500179185023, "grad_norm": 2.454061985015869, "learning_rate": 3.970629567172569e-05, "loss": 0.9729, "step": 5652 }, { "epoch": 0.2383371629740498, "grad_norm": 1.509915828704834, "learning_rate": 3.9713322091062396e-05, "loss": 0.5922, "step": 5653 }, { "epoch": 0.23837932415624935, "grad_norm": 0.5129079222679138, "learning_rate": 3.97203485103991e-05, "loss": 0.2961, "step": 5654 }, { "epoch": 0.2384214853384489, "grad_norm": 0.5616608262062073, "learning_rate": 3.972737492973581e-05, "loss": 0.2019, "step": 5655 }, { "epoch": 0.23846364652064844, "grad_norm": 0.5511250495910645, "learning_rate": 3.9734401349072516e-05, "loss": 0.6294, "step": 5656 }, { "epoch": 0.238505807702848, "grad_norm": 0.4009582996368408, "learning_rate": 3.974142776840922e-05, "loss": 0.1207, "step": 5657 }, { "epoch": 0.23854796888504753, "grad_norm": 0.5010126829147339, "learning_rate": 3.9748454187745925e-05, "loss": 0.4995, "step": 5658 }, { "epoch": 0.23859013006724708, "grad_norm": 1.2455873489379883, "learning_rate": 3.9755480607082636e-05, "loss": 0.9684, "step": 5659 }, { "epoch": 0.23863229124944663, "grad_norm": 0.5592502355575562, "learning_rate": 3.976250702641934e-05, "loss": 0.4559, "step": 5660 }, { "epoch": 0.23867445243164617, "grad_norm": 0.5679534673690796, "learning_rate": 3.9769533445756045e-05, "loss": 0.2693, "step": 5661 }, { "epoch": 0.23871661361384572, "grad_norm": 0.7396819591522217, "learning_rate": 3.977655986509275e-05, "loss": 0.6733, "step": 5662 }, { "epoch": 0.2387587747960453, "grad_norm": 0.5447171926498413, "learning_rate": 3.978358628442946e-05, "loss": 0.1941, "step": 5663 }, { "epoch": 0.23880093597824484, "grad_norm": 0.5662609934806824, "learning_rate": 3.979061270376616e-05, "loss": 0.4957, "step": 5664 }, { "epoch": 0.23884309716044438, "grad_norm": 0.5557995438575745, "learning_rate": 3.979763912310287e-05, "loss": 0.1644, "step": 5665 }, { "epoch": 0.23888525834264393, "grad_norm": 0.5905865430831909, "learning_rate": 3.9804665542439574e-05, "loss": 0.1673, "step": 5666 }, { "epoch": 0.23892741952484348, "grad_norm": 0.701987087726593, "learning_rate": 3.9811691961776285e-05, "loss": 0.2152, "step": 5667 }, { "epoch": 0.23896958070704302, "grad_norm": 0.3471778929233551, "learning_rate": 3.981871838111298e-05, "loss": 0.1, "step": 5668 }, { "epoch": 0.23901174188924257, "grad_norm": 0.7415712475776672, "learning_rate": 3.9825744800449694e-05, "loss": 0.5051, "step": 5669 }, { "epoch": 0.23905390307144211, "grad_norm": 2.4398107528686523, "learning_rate": 3.98327712197864e-05, "loss": 1.096, "step": 5670 }, { "epoch": 0.23909606425364166, "grad_norm": 0.6772620677947998, "learning_rate": 3.98397976391231e-05, "loss": 0.4822, "step": 5671 }, { "epoch": 0.23913822543584123, "grad_norm": 0.49821746349334717, "learning_rate": 3.984682405845981e-05, "loss": 0.165, "step": 5672 }, { "epoch": 0.23918038661804078, "grad_norm": 0.8091294765472412, "learning_rate": 3.985385047779652e-05, "loss": 0.2786, "step": 5673 }, { "epoch": 0.23922254780024033, "grad_norm": 0.7483667731285095, "learning_rate": 3.986087689713322e-05, "loss": 0.7825, "step": 5674 }, { "epoch": 0.23926470898243987, "grad_norm": 0.2991989254951477, "learning_rate": 3.986790331646993e-05, "loss": 0.0702, "step": 5675 }, { "epoch": 0.23930687016463942, "grad_norm": 0.5333059430122375, "learning_rate": 3.987492973580664e-05, "loss": 0.1836, "step": 5676 }, { "epoch": 0.23934903134683896, "grad_norm": 0.5226123929023743, "learning_rate": 3.988195615514334e-05, "loss": 0.4049, "step": 5677 }, { "epoch": 0.2393911925290385, "grad_norm": 0.3703199028968811, "learning_rate": 3.988898257448005e-05, "loss": 0.1524, "step": 5678 }, { "epoch": 0.23943335371123806, "grad_norm": 1.10942804813385, "learning_rate": 3.989600899381675e-05, "loss": 0.2032, "step": 5679 }, { "epoch": 0.2394755148934376, "grad_norm": 0.8005883097648621, "learning_rate": 3.990303541315346e-05, "loss": 0.3382, "step": 5680 }, { "epoch": 0.23951767607563715, "grad_norm": 0.7795405387878418, "learning_rate": 3.991006183249017e-05, "loss": 0.3128, "step": 5681 }, { "epoch": 0.23955983725783672, "grad_norm": 0.7339290976524353, "learning_rate": 3.991708825182687e-05, "loss": 0.1069, "step": 5682 }, { "epoch": 0.23960199844003627, "grad_norm": 0.5331852436065674, "learning_rate": 3.9924114671163575e-05, "loss": 0.2783, "step": 5683 }, { "epoch": 0.23964415962223581, "grad_norm": 1.3408448696136475, "learning_rate": 3.993114109050029e-05, "loss": 0.3623, "step": 5684 }, { "epoch": 0.23968632080443536, "grad_norm": 0.4977969229221344, "learning_rate": 3.993816750983699e-05, "loss": 0.5332, "step": 5685 }, { "epoch": 0.2397284819866349, "grad_norm": 0.7971481680870056, "learning_rate": 3.9945193929173695e-05, "loss": 0.4312, "step": 5686 }, { "epoch": 0.23977064316883445, "grad_norm": 0.7250139117240906, "learning_rate": 3.99522203485104e-05, "loss": 0.5687, "step": 5687 }, { "epoch": 0.239812804351034, "grad_norm": 0.6636866927146912, "learning_rate": 3.995924676784711e-05, "loss": 0.4797, "step": 5688 }, { "epoch": 0.23985496553323354, "grad_norm": 1.2382395267486572, "learning_rate": 3.996627318718381e-05, "loss": 0.9761, "step": 5689 }, { "epoch": 0.2398971267154331, "grad_norm": 0.458751916885376, "learning_rate": 3.997329960652052e-05, "loss": 0.208, "step": 5690 }, { "epoch": 0.23993928789763264, "grad_norm": 0.6270509958267212, "learning_rate": 3.9980326025857224e-05, "loss": 0.1145, "step": 5691 }, { "epoch": 0.2399814490798322, "grad_norm": 1.2097285985946655, "learning_rate": 3.998735244519393e-05, "loss": 0.411, "step": 5692 }, { "epoch": 0.24002361026203176, "grad_norm": 0.5050455927848816, "learning_rate": 3.999437886453063e-05, "loss": 0.146, "step": 5693 }, { "epoch": 0.2400657714442313, "grad_norm": 0.9666705131530762, "learning_rate": 4.0001405283867344e-05, "loss": 0.1883, "step": 5694 }, { "epoch": 0.24010793262643085, "grad_norm": 0.6740818023681641, "learning_rate": 4.000843170320405e-05, "loss": 0.2507, "step": 5695 }, { "epoch": 0.2401500938086304, "grad_norm": 0.9211426973342896, "learning_rate": 4.001545812254075e-05, "loss": 0.4866, "step": 5696 }, { "epoch": 0.24019225499082994, "grad_norm": 1.210163950920105, "learning_rate": 4.002248454187746e-05, "loss": 1.1669, "step": 5697 }, { "epoch": 0.24023441617302949, "grad_norm": 1.1117902994155884, "learning_rate": 4.002951096121417e-05, "loss": 0.4133, "step": 5698 }, { "epoch": 0.24027657735522903, "grad_norm": 0.7841476202011108, "learning_rate": 4.003653738055087e-05, "loss": 0.2866, "step": 5699 }, { "epoch": 0.24031873853742858, "grad_norm": 0.800288736820221, "learning_rate": 4.004356379988758e-05, "loss": 0.5371, "step": 5700 }, { "epoch": 0.24036089971962815, "grad_norm": 0.6877517700195312, "learning_rate": 4.005059021922429e-05, "loss": 0.3636, "step": 5701 }, { "epoch": 0.2404030609018277, "grad_norm": 0.6983213424682617, "learning_rate": 4.005761663856099e-05, "loss": 0.2787, "step": 5702 }, { "epoch": 0.24044522208402724, "grad_norm": 1.148699402809143, "learning_rate": 4.00646430578977e-05, "loss": 1.0059, "step": 5703 }, { "epoch": 0.2404873832662268, "grad_norm": 0.9829616546630859, "learning_rate": 4.00716694772344e-05, "loss": 0.8557, "step": 5704 }, { "epoch": 0.24052954444842634, "grad_norm": 0.5875556468963623, "learning_rate": 4.007869589657111e-05, "loss": 0.6628, "step": 5705 }, { "epoch": 0.24057170563062588, "grad_norm": 0.47168347239494324, "learning_rate": 4.008572231590782e-05, "loss": 0.1623, "step": 5706 }, { "epoch": 0.24061386681282543, "grad_norm": 0.8082226514816284, "learning_rate": 4.009274873524452e-05, "loss": 0.2015, "step": 5707 }, { "epoch": 0.24065602799502497, "grad_norm": 0.45178544521331787, "learning_rate": 4.0099775154581226e-05, "loss": 0.1764, "step": 5708 }, { "epoch": 0.24069818917722452, "grad_norm": 0.6722436547279358, "learning_rate": 4.010680157391794e-05, "loss": 0.7239, "step": 5709 }, { "epoch": 0.24074035035942407, "grad_norm": 0.6968451142311096, "learning_rate": 4.0113827993254635e-05, "loss": 0.5658, "step": 5710 }, { "epoch": 0.24078251154162364, "grad_norm": 0.4693848192691803, "learning_rate": 4.0120854412591346e-05, "loss": 0.2227, "step": 5711 }, { "epoch": 0.24082467272382319, "grad_norm": 0.42871296405792236, "learning_rate": 4.012788083192805e-05, "loss": 0.2055, "step": 5712 }, { "epoch": 0.24086683390602273, "grad_norm": 0.647282063961029, "learning_rate": 4.013490725126476e-05, "loss": 0.6703, "step": 5713 }, { "epoch": 0.24090899508822228, "grad_norm": 1.4538224935531616, "learning_rate": 4.014193367060146e-05, "loss": 0.4151, "step": 5714 }, { "epoch": 0.24095115627042182, "grad_norm": 48670.9453125, "learning_rate": 4.014896008993817e-05, "loss": 0.1138, "step": 5715 }, { "epoch": 0.24099331745262137, "grad_norm": 0.4102419316768646, "learning_rate": 4.0155986509274875e-05, "loss": 0.1067, "step": 5716 }, { "epoch": 0.24103547863482092, "grad_norm": 1.508682131767273, "learning_rate": 4.016301292861158e-05, "loss": 0.5459, "step": 5717 }, { "epoch": 0.24107763981702046, "grad_norm": 0.5228487849235535, "learning_rate": 4.0170039347948284e-05, "loss": 0.579, "step": 5718 }, { "epoch": 0.24111980099922, "grad_norm": 0.6010459065437317, "learning_rate": 4.0177065767284995e-05, "loss": 0.253, "step": 5719 }, { "epoch": 0.24116196218141955, "grad_norm": 0.9310987591743469, "learning_rate": 4.01840921866217e-05, "loss": 0.6137, "step": 5720 }, { "epoch": 0.24120412336361913, "grad_norm": 0.9838120341300964, "learning_rate": 4.0191118605958403e-05, "loss": 0.4641, "step": 5721 }, { "epoch": 0.24124628454581867, "grad_norm": 0.4336284101009369, "learning_rate": 4.0198145025295115e-05, "loss": 0.4602, "step": 5722 }, { "epoch": 0.24128844572801822, "grad_norm": 0.37572136521339417, "learning_rate": 4.020517144463182e-05, "loss": 0.1683, "step": 5723 }, { "epoch": 0.24133060691021777, "grad_norm": 0.612616777420044, "learning_rate": 4.0212197863968523e-05, "loss": 1.6524, "step": 5724 }, { "epoch": 0.2413727680924173, "grad_norm": 0.47171449661254883, "learning_rate": 4.021922428330523e-05, "loss": 0.2273, "step": 5725 }, { "epoch": 0.24141492927461686, "grad_norm": 0.7407134175300598, "learning_rate": 4.022625070264194e-05, "loss": 0.5605, "step": 5726 }, { "epoch": 0.2414570904568164, "grad_norm": 1.2258849143981934, "learning_rate": 4.0233277121978643e-05, "loss": 1.1072, "step": 5727 }, { "epoch": 0.24149925163901595, "grad_norm": 0.7743741273880005, "learning_rate": 4.024030354131535e-05, "loss": 0.4499, "step": 5728 }, { "epoch": 0.2415414128212155, "grad_norm": 0.7613524794578552, "learning_rate": 4.024732996065205e-05, "loss": 0.361, "step": 5729 }, { "epoch": 0.24158357400341507, "grad_norm": 0.7479456067085266, "learning_rate": 4.025435637998876e-05, "loss": 0.2136, "step": 5730 }, { "epoch": 0.24162573518561462, "grad_norm": 0.4537017047405243, "learning_rate": 4.026138279932547e-05, "loss": 0.2671, "step": 5731 }, { "epoch": 0.24166789636781416, "grad_norm": 0.5253421068191528, "learning_rate": 4.026840921866217e-05, "loss": 0.2224, "step": 5732 }, { "epoch": 0.2417100575500137, "grad_norm": 0.7155199646949768, "learning_rate": 4.0275435637998877e-05, "loss": 0.1765, "step": 5733 }, { "epoch": 0.24175221873221325, "grad_norm": 0.8906198143959045, "learning_rate": 4.028246205733559e-05, "loss": 0.3102, "step": 5734 }, { "epoch": 0.2417943799144128, "grad_norm": 0.5912546515464783, "learning_rate": 4.0289488476672285e-05, "loss": 0.1533, "step": 5735 }, { "epoch": 0.24183654109661235, "grad_norm": 0.3937694728374481, "learning_rate": 4.0296514896008997e-05, "loss": 0.1019, "step": 5736 }, { "epoch": 0.2418787022788119, "grad_norm": 0.5857095122337341, "learning_rate": 4.03035413153457e-05, "loss": 0.2092, "step": 5737 }, { "epoch": 0.24192086346101144, "grad_norm": 0.7951632142066956, "learning_rate": 4.0310567734682405e-05, "loss": 0.6226, "step": 5738 }, { "epoch": 0.24196302464321098, "grad_norm": 0.5157783627510071, "learning_rate": 4.031759415401911e-05, "loss": 0.1517, "step": 5739 }, { "epoch": 0.24200518582541056, "grad_norm": 0.9734373092651367, "learning_rate": 4.032462057335582e-05, "loss": 1.0816, "step": 5740 }, { "epoch": 0.2420473470076101, "grad_norm": 0.5662264227867126, "learning_rate": 4.0331646992692525e-05, "loss": 0.5091, "step": 5741 }, { "epoch": 0.24208950818980965, "grad_norm": 0.7851970791816711, "learning_rate": 4.033867341202923e-05, "loss": 0.7034, "step": 5742 }, { "epoch": 0.2421316693720092, "grad_norm": 0.657234251499176, "learning_rate": 4.0345699831365934e-05, "loss": 0.5014, "step": 5743 }, { "epoch": 0.24217383055420874, "grad_norm": 0.6100230813026428, "learning_rate": 4.0352726250702645e-05, "loss": 0.2214, "step": 5744 }, { "epoch": 0.2422159917364083, "grad_norm": 0.82221919298172, "learning_rate": 4.035975267003935e-05, "loss": 0.4524, "step": 5745 }, { "epoch": 0.24225815291860783, "grad_norm": 0.35813820362091064, "learning_rate": 4.0366779089376054e-05, "loss": 0.1967, "step": 5746 }, { "epoch": 0.24230031410080738, "grad_norm": 0.6018133759498596, "learning_rate": 4.0373805508712765e-05, "loss": 0.189, "step": 5747 }, { "epoch": 0.24234247528300693, "grad_norm": 0.5514258146286011, "learning_rate": 4.038083192804947e-05, "loss": 0.184, "step": 5748 }, { "epoch": 0.24238463646520647, "grad_norm": 0.6392618417739868, "learning_rate": 4.0387858347386174e-05, "loss": 0.1937, "step": 5749 }, { "epoch": 0.24242679764740604, "grad_norm": 1.04008948802948, "learning_rate": 4.039488476672288e-05, "loss": 1.1882, "step": 5750 }, { "epoch": 0.2424689588296056, "grad_norm": 0.6489360332489014, "learning_rate": 4.040191118605959e-05, "loss": 0.2553, "step": 5751 }, { "epoch": 0.24251112001180514, "grad_norm": 0.8805809617042542, "learning_rate": 4.0408937605396294e-05, "loss": 0.588, "step": 5752 }, { "epoch": 0.24255328119400468, "grad_norm": 0.8461984395980835, "learning_rate": 4.0415964024733e-05, "loss": 0.6329, "step": 5753 }, { "epoch": 0.24259544237620423, "grad_norm": 0.6493628621101379, "learning_rate": 4.04229904440697e-05, "loss": 0.3066, "step": 5754 }, { "epoch": 0.24263760355840377, "grad_norm": 0.5271475315093994, "learning_rate": 4.0430016863406414e-05, "loss": 0.2711, "step": 5755 }, { "epoch": 0.24267976474060332, "grad_norm": 0.6522708535194397, "learning_rate": 4.043704328274311e-05, "loss": 0.2554, "step": 5756 }, { "epoch": 0.24272192592280287, "grad_norm": 0.699528157711029, "learning_rate": 4.044406970207982e-05, "loss": 0.3204, "step": 5757 }, { "epoch": 0.2427640871050024, "grad_norm": 0.7372255921363831, "learning_rate": 4.045109612141653e-05, "loss": 0.1327, "step": 5758 }, { "epoch": 0.242806248287202, "grad_norm": 0.5221070647239685, "learning_rate": 4.045812254075324e-05, "loss": 0.0798, "step": 5759 }, { "epoch": 0.24284840946940153, "grad_norm": 0.8950637578964233, "learning_rate": 4.0465148960089936e-05, "loss": 0.4261, "step": 5760 }, { "epoch": 0.24289057065160108, "grad_norm": 0.3716680109500885, "learning_rate": 4.047217537942665e-05, "loss": 0.123, "step": 5761 }, { "epoch": 0.24293273183380062, "grad_norm": 0.7144090533256531, "learning_rate": 4.047920179876335e-05, "loss": 0.7268, "step": 5762 }, { "epoch": 0.24297489301600017, "grad_norm": 0.65850430727005, "learning_rate": 4.0486228218100056e-05, "loss": 0.1636, "step": 5763 }, { "epoch": 0.24301705419819972, "grad_norm": 0.6463544368743896, "learning_rate": 4.049325463743676e-05, "loss": 0.5782, "step": 5764 }, { "epoch": 0.24305921538039926, "grad_norm": 0.513511598110199, "learning_rate": 4.050028105677347e-05, "loss": 0.1226, "step": 5765 }, { "epoch": 0.2431013765625988, "grad_norm": 0.7320073843002319, "learning_rate": 4.0507307476110176e-05, "loss": 0.5031, "step": 5766 }, { "epoch": 0.24314353774479835, "grad_norm": 1.3558900356292725, "learning_rate": 4.051433389544688e-05, "loss": 0.4052, "step": 5767 }, { "epoch": 0.2431856989269979, "grad_norm": 0.48678484559059143, "learning_rate": 4.052136031478359e-05, "loss": 0.1963, "step": 5768 }, { "epoch": 0.24322786010919747, "grad_norm": 0.41852444410324097, "learning_rate": 4.0528386734120296e-05, "loss": 0.0948, "step": 5769 }, { "epoch": 0.24327002129139702, "grad_norm": 0.46657219529151917, "learning_rate": 4.0535413153457e-05, "loss": 0.1848, "step": 5770 }, { "epoch": 0.24331218247359657, "grad_norm": 0.7649263739585876, "learning_rate": 4.0542439572793705e-05, "loss": 0.5466, "step": 5771 }, { "epoch": 0.2433543436557961, "grad_norm": 0.7144366502761841, "learning_rate": 4.0549465992130416e-05, "loss": 0.3749, "step": 5772 }, { "epoch": 0.24339650483799566, "grad_norm": 0.47964468598365784, "learning_rate": 4.055649241146712e-05, "loss": 0.1808, "step": 5773 }, { "epoch": 0.2434386660201952, "grad_norm": 0.5855085253715515, "learning_rate": 4.0563518830803825e-05, "loss": 0.1772, "step": 5774 }, { "epoch": 0.24348082720239475, "grad_norm": 0.6568978428840637, "learning_rate": 4.057054525014053e-05, "loss": 0.17, "step": 5775 }, { "epoch": 0.2435229883845943, "grad_norm": 1.7654097080230713, "learning_rate": 4.057757166947724e-05, "loss": 0.8316, "step": 5776 }, { "epoch": 0.24356514956679384, "grad_norm": 0.6680094003677368, "learning_rate": 4.058459808881394e-05, "loss": 0.2561, "step": 5777 }, { "epoch": 0.2436073107489934, "grad_norm": 0.5320780873298645, "learning_rate": 4.059162450815065e-05, "loss": 0.1948, "step": 5778 }, { "epoch": 0.24364947193119296, "grad_norm": 0.6101425886154175, "learning_rate": 4.059865092748735e-05, "loss": 0.5402, "step": 5779 }, { "epoch": 0.2436916331133925, "grad_norm": 1.0016783475875854, "learning_rate": 4.0605677346824064e-05, "loss": 0.16, "step": 5780 }, { "epoch": 0.24373379429559205, "grad_norm": 0.64484041929245, "learning_rate": 4.061270376616076e-05, "loss": 0.7521, "step": 5781 }, { "epoch": 0.2437759554777916, "grad_norm": 0.9263688921928406, "learning_rate": 4.061973018549747e-05, "loss": 0.2504, "step": 5782 }, { "epoch": 0.24381811665999115, "grad_norm": 1.5981192588806152, "learning_rate": 4.062675660483418e-05, "loss": 0.338, "step": 5783 }, { "epoch": 0.2438602778421907, "grad_norm": 0.6799440979957581, "learning_rate": 4.063378302417088e-05, "loss": 0.5713, "step": 5784 }, { "epoch": 0.24390243902439024, "grad_norm": 0.7297481298446655, "learning_rate": 4.0640809443507586e-05, "loss": 0.2678, "step": 5785 }, { "epoch": 0.24394460020658978, "grad_norm": 0.6623439192771912, "learning_rate": 4.06478358628443e-05, "loss": 0.6551, "step": 5786 }, { "epoch": 0.24398676138878933, "grad_norm": 0.5707394480705261, "learning_rate": 4.0654862282181e-05, "loss": 0.1968, "step": 5787 }, { "epoch": 0.2440289225709889, "grad_norm": 0.5892679691314697, "learning_rate": 4.0661888701517706e-05, "loss": 0.3283, "step": 5788 }, { "epoch": 0.24407108375318845, "grad_norm": 0.652779757976532, "learning_rate": 4.066891512085442e-05, "loss": 0.1863, "step": 5789 }, { "epoch": 0.244113244935388, "grad_norm": 0.621402382850647, "learning_rate": 4.067594154019112e-05, "loss": 0.1216, "step": 5790 }, { "epoch": 0.24415540611758754, "grad_norm": 1.7166011333465576, "learning_rate": 4.0682967959527826e-05, "loss": 0.2586, "step": 5791 }, { "epoch": 0.2441975672997871, "grad_norm": 0.4348773956298828, "learning_rate": 4.068999437886453e-05, "loss": 0.0836, "step": 5792 }, { "epoch": 0.24423972848198663, "grad_norm": 0.7550461888313293, "learning_rate": 4.069702079820124e-05, "loss": 0.6322, "step": 5793 }, { "epoch": 0.24428188966418618, "grad_norm": 0.7550560235977173, "learning_rate": 4.0704047217537946e-05, "loss": 0.2049, "step": 5794 }, { "epoch": 0.24432405084638573, "grad_norm": 0.46542325615882874, "learning_rate": 4.071107363687465e-05, "loss": 0.1489, "step": 5795 }, { "epoch": 0.24436621202858527, "grad_norm": 0.6813238263130188, "learning_rate": 4.0718100056211355e-05, "loss": 0.6599, "step": 5796 }, { "epoch": 0.24440837321078482, "grad_norm": 0.5846788287162781, "learning_rate": 4.0725126475548066e-05, "loss": 0.3884, "step": 5797 }, { "epoch": 0.2444505343929844, "grad_norm": 0.4658152461051941, "learning_rate": 4.073215289488477e-05, "loss": 0.1398, "step": 5798 }, { "epoch": 0.24449269557518394, "grad_norm": 0.4801306426525116, "learning_rate": 4.0739179314221475e-05, "loss": 0.1072, "step": 5799 }, { "epoch": 0.24453485675738348, "grad_norm": 0.6834713816642761, "learning_rate": 4.074620573355818e-05, "loss": 0.2361, "step": 5800 }, { "epoch": 0.24457701793958303, "grad_norm": 0.723422646522522, "learning_rate": 4.075323215289489e-05, "loss": 0.1969, "step": 5801 }, { "epoch": 0.24461917912178258, "grad_norm": 0.6701083183288574, "learning_rate": 4.076025857223159e-05, "loss": 0.8404, "step": 5802 }, { "epoch": 0.24466134030398212, "grad_norm": 0.41305580735206604, "learning_rate": 4.07672849915683e-05, "loss": 0.1193, "step": 5803 }, { "epoch": 0.24470350148618167, "grad_norm": 0.5086827278137207, "learning_rate": 4.0774311410905004e-05, "loss": 0.1172, "step": 5804 }, { "epoch": 0.24474566266838121, "grad_norm": 0.9625945687294006, "learning_rate": 4.078133783024171e-05, "loss": 0.3048, "step": 5805 }, { "epoch": 0.24478782385058076, "grad_norm": 0.4478550851345062, "learning_rate": 4.078836424957841e-05, "loss": 0.1075, "step": 5806 }, { "epoch": 0.2448299850327803, "grad_norm": 0.5610047578811646, "learning_rate": 4.0795390668915124e-05, "loss": 0.2178, "step": 5807 }, { "epoch": 0.24487214621497988, "grad_norm": 0.6111399531364441, "learning_rate": 4.080241708825183e-05, "loss": 0.4822, "step": 5808 }, { "epoch": 0.24491430739717943, "grad_norm": 0.9898868203163147, "learning_rate": 4.080944350758853e-05, "loss": 1.122, "step": 5809 }, { "epoch": 0.24495646857937897, "grad_norm": 0.47617578506469727, "learning_rate": 4.081646992692524e-05, "loss": 0.11, "step": 5810 }, { "epoch": 0.24499862976157852, "grad_norm": 0.45956143736839294, "learning_rate": 4.082349634626195e-05, "loss": 0.1898, "step": 5811 }, { "epoch": 0.24504079094377806, "grad_norm": 0.6246912479400635, "learning_rate": 4.083052276559865e-05, "loss": 0.1671, "step": 5812 }, { "epoch": 0.2450829521259776, "grad_norm": 0.7082919478416443, "learning_rate": 4.083754918493536e-05, "loss": 0.3113, "step": 5813 }, { "epoch": 0.24512511330817716, "grad_norm": 0.5380542278289795, "learning_rate": 4.084457560427207e-05, "loss": 0.2306, "step": 5814 }, { "epoch": 0.2451672744903767, "grad_norm": 0.8167911171913147, "learning_rate": 4.085160202360877e-05, "loss": 0.3813, "step": 5815 }, { "epoch": 0.24520943567257625, "grad_norm": 0.6531877517700195, "learning_rate": 4.085862844294548e-05, "loss": 0.4129, "step": 5816 }, { "epoch": 0.2452515968547758, "grad_norm": 0.6038156151771545, "learning_rate": 4.086565486228218e-05, "loss": 0.1393, "step": 5817 }, { "epoch": 0.24529375803697537, "grad_norm": 1.1550792455673218, "learning_rate": 4.087268128161889e-05, "loss": 1.2962, "step": 5818 }, { "epoch": 0.2453359192191749, "grad_norm": 0.5439779758453369, "learning_rate": 4.08797077009556e-05, "loss": 0.324, "step": 5819 }, { "epoch": 0.24537808040137446, "grad_norm": 0.4787807762622833, "learning_rate": 4.08867341202923e-05, "loss": 0.1746, "step": 5820 }, { "epoch": 0.245420241583574, "grad_norm": 1.3548465967178345, "learning_rate": 4.0893760539629006e-05, "loss": 1.001, "step": 5821 }, { "epoch": 0.24546240276577355, "grad_norm": 1.3052387237548828, "learning_rate": 4.090078695896572e-05, "loss": 0.3893, "step": 5822 }, { "epoch": 0.2455045639479731, "grad_norm": 0.6659852266311646, "learning_rate": 4.0907813378302414e-05, "loss": 0.2933, "step": 5823 }, { "epoch": 0.24554672513017264, "grad_norm": 0.33528199791908264, "learning_rate": 4.0914839797639126e-05, "loss": 0.1467, "step": 5824 }, { "epoch": 0.2455888863123722, "grad_norm": 0.4462713599205017, "learning_rate": 4.092186621697583e-05, "loss": 0.3073, "step": 5825 }, { "epoch": 0.24563104749457174, "grad_norm": 0.37475886940956116, "learning_rate": 4.092889263631254e-05, "loss": 0.1033, "step": 5826 }, { "epoch": 0.2456732086767713, "grad_norm": 0.6632177829742432, "learning_rate": 4.093591905564924e-05, "loss": 0.3659, "step": 5827 }, { "epoch": 0.24571536985897086, "grad_norm": 1.154565453529358, "learning_rate": 4.094294547498595e-05, "loss": 0.4739, "step": 5828 }, { "epoch": 0.2457575310411704, "grad_norm": 0.5704101324081421, "learning_rate": 4.0949971894322654e-05, "loss": 0.2885, "step": 5829 }, { "epoch": 0.24579969222336995, "grad_norm": 0.4930320382118225, "learning_rate": 4.095699831365936e-05, "loss": 0.3649, "step": 5830 }, { "epoch": 0.2458418534055695, "grad_norm": 0.7600178122520447, "learning_rate": 4.096402473299606e-05, "loss": 0.6849, "step": 5831 }, { "epoch": 0.24588401458776904, "grad_norm": 1.236137866973877, "learning_rate": 4.0971051152332774e-05, "loss": 0.3686, "step": 5832 }, { "epoch": 0.24592617576996859, "grad_norm": 0.323519766330719, "learning_rate": 4.097807757166948e-05, "loss": 0.0927, "step": 5833 }, { "epoch": 0.24596833695216813, "grad_norm": 2.752065420150757, "learning_rate": 4.098510399100618e-05, "loss": 0.712, "step": 5834 }, { "epoch": 0.24601049813436768, "grad_norm": 0.7883109450340271, "learning_rate": 4.0992130410342894e-05, "loss": 0.4262, "step": 5835 }, { "epoch": 0.24605265931656722, "grad_norm": 1.3867040872573853, "learning_rate": 4.09991568296796e-05, "loss": 1.2897, "step": 5836 }, { "epoch": 0.2460948204987668, "grad_norm": 1.3473079204559326, "learning_rate": 4.10061832490163e-05, "loss": 0.2609, "step": 5837 }, { "epoch": 0.24613698168096634, "grad_norm": 0.7332273125648499, "learning_rate": 4.101320966835301e-05, "loss": 0.6946, "step": 5838 }, { "epoch": 0.2461791428631659, "grad_norm": 0.6716182827949524, "learning_rate": 4.102023608768972e-05, "loss": 0.4358, "step": 5839 }, { "epoch": 0.24622130404536544, "grad_norm": 0.7173119783401489, "learning_rate": 4.102726250702642e-05, "loss": 0.2102, "step": 5840 }, { "epoch": 0.24626346522756498, "grad_norm": 0.8590998649597168, "learning_rate": 4.103428892636313e-05, "loss": 0.2874, "step": 5841 }, { "epoch": 0.24630562640976453, "grad_norm": 0.6626501083374023, "learning_rate": 4.104131534569983e-05, "loss": 1.4386, "step": 5842 }, { "epoch": 0.24634778759196407, "grad_norm": 0.5103005170822144, "learning_rate": 4.104834176503654e-05, "loss": 0.3622, "step": 5843 }, { "epoch": 0.24638994877416362, "grad_norm": 0.8678844571113586, "learning_rate": 4.105536818437325e-05, "loss": 0.1239, "step": 5844 }, { "epoch": 0.24643210995636317, "grad_norm": 1.3811312913894653, "learning_rate": 4.106239460370995e-05, "loss": 0.6638, "step": 5845 }, { "epoch": 0.2464742711385627, "grad_norm": 0.8519685864448547, "learning_rate": 4.1069421023046656e-05, "loss": 0.1906, "step": 5846 }, { "epoch": 0.24651643232076229, "grad_norm": 0.7156473994255066, "learning_rate": 4.107644744238337e-05, "loss": 0.5311, "step": 5847 }, { "epoch": 0.24655859350296183, "grad_norm": 0.2874400317668915, "learning_rate": 4.1083473861720065e-05, "loss": 0.1447, "step": 5848 }, { "epoch": 0.24660075468516138, "grad_norm": 0.5991911292076111, "learning_rate": 4.1090500281056776e-05, "loss": 0.2283, "step": 5849 }, { "epoch": 0.24664291586736092, "grad_norm": 0.8224478363990784, "learning_rate": 4.109752670039348e-05, "loss": 0.1138, "step": 5850 }, { "epoch": 0.24668507704956047, "grad_norm": 0.7810804843902588, "learning_rate": 4.1104553119730185e-05, "loss": 0.5517, "step": 5851 }, { "epoch": 0.24672723823176002, "grad_norm": 0.5330067873001099, "learning_rate": 4.111157953906689e-05, "loss": 0.1261, "step": 5852 }, { "epoch": 0.24676939941395956, "grad_norm": 1.2651921510696411, "learning_rate": 4.11186059584036e-05, "loss": 1.0741, "step": 5853 }, { "epoch": 0.2468115605961591, "grad_norm": 1.2294292449951172, "learning_rate": 4.1125632377740305e-05, "loss": 0.9743, "step": 5854 }, { "epoch": 0.24685372177835865, "grad_norm": 0.8867425322532654, "learning_rate": 4.113265879707701e-05, "loss": 0.2729, "step": 5855 }, { "epoch": 0.24689588296055823, "grad_norm": 0.5289014577865601, "learning_rate": 4.113968521641372e-05, "loss": 0.1286, "step": 5856 }, { "epoch": 0.24693804414275777, "grad_norm": 1.8561925888061523, "learning_rate": 4.1146711635750425e-05, "loss": 0.5463, "step": 5857 }, { "epoch": 0.24698020532495732, "grad_norm": 0.5163257122039795, "learning_rate": 4.115373805508713e-05, "loss": 0.1295, "step": 5858 }, { "epoch": 0.24702236650715687, "grad_norm": 1.9233148097991943, "learning_rate": 4.1160764474423834e-05, "loss": 0.835, "step": 5859 }, { "epoch": 0.2470645276893564, "grad_norm": 0.7134460806846619, "learning_rate": 4.1167790893760545e-05, "loss": 1.0127, "step": 5860 }, { "epoch": 0.24710668887155596, "grad_norm": 0.7624219059944153, "learning_rate": 4.117481731309725e-05, "loss": 0.1209, "step": 5861 }, { "epoch": 0.2471488500537555, "grad_norm": 0.8105216026306152, "learning_rate": 4.1181843732433954e-05, "loss": 0.6151, "step": 5862 }, { "epoch": 0.24719101123595505, "grad_norm": 0.5777645707130432, "learning_rate": 4.118887015177066e-05, "loss": 0.4241, "step": 5863 }, { "epoch": 0.2472331724181546, "grad_norm": 0.650534987449646, "learning_rate": 4.119589657110737e-05, "loss": 0.679, "step": 5864 }, { "epoch": 0.24727533360035414, "grad_norm": 1.182286262512207, "learning_rate": 4.1202922990444074e-05, "loss": 0.6015, "step": 5865 }, { "epoch": 0.24731749478255372, "grad_norm": 0.5160400867462158, "learning_rate": 4.120994940978078e-05, "loss": 0.4927, "step": 5866 }, { "epoch": 0.24735965596475326, "grad_norm": 0.5233809947967529, "learning_rate": 4.121697582911748e-05, "loss": 0.1447, "step": 5867 }, { "epoch": 0.2474018171469528, "grad_norm": 1.2684547901153564, "learning_rate": 4.1224002248454194e-05, "loss": 0.9347, "step": 5868 }, { "epoch": 0.24744397832915235, "grad_norm": 0.5151516795158386, "learning_rate": 4.123102866779089e-05, "loss": 0.2901, "step": 5869 }, { "epoch": 0.2474861395113519, "grad_norm": 0.5127412676811218, "learning_rate": 4.12380550871276e-05, "loss": 0.1683, "step": 5870 }, { "epoch": 0.24752830069355145, "grad_norm": 0.704382598400116, "learning_rate": 4.124508150646431e-05, "loss": 0.5328, "step": 5871 }, { "epoch": 0.247570461875751, "grad_norm": 1.8243097066879272, "learning_rate": 4.125210792580102e-05, "loss": 1.0426, "step": 5872 }, { "epoch": 0.24761262305795054, "grad_norm": 1.960957646369934, "learning_rate": 4.1259134345137715e-05, "loss": 0.185, "step": 5873 }, { "epoch": 0.24765478424015008, "grad_norm": 0.6931017637252808, "learning_rate": 4.126616076447443e-05, "loss": 0.1974, "step": 5874 }, { "epoch": 0.24769694542234963, "grad_norm": 1.1655665636062622, "learning_rate": 4.127318718381113e-05, "loss": 1.1552, "step": 5875 }, { "epoch": 0.2477391066045492, "grad_norm": 0.6758150458335876, "learning_rate": 4.1280213603147835e-05, "loss": 0.6165, "step": 5876 }, { "epoch": 0.24778126778674875, "grad_norm": 0.5899762511253357, "learning_rate": 4.128724002248454e-05, "loss": 0.2605, "step": 5877 }, { "epoch": 0.2478234289689483, "grad_norm": 0.6297777891159058, "learning_rate": 4.129426644182125e-05, "loss": 0.5816, "step": 5878 }, { "epoch": 0.24786559015114784, "grad_norm": 1.0424261093139648, "learning_rate": 4.1301292861157955e-05, "loss": 1.0698, "step": 5879 }, { "epoch": 0.2479077513333474, "grad_norm": 0.6475388407707214, "learning_rate": 4.130831928049466e-05, "loss": 0.5313, "step": 5880 }, { "epoch": 0.24794991251554693, "grad_norm": 0.61393803358078, "learning_rate": 4.131534569983137e-05, "loss": 0.2099, "step": 5881 }, { "epoch": 0.24799207369774648, "grad_norm": 0.46919959783554077, "learning_rate": 4.1322372119168075e-05, "loss": 0.1912, "step": 5882 }, { "epoch": 0.24803423487994603, "grad_norm": 0.7891495227813721, "learning_rate": 4.132939853850478e-05, "loss": 0.1781, "step": 5883 }, { "epoch": 0.24807639606214557, "grad_norm": 1.228431224822998, "learning_rate": 4.1336424957841484e-05, "loss": 1.0458, "step": 5884 }, { "epoch": 0.24811855724434514, "grad_norm": 0.5361713767051697, "learning_rate": 4.1343451377178195e-05, "loss": 0.2352, "step": 5885 }, { "epoch": 0.2481607184265447, "grad_norm": 0.5549903512001038, "learning_rate": 4.13504777965149e-05, "loss": 0.3961, "step": 5886 }, { "epoch": 0.24820287960874424, "grad_norm": 2.2991065979003906, "learning_rate": 4.1357504215851604e-05, "loss": 0.622, "step": 5887 }, { "epoch": 0.24824504079094378, "grad_norm": 0.5611404180526733, "learning_rate": 4.136453063518831e-05, "loss": 0.2217, "step": 5888 }, { "epoch": 0.24828720197314333, "grad_norm": 0.39228227734565735, "learning_rate": 4.137155705452502e-05, "loss": 0.1039, "step": 5889 }, { "epoch": 0.24832936315534287, "grad_norm": 0.7028424143791199, "learning_rate": 4.137858347386172e-05, "loss": 0.5806, "step": 5890 }, { "epoch": 0.24837152433754242, "grad_norm": 0.7047517895698547, "learning_rate": 4.138560989319843e-05, "loss": 0.7524, "step": 5891 }, { "epoch": 0.24841368551974197, "grad_norm": 0.42066454887390137, "learning_rate": 4.139263631253513e-05, "loss": 0.1999, "step": 5892 }, { "epoch": 0.2484558467019415, "grad_norm": 1.2869335412979126, "learning_rate": 4.1399662731871844e-05, "loss": 0.5226, "step": 5893 }, { "epoch": 0.24849800788414106, "grad_norm": 1.1342134475708008, "learning_rate": 4.140668915120854e-05, "loss": 1.0005, "step": 5894 }, { "epoch": 0.24854016906634063, "grad_norm": 0.5587583780288696, "learning_rate": 4.141371557054525e-05, "loss": 0.4398, "step": 5895 }, { "epoch": 0.24858233024854018, "grad_norm": 0.39602071046829224, "learning_rate": 4.142074198988196e-05, "loss": 0.1735, "step": 5896 }, { "epoch": 0.24862449143073972, "grad_norm": 0.7093988060951233, "learning_rate": 4.142776840921866e-05, "loss": 0.3428, "step": 5897 }, { "epoch": 0.24866665261293927, "grad_norm": 0.5153607726097107, "learning_rate": 4.1434794828555366e-05, "loss": 0.2274, "step": 5898 }, { "epoch": 0.24870881379513882, "grad_norm": 0.327527791261673, "learning_rate": 4.144182124789208e-05, "loss": 0.0657, "step": 5899 }, { "epoch": 0.24875097497733836, "grad_norm": 0.34976834058761597, "learning_rate": 4.144884766722878e-05, "loss": 0.1499, "step": 5900 }, { "epoch": 0.2487931361595379, "grad_norm": 0.564477264881134, "learning_rate": 4.1455874086565486e-05, "loss": 0.5159, "step": 5901 }, { "epoch": 0.24883529734173745, "grad_norm": 0.6492339372634888, "learning_rate": 4.14629005059022e-05, "loss": 0.3382, "step": 5902 }, { "epoch": 0.248877458523937, "grad_norm": 0.4516836106777191, "learning_rate": 4.14699269252389e-05, "loss": 0.4534, "step": 5903 }, { "epoch": 0.24891961970613655, "grad_norm": 0.369832843542099, "learning_rate": 4.1476953344575606e-05, "loss": 0.1683, "step": 5904 }, { "epoch": 0.24896178088833612, "grad_norm": 0.4575158953666687, "learning_rate": 4.148397976391231e-05, "loss": 0.1566, "step": 5905 }, { "epoch": 0.24900394207053567, "grad_norm": 1.7801032066345215, "learning_rate": 4.149100618324902e-05, "loss": 0.8749, "step": 5906 }, { "epoch": 0.2490461032527352, "grad_norm": 0.6646029353141785, "learning_rate": 4.1498032602585726e-05, "loss": 0.2538, "step": 5907 }, { "epoch": 0.24908826443493476, "grad_norm": 0.35691842436790466, "learning_rate": 4.150505902192243e-05, "loss": 0.1433, "step": 5908 }, { "epoch": 0.2491304256171343, "grad_norm": 0.8908524513244629, "learning_rate": 4.1512085441259135e-05, "loss": 0.61, "step": 5909 }, { "epoch": 0.24917258679933385, "grad_norm": 0.43504369258880615, "learning_rate": 4.1519111860595846e-05, "loss": 0.2897, "step": 5910 }, { "epoch": 0.2492147479815334, "grad_norm": 0.7352142930030823, "learning_rate": 4.152613827993255e-05, "loss": 0.2418, "step": 5911 }, { "epoch": 0.24925690916373294, "grad_norm": 0.6129283308982849, "learning_rate": 4.1533164699269255e-05, "loss": 0.5941, "step": 5912 }, { "epoch": 0.2492990703459325, "grad_norm": 0.40463176369667053, "learning_rate": 4.154019111860596e-05, "loss": 0.1358, "step": 5913 }, { "epoch": 0.24934123152813206, "grad_norm": 1.3319271802902222, "learning_rate": 4.154721753794267e-05, "loss": 0.3075, "step": 5914 }, { "epoch": 0.2493833927103316, "grad_norm": 0.4482915997505188, "learning_rate": 4.155424395727937e-05, "loss": 0.2531, "step": 5915 }, { "epoch": 0.24942555389253115, "grad_norm": 0.4379643499851227, "learning_rate": 4.156127037661608e-05, "loss": 0.2669, "step": 5916 }, { "epoch": 0.2494677150747307, "grad_norm": 0.5129122734069824, "learning_rate": 4.1568296795952783e-05, "loss": 0.1863, "step": 5917 }, { "epoch": 0.24950987625693025, "grad_norm": 0.4845869541168213, "learning_rate": 4.157532321528949e-05, "loss": 0.1534, "step": 5918 }, { "epoch": 0.2495520374391298, "grad_norm": 0.8049329519271851, "learning_rate": 4.158234963462619e-05, "loss": 0.2413, "step": 5919 }, { "epoch": 0.24959419862132934, "grad_norm": 1.2940810918807983, "learning_rate": 4.15893760539629e-05, "loss": 1.0562, "step": 5920 }, { "epoch": 0.24963635980352888, "grad_norm": 0.855796217918396, "learning_rate": 4.159640247329961e-05, "loss": 0.7538, "step": 5921 }, { "epoch": 0.24967852098572843, "grad_norm": 0.617790699005127, "learning_rate": 4.160342889263631e-05, "loss": 0.2308, "step": 5922 }, { "epoch": 0.24972068216792798, "grad_norm": 0.6949178576469421, "learning_rate": 4.1610455311973017e-05, "loss": 0.6512, "step": 5923 }, { "epoch": 0.24976284335012755, "grad_norm": 0.405220091342926, "learning_rate": 4.161748173130973e-05, "loss": 0.1224, "step": 5924 }, { "epoch": 0.2498050045323271, "grad_norm": 0.6060587167739868, "learning_rate": 4.162450815064643e-05, "loss": 0.1807, "step": 5925 }, { "epoch": 0.24984716571452664, "grad_norm": 0.5875492095947266, "learning_rate": 4.1631534569983137e-05, "loss": 0.726, "step": 5926 }, { "epoch": 0.2498893268967262, "grad_norm": 0.3633253574371338, "learning_rate": 4.163856098931985e-05, "loss": 0.2569, "step": 5927 }, { "epoch": 0.24993148807892573, "grad_norm": 2.0378143787384033, "learning_rate": 4.164558740865655e-05, "loss": 0.6678, "step": 5928 }, { "epoch": 0.24997364926112528, "grad_norm": 0.49284306168556213, "learning_rate": 4.1652613827993256e-05, "loss": 0.2048, "step": 5929 }, { "epoch": 0.2500158104433248, "grad_norm": 0.8111305236816406, "learning_rate": 4.165964024732996e-05, "loss": 0.3071, "step": 5930 }, { "epoch": 0.2500579716255244, "grad_norm": 0.40117043256759644, "learning_rate": 4.166666666666667e-05, "loss": 0.1113, "step": 5931 }, { "epoch": 0.2501001328077239, "grad_norm": 1.274897813796997, "learning_rate": 4.1673693086003376e-05, "loss": 0.7989, "step": 5932 }, { "epoch": 0.2501422939899235, "grad_norm": 0.681252121925354, "learning_rate": 4.168071950534008e-05, "loss": 0.2345, "step": 5933 }, { "epoch": 0.250184455172123, "grad_norm": 0.6838977336883545, "learning_rate": 4.1687745924676785e-05, "loss": 0.6528, "step": 5934 }, { "epoch": 0.2502266163543226, "grad_norm": 0.6653366684913635, "learning_rate": 4.1694772344013496e-05, "loss": 0.4092, "step": 5935 }, { "epoch": 0.2502687775365221, "grad_norm": 0.48381277918815613, "learning_rate": 4.1701798763350194e-05, "loss": 0.3107, "step": 5936 }, { "epoch": 0.2503109387187217, "grad_norm": 0.37675046920776367, "learning_rate": 4.1708825182686905e-05, "loss": 0.1183, "step": 5937 }, { "epoch": 0.2503530999009212, "grad_norm": 0.5463252067565918, "learning_rate": 4.171585160202361e-05, "loss": 0.2433, "step": 5938 }, { "epoch": 0.25039526108312077, "grad_norm": 0.6512745022773743, "learning_rate": 4.172287802136032e-05, "loss": 0.6731, "step": 5939 }, { "epoch": 0.25043742226532034, "grad_norm": 0.6820567846298218, "learning_rate": 4.172990444069702e-05, "loss": 0.1767, "step": 5940 }, { "epoch": 0.25047958344751986, "grad_norm": 0.7932310104370117, "learning_rate": 4.173693086003373e-05, "loss": 1.1636, "step": 5941 }, { "epoch": 0.25052174462971943, "grad_norm": 0.8778334856033325, "learning_rate": 4.1743957279370434e-05, "loss": 1.0992, "step": 5942 }, { "epoch": 0.25056390581191895, "grad_norm": 0.6060431003570557, "learning_rate": 4.175098369870714e-05, "loss": 0.5169, "step": 5943 }, { "epoch": 0.2506060669941185, "grad_norm": 0.6408033967018127, "learning_rate": 4.175801011804384e-05, "loss": 0.6206, "step": 5944 }, { "epoch": 0.25064822817631804, "grad_norm": 0.6051128506660461, "learning_rate": 4.1765036537380554e-05, "loss": 0.5609, "step": 5945 }, { "epoch": 0.2506903893585176, "grad_norm": 0.591212272644043, "learning_rate": 4.177206295671726e-05, "loss": 0.1986, "step": 5946 }, { "epoch": 0.25073255054071714, "grad_norm": 0.3556881248950958, "learning_rate": 4.177908937605396e-05, "loss": 0.1146, "step": 5947 }, { "epoch": 0.2507747117229167, "grad_norm": 0.37531399726867676, "learning_rate": 4.1786115795390674e-05, "loss": 0.0779, "step": 5948 }, { "epoch": 0.2508168729051163, "grad_norm": 0.5368701815605164, "learning_rate": 4.179314221472738e-05, "loss": 0.2886, "step": 5949 }, { "epoch": 0.2508590340873158, "grad_norm": 0.6294023990631104, "learning_rate": 4.180016863406408e-05, "loss": 0.3714, "step": 5950 }, { "epoch": 0.2509011952695154, "grad_norm": 0.7685365676879883, "learning_rate": 4.180719505340079e-05, "loss": 0.4886, "step": 5951 }, { "epoch": 0.2509433564517149, "grad_norm": 0.6194131970405579, "learning_rate": 4.18142214727375e-05, "loss": 0.2739, "step": 5952 }, { "epoch": 0.25098551763391447, "grad_norm": 0.5306821465492249, "learning_rate": 4.18212478920742e-05, "loss": 0.1265, "step": 5953 }, { "epoch": 0.251027678816114, "grad_norm": 0.7472935914993286, "learning_rate": 4.182827431141091e-05, "loss": 0.5311, "step": 5954 }, { "epoch": 0.25106983999831356, "grad_norm": 0.5574510097503662, "learning_rate": 4.183530073074761e-05, "loss": 0.217, "step": 5955 }, { "epoch": 0.2511120011805131, "grad_norm": 0.4028880000114441, "learning_rate": 4.184232715008432e-05, "loss": 0.1636, "step": 5956 }, { "epoch": 0.25115416236271265, "grad_norm": 0.5753700733184814, "learning_rate": 4.184935356942103e-05, "loss": 0.1682, "step": 5957 }, { "epoch": 0.2511963235449122, "grad_norm": 0.9481659531593323, "learning_rate": 4.185637998875773e-05, "loss": 1.1275, "step": 5958 }, { "epoch": 0.25123848472711174, "grad_norm": 0.4247422218322754, "learning_rate": 4.1863406408094436e-05, "loss": 0.22, "step": 5959 }, { "epoch": 0.2512806459093113, "grad_norm": 0.6327815651893616, "learning_rate": 4.187043282743115e-05, "loss": 0.3532, "step": 5960 }, { "epoch": 0.25132280709151084, "grad_norm": 0.8430453538894653, "learning_rate": 4.1877459246767845e-05, "loss": 0.2009, "step": 5961 }, { "epoch": 0.2513649682737104, "grad_norm": 0.6102507710456848, "learning_rate": 4.1884485666104556e-05, "loss": 0.3904, "step": 5962 }, { "epoch": 0.25140712945590993, "grad_norm": 0.3670629858970642, "learning_rate": 4.189151208544126e-05, "loss": 0.1423, "step": 5963 }, { "epoch": 0.2514492906381095, "grad_norm": 0.6953288316726685, "learning_rate": 4.1898538504777965e-05, "loss": 0.1607, "step": 5964 }, { "epoch": 0.251491451820309, "grad_norm": 0.8314482569694519, "learning_rate": 4.190556492411467e-05, "loss": 0.294, "step": 5965 }, { "epoch": 0.2515336130025086, "grad_norm": 0.498038649559021, "learning_rate": 4.191259134345138e-05, "loss": 0.3055, "step": 5966 }, { "epoch": 0.2515757741847081, "grad_norm": 0.6187154650688171, "learning_rate": 4.1919617762788084e-05, "loss": 0.1582, "step": 5967 }, { "epoch": 0.2516179353669077, "grad_norm": 0.6623698472976685, "learning_rate": 4.192664418212479e-05, "loss": 0.1332, "step": 5968 }, { "epoch": 0.25166009654910726, "grad_norm": 1.0380946397781372, "learning_rate": 4.19336706014615e-05, "loss": 1.0367, "step": 5969 }, { "epoch": 0.2517022577313068, "grad_norm": 0.5995873212814331, "learning_rate": 4.1940697020798204e-05, "loss": 0.2029, "step": 5970 }, { "epoch": 0.25174441891350635, "grad_norm": 0.5208089351654053, "learning_rate": 4.194772344013491e-05, "loss": 0.3191, "step": 5971 }, { "epoch": 0.25178658009570587, "grad_norm": 0.7518671751022339, "learning_rate": 4.195474985947161e-05, "loss": 0.2895, "step": 5972 }, { "epoch": 0.25182874127790544, "grad_norm": 1.0136909484863281, "learning_rate": 4.1961776278808324e-05, "loss": 0.4909, "step": 5973 }, { "epoch": 0.25187090246010496, "grad_norm": 0.5991634726524353, "learning_rate": 4.196880269814503e-05, "loss": 0.2107, "step": 5974 }, { "epoch": 0.25191306364230454, "grad_norm": 0.5339450836181641, "learning_rate": 4.197582911748173e-05, "loss": 0.2217, "step": 5975 }, { "epoch": 0.25195522482450405, "grad_norm": 0.7023271322250366, "learning_rate": 4.198285553681844e-05, "loss": 0.3938, "step": 5976 }, { "epoch": 0.2519973860067036, "grad_norm": 0.8086064457893372, "learning_rate": 4.198988195615515e-05, "loss": 0.7571, "step": 5977 }, { "epoch": 0.2520395471889032, "grad_norm": 0.6127766966819763, "learning_rate": 4.199690837549185e-05, "loss": 0.4366, "step": 5978 }, { "epoch": 0.2520817083711027, "grad_norm": 0.6708973050117493, "learning_rate": 4.200393479482856e-05, "loss": 0.516, "step": 5979 }, { "epoch": 0.2521238695533023, "grad_norm": 0.9116885662078857, "learning_rate": 4.201096121416526e-05, "loss": 0.623, "step": 5980 }, { "epoch": 0.2521660307355018, "grad_norm": 0.40409520268440247, "learning_rate": 4.201798763350197e-05, "loss": 0.1509, "step": 5981 }, { "epoch": 0.2522081919177014, "grad_norm": 0.6264949440956116, "learning_rate": 4.202501405283867e-05, "loss": 0.4941, "step": 5982 }, { "epoch": 0.2522503530999009, "grad_norm": 0.6314494013786316, "learning_rate": 4.203204047217538e-05, "loss": 0.3121, "step": 5983 }, { "epoch": 0.2522925142821005, "grad_norm": 0.3840036988258362, "learning_rate": 4.2039066891512086e-05, "loss": 0.0986, "step": 5984 }, { "epoch": 0.2523346754643, "grad_norm": 0.4673832952976227, "learning_rate": 4.20460933108488e-05, "loss": 0.1483, "step": 5985 }, { "epoch": 0.25237683664649957, "grad_norm": 0.9069986343383789, "learning_rate": 4.2053119730185495e-05, "loss": 0.2139, "step": 5986 }, { "epoch": 0.25241899782869914, "grad_norm": 0.7972880005836487, "learning_rate": 4.2060146149522206e-05, "loss": 0.4828, "step": 5987 }, { "epoch": 0.25246115901089866, "grad_norm": 0.6194260120391846, "learning_rate": 4.206717256885891e-05, "loss": 0.1651, "step": 5988 }, { "epoch": 0.25250332019309824, "grad_norm": 0.47561773657798767, "learning_rate": 4.2074198988195615e-05, "loss": 0.1729, "step": 5989 }, { "epoch": 0.25254548137529775, "grad_norm": 1.7946091890335083, "learning_rate": 4.208122540753232e-05, "loss": 0.168, "step": 5990 }, { "epoch": 0.2525876425574973, "grad_norm": 0.6715545654296875, "learning_rate": 4.208825182686903e-05, "loss": 0.7338, "step": 5991 }, { "epoch": 0.25262980373969685, "grad_norm": 0.7400501370429993, "learning_rate": 4.2095278246205735e-05, "loss": 0.7211, "step": 5992 }, { "epoch": 0.2526719649218964, "grad_norm": 0.7517655491828918, "learning_rate": 4.210230466554244e-05, "loss": 0.3123, "step": 5993 }, { "epoch": 0.25271412610409594, "grad_norm": 0.4020538628101349, "learning_rate": 4.210933108487915e-05, "loss": 0.1087, "step": 5994 }, { "epoch": 0.2527562872862955, "grad_norm": 0.8143284320831299, "learning_rate": 4.2116357504215855e-05, "loss": 0.2617, "step": 5995 }, { "epoch": 0.25279844846849503, "grad_norm": 0.3830530643463135, "learning_rate": 4.212338392355256e-05, "loss": 0.1969, "step": 5996 }, { "epoch": 0.2528406096506946, "grad_norm": 1.165500283241272, "learning_rate": 4.2130410342889264e-05, "loss": 0.2833, "step": 5997 }, { "epoch": 0.2528827708328942, "grad_norm": 0.4091595411300659, "learning_rate": 4.2137436762225975e-05, "loss": 0.1666, "step": 5998 }, { "epoch": 0.2529249320150937, "grad_norm": 0.5689324736595154, "learning_rate": 4.214446318156268e-05, "loss": 0.1202, "step": 5999 }, { "epoch": 0.25296709319729327, "grad_norm": 0.5774123072624207, "learning_rate": 4.2151489600899384e-05, "loss": 0.5067, "step": 6000 }, { "epoch": 0.2530092543794928, "grad_norm": 0.4177442193031311, "learning_rate": 4.215851602023609e-05, "loss": 0.1601, "step": 6001 }, { "epoch": 0.25305141556169236, "grad_norm": 0.5716189742088318, "learning_rate": 4.21655424395728e-05, "loss": 0.4607, "step": 6002 }, { "epoch": 0.2530935767438919, "grad_norm": 1.128085732460022, "learning_rate": 4.21725688589095e-05, "loss": 1.1664, "step": 6003 }, { "epoch": 0.25313573792609145, "grad_norm": 0.42114731669425964, "learning_rate": 4.217959527824621e-05, "loss": 0.0969, "step": 6004 }, { "epoch": 0.25317789910829097, "grad_norm": 0.5499265193939209, "learning_rate": 4.218662169758291e-05, "loss": 0.4181, "step": 6005 }, { "epoch": 0.25322006029049055, "grad_norm": 1.4948829412460327, "learning_rate": 4.2193648116919624e-05, "loss": 0.652, "step": 6006 }, { "epoch": 0.2532622214726901, "grad_norm": 0.7228555083274841, "learning_rate": 4.220067453625632e-05, "loss": 0.6657, "step": 6007 }, { "epoch": 0.25330438265488964, "grad_norm": 0.5104141235351562, "learning_rate": 4.220770095559303e-05, "loss": 0.0861, "step": 6008 }, { "epoch": 0.2533465438370892, "grad_norm": 0.49095118045806885, "learning_rate": 4.221472737492974e-05, "loss": 0.1858, "step": 6009 }, { "epoch": 0.25338870501928873, "grad_norm": 0.8713549375534058, "learning_rate": 4.222175379426644e-05, "loss": 0.2696, "step": 6010 }, { "epoch": 0.2534308662014883, "grad_norm": 0.6772237420082092, "learning_rate": 4.2228780213603146e-05, "loss": 0.9266, "step": 6011 }, { "epoch": 0.2534730273836878, "grad_norm": 1.9829775094985962, "learning_rate": 4.223580663293986e-05, "loss": 0.9867, "step": 6012 }, { "epoch": 0.2535151885658874, "grad_norm": 0.7231395840644836, "learning_rate": 4.224283305227656e-05, "loss": 0.4091, "step": 6013 }, { "epoch": 0.2535573497480869, "grad_norm": 0.851216197013855, "learning_rate": 4.2249859471613266e-05, "loss": 0.4673, "step": 6014 }, { "epoch": 0.2535995109302865, "grad_norm": 0.6527634859085083, "learning_rate": 4.225688589094998e-05, "loss": 0.3433, "step": 6015 }, { "epoch": 0.25364167211248606, "grad_norm": 0.552006721496582, "learning_rate": 4.226391231028668e-05, "loss": 0.2559, "step": 6016 }, { "epoch": 0.2536838332946856, "grad_norm": 1.1272544860839844, "learning_rate": 4.2270938729623386e-05, "loss": 0.9081, "step": 6017 }, { "epoch": 0.25372599447688515, "grad_norm": 0.564605176448822, "learning_rate": 4.227796514896009e-05, "loss": 0.3386, "step": 6018 }, { "epoch": 0.25376815565908467, "grad_norm": 0.33032411336898804, "learning_rate": 4.22849915682968e-05, "loss": 0.153, "step": 6019 }, { "epoch": 0.25381031684128424, "grad_norm": 0.5101647973060608, "learning_rate": 4.2292017987633506e-05, "loss": 0.3582, "step": 6020 }, { "epoch": 0.25385247802348376, "grad_norm": 0.7899696230888367, "learning_rate": 4.229904440697021e-05, "loss": 1.0652, "step": 6021 }, { "epoch": 0.25389463920568334, "grad_norm": 2.6899940967559814, "learning_rate": 4.2306070826306914e-05, "loss": 0.683, "step": 6022 }, { "epoch": 0.25393680038788286, "grad_norm": 0.3903101682662964, "learning_rate": 4.2313097245643625e-05, "loss": 0.1295, "step": 6023 }, { "epoch": 0.25397896157008243, "grad_norm": 0.49988430738449097, "learning_rate": 4.232012366498033e-05, "loss": 0.1825, "step": 6024 }, { "epoch": 0.25402112275228195, "grad_norm": 3.3883795738220215, "learning_rate": 4.2327150084317034e-05, "loss": 0.8887, "step": 6025 }, { "epoch": 0.2540632839344815, "grad_norm": 0.7299314141273499, "learning_rate": 4.233417650365374e-05, "loss": 0.1554, "step": 6026 }, { "epoch": 0.2541054451166811, "grad_norm": 0.5432319641113281, "learning_rate": 4.234120292299045e-05, "loss": 0.1074, "step": 6027 }, { "epoch": 0.2541476062988806, "grad_norm": 0.4661187529563904, "learning_rate": 4.234822934232715e-05, "loss": 0.0812, "step": 6028 }, { "epoch": 0.2541897674810802, "grad_norm": 2.4910888671875, "learning_rate": 4.235525576166386e-05, "loss": 0.5455, "step": 6029 }, { "epoch": 0.2542319286632797, "grad_norm": 0.6543260812759399, "learning_rate": 4.236228218100056e-05, "loss": 0.4845, "step": 6030 }, { "epoch": 0.2542740898454793, "grad_norm": 0.4369950592517853, "learning_rate": 4.236930860033727e-05, "loss": 0.1973, "step": 6031 }, { "epoch": 0.2543162510276788, "grad_norm": 0.5941952466964722, "learning_rate": 4.237633501967397e-05, "loss": 0.2044, "step": 6032 }, { "epoch": 0.25435841220987837, "grad_norm": 0.6496599316596985, "learning_rate": 4.238336143901068e-05, "loss": 0.6554, "step": 6033 }, { "epoch": 0.2544005733920779, "grad_norm": 0.6225771903991699, "learning_rate": 4.239038785834739e-05, "loss": 0.4167, "step": 6034 }, { "epoch": 0.25444273457427746, "grad_norm": 0.5036256313323975, "learning_rate": 4.239741427768409e-05, "loss": 0.1801, "step": 6035 }, { "epoch": 0.25448489575647704, "grad_norm": 0.8839945793151855, "learning_rate": 4.2404440697020796e-05, "loss": 0.3042, "step": 6036 }, { "epoch": 0.25452705693867655, "grad_norm": 0.5007217526435852, "learning_rate": 4.241146711635751e-05, "loss": 0.2131, "step": 6037 }, { "epoch": 0.25456921812087613, "grad_norm": 0.9845647215843201, "learning_rate": 4.241849353569421e-05, "loss": 1.0449, "step": 6038 }, { "epoch": 0.25461137930307565, "grad_norm": 1.0172652006149292, "learning_rate": 4.2425519955030916e-05, "loss": 0.3793, "step": 6039 }, { "epoch": 0.2546535404852752, "grad_norm": 1.954176664352417, "learning_rate": 4.243254637436763e-05, "loss": 0.2005, "step": 6040 }, { "epoch": 0.25469570166747474, "grad_norm": 0.4074998199939728, "learning_rate": 4.243957279370433e-05, "loss": 0.1318, "step": 6041 }, { "epoch": 0.2547378628496743, "grad_norm": 0.9405093193054199, "learning_rate": 4.2446599213041036e-05, "loss": 0.8518, "step": 6042 }, { "epoch": 0.25478002403187383, "grad_norm": 0.4545576572418213, "learning_rate": 4.245362563237774e-05, "loss": 0.1362, "step": 6043 }, { "epoch": 0.2548221852140734, "grad_norm": 0.5425360202789307, "learning_rate": 4.246065205171445e-05, "loss": 0.1094, "step": 6044 }, { "epoch": 0.254864346396273, "grad_norm": 0.5126540064811707, "learning_rate": 4.2467678471051156e-05, "loss": 0.1605, "step": 6045 }, { "epoch": 0.2549065075784725, "grad_norm": 0.7513399720191956, "learning_rate": 4.247470489038786e-05, "loss": 0.4687, "step": 6046 }, { "epoch": 0.25494866876067207, "grad_norm": 0.46182921528816223, "learning_rate": 4.2481731309724565e-05, "loss": 0.2508, "step": 6047 }, { "epoch": 0.2549908299428716, "grad_norm": 2.7846615314483643, "learning_rate": 4.2488757729061276e-05, "loss": 0.6895, "step": 6048 }, { "epoch": 0.25503299112507116, "grad_norm": 1.2395988702774048, "learning_rate": 4.2495784148397974e-05, "loss": 1.0829, "step": 6049 }, { "epoch": 0.2550751523072707, "grad_norm": 0.7005373239517212, "learning_rate": 4.2502810567734685e-05, "loss": 0.4868, "step": 6050 }, { "epoch": 0.25511731348947025, "grad_norm": 0.6734073162078857, "learning_rate": 4.250983698707139e-05, "loss": 0.1521, "step": 6051 }, { "epoch": 0.2551594746716698, "grad_norm": 0.34308069944381714, "learning_rate": 4.25168634064081e-05, "loss": 0.0734, "step": 6052 }, { "epoch": 0.25520163585386935, "grad_norm": 0.5897279977798462, "learning_rate": 4.25238898257448e-05, "loss": 0.1947, "step": 6053 }, { "epoch": 0.25524379703606886, "grad_norm": 0.840733528137207, "learning_rate": 4.253091624508151e-05, "loss": 0.782, "step": 6054 }, { "epoch": 0.25528595821826844, "grad_norm": 0.7426961064338684, "learning_rate": 4.2537942664418214e-05, "loss": 0.2862, "step": 6055 }, { "epoch": 0.255328119400468, "grad_norm": 0.5085976123809814, "learning_rate": 4.254496908375492e-05, "loss": 0.1148, "step": 6056 }, { "epoch": 0.25537028058266753, "grad_norm": 1.3422874212265015, "learning_rate": 4.255199550309162e-05, "loss": 1.0705, "step": 6057 }, { "epoch": 0.2554124417648671, "grad_norm": 0.8567607998847961, "learning_rate": 4.2559021922428334e-05, "loss": 0.277, "step": 6058 }, { "epoch": 0.2554546029470666, "grad_norm": 1.4038755893707275, "learning_rate": 4.256604834176504e-05, "loss": 1.0115, "step": 6059 }, { "epoch": 0.2554967641292662, "grad_norm": 0.5292350053787231, "learning_rate": 4.257307476110174e-05, "loss": 0.2505, "step": 6060 }, { "epoch": 0.2555389253114657, "grad_norm": 0.7482162714004517, "learning_rate": 4.2580101180438453e-05, "loss": 0.5171, "step": 6061 }, { "epoch": 0.2555810864936653, "grad_norm": 4.674190998077393, "learning_rate": 4.258712759977516e-05, "loss": 1.0183, "step": 6062 }, { "epoch": 0.2556232476758648, "grad_norm": 0.7918773889541626, "learning_rate": 4.259415401911186e-05, "loss": 0.2043, "step": 6063 }, { "epoch": 0.2556654088580644, "grad_norm": 0.3724330961704254, "learning_rate": 4.260118043844857e-05, "loss": 0.0732, "step": 6064 }, { "epoch": 0.25570757004026395, "grad_norm": 0.9877192378044128, "learning_rate": 4.260820685778528e-05, "loss": 0.5365, "step": 6065 }, { "epoch": 0.25574973122246347, "grad_norm": 0.5140203833580017, "learning_rate": 4.261523327712198e-05, "loss": 0.143, "step": 6066 }, { "epoch": 0.25579189240466305, "grad_norm": 0.762412965297699, "learning_rate": 4.262225969645869e-05, "loss": 0.6128, "step": 6067 }, { "epoch": 0.25583405358686256, "grad_norm": 0.6113821864128113, "learning_rate": 4.262928611579539e-05, "loss": 0.4667, "step": 6068 }, { "epoch": 0.25587621476906214, "grad_norm": 0.4868859052658081, "learning_rate": 4.26363125351321e-05, "loss": 0.2315, "step": 6069 }, { "epoch": 0.25591837595126166, "grad_norm": 0.6376557946205139, "learning_rate": 4.2643338954468807e-05, "loss": 0.103, "step": 6070 }, { "epoch": 0.25596053713346123, "grad_norm": 1.078087329864502, "learning_rate": 4.265036537380551e-05, "loss": 0.2949, "step": 6071 }, { "epoch": 0.25600269831566075, "grad_norm": 1.0352121591567993, "learning_rate": 4.2657391793142215e-05, "loss": 0.2602, "step": 6072 }, { "epoch": 0.2560448594978603, "grad_norm": 0.494415819644928, "learning_rate": 4.2664418212478927e-05, "loss": 0.1956, "step": 6073 }, { "epoch": 0.2560870206800599, "grad_norm": 0.7396073341369629, "learning_rate": 4.2671444631815624e-05, "loss": 0.3304, "step": 6074 }, { "epoch": 0.2561291818622594, "grad_norm": 0.6578773856163025, "learning_rate": 4.2678471051152335e-05, "loss": 0.2025, "step": 6075 }, { "epoch": 0.256171343044459, "grad_norm": 0.5348513126373291, "learning_rate": 4.268549747048904e-05, "loss": 0.1988, "step": 6076 }, { "epoch": 0.2562135042266585, "grad_norm": 0.5887450575828552, "learning_rate": 4.2692523889825744e-05, "loss": 0.1268, "step": 6077 }, { "epoch": 0.2562556654088581, "grad_norm": 2.227935791015625, "learning_rate": 4.269955030916245e-05, "loss": 0.2319, "step": 6078 }, { "epoch": 0.2562978265910576, "grad_norm": 2.09114670753479, "learning_rate": 4.270657672849916e-05, "loss": 0.9884, "step": 6079 }, { "epoch": 0.25633998777325717, "grad_norm": 0.7832377552986145, "learning_rate": 4.2713603147835864e-05, "loss": 0.2587, "step": 6080 }, { "epoch": 0.2563821489554567, "grad_norm": 0.48396724462509155, "learning_rate": 4.272062956717257e-05, "loss": 0.0959, "step": 6081 }, { "epoch": 0.25642431013765626, "grad_norm": 0.6943108439445496, "learning_rate": 4.272765598650928e-05, "loss": 0.1957, "step": 6082 }, { "epoch": 0.2564664713198558, "grad_norm": 1.0088732242584229, "learning_rate": 4.2734682405845984e-05, "loss": 0.812, "step": 6083 }, { "epoch": 0.25650863250205536, "grad_norm": 1.079724907875061, "learning_rate": 4.274170882518269e-05, "loss": 0.8249, "step": 6084 }, { "epoch": 0.25655079368425493, "grad_norm": 0.699137270450592, "learning_rate": 4.274873524451939e-05, "loss": 0.0993, "step": 6085 }, { "epoch": 0.25659295486645445, "grad_norm": 0.8931631445884705, "learning_rate": 4.2755761663856104e-05, "loss": 0.3606, "step": 6086 }, { "epoch": 0.256635116048654, "grad_norm": 0.6027272343635559, "learning_rate": 4.276278808319281e-05, "loss": 0.3617, "step": 6087 }, { "epoch": 0.25667727723085354, "grad_norm": 0.6244955658912659, "learning_rate": 4.276981450252951e-05, "loss": 0.2374, "step": 6088 }, { "epoch": 0.2567194384130531, "grad_norm": 0.5786457657814026, "learning_rate": 4.277684092186622e-05, "loss": 0.4508, "step": 6089 }, { "epoch": 0.25676159959525263, "grad_norm": 0.6868933439254761, "learning_rate": 4.278386734120293e-05, "loss": 0.3247, "step": 6090 }, { "epoch": 0.2568037607774522, "grad_norm": 0.4441527724266052, "learning_rate": 4.279089376053963e-05, "loss": 0.2091, "step": 6091 }, { "epoch": 0.2568459219596517, "grad_norm": 0.5411871671676636, "learning_rate": 4.279792017987634e-05, "loss": 0.1569, "step": 6092 }, { "epoch": 0.2568880831418513, "grad_norm": 0.5619373321533203, "learning_rate": 4.280494659921304e-05, "loss": 0.1631, "step": 6093 }, { "epoch": 0.25693024432405087, "grad_norm": 0.8019326329231262, "learning_rate": 4.281197301854975e-05, "loss": 0.2057, "step": 6094 }, { "epoch": 0.2569724055062504, "grad_norm": 0.6831398010253906, "learning_rate": 4.281899943788645e-05, "loss": 0.61, "step": 6095 }, { "epoch": 0.25701456668844996, "grad_norm": 0.4507567882537842, "learning_rate": 4.282602585722316e-05, "loss": 0.1497, "step": 6096 }, { "epoch": 0.2570567278706495, "grad_norm": 0.9410064220428467, "learning_rate": 4.2833052276559866e-05, "loss": 0.9363, "step": 6097 }, { "epoch": 0.25709888905284906, "grad_norm": 0.49321287870407104, "learning_rate": 4.284007869589658e-05, "loss": 0.1923, "step": 6098 }, { "epoch": 0.2571410502350486, "grad_norm": 0.5389814376831055, "learning_rate": 4.2847105115233275e-05, "loss": 0.2253, "step": 6099 }, { "epoch": 0.25718321141724815, "grad_norm": 0.7741926312446594, "learning_rate": 4.2854131534569986e-05, "loss": 0.2536, "step": 6100 }, { "epoch": 0.25722537259944767, "grad_norm": 0.4332684874534607, "learning_rate": 4.286115795390669e-05, "loss": 0.2847, "step": 6101 }, { "epoch": 0.25726753378164724, "grad_norm": 0.7565579414367676, "learning_rate": 4.2868184373243395e-05, "loss": 0.3219, "step": 6102 }, { "epoch": 0.2573096949638468, "grad_norm": 0.6015923023223877, "learning_rate": 4.28752107925801e-05, "loss": 0.6181, "step": 6103 }, { "epoch": 0.25735185614604633, "grad_norm": 0.4818202555179596, "learning_rate": 4.288223721191681e-05, "loss": 0.1231, "step": 6104 }, { "epoch": 0.2573940173282459, "grad_norm": 0.6409415602684021, "learning_rate": 4.2889263631253515e-05, "loss": 0.4547, "step": 6105 }, { "epoch": 0.2574361785104454, "grad_norm": 0.9413270354270935, "learning_rate": 4.289629005059022e-05, "loss": 0.4491, "step": 6106 }, { "epoch": 0.257478339692645, "grad_norm": 0.6229709386825562, "learning_rate": 4.290331646992693e-05, "loss": 0.1561, "step": 6107 }, { "epoch": 0.2575205008748445, "grad_norm": 0.6323816180229187, "learning_rate": 4.2910342889263635e-05, "loss": 0.2452, "step": 6108 }, { "epoch": 0.2575626620570441, "grad_norm": 0.6779490113258362, "learning_rate": 4.291736930860034e-05, "loss": 0.2589, "step": 6109 }, { "epoch": 0.2576048232392436, "grad_norm": 0.9701946377754211, "learning_rate": 4.2924395727937043e-05, "loss": 0.821, "step": 6110 }, { "epoch": 0.2576469844214432, "grad_norm": 0.6069554686546326, "learning_rate": 4.2931422147273755e-05, "loss": 0.619, "step": 6111 }, { "epoch": 0.2576891456036427, "grad_norm": 0.9656544923782349, "learning_rate": 4.293844856661046e-05, "loss": 1.0088, "step": 6112 }, { "epoch": 0.2577313067858423, "grad_norm": 0.7595126032829285, "learning_rate": 4.294547498594716e-05, "loss": 0.3769, "step": 6113 }, { "epoch": 0.25777346796804185, "grad_norm": 1.1990482807159424, "learning_rate": 4.295250140528387e-05, "loss": 0.3124, "step": 6114 }, { "epoch": 0.25781562915024137, "grad_norm": 0.7363631129264832, "learning_rate": 4.295952782462058e-05, "loss": 0.3017, "step": 6115 }, { "epoch": 0.25785779033244094, "grad_norm": 0.9793179035186768, "learning_rate": 4.2966554243957277e-05, "loss": 0.427, "step": 6116 }, { "epoch": 0.25789995151464046, "grad_norm": 0.6182034015655518, "learning_rate": 4.297358066329399e-05, "loss": 0.4539, "step": 6117 }, { "epoch": 0.25794211269684003, "grad_norm": 0.45440128445625305, "learning_rate": 4.298060708263069e-05, "loss": 0.1518, "step": 6118 }, { "epoch": 0.25798427387903955, "grad_norm": 0.8778557777404785, "learning_rate": 4.29876335019674e-05, "loss": 0.3352, "step": 6119 }, { "epoch": 0.2580264350612391, "grad_norm": 0.4972408711910248, "learning_rate": 4.29946599213041e-05, "loss": 0.1361, "step": 6120 }, { "epoch": 0.25806859624343864, "grad_norm": 0.5391160845756531, "learning_rate": 4.300168634064081e-05, "loss": 0.1267, "step": 6121 }, { "epoch": 0.2581107574256382, "grad_norm": 0.46396222710609436, "learning_rate": 4.3008712759977516e-05, "loss": 0.0902, "step": 6122 }, { "epoch": 0.2581529186078378, "grad_norm": 0.42562803626060486, "learning_rate": 4.301573917931422e-05, "loss": 0.1078, "step": 6123 }, { "epoch": 0.2581950797900373, "grad_norm": 1.3617548942565918, "learning_rate": 4.3022765598650925e-05, "loss": 1.1371, "step": 6124 }, { "epoch": 0.2582372409722369, "grad_norm": 0.5528624653816223, "learning_rate": 4.3029792017987636e-05, "loss": 0.5813, "step": 6125 }, { "epoch": 0.2582794021544364, "grad_norm": 0.766695499420166, "learning_rate": 4.303681843732434e-05, "loss": 0.7526, "step": 6126 }, { "epoch": 0.258321563336636, "grad_norm": 0.40414440631866455, "learning_rate": 4.3043844856661045e-05, "loss": 0.1136, "step": 6127 }, { "epoch": 0.2583637245188355, "grad_norm": 0.6106020212173462, "learning_rate": 4.3050871275997756e-05, "loss": 0.423, "step": 6128 }, { "epoch": 0.25840588570103507, "grad_norm": 0.5966721773147583, "learning_rate": 4.305789769533446e-05, "loss": 0.2453, "step": 6129 }, { "epoch": 0.2584480468832346, "grad_norm": 0.44386744499206543, "learning_rate": 4.3064924114671165e-05, "loss": 0.1086, "step": 6130 }, { "epoch": 0.25849020806543416, "grad_norm": 0.40560540556907654, "learning_rate": 4.307195053400787e-05, "loss": 0.1063, "step": 6131 }, { "epoch": 0.25853236924763373, "grad_norm": 0.4687374532222748, "learning_rate": 4.307897695334458e-05, "loss": 0.1747, "step": 6132 }, { "epoch": 0.25857453042983325, "grad_norm": 0.573520302772522, "learning_rate": 4.3086003372681285e-05, "loss": 0.1383, "step": 6133 }, { "epoch": 0.2586166916120328, "grad_norm": 0.6870056986808777, "learning_rate": 4.309302979201799e-05, "loss": 0.1798, "step": 6134 }, { "epoch": 0.25865885279423234, "grad_norm": 0.4528743624687195, "learning_rate": 4.3100056211354694e-05, "loss": 0.1744, "step": 6135 }, { "epoch": 0.2587010139764319, "grad_norm": 0.5878273844718933, "learning_rate": 4.3107082630691405e-05, "loss": 0.2613, "step": 6136 }, { "epoch": 0.25874317515863143, "grad_norm": 0.41768163442611694, "learning_rate": 4.311410905002811e-05, "loss": 0.1501, "step": 6137 }, { "epoch": 0.258785336340831, "grad_norm": 0.7866160273551941, "learning_rate": 4.3121135469364814e-05, "loss": 0.5686, "step": 6138 }, { "epoch": 0.2588274975230305, "grad_norm": 0.37427568435668945, "learning_rate": 4.312816188870152e-05, "loss": 0.1009, "step": 6139 }, { "epoch": 0.2588696587052301, "grad_norm": 0.5761900544166565, "learning_rate": 4.313518830803823e-05, "loss": 0.3591, "step": 6140 }, { "epoch": 0.2589118198874296, "grad_norm": 0.8679367899894714, "learning_rate": 4.314221472737493e-05, "loss": 1.1508, "step": 6141 }, { "epoch": 0.2589539810696292, "grad_norm": 0.6367514133453369, "learning_rate": 4.314924114671164e-05, "loss": 0.2102, "step": 6142 }, { "epoch": 0.25899614225182876, "grad_norm": 0.49910449981689453, "learning_rate": 4.315626756604834e-05, "loss": 0.1117, "step": 6143 }, { "epoch": 0.2590383034340283, "grad_norm": 0.4950942397117615, "learning_rate": 4.3163293985385054e-05, "loss": 0.4687, "step": 6144 }, { "epoch": 0.25908046461622786, "grad_norm": 0.993996798992157, "learning_rate": 4.317032040472175e-05, "loss": 0.1902, "step": 6145 }, { "epoch": 0.2591226257984274, "grad_norm": 1.0831692218780518, "learning_rate": 4.317734682405846e-05, "loss": 0.2843, "step": 6146 }, { "epoch": 0.25916478698062695, "grad_norm": 0.8316451907157898, "learning_rate": 4.318437324339517e-05, "loss": 1.0679, "step": 6147 }, { "epoch": 0.25920694816282647, "grad_norm": 4.263582706451416, "learning_rate": 4.319139966273187e-05, "loss": 0.4225, "step": 6148 }, { "epoch": 0.25924910934502604, "grad_norm": 0.5556440949440002, "learning_rate": 4.319842608206858e-05, "loss": 0.1422, "step": 6149 }, { "epoch": 0.25929127052722556, "grad_norm": 0.5722894668579102, "learning_rate": 4.320545250140529e-05, "loss": 0.2203, "step": 6150 }, { "epoch": 0.25933343170942513, "grad_norm": 0.712559163570404, "learning_rate": 4.321247892074199e-05, "loss": 0.6298, "step": 6151 }, { "epoch": 0.2593755928916247, "grad_norm": 0.7332544922828674, "learning_rate": 4.3219505340078696e-05, "loss": 0.2661, "step": 6152 }, { "epoch": 0.2594177540738242, "grad_norm": 0.6329300403594971, "learning_rate": 4.322653175941541e-05, "loss": 0.315, "step": 6153 }, { "epoch": 0.2594599152560238, "grad_norm": 0.6288333535194397, "learning_rate": 4.323355817875211e-05, "loss": 0.7548, "step": 6154 }, { "epoch": 0.2595020764382233, "grad_norm": 1.3551644086837769, "learning_rate": 4.3240584598088816e-05, "loss": 0.3165, "step": 6155 }, { "epoch": 0.2595442376204229, "grad_norm": 1.0474234819412231, "learning_rate": 4.324761101742552e-05, "loss": 0.9177, "step": 6156 }, { "epoch": 0.2595863988026224, "grad_norm": 0.7916163802146912, "learning_rate": 4.325463743676223e-05, "loss": 0.2311, "step": 6157 }, { "epoch": 0.259628559984822, "grad_norm": 0.48133957386016846, "learning_rate": 4.3261663856098936e-05, "loss": 0.0808, "step": 6158 }, { "epoch": 0.2596707211670215, "grad_norm": 1.4925048351287842, "learning_rate": 4.326869027543564e-05, "loss": 0.3717, "step": 6159 }, { "epoch": 0.2597128823492211, "grad_norm": 0.8829246163368225, "learning_rate": 4.3275716694772344e-05, "loss": 0.6201, "step": 6160 }, { "epoch": 0.25975504353142065, "grad_norm": 0.6437954306602478, "learning_rate": 4.3282743114109056e-05, "loss": 0.4076, "step": 6161 }, { "epoch": 0.25979720471362017, "grad_norm": 1.0409834384918213, "learning_rate": 4.328976953344575e-05, "loss": 0.7107, "step": 6162 }, { "epoch": 0.25983936589581974, "grad_norm": 0.3345072567462921, "learning_rate": 4.3296795952782464e-05, "loss": 0.0881, "step": 6163 }, { "epoch": 0.25988152707801926, "grad_norm": 0.36016473174095154, "learning_rate": 4.330382237211917e-05, "loss": 0.085, "step": 6164 }, { "epoch": 0.25992368826021883, "grad_norm": 0.3152700364589691, "learning_rate": 4.331084879145588e-05, "loss": 0.1188, "step": 6165 }, { "epoch": 0.25996584944241835, "grad_norm": 1.287679672241211, "learning_rate": 4.331787521079258e-05, "loss": 0.3937, "step": 6166 }, { "epoch": 0.2600080106246179, "grad_norm": 0.29730096459388733, "learning_rate": 4.332490163012929e-05, "loss": 0.1458, "step": 6167 }, { "epoch": 0.26005017180681744, "grad_norm": 0.6225937008857727, "learning_rate": 4.333192804946599e-05, "loss": 0.1368, "step": 6168 }, { "epoch": 0.260092332989017, "grad_norm": 1.6447505950927734, "learning_rate": 4.33389544688027e-05, "loss": 1.0367, "step": 6169 }, { "epoch": 0.26013449417121653, "grad_norm": 0.34474390745162964, "learning_rate": 4.33459808881394e-05, "loss": 0.093, "step": 6170 }, { "epoch": 0.2601766553534161, "grad_norm": 1.0177613496780396, "learning_rate": 4.335300730747611e-05, "loss": 0.9903, "step": 6171 }, { "epoch": 0.2602188165356157, "grad_norm": 0.6033234000205994, "learning_rate": 4.3360033726812824e-05, "loss": 0.2043, "step": 6172 }, { "epoch": 0.2602609777178152, "grad_norm": 0.568184494972229, "learning_rate": 4.336706014614952e-05, "loss": 0.1069, "step": 6173 }, { "epoch": 0.2603031389000148, "grad_norm": 0.7454269528388977, "learning_rate": 4.337408656548623e-05, "loss": 0.5546, "step": 6174 }, { "epoch": 0.2603453000822143, "grad_norm": 0.6168844699859619, "learning_rate": 4.338111298482294e-05, "loss": 0.2162, "step": 6175 }, { "epoch": 0.26038746126441387, "grad_norm": 4.011287212371826, "learning_rate": 4.338813940415964e-05, "loss": 0.697, "step": 6176 }, { "epoch": 0.2604296224466134, "grad_norm": 1.0238627195358276, "learning_rate": 4.3395165823496346e-05, "loss": 0.7314, "step": 6177 }, { "epoch": 0.26047178362881296, "grad_norm": 0.6721169948577881, "learning_rate": 4.340219224283306e-05, "loss": 0.7182, "step": 6178 }, { "epoch": 0.2605139448110125, "grad_norm": 0.39017364382743835, "learning_rate": 4.340921866216976e-05, "loss": 0.1779, "step": 6179 }, { "epoch": 0.26055610599321205, "grad_norm": 0.962357223033905, "learning_rate": 4.3416245081506466e-05, "loss": 1.0583, "step": 6180 }, { "epoch": 0.2605982671754116, "grad_norm": 0.8117133378982544, "learning_rate": 4.342327150084317e-05, "loss": 0.2797, "step": 6181 }, { "epoch": 0.26064042835761114, "grad_norm": 0.413066029548645, "learning_rate": 4.343029792017988e-05, "loss": 0.0872, "step": 6182 }, { "epoch": 0.2606825895398107, "grad_norm": 0.7779135704040527, "learning_rate": 4.3437324339516586e-05, "loss": 0.4986, "step": 6183 }, { "epoch": 0.26072475072201023, "grad_norm": 0.7376607060432434, "learning_rate": 4.344435075885329e-05, "loss": 0.1389, "step": 6184 }, { "epoch": 0.2607669119042098, "grad_norm": 0.6389741897583008, "learning_rate": 4.3451377178189995e-05, "loss": 0.6, "step": 6185 }, { "epoch": 0.2608090730864093, "grad_norm": 0.5328117609024048, "learning_rate": 4.3458403597526706e-05, "loss": 0.1265, "step": 6186 }, { "epoch": 0.2608512342686089, "grad_norm": 1.619864821434021, "learning_rate": 4.3465430016863404e-05, "loss": 0.7787, "step": 6187 }, { "epoch": 0.2608933954508084, "grad_norm": 0.7214027643203735, "learning_rate": 4.3472456436200115e-05, "loss": 0.665, "step": 6188 }, { "epoch": 0.260935556633008, "grad_norm": 0.8125354051589966, "learning_rate": 4.347948285553682e-05, "loss": 0.7243, "step": 6189 }, { "epoch": 0.26097771781520757, "grad_norm": 0.7750635743141174, "learning_rate": 4.3486509274873524e-05, "loss": 0.5739, "step": 6190 }, { "epoch": 0.2610198789974071, "grad_norm": 0.43327173590660095, "learning_rate": 4.349353569421023e-05, "loss": 0.1297, "step": 6191 }, { "epoch": 0.26106204017960666, "grad_norm": 0.8057484030723572, "learning_rate": 4.350056211354694e-05, "loss": 0.1795, "step": 6192 }, { "epoch": 0.2611042013618062, "grad_norm": 1.15708589553833, "learning_rate": 4.3507588532883644e-05, "loss": 0.1807, "step": 6193 }, { "epoch": 0.26114636254400575, "grad_norm": 0.5434778332710266, "learning_rate": 4.351461495222035e-05, "loss": 0.4374, "step": 6194 }, { "epoch": 0.26118852372620527, "grad_norm": 0.745296835899353, "learning_rate": 4.352164137155706e-05, "loss": 0.2394, "step": 6195 }, { "epoch": 0.26123068490840484, "grad_norm": 0.5007538199424744, "learning_rate": 4.3528667790893764e-05, "loss": 0.163, "step": 6196 }, { "epoch": 0.26127284609060436, "grad_norm": 0.7754333019256592, "learning_rate": 4.353569421023047e-05, "loss": 0.1435, "step": 6197 }, { "epoch": 0.26131500727280393, "grad_norm": 0.8480284810066223, "learning_rate": 4.354272062956717e-05, "loss": 0.3061, "step": 6198 }, { "epoch": 0.26135716845500345, "grad_norm": 0.5498793125152588, "learning_rate": 4.3549747048903884e-05, "loss": 0.1733, "step": 6199 }, { "epoch": 0.261399329637203, "grad_norm": 0.5904689431190491, "learning_rate": 4.355677346824059e-05, "loss": 0.1981, "step": 6200 }, { "epoch": 0.2614414908194026, "grad_norm": 1.7955491542816162, "learning_rate": 4.356379988757729e-05, "loss": 0.3093, "step": 6201 }, { "epoch": 0.2614836520016021, "grad_norm": 1.0500712394714355, "learning_rate": 4.3570826306914e-05, "loss": 0.4393, "step": 6202 }, { "epoch": 0.2615258131838017, "grad_norm": 1.011404037475586, "learning_rate": 4.357785272625071e-05, "loss": 0.44, "step": 6203 }, { "epoch": 0.2615679743660012, "grad_norm": 0.4892173707485199, "learning_rate": 4.358487914558741e-05, "loss": 0.0929, "step": 6204 }, { "epoch": 0.2616101355482008, "grad_norm": 0.571679949760437, "learning_rate": 4.359190556492412e-05, "loss": 0.2369, "step": 6205 }, { "epoch": 0.2616522967304003, "grad_norm": 0.6165667772293091, "learning_rate": 4.359893198426082e-05, "loss": 0.2268, "step": 6206 }, { "epoch": 0.2616944579125999, "grad_norm": 0.9580693244934082, "learning_rate": 4.360595840359753e-05, "loss": 0.7623, "step": 6207 }, { "epoch": 0.2617366190947994, "grad_norm": 0.5724450349807739, "learning_rate": 4.361298482293423e-05, "loss": 0.5972, "step": 6208 }, { "epoch": 0.26177878027699897, "grad_norm": 0.45004338026046753, "learning_rate": 4.362001124227094e-05, "loss": 0.241, "step": 6209 }, { "epoch": 0.26182094145919854, "grad_norm": 0.4824538826942444, "learning_rate": 4.3627037661607646e-05, "loss": 0.157, "step": 6210 }, { "epoch": 0.26186310264139806, "grad_norm": 1.3132401704788208, "learning_rate": 4.363406408094436e-05, "loss": 0.5727, "step": 6211 }, { "epoch": 0.26190526382359763, "grad_norm": 0.7067558765411377, "learning_rate": 4.3641090500281054e-05, "loss": 0.1436, "step": 6212 }, { "epoch": 0.26194742500579715, "grad_norm": 0.3667779564857483, "learning_rate": 4.3648116919617765e-05, "loss": 0.1599, "step": 6213 }, { "epoch": 0.2619895861879967, "grad_norm": 0.5880251526832581, "learning_rate": 4.365514333895447e-05, "loss": 0.2979, "step": 6214 }, { "epoch": 0.26203174737019624, "grad_norm": 0.6852615475654602, "learning_rate": 4.3662169758291174e-05, "loss": 0.5482, "step": 6215 }, { "epoch": 0.2620739085523958, "grad_norm": 0.7193609476089478, "learning_rate": 4.366919617762788e-05, "loss": 0.4081, "step": 6216 }, { "epoch": 0.26211606973459534, "grad_norm": 0.7433097958564758, "learning_rate": 4.367622259696459e-05, "loss": 0.5893, "step": 6217 }, { "epoch": 0.2621582309167949, "grad_norm": 0.5453566312789917, "learning_rate": 4.3683249016301294e-05, "loss": 0.2901, "step": 6218 }, { "epoch": 0.26220039209899443, "grad_norm": 1.1161154508590698, "learning_rate": 4.3690275435638e-05, "loss": 1.1622, "step": 6219 }, { "epoch": 0.262242553281194, "grad_norm": 0.4446624517440796, "learning_rate": 4.369730185497471e-05, "loss": 0.117, "step": 6220 }, { "epoch": 0.2622847144633936, "grad_norm": 0.7536635398864746, "learning_rate": 4.3704328274311414e-05, "loss": 0.2979, "step": 6221 }, { "epoch": 0.2623268756455931, "grad_norm": 1.1571385860443115, "learning_rate": 4.371135469364812e-05, "loss": 0.9734, "step": 6222 }, { "epoch": 0.26236903682779267, "grad_norm": 0.46918559074401855, "learning_rate": 4.371838111298482e-05, "loss": 0.1597, "step": 6223 }, { "epoch": 0.2624111980099922, "grad_norm": 0.33259138464927673, "learning_rate": 4.3725407532321534e-05, "loss": 0.0761, "step": 6224 }, { "epoch": 0.26245335919219176, "grad_norm": 0.6346290111541748, "learning_rate": 4.373243395165824e-05, "loss": 0.4708, "step": 6225 }, { "epoch": 0.2624955203743913, "grad_norm": 0.7027374505996704, "learning_rate": 4.373946037099494e-05, "loss": 0.7327, "step": 6226 }, { "epoch": 0.26253768155659085, "grad_norm": 0.40880271792411804, "learning_rate": 4.374648679033165e-05, "loss": 0.1932, "step": 6227 }, { "epoch": 0.26257984273879037, "grad_norm": 0.5015217065811157, "learning_rate": 4.375351320966836e-05, "loss": 0.1557, "step": 6228 }, { "epoch": 0.26262200392098994, "grad_norm": 0.3757336437702179, "learning_rate": 4.376053962900506e-05, "loss": 0.1167, "step": 6229 }, { "epoch": 0.2626641651031895, "grad_norm": 0.626047670841217, "learning_rate": 4.376756604834177e-05, "loss": 0.308, "step": 6230 }, { "epoch": 0.26270632628538904, "grad_norm": 1.404371976852417, "learning_rate": 4.377459246767847e-05, "loss": 0.4821, "step": 6231 }, { "epoch": 0.2627484874675886, "grad_norm": 0.7082920670509338, "learning_rate": 4.378161888701518e-05, "loss": 0.3397, "step": 6232 }, { "epoch": 0.26279064864978813, "grad_norm": 0.34191495180130005, "learning_rate": 4.378864530635188e-05, "loss": 0.1342, "step": 6233 }, { "epoch": 0.2628328098319877, "grad_norm": 0.5055540800094604, "learning_rate": 4.379567172568859e-05, "loss": 0.1213, "step": 6234 }, { "epoch": 0.2628749710141872, "grad_norm": 0.45737284421920776, "learning_rate": 4.3802698145025296e-05, "loss": 0.0878, "step": 6235 }, { "epoch": 0.2629171321963868, "grad_norm": 0.4912269413471222, "learning_rate": 4.3809724564362e-05, "loss": 0.3319, "step": 6236 }, { "epoch": 0.2629592933785863, "grad_norm": 0.6612973809242249, "learning_rate": 4.3816750983698705e-05, "loss": 0.1773, "step": 6237 }, { "epoch": 0.2630014545607859, "grad_norm": 0.35656389594078064, "learning_rate": 4.3823777403035416e-05, "loss": 0.1307, "step": 6238 }, { "epoch": 0.26304361574298546, "grad_norm": 0.3542029857635498, "learning_rate": 4.383080382237213e-05, "loss": 0.1576, "step": 6239 }, { "epoch": 0.263085776925185, "grad_norm": 0.5090293884277344, "learning_rate": 4.3837830241708825e-05, "loss": 0.3445, "step": 6240 }, { "epoch": 0.26312793810738455, "grad_norm": 0.732986330986023, "learning_rate": 4.3844856661045536e-05, "loss": 0.305, "step": 6241 }, { "epoch": 0.26317009928958407, "grad_norm": 0.702196478843689, "learning_rate": 4.385188308038224e-05, "loss": 0.2777, "step": 6242 }, { "epoch": 0.26321226047178364, "grad_norm": 1.6900070905685425, "learning_rate": 4.3858909499718945e-05, "loss": 0.5337, "step": 6243 }, { "epoch": 0.26325442165398316, "grad_norm": 0.680332601070404, "learning_rate": 4.386593591905565e-05, "loss": 0.5741, "step": 6244 }, { "epoch": 0.26329658283618274, "grad_norm": 1.427438735961914, "learning_rate": 4.387296233839236e-05, "loss": 0.9492, "step": 6245 }, { "epoch": 0.26333874401838225, "grad_norm": 0.967252254486084, "learning_rate": 4.3879988757729065e-05, "loss": 0.4903, "step": 6246 }, { "epoch": 0.2633809052005818, "grad_norm": 0.6432284712791443, "learning_rate": 4.388701517706577e-05, "loss": 0.1676, "step": 6247 }, { "epoch": 0.26342306638278135, "grad_norm": 0.49446073174476624, "learning_rate": 4.3894041596402474e-05, "loss": 0.2402, "step": 6248 }, { "epoch": 0.2634652275649809, "grad_norm": 0.7638348937034607, "learning_rate": 4.3901068015739185e-05, "loss": 0.2782, "step": 6249 }, { "epoch": 0.2635073887471805, "grad_norm": 0.6995354890823364, "learning_rate": 4.390809443507589e-05, "loss": 0.746, "step": 6250 }, { "epoch": 0.26354954992938, "grad_norm": 0.7253419756889343, "learning_rate": 4.3915120854412593e-05, "loss": 0.1805, "step": 6251 }, { "epoch": 0.2635917111115796, "grad_norm": 0.9451385140419006, "learning_rate": 4.39221472737493e-05, "loss": 0.1783, "step": 6252 }, { "epoch": 0.2636338722937791, "grad_norm": 1.004997730255127, "learning_rate": 4.392917369308601e-05, "loss": 1.306, "step": 6253 }, { "epoch": 0.2636760334759787, "grad_norm": 0.5980440378189087, "learning_rate": 4.393620011242271e-05, "loss": 0.9599, "step": 6254 }, { "epoch": 0.2637181946581782, "grad_norm": 0.5703901648521423, "learning_rate": 4.394322653175942e-05, "loss": 0.1579, "step": 6255 }, { "epoch": 0.26376035584037777, "grad_norm": 0.5059590935707092, "learning_rate": 4.395025295109612e-05, "loss": 0.1186, "step": 6256 }, { "epoch": 0.2638025170225773, "grad_norm": 0.5170095562934875, "learning_rate": 4.3957279370432833e-05, "loss": 0.1717, "step": 6257 }, { "epoch": 0.26384467820477686, "grad_norm": 0.64984130859375, "learning_rate": 4.396430578976953e-05, "loss": 0.6491, "step": 6258 }, { "epoch": 0.26388683938697644, "grad_norm": 0.3775659203529358, "learning_rate": 4.397133220910624e-05, "loss": 0.0965, "step": 6259 }, { "epoch": 0.26392900056917595, "grad_norm": 0.7340341806411743, "learning_rate": 4.3978358628442947e-05, "loss": 0.2753, "step": 6260 }, { "epoch": 0.2639711617513755, "grad_norm": 0.5286588072776794, "learning_rate": 4.398538504777965e-05, "loss": 0.1829, "step": 6261 }, { "epoch": 0.26401332293357505, "grad_norm": 1.1425999402999878, "learning_rate": 4.399241146711636e-05, "loss": 0.514, "step": 6262 }, { "epoch": 0.2640554841157746, "grad_norm": 0.48340773582458496, "learning_rate": 4.3999437886453067e-05, "loss": 0.1271, "step": 6263 }, { "epoch": 0.26409764529797414, "grad_norm": 0.6284737586975098, "learning_rate": 4.400646430578977e-05, "loss": 0.6506, "step": 6264 }, { "epoch": 0.2641398064801737, "grad_norm": 0.9772236347198486, "learning_rate": 4.4013490725126475e-05, "loss": 0.7788, "step": 6265 }, { "epoch": 0.26418196766237323, "grad_norm": 0.5396114587783813, "learning_rate": 4.4020517144463187e-05, "loss": 0.1838, "step": 6266 }, { "epoch": 0.2642241288445728, "grad_norm": 0.9150129556655884, "learning_rate": 4.402754356379989e-05, "loss": 0.8292, "step": 6267 }, { "epoch": 0.2642662900267724, "grad_norm": 1.0769577026367188, "learning_rate": 4.4034569983136595e-05, "loss": 0.3537, "step": 6268 }, { "epoch": 0.2643084512089719, "grad_norm": 0.40994158387184143, "learning_rate": 4.40415964024733e-05, "loss": 0.1343, "step": 6269 }, { "epoch": 0.26435061239117147, "grad_norm": 1.7152823209762573, "learning_rate": 4.404862282181001e-05, "loss": 0.864, "step": 6270 }, { "epoch": 0.264392773573371, "grad_norm": 1.1154643297195435, "learning_rate": 4.4055649241146715e-05, "loss": 1.1655, "step": 6271 }, { "epoch": 0.26443493475557056, "grad_norm": 0.6487274169921875, "learning_rate": 4.406267566048342e-05, "loss": 0.3029, "step": 6272 }, { "epoch": 0.2644770959377701, "grad_norm": 0.3385303020477295, "learning_rate": 4.4069702079820124e-05, "loss": 0.135, "step": 6273 }, { "epoch": 0.26451925711996965, "grad_norm": 0.6389906406402588, "learning_rate": 4.4076728499156835e-05, "loss": 0.6799, "step": 6274 }, { "epoch": 0.26456141830216917, "grad_norm": 0.5780599117279053, "learning_rate": 4.408375491849353e-05, "loss": 0.1982, "step": 6275 }, { "epoch": 0.26460357948436874, "grad_norm": 0.6962350606918335, "learning_rate": 4.4090781337830244e-05, "loss": 0.4782, "step": 6276 }, { "epoch": 0.26464574066656826, "grad_norm": 1.2556253671646118, "learning_rate": 4.409780775716695e-05, "loss": 0.9046, "step": 6277 }, { "epoch": 0.26468790184876784, "grad_norm": 0.6819780468940735, "learning_rate": 4.410483417650366e-05, "loss": 0.1604, "step": 6278 }, { "epoch": 0.2647300630309674, "grad_norm": 0.5393187999725342, "learning_rate": 4.411186059584036e-05, "loss": 0.2365, "step": 6279 }, { "epoch": 0.26477222421316693, "grad_norm": 1.7328318357467651, "learning_rate": 4.411888701517707e-05, "loss": 0.4242, "step": 6280 }, { "epoch": 0.2648143853953665, "grad_norm": 0.8841685652732849, "learning_rate": 4.412591343451377e-05, "loss": 0.3704, "step": 6281 }, { "epoch": 0.264856546577566, "grad_norm": 0.5664489269256592, "learning_rate": 4.413293985385048e-05, "loss": 0.2517, "step": 6282 }, { "epoch": 0.2648987077597656, "grad_norm": 0.4166731834411621, "learning_rate": 4.413996627318718e-05, "loss": 0.1294, "step": 6283 }, { "epoch": 0.2649408689419651, "grad_norm": 0.5324485898017883, "learning_rate": 4.414699269252389e-05, "loss": 0.1979, "step": 6284 }, { "epoch": 0.2649830301241647, "grad_norm": 0.6699258685112, "learning_rate": 4.4154019111860604e-05, "loss": 0.0976, "step": 6285 }, { "epoch": 0.2650251913063642, "grad_norm": 0.4618164598941803, "learning_rate": 4.41610455311973e-05, "loss": 0.1888, "step": 6286 }, { "epoch": 0.2650673524885638, "grad_norm": 1.4975457191467285, "learning_rate": 4.416807195053401e-05, "loss": 0.4682, "step": 6287 }, { "epoch": 0.26510951367076335, "grad_norm": 0.6265299916267395, "learning_rate": 4.417509836987072e-05, "loss": 0.4082, "step": 6288 }, { "epoch": 0.26515167485296287, "grad_norm": 0.49243709444999695, "learning_rate": 4.418212478920742e-05, "loss": 0.1787, "step": 6289 }, { "epoch": 0.26519383603516244, "grad_norm": 0.4548352360725403, "learning_rate": 4.4189151208544126e-05, "loss": 0.13, "step": 6290 }, { "epoch": 0.26523599721736196, "grad_norm": 0.8123786449432373, "learning_rate": 4.419617762788084e-05, "loss": 0.4521, "step": 6291 }, { "epoch": 0.26527815839956154, "grad_norm": 0.5613874793052673, "learning_rate": 4.420320404721754e-05, "loss": 0.4562, "step": 6292 }, { "epoch": 0.26532031958176105, "grad_norm": 0.44280746579170227, "learning_rate": 4.4210230466554246e-05, "loss": 0.1413, "step": 6293 }, { "epoch": 0.26536248076396063, "grad_norm": 0.5983189344406128, "learning_rate": 4.421725688589095e-05, "loss": 0.5449, "step": 6294 }, { "epoch": 0.26540464194616015, "grad_norm": 0.5602157711982727, "learning_rate": 4.422428330522766e-05, "loss": 0.1937, "step": 6295 }, { "epoch": 0.2654468031283597, "grad_norm": 1.141709327697754, "learning_rate": 4.4231309724564366e-05, "loss": 1.0938, "step": 6296 }, { "epoch": 0.2654889643105593, "grad_norm": 1.1842409372329712, "learning_rate": 4.423833614390107e-05, "loss": 0.7494, "step": 6297 }, { "epoch": 0.2655311254927588, "grad_norm": 0.552038311958313, "learning_rate": 4.4245362563237775e-05, "loss": 0.1352, "step": 6298 }, { "epoch": 0.2655732866749584, "grad_norm": 0.6426816582679749, "learning_rate": 4.4252388982574486e-05, "loss": 0.2758, "step": 6299 }, { "epoch": 0.2656154478571579, "grad_norm": 1.8000632524490356, "learning_rate": 4.4259415401911183e-05, "loss": 0.9428, "step": 6300 }, { "epoch": 0.2656576090393575, "grad_norm": 0.6136403679847717, "learning_rate": 4.4266441821247895e-05, "loss": 0.1443, "step": 6301 }, { "epoch": 0.265699770221557, "grad_norm": 1.4446946382522583, "learning_rate": 4.42734682405846e-05, "loss": 1.1074, "step": 6302 }, { "epoch": 0.26574193140375657, "grad_norm": 0.5195522904396057, "learning_rate": 4.42804946599213e-05, "loss": 0.5915, "step": 6303 }, { "epoch": 0.2657840925859561, "grad_norm": 0.4776391386985779, "learning_rate": 4.428752107925801e-05, "loss": 0.2957, "step": 6304 }, { "epoch": 0.26582625376815566, "grad_norm": 0.5382117629051208, "learning_rate": 4.429454749859472e-05, "loss": 0.1008, "step": 6305 }, { "epoch": 0.2658684149503552, "grad_norm": 0.9559544920921326, "learning_rate": 4.430157391793142e-05, "loss": 0.4561, "step": 6306 }, { "epoch": 0.26591057613255475, "grad_norm": 0.7047770023345947, "learning_rate": 4.430860033726813e-05, "loss": 0.303, "step": 6307 }, { "epoch": 0.26595273731475433, "grad_norm": 0.47551465034484863, "learning_rate": 4.431562675660484e-05, "loss": 0.1362, "step": 6308 }, { "epoch": 0.26599489849695385, "grad_norm": 0.8636555075645447, "learning_rate": 4.432265317594154e-05, "loss": 0.3069, "step": 6309 }, { "epoch": 0.2660370596791534, "grad_norm": 0.4108597934246063, "learning_rate": 4.432967959527825e-05, "loss": 0.2032, "step": 6310 }, { "epoch": 0.26607922086135294, "grad_norm": 0.7024213075637817, "learning_rate": 4.433670601461495e-05, "loss": 0.325, "step": 6311 }, { "epoch": 0.2661213820435525, "grad_norm": 1.2267180681228638, "learning_rate": 4.434373243395166e-05, "loss": 1.0724, "step": 6312 }, { "epoch": 0.26616354322575203, "grad_norm": 0.42108574509620667, "learning_rate": 4.435075885328837e-05, "loss": 0.1266, "step": 6313 }, { "epoch": 0.2662057044079516, "grad_norm": 1.6708208322525024, "learning_rate": 4.435778527262507e-05, "loss": 0.6152, "step": 6314 }, { "epoch": 0.2662478655901511, "grad_norm": 0.5485541224479675, "learning_rate": 4.4364811691961776e-05, "loss": 0.1383, "step": 6315 }, { "epoch": 0.2662900267723507, "grad_norm": 0.5635025501251221, "learning_rate": 4.437183811129849e-05, "loss": 0.3914, "step": 6316 }, { "epoch": 0.26633218795455027, "grad_norm": 0.5090409517288208, "learning_rate": 4.437886453063519e-05, "loss": 0.1985, "step": 6317 }, { "epoch": 0.2663743491367498, "grad_norm": 0.4051845371723175, "learning_rate": 4.4385890949971896e-05, "loss": 0.0994, "step": 6318 }, { "epoch": 0.26641651031894936, "grad_norm": 0.7409703731536865, "learning_rate": 4.43929173693086e-05, "loss": 0.3002, "step": 6319 }, { "epoch": 0.2664586715011489, "grad_norm": 0.6858595013618469, "learning_rate": 4.439994378864531e-05, "loss": 0.2461, "step": 6320 }, { "epoch": 0.26650083268334845, "grad_norm": 0.37197497487068176, "learning_rate": 4.440697020798201e-05, "loss": 0.1325, "step": 6321 }, { "epoch": 0.266542993865548, "grad_norm": 1.2208342552185059, "learning_rate": 4.441399662731872e-05, "loss": 1.173, "step": 6322 }, { "epoch": 0.26658515504774755, "grad_norm": 1.2224562168121338, "learning_rate": 4.4421023046655425e-05, "loss": 1.1742, "step": 6323 }, { "epoch": 0.26662731622994706, "grad_norm": 0.8582702279090881, "learning_rate": 4.4428049465992136e-05, "loss": 0.5836, "step": 6324 }, { "epoch": 0.26666947741214664, "grad_norm": 0.3634903132915497, "learning_rate": 4.4435075885328834e-05, "loss": 0.1152, "step": 6325 }, { "epoch": 0.2667116385943462, "grad_norm": 0.3170793652534485, "learning_rate": 4.4442102304665545e-05, "loss": 0.1555, "step": 6326 }, { "epoch": 0.26675379977654573, "grad_norm": 0.9218469858169556, "learning_rate": 4.444912872400225e-05, "loss": 1.058, "step": 6327 }, { "epoch": 0.2667959609587453, "grad_norm": 0.584628701210022, "learning_rate": 4.4456155143338954e-05, "loss": 0.8043, "step": 6328 }, { "epoch": 0.2668381221409448, "grad_norm": 0.38359731435775757, "learning_rate": 4.4463181562675665e-05, "loss": 0.1521, "step": 6329 }, { "epoch": 0.2668802833231444, "grad_norm": 0.5626555681228638, "learning_rate": 4.447020798201237e-05, "loss": 0.3597, "step": 6330 }, { "epoch": 0.2669224445053439, "grad_norm": 0.8624475002288818, "learning_rate": 4.4477234401349074e-05, "loss": 0.8996, "step": 6331 }, { "epoch": 0.2669646056875435, "grad_norm": 0.4259711503982544, "learning_rate": 4.448426082068578e-05, "loss": 0.2665, "step": 6332 }, { "epoch": 0.267006766869743, "grad_norm": 0.45695310831069946, "learning_rate": 4.449128724002249e-05, "loss": 0.1153, "step": 6333 }, { "epoch": 0.2670489280519426, "grad_norm": 1.953935980796814, "learning_rate": 4.4498313659359194e-05, "loss": 0.3524, "step": 6334 }, { "epoch": 0.2670910892341421, "grad_norm": 0.5570676326751709, "learning_rate": 4.45053400786959e-05, "loss": 0.4353, "step": 6335 }, { "epoch": 0.26713325041634167, "grad_norm": 0.3597100079059601, "learning_rate": 4.45123664980326e-05, "loss": 0.1433, "step": 6336 }, { "epoch": 0.26717541159854125, "grad_norm": 0.40474697947502136, "learning_rate": 4.4519392917369314e-05, "loss": 0.1425, "step": 6337 }, { "epoch": 0.26721757278074076, "grad_norm": 0.8432857990264893, "learning_rate": 4.452641933670602e-05, "loss": 0.3461, "step": 6338 }, { "epoch": 0.26725973396294034, "grad_norm": 0.29960763454437256, "learning_rate": 4.453344575604272e-05, "loss": 0.0731, "step": 6339 }, { "epoch": 0.26730189514513986, "grad_norm": 0.35805097222328186, "learning_rate": 4.454047217537943e-05, "loss": 0.1255, "step": 6340 }, { "epoch": 0.26734405632733943, "grad_norm": 0.7296116948127747, "learning_rate": 4.454749859471614e-05, "loss": 0.1506, "step": 6341 }, { "epoch": 0.26738621750953895, "grad_norm": 0.6413480043411255, "learning_rate": 4.455452501405284e-05, "loss": 0.1885, "step": 6342 }, { "epoch": 0.2674283786917385, "grad_norm": 0.4824412763118744, "learning_rate": 4.456155143338955e-05, "loss": 0.1514, "step": 6343 }, { "epoch": 0.26747053987393804, "grad_norm": 0.587591290473938, "learning_rate": 4.456857785272625e-05, "loss": 0.4028, "step": 6344 }, { "epoch": 0.2675127010561376, "grad_norm": 0.7083358764648438, "learning_rate": 4.457560427206296e-05, "loss": 0.205, "step": 6345 }, { "epoch": 0.2675548622383372, "grad_norm": 0.6346138715744019, "learning_rate": 4.458263069139966e-05, "loss": 0.6021, "step": 6346 }, { "epoch": 0.2675970234205367, "grad_norm": 1.0616531372070312, "learning_rate": 4.458965711073637e-05, "loss": 0.8861, "step": 6347 }, { "epoch": 0.2676391846027363, "grad_norm": 0.6760595440864563, "learning_rate": 4.4596683530073076e-05, "loss": 0.1824, "step": 6348 }, { "epoch": 0.2676813457849358, "grad_norm": 0.5241246223449707, "learning_rate": 4.460370994940978e-05, "loss": 0.3818, "step": 6349 }, { "epoch": 0.26772350696713537, "grad_norm": 0.7978575229644775, "learning_rate": 4.4610736368746484e-05, "loss": 0.2582, "step": 6350 }, { "epoch": 0.2677656681493349, "grad_norm": 0.38369742035865784, "learning_rate": 4.4617762788083196e-05, "loss": 0.1473, "step": 6351 }, { "epoch": 0.26780782933153446, "grad_norm": 0.48793262243270874, "learning_rate": 4.462478920741991e-05, "loss": 0.14, "step": 6352 }, { "epoch": 0.267849990513734, "grad_norm": 0.6356562972068787, "learning_rate": 4.4631815626756604e-05, "loss": 0.1617, "step": 6353 }, { "epoch": 0.26789215169593356, "grad_norm": 1.2712277173995972, "learning_rate": 4.4638842046093316e-05, "loss": 1.1804, "step": 6354 }, { "epoch": 0.26793431287813313, "grad_norm": 0.6525891423225403, "learning_rate": 4.464586846543002e-05, "loss": 0.2429, "step": 6355 }, { "epoch": 0.26797647406033265, "grad_norm": 0.9476771950721741, "learning_rate": 4.4652894884766724e-05, "loss": 0.2656, "step": 6356 }, { "epoch": 0.2680186352425322, "grad_norm": 0.662155032157898, "learning_rate": 4.465992130410343e-05, "loss": 0.2357, "step": 6357 }, { "epoch": 0.26806079642473174, "grad_norm": 1.5079437494277954, "learning_rate": 4.466694772344014e-05, "loss": 0.4478, "step": 6358 }, { "epoch": 0.2681029576069313, "grad_norm": 0.7648508548736572, "learning_rate": 4.4673974142776844e-05, "loss": 0.2327, "step": 6359 }, { "epoch": 0.26814511878913083, "grad_norm": 0.3425125777721405, "learning_rate": 4.468100056211355e-05, "loss": 0.1237, "step": 6360 }, { "epoch": 0.2681872799713304, "grad_norm": 0.7451330423355103, "learning_rate": 4.468802698145025e-05, "loss": 0.8273, "step": 6361 }, { "epoch": 0.2682294411535299, "grad_norm": 0.7177273035049438, "learning_rate": 4.4695053400786964e-05, "loss": 0.199, "step": 6362 }, { "epoch": 0.2682716023357295, "grad_norm": 0.47107064723968506, "learning_rate": 4.470207982012367e-05, "loss": 0.188, "step": 6363 }, { "epoch": 0.268313763517929, "grad_norm": 1.0219677686691284, "learning_rate": 4.470910623946037e-05, "loss": 0.1549, "step": 6364 }, { "epoch": 0.2683559247001286, "grad_norm": 0.8164552450180054, "learning_rate": 4.471613265879708e-05, "loss": 0.5757, "step": 6365 }, { "epoch": 0.26839808588232816, "grad_norm": 0.7482969760894775, "learning_rate": 4.472315907813379e-05, "loss": 0.2406, "step": 6366 }, { "epoch": 0.2684402470645277, "grad_norm": 0.6901758909225464, "learning_rate": 4.4730185497470486e-05, "loss": 0.21, "step": 6367 }, { "epoch": 0.26848240824672726, "grad_norm": 0.8635175824165344, "learning_rate": 4.47372119168072e-05, "loss": 0.7926, "step": 6368 }, { "epoch": 0.2685245694289268, "grad_norm": 0.5773311257362366, "learning_rate": 4.47442383361439e-05, "loss": 0.3119, "step": 6369 }, { "epoch": 0.26856673061112635, "grad_norm": 0.3623211681842804, "learning_rate": 4.475126475548061e-05, "loss": 0.1131, "step": 6370 }, { "epoch": 0.26860889179332587, "grad_norm": 0.672609269618988, "learning_rate": 4.475829117481731e-05, "loss": 0.6325, "step": 6371 }, { "epoch": 0.26865105297552544, "grad_norm": 0.9940078854560852, "learning_rate": 4.476531759415402e-05, "loss": 0.7472, "step": 6372 }, { "epoch": 0.26869321415772496, "grad_norm": 2.493025779724121, "learning_rate": 4.4772344013490726e-05, "loss": 0.6899, "step": 6373 }, { "epoch": 0.26873537533992453, "grad_norm": 0.38895031809806824, "learning_rate": 4.477937043282743e-05, "loss": 0.1329, "step": 6374 }, { "epoch": 0.2687775365221241, "grad_norm": 0.669430136680603, "learning_rate": 4.478639685216414e-05, "loss": 0.3937, "step": 6375 }, { "epoch": 0.2688196977043236, "grad_norm": 0.42814555764198303, "learning_rate": 4.4793423271500846e-05, "loss": 0.0867, "step": 6376 }, { "epoch": 0.2688618588865232, "grad_norm": 1.2572553157806396, "learning_rate": 4.480044969083755e-05, "loss": 1.2274, "step": 6377 }, { "epoch": 0.2689040200687227, "grad_norm": 0.6885879039764404, "learning_rate": 4.4807476110174255e-05, "loss": 0.3763, "step": 6378 }, { "epoch": 0.2689461812509223, "grad_norm": 0.9245510697364807, "learning_rate": 4.4814502529510966e-05, "loss": 0.2419, "step": 6379 }, { "epoch": 0.2689883424331218, "grad_norm": 0.7999501824378967, "learning_rate": 4.482152894884767e-05, "loss": 0.5215, "step": 6380 }, { "epoch": 0.2690305036153214, "grad_norm": 0.6052092909812927, "learning_rate": 4.4828555368184375e-05, "loss": 0.239, "step": 6381 }, { "epoch": 0.2690726647975209, "grad_norm": 0.503582239151001, "learning_rate": 4.483558178752108e-05, "loss": 0.2101, "step": 6382 }, { "epoch": 0.2691148259797205, "grad_norm": 0.4289797842502594, "learning_rate": 4.484260820685779e-05, "loss": 0.1357, "step": 6383 }, { "epoch": 0.26915698716192005, "grad_norm": 0.520807147026062, "learning_rate": 4.4849634626194495e-05, "loss": 0.204, "step": 6384 }, { "epoch": 0.26919914834411957, "grad_norm": 0.5318443179130554, "learning_rate": 4.48566610455312e-05, "loss": 0.2017, "step": 6385 }, { "epoch": 0.26924130952631914, "grad_norm": 0.6471904516220093, "learning_rate": 4.4863687464867904e-05, "loss": 0.4297, "step": 6386 }, { "epoch": 0.26928347070851866, "grad_norm": 0.9878613352775574, "learning_rate": 4.4870713884204615e-05, "loss": 0.2035, "step": 6387 }, { "epoch": 0.26932563189071823, "grad_norm": 0.561484694480896, "learning_rate": 4.487774030354131e-05, "loss": 0.4028, "step": 6388 }, { "epoch": 0.26936779307291775, "grad_norm": 1.7669163942337036, "learning_rate": 4.4884766722878024e-05, "loss": 0.981, "step": 6389 }, { "epoch": 0.2694099542551173, "grad_norm": 0.6278819441795349, "learning_rate": 4.489179314221473e-05, "loss": 0.5605, "step": 6390 }, { "epoch": 0.26945211543731684, "grad_norm": 0.7105696201324463, "learning_rate": 4.489881956155144e-05, "loss": 0.2352, "step": 6391 }, { "epoch": 0.2694942766195164, "grad_norm": 0.7186740040779114, "learning_rate": 4.490584598088814e-05, "loss": 0.5021, "step": 6392 }, { "epoch": 0.26953643780171593, "grad_norm": 0.7928827404975891, "learning_rate": 4.491287240022485e-05, "loss": 0.533, "step": 6393 }, { "epoch": 0.2695785989839155, "grad_norm": 0.7693896889686584, "learning_rate": 4.491989881956155e-05, "loss": 0.1271, "step": 6394 }, { "epoch": 0.2696207601661151, "grad_norm": 0.7979087829589844, "learning_rate": 4.492692523889826e-05, "loss": 0.3728, "step": 6395 }, { "epoch": 0.2696629213483146, "grad_norm": 0.3626810908317566, "learning_rate": 4.493395165823496e-05, "loss": 0.0966, "step": 6396 }, { "epoch": 0.2697050825305142, "grad_norm": 0.658073902130127, "learning_rate": 4.494097807757167e-05, "loss": 0.2338, "step": 6397 }, { "epoch": 0.2697472437127137, "grad_norm": 0.5075638890266418, "learning_rate": 4.4948004496908384e-05, "loss": 0.1608, "step": 6398 }, { "epoch": 0.26978940489491327, "grad_norm": 0.3910248279571533, "learning_rate": 4.495503091624508e-05, "loss": 0.1, "step": 6399 }, { "epoch": 0.2698315660771128, "grad_norm": 0.7513812184333801, "learning_rate": 4.496205733558179e-05, "loss": 0.37, "step": 6400 }, { "epoch": 0.26987372725931236, "grad_norm": 0.658964216709137, "learning_rate": 4.49690837549185e-05, "loss": 0.1782, "step": 6401 }, { "epoch": 0.2699158884415119, "grad_norm": 0.6623678207397461, "learning_rate": 4.49761101742552e-05, "loss": 0.5695, "step": 6402 }, { "epoch": 0.26995804962371145, "grad_norm": 0.39869627356529236, "learning_rate": 4.4983136593591906e-05, "loss": 0.0969, "step": 6403 }, { "epoch": 0.270000210805911, "grad_norm": 0.567156195640564, "learning_rate": 4.499016301292862e-05, "loss": 0.1421, "step": 6404 }, { "epoch": 0.27004237198811054, "grad_norm": 0.5963698625564575, "learning_rate": 4.499718943226532e-05, "loss": 0.1106, "step": 6405 }, { "epoch": 0.2700845331703101, "grad_norm": 0.6565508842468262, "learning_rate": 4.5004215851602025e-05, "loss": 0.1213, "step": 6406 }, { "epoch": 0.27012669435250963, "grad_norm": 1.3139559030532837, "learning_rate": 4.501124227093873e-05, "loss": 1.1004, "step": 6407 }, { "epoch": 0.2701688555347092, "grad_norm": 0.4083258807659149, "learning_rate": 4.501826869027544e-05, "loss": 0.175, "step": 6408 }, { "epoch": 0.2702110167169087, "grad_norm": 0.31360745429992676, "learning_rate": 4.5025295109612145e-05, "loss": 0.096, "step": 6409 }, { "epoch": 0.2702531778991083, "grad_norm": 1.0267598628997803, "learning_rate": 4.503232152894885e-05, "loss": 0.1616, "step": 6410 }, { "epoch": 0.2702953390813078, "grad_norm": 0.523531973361969, "learning_rate": 4.5039347948285554e-05, "loss": 0.1137, "step": 6411 }, { "epoch": 0.2703375002635074, "grad_norm": 1.0005356073379517, "learning_rate": 4.5046374367622265e-05, "loss": 0.5669, "step": 6412 }, { "epoch": 0.27037966144570696, "grad_norm": 0.7194774150848389, "learning_rate": 4.505340078695896e-05, "loss": 0.3012, "step": 6413 }, { "epoch": 0.2704218226279065, "grad_norm": 0.6870586276054382, "learning_rate": 4.5060427206295674e-05, "loss": 0.7498, "step": 6414 }, { "epoch": 0.27046398381010606, "grad_norm": 0.5722192525863647, "learning_rate": 4.506745362563238e-05, "loss": 0.42, "step": 6415 }, { "epoch": 0.2705061449923056, "grad_norm": 1.2299331426620483, "learning_rate": 4.507448004496908e-05, "loss": 0.9844, "step": 6416 }, { "epoch": 0.27054830617450515, "grad_norm": 0.48941904306411743, "learning_rate": 4.508150646430579e-05, "loss": 0.1305, "step": 6417 }, { "epoch": 0.27059046735670467, "grad_norm": 0.6179946064949036, "learning_rate": 4.50885328836425e-05, "loss": 0.4901, "step": 6418 }, { "epoch": 0.27063262853890424, "grad_norm": 0.5506634712219238, "learning_rate": 4.509555930297921e-05, "loss": 0.583, "step": 6419 }, { "epoch": 0.27067478972110376, "grad_norm": 0.5133177042007446, "learning_rate": 4.510258572231591e-05, "loss": 0.1387, "step": 6420 }, { "epoch": 0.27071695090330333, "grad_norm": 0.8063769340515137, "learning_rate": 4.510961214165262e-05, "loss": 0.3025, "step": 6421 }, { "epoch": 0.27075911208550285, "grad_norm": 0.7612944841384888, "learning_rate": 4.511663856098932e-05, "loss": 0.3415, "step": 6422 }, { "epoch": 0.2708012732677024, "grad_norm": 0.5497949719429016, "learning_rate": 4.512366498032603e-05, "loss": 0.2601, "step": 6423 }, { "epoch": 0.270843434449902, "grad_norm": 0.5294073820114136, "learning_rate": 4.513069139966273e-05, "loss": 0.1255, "step": 6424 }, { "epoch": 0.2708855956321015, "grad_norm": 2.622535467147827, "learning_rate": 4.513771781899944e-05, "loss": 0.6232, "step": 6425 }, { "epoch": 0.2709277568143011, "grad_norm": 0.9052828550338745, "learning_rate": 4.514474423833615e-05, "loss": 0.3793, "step": 6426 }, { "epoch": 0.2709699179965006, "grad_norm": 0.41948986053466797, "learning_rate": 4.515177065767285e-05, "loss": 0.1084, "step": 6427 }, { "epoch": 0.2710120791787002, "grad_norm": 0.37749552726745605, "learning_rate": 4.5158797077009556e-05, "loss": 0.1097, "step": 6428 }, { "epoch": 0.2710542403608997, "grad_norm": 0.4684703052043915, "learning_rate": 4.516582349634627e-05, "loss": 0.1065, "step": 6429 }, { "epoch": 0.2710964015430993, "grad_norm": 2.0402028560638428, "learning_rate": 4.517284991568297e-05, "loss": 0.7179, "step": 6430 }, { "epoch": 0.2711385627252988, "grad_norm": 0.5138421058654785, "learning_rate": 4.5179876335019676e-05, "loss": 0.2373, "step": 6431 }, { "epoch": 0.27118072390749837, "grad_norm": 0.6453336477279663, "learning_rate": 4.518690275435638e-05, "loss": 0.2537, "step": 6432 }, { "epoch": 0.27122288508969794, "grad_norm": 0.7967948317527771, "learning_rate": 4.519392917369309e-05, "loss": 0.215, "step": 6433 }, { "epoch": 0.27126504627189746, "grad_norm": 0.844662606716156, "learning_rate": 4.520095559302979e-05, "loss": 0.1594, "step": 6434 }, { "epoch": 0.27130720745409703, "grad_norm": 1.8966525793075562, "learning_rate": 4.52079820123665e-05, "loss": 0.5798, "step": 6435 }, { "epoch": 0.27134936863629655, "grad_norm": 0.7581397294998169, "learning_rate": 4.5215008431703205e-05, "loss": 0.3648, "step": 6436 }, { "epoch": 0.2713915298184961, "grad_norm": 0.8732220530509949, "learning_rate": 4.5222034851039916e-05, "loss": 0.2452, "step": 6437 }, { "epoch": 0.27143369100069564, "grad_norm": 0.7928649187088013, "learning_rate": 4.5229061270376614e-05, "loss": 0.2372, "step": 6438 }, { "epoch": 0.2714758521828952, "grad_norm": 1.1132643222808838, "learning_rate": 4.5236087689713325e-05, "loss": 0.167, "step": 6439 }, { "epoch": 0.27151801336509473, "grad_norm": 0.5208905935287476, "learning_rate": 4.524311410905003e-05, "loss": 0.0937, "step": 6440 }, { "epoch": 0.2715601745472943, "grad_norm": 0.6465511918067932, "learning_rate": 4.5250140528386734e-05, "loss": 0.6088, "step": 6441 }, { "epoch": 0.2716023357294939, "grad_norm": 1.2100989818572998, "learning_rate": 4.5257166947723445e-05, "loss": 0.2814, "step": 6442 }, { "epoch": 0.2716444969116934, "grad_norm": 0.42971351742744446, "learning_rate": 4.526419336706015e-05, "loss": 0.1986, "step": 6443 }, { "epoch": 0.271686658093893, "grad_norm": 0.9888355135917664, "learning_rate": 4.5271219786396853e-05, "loss": 0.8358, "step": 6444 }, { "epoch": 0.2717288192760925, "grad_norm": 0.7724478244781494, "learning_rate": 4.527824620573356e-05, "loss": 0.3091, "step": 6445 }, { "epoch": 0.27177098045829207, "grad_norm": 0.6546334624290466, "learning_rate": 4.528527262507027e-05, "loss": 0.1016, "step": 6446 }, { "epoch": 0.2718131416404916, "grad_norm": 0.7639857530593872, "learning_rate": 4.5292299044406973e-05, "loss": 0.4168, "step": 6447 }, { "epoch": 0.27185530282269116, "grad_norm": 0.5288860201835632, "learning_rate": 4.529932546374368e-05, "loss": 0.1024, "step": 6448 }, { "epoch": 0.2718974640048907, "grad_norm": 0.7203040719032288, "learning_rate": 4.530635188308038e-05, "loss": 0.5852, "step": 6449 }, { "epoch": 0.27193962518709025, "grad_norm": 0.4330577850341797, "learning_rate": 4.5313378302417093e-05, "loss": 0.0823, "step": 6450 }, { "epoch": 0.27198178636928977, "grad_norm": 0.7217789888381958, "learning_rate": 4.53204047217538e-05, "loss": 0.2262, "step": 6451 }, { "epoch": 0.27202394755148934, "grad_norm": 1.5188333988189697, "learning_rate": 4.53274311410905e-05, "loss": 0.7022, "step": 6452 }, { "epoch": 0.2720661087336889, "grad_norm": 0.6533438563346863, "learning_rate": 4.5334457560427207e-05, "loss": 0.6261, "step": 6453 }, { "epoch": 0.27210826991588843, "grad_norm": 1.2025048732757568, "learning_rate": 4.534148397976392e-05, "loss": 0.6535, "step": 6454 }, { "epoch": 0.272150431098088, "grad_norm": 0.6002328395843506, "learning_rate": 4.534851039910062e-05, "loss": 0.2273, "step": 6455 }, { "epoch": 0.2721925922802875, "grad_norm": 1.0835001468658447, "learning_rate": 4.5355536818437327e-05, "loss": 0.1928, "step": 6456 }, { "epoch": 0.2722347534624871, "grad_norm": 0.4240284562110901, "learning_rate": 4.536256323777403e-05, "loss": 0.1468, "step": 6457 }, { "epoch": 0.2722769146446866, "grad_norm": 1.0058214664459229, "learning_rate": 4.536958965711074e-05, "loss": 0.5868, "step": 6458 }, { "epoch": 0.2723190758268862, "grad_norm": 0.4804863929748535, "learning_rate": 4.537661607644744e-05, "loss": 0.2279, "step": 6459 }, { "epoch": 0.2723612370090857, "grad_norm": 0.7216521501541138, "learning_rate": 4.538364249578415e-05, "loss": 0.5731, "step": 6460 }, { "epoch": 0.2724033981912853, "grad_norm": 0.9717541933059692, "learning_rate": 4.5390668915120855e-05, "loss": 0.2102, "step": 6461 }, { "epoch": 0.27244555937348486, "grad_norm": 0.808340311050415, "learning_rate": 4.539769533445756e-05, "loss": 0.4016, "step": 6462 }, { "epoch": 0.2724877205556844, "grad_norm": 0.9821809530258179, "learning_rate": 4.5404721753794264e-05, "loss": 0.1846, "step": 6463 }, { "epoch": 0.27252988173788395, "grad_norm": 0.45851626992225647, "learning_rate": 4.5411748173130975e-05, "loss": 0.0996, "step": 6464 }, { "epoch": 0.27257204292008347, "grad_norm": 0.4292161464691162, "learning_rate": 4.5418774592467686e-05, "loss": 0.1722, "step": 6465 }, { "epoch": 0.27261420410228304, "grad_norm": 0.6823869347572327, "learning_rate": 4.5425801011804384e-05, "loss": 0.2625, "step": 6466 }, { "epoch": 0.27265636528448256, "grad_norm": 0.5615660548210144, "learning_rate": 4.5432827431141095e-05, "loss": 0.2793, "step": 6467 }, { "epoch": 0.27269852646668213, "grad_norm": 0.7509360313415527, "learning_rate": 4.54398538504778e-05, "loss": 0.2057, "step": 6468 }, { "epoch": 0.27274068764888165, "grad_norm": 0.865070641040802, "learning_rate": 4.5446880269814504e-05, "loss": 0.4878, "step": 6469 }, { "epoch": 0.2727828488310812, "grad_norm": 0.5214992761611938, "learning_rate": 4.545390668915121e-05, "loss": 0.1557, "step": 6470 }, { "epoch": 0.2728250100132808, "grad_norm": 0.9904679656028748, "learning_rate": 4.546093310848792e-05, "loss": 1.0502, "step": 6471 }, { "epoch": 0.2728671711954803, "grad_norm": 1.538470983505249, "learning_rate": 4.5467959527824624e-05, "loss": 0.9459, "step": 6472 }, { "epoch": 0.2729093323776799, "grad_norm": 0.6044161319732666, "learning_rate": 4.547498594716133e-05, "loss": 0.3657, "step": 6473 }, { "epoch": 0.2729514935598794, "grad_norm": 1.2299563884735107, "learning_rate": 4.548201236649803e-05, "loss": 1.1198, "step": 6474 }, { "epoch": 0.272993654742079, "grad_norm": 0.8156614303588867, "learning_rate": 4.5489038785834744e-05, "loss": 0.3264, "step": 6475 }, { "epoch": 0.2730358159242785, "grad_norm": 0.6238086819648743, "learning_rate": 4.549606520517145e-05, "loss": 0.1855, "step": 6476 }, { "epoch": 0.2730779771064781, "grad_norm": 1.1900862455368042, "learning_rate": 4.550309162450815e-05, "loss": 0.7617, "step": 6477 }, { "epoch": 0.2731201382886776, "grad_norm": 0.6598861813545227, "learning_rate": 4.551011804384486e-05, "loss": 0.2101, "step": 6478 }, { "epoch": 0.27316229947087717, "grad_norm": 1.0201274156570435, "learning_rate": 4.551714446318157e-05, "loss": 0.9151, "step": 6479 }, { "epoch": 0.2732044606530767, "grad_norm": 0.9843476414680481, "learning_rate": 4.5524170882518266e-05, "loss": 0.3008, "step": 6480 }, { "epoch": 0.27324662183527626, "grad_norm": 1.0981042385101318, "learning_rate": 4.553119730185498e-05, "loss": 0.3794, "step": 6481 }, { "epoch": 0.27328878301747583, "grad_norm": 0.5094210505485535, "learning_rate": 4.553822372119168e-05, "loss": 0.3043, "step": 6482 }, { "epoch": 0.27333094419967535, "grad_norm": 0.9654976725578308, "learning_rate": 4.554525014052839e-05, "loss": 0.6758, "step": 6483 }, { "epoch": 0.2733731053818749, "grad_norm": 0.44405895471572876, "learning_rate": 4.555227655986509e-05, "loss": 0.1095, "step": 6484 }, { "epoch": 0.27341526656407444, "grad_norm": 0.36650344729423523, "learning_rate": 4.55593029792018e-05, "loss": 0.0878, "step": 6485 }, { "epoch": 0.273457427746274, "grad_norm": 1.2087113857269287, "learning_rate": 4.5566329398538506e-05, "loss": 0.9677, "step": 6486 }, { "epoch": 0.27349958892847354, "grad_norm": 2.8171377182006836, "learning_rate": 4.557335581787521e-05, "loss": 0.3227, "step": 6487 }, { "epoch": 0.2735417501106731, "grad_norm": 0.8158521056175232, "learning_rate": 4.558038223721192e-05, "loss": 0.2521, "step": 6488 }, { "epoch": 0.27358391129287263, "grad_norm": 1.096139907836914, "learning_rate": 4.5587408656548626e-05, "loss": 0.973, "step": 6489 }, { "epoch": 0.2736260724750722, "grad_norm": 0.42236730456352234, "learning_rate": 4.559443507588533e-05, "loss": 0.178, "step": 6490 }, { "epoch": 0.2736682336572718, "grad_norm": 0.6863534450531006, "learning_rate": 4.5601461495222035e-05, "loss": 0.6197, "step": 6491 }, { "epoch": 0.2737103948394713, "grad_norm": 2.402893304824829, "learning_rate": 4.5608487914558746e-05, "loss": 0.3129, "step": 6492 }, { "epoch": 0.27375255602167087, "grad_norm": 0.8729544878005981, "learning_rate": 4.561551433389545e-05, "loss": 0.3181, "step": 6493 }, { "epoch": 0.2737947172038704, "grad_norm": 3.6315066814422607, "learning_rate": 4.5622540753232155e-05, "loss": 0.3482, "step": 6494 }, { "epoch": 0.27383687838606996, "grad_norm": 0.6527910232543945, "learning_rate": 4.562956717256886e-05, "loss": 0.4731, "step": 6495 }, { "epoch": 0.2738790395682695, "grad_norm": 0.4792468249797821, "learning_rate": 4.563659359190557e-05, "loss": 0.4732, "step": 6496 }, { "epoch": 0.27392120075046905, "grad_norm": 0.6757930517196655, "learning_rate": 4.5643620011242275e-05, "loss": 0.5645, "step": 6497 }, { "epoch": 0.27396336193266857, "grad_norm": 0.803403913974762, "learning_rate": 4.565064643057898e-05, "loss": 0.2699, "step": 6498 }, { "epoch": 0.27400552311486814, "grad_norm": 0.6849805116653442, "learning_rate": 4.565767284991568e-05, "loss": 0.1903, "step": 6499 }, { "epoch": 0.2740476842970677, "grad_norm": 1.260204553604126, "learning_rate": 4.5664699269252394e-05, "loss": 0.413, "step": 6500 }, { "epoch": 0.27408984547926724, "grad_norm": 0.7487557530403137, "learning_rate": 4.567172568858909e-05, "loss": 0.1758, "step": 6501 }, { "epoch": 0.2741320066614668, "grad_norm": 0.685341477394104, "learning_rate": 4.56787521079258e-05, "loss": 0.7476, "step": 6502 }, { "epoch": 0.2741741678436663, "grad_norm": 0.5056694746017456, "learning_rate": 4.568577852726251e-05, "loss": 0.1077, "step": 6503 }, { "epoch": 0.2742163290258659, "grad_norm": 0.7067655920982361, "learning_rate": 4.569280494659922e-05, "loss": 0.3392, "step": 6504 }, { "epoch": 0.2742584902080654, "grad_norm": 0.4310816526412964, "learning_rate": 4.5699831365935916e-05, "loss": 0.1017, "step": 6505 }, { "epoch": 0.274300651390265, "grad_norm": 1.112338900566101, "learning_rate": 4.570685778527263e-05, "loss": 0.6568, "step": 6506 }, { "epoch": 0.2743428125724645, "grad_norm": 1.0459717512130737, "learning_rate": 4.571388420460933e-05, "loss": 0.5297, "step": 6507 }, { "epoch": 0.2743849737546641, "grad_norm": 0.6146926879882812, "learning_rate": 4.5720910623946036e-05, "loss": 0.4479, "step": 6508 }, { "epoch": 0.2744271349368636, "grad_norm": 0.6625178456306458, "learning_rate": 4.572793704328275e-05, "loss": 0.1251, "step": 6509 }, { "epoch": 0.2744692961190632, "grad_norm": 0.44002780318260193, "learning_rate": 4.573496346261945e-05, "loss": 0.1898, "step": 6510 }, { "epoch": 0.27451145730126275, "grad_norm": 0.7120928168296814, "learning_rate": 4.574198988195616e-05, "loss": 0.5248, "step": 6511 }, { "epoch": 0.27455361848346227, "grad_norm": 0.6662625074386597, "learning_rate": 4.574901630129286e-05, "loss": 0.4349, "step": 6512 }, { "epoch": 0.27459577966566184, "grad_norm": 1.0698139667510986, "learning_rate": 4.575604272062957e-05, "loss": 0.2101, "step": 6513 }, { "epoch": 0.27463794084786136, "grad_norm": 0.6301933526992798, "learning_rate": 4.5763069139966276e-05, "loss": 0.3097, "step": 6514 }, { "epoch": 0.27468010203006094, "grad_norm": 0.4213171601295471, "learning_rate": 4.577009555930298e-05, "loss": 0.0935, "step": 6515 }, { "epoch": 0.27472226321226045, "grad_norm": 0.5009748935699463, "learning_rate": 4.5777121978639685e-05, "loss": 0.3193, "step": 6516 }, { "epoch": 0.27476442439446, "grad_norm": 0.6277568340301514, "learning_rate": 4.5784148397976396e-05, "loss": 0.1737, "step": 6517 }, { "epoch": 0.27480658557665955, "grad_norm": 0.7603915929794312, "learning_rate": 4.57911748173131e-05, "loss": 0.8319, "step": 6518 }, { "epoch": 0.2748487467588591, "grad_norm": 0.9397076368331909, "learning_rate": 4.5798201236649805e-05, "loss": 1.0446, "step": 6519 }, { "epoch": 0.2748909079410587, "grad_norm": 0.7787938117980957, "learning_rate": 4.580522765598651e-05, "loss": 0.6301, "step": 6520 }, { "epoch": 0.2749330691232582, "grad_norm": 0.7628875970840454, "learning_rate": 4.581225407532322e-05, "loss": 0.2329, "step": 6521 }, { "epoch": 0.2749752303054578, "grad_norm": 0.4134522080421448, "learning_rate": 4.5819280494659925e-05, "loss": 0.1183, "step": 6522 }, { "epoch": 0.2750173914876573, "grad_norm": 1.8497941493988037, "learning_rate": 4.582630691399663e-05, "loss": 0.4462, "step": 6523 }, { "epoch": 0.2750595526698569, "grad_norm": 0.4605115056037903, "learning_rate": 4.5833333333333334e-05, "loss": 0.1402, "step": 6524 }, { "epoch": 0.2751017138520564, "grad_norm": 0.497455358505249, "learning_rate": 4.5840359752670045e-05, "loss": 0.346, "step": 6525 }, { "epoch": 0.27514387503425597, "grad_norm": 1.9977328777313232, "learning_rate": 4.584738617200674e-05, "loss": 0.6127, "step": 6526 }, { "epoch": 0.2751860362164555, "grad_norm": 0.4578629434108734, "learning_rate": 4.5854412591343454e-05, "loss": 0.1795, "step": 6527 }, { "epoch": 0.27522819739865506, "grad_norm": 0.6945170164108276, "learning_rate": 4.586143901068016e-05, "loss": 0.1458, "step": 6528 }, { "epoch": 0.2752703585808546, "grad_norm": 0.6393390893936157, "learning_rate": 4.586846543001686e-05, "loss": 0.4085, "step": 6529 }, { "epoch": 0.27531251976305415, "grad_norm": 1.0475155115127563, "learning_rate": 4.587549184935357e-05, "loss": 0.1851, "step": 6530 }, { "epoch": 0.2753546809452537, "grad_norm": 3.496074914932251, "learning_rate": 4.588251826869028e-05, "loss": 1.0379, "step": 6531 }, { "epoch": 0.27539684212745325, "grad_norm": 0.6843187212944031, "learning_rate": 4.588954468802699e-05, "loss": 0.4767, "step": 6532 }, { "epoch": 0.2754390033096528, "grad_norm": 0.38024064898490906, "learning_rate": 4.589657110736369e-05, "loss": 0.1477, "step": 6533 }, { "epoch": 0.27548116449185234, "grad_norm": 0.9695206880569458, "learning_rate": 4.59035975267004e-05, "loss": 0.3328, "step": 6534 }, { "epoch": 0.2755233256740519, "grad_norm": 0.8191991448402405, "learning_rate": 4.59106239460371e-05, "loss": 0.2446, "step": 6535 }, { "epoch": 0.27556548685625143, "grad_norm": 1.346442461013794, "learning_rate": 4.591765036537381e-05, "loss": 0.8188, "step": 6536 }, { "epoch": 0.275607648038451, "grad_norm": 0.8874176740646362, "learning_rate": 4.592467678471051e-05, "loss": 0.3881, "step": 6537 }, { "epoch": 0.2756498092206505, "grad_norm": 0.4105652868747711, "learning_rate": 4.593170320404722e-05, "loss": 0.11, "step": 6538 }, { "epoch": 0.2756919704028501, "grad_norm": 1.107035756111145, "learning_rate": 4.593872962338393e-05, "loss": 0.2457, "step": 6539 }, { "epoch": 0.27573413158504967, "grad_norm": 1.4200369119644165, "learning_rate": 4.594575604272063e-05, "loss": 0.8546, "step": 6540 }, { "epoch": 0.2757762927672492, "grad_norm": 0.37804996967315674, "learning_rate": 4.5952782462057336e-05, "loss": 0.1762, "step": 6541 }, { "epoch": 0.27581845394944876, "grad_norm": 0.6821087002754211, "learning_rate": 4.595980888139405e-05, "loss": 0.6754, "step": 6542 }, { "epoch": 0.2758606151316483, "grad_norm": 0.4162692129611969, "learning_rate": 4.596683530073075e-05, "loss": 0.1702, "step": 6543 }, { "epoch": 0.27590277631384785, "grad_norm": 0.7210974097251892, "learning_rate": 4.5973861720067456e-05, "loss": 0.5846, "step": 6544 }, { "epoch": 0.27594493749604737, "grad_norm": 0.6484482288360596, "learning_rate": 4.598088813940416e-05, "loss": 0.1968, "step": 6545 }, { "epoch": 0.27598709867824694, "grad_norm": 0.5913231372833252, "learning_rate": 4.598791455874087e-05, "loss": 0.2978, "step": 6546 }, { "epoch": 0.27602925986044646, "grad_norm": 0.40008023381233215, "learning_rate": 4.599494097807757e-05, "loss": 0.1793, "step": 6547 }, { "epoch": 0.27607142104264604, "grad_norm": 0.5120938420295715, "learning_rate": 4.600196739741428e-05, "loss": 0.1692, "step": 6548 }, { "epoch": 0.2761135822248456, "grad_norm": 1.2327402830123901, "learning_rate": 4.6008993816750984e-05, "loss": 0.5677, "step": 6549 }, { "epoch": 0.27615574340704513, "grad_norm": 0.5883049368858337, "learning_rate": 4.6016020236087696e-05, "loss": 0.162, "step": 6550 }, { "epoch": 0.2761979045892447, "grad_norm": 0.3800859749317169, "learning_rate": 4.602304665542439e-05, "loss": 0.1027, "step": 6551 }, { "epoch": 0.2762400657714442, "grad_norm": 0.6916140913963318, "learning_rate": 4.6030073074761104e-05, "loss": 0.2301, "step": 6552 }, { "epoch": 0.2762822269536438, "grad_norm": 0.916190505027771, "learning_rate": 4.603709949409781e-05, "loss": 0.0877, "step": 6553 }, { "epoch": 0.2763243881358433, "grad_norm": 0.7682929635047913, "learning_rate": 4.604412591343451e-05, "loss": 0.2332, "step": 6554 }, { "epoch": 0.2763665493180429, "grad_norm": 0.47596803307533264, "learning_rate": 4.6051152332771224e-05, "loss": 0.2816, "step": 6555 }, { "epoch": 0.2764087105002424, "grad_norm": 0.46182501316070557, "learning_rate": 4.605817875210793e-05, "loss": 0.1281, "step": 6556 }, { "epoch": 0.276450871682442, "grad_norm": 0.6358615756034851, "learning_rate": 4.606520517144463e-05, "loss": 0.5686, "step": 6557 }, { "epoch": 0.2764930328646415, "grad_norm": 1.1744142770767212, "learning_rate": 4.607223159078134e-05, "loss": 1.0324, "step": 6558 }, { "epoch": 0.27653519404684107, "grad_norm": 0.4534493386745453, "learning_rate": 4.607925801011805e-05, "loss": 0.1853, "step": 6559 }, { "epoch": 0.27657735522904064, "grad_norm": 2.5589518547058105, "learning_rate": 4.608628442945475e-05, "loss": 0.7597, "step": 6560 }, { "epoch": 0.27661951641124016, "grad_norm": 0.6754030585289001, "learning_rate": 4.609331084879146e-05, "loss": 0.2665, "step": 6561 }, { "epoch": 0.27666167759343974, "grad_norm": 0.3276345729827881, "learning_rate": 4.610033726812816e-05, "loss": 0.0918, "step": 6562 }, { "epoch": 0.27670383877563925, "grad_norm": 0.6219237446784973, "learning_rate": 4.610736368746487e-05, "loss": 0.1976, "step": 6563 }, { "epoch": 0.27674599995783883, "grad_norm": 1.527159333229065, "learning_rate": 4.611439010680158e-05, "loss": 1.197, "step": 6564 }, { "epoch": 0.27678816114003835, "grad_norm": 0.3767247796058655, "learning_rate": 4.612141652613828e-05, "loss": 0.1293, "step": 6565 }, { "epoch": 0.2768303223222379, "grad_norm": 0.31641459465026855, "learning_rate": 4.6128442945474986e-05, "loss": 0.0889, "step": 6566 }, { "epoch": 0.27687248350443744, "grad_norm": 0.9414916634559631, "learning_rate": 4.61354693648117e-05, "loss": 0.9586, "step": 6567 }, { "epoch": 0.276914644686637, "grad_norm": 0.8598706126213074, "learning_rate": 4.61424957841484e-05, "loss": 0.5474, "step": 6568 }, { "epoch": 0.2769568058688366, "grad_norm": 0.7854752540588379, "learning_rate": 4.6149522203485106e-05, "loss": 1.0592, "step": 6569 }, { "epoch": 0.2769989670510361, "grad_norm": 0.34228524565696716, "learning_rate": 4.615654862282181e-05, "loss": 0.1283, "step": 6570 }, { "epoch": 0.2770411282332357, "grad_norm": 0.72988361120224, "learning_rate": 4.616357504215852e-05, "loss": 0.1986, "step": 6571 }, { "epoch": 0.2770832894154352, "grad_norm": 0.30073657631874084, "learning_rate": 4.617060146149522e-05, "loss": 0.1003, "step": 6572 }, { "epoch": 0.27712545059763477, "grad_norm": 1.088366150856018, "learning_rate": 4.617762788083193e-05, "loss": 0.375, "step": 6573 }, { "epoch": 0.2771676117798343, "grad_norm": 0.4155430793762207, "learning_rate": 4.6184654300168635e-05, "loss": 0.0993, "step": 6574 }, { "epoch": 0.27720977296203386, "grad_norm": 1.2078384160995483, "learning_rate": 4.619168071950534e-05, "loss": 1.1085, "step": 6575 }, { "epoch": 0.2772519341442334, "grad_norm": 0.6435655355453491, "learning_rate": 4.6198707138842044e-05, "loss": 0.5648, "step": 6576 }, { "epoch": 0.27729409532643295, "grad_norm": 0.6413850784301758, "learning_rate": 4.6205733558178755e-05, "loss": 0.3612, "step": 6577 }, { "epoch": 0.27733625650863253, "grad_norm": 0.5095108151435852, "learning_rate": 4.6212759977515466e-05, "loss": 0.3008, "step": 6578 }, { "epoch": 0.27737841769083205, "grad_norm": 0.7787865996360779, "learning_rate": 4.6219786396852164e-05, "loss": 0.4996, "step": 6579 }, { "epoch": 0.2774205788730316, "grad_norm": 0.8517545461654663, "learning_rate": 4.6226812816188875e-05, "loss": 0.16, "step": 6580 }, { "epoch": 0.27746274005523114, "grad_norm": 0.5615442395210266, "learning_rate": 4.623383923552558e-05, "loss": 0.502, "step": 6581 }, { "epoch": 0.2775049012374307, "grad_norm": 0.9189218282699585, "learning_rate": 4.6240865654862284e-05, "loss": 0.2649, "step": 6582 }, { "epoch": 0.27754706241963023, "grad_norm": 0.7035237550735474, "learning_rate": 4.624789207419899e-05, "loss": 0.5956, "step": 6583 }, { "epoch": 0.2775892236018298, "grad_norm": 0.37918415665626526, "learning_rate": 4.62549184935357e-05, "loss": 0.1404, "step": 6584 }, { "epoch": 0.2776313847840293, "grad_norm": 0.4876538813114166, "learning_rate": 4.6261944912872404e-05, "loss": 0.2849, "step": 6585 }, { "epoch": 0.2776735459662289, "grad_norm": 0.3842571973800659, "learning_rate": 4.626897133220911e-05, "loss": 0.0697, "step": 6586 }, { "epoch": 0.2777157071484284, "grad_norm": 0.32572031021118164, "learning_rate": 4.627599775154581e-05, "loss": 0.1164, "step": 6587 }, { "epoch": 0.277757868330628, "grad_norm": 0.6071306467056274, "learning_rate": 4.6283024170882524e-05, "loss": 0.1587, "step": 6588 }, { "epoch": 0.27780002951282756, "grad_norm": 0.6739591360092163, "learning_rate": 4.629005059021923e-05, "loss": 0.2529, "step": 6589 }, { "epoch": 0.2778421906950271, "grad_norm": 1.3043220043182373, "learning_rate": 4.629707700955593e-05, "loss": 0.3312, "step": 6590 }, { "epoch": 0.27788435187722665, "grad_norm": 1.2770984172821045, "learning_rate": 4.630410342889264e-05, "loss": 1.0004, "step": 6591 }, { "epoch": 0.2779265130594262, "grad_norm": 0.7120818495750427, "learning_rate": 4.631112984822935e-05, "loss": 0.4947, "step": 6592 }, { "epoch": 0.27796867424162575, "grad_norm": 0.5640085935592651, "learning_rate": 4.6318156267566046e-05, "loss": 0.1858, "step": 6593 }, { "epoch": 0.27801083542382526, "grad_norm": 0.5656918883323669, "learning_rate": 4.632518268690276e-05, "loss": 0.6028, "step": 6594 }, { "epoch": 0.27805299660602484, "grad_norm": 0.6318227052688599, "learning_rate": 4.633220910623946e-05, "loss": 0.6716, "step": 6595 }, { "epoch": 0.27809515778822436, "grad_norm": 1.639609456062317, "learning_rate": 4.633923552557617e-05, "loss": 0.5196, "step": 6596 }, { "epoch": 0.27813731897042393, "grad_norm": 0.7802035212516785, "learning_rate": 4.634626194491287e-05, "loss": 0.2064, "step": 6597 }, { "epoch": 0.2781794801526235, "grad_norm": 0.39439746737480164, "learning_rate": 4.635328836424958e-05, "loss": 0.1017, "step": 6598 }, { "epoch": 0.278221641334823, "grad_norm": 0.6226504445075989, "learning_rate": 4.6360314783586285e-05, "loss": 0.5968, "step": 6599 }, { "epoch": 0.2782638025170226, "grad_norm": 0.6372654438018799, "learning_rate": 4.636734120292299e-05, "loss": 0.502, "step": 6600 }, { "epoch": 0.2783059636992221, "grad_norm": 0.5626714825630188, "learning_rate": 4.63743676222597e-05, "loss": 0.2147, "step": 6601 }, { "epoch": 0.2783481248814217, "grad_norm": 0.5754001140594482, "learning_rate": 4.6381394041596405e-05, "loss": 0.101, "step": 6602 }, { "epoch": 0.2783902860636212, "grad_norm": 0.45146918296813965, "learning_rate": 4.638842046093311e-05, "loss": 0.1848, "step": 6603 }, { "epoch": 0.2784324472458208, "grad_norm": 0.5504632592201233, "learning_rate": 4.6395446880269814e-05, "loss": 0.2089, "step": 6604 }, { "epoch": 0.2784746084280203, "grad_norm": 0.6329428553581238, "learning_rate": 4.6402473299606525e-05, "loss": 0.4011, "step": 6605 }, { "epoch": 0.27851676961021987, "grad_norm": 1.0527845621109009, "learning_rate": 4.640949971894323e-05, "loss": 1.0869, "step": 6606 }, { "epoch": 0.27855893079241945, "grad_norm": 0.8021690249443054, "learning_rate": 4.6416526138279934e-05, "loss": 0.1473, "step": 6607 }, { "epoch": 0.27860109197461896, "grad_norm": 0.7299266457557678, "learning_rate": 4.642355255761664e-05, "loss": 0.2321, "step": 6608 }, { "epoch": 0.27864325315681854, "grad_norm": 0.7610882520675659, "learning_rate": 4.643057897695335e-05, "loss": 0.5569, "step": 6609 }, { "epoch": 0.27868541433901806, "grad_norm": 1.3395533561706543, "learning_rate": 4.6437605396290054e-05, "loss": 0.6227, "step": 6610 }, { "epoch": 0.27872757552121763, "grad_norm": 0.584578812122345, "learning_rate": 4.644463181562676e-05, "loss": 0.1799, "step": 6611 }, { "epoch": 0.27876973670341715, "grad_norm": 0.6431535482406616, "learning_rate": 4.645165823496346e-05, "loss": 0.2276, "step": 6612 }, { "epoch": 0.2788118978856167, "grad_norm": 0.5847049355506897, "learning_rate": 4.6458684654300174e-05, "loss": 0.1164, "step": 6613 }, { "epoch": 0.27885405906781624, "grad_norm": 0.8716516494750977, "learning_rate": 4.646571107363687e-05, "loss": 0.4437, "step": 6614 }, { "epoch": 0.2788962202500158, "grad_norm": 0.3393620550632477, "learning_rate": 4.647273749297358e-05, "loss": 0.091, "step": 6615 }, { "epoch": 0.27893838143221533, "grad_norm": 0.3528255224227905, "learning_rate": 4.647976391231029e-05, "loss": 0.0915, "step": 6616 }, { "epoch": 0.2789805426144149, "grad_norm": 0.5531798005104065, "learning_rate": 4.6486790331647e-05, "loss": 0.243, "step": 6617 }, { "epoch": 0.2790227037966145, "grad_norm": 0.8952436447143555, "learning_rate": 4.6493816750983696e-05, "loss": 0.2625, "step": 6618 }, { "epoch": 0.279064864978814, "grad_norm": 0.9593549370765686, "learning_rate": 4.650084317032041e-05, "loss": 0.6008, "step": 6619 }, { "epoch": 0.27910702616101357, "grad_norm": 1.991333246231079, "learning_rate": 4.650786958965711e-05, "loss": 0.6856, "step": 6620 }, { "epoch": 0.2791491873432131, "grad_norm": 0.7512890696525574, "learning_rate": 4.6514896008993816e-05, "loss": 0.4095, "step": 6621 }, { "epoch": 0.27919134852541266, "grad_norm": 0.49555736780166626, "learning_rate": 4.652192242833053e-05, "loss": 0.2482, "step": 6622 }, { "epoch": 0.2792335097076122, "grad_norm": 0.6640405654907227, "learning_rate": 4.652894884766723e-05, "loss": 0.1738, "step": 6623 }, { "epoch": 0.27927567088981176, "grad_norm": 0.5294070839881897, "learning_rate": 4.653597526700394e-05, "loss": 0.5433, "step": 6624 }, { "epoch": 0.2793178320720113, "grad_norm": 3.7861204147338867, "learning_rate": 4.654300168634064e-05, "loss": 1.1194, "step": 6625 }, { "epoch": 0.27935999325421085, "grad_norm": 0.7511841058731079, "learning_rate": 4.655002810567735e-05, "loss": 0.1676, "step": 6626 }, { "epoch": 0.2794021544364104, "grad_norm": 0.9478866457939148, "learning_rate": 4.6557054525014056e-05, "loss": 0.2297, "step": 6627 }, { "epoch": 0.27944431561860994, "grad_norm": 0.453695148229599, "learning_rate": 4.656408094435076e-05, "loss": 0.0947, "step": 6628 }, { "epoch": 0.2794864768008095, "grad_norm": 0.7043462991714478, "learning_rate": 4.6571107363687465e-05, "loss": 0.1507, "step": 6629 }, { "epoch": 0.27952863798300903, "grad_norm": 0.5550663471221924, "learning_rate": 4.6578133783024176e-05, "loss": 0.3629, "step": 6630 }, { "epoch": 0.2795707991652086, "grad_norm": 0.43578681349754333, "learning_rate": 4.658516020236088e-05, "loss": 0.1081, "step": 6631 }, { "epoch": 0.2796129603474081, "grad_norm": 1.0468930006027222, "learning_rate": 4.6592186621697585e-05, "loss": 0.6465, "step": 6632 }, { "epoch": 0.2796551215296077, "grad_norm": 0.3329147696495056, "learning_rate": 4.659921304103429e-05, "loss": 0.1002, "step": 6633 }, { "epoch": 0.2796972827118072, "grad_norm": 0.9417448043823242, "learning_rate": 4.6606239460371e-05, "loss": 0.2368, "step": 6634 }, { "epoch": 0.2797394438940068, "grad_norm": 0.42632365226745605, "learning_rate": 4.6613265879707705e-05, "loss": 0.1303, "step": 6635 }, { "epoch": 0.27978160507620636, "grad_norm": 0.7775906324386597, "learning_rate": 4.662029229904441e-05, "loss": 0.4178, "step": 6636 }, { "epoch": 0.2798237662584059, "grad_norm": 1.0565615892410278, "learning_rate": 4.6627318718381113e-05, "loss": 1.0098, "step": 6637 }, { "epoch": 0.27986592744060546, "grad_norm": 0.41260242462158203, "learning_rate": 4.6634345137717825e-05, "loss": 0.2112, "step": 6638 }, { "epoch": 0.279908088622805, "grad_norm": 1.637242317199707, "learning_rate": 4.664137155705452e-05, "loss": 0.6778, "step": 6639 }, { "epoch": 0.27995024980500455, "grad_norm": 0.5846605896949768, "learning_rate": 4.6648397976391233e-05, "loss": 0.1538, "step": 6640 }, { "epoch": 0.27999241098720407, "grad_norm": 2.217982053756714, "learning_rate": 4.665542439572794e-05, "loss": 0.2182, "step": 6641 }, { "epoch": 0.28003457216940364, "grad_norm": 0.3414384722709656, "learning_rate": 4.666245081506464e-05, "loss": 0.1197, "step": 6642 }, { "epoch": 0.28007673335160316, "grad_norm": 0.947547435760498, "learning_rate": 4.6669477234401347e-05, "loss": 0.3846, "step": 6643 }, { "epoch": 0.28011889453380273, "grad_norm": 1.1631767749786377, "learning_rate": 4.667650365373806e-05, "loss": 0.4176, "step": 6644 }, { "epoch": 0.28016105571600225, "grad_norm": 0.6929599642753601, "learning_rate": 4.668353007307477e-05, "loss": 0.4912, "step": 6645 }, { "epoch": 0.2802032168982018, "grad_norm": 0.6382838487625122, "learning_rate": 4.6690556492411467e-05, "loss": 0.6969, "step": 6646 }, { "epoch": 0.2802453780804014, "grad_norm": 0.38130733370780945, "learning_rate": 4.669758291174818e-05, "loss": 0.1591, "step": 6647 }, { "epoch": 0.2802875392626009, "grad_norm": 0.7176529169082642, "learning_rate": 4.670460933108488e-05, "loss": 0.1971, "step": 6648 }, { "epoch": 0.2803297004448005, "grad_norm": 0.5980280637741089, "learning_rate": 4.6711635750421587e-05, "loss": 0.3918, "step": 6649 }, { "epoch": 0.280371861627, "grad_norm": 0.6001821756362915, "learning_rate": 4.671866216975829e-05, "loss": 0.647, "step": 6650 }, { "epoch": 0.2804140228091996, "grad_norm": 0.8208985328674316, "learning_rate": 4.6725688589095e-05, "loss": 0.2079, "step": 6651 }, { "epoch": 0.2804561839913991, "grad_norm": 0.4086107611656189, "learning_rate": 4.6732715008431706e-05, "loss": 0.1622, "step": 6652 }, { "epoch": 0.2804983451735987, "grad_norm": 1.6059497594833374, "learning_rate": 4.673974142776841e-05, "loss": 0.8971, "step": 6653 }, { "epoch": 0.2805405063557982, "grad_norm": 0.4396562874317169, "learning_rate": 4.6746767847105115e-05, "loss": 0.2171, "step": 6654 }, { "epoch": 0.28058266753799777, "grad_norm": 0.5119140148162842, "learning_rate": 4.6753794266441826e-05, "loss": 0.1373, "step": 6655 }, { "epoch": 0.28062482872019734, "grad_norm": 0.8195885419845581, "learning_rate": 4.676082068577853e-05, "loss": 0.8306, "step": 6656 }, { "epoch": 0.28066698990239686, "grad_norm": 0.49934065341949463, "learning_rate": 4.6767847105115235e-05, "loss": 0.1966, "step": 6657 }, { "epoch": 0.28070915108459643, "grad_norm": 0.5188796520233154, "learning_rate": 4.677487352445194e-05, "loss": 0.468, "step": 6658 }, { "epoch": 0.28075131226679595, "grad_norm": 0.51842200756073, "learning_rate": 4.678189994378865e-05, "loss": 0.3822, "step": 6659 }, { "epoch": 0.2807934734489955, "grad_norm": 0.4048662483692169, "learning_rate": 4.678892636312535e-05, "loss": 0.0892, "step": 6660 }, { "epoch": 0.28083563463119504, "grad_norm": 0.6109966039657593, "learning_rate": 4.679595278246206e-05, "loss": 0.5114, "step": 6661 }, { "epoch": 0.2808777958133946, "grad_norm": 0.7834191918373108, "learning_rate": 4.6802979201798764e-05, "loss": 0.2685, "step": 6662 }, { "epoch": 0.28091995699559413, "grad_norm": 0.8476229310035706, "learning_rate": 4.6810005621135475e-05, "loss": 0.2738, "step": 6663 }, { "epoch": 0.2809621181777937, "grad_norm": 0.7021121382713318, "learning_rate": 4.681703204047217e-05, "loss": 0.5167, "step": 6664 }, { "epoch": 0.2810042793599933, "grad_norm": 0.7864976525306702, "learning_rate": 4.6824058459808884e-05, "loss": 0.5067, "step": 6665 }, { "epoch": 0.2810464405421928, "grad_norm": 0.8266788125038147, "learning_rate": 4.683108487914559e-05, "loss": 0.1367, "step": 6666 }, { "epoch": 0.2810886017243924, "grad_norm": 1.2244552373886108, "learning_rate": 4.683811129848229e-05, "loss": 0.2706, "step": 6667 }, { "epoch": 0.2811307629065919, "grad_norm": 0.45950236916542053, "learning_rate": 4.6845137717819004e-05, "loss": 0.165, "step": 6668 }, { "epoch": 0.28117292408879146, "grad_norm": 0.560843288898468, "learning_rate": 4.685216413715571e-05, "loss": 0.598, "step": 6669 }, { "epoch": 0.281215085270991, "grad_norm": 0.32288920879364014, "learning_rate": 4.685919055649241e-05, "loss": 0.1188, "step": 6670 }, { "epoch": 0.28125724645319056, "grad_norm": 0.5464332699775696, "learning_rate": 4.686621697582912e-05, "loss": 0.5224, "step": 6671 }, { "epoch": 0.2812994076353901, "grad_norm": 0.8125031590461731, "learning_rate": 4.687324339516583e-05, "loss": 0.3833, "step": 6672 }, { "epoch": 0.28134156881758965, "grad_norm": 0.45992112159729004, "learning_rate": 4.688026981450253e-05, "loss": 0.1056, "step": 6673 }, { "epoch": 0.28138372999978917, "grad_norm": 0.49201005697250366, "learning_rate": 4.688729623383924e-05, "loss": 0.2039, "step": 6674 }, { "epoch": 0.28142589118198874, "grad_norm": 0.7715864181518555, "learning_rate": 4.689432265317594e-05, "loss": 0.8121, "step": 6675 }, { "epoch": 0.2814680523641883, "grad_norm": 0.8078351020812988, "learning_rate": 4.690134907251265e-05, "loss": 0.3999, "step": 6676 }, { "epoch": 0.28151021354638783, "grad_norm": 0.6697413325309753, "learning_rate": 4.690837549184936e-05, "loss": 0.1898, "step": 6677 }, { "epoch": 0.2815523747285874, "grad_norm": 0.4577898681163788, "learning_rate": 4.691540191118606e-05, "loss": 0.1934, "step": 6678 }, { "epoch": 0.2815945359107869, "grad_norm": 0.7398058176040649, "learning_rate": 4.6922428330522766e-05, "loss": 0.2467, "step": 6679 }, { "epoch": 0.2816366970929865, "grad_norm": 1.4476799964904785, "learning_rate": 4.692945474985948e-05, "loss": 1.0913, "step": 6680 }, { "epoch": 0.281678858275186, "grad_norm": 0.712788462638855, "learning_rate": 4.693648116919618e-05, "loss": 0.6619, "step": 6681 }, { "epoch": 0.2817210194573856, "grad_norm": 0.4449765384197235, "learning_rate": 4.6943507588532886e-05, "loss": 0.1376, "step": 6682 }, { "epoch": 0.2817631806395851, "grad_norm": 0.36259981989860535, "learning_rate": 4.695053400786959e-05, "loss": 0.1337, "step": 6683 }, { "epoch": 0.2818053418217847, "grad_norm": 2.5473814010620117, "learning_rate": 4.69575604272063e-05, "loss": 0.7213, "step": 6684 }, { "epoch": 0.28184750300398426, "grad_norm": 0.378101646900177, "learning_rate": 4.6964586846543e-05, "loss": 0.1015, "step": 6685 }, { "epoch": 0.2818896641861838, "grad_norm": 0.4859020411968231, "learning_rate": 4.697161326587971e-05, "loss": 0.1354, "step": 6686 }, { "epoch": 0.28193182536838335, "grad_norm": 1.140804409980774, "learning_rate": 4.6978639685216415e-05, "loss": 0.3627, "step": 6687 }, { "epoch": 0.28197398655058287, "grad_norm": 0.6632575392723083, "learning_rate": 4.698566610455312e-05, "loss": 0.6431, "step": 6688 }, { "epoch": 0.28201614773278244, "grad_norm": 0.34713074564933777, "learning_rate": 4.699269252388982e-05, "loss": 0.1636, "step": 6689 }, { "epoch": 0.28205830891498196, "grad_norm": 2.8155558109283447, "learning_rate": 4.6999718943226534e-05, "loss": 0.8382, "step": 6690 }, { "epoch": 0.28210047009718153, "grad_norm": 0.9340588450431824, "learning_rate": 4.7006745362563246e-05, "loss": 0.4587, "step": 6691 }, { "epoch": 0.28214263127938105, "grad_norm": 0.2774790823459625, "learning_rate": 4.701377178189994e-05, "loss": 0.1303, "step": 6692 }, { "epoch": 0.2821847924615806, "grad_norm": 0.25628575682640076, "learning_rate": 4.7020798201236654e-05, "loss": 0.0872, "step": 6693 }, { "epoch": 0.2822269536437802, "grad_norm": 0.519151508808136, "learning_rate": 4.702782462057336e-05, "loss": 0.0757, "step": 6694 }, { "epoch": 0.2822691148259797, "grad_norm": 1.2397985458374023, "learning_rate": 4.703485103991006e-05, "loss": 1.175, "step": 6695 }, { "epoch": 0.2823112760081793, "grad_norm": 0.3397352397441864, "learning_rate": 4.704187745924677e-05, "loss": 0.0952, "step": 6696 }, { "epoch": 0.2823534371903788, "grad_norm": 0.5347598791122437, "learning_rate": 4.704890387858348e-05, "loss": 0.223, "step": 6697 }, { "epoch": 0.2823955983725784, "grad_norm": 0.5592583417892456, "learning_rate": 4.705593029792018e-05, "loss": 0.1374, "step": 6698 }, { "epoch": 0.2824377595547779, "grad_norm": 0.3630213141441345, "learning_rate": 4.706295671725689e-05, "loss": 0.088, "step": 6699 }, { "epoch": 0.2824799207369775, "grad_norm": 0.8324704170227051, "learning_rate": 4.706998313659359e-05, "loss": 0.1965, "step": 6700 }, { "epoch": 0.282522081919177, "grad_norm": 1.0367788076400757, "learning_rate": 4.70770095559303e-05, "loss": 0.3937, "step": 6701 }, { "epoch": 0.28256424310137657, "grad_norm": 0.5774011611938477, "learning_rate": 4.708403597526701e-05, "loss": 0.4139, "step": 6702 }, { "epoch": 0.2826064042835761, "grad_norm": 1.1119866371154785, "learning_rate": 4.709106239460371e-05, "loss": 0.6088, "step": 6703 }, { "epoch": 0.28264856546577566, "grad_norm": 3.176835298538208, "learning_rate": 4.7098088813940416e-05, "loss": 1.2747, "step": 6704 }, { "epoch": 0.28269072664797523, "grad_norm": 0.7770178318023682, "learning_rate": 4.710511523327713e-05, "loss": 0.2293, "step": 6705 }, { "epoch": 0.28273288783017475, "grad_norm": 0.8431422710418701, "learning_rate": 4.7112141652613825e-05, "loss": 0.569, "step": 6706 }, { "epoch": 0.2827750490123743, "grad_norm": 0.9966432452201843, "learning_rate": 4.7119168071950536e-05, "loss": 0.761, "step": 6707 }, { "epoch": 0.28281721019457384, "grad_norm": 0.5442252159118652, "learning_rate": 4.712619449128724e-05, "loss": 0.3509, "step": 6708 }, { "epoch": 0.2828593713767734, "grad_norm": 0.8613929152488708, "learning_rate": 4.713322091062395e-05, "loss": 0.6538, "step": 6709 }, { "epoch": 0.28290153255897293, "grad_norm": 0.5120602250099182, "learning_rate": 4.714024732996065e-05, "loss": 0.101, "step": 6710 }, { "epoch": 0.2829436937411725, "grad_norm": 0.3525436818599701, "learning_rate": 4.714727374929736e-05, "loss": 0.117, "step": 6711 }, { "epoch": 0.282985854923372, "grad_norm": 0.5436009168624878, "learning_rate": 4.715430016863407e-05, "loss": 0.1929, "step": 6712 }, { "epoch": 0.2830280161055716, "grad_norm": 0.5846945643424988, "learning_rate": 4.716132658797077e-05, "loss": 0.4782, "step": 6713 }, { "epoch": 0.2830701772877712, "grad_norm": 0.4954264163970947, "learning_rate": 4.716835300730748e-05, "loss": 0.1215, "step": 6714 }, { "epoch": 0.2831123384699707, "grad_norm": 0.5862001180648804, "learning_rate": 4.7175379426644185e-05, "loss": 0.21, "step": 6715 }, { "epoch": 0.28315449965217027, "grad_norm": 0.44315582513809204, "learning_rate": 4.718240584598089e-05, "loss": 0.1246, "step": 6716 }, { "epoch": 0.2831966608343698, "grad_norm": 0.8030522465705872, "learning_rate": 4.7189432265317594e-05, "loss": 0.41, "step": 6717 }, { "epoch": 0.28323882201656936, "grad_norm": 1.022310495376587, "learning_rate": 4.7196458684654305e-05, "loss": 0.1897, "step": 6718 }, { "epoch": 0.2832809831987689, "grad_norm": 0.780165433883667, "learning_rate": 4.720348510399101e-05, "loss": 0.2484, "step": 6719 }, { "epoch": 0.28332314438096845, "grad_norm": 0.35456135869026184, "learning_rate": 4.7210511523327714e-05, "loss": 0.1717, "step": 6720 }, { "epoch": 0.28336530556316797, "grad_norm": 0.4551668167114258, "learning_rate": 4.721753794266442e-05, "loss": 0.0792, "step": 6721 }, { "epoch": 0.28340746674536754, "grad_norm": 12.942176818847656, "learning_rate": 4.722456436200113e-05, "loss": 0.3254, "step": 6722 }, { "epoch": 0.2834496279275671, "grad_norm": 1.4690310955047607, "learning_rate": 4.7231590781337834e-05, "loss": 0.2884, "step": 6723 }, { "epoch": 0.28349178910976663, "grad_norm": 0.36789241433143616, "learning_rate": 4.723861720067454e-05, "loss": 0.1545, "step": 6724 }, { "epoch": 0.2835339502919662, "grad_norm": 0.8256356716156006, "learning_rate": 4.724564362001124e-05, "loss": 0.2017, "step": 6725 }, { "epoch": 0.2835761114741657, "grad_norm": 0.5257163643836975, "learning_rate": 4.7252670039347954e-05, "loss": 0.094, "step": 6726 }, { "epoch": 0.2836182726563653, "grad_norm": 0.7421267032623291, "learning_rate": 4.725969645868465e-05, "loss": 0.5965, "step": 6727 }, { "epoch": 0.2836604338385648, "grad_norm": 0.8407439589500427, "learning_rate": 4.726672287802136e-05, "loss": 0.322, "step": 6728 }, { "epoch": 0.2837025950207644, "grad_norm": 1.5727651119232178, "learning_rate": 4.727374929735807e-05, "loss": 0.4621, "step": 6729 }, { "epoch": 0.2837447562029639, "grad_norm": 0.4421077370643616, "learning_rate": 4.728077571669478e-05, "loss": 0.1616, "step": 6730 }, { "epoch": 0.2837869173851635, "grad_norm": 1.6873891353607178, "learning_rate": 4.7287802136031476e-05, "loss": 0.6281, "step": 6731 }, { "epoch": 0.283829078567363, "grad_norm": 0.6552485227584839, "learning_rate": 4.729482855536819e-05, "loss": 0.4184, "step": 6732 }, { "epoch": 0.2838712397495626, "grad_norm": 0.5065708160400391, "learning_rate": 4.730185497470489e-05, "loss": 0.1818, "step": 6733 }, { "epoch": 0.28391340093176215, "grad_norm": 1.1320310831069946, "learning_rate": 4.7308881394041596e-05, "loss": 0.2283, "step": 6734 }, { "epoch": 0.28395556211396167, "grad_norm": 0.785331666469574, "learning_rate": 4.731590781337831e-05, "loss": 0.1649, "step": 6735 }, { "epoch": 0.28399772329616124, "grad_norm": 0.7540664672851562, "learning_rate": 4.732293423271501e-05, "loss": 0.1848, "step": 6736 }, { "epoch": 0.28403988447836076, "grad_norm": 0.521450936794281, "learning_rate": 4.732996065205172e-05, "loss": 0.3801, "step": 6737 }, { "epoch": 0.28408204566056033, "grad_norm": 0.6462383270263672, "learning_rate": 4.733698707138842e-05, "loss": 0.5717, "step": 6738 }, { "epoch": 0.28412420684275985, "grad_norm": 0.691025972366333, "learning_rate": 4.734401349072513e-05, "loss": 0.5497, "step": 6739 }, { "epoch": 0.2841663680249594, "grad_norm": 2.52242112159729, "learning_rate": 4.7351039910061836e-05, "loss": 1.1048, "step": 6740 }, { "epoch": 0.28420852920715894, "grad_norm": 0.5904001593589783, "learning_rate": 4.735806632939854e-05, "loss": 0.483, "step": 6741 }, { "epoch": 0.2842506903893585, "grad_norm": 0.5492624044418335, "learning_rate": 4.7365092748735244e-05, "loss": 0.2609, "step": 6742 }, { "epoch": 0.2842928515715581, "grad_norm": 0.6425537467002869, "learning_rate": 4.7372119168071956e-05, "loss": 0.3393, "step": 6743 }, { "epoch": 0.2843350127537576, "grad_norm": 3.1746795177459717, "learning_rate": 4.737914558740866e-05, "loss": 0.3994, "step": 6744 }, { "epoch": 0.2843771739359572, "grad_norm": 0.9027732014656067, "learning_rate": 4.7386172006745364e-05, "loss": 0.4732, "step": 6745 }, { "epoch": 0.2844193351181567, "grad_norm": 0.8227939009666443, "learning_rate": 4.739319842608207e-05, "loss": 0.2217, "step": 6746 }, { "epoch": 0.2844614963003563, "grad_norm": 0.31083959341049194, "learning_rate": 4.740022484541878e-05, "loss": 0.0708, "step": 6747 }, { "epoch": 0.2845036574825558, "grad_norm": 1.454559326171875, "learning_rate": 4.7407251264755484e-05, "loss": 0.3148, "step": 6748 }, { "epoch": 0.28454581866475537, "grad_norm": 0.5947179198265076, "learning_rate": 4.741427768409219e-05, "loss": 0.1947, "step": 6749 }, { "epoch": 0.2845879798469549, "grad_norm": 0.4023536145687103, "learning_rate": 4.742130410342889e-05, "loss": 0.0936, "step": 6750 }, { "epoch": 0.28463014102915446, "grad_norm": 0.4789063036441803, "learning_rate": 4.7428330522765604e-05, "loss": 0.2171, "step": 6751 }, { "epoch": 0.28467230221135403, "grad_norm": 0.6375532150268555, "learning_rate": 4.74353569421023e-05, "loss": 0.1955, "step": 6752 }, { "epoch": 0.28471446339355355, "grad_norm": 0.33809778094291687, "learning_rate": 4.744238336143901e-05, "loss": 0.0921, "step": 6753 }, { "epoch": 0.2847566245757531, "grad_norm": 0.9410239458084106, "learning_rate": 4.744940978077572e-05, "loss": 0.9883, "step": 6754 }, { "epoch": 0.28479878575795264, "grad_norm": 0.32534629106521606, "learning_rate": 4.745643620011242e-05, "loss": 0.0808, "step": 6755 }, { "epoch": 0.2848409469401522, "grad_norm": 0.830372154712677, "learning_rate": 4.7463462619449126e-05, "loss": 0.2191, "step": 6756 }, { "epoch": 0.28488310812235174, "grad_norm": 0.7260542511940002, "learning_rate": 4.747048903878584e-05, "loss": 0.5145, "step": 6757 }, { "epoch": 0.2849252693045513, "grad_norm": 0.6415862441062927, "learning_rate": 4.747751545812255e-05, "loss": 0.5387, "step": 6758 }, { "epoch": 0.28496743048675083, "grad_norm": 0.6748367547988892, "learning_rate": 4.7484541877459246e-05, "loss": 0.6644, "step": 6759 }, { "epoch": 0.2850095916689504, "grad_norm": 0.7212662696838379, "learning_rate": 4.749156829679596e-05, "loss": 0.2035, "step": 6760 }, { "epoch": 0.2850517528511499, "grad_norm": 0.885661780834198, "learning_rate": 4.749859471613266e-05, "loss": 0.4682, "step": 6761 }, { "epoch": 0.2850939140333495, "grad_norm": 0.6006132960319519, "learning_rate": 4.7505621135469366e-05, "loss": 0.4648, "step": 6762 }, { "epoch": 0.28513607521554907, "grad_norm": 0.9612435698509216, "learning_rate": 4.751264755480607e-05, "loss": 0.1933, "step": 6763 }, { "epoch": 0.2851782363977486, "grad_norm": 0.5518901348114014, "learning_rate": 4.751967397414278e-05, "loss": 0.121, "step": 6764 }, { "epoch": 0.28522039757994816, "grad_norm": 1.2137446403503418, "learning_rate": 4.7526700393479486e-05, "loss": 0.7425, "step": 6765 }, { "epoch": 0.2852625587621477, "grad_norm": 0.4325915575027466, "learning_rate": 4.753372681281619e-05, "loss": 0.2495, "step": 6766 }, { "epoch": 0.28530471994434725, "grad_norm": 0.6859489679336548, "learning_rate": 4.7540753232152895e-05, "loss": 0.2146, "step": 6767 }, { "epoch": 0.28534688112654677, "grad_norm": 0.5020226240158081, "learning_rate": 4.7547779651489606e-05, "loss": 0.1497, "step": 6768 }, { "epoch": 0.28538904230874634, "grad_norm": 0.9379441142082214, "learning_rate": 4.755480607082631e-05, "loss": 0.3692, "step": 6769 }, { "epoch": 0.28543120349094586, "grad_norm": 0.8436119556427002, "learning_rate": 4.7561832490163015e-05, "loss": 0.1776, "step": 6770 }, { "epoch": 0.28547336467314544, "grad_norm": 0.44468528032302856, "learning_rate": 4.756885890949972e-05, "loss": 0.0721, "step": 6771 }, { "epoch": 0.285515525855345, "grad_norm": 0.6442347168922424, "learning_rate": 4.757588532883643e-05, "loss": 0.3945, "step": 6772 }, { "epoch": 0.2855576870375445, "grad_norm": 0.6051396727561951, "learning_rate": 4.758291174817313e-05, "loss": 0.3059, "step": 6773 }, { "epoch": 0.2855998482197441, "grad_norm": 0.5007705688476562, "learning_rate": 4.758993816750984e-05, "loss": 0.1463, "step": 6774 }, { "epoch": 0.2856420094019436, "grad_norm": 0.7553417682647705, "learning_rate": 4.7596964586846544e-05, "loss": 0.6478, "step": 6775 }, { "epoch": 0.2856841705841432, "grad_norm": 2.049689292907715, "learning_rate": 4.7603991006183255e-05, "loss": 0.7284, "step": 6776 }, { "epoch": 0.2857263317663427, "grad_norm": 0.9901664853096008, "learning_rate": 4.761101742551995e-05, "loss": 0.9249, "step": 6777 }, { "epoch": 0.2857684929485423, "grad_norm": 0.8319377303123474, "learning_rate": 4.7618043844856664e-05, "loss": 0.4141, "step": 6778 }, { "epoch": 0.2858106541307418, "grad_norm": 1.632080316543579, "learning_rate": 4.762507026419337e-05, "loss": 0.9863, "step": 6779 }, { "epoch": 0.2858528153129414, "grad_norm": 0.8054706454277039, "learning_rate": 4.763209668353007e-05, "loss": 0.3353, "step": 6780 }, { "epoch": 0.28589497649514095, "grad_norm": 0.7713080048561096, "learning_rate": 4.7639123102866784e-05, "loss": 0.2707, "step": 6781 }, { "epoch": 0.28593713767734047, "grad_norm": 0.5632901191711426, "learning_rate": 4.764614952220349e-05, "loss": 0.5039, "step": 6782 }, { "epoch": 0.28597929885954004, "grad_norm": 0.754794716835022, "learning_rate": 4.76531759415402e-05, "loss": 0.1802, "step": 6783 }, { "epoch": 0.28602146004173956, "grad_norm": 0.46271243691444397, "learning_rate": 4.76602023608769e-05, "loss": 0.1505, "step": 6784 }, { "epoch": 0.28606362122393914, "grad_norm": 0.3983902335166931, "learning_rate": 4.766722878021361e-05, "loss": 0.1309, "step": 6785 }, { "epoch": 0.28610578240613865, "grad_norm": 0.7557656764984131, "learning_rate": 4.767425519955031e-05, "loss": 0.3768, "step": 6786 }, { "epoch": 0.2861479435883382, "grad_norm": 0.5441089272499084, "learning_rate": 4.768128161888702e-05, "loss": 0.2623, "step": 6787 }, { "epoch": 0.28619010477053775, "grad_norm": 1.4101881980895996, "learning_rate": 4.768830803822372e-05, "loss": 1.0926, "step": 6788 }, { "epoch": 0.2862322659527373, "grad_norm": 0.8265970349311829, "learning_rate": 4.769533445756043e-05, "loss": 0.5404, "step": 6789 }, { "epoch": 0.28627442713493684, "grad_norm": 1.6788034439086914, "learning_rate": 4.770236087689714e-05, "loss": 0.7272, "step": 6790 }, { "epoch": 0.2863165883171364, "grad_norm": 0.39628592133522034, "learning_rate": 4.770938729623384e-05, "loss": 0.1169, "step": 6791 }, { "epoch": 0.286358749499336, "grad_norm": 0.6496573090553284, "learning_rate": 4.7716413715570545e-05, "loss": 0.4778, "step": 6792 }, { "epoch": 0.2864009106815355, "grad_norm": 1.3023022413253784, "learning_rate": 4.7723440134907257e-05, "loss": 0.2865, "step": 6793 }, { "epoch": 0.2864430718637351, "grad_norm": 0.6239800453186035, "learning_rate": 4.773046655424396e-05, "loss": 0.6521, "step": 6794 }, { "epoch": 0.2864852330459346, "grad_norm": 0.7644243240356445, "learning_rate": 4.7737492973580665e-05, "loss": 0.2412, "step": 6795 }, { "epoch": 0.28652739422813417, "grad_norm": 0.6053683161735535, "learning_rate": 4.774451939291737e-05, "loss": 0.6854, "step": 6796 }, { "epoch": 0.2865695554103337, "grad_norm": 0.9708220958709717, "learning_rate": 4.775154581225408e-05, "loss": 0.2634, "step": 6797 }, { "epoch": 0.28661171659253326, "grad_norm": 2.054795265197754, "learning_rate": 4.775857223159078e-05, "loss": 0.6584, "step": 6798 }, { "epoch": 0.2866538777747328, "grad_norm": 0.7621669173240662, "learning_rate": 4.776559865092749e-05, "loss": 0.8632, "step": 6799 }, { "epoch": 0.28669603895693235, "grad_norm": 0.6990841627120972, "learning_rate": 4.7772625070264194e-05, "loss": 0.3355, "step": 6800 }, { "epoch": 0.2867382001391319, "grad_norm": 2.037100315093994, "learning_rate": 4.77796514896009e-05, "loss": 0.8696, "step": 6801 }, { "epoch": 0.28678036132133145, "grad_norm": 0.44465330243110657, "learning_rate": 4.778667790893761e-05, "loss": 0.0924, "step": 6802 }, { "epoch": 0.286822522503531, "grad_norm": 0.2840925455093384, "learning_rate": 4.7793704328274314e-05, "loss": 0.1191, "step": 6803 }, { "epoch": 0.28686468368573054, "grad_norm": 3.235287666320801, "learning_rate": 4.7800730747611025e-05, "loss": 0.6203, "step": 6804 }, { "epoch": 0.2869068448679301, "grad_norm": 1.857740044593811, "learning_rate": 4.780775716694772e-05, "loss": 0.6561, "step": 6805 }, { "epoch": 0.28694900605012963, "grad_norm": 0.48623210191726685, "learning_rate": 4.7814783586284434e-05, "loss": 0.4287, "step": 6806 }, { "epoch": 0.2869911672323292, "grad_norm": 0.6245070099830627, "learning_rate": 4.782181000562114e-05, "loss": 0.4823, "step": 6807 }, { "epoch": 0.2870333284145287, "grad_norm": 0.7084903717041016, "learning_rate": 4.782883642495784e-05, "loss": 0.2165, "step": 6808 }, { "epoch": 0.2870754895967283, "grad_norm": 0.32268884778022766, "learning_rate": 4.783586284429455e-05, "loss": 0.1263, "step": 6809 }, { "epoch": 0.28711765077892787, "grad_norm": 0.9381173253059387, "learning_rate": 4.784288926363126e-05, "loss": 0.9337, "step": 6810 }, { "epoch": 0.2871598119611274, "grad_norm": 0.8506689071655273, "learning_rate": 4.784991568296796e-05, "loss": 0.2845, "step": 6811 }, { "epoch": 0.28720197314332696, "grad_norm": 1.032824158668518, "learning_rate": 4.785694210230467e-05, "loss": 0.1946, "step": 6812 }, { "epoch": 0.2872441343255265, "grad_norm": 0.9491596221923828, "learning_rate": 4.786396852164137e-05, "loss": 0.4386, "step": 6813 }, { "epoch": 0.28728629550772605, "grad_norm": 0.3611345887184143, "learning_rate": 4.787099494097808e-05, "loss": 0.0792, "step": 6814 }, { "epoch": 0.28732845668992557, "grad_norm": 0.630133867263794, "learning_rate": 4.787802136031479e-05, "loss": 0.1339, "step": 6815 }, { "epoch": 0.28737061787212514, "grad_norm": 0.6009783148765564, "learning_rate": 4.788504777965149e-05, "loss": 0.4116, "step": 6816 }, { "epoch": 0.28741277905432466, "grad_norm": 0.6386522054672241, "learning_rate": 4.7892074198988196e-05, "loss": 0.0719, "step": 6817 }, { "epoch": 0.28745494023652424, "grad_norm": 0.6972795724868774, "learning_rate": 4.789910061832491e-05, "loss": 0.4898, "step": 6818 }, { "epoch": 0.28749710141872376, "grad_norm": 0.36232316493988037, "learning_rate": 4.7906127037661605e-05, "loss": 0.1162, "step": 6819 }, { "epoch": 0.28753926260092333, "grad_norm": 0.6446795463562012, "learning_rate": 4.7913153456998316e-05, "loss": 0.1791, "step": 6820 }, { "epoch": 0.2875814237831229, "grad_norm": 0.3839791417121887, "learning_rate": 4.792017987633502e-05, "loss": 0.1301, "step": 6821 }, { "epoch": 0.2876235849653224, "grad_norm": 0.5732271671295166, "learning_rate": 4.792720629567173e-05, "loss": 0.2849, "step": 6822 }, { "epoch": 0.287665746147522, "grad_norm": 0.5140169262886047, "learning_rate": 4.793423271500843e-05, "loss": 0.1179, "step": 6823 }, { "epoch": 0.2877079073297215, "grad_norm": 0.5673938989639282, "learning_rate": 4.794125913434514e-05, "loss": 0.5712, "step": 6824 }, { "epoch": 0.2877500685119211, "grad_norm": 0.6247557997703552, "learning_rate": 4.794828555368185e-05, "loss": 0.1518, "step": 6825 }, { "epoch": 0.2877922296941206, "grad_norm": 0.431456059217453, "learning_rate": 4.795531197301855e-05, "loss": 0.1979, "step": 6826 }, { "epoch": 0.2878343908763202, "grad_norm": 2.0421676635742188, "learning_rate": 4.796233839235526e-05, "loss": 0.7386, "step": 6827 }, { "epoch": 0.2878765520585197, "grad_norm": 0.8339139819145203, "learning_rate": 4.7969364811691965e-05, "loss": 0.2754, "step": 6828 }, { "epoch": 0.28791871324071927, "grad_norm": 0.525452196598053, "learning_rate": 4.797639123102867e-05, "loss": 0.1966, "step": 6829 }, { "epoch": 0.28796087442291884, "grad_norm": 0.3737293481826782, "learning_rate": 4.7983417650365373e-05, "loss": 0.1025, "step": 6830 }, { "epoch": 0.28800303560511836, "grad_norm": 0.41282686591148376, "learning_rate": 4.7990444069702085e-05, "loss": 0.1566, "step": 6831 }, { "epoch": 0.28804519678731794, "grad_norm": 0.6285820603370667, "learning_rate": 4.799747048903879e-05, "loss": 0.5725, "step": 6832 }, { "epoch": 0.28808735796951745, "grad_norm": 1.108169436454773, "learning_rate": 4.800449690837549e-05, "loss": 0.5906, "step": 6833 }, { "epoch": 0.28812951915171703, "grad_norm": 1.3964293003082275, "learning_rate": 4.80115233277122e-05, "loss": 0.5624, "step": 6834 }, { "epoch": 0.28817168033391655, "grad_norm": 0.6406632661819458, "learning_rate": 4.801854974704891e-05, "loss": 0.3046, "step": 6835 }, { "epoch": 0.2882138415161161, "grad_norm": 0.4253230392932892, "learning_rate": 4.802557616638561e-05, "loss": 0.2441, "step": 6836 }, { "epoch": 0.28825600269831564, "grad_norm": 1.106972575187683, "learning_rate": 4.803260258572232e-05, "loss": 0.9267, "step": 6837 }, { "epoch": 0.2882981638805152, "grad_norm": 0.34005919098854065, "learning_rate": 4.803962900505902e-05, "loss": 0.1262, "step": 6838 }, { "epoch": 0.2883403250627148, "grad_norm": 0.5963737964630127, "learning_rate": 4.804665542439573e-05, "loss": 0.0931, "step": 6839 }, { "epoch": 0.2883824862449143, "grad_norm": 0.6860076189041138, "learning_rate": 4.805368184373243e-05, "loss": 0.1415, "step": 6840 }, { "epoch": 0.2884246474271139, "grad_norm": 1.0590275526046753, "learning_rate": 4.806070826306914e-05, "loss": 0.1858, "step": 6841 }, { "epoch": 0.2884668086093134, "grad_norm": 0.7485301494598389, "learning_rate": 4.8067734682405846e-05, "loss": 0.3997, "step": 6842 }, { "epoch": 0.28850896979151297, "grad_norm": 0.5079236030578613, "learning_rate": 4.807476110174256e-05, "loss": 0.1392, "step": 6843 }, { "epoch": 0.2885511309737125, "grad_norm": 1.7357118129730225, "learning_rate": 4.8081787521079255e-05, "loss": 0.4004, "step": 6844 }, { "epoch": 0.28859329215591206, "grad_norm": 0.8529473543167114, "learning_rate": 4.8088813940415966e-05, "loss": 0.4147, "step": 6845 }, { "epoch": 0.2886354533381116, "grad_norm": 0.6556881070137024, "learning_rate": 4.809584035975267e-05, "loss": 0.5321, "step": 6846 }, { "epoch": 0.28867761452031115, "grad_norm": 0.33458957076072693, "learning_rate": 4.8102866779089375e-05, "loss": 0.1119, "step": 6847 }, { "epoch": 0.2887197757025107, "grad_norm": 0.56710284948349, "learning_rate": 4.8109893198426086e-05, "loss": 0.1641, "step": 6848 }, { "epoch": 0.28876193688471025, "grad_norm": 1.0099906921386719, "learning_rate": 4.811691961776279e-05, "loss": 0.3203, "step": 6849 }, { "epoch": 0.2888040980669098, "grad_norm": 1.0376588106155396, "learning_rate": 4.81239460370995e-05, "loss": 0.4314, "step": 6850 }, { "epoch": 0.28884625924910934, "grad_norm": 0.5648683905601501, "learning_rate": 4.81309724564362e-05, "loss": 0.1981, "step": 6851 }, { "epoch": 0.2888884204313089, "grad_norm": 0.6451598405838013, "learning_rate": 4.813799887577291e-05, "loss": 0.2541, "step": 6852 }, { "epoch": 0.28893058161350843, "grad_norm": 6.130153656005859, "learning_rate": 4.8145025295109615e-05, "loss": 0.5683, "step": 6853 }, { "epoch": 0.288972742795708, "grad_norm": 0.5197399258613586, "learning_rate": 4.815205171444632e-05, "loss": 0.2926, "step": 6854 }, { "epoch": 0.2890149039779075, "grad_norm": 0.5970582365989685, "learning_rate": 4.8159078133783024e-05, "loss": 0.4039, "step": 6855 }, { "epoch": 0.2890570651601071, "grad_norm": 0.6129202842712402, "learning_rate": 4.8166104553119735e-05, "loss": 0.71, "step": 6856 }, { "epoch": 0.2890992263423066, "grad_norm": 0.6304472088813782, "learning_rate": 4.817313097245644e-05, "loss": 0.2526, "step": 6857 }, { "epoch": 0.2891413875245062, "grad_norm": 0.8726878762245178, "learning_rate": 4.8180157391793144e-05, "loss": 0.1913, "step": 6858 }, { "epoch": 0.28918354870670576, "grad_norm": 1.0462192296981812, "learning_rate": 4.818718381112985e-05, "loss": 1.068, "step": 6859 }, { "epoch": 0.2892257098889053, "grad_norm": 0.7679135799407959, "learning_rate": 4.819421023046656e-05, "loss": 0.2346, "step": 6860 }, { "epoch": 0.28926787107110485, "grad_norm": 0.612008810043335, "learning_rate": 4.8201236649803264e-05, "loss": 0.2399, "step": 6861 }, { "epoch": 0.28931003225330437, "grad_norm": 0.6129389405250549, "learning_rate": 4.820826306913997e-05, "loss": 0.6671, "step": 6862 }, { "epoch": 0.28935219343550395, "grad_norm": 0.8022855520248413, "learning_rate": 4.821528948847667e-05, "loss": 0.2663, "step": 6863 }, { "epoch": 0.28939435461770346, "grad_norm": 0.45062556862831116, "learning_rate": 4.8222315907813384e-05, "loss": 0.4002, "step": 6864 }, { "epoch": 0.28943651579990304, "grad_norm": 0.8176048994064331, "learning_rate": 4.822934232715008e-05, "loss": 0.6606, "step": 6865 }, { "epoch": 0.28947867698210256, "grad_norm": 0.6772136092185974, "learning_rate": 4.823636874648679e-05, "loss": 0.7253, "step": 6866 }, { "epoch": 0.28952083816430213, "grad_norm": 0.7823106646537781, "learning_rate": 4.82433951658235e-05, "loss": 0.5539, "step": 6867 }, { "epoch": 0.28956299934650165, "grad_norm": 0.6210682988166809, "learning_rate": 4.825042158516021e-05, "loss": 0.5307, "step": 6868 }, { "epoch": 0.2896051605287012, "grad_norm": 0.85365229845047, "learning_rate": 4.8257448004496906e-05, "loss": 0.3655, "step": 6869 }, { "epoch": 0.2896473217109008, "grad_norm": 0.7595174908638, "learning_rate": 4.826447442383362e-05, "loss": 0.2147, "step": 6870 }, { "epoch": 0.2896894828931003, "grad_norm": 1.0239391326904297, "learning_rate": 4.827150084317033e-05, "loss": 1.0217, "step": 6871 }, { "epoch": 0.2897316440752999, "grad_norm": 1.3656154870986938, "learning_rate": 4.8278527262507026e-05, "loss": 0.3908, "step": 6872 }, { "epoch": 0.2897738052574994, "grad_norm": 0.580905020236969, "learning_rate": 4.828555368184374e-05, "loss": 0.4752, "step": 6873 }, { "epoch": 0.289815966439699, "grad_norm": 0.6272801756858826, "learning_rate": 4.829258010118044e-05, "loss": 0.4967, "step": 6874 }, { "epoch": 0.2898581276218985, "grad_norm": 0.5622564554214478, "learning_rate": 4.8299606520517146e-05, "loss": 0.2322, "step": 6875 }, { "epoch": 0.28990028880409807, "grad_norm": 0.6979647278785706, "learning_rate": 4.830663293985385e-05, "loss": 0.7374, "step": 6876 }, { "epoch": 0.2899424499862976, "grad_norm": 0.9645177721977234, "learning_rate": 4.831365935919056e-05, "loss": 0.952, "step": 6877 }, { "epoch": 0.28998461116849716, "grad_norm": 0.9323081374168396, "learning_rate": 4.8320685778527266e-05, "loss": 1.1421, "step": 6878 }, { "epoch": 0.29002677235069674, "grad_norm": 0.5985130071640015, "learning_rate": 4.832771219786397e-05, "loss": 0.194, "step": 6879 }, { "epoch": 0.29006893353289626, "grad_norm": 0.5932117104530334, "learning_rate": 4.8334738617200674e-05, "loss": 0.5617, "step": 6880 }, { "epoch": 0.29011109471509583, "grad_norm": 0.6661390066146851, "learning_rate": 4.8341765036537386e-05, "loss": 0.4488, "step": 6881 }, { "epoch": 0.29015325589729535, "grad_norm": 0.9369485974311829, "learning_rate": 4.834879145587409e-05, "loss": 0.4091, "step": 6882 }, { "epoch": 0.2901954170794949, "grad_norm": 0.7559217214584351, "learning_rate": 4.8355817875210794e-05, "loss": 0.1833, "step": 6883 }, { "epoch": 0.29023757826169444, "grad_norm": 0.5395531058311462, "learning_rate": 4.83628442945475e-05, "loss": 0.1794, "step": 6884 }, { "epoch": 0.290279739443894, "grad_norm": 0.6564905047416687, "learning_rate": 4.836987071388421e-05, "loss": 0.4704, "step": 6885 }, { "epoch": 0.29032190062609353, "grad_norm": 0.5146715044975281, "learning_rate": 4.837689713322091e-05, "loss": 0.2963, "step": 6886 }, { "epoch": 0.2903640618082931, "grad_norm": 0.5320271849632263, "learning_rate": 4.838392355255762e-05, "loss": 0.315, "step": 6887 }, { "epoch": 0.2904062229904927, "grad_norm": 0.5704553127288818, "learning_rate": 4.839094997189432e-05, "loss": 0.14, "step": 6888 }, { "epoch": 0.2904483841726922, "grad_norm": 0.9665951728820801, "learning_rate": 4.8397976391231034e-05, "loss": 0.2527, "step": 6889 }, { "epoch": 0.29049054535489177, "grad_norm": 0.5133302807807922, "learning_rate": 4.840500281056773e-05, "loss": 0.0867, "step": 6890 }, { "epoch": 0.2905327065370913, "grad_norm": 0.879673182964325, "learning_rate": 4.841202922990444e-05, "loss": 0.62, "step": 6891 }, { "epoch": 0.29057486771929086, "grad_norm": 0.4217681884765625, "learning_rate": 4.8419055649241154e-05, "loss": 0.0971, "step": 6892 }, { "epoch": 0.2906170289014904, "grad_norm": 0.41723349690437317, "learning_rate": 4.842608206857785e-05, "loss": 0.0911, "step": 6893 }, { "epoch": 0.29065919008368996, "grad_norm": 1.8850574493408203, "learning_rate": 4.843310848791456e-05, "loss": 0.5268, "step": 6894 }, { "epoch": 0.2907013512658895, "grad_norm": 0.8169699311256409, "learning_rate": 4.844013490725127e-05, "loss": 0.611, "step": 6895 }, { "epoch": 0.29074351244808905, "grad_norm": 0.6561939120292664, "learning_rate": 4.844716132658798e-05, "loss": 0.6665, "step": 6896 }, { "epoch": 0.29078567363028857, "grad_norm": 1.798453450202942, "learning_rate": 4.8454187745924676e-05, "loss": 0.4362, "step": 6897 }, { "epoch": 0.29082783481248814, "grad_norm": 0.6696305871009827, "learning_rate": 4.846121416526139e-05, "loss": 0.1462, "step": 6898 }, { "epoch": 0.2908699959946877, "grad_norm": 0.6376819014549255, "learning_rate": 4.846824058459809e-05, "loss": 0.2251, "step": 6899 }, { "epoch": 0.29091215717688723, "grad_norm": 2.4123761653900146, "learning_rate": 4.8475267003934796e-05, "loss": 1.258, "step": 6900 }, { "epoch": 0.2909543183590868, "grad_norm": 0.5175679326057434, "learning_rate": 4.84822934232715e-05, "loss": 0.1401, "step": 6901 }, { "epoch": 0.2909964795412863, "grad_norm": 3.508561134338379, "learning_rate": 4.848931984260821e-05, "loss": 0.4398, "step": 6902 }, { "epoch": 0.2910386407234859, "grad_norm": 0.5074993968009949, "learning_rate": 4.8496346261944916e-05, "loss": 0.1171, "step": 6903 }, { "epoch": 0.2910808019056854, "grad_norm": 0.847979724407196, "learning_rate": 4.850337268128162e-05, "loss": 0.9578, "step": 6904 }, { "epoch": 0.291122963087885, "grad_norm": 0.3798757791519165, "learning_rate": 4.8510399100618325e-05, "loss": 0.1363, "step": 6905 }, { "epoch": 0.2911651242700845, "grad_norm": 0.5914956331253052, "learning_rate": 4.8517425519955036e-05, "loss": 0.1851, "step": 6906 }, { "epoch": 0.2912072854522841, "grad_norm": 0.5850551128387451, "learning_rate": 4.852445193929174e-05, "loss": 0.2166, "step": 6907 }, { "epoch": 0.29124944663448366, "grad_norm": 0.6435558199882507, "learning_rate": 4.8531478358628445e-05, "loss": 0.4852, "step": 6908 }, { "epoch": 0.2912916078166832, "grad_norm": 0.5513253211975098, "learning_rate": 4.853850477796515e-05, "loss": 0.0931, "step": 6909 }, { "epoch": 0.29133376899888275, "grad_norm": 0.44381821155548096, "learning_rate": 4.854553119730186e-05, "loss": 0.104, "step": 6910 }, { "epoch": 0.29137593018108227, "grad_norm": 0.8898392915725708, "learning_rate": 4.855255761663856e-05, "loss": 0.7054, "step": 6911 }, { "epoch": 0.29141809136328184, "grad_norm": 0.4136263132095337, "learning_rate": 4.855958403597527e-05, "loss": 0.1201, "step": 6912 }, { "epoch": 0.29146025254548136, "grad_norm": 0.6096224784851074, "learning_rate": 4.8566610455311974e-05, "loss": 0.5807, "step": 6913 }, { "epoch": 0.29150241372768093, "grad_norm": 1.733804702758789, "learning_rate": 4.857363687464868e-05, "loss": 0.9422, "step": 6914 }, { "epoch": 0.29154457490988045, "grad_norm": 0.2770509421825409, "learning_rate": 4.858066329398539e-05, "loss": 0.0815, "step": 6915 }, { "epoch": 0.29158673609208, "grad_norm": 1.2923858165740967, "learning_rate": 4.8587689713322094e-05, "loss": 1.0464, "step": 6916 }, { "epoch": 0.2916288972742796, "grad_norm": 3.809277057647705, "learning_rate": 4.8594716132658805e-05, "loss": 1.1679, "step": 6917 }, { "epoch": 0.2916710584564791, "grad_norm": 0.5662473440170288, "learning_rate": 4.86017425519955e-05, "loss": 0.1402, "step": 6918 }, { "epoch": 0.2917132196386787, "grad_norm": 1.715061902999878, "learning_rate": 4.8608768971332214e-05, "loss": 1.0989, "step": 6919 }, { "epoch": 0.2917553808208782, "grad_norm": 0.7270501852035522, "learning_rate": 4.861579539066892e-05, "loss": 0.5495, "step": 6920 }, { "epoch": 0.2917975420030778, "grad_norm": 1.1231662034988403, "learning_rate": 4.862282181000562e-05, "loss": 0.1817, "step": 6921 }, { "epoch": 0.2918397031852773, "grad_norm": 0.5427102446556091, "learning_rate": 4.862984822934233e-05, "loss": 0.1431, "step": 6922 }, { "epoch": 0.2918818643674769, "grad_norm": 1.2792307138442993, "learning_rate": 4.863687464867904e-05, "loss": 1.0349, "step": 6923 }, { "epoch": 0.2919240255496764, "grad_norm": 0.29548168182373047, "learning_rate": 4.864390106801574e-05, "loss": 0.0942, "step": 6924 }, { "epoch": 0.29196618673187597, "grad_norm": 0.9529247283935547, "learning_rate": 4.865092748735245e-05, "loss": 0.7125, "step": 6925 }, { "epoch": 0.2920083479140755, "grad_norm": 0.5690625309944153, "learning_rate": 4.865795390668915e-05, "loss": 0.1481, "step": 6926 }, { "epoch": 0.29205050909627506, "grad_norm": 0.397566556930542, "learning_rate": 4.866498032602586e-05, "loss": 0.0983, "step": 6927 }, { "epoch": 0.29209267027847463, "grad_norm": 0.4828674793243408, "learning_rate": 4.867200674536257e-05, "loss": 0.1973, "step": 6928 }, { "epoch": 0.29213483146067415, "grad_norm": 1.043542504310608, "learning_rate": 4.867903316469927e-05, "loss": 0.6782, "step": 6929 }, { "epoch": 0.2921769926428737, "grad_norm": 0.842239260673523, "learning_rate": 4.8686059584035976e-05, "loss": 0.334, "step": 6930 }, { "epoch": 0.29221915382507324, "grad_norm": 0.579984724521637, "learning_rate": 4.869308600337269e-05, "loss": 0.1334, "step": 6931 }, { "epoch": 0.2922613150072728, "grad_norm": 1.1209907531738281, "learning_rate": 4.8700112422709384e-05, "loss": 0.4632, "step": 6932 }, { "epoch": 0.29230347618947233, "grad_norm": 0.6062950491905212, "learning_rate": 4.8707138842046096e-05, "loss": 0.2058, "step": 6933 }, { "epoch": 0.2923456373716719, "grad_norm": 0.612022876739502, "learning_rate": 4.87141652613828e-05, "loss": 0.3474, "step": 6934 }, { "epoch": 0.2923877985538714, "grad_norm": 0.634057879447937, "learning_rate": 4.872119168071951e-05, "loss": 0.5835, "step": 6935 }, { "epoch": 0.292429959736071, "grad_norm": 0.8613002300262451, "learning_rate": 4.872821810005621e-05, "loss": 0.4787, "step": 6936 }, { "epoch": 0.2924721209182706, "grad_norm": 0.8492224216461182, "learning_rate": 4.873524451939292e-05, "loss": 0.2647, "step": 6937 }, { "epoch": 0.2925142821004701, "grad_norm": 0.7896605134010315, "learning_rate": 4.874227093872963e-05, "loss": 0.4364, "step": 6938 }, { "epoch": 0.29255644328266966, "grad_norm": 0.5980191826820374, "learning_rate": 4.874929735806633e-05, "loss": 0.1553, "step": 6939 }, { "epoch": 0.2925986044648692, "grad_norm": 1.7845669984817505, "learning_rate": 4.875632377740304e-05, "loss": 0.5691, "step": 6940 }, { "epoch": 0.29264076564706876, "grad_norm": 1.6900001764297485, "learning_rate": 4.8763350196739744e-05, "loss": 0.4664, "step": 6941 }, { "epoch": 0.2926829268292683, "grad_norm": 0.3733469843864441, "learning_rate": 4.877037661607645e-05, "loss": 0.1725, "step": 6942 }, { "epoch": 0.29272508801146785, "grad_norm": 0.6001032590866089, "learning_rate": 4.877740303541315e-05, "loss": 0.5445, "step": 6943 }, { "epoch": 0.29276724919366737, "grad_norm": 0.4485958516597748, "learning_rate": 4.8784429454749864e-05, "loss": 0.1272, "step": 6944 }, { "epoch": 0.29280941037586694, "grad_norm": 1.0656840801239014, "learning_rate": 4.879145587408657e-05, "loss": 0.1969, "step": 6945 }, { "epoch": 0.2928515715580665, "grad_norm": 0.4017028510570526, "learning_rate": 4.879848229342327e-05, "loss": 0.089, "step": 6946 }, { "epoch": 0.29289373274026603, "grad_norm": 0.893657922744751, "learning_rate": 4.880550871275998e-05, "loss": 0.7192, "step": 6947 }, { "epoch": 0.2929358939224656, "grad_norm": 1.0865038633346558, "learning_rate": 4.881253513209669e-05, "loss": 1.1168, "step": 6948 }, { "epoch": 0.2929780551046651, "grad_norm": 1.7811936140060425, "learning_rate": 4.881956155143339e-05, "loss": 0.8738, "step": 6949 }, { "epoch": 0.2930202162868647, "grad_norm": 0.6738539338111877, "learning_rate": 4.88265879707701e-05, "loss": 0.6323, "step": 6950 }, { "epoch": 0.2930623774690642, "grad_norm": 0.7207532525062561, "learning_rate": 4.88336143901068e-05, "loss": 0.4316, "step": 6951 }, { "epoch": 0.2931045386512638, "grad_norm": 0.9322458505630493, "learning_rate": 4.884064080944351e-05, "loss": 1.2532, "step": 6952 }, { "epoch": 0.2931466998334633, "grad_norm": 0.7351142764091492, "learning_rate": 4.884766722878022e-05, "loss": 0.674, "step": 6953 }, { "epoch": 0.2931888610156629, "grad_norm": 0.7669952511787415, "learning_rate": 4.885469364811692e-05, "loss": 0.2687, "step": 6954 }, { "epoch": 0.2932310221978624, "grad_norm": 2.9810216426849365, "learning_rate": 4.8861720067453626e-05, "loss": 1.1264, "step": 6955 }, { "epoch": 0.293273183380062, "grad_norm": 1.1354172229766846, "learning_rate": 4.886874648679034e-05, "loss": 0.4481, "step": 6956 }, { "epoch": 0.29331534456226155, "grad_norm": 0.6875216364860535, "learning_rate": 4.8875772906127035e-05, "loss": 0.238, "step": 6957 }, { "epoch": 0.29335750574446107, "grad_norm": 0.7288676500320435, "learning_rate": 4.8882799325463746e-05, "loss": 0.4465, "step": 6958 }, { "epoch": 0.29339966692666064, "grad_norm": 1.1802787780761719, "learning_rate": 4.888982574480045e-05, "loss": 1.0578, "step": 6959 }, { "epoch": 0.29344182810886016, "grad_norm": 0.6529518365859985, "learning_rate": 4.8896852164137155e-05, "loss": 0.5848, "step": 6960 }, { "epoch": 0.29348398929105973, "grad_norm": 1.3979346752166748, "learning_rate": 4.8903878583473866e-05, "loss": 0.2866, "step": 6961 }, { "epoch": 0.29352615047325925, "grad_norm": 0.8326793909072876, "learning_rate": 4.891090500281057e-05, "loss": 0.152, "step": 6962 }, { "epoch": 0.2935683116554588, "grad_norm": 0.9706668257713318, "learning_rate": 4.891793142214728e-05, "loss": 0.576, "step": 6963 }, { "epoch": 0.29361047283765834, "grad_norm": 0.7011312246322632, "learning_rate": 4.892495784148398e-05, "loss": 0.2139, "step": 6964 }, { "epoch": 0.2936526340198579, "grad_norm": 0.6657605171203613, "learning_rate": 4.893198426082069e-05, "loss": 0.2355, "step": 6965 }, { "epoch": 0.2936947952020575, "grad_norm": 0.45804184675216675, "learning_rate": 4.8939010680157395e-05, "loss": 0.1594, "step": 6966 }, { "epoch": 0.293736956384257, "grad_norm": 1.0085043907165527, "learning_rate": 4.89460370994941e-05, "loss": 0.2559, "step": 6967 }, { "epoch": 0.2937791175664566, "grad_norm": 0.5581660270690918, "learning_rate": 4.8953063518830804e-05, "loss": 0.0984, "step": 6968 }, { "epoch": 0.2938212787486561, "grad_norm": 0.5426018834114075, "learning_rate": 4.8960089938167515e-05, "loss": 0.3417, "step": 6969 }, { "epoch": 0.2938634399308557, "grad_norm": 0.543735921382904, "learning_rate": 4.896711635750422e-05, "loss": 0.4264, "step": 6970 }, { "epoch": 0.2939056011130552, "grad_norm": 0.4210697412490845, "learning_rate": 4.8974142776840924e-05, "loss": 0.1093, "step": 6971 }, { "epoch": 0.29394776229525477, "grad_norm": 1.0518848896026611, "learning_rate": 4.898116919617763e-05, "loss": 0.7158, "step": 6972 }, { "epoch": 0.2939899234774543, "grad_norm": 0.6354512572288513, "learning_rate": 4.898819561551434e-05, "loss": 0.2118, "step": 6973 }, { "epoch": 0.29403208465965386, "grad_norm": 0.6966797709465027, "learning_rate": 4.8995222034851043e-05, "loss": 0.1213, "step": 6974 }, { "epoch": 0.29407424584185343, "grad_norm": 0.4644117057323456, "learning_rate": 4.900224845418775e-05, "loss": 0.1799, "step": 6975 }, { "epoch": 0.29411640702405295, "grad_norm": 0.5351444482803345, "learning_rate": 4.900927487352445e-05, "loss": 0.18, "step": 6976 }, { "epoch": 0.2941585682062525, "grad_norm": 0.6582663655281067, "learning_rate": 4.9016301292861163e-05, "loss": 0.5828, "step": 6977 }, { "epoch": 0.29420072938845204, "grad_norm": 0.48191750049591064, "learning_rate": 4.902332771219786e-05, "loss": 0.1288, "step": 6978 }, { "epoch": 0.2942428905706516, "grad_norm": 1.1284674406051636, "learning_rate": 4.903035413153457e-05, "loss": 0.3502, "step": 6979 }, { "epoch": 0.29428505175285113, "grad_norm": 0.5807181596755981, "learning_rate": 4.903738055087128e-05, "loss": 0.124, "step": 6980 }, { "epoch": 0.2943272129350507, "grad_norm": 1.289544701576233, "learning_rate": 4.904440697020799e-05, "loss": 0.7064, "step": 6981 }, { "epoch": 0.2943693741172502, "grad_norm": 0.7024269104003906, "learning_rate": 4.905143338954469e-05, "loss": 0.5175, "step": 6982 }, { "epoch": 0.2944115352994498, "grad_norm": 0.48337626457214355, "learning_rate": 4.9058459808881397e-05, "loss": 0.0956, "step": 6983 }, { "epoch": 0.2944536964816493, "grad_norm": 0.7300885915756226, "learning_rate": 4.906548622821811e-05, "loss": 0.2687, "step": 6984 }, { "epoch": 0.2944958576638489, "grad_norm": 0.7063689827919006, "learning_rate": 4.9072512647554805e-05, "loss": 0.1415, "step": 6985 }, { "epoch": 0.29453801884604847, "grad_norm": 0.4058975279331207, "learning_rate": 4.9079539066891517e-05, "loss": 0.1226, "step": 6986 }, { "epoch": 0.294580180028248, "grad_norm": 0.7435336709022522, "learning_rate": 4.908656548622822e-05, "loss": 0.1357, "step": 6987 }, { "epoch": 0.29462234121044756, "grad_norm": 1.0952703952789307, "learning_rate": 4.9093591905564925e-05, "loss": 0.5343, "step": 6988 }, { "epoch": 0.2946645023926471, "grad_norm": 1.6807769536972046, "learning_rate": 4.910061832490163e-05, "loss": 0.3289, "step": 6989 }, { "epoch": 0.29470666357484665, "grad_norm": 1.0455206632614136, "learning_rate": 4.910764474423834e-05, "loss": 0.51, "step": 6990 }, { "epoch": 0.29474882475704617, "grad_norm": 1.354097843170166, "learning_rate": 4.9114671163575045e-05, "loss": 0.7404, "step": 6991 }, { "epoch": 0.29479098593924574, "grad_norm": 0.6981687545776367, "learning_rate": 4.912169758291175e-05, "loss": 0.3443, "step": 6992 }, { "epoch": 0.29483314712144526, "grad_norm": 1.0650498867034912, "learning_rate": 4.9128724002248454e-05, "loss": 0.3348, "step": 6993 }, { "epoch": 0.29487530830364483, "grad_norm": 0.904941201210022, "learning_rate": 4.9135750421585165e-05, "loss": 0.395, "step": 6994 }, { "epoch": 0.2949174694858444, "grad_norm": 1.3440133333206177, "learning_rate": 4.914277684092187e-05, "loss": 1.0543, "step": 6995 }, { "epoch": 0.2949596306680439, "grad_norm": 0.7502921223640442, "learning_rate": 4.9149803260258574e-05, "loss": 0.1477, "step": 6996 }, { "epoch": 0.2950017918502435, "grad_norm": 0.8190930485725403, "learning_rate": 4.915682967959528e-05, "loss": 0.3369, "step": 6997 }, { "epoch": 0.295043953032443, "grad_norm": 0.42456403374671936, "learning_rate": 4.916385609893199e-05, "loss": 0.2132, "step": 6998 }, { "epoch": 0.2950861142146426, "grad_norm": 0.4542069137096405, "learning_rate": 4.917088251826869e-05, "loss": 0.2295, "step": 6999 }, { "epoch": 0.2951282753968421, "grad_norm": 0.40750986337661743, "learning_rate": 4.91779089376054e-05, "loss": 0.1356, "step": 7000 }, { "epoch": 0.2951704365790417, "grad_norm": 1.2558867931365967, "learning_rate": 4.91849353569421e-05, "loss": 0.9537, "step": 7001 }, { "epoch": 0.2952125977612412, "grad_norm": 0.30823537707328796, "learning_rate": 4.9191961776278814e-05, "loss": 0.0837, "step": 7002 }, { "epoch": 0.2952547589434408, "grad_norm": 0.5343790054321289, "learning_rate": 4.919898819561551e-05, "loss": 0.2922, "step": 7003 }, { "epoch": 0.29529692012564035, "grad_norm": 0.6088089942932129, "learning_rate": 4.920601461495222e-05, "loss": 0.2369, "step": 7004 }, { "epoch": 0.29533908130783987, "grad_norm": 0.644980788230896, "learning_rate": 4.9213041034288934e-05, "loss": 0.3452, "step": 7005 }, { "epoch": 0.29538124249003944, "grad_norm": 0.5204461216926575, "learning_rate": 4.922006745362563e-05, "loss": 0.3279, "step": 7006 }, { "epoch": 0.29542340367223896, "grad_norm": 0.38679856061935425, "learning_rate": 4.922709387296234e-05, "loss": 0.1999, "step": 7007 }, { "epoch": 0.29546556485443853, "grad_norm": 0.4803408086299896, "learning_rate": 4.923412029229905e-05, "loss": 0.1821, "step": 7008 }, { "epoch": 0.29550772603663805, "grad_norm": 1.0560104846954346, "learning_rate": 4.924114671163576e-05, "loss": 0.6372, "step": 7009 }, { "epoch": 0.2955498872188376, "grad_norm": 0.5563862323760986, "learning_rate": 4.9248173130972456e-05, "loss": 0.1947, "step": 7010 }, { "epoch": 0.29559204840103714, "grad_norm": 0.5171571969985962, "learning_rate": 4.925519955030917e-05, "loss": 0.0985, "step": 7011 }, { "epoch": 0.2956342095832367, "grad_norm": 0.54107266664505, "learning_rate": 4.926222596964587e-05, "loss": 0.3139, "step": 7012 }, { "epoch": 0.29567637076543624, "grad_norm": 1.0132254362106323, "learning_rate": 4.9269252388982576e-05, "loss": 0.1273, "step": 7013 }, { "epoch": 0.2957185319476358, "grad_norm": 1.1015262603759766, "learning_rate": 4.927627880831928e-05, "loss": 1.0444, "step": 7014 }, { "epoch": 0.2957606931298354, "grad_norm": 0.7588135600090027, "learning_rate": 4.928330522765599e-05, "loss": 0.3732, "step": 7015 }, { "epoch": 0.2958028543120349, "grad_norm": 0.9044573903083801, "learning_rate": 4.9290331646992696e-05, "loss": 0.6612, "step": 7016 }, { "epoch": 0.2958450154942345, "grad_norm": 0.5762934684753418, "learning_rate": 4.92973580663294e-05, "loss": 0.1579, "step": 7017 }, { "epoch": 0.295887176676434, "grad_norm": 0.48373618721961975, "learning_rate": 4.9304384485666105e-05, "loss": 0.1747, "step": 7018 }, { "epoch": 0.29592933785863357, "grad_norm": 0.4218375086784363, "learning_rate": 4.9311410905002816e-05, "loss": 0.1251, "step": 7019 }, { "epoch": 0.2959714990408331, "grad_norm": 0.7901890873908997, "learning_rate": 4.931843732433952e-05, "loss": 0.3031, "step": 7020 }, { "epoch": 0.29601366022303266, "grad_norm": 0.6691250205039978, "learning_rate": 4.9325463743676225e-05, "loss": 0.5996, "step": 7021 }, { "epoch": 0.2960558214052322, "grad_norm": 0.5170177221298218, "learning_rate": 4.933249016301293e-05, "loss": 0.2722, "step": 7022 }, { "epoch": 0.29609798258743175, "grad_norm": 0.3311590552330017, "learning_rate": 4.933951658234964e-05, "loss": 0.1115, "step": 7023 }, { "epoch": 0.2961401437696313, "grad_norm": 0.8134705424308777, "learning_rate": 4.934654300168634e-05, "loss": 0.3849, "step": 7024 }, { "epoch": 0.29618230495183084, "grad_norm": 0.5319298505783081, "learning_rate": 4.935356942102305e-05, "loss": 0.1139, "step": 7025 }, { "epoch": 0.2962244661340304, "grad_norm": 0.6199604868888855, "learning_rate": 4.936059584035975e-05, "loss": 0.1694, "step": 7026 }, { "epoch": 0.29626662731622994, "grad_norm": 4.149291038513184, "learning_rate": 4.936762225969646e-05, "loss": 1.0162, "step": 7027 }, { "epoch": 0.2963087884984295, "grad_norm": 0.9637707471847534, "learning_rate": 4.937464867903317e-05, "loss": 0.3487, "step": 7028 }, { "epoch": 0.29635094968062903, "grad_norm": 2.974580764770508, "learning_rate": 4.938167509836987e-05, "loss": 0.5553, "step": 7029 }, { "epoch": 0.2963931108628286, "grad_norm": 1.7859163284301758, "learning_rate": 4.9388701517706584e-05, "loss": 0.4327, "step": 7030 }, { "epoch": 0.2964352720450281, "grad_norm": 0.5866581797599792, "learning_rate": 4.939572793704328e-05, "loss": 0.4388, "step": 7031 }, { "epoch": 0.2964774332272277, "grad_norm": 0.6078464388847351, "learning_rate": 4.940275435637999e-05, "loss": 0.7946, "step": 7032 }, { "epoch": 0.29651959440942727, "grad_norm": 2.66223406791687, "learning_rate": 4.94097807757167e-05, "loss": 0.7319, "step": 7033 }, { "epoch": 0.2965617555916268, "grad_norm": 0.9528329372406006, "learning_rate": 4.94168071950534e-05, "loss": 1.0979, "step": 7034 }, { "epoch": 0.29660391677382636, "grad_norm": 0.5432723164558411, "learning_rate": 4.9423833614390106e-05, "loss": 0.1548, "step": 7035 }, { "epoch": 0.2966460779560259, "grad_norm": 1.843217372894287, "learning_rate": 4.943086003372682e-05, "loss": 0.4879, "step": 7036 }, { "epoch": 0.29668823913822545, "grad_norm": 0.8414170742034912, "learning_rate": 4.943788645306352e-05, "loss": 0.3638, "step": 7037 }, { "epoch": 0.29673040032042497, "grad_norm": 0.9999006390571594, "learning_rate": 4.9444912872400226e-05, "loss": 0.3417, "step": 7038 }, { "epoch": 0.29677256150262454, "grad_norm": 0.523674726486206, "learning_rate": 4.945193929173693e-05, "loss": 0.1282, "step": 7039 }, { "epoch": 0.29681472268482406, "grad_norm": 0.7568109035491943, "learning_rate": 4.945896571107364e-05, "loss": 0.2414, "step": 7040 }, { "epoch": 0.29685688386702364, "grad_norm": 0.895452082157135, "learning_rate": 4.9465992130410346e-05, "loss": 0.2599, "step": 7041 }, { "epoch": 0.29689904504922315, "grad_norm": 0.7327632904052734, "learning_rate": 4.947301854974705e-05, "loss": 0.4183, "step": 7042 }, { "epoch": 0.2969412062314227, "grad_norm": 0.7991166710853577, "learning_rate": 4.9480044969083755e-05, "loss": 0.1413, "step": 7043 }, { "epoch": 0.2969833674136223, "grad_norm": 0.7051597833633423, "learning_rate": 4.9487071388420466e-05, "loss": 0.2403, "step": 7044 }, { "epoch": 0.2970255285958218, "grad_norm": 0.6673212051391602, "learning_rate": 4.9494097807757164e-05, "loss": 0.335, "step": 7045 }, { "epoch": 0.2970676897780214, "grad_norm": 0.6538505554199219, "learning_rate": 4.9501124227093875e-05, "loss": 0.1658, "step": 7046 }, { "epoch": 0.2971098509602209, "grad_norm": 0.7170931100845337, "learning_rate": 4.950815064643058e-05, "loss": 0.3417, "step": 7047 }, { "epoch": 0.2971520121424205, "grad_norm": 1.9038190841674805, "learning_rate": 4.951517706576729e-05, "loss": 0.6215, "step": 7048 }, { "epoch": 0.29719417332462, "grad_norm": 0.6016217470169067, "learning_rate": 4.952220348510399e-05, "loss": 0.2522, "step": 7049 }, { "epoch": 0.2972363345068196, "grad_norm": 0.48973628878593445, "learning_rate": 4.95292299044407e-05, "loss": 0.1066, "step": 7050 }, { "epoch": 0.2972784956890191, "grad_norm": 1.1950396299362183, "learning_rate": 4.953625632377741e-05, "loss": 0.8915, "step": 7051 }, { "epoch": 0.29732065687121867, "grad_norm": 0.8276901245117188, "learning_rate": 4.954328274311411e-05, "loss": 0.6011, "step": 7052 }, { "epoch": 0.29736281805341824, "grad_norm": 0.6044728755950928, "learning_rate": 4.955030916245082e-05, "loss": 0.3251, "step": 7053 }, { "epoch": 0.29740497923561776, "grad_norm": 0.4626674950122833, "learning_rate": 4.9557335581787524e-05, "loss": 0.1216, "step": 7054 }, { "epoch": 0.29744714041781734, "grad_norm": 1.3941869735717773, "learning_rate": 4.956436200112423e-05, "loss": 0.905, "step": 7055 }, { "epoch": 0.29748930160001685, "grad_norm": 0.7412351369857788, "learning_rate": 4.957138842046093e-05, "loss": 0.2293, "step": 7056 }, { "epoch": 0.2975314627822164, "grad_norm": 0.7919116020202637, "learning_rate": 4.9578414839797644e-05, "loss": 0.2169, "step": 7057 }, { "epoch": 0.29757362396441595, "grad_norm": 0.6517002582550049, "learning_rate": 4.958544125913435e-05, "loss": 0.0957, "step": 7058 }, { "epoch": 0.2976157851466155, "grad_norm": 0.8154348731040955, "learning_rate": 4.959246767847105e-05, "loss": 0.7201, "step": 7059 }, { "epoch": 0.29765794632881504, "grad_norm": 3.0242788791656494, "learning_rate": 4.959949409780776e-05, "loss": 0.8352, "step": 7060 }, { "epoch": 0.2977001075110146, "grad_norm": 0.7026412487030029, "learning_rate": 4.960652051714447e-05, "loss": 0.1172, "step": 7061 }, { "epoch": 0.2977422686932142, "grad_norm": 0.567901074886322, "learning_rate": 4.961354693648117e-05, "loss": 0.1823, "step": 7062 }, { "epoch": 0.2977844298754137, "grad_norm": 0.4475407302379608, "learning_rate": 4.962057335581788e-05, "loss": 0.1291, "step": 7063 }, { "epoch": 0.2978265910576133, "grad_norm": 0.6995495557785034, "learning_rate": 4.962759977515458e-05, "loss": 0.2605, "step": 7064 }, { "epoch": 0.2978687522398128, "grad_norm": 0.6045640707015991, "learning_rate": 4.963462619449129e-05, "loss": 0.4941, "step": 7065 }, { "epoch": 0.29791091342201237, "grad_norm": 1.2016137838363647, "learning_rate": 4.9641652613828e-05, "loss": 0.4489, "step": 7066 }, { "epoch": 0.2979530746042119, "grad_norm": 0.6011641621589661, "learning_rate": 4.96486790331647e-05, "loss": 0.2885, "step": 7067 }, { "epoch": 0.29799523578641146, "grad_norm": 0.5825918912887573, "learning_rate": 4.9655705452501406e-05, "loss": 0.2469, "step": 7068 }, { "epoch": 0.298037396968611, "grad_norm": 1.2532228231430054, "learning_rate": 4.966273187183812e-05, "loss": 1.1183, "step": 7069 }, { "epoch": 0.29807955815081055, "grad_norm": 0.714523434638977, "learning_rate": 4.9669758291174815e-05, "loss": 0.3068, "step": 7070 }, { "epoch": 0.29812171933301007, "grad_norm": 0.38737380504608154, "learning_rate": 4.9676784710511526e-05, "loss": 0.2131, "step": 7071 }, { "epoch": 0.29816388051520964, "grad_norm": 0.7063922882080078, "learning_rate": 4.968381112984824e-05, "loss": 0.2045, "step": 7072 }, { "epoch": 0.2982060416974092, "grad_norm": 0.4872715175151825, "learning_rate": 4.9690837549184934e-05, "loss": 0.1568, "step": 7073 }, { "epoch": 0.29824820287960874, "grad_norm": 0.4019731879234314, "learning_rate": 4.9697863968521646e-05, "loss": 0.1269, "step": 7074 }, { "epoch": 0.2982903640618083, "grad_norm": 1.850131869316101, "learning_rate": 4.970489038785835e-05, "loss": 1.0742, "step": 7075 }, { "epoch": 0.29833252524400783, "grad_norm": 0.852631688117981, "learning_rate": 4.971191680719506e-05, "loss": 0.5442, "step": 7076 }, { "epoch": 0.2983746864262074, "grad_norm": 0.847002387046814, "learning_rate": 4.971894322653176e-05, "loss": 0.9743, "step": 7077 }, { "epoch": 0.2984168476084069, "grad_norm": 1.1968485116958618, "learning_rate": 4.972596964586847e-05, "loss": 1.2134, "step": 7078 }, { "epoch": 0.2984590087906065, "grad_norm": 6.715907573699951, "learning_rate": 4.9732996065205174e-05, "loss": 0.5146, "step": 7079 }, { "epoch": 0.298501169972806, "grad_norm": 1.087207317352295, "learning_rate": 4.974002248454188e-05, "loss": 0.7338, "step": 7080 }, { "epoch": 0.2985433311550056, "grad_norm": 0.6360328197479248, "learning_rate": 4.974704890387858e-05, "loss": 0.4776, "step": 7081 }, { "epoch": 0.29858549233720516, "grad_norm": 1.078463077545166, "learning_rate": 4.9754075323215294e-05, "loss": 0.0927, "step": 7082 }, { "epoch": 0.2986276535194047, "grad_norm": 3.818680763244629, "learning_rate": 4.9761101742552e-05, "loss": 0.1355, "step": 7083 }, { "epoch": 0.29866981470160425, "grad_norm": 0.9080875515937805, "learning_rate": 4.97681281618887e-05, "loss": 0.1165, "step": 7084 }, { "epoch": 0.29871197588380377, "grad_norm": 0.6846916675567627, "learning_rate": 4.977515458122541e-05, "loss": 0.6043, "step": 7085 }, { "epoch": 0.29875413706600334, "grad_norm": 0.44891712069511414, "learning_rate": 4.978218100056212e-05, "loss": 0.1378, "step": 7086 }, { "epoch": 0.29879629824820286, "grad_norm": 0.5640862584114075, "learning_rate": 4.978920741989882e-05, "loss": 0.5669, "step": 7087 }, { "epoch": 0.29883845943040244, "grad_norm": 1.1055351495742798, "learning_rate": 4.979623383923553e-05, "loss": 0.2254, "step": 7088 }, { "epoch": 0.29888062061260195, "grad_norm": 0.42678186297416687, "learning_rate": 4.980326025857223e-05, "loss": 0.1463, "step": 7089 }, { "epoch": 0.29892278179480153, "grad_norm": 0.4490882456302643, "learning_rate": 4.981028667790894e-05, "loss": 0.137, "step": 7090 }, { "epoch": 0.2989649429770011, "grad_norm": 1.0444844961166382, "learning_rate": 4.981731309724564e-05, "loss": 0.579, "step": 7091 }, { "epoch": 0.2990071041592006, "grad_norm": 0.6361350417137146, "learning_rate": 4.982433951658235e-05, "loss": 0.3092, "step": 7092 }, { "epoch": 0.2990492653414002, "grad_norm": 0.7934025526046753, "learning_rate": 4.9831365935919056e-05, "loss": 0.4335, "step": 7093 }, { "epoch": 0.2990914265235997, "grad_norm": 0.749711811542511, "learning_rate": 4.983839235525577e-05, "loss": 0.1879, "step": 7094 }, { "epoch": 0.2991335877057993, "grad_norm": 0.7145419120788574, "learning_rate": 4.984541877459247e-05, "loss": 0.5712, "step": 7095 }, { "epoch": 0.2991757488879988, "grad_norm": 0.40045928955078125, "learning_rate": 4.9852445193929176e-05, "loss": 0.1249, "step": 7096 }, { "epoch": 0.2992179100701984, "grad_norm": 0.663623034954071, "learning_rate": 4.985947161326589e-05, "loss": 0.672, "step": 7097 }, { "epoch": 0.2992600712523979, "grad_norm": 0.7442349195480347, "learning_rate": 4.9866498032602585e-05, "loss": 0.2516, "step": 7098 }, { "epoch": 0.29930223243459747, "grad_norm": 0.4689808189868927, "learning_rate": 4.9873524451939296e-05, "loss": 0.1263, "step": 7099 }, { "epoch": 0.299344393616797, "grad_norm": 1.0267888307571411, "learning_rate": 4.9880550871276e-05, "loss": 0.873, "step": 7100 }, { "epoch": 0.29938655479899656, "grad_norm": 0.4458671510219574, "learning_rate": 4.9887577290612705e-05, "loss": 0.1884, "step": 7101 }, { "epoch": 0.29942871598119614, "grad_norm": 0.598825216293335, "learning_rate": 4.989460370994941e-05, "loss": 0.2486, "step": 7102 }, { "epoch": 0.29947087716339565, "grad_norm": 0.7021783590316772, "learning_rate": 4.990163012928612e-05, "loss": 0.7109, "step": 7103 }, { "epoch": 0.29951303834559523, "grad_norm": 0.6083566546440125, "learning_rate": 4.9908656548622825e-05, "loss": 0.2126, "step": 7104 }, { "epoch": 0.29955519952779475, "grad_norm": 0.7011376619338989, "learning_rate": 4.991568296795953e-05, "loss": 0.5834, "step": 7105 }, { "epoch": 0.2995973607099943, "grad_norm": 0.6472908854484558, "learning_rate": 4.9922709387296234e-05, "loss": 0.2047, "step": 7106 }, { "epoch": 0.29963952189219384, "grad_norm": 0.946202278137207, "learning_rate": 4.9929735806632945e-05, "loss": 0.144, "step": 7107 }, { "epoch": 0.2996816830743934, "grad_norm": 0.7829656600952148, "learning_rate": 4.993676222596965e-05, "loss": 1.1537, "step": 7108 }, { "epoch": 0.29972384425659293, "grad_norm": 0.6109324097633362, "learning_rate": 4.9943788645306354e-05, "loss": 0.6041, "step": 7109 }, { "epoch": 0.2997660054387925, "grad_norm": 0.35609176754951477, "learning_rate": 4.995081506464306e-05, "loss": 0.1283, "step": 7110 }, { "epoch": 0.2998081666209921, "grad_norm": 0.3974577486515045, "learning_rate": 4.995784148397977e-05, "loss": 0.1248, "step": 7111 }, { "epoch": 0.2998503278031916, "grad_norm": 0.9532443284988403, "learning_rate": 4.996486790331647e-05, "loss": 0.7945, "step": 7112 }, { "epoch": 0.29989248898539117, "grad_norm": 0.5616841912269592, "learning_rate": 4.997189432265318e-05, "loss": 0.6552, "step": 7113 }, { "epoch": 0.2999346501675907, "grad_norm": 1.6864776611328125, "learning_rate": 4.997892074198988e-05, "loss": 0.5717, "step": 7114 }, { "epoch": 0.29997681134979026, "grad_norm": 0.4930451512336731, "learning_rate": 4.9985947161326594e-05, "loss": 0.1231, "step": 7115 }, { "epoch": 0.3000189725319898, "grad_norm": 0.6978693008422852, "learning_rate": 4.999297358066329e-05, "loss": 0.2422, "step": 7116 }, { "epoch": 0.30006113371418935, "grad_norm": 0.595024585723877, "learning_rate": 5e-05, "loss": 0.2011, "step": 7117 }, { "epoch": 0.3001032948963889, "grad_norm": 0.9909893870353699, "learning_rate": 4.999921921359193e-05, "loss": 0.346, "step": 7118 }, { "epoch": 0.30014545607858845, "grad_norm": 0.48317191004753113, "learning_rate": 4.9998438427183865e-05, "loss": 0.213, "step": 7119 }, { "epoch": 0.300187617260788, "grad_norm": 0.5578837394714355, "learning_rate": 4.999765764077579e-05, "loss": 0.5628, "step": 7120 }, { "epoch": 0.30022977844298754, "grad_norm": 0.5283677577972412, "learning_rate": 4.999687685436772e-05, "loss": 0.1143, "step": 7121 }, { "epoch": 0.3002719396251871, "grad_norm": 0.7304375171661377, "learning_rate": 4.999609606795965e-05, "loss": 0.4895, "step": 7122 }, { "epoch": 0.30031410080738663, "grad_norm": 0.6818676590919495, "learning_rate": 4.999531528155158e-05, "loss": 0.5414, "step": 7123 }, { "epoch": 0.3003562619895862, "grad_norm": 0.7232560515403748, "learning_rate": 4.9994534495143514e-05, "loss": 0.4734, "step": 7124 }, { "epoch": 0.3003984231717857, "grad_norm": 0.48044419288635254, "learning_rate": 4.999375370873544e-05, "loss": 0.5073, "step": 7125 }, { "epoch": 0.3004405843539853, "grad_norm": 0.3891483247280121, "learning_rate": 4.999297292232737e-05, "loss": 0.0722, "step": 7126 }, { "epoch": 0.3004827455361848, "grad_norm": 0.5602971315383911, "learning_rate": 4.99921921359193e-05, "loss": 0.2502, "step": 7127 }, { "epoch": 0.3005249067183844, "grad_norm": 0.925534725189209, "learning_rate": 4.999141134951123e-05, "loss": 0.7821, "step": 7128 }, { "epoch": 0.3005670679005839, "grad_norm": 0.4991223216056824, "learning_rate": 4.999063056310316e-05, "loss": 0.1517, "step": 7129 }, { "epoch": 0.3006092290827835, "grad_norm": 0.3779987096786499, "learning_rate": 4.998984977669509e-05, "loss": 0.0728, "step": 7130 }, { "epoch": 0.30065139026498305, "grad_norm": 0.2769637107849121, "learning_rate": 4.998906899028702e-05, "loss": 0.1028, "step": 7131 }, { "epoch": 0.30069355144718257, "grad_norm": 0.7140936255455017, "learning_rate": 4.9988288203878944e-05, "loss": 0.1896, "step": 7132 }, { "epoch": 0.30073571262938215, "grad_norm": 0.42897549271583557, "learning_rate": 4.998750741747088e-05, "loss": 0.1056, "step": 7133 }, { "epoch": 0.30077787381158166, "grad_norm": 0.5621034502983093, "learning_rate": 4.998672663106281e-05, "loss": 0.6157, "step": 7134 }, { "epoch": 0.30082003499378124, "grad_norm": 0.5831985473632812, "learning_rate": 4.998594584465474e-05, "loss": 0.384, "step": 7135 }, { "epoch": 0.30086219617598076, "grad_norm": 1.5750963687896729, "learning_rate": 4.998516505824667e-05, "loss": 0.6768, "step": 7136 }, { "epoch": 0.30090435735818033, "grad_norm": 1.4145740270614624, "learning_rate": 4.9984384271838594e-05, "loss": 0.1436, "step": 7137 }, { "epoch": 0.30094651854037985, "grad_norm": 0.6707241535186768, "learning_rate": 4.998360348543053e-05, "loss": 0.6163, "step": 7138 }, { "epoch": 0.3009886797225794, "grad_norm": 0.5787558555603027, "learning_rate": 4.9982822699022456e-05, "loss": 0.2826, "step": 7139 }, { "epoch": 0.301030840904779, "grad_norm": 2.8204803466796875, "learning_rate": 4.998204191261439e-05, "loss": 0.8921, "step": 7140 }, { "epoch": 0.3010730020869785, "grad_norm": 0.48638537526130676, "learning_rate": 4.998126112620632e-05, "loss": 0.189, "step": 7141 }, { "epoch": 0.3011151632691781, "grad_norm": 1.0318716764450073, "learning_rate": 4.998048033979825e-05, "loss": 1.1949, "step": 7142 }, { "epoch": 0.3011573244513776, "grad_norm": 3.951474189758301, "learning_rate": 4.9979699553390174e-05, "loss": 0.213, "step": 7143 }, { "epoch": 0.3011994856335772, "grad_norm": 1.2217299938201904, "learning_rate": 4.9978918766982106e-05, "loss": 0.3292, "step": 7144 }, { "epoch": 0.3012416468157767, "grad_norm": 0.6597879528999329, "learning_rate": 4.997813798057404e-05, "loss": 0.6567, "step": 7145 }, { "epoch": 0.30128380799797627, "grad_norm": 0.5825934410095215, "learning_rate": 4.997735719416597e-05, "loss": 0.1928, "step": 7146 }, { "epoch": 0.3013259691801758, "grad_norm": 0.35914093255996704, "learning_rate": 4.99765764077579e-05, "loss": 0.1888, "step": 7147 }, { "epoch": 0.30136813036237536, "grad_norm": 25425260.0, "learning_rate": 4.9975795621349824e-05, "loss": 0.2539, "step": 7148 }, { "epoch": 0.30141029154457494, "grad_norm": 0.410613477230072, "learning_rate": 4.9975014834941755e-05, "loss": 0.1208, "step": 7149 }, { "epoch": 0.30145245272677446, "grad_norm": 0.6474750638008118, "learning_rate": 4.9974234048533687e-05, "loss": 0.4763, "step": 7150 }, { "epoch": 0.30149461390897403, "grad_norm": 0.6835266947746277, "learning_rate": 4.997345326212561e-05, "loss": 0.668, "step": 7151 }, { "epoch": 0.30153677509117355, "grad_norm": 1.0758243799209595, "learning_rate": 4.997267247571755e-05, "loss": 0.9434, "step": 7152 }, { "epoch": 0.3015789362733731, "grad_norm": 0.7410209774971008, "learning_rate": 4.9971891689309474e-05, "loss": 0.2006, "step": 7153 }, { "epoch": 0.30162109745557264, "grad_norm": 0.6371145248413086, "learning_rate": 4.9971110902901405e-05, "loss": 0.1289, "step": 7154 }, { "epoch": 0.3016632586377722, "grad_norm": 0.4797098636627197, "learning_rate": 4.9970330116493336e-05, "loss": 0.2082, "step": 7155 }, { "epoch": 0.30170541981997173, "grad_norm": 1.1009529829025269, "learning_rate": 4.996954933008526e-05, "loss": 0.8556, "step": 7156 }, { "epoch": 0.3017475810021713, "grad_norm": 0.559968888759613, "learning_rate": 4.996876854367719e-05, "loss": 0.6108, "step": 7157 }, { "epoch": 0.3017897421843708, "grad_norm": 0.6445330381393433, "learning_rate": 4.996798775726912e-05, "loss": 0.2255, "step": 7158 }, { "epoch": 0.3018319033665704, "grad_norm": 0.7048972845077515, "learning_rate": 4.9967206970861054e-05, "loss": 0.5404, "step": 7159 }, { "epoch": 0.30187406454876997, "grad_norm": 0.6499255895614624, "learning_rate": 4.9966426184452986e-05, "loss": 0.1543, "step": 7160 }, { "epoch": 0.3019162257309695, "grad_norm": 0.5610412359237671, "learning_rate": 4.996564539804492e-05, "loss": 0.2084, "step": 7161 }, { "epoch": 0.30195838691316906, "grad_norm": 0.8919574618339539, "learning_rate": 4.996486461163684e-05, "loss": 0.4325, "step": 7162 }, { "epoch": 0.3020005480953686, "grad_norm": 0.4960039556026459, "learning_rate": 4.996408382522877e-05, "loss": 0.1533, "step": 7163 }, { "epoch": 0.30204270927756816, "grad_norm": 0.7217344045639038, "learning_rate": 4.9963303038820704e-05, "loss": 0.7361, "step": 7164 }, { "epoch": 0.3020848704597677, "grad_norm": 0.4481985569000244, "learning_rate": 4.996252225241263e-05, "loss": 0.0983, "step": 7165 }, { "epoch": 0.30212703164196725, "grad_norm": 0.8512209057807922, "learning_rate": 4.9961741466004566e-05, "loss": 0.9537, "step": 7166 }, { "epoch": 0.30216919282416677, "grad_norm": 0.48084700107574463, "learning_rate": 4.996096067959649e-05, "loss": 0.135, "step": 7167 }, { "epoch": 0.30221135400636634, "grad_norm": 0.7259600162506104, "learning_rate": 4.996017989318842e-05, "loss": 0.3385, "step": 7168 }, { "epoch": 0.3022535151885659, "grad_norm": 2.925039529800415, "learning_rate": 4.995939910678035e-05, "loss": 0.7718, "step": 7169 }, { "epoch": 0.30229567637076543, "grad_norm": 0.37920111417770386, "learning_rate": 4.995861832037228e-05, "loss": 0.1756, "step": 7170 }, { "epoch": 0.302337837552965, "grad_norm": 0.5142984390258789, "learning_rate": 4.9957837533964216e-05, "loss": 0.1212, "step": 7171 }, { "epoch": 0.3023799987351645, "grad_norm": 0.2793532609939575, "learning_rate": 4.995705674755614e-05, "loss": 0.1181, "step": 7172 }, { "epoch": 0.3024221599173641, "grad_norm": 0.35838234424591064, "learning_rate": 4.995627596114807e-05, "loss": 0.0726, "step": 7173 }, { "epoch": 0.3024643210995636, "grad_norm": 0.9282586574554443, "learning_rate": 4.995549517474e-05, "loss": 1.1896, "step": 7174 }, { "epoch": 0.3025064822817632, "grad_norm": 0.32163745164871216, "learning_rate": 4.995471438833193e-05, "loss": 0.1039, "step": 7175 }, { "epoch": 0.3025486434639627, "grad_norm": 0.42116406559944153, "learning_rate": 4.995393360192386e-05, "loss": 0.0981, "step": 7176 }, { "epoch": 0.3025908046461623, "grad_norm": 2.7661402225494385, "learning_rate": 4.995315281551579e-05, "loss": 0.9314, "step": 7177 }, { "epoch": 0.3026329658283618, "grad_norm": 1.4529147148132324, "learning_rate": 4.995237202910772e-05, "loss": 0.8898, "step": 7178 }, { "epoch": 0.3026751270105614, "grad_norm": 0.5707547664642334, "learning_rate": 4.9951591242699646e-05, "loss": 0.1724, "step": 7179 }, { "epoch": 0.30271728819276095, "grad_norm": 2.5770397186279297, "learning_rate": 4.995081045629158e-05, "loss": 0.4454, "step": 7180 }, { "epoch": 0.30275944937496047, "grad_norm": 0.5251772403717041, "learning_rate": 4.995002966988351e-05, "loss": 0.2756, "step": 7181 }, { "epoch": 0.30280161055716004, "grad_norm": 0.6179969906806946, "learning_rate": 4.994924888347544e-05, "loss": 0.1047, "step": 7182 }, { "epoch": 0.30284377173935956, "grad_norm": 0.7943978309631348, "learning_rate": 4.994846809706737e-05, "loss": 0.3142, "step": 7183 }, { "epoch": 0.30288593292155913, "grad_norm": 0.9743282794952393, "learning_rate": 4.9947687310659295e-05, "loss": 0.2571, "step": 7184 }, { "epoch": 0.30292809410375865, "grad_norm": 1.2241560220718384, "learning_rate": 4.994690652425123e-05, "loss": 0.1858, "step": 7185 }, { "epoch": 0.3029702552859582, "grad_norm": 0.7352146506309509, "learning_rate": 4.994612573784316e-05, "loss": 0.429, "step": 7186 }, { "epoch": 0.30301241646815774, "grad_norm": 0.9386407732963562, "learning_rate": 4.994534495143509e-05, "loss": 1.108, "step": 7187 }, { "epoch": 0.3030545776503573, "grad_norm": 1.1788069009780884, "learning_rate": 4.994456416502702e-05, "loss": 0.3188, "step": 7188 }, { "epoch": 0.3030967388325569, "grad_norm": 1.501635193824768, "learning_rate": 4.9943783378618945e-05, "loss": 0.8544, "step": 7189 }, { "epoch": 0.3031389000147564, "grad_norm": 0.3855392038822174, "learning_rate": 4.9943002592210876e-05, "loss": 0.091, "step": 7190 }, { "epoch": 0.303181061196956, "grad_norm": 0.4208660125732422, "learning_rate": 4.994222180580281e-05, "loss": 0.0848, "step": 7191 }, { "epoch": 0.3032232223791555, "grad_norm": 0.5853855013847351, "learning_rate": 4.994144101939474e-05, "loss": 0.1678, "step": 7192 }, { "epoch": 0.3032653835613551, "grad_norm": 2.050452709197998, "learning_rate": 4.994066023298667e-05, "loss": 0.4514, "step": 7193 }, { "epoch": 0.3033075447435546, "grad_norm": 0.6621651649475098, "learning_rate": 4.9939879446578594e-05, "loss": 0.5707, "step": 7194 }, { "epoch": 0.30334970592575417, "grad_norm": 1.031227707862854, "learning_rate": 4.9939098660170525e-05, "loss": 1.0075, "step": 7195 }, { "epoch": 0.3033918671079537, "grad_norm": 0.4034630358219147, "learning_rate": 4.993831787376246e-05, "loss": 0.1109, "step": 7196 }, { "epoch": 0.30343402829015326, "grad_norm": 0.8557604551315308, "learning_rate": 4.993753708735439e-05, "loss": 0.3185, "step": 7197 }, { "epoch": 0.30347618947235283, "grad_norm": 1.2060203552246094, "learning_rate": 4.993675630094631e-05, "loss": 0.5277, "step": 7198 }, { "epoch": 0.30351835065455235, "grad_norm": 0.5330325961112976, "learning_rate": 4.9935975514538244e-05, "loss": 0.2178, "step": 7199 }, { "epoch": 0.3035605118367519, "grad_norm": 5.763969898223877, "learning_rate": 4.9935194728130175e-05, "loss": 0.562, "step": 7200 }, { "epoch": 0.30360267301895144, "grad_norm": 0.650043249130249, "learning_rate": 4.99344139417221e-05, "loss": 0.486, "step": 7201 }, { "epoch": 0.303644834201151, "grad_norm": 0.4260926842689514, "learning_rate": 4.993363315531404e-05, "loss": 0.0998, "step": 7202 }, { "epoch": 0.30368699538335053, "grad_norm": 2.2721073627471924, "learning_rate": 4.993285236890596e-05, "loss": 0.6341, "step": 7203 }, { "epoch": 0.3037291565655501, "grad_norm": 1.1193804740905762, "learning_rate": 4.99320715824979e-05, "loss": 0.9869, "step": 7204 }, { "epoch": 0.3037713177477496, "grad_norm": 0.5042324662208557, "learning_rate": 4.9931290796089825e-05, "loss": 0.1165, "step": 7205 }, { "epoch": 0.3038134789299492, "grad_norm": 0.6002709269523621, "learning_rate": 4.9930510009681756e-05, "loss": 0.5149, "step": 7206 }, { "epoch": 0.3038556401121487, "grad_norm": 0.9536521434783936, "learning_rate": 4.992972922327369e-05, "loss": 0.3416, "step": 7207 }, { "epoch": 0.3038978012943483, "grad_norm": 0.7302177548408508, "learning_rate": 4.992894843686561e-05, "loss": 0.1782, "step": 7208 }, { "epoch": 0.30393996247654786, "grad_norm": 0.27552032470703125, "learning_rate": 4.992816765045754e-05, "loss": 0.0876, "step": 7209 }, { "epoch": 0.3039821236587474, "grad_norm": 0.5309656262397766, "learning_rate": 4.9927386864049474e-05, "loss": 0.1777, "step": 7210 }, { "epoch": 0.30402428484094696, "grad_norm": 0.288835346698761, "learning_rate": 4.9926606077641405e-05, "loss": 0.0702, "step": 7211 }, { "epoch": 0.3040664460231465, "grad_norm": 2.364238977432251, "learning_rate": 4.992582529123333e-05, "loss": 0.5889, "step": 7212 }, { "epoch": 0.30410860720534605, "grad_norm": 0.47950732707977295, "learning_rate": 4.992504450482526e-05, "loss": 0.3224, "step": 7213 }, { "epoch": 0.30415076838754557, "grad_norm": 1.562487244606018, "learning_rate": 4.992426371841719e-05, "loss": 1.0614, "step": 7214 }, { "epoch": 0.30419292956974514, "grad_norm": 0.36740779876708984, "learning_rate": 4.9923482932009124e-05, "loss": 0.1292, "step": 7215 }, { "epoch": 0.30423509075194466, "grad_norm": 1.5671666860580444, "learning_rate": 4.9922702145601055e-05, "loss": 0.1798, "step": 7216 }, { "epoch": 0.30427725193414423, "grad_norm": 0.44421276450157166, "learning_rate": 4.992192135919298e-05, "loss": 0.1818, "step": 7217 }, { "epoch": 0.3043194131163438, "grad_norm": 1.2248731851577759, "learning_rate": 4.992114057278491e-05, "loss": 1.1838, "step": 7218 }, { "epoch": 0.3043615742985433, "grad_norm": 1.0036873817443848, "learning_rate": 4.992035978637684e-05, "loss": 0.2164, "step": 7219 }, { "epoch": 0.3044037354807429, "grad_norm": 0.6183412671089172, "learning_rate": 4.9919578999968766e-05, "loss": 0.4329, "step": 7220 }, { "epoch": 0.3044458966629424, "grad_norm": 0.5637476444244385, "learning_rate": 4.9918798213560704e-05, "loss": 0.2713, "step": 7221 }, { "epoch": 0.304488057845142, "grad_norm": 0.5769606828689575, "learning_rate": 4.991801742715263e-05, "loss": 0.2994, "step": 7222 }, { "epoch": 0.3045302190273415, "grad_norm": 0.930321991443634, "learning_rate": 4.991723664074456e-05, "loss": 0.2496, "step": 7223 }, { "epoch": 0.3045723802095411, "grad_norm": 1.0097532272338867, "learning_rate": 4.991645585433649e-05, "loss": 0.5343, "step": 7224 }, { "epoch": 0.3046145413917406, "grad_norm": 0.918471097946167, "learning_rate": 4.9915675067928416e-05, "loss": 0.3539, "step": 7225 }, { "epoch": 0.3046567025739402, "grad_norm": 0.769752562046051, "learning_rate": 4.9914894281520354e-05, "loss": 0.1786, "step": 7226 }, { "epoch": 0.30469886375613975, "grad_norm": 1.476557970046997, "learning_rate": 4.991411349511228e-05, "loss": 0.4254, "step": 7227 }, { "epoch": 0.30474102493833927, "grad_norm": 1.0274972915649414, "learning_rate": 4.991333270870421e-05, "loss": 0.2766, "step": 7228 }, { "epoch": 0.30478318612053884, "grad_norm": 0.6899006962776184, "learning_rate": 4.991255192229614e-05, "loss": 0.143, "step": 7229 }, { "epoch": 0.30482534730273836, "grad_norm": 6.751684665679932, "learning_rate": 4.991177113588807e-05, "loss": 0.1323, "step": 7230 }, { "epoch": 0.30486750848493793, "grad_norm": 1.4292073249816895, "learning_rate": 4.991099034948e-05, "loss": 0.9856, "step": 7231 }, { "epoch": 0.30490966966713745, "grad_norm": 1.9246559143066406, "learning_rate": 4.991020956307193e-05, "loss": 0.706, "step": 7232 }, { "epoch": 0.304951830849337, "grad_norm": 0.5590506196022034, "learning_rate": 4.990942877666386e-05, "loss": 0.1207, "step": 7233 }, { "epoch": 0.30499399203153654, "grad_norm": 0.643649160861969, "learning_rate": 4.9908647990255784e-05, "loss": 0.6589, "step": 7234 }, { "epoch": 0.3050361532137361, "grad_norm": 0.5814735293388367, "learning_rate": 4.990786720384772e-05, "loss": 0.1657, "step": 7235 }, { "epoch": 0.30507831439593563, "grad_norm": 0.601168155670166, "learning_rate": 4.9907086417439646e-05, "loss": 0.5882, "step": 7236 }, { "epoch": 0.3051204755781352, "grad_norm": 0.6302869319915771, "learning_rate": 4.990630563103158e-05, "loss": 0.2412, "step": 7237 }, { "epoch": 0.3051626367603348, "grad_norm": 1.339994192123413, "learning_rate": 4.990552484462351e-05, "loss": 0.1343, "step": 7238 }, { "epoch": 0.3052047979425343, "grad_norm": 1.4304801225662231, "learning_rate": 4.990474405821543e-05, "loss": 0.496, "step": 7239 }, { "epoch": 0.3052469591247339, "grad_norm": 0.9361984729766846, "learning_rate": 4.990396327180737e-05, "loss": 0.1879, "step": 7240 }, { "epoch": 0.3052891203069334, "grad_norm": 1.3705884218215942, "learning_rate": 4.9903182485399296e-05, "loss": 1.2712, "step": 7241 }, { "epoch": 0.30533128148913297, "grad_norm": 0.6274212598800659, "learning_rate": 4.990240169899123e-05, "loss": 0.2156, "step": 7242 }, { "epoch": 0.3053734426713325, "grad_norm": 0.8027297258377075, "learning_rate": 4.990162091258316e-05, "loss": 0.2943, "step": 7243 }, { "epoch": 0.30541560385353206, "grad_norm": 0.39214491844177246, "learning_rate": 4.990084012617508e-05, "loss": 0.0826, "step": 7244 }, { "epoch": 0.3054577650357316, "grad_norm": 0.9526845216751099, "learning_rate": 4.9900059339767014e-05, "loss": 0.575, "step": 7245 }, { "epoch": 0.30549992621793115, "grad_norm": 1.0063985586166382, "learning_rate": 4.9899278553358945e-05, "loss": 0.1699, "step": 7246 }, { "epoch": 0.3055420874001307, "grad_norm": 0.45174118876457214, "learning_rate": 4.9898497766950876e-05, "loss": 0.2918, "step": 7247 }, { "epoch": 0.30558424858233024, "grad_norm": 0.6932430863380432, "learning_rate": 4.98977169805428e-05, "loss": 0.4897, "step": 7248 }, { "epoch": 0.3056264097645298, "grad_norm": 0.667625904083252, "learning_rate": 4.989693619413474e-05, "loss": 0.1525, "step": 7249 }, { "epoch": 0.30566857094672933, "grad_norm": 1.315901756286621, "learning_rate": 4.9896155407726663e-05, "loss": 0.4251, "step": 7250 }, { "epoch": 0.3057107321289289, "grad_norm": 0.2803805470466614, "learning_rate": 4.9895374621318595e-05, "loss": 0.1356, "step": 7251 }, { "epoch": 0.3057528933111284, "grad_norm": 0.5937181711196899, "learning_rate": 4.9894593834910526e-05, "loss": 0.1806, "step": 7252 }, { "epoch": 0.305795054493328, "grad_norm": 0.36009249091148376, "learning_rate": 4.989381304850245e-05, "loss": 0.0947, "step": 7253 }, { "epoch": 0.3058372156755275, "grad_norm": 0.8955678343772888, "learning_rate": 4.989303226209439e-05, "loss": 0.3026, "step": 7254 }, { "epoch": 0.3058793768577271, "grad_norm": 1.7204563617706299, "learning_rate": 4.989225147568631e-05, "loss": 0.2817, "step": 7255 }, { "epoch": 0.30592153803992667, "grad_norm": 0.5730254054069519, "learning_rate": 4.9891470689278244e-05, "loss": 0.4459, "step": 7256 }, { "epoch": 0.3059636992221262, "grad_norm": 0.5742802023887634, "learning_rate": 4.9890689902870175e-05, "loss": 0.6715, "step": 7257 }, { "epoch": 0.30600586040432576, "grad_norm": 1.6286909580230713, "learning_rate": 4.98899091164621e-05, "loss": 0.5623, "step": 7258 }, { "epoch": 0.3060480215865253, "grad_norm": 0.6690100431442261, "learning_rate": 4.988912833005403e-05, "loss": 0.4819, "step": 7259 }, { "epoch": 0.30609018276872485, "grad_norm": 0.7763960361480713, "learning_rate": 4.988834754364596e-05, "loss": 0.3143, "step": 7260 }, { "epoch": 0.30613234395092437, "grad_norm": 0.5760533213615417, "learning_rate": 4.9887566757237894e-05, "loss": 0.1615, "step": 7261 }, { "epoch": 0.30617450513312394, "grad_norm": 0.45760560035705566, "learning_rate": 4.9886785970829825e-05, "loss": 0.1458, "step": 7262 }, { "epoch": 0.30621666631532346, "grad_norm": 1.0637624263763428, "learning_rate": 4.988600518442175e-05, "loss": 0.7968, "step": 7263 }, { "epoch": 0.30625882749752303, "grad_norm": 0.6101764440536499, "learning_rate": 4.988522439801368e-05, "loss": 0.3878, "step": 7264 }, { "epoch": 0.30630098867972255, "grad_norm": 0.40867847204208374, "learning_rate": 4.988444361160561e-05, "loss": 0.108, "step": 7265 }, { "epoch": 0.3063431498619221, "grad_norm": 0.6233656406402588, "learning_rate": 4.988366282519754e-05, "loss": 0.3015, "step": 7266 }, { "epoch": 0.3063853110441217, "grad_norm": 5.1032867431640625, "learning_rate": 4.988288203878947e-05, "loss": 0.2202, "step": 7267 }, { "epoch": 0.3064274722263212, "grad_norm": 0.41986027359962463, "learning_rate": 4.9882101252381406e-05, "loss": 0.0894, "step": 7268 }, { "epoch": 0.3064696334085208, "grad_norm": 0.6107840538024902, "learning_rate": 4.988132046597333e-05, "loss": 0.5368, "step": 7269 }, { "epoch": 0.3065117945907203, "grad_norm": 0.6182718276977539, "learning_rate": 4.9880539679565255e-05, "loss": 0.1755, "step": 7270 }, { "epoch": 0.3065539557729199, "grad_norm": 0.9174785017967224, "learning_rate": 4.987975889315719e-05, "loss": 0.6731, "step": 7271 }, { "epoch": 0.3065961169551194, "grad_norm": 0.38231056928634644, "learning_rate": 4.987897810674912e-05, "loss": 0.1102, "step": 7272 }, { "epoch": 0.306638278137319, "grad_norm": 2.9112584590911865, "learning_rate": 4.9878197320341055e-05, "loss": 0.2629, "step": 7273 }, { "epoch": 0.3066804393195185, "grad_norm": 1.158910870552063, "learning_rate": 4.987741653393298e-05, "loss": 1.1326, "step": 7274 }, { "epoch": 0.30672260050171807, "grad_norm": 1.130980372428894, "learning_rate": 4.987663574752491e-05, "loss": 0.369, "step": 7275 }, { "epoch": 0.30676476168391764, "grad_norm": 1.483099341392517, "learning_rate": 4.987585496111684e-05, "loss": 0.8882, "step": 7276 }, { "epoch": 0.30680692286611716, "grad_norm": 1.0907150506973267, "learning_rate": 4.987507417470877e-05, "loss": 1.2273, "step": 7277 }, { "epoch": 0.30684908404831673, "grad_norm": 0.5583686828613281, "learning_rate": 4.98742933883007e-05, "loss": 0.141, "step": 7278 }, { "epoch": 0.30689124523051625, "grad_norm": 0.5889118313789368, "learning_rate": 4.987351260189263e-05, "loss": 0.1139, "step": 7279 }, { "epoch": 0.3069334064127158, "grad_norm": 0.47138091921806335, "learning_rate": 4.987273181548456e-05, "loss": 0.2103, "step": 7280 }, { "epoch": 0.30697556759491534, "grad_norm": 0.7218319177627563, "learning_rate": 4.9871951029076485e-05, "loss": 0.1642, "step": 7281 }, { "epoch": 0.3070177287771149, "grad_norm": 0.9997347593307495, "learning_rate": 4.9871170242668416e-05, "loss": 0.3528, "step": 7282 }, { "epoch": 0.30705988995931444, "grad_norm": 1.1698449850082397, "learning_rate": 4.987038945626035e-05, "loss": 0.3783, "step": 7283 }, { "epoch": 0.307102051141514, "grad_norm": 0.5401870608329773, "learning_rate": 4.986960866985228e-05, "loss": 0.1466, "step": 7284 }, { "epoch": 0.3071442123237136, "grad_norm": 0.562680184841156, "learning_rate": 4.986882788344421e-05, "loss": 0.1852, "step": 7285 }, { "epoch": 0.3071863735059131, "grad_norm": 0.6719878315925598, "learning_rate": 4.9868047097036135e-05, "loss": 0.2439, "step": 7286 }, { "epoch": 0.3072285346881127, "grad_norm": 0.5599468946456909, "learning_rate": 4.9867266310628066e-05, "loss": 0.5095, "step": 7287 }, { "epoch": 0.3072706958703122, "grad_norm": 0.7367905378341675, "learning_rate": 4.986648552422e-05, "loss": 0.4169, "step": 7288 }, { "epoch": 0.30731285705251177, "grad_norm": 0.5195437669754028, "learning_rate": 4.986570473781192e-05, "loss": 0.3261, "step": 7289 }, { "epoch": 0.3073550182347113, "grad_norm": 0.727852463722229, "learning_rate": 4.986492395140386e-05, "loss": 0.4911, "step": 7290 }, { "epoch": 0.30739717941691086, "grad_norm": 0.8549585342407227, "learning_rate": 4.9864143164995784e-05, "loss": 0.2313, "step": 7291 }, { "epoch": 0.3074393405991104, "grad_norm": 0.8238479495048523, "learning_rate": 4.9863362378587715e-05, "loss": 0.5856, "step": 7292 }, { "epoch": 0.30748150178130995, "grad_norm": 2.4180850982666016, "learning_rate": 4.986258159217965e-05, "loss": 0.797, "step": 7293 }, { "epoch": 0.30752366296350947, "grad_norm": 0.6499332189559937, "learning_rate": 4.986180080577158e-05, "loss": 0.5903, "step": 7294 }, { "epoch": 0.30756582414570904, "grad_norm": 0.43270576000213623, "learning_rate": 4.986102001936351e-05, "loss": 0.0877, "step": 7295 }, { "epoch": 0.3076079853279086, "grad_norm": 0.5948163270950317, "learning_rate": 4.9860239232955434e-05, "loss": 0.2127, "step": 7296 }, { "epoch": 0.30765014651010814, "grad_norm": 0.777409017086029, "learning_rate": 4.9859458446547365e-05, "loss": 0.3288, "step": 7297 }, { "epoch": 0.3076923076923077, "grad_norm": 0.48014962673187256, "learning_rate": 4.9858677660139296e-05, "loss": 0.1481, "step": 7298 }, { "epoch": 0.3077344688745072, "grad_norm": 1.1473004817962646, "learning_rate": 4.985789687373123e-05, "loss": 0.5578, "step": 7299 }, { "epoch": 0.3077766300567068, "grad_norm": 0.6048406362533569, "learning_rate": 4.985711608732315e-05, "loss": 0.3215, "step": 7300 }, { "epoch": 0.3078187912389063, "grad_norm": 0.7616890072822571, "learning_rate": 4.985633530091508e-05, "loss": 0.5551, "step": 7301 }, { "epoch": 0.3078609524211059, "grad_norm": 0.4749971926212311, "learning_rate": 4.9855554514507014e-05, "loss": 0.3011, "step": 7302 }, { "epoch": 0.3079031136033054, "grad_norm": 0.4574067294597626, "learning_rate": 4.985477372809894e-05, "loss": 0.0635, "step": 7303 }, { "epoch": 0.307945274785505, "grad_norm": 0.8404327034950256, "learning_rate": 4.985399294169088e-05, "loss": 0.1981, "step": 7304 }, { "epoch": 0.30798743596770456, "grad_norm": 0.6325488090515137, "learning_rate": 4.98532121552828e-05, "loss": 0.5024, "step": 7305 }, { "epoch": 0.3080295971499041, "grad_norm": 15.246438026428223, "learning_rate": 4.985243136887473e-05, "loss": 1.0611, "step": 7306 }, { "epoch": 0.30807175833210365, "grad_norm": 0.4248404800891876, "learning_rate": 4.9851650582466664e-05, "loss": 0.1472, "step": 7307 }, { "epoch": 0.30811391951430317, "grad_norm": 0.9798128604888916, "learning_rate": 4.985086979605859e-05, "loss": 0.9476, "step": 7308 }, { "epoch": 0.30815608069650274, "grad_norm": 0.6632372736930847, "learning_rate": 4.9850089009650526e-05, "loss": 0.3142, "step": 7309 }, { "epoch": 0.30819824187870226, "grad_norm": 0.47884103655815125, "learning_rate": 4.984930822324245e-05, "loss": 0.1688, "step": 7310 }, { "epoch": 0.30824040306090184, "grad_norm": 0.6299852728843689, "learning_rate": 4.984852743683438e-05, "loss": 0.6059, "step": 7311 }, { "epoch": 0.30828256424310135, "grad_norm": 1.1519166231155396, "learning_rate": 4.9847746650426313e-05, "loss": 1.058, "step": 7312 }, { "epoch": 0.3083247254253009, "grad_norm": 1.194234848022461, "learning_rate": 4.984696586401824e-05, "loss": 0.5613, "step": 7313 }, { "epoch": 0.3083668866075005, "grad_norm": 1.038270354270935, "learning_rate": 4.984618507761017e-05, "loss": 0.3973, "step": 7314 }, { "epoch": 0.3084090477897, "grad_norm": 0.3868010342121124, "learning_rate": 4.98454042912021e-05, "loss": 0.1229, "step": 7315 }, { "epoch": 0.3084512089718996, "grad_norm": 0.7080903649330139, "learning_rate": 4.984462350479403e-05, "loss": 0.3222, "step": 7316 }, { "epoch": 0.3084933701540991, "grad_norm": 0.3932863473892212, "learning_rate": 4.9843842718385956e-05, "loss": 0.1174, "step": 7317 }, { "epoch": 0.3085355313362987, "grad_norm": 0.6519829034805298, "learning_rate": 4.9843061931977894e-05, "loss": 0.2577, "step": 7318 }, { "epoch": 0.3085776925184982, "grad_norm": 0.4201031029224396, "learning_rate": 4.984228114556982e-05, "loss": 0.1074, "step": 7319 }, { "epoch": 0.3086198537006978, "grad_norm": 0.7725379467010498, "learning_rate": 4.984150035916175e-05, "loss": 0.1059, "step": 7320 }, { "epoch": 0.3086620148828973, "grad_norm": 2.17915940284729, "learning_rate": 4.984071957275368e-05, "loss": 0.5948, "step": 7321 }, { "epoch": 0.30870417606509687, "grad_norm": 0.5240865349769592, "learning_rate": 4.9839938786345606e-05, "loss": 0.2227, "step": 7322 }, { "epoch": 0.3087463372472964, "grad_norm": 0.6629361510276794, "learning_rate": 4.9839157999937544e-05, "loss": 0.7866, "step": 7323 }, { "epoch": 0.30878849842949596, "grad_norm": 0.6139956116676331, "learning_rate": 4.983837721352947e-05, "loss": 0.5795, "step": 7324 }, { "epoch": 0.30883065961169553, "grad_norm": 0.5211203694343567, "learning_rate": 4.98375964271214e-05, "loss": 0.3547, "step": 7325 }, { "epoch": 0.30887282079389505, "grad_norm": 1.0110855102539062, "learning_rate": 4.983681564071333e-05, "loss": 0.7211, "step": 7326 }, { "epoch": 0.3089149819760946, "grad_norm": 0.9746285676956177, "learning_rate": 4.9836034854305255e-05, "loss": 0.4452, "step": 7327 }, { "epoch": 0.30895714315829415, "grad_norm": 0.743437647819519, "learning_rate": 4.9835254067897187e-05, "loss": 0.6398, "step": 7328 }, { "epoch": 0.3089993043404937, "grad_norm": 1.9655472040176392, "learning_rate": 4.983447328148912e-05, "loss": 0.965, "step": 7329 }, { "epoch": 0.30904146552269324, "grad_norm": 0.6540563106536865, "learning_rate": 4.983369249508105e-05, "loss": 0.7325, "step": 7330 }, { "epoch": 0.3090836267048928, "grad_norm": 0.8957778811454773, "learning_rate": 4.983291170867298e-05, "loss": 0.5301, "step": 7331 }, { "epoch": 0.30912578788709233, "grad_norm": 0.5001277923583984, "learning_rate": 4.9832130922264905e-05, "loss": 0.1037, "step": 7332 }, { "epoch": 0.3091679490692919, "grad_norm": 0.32819047570228577, "learning_rate": 4.9831350135856836e-05, "loss": 0.1399, "step": 7333 }, { "epoch": 0.3092101102514915, "grad_norm": 0.5059475302696228, "learning_rate": 4.983056934944877e-05, "loss": 0.1769, "step": 7334 }, { "epoch": 0.309252271433691, "grad_norm": 0.6028721928596497, "learning_rate": 4.98297885630407e-05, "loss": 0.638, "step": 7335 }, { "epoch": 0.30929443261589057, "grad_norm": 0.5276830792427063, "learning_rate": 4.982900777663262e-05, "loss": 0.2203, "step": 7336 }, { "epoch": 0.3093365937980901, "grad_norm": 0.8501582145690918, "learning_rate": 4.982822699022456e-05, "loss": 0.3584, "step": 7337 }, { "epoch": 0.30937875498028966, "grad_norm": 0.9470238089561462, "learning_rate": 4.9827446203816486e-05, "loss": 0.2643, "step": 7338 }, { "epoch": 0.3094209161624892, "grad_norm": 0.47941362857818604, "learning_rate": 4.982666541740842e-05, "loss": 0.2102, "step": 7339 }, { "epoch": 0.30946307734468875, "grad_norm": 1.5316853523254395, "learning_rate": 4.982588463100035e-05, "loss": 1.1165, "step": 7340 }, { "epoch": 0.30950523852688827, "grad_norm": 0.5291728973388672, "learning_rate": 4.982510384459227e-05, "loss": 0.1549, "step": 7341 }, { "epoch": 0.30954739970908784, "grad_norm": 0.49686646461486816, "learning_rate": 4.982432305818421e-05, "loss": 0.1362, "step": 7342 }, { "epoch": 0.3095895608912874, "grad_norm": 0.5368126034736633, "learning_rate": 4.9823542271776135e-05, "loss": 0.4252, "step": 7343 }, { "epoch": 0.30963172207348694, "grad_norm": 0.9058696627616882, "learning_rate": 4.9822761485368066e-05, "loss": 0.1674, "step": 7344 }, { "epoch": 0.3096738832556865, "grad_norm": 0.787880003452301, "learning_rate": 4.982198069896e-05, "loss": 0.1036, "step": 7345 }, { "epoch": 0.30971604443788603, "grad_norm": 2.0021824836730957, "learning_rate": 4.982119991255192e-05, "loss": 0.3227, "step": 7346 }, { "epoch": 0.3097582056200856, "grad_norm": 0.751654863357544, "learning_rate": 4.982041912614385e-05, "loss": 0.4592, "step": 7347 }, { "epoch": 0.3098003668022851, "grad_norm": 4.222126007080078, "learning_rate": 4.9819638339735785e-05, "loss": 0.6581, "step": 7348 }, { "epoch": 0.3098425279844847, "grad_norm": 0.6190034747123718, "learning_rate": 4.9818857553327716e-05, "loss": 0.2619, "step": 7349 }, { "epoch": 0.3098846891666842, "grad_norm": 1.1339341402053833, "learning_rate": 4.981807676691964e-05, "loss": 1.0941, "step": 7350 }, { "epoch": 0.3099268503488838, "grad_norm": 0.7093879580497742, "learning_rate": 4.981729598051157e-05, "loss": 1.1189, "step": 7351 }, { "epoch": 0.3099690115310833, "grad_norm": 1.7401368618011475, "learning_rate": 4.98165151941035e-05, "loss": 1.0941, "step": 7352 }, { "epoch": 0.3100111727132829, "grad_norm": 0.5379844903945923, "learning_rate": 4.9815734407695434e-05, "loss": 0.4778, "step": 7353 }, { "epoch": 0.31005333389548245, "grad_norm": 0.6927582025527954, "learning_rate": 4.9814953621287365e-05, "loss": 0.6404, "step": 7354 }, { "epoch": 0.31009549507768197, "grad_norm": 2.0261151790618896, "learning_rate": 4.981417283487929e-05, "loss": 0.513, "step": 7355 }, { "epoch": 0.31013765625988154, "grad_norm": 0.5914562940597534, "learning_rate": 4.981339204847123e-05, "loss": 0.1791, "step": 7356 }, { "epoch": 0.31017981744208106, "grad_norm": 0.7465640902519226, "learning_rate": 4.981261126206315e-05, "loss": 0.86, "step": 7357 }, { "epoch": 0.31022197862428064, "grad_norm": 5.471663475036621, "learning_rate": 4.981183047565508e-05, "loss": 0.1198, "step": 7358 }, { "epoch": 0.31026413980648015, "grad_norm": 0.4492192268371582, "learning_rate": 4.9811049689247015e-05, "loss": 0.3839, "step": 7359 }, { "epoch": 0.31030630098867973, "grad_norm": 0.37301093339920044, "learning_rate": 4.981026890283894e-05, "loss": 0.2217, "step": 7360 }, { "epoch": 0.31034846217087925, "grad_norm": 0.6149757504463196, "learning_rate": 4.980948811643087e-05, "loss": 0.4267, "step": 7361 }, { "epoch": 0.3103906233530788, "grad_norm": 0.7378325462341309, "learning_rate": 4.98087073300228e-05, "loss": 0.1586, "step": 7362 }, { "epoch": 0.3104327845352784, "grad_norm": 1.4971495866775513, "learning_rate": 4.980792654361473e-05, "loss": 0.9839, "step": 7363 }, { "epoch": 0.3104749457174779, "grad_norm": 0.6625845432281494, "learning_rate": 4.9807145757206664e-05, "loss": 0.7005, "step": 7364 }, { "epoch": 0.3105171068996775, "grad_norm": 0.6023096442222595, "learning_rate": 4.980636497079859e-05, "loss": 0.3052, "step": 7365 }, { "epoch": 0.310559268081877, "grad_norm": 0.613728940486908, "learning_rate": 4.980558418439052e-05, "loss": 0.3513, "step": 7366 }, { "epoch": 0.3106014292640766, "grad_norm": 0.6638384461402893, "learning_rate": 4.980480339798245e-05, "loss": 0.2244, "step": 7367 }, { "epoch": 0.3106435904462761, "grad_norm": 0.9158965349197388, "learning_rate": 4.980402261157438e-05, "loss": 0.2923, "step": 7368 }, { "epoch": 0.31068575162847567, "grad_norm": 0.4929923415184021, "learning_rate": 4.980324182516631e-05, "loss": 0.1596, "step": 7369 }, { "epoch": 0.3107279128106752, "grad_norm": 0.5832139253616333, "learning_rate": 4.980246103875824e-05, "loss": 0.3085, "step": 7370 }, { "epoch": 0.31077007399287476, "grad_norm": 1.2719762325286865, "learning_rate": 4.980168025235017e-05, "loss": 0.2381, "step": 7371 }, { "epoch": 0.31081223517507434, "grad_norm": 0.9712891578674316, "learning_rate": 4.9800899465942094e-05, "loss": 0.255, "step": 7372 }, { "epoch": 0.31085439635727385, "grad_norm": 1.258510947227478, "learning_rate": 4.980011867953403e-05, "loss": 0.6791, "step": 7373 }, { "epoch": 0.31089655753947343, "grad_norm": 1.2588434219360352, "learning_rate": 4.979933789312596e-05, "loss": 0.4389, "step": 7374 }, { "epoch": 0.31093871872167295, "grad_norm": 0.3586653769016266, "learning_rate": 4.979855710671789e-05, "loss": 0.1281, "step": 7375 }, { "epoch": 0.3109808799038725, "grad_norm": 0.38792818784713745, "learning_rate": 4.979777632030982e-05, "loss": 0.116, "step": 7376 }, { "epoch": 0.31102304108607204, "grad_norm": 1.0098776817321777, "learning_rate": 4.9796995533901744e-05, "loss": 0.5869, "step": 7377 }, { "epoch": 0.3110652022682716, "grad_norm": 0.8230668902397156, "learning_rate": 4.979621474749368e-05, "loss": 0.4772, "step": 7378 }, { "epoch": 0.31110736345047113, "grad_norm": 1.1538360118865967, "learning_rate": 4.9795433961085606e-05, "loss": 0.2273, "step": 7379 }, { "epoch": 0.3111495246326707, "grad_norm": 0.46748822927474976, "learning_rate": 4.979465317467754e-05, "loss": 0.2113, "step": 7380 }, { "epoch": 0.3111916858148702, "grad_norm": 0.4523391127586365, "learning_rate": 4.979387238826947e-05, "loss": 0.3129, "step": 7381 }, { "epoch": 0.3112338469970698, "grad_norm": 0.3432166278362274, "learning_rate": 4.97930916018614e-05, "loss": 0.1184, "step": 7382 }, { "epoch": 0.31127600817926937, "grad_norm": 0.5836549997329712, "learning_rate": 4.9792310815453324e-05, "loss": 0.5912, "step": 7383 }, { "epoch": 0.3113181693614689, "grad_norm": 0.32303857803344727, "learning_rate": 4.9791530029045256e-05, "loss": 0.0926, "step": 7384 }, { "epoch": 0.31136033054366846, "grad_norm": 2.038456439971924, "learning_rate": 4.979074924263719e-05, "loss": 0.8383, "step": 7385 }, { "epoch": 0.311402491725868, "grad_norm": 0.8159652352333069, "learning_rate": 4.978996845622912e-05, "loss": 0.1673, "step": 7386 }, { "epoch": 0.31144465290806755, "grad_norm": 1.234675645828247, "learning_rate": 4.978918766982105e-05, "loss": 0.1625, "step": 7387 }, { "epoch": 0.3114868140902671, "grad_norm": 0.5858327746391296, "learning_rate": 4.9788406883412974e-05, "loss": 0.6604, "step": 7388 }, { "epoch": 0.31152897527246665, "grad_norm": 0.8909353613853455, "learning_rate": 4.9787626097004905e-05, "loss": 0.5007, "step": 7389 }, { "epoch": 0.31157113645466616, "grad_norm": 0.62091463804245, "learning_rate": 4.9786845310596837e-05, "loss": 0.2068, "step": 7390 }, { "epoch": 0.31161329763686574, "grad_norm": 0.9866488575935364, "learning_rate": 4.978606452418876e-05, "loss": 0.5277, "step": 7391 }, { "epoch": 0.3116554588190653, "grad_norm": 0.7409759163856506, "learning_rate": 4.97852837377807e-05, "loss": 0.1595, "step": 7392 }, { "epoch": 0.31169762000126483, "grad_norm": 1.0196253061294556, "learning_rate": 4.9784502951372624e-05, "loss": 0.164, "step": 7393 }, { "epoch": 0.3117397811834644, "grad_norm": 0.366299033164978, "learning_rate": 4.9783722164964555e-05, "loss": 0.0988, "step": 7394 }, { "epoch": 0.3117819423656639, "grad_norm": 0.6329078078269958, "learning_rate": 4.9782941378556486e-05, "loss": 0.6171, "step": 7395 }, { "epoch": 0.3118241035478635, "grad_norm": 0.5615409016609192, "learning_rate": 4.978216059214841e-05, "loss": 0.1855, "step": 7396 }, { "epoch": 0.311866264730063, "grad_norm": 0.5160951018333435, "learning_rate": 4.978137980574034e-05, "loss": 0.1649, "step": 7397 }, { "epoch": 0.3119084259122626, "grad_norm": 0.47677475214004517, "learning_rate": 4.978059901933227e-05, "loss": 0.1717, "step": 7398 }, { "epoch": 0.3119505870944621, "grad_norm": 1.0693538188934326, "learning_rate": 4.9779818232924204e-05, "loss": 0.6017, "step": 7399 }, { "epoch": 0.3119927482766617, "grad_norm": 0.7265190482139587, "learning_rate": 4.9779037446516136e-05, "loss": 0.2005, "step": 7400 }, { "epoch": 0.31203490945886125, "grad_norm": 0.5878857374191284, "learning_rate": 4.977825666010807e-05, "loss": 0.1783, "step": 7401 }, { "epoch": 0.31207707064106077, "grad_norm": 0.7172933220863342, "learning_rate": 4.977747587369999e-05, "loss": 0.2392, "step": 7402 }, { "epoch": 0.31211923182326035, "grad_norm": 2.1055192947387695, "learning_rate": 4.977669508729192e-05, "loss": 0.9048, "step": 7403 }, { "epoch": 0.31216139300545986, "grad_norm": 7.124802112579346, "learning_rate": 4.9775914300883854e-05, "loss": 0.3421, "step": 7404 }, { "epoch": 0.31220355418765944, "grad_norm": 0.8219845294952393, "learning_rate": 4.977513351447578e-05, "loss": 0.2697, "step": 7405 }, { "epoch": 0.31224571536985896, "grad_norm": 0.23137237131595612, "learning_rate": 4.9774352728067716e-05, "loss": 0.0743, "step": 7406 }, { "epoch": 0.31228787655205853, "grad_norm": 0.5465040802955627, "learning_rate": 4.977357194165964e-05, "loss": 0.1436, "step": 7407 }, { "epoch": 0.31233003773425805, "grad_norm": 0.7622815370559692, "learning_rate": 4.977279115525157e-05, "loss": 0.6409, "step": 7408 }, { "epoch": 0.3123721989164576, "grad_norm": 0.6293045878410339, "learning_rate": 4.97720103688435e-05, "loss": 0.083, "step": 7409 }, { "epoch": 0.31241436009865714, "grad_norm": 0.5946604013442993, "learning_rate": 4.977122958243543e-05, "loss": 0.1949, "step": 7410 }, { "epoch": 0.3124565212808567, "grad_norm": 0.3505973219871521, "learning_rate": 4.9770448796027366e-05, "loss": 0.0956, "step": 7411 }, { "epoch": 0.3124986824630563, "grad_norm": 0.4026801288127899, "learning_rate": 4.976966800961929e-05, "loss": 0.4092, "step": 7412 }, { "epoch": 0.3125408436452558, "grad_norm": 1.7664768695831299, "learning_rate": 4.976888722321122e-05, "loss": 0.4285, "step": 7413 }, { "epoch": 0.3125830048274554, "grad_norm": 0.4118407070636749, "learning_rate": 4.976810643680315e-05, "loss": 0.1488, "step": 7414 }, { "epoch": 0.3126251660096549, "grad_norm": 0.42180177569389343, "learning_rate": 4.976732565039508e-05, "loss": 0.0811, "step": 7415 }, { "epoch": 0.31266732719185447, "grad_norm": 0.3554933965206146, "learning_rate": 4.976654486398701e-05, "loss": 0.1153, "step": 7416 }, { "epoch": 0.312709488374054, "grad_norm": 1.2534043788909912, "learning_rate": 4.976576407757894e-05, "loss": 0.2399, "step": 7417 }, { "epoch": 0.31275164955625356, "grad_norm": 0.6492864489555359, "learning_rate": 4.976498329117087e-05, "loss": 0.6447, "step": 7418 }, { "epoch": 0.3127938107384531, "grad_norm": 2.76101016998291, "learning_rate": 4.9764202504762796e-05, "loss": 0.5226, "step": 7419 }, { "epoch": 0.31283597192065266, "grad_norm": 0.7576063275337219, "learning_rate": 4.976342171835473e-05, "loss": 0.1765, "step": 7420 }, { "epoch": 0.31287813310285223, "grad_norm": 0.7642138004302979, "learning_rate": 4.976264093194666e-05, "loss": 0.4604, "step": 7421 }, { "epoch": 0.31292029428505175, "grad_norm": 0.5912198424339294, "learning_rate": 4.976186014553859e-05, "loss": 0.1182, "step": 7422 }, { "epoch": 0.3129624554672513, "grad_norm": 0.2976185381412506, "learning_rate": 4.976107935913052e-05, "loss": 0.1123, "step": 7423 }, { "epoch": 0.31300461664945084, "grad_norm": 0.8649865984916687, "learning_rate": 4.9760298572722445e-05, "loss": 0.5743, "step": 7424 }, { "epoch": 0.3130467778316504, "grad_norm": 0.38361695408821106, "learning_rate": 4.975951778631438e-05, "loss": 0.1327, "step": 7425 }, { "epoch": 0.31308893901384993, "grad_norm": 2.192126512527466, "learning_rate": 4.975873699990631e-05, "loss": 1.1272, "step": 7426 }, { "epoch": 0.3131311001960495, "grad_norm": 3.9383630752563477, "learning_rate": 4.975795621349824e-05, "loss": 0.5482, "step": 7427 }, { "epoch": 0.313173261378249, "grad_norm": 0.7183594703674316, "learning_rate": 4.975717542709017e-05, "loss": 0.1544, "step": 7428 }, { "epoch": 0.3132154225604486, "grad_norm": 0.7689663767814636, "learning_rate": 4.9756394640682095e-05, "loss": 0.5159, "step": 7429 }, { "epoch": 0.31325758374264817, "grad_norm": 0.5283539295196533, "learning_rate": 4.9755613854274026e-05, "loss": 0.2346, "step": 7430 }, { "epoch": 0.3132997449248477, "grad_norm": 0.6981973052024841, "learning_rate": 4.975483306786596e-05, "loss": 0.2364, "step": 7431 }, { "epoch": 0.31334190610704726, "grad_norm": 0.9903140068054199, "learning_rate": 4.975405228145789e-05, "loss": 0.3392, "step": 7432 }, { "epoch": 0.3133840672892468, "grad_norm": 1.0620148181915283, "learning_rate": 4.975327149504982e-05, "loss": 0.2423, "step": 7433 }, { "epoch": 0.31342622847144636, "grad_norm": 0.5028334259986877, "learning_rate": 4.9752490708641744e-05, "loss": 0.1189, "step": 7434 }, { "epoch": 0.3134683896536459, "grad_norm": 0.8216344714164734, "learning_rate": 4.9751709922233675e-05, "loss": 0.5068, "step": 7435 }, { "epoch": 0.31351055083584545, "grad_norm": 1.0724025964736938, "learning_rate": 4.975092913582561e-05, "loss": 0.237, "step": 7436 }, { "epoch": 0.31355271201804497, "grad_norm": 0.643614649772644, "learning_rate": 4.975014834941754e-05, "loss": 0.5956, "step": 7437 }, { "epoch": 0.31359487320024454, "grad_norm": 0.9833900332450867, "learning_rate": 4.974936756300946e-05, "loss": 0.1633, "step": 7438 }, { "epoch": 0.31363703438244406, "grad_norm": 0.7330596446990967, "learning_rate": 4.9748586776601394e-05, "loss": 0.5147, "step": 7439 }, { "epoch": 0.31367919556464363, "grad_norm": 0.5808245539665222, "learning_rate": 4.9747805990193325e-05, "loss": 0.5513, "step": 7440 }, { "epoch": 0.3137213567468432, "grad_norm": 0.6119595766067505, "learning_rate": 4.974702520378525e-05, "loss": 0.3867, "step": 7441 }, { "epoch": 0.3137635179290427, "grad_norm": 1.250051736831665, "learning_rate": 4.974624441737719e-05, "loss": 1.0128, "step": 7442 }, { "epoch": 0.3138056791112423, "grad_norm": 0.7736785411834717, "learning_rate": 4.974546363096911e-05, "loss": 0.4077, "step": 7443 }, { "epoch": 0.3138478402934418, "grad_norm": 1.143507719039917, "learning_rate": 4.974468284456105e-05, "loss": 0.1201, "step": 7444 }, { "epoch": 0.3138900014756414, "grad_norm": 2.0376136302948, "learning_rate": 4.9743902058152975e-05, "loss": 0.5578, "step": 7445 }, { "epoch": 0.3139321626578409, "grad_norm": 0.5436827540397644, "learning_rate": 4.9743121271744906e-05, "loss": 0.4906, "step": 7446 }, { "epoch": 0.3139743238400405, "grad_norm": 1.5805742740631104, "learning_rate": 4.974234048533684e-05, "loss": 0.6887, "step": 7447 }, { "epoch": 0.31401648502224, "grad_norm": 0.6701764464378357, "learning_rate": 4.974155969892876e-05, "loss": 0.2669, "step": 7448 }, { "epoch": 0.3140586462044396, "grad_norm": 1.7213096618652344, "learning_rate": 4.974077891252069e-05, "loss": 0.2486, "step": 7449 }, { "epoch": 0.31410080738663915, "grad_norm": 0.2883455455303192, "learning_rate": 4.9739998126112624e-05, "loss": 0.0819, "step": 7450 }, { "epoch": 0.31414296856883867, "grad_norm": 0.8456642031669617, "learning_rate": 4.9739217339704555e-05, "loss": 0.2362, "step": 7451 }, { "epoch": 0.31418512975103824, "grad_norm": 0.7722310423851013, "learning_rate": 4.973843655329648e-05, "loss": 0.3117, "step": 7452 }, { "epoch": 0.31422729093323776, "grad_norm": 0.4684138894081116, "learning_rate": 4.973765576688841e-05, "loss": 0.1928, "step": 7453 }, { "epoch": 0.31426945211543733, "grad_norm": 0.9729281067848206, "learning_rate": 4.973687498048034e-05, "loss": 0.9488, "step": 7454 }, { "epoch": 0.31431161329763685, "grad_norm": 0.8195375204086304, "learning_rate": 4.9736094194072274e-05, "loss": 0.7168, "step": 7455 }, { "epoch": 0.3143537744798364, "grad_norm": 0.6616782546043396, "learning_rate": 4.9735313407664205e-05, "loss": 0.1658, "step": 7456 }, { "epoch": 0.31439593566203594, "grad_norm": 0.7084793448448181, "learning_rate": 4.973453262125613e-05, "loss": 0.2995, "step": 7457 }, { "epoch": 0.3144380968442355, "grad_norm": 2.3319013118743896, "learning_rate": 4.973375183484806e-05, "loss": 1.165, "step": 7458 }, { "epoch": 0.3144802580264351, "grad_norm": 0.3117603361606598, "learning_rate": 4.973297104843999e-05, "loss": 0.1526, "step": 7459 }, { "epoch": 0.3145224192086346, "grad_norm": 1.4241353273391724, "learning_rate": 4.9732190262031916e-05, "loss": 0.374, "step": 7460 }, { "epoch": 0.3145645803908342, "grad_norm": 0.7663248181343079, "learning_rate": 4.9731409475623854e-05, "loss": 0.221, "step": 7461 }, { "epoch": 0.3146067415730337, "grad_norm": 1.604125738143921, "learning_rate": 4.973062868921578e-05, "loss": 1.0904, "step": 7462 }, { "epoch": 0.3146489027552333, "grad_norm": 1.258020281791687, "learning_rate": 4.972984790280771e-05, "loss": 0.7044, "step": 7463 }, { "epoch": 0.3146910639374328, "grad_norm": 0.7852577567100525, "learning_rate": 4.972906711639964e-05, "loss": 0.6089, "step": 7464 }, { "epoch": 0.31473322511963236, "grad_norm": 1.5852376222610474, "learning_rate": 4.9728286329991566e-05, "loss": 0.3405, "step": 7465 }, { "epoch": 0.3147753863018319, "grad_norm": 0.3444455862045288, "learning_rate": 4.9727505543583504e-05, "loss": 0.0842, "step": 7466 }, { "epoch": 0.31481754748403146, "grad_norm": 0.2885657548904419, "learning_rate": 4.972672475717543e-05, "loss": 0.0915, "step": 7467 }, { "epoch": 0.314859708666231, "grad_norm": 0.5658945441246033, "learning_rate": 4.972594397076736e-05, "loss": 0.3398, "step": 7468 }, { "epoch": 0.31490186984843055, "grad_norm": 0.8077555894851685, "learning_rate": 4.972516318435929e-05, "loss": 0.125, "step": 7469 }, { "epoch": 0.3149440310306301, "grad_norm": 0.9568435549736023, "learning_rate": 4.972438239795122e-05, "loss": 0.1916, "step": 7470 }, { "epoch": 0.31498619221282964, "grad_norm": 0.48458656668663025, "learning_rate": 4.9723601611543147e-05, "loss": 0.2589, "step": 7471 }, { "epoch": 0.3150283533950292, "grad_norm": 0.5101461410522461, "learning_rate": 4.972282082513508e-05, "loss": 0.1106, "step": 7472 }, { "epoch": 0.31507051457722873, "grad_norm": 0.8014541268348694, "learning_rate": 4.972204003872701e-05, "loss": 0.4512, "step": 7473 }, { "epoch": 0.3151126757594283, "grad_norm": 1.152951955795288, "learning_rate": 4.9721259252318934e-05, "loss": 0.1731, "step": 7474 }, { "epoch": 0.3151548369416278, "grad_norm": 0.664989709854126, "learning_rate": 4.972047846591087e-05, "loss": 0.6616, "step": 7475 }, { "epoch": 0.3151969981238274, "grad_norm": 1.767228364944458, "learning_rate": 4.9719697679502796e-05, "loss": 0.9784, "step": 7476 }, { "epoch": 0.3152391593060269, "grad_norm": 0.3732711672782898, "learning_rate": 4.971891689309473e-05, "loss": 0.1468, "step": 7477 }, { "epoch": 0.3152813204882265, "grad_norm": 0.5688472390174866, "learning_rate": 4.971813610668666e-05, "loss": 0.4556, "step": 7478 }, { "epoch": 0.31532348167042606, "grad_norm": 0.3739578425884247, "learning_rate": 4.971735532027858e-05, "loss": 0.1291, "step": 7479 }, { "epoch": 0.3153656428526256, "grad_norm": 0.5275723934173584, "learning_rate": 4.971657453387052e-05, "loss": 0.1499, "step": 7480 }, { "epoch": 0.31540780403482516, "grad_norm": 0.4393303394317627, "learning_rate": 4.9715793747462446e-05, "loss": 0.1888, "step": 7481 }, { "epoch": 0.3154499652170247, "grad_norm": 0.726729154586792, "learning_rate": 4.971501296105438e-05, "loss": 0.4439, "step": 7482 }, { "epoch": 0.31549212639922425, "grad_norm": 0.7275832891464233, "learning_rate": 4.971423217464631e-05, "loss": 0.626, "step": 7483 }, { "epoch": 0.31553428758142377, "grad_norm": 0.9625864028930664, "learning_rate": 4.971345138823823e-05, "loss": 0.7377, "step": 7484 }, { "epoch": 0.31557644876362334, "grad_norm": 1.216604471206665, "learning_rate": 4.9712670601830164e-05, "loss": 0.276, "step": 7485 }, { "epoch": 0.31561860994582286, "grad_norm": 1.0304571390151978, "learning_rate": 4.9711889815422095e-05, "loss": 0.3401, "step": 7486 }, { "epoch": 0.31566077112802243, "grad_norm": 0.639548659324646, "learning_rate": 4.9711109029014026e-05, "loss": 0.3938, "step": 7487 }, { "epoch": 0.31570293231022195, "grad_norm": 0.7457719445228577, "learning_rate": 4.971032824260595e-05, "loss": 0.4122, "step": 7488 }, { "epoch": 0.3157450934924215, "grad_norm": 10.84555721282959, "learning_rate": 4.970954745619789e-05, "loss": 0.1388, "step": 7489 }, { "epoch": 0.3157872546746211, "grad_norm": 0.43018537759780884, "learning_rate": 4.9708766669789813e-05, "loss": 0.2194, "step": 7490 }, { "epoch": 0.3158294158568206, "grad_norm": 0.4613083600997925, "learning_rate": 4.9707985883381745e-05, "loss": 0.1693, "step": 7491 }, { "epoch": 0.3158715770390202, "grad_norm": 0.6272141933441162, "learning_rate": 4.9707205096973676e-05, "loss": 0.5183, "step": 7492 }, { "epoch": 0.3159137382212197, "grad_norm": 0.7772790789604187, "learning_rate": 4.97064243105656e-05, "loss": 0.589, "step": 7493 }, { "epoch": 0.3159558994034193, "grad_norm": 0.5661923885345459, "learning_rate": 4.970564352415754e-05, "loss": 0.5908, "step": 7494 }, { "epoch": 0.3159980605856188, "grad_norm": 0.4282901883125305, "learning_rate": 4.970486273774946e-05, "loss": 0.1533, "step": 7495 }, { "epoch": 0.3160402217678184, "grad_norm": 1.2892993688583374, "learning_rate": 4.9704081951341394e-05, "loss": 0.3172, "step": 7496 }, { "epoch": 0.3160823829500179, "grad_norm": 0.6988453269004822, "learning_rate": 4.9703301164933325e-05, "loss": 0.3427, "step": 7497 }, { "epoch": 0.31612454413221747, "grad_norm": 1.0176929235458374, "learning_rate": 4.970252037852525e-05, "loss": 0.1883, "step": 7498 }, { "epoch": 0.31616670531441704, "grad_norm": 0.788316547870636, "learning_rate": 4.970173959211718e-05, "loss": 0.2544, "step": 7499 }, { "epoch": 0.31620886649661656, "grad_norm": 1.2757824659347534, "learning_rate": 4.970095880570911e-05, "loss": 1.0208, "step": 7500 }, { "epoch": 0.31625102767881613, "grad_norm": 0.5032047629356384, "learning_rate": 4.9700178019301044e-05, "loss": 0.1336, "step": 7501 }, { "epoch": 0.31629318886101565, "grad_norm": 0.8490986824035645, "learning_rate": 4.9699397232892975e-05, "loss": 0.1858, "step": 7502 }, { "epoch": 0.3163353500432152, "grad_norm": 0.7827265858650208, "learning_rate": 4.96986164464849e-05, "loss": 0.1693, "step": 7503 }, { "epoch": 0.31637751122541474, "grad_norm": 0.5583136677742004, "learning_rate": 4.969783566007683e-05, "loss": 0.1926, "step": 7504 }, { "epoch": 0.3164196724076143, "grad_norm": 1.5489495992660522, "learning_rate": 4.969705487366876e-05, "loss": 0.5337, "step": 7505 }, { "epoch": 0.31646183358981383, "grad_norm": 1.4138537645339966, "learning_rate": 4.969627408726069e-05, "loss": 0.1571, "step": 7506 }, { "epoch": 0.3165039947720134, "grad_norm": 0.8502295613288879, "learning_rate": 4.969549330085262e-05, "loss": 0.1559, "step": 7507 }, { "epoch": 0.316546155954213, "grad_norm": 2.1276800632476807, "learning_rate": 4.9694712514444556e-05, "loss": 0.7711, "step": 7508 }, { "epoch": 0.3165883171364125, "grad_norm": 0.5596317648887634, "learning_rate": 4.969393172803648e-05, "loss": 0.2342, "step": 7509 }, { "epoch": 0.3166304783186121, "grad_norm": 0.7909955978393555, "learning_rate": 4.9693150941628405e-05, "loss": 0.2328, "step": 7510 }, { "epoch": 0.3166726395008116, "grad_norm": 0.3848377764225006, "learning_rate": 4.969237015522034e-05, "loss": 0.1834, "step": 7511 }, { "epoch": 0.31671480068301117, "grad_norm": 0.36012789607048035, "learning_rate": 4.969158936881227e-05, "loss": 0.1137, "step": 7512 }, { "epoch": 0.3167569618652107, "grad_norm": 0.5981179475784302, "learning_rate": 4.9690808582404205e-05, "loss": 0.5496, "step": 7513 }, { "epoch": 0.31679912304741026, "grad_norm": 0.6942450404167175, "learning_rate": 4.969002779599613e-05, "loss": 0.3249, "step": 7514 }, { "epoch": 0.3168412842296098, "grad_norm": 1.029809594154358, "learning_rate": 4.968924700958806e-05, "loss": 0.2599, "step": 7515 }, { "epoch": 0.31688344541180935, "grad_norm": 0.4270893931388855, "learning_rate": 4.968846622317999e-05, "loss": 0.1037, "step": 7516 }, { "epoch": 0.31692560659400887, "grad_norm": 0.684688925743103, "learning_rate": 4.968768543677192e-05, "loss": 0.4534, "step": 7517 }, { "epoch": 0.31696776777620844, "grad_norm": 1.159080982208252, "learning_rate": 4.968690465036385e-05, "loss": 0.8668, "step": 7518 }, { "epoch": 0.317009928958408, "grad_norm": 0.5026242733001709, "learning_rate": 4.968612386395578e-05, "loss": 0.1231, "step": 7519 }, { "epoch": 0.31705209014060753, "grad_norm": 0.6594807505607605, "learning_rate": 4.968534307754771e-05, "loss": 0.4549, "step": 7520 }, { "epoch": 0.3170942513228071, "grad_norm": 1.5901532173156738, "learning_rate": 4.9684562291139635e-05, "loss": 0.6947, "step": 7521 }, { "epoch": 0.3171364125050066, "grad_norm": 0.48920121788978577, "learning_rate": 4.9683781504731566e-05, "loss": 0.2333, "step": 7522 }, { "epoch": 0.3171785736872062, "grad_norm": 0.7721165418624878, "learning_rate": 4.96830007183235e-05, "loss": 0.1547, "step": 7523 }, { "epoch": 0.3172207348694057, "grad_norm": 0.7219516634941101, "learning_rate": 4.968221993191543e-05, "loss": 0.6595, "step": 7524 }, { "epoch": 0.3172628960516053, "grad_norm": 1.7347166538238525, "learning_rate": 4.968143914550736e-05, "loss": 1.0659, "step": 7525 }, { "epoch": 0.3173050572338048, "grad_norm": 0.5281823873519897, "learning_rate": 4.9680658359099285e-05, "loss": 0.2116, "step": 7526 }, { "epoch": 0.3173472184160044, "grad_norm": 0.6219302415847778, "learning_rate": 4.9679877572691216e-05, "loss": 0.5171, "step": 7527 }, { "epoch": 0.31738937959820396, "grad_norm": 0.6347947716712952, "learning_rate": 4.967909678628315e-05, "loss": 0.5022, "step": 7528 }, { "epoch": 0.3174315407804035, "grad_norm": 0.7337154746055603, "learning_rate": 4.967831599987507e-05, "loss": 0.2348, "step": 7529 }, { "epoch": 0.31747370196260305, "grad_norm": 0.4548053443431854, "learning_rate": 4.967753521346701e-05, "loss": 0.2968, "step": 7530 }, { "epoch": 0.31751586314480257, "grad_norm": 0.5477220416069031, "learning_rate": 4.9676754427058934e-05, "loss": 0.5069, "step": 7531 }, { "epoch": 0.31755802432700214, "grad_norm": 0.713554859161377, "learning_rate": 4.9675973640650865e-05, "loss": 0.1673, "step": 7532 }, { "epoch": 0.31760018550920166, "grad_norm": 0.7669718861579895, "learning_rate": 4.96751928542428e-05, "loss": 0.1597, "step": 7533 }, { "epoch": 0.31764234669140123, "grad_norm": 3.522759437561035, "learning_rate": 4.967441206783473e-05, "loss": 0.1078, "step": 7534 }, { "epoch": 0.31768450787360075, "grad_norm": 0.4969756007194519, "learning_rate": 4.967363128142666e-05, "loss": 0.2655, "step": 7535 }, { "epoch": 0.3177266690558003, "grad_norm": 3.2893593311309814, "learning_rate": 4.9672850495018584e-05, "loss": 0.4322, "step": 7536 }, { "epoch": 0.3177688302379999, "grad_norm": 1.2042288780212402, "learning_rate": 4.9672069708610515e-05, "loss": 0.8958, "step": 7537 }, { "epoch": 0.3178109914201994, "grad_norm": 0.6042218804359436, "learning_rate": 4.9671288922202446e-05, "loss": 0.1464, "step": 7538 }, { "epoch": 0.317853152602399, "grad_norm": 0.6230404376983643, "learning_rate": 4.967050813579438e-05, "loss": 0.1349, "step": 7539 }, { "epoch": 0.3178953137845985, "grad_norm": 0.6494224667549133, "learning_rate": 4.96697273493863e-05, "loss": 0.2251, "step": 7540 }, { "epoch": 0.3179374749667981, "grad_norm": 0.7012438774108887, "learning_rate": 4.966894656297823e-05, "loss": 0.1635, "step": 7541 }, { "epoch": 0.3179796361489976, "grad_norm": 0.5733148455619812, "learning_rate": 4.9668165776570164e-05, "loss": 0.5593, "step": 7542 }, { "epoch": 0.3180217973311972, "grad_norm": 2.03416109085083, "learning_rate": 4.966738499016209e-05, "loss": 0.8765, "step": 7543 }, { "epoch": 0.3180639585133967, "grad_norm": 0.5987300276756287, "learning_rate": 4.966660420375403e-05, "loss": 0.1124, "step": 7544 }, { "epoch": 0.31810611969559627, "grad_norm": 1.1106184720993042, "learning_rate": 4.966582341734595e-05, "loss": 0.9661, "step": 7545 }, { "epoch": 0.3181482808777958, "grad_norm": 0.5323086380958557, "learning_rate": 4.966504263093788e-05, "loss": 0.5019, "step": 7546 }, { "epoch": 0.31819044205999536, "grad_norm": 0.36030951142311096, "learning_rate": 4.9664261844529814e-05, "loss": 0.1148, "step": 7547 }, { "epoch": 0.31823260324219493, "grad_norm": 0.715556800365448, "learning_rate": 4.966348105812174e-05, "loss": 0.1161, "step": 7548 }, { "epoch": 0.31827476442439445, "grad_norm": 0.6653590202331543, "learning_rate": 4.9662700271713676e-05, "loss": 0.2134, "step": 7549 }, { "epoch": 0.318316925606594, "grad_norm": 3.4779393672943115, "learning_rate": 4.96619194853056e-05, "loss": 0.8497, "step": 7550 }, { "epoch": 0.31835908678879354, "grad_norm": 1.034873604774475, "learning_rate": 4.966113869889753e-05, "loss": 0.5055, "step": 7551 }, { "epoch": 0.3184012479709931, "grad_norm": 0.603668212890625, "learning_rate": 4.9660357912489463e-05, "loss": 0.5475, "step": 7552 }, { "epoch": 0.31844340915319264, "grad_norm": 0.8274823427200317, "learning_rate": 4.9659577126081395e-05, "loss": 0.4343, "step": 7553 }, { "epoch": 0.3184855703353922, "grad_norm": 0.49285539984703064, "learning_rate": 4.965879633967332e-05, "loss": 0.3206, "step": 7554 }, { "epoch": 0.31852773151759173, "grad_norm": 0.6695259213447571, "learning_rate": 4.965801555326525e-05, "loss": 0.7361, "step": 7555 }, { "epoch": 0.3185698926997913, "grad_norm": 9.270378112792969, "learning_rate": 4.965723476685718e-05, "loss": 0.2158, "step": 7556 }, { "epoch": 0.3186120538819909, "grad_norm": 1.0417948961257935, "learning_rate": 4.9656453980449106e-05, "loss": 0.7615, "step": 7557 }, { "epoch": 0.3186542150641904, "grad_norm": 0.4772723913192749, "learning_rate": 4.9655673194041044e-05, "loss": 0.2156, "step": 7558 }, { "epoch": 0.31869637624638997, "grad_norm": 0.7545613646507263, "learning_rate": 4.965489240763297e-05, "loss": 0.101, "step": 7559 }, { "epoch": 0.3187385374285895, "grad_norm": 0.7614084482192993, "learning_rate": 4.96541116212249e-05, "loss": 0.293, "step": 7560 }, { "epoch": 0.31878069861078906, "grad_norm": 1.5774955749511719, "learning_rate": 4.965333083481683e-05, "loss": 0.9415, "step": 7561 }, { "epoch": 0.3188228597929886, "grad_norm": 0.5689199566841125, "learning_rate": 4.9652550048408756e-05, "loss": 0.2639, "step": 7562 }, { "epoch": 0.31886502097518815, "grad_norm": 1.778674602508545, "learning_rate": 4.9651769262000694e-05, "loss": 0.476, "step": 7563 }, { "epoch": 0.31890718215738767, "grad_norm": 0.6827443242073059, "learning_rate": 4.965098847559262e-05, "loss": 0.3774, "step": 7564 }, { "epoch": 0.31894934333958724, "grad_norm": 0.88897705078125, "learning_rate": 4.965020768918455e-05, "loss": 0.9194, "step": 7565 }, { "epoch": 0.3189915045217868, "grad_norm": 1.1892282962799072, "learning_rate": 4.964942690277648e-05, "loss": 0.9065, "step": 7566 }, { "epoch": 0.31903366570398634, "grad_norm": 3.0558481216430664, "learning_rate": 4.9648646116368405e-05, "loss": 1.216, "step": 7567 }, { "epoch": 0.3190758268861859, "grad_norm": 4.903900623321533, "learning_rate": 4.9647865329960337e-05, "loss": 0.147, "step": 7568 }, { "epoch": 0.3191179880683854, "grad_norm": 0.7892494201660156, "learning_rate": 4.964708454355227e-05, "loss": 0.2764, "step": 7569 }, { "epoch": 0.319160149250585, "grad_norm": 1.3696013689041138, "learning_rate": 4.96463037571442e-05, "loss": 1.3817, "step": 7570 }, { "epoch": 0.3192023104327845, "grad_norm": 0.8516808152198792, "learning_rate": 4.964552297073613e-05, "loss": 1.1097, "step": 7571 }, { "epoch": 0.3192444716149841, "grad_norm": 0.6897605657577515, "learning_rate": 4.9644742184328055e-05, "loss": 0.1633, "step": 7572 }, { "epoch": 0.3192866327971836, "grad_norm": 0.43224892020225525, "learning_rate": 4.9643961397919986e-05, "loss": 0.1365, "step": 7573 }, { "epoch": 0.3193287939793832, "grad_norm": 1.3007193803787231, "learning_rate": 4.964318061151192e-05, "loss": 0.2091, "step": 7574 }, { "epoch": 0.3193709551615827, "grad_norm": 0.4700179994106293, "learning_rate": 4.964239982510385e-05, "loss": 0.1893, "step": 7575 }, { "epoch": 0.3194131163437823, "grad_norm": 0.3448899984359741, "learning_rate": 4.964161903869577e-05, "loss": 0.1218, "step": 7576 }, { "epoch": 0.31945527752598185, "grad_norm": 0.63805091381073, "learning_rate": 4.964083825228771e-05, "loss": 0.3883, "step": 7577 }, { "epoch": 0.31949743870818137, "grad_norm": 1.8907123804092407, "learning_rate": 4.9640057465879636e-05, "loss": 0.847, "step": 7578 }, { "epoch": 0.31953959989038094, "grad_norm": 1.039634346961975, "learning_rate": 4.963927667947157e-05, "loss": 0.1467, "step": 7579 }, { "epoch": 0.31958176107258046, "grad_norm": 1.1819982528686523, "learning_rate": 4.96384958930635e-05, "loss": 0.2856, "step": 7580 }, { "epoch": 0.31962392225478004, "grad_norm": 1.3843145370483398, "learning_rate": 4.963771510665542e-05, "loss": 1.0682, "step": 7581 }, { "epoch": 0.31966608343697955, "grad_norm": 1.389547348022461, "learning_rate": 4.963693432024736e-05, "loss": 0.874, "step": 7582 }, { "epoch": 0.3197082446191791, "grad_norm": 0.6258354783058167, "learning_rate": 4.9636153533839285e-05, "loss": 0.5901, "step": 7583 }, { "epoch": 0.31975040580137865, "grad_norm": 1.0194568634033203, "learning_rate": 4.9635372747431216e-05, "loss": 0.0736, "step": 7584 }, { "epoch": 0.3197925669835782, "grad_norm": 0.817176878452301, "learning_rate": 4.963459196102315e-05, "loss": 0.2114, "step": 7585 }, { "epoch": 0.3198347281657778, "grad_norm": 5.667961597442627, "learning_rate": 4.963381117461507e-05, "loss": 0.1166, "step": 7586 }, { "epoch": 0.3198768893479773, "grad_norm": 0.46179381012916565, "learning_rate": 4.9633030388207e-05, "loss": 0.1296, "step": 7587 }, { "epoch": 0.3199190505301769, "grad_norm": 1.4030051231384277, "learning_rate": 4.9632249601798935e-05, "loss": 0.4256, "step": 7588 }, { "epoch": 0.3199612117123764, "grad_norm": 0.7672938108444214, "learning_rate": 4.9631468815390866e-05, "loss": 0.3469, "step": 7589 }, { "epoch": 0.320003372894576, "grad_norm": 3.6712303161621094, "learning_rate": 4.963068802898279e-05, "loss": 1.0481, "step": 7590 }, { "epoch": 0.3200455340767755, "grad_norm": 1.2141425609588623, "learning_rate": 4.962990724257472e-05, "loss": 0.587, "step": 7591 }, { "epoch": 0.32008769525897507, "grad_norm": 3.090832471847534, "learning_rate": 4.962912645616665e-05, "loss": 1.6861, "step": 7592 }, { "epoch": 0.3201298564411746, "grad_norm": 1.723828911781311, "learning_rate": 4.9628345669758584e-05, "loss": 0.645, "step": 7593 }, { "epoch": 0.32017201762337416, "grad_norm": 0.6652144193649292, "learning_rate": 4.9627564883350515e-05, "loss": 0.165, "step": 7594 }, { "epoch": 0.32021417880557373, "grad_norm": 0.3697267770767212, "learning_rate": 4.962678409694244e-05, "loss": 0.0889, "step": 7595 }, { "epoch": 0.32025633998777325, "grad_norm": 0.9362396597862244, "learning_rate": 4.962600331053438e-05, "loss": 0.2081, "step": 7596 }, { "epoch": 0.3202985011699728, "grad_norm": 0.47456541657447815, "learning_rate": 4.96252225241263e-05, "loss": 0.2913, "step": 7597 }, { "epoch": 0.32034066235217235, "grad_norm": 0.473869651556015, "learning_rate": 4.962444173771823e-05, "loss": 0.0999, "step": 7598 }, { "epoch": 0.3203828235343719, "grad_norm": 0.3059614896774292, "learning_rate": 4.9623660951310165e-05, "loss": 0.0722, "step": 7599 }, { "epoch": 0.32042498471657144, "grad_norm": 4.915802478790283, "learning_rate": 4.962288016490209e-05, "loss": 0.1023, "step": 7600 }, { "epoch": 0.320467145898771, "grad_norm": 0.4842813014984131, "learning_rate": 4.962209937849402e-05, "loss": 0.2055, "step": 7601 }, { "epoch": 0.32050930708097053, "grad_norm": 1.4646937847137451, "learning_rate": 4.962131859208595e-05, "loss": 1.1452, "step": 7602 }, { "epoch": 0.3205514682631701, "grad_norm": 0.7019630670547485, "learning_rate": 4.962053780567788e-05, "loss": 0.2384, "step": 7603 }, { "epoch": 0.3205936294453696, "grad_norm": 0.48829391598701477, "learning_rate": 4.9619757019269814e-05, "loss": 0.1572, "step": 7604 }, { "epoch": 0.3206357906275692, "grad_norm": 0.657062292098999, "learning_rate": 4.961897623286174e-05, "loss": 0.6428, "step": 7605 }, { "epoch": 0.32067795180976877, "grad_norm": 2.1735527515411377, "learning_rate": 4.961819544645367e-05, "loss": 0.5431, "step": 7606 }, { "epoch": 0.3207201129919683, "grad_norm": 0.5575162172317505, "learning_rate": 4.96174146600456e-05, "loss": 0.0837, "step": 7607 }, { "epoch": 0.32076227417416786, "grad_norm": 0.7482373118400574, "learning_rate": 4.961663387363753e-05, "loss": 0.1846, "step": 7608 }, { "epoch": 0.3208044353563674, "grad_norm": 0.435262531042099, "learning_rate": 4.961585308722946e-05, "loss": 0.313, "step": 7609 }, { "epoch": 0.32084659653856695, "grad_norm": 0.46606773138046265, "learning_rate": 4.961507230082139e-05, "loss": 0.2015, "step": 7610 }, { "epoch": 0.32088875772076647, "grad_norm": 0.856580376625061, "learning_rate": 4.961429151441332e-05, "loss": 0.3932, "step": 7611 }, { "epoch": 0.32093091890296604, "grad_norm": 0.2951192259788513, "learning_rate": 4.9613510728005244e-05, "loss": 0.11, "step": 7612 }, { "epoch": 0.32097308008516556, "grad_norm": 0.24710245430469513, "learning_rate": 4.961272994159718e-05, "loss": 0.0799, "step": 7613 }, { "epoch": 0.32101524126736514, "grad_norm": 0.4836794137954712, "learning_rate": 4.961194915518911e-05, "loss": 0.1795, "step": 7614 }, { "epoch": 0.3210574024495647, "grad_norm": 0.7431066632270813, "learning_rate": 4.961116836878104e-05, "loss": 0.2533, "step": 7615 }, { "epoch": 0.32109956363176423, "grad_norm": 0.7434492707252502, "learning_rate": 4.961038758237297e-05, "loss": 0.758, "step": 7616 }, { "epoch": 0.3211417248139638, "grad_norm": 1.348528265953064, "learning_rate": 4.9609606795964894e-05, "loss": 0.1286, "step": 7617 }, { "epoch": 0.3211838859961633, "grad_norm": 0.4608150124549866, "learning_rate": 4.960882600955683e-05, "loss": 0.1929, "step": 7618 }, { "epoch": 0.3212260471783629, "grad_norm": 0.9359932541847229, "learning_rate": 4.9608045223148756e-05, "loss": 0.2881, "step": 7619 }, { "epoch": 0.3212682083605624, "grad_norm": 0.5022112727165222, "learning_rate": 4.960726443674069e-05, "loss": 0.294, "step": 7620 }, { "epoch": 0.321310369542762, "grad_norm": 0.7698765397071838, "learning_rate": 4.960648365033262e-05, "loss": 0.5921, "step": 7621 }, { "epoch": 0.3213525307249615, "grad_norm": 0.410596638917923, "learning_rate": 4.960570286392455e-05, "loss": 0.2094, "step": 7622 }, { "epoch": 0.3213946919071611, "grad_norm": 0.7287007570266724, "learning_rate": 4.9604922077516474e-05, "loss": 0.1518, "step": 7623 }, { "epoch": 0.32143685308936065, "grad_norm": 1.1832835674285889, "learning_rate": 4.9604141291108406e-05, "loss": 1.0184, "step": 7624 }, { "epoch": 0.32147901427156017, "grad_norm": 0.37950196862220764, "learning_rate": 4.960336050470034e-05, "loss": 0.1238, "step": 7625 }, { "epoch": 0.32152117545375974, "grad_norm": 0.8182227611541748, "learning_rate": 4.960257971829226e-05, "loss": 0.1914, "step": 7626 }, { "epoch": 0.32156333663595926, "grad_norm": 0.32769837975502014, "learning_rate": 4.96017989318842e-05, "loss": 0.061, "step": 7627 }, { "epoch": 0.32160549781815884, "grad_norm": 1.0574233531951904, "learning_rate": 4.9601018145476124e-05, "loss": 1.0766, "step": 7628 }, { "epoch": 0.32164765900035835, "grad_norm": 0.6348622441291809, "learning_rate": 4.9600237359068055e-05, "loss": 0.3171, "step": 7629 }, { "epoch": 0.32168982018255793, "grad_norm": 0.8782979846000671, "learning_rate": 4.9599456572659987e-05, "loss": 0.4082, "step": 7630 }, { "epoch": 0.32173198136475745, "grad_norm": 1.6481122970581055, "learning_rate": 4.959867578625191e-05, "loss": 0.4716, "step": 7631 }, { "epoch": 0.321774142546957, "grad_norm": 0.5602027177810669, "learning_rate": 4.959789499984385e-05, "loss": 0.235, "step": 7632 }, { "epoch": 0.32181630372915654, "grad_norm": 0.6471146941184998, "learning_rate": 4.9597114213435774e-05, "loss": 0.5721, "step": 7633 }, { "epoch": 0.3218584649113561, "grad_norm": 0.9993993639945984, "learning_rate": 4.9596333427027705e-05, "loss": 0.8313, "step": 7634 }, { "epoch": 0.3219006260935557, "grad_norm": 0.8422398567199707, "learning_rate": 4.9595552640619636e-05, "loss": 0.7122, "step": 7635 }, { "epoch": 0.3219427872757552, "grad_norm": 0.539405882358551, "learning_rate": 4.959477185421156e-05, "loss": 0.2181, "step": 7636 }, { "epoch": 0.3219849484579548, "grad_norm": 0.575857937335968, "learning_rate": 4.959399106780349e-05, "loss": 0.1919, "step": 7637 }, { "epoch": 0.3220271096401543, "grad_norm": 0.942453920841217, "learning_rate": 4.959321028139542e-05, "loss": 0.2341, "step": 7638 }, { "epoch": 0.32206927082235387, "grad_norm": 0.6575480699539185, "learning_rate": 4.9592429494987354e-05, "loss": 0.1627, "step": 7639 }, { "epoch": 0.3221114320045534, "grad_norm": 0.477560818195343, "learning_rate": 4.9591648708579286e-05, "loss": 0.1682, "step": 7640 }, { "epoch": 0.32215359318675296, "grad_norm": 0.677237868309021, "learning_rate": 4.959086792217122e-05, "loss": 0.6537, "step": 7641 }, { "epoch": 0.3221957543689525, "grad_norm": 0.7888278365135193, "learning_rate": 4.959008713576314e-05, "loss": 0.4515, "step": 7642 }, { "epoch": 0.32223791555115205, "grad_norm": 0.7383669018745422, "learning_rate": 4.958930634935507e-05, "loss": 0.174, "step": 7643 }, { "epoch": 0.32228007673335163, "grad_norm": 0.8684551119804382, "learning_rate": 4.9588525562947004e-05, "loss": 0.2181, "step": 7644 }, { "epoch": 0.32232223791555115, "grad_norm": 1.0142192840576172, "learning_rate": 4.958774477653893e-05, "loss": 1.0247, "step": 7645 }, { "epoch": 0.3223643990977507, "grad_norm": 0.3417469561100006, "learning_rate": 4.9586963990130866e-05, "loss": 0.0692, "step": 7646 }, { "epoch": 0.32240656027995024, "grad_norm": 0.4535544812679291, "learning_rate": 4.958618320372279e-05, "loss": 0.1206, "step": 7647 }, { "epoch": 0.3224487214621498, "grad_norm": 0.6152970194816589, "learning_rate": 4.958540241731472e-05, "loss": 0.1586, "step": 7648 }, { "epoch": 0.32249088264434933, "grad_norm": 0.6185925006866455, "learning_rate": 4.958462163090665e-05, "loss": 0.4437, "step": 7649 }, { "epoch": 0.3225330438265489, "grad_norm": 1.2132344245910645, "learning_rate": 4.958384084449858e-05, "loss": 0.96, "step": 7650 }, { "epoch": 0.3225752050087484, "grad_norm": 0.755391538143158, "learning_rate": 4.9583060058090516e-05, "loss": 0.6126, "step": 7651 }, { "epoch": 0.322617366190948, "grad_norm": 0.5101978182792664, "learning_rate": 4.958227927168244e-05, "loss": 0.4588, "step": 7652 }, { "epoch": 0.32265952737314757, "grad_norm": 0.5262719988822937, "learning_rate": 4.958149848527437e-05, "loss": 0.2061, "step": 7653 }, { "epoch": 0.3227016885553471, "grad_norm": 0.5412322878837585, "learning_rate": 4.95807176988663e-05, "loss": 0.4955, "step": 7654 }, { "epoch": 0.32274384973754666, "grad_norm": 0.2648876905441284, "learning_rate": 4.957993691245823e-05, "loss": 0.0674, "step": 7655 }, { "epoch": 0.3227860109197462, "grad_norm": 0.5592384934425354, "learning_rate": 4.957915612605016e-05, "loss": 0.1434, "step": 7656 }, { "epoch": 0.32282817210194575, "grad_norm": 3.3663218021392822, "learning_rate": 4.957837533964209e-05, "loss": 0.1768, "step": 7657 }, { "epoch": 0.32287033328414527, "grad_norm": 2.6898393630981445, "learning_rate": 4.957759455323402e-05, "loss": 0.586, "step": 7658 }, { "epoch": 0.32291249446634485, "grad_norm": 1.0616713762283325, "learning_rate": 4.9576813766825946e-05, "loss": 0.9644, "step": 7659 }, { "epoch": 0.32295465564854436, "grad_norm": 0.3976234793663025, "learning_rate": 4.957603298041788e-05, "loss": 0.1057, "step": 7660 }, { "epoch": 0.32299681683074394, "grad_norm": 0.5992975831031799, "learning_rate": 4.957525219400981e-05, "loss": 0.3174, "step": 7661 }, { "epoch": 0.32303897801294346, "grad_norm": 0.4102523624897003, "learning_rate": 4.957447140760174e-05, "loss": 0.1019, "step": 7662 }, { "epoch": 0.32308113919514303, "grad_norm": 0.61186283826828, "learning_rate": 4.957369062119367e-05, "loss": 0.185, "step": 7663 }, { "epoch": 0.3231233003773426, "grad_norm": 0.41047409176826477, "learning_rate": 4.9572909834785595e-05, "loss": 0.1199, "step": 7664 }, { "epoch": 0.3231654615595421, "grad_norm": 1.5494533777236938, "learning_rate": 4.957212904837753e-05, "loss": 0.3914, "step": 7665 }, { "epoch": 0.3232076227417417, "grad_norm": 8.055904388427734, "learning_rate": 4.957134826196946e-05, "loss": 0.1479, "step": 7666 }, { "epoch": 0.3232497839239412, "grad_norm": 1.9449515342712402, "learning_rate": 4.957056747556139e-05, "loss": 1.1141, "step": 7667 }, { "epoch": 0.3232919451061408, "grad_norm": 0.805310845375061, "learning_rate": 4.956978668915332e-05, "loss": 0.2326, "step": 7668 }, { "epoch": 0.3233341062883403, "grad_norm": 0.6370805501937866, "learning_rate": 4.9569005902745245e-05, "loss": 0.3756, "step": 7669 }, { "epoch": 0.3233762674705399, "grad_norm": 0.6262344717979431, "learning_rate": 4.9568225116337176e-05, "loss": 0.4914, "step": 7670 }, { "epoch": 0.3234184286527394, "grad_norm": 1.1675755977630615, "learning_rate": 4.956744432992911e-05, "loss": 0.2825, "step": 7671 }, { "epoch": 0.32346058983493897, "grad_norm": 0.8306676745414734, "learning_rate": 4.956666354352104e-05, "loss": 0.2042, "step": 7672 }, { "epoch": 0.32350275101713855, "grad_norm": 0.6251800656318665, "learning_rate": 4.956588275711297e-05, "loss": 0.1593, "step": 7673 }, { "epoch": 0.32354491219933806, "grad_norm": 0.5277538299560547, "learning_rate": 4.9565101970704894e-05, "loss": 0.1798, "step": 7674 }, { "epoch": 0.32358707338153764, "grad_norm": 1.3981246948242188, "learning_rate": 4.9564321184296825e-05, "loss": 0.9724, "step": 7675 }, { "epoch": 0.32362923456373716, "grad_norm": 0.8527978658676147, "learning_rate": 4.956354039788876e-05, "loss": 0.1349, "step": 7676 }, { "epoch": 0.32367139574593673, "grad_norm": 0.5486540198326111, "learning_rate": 4.956275961148069e-05, "loss": 0.442, "step": 7677 }, { "epoch": 0.32371355692813625, "grad_norm": 0.28365427255630493, "learning_rate": 4.956197882507261e-05, "loss": 0.1058, "step": 7678 }, { "epoch": 0.3237557181103358, "grad_norm": 0.4608331620693207, "learning_rate": 4.9561198038664544e-05, "loss": 0.1162, "step": 7679 }, { "epoch": 0.32379787929253534, "grad_norm": 0.5101648569107056, "learning_rate": 4.9560417252256475e-05, "loss": 0.2698, "step": 7680 }, { "epoch": 0.3238400404747349, "grad_norm": 0.44894346594810486, "learning_rate": 4.95596364658484e-05, "loss": 0.1418, "step": 7681 }, { "epoch": 0.3238822016569345, "grad_norm": 1.1822582483291626, "learning_rate": 4.955885567944034e-05, "loss": 0.4861, "step": 7682 }, { "epoch": 0.323924362839134, "grad_norm": 0.738073468208313, "learning_rate": 4.955807489303226e-05, "loss": 0.4598, "step": 7683 }, { "epoch": 0.3239665240213336, "grad_norm": 0.4588068425655365, "learning_rate": 4.95572941066242e-05, "loss": 0.1552, "step": 7684 }, { "epoch": 0.3240086852035331, "grad_norm": 0.6337651014328003, "learning_rate": 4.9556513320216125e-05, "loss": 0.11, "step": 7685 }, { "epoch": 0.32405084638573267, "grad_norm": 0.7184485197067261, "learning_rate": 4.9555732533808056e-05, "loss": 0.0479, "step": 7686 }, { "epoch": 0.3240930075679322, "grad_norm": 0.2936760187149048, "learning_rate": 4.955495174739999e-05, "loss": 0.0974, "step": 7687 }, { "epoch": 0.32413516875013176, "grad_norm": 0.5261553525924683, "learning_rate": 4.955417096099191e-05, "loss": 0.5688, "step": 7688 }, { "epoch": 0.3241773299323313, "grad_norm": 0.5327452421188354, "learning_rate": 4.955339017458384e-05, "loss": 0.2346, "step": 7689 }, { "epoch": 0.32421949111453086, "grad_norm": 0.5477545857429504, "learning_rate": 4.9552609388175774e-05, "loss": 0.1354, "step": 7690 }, { "epoch": 0.3242616522967304, "grad_norm": 0.837405264377594, "learning_rate": 4.9551828601767705e-05, "loss": 0.4799, "step": 7691 }, { "epoch": 0.32430381347892995, "grad_norm": 1.4849594831466675, "learning_rate": 4.955104781535963e-05, "loss": 0.926, "step": 7692 }, { "epoch": 0.3243459746611295, "grad_norm": 0.7022086381912231, "learning_rate": 4.955026702895156e-05, "loss": 0.4134, "step": 7693 }, { "epoch": 0.32438813584332904, "grad_norm": 0.6589395999908447, "learning_rate": 4.954948624254349e-05, "loss": 0.5004, "step": 7694 }, { "epoch": 0.3244302970255286, "grad_norm": 0.4774872660636902, "learning_rate": 4.9548705456135424e-05, "loss": 0.2416, "step": 7695 }, { "epoch": 0.32447245820772813, "grad_norm": 0.5801798105239868, "learning_rate": 4.9547924669727355e-05, "loss": 0.436, "step": 7696 }, { "epoch": 0.3245146193899277, "grad_norm": 2.8217153549194336, "learning_rate": 4.954714388331928e-05, "loss": 0.1471, "step": 7697 }, { "epoch": 0.3245567805721272, "grad_norm": 1.448285460472107, "learning_rate": 4.954636309691121e-05, "loss": 0.4893, "step": 7698 }, { "epoch": 0.3245989417543268, "grad_norm": 1.2561885118484497, "learning_rate": 4.954558231050314e-05, "loss": 0.349, "step": 7699 }, { "epoch": 0.3246411029365263, "grad_norm": 0.5029399394989014, "learning_rate": 4.9544801524095066e-05, "loss": 0.2459, "step": 7700 }, { "epoch": 0.3246832641187259, "grad_norm": 1.0767314434051514, "learning_rate": 4.9544020737687004e-05, "loss": 0.4837, "step": 7701 }, { "epoch": 0.32472542530092546, "grad_norm": 1.1733858585357666, "learning_rate": 4.954323995127893e-05, "loss": 0.1562, "step": 7702 }, { "epoch": 0.324767586483125, "grad_norm": 0.3373642563819885, "learning_rate": 4.954245916487086e-05, "loss": 0.0667, "step": 7703 }, { "epoch": 0.32480974766532456, "grad_norm": 3.768956422805786, "learning_rate": 4.954167837846279e-05, "loss": 1.0545, "step": 7704 }, { "epoch": 0.3248519088475241, "grad_norm": 1.6635867357254028, "learning_rate": 4.9540897592054716e-05, "loss": 0.8529, "step": 7705 }, { "epoch": 0.32489407002972365, "grad_norm": 0.4086071848869324, "learning_rate": 4.954011680564665e-05, "loss": 0.1544, "step": 7706 }, { "epoch": 0.32493623121192317, "grad_norm": 0.5988672375679016, "learning_rate": 4.953933601923858e-05, "loss": 0.1795, "step": 7707 }, { "epoch": 0.32497839239412274, "grad_norm": 1.3378254175186157, "learning_rate": 4.953855523283051e-05, "loss": 0.9399, "step": 7708 }, { "epoch": 0.32502055357632226, "grad_norm": 0.5788483619689941, "learning_rate": 4.953777444642244e-05, "loss": 0.1421, "step": 7709 }, { "epoch": 0.32506271475852183, "grad_norm": 0.5211089253425598, "learning_rate": 4.953699366001437e-05, "loss": 0.2969, "step": 7710 }, { "epoch": 0.3251048759407214, "grad_norm": 0.6629891395568848, "learning_rate": 4.9536212873606297e-05, "loss": 0.4831, "step": 7711 }, { "epoch": 0.3251470371229209, "grad_norm": 0.8659355044364929, "learning_rate": 4.953543208719823e-05, "loss": 0.6017, "step": 7712 }, { "epoch": 0.3251891983051205, "grad_norm": 0.7649155855178833, "learning_rate": 4.953465130079016e-05, "loss": 0.4108, "step": 7713 }, { "epoch": 0.32523135948732, "grad_norm": 0.4482557773590088, "learning_rate": 4.9533870514382084e-05, "loss": 0.0853, "step": 7714 }, { "epoch": 0.3252735206695196, "grad_norm": 0.5533617734909058, "learning_rate": 4.953308972797402e-05, "loss": 0.1427, "step": 7715 }, { "epoch": 0.3253156818517191, "grad_norm": 0.8548253774642944, "learning_rate": 4.9532308941565946e-05, "loss": 0.168, "step": 7716 }, { "epoch": 0.3253578430339187, "grad_norm": 0.43968045711517334, "learning_rate": 4.953152815515788e-05, "loss": 0.1191, "step": 7717 }, { "epoch": 0.3254000042161182, "grad_norm": 1.2020554542541504, "learning_rate": 4.953074736874981e-05, "loss": 0.1642, "step": 7718 }, { "epoch": 0.3254421653983178, "grad_norm": 0.6536654233932495, "learning_rate": 4.952996658234173e-05, "loss": 0.1662, "step": 7719 }, { "epoch": 0.3254843265805173, "grad_norm": 1.0045217275619507, "learning_rate": 4.952918579593367e-05, "loss": 0.1643, "step": 7720 }, { "epoch": 0.32552648776271687, "grad_norm": 10.117094039916992, "learning_rate": 4.9528405009525596e-05, "loss": 0.1969, "step": 7721 }, { "epoch": 0.32556864894491644, "grad_norm": 0.45755502581596375, "learning_rate": 4.952762422311753e-05, "loss": 0.1535, "step": 7722 }, { "epoch": 0.32561081012711596, "grad_norm": 2.4261634349823, "learning_rate": 4.952684343670946e-05, "loss": 1.1602, "step": 7723 }, { "epoch": 0.32565297130931553, "grad_norm": 0.4339160621166229, "learning_rate": 4.952606265030138e-05, "loss": 0.1828, "step": 7724 }, { "epoch": 0.32569513249151505, "grad_norm": 1.4835774898529053, "learning_rate": 4.9525281863893314e-05, "loss": 0.4843, "step": 7725 }, { "epoch": 0.3257372936737146, "grad_norm": 0.9609315991401672, "learning_rate": 4.9524501077485245e-05, "loss": 0.66, "step": 7726 }, { "epoch": 0.32577945485591414, "grad_norm": 0.9835946559906006, "learning_rate": 4.9523720291077176e-05, "loss": 0.1977, "step": 7727 }, { "epoch": 0.3258216160381137, "grad_norm": 1.0541512966156006, "learning_rate": 4.95229395046691e-05, "loss": 0.2366, "step": 7728 }, { "epoch": 0.32586377722031323, "grad_norm": 1.4405897855758667, "learning_rate": 4.952215871826104e-05, "loss": 1.1439, "step": 7729 }, { "epoch": 0.3259059384025128, "grad_norm": 0.4212966561317444, "learning_rate": 4.9521377931852963e-05, "loss": 0.1072, "step": 7730 }, { "epoch": 0.3259480995847124, "grad_norm": 0.8631520867347717, "learning_rate": 4.9520597145444895e-05, "loss": 0.2235, "step": 7731 }, { "epoch": 0.3259902607669119, "grad_norm": 0.4014129340648651, "learning_rate": 4.9519816359036826e-05, "loss": 0.1602, "step": 7732 }, { "epoch": 0.3260324219491115, "grad_norm": 0.5103707909584045, "learning_rate": 4.951903557262875e-05, "loss": 0.132, "step": 7733 }, { "epoch": 0.326074583131311, "grad_norm": 0.8724700808525085, "learning_rate": 4.951825478622069e-05, "loss": 0.782, "step": 7734 }, { "epoch": 0.32611674431351056, "grad_norm": 0.7988268733024597, "learning_rate": 4.951747399981261e-05, "loss": 0.4301, "step": 7735 }, { "epoch": 0.3261589054957101, "grad_norm": 0.782303512096405, "learning_rate": 4.9516693213404544e-05, "loss": 0.5813, "step": 7736 }, { "epoch": 0.32620106667790966, "grad_norm": 0.6851363778114319, "learning_rate": 4.9515912426996475e-05, "loss": 0.4229, "step": 7737 }, { "epoch": 0.3262432278601092, "grad_norm": 0.8912496566772461, "learning_rate": 4.95151316405884e-05, "loss": 0.2103, "step": 7738 }, { "epoch": 0.32628538904230875, "grad_norm": 0.7187910079956055, "learning_rate": 4.951435085418033e-05, "loss": 0.0796, "step": 7739 }, { "epoch": 0.3263275502245083, "grad_norm": 0.5624848008155823, "learning_rate": 4.951357006777226e-05, "loss": 0.2361, "step": 7740 }, { "epoch": 0.32636971140670784, "grad_norm": 7.332226753234863, "learning_rate": 4.9512789281364194e-05, "loss": 0.9981, "step": 7741 }, { "epoch": 0.3264118725889074, "grad_norm": 0.637561559677124, "learning_rate": 4.9512008494956125e-05, "loss": 0.1115, "step": 7742 }, { "epoch": 0.32645403377110693, "grad_norm": 1.1404297351837158, "learning_rate": 4.951122770854805e-05, "loss": 1.0626, "step": 7743 }, { "epoch": 0.3264961949533065, "grad_norm": 1.5341506004333496, "learning_rate": 4.951044692213998e-05, "loss": 0.528, "step": 7744 }, { "epoch": 0.326538356135506, "grad_norm": 1.3633365631103516, "learning_rate": 4.950966613573191e-05, "loss": 0.1745, "step": 7745 }, { "epoch": 0.3265805173177056, "grad_norm": 0.7971922159194946, "learning_rate": 4.950888534932384e-05, "loss": 0.2328, "step": 7746 }, { "epoch": 0.3266226784999051, "grad_norm": 0.6481251120567322, "learning_rate": 4.950810456291577e-05, "loss": 0.6259, "step": 7747 }, { "epoch": 0.3266648396821047, "grad_norm": 0.7923621535301208, "learning_rate": 4.9507323776507706e-05, "loss": 0.2071, "step": 7748 }, { "epoch": 0.3267070008643042, "grad_norm": 0.6200310587882996, "learning_rate": 4.950654299009963e-05, "loss": 0.1488, "step": 7749 }, { "epoch": 0.3267491620465038, "grad_norm": 0.3205885589122772, "learning_rate": 4.9505762203691555e-05, "loss": 0.0762, "step": 7750 }, { "epoch": 0.32679132322870336, "grad_norm": 1.1773779392242432, "learning_rate": 4.950498141728349e-05, "loss": 0.1488, "step": 7751 }, { "epoch": 0.3268334844109029, "grad_norm": 0.6881730556488037, "learning_rate": 4.950420063087542e-05, "loss": 0.3204, "step": 7752 }, { "epoch": 0.32687564559310245, "grad_norm": 0.8600224852561951, "learning_rate": 4.9503419844467355e-05, "loss": 0.7083, "step": 7753 }, { "epoch": 0.32691780677530197, "grad_norm": 0.33727002143859863, "learning_rate": 4.950263905805928e-05, "loss": 0.1387, "step": 7754 }, { "epoch": 0.32695996795750154, "grad_norm": 0.5828532576560974, "learning_rate": 4.950185827165121e-05, "loss": 0.1566, "step": 7755 }, { "epoch": 0.32700212913970106, "grad_norm": 0.4685550928115845, "learning_rate": 4.950107748524314e-05, "loss": 0.2242, "step": 7756 }, { "epoch": 0.32704429032190063, "grad_norm": 1.1252423524856567, "learning_rate": 4.950029669883507e-05, "loss": 0.2992, "step": 7757 }, { "epoch": 0.32708645150410015, "grad_norm": 0.5504000186920166, "learning_rate": 4.9499515912427e-05, "loss": 0.3723, "step": 7758 }, { "epoch": 0.3271286126862997, "grad_norm": 0.9689599871635437, "learning_rate": 4.949873512601893e-05, "loss": 0.1608, "step": 7759 }, { "epoch": 0.3271707738684993, "grad_norm": 1.5899749994277954, "learning_rate": 4.949795433961086e-05, "loss": 0.6735, "step": 7760 }, { "epoch": 0.3272129350506988, "grad_norm": 0.4682827889919281, "learning_rate": 4.9497173553202785e-05, "loss": 0.2141, "step": 7761 }, { "epoch": 0.3272550962328984, "grad_norm": 0.8535161018371582, "learning_rate": 4.9496392766794716e-05, "loss": 0.2693, "step": 7762 }, { "epoch": 0.3272972574150979, "grad_norm": 0.6750187873840332, "learning_rate": 4.949561198038665e-05, "loss": 0.2413, "step": 7763 }, { "epoch": 0.3273394185972975, "grad_norm": 0.7130934596061707, "learning_rate": 4.949483119397858e-05, "loss": 0.1241, "step": 7764 }, { "epoch": 0.327381579779497, "grad_norm": 0.35387420654296875, "learning_rate": 4.949405040757051e-05, "loss": 0.0895, "step": 7765 }, { "epoch": 0.3274237409616966, "grad_norm": 0.6790843605995178, "learning_rate": 4.9493269621162435e-05, "loss": 0.271, "step": 7766 }, { "epoch": 0.3274659021438961, "grad_norm": 0.7081647515296936, "learning_rate": 4.9492488834754366e-05, "loss": 0.1609, "step": 7767 }, { "epoch": 0.32750806332609567, "grad_norm": 0.48549360036849976, "learning_rate": 4.94917080483463e-05, "loss": 0.2121, "step": 7768 }, { "epoch": 0.32755022450829524, "grad_norm": 1.033087968826294, "learning_rate": 4.949092726193822e-05, "loss": 0.2749, "step": 7769 }, { "epoch": 0.32759238569049476, "grad_norm": 0.611421525478363, "learning_rate": 4.949014647553016e-05, "loss": 0.6008, "step": 7770 }, { "epoch": 0.32763454687269433, "grad_norm": 1.0020118951797485, "learning_rate": 4.9489365689122084e-05, "loss": 0.1185, "step": 7771 }, { "epoch": 0.32767670805489385, "grad_norm": 0.9948189854621887, "learning_rate": 4.9488584902714015e-05, "loss": 1.1526, "step": 7772 }, { "epoch": 0.3277188692370934, "grad_norm": 0.37761324644088745, "learning_rate": 4.948780411630595e-05, "loss": 0.1483, "step": 7773 }, { "epoch": 0.32776103041929294, "grad_norm": 2.4553656578063965, "learning_rate": 4.948702332989788e-05, "loss": 0.6213, "step": 7774 }, { "epoch": 0.3278031916014925, "grad_norm": 1.0195423364639282, "learning_rate": 4.948624254348981e-05, "loss": 0.8376, "step": 7775 }, { "epoch": 0.32784535278369203, "grad_norm": 11.742402076721191, "learning_rate": 4.9485461757081734e-05, "loss": 0.594, "step": 7776 }, { "epoch": 0.3278875139658916, "grad_norm": 0.5068666934967041, "learning_rate": 4.9484680970673665e-05, "loss": 0.1075, "step": 7777 }, { "epoch": 0.3279296751480911, "grad_norm": 0.3321482241153717, "learning_rate": 4.9483900184265596e-05, "loss": 0.0892, "step": 7778 }, { "epoch": 0.3279718363302907, "grad_norm": 0.36461925506591797, "learning_rate": 4.948311939785753e-05, "loss": 0.1607, "step": 7779 }, { "epoch": 0.3280139975124903, "grad_norm": 0.5331071615219116, "learning_rate": 4.948233861144945e-05, "loss": 0.4274, "step": 7780 }, { "epoch": 0.3280561586946898, "grad_norm": 0.9388265013694763, "learning_rate": 4.948155782504138e-05, "loss": 0.1716, "step": 7781 }, { "epoch": 0.32809831987688937, "grad_norm": 0.6089946627616882, "learning_rate": 4.9480777038633314e-05, "loss": 0.4521, "step": 7782 }, { "epoch": 0.3281404810590889, "grad_norm": 0.7457163333892822, "learning_rate": 4.947999625222524e-05, "loss": 0.2678, "step": 7783 }, { "epoch": 0.32818264224128846, "grad_norm": 0.9190959930419922, "learning_rate": 4.947921546581718e-05, "loss": 1.1648, "step": 7784 }, { "epoch": 0.328224803423488, "grad_norm": 0.3974487781524658, "learning_rate": 4.94784346794091e-05, "loss": 0.157, "step": 7785 }, { "epoch": 0.32826696460568755, "grad_norm": 0.8199060559272766, "learning_rate": 4.947765389300103e-05, "loss": 0.4256, "step": 7786 }, { "epoch": 0.32830912578788707, "grad_norm": 3.0052101612091064, "learning_rate": 4.9476873106592964e-05, "loss": 0.2793, "step": 7787 }, { "epoch": 0.32835128697008664, "grad_norm": 0.753282904624939, "learning_rate": 4.947609232018489e-05, "loss": 0.5024, "step": 7788 }, { "epoch": 0.3283934481522862, "grad_norm": 0.35263511538505554, "learning_rate": 4.9475311533776826e-05, "loss": 0.0873, "step": 7789 }, { "epoch": 0.32843560933448573, "grad_norm": 0.46755504608154297, "learning_rate": 4.947453074736875e-05, "loss": 0.1733, "step": 7790 }, { "epoch": 0.3284777705166853, "grad_norm": 0.6195061206817627, "learning_rate": 4.947374996096068e-05, "loss": 0.4791, "step": 7791 }, { "epoch": 0.3285199316988848, "grad_norm": 2.3212203979492188, "learning_rate": 4.9472969174552613e-05, "loss": 0.3489, "step": 7792 }, { "epoch": 0.3285620928810844, "grad_norm": 1.1213715076446533, "learning_rate": 4.9472188388144545e-05, "loss": 1.1073, "step": 7793 }, { "epoch": 0.3286042540632839, "grad_norm": 1.661954641342163, "learning_rate": 4.947140760173647e-05, "loss": 0.7431, "step": 7794 }, { "epoch": 0.3286464152454835, "grad_norm": 1.1441097259521484, "learning_rate": 4.94706268153284e-05, "loss": 0.3632, "step": 7795 }, { "epoch": 0.328688576427683, "grad_norm": 0.742641806602478, "learning_rate": 4.946984602892033e-05, "loss": 0.2241, "step": 7796 }, { "epoch": 0.3287307376098826, "grad_norm": 0.7801170945167542, "learning_rate": 4.9469065242512256e-05, "loss": 0.3746, "step": 7797 }, { "epoch": 0.32877289879208216, "grad_norm": 0.6143150329589844, "learning_rate": 4.9468284456104194e-05, "loss": 0.6477, "step": 7798 }, { "epoch": 0.3288150599742817, "grad_norm": 0.7976833581924438, "learning_rate": 4.946750366969612e-05, "loss": 0.3103, "step": 7799 }, { "epoch": 0.32885722115648125, "grad_norm": 2.343820571899414, "learning_rate": 4.946672288328805e-05, "loss": 0.64, "step": 7800 }, { "epoch": 0.32889938233868077, "grad_norm": 0.6980390548706055, "learning_rate": 4.946594209687998e-05, "loss": 0.7348, "step": 7801 }, { "epoch": 0.32894154352088034, "grad_norm": 0.5874543190002441, "learning_rate": 4.9465161310471906e-05, "loss": 0.384, "step": 7802 }, { "epoch": 0.32898370470307986, "grad_norm": 0.4716785252094269, "learning_rate": 4.9464380524063844e-05, "loss": 0.1018, "step": 7803 }, { "epoch": 0.32902586588527943, "grad_norm": 0.6887405514717102, "learning_rate": 4.946359973765577e-05, "loss": 0.4528, "step": 7804 }, { "epoch": 0.32906802706747895, "grad_norm": 0.3517626225948334, "learning_rate": 4.94628189512477e-05, "loss": 0.0914, "step": 7805 }, { "epoch": 0.3291101882496785, "grad_norm": 0.4899471700191498, "learning_rate": 4.946203816483963e-05, "loss": 0.1483, "step": 7806 }, { "epoch": 0.32915234943187804, "grad_norm": 2.198338270187378, "learning_rate": 4.9461257378431555e-05, "loss": 0.3353, "step": 7807 }, { "epoch": 0.3291945106140776, "grad_norm": 0.6254680752754211, "learning_rate": 4.9460476592023487e-05, "loss": 0.4915, "step": 7808 }, { "epoch": 0.3292366717962772, "grad_norm": 0.8869178891181946, "learning_rate": 4.945969580561542e-05, "loss": 0.1273, "step": 7809 }, { "epoch": 0.3292788329784767, "grad_norm": 1.078611135482788, "learning_rate": 4.945891501920735e-05, "loss": 0.5957, "step": 7810 }, { "epoch": 0.3293209941606763, "grad_norm": 0.43248486518859863, "learning_rate": 4.945813423279928e-05, "loss": 0.1188, "step": 7811 }, { "epoch": 0.3293631553428758, "grad_norm": 0.7707790732383728, "learning_rate": 4.9457353446391205e-05, "loss": 0.3878, "step": 7812 }, { "epoch": 0.3294053165250754, "grad_norm": 0.5990008115768433, "learning_rate": 4.9456572659983136e-05, "loss": 0.8386, "step": 7813 }, { "epoch": 0.3294474777072749, "grad_norm": 0.5918191075325012, "learning_rate": 4.945579187357507e-05, "loss": 0.3269, "step": 7814 }, { "epoch": 0.32948963888947447, "grad_norm": 0.326090008020401, "learning_rate": 4.9455011087167e-05, "loss": 0.12, "step": 7815 }, { "epoch": 0.329531800071674, "grad_norm": 0.8728682994842529, "learning_rate": 4.945423030075892e-05, "loss": 0.3076, "step": 7816 }, { "epoch": 0.32957396125387356, "grad_norm": 0.7142723798751831, "learning_rate": 4.945344951435086e-05, "loss": 0.3377, "step": 7817 }, { "epoch": 0.32961612243607313, "grad_norm": 1.0129585266113281, "learning_rate": 4.9452668727942786e-05, "loss": 0.2324, "step": 7818 }, { "epoch": 0.32965828361827265, "grad_norm": 0.8852213025093079, "learning_rate": 4.945188794153472e-05, "loss": 0.3722, "step": 7819 }, { "epoch": 0.3297004448004722, "grad_norm": 0.704423189163208, "learning_rate": 4.945110715512665e-05, "loss": 0.5734, "step": 7820 }, { "epoch": 0.32974260598267174, "grad_norm": 1.224678874015808, "learning_rate": 4.945032636871857e-05, "loss": 0.5865, "step": 7821 }, { "epoch": 0.3297847671648713, "grad_norm": 0.7739362716674805, "learning_rate": 4.944954558231051e-05, "loss": 0.1661, "step": 7822 }, { "epoch": 0.32982692834707084, "grad_norm": 0.4935155510902405, "learning_rate": 4.9448764795902435e-05, "loss": 0.1282, "step": 7823 }, { "epoch": 0.3298690895292704, "grad_norm": 1.5703606605529785, "learning_rate": 4.9447984009494366e-05, "loss": 0.3181, "step": 7824 }, { "epoch": 0.3299112507114699, "grad_norm": 0.5240760445594788, "learning_rate": 4.94472032230863e-05, "loss": 0.199, "step": 7825 }, { "epoch": 0.3299534118936695, "grad_norm": 0.6262580156326294, "learning_rate": 4.944642243667822e-05, "loss": 0.2213, "step": 7826 }, { "epoch": 0.329995573075869, "grad_norm": 1.0094940662384033, "learning_rate": 4.944564165027015e-05, "loss": 0.1656, "step": 7827 }, { "epoch": 0.3300377342580686, "grad_norm": 1.5497225522994995, "learning_rate": 4.9444860863862085e-05, "loss": 0.1499, "step": 7828 }, { "epoch": 0.33007989544026817, "grad_norm": 0.6628398895263672, "learning_rate": 4.9444080077454016e-05, "loss": 0.2712, "step": 7829 }, { "epoch": 0.3301220566224677, "grad_norm": 0.7611330151557922, "learning_rate": 4.944329929104594e-05, "loss": 0.3443, "step": 7830 }, { "epoch": 0.33016421780466726, "grad_norm": 0.9376664757728577, "learning_rate": 4.944251850463787e-05, "loss": 0.3002, "step": 7831 }, { "epoch": 0.3302063789868668, "grad_norm": 13.13972282409668, "learning_rate": 4.94417377182298e-05, "loss": 0.5733, "step": 7832 }, { "epoch": 0.33024854016906635, "grad_norm": 0.396177738904953, "learning_rate": 4.9440956931821734e-05, "loss": 0.1772, "step": 7833 }, { "epoch": 0.33029070135126587, "grad_norm": 24.953044891357422, "learning_rate": 4.9440176145413665e-05, "loss": 0.4582, "step": 7834 }, { "epoch": 0.33033286253346544, "grad_norm": 0.3548722267150879, "learning_rate": 4.943939535900559e-05, "loss": 0.2327, "step": 7835 }, { "epoch": 0.33037502371566496, "grad_norm": 1.3218175172805786, "learning_rate": 4.943861457259753e-05, "loss": 0.3732, "step": 7836 }, { "epoch": 0.33041718489786454, "grad_norm": 1.214735507965088, "learning_rate": 4.943783378618945e-05, "loss": 0.4381, "step": 7837 }, { "epoch": 0.3304593460800641, "grad_norm": 0.48727190494537354, "learning_rate": 4.9437052999781384e-05, "loss": 0.1583, "step": 7838 }, { "epoch": 0.3305015072622636, "grad_norm": 0.40766045451164246, "learning_rate": 4.9436272213373315e-05, "loss": 0.0847, "step": 7839 }, { "epoch": 0.3305436684444632, "grad_norm": 0.755052387714386, "learning_rate": 4.943549142696524e-05, "loss": 0.362, "step": 7840 }, { "epoch": 0.3305858296266627, "grad_norm": 1.0171749591827393, "learning_rate": 4.943471064055717e-05, "loss": 1.0228, "step": 7841 }, { "epoch": 0.3306279908088623, "grad_norm": 0.6687262058258057, "learning_rate": 4.94339298541491e-05, "loss": 0.0896, "step": 7842 }, { "epoch": 0.3306701519910618, "grad_norm": 4.41825008392334, "learning_rate": 4.943314906774103e-05, "loss": 1.01, "step": 7843 }, { "epoch": 0.3307123131732614, "grad_norm": 0.3557393550872803, "learning_rate": 4.9432368281332964e-05, "loss": 0.1561, "step": 7844 }, { "epoch": 0.3307544743554609, "grad_norm": 1.147192358970642, "learning_rate": 4.943158749492489e-05, "loss": 0.9907, "step": 7845 }, { "epoch": 0.3307966355376605, "grad_norm": 0.8173346519470215, "learning_rate": 4.943080670851682e-05, "loss": 0.2682, "step": 7846 }, { "epoch": 0.33083879671986005, "grad_norm": 1.5202341079711914, "learning_rate": 4.943002592210875e-05, "loss": 0.481, "step": 7847 }, { "epoch": 0.33088095790205957, "grad_norm": 1.1077195405960083, "learning_rate": 4.942924513570068e-05, "loss": 0.6624, "step": 7848 }, { "epoch": 0.33092311908425914, "grad_norm": 1.5313735008239746, "learning_rate": 4.942846434929261e-05, "loss": 0.8206, "step": 7849 }, { "epoch": 0.33096528026645866, "grad_norm": 0.34691449999809265, "learning_rate": 4.942768356288454e-05, "loss": 0.1848, "step": 7850 }, { "epoch": 0.33100744144865824, "grad_norm": 0.4691162705421448, "learning_rate": 4.942690277647647e-05, "loss": 0.1377, "step": 7851 }, { "epoch": 0.33104960263085775, "grad_norm": 0.5548700094223022, "learning_rate": 4.9426121990068394e-05, "loss": 0.1062, "step": 7852 }, { "epoch": 0.3310917638130573, "grad_norm": 0.7048453688621521, "learning_rate": 4.942534120366033e-05, "loss": 0.12, "step": 7853 }, { "epoch": 0.33113392499525685, "grad_norm": 1.0048472881317139, "learning_rate": 4.942456041725226e-05, "loss": 0.3007, "step": 7854 }, { "epoch": 0.3311760861774564, "grad_norm": 0.6513274908065796, "learning_rate": 4.9423779630844195e-05, "loss": 0.6514, "step": 7855 }, { "epoch": 0.33121824735965594, "grad_norm": 0.5730298757553101, "learning_rate": 4.942299884443612e-05, "loss": 0.1207, "step": 7856 }, { "epoch": 0.3312604085418555, "grad_norm": 0.4851478040218353, "learning_rate": 4.9422218058028044e-05, "loss": 0.1404, "step": 7857 }, { "epoch": 0.3313025697240551, "grad_norm": 0.8848634362220764, "learning_rate": 4.942143727161998e-05, "loss": 0.2845, "step": 7858 }, { "epoch": 0.3313447309062546, "grad_norm": 0.5875973105430603, "learning_rate": 4.9420656485211906e-05, "loss": 0.3427, "step": 7859 }, { "epoch": 0.3313868920884542, "grad_norm": 0.7425317168235779, "learning_rate": 4.941987569880384e-05, "loss": 0.1225, "step": 7860 }, { "epoch": 0.3314290532706537, "grad_norm": 0.5845487117767334, "learning_rate": 4.941909491239577e-05, "loss": 0.4541, "step": 7861 }, { "epoch": 0.33147121445285327, "grad_norm": 0.8791206479072571, "learning_rate": 4.94183141259877e-05, "loss": 0.6107, "step": 7862 }, { "epoch": 0.3315133756350528, "grad_norm": 7.171413898468018, "learning_rate": 4.9417533339579624e-05, "loss": 0.1568, "step": 7863 }, { "epoch": 0.33155553681725236, "grad_norm": 0.4369153678417206, "learning_rate": 4.9416752553171556e-05, "loss": 0.0662, "step": 7864 }, { "epoch": 0.3315976979994519, "grad_norm": 0.48050686717033386, "learning_rate": 4.941597176676349e-05, "loss": 0.1582, "step": 7865 }, { "epoch": 0.33163985918165145, "grad_norm": 0.885890543460846, "learning_rate": 4.941519098035541e-05, "loss": 0.5091, "step": 7866 }, { "epoch": 0.331682020363851, "grad_norm": 0.8349555730819702, "learning_rate": 4.941441019394735e-05, "loss": 0.5553, "step": 7867 }, { "epoch": 0.33172418154605054, "grad_norm": 1.1822818517684937, "learning_rate": 4.9413629407539274e-05, "loss": 1.1817, "step": 7868 }, { "epoch": 0.3317663427282501, "grad_norm": 0.42631545662879944, "learning_rate": 4.9412848621131205e-05, "loss": 0.1621, "step": 7869 }, { "epoch": 0.33180850391044964, "grad_norm": 0.6190347075462341, "learning_rate": 4.9412067834723137e-05, "loss": 0.5086, "step": 7870 }, { "epoch": 0.3318506650926492, "grad_norm": 0.6934875249862671, "learning_rate": 4.941128704831506e-05, "loss": 0.3822, "step": 7871 }, { "epoch": 0.33189282627484873, "grad_norm": 0.44118550419807434, "learning_rate": 4.9410506261907e-05, "loss": 0.2591, "step": 7872 }, { "epoch": 0.3319349874570483, "grad_norm": 0.7047362923622131, "learning_rate": 4.9409725475498924e-05, "loss": 0.4959, "step": 7873 }, { "epoch": 0.3319771486392478, "grad_norm": 0.9433783888816833, "learning_rate": 4.9408944689090855e-05, "loss": 0.1975, "step": 7874 }, { "epoch": 0.3320193098214474, "grad_norm": 0.36733174324035645, "learning_rate": 4.9408163902682786e-05, "loss": 0.1233, "step": 7875 }, { "epoch": 0.33206147100364697, "grad_norm": 1.7191184759140015, "learning_rate": 4.940738311627471e-05, "loss": 0.9888, "step": 7876 }, { "epoch": 0.3321036321858465, "grad_norm": 1.8084502220153809, "learning_rate": 4.940660232986664e-05, "loss": 0.9906, "step": 7877 }, { "epoch": 0.33214579336804606, "grad_norm": 1.129390001296997, "learning_rate": 4.940582154345857e-05, "loss": 0.7585, "step": 7878 }, { "epoch": 0.3321879545502456, "grad_norm": 0.5744965076446533, "learning_rate": 4.9405040757050504e-05, "loss": 0.3864, "step": 7879 }, { "epoch": 0.33223011573244515, "grad_norm": 1.2926350831985474, "learning_rate": 4.9404259970642436e-05, "loss": 0.6455, "step": 7880 }, { "epoch": 0.33227227691464467, "grad_norm": 1.3918352127075195, "learning_rate": 4.940347918423437e-05, "loss": 1.2569, "step": 7881 }, { "epoch": 0.33231443809684424, "grad_norm": 17.077733993530273, "learning_rate": 4.940269839782629e-05, "loss": 0.5798, "step": 7882 }, { "epoch": 0.33235659927904376, "grad_norm": 0.40465712547302246, "learning_rate": 4.940191761141822e-05, "loss": 0.077, "step": 7883 }, { "epoch": 0.33239876046124334, "grad_norm": 0.3465574085712433, "learning_rate": 4.9401136825010154e-05, "loss": 0.0865, "step": 7884 }, { "epoch": 0.33244092164344285, "grad_norm": 0.6791514158248901, "learning_rate": 4.940035603860208e-05, "loss": 0.531, "step": 7885 }, { "epoch": 0.33248308282564243, "grad_norm": 0.3949412703514099, "learning_rate": 4.9399575252194016e-05, "loss": 0.0803, "step": 7886 }, { "epoch": 0.332525244007842, "grad_norm": 0.8681821823120117, "learning_rate": 4.939879446578594e-05, "loss": 0.3016, "step": 7887 }, { "epoch": 0.3325674051900415, "grad_norm": 0.3869817852973938, "learning_rate": 4.939801367937787e-05, "loss": 0.1171, "step": 7888 }, { "epoch": 0.3326095663722411, "grad_norm": 13.972356796264648, "learning_rate": 4.93972328929698e-05, "loss": 1.0208, "step": 7889 }, { "epoch": 0.3326517275544406, "grad_norm": 0.41020938754081726, "learning_rate": 4.939645210656173e-05, "loss": 0.1988, "step": 7890 }, { "epoch": 0.3326938887366402, "grad_norm": 0.6322850584983826, "learning_rate": 4.9395671320153666e-05, "loss": 0.1418, "step": 7891 }, { "epoch": 0.3327360499188397, "grad_norm": 0.33901503682136536, "learning_rate": 4.939489053374559e-05, "loss": 0.1346, "step": 7892 }, { "epoch": 0.3327782111010393, "grad_norm": 0.3790864944458008, "learning_rate": 4.939410974733752e-05, "loss": 0.1263, "step": 7893 }, { "epoch": 0.3328203722832388, "grad_norm": 0.5969566702842712, "learning_rate": 4.939332896092945e-05, "loss": 0.6597, "step": 7894 }, { "epoch": 0.33286253346543837, "grad_norm": 0.5170816779136658, "learning_rate": 4.939254817452138e-05, "loss": 0.2011, "step": 7895 }, { "epoch": 0.33290469464763794, "grad_norm": 0.6832467317581177, "learning_rate": 4.939176738811331e-05, "loss": 0.1775, "step": 7896 }, { "epoch": 0.33294685582983746, "grad_norm": 1.1173168420791626, "learning_rate": 4.939098660170524e-05, "loss": 0.1476, "step": 7897 }, { "epoch": 0.33298901701203704, "grad_norm": 0.5132502913475037, "learning_rate": 4.939020581529717e-05, "loss": 0.4874, "step": 7898 }, { "epoch": 0.33303117819423655, "grad_norm": 0.686306893825531, "learning_rate": 4.9389425028889096e-05, "loss": 0.2968, "step": 7899 }, { "epoch": 0.33307333937643613, "grad_norm": 1.3839553594589233, "learning_rate": 4.938864424248103e-05, "loss": 0.3035, "step": 7900 }, { "epoch": 0.33311550055863565, "grad_norm": 0.7998965978622437, "learning_rate": 4.938786345607296e-05, "loss": 0.631, "step": 7901 }, { "epoch": 0.3331576617408352, "grad_norm": 0.5900026559829712, "learning_rate": 4.938708266966489e-05, "loss": 0.3437, "step": 7902 }, { "epoch": 0.33319982292303474, "grad_norm": 0.352317750453949, "learning_rate": 4.938630188325682e-05, "loss": 0.1632, "step": 7903 }, { "epoch": 0.3332419841052343, "grad_norm": 1.183040738105774, "learning_rate": 4.9385521096848745e-05, "loss": 1.0156, "step": 7904 }, { "epoch": 0.3332841452874339, "grad_norm": 1.7965813875198364, "learning_rate": 4.938474031044068e-05, "loss": 0.2696, "step": 7905 }, { "epoch": 0.3333263064696334, "grad_norm": 0.8034742474555969, "learning_rate": 4.938395952403261e-05, "loss": 0.454, "step": 7906 }, { "epoch": 0.333368467651833, "grad_norm": 1.0662953853607178, "learning_rate": 4.938317873762454e-05, "loss": 0.6726, "step": 7907 }, { "epoch": 0.3334106288340325, "grad_norm": 0.5087801814079285, "learning_rate": 4.938239795121647e-05, "loss": 0.1708, "step": 7908 }, { "epoch": 0.33345279001623207, "grad_norm": 1.319564700126648, "learning_rate": 4.9381617164808395e-05, "loss": 0.6727, "step": 7909 }, { "epoch": 0.3334949511984316, "grad_norm": 0.3343695402145386, "learning_rate": 4.9380836378400326e-05, "loss": 0.114, "step": 7910 }, { "epoch": 0.33353711238063116, "grad_norm": 0.38553228974342346, "learning_rate": 4.938005559199226e-05, "loss": 0.1921, "step": 7911 }, { "epoch": 0.3335792735628307, "grad_norm": 0.9206613898277283, "learning_rate": 4.937927480558419e-05, "loss": 0.551, "step": 7912 }, { "epoch": 0.33362143474503025, "grad_norm": 1.2156800031661987, "learning_rate": 4.937849401917612e-05, "loss": 1.1931, "step": 7913 }, { "epoch": 0.3336635959272298, "grad_norm": 0.7175296545028687, "learning_rate": 4.9377713232768044e-05, "loss": 0.2528, "step": 7914 }, { "epoch": 0.33370575710942935, "grad_norm": 0.4219092130661011, "learning_rate": 4.9376932446359975e-05, "loss": 0.1477, "step": 7915 }, { "epoch": 0.3337479182916289, "grad_norm": 1.0422184467315674, "learning_rate": 4.937615165995191e-05, "loss": 0.4233, "step": 7916 }, { "epoch": 0.33379007947382844, "grad_norm": 1.6777466535568237, "learning_rate": 4.937537087354384e-05, "loss": 0.1041, "step": 7917 }, { "epoch": 0.333832240656028, "grad_norm": 0.48388785123825073, "learning_rate": 4.937459008713576e-05, "loss": 0.1635, "step": 7918 }, { "epoch": 0.33387440183822753, "grad_norm": 0.6064760088920593, "learning_rate": 4.9373809300727694e-05, "loss": 0.1665, "step": 7919 }, { "epoch": 0.3339165630204271, "grad_norm": 1.186261773109436, "learning_rate": 4.9373028514319625e-05, "loss": 0.371, "step": 7920 }, { "epoch": 0.3339587242026266, "grad_norm": 0.7429848313331604, "learning_rate": 4.937224772791155e-05, "loss": 0.1998, "step": 7921 }, { "epoch": 0.3340008853848262, "grad_norm": 0.8839889764785767, "learning_rate": 4.937146694150349e-05, "loss": 0.4698, "step": 7922 }, { "epoch": 0.3340430465670257, "grad_norm": 14.038392066955566, "learning_rate": 4.937068615509541e-05, "loss": 1.0265, "step": 7923 }, { "epoch": 0.3340852077492253, "grad_norm": 0.6107796430587769, "learning_rate": 4.936990536868735e-05, "loss": 0.1177, "step": 7924 }, { "epoch": 0.33412736893142486, "grad_norm": 0.531558096408844, "learning_rate": 4.9369124582279275e-05, "loss": 0.1951, "step": 7925 }, { "epoch": 0.3341695301136244, "grad_norm": 0.697104811668396, "learning_rate": 4.9368343795871206e-05, "loss": 0.4949, "step": 7926 }, { "epoch": 0.33421169129582395, "grad_norm": 0.5976822376251221, "learning_rate": 4.936756300946314e-05, "loss": 0.1153, "step": 7927 }, { "epoch": 0.33425385247802347, "grad_norm": 0.3902406096458435, "learning_rate": 4.936678222305506e-05, "loss": 0.1074, "step": 7928 }, { "epoch": 0.33429601366022305, "grad_norm": 2.285006523132324, "learning_rate": 4.936600143664699e-05, "loss": 0.6753, "step": 7929 }, { "epoch": 0.33433817484242256, "grad_norm": 1.5795060396194458, "learning_rate": 4.9365220650238924e-05, "loss": 0.1067, "step": 7930 }, { "epoch": 0.33438033602462214, "grad_norm": 0.697750449180603, "learning_rate": 4.9364439863830855e-05, "loss": 0.1981, "step": 7931 }, { "epoch": 0.33442249720682166, "grad_norm": 0.7966853380203247, "learning_rate": 4.936365907742278e-05, "loss": 0.4318, "step": 7932 }, { "epoch": 0.33446465838902123, "grad_norm": 0.7046104073524475, "learning_rate": 4.936287829101471e-05, "loss": 0.6419, "step": 7933 }, { "epoch": 0.3345068195712208, "grad_norm": 0.5630910992622375, "learning_rate": 4.936209750460664e-05, "loss": 0.1666, "step": 7934 }, { "epoch": 0.3345489807534203, "grad_norm": 0.8013085722923279, "learning_rate": 4.9361316718198574e-05, "loss": 0.5022, "step": 7935 }, { "epoch": 0.3345911419356199, "grad_norm": 0.3908577561378479, "learning_rate": 4.9360535931790505e-05, "loss": 0.0819, "step": 7936 }, { "epoch": 0.3346333031178194, "grad_norm": 1.0127315521240234, "learning_rate": 4.935975514538243e-05, "loss": 0.1793, "step": 7937 }, { "epoch": 0.334675464300019, "grad_norm": 0.9583228826522827, "learning_rate": 4.935897435897436e-05, "loss": 0.4728, "step": 7938 }, { "epoch": 0.3347176254822185, "grad_norm": 0.42275604605674744, "learning_rate": 4.935819357256629e-05, "loss": 0.1607, "step": 7939 }, { "epoch": 0.3347597866644181, "grad_norm": 0.5117814540863037, "learning_rate": 4.9357412786158216e-05, "loss": 0.2765, "step": 7940 }, { "epoch": 0.3348019478466176, "grad_norm": 0.6135039925575256, "learning_rate": 4.9356631999750154e-05, "loss": 0.3738, "step": 7941 }, { "epoch": 0.33484410902881717, "grad_norm": 1.1026678085327148, "learning_rate": 4.935585121334208e-05, "loss": 0.5752, "step": 7942 }, { "epoch": 0.3348862702110167, "grad_norm": 1.5599263906478882, "learning_rate": 4.935507042693401e-05, "loss": 0.2818, "step": 7943 }, { "epoch": 0.33492843139321626, "grad_norm": 0.41297951340675354, "learning_rate": 4.935428964052594e-05, "loss": 0.0801, "step": 7944 }, { "epoch": 0.33497059257541584, "grad_norm": 1.0698648691177368, "learning_rate": 4.9353508854117866e-05, "loss": 1.0752, "step": 7945 }, { "epoch": 0.33501275375761536, "grad_norm": 0.8489609360694885, "learning_rate": 4.93527280677098e-05, "loss": 0.4144, "step": 7946 }, { "epoch": 0.33505491493981493, "grad_norm": 0.9973334670066833, "learning_rate": 4.935194728130173e-05, "loss": 0.4364, "step": 7947 }, { "epoch": 0.33509707612201445, "grad_norm": 0.6387335658073425, "learning_rate": 4.935116649489366e-05, "loss": 0.2001, "step": 7948 }, { "epoch": 0.335139237304214, "grad_norm": 0.7937806248664856, "learning_rate": 4.935038570848559e-05, "loss": 0.5117, "step": 7949 }, { "epoch": 0.33518139848641354, "grad_norm": 0.5286217927932739, "learning_rate": 4.934960492207752e-05, "loss": 0.1448, "step": 7950 }, { "epoch": 0.3352235596686131, "grad_norm": 1.2812927961349487, "learning_rate": 4.9348824135669447e-05, "loss": 0.882, "step": 7951 }, { "epoch": 0.33526572085081263, "grad_norm": 2480581632.0, "learning_rate": 4.934804334926138e-05, "loss": 0.1737, "step": 7952 }, { "epoch": 0.3353078820330122, "grad_norm": 1.9164408445358276, "learning_rate": 4.934726256285331e-05, "loss": 0.3714, "step": 7953 }, { "epoch": 0.3353500432152118, "grad_norm": 1.3230687379837036, "learning_rate": 4.9346481776445234e-05, "loss": 0.4443, "step": 7954 }, { "epoch": 0.3353922043974113, "grad_norm": 3.3225889205932617, "learning_rate": 4.934570099003717e-05, "loss": 0.4215, "step": 7955 }, { "epoch": 0.33543436557961087, "grad_norm": 0.6551787257194519, "learning_rate": 4.9344920203629096e-05, "loss": 0.1258, "step": 7956 }, { "epoch": 0.3354765267618104, "grad_norm": 0.7088029384613037, "learning_rate": 4.934413941722103e-05, "loss": 0.1359, "step": 7957 }, { "epoch": 0.33551868794400996, "grad_norm": 0.7837749123573303, "learning_rate": 4.934335863081296e-05, "loss": 0.6603, "step": 7958 }, { "epoch": 0.3355608491262095, "grad_norm": 2.298382520675659, "learning_rate": 4.934257784440488e-05, "loss": 0.1501, "step": 7959 }, { "epoch": 0.33560301030840906, "grad_norm": 0.6047143340110779, "learning_rate": 4.934179705799682e-05, "loss": 0.5412, "step": 7960 }, { "epoch": 0.3356451714906086, "grad_norm": 0.5950296521186829, "learning_rate": 4.9341016271588746e-05, "loss": 0.6382, "step": 7961 }, { "epoch": 0.33568733267280815, "grad_norm": 0.5236849188804626, "learning_rate": 4.934023548518068e-05, "loss": 0.4029, "step": 7962 }, { "epoch": 0.3357294938550077, "grad_norm": 0.5434891581535339, "learning_rate": 4.933945469877261e-05, "loss": 0.1651, "step": 7963 }, { "epoch": 0.33577165503720724, "grad_norm": 2.5894484519958496, "learning_rate": 4.933867391236453e-05, "loss": 0.7433, "step": 7964 }, { "epoch": 0.3358138162194068, "grad_norm": 0.5044467449188232, "learning_rate": 4.9337893125956464e-05, "loss": 0.4521, "step": 7965 }, { "epoch": 0.33585597740160633, "grad_norm": 0.6607158184051514, "learning_rate": 4.9337112339548395e-05, "loss": 0.1904, "step": 7966 }, { "epoch": 0.3358981385838059, "grad_norm": 0.745089590549469, "learning_rate": 4.9336331553140326e-05, "loss": 0.5988, "step": 7967 }, { "epoch": 0.3359402997660054, "grad_norm": 1.335058569908142, "learning_rate": 4.933555076673225e-05, "loss": 0.5372, "step": 7968 }, { "epoch": 0.335982460948205, "grad_norm": 2.6833624839782715, "learning_rate": 4.933476998032419e-05, "loss": 0.8424, "step": 7969 }, { "epoch": 0.3360246221304045, "grad_norm": 1.0213993787765503, "learning_rate": 4.9333989193916113e-05, "loss": 0.3867, "step": 7970 }, { "epoch": 0.3360667833126041, "grad_norm": 0.6371411681175232, "learning_rate": 4.9333208407508045e-05, "loss": 0.0666, "step": 7971 }, { "epoch": 0.3361089444948036, "grad_norm": 0.5816647410392761, "learning_rate": 4.9332427621099976e-05, "loss": 0.1534, "step": 7972 }, { "epoch": 0.3361511056770032, "grad_norm": 0.9071958661079407, "learning_rate": 4.93316468346919e-05, "loss": 0.3192, "step": 7973 }, { "epoch": 0.33619326685920276, "grad_norm": 0.6839936971664429, "learning_rate": 4.933086604828384e-05, "loss": 0.6004, "step": 7974 }, { "epoch": 0.3362354280414023, "grad_norm": 0.8348408937454224, "learning_rate": 4.933008526187576e-05, "loss": 0.3321, "step": 7975 }, { "epoch": 0.33627758922360185, "grad_norm": 0.4890582263469696, "learning_rate": 4.9329304475467694e-05, "loss": 0.2145, "step": 7976 }, { "epoch": 0.33631975040580137, "grad_norm": 0.8372754454612732, "learning_rate": 4.9328523689059625e-05, "loss": 0.6105, "step": 7977 }, { "epoch": 0.33636191158800094, "grad_norm": 1.9175634384155273, "learning_rate": 4.932774290265155e-05, "loss": 0.8443, "step": 7978 }, { "epoch": 0.33640407277020046, "grad_norm": 0.497829407453537, "learning_rate": 4.932696211624348e-05, "loss": 0.1196, "step": 7979 }, { "epoch": 0.33644623395240003, "grad_norm": 1.0778145790100098, "learning_rate": 4.932618132983541e-05, "loss": 0.3845, "step": 7980 }, { "epoch": 0.33648839513459955, "grad_norm": 0.7474990487098694, "learning_rate": 4.9325400543427344e-05, "loss": 0.2287, "step": 7981 }, { "epoch": 0.3365305563167991, "grad_norm": 0.4773689806461334, "learning_rate": 4.9324619757019275e-05, "loss": 0.385, "step": 7982 }, { "epoch": 0.3365727174989987, "grad_norm": 1.119624376296997, "learning_rate": 4.93238389706112e-05, "loss": 0.4542, "step": 7983 }, { "epoch": 0.3366148786811982, "grad_norm": 0.6829540133476257, "learning_rate": 4.932305818420313e-05, "loss": 0.3107, "step": 7984 }, { "epoch": 0.3366570398633978, "grad_norm": 1.233400821685791, "learning_rate": 4.932227739779506e-05, "loss": 0.1797, "step": 7985 }, { "epoch": 0.3366992010455973, "grad_norm": 0.6191327571868896, "learning_rate": 4.932149661138699e-05, "loss": 0.1542, "step": 7986 }, { "epoch": 0.3367413622277969, "grad_norm": 0.5072474479675293, "learning_rate": 4.932071582497892e-05, "loss": 0.1121, "step": 7987 }, { "epoch": 0.3367835234099964, "grad_norm": 0.2999827265739441, "learning_rate": 4.9319935038570856e-05, "loss": 0.1107, "step": 7988 }, { "epoch": 0.336825684592196, "grad_norm": 0.6178728342056274, "learning_rate": 4.931915425216278e-05, "loss": 0.2523, "step": 7989 }, { "epoch": 0.3368678457743955, "grad_norm": 0.5007992386817932, "learning_rate": 4.9318373465754705e-05, "loss": 0.1777, "step": 7990 }, { "epoch": 0.33691000695659507, "grad_norm": 0.558999240398407, "learning_rate": 4.931759267934664e-05, "loss": 0.0934, "step": 7991 }, { "epoch": 0.33695216813879464, "grad_norm": 0.5053093433380127, "learning_rate": 4.931681189293857e-05, "loss": 0.3314, "step": 7992 }, { "epoch": 0.33699432932099416, "grad_norm": 0.5024606585502625, "learning_rate": 4.9316031106530505e-05, "loss": 0.1459, "step": 7993 }, { "epoch": 0.33703649050319373, "grad_norm": 0.3884508013725281, "learning_rate": 4.931525032012243e-05, "loss": 0.1523, "step": 7994 }, { "epoch": 0.33707865168539325, "grad_norm": 3.8449904918670654, "learning_rate": 4.931446953371436e-05, "loss": 1.0733, "step": 7995 }, { "epoch": 0.3371208128675928, "grad_norm": 1.2606287002563477, "learning_rate": 4.931368874730629e-05, "loss": 0.186, "step": 7996 }, { "epoch": 0.33716297404979234, "grad_norm": 0.9674380421638489, "learning_rate": 4.931290796089822e-05, "loss": 0.5335, "step": 7997 }, { "epoch": 0.3372051352319919, "grad_norm": 0.6425155401229858, "learning_rate": 4.931212717449015e-05, "loss": 0.6155, "step": 7998 }, { "epoch": 0.33724729641419143, "grad_norm": 0.8007364273071289, "learning_rate": 4.931134638808208e-05, "loss": 0.4497, "step": 7999 }, { "epoch": 0.337289457596391, "grad_norm": 0.49576041102409363, "learning_rate": 4.931056560167401e-05, "loss": 0.0817, "step": 8000 }, { "epoch": 0.3373316187785905, "grad_norm": 0.800419270992279, "learning_rate": 4.9309784815265935e-05, "loss": 0.4994, "step": 8001 }, { "epoch": 0.3373737799607901, "grad_norm": 0.8685582876205444, "learning_rate": 4.9309004028857866e-05, "loss": 0.1695, "step": 8002 }, { "epoch": 0.3374159411429897, "grad_norm": 1.7108274698257446, "learning_rate": 4.93082232424498e-05, "loss": 0.5146, "step": 8003 }, { "epoch": 0.3374581023251892, "grad_norm": 1.302804946899414, "learning_rate": 4.930744245604173e-05, "loss": 1.1209, "step": 8004 }, { "epoch": 0.33750026350738876, "grad_norm": 0.5441687703132629, "learning_rate": 4.930666166963366e-05, "loss": 0.2857, "step": 8005 }, { "epoch": 0.3375424246895883, "grad_norm": 1.3301469087600708, "learning_rate": 4.9305880883225585e-05, "loss": 0.1096, "step": 8006 }, { "epoch": 0.33758458587178786, "grad_norm": 1.5240806341171265, "learning_rate": 4.9305100096817516e-05, "loss": 0.4283, "step": 8007 }, { "epoch": 0.3376267470539874, "grad_norm": 1.2191306352615356, "learning_rate": 4.930431931040945e-05, "loss": 0.1944, "step": 8008 }, { "epoch": 0.33766890823618695, "grad_norm": 0.49700382351875305, "learning_rate": 4.930353852400137e-05, "loss": 0.163, "step": 8009 }, { "epoch": 0.33771106941838647, "grad_norm": 1.2446614503860474, "learning_rate": 4.930275773759331e-05, "loss": 0.493, "step": 8010 }, { "epoch": 0.33775323060058604, "grad_norm": 0.7614259719848633, "learning_rate": 4.9301976951185234e-05, "loss": 0.1358, "step": 8011 }, { "epoch": 0.3377953917827856, "grad_norm": 0.6053628325462341, "learning_rate": 4.9301196164777165e-05, "loss": 0.1303, "step": 8012 }, { "epoch": 0.33783755296498513, "grad_norm": 0.6616206169128418, "learning_rate": 4.93004153783691e-05, "loss": 0.3308, "step": 8013 }, { "epoch": 0.3378797141471847, "grad_norm": 2.048394203186035, "learning_rate": 4.929963459196103e-05, "loss": 0.1492, "step": 8014 }, { "epoch": 0.3379218753293842, "grad_norm": 0.6805393099784851, "learning_rate": 4.929885380555296e-05, "loss": 0.7163, "step": 8015 }, { "epoch": 0.3379640365115838, "grad_norm": 4.98323392868042, "learning_rate": 4.9298073019144884e-05, "loss": 0.5571, "step": 8016 }, { "epoch": 0.3380061976937833, "grad_norm": 0.6929663419723511, "learning_rate": 4.9297292232736815e-05, "loss": 0.4629, "step": 8017 }, { "epoch": 0.3380483588759829, "grad_norm": 0.5970261693000793, "learning_rate": 4.9296511446328746e-05, "loss": 0.1212, "step": 8018 }, { "epoch": 0.3380905200581824, "grad_norm": 1.5088454484939575, "learning_rate": 4.929573065992068e-05, "loss": 0.2836, "step": 8019 }, { "epoch": 0.338132681240382, "grad_norm": 0.7305095791816711, "learning_rate": 4.92949498735126e-05, "loss": 0.249, "step": 8020 }, { "epoch": 0.33817484242258156, "grad_norm": 0.5315858721733093, "learning_rate": 4.929416908710453e-05, "loss": 0.2073, "step": 8021 }, { "epoch": 0.3382170036047811, "grad_norm": 0.8740024566650391, "learning_rate": 4.9293388300696464e-05, "loss": 0.2025, "step": 8022 }, { "epoch": 0.33825916478698065, "grad_norm": 0.633019745349884, "learning_rate": 4.929260751428839e-05, "loss": 0.2819, "step": 8023 }, { "epoch": 0.33830132596918017, "grad_norm": 0.9529792666435242, "learning_rate": 4.929182672788033e-05, "loss": 0.9943, "step": 8024 }, { "epoch": 0.33834348715137974, "grad_norm": 1.5570836067199707, "learning_rate": 4.929104594147225e-05, "loss": 0.7136, "step": 8025 }, { "epoch": 0.33838564833357926, "grad_norm": 1.2171895503997803, "learning_rate": 4.929026515506418e-05, "loss": 0.7667, "step": 8026 }, { "epoch": 0.33842780951577883, "grad_norm": 0.9491652846336365, "learning_rate": 4.9289484368656114e-05, "loss": 0.7426, "step": 8027 }, { "epoch": 0.33846997069797835, "grad_norm": 0.606768012046814, "learning_rate": 4.928870358224804e-05, "loss": 0.6418, "step": 8028 }, { "epoch": 0.3385121318801779, "grad_norm": 0.7318097352981567, "learning_rate": 4.9287922795839976e-05, "loss": 0.2925, "step": 8029 }, { "epoch": 0.33855429306237744, "grad_norm": 0.8579711318016052, "learning_rate": 4.92871420094319e-05, "loss": 0.5242, "step": 8030 }, { "epoch": 0.338596454244577, "grad_norm": 2.2017154693603516, "learning_rate": 4.928636122302383e-05, "loss": 0.1548, "step": 8031 }, { "epoch": 0.3386386154267766, "grad_norm": 1.9234338998794556, "learning_rate": 4.9285580436615763e-05, "loss": 0.383, "step": 8032 }, { "epoch": 0.3386807766089761, "grad_norm": 0.8581458330154419, "learning_rate": 4.9284799650207695e-05, "loss": 0.5446, "step": 8033 }, { "epoch": 0.3387229377911757, "grad_norm": 0.3298199474811554, "learning_rate": 4.928401886379962e-05, "loss": 0.13, "step": 8034 }, { "epoch": 0.3387650989733752, "grad_norm": 0.8445586562156677, "learning_rate": 4.928323807739155e-05, "loss": 0.157, "step": 8035 }, { "epoch": 0.3388072601555748, "grad_norm": 0.580739438533783, "learning_rate": 4.928245729098348e-05, "loss": 0.1496, "step": 8036 }, { "epoch": 0.3388494213377743, "grad_norm": 0.4698554277420044, "learning_rate": 4.9281676504575406e-05, "loss": 0.3235, "step": 8037 }, { "epoch": 0.33889158251997387, "grad_norm": 0.5507820248603821, "learning_rate": 4.9280895718167344e-05, "loss": 0.1792, "step": 8038 }, { "epoch": 0.3389337437021734, "grad_norm": 1.2683531045913696, "learning_rate": 4.928011493175927e-05, "loss": 0.7193, "step": 8039 }, { "epoch": 0.33897590488437296, "grad_norm": 0.42009201645851135, "learning_rate": 4.92793341453512e-05, "loss": 0.109, "step": 8040 }, { "epoch": 0.33901806606657253, "grad_norm": 1.2064917087554932, "learning_rate": 4.927855335894313e-05, "loss": 0.2621, "step": 8041 }, { "epoch": 0.33906022724877205, "grad_norm": 0.3458399772644043, "learning_rate": 4.9277772572535056e-05, "loss": 0.1595, "step": 8042 }, { "epoch": 0.3391023884309716, "grad_norm": 1.20820152759552, "learning_rate": 4.9276991786126994e-05, "loss": 0.2707, "step": 8043 }, { "epoch": 0.33914454961317114, "grad_norm": 2.2263615131378174, "learning_rate": 4.927621099971892e-05, "loss": 1.2377, "step": 8044 }, { "epoch": 0.3391867107953707, "grad_norm": 0.5407209992408752, "learning_rate": 4.927543021331085e-05, "loss": 0.1888, "step": 8045 }, { "epoch": 0.33922887197757023, "grad_norm": 1.0509341955184937, "learning_rate": 4.927464942690278e-05, "loss": 0.2215, "step": 8046 }, { "epoch": 0.3392710331597698, "grad_norm": 0.47914162278175354, "learning_rate": 4.9273868640494705e-05, "loss": 0.1246, "step": 8047 }, { "epoch": 0.3393131943419693, "grad_norm": 0.48964783549308777, "learning_rate": 4.9273087854086637e-05, "loss": 0.3779, "step": 8048 }, { "epoch": 0.3393553555241689, "grad_norm": 0.6655411720275879, "learning_rate": 4.927230706767857e-05, "loss": 0.5451, "step": 8049 }, { "epoch": 0.3393975167063685, "grad_norm": 1.1697160005569458, "learning_rate": 4.92715262812705e-05, "loss": 0.4304, "step": 8050 }, { "epoch": 0.339439677888568, "grad_norm": 1.9257251024246216, "learning_rate": 4.927074549486243e-05, "loss": 0.3013, "step": 8051 }, { "epoch": 0.33948183907076757, "grad_norm": 1.2407944202423096, "learning_rate": 4.9269964708454355e-05, "loss": 0.7552, "step": 8052 }, { "epoch": 0.3395240002529671, "grad_norm": 0.6360632181167603, "learning_rate": 4.9269183922046286e-05, "loss": 0.4929, "step": 8053 }, { "epoch": 0.33956616143516666, "grad_norm": 0.7620409727096558, "learning_rate": 4.926840313563822e-05, "loss": 0.5496, "step": 8054 }, { "epoch": 0.3396083226173662, "grad_norm": 0.8624529242515564, "learning_rate": 4.926762234923015e-05, "loss": 0.4179, "step": 8055 }, { "epoch": 0.33965048379956575, "grad_norm": 0.8658126592636108, "learning_rate": 4.926684156282207e-05, "loss": 0.5903, "step": 8056 }, { "epoch": 0.33969264498176527, "grad_norm": 1.8436909914016724, "learning_rate": 4.926606077641401e-05, "loss": 0.7677, "step": 8057 }, { "epoch": 0.33973480616396484, "grad_norm": 0.45400768518447876, "learning_rate": 4.9265279990005936e-05, "loss": 0.1748, "step": 8058 }, { "epoch": 0.33977696734616436, "grad_norm": 0.5437358617782593, "learning_rate": 4.926449920359787e-05, "loss": 0.3765, "step": 8059 }, { "epoch": 0.33981912852836393, "grad_norm": 1.2245768308639526, "learning_rate": 4.92637184171898e-05, "loss": 0.2123, "step": 8060 }, { "epoch": 0.3398612897105635, "grad_norm": 2.3931286334991455, "learning_rate": 4.926293763078172e-05, "loss": 0.2619, "step": 8061 }, { "epoch": 0.339903450892763, "grad_norm": 0.5671524405479431, "learning_rate": 4.926215684437366e-05, "loss": 0.0661, "step": 8062 }, { "epoch": 0.3399456120749626, "grad_norm": 0.8486682772636414, "learning_rate": 4.9261376057965585e-05, "loss": 0.8712, "step": 8063 }, { "epoch": 0.3399877732571621, "grad_norm": 0.6066863536834717, "learning_rate": 4.9260595271557516e-05, "loss": 0.3762, "step": 8064 }, { "epoch": 0.3400299344393617, "grad_norm": 0.4138808250427246, "learning_rate": 4.925981448514945e-05, "loss": 0.0915, "step": 8065 }, { "epoch": 0.3400720956215612, "grad_norm": 0.791409969329834, "learning_rate": 4.925903369874137e-05, "loss": 0.3235, "step": 8066 }, { "epoch": 0.3401142568037608, "grad_norm": 0.8006197214126587, "learning_rate": 4.92582529123333e-05, "loss": 0.2911, "step": 8067 }, { "epoch": 0.3401564179859603, "grad_norm": 0.6074906587600708, "learning_rate": 4.9257472125925235e-05, "loss": 0.2195, "step": 8068 }, { "epoch": 0.3401985791681599, "grad_norm": 0.5505651235580444, "learning_rate": 4.9256691339517166e-05, "loss": 0.2177, "step": 8069 }, { "epoch": 0.34024074035035945, "grad_norm": 0.6122113466262817, "learning_rate": 4.925591055310909e-05, "loss": 0.6914, "step": 8070 }, { "epoch": 0.34028290153255897, "grad_norm": 0.5486038327217102, "learning_rate": 4.925512976670102e-05, "loss": 0.1748, "step": 8071 }, { "epoch": 0.34032506271475854, "grad_norm": 0.47722068428993225, "learning_rate": 4.925434898029295e-05, "loss": 0.1336, "step": 8072 }, { "epoch": 0.34036722389695806, "grad_norm": 0.9848927855491638, "learning_rate": 4.9253568193884884e-05, "loss": 0.8558, "step": 8073 }, { "epoch": 0.34040938507915763, "grad_norm": 0.8240615725517273, "learning_rate": 4.9252787407476815e-05, "loss": 0.3697, "step": 8074 }, { "epoch": 0.34045154626135715, "grad_norm": 0.4492964446544647, "learning_rate": 4.925200662106874e-05, "loss": 0.4373, "step": 8075 }, { "epoch": 0.3404937074435567, "grad_norm": 1.4320497512817383, "learning_rate": 4.925122583466068e-05, "loss": 0.5298, "step": 8076 }, { "epoch": 0.34053586862575624, "grad_norm": 0.5134016871452332, "learning_rate": 4.92504450482526e-05, "loss": 0.6569, "step": 8077 }, { "epoch": 0.3405780298079558, "grad_norm": 1.0817056894302368, "learning_rate": 4.9249664261844534e-05, "loss": 0.1467, "step": 8078 }, { "epoch": 0.3406201909901554, "grad_norm": 1.0260286331176758, "learning_rate": 4.9248883475436465e-05, "loss": 0.4317, "step": 8079 }, { "epoch": 0.3406623521723549, "grad_norm": 0.6888648867607117, "learning_rate": 4.924810268902839e-05, "loss": 0.2769, "step": 8080 }, { "epoch": 0.3407045133545545, "grad_norm": 0.3529675304889679, "learning_rate": 4.924732190262032e-05, "loss": 0.1379, "step": 8081 }, { "epoch": 0.340746674536754, "grad_norm": 0.7197890877723694, "learning_rate": 4.924654111621225e-05, "loss": 0.2134, "step": 8082 }, { "epoch": 0.3407888357189536, "grad_norm": 0.5730918645858765, "learning_rate": 4.924576032980418e-05, "loss": 0.2201, "step": 8083 }, { "epoch": 0.3408309969011531, "grad_norm": 1.0391291379928589, "learning_rate": 4.9244979543396114e-05, "loss": 1.1826, "step": 8084 }, { "epoch": 0.34087315808335267, "grad_norm": 0.35664528608322144, "learning_rate": 4.924419875698804e-05, "loss": 0.1146, "step": 8085 }, { "epoch": 0.3409153192655522, "grad_norm": 0.5737833976745605, "learning_rate": 4.924341797057997e-05, "loss": 0.1658, "step": 8086 }, { "epoch": 0.34095748044775176, "grad_norm": 0.8127269148826599, "learning_rate": 4.92426371841719e-05, "loss": 0.5632, "step": 8087 }, { "epoch": 0.3409996416299513, "grad_norm": 0.567973792552948, "learning_rate": 4.924185639776383e-05, "loss": 0.2403, "step": 8088 }, { "epoch": 0.34104180281215085, "grad_norm": 0.8700563311576843, "learning_rate": 4.924107561135576e-05, "loss": 0.5786, "step": 8089 }, { "epoch": 0.3410839639943504, "grad_norm": 1.0254383087158203, "learning_rate": 4.924029482494769e-05, "loss": 0.5461, "step": 8090 }, { "epoch": 0.34112612517654994, "grad_norm": 1.1710824966430664, "learning_rate": 4.923951403853962e-05, "loss": 0.3728, "step": 8091 }, { "epoch": 0.3411682863587495, "grad_norm": 1.398705005645752, "learning_rate": 4.9238733252131544e-05, "loss": 0.4069, "step": 8092 }, { "epoch": 0.34121044754094904, "grad_norm": 0.768267035484314, "learning_rate": 4.923795246572348e-05, "loss": 0.2647, "step": 8093 }, { "epoch": 0.3412526087231486, "grad_norm": 0.372043251991272, "learning_rate": 4.923717167931541e-05, "loss": 0.0947, "step": 8094 }, { "epoch": 0.3412947699053481, "grad_norm": 0.49565643072128296, "learning_rate": 4.923639089290734e-05, "loss": 0.1777, "step": 8095 }, { "epoch": 0.3413369310875477, "grad_norm": 1.55570387840271, "learning_rate": 4.923561010649927e-05, "loss": 0.5352, "step": 8096 }, { "epoch": 0.3413790922697472, "grad_norm": 10.624991416931152, "learning_rate": 4.9234829320091194e-05, "loss": 0.2953, "step": 8097 }, { "epoch": 0.3414212534519468, "grad_norm": 0.5462278723716736, "learning_rate": 4.923404853368313e-05, "loss": 0.1891, "step": 8098 }, { "epoch": 0.34146341463414637, "grad_norm": 0.6385743618011475, "learning_rate": 4.9233267747275056e-05, "loss": 0.1346, "step": 8099 }, { "epoch": 0.3415055758163459, "grad_norm": 0.9954359531402588, "learning_rate": 4.923248696086699e-05, "loss": 0.5177, "step": 8100 }, { "epoch": 0.34154773699854546, "grad_norm": 0.584135890007019, "learning_rate": 4.923170617445892e-05, "loss": 0.1905, "step": 8101 }, { "epoch": 0.341589898180745, "grad_norm": 0.8747541308403015, "learning_rate": 4.923092538805085e-05, "loss": 1.0173, "step": 8102 }, { "epoch": 0.34163205936294455, "grad_norm": 0.3596372604370117, "learning_rate": 4.9230144601642774e-05, "loss": 0.1019, "step": 8103 }, { "epoch": 0.34167422054514407, "grad_norm": 0.512677788734436, "learning_rate": 4.9229363815234706e-05, "loss": 0.2131, "step": 8104 }, { "epoch": 0.34171638172734364, "grad_norm": 0.7778959274291992, "learning_rate": 4.922858302882664e-05, "loss": 0.3908, "step": 8105 }, { "epoch": 0.34175854290954316, "grad_norm": 0.6080349087715149, "learning_rate": 4.922780224241856e-05, "loss": 0.5249, "step": 8106 }, { "epoch": 0.34180070409174274, "grad_norm": 0.892920732498169, "learning_rate": 4.92270214560105e-05, "loss": 0.4467, "step": 8107 }, { "epoch": 0.3418428652739423, "grad_norm": 1.1233829259872437, "learning_rate": 4.9226240669602424e-05, "loss": 1.1038, "step": 8108 }, { "epoch": 0.3418850264561418, "grad_norm": 0.607464075088501, "learning_rate": 4.9225459883194355e-05, "loss": 0.165, "step": 8109 }, { "epoch": 0.3419271876383414, "grad_norm": 0.6703691482543945, "learning_rate": 4.9224679096786287e-05, "loss": 0.1324, "step": 8110 }, { "epoch": 0.3419693488205409, "grad_norm": 0.6978781223297119, "learning_rate": 4.922389831037821e-05, "loss": 0.5854, "step": 8111 }, { "epoch": 0.3420115100027405, "grad_norm": 0.44588613510131836, "learning_rate": 4.922311752397015e-05, "loss": 0.1448, "step": 8112 }, { "epoch": 0.34205367118494, "grad_norm": 1.1374404430389404, "learning_rate": 4.9222336737562074e-05, "loss": 1.0715, "step": 8113 }, { "epoch": 0.3420958323671396, "grad_norm": 1.658758521080017, "learning_rate": 4.9221555951154005e-05, "loss": 0.3857, "step": 8114 }, { "epoch": 0.3421379935493391, "grad_norm": 1.2406038045883179, "learning_rate": 4.9220775164745936e-05, "loss": 0.2207, "step": 8115 }, { "epoch": 0.3421801547315387, "grad_norm": 0.6087661385536194, "learning_rate": 4.921999437833786e-05, "loss": 0.1957, "step": 8116 }, { "epoch": 0.3422223159137382, "grad_norm": 0.5163567066192627, "learning_rate": 4.921921359192979e-05, "loss": 0.1332, "step": 8117 }, { "epoch": 0.34226447709593777, "grad_norm": 0.8734192252159119, "learning_rate": 4.921843280552172e-05, "loss": 0.3908, "step": 8118 }, { "epoch": 0.34230663827813734, "grad_norm": 0.6130973100662231, "learning_rate": 4.9217652019113654e-05, "loss": 0.3168, "step": 8119 }, { "epoch": 0.34234879946033686, "grad_norm": 1.2662277221679688, "learning_rate": 4.9216871232705586e-05, "loss": 0.2177, "step": 8120 }, { "epoch": 0.34239096064253643, "grad_norm": 1.0178054571151733, "learning_rate": 4.921609044629752e-05, "loss": 0.2435, "step": 8121 }, { "epoch": 0.34243312182473595, "grad_norm": 0.6664835810661316, "learning_rate": 4.921530965988944e-05, "loss": 0.5877, "step": 8122 }, { "epoch": 0.3424752830069355, "grad_norm": 1.9036879539489746, "learning_rate": 4.921452887348137e-05, "loss": 0.617, "step": 8123 }, { "epoch": 0.34251744418913505, "grad_norm": 0.4981796145439148, "learning_rate": 4.9213748087073304e-05, "loss": 0.2115, "step": 8124 }, { "epoch": 0.3425596053713346, "grad_norm": 2.3043391704559326, "learning_rate": 4.921296730066523e-05, "loss": 0.6377, "step": 8125 }, { "epoch": 0.34260176655353414, "grad_norm": 0.905528724193573, "learning_rate": 4.9212186514257166e-05, "loss": 1.0414, "step": 8126 }, { "epoch": 0.3426439277357337, "grad_norm": 1.0129384994506836, "learning_rate": 4.921140572784909e-05, "loss": 0.2981, "step": 8127 }, { "epoch": 0.3426860889179333, "grad_norm": 0.8945870995521545, "learning_rate": 4.921062494144102e-05, "loss": 0.6819, "step": 8128 }, { "epoch": 0.3427282501001328, "grad_norm": 0.5153387784957886, "learning_rate": 4.920984415503295e-05, "loss": 0.1636, "step": 8129 }, { "epoch": 0.3427704112823324, "grad_norm": 0.6875903010368347, "learning_rate": 4.920906336862488e-05, "loss": 0.093, "step": 8130 }, { "epoch": 0.3428125724645319, "grad_norm": 0.8099156618118286, "learning_rate": 4.9208282582216816e-05, "loss": 0.2419, "step": 8131 }, { "epoch": 0.34285473364673147, "grad_norm": 0.7206990122795105, "learning_rate": 4.920750179580874e-05, "loss": 0.1731, "step": 8132 }, { "epoch": 0.342896894828931, "grad_norm": 7.509542942047119, "learning_rate": 4.920672100940067e-05, "loss": 1.2231, "step": 8133 }, { "epoch": 0.34293905601113056, "grad_norm": 0.8200409412384033, "learning_rate": 4.92059402229926e-05, "loss": 0.3497, "step": 8134 }, { "epoch": 0.3429812171933301, "grad_norm": 1.192548155784607, "learning_rate": 4.920515943658453e-05, "loss": 0.9644, "step": 8135 }, { "epoch": 0.34302337837552965, "grad_norm": 0.2878390848636627, "learning_rate": 4.920437865017646e-05, "loss": 0.0666, "step": 8136 }, { "epoch": 0.34306553955772917, "grad_norm": 1.2640140056610107, "learning_rate": 4.920359786376839e-05, "loss": 0.1523, "step": 8137 }, { "epoch": 0.34310770073992874, "grad_norm": 1.0383695363998413, "learning_rate": 4.920281707736032e-05, "loss": 0.5249, "step": 8138 }, { "epoch": 0.3431498619221283, "grad_norm": 2.2894670963287354, "learning_rate": 4.9202036290952246e-05, "loss": 0.158, "step": 8139 }, { "epoch": 0.34319202310432784, "grad_norm": 0.6366100311279297, "learning_rate": 4.9201255504544184e-05, "loss": 0.5996, "step": 8140 }, { "epoch": 0.3432341842865274, "grad_norm": 0.606626570224762, "learning_rate": 4.920047471813611e-05, "loss": 0.2042, "step": 8141 }, { "epoch": 0.34327634546872693, "grad_norm": 0.9369965195655823, "learning_rate": 4.919969393172804e-05, "loss": 0.6103, "step": 8142 }, { "epoch": 0.3433185066509265, "grad_norm": 1.001081943511963, "learning_rate": 4.919891314531997e-05, "loss": 0.8189, "step": 8143 }, { "epoch": 0.343360667833126, "grad_norm": 1.1050695180892944, "learning_rate": 4.9198132358911895e-05, "loss": 0.1075, "step": 8144 }, { "epoch": 0.3434028290153256, "grad_norm": 0.6880106329917908, "learning_rate": 4.919735157250383e-05, "loss": 0.5634, "step": 8145 }, { "epoch": 0.3434449901975251, "grad_norm": 0.5617035031318665, "learning_rate": 4.919657078609576e-05, "loss": 0.4155, "step": 8146 }, { "epoch": 0.3434871513797247, "grad_norm": 1.0263760089874268, "learning_rate": 4.919578999968769e-05, "loss": 0.895, "step": 8147 }, { "epoch": 0.34352931256192426, "grad_norm": 1.137422800064087, "learning_rate": 4.919500921327962e-05, "loss": 0.3889, "step": 8148 }, { "epoch": 0.3435714737441238, "grad_norm": 1.4523040056228638, "learning_rate": 4.9194228426871545e-05, "loss": 0.3635, "step": 8149 }, { "epoch": 0.34361363492632335, "grad_norm": 0.7915636301040649, "learning_rate": 4.9193447640463476e-05, "loss": 0.5471, "step": 8150 }, { "epoch": 0.34365579610852287, "grad_norm": 1.4451078176498413, "learning_rate": 4.919266685405541e-05, "loss": 0.3133, "step": 8151 }, { "epoch": 0.34369795729072244, "grad_norm": 1.8475862741470337, "learning_rate": 4.919188606764734e-05, "loss": 0.5639, "step": 8152 }, { "epoch": 0.34374011847292196, "grad_norm": 0.49956876039505005, "learning_rate": 4.919110528123927e-05, "loss": 0.3003, "step": 8153 }, { "epoch": 0.34378227965512154, "grad_norm": 0.8117635250091553, "learning_rate": 4.9190324494831194e-05, "loss": 0.2531, "step": 8154 }, { "epoch": 0.34382444083732105, "grad_norm": 0.7829174399375916, "learning_rate": 4.9189543708423125e-05, "loss": 0.6577, "step": 8155 }, { "epoch": 0.34386660201952063, "grad_norm": 0.583089292049408, "learning_rate": 4.918876292201506e-05, "loss": 0.0999, "step": 8156 }, { "epoch": 0.3439087632017202, "grad_norm": 0.4460810720920563, "learning_rate": 4.918798213560699e-05, "loss": 0.4679, "step": 8157 }, { "epoch": 0.3439509243839197, "grad_norm": 0.5740422010421753, "learning_rate": 4.918720134919891e-05, "loss": 0.0713, "step": 8158 }, { "epoch": 0.3439930855661193, "grad_norm": 0.8356308341026306, "learning_rate": 4.9186420562790844e-05, "loss": 0.2069, "step": 8159 }, { "epoch": 0.3440352467483188, "grad_norm": 0.2725408673286438, "learning_rate": 4.9185639776382775e-05, "loss": 0.0732, "step": 8160 }, { "epoch": 0.3440774079305184, "grad_norm": 1.2562973499298096, "learning_rate": 4.91848589899747e-05, "loss": 1.0866, "step": 8161 }, { "epoch": 0.3441195691127179, "grad_norm": 1.0177119970321655, "learning_rate": 4.918407820356664e-05, "loss": 0.6577, "step": 8162 }, { "epoch": 0.3441617302949175, "grad_norm": 0.5339298248291016, "learning_rate": 4.918329741715856e-05, "loss": 0.1714, "step": 8163 }, { "epoch": 0.344203891477117, "grad_norm": 0.4724736213684082, "learning_rate": 4.91825166307505e-05, "loss": 0.1036, "step": 8164 }, { "epoch": 0.34424605265931657, "grad_norm": 0.7562977075576782, "learning_rate": 4.9181735844342425e-05, "loss": 0.5587, "step": 8165 }, { "epoch": 0.3442882138415161, "grad_norm": 0.8705638647079468, "learning_rate": 4.9180955057934356e-05, "loss": 0.2102, "step": 8166 }, { "epoch": 0.34433037502371566, "grad_norm": 1.470144271850586, "learning_rate": 4.918017427152629e-05, "loss": 0.1589, "step": 8167 }, { "epoch": 0.34437253620591524, "grad_norm": 0.7573671936988831, "learning_rate": 4.917939348511821e-05, "loss": 0.1163, "step": 8168 }, { "epoch": 0.34441469738811475, "grad_norm": 1.0721402168273926, "learning_rate": 4.917861269871014e-05, "loss": 0.1455, "step": 8169 }, { "epoch": 0.34445685857031433, "grad_norm": 1.177837610244751, "learning_rate": 4.9177831912302074e-05, "loss": 1.0926, "step": 8170 }, { "epoch": 0.34449901975251385, "grad_norm": 1.5538066625595093, "learning_rate": 4.9177051125894005e-05, "loss": 0.4751, "step": 8171 }, { "epoch": 0.3445411809347134, "grad_norm": 1.1171818971633911, "learning_rate": 4.917627033948593e-05, "loss": 0.3211, "step": 8172 }, { "epoch": 0.34458334211691294, "grad_norm": 1.0794259309768677, "learning_rate": 4.917548955307786e-05, "loss": 0.5245, "step": 8173 }, { "epoch": 0.3446255032991125, "grad_norm": 0.388496458530426, "learning_rate": 4.917470876666979e-05, "loss": 0.159, "step": 8174 }, { "epoch": 0.34466766448131203, "grad_norm": 0.5551720261573792, "learning_rate": 4.917392798026172e-05, "loss": 0.4287, "step": 8175 }, { "epoch": 0.3447098256635116, "grad_norm": 0.562686562538147, "learning_rate": 4.9173147193853655e-05, "loss": 0.2239, "step": 8176 }, { "epoch": 0.3447519868457112, "grad_norm": 1.0258610248565674, "learning_rate": 4.917236640744558e-05, "loss": 0.3382, "step": 8177 }, { "epoch": 0.3447941480279107, "grad_norm": 0.4966539144515991, "learning_rate": 4.917158562103751e-05, "loss": 0.1501, "step": 8178 }, { "epoch": 0.34483630921011027, "grad_norm": 0.8372641205787659, "learning_rate": 4.917080483462944e-05, "loss": 0.1162, "step": 8179 }, { "epoch": 0.3448784703923098, "grad_norm": 0.6400594711303711, "learning_rate": 4.9170024048221366e-05, "loss": 0.5899, "step": 8180 }, { "epoch": 0.34492063157450936, "grad_norm": 1.2871953248977661, "learning_rate": 4.9169243261813304e-05, "loss": 0.5004, "step": 8181 }, { "epoch": 0.3449627927567089, "grad_norm": 0.4919203221797943, "learning_rate": 4.916846247540523e-05, "loss": 0.092, "step": 8182 }, { "epoch": 0.34500495393890845, "grad_norm": 0.8764297366142273, "learning_rate": 4.916768168899716e-05, "loss": 0.1774, "step": 8183 }, { "epoch": 0.345047115121108, "grad_norm": 0.8448684811592102, "learning_rate": 4.916690090258909e-05, "loss": 0.221, "step": 8184 }, { "epoch": 0.34508927630330755, "grad_norm": 1.4795928001403809, "learning_rate": 4.9166120116181016e-05, "loss": 0.2148, "step": 8185 }, { "epoch": 0.3451314374855071, "grad_norm": 0.5512458086013794, "learning_rate": 4.916533932977295e-05, "loss": 0.2547, "step": 8186 }, { "epoch": 0.34517359866770664, "grad_norm": 0.9922723174095154, "learning_rate": 4.916455854336488e-05, "loss": 0.1519, "step": 8187 }, { "epoch": 0.3452157598499062, "grad_norm": 1.1275241374969482, "learning_rate": 4.916377775695681e-05, "loss": 0.3859, "step": 8188 }, { "epoch": 0.34525792103210573, "grad_norm": 1.2965975999832153, "learning_rate": 4.916299697054874e-05, "loss": 0.5404, "step": 8189 }, { "epoch": 0.3453000822143053, "grad_norm": 2.307347536087036, "learning_rate": 4.916221618414067e-05, "loss": 0.5501, "step": 8190 }, { "epoch": 0.3453422433965048, "grad_norm": 0.697571337223053, "learning_rate": 4.9161435397732597e-05, "loss": 0.5928, "step": 8191 }, { "epoch": 0.3453844045787044, "grad_norm": 0.9145248532295227, "learning_rate": 4.916065461132453e-05, "loss": 0.6547, "step": 8192 }, { "epoch": 0.3454265657609039, "grad_norm": 45.65594482421875, "learning_rate": 4.915987382491646e-05, "loss": 0.1385, "step": 8193 }, { "epoch": 0.3454687269431035, "grad_norm": 1.0137875080108643, "learning_rate": 4.9159093038508384e-05, "loss": 0.8729, "step": 8194 }, { "epoch": 0.345510888125303, "grad_norm": 0.5952677726745605, "learning_rate": 4.915831225210032e-05, "loss": 0.1224, "step": 8195 }, { "epoch": 0.3455530493075026, "grad_norm": 0.6662888526916504, "learning_rate": 4.9157531465692246e-05, "loss": 0.1407, "step": 8196 }, { "epoch": 0.34559521048970215, "grad_norm": 0.8315304517745972, "learning_rate": 4.915675067928418e-05, "loss": 0.6937, "step": 8197 }, { "epoch": 0.34563737167190167, "grad_norm": 0.7999536991119385, "learning_rate": 4.915596989287611e-05, "loss": 0.6041, "step": 8198 }, { "epoch": 0.34567953285410125, "grad_norm": 0.669254720211029, "learning_rate": 4.915518910646803e-05, "loss": 0.0991, "step": 8199 }, { "epoch": 0.34572169403630076, "grad_norm": 0.9090996980667114, "learning_rate": 4.915440832005997e-05, "loss": 0.2413, "step": 8200 }, { "epoch": 0.34576385521850034, "grad_norm": 0.7121817469596863, "learning_rate": 4.9153627533651896e-05, "loss": 0.6363, "step": 8201 }, { "epoch": 0.34580601640069986, "grad_norm": 0.6684048771858215, "learning_rate": 4.915284674724383e-05, "loss": 0.0959, "step": 8202 }, { "epoch": 0.34584817758289943, "grad_norm": 0.7375221848487854, "learning_rate": 4.915206596083576e-05, "loss": 0.6426, "step": 8203 }, { "epoch": 0.34589033876509895, "grad_norm": 0.6599108576774597, "learning_rate": 4.915128517442768e-05, "loss": 0.5551, "step": 8204 }, { "epoch": 0.3459324999472985, "grad_norm": 0.882614254951477, "learning_rate": 4.9150504388019614e-05, "loss": 0.6284, "step": 8205 }, { "epoch": 0.3459746611294981, "grad_norm": 3.1460535526275635, "learning_rate": 4.9149723601611545e-05, "loss": 0.9597, "step": 8206 }, { "epoch": 0.3460168223116976, "grad_norm": 0.4566066563129425, "learning_rate": 4.9148942815203476e-05, "loss": 0.3218, "step": 8207 }, { "epoch": 0.3460589834938972, "grad_norm": 1.1702501773834229, "learning_rate": 4.91481620287954e-05, "loss": 0.3546, "step": 8208 }, { "epoch": 0.3461011446760967, "grad_norm": 5.342572212219238, "learning_rate": 4.914738124238734e-05, "loss": 0.169, "step": 8209 }, { "epoch": 0.3461433058582963, "grad_norm": 0.5942874550819397, "learning_rate": 4.9146600455979263e-05, "loss": 0.3521, "step": 8210 }, { "epoch": 0.3461854670404958, "grad_norm": 0.6043810248374939, "learning_rate": 4.9145819669571195e-05, "loss": 0.1681, "step": 8211 }, { "epoch": 0.34622762822269537, "grad_norm": 0.41327983140945435, "learning_rate": 4.9145038883163126e-05, "loss": 0.1524, "step": 8212 }, { "epoch": 0.3462697894048949, "grad_norm": 0.4030357301235199, "learning_rate": 4.914425809675505e-05, "loss": 0.1939, "step": 8213 }, { "epoch": 0.34631195058709446, "grad_norm": 0.6501353979110718, "learning_rate": 4.914347731034699e-05, "loss": 0.1442, "step": 8214 }, { "epoch": 0.34635411176929404, "grad_norm": 0.609803318977356, "learning_rate": 4.914269652393891e-05, "loss": 0.1929, "step": 8215 }, { "epoch": 0.34639627295149356, "grad_norm": 0.4814116358757019, "learning_rate": 4.9141915737530844e-05, "loss": 0.1238, "step": 8216 }, { "epoch": 0.34643843413369313, "grad_norm": 1.164642333984375, "learning_rate": 4.9141134951122775e-05, "loss": 0.25, "step": 8217 }, { "epoch": 0.34648059531589265, "grad_norm": 0.4184390902519226, "learning_rate": 4.91403541647147e-05, "loss": 0.0803, "step": 8218 }, { "epoch": 0.3465227564980922, "grad_norm": 0.37780025601387024, "learning_rate": 4.913957337830663e-05, "loss": 0.0997, "step": 8219 }, { "epoch": 0.34656491768029174, "grad_norm": 0.9251286387443542, "learning_rate": 4.913879259189856e-05, "loss": 0.1553, "step": 8220 }, { "epoch": 0.3466070788624913, "grad_norm": 1.1087037324905396, "learning_rate": 4.9138011805490494e-05, "loss": 0.9015, "step": 8221 }, { "epoch": 0.34664924004469083, "grad_norm": 0.578328013420105, "learning_rate": 4.9137231019082425e-05, "loss": 0.1736, "step": 8222 }, { "epoch": 0.3466914012268904, "grad_norm": 0.413242369890213, "learning_rate": 4.913645023267435e-05, "loss": 0.1578, "step": 8223 }, { "epoch": 0.3467335624090899, "grad_norm": 0.8716453313827515, "learning_rate": 4.913566944626628e-05, "loss": 0.6742, "step": 8224 }, { "epoch": 0.3467757235912895, "grad_norm": 0.6573629975318909, "learning_rate": 4.913488865985821e-05, "loss": 0.1632, "step": 8225 }, { "epoch": 0.34681788477348907, "grad_norm": 2.080585479736328, "learning_rate": 4.913410787345014e-05, "loss": 0.2822, "step": 8226 }, { "epoch": 0.3468600459556886, "grad_norm": 1.2075209617614746, "learning_rate": 4.913332708704207e-05, "loss": 0.4682, "step": 8227 }, { "epoch": 0.34690220713788816, "grad_norm": 1.1032354831695557, "learning_rate": 4.9132546300634006e-05, "loss": 1.1042, "step": 8228 }, { "epoch": 0.3469443683200877, "grad_norm": 0.6572443246841431, "learning_rate": 4.913176551422593e-05, "loss": 0.5329, "step": 8229 }, { "epoch": 0.34698652950228726, "grad_norm": 0.6080334186553955, "learning_rate": 4.9130984727817855e-05, "loss": 0.2756, "step": 8230 }, { "epoch": 0.3470286906844868, "grad_norm": 0.7561857104301453, "learning_rate": 4.913020394140979e-05, "loss": 0.5458, "step": 8231 }, { "epoch": 0.34707085186668635, "grad_norm": 1.2226612567901611, "learning_rate": 4.912942315500172e-05, "loss": 1.0964, "step": 8232 }, { "epoch": 0.34711301304888587, "grad_norm": 3.1134204864501953, "learning_rate": 4.9128642368593655e-05, "loss": 0.7899, "step": 8233 }, { "epoch": 0.34715517423108544, "grad_norm": 0.5604845881462097, "learning_rate": 4.912786158218558e-05, "loss": 0.1881, "step": 8234 }, { "epoch": 0.347197335413285, "grad_norm": 0.5324761867523193, "learning_rate": 4.912708079577751e-05, "loss": 0.2265, "step": 8235 }, { "epoch": 0.34723949659548453, "grad_norm": 0.32806456089019775, "learning_rate": 4.912630000936944e-05, "loss": 0.102, "step": 8236 }, { "epoch": 0.3472816577776841, "grad_norm": 2.007974624633789, "learning_rate": 4.912551922296137e-05, "loss": 0.523, "step": 8237 }, { "epoch": 0.3473238189598836, "grad_norm": 1.7843787670135498, "learning_rate": 4.91247384365533e-05, "loss": 0.4504, "step": 8238 }, { "epoch": 0.3473659801420832, "grad_norm": 0.3316602408885956, "learning_rate": 4.912395765014523e-05, "loss": 0.166, "step": 8239 }, { "epoch": 0.3474081413242827, "grad_norm": 0.32845985889434814, "learning_rate": 4.912317686373716e-05, "loss": 0.0906, "step": 8240 }, { "epoch": 0.3474503025064823, "grad_norm": 0.4889596998691559, "learning_rate": 4.9122396077329085e-05, "loss": 0.1451, "step": 8241 }, { "epoch": 0.3474924636886818, "grad_norm": 0.548671305179596, "learning_rate": 4.9121615290921016e-05, "loss": 0.2418, "step": 8242 }, { "epoch": 0.3475346248708814, "grad_norm": 0.41443145275115967, "learning_rate": 4.912083450451295e-05, "loss": 0.1377, "step": 8243 }, { "epoch": 0.34757678605308095, "grad_norm": 0.9150800704956055, "learning_rate": 4.912005371810488e-05, "loss": 0.2395, "step": 8244 }, { "epoch": 0.3476189472352805, "grad_norm": 0.6230708956718445, "learning_rate": 4.911927293169681e-05, "loss": 0.2039, "step": 8245 }, { "epoch": 0.34766110841748005, "grad_norm": 0.6320980191230774, "learning_rate": 4.9118492145288735e-05, "loss": 0.3163, "step": 8246 }, { "epoch": 0.34770326959967957, "grad_norm": 0.9163508415222168, "learning_rate": 4.9117711358880666e-05, "loss": 0.7313, "step": 8247 }, { "epoch": 0.34774543078187914, "grad_norm": 0.5322421789169312, "learning_rate": 4.91169305724726e-05, "loss": 0.2128, "step": 8248 }, { "epoch": 0.34778759196407866, "grad_norm": 0.4777228534221649, "learning_rate": 4.911614978606452e-05, "loss": 0.1345, "step": 8249 }, { "epoch": 0.34782975314627823, "grad_norm": 0.7446433305740356, "learning_rate": 4.911536899965646e-05, "loss": 0.5651, "step": 8250 }, { "epoch": 0.34787191432847775, "grad_norm": 0.9395846128463745, "learning_rate": 4.9114588213248384e-05, "loss": 0.3371, "step": 8251 }, { "epoch": 0.3479140755106773, "grad_norm": 0.5683501958847046, "learning_rate": 4.9113807426840315e-05, "loss": 0.1469, "step": 8252 }, { "epoch": 0.34795623669287684, "grad_norm": 0.6007663607597351, "learning_rate": 4.911302664043225e-05, "loss": 0.6798, "step": 8253 }, { "epoch": 0.3479983978750764, "grad_norm": 0.46255412697792053, "learning_rate": 4.911224585402418e-05, "loss": 0.189, "step": 8254 }, { "epoch": 0.348040559057276, "grad_norm": 2.0718350410461426, "learning_rate": 4.91114650676161e-05, "loss": 0.9757, "step": 8255 }, { "epoch": 0.3480827202394755, "grad_norm": 0.45408859848976135, "learning_rate": 4.9110684281208034e-05, "loss": 0.1364, "step": 8256 }, { "epoch": 0.3481248814216751, "grad_norm": 0.5561493635177612, "learning_rate": 4.9109903494799965e-05, "loss": 0.245, "step": 8257 }, { "epoch": 0.3481670426038746, "grad_norm": 0.6048231720924377, "learning_rate": 4.9109122708391896e-05, "loss": 0.0842, "step": 8258 }, { "epoch": 0.3482092037860742, "grad_norm": 0.7203450202941895, "learning_rate": 4.910834192198383e-05, "loss": 0.2755, "step": 8259 }, { "epoch": 0.3482513649682737, "grad_norm": 0.8980373740196228, "learning_rate": 4.910756113557575e-05, "loss": 0.1643, "step": 8260 }, { "epoch": 0.34829352615047326, "grad_norm": 0.7202624678611755, "learning_rate": 4.910678034916768e-05, "loss": 0.6007, "step": 8261 }, { "epoch": 0.3483356873326728, "grad_norm": 0.5288437604904175, "learning_rate": 4.9105999562759614e-05, "loss": 0.2397, "step": 8262 }, { "epoch": 0.34837784851487236, "grad_norm": 0.618811845779419, "learning_rate": 4.910521877635154e-05, "loss": 0.3137, "step": 8263 }, { "epoch": 0.34842000969707193, "grad_norm": 0.6066488027572632, "learning_rate": 4.910443798994348e-05, "loss": 0.5012, "step": 8264 }, { "epoch": 0.34846217087927145, "grad_norm": 0.9404746890068054, "learning_rate": 4.91036572035354e-05, "loss": 0.1636, "step": 8265 }, { "epoch": 0.348504332061471, "grad_norm": 0.6437553763389587, "learning_rate": 4.910287641712733e-05, "loss": 0.4032, "step": 8266 }, { "epoch": 0.34854649324367054, "grad_norm": 3.887620210647583, "learning_rate": 4.9102095630719264e-05, "loss": 1.1301, "step": 8267 }, { "epoch": 0.3485886544258701, "grad_norm": 0.6287360191345215, "learning_rate": 4.910131484431119e-05, "loss": 0.4029, "step": 8268 }, { "epoch": 0.34863081560806963, "grad_norm": 0.7080799341201782, "learning_rate": 4.9100534057903126e-05, "loss": 0.2101, "step": 8269 }, { "epoch": 0.3486729767902692, "grad_norm": 1.1026835441589355, "learning_rate": 4.909975327149505e-05, "loss": 0.9317, "step": 8270 }, { "epoch": 0.3487151379724687, "grad_norm": 1.4312191009521484, "learning_rate": 4.909897248508698e-05, "loss": 1.0219, "step": 8271 }, { "epoch": 0.3487572991546683, "grad_norm": 0.890484631061554, "learning_rate": 4.9098191698678913e-05, "loss": 1.0449, "step": 8272 }, { "epoch": 0.3487994603368679, "grad_norm": 0.6403425335884094, "learning_rate": 4.9097410912270845e-05, "loss": 0.7157, "step": 8273 }, { "epoch": 0.3488416215190674, "grad_norm": 1.3113322257995605, "learning_rate": 4.909663012586277e-05, "loss": 0.1814, "step": 8274 }, { "epoch": 0.34888378270126696, "grad_norm": 1.0481042861938477, "learning_rate": 4.90958493394547e-05, "loss": 0.1928, "step": 8275 }, { "epoch": 0.3489259438834665, "grad_norm": 0.6375716924667358, "learning_rate": 4.909506855304663e-05, "loss": 0.1824, "step": 8276 }, { "epoch": 0.34896810506566606, "grad_norm": 0.8790286183357239, "learning_rate": 4.9094287766638556e-05, "loss": 0.1176, "step": 8277 }, { "epoch": 0.3490102662478656, "grad_norm": 0.36752915382385254, "learning_rate": 4.9093506980230494e-05, "loss": 0.1277, "step": 8278 }, { "epoch": 0.34905242743006515, "grad_norm": 8.855587005615234, "learning_rate": 4.909272619382242e-05, "loss": 0.1905, "step": 8279 }, { "epoch": 0.34909458861226467, "grad_norm": 0.5440651774406433, "learning_rate": 4.909194540741435e-05, "loss": 0.1894, "step": 8280 }, { "epoch": 0.34913674979446424, "grad_norm": 0.6741387844085693, "learning_rate": 4.909116462100628e-05, "loss": 0.2355, "step": 8281 }, { "epoch": 0.34917891097666376, "grad_norm": 0.6678037643432617, "learning_rate": 4.9090383834598206e-05, "loss": 0.1139, "step": 8282 }, { "epoch": 0.34922107215886333, "grad_norm": 0.427477091550827, "learning_rate": 4.9089603048190144e-05, "loss": 0.2321, "step": 8283 }, { "epoch": 0.3492632333410629, "grad_norm": 1.2891936302185059, "learning_rate": 4.908882226178207e-05, "loss": 0.1425, "step": 8284 }, { "epoch": 0.3493053945232624, "grad_norm": 0.46677565574645996, "learning_rate": 4.9088041475374e-05, "loss": 0.1335, "step": 8285 }, { "epoch": 0.349347555705462, "grad_norm": 0.7946088314056396, "learning_rate": 4.908726068896593e-05, "loss": 0.42, "step": 8286 }, { "epoch": 0.3493897168876615, "grad_norm": 0.39591166377067566, "learning_rate": 4.9086479902557855e-05, "loss": 0.1025, "step": 8287 }, { "epoch": 0.3494318780698611, "grad_norm": 0.45990070700645447, "learning_rate": 4.9085699116149787e-05, "loss": 0.2043, "step": 8288 }, { "epoch": 0.3494740392520606, "grad_norm": 0.5253873467445374, "learning_rate": 4.908491832974172e-05, "loss": 0.154, "step": 8289 }, { "epoch": 0.3495162004342602, "grad_norm": 0.719657301902771, "learning_rate": 4.908413754333365e-05, "loss": 0.5276, "step": 8290 }, { "epoch": 0.3495583616164597, "grad_norm": 0.4760863184928894, "learning_rate": 4.908335675692558e-05, "loss": 0.0907, "step": 8291 }, { "epoch": 0.3496005227986593, "grad_norm": 0.9849464297294617, "learning_rate": 4.9082575970517505e-05, "loss": 0.394, "step": 8292 }, { "epoch": 0.34964268398085885, "grad_norm": 1.1508723497390747, "learning_rate": 4.9081795184109436e-05, "loss": 0.2992, "step": 8293 }, { "epoch": 0.34968484516305837, "grad_norm": 0.4154919981956482, "learning_rate": 4.908101439770137e-05, "loss": 0.1525, "step": 8294 }, { "epoch": 0.34972700634525794, "grad_norm": 0.8358413577079773, "learning_rate": 4.90802336112933e-05, "loss": 0.6202, "step": 8295 }, { "epoch": 0.34976916752745746, "grad_norm": 0.6180986762046814, "learning_rate": 4.907945282488522e-05, "loss": 0.0857, "step": 8296 }, { "epoch": 0.34981132870965703, "grad_norm": 1.215909719467163, "learning_rate": 4.907867203847716e-05, "loss": 1.1235, "step": 8297 }, { "epoch": 0.34985348989185655, "grad_norm": 3.001148223876953, "learning_rate": 4.9077891252069086e-05, "loss": 0.8113, "step": 8298 }, { "epoch": 0.3498956510740561, "grad_norm": 0.7796746492385864, "learning_rate": 4.907711046566102e-05, "loss": 0.588, "step": 8299 }, { "epoch": 0.34993781225625564, "grad_norm": 0.42084670066833496, "learning_rate": 4.907632967925295e-05, "loss": 0.1418, "step": 8300 }, { "epoch": 0.3499799734384552, "grad_norm": 1.7062276601791382, "learning_rate": 4.907554889284487e-05, "loss": 1.0848, "step": 8301 }, { "epoch": 0.3500221346206548, "grad_norm": 7.940903186798096, "learning_rate": 4.907476810643681e-05, "loss": 0.1826, "step": 8302 }, { "epoch": 0.3500642958028543, "grad_norm": 1.1097238063812256, "learning_rate": 4.9073987320028735e-05, "loss": 0.95, "step": 8303 }, { "epoch": 0.3501064569850539, "grad_norm": 0.34676647186279297, "learning_rate": 4.9073206533620666e-05, "loss": 0.1226, "step": 8304 }, { "epoch": 0.3501486181672534, "grad_norm": 0.5250118374824524, "learning_rate": 4.90724257472126e-05, "loss": 0.5776, "step": 8305 }, { "epoch": 0.350190779349453, "grad_norm": 1.7831671237945557, "learning_rate": 4.907164496080452e-05, "loss": 0.923, "step": 8306 }, { "epoch": 0.3502329405316525, "grad_norm": 0.6019322276115417, "learning_rate": 4.907086417439645e-05, "loss": 0.6392, "step": 8307 }, { "epoch": 0.35027510171385207, "grad_norm": 0.3853985667228699, "learning_rate": 4.9070083387988385e-05, "loss": 0.1581, "step": 8308 }, { "epoch": 0.3503172628960516, "grad_norm": 0.6053529381752014, "learning_rate": 4.9069302601580316e-05, "loss": 0.6268, "step": 8309 }, { "epoch": 0.35035942407825116, "grad_norm": 0.638621985912323, "learning_rate": 4.906852181517224e-05, "loss": 0.8565, "step": 8310 }, { "epoch": 0.3504015852604507, "grad_norm": 0.7425834536552429, "learning_rate": 4.906774102876417e-05, "loss": 0.1246, "step": 8311 }, { "epoch": 0.35044374644265025, "grad_norm": 1.0776597261428833, "learning_rate": 4.90669602423561e-05, "loss": 0.1713, "step": 8312 }, { "epoch": 0.3504859076248498, "grad_norm": 0.7136482000350952, "learning_rate": 4.9066179455948034e-05, "loss": 0.1712, "step": 8313 }, { "epoch": 0.35052806880704934, "grad_norm": 0.5790529251098633, "learning_rate": 4.9065398669539965e-05, "loss": 0.4928, "step": 8314 }, { "epoch": 0.3505702299892489, "grad_norm": 1.7941505908966064, "learning_rate": 4.906461788313189e-05, "loss": 1.0639, "step": 8315 }, { "epoch": 0.35061239117144843, "grad_norm": 1.315393328666687, "learning_rate": 4.906383709672383e-05, "loss": 0.9779, "step": 8316 }, { "epoch": 0.350654552353648, "grad_norm": 0.7305330634117126, "learning_rate": 4.906305631031575e-05, "loss": 0.2984, "step": 8317 }, { "epoch": 0.3506967135358475, "grad_norm": 1.3179292678833008, "learning_rate": 4.9062275523907684e-05, "loss": 0.837, "step": 8318 }, { "epoch": 0.3507388747180471, "grad_norm": 0.3374815285205841, "learning_rate": 4.9061494737499615e-05, "loss": 0.1392, "step": 8319 }, { "epoch": 0.3507810359002466, "grad_norm": 0.7863338589668274, "learning_rate": 4.906071395109154e-05, "loss": 0.2705, "step": 8320 }, { "epoch": 0.3508231970824462, "grad_norm": 13.528050422668457, "learning_rate": 4.905993316468347e-05, "loss": 0.5595, "step": 8321 }, { "epoch": 0.35086535826464577, "grad_norm": 0.5824559926986694, "learning_rate": 4.90591523782754e-05, "loss": 0.4042, "step": 8322 }, { "epoch": 0.3509075194468453, "grad_norm": 1.545914888381958, "learning_rate": 4.905837159186733e-05, "loss": 0.8613, "step": 8323 }, { "epoch": 0.35094968062904486, "grad_norm": 0.6358683705329895, "learning_rate": 4.9057590805459264e-05, "loss": 0.3065, "step": 8324 }, { "epoch": 0.3509918418112444, "grad_norm": 0.6922004222869873, "learning_rate": 4.905681001905119e-05, "loss": 0.5904, "step": 8325 }, { "epoch": 0.35103400299344395, "grad_norm": 1.0303875207901, "learning_rate": 4.905602923264312e-05, "loss": 0.1281, "step": 8326 }, { "epoch": 0.35107616417564347, "grad_norm": 1.3805679082870483, "learning_rate": 4.905524844623505e-05, "loss": 1.1059, "step": 8327 }, { "epoch": 0.35111832535784304, "grad_norm": 1.2219979763031006, "learning_rate": 4.905446765982698e-05, "loss": 0.9043, "step": 8328 }, { "epoch": 0.35116048654004256, "grad_norm": 1.0000758171081543, "learning_rate": 4.905368687341891e-05, "loss": 0.2067, "step": 8329 }, { "epoch": 0.35120264772224213, "grad_norm": 0.42114633321762085, "learning_rate": 4.905290608701084e-05, "loss": 0.1113, "step": 8330 }, { "epoch": 0.3512448089044417, "grad_norm": 0.8087896704673767, "learning_rate": 4.905212530060277e-05, "loss": 0.2275, "step": 8331 }, { "epoch": 0.3512869700866412, "grad_norm": 4.380801200866699, "learning_rate": 4.9051344514194694e-05, "loss": 0.3333, "step": 8332 }, { "epoch": 0.3513291312688408, "grad_norm": 0.8893001079559326, "learning_rate": 4.905056372778663e-05, "loss": 0.461, "step": 8333 }, { "epoch": 0.3513712924510403, "grad_norm": 0.495584100484848, "learning_rate": 4.904978294137856e-05, "loss": 0.1472, "step": 8334 }, { "epoch": 0.3514134536332399, "grad_norm": 0.7688215970993042, "learning_rate": 4.904900215497049e-05, "loss": 0.2567, "step": 8335 }, { "epoch": 0.3514556148154394, "grad_norm": 1.1640738248825073, "learning_rate": 4.904822136856242e-05, "loss": 0.6249, "step": 8336 }, { "epoch": 0.351497775997639, "grad_norm": 1.069346308708191, "learning_rate": 4.9047440582154344e-05, "loss": 0.392, "step": 8337 }, { "epoch": 0.3515399371798385, "grad_norm": 0.36234599351882935, "learning_rate": 4.904665979574628e-05, "loss": 0.0885, "step": 8338 }, { "epoch": 0.3515820983620381, "grad_norm": 23.85769271850586, "learning_rate": 4.9045879009338206e-05, "loss": 0.1895, "step": 8339 }, { "epoch": 0.3516242595442376, "grad_norm": 0.4708002507686615, "learning_rate": 4.904509822293014e-05, "loss": 0.208, "step": 8340 }, { "epoch": 0.35166642072643717, "grad_norm": 0.7719089388847351, "learning_rate": 4.904431743652207e-05, "loss": 0.2716, "step": 8341 }, { "epoch": 0.35170858190863674, "grad_norm": 6.531362056732178, "learning_rate": 4.9043536650114e-05, "loss": 0.9729, "step": 8342 }, { "epoch": 0.35175074309083626, "grad_norm": 0.4192299544811249, "learning_rate": 4.9042755863705924e-05, "loss": 0.0912, "step": 8343 }, { "epoch": 0.35179290427303583, "grad_norm": 0.47181421518325806, "learning_rate": 4.9041975077297856e-05, "loss": 0.0671, "step": 8344 }, { "epoch": 0.35183506545523535, "grad_norm": 0.36702802777290344, "learning_rate": 4.904119429088979e-05, "loss": 0.1376, "step": 8345 }, { "epoch": 0.3518772266374349, "grad_norm": 0.5542523860931396, "learning_rate": 4.904041350448171e-05, "loss": 0.1737, "step": 8346 }, { "epoch": 0.35191938781963444, "grad_norm": 1.4429131746292114, "learning_rate": 4.903963271807365e-05, "loss": 0.293, "step": 8347 }, { "epoch": 0.351961549001834, "grad_norm": 0.7045050859451294, "learning_rate": 4.9038851931665574e-05, "loss": 0.8226, "step": 8348 }, { "epoch": 0.35200371018403354, "grad_norm": 0.45016053318977356, "learning_rate": 4.9038071145257505e-05, "loss": 0.0981, "step": 8349 }, { "epoch": 0.3520458713662331, "grad_norm": 0.6664035320281982, "learning_rate": 4.9037290358849437e-05, "loss": 0.1557, "step": 8350 }, { "epoch": 0.3520880325484327, "grad_norm": 1.6465368270874023, "learning_rate": 4.903650957244136e-05, "loss": 1.1735, "step": 8351 }, { "epoch": 0.3521301937306322, "grad_norm": 1.2823779582977295, "learning_rate": 4.90357287860333e-05, "loss": 1.1154, "step": 8352 }, { "epoch": 0.3521723549128318, "grad_norm": 0.4143024981021881, "learning_rate": 4.9034947999625224e-05, "loss": 0.1348, "step": 8353 }, { "epoch": 0.3522145160950313, "grad_norm": 2.535191774368286, "learning_rate": 4.9034167213217155e-05, "loss": 0.3656, "step": 8354 }, { "epoch": 0.35225667727723087, "grad_norm": 13.136282920837402, "learning_rate": 4.9033386426809086e-05, "loss": 0.4166, "step": 8355 }, { "epoch": 0.3522988384594304, "grad_norm": 1.4109877347946167, "learning_rate": 4.903260564040101e-05, "loss": 0.6378, "step": 8356 }, { "epoch": 0.35234099964162996, "grad_norm": 0.7169075608253479, "learning_rate": 4.903182485399294e-05, "loss": 0.2799, "step": 8357 }, { "epoch": 0.3523831608238295, "grad_norm": 15.275254249572754, "learning_rate": 4.903104406758487e-05, "loss": 0.0763, "step": 8358 }, { "epoch": 0.35242532200602905, "grad_norm": 1.125933051109314, "learning_rate": 4.9030263281176804e-05, "loss": 0.2246, "step": 8359 }, { "epoch": 0.3524674831882286, "grad_norm": 0.6914225220680237, "learning_rate": 4.9029482494768736e-05, "loss": 0.1637, "step": 8360 }, { "epoch": 0.35250964437042814, "grad_norm": 0.4557643532752991, "learning_rate": 4.902870170836067e-05, "loss": 0.1324, "step": 8361 }, { "epoch": 0.3525518055526277, "grad_norm": 0.7578620910644531, "learning_rate": 4.902792092195259e-05, "loss": 0.1597, "step": 8362 }, { "epoch": 0.35259396673482724, "grad_norm": 0.5032753944396973, "learning_rate": 4.902714013554452e-05, "loss": 0.4308, "step": 8363 }, { "epoch": 0.3526361279170268, "grad_norm": 1.4237734079360962, "learning_rate": 4.9026359349136454e-05, "loss": 0.2847, "step": 8364 }, { "epoch": 0.3526782890992263, "grad_norm": 0.8011414408683777, "learning_rate": 4.902557856272838e-05, "loss": 0.8175, "step": 8365 }, { "epoch": 0.3527204502814259, "grad_norm": 0.5776323676109314, "learning_rate": 4.9024797776320316e-05, "loss": 0.5449, "step": 8366 }, { "epoch": 0.3527626114636254, "grad_norm": 0.6056849956512451, "learning_rate": 4.902401698991224e-05, "loss": 0.54, "step": 8367 }, { "epoch": 0.352804772645825, "grad_norm": 0.3055216372013092, "learning_rate": 4.902323620350417e-05, "loss": 0.1177, "step": 8368 }, { "epoch": 0.3528469338280245, "grad_norm": 0.8398019671440125, "learning_rate": 4.90224554170961e-05, "loss": 0.5647, "step": 8369 }, { "epoch": 0.3528890950102241, "grad_norm": 1.0430591106414795, "learning_rate": 4.902167463068803e-05, "loss": 1.066, "step": 8370 }, { "epoch": 0.35293125619242366, "grad_norm": 1.464484453201294, "learning_rate": 4.9020893844279966e-05, "loss": 1.0283, "step": 8371 }, { "epoch": 0.3529734173746232, "grad_norm": 0.4658258557319641, "learning_rate": 4.902011305787189e-05, "loss": 0.1267, "step": 8372 }, { "epoch": 0.35301557855682275, "grad_norm": 0.3835509717464447, "learning_rate": 4.901933227146382e-05, "loss": 0.1199, "step": 8373 }, { "epoch": 0.35305773973902227, "grad_norm": 0.7939370274543762, "learning_rate": 4.901855148505575e-05, "loss": 0.1669, "step": 8374 }, { "epoch": 0.35309990092122184, "grad_norm": 0.43747425079345703, "learning_rate": 4.901777069864768e-05, "loss": 0.1538, "step": 8375 }, { "epoch": 0.35314206210342136, "grad_norm": 0.41719964146614075, "learning_rate": 4.901698991223961e-05, "loss": 0.2386, "step": 8376 }, { "epoch": 0.35318422328562094, "grad_norm": 0.8568476438522339, "learning_rate": 4.901620912583154e-05, "loss": 0.2615, "step": 8377 }, { "epoch": 0.35322638446782045, "grad_norm": 0.39067551493644714, "learning_rate": 4.901542833942347e-05, "loss": 0.108, "step": 8378 }, { "epoch": 0.35326854565002, "grad_norm": 0.47794899344444275, "learning_rate": 4.9014647553015396e-05, "loss": 0.1895, "step": 8379 }, { "epoch": 0.3533107068322196, "grad_norm": 1.5772881507873535, "learning_rate": 4.9013866766607334e-05, "loss": 0.1498, "step": 8380 }, { "epoch": 0.3533528680144191, "grad_norm": 0.9944279193878174, "learning_rate": 4.901308598019926e-05, "loss": 0.1096, "step": 8381 }, { "epoch": 0.3533950291966187, "grad_norm": 2.1836211681365967, "learning_rate": 4.901230519379119e-05, "loss": 0.9879, "step": 8382 }, { "epoch": 0.3534371903788182, "grad_norm": 2.0391833782196045, "learning_rate": 4.901152440738312e-05, "loss": 0.9093, "step": 8383 }, { "epoch": 0.3534793515610178, "grad_norm": 0.7175149917602539, "learning_rate": 4.9010743620975045e-05, "loss": 0.178, "step": 8384 }, { "epoch": 0.3535215127432173, "grad_norm": 0.823894739151001, "learning_rate": 4.900996283456698e-05, "loss": 0.2852, "step": 8385 }, { "epoch": 0.3535636739254169, "grad_norm": 0.6758971214294434, "learning_rate": 4.900918204815891e-05, "loss": 0.3559, "step": 8386 }, { "epoch": 0.3536058351076164, "grad_norm": 0.6357594728469849, "learning_rate": 4.900840126175084e-05, "loss": 0.4613, "step": 8387 }, { "epoch": 0.35364799628981597, "grad_norm": 0.48769137263298035, "learning_rate": 4.900762047534277e-05, "loss": 0.1719, "step": 8388 }, { "epoch": 0.35369015747201554, "grad_norm": 0.5856233239173889, "learning_rate": 4.9006839688934695e-05, "loss": 0.6214, "step": 8389 }, { "epoch": 0.35373231865421506, "grad_norm": 0.5596374869346619, "learning_rate": 4.9006058902526626e-05, "loss": 0.2051, "step": 8390 }, { "epoch": 0.35377447983641463, "grad_norm": 0.8479266166687012, "learning_rate": 4.900527811611856e-05, "loss": 0.4649, "step": 8391 }, { "epoch": 0.35381664101861415, "grad_norm": 1.9642220735549927, "learning_rate": 4.900449732971049e-05, "loss": 0.2295, "step": 8392 }, { "epoch": 0.3538588022008137, "grad_norm": 0.7530664205551147, "learning_rate": 4.900371654330242e-05, "loss": 0.1953, "step": 8393 }, { "epoch": 0.35390096338301325, "grad_norm": 0.5801214575767517, "learning_rate": 4.9002935756894344e-05, "loss": 0.0972, "step": 8394 }, { "epoch": 0.3539431245652128, "grad_norm": 0.6642543077468872, "learning_rate": 4.9002154970486275e-05, "loss": 0.1721, "step": 8395 }, { "epoch": 0.35398528574741234, "grad_norm": 0.6693850159645081, "learning_rate": 4.900137418407821e-05, "loss": 0.1598, "step": 8396 }, { "epoch": 0.3540274469296119, "grad_norm": 1.2497717142105103, "learning_rate": 4.900059339767014e-05, "loss": 0.2901, "step": 8397 }, { "epoch": 0.35406960811181143, "grad_norm": 0.5536201000213623, "learning_rate": 4.899981261126206e-05, "loss": 0.6697, "step": 8398 }, { "epoch": 0.354111769294011, "grad_norm": 0.5434826016426086, "learning_rate": 4.8999031824853994e-05, "loss": 0.5249, "step": 8399 }, { "epoch": 0.3541539304762106, "grad_norm": 0.7641621232032776, "learning_rate": 4.8998251038445925e-05, "loss": 0.6058, "step": 8400 }, { "epoch": 0.3541960916584101, "grad_norm": 0.5294919013977051, "learning_rate": 4.899747025203785e-05, "loss": 0.2088, "step": 8401 }, { "epoch": 0.35423825284060967, "grad_norm": 0.7556337118148804, "learning_rate": 4.899668946562979e-05, "loss": 0.1832, "step": 8402 }, { "epoch": 0.3542804140228092, "grad_norm": 5.312205791473389, "learning_rate": 4.899590867922171e-05, "loss": 0.4783, "step": 8403 }, { "epoch": 0.35432257520500876, "grad_norm": 1.0061163902282715, "learning_rate": 4.899512789281365e-05, "loss": 0.1479, "step": 8404 }, { "epoch": 0.3543647363872083, "grad_norm": 1.1280280351638794, "learning_rate": 4.8994347106405575e-05, "loss": 0.955, "step": 8405 }, { "epoch": 0.35440689756940785, "grad_norm": 0.6791980862617493, "learning_rate": 4.8993566319997506e-05, "loss": 0.6264, "step": 8406 }, { "epoch": 0.35444905875160737, "grad_norm": 0.46700620651245117, "learning_rate": 4.899278553358944e-05, "loss": 0.0824, "step": 8407 }, { "epoch": 0.35449121993380694, "grad_norm": 0.8588984608650208, "learning_rate": 4.899200474718136e-05, "loss": 0.564, "step": 8408 }, { "epoch": 0.3545333811160065, "grad_norm": 1.1977756023406982, "learning_rate": 4.899122396077329e-05, "loss": 0.3951, "step": 8409 }, { "epoch": 0.35457554229820604, "grad_norm": 1.088083028793335, "learning_rate": 4.8990443174365224e-05, "loss": 1.0343, "step": 8410 }, { "epoch": 0.3546177034804056, "grad_norm": 0.5217148065567017, "learning_rate": 4.8989662387957155e-05, "loss": 0.1581, "step": 8411 }, { "epoch": 0.35465986466260513, "grad_norm": 0.770867645740509, "learning_rate": 4.898888160154908e-05, "loss": 0.1829, "step": 8412 }, { "epoch": 0.3547020258448047, "grad_norm": 0.6045000553131104, "learning_rate": 4.898810081514101e-05, "loss": 0.2982, "step": 8413 }, { "epoch": 0.3547441870270042, "grad_norm": 0.7236652374267578, "learning_rate": 4.898732002873294e-05, "loss": 0.1205, "step": 8414 }, { "epoch": 0.3547863482092038, "grad_norm": 0.9191716909408569, "learning_rate": 4.898653924232487e-05, "loss": 0.1305, "step": 8415 }, { "epoch": 0.3548285093914033, "grad_norm": 0.6973273158073425, "learning_rate": 4.8985758455916805e-05, "loss": 0.4443, "step": 8416 }, { "epoch": 0.3548706705736029, "grad_norm": 1.1601923704147339, "learning_rate": 4.898497766950873e-05, "loss": 0.4523, "step": 8417 }, { "epoch": 0.35491283175580246, "grad_norm": 1.5745205879211426, "learning_rate": 4.898419688310066e-05, "loss": 0.4025, "step": 8418 }, { "epoch": 0.354954992938002, "grad_norm": 0.8395532369613647, "learning_rate": 4.898341609669259e-05, "loss": 0.5917, "step": 8419 }, { "epoch": 0.35499715412020155, "grad_norm": 0.6461135149002075, "learning_rate": 4.8982635310284516e-05, "loss": 0.2638, "step": 8420 }, { "epoch": 0.35503931530240107, "grad_norm": 0.7624237537384033, "learning_rate": 4.8981854523876454e-05, "loss": 0.1442, "step": 8421 }, { "epoch": 0.35508147648460064, "grad_norm": 1.3974686861038208, "learning_rate": 4.898107373746838e-05, "loss": 0.5285, "step": 8422 }, { "epoch": 0.35512363766680016, "grad_norm": 0.5548469424247742, "learning_rate": 4.898029295106031e-05, "loss": 0.221, "step": 8423 }, { "epoch": 0.35516579884899974, "grad_norm": 0.7170133590698242, "learning_rate": 4.897951216465224e-05, "loss": 0.2056, "step": 8424 }, { "epoch": 0.35520796003119925, "grad_norm": 0.8360167741775513, "learning_rate": 4.897873137824417e-05, "loss": 0.3032, "step": 8425 }, { "epoch": 0.35525012121339883, "grad_norm": 0.8076806664466858, "learning_rate": 4.89779505918361e-05, "loss": 1.1101, "step": 8426 }, { "epoch": 0.35529228239559835, "grad_norm": 0.40313124656677246, "learning_rate": 4.897716980542803e-05, "loss": 0.1237, "step": 8427 }, { "epoch": 0.3553344435777979, "grad_norm": 0.6150571703910828, "learning_rate": 4.897638901901996e-05, "loss": 0.1174, "step": 8428 }, { "epoch": 0.3553766047599975, "grad_norm": 0.9932196140289307, "learning_rate": 4.897560823261189e-05, "loss": 1.3439, "step": 8429 }, { "epoch": 0.355418765942197, "grad_norm": 1.6642768383026123, "learning_rate": 4.897482744620382e-05, "loss": 0.5254, "step": 8430 }, { "epoch": 0.3554609271243966, "grad_norm": 0.8881227374076843, "learning_rate": 4.8974046659795747e-05, "loss": 0.2401, "step": 8431 }, { "epoch": 0.3555030883065961, "grad_norm": 0.8767439126968384, "learning_rate": 4.897326587338768e-05, "loss": 0.2661, "step": 8432 }, { "epoch": 0.3555452494887957, "grad_norm": 0.947659969329834, "learning_rate": 4.897248508697961e-05, "loss": 0.1084, "step": 8433 }, { "epoch": 0.3555874106709952, "grad_norm": 0.5537811517715454, "learning_rate": 4.8971704300571534e-05, "loss": 0.2315, "step": 8434 }, { "epoch": 0.35562957185319477, "grad_norm": 0.44123733043670654, "learning_rate": 4.897092351416347e-05, "loss": 0.154, "step": 8435 }, { "epoch": 0.3556717330353943, "grad_norm": 0.6138821244239807, "learning_rate": 4.8970142727755396e-05, "loss": 0.1213, "step": 8436 }, { "epoch": 0.35571389421759386, "grad_norm": 1.203760027885437, "learning_rate": 4.896936194134733e-05, "loss": 0.3076, "step": 8437 }, { "epoch": 0.35575605539979344, "grad_norm": 0.5225828289985657, "learning_rate": 4.896858115493926e-05, "loss": 0.1529, "step": 8438 }, { "epoch": 0.35579821658199295, "grad_norm": 0.6596587300300598, "learning_rate": 4.896780036853118e-05, "loss": 0.1384, "step": 8439 }, { "epoch": 0.35584037776419253, "grad_norm": 1.9582757949829102, "learning_rate": 4.896701958212312e-05, "loss": 0.2304, "step": 8440 }, { "epoch": 0.35588253894639205, "grad_norm": 0.5955097079277039, "learning_rate": 4.8966238795715046e-05, "loss": 0.1886, "step": 8441 }, { "epoch": 0.3559247001285916, "grad_norm": 0.7617220282554626, "learning_rate": 4.896545800930698e-05, "loss": 0.4601, "step": 8442 }, { "epoch": 0.35596686131079114, "grad_norm": 0.8740420341491699, "learning_rate": 4.896467722289891e-05, "loss": 0.231, "step": 8443 }, { "epoch": 0.3560090224929907, "grad_norm": 1.9373087882995605, "learning_rate": 4.896389643649083e-05, "loss": 0.2719, "step": 8444 }, { "epoch": 0.35605118367519023, "grad_norm": 4.136258125305176, "learning_rate": 4.8963115650082764e-05, "loss": 0.9291, "step": 8445 }, { "epoch": 0.3560933448573898, "grad_norm": 0.5368419885635376, "learning_rate": 4.8962334863674695e-05, "loss": 0.2585, "step": 8446 }, { "epoch": 0.3561355060395893, "grad_norm": 0.5425662398338318, "learning_rate": 4.8961554077266626e-05, "loss": 0.1181, "step": 8447 }, { "epoch": 0.3561776672217889, "grad_norm": 0.741371750831604, "learning_rate": 4.896077329085855e-05, "loss": 0.5731, "step": 8448 }, { "epoch": 0.35621982840398847, "grad_norm": 0.6287175416946411, "learning_rate": 4.895999250445049e-05, "loss": 0.1356, "step": 8449 }, { "epoch": 0.356261989586188, "grad_norm": 0.8774129748344421, "learning_rate": 4.8959211718042413e-05, "loss": 0.6102, "step": 8450 }, { "epoch": 0.35630415076838756, "grad_norm": 98.88182830810547, "learning_rate": 4.8958430931634345e-05, "loss": 0.1251, "step": 8451 }, { "epoch": 0.3563463119505871, "grad_norm": 0.7636194229125977, "learning_rate": 4.8957650145226276e-05, "loss": 0.6441, "step": 8452 }, { "epoch": 0.35638847313278665, "grad_norm": 1.7149606943130493, "learning_rate": 4.89568693588182e-05, "loss": 0.8957, "step": 8453 }, { "epoch": 0.35643063431498617, "grad_norm": 0.26532113552093506, "learning_rate": 4.895608857241014e-05, "loss": 0.1598, "step": 8454 }, { "epoch": 0.35647279549718575, "grad_norm": 0.5809234976768494, "learning_rate": 4.895530778600206e-05, "loss": 0.1915, "step": 8455 }, { "epoch": 0.35651495667938526, "grad_norm": 1.062361478805542, "learning_rate": 4.8954526999593994e-05, "loss": 1.0708, "step": 8456 }, { "epoch": 0.35655711786158484, "grad_norm": 1.2252956628799438, "learning_rate": 4.8953746213185925e-05, "loss": 0.5788, "step": 8457 }, { "epoch": 0.3565992790437844, "grad_norm": 1.0050297975540161, "learning_rate": 4.895296542677785e-05, "loss": 1.0837, "step": 8458 }, { "epoch": 0.35664144022598393, "grad_norm": 0.4586888253688812, "learning_rate": 4.895218464036978e-05, "loss": 0.1371, "step": 8459 }, { "epoch": 0.3566836014081835, "grad_norm": 1.455586552619934, "learning_rate": 4.895140385396171e-05, "loss": 0.3475, "step": 8460 }, { "epoch": 0.356725762590383, "grad_norm": 1.176149845123291, "learning_rate": 4.8950623067553644e-05, "loss": 0.1184, "step": 8461 }, { "epoch": 0.3567679237725826, "grad_norm": 0.49096041917800903, "learning_rate": 4.8949842281145575e-05, "loss": 0.1498, "step": 8462 }, { "epoch": 0.3568100849547821, "grad_norm": 0.8193782567977905, "learning_rate": 4.89490614947375e-05, "loss": 0.3708, "step": 8463 }, { "epoch": 0.3568522461369817, "grad_norm": 1.521510362625122, "learning_rate": 4.894828070832943e-05, "loss": 1.0051, "step": 8464 }, { "epoch": 0.3568944073191812, "grad_norm": 1.1401400566101074, "learning_rate": 4.894749992192136e-05, "loss": 0.2278, "step": 8465 }, { "epoch": 0.3569365685013808, "grad_norm": 0.3995422124862671, "learning_rate": 4.894671913551329e-05, "loss": 0.0813, "step": 8466 }, { "epoch": 0.35697872968358035, "grad_norm": 1.1287795305252075, "learning_rate": 4.894593834910522e-05, "loss": 0.9658, "step": 8467 }, { "epoch": 0.35702089086577987, "grad_norm": 2.0914320945739746, "learning_rate": 4.8945157562697156e-05, "loss": 0.2867, "step": 8468 }, { "epoch": 0.35706305204797945, "grad_norm": 0.5742040872573853, "learning_rate": 4.894437677628908e-05, "loss": 0.6378, "step": 8469 }, { "epoch": 0.35710521323017896, "grad_norm": 1.2856545448303223, "learning_rate": 4.8943595989881005e-05, "loss": 0.3088, "step": 8470 }, { "epoch": 0.35714737441237854, "grad_norm": 520.3074951171875, "learning_rate": 4.894281520347294e-05, "loss": 0.1673, "step": 8471 }, { "epoch": 0.35718953559457806, "grad_norm": 1.7873424291610718, "learning_rate": 4.894203441706487e-05, "loss": 0.9538, "step": 8472 }, { "epoch": 0.35723169677677763, "grad_norm": 1.004209280014038, "learning_rate": 4.8941253630656805e-05, "loss": 0.4757, "step": 8473 }, { "epoch": 0.35727385795897715, "grad_norm": 0.7369362711906433, "learning_rate": 4.894047284424873e-05, "loss": 0.8049, "step": 8474 }, { "epoch": 0.3573160191411767, "grad_norm": 0.5316228866577148, "learning_rate": 4.893969205784066e-05, "loss": 0.1594, "step": 8475 }, { "epoch": 0.35735818032337624, "grad_norm": 0.9402765035629272, "learning_rate": 4.893891127143259e-05, "loss": 0.0935, "step": 8476 }, { "epoch": 0.3574003415055758, "grad_norm": 0.3678923547267914, "learning_rate": 4.893813048502452e-05, "loss": 0.1047, "step": 8477 }, { "epoch": 0.3574425026877754, "grad_norm": 0.9205108880996704, "learning_rate": 4.893734969861645e-05, "loss": 0.1411, "step": 8478 }, { "epoch": 0.3574846638699749, "grad_norm": 2.050278663635254, "learning_rate": 4.893656891220838e-05, "loss": 0.6499, "step": 8479 }, { "epoch": 0.3575268250521745, "grad_norm": 0.5537700653076172, "learning_rate": 4.893578812580031e-05, "loss": 0.2036, "step": 8480 }, { "epoch": 0.357568986234374, "grad_norm": 0.38357728719711304, "learning_rate": 4.8935007339392235e-05, "loss": 0.1704, "step": 8481 }, { "epoch": 0.35761114741657357, "grad_norm": 1.5829964876174927, "learning_rate": 4.8934226552984166e-05, "loss": 0.4314, "step": 8482 }, { "epoch": 0.3576533085987731, "grad_norm": 2.271421194076538, "learning_rate": 4.89334457665761e-05, "loss": 0.2901, "step": 8483 }, { "epoch": 0.35769546978097266, "grad_norm": 1.9813357591629028, "learning_rate": 4.893266498016802e-05, "loss": 0.3028, "step": 8484 }, { "epoch": 0.3577376309631722, "grad_norm": 1.2413115501403809, "learning_rate": 4.893188419375996e-05, "loss": 0.4773, "step": 8485 }, { "epoch": 0.35777979214537176, "grad_norm": 1.4214105606079102, "learning_rate": 4.8931103407351885e-05, "loss": 0.3362, "step": 8486 }, { "epoch": 0.35782195332757133, "grad_norm": 2.3493118286132812, "learning_rate": 4.893032262094382e-05, "loss": 0.5045, "step": 8487 }, { "epoch": 0.35786411450977085, "grad_norm": 0.7664677500724792, "learning_rate": 4.892954183453575e-05, "loss": 0.2822, "step": 8488 }, { "epoch": 0.3579062756919704, "grad_norm": 1.5157994031906128, "learning_rate": 4.892876104812767e-05, "loss": 0.2323, "step": 8489 }, { "epoch": 0.35794843687416994, "grad_norm": 1.20449960231781, "learning_rate": 4.892798026171961e-05, "loss": 0.6002, "step": 8490 }, { "epoch": 0.3579905980563695, "grad_norm": 0.8046810626983643, "learning_rate": 4.8927199475311534e-05, "loss": 0.5627, "step": 8491 }, { "epoch": 0.35803275923856903, "grad_norm": 0.7642922401428223, "learning_rate": 4.8926418688903465e-05, "loss": 0.1438, "step": 8492 }, { "epoch": 0.3580749204207686, "grad_norm": 0.6648154258728027, "learning_rate": 4.89256379024954e-05, "loss": 0.5497, "step": 8493 }, { "epoch": 0.3581170816029681, "grad_norm": 0.5597543716430664, "learning_rate": 4.892485711608733e-05, "loss": 0.1571, "step": 8494 }, { "epoch": 0.3581592427851677, "grad_norm": 1.0865709781646729, "learning_rate": 4.892407632967925e-05, "loss": 0.5992, "step": 8495 }, { "epoch": 0.35820140396736727, "grad_norm": 1.026309609413147, "learning_rate": 4.8923295543271184e-05, "loss": 0.3182, "step": 8496 }, { "epoch": 0.3582435651495668, "grad_norm": 0.5257204174995422, "learning_rate": 4.8922514756863115e-05, "loss": 0.2344, "step": 8497 }, { "epoch": 0.35828572633176636, "grad_norm": 0.39259040355682373, "learning_rate": 4.8921733970455046e-05, "loss": 0.1138, "step": 8498 }, { "epoch": 0.3583278875139659, "grad_norm": 0.8112622499465942, "learning_rate": 4.892095318404698e-05, "loss": 0.5266, "step": 8499 }, { "epoch": 0.35837004869616546, "grad_norm": 0.45037782192230225, "learning_rate": 4.89201723976389e-05, "loss": 0.1564, "step": 8500 }, { "epoch": 0.358412209878365, "grad_norm": 1.1434743404388428, "learning_rate": 4.891939161123083e-05, "loss": 0.4347, "step": 8501 }, { "epoch": 0.35845437106056455, "grad_norm": 0.7093347907066345, "learning_rate": 4.8918610824822764e-05, "loss": 0.2133, "step": 8502 }, { "epoch": 0.35849653224276407, "grad_norm": 257.6149597167969, "learning_rate": 4.891783003841469e-05, "loss": 0.2463, "step": 8503 }, { "epoch": 0.35853869342496364, "grad_norm": 2.124586582183838, "learning_rate": 4.891704925200663e-05, "loss": 0.9941, "step": 8504 }, { "epoch": 0.35858085460716316, "grad_norm": 0.6008121371269226, "learning_rate": 4.891626846559855e-05, "loss": 0.2118, "step": 8505 }, { "epoch": 0.35862301578936273, "grad_norm": 1.0503246784210205, "learning_rate": 4.891548767919048e-05, "loss": 1.177, "step": 8506 }, { "epoch": 0.3586651769715623, "grad_norm": 0.5413902401924133, "learning_rate": 4.8914706892782414e-05, "loss": 0.209, "step": 8507 }, { "epoch": 0.3587073381537618, "grad_norm": 0.768659770488739, "learning_rate": 4.891392610637434e-05, "loss": 0.2089, "step": 8508 }, { "epoch": 0.3587494993359614, "grad_norm": 1.0754703283309937, "learning_rate": 4.8913145319966276e-05, "loss": 0.1513, "step": 8509 }, { "epoch": 0.3587916605181609, "grad_norm": 1.9422625303268433, "learning_rate": 4.89123645335582e-05, "loss": 1.0258, "step": 8510 }, { "epoch": 0.3588338217003605, "grad_norm": 0.9945532083511353, "learning_rate": 4.891158374715013e-05, "loss": 0.3386, "step": 8511 }, { "epoch": 0.35887598288256, "grad_norm": 0.3742380142211914, "learning_rate": 4.8910802960742063e-05, "loss": 0.1006, "step": 8512 }, { "epoch": 0.3589181440647596, "grad_norm": 0.9563780426979065, "learning_rate": 4.8910022174333995e-05, "loss": 0.2249, "step": 8513 }, { "epoch": 0.3589603052469591, "grad_norm": 0.8689449429512024, "learning_rate": 4.890924138792592e-05, "loss": 0.4561, "step": 8514 }, { "epoch": 0.3590024664291587, "grad_norm": 0.5180908441543579, "learning_rate": 4.890846060151785e-05, "loss": 0.2365, "step": 8515 }, { "epoch": 0.35904462761135825, "grad_norm": 0.7807571887969971, "learning_rate": 4.890767981510978e-05, "loss": 0.3107, "step": 8516 }, { "epoch": 0.35908678879355777, "grad_norm": 0.43171635270118713, "learning_rate": 4.8906899028701706e-05, "loss": 0.2162, "step": 8517 }, { "epoch": 0.35912894997575734, "grad_norm": 1.68283212184906, "learning_rate": 4.8906118242293644e-05, "loss": 0.1804, "step": 8518 }, { "epoch": 0.35917111115795686, "grad_norm": 0.6337000727653503, "learning_rate": 4.890533745588557e-05, "loss": 0.3789, "step": 8519 }, { "epoch": 0.35921327234015643, "grad_norm": 0.7257782816886902, "learning_rate": 4.89045566694775e-05, "loss": 0.2233, "step": 8520 }, { "epoch": 0.35925543352235595, "grad_norm": 1.0682129859924316, "learning_rate": 4.890377588306943e-05, "loss": 0.3171, "step": 8521 }, { "epoch": 0.3592975947045555, "grad_norm": 2.3227808475494385, "learning_rate": 4.8902995096661356e-05, "loss": 0.3272, "step": 8522 }, { "epoch": 0.35933975588675504, "grad_norm": 0.5549852252006531, "learning_rate": 4.8902214310253294e-05, "loss": 0.189, "step": 8523 }, { "epoch": 0.3593819170689546, "grad_norm": 0.6280859112739563, "learning_rate": 4.890143352384522e-05, "loss": 0.0898, "step": 8524 }, { "epoch": 0.3594240782511542, "grad_norm": 1.201102614402771, "learning_rate": 4.890065273743715e-05, "loss": 0.2799, "step": 8525 }, { "epoch": 0.3594662394333537, "grad_norm": 2.266469717025757, "learning_rate": 4.889987195102908e-05, "loss": 0.8837, "step": 8526 }, { "epoch": 0.3595084006155533, "grad_norm": 1.3236645460128784, "learning_rate": 4.8899091164621005e-05, "loss": 0.6731, "step": 8527 }, { "epoch": 0.3595505617977528, "grad_norm": 0.6619409918785095, "learning_rate": 4.8898310378212937e-05, "loss": 0.3165, "step": 8528 }, { "epoch": 0.3595927229799524, "grad_norm": 0.6517624855041504, "learning_rate": 4.889752959180487e-05, "loss": 0.1753, "step": 8529 }, { "epoch": 0.3596348841621519, "grad_norm": 0.8166934251785278, "learning_rate": 4.88967488053968e-05, "loss": 0.125, "step": 8530 }, { "epoch": 0.35967704534435146, "grad_norm": 1.0969772338867188, "learning_rate": 4.889596801898873e-05, "loss": 0.2353, "step": 8531 }, { "epoch": 0.359719206526551, "grad_norm": 0.4432434141635895, "learning_rate": 4.8895187232580655e-05, "loss": 0.1451, "step": 8532 }, { "epoch": 0.35976136770875056, "grad_norm": 0.6984183192253113, "learning_rate": 4.8894406446172586e-05, "loss": 0.4984, "step": 8533 }, { "epoch": 0.3598035288909501, "grad_norm": 1.5161114931106567, "learning_rate": 4.889362565976452e-05, "loss": 0.2256, "step": 8534 }, { "epoch": 0.35984569007314965, "grad_norm": 0.7589613199234009, "learning_rate": 4.889284487335645e-05, "loss": 0.3144, "step": 8535 }, { "epoch": 0.3598878512553492, "grad_norm": 10.369504928588867, "learning_rate": 4.889206408694837e-05, "loss": 0.2027, "step": 8536 }, { "epoch": 0.35993001243754874, "grad_norm": 0.6952071189880371, "learning_rate": 4.889128330054031e-05, "loss": 0.2878, "step": 8537 }, { "epoch": 0.3599721736197483, "grad_norm": 0.42146655917167664, "learning_rate": 4.8890502514132236e-05, "loss": 0.1156, "step": 8538 }, { "epoch": 0.36001433480194783, "grad_norm": 0.6094887256622314, "learning_rate": 4.888972172772417e-05, "loss": 0.17, "step": 8539 }, { "epoch": 0.3600564959841474, "grad_norm": 0.7020875215530396, "learning_rate": 4.88889409413161e-05, "loss": 0.3917, "step": 8540 }, { "epoch": 0.3600986571663469, "grad_norm": 0.7111349701881409, "learning_rate": 4.888816015490802e-05, "loss": 0.2349, "step": 8541 }, { "epoch": 0.3601408183485465, "grad_norm": 0.48481321334838867, "learning_rate": 4.888737936849996e-05, "loss": 0.2093, "step": 8542 }, { "epoch": 0.360182979530746, "grad_norm": 1.1232043504714966, "learning_rate": 4.8886598582091885e-05, "loss": 1.1666, "step": 8543 }, { "epoch": 0.3602251407129456, "grad_norm": 0.7317917346954346, "learning_rate": 4.8885817795683816e-05, "loss": 0.2178, "step": 8544 }, { "epoch": 0.36026730189514516, "grad_norm": 0.8661327362060547, "learning_rate": 4.888503700927575e-05, "loss": 0.5035, "step": 8545 }, { "epoch": 0.3603094630773447, "grad_norm": 0.90767502784729, "learning_rate": 4.888425622286767e-05, "loss": 0.4237, "step": 8546 }, { "epoch": 0.36035162425954426, "grad_norm": 3.482112169265747, "learning_rate": 4.88834754364596e-05, "loss": 1.1359, "step": 8547 }, { "epoch": 0.3603937854417438, "grad_norm": 0.877215564250946, "learning_rate": 4.8882694650051535e-05, "loss": 0.6978, "step": 8548 }, { "epoch": 0.36043594662394335, "grad_norm": 0.6506308913230896, "learning_rate": 4.8881913863643466e-05, "loss": 0.617, "step": 8549 }, { "epoch": 0.36047810780614287, "grad_norm": 1.7835766077041626, "learning_rate": 4.888113307723539e-05, "loss": 0.5602, "step": 8550 }, { "epoch": 0.36052026898834244, "grad_norm": 0.7116007804870605, "learning_rate": 4.888035229082732e-05, "loss": 0.2237, "step": 8551 }, { "epoch": 0.36056243017054196, "grad_norm": 0.6817567348480225, "learning_rate": 4.887957150441925e-05, "loss": 0.3438, "step": 8552 }, { "epoch": 0.36060459135274153, "grad_norm": 0.3646974563598633, "learning_rate": 4.8878790718011184e-05, "loss": 0.147, "step": 8553 }, { "epoch": 0.3606467525349411, "grad_norm": 0.9693772196769714, "learning_rate": 4.8878009931603115e-05, "loss": 0.357, "step": 8554 }, { "epoch": 0.3606889137171406, "grad_norm": 0.49495694041252136, "learning_rate": 4.887722914519504e-05, "loss": 0.1371, "step": 8555 }, { "epoch": 0.3607310748993402, "grad_norm": 0.9937717914581299, "learning_rate": 4.887644835878698e-05, "loss": 0.1806, "step": 8556 }, { "epoch": 0.3607732360815397, "grad_norm": 1.522727370262146, "learning_rate": 4.88756675723789e-05, "loss": 0.2704, "step": 8557 }, { "epoch": 0.3608153972637393, "grad_norm": 0.7705485224723816, "learning_rate": 4.8874886785970834e-05, "loss": 0.6738, "step": 8558 }, { "epoch": 0.3608575584459388, "grad_norm": 0.31009283661842346, "learning_rate": 4.8874105999562765e-05, "loss": 0.1124, "step": 8559 }, { "epoch": 0.3608997196281384, "grad_norm": 0.6140764951705933, "learning_rate": 4.887332521315469e-05, "loss": 0.2224, "step": 8560 }, { "epoch": 0.3609418808103379, "grad_norm": 0.6200506687164307, "learning_rate": 4.887254442674662e-05, "loss": 0.3228, "step": 8561 }, { "epoch": 0.3609840419925375, "grad_norm": 0.6806830167770386, "learning_rate": 4.887176364033855e-05, "loss": 0.1906, "step": 8562 }, { "epoch": 0.361026203174737, "grad_norm": 0.8949986100196838, "learning_rate": 4.887098285393048e-05, "loss": 0.199, "step": 8563 }, { "epoch": 0.36106836435693657, "grad_norm": 0.6428363919258118, "learning_rate": 4.887020206752241e-05, "loss": 0.3111, "step": 8564 }, { "epoch": 0.36111052553913614, "grad_norm": 0.9546583890914917, "learning_rate": 4.886942128111434e-05, "loss": 0.2216, "step": 8565 }, { "epoch": 0.36115268672133566, "grad_norm": 0.4323709011077881, "learning_rate": 4.886864049470627e-05, "loss": 0.2152, "step": 8566 }, { "epoch": 0.36119484790353523, "grad_norm": 0.4964379370212555, "learning_rate": 4.88678597082982e-05, "loss": 0.4828, "step": 8567 }, { "epoch": 0.36123700908573475, "grad_norm": 0.8037756085395813, "learning_rate": 4.886707892189013e-05, "loss": 0.1862, "step": 8568 }, { "epoch": 0.3612791702679343, "grad_norm": 3.591562509536743, "learning_rate": 4.886629813548206e-05, "loss": 0.4868, "step": 8569 }, { "epoch": 0.36132133145013384, "grad_norm": 0.8842632174491882, "learning_rate": 4.886551734907399e-05, "loss": 0.1414, "step": 8570 }, { "epoch": 0.3613634926323334, "grad_norm": 0.5357716679573059, "learning_rate": 4.886473656266592e-05, "loss": 0.5696, "step": 8571 }, { "epoch": 0.36140565381453293, "grad_norm": 0.6394959688186646, "learning_rate": 4.8863955776257844e-05, "loss": 0.1811, "step": 8572 }, { "epoch": 0.3614478149967325, "grad_norm": 0.8259359002113342, "learning_rate": 4.886317498984978e-05, "loss": 0.7898, "step": 8573 }, { "epoch": 0.3614899761789321, "grad_norm": 0.880675733089447, "learning_rate": 4.886239420344171e-05, "loss": 0.4298, "step": 8574 }, { "epoch": 0.3615321373611316, "grad_norm": 1.9199954271316528, "learning_rate": 4.886161341703364e-05, "loss": 0.4597, "step": 8575 }, { "epoch": 0.3615742985433312, "grad_norm": 0.26115304231643677, "learning_rate": 4.886083263062557e-05, "loss": 0.0783, "step": 8576 }, { "epoch": 0.3616164597255307, "grad_norm": 1.629443883895874, "learning_rate": 4.8860051844217494e-05, "loss": 0.5655, "step": 8577 }, { "epoch": 0.36165862090773027, "grad_norm": 0.7612066864967346, "learning_rate": 4.885927105780943e-05, "loss": 0.1598, "step": 8578 }, { "epoch": 0.3617007820899298, "grad_norm": 0.56119304895401, "learning_rate": 4.8858490271401356e-05, "loss": 0.1184, "step": 8579 }, { "epoch": 0.36174294327212936, "grad_norm": 0.5274698734283447, "learning_rate": 4.885770948499329e-05, "loss": 0.1393, "step": 8580 }, { "epoch": 0.3617851044543289, "grad_norm": 0.9771499633789062, "learning_rate": 4.885692869858522e-05, "loss": 0.7585, "step": 8581 }, { "epoch": 0.36182726563652845, "grad_norm": 0.6587379574775696, "learning_rate": 4.885614791217715e-05, "loss": 0.5347, "step": 8582 }, { "epoch": 0.361869426818728, "grad_norm": 1.294608473777771, "learning_rate": 4.8855367125769074e-05, "loss": 0.8945, "step": 8583 }, { "epoch": 0.36191158800092754, "grad_norm": 0.3135990500450134, "learning_rate": 4.8854586339361006e-05, "loss": 0.1109, "step": 8584 }, { "epoch": 0.3619537491831271, "grad_norm": 1.5058434009552002, "learning_rate": 4.885380555295294e-05, "loss": 0.236, "step": 8585 }, { "epoch": 0.36199591036532663, "grad_norm": 0.6493469476699829, "learning_rate": 4.885302476654486e-05, "loss": 0.4179, "step": 8586 }, { "epoch": 0.3620380715475262, "grad_norm": 0.28206732869148254, "learning_rate": 4.88522439801368e-05, "loss": 0.1172, "step": 8587 }, { "epoch": 0.3620802327297257, "grad_norm": 1.0371910333633423, "learning_rate": 4.8851463193728724e-05, "loss": 0.0927, "step": 8588 }, { "epoch": 0.3621223939119253, "grad_norm": 0.7583909034729004, "learning_rate": 4.8850682407320655e-05, "loss": 0.4659, "step": 8589 }, { "epoch": 0.3621645550941248, "grad_norm": 8.403396606445312, "learning_rate": 4.8849901620912587e-05, "loss": 0.5297, "step": 8590 }, { "epoch": 0.3622067162763244, "grad_norm": 0.5150717496871948, "learning_rate": 4.884912083450451e-05, "loss": 0.1326, "step": 8591 }, { "epoch": 0.3622488774585239, "grad_norm": 1.6164594888687134, "learning_rate": 4.884834004809645e-05, "loss": 0.3711, "step": 8592 }, { "epoch": 0.3622910386407235, "grad_norm": 0.7399498224258423, "learning_rate": 4.8847559261688374e-05, "loss": 0.1492, "step": 8593 }, { "epoch": 0.36233319982292306, "grad_norm": 0.8798370957374573, "learning_rate": 4.8846778475280305e-05, "loss": 0.8354, "step": 8594 }, { "epoch": 0.3623753610051226, "grad_norm": 0.7634314298629761, "learning_rate": 4.8845997688872236e-05, "loss": 0.264, "step": 8595 }, { "epoch": 0.36241752218732215, "grad_norm": 0.44299355149269104, "learning_rate": 4.884521690246416e-05, "loss": 0.1356, "step": 8596 }, { "epoch": 0.36245968336952167, "grad_norm": 0.531291127204895, "learning_rate": 4.884443611605609e-05, "loss": 0.1329, "step": 8597 }, { "epoch": 0.36250184455172124, "grad_norm": 8.321599960327148, "learning_rate": 4.884365532964802e-05, "loss": 0.732, "step": 8598 }, { "epoch": 0.36254400573392076, "grad_norm": 24.16192626953125, "learning_rate": 4.8842874543239954e-05, "loss": 0.6144, "step": 8599 }, { "epoch": 0.36258616691612033, "grad_norm": 1.0705842971801758, "learning_rate": 4.8842093756831886e-05, "loss": 0.2583, "step": 8600 }, { "epoch": 0.36262832809831985, "grad_norm": 1.4679242372512817, "learning_rate": 4.884131297042382e-05, "loss": 0.5923, "step": 8601 }, { "epoch": 0.3626704892805194, "grad_norm": 0.6218234300613403, "learning_rate": 4.884053218401574e-05, "loss": 0.1251, "step": 8602 }, { "epoch": 0.362712650462719, "grad_norm": 0.8694786429405212, "learning_rate": 4.883975139760767e-05, "loss": 0.173, "step": 8603 }, { "epoch": 0.3627548116449185, "grad_norm": 2.0603365898132324, "learning_rate": 4.8838970611199604e-05, "loss": 0.3764, "step": 8604 }, { "epoch": 0.3627969728271181, "grad_norm": 0.6454135179519653, "learning_rate": 4.883818982479153e-05, "loss": 0.2602, "step": 8605 }, { "epoch": 0.3628391340093176, "grad_norm": 0.3865707814693451, "learning_rate": 4.8837409038383466e-05, "loss": 0.1299, "step": 8606 }, { "epoch": 0.3628812951915172, "grad_norm": 0.9192371368408203, "learning_rate": 4.883662825197539e-05, "loss": 0.5983, "step": 8607 }, { "epoch": 0.3629234563737167, "grad_norm": 1.3307613134384155, "learning_rate": 4.883584746556732e-05, "loss": 0.1867, "step": 8608 }, { "epoch": 0.3629656175559163, "grad_norm": 0.7346922159194946, "learning_rate": 4.883506667915925e-05, "loss": 0.2827, "step": 8609 }, { "epoch": 0.3630077787381158, "grad_norm": 0.9206390380859375, "learning_rate": 4.883428589275118e-05, "loss": 0.1214, "step": 8610 }, { "epoch": 0.36304993992031537, "grad_norm": 0.7730165719985962, "learning_rate": 4.8833505106343116e-05, "loss": 0.623, "step": 8611 }, { "epoch": 0.36309210110251494, "grad_norm": 0.5112051963806152, "learning_rate": 4.883272431993504e-05, "loss": 0.3151, "step": 8612 }, { "epoch": 0.36313426228471446, "grad_norm": 0.6413308382034302, "learning_rate": 4.883194353352697e-05, "loss": 0.2786, "step": 8613 }, { "epoch": 0.36317642346691403, "grad_norm": 0.8808267712593079, "learning_rate": 4.88311627471189e-05, "loss": 0.2971, "step": 8614 }, { "epoch": 0.36321858464911355, "grad_norm": 1.3349121809005737, "learning_rate": 4.883038196071083e-05, "loss": 0.8657, "step": 8615 }, { "epoch": 0.3632607458313131, "grad_norm": 1.3528378009796143, "learning_rate": 4.882960117430276e-05, "loss": 0.224, "step": 8616 }, { "epoch": 0.36330290701351264, "grad_norm": 3.182300329208374, "learning_rate": 4.882882038789469e-05, "loss": 1.1317, "step": 8617 }, { "epoch": 0.3633450681957122, "grad_norm": 1.0839815139770508, "learning_rate": 4.882803960148662e-05, "loss": 0.2065, "step": 8618 }, { "epoch": 0.36338722937791174, "grad_norm": 0.6822651028633118, "learning_rate": 4.8827258815078546e-05, "loss": 0.6303, "step": 8619 }, { "epoch": 0.3634293905601113, "grad_norm": 1.0069984197616577, "learning_rate": 4.8826478028670484e-05, "loss": 0.2519, "step": 8620 }, { "epoch": 0.3634715517423108, "grad_norm": 0.4583740830421448, "learning_rate": 4.882569724226241e-05, "loss": 0.1909, "step": 8621 }, { "epoch": 0.3635137129245104, "grad_norm": 2.3029043674468994, "learning_rate": 4.882491645585434e-05, "loss": 0.1284, "step": 8622 }, { "epoch": 0.36355587410671, "grad_norm": 0.33170285820961, "learning_rate": 4.882413566944627e-05, "loss": 0.0951, "step": 8623 }, { "epoch": 0.3635980352889095, "grad_norm": 0.3467293977737427, "learning_rate": 4.8823354883038195e-05, "loss": 0.1167, "step": 8624 }, { "epoch": 0.36364019647110907, "grad_norm": 2.279097080230713, "learning_rate": 4.882257409663013e-05, "loss": 0.6788, "step": 8625 }, { "epoch": 0.3636823576533086, "grad_norm": 1.5964851379394531, "learning_rate": 4.882179331022206e-05, "loss": 0.6314, "step": 8626 }, { "epoch": 0.36372451883550816, "grad_norm": 0.6779624223709106, "learning_rate": 4.882101252381399e-05, "loss": 0.1447, "step": 8627 }, { "epoch": 0.3637666800177077, "grad_norm": 1.3013607263565063, "learning_rate": 4.882023173740592e-05, "loss": 0.9292, "step": 8628 }, { "epoch": 0.36380884119990725, "grad_norm": 1.2069462537765503, "learning_rate": 4.8819450950997845e-05, "loss": 1.021, "step": 8629 }, { "epoch": 0.36385100238210677, "grad_norm": 0.623174786567688, "learning_rate": 4.8818670164589776e-05, "loss": 0.6227, "step": 8630 }, { "epoch": 0.36389316356430634, "grad_norm": 0.5764764547348022, "learning_rate": 4.881788937818171e-05, "loss": 0.1805, "step": 8631 }, { "epoch": 0.3639353247465059, "grad_norm": 0.8344529271125793, "learning_rate": 4.881710859177364e-05, "loss": 1.0765, "step": 8632 }, { "epoch": 0.36397748592870544, "grad_norm": 0.7552489638328552, "learning_rate": 4.881632780536557e-05, "loss": 0.221, "step": 8633 }, { "epoch": 0.364019647110905, "grad_norm": 0.7191752791404724, "learning_rate": 4.8815547018957494e-05, "loss": 0.6548, "step": 8634 }, { "epoch": 0.3640618082931045, "grad_norm": 1.4093639850616455, "learning_rate": 4.8814766232549425e-05, "loss": 0.1757, "step": 8635 }, { "epoch": 0.3641039694753041, "grad_norm": 0.33308109641075134, "learning_rate": 4.881398544614136e-05, "loss": 0.0932, "step": 8636 }, { "epoch": 0.3641461306575036, "grad_norm": 1.0908725261688232, "learning_rate": 4.881320465973329e-05, "loss": 0.1753, "step": 8637 }, { "epoch": 0.3641882918397032, "grad_norm": 0.6349830031394958, "learning_rate": 4.881242387332521e-05, "loss": 0.6788, "step": 8638 }, { "epoch": 0.3642304530219027, "grad_norm": 0.8607390522956848, "learning_rate": 4.8811643086917144e-05, "loss": 0.1054, "step": 8639 }, { "epoch": 0.3642726142041023, "grad_norm": 0.6949175000190735, "learning_rate": 4.8810862300509075e-05, "loss": 0.2543, "step": 8640 }, { "epoch": 0.36431477538630186, "grad_norm": 0.8316410779953003, "learning_rate": 4.8810081514101e-05, "loss": 0.577, "step": 8641 }, { "epoch": 0.3643569365685014, "grad_norm": 0.5626307725906372, "learning_rate": 4.880930072769294e-05, "loss": 0.3742, "step": 8642 }, { "epoch": 0.36439909775070095, "grad_norm": 1.5231469869613647, "learning_rate": 4.880851994128486e-05, "loss": 1.0171, "step": 8643 }, { "epoch": 0.36444125893290047, "grad_norm": 1.303889274597168, "learning_rate": 4.880773915487679e-05, "loss": 0.6918, "step": 8644 }, { "epoch": 0.36448342011510004, "grad_norm": 1.2574272155761719, "learning_rate": 4.8806958368468725e-05, "loss": 0.5269, "step": 8645 }, { "epoch": 0.36452558129729956, "grad_norm": 0.5827775001525879, "learning_rate": 4.8806177582060656e-05, "loss": 0.1634, "step": 8646 }, { "epoch": 0.36456774247949914, "grad_norm": 0.9561557173728943, "learning_rate": 4.880539679565259e-05, "loss": 0.3942, "step": 8647 }, { "epoch": 0.36460990366169865, "grad_norm": 0.6884466409683228, "learning_rate": 4.880461600924451e-05, "loss": 0.3385, "step": 8648 }, { "epoch": 0.3646520648438982, "grad_norm": 0.4955238103866577, "learning_rate": 4.880383522283644e-05, "loss": 0.1424, "step": 8649 }, { "epoch": 0.36469422602609775, "grad_norm": 0.5137329697608948, "learning_rate": 4.8803054436428374e-05, "loss": 0.3715, "step": 8650 }, { "epoch": 0.3647363872082973, "grad_norm": 0.8625579476356506, "learning_rate": 4.8802273650020305e-05, "loss": 0.2685, "step": 8651 }, { "epoch": 0.3647785483904969, "grad_norm": 0.5895360708236694, "learning_rate": 4.880149286361223e-05, "loss": 0.5211, "step": 8652 }, { "epoch": 0.3648207095726964, "grad_norm": 2.100494861602783, "learning_rate": 4.880071207720416e-05, "loss": 0.8239, "step": 8653 }, { "epoch": 0.364862870754896, "grad_norm": 0.8845917582511902, "learning_rate": 4.879993129079609e-05, "loss": 0.2765, "step": 8654 }, { "epoch": 0.3649050319370955, "grad_norm": 1.0641217231750488, "learning_rate": 4.879915050438802e-05, "loss": 0.9679, "step": 8655 }, { "epoch": 0.3649471931192951, "grad_norm": 1.336319088935852, "learning_rate": 4.8798369717979955e-05, "loss": 0.4907, "step": 8656 }, { "epoch": 0.3649893543014946, "grad_norm": 4.681211471557617, "learning_rate": 4.879758893157188e-05, "loss": 0.706, "step": 8657 }, { "epoch": 0.36503151548369417, "grad_norm": 0.93281489610672, "learning_rate": 4.879680814516381e-05, "loss": 0.1553, "step": 8658 }, { "epoch": 0.3650736766658937, "grad_norm": 0.5147984623908997, "learning_rate": 4.879602735875574e-05, "loss": 0.1964, "step": 8659 }, { "epoch": 0.36511583784809326, "grad_norm": 1.400288701057434, "learning_rate": 4.8795246572347666e-05, "loss": 0.7267, "step": 8660 }, { "epoch": 0.36515799903029283, "grad_norm": 1.353226661682129, "learning_rate": 4.8794465785939604e-05, "loss": 0.4156, "step": 8661 }, { "epoch": 0.36520016021249235, "grad_norm": 0.6488979458808899, "learning_rate": 4.879368499953153e-05, "loss": 0.1537, "step": 8662 }, { "epoch": 0.3652423213946919, "grad_norm": 0.7921510934829712, "learning_rate": 4.879290421312346e-05, "loss": 0.185, "step": 8663 }, { "epoch": 0.36528448257689144, "grad_norm": 0.4114150106906891, "learning_rate": 4.879212342671539e-05, "loss": 0.1425, "step": 8664 }, { "epoch": 0.365326643759091, "grad_norm": 0.5562102794647217, "learning_rate": 4.879134264030732e-05, "loss": 0.0957, "step": 8665 }, { "epoch": 0.36536880494129054, "grad_norm": 0.8817355632781982, "learning_rate": 4.879056185389925e-05, "loss": 0.439, "step": 8666 }, { "epoch": 0.3654109661234901, "grad_norm": 0.6652684807777405, "learning_rate": 4.878978106749118e-05, "loss": 0.2021, "step": 8667 }, { "epoch": 0.36545312730568963, "grad_norm": 0.554481565952301, "learning_rate": 4.878900028108311e-05, "loss": 0.1336, "step": 8668 }, { "epoch": 0.3654952884878892, "grad_norm": 0.6562888026237488, "learning_rate": 4.878821949467504e-05, "loss": 0.1964, "step": 8669 }, { "epoch": 0.3655374496700888, "grad_norm": 0.6893541812896729, "learning_rate": 4.878743870826697e-05, "loss": 0.1642, "step": 8670 }, { "epoch": 0.3655796108522883, "grad_norm": 0.4608761966228485, "learning_rate": 4.8786657921858897e-05, "loss": 0.1482, "step": 8671 }, { "epoch": 0.36562177203448787, "grad_norm": 0.7481998801231384, "learning_rate": 4.878587713545083e-05, "loss": 0.0725, "step": 8672 }, { "epoch": 0.3656639332166874, "grad_norm": 0.4456314444541931, "learning_rate": 4.878509634904276e-05, "loss": 0.1075, "step": 8673 }, { "epoch": 0.36570609439888696, "grad_norm": 0.6883566379547119, "learning_rate": 4.8784315562634684e-05, "loss": 0.5582, "step": 8674 }, { "epoch": 0.3657482555810865, "grad_norm": 8.79545783996582, "learning_rate": 4.878353477622662e-05, "loss": 0.2012, "step": 8675 }, { "epoch": 0.36579041676328605, "grad_norm": 0.6889982223510742, "learning_rate": 4.8782753989818546e-05, "loss": 0.4263, "step": 8676 }, { "epoch": 0.36583257794548557, "grad_norm": 0.4782330393791199, "learning_rate": 4.878197320341048e-05, "loss": 0.083, "step": 8677 }, { "epoch": 0.36587473912768514, "grad_norm": 1.4989339113235474, "learning_rate": 4.878119241700241e-05, "loss": 0.2246, "step": 8678 }, { "epoch": 0.36591690030988466, "grad_norm": 1.864188313484192, "learning_rate": 4.878041163059433e-05, "loss": 0.5644, "step": 8679 }, { "epoch": 0.36595906149208424, "grad_norm": 0.7807154059410095, "learning_rate": 4.877963084418627e-05, "loss": 0.2059, "step": 8680 }, { "epoch": 0.3660012226742838, "grad_norm": 1.0456231832504272, "learning_rate": 4.8778850057778196e-05, "loss": 0.7015, "step": 8681 }, { "epoch": 0.36604338385648333, "grad_norm": 1.569926142692566, "learning_rate": 4.877806927137013e-05, "loss": 0.3166, "step": 8682 }, { "epoch": 0.3660855450386829, "grad_norm": 0.6439491510391235, "learning_rate": 4.877728848496206e-05, "loss": 0.1972, "step": 8683 }, { "epoch": 0.3661277062208824, "grad_norm": 0.4989711046218872, "learning_rate": 4.877650769855398e-05, "loss": 0.1855, "step": 8684 }, { "epoch": 0.366169867403082, "grad_norm": 1.1495851278305054, "learning_rate": 4.8775726912145914e-05, "loss": 0.766, "step": 8685 }, { "epoch": 0.3662120285852815, "grad_norm": 0.4633428752422333, "learning_rate": 4.8774946125737845e-05, "loss": 0.167, "step": 8686 }, { "epoch": 0.3662541897674811, "grad_norm": 1.723618984222412, "learning_rate": 4.8774165339329776e-05, "loss": 0.6175, "step": 8687 }, { "epoch": 0.3662963509496806, "grad_norm": 0.3508679270744324, "learning_rate": 4.87733845529217e-05, "loss": 0.1738, "step": 8688 }, { "epoch": 0.3663385121318802, "grad_norm": 0.5521854162216187, "learning_rate": 4.877260376651364e-05, "loss": 0.1057, "step": 8689 }, { "epoch": 0.36638067331407975, "grad_norm": 2.6673028469085693, "learning_rate": 4.8771822980105563e-05, "loss": 0.8346, "step": 8690 }, { "epoch": 0.36642283449627927, "grad_norm": 0.6534085869789124, "learning_rate": 4.8771042193697495e-05, "loss": 0.7021, "step": 8691 }, { "epoch": 0.36646499567847884, "grad_norm": 0.47873127460479736, "learning_rate": 4.8770261407289426e-05, "loss": 0.2047, "step": 8692 }, { "epoch": 0.36650715686067836, "grad_norm": 1.667364239692688, "learning_rate": 4.876948062088135e-05, "loss": 0.4882, "step": 8693 }, { "epoch": 0.36654931804287794, "grad_norm": 1.153498649597168, "learning_rate": 4.876869983447329e-05, "loss": 0.3194, "step": 8694 }, { "epoch": 0.36659147922507745, "grad_norm": 0.5226532220840454, "learning_rate": 4.876791904806521e-05, "loss": 0.4297, "step": 8695 }, { "epoch": 0.36663364040727703, "grad_norm": 0.36838167905807495, "learning_rate": 4.8767138261657144e-05, "loss": 0.1055, "step": 8696 }, { "epoch": 0.36667580158947655, "grad_norm": 0.976454496383667, "learning_rate": 4.8766357475249075e-05, "loss": 0.2004, "step": 8697 }, { "epoch": 0.3667179627716761, "grad_norm": 1.2299773693084717, "learning_rate": 4.8765576688841e-05, "loss": 0.6125, "step": 8698 }, { "epoch": 0.3667601239538757, "grad_norm": 1.7905471324920654, "learning_rate": 4.876479590243293e-05, "loss": 0.4233, "step": 8699 }, { "epoch": 0.3668022851360752, "grad_norm": 0.9784020185470581, "learning_rate": 4.876401511602486e-05, "loss": 0.1888, "step": 8700 }, { "epoch": 0.3668444463182748, "grad_norm": 1.4045590162277222, "learning_rate": 4.8763234329616794e-05, "loss": 0.5701, "step": 8701 }, { "epoch": 0.3668866075004743, "grad_norm": 0.6195655465126038, "learning_rate": 4.8762453543208725e-05, "loss": 0.5725, "step": 8702 }, { "epoch": 0.3669287686826739, "grad_norm": 0.7196131348609924, "learning_rate": 4.876167275680065e-05, "loss": 0.1518, "step": 8703 }, { "epoch": 0.3669709298648734, "grad_norm": 0.617165207862854, "learning_rate": 4.876089197039258e-05, "loss": 0.1746, "step": 8704 }, { "epoch": 0.36701309104707297, "grad_norm": 0.7002089023590088, "learning_rate": 4.876011118398451e-05, "loss": 0.2538, "step": 8705 }, { "epoch": 0.3670552522292725, "grad_norm": 0.47908657789230347, "learning_rate": 4.875933039757644e-05, "loss": 0.1468, "step": 8706 }, { "epoch": 0.36709741341147206, "grad_norm": 0.9212803840637207, "learning_rate": 4.875854961116837e-05, "loss": 0.5945, "step": 8707 }, { "epoch": 0.3671395745936716, "grad_norm": 3.3330368995666504, "learning_rate": 4.8757768824760306e-05, "loss": 0.4651, "step": 8708 }, { "epoch": 0.36718173577587115, "grad_norm": 1.5899220705032349, "learning_rate": 4.875698803835223e-05, "loss": 0.2816, "step": 8709 }, { "epoch": 0.36722389695807073, "grad_norm": 0.3985673189163208, "learning_rate": 4.875620725194416e-05, "loss": 0.108, "step": 8710 }, { "epoch": 0.36726605814027025, "grad_norm": 0.44887229800224304, "learning_rate": 4.875542646553609e-05, "loss": 0.1866, "step": 8711 }, { "epoch": 0.3673082193224698, "grad_norm": 0.33637598156929016, "learning_rate": 4.875464567912802e-05, "loss": 0.12, "step": 8712 }, { "epoch": 0.36735038050466934, "grad_norm": 1.4004747867584229, "learning_rate": 4.8753864892719955e-05, "loss": 0.9151, "step": 8713 }, { "epoch": 0.3673925416868689, "grad_norm": 0.3461603820323944, "learning_rate": 4.875308410631188e-05, "loss": 0.0957, "step": 8714 }, { "epoch": 0.36743470286906843, "grad_norm": 0.8916537165641785, "learning_rate": 4.875230331990381e-05, "loss": 0.3993, "step": 8715 }, { "epoch": 0.367476864051268, "grad_norm": 0.24235039949417114, "learning_rate": 4.875152253349574e-05, "loss": 0.0582, "step": 8716 }, { "epoch": 0.3675190252334675, "grad_norm": 1.9336833953857422, "learning_rate": 4.875074174708767e-05, "loss": 0.2583, "step": 8717 }, { "epoch": 0.3675611864156671, "grad_norm": 0.6539999842643738, "learning_rate": 4.87499609606796e-05, "loss": 0.131, "step": 8718 }, { "epoch": 0.36760334759786667, "grad_norm": 1.1287161111831665, "learning_rate": 4.874918017427153e-05, "loss": 0.3325, "step": 8719 }, { "epoch": 0.3676455087800662, "grad_norm": 0.3464054763317108, "learning_rate": 4.874839938786346e-05, "loss": 0.1442, "step": 8720 }, { "epoch": 0.36768766996226576, "grad_norm": 1.9688652753829956, "learning_rate": 4.8747618601455385e-05, "loss": 0.4222, "step": 8721 }, { "epoch": 0.3677298311444653, "grad_norm": 0.5111423134803772, "learning_rate": 4.8746837815047316e-05, "loss": 0.23, "step": 8722 }, { "epoch": 0.36777199232666485, "grad_norm": 1.2224830389022827, "learning_rate": 4.874605702863925e-05, "loss": 0.6247, "step": 8723 }, { "epoch": 0.36781415350886437, "grad_norm": 0.9536195993423462, "learning_rate": 4.874527624223117e-05, "loss": 0.2534, "step": 8724 }, { "epoch": 0.36785631469106395, "grad_norm": 0.7524387836456299, "learning_rate": 4.874449545582311e-05, "loss": 0.2205, "step": 8725 }, { "epoch": 0.36789847587326346, "grad_norm": 0.9760278463363647, "learning_rate": 4.8743714669415035e-05, "loss": 0.1316, "step": 8726 }, { "epoch": 0.36794063705546304, "grad_norm": 0.5949402451515198, "learning_rate": 4.874293388300697e-05, "loss": 0.1302, "step": 8727 }, { "epoch": 0.3679827982376626, "grad_norm": 0.8422734141349792, "learning_rate": 4.87421530965989e-05, "loss": 0.2477, "step": 8728 }, { "epoch": 0.36802495941986213, "grad_norm": 0.9871845841407776, "learning_rate": 4.874137231019082e-05, "loss": 0.6828, "step": 8729 }, { "epoch": 0.3680671206020617, "grad_norm": 0.26743799448013306, "learning_rate": 4.874059152378276e-05, "loss": 0.0978, "step": 8730 }, { "epoch": 0.3681092817842612, "grad_norm": 0.4593167304992676, "learning_rate": 4.8739810737374684e-05, "loss": 0.1162, "step": 8731 }, { "epoch": 0.3681514429664608, "grad_norm": 1.1286016702651978, "learning_rate": 4.8739029950966615e-05, "loss": 1.0536, "step": 8732 }, { "epoch": 0.3681936041486603, "grad_norm": 0.3371836841106415, "learning_rate": 4.873824916455855e-05, "loss": 0.1251, "step": 8733 }, { "epoch": 0.3682357653308599, "grad_norm": 0.3984753489494324, "learning_rate": 4.873746837815048e-05, "loss": 0.0773, "step": 8734 }, { "epoch": 0.3682779265130594, "grad_norm": 0.3799370229244232, "learning_rate": 4.87366875917424e-05, "loss": 0.1774, "step": 8735 }, { "epoch": 0.368320087695259, "grad_norm": 0.5186516046524048, "learning_rate": 4.8735906805334334e-05, "loss": 0.0929, "step": 8736 }, { "epoch": 0.3683622488774585, "grad_norm": 0.935687780380249, "learning_rate": 4.8735126018926265e-05, "loss": 0.5367, "step": 8737 }, { "epoch": 0.36840441005965807, "grad_norm": 0.3964940011501312, "learning_rate": 4.8734345232518196e-05, "loss": 0.1438, "step": 8738 }, { "epoch": 0.36844657124185765, "grad_norm": 0.9719165563583374, "learning_rate": 4.873356444611013e-05, "loss": 0.4687, "step": 8739 }, { "epoch": 0.36848873242405716, "grad_norm": 0.6220455169677734, "learning_rate": 4.873278365970205e-05, "loss": 0.1724, "step": 8740 }, { "epoch": 0.36853089360625674, "grad_norm": 0.3963317573070526, "learning_rate": 4.873200287329398e-05, "loss": 0.1514, "step": 8741 }, { "epoch": 0.36857305478845626, "grad_norm": 1.1900337934494019, "learning_rate": 4.8731222086885914e-05, "loss": 0.4822, "step": 8742 }, { "epoch": 0.36861521597065583, "grad_norm": 0.529988706111908, "learning_rate": 4.873044130047784e-05, "loss": 0.2227, "step": 8743 }, { "epoch": 0.36865737715285535, "grad_norm": 0.2883811891078949, "learning_rate": 4.872966051406978e-05, "loss": 0.1167, "step": 8744 }, { "epoch": 0.3686995383350549, "grad_norm": 0.43196529150009155, "learning_rate": 4.87288797276617e-05, "loss": 0.1332, "step": 8745 }, { "epoch": 0.36874169951725444, "grad_norm": 0.4947945475578308, "learning_rate": 4.872809894125363e-05, "loss": 0.167, "step": 8746 }, { "epoch": 0.368783860699454, "grad_norm": 0.5467376708984375, "learning_rate": 4.8727318154845564e-05, "loss": 0.168, "step": 8747 }, { "epoch": 0.3688260218816536, "grad_norm": 0.42213523387908936, "learning_rate": 4.872653736843749e-05, "loss": 0.28, "step": 8748 }, { "epoch": 0.3688681830638531, "grad_norm": 0.727408766746521, "learning_rate": 4.8725756582029426e-05, "loss": 0.2522, "step": 8749 }, { "epoch": 0.3689103442460527, "grad_norm": 175.02528381347656, "learning_rate": 4.872497579562135e-05, "loss": 0.866, "step": 8750 }, { "epoch": 0.3689525054282522, "grad_norm": 0.7730734944343567, "learning_rate": 4.872419500921328e-05, "loss": 0.6142, "step": 8751 }, { "epoch": 0.36899466661045177, "grad_norm": 0.36320483684539795, "learning_rate": 4.8723414222805213e-05, "loss": 0.1935, "step": 8752 }, { "epoch": 0.3690368277926513, "grad_norm": 1.4480328559875488, "learning_rate": 4.8722633436397145e-05, "loss": 0.1746, "step": 8753 }, { "epoch": 0.36907898897485086, "grad_norm": 0.5696863532066345, "learning_rate": 4.872185264998907e-05, "loss": 0.1512, "step": 8754 }, { "epoch": 0.3691211501570504, "grad_norm": 0.6480450630187988, "learning_rate": 4.8721071863581e-05, "loss": 0.311, "step": 8755 }, { "epoch": 0.36916331133924996, "grad_norm": 0.4840180277824402, "learning_rate": 4.872029107717293e-05, "loss": 0.1343, "step": 8756 }, { "epoch": 0.3692054725214495, "grad_norm": 0.42309120297431946, "learning_rate": 4.8719510290764856e-05, "loss": 0.1906, "step": 8757 }, { "epoch": 0.36924763370364905, "grad_norm": 1.4278204441070557, "learning_rate": 4.8718729504356794e-05, "loss": 0.6591, "step": 8758 }, { "epoch": 0.3692897948858486, "grad_norm": 1.1749656200408936, "learning_rate": 4.871794871794872e-05, "loss": 1.1644, "step": 8759 }, { "epoch": 0.36933195606804814, "grad_norm": 0.7186703085899353, "learning_rate": 4.871716793154065e-05, "loss": 0.2031, "step": 8760 }, { "epoch": 0.3693741172502477, "grad_norm": 1.6047953367233276, "learning_rate": 4.871638714513258e-05, "loss": 0.2508, "step": 8761 }, { "epoch": 0.36941627843244723, "grad_norm": 0.6287450194358826, "learning_rate": 4.8715606358724506e-05, "loss": 0.2694, "step": 8762 }, { "epoch": 0.3694584396146468, "grad_norm": 0.6733883023262024, "learning_rate": 4.8714825572316444e-05, "loss": 0.3004, "step": 8763 }, { "epoch": 0.3695006007968463, "grad_norm": 0.7354693412780762, "learning_rate": 4.871404478590837e-05, "loss": 0.2489, "step": 8764 }, { "epoch": 0.3695427619790459, "grad_norm": 0.576468825340271, "learning_rate": 4.87132639995003e-05, "loss": 0.518, "step": 8765 }, { "epoch": 0.3695849231612454, "grad_norm": 0.5799363851547241, "learning_rate": 4.871248321309223e-05, "loss": 0.5683, "step": 8766 }, { "epoch": 0.369627084343445, "grad_norm": 0.395187109708786, "learning_rate": 4.8711702426684155e-05, "loss": 0.1266, "step": 8767 }, { "epoch": 0.36966924552564456, "grad_norm": 0.8396881818771362, "learning_rate": 4.8710921640276087e-05, "loss": 0.1898, "step": 8768 }, { "epoch": 0.3697114067078441, "grad_norm": 0.4624282121658325, "learning_rate": 4.871014085386802e-05, "loss": 0.1098, "step": 8769 }, { "epoch": 0.36975356789004366, "grad_norm": 0.5887886881828308, "learning_rate": 4.870936006745995e-05, "loss": 0.0627, "step": 8770 }, { "epoch": 0.3697957290722432, "grad_norm": 0.4104263186454773, "learning_rate": 4.870857928105188e-05, "loss": 0.0922, "step": 8771 }, { "epoch": 0.36983789025444275, "grad_norm": 0.27649131417274475, "learning_rate": 4.870779849464381e-05, "loss": 0.0853, "step": 8772 }, { "epoch": 0.36988005143664227, "grad_norm": 1.172182321548462, "learning_rate": 4.8707017708235736e-05, "loss": 0.9745, "step": 8773 }, { "epoch": 0.36992221261884184, "grad_norm": 0.6979789733886719, "learning_rate": 4.870623692182767e-05, "loss": 0.7208, "step": 8774 }, { "epoch": 0.36996437380104136, "grad_norm": 2.3178515434265137, "learning_rate": 4.87054561354196e-05, "loss": 0.218, "step": 8775 }, { "epoch": 0.37000653498324093, "grad_norm": 0.6308692693710327, "learning_rate": 4.870467534901152e-05, "loss": 0.3983, "step": 8776 }, { "epoch": 0.3700486961654405, "grad_norm": 0.8521643877029419, "learning_rate": 4.870389456260346e-05, "loss": 0.2791, "step": 8777 }, { "epoch": 0.37009085734764, "grad_norm": 0.5856181383132935, "learning_rate": 4.8703113776195386e-05, "loss": 0.2427, "step": 8778 }, { "epoch": 0.3701330185298396, "grad_norm": 1.0642201900482178, "learning_rate": 4.870233298978732e-05, "loss": 0.2326, "step": 8779 }, { "epoch": 0.3701751797120391, "grad_norm": 1.195270299911499, "learning_rate": 4.870155220337925e-05, "loss": 0.1588, "step": 8780 }, { "epoch": 0.3702173408942387, "grad_norm": 0.5407750606536865, "learning_rate": 4.870077141697117e-05, "loss": 0.1098, "step": 8781 }, { "epoch": 0.3702595020764382, "grad_norm": 0.29442712664604187, "learning_rate": 4.869999063056311e-05, "loss": 0.1268, "step": 8782 }, { "epoch": 0.3703016632586378, "grad_norm": 2.0567626953125, "learning_rate": 4.8699209844155035e-05, "loss": 0.6127, "step": 8783 }, { "epoch": 0.3703438244408373, "grad_norm": 1.5026640892028809, "learning_rate": 4.8698429057746966e-05, "loss": 0.2655, "step": 8784 }, { "epoch": 0.3703859856230369, "grad_norm": 0.43916863203048706, "learning_rate": 4.86976482713389e-05, "loss": 0.1262, "step": 8785 }, { "epoch": 0.3704281468052364, "grad_norm": 0.5079506039619446, "learning_rate": 4.869686748493082e-05, "loss": 0.1817, "step": 8786 }, { "epoch": 0.37047030798743597, "grad_norm": 0.3895459771156311, "learning_rate": 4.869608669852275e-05, "loss": 0.1007, "step": 8787 }, { "epoch": 0.37051246916963554, "grad_norm": 2.631183385848999, "learning_rate": 4.8695305912114685e-05, "loss": 0.8466, "step": 8788 }, { "epoch": 0.37055463035183506, "grad_norm": 0.7870593070983887, "learning_rate": 4.8694525125706616e-05, "loss": 0.097, "step": 8789 }, { "epoch": 0.37059679153403463, "grad_norm": 1.5185362100601196, "learning_rate": 4.869374433929854e-05, "loss": 0.6805, "step": 8790 }, { "epoch": 0.37063895271623415, "grad_norm": 0.452287495136261, "learning_rate": 4.869296355289047e-05, "loss": 0.1371, "step": 8791 }, { "epoch": 0.3706811138984337, "grad_norm": 0.38088566064834595, "learning_rate": 4.86921827664824e-05, "loss": 0.1481, "step": 8792 }, { "epoch": 0.37072327508063324, "grad_norm": 0.7545584440231323, "learning_rate": 4.8691401980074334e-05, "loss": 0.4336, "step": 8793 }, { "epoch": 0.3707654362628328, "grad_norm": 27.57431411743164, "learning_rate": 4.8690621193666265e-05, "loss": 1.0516, "step": 8794 }, { "epoch": 0.37080759744503233, "grad_norm": 2.739241361618042, "learning_rate": 4.868984040725819e-05, "loss": 1.0766, "step": 8795 }, { "epoch": 0.3708497586272319, "grad_norm": 2.8710522651672363, "learning_rate": 4.868905962085013e-05, "loss": 1.0088, "step": 8796 }, { "epoch": 0.3708919198094315, "grad_norm": 1.3312890529632568, "learning_rate": 4.868827883444205e-05, "loss": 0.5256, "step": 8797 }, { "epoch": 0.370934080991631, "grad_norm": 0.34560826420783997, "learning_rate": 4.8687498048033984e-05, "loss": 0.1046, "step": 8798 }, { "epoch": 0.3709762421738306, "grad_norm": 0.746191143989563, "learning_rate": 4.8686717261625915e-05, "loss": 0.2293, "step": 8799 }, { "epoch": 0.3710184033560301, "grad_norm": 0.47933468222618103, "learning_rate": 4.868593647521784e-05, "loss": 0.1666, "step": 8800 }, { "epoch": 0.37106056453822966, "grad_norm": 0.7641324996948242, "learning_rate": 4.868515568880977e-05, "loss": 0.2922, "step": 8801 }, { "epoch": 0.3711027257204292, "grad_norm": 1.7198829650878906, "learning_rate": 4.86843749024017e-05, "loss": 1.1457, "step": 8802 }, { "epoch": 0.37114488690262876, "grad_norm": 0.7909870743751526, "learning_rate": 4.868359411599363e-05, "loss": 0.6186, "step": 8803 }, { "epoch": 0.3711870480848283, "grad_norm": 0.5490169525146484, "learning_rate": 4.868281332958556e-05, "loss": 0.1625, "step": 8804 }, { "epoch": 0.37122920926702785, "grad_norm": 0.6332014203071594, "learning_rate": 4.868203254317749e-05, "loss": 0.0758, "step": 8805 }, { "epoch": 0.3712713704492274, "grad_norm": 1.284060001373291, "learning_rate": 4.868125175676942e-05, "loss": 0.532, "step": 8806 }, { "epoch": 0.37131353163142694, "grad_norm": 0.6286680102348328, "learning_rate": 4.868047097036135e-05, "loss": 0.1255, "step": 8807 }, { "epoch": 0.3713556928136265, "grad_norm": 0.5426868796348572, "learning_rate": 4.867969018395328e-05, "loss": 0.2403, "step": 8808 }, { "epoch": 0.37139785399582603, "grad_norm": 1.5503758192062378, "learning_rate": 4.867890939754521e-05, "loss": 0.3607, "step": 8809 }, { "epoch": 0.3714400151780256, "grad_norm": 0.633810818195343, "learning_rate": 4.867812861113714e-05, "loss": 0.1858, "step": 8810 }, { "epoch": 0.3714821763602251, "grad_norm": 2.8526151180267334, "learning_rate": 4.867734782472907e-05, "loss": 0.8405, "step": 8811 }, { "epoch": 0.3715243375424247, "grad_norm": 1.7390854358673096, "learning_rate": 4.8676567038320994e-05, "loss": 1.1061, "step": 8812 }, { "epoch": 0.3715664987246242, "grad_norm": 0.6189761161804199, "learning_rate": 4.867578625191293e-05, "loss": 0.6888, "step": 8813 }, { "epoch": 0.3716086599068238, "grad_norm": 0.6939609050750732, "learning_rate": 4.867500546550486e-05, "loss": 0.2845, "step": 8814 }, { "epoch": 0.3716508210890233, "grad_norm": 0.34609320759773254, "learning_rate": 4.867422467909679e-05, "loss": 0.1762, "step": 8815 }, { "epoch": 0.3716929822712229, "grad_norm": 0.40069887042045593, "learning_rate": 4.867344389268872e-05, "loss": 0.1703, "step": 8816 }, { "epoch": 0.37173514345342246, "grad_norm": 0.30901455879211426, "learning_rate": 4.8672663106280644e-05, "loss": 0.1401, "step": 8817 }, { "epoch": 0.371777304635622, "grad_norm": 3.1678643226623535, "learning_rate": 4.867188231987258e-05, "loss": 0.2093, "step": 8818 }, { "epoch": 0.37181946581782155, "grad_norm": 0.4588203430175781, "learning_rate": 4.8671101533464506e-05, "loss": 0.232, "step": 8819 }, { "epoch": 0.37186162700002107, "grad_norm": 0.506900429725647, "learning_rate": 4.867032074705644e-05, "loss": 0.1957, "step": 8820 }, { "epoch": 0.37190378818222064, "grad_norm": 0.6247820854187012, "learning_rate": 4.866953996064837e-05, "loss": 0.1681, "step": 8821 }, { "epoch": 0.37194594936442016, "grad_norm": 1.1287951469421387, "learning_rate": 4.86687591742403e-05, "loss": 1.0252, "step": 8822 }, { "epoch": 0.37198811054661973, "grad_norm": 1.0893369913101196, "learning_rate": 4.8667978387832224e-05, "loss": 0.1634, "step": 8823 }, { "epoch": 0.37203027172881925, "grad_norm": 0.2655270993709564, "learning_rate": 4.8667197601424156e-05, "loss": 0.1151, "step": 8824 }, { "epoch": 0.3720724329110188, "grad_norm": 0.7966128587722778, "learning_rate": 4.866641681501609e-05, "loss": 0.2369, "step": 8825 }, { "epoch": 0.3721145940932184, "grad_norm": 0.801884651184082, "learning_rate": 4.866563602860801e-05, "loss": 0.4398, "step": 8826 }, { "epoch": 0.3721567552754179, "grad_norm": 0.516645073890686, "learning_rate": 4.866485524219995e-05, "loss": 0.1034, "step": 8827 }, { "epoch": 0.3721989164576175, "grad_norm": 0.5262885093688965, "learning_rate": 4.8664074455791874e-05, "loss": 0.3161, "step": 8828 }, { "epoch": 0.372241077639817, "grad_norm": 1.3700326681137085, "learning_rate": 4.8663293669383805e-05, "loss": 0.5047, "step": 8829 }, { "epoch": 0.3722832388220166, "grad_norm": 0.4032430052757263, "learning_rate": 4.8662512882975737e-05, "loss": 0.1283, "step": 8830 }, { "epoch": 0.3723254000042161, "grad_norm": 0.29357555508613586, "learning_rate": 4.866173209656766e-05, "loss": 0.084, "step": 8831 }, { "epoch": 0.3723675611864157, "grad_norm": 0.27053868770599365, "learning_rate": 4.86609513101596e-05, "loss": 0.0633, "step": 8832 }, { "epoch": 0.3724097223686152, "grad_norm": 0.8019883036613464, "learning_rate": 4.8660170523751524e-05, "loss": 0.6317, "step": 8833 }, { "epoch": 0.37245188355081477, "grad_norm": 1.5154603719711304, "learning_rate": 4.8659389737343455e-05, "loss": 0.2236, "step": 8834 }, { "epoch": 0.37249404473301434, "grad_norm": 0.5812084078788757, "learning_rate": 4.8658608950935386e-05, "loss": 0.2154, "step": 8835 }, { "epoch": 0.37253620591521386, "grad_norm": 0.6646406054496765, "learning_rate": 4.865782816452731e-05, "loss": 0.5921, "step": 8836 }, { "epoch": 0.37257836709741343, "grad_norm": 1.376340627670288, "learning_rate": 4.865704737811924e-05, "loss": 1.0463, "step": 8837 }, { "epoch": 0.37262052827961295, "grad_norm": 0.7292866110801697, "learning_rate": 4.865626659171117e-05, "loss": 0.1849, "step": 8838 }, { "epoch": 0.3726626894618125, "grad_norm": 1.1850571632385254, "learning_rate": 4.8655485805303104e-05, "loss": 0.8433, "step": 8839 }, { "epoch": 0.37270485064401204, "grad_norm": 0.8401421904563904, "learning_rate": 4.8654705018895036e-05, "loss": 0.156, "step": 8840 }, { "epoch": 0.3727470118262116, "grad_norm": 0.9577922821044922, "learning_rate": 4.865392423248697e-05, "loss": 0.3393, "step": 8841 }, { "epoch": 0.37278917300841113, "grad_norm": 0.3555125892162323, "learning_rate": 4.865314344607889e-05, "loss": 0.1935, "step": 8842 }, { "epoch": 0.3728313341906107, "grad_norm": 0.6603702902793884, "learning_rate": 4.865236265967082e-05, "loss": 0.4338, "step": 8843 }, { "epoch": 0.3728734953728102, "grad_norm": 0.485948383808136, "learning_rate": 4.8651581873262754e-05, "loss": 0.2479, "step": 8844 }, { "epoch": 0.3729156565550098, "grad_norm": 0.8962174654006958, "learning_rate": 4.865080108685468e-05, "loss": 0.5457, "step": 8845 }, { "epoch": 0.3729578177372094, "grad_norm": 0.8255180716514587, "learning_rate": 4.8650020300446616e-05, "loss": 0.586, "step": 8846 }, { "epoch": 0.3729999789194089, "grad_norm": 1.572387933731079, "learning_rate": 4.864923951403854e-05, "loss": 1.0259, "step": 8847 }, { "epoch": 0.37304214010160847, "grad_norm": 1.1090717315673828, "learning_rate": 4.864845872763047e-05, "loss": 0.7381, "step": 8848 }, { "epoch": 0.373084301283808, "grad_norm": 0.7042994499206543, "learning_rate": 4.86476779412224e-05, "loss": 0.4739, "step": 8849 }, { "epoch": 0.37312646246600756, "grad_norm": 0.4780246913433075, "learning_rate": 4.864689715481433e-05, "loss": 0.147, "step": 8850 }, { "epoch": 0.3731686236482071, "grad_norm": 4.264828681945801, "learning_rate": 4.8646116368406266e-05, "loss": 0.7075, "step": 8851 }, { "epoch": 0.37321078483040665, "grad_norm": 2.0255496501922607, "learning_rate": 4.864533558199819e-05, "loss": 1.0038, "step": 8852 }, { "epoch": 0.37325294601260617, "grad_norm": 0.7905975580215454, "learning_rate": 4.864455479559012e-05, "loss": 0.1267, "step": 8853 }, { "epoch": 0.37329510719480574, "grad_norm": 0.661480724811554, "learning_rate": 4.864377400918205e-05, "loss": 0.2051, "step": 8854 }, { "epoch": 0.3733372683770053, "grad_norm": 0.6949150562286377, "learning_rate": 4.864299322277398e-05, "loss": 0.5877, "step": 8855 }, { "epoch": 0.37337942955920483, "grad_norm": 0.7160041928291321, "learning_rate": 4.864221243636591e-05, "loss": 0.1854, "step": 8856 }, { "epoch": 0.3734215907414044, "grad_norm": 1.4620630741119385, "learning_rate": 4.864143164995784e-05, "loss": 0.1447, "step": 8857 }, { "epoch": 0.3734637519236039, "grad_norm": 0.3702944815158844, "learning_rate": 4.864065086354977e-05, "loss": 0.115, "step": 8858 }, { "epoch": 0.3735059131058035, "grad_norm": 1.0183435678482056, "learning_rate": 4.8639870077141696e-05, "loss": 0.1677, "step": 8859 }, { "epoch": 0.373548074288003, "grad_norm": 1.3524070978164673, "learning_rate": 4.8639089290733634e-05, "loss": 0.3312, "step": 8860 }, { "epoch": 0.3735902354702026, "grad_norm": 0.3436811566352844, "learning_rate": 4.863830850432556e-05, "loss": 0.106, "step": 8861 }, { "epoch": 0.3736323966524021, "grad_norm": 2.122218132019043, "learning_rate": 4.863752771791749e-05, "loss": 0.5041, "step": 8862 }, { "epoch": 0.3736745578346017, "grad_norm": 0.610621452331543, "learning_rate": 4.863674693150942e-05, "loss": 0.4896, "step": 8863 }, { "epoch": 0.37371671901680126, "grad_norm": 0.7339024543762207, "learning_rate": 4.8635966145101345e-05, "loss": 0.1913, "step": 8864 }, { "epoch": 0.3737588801990008, "grad_norm": 0.4392842948436737, "learning_rate": 4.863518535869328e-05, "loss": 0.3805, "step": 8865 }, { "epoch": 0.37380104138120035, "grad_norm": 0.7763556241989136, "learning_rate": 4.863440457228521e-05, "loss": 0.0839, "step": 8866 }, { "epoch": 0.37384320256339987, "grad_norm": 1.4088258743286133, "learning_rate": 4.863362378587714e-05, "loss": 0.2741, "step": 8867 }, { "epoch": 0.37388536374559944, "grad_norm": 1.232587218284607, "learning_rate": 4.863284299946907e-05, "loss": 0.0825, "step": 8868 }, { "epoch": 0.37392752492779896, "grad_norm": 1.0900137424468994, "learning_rate": 4.8632062213060995e-05, "loss": 0.1707, "step": 8869 }, { "epoch": 0.37396968610999853, "grad_norm": 0.6142702698707581, "learning_rate": 4.8631281426652926e-05, "loss": 0.4978, "step": 8870 }, { "epoch": 0.37401184729219805, "grad_norm": 1.9042789936065674, "learning_rate": 4.863050064024486e-05, "loss": 0.3332, "step": 8871 }, { "epoch": 0.3740540084743976, "grad_norm": 1.2859556674957275, "learning_rate": 4.862971985383679e-05, "loss": 0.2761, "step": 8872 }, { "epoch": 0.37409616965659714, "grad_norm": 0.8684176206588745, "learning_rate": 4.862893906742872e-05, "loss": 0.4917, "step": 8873 }, { "epoch": 0.3741383308387967, "grad_norm": 0.422676682472229, "learning_rate": 4.8628158281020644e-05, "loss": 0.1014, "step": 8874 }, { "epoch": 0.3741804920209963, "grad_norm": 3.7986748218536377, "learning_rate": 4.8627377494612575e-05, "loss": 1.0938, "step": 8875 }, { "epoch": 0.3742226532031958, "grad_norm": 0.8227003216743469, "learning_rate": 4.862659670820451e-05, "loss": 0.1854, "step": 8876 }, { "epoch": 0.3742648143853954, "grad_norm": 1.1879677772521973, "learning_rate": 4.862581592179644e-05, "loss": 0.4036, "step": 8877 }, { "epoch": 0.3743069755675949, "grad_norm": 0.6648616790771484, "learning_rate": 4.862503513538836e-05, "loss": 0.1067, "step": 8878 }, { "epoch": 0.3743491367497945, "grad_norm": 0.9698732495307922, "learning_rate": 4.8624254348980294e-05, "loss": 0.3963, "step": 8879 }, { "epoch": 0.374391297931994, "grad_norm": 1.6758633852005005, "learning_rate": 4.8623473562572225e-05, "loss": 0.4089, "step": 8880 }, { "epoch": 0.37443345911419357, "grad_norm": 14.539209365844727, "learning_rate": 4.862269277616415e-05, "loss": 0.1509, "step": 8881 }, { "epoch": 0.3744756202963931, "grad_norm": 1.9053895473480225, "learning_rate": 4.862191198975609e-05, "loss": 0.6516, "step": 8882 }, { "epoch": 0.37451778147859266, "grad_norm": 0.9488234519958496, "learning_rate": 4.862113120334801e-05, "loss": 0.2149, "step": 8883 }, { "epoch": 0.37455994266079223, "grad_norm": 17.06853485107422, "learning_rate": 4.862035041693994e-05, "loss": 0.3785, "step": 8884 }, { "epoch": 0.37460210384299175, "grad_norm": 14.827667236328125, "learning_rate": 4.8619569630531875e-05, "loss": 0.399, "step": 8885 }, { "epoch": 0.3746442650251913, "grad_norm": 0.43597689270973206, "learning_rate": 4.8618788844123806e-05, "loss": 0.1499, "step": 8886 }, { "epoch": 0.37468642620739084, "grad_norm": 0.9880202412605286, "learning_rate": 4.861800805771574e-05, "loss": 0.5513, "step": 8887 }, { "epoch": 0.3747285873895904, "grad_norm": 2.7238595485687256, "learning_rate": 4.861722727130766e-05, "loss": 0.5798, "step": 8888 }, { "epoch": 0.37477074857178994, "grad_norm": 0.37299853563308716, "learning_rate": 4.861644648489959e-05, "loss": 0.1532, "step": 8889 }, { "epoch": 0.3748129097539895, "grad_norm": 1.513210654258728, "learning_rate": 4.8615665698491524e-05, "loss": 0.2345, "step": 8890 }, { "epoch": 0.374855070936189, "grad_norm": 0.9357491135597229, "learning_rate": 4.8614884912083455e-05, "loss": 0.5828, "step": 8891 }, { "epoch": 0.3748972321183886, "grad_norm": 0.4298141896724701, "learning_rate": 4.861410412567538e-05, "loss": 0.1417, "step": 8892 }, { "epoch": 0.3749393933005882, "grad_norm": 0.8516545295715332, "learning_rate": 4.861332333926731e-05, "loss": 0.3377, "step": 8893 }, { "epoch": 0.3749815544827877, "grad_norm": 0.6066991686820984, "learning_rate": 4.861254255285924e-05, "loss": 0.5572, "step": 8894 }, { "epoch": 0.37502371566498727, "grad_norm": 0.5134674906730652, "learning_rate": 4.861176176645117e-05, "loss": 0.1602, "step": 8895 }, { "epoch": 0.3750658768471868, "grad_norm": 0.8843138813972473, "learning_rate": 4.8610980980043105e-05, "loss": 0.2729, "step": 8896 }, { "epoch": 0.37510803802938636, "grad_norm": 0.6966291069984436, "learning_rate": 4.861020019363503e-05, "loss": 0.555, "step": 8897 }, { "epoch": 0.3751501992115859, "grad_norm": 0.5299340486526489, "learning_rate": 4.860941940722696e-05, "loss": 0.107, "step": 8898 }, { "epoch": 0.37519236039378545, "grad_norm": 2.1123697757720947, "learning_rate": 4.860863862081889e-05, "loss": 0.4195, "step": 8899 }, { "epoch": 0.37523452157598497, "grad_norm": 0.5283926725387573, "learning_rate": 4.8607857834410816e-05, "loss": 0.2016, "step": 8900 }, { "epoch": 0.37527668275818454, "grad_norm": 1394830.5, "learning_rate": 4.8607077048002754e-05, "loss": 0.6083, "step": 8901 }, { "epoch": 0.37531884394038406, "grad_norm": 0.4164530038833618, "learning_rate": 4.860629626159468e-05, "loss": 0.0938, "step": 8902 }, { "epoch": 0.37536100512258364, "grad_norm": 0.2946752905845642, "learning_rate": 4.860551547518661e-05, "loss": 0.0603, "step": 8903 }, { "epoch": 0.3754031663047832, "grad_norm": 2.1391329765319824, "learning_rate": 4.860473468877854e-05, "loss": 0.3222, "step": 8904 }, { "epoch": 0.3754453274869827, "grad_norm": 0.4802580773830414, "learning_rate": 4.860395390237047e-05, "loss": 0.3063, "step": 8905 }, { "epoch": 0.3754874886691823, "grad_norm": 0.5357673168182373, "learning_rate": 4.86031731159624e-05, "loss": 0.1268, "step": 8906 }, { "epoch": 0.3755296498513818, "grad_norm": 1.0033788681030273, "learning_rate": 4.860239232955433e-05, "loss": 1.1423, "step": 8907 }, { "epoch": 0.3755718110335814, "grad_norm": 0.44453200697898865, "learning_rate": 4.860161154314626e-05, "loss": 0.1342, "step": 8908 }, { "epoch": 0.3756139722157809, "grad_norm": 4.453742504119873, "learning_rate": 4.860083075673819e-05, "loss": 0.8174, "step": 8909 }, { "epoch": 0.3756561333979805, "grad_norm": 0.7493364214897156, "learning_rate": 4.860004997033012e-05, "loss": 0.0959, "step": 8910 }, { "epoch": 0.37569829458018, "grad_norm": 0.6098311543464661, "learning_rate": 4.8599269183922047e-05, "loss": 0.1896, "step": 8911 }, { "epoch": 0.3757404557623796, "grad_norm": 0.5813335180282593, "learning_rate": 4.859848839751398e-05, "loss": 0.1438, "step": 8912 }, { "epoch": 0.37578261694457915, "grad_norm": 0.33250299096107483, "learning_rate": 4.859770761110591e-05, "loss": 0.1166, "step": 8913 }, { "epoch": 0.37582477812677867, "grad_norm": 0.5233846306800842, "learning_rate": 4.8596926824697834e-05, "loss": 0.1469, "step": 8914 }, { "epoch": 0.37586693930897824, "grad_norm": 0.8839681148529053, "learning_rate": 4.859614603828977e-05, "loss": 0.2023, "step": 8915 }, { "epoch": 0.37590910049117776, "grad_norm": 0.4025192856788635, "learning_rate": 4.8595365251881696e-05, "loss": 0.0997, "step": 8916 }, { "epoch": 0.37595126167337733, "grad_norm": 1.7359074354171753, "learning_rate": 4.859458446547363e-05, "loss": 0.5037, "step": 8917 }, { "epoch": 0.37599342285557685, "grad_norm": 0.5549665689468384, "learning_rate": 4.859380367906556e-05, "loss": 0.0852, "step": 8918 }, { "epoch": 0.3760355840377764, "grad_norm": 0.9939571022987366, "learning_rate": 4.859302289265748e-05, "loss": 0.9498, "step": 8919 }, { "epoch": 0.37607774521997595, "grad_norm": 0.5052118301391602, "learning_rate": 4.859224210624942e-05, "loss": 0.2538, "step": 8920 }, { "epoch": 0.3761199064021755, "grad_norm": 0.35097163915634155, "learning_rate": 4.8591461319841346e-05, "loss": 0.0859, "step": 8921 }, { "epoch": 0.3761620675843751, "grad_norm": 0.9986987709999084, "learning_rate": 4.859068053343328e-05, "loss": 0.5694, "step": 8922 }, { "epoch": 0.3762042287665746, "grad_norm": 0.30290868878364563, "learning_rate": 4.858989974702521e-05, "loss": 0.0937, "step": 8923 }, { "epoch": 0.3762463899487742, "grad_norm": 0.651939332485199, "learning_rate": 4.858911896061713e-05, "loss": 0.2299, "step": 8924 }, { "epoch": 0.3762885511309737, "grad_norm": 3.1118037700653076, "learning_rate": 4.8588338174209064e-05, "loss": 0.6279, "step": 8925 }, { "epoch": 0.3763307123131733, "grad_norm": 0.856381893157959, "learning_rate": 4.8587557387800995e-05, "loss": 0.3062, "step": 8926 }, { "epoch": 0.3763728734953728, "grad_norm": 1.3003836870193481, "learning_rate": 4.8586776601392926e-05, "loss": 0.3714, "step": 8927 }, { "epoch": 0.37641503467757237, "grad_norm": 0.7700074315071106, "learning_rate": 4.858599581498485e-05, "loss": 0.2588, "step": 8928 }, { "epoch": 0.3764571958597719, "grad_norm": 0.717318058013916, "learning_rate": 4.858521502857679e-05, "loss": 0.147, "step": 8929 }, { "epoch": 0.37649935704197146, "grad_norm": 0.7910098433494568, "learning_rate": 4.8584434242168713e-05, "loss": 0.1535, "step": 8930 }, { "epoch": 0.376541518224171, "grad_norm": 0.9176425933837891, "learning_rate": 4.8583653455760645e-05, "loss": 0.3722, "step": 8931 }, { "epoch": 0.37658367940637055, "grad_norm": 1.094254493713379, "learning_rate": 4.8582872669352576e-05, "loss": 0.7786, "step": 8932 }, { "epoch": 0.3766258405885701, "grad_norm": 0.3582988381385803, "learning_rate": 4.85820918829445e-05, "loss": 0.0362, "step": 8933 }, { "epoch": 0.37666800177076964, "grad_norm": 0.5600628852844238, "learning_rate": 4.858131109653644e-05, "loss": 0.0797, "step": 8934 }, { "epoch": 0.3767101629529692, "grad_norm": 0.9363256096839905, "learning_rate": 4.858053031012836e-05, "loss": 0.2235, "step": 8935 }, { "epoch": 0.37675232413516874, "grad_norm": 0.790461003780365, "learning_rate": 4.8579749523720294e-05, "loss": 0.1055, "step": 8936 }, { "epoch": 0.3767944853173683, "grad_norm": 0.5900467038154602, "learning_rate": 4.8578968737312225e-05, "loss": 0.5565, "step": 8937 }, { "epoch": 0.37683664649956783, "grad_norm": 0.5369819402694702, "learning_rate": 4.857818795090415e-05, "loss": 0.205, "step": 8938 }, { "epoch": 0.3768788076817674, "grad_norm": 0.277014821767807, "learning_rate": 4.857740716449608e-05, "loss": 0.0805, "step": 8939 }, { "epoch": 0.3769209688639669, "grad_norm": 1.6621510982513428, "learning_rate": 4.857662637808801e-05, "loss": 0.8331, "step": 8940 }, { "epoch": 0.3769631300461665, "grad_norm": 0.49743354320526123, "learning_rate": 4.8575845591679944e-05, "loss": 0.1779, "step": 8941 }, { "epoch": 0.37700529122836607, "grad_norm": 0.5649973154067993, "learning_rate": 4.8575064805271875e-05, "loss": 0.0893, "step": 8942 }, { "epoch": 0.3770474524105656, "grad_norm": 1.541007161140442, "learning_rate": 4.85742840188638e-05, "loss": 0.938, "step": 8943 }, { "epoch": 0.37708961359276516, "grad_norm": 2.198394775390625, "learning_rate": 4.857350323245573e-05, "loss": 0.8698, "step": 8944 }, { "epoch": 0.3771317747749647, "grad_norm": 0.5430055260658264, "learning_rate": 4.857272244604766e-05, "loss": 0.1847, "step": 8945 }, { "epoch": 0.37717393595716425, "grad_norm": 1.4501608610153198, "learning_rate": 4.857194165963959e-05, "loss": 0.728, "step": 8946 }, { "epoch": 0.37721609713936377, "grad_norm": 171.70529174804688, "learning_rate": 4.857116087323152e-05, "loss": 1.4938, "step": 8947 }, { "epoch": 0.37725825832156334, "grad_norm": 0.567425012588501, "learning_rate": 4.8570380086823456e-05, "loss": 0.5438, "step": 8948 }, { "epoch": 0.37730041950376286, "grad_norm": 0.5857519507408142, "learning_rate": 4.856959930041538e-05, "loss": 0.5931, "step": 8949 }, { "epoch": 0.37734258068596244, "grad_norm": 1.9688307046890259, "learning_rate": 4.856881851400731e-05, "loss": 0.1849, "step": 8950 }, { "epoch": 0.377384741868162, "grad_norm": 1.9863693714141846, "learning_rate": 4.856803772759924e-05, "loss": 1.1743, "step": 8951 }, { "epoch": 0.37742690305036153, "grad_norm": 0.720667839050293, "learning_rate": 4.856725694119117e-05, "loss": 0.3541, "step": 8952 }, { "epoch": 0.3774690642325611, "grad_norm": 0.2966729700565338, "learning_rate": 4.85664761547831e-05, "loss": 0.0865, "step": 8953 }, { "epoch": 0.3775112254147606, "grad_norm": 0.6698333024978638, "learning_rate": 4.856569536837503e-05, "loss": 0.4808, "step": 8954 }, { "epoch": 0.3775533865969602, "grad_norm": 0.8997393846511841, "learning_rate": 4.856491458196696e-05, "loss": 0.5866, "step": 8955 }, { "epoch": 0.3775955477791597, "grad_norm": 9.040731430053711, "learning_rate": 4.856413379555889e-05, "loss": 0.1733, "step": 8956 }, { "epoch": 0.3776377089613593, "grad_norm": 1.8739615678787231, "learning_rate": 4.856335300915082e-05, "loss": 0.4947, "step": 8957 }, { "epoch": 0.3776798701435588, "grad_norm": 1.1317001581192017, "learning_rate": 4.856257222274275e-05, "loss": 0.9172, "step": 8958 }, { "epoch": 0.3777220313257584, "grad_norm": 1.7288715839385986, "learning_rate": 4.856179143633468e-05, "loss": 0.1922, "step": 8959 }, { "epoch": 0.3777641925079579, "grad_norm": 0.3826458156108856, "learning_rate": 4.856101064992661e-05, "loss": 0.0958, "step": 8960 }, { "epoch": 0.37780635369015747, "grad_norm": 0.26985034346580505, "learning_rate": 4.8560229863518535e-05, "loss": 0.0673, "step": 8961 }, { "epoch": 0.37784851487235704, "grad_norm": 3.0401008129119873, "learning_rate": 4.8559449077110466e-05, "loss": 0.5389, "step": 8962 }, { "epoch": 0.37789067605455656, "grad_norm": 3.3346211910247803, "learning_rate": 4.85586682907024e-05, "loss": 0.1501, "step": 8963 }, { "epoch": 0.37793283723675614, "grad_norm": 1.2040361166000366, "learning_rate": 4.855788750429432e-05, "loss": 1.022, "step": 8964 }, { "epoch": 0.37797499841895565, "grad_norm": 0.6765183806419373, "learning_rate": 4.855710671788626e-05, "loss": 0.3184, "step": 8965 }, { "epoch": 0.37801715960115523, "grad_norm": 0.7378234267234802, "learning_rate": 4.8556325931478185e-05, "loss": 0.2402, "step": 8966 }, { "epoch": 0.37805932078335475, "grad_norm": 1.4588874578475952, "learning_rate": 4.855554514507012e-05, "loss": 0.4724, "step": 8967 }, { "epoch": 0.3781014819655543, "grad_norm": 0.3208768367767334, "learning_rate": 4.855476435866205e-05, "loss": 0.0678, "step": 8968 }, { "epoch": 0.37814364314775384, "grad_norm": 0.73331618309021, "learning_rate": 4.855398357225397e-05, "loss": 0.2058, "step": 8969 }, { "epoch": 0.3781858043299534, "grad_norm": 0.6028226017951965, "learning_rate": 4.855320278584591e-05, "loss": 0.1263, "step": 8970 }, { "epoch": 0.378227965512153, "grad_norm": 0.6038063764572144, "learning_rate": 4.8552421999437834e-05, "loss": 0.3408, "step": 8971 }, { "epoch": 0.3782701266943525, "grad_norm": 0.585770845413208, "learning_rate": 4.8551641213029765e-05, "loss": 0.189, "step": 8972 }, { "epoch": 0.3783122878765521, "grad_norm": 6.643801689147949, "learning_rate": 4.85508604266217e-05, "loss": 0.4211, "step": 8973 }, { "epoch": 0.3783544490587516, "grad_norm": 1.2551372051239014, "learning_rate": 4.855007964021363e-05, "loss": 0.9379, "step": 8974 }, { "epoch": 0.37839661024095117, "grad_norm": 0.999953031539917, "learning_rate": 4.854929885380555e-05, "loss": 0.1321, "step": 8975 }, { "epoch": 0.3784387714231507, "grad_norm": 0.8755267262458801, "learning_rate": 4.8548518067397484e-05, "loss": 0.5773, "step": 8976 }, { "epoch": 0.37848093260535026, "grad_norm": 1.2045444250106812, "learning_rate": 4.8547737280989415e-05, "loss": 0.2085, "step": 8977 }, { "epoch": 0.3785230937875498, "grad_norm": 0.6654667258262634, "learning_rate": 4.8546956494581346e-05, "loss": 0.2013, "step": 8978 }, { "epoch": 0.37856525496974935, "grad_norm": 1.3895703554153442, "learning_rate": 4.854617570817328e-05, "loss": 0.5591, "step": 8979 }, { "epoch": 0.37860741615194893, "grad_norm": 0.5381752252578735, "learning_rate": 4.85453949217652e-05, "loss": 0.1779, "step": 8980 }, { "epoch": 0.37864957733414845, "grad_norm": 1.013837218284607, "learning_rate": 4.854461413535713e-05, "loss": 0.3217, "step": 8981 }, { "epoch": 0.378691738516348, "grad_norm": 0.8194629549980164, "learning_rate": 4.8543833348949064e-05, "loss": 0.7607, "step": 8982 }, { "epoch": 0.37873389969854754, "grad_norm": 2.273028612136841, "learning_rate": 4.854305256254099e-05, "loss": 0.2204, "step": 8983 }, { "epoch": 0.3787760608807471, "grad_norm": 0.4145943522453308, "learning_rate": 4.854227177613293e-05, "loss": 0.0989, "step": 8984 }, { "epoch": 0.37881822206294663, "grad_norm": 0.7800818681716919, "learning_rate": 4.854149098972485e-05, "loss": 0.0967, "step": 8985 }, { "epoch": 0.3788603832451462, "grad_norm": 0.452155202627182, "learning_rate": 4.854071020331678e-05, "loss": 0.134, "step": 8986 }, { "epoch": 0.3789025444273457, "grad_norm": 0.5036270618438721, "learning_rate": 4.8539929416908714e-05, "loss": 0.1574, "step": 8987 }, { "epoch": 0.3789447056095453, "grad_norm": 1.4131489992141724, "learning_rate": 4.853914863050064e-05, "loss": 0.163, "step": 8988 }, { "epoch": 0.3789868667917448, "grad_norm": 0.5518298745155334, "learning_rate": 4.8538367844092576e-05, "loss": 0.0832, "step": 8989 }, { "epoch": 0.3790290279739444, "grad_norm": 4.674736022949219, "learning_rate": 4.85375870576845e-05, "loss": 1.2315, "step": 8990 }, { "epoch": 0.37907118915614396, "grad_norm": 1.3765442371368408, "learning_rate": 4.853680627127643e-05, "loss": 0.215, "step": 8991 }, { "epoch": 0.3791133503383435, "grad_norm": 0.786905825138092, "learning_rate": 4.8536025484868363e-05, "loss": 0.2802, "step": 8992 }, { "epoch": 0.37915551152054305, "grad_norm": 0.8068037629127502, "learning_rate": 4.8535244698460295e-05, "loss": 0.1526, "step": 8993 }, { "epoch": 0.37919767270274257, "grad_norm": 0.4098544418811798, "learning_rate": 4.853446391205222e-05, "loss": 0.1231, "step": 8994 }, { "epoch": 0.37923983388494215, "grad_norm": 0.5915040969848633, "learning_rate": 4.853368312564415e-05, "loss": 0.1331, "step": 8995 }, { "epoch": 0.37928199506714166, "grad_norm": 0.5481694936752319, "learning_rate": 4.853290233923608e-05, "loss": 0.2084, "step": 8996 }, { "epoch": 0.37932415624934124, "grad_norm": 2.9582200050354004, "learning_rate": 4.8532121552828006e-05, "loss": 0.5876, "step": 8997 }, { "epoch": 0.37936631743154076, "grad_norm": 5.025558948516846, "learning_rate": 4.8531340766419944e-05, "loss": 0.5064, "step": 8998 }, { "epoch": 0.37940847861374033, "grad_norm": 0.5217496156692505, "learning_rate": 4.853055998001187e-05, "loss": 0.3391, "step": 8999 }, { "epoch": 0.3794506397959399, "grad_norm": 4.1686201095581055, "learning_rate": 4.85297791936038e-05, "loss": 0.9149, "step": 9000 }, { "epoch": 0.3794928009781394, "grad_norm": 0.8576016426086426, "learning_rate": 4.852899840719573e-05, "loss": 0.6201, "step": 9001 }, { "epoch": 0.379534962160339, "grad_norm": 0.4388183653354645, "learning_rate": 4.8528217620787656e-05, "loss": 0.1428, "step": 9002 }, { "epoch": 0.3795771233425385, "grad_norm": 1.0867719650268555, "learning_rate": 4.8527436834379594e-05, "loss": 0.8152, "step": 9003 }, { "epoch": 0.3796192845247381, "grad_norm": 2.127243757247925, "learning_rate": 4.852665604797152e-05, "loss": 0.2457, "step": 9004 }, { "epoch": 0.3796614457069376, "grad_norm": 0.668100118637085, "learning_rate": 4.852587526156345e-05, "loss": 0.1439, "step": 9005 }, { "epoch": 0.3797036068891372, "grad_norm": 1.5328309535980225, "learning_rate": 4.852509447515538e-05, "loss": 0.4663, "step": 9006 }, { "epoch": 0.3797457680713367, "grad_norm": 1.1455084085464478, "learning_rate": 4.8524313688747305e-05, "loss": 0.1985, "step": 9007 }, { "epoch": 0.37978792925353627, "grad_norm": 0.7798958420753479, "learning_rate": 4.8523532902339237e-05, "loss": 0.4785, "step": 9008 }, { "epoch": 0.37983009043573585, "grad_norm": 0.6592609286308289, "learning_rate": 4.852275211593117e-05, "loss": 0.1899, "step": 9009 }, { "epoch": 0.37987225161793536, "grad_norm": 1.5101120471954346, "learning_rate": 4.85219713295231e-05, "loss": 0.1806, "step": 9010 }, { "epoch": 0.37991441280013494, "grad_norm": 0.42839911580085754, "learning_rate": 4.852119054311503e-05, "loss": 0.1444, "step": 9011 }, { "epoch": 0.37995657398233446, "grad_norm": 0.4028065502643585, "learning_rate": 4.852040975670696e-05, "loss": 0.091, "step": 9012 }, { "epoch": 0.37999873516453403, "grad_norm": 0.9540210962295532, "learning_rate": 4.8519628970298886e-05, "loss": 0.2659, "step": 9013 }, { "epoch": 0.38004089634673355, "grad_norm": 1.6368002891540527, "learning_rate": 4.851884818389082e-05, "loss": 0.4268, "step": 9014 }, { "epoch": 0.3800830575289331, "grad_norm": 0.7849382162094116, "learning_rate": 4.851806739748275e-05, "loss": 0.906, "step": 9015 }, { "epoch": 0.38012521871113264, "grad_norm": 0.8576874136924744, "learning_rate": 4.851728661107467e-05, "loss": 0.176, "step": 9016 }, { "epoch": 0.3801673798933322, "grad_norm": 0.4530232548713684, "learning_rate": 4.851650582466661e-05, "loss": 0.1283, "step": 9017 }, { "epoch": 0.38020954107553173, "grad_norm": 0.6589853167533875, "learning_rate": 4.8515725038258536e-05, "loss": 0.1731, "step": 9018 }, { "epoch": 0.3802517022577313, "grad_norm": 0.6452889442443848, "learning_rate": 4.851494425185047e-05, "loss": 0.1502, "step": 9019 }, { "epoch": 0.3802938634399309, "grad_norm": 0.6090275645256042, "learning_rate": 4.85141634654424e-05, "loss": 0.1993, "step": 9020 }, { "epoch": 0.3803360246221304, "grad_norm": 0.7136497497558594, "learning_rate": 4.851338267903432e-05, "loss": 0.5165, "step": 9021 }, { "epoch": 0.38037818580432997, "grad_norm": 0.30648401379585266, "learning_rate": 4.851260189262626e-05, "loss": 0.0818, "step": 9022 }, { "epoch": 0.3804203469865295, "grad_norm": 1.4354270696640015, "learning_rate": 4.8511821106218185e-05, "loss": 1.1444, "step": 9023 }, { "epoch": 0.38046250816872906, "grad_norm": 1.0757572650909424, "learning_rate": 4.8511040319810116e-05, "loss": 1.0042, "step": 9024 }, { "epoch": 0.3805046693509286, "grad_norm": 2.358799695968628, "learning_rate": 4.851025953340205e-05, "loss": 0.9906, "step": 9025 }, { "epoch": 0.38054683053312816, "grad_norm": 0.7793270945549011, "learning_rate": 4.850947874699397e-05, "loss": 0.2525, "step": 9026 }, { "epoch": 0.3805889917153277, "grad_norm": 1.3725495338439941, "learning_rate": 4.85086979605859e-05, "loss": 1.0564, "step": 9027 }, { "epoch": 0.38063115289752725, "grad_norm": 3.6746556758880615, "learning_rate": 4.8507917174177835e-05, "loss": 0.7996, "step": 9028 }, { "epoch": 0.3806733140797268, "grad_norm": 0.6929364204406738, "learning_rate": 4.8507136387769766e-05, "loss": 0.2752, "step": 9029 }, { "epoch": 0.38071547526192634, "grad_norm": 0.5892654061317444, "learning_rate": 4.850635560136169e-05, "loss": 0.1387, "step": 9030 }, { "epoch": 0.3807576364441259, "grad_norm": 0.4026784300804138, "learning_rate": 4.850557481495362e-05, "loss": 0.1596, "step": 9031 }, { "epoch": 0.38079979762632543, "grad_norm": 0.8274335265159607, "learning_rate": 4.850479402854555e-05, "loss": 0.1996, "step": 9032 }, { "epoch": 0.380841958808525, "grad_norm": 7.258777141571045, "learning_rate": 4.850401324213748e-05, "loss": 0.3129, "step": 9033 }, { "epoch": 0.3808841199907245, "grad_norm": 0.8737397193908691, "learning_rate": 4.8503232455729415e-05, "loss": 0.5356, "step": 9034 }, { "epoch": 0.3809262811729241, "grad_norm": 1.2895121574401855, "learning_rate": 4.850245166932134e-05, "loss": 0.2549, "step": 9035 }, { "epoch": 0.3809684423551236, "grad_norm": 1.4116723537445068, "learning_rate": 4.850167088291328e-05, "loss": 0.2586, "step": 9036 }, { "epoch": 0.3810106035373232, "grad_norm": 0.949880838394165, "learning_rate": 4.85008900965052e-05, "loss": 0.5135, "step": 9037 }, { "epoch": 0.38105276471952276, "grad_norm": 0.7817506790161133, "learning_rate": 4.8500109310097134e-05, "loss": 0.1834, "step": 9038 }, { "epoch": 0.3810949259017223, "grad_norm": 0.8739628791809082, "learning_rate": 4.8499328523689065e-05, "loss": 0.9438, "step": 9039 }, { "epoch": 0.38113708708392185, "grad_norm": 1.3207870721817017, "learning_rate": 4.849854773728099e-05, "loss": 0.2902, "step": 9040 }, { "epoch": 0.3811792482661214, "grad_norm": 1.275718331336975, "learning_rate": 4.849776695087292e-05, "loss": 0.1557, "step": 9041 }, { "epoch": 0.38122140944832095, "grad_norm": 0.6094449758529663, "learning_rate": 4.849698616446485e-05, "loss": 0.1906, "step": 9042 }, { "epoch": 0.38126357063052047, "grad_norm": 1.0828245878219604, "learning_rate": 4.849620537805678e-05, "loss": 0.9876, "step": 9043 }, { "epoch": 0.38130573181272004, "grad_norm": 0.7766636610031128, "learning_rate": 4.849542459164871e-05, "loss": 0.1701, "step": 9044 }, { "epoch": 0.38134789299491956, "grad_norm": 1.3841354846954346, "learning_rate": 4.849464380524064e-05, "loss": 0.5967, "step": 9045 }, { "epoch": 0.38139005417711913, "grad_norm": 0.4586750268936157, "learning_rate": 4.849386301883257e-05, "loss": 0.231, "step": 9046 }, { "epoch": 0.38143221535931865, "grad_norm": 0.663866400718689, "learning_rate": 4.84930822324245e-05, "loss": 0.2848, "step": 9047 }, { "epoch": 0.3814743765415182, "grad_norm": 0.5517528653144836, "learning_rate": 4.849230144601643e-05, "loss": 0.1447, "step": 9048 }, { "epoch": 0.3815165377237178, "grad_norm": 0.6582919359207153, "learning_rate": 4.849152065960836e-05, "loss": 0.3012, "step": 9049 }, { "epoch": 0.3815586989059173, "grad_norm": 10.532134056091309, "learning_rate": 4.849073987320029e-05, "loss": 0.0988, "step": 9050 }, { "epoch": 0.3816008600881169, "grad_norm": 0.37075334787368774, "learning_rate": 4.848995908679222e-05, "loss": 0.1736, "step": 9051 }, { "epoch": 0.3816430212703164, "grad_norm": 0.8113101124763489, "learning_rate": 4.8489178300384144e-05, "loss": 0.2366, "step": 9052 }, { "epoch": 0.381685182452516, "grad_norm": 1.0872122049331665, "learning_rate": 4.848839751397608e-05, "loss": 0.1256, "step": 9053 }, { "epoch": 0.3817273436347155, "grad_norm": 0.8247434496879578, "learning_rate": 4.848761672756801e-05, "loss": 0.5161, "step": 9054 }, { "epoch": 0.3817695048169151, "grad_norm": 1.7131717205047607, "learning_rate": 4.848683594115994e-05, "loss": 0.4391, "step": 9055 }, { "epoch": 0.3818116659991146, "grad_norm": 0.9311027526855469, "learning_rate": 4.848605515475187e-05, "loss": 1.0526, "step": 9056 }, { "epoch": 0.38185382718131416, "grad_norm": 0.9552420377731323, "learning_rate": 4.84852743683438e-05, "loss": 0.9303, "step": 9057 }, { "epoch": 0.38189598836351374, "grad_norm": 0.6114863753318787, "learning_rate": 4.848449358193573e-05, "loss": 0.7856, "step": 9058 }, { "epoch": 0.38193814954571326, "grad_norm": 1.0969727039337158, "learning_rate": 4.8483712795527656e-05, "loss": 0.5516, "step": 9059 }, { "epoch": 0.38198031072791283, "grad_norm": 0.6824325323104858, "learning_rate": 4.848293200911959e-05, "loss": 0.5717, "step": 9060 }, { "epoch": 0.38202247191011235, "grad_norm": 0.40947338938713074, "learning_rate": 4.848215122271152e-05, "loss": 0.0847, "step": 9061 }, { "epoch": 0.3820646330923119, "grad_norm": 1.013066291809082, "learning_rate": 4.848137043630345e-05, "loss": 0.5654, "step": 9062 }, { "epoch": 0.38210679427451144, "grad_norm": 0.3123890161514282, "learning_rate": 4.8480589649895374e-05, "loss": 0.1205, "step": 9063 }, { "epoch": 0.382148955456711, "grad_norm": 1.0674632787704468, "learning_rate": 4.8479808863487306e-05, "loss": 0.1388, "step": 9064 }, { "epoch": 0.38219111663891053, "grad_norm": 0.38075658679008484, "learning_rate": 4.847902807707924e-05, "loss": 0.1633, "step": 9065 }, { "epoch": 0.3822332778211101, "grad_norm": 0.24646475911140442, "learning_rate": 4.847824729067116e-05, "loss": 0.0622, "step": 9066 }, { "epoch": 0.3822754390033097, "grad_norm": 0.9484783411026001, "learning_rate": 4.84774665042631e-05, "loss": 0.5983, "step": 9067 }, { "epoch": 0.3823176001855092, "grad_norm": 0.9699220061302185, "learning_rate": 4.8476685717855024e-05, "loss": 0.3672, "step": 9068 }, { "epoch": 0.3823597613677088, "grad_norm": 0.4609724283218384, "learning_rate": 4.8475904931446955e-05, "loss": 0.3331, "step": 9069 }, { "epoch": 0.3824019225499083, "grad_norm": 0.4010120928287506, "learning_rate": 4.8475124145038887e-05, "loss": 0.1293, "step": 9070 }, { "epoch": 0.38244408373210786, "grad_norm": 0.5984978675842285, "learning_rate": 4.847434335863081e-05, "loss": 0.5538, "step": 9071 }, { "epoch": 0.3824862449143074, "grad_norm": 0.40562641620635986, "learning_rate": 4.847356257222275e-05, "loss": 0.1188, "step": 9072 }, { "epoch": 0.38252840609650696, "grad_norm": 0.45297688245773315, "learning_rate": 4.8472781785814674e-05, "loss": 0.2, "step": 9073 }, { "epoch": 0.3825705672787065, "grad_norm": 0.4682612717151642, "learning_rate": 4.8472000999406605e-05, "loss": 0.203, "step": 9074 }, { "epoch": 0.38261272846090605, "grad_norm": 0.5063762068748474, "learning_rate": 4.8471220212998536e-05, "loss": 0.0973, "step": 9075 }, { "epoch": 0.38265488964310557, "grad_norm": 0.6891642808914185, "learning_rate": 4.847043942659046e-05, "loss": 0.3684, "step": 9076 }, { "epoch": 0.38269705082530514, "grad_norm": 0.9383666515350342, "learning_rate": 4.846965864018239e-05, "loss": 0.4688, "step": 9077 }, { "epoch": 0.3827392120075047, "grad_norm": 0.593911349773407, "learning_rate": 4.846887785377432e-05, "loss": 0.1307, "step": 9078 }, { "epoch": 0.38278137318970423, "grad_norm": 1.0547125339508057, "learning_rate": 4.8468097067366254e-05, "loss": 0.2715, "step": 9079 }, { "epoch": 0.3828235343719038, "grad_norm": 0.5145197510719299, "learning_rate": 4.8467316280958186e-05, "loss": 0.1078, "step": 9080 }, { "epoch": 0.3828656955541033, "grad_norm": 0.3004433214664459, "learning_rate": 4.846653549455012e-05, "loss": 0.0725, "step": 9081 }, { "epoch": 0.3829078567363029, "grad_norm": 0.4841165244579315, "learning_rate": 4.846575470814204e-05, "loss": 0.1769, "step": 9082 }, { "epoch": 0.3829500179185024, "grad_norm": 0.46445393562316895, "learning_rate": 4.846497392173397e-05, "loss": 0.1809, "step": 9083 }, { "epoch": 0.382992179100702, "grad_norm": 2.0794572830200195, "learning_rate": 4.8464193135325904e-05, "loss": 0.7502, "step": 9084 }, { "epoch": 0.3830343402829015, "grad_norm": 0.7920101881027222, "learning_rate": 4.846341234891783e-05, "loss": 0.2655, "step": 9085 }, { "epoch": 0.3830765014651011, "grad_norm": 0.9812652468681335, "learning_rate": 4.8462631562509766e-05, "loss": 0.2528, "step": 9086 }, { "epoch": 0.38311866264730066, "grad_norm": 0.5493429899215698, "learning_rate": 4.846185077610169e-05, "loss": 0.22, "step": 9087 }, { "epoch": 0.3831608238295002, "grad_norm": 0.5269830226898193, "learning_rate": 4.846106998969362e-05, "loss": 0.2367, "step": 9088 }, { "epoch": 0.38320298501169975, "grad_norm": 1.9009863138198853, "learning_rate": 4.846028920328555e-05, "loss": 0.5332, "step": 9089 }, { "epoch": 0.38324514619389927, "grad_norm": 1.0567511320114136, "learning_rate": 4.845950841687748e-05, "loss": 0.9457, "step": 9090 }, { "epoch": 0.38328730737609884, "grad_norm": 1.4893081188201904, "learning_rate": 4.8458727630469416e-05, "loss": 0.5193, "step": 9091 }, { "epoch": 0.38332946855829836, "grad_norm": 0.8849568367004395, "learning_rate": 4.845794684406134e-05, "loss": 0.5431, "step": 9092 }, { "epoch": 0.38337162974049793, "grad_norm": 1.2168705463409424, "learning_rate": 4.845716605765327e-05, "loss": 0.7599, "step": 9093 }, { "epoch": 0.38341379092269745, "grad_norm": 1.1713773012161255, "learning_rate": 4.84563852712452e-05, "loss": 0.5365, "step": 9094 }, { "epoch": 0.383455952104897, "grad_norm": 1.2040979862213135, "learning_rate": 4.845560448483713e-05, "loss": 0.1432, "step": 9095 }, { "epoch": 0.38349811328709654, "grad_norm": 0.4593879282474518, "learning_rate": 4.845482369842906e-05, "loss": 0.1268, "step": 9096 }, { "epoch": 0.3835402744692961, "grad_norm": 0.3649628460407257, "learning_rate": 4.845404291202099e-05, "loss": 0.0813, "step": 9097 }, { "epoch": 0.3835824356514957, "grad_norm": 0.5369668006896973, "learning_rate": 4.845326212561292e-05, "loss": 0.2171, "step": 9098 }, { "epoch": 0.3836245968336952, "grad_norm": 1.163530945777893, "learning_rate": 4.8452481339204846e-05, "loss": 0.1568, "step": 9099 }, { "epoch": 0.3836667580158948, "grad_norm": 1.150615930557251, "learning_rate": 4.8451700552796784e-05, "loss": 1.1658, "step": 9100 }, { "epoch": 0.3837089191980943, "grad_norm": 0.7454409599304199, "learning_rate": 4.845091976638871e-05, "loss": 0.5709, "step": 9101 }, { "epoch": 0.3837510803802939, "grad_norm": 0.709385097026825, "learning_rate": 4.845013897998064e-05, "loss": 0.6573, "step": 9102 }, { "epoch": 0.3837932415624934, "grad_norm": 47.177120208740234, "learning_rate": 4.844935819357257e-05, "loss": 1.0006, "step": 9103 }, { "epoch": 0.38383540274469297, "grad_norm": 2.3974504470825195, "learning_rate": 4.8448577407164495e-05, "loss": 0.2784, "step": 9104 }, { "epoch": 0.3838775639268925, "grad_norm": 0.321808785200119, "learning_rate": 4.844779662075643e-05, "loss": 0.0904, "step": 9105 }, { "epoch": 0.38391972510909206, "grad_norm": 0.8978360891342163, "learning_rate": 4.844701583434836e-05, "loss": 0.1454, "step": 9106 }, { "epoch": 0.38396188629129163, "grad_norm": 2.040972948074341, "learning_rate": 4.844623504794029e-05, "loss": 0.641, "step": 9107 }, { "epoch": 0.38400404747349115, "grad_norm": 1.0583044290542603, "learning_rate": 4.844545426153222e-05, "loss": 1.0033, "step": 9108 }, { "epoch": 0.3840462086556907, "grad_norm": 10.365439414978027, "learning_rate": 4.8444673475124145e-05, "loss": 0.198, "step": 9109 }, { "epoch": 0.38408836983789024, "grad_norm": 0.568805456161499, "learning_rate": 4.8443892688716076e-05, "loss": 0.1755, "step": 9110 }, { "epoch": 0.3841305310200898, "grad_norm": 0.4516443908214569, "learning_rate": 4.844311190230801e-05, "loss": 0.1948, "step": 9111 }, { "epoch": 0.38417269220228933, "grad_norm": 0.33696630597114563, "learning_rate": 4.844233111589994e-05, "loss": 0.0781, "step": 9112 }, { "epoch": 0.3842148533844889, "grad_norm": 1.327772855758667, "learning_rate": 4.844155032949186e-05, "loss": 0.1875, "step": 9113 }, { "epoch": 0.3842570145666884, "grad_norm": 0.5181577801704407, "learning_rate": 4.8440769543083794e-05, "loss": 0.1163, "step": 9114 }, { "epoch": 0.384299175748888, "grad_norm": 0.5060354471206665, "learning_rate": 4.8439988756675725e-05, "loss": 0.1578, "step": 9115 }, { "epoch": 0.3843413369310876, "grad_norm": 0.5796499252319336, "learning_rate": 4.843920797026766e-05, "loss": 0.1128, "step": 9116 }, { "epoch": 0.3843834981132871, "grad_norm": 0.46316421031951904, "learning_rate": 4.843842718385959e-05, "loss": 0.1062, "step": 9117 }, { "epoch": 0.38442565929548667, "grad_norm": 0.8436246514320374, "learning_rate": 4.843764639745151e-05, "loss": 0.1176, "step": 9118 }, { "epoch": 0.3844678204776862, "grad_norm": 1.4713581800460815, "learning_rate": 4.843686561104345e-05, "loss": 1.0306, "step": 9119 }, { "epoch": 0.38450998165988576, "grad_norm": 0.9507941007614136, "learning_rate": 4.8436084824635375e-05, "loss": 0.39, "step": 9120 }, { "epoch": 0.3845521428420853, "grad_norm": 1.1664130687713623, "learning_rate": 4.84353040382273e-05, "loss": 0.3272, "step": 9121 }, { "epoch": 0.38459430402428485, "grad_norm": 0.5342376232147217, "learning_rate": 4.843452325181924e-05, "loss": 0.1823, "step": 9122 }, { "epoch": 0.38463646520648437, "grad_norm": 1.364142656326294, "learning_rate": 4.843374246541116e-05, "loss": 0.6304, "step": 9123 }, { "epoch": 0.38467862638868394, "grad_norm": 1.8872742652893066, "learning_rate": 4.843296167900309e-05, "loss": 1.0621, "step": 9124 }, { "epoch": 0.38472078757088346, "grad_norm": 0.5645992159843445, "learning_rate": 4.8432180892595024e-05, "loss": 0.1604, "step": 9125 }, { "epoch": 0.38476294875308303, "grad_norm": 0.5953753590583801, "learning_rate": 4.8431400106186956e-05, "loss": 0.3039, "step": 9126 }, { "epoch": 0.3848051099352826, "grad_norm": 1.4192560911178589, "learning_rate": 4.843061931977889e-05, "loss": 1.1884, "step": 9127 }, { "epoch": 0.3848472711174821, "grad_norm": 4.4041948318481445, "learning_rate": 4.842983853337081e-05, "loss": 0.4127, "step": 9128 }, { "epoch": 0.3848894322996817, "grad_norm": 0.7934144139289856, "learning_rate": 4.842905774696274e-05, "loss": 0.59, "step": 9129 }, { "epoch": 0.3849315934818812, "grad_norm": 1.0323649644851685, "learning_rate": 4.8428276960554674e-05, "loss": 0.1475, "step": 9130 }, { "epoch": 0.3849737546640808, "grad_norm": 0.5795334577560425, "learning_rate": 4.8427496174146605e-05, "loss": 0.1222, "step": 9131 }, { "epoch": 0.3850159158462803, "grad_norm": 1.0284228324890137, "learning_rate": 4.842671538773853e-05, "loss": 0.1317, "step": 9132 }, { "epoch": 0.3850580770284799, "grad_norm": 0.49738240242004395, "learning_rate": 4.842593460133046e-05, "loss": 0.2432, "step": 9133 }, { "epoch": 0.3851002382106794, "grad_norm": 0.7689082622528076, "learning_rate": 4.842515381492239e-05, "loss": 0.1064, "step": 9134 }, { "epoch": 0.385142399392879, "grad_norm": 0.5841915011405945, "learning_rate": 4.842437302851432e-05, "loss": 0.4333, "step": 9135 }, { "epoch": 0.38518456057507855, "grad_norm": 0.6217490434646606, "learning_rate": 4.8423592242106255e-05, "loss": 0.5948, "step": 9136 }, { "epoch": 0.38522672175727807, "grad_norm": 1.820229172706604, "learning_rate": 4.842281145569818e-05, "loss": 0.615, "step": 9137 }, { "epoch": 0.38526888293947764, "grad_norm": 1.5941615104675293, "learning_rate": 4.842203066929011e-05, "loss": 0.412, "step": 9138 }, { "epoch": 0.38531104412167716, "grad_norm": 0.6479049324989319, "learning_rate": 4.842124988288204e-05, "loss": 0.5706, "step": 9139 }, { "epoch": 0.38535320530387673, "grad_norm": 2.2348484992980957, "learning_rate": 4.8420469096473966e-05, "loss": 0.1833, "step": 9140 }, { "epoch": 0.38539536648607625, "grad_norm": 2.2124733924865723, "learning_rate": 4.8419688310065904e-05, "loss": 0.523, "step": 9141 }, { "epoch": 0.3854375276682758, "grad_norm": 1.2866028547286987, "learning_rate": 4.841890752365783e-05, "loss": 0.3173, "step": 9142 }, { "epoch": 0.38547968885047534, "grad_norm": 0.8770689964294434, "learning_rate": 4.841812673724976e-05, "loss": 0.1453, "step": 9143 }, { "epoch": 0.3855218500326749, "grad_norm": 1.6017390489578247, "learning_rate": 4.841734595084169e-05, "loss": 0.5872, "step": 9144 }, { "epoch": 0.3855640112148745, "grad_norm": 2.0729174613952637, "learning_rate": 4.841656516443362e-05, "loss": 0.5284, "step": 9145 }, { "epoch": 0.385606172397074, "grad_norm": 1.4669064283370972, "learning_rate": 4.841578437802555e-05, "loss": 0.4023, "step": 9146 }, { "epoch": 0.3856483335792736, "grad_norm": 2.0483784675598145, "learning_rate": 4.841500359161748e-05, "loss": 0.1398, "step": 9147 }, { "epoch": 0.3856904947614731, "grad_norm": 1.1429189443588257, "learning_rate": 4.841422280520941e-05, "loss": 0.6338, "step": 9148 }, { "epoch": 0.3857326559436727, "grad_norm": 1.88908851146698, "learning_rate": 4.841344201880134e-05, "loss": 0.5295, "step": 9149 }, { "epoch": 0.3857748171258722, "grad_norm": 0.6186453104019165, "learning_rate": 4.841266123239327e-05, "loss": 0.3846, "step": 9150 }, { "epoch": 0.38581697830807177, "grad_norm": 0.5179184675216675, "learning_rate": 4.8411880445985197e-05, "loss": 0.1163, "step": 9151 }, { "epoch": 0.3858591394902713, "grad_norm": 1.6465024948120117, "learning_rate": 4.841109965957713e-05, "loss": 0.6955, "step": 9152 }, { "epoch": 0.38590130067247086, "grad_norm": 1.17034912109375, "learning_rate": 4.841031887316906e-05, "loss": 0.2883, "step": 9153 }, { "epoch": 0.3859434618546704, "grad_norm": 0.28850945830345154, "learning_rate": 4.8409538086760984e-05, "loss": 0.0932, "step": 9154 }, { "epoch": 0.38598562303686995, "grad_norm": 0.780225932598114, "learning_rate": 4.840875730035292e-05, "loss": 0.6688, "step": 9155 }, { "epoch": 0.3860277842190695, "grad_norm": 1.4329289197921753, "learning_rate": 4.8407976513944846e-05, "loss": 0.2037, "step": 9156 }, { "epoch": 0.38606994540126904, "grad_norm": 0.46557000279426575, "learning_rate": 4.840719572753678e-05, "loss": 0.1858, "step": 9157 }, { "epoch": 0.3861121065834686, "grad_norm": 0.3557009994983673, "learning_rate": 4.840641494112871e-05, "loss": 0.1237, "step": 9158 }, { "epoch": 0.38615426776566814, "grad_norm": 1.8860923051834106, "learning_rate": 4.840563415472063e-05, "loss": 0.5905, "step": 9159 }, { "epoch": 0.3861964289478677, "grad_norm": 0.5600783228874207, "learning_rate": 4.840485336831257e-05, "loss": 0.1027, "step": 9160 }, { "epoch": 0.3862385901300672, "grad_norm": 0.6857343316078186, "learning_rate": 4.8404072581904496e-05, "loss": 0.5601, "step": 9161 }, { "epoch": 0.3862807513122668, "grad_norm": 0.5374459028244019, "learning_rate": 4.840329179549643e-05, "loss": 0.1729, "step": 9162 }, { "epoch": 0.3863229124944663, "grad_norm": 0.36771753430366516, "learning_rate": 4.840251100908836e-05, "loss": 0.1496, "step": 9163 }, { "epoch": 0.3863650736766659, "grad_norm": 58.38151931762695, "learning_rate": 4.840173022268028e-05, "loss": 0.2328, "step": 9164 }, { "epoch": 0.38640723485886547, "grad_norm": 0.3940397799015045, "learning_rate": 4.8400949436272214e-05, "loss": 0.1513, "step": 9165 }, { "epoch": 0.386449396041065, "grad_norm": 0.558964192867279, "learning_rate": 4.8400168649864145e-05, "loss": 0.1361, "step": 9166 }, { "epoch": 0.38649155722326456, "grad_norm": 1.8085721731185913, "learning_rate": 4.8399387863456076e-05, "loss": 1.1699, "step": 9167 }, { "epoch": 0.3865337184054641, "grad_norm": 1.0383808612823486, "learning_rate": 4.8398607077048e-05, "loss": 0.1096, "step": 9168 }, { "epoch": 0.38657587958766365, "grad_norm": 1.0375889539718628, "learning_rate": 4.839782629063994e-05, "loss": 1.1533, "step": 9169 }, { "epoch": 0.38661804076986317, "grad_norm": 0.6264773011207581, "learning_rate": 4.8397045504231863e-05, "loss": 0.7336, "step": 9170 }, { "epoch": 0.38666020195206274, "grad_norm": 0.5695597529411316, "learning_rate": 4.8396264717823795e-05, "loss": 0.5212, "step": 9171 }, { "epoch": 0.38670236313426226, "grad_norm": 0.49176332354545593, "learning_rate": 4.8395483931415726e-05, "loss": 0.0827, "step": 9172 }, { "epoch": 0.38674452431646184, "grad_norm": 1.193483591079712, "learning_rate": 4.839470314500765e-05, "loss": 0.3026, "step": 9173 }, { "epoch": 0.3867866854986614, "grad_norm": 1.1446655988693237, "learning_rate": 4.839392235859959e-05, "loss": 0.3655, "step": 9174 }, { "epoch": 0.3868288466808609, "grad_norm": 0.9520356059074402, "learning_rate": 4.839314157219151e-05, "loss": 0.4595, "step": 9175 }, { "epoch": 0.3868710078630605, "grad_norm": 0.4165317416191101, "learning_rate": 4.8392360785783444e-05, "loss": 0.1076, "step": 9176 }, { "epoch": 0.38691316904526, "grad_norm": 1.6190377473831177, "learning_rate": 4.8391579999375375e-05, "loss": 0.2909, "step": 9177 }, { "epoch": 0.3869553302274596, "grad_norm": 0.4424947202205658, "learning_rate": 4.83907992129673e-05, "loss": 0.1099, "step": 9178 }, { "epoch": 0.3869974914096591, "grad_norm": 13.296552658081055, "learning_rate": 4.839001842655923e-05, "loss": 0.1799, "step": 9179 }, { "epoch": 0.3870396525918587, "grad_norm": 0.7493782639503479, "learning_rate": 4.838923764015116e-05, "loss": 0.227, "step": 9180 }, { "epoch": 0.3870818137740582, "grad_norm": 2.2220821380615234, "learning_rate": 4.8388456853743094e-05, "loss": 0.332, "step": 9181 }, { "epoch": 0.3871239749562578, "grad_norm": 1.1148737668991089, "learning_rate": 4.8387676067335025e-05, "loss": 0.2466, "step": 9182 }, { "epoch": 0.3871661361384573, "grad_norm": 0.34297412633895874, "learning_rate": 4.838689528092695e-05, "loss": 0.1074, "step": 9183 }, { "epoch": 0.38720829732065687, "grad_norm": 3.951399326324463, "learning_rate": 4.838611449451888e-05, "loss": 0.6612, "step": 9184 }, { "epoch": 0.38725045850285644, "grad_norm": 0.6507620215415955, "learning_rate": 4.838533370811081e-05, "loss": 0.2226, "step": 9185 }, { "epoch": 0.38729261968505596, "grad_norm": 0.5991932153701782, "learning_rate": 4.838455292170274e-05, "loss": 0.2051, "step": 9186 }, { "epoch": 0.38733478086725553, "grad_norm": 1.1199233531951904, "learning_rate": 4.838377213529467e-05, "loss": 0.4171, "step": 9187 }, { "epoch": 0.38737694204945505, "grad_norm": 0.6268013715744019, "learning_rate": 4.8382991348886606e-05, "loss": 0.5698, "step": 9188 }, { "epoch": 0.3874191032316546, "grad_norm": 1.4801427125930786, "learning_rate": 4.838221056247853e-05, "loss": 0.1803, "step": 9189 }, { "epoch": 0.38746126441385415, "grad_norm": 0.8084945678710938, "learning_rate": 4.838142977607046e-05, "loss": 0.1328, "step": 9190 }, { "epoch": 0.3875034255960537, "grad_norm": 0.8424553871154785, "learning_rate": 4.838064898966239e-05, "loss": 0.3261, "step": 9191 }, { "epoch": 0.38754558677825324, "grad_norm": 1.0776126384735107, "learning_rate": 4.837986820325432e-05, "loss": 0.2534, "step": 9192 }, { "epoch": 0.3875877479604528, "grad_norm": 0.9962378144264221, "learning_rate": 4.837908741684625e-05, "loss": 0.2022, "step": 9193 }, { "epoch": 0.3876299091426524, "grad_norm": 0.4838554263114929, "learning_rate": 4.837830663043818e-05, "loss": 0.111, "step": 9194 }, { "epoch": 0.3876720703248519, "grad_norm": 0.5528291463851929, "learning_rate": 4.837752584403011e-05, "loss": 0.0782, "step": 9195 }, { "epoch": 0.3877142315070515, "grad_norm": 1.5313279628753662, "learning_rate": 4.837674505762204e-05, "loss": 0.2701, "step": 9196 }, { "epoch": 0.387756392689251, "grad_norm": 1.3869404792785645, "learning_rate": 4.837596427121397e-05, "loss": 0.6994, "step": 9197 }, { "epoch": 0.38779855387145057, "grad_norm": 1.0726978778839111, "learning_rate": 4.83751834848059e-05, "loss": 1.0251, "step": 9198 }, { "epoch": 0.3878407150536501, "grad_norm": 0.9276685118675232, "learning_rate": 4.837440269839783e-05, "loss": 0.2578, "step": 9199 }, { "epoch": 0.38788287623584966, "grad_norm": 0.672326385974884, "learning_rate": 4.837362191198976e-05, "loss": 0.5387, "step": 9200 }, { "epoch": 0.3879250374180492, "grad_norm": 0.3603985905647278, "learning_rate": 4.8372841125581685e-05, "loss": 0.1328, "step": 9201 }, { "epoch": 0.38796719860024875, "grad_norm": 1.4769401550292969, "learning_rate": 4.8372060339173616e-05, "loss": 0.3635, "step": 9202 }, { "epoch": 0.3880093597824483, "grad_norm": 0.7073140144348145, "learning_rate": 4.837127955276555e-05, "loss": 0.3597, "step": 9203 }, { "epoch": 0.38805152096464784, "grad_norm": 0.5073933601379395, "learning_rate": 4.837049876635747e-05, "loss": 0.1167, "step": 9204 }, { "epoch": 0.3880936821468474, "grad_norm": 1.6407804489135742, "learning_rate": 4.836971797994941e-05, "loss": 0.4803, "step": 9205 }, { "epoch": 0.38813584332904694, "grad_norm": 1.2151453495025635, "learning_rate": 4.8368937193541335e-05, "loss": 0.6477, "step": 9206 }, { "epoch": 0.3881780045112465, "grad_norm": 1.050399661064148, "learning_rate": 4.836815640713327e-05, "loss": 1.0262, "step": 9207 }, { "epoch": 0.38822016569344603, "grad_norm": 0.8344534635543823, "learning_rate": 4.83673756207252e-05, "loss": 0.1769, "step": 9208 }, { "epoch": 0.3882623268756456, "grad_norm": 0.48134392499923706, "learning_rate": 4.836659483431712e-05, "loss": 0.1402, "step": 9209 }, { "epoch": 0.3883044880578451, "grad_norm": 0.8090944886207581, "learning_rate": 4.836581404790906e-05, "loss": 0.1429, "step": 9210 }, { "epoch": 0.3883466492400447, "grad_norm": 1.9412033557891846, "learning_rate": 4.8365033261500984e-05, "loss": 0.3303, "step": 9211 }, { "epoch": 0.3883888104222442, "grad_norm": 3.582683563232422, "learning_rate": 4.8364252475092915e-05, "loss": 0.3672, "step": 9212 }, { "epoch": 0.3884309716044438, "grad_norm": 1.109204649925232, "learning_rate": 4.836347168868485e-05, "loss": 0.2125, "step": 9213 }, { "epoch": 0.38847313278664336, "grad_norm": 0.7765980362892151, "learning_rate": 4.836269090227678e-05, "loss": 0.2341, "step": 9214 }, { "epoch": 0.3885152939688429, "grad_norm": 1.0524169206619263, "learning_rate": 4.83619101158687e-05, "loss": 0.9549, "step": 9215 }, { "epoch": 0.38855745515104245, "grad_norm": 0.3284972012042999, "learning_rate": 4.8361129329460634e-05, "loss": 0.0902, "step": 9216 }, { "epoch": 0.38859961633324197, "grad_norm": 0.3591338098049164, "learning_rate": 4.8360348543052565e-05, "loss": 0.1727, "step": 9217 }, { "epoch": 0.38864177751544154, "grad_norm": 0.7618734836578369, "learning_rate": 4.8359567756644496e-05, "loss": 0.1947, "step": 9218 }, { "epoch": 0.38868393869764106, "grad_norm": 0.5081849694252014, "learning_rate": 4.835878697023643e-05, "loss": 0.1445, "step": 9219 }, { "epoch": 0.38872609987984064, "grad_norm": 0.690281867980957, "learning_rate": 4.835800618382835e-05, "loss": 0.2808, "step": 9220 }, { "epoch": 0.38876826106204015, "grad_norm": 0.7139136791229248, "learning_rate": 4.835722539742028e-05, "loss": 0.3354, "step": 9221 }, { "epoch": 0.38881042224423973, "grad_norm": 0.6380494832992554, "learning_rate": 4.8356444611012214e-05, "loss": 0.2478, "step": 9222 }, { "epoch": 0.3888525834264393, "grad_norm": 0.550300121307373, "learning_rate": 4.835566382460414e-05, "loss": 0.1118, "step": 9223 }, { "epoch": 0.3888947446086388, "grad_norm": 0.683299720287323, "learning_rate": 4.835488303819608e-05, "loss": 0.2422, "step": 9224 }, { "epoch": 0.3889369057908384, "grad_norm": 0.8429284691810608, "learning_rate": 4.8354102251788e-05, "loss": 0.5713, "step": 9225 }, { "epoch": 0.3889790669730379, "grad_norm": 1.1938464641571045, "learning_rate": 4.835332146537993e-05, "loss": 0.7923, "step": 9226 }, { "epoch": 0.3890212281552375, "grad_norm": 1.173506498336792, "learning_rate": 4.8352540678971864e-05, "loss": 0.6648, "step": 9227 }, { "epoch": 0.389063389337437, "grad_norm": 0.42036110162734985, "learning_rate": 4.835175989256379e-05, "loss": 0.171, "step": 9228 }, { "epoch": 0.3891055505196366, "grad_norm": 0.3118346035480499, "learning_rate": 4.8350979106155726e-05, "loss": 0.1459, "step": 9229 }, { "epoch": 0.3891477117018361, "grad_norm": 0.7412299513816833, "learning_rate": 4.835019831974765e-05, "loss": 0.1152, "step": 9230 }, { "epoch": 0.38918987288403567, "grad_norm": 0.401467502117157, "learning_rate": 4.834941753333958e-05, "loss": 0.1581, "step": 9231 }, { "epoch": 0.38923203406623524, "grad_norm": 1.2515565156936646, "learning_rate": 4.8348636746931513e-05, "loss": 0.278, "step": 9232 }, { "epoch": 0.38927419524843476, "grad_norm": 2.4081685543060303, "learning_rate": 4.8347855960523445e-05, "loss": 0.274, "step": 9233 }, { "epoch": 0.38931635643063434, "grad_norm": 0.8750820159912109, "learning_rate": 4.834707517411537e-05, "loss": 0.3071, "step": 9234 }, { "epoch": 0.38935851761283385, "grad_norm": 0.9688261151313782, "learning_rate": 4.83462943877073e-05, "loss": 0.2729, "step": 9235 }, { "epoch": 0.38940067879503343, "grad_norm": 0.42965278029441833, "learning_rate": 4.834551360129923e-05, "loss": 0.1054, "step": 9236 }, { "epoch": 0.38944283997723295, "grad_norm": 1.590417742729187, "learning_rate": 4.8344732814891156e-05, "loss": 0.4739, "step": 9237 }, { "epoch": 0.3894850011594325, "grad_norm": 0.9469932317733765, "learning_rate": 4.8343952028483094e-05, "loss": 0.3868, "step": 9238 }, { "epoch": 0.38952716234163204, "grad_norm": 0.8050105571746826, "learning_rate": 4.834317124207502e-05, "loss": 0.3533, "step": 9239 }, { "epoch": 0.3895693235238316, "grad_norm": 0.7032774090766907, "learning_rate": 4.834239045566695e-05, "loss": 0.339, "step": 9240 }, { "epoch": 0.38961148470603113, "grad_norm": 0.470572829246521, "learning_rate": 4.834160966925888e-05, "loss": 0.1723, "step": 9241 }, { "epoch": 0.3896536458882307, "grad_norm": 1.3069746494293213, "learning_rate": 4.8340828882850806e-05, "loss": 0.246, "step": 9242 }, { "epoch": 0.3896958070704303, "grad_norm": 0.39393386244773865, "learning_rate": 4.8340048096442744e-05, "loss": 0.0925, "step": 9243 }, { "epoch": 0.3897379682526298, "grad_norm": 0.444787859916687, "learning_rate": 4.833926731003467e-05, "loss": 0.1489, "step": 9244 }, { "epoch": 0.38978012943482937, "grad_norm": 1.645636796951294, "learning_rate": 4.83384865236266e-05, "loss": 0.1458, "step": 9245 }, { "epoch": 0.3898222906170289, "grad_norm": 0.41561928391456604, "learning_rate": 4.833770573721853e-05, "loss": 0.137, "step": 9246 }, { "epoch": 0.38986445179922846, "grad_norm": 0.7634973526000977, "learning_rate": 4.8336924950810455e-05, "loss": 0.4956, "step": 9247 }, { "epoch": 0.389906612981428, "grad_norm": 0.6454463601112366, "learning_rate": 4.8336144164402387e-05, "loss": 0.1854, "step": 9248 }, { "epoch": 0.38994877416362755, "grad_norm": 0.8592225909233093, "learning_rate": 4.833536337799432e-05, "loss": 0.4143, "step": 9249 }, { "epoch": 0.38999093534582707, "grad_norm": 643.2132568359375, "learning_rate": 4.833458259158625e-05, "loss": 0.7777, "step": 9250 }, { "epoch": 0.39003309652802665, "grad_norm": 1.7139086723327637, "learning_rate": 4.833380180517818e-05, "loss": 0.1854, "step": 9251 }, { "epoch": 0.3900752577102262, "grad_norm": 1.6230671405792236, "learning_rate": 4.833302101877011e-05, "loss": 0.4814, "step": 9252 }, { "epoch": 0.39011741889242574, "grad_norm": 1.481667160987854, "learning_rate": 4.8332240232362036e-05, "loss": 0.1958, "step": 9253 }, { "epoch": 0.3901595800746253, "grad_norm": 0.8192741870880127, "learning_rate": 4.833145944595397e-05, "loss": 0.3198, "step": 9254 }, { "epoch": 0.39020174125682483, "grad_norm": 1.3222196102142334, "learning_rate": 4.83306786595459e-05, "loss": 0.539, "step": 9255 }, { "epoch": 0.3902439024390244, "grad_norm": 3.304877281188965, "learning_rate": 4.832989787313782e-05, "loss": 0.2424, "step": 9256 }, { "epoch": 0.3902860636212239, "grad_norm": 0.5343407988548279, "learning_rate": 4.832911708672976e-05, "loss": 0.509, "step": 9257 }, { "epoch": 0.3903282248034235, "grad_norm": 1.0215373039245605, "learning_rate": 4.8328336300321686e-05, "loss": 0.9486, "step": 9258 }, { "epoch": 0.390370385985623, "grad_norm": 0.7311291694641113, "learning_rate": 4.832755551391362e-05, "loss": 0.093, "step": 9259 }, { "epoch": 0.3904125471678226, "grad_norm": 1.2458887100219727, "learning_rate": 4.832677472750555e-05, "loss": 0.5488, "step": 9260 }, { "epoch": 0.39045470835002216, "grad_norm": 0.9685040712356567, "learning_rate": 4.832599394109747e-05, "loss": 0.2982, "step": 9261 }, { "epoch": 0.3904968695322217, "grad_norm": 0.44944337010383606, "learning_rate": 4.832521315468941e-05, "loss": 0.0814, "step": 9262 }, { "epoch": 0.39053903071442125, "grad_norm": 0.42092418670654297, "learning_rate": 4.8324432368281335e-05, "loss": 0.1077, "step": 9263 }, { "epoch": 0.39058119189662077, "grad_norm": 1.3137638568878174, "learning_rate": 4.8323651581873266e-05, "loss": 0.2959, "step": 9264 }, { "epoch": 0.39062335307882035, "grad_norm": 0.802706241607666, "learning_rate": 4.83228707954652e-05, "loss": 0.1431, "step": 9265 }, { "epoch": 0.39066551426101986, "grad_norm": 0.9078596234321594, "learning_rate": 4.832209000905712e-05, "loss": 0.9876, "step": 9266 }, { "epoch": 0.39070767544321944, "grad_norm": 0.7874327898025513, "learning_rate": 4.832130922264905e-05, "loss": 0.437, "step": 9267 }, { "epoch": 0.39074983662541896, "grad_norm": 0.4468691647052765, "learning_rate": 4.8320528436240985e-05, "loss": 0.454, "step": 9268 }, { "epoch": 0.39079199780761853, "grad_norm": 2.201119899749756, "learning_rate": 4.8319747649832916e-05, "loss": 0.8163, "step": 9269 }, { "epoch": 0.39083415898981805, "grad_norm": 2.521888494491577, "learning_rate": 4.831896686342484e-05, "loss": 0.7294, "step": 9270 }, { "epoch": 0.3908763201720176, "grad_norm": 0.5447379946708679, "learning_rate": 4.831818607701677e-05, "loss": 0.138, "step": 9271 }, { "epoch": 0.3909184813542172, "grad_norm": 0.7801186442375183, "learning_rate": 4.83174052906087e-05, "loss": 0.1732, "step": 9272 }, { "epoch": 0.3909606425364167, "grad_norm": 0.7655255198478699, "learning_rate": 4.831662450420063e-05, "loss": 0.441, "step": 9273 }, { "epoch": 0.3910028037186163, "grad_norm": 0.8716804385185242, "learning_rate": 4.8315843717792565e-05, "loss": 0.1884, "step": 9274 }, { "epoch": 0.3910449649008158, "grad_norm": 0.9341549277305603, "learning_rate": 4.831506293138449e-05, "loss": 0.4615, "step": 9275 }, { "epoch": 0.3910871260830154, "grad_norm": 0.7125199437141418, "learning_rate": 4.831428214497643e-05, "loss": 0.7028, "step": 9276 }, { "epoch": 0.3911292872652149, "grad_norm": 1.826157569885254, "learning_rate": 4.831350135856835e-05, "loss": 0.8244, "step": 9277 }, { "epoch": 0.39117144844741447, "grad_norm": 0.4650065004825592, "learning_rate": 4.8312720572160284e-05, "loss": 0.1076, "step": 9278 }, { "epoch": 0.391213609629614, "grad_norm": 1.055044174194336, "learning_rate": 4.8311939785752215e-05, "loss": 0.2828, "step": 9279 }, { "epoch": 0.39125577081181356, "grad_norm": 1.5103843212127686, "learning_rate": 4.831115899934414e-05, "loss": 0.4764, "step": 9280 }, { "epoch": 0.39129793199401314, "grad_norm": 0.6186622977256775, "learning_rate": 4.831037821293607e-05, "loss": 0.2168, "step": 9281 }, { "epoch": 0.39134009317621266, "grad_norm": 0.3163295388221741, "learning_rate": 4.8309597426528e-05, "loss": 0.0873, "step": 9282 }, { "epoch": 0.39138225435841223, "grad_norm": 0.5981609225273132, "learning_rate": 4.830881664011993e-05, "loss": 0.2487, "step": 9283 }, { "epoch": 0.39142441554061175, "grad_norm": 0.6638497710227966, "learning_rate": 4.830803585371186e-05, "loss": 0.6022, "step": 9284 }, { "epoch": 0.3914665767228113, "grad_norm": 0.6521099805831909, "learning_rate": 4.830725506730379e-05, "loss": 0.4921, "step": 9285 }, { "epoch": 0.39150873790501084, "grad_norm": 1.5701115131378174, "learning_rate": 4.830647428089572e-05, "loss": 0.1319, "step": 9286 }, { "epoch": 0.3915508990872104, "grad_norm": 1.0553758144378662, "learning_rate": 4.830569349448765e-05, "loss": 0.2842, "step": 9287 }, { "epoch": 0.39159306026940993, "grad_norm": 1.1190123558044434, "learning_rate": 4.830491270807958e-05, "loss": 0.1735, "step": 9288 }, { "epoch": 0.3916352214516095, "grad_norm": 0.8404326438903809, "learning_rate": 4.830413192167151e-05, "loss": 0.1779, "step": 9289 }, { "epoch": 0.3916773826338091, "grad_norm": 3.274677038192749, "learning_rate": 4.830335113526344e-05, "loss": 0.5292, "step": 9290 }, { "epoch": 0.3917195438160086, "grad_norm": 1.0588420629501343, "learning_rate": 4.830257034885537e-05, "loss": 0.6233, "step": 9291 }, { "epoch": 0.39176170499820817, "grad_norm": 1.2550849914550781, "learning_rate": 4.8301789562447294e-05, "loss": 1.0345, "step": 9292 }, { "epoch": 0.3918038661804077, "grad_norm": 4.167516708374023, "learning_rate": 4.830100877603923e-05, "loss": 0.9828, "step": 9293 }, { "epoch": 0.39184602736260726, "grad_norm": 31.890207290649414, "learning_rate": 4.830022798963116e-05, "loss": 0.9087, "step": 9294 }, { "epoch": 0.3918881885448068, "grad_norm": 336.881103515625, "learning_rate": 4.829944720322309e-05, "loss": 1.8935, "step": 9295 }, { "epoch": 0.39193034972700636, "grad_norm": 8.398245811462402, "learning_rate": 4.829866641681502e-05, "loss": 0.5082, "step": 9296 }, { "epoch": 0.3919725109092059, "grad_norm": 137.75608825683594, "learning_rate": 4.829788563040695e-05, "loss": 2.1617, "step": 9297 }, { "epoch": 0.39201467209140545, "grad_norm": 520.9969482421875, "learning_rate": 4.829710484399888e-05, "loss": 4.0259, "step": 9298 }, { "epoch": 0.39205683327360497, "grad_norm": 49.06074523925781, "learning_rate": 4.8296324057590806e-05, "loss": 0.9065, "step": 9299 }, { "epoch": 0.39209899445580454, "grad_norm": 169.3170928955078, "learning_rate": 4.829554327118274e-05, "loss": 1.7778, "step": 9300 }, { "epoch": 0.3921411556380041, "grad_norm": 74.79087829589844, "learning_rate": 4.829476248477467e-05, "loss": 0.5427, "step": 9301 }, { "epoch": 0.39218331682020363, "grad_norm": 0.7243114709854126, "learning_rate": 4.82939816983666e-05, "loss": 0.1181, "step": 9302 }, { "epoch": 0.3922254780024032, "grad_norm": 0.7931376099586487, "learning_rate": 4.8293200911958524e-05, "loss": 0.111, "step": 9303 }, { "epoch": 0.3922676391846027, "grad_norm": 0.538605272769928, "learning_rate": 4.8292420125550456e-05, "loss": 0.2113, "step": 9304 }, { "epoch": 0.3923098003668023, "grad_norm": 1.6947681903839111, "learning_rate": 4.829163933914239e-05, "loss": 0.5242, "step": 9305 }, { "epoch": 0.3923519615490018, "grad_norm": 0.8261560797691345, "learning_rate": 4.829085855273431e-05, "loss": 0.2083, "step": 9306 }, { "epoch": 0.3923941227312014, "grad_norm": 1.1553966999053955, "learning_rate": 4.829007776632625e-05, "loss": 0.1782, "step": 9307 }, { "epoch": 0.3924362839134009, "grad_norm": 0.9569820761680603, "learning_rate": 4.8289296979918174e-05, "loss": 0.3329, "step": 9308 }, { "epoch": 0.3924784450956005, "grad_norm": 1.5119330883026123, "learning_rate": 4.8288516193510105e-05, "loss": 1.0392, "step": 9309 }, { "epoch": 0.39252060627780005, "grad_norm": 0.8896615505218506, "learning_rate": 4.8287735407102037e-05, "loss": 0.0936, "step": 9310 }, { "epoch": 0.3925627674599996, "grad_norm": 3.180232286453247, "learning_rate": 4.828695462069396e-05, "loss": 1.1251, "step": 9311 }, { "epoch": 0.39260492864219915, "grad_norm": 26.39525032043457, "learning_rate": 4.82861738342859e-05, "loss": 0.8442, "step": 9312 }, { "epoch": 0.39264708982439867, "grad_norm": 0.45315101742744446, "learning_rate": 4.8285393047877824e-05, "loss": 0.0813, "step": 9313 }, { "epoch": 0.39268925100659824, "grad_norm": 0.5625874400138855, "learning_rate": 4.8284612261469755e-05, "loss": 0.1573, "step": 9314 }, { "epoch": 0.39273141218879776, "grad_norm": 1.4421963691711426, "learning_rate": 4.8283831475061686e-05, "loss": 0.9643, "step": 9315 }, { "epoch": 0.39277357337099733, "grad_norm": 0.8489149212837219, "learning_rate": 4.828305068865361e-05, "loss": 0.2051, "step": 9316 }, { "epoch": 0.39281573455319685, "grad_norm": 0.8702167868614197, "learning_rate": 4.828226990224554e-05, "loss": 0.6024, "step": 9317 }, { "epoch": 0.3928578957353964, "grad_norm": 0.5232473611831665, "learning_rate": 4.828148911583747e-05, "loss": 0.1084, "step": 9318 }, { "epoch": 0.392900056917596, "grad_norm": 0.504119873046875, "learning_rate": 4.8280708329429404e-05, "loss": 0.1704, "step": 9319 }, { "epoch": 0.3929422180997955, "grad_norm": 0.8971471190452576, "learning_rate": 4.8279927543021336e-05, "loss": 0.9745, "step": 9320 }, { "epoch": 0.3929843792819951, "grad_norm": 0.7617174386978149, "learning_rate": 4.827914675661327e-05, "loss": 0.2898, "step": 9321 }, { "epoch": 0.3930265404641946, "grad_norm": 0.7924504280090332, "learning_rate": 4.827836597020519e-05, "loss": 0.5197, "step": 9322 }, { "epoch": 0.3930687016463942, "grad_norm": 0.5967532992362976, "learning_rate": 4.827758518379712e-05, "loss": 0.5082, "step": 9323 }, { "epoch": 0.3931108628285937, "grad_norm": 1.5741238594055176, "learning_rate": 4.8276804397389054e-05, "loss": 0.4261, "step": 9324 }, { "epoch": 0.3931530240107933, "grad_norm": 1.8617119789123535, "learning_rate": 4.827602361098098e-05, "loss": 0.9616, "step": 9325 }, { "epoch": 0.3931951851929928, "grad_norm": 0.7325236201286316, "learning_rate": 4.8275242824572916e-05, "loss": 0.306, "step": 9326 }, { "epoch": 0.39323734637519236, "grad_norm": 3.2720489501953125, "learning_rate": 4.827446203816484e-05, "loss": 0.5606, "step": 9327 }, { "epoch": 0.3932795075573919, "grad_norm": 0.952178955078125, "learning_rate": 4.827368125175677e-05, "loss": 0.1249, "step": 9328 }, { "epoch": 0.39332166873959146, "grad_norm": 0.718459963798523, "learning_rate": 4.82729004653487e-05, "loss": 0.221, "step": 9329 }, { "epoch": 0.39336382992179103, "grad_norm": 0.7135993838310242, "learning_rate": 4.827211967894063e-05, "loss": 0.6115, "step": 9330 }, { "epoch": 0.39340599110399055, "grad_norm": 0.6099116206169128, "learning_rate": 4.8271338892532566e-05, "loss": 0.1878, "step": 9331 }, { "epoch": 0.3934481522861901, "grad_norm": 1.7975276708602905, "learning_rate": 4.827055810612449e-05, "loss": 0.7405, "step": 9332 }, { "epoch": 0.39349031346838964, "grad_norm": 0.7055613994598389, "learning_rate": 4.826977731971642e-05, "loss": 0.448, "step": 9333 }, { "epoch": 0.3935324746505892, "grad_norm": 2.318915843963623, "learning_rate": 4.826899653330835e-05, "loss": 0.313, "step": 9334 }, { "epoch": 0.39357463583278873, "grad_norm": 4.409668922424316, "learning_rate": 4.826821574690028e-05, "loss": 0.3765, "step": 9335 }, { "epoch": 0.3936167970149883, "grad_norm": 0.666257381439209, "learning_rate": 4.826743496049221e-05, "loss": 0.2083, "step": 9336 }, { "epoch": 0.3936589581971878, "grad_norm": 1.6421279907226562, "learning_rate": 4.826665417408414e-05, "loss": 0.6806, "step": 9337 }, { "epoch": 0.3937011193793874, "grad_norm": 0.37302201986312866, "learning_rate": 4.826587338767607e-05, "loss": 0.1238, "step": 9338 }, { "epoch": 0.393743280561587, "grad_norm": 0.6169028282165527, "learning_rate": 4.8265092601267996e-05, "loss": 0.1175, "step": 9339 }, { "epoch": 0.3937854417437865, "grad_norm": 0.5568733811378479, "learning_rate": 4.8264311814859934e-05, "loss": 0.1255, "step": 9340 }, { "epoch": 0.39382760292598606, "grad_norm": 0.6435645222663879, "learning_rate": 4.826353102845186e-05, "loss": 0.5712, "step": 9341 }, { "epoch": 0.3938697641081856, "grad_norm": 0.658327579498291, "learning_rate": 4.826275024204379e-05, "loss": 0.1317, "step": 9342 }, { "epoch": 0.39391192529038516, "grad_norm": 4.36601448059082, "learning_rate": 4.826196945563572e-05, "loss": 0.2339, "step": 9343 }, { "epoch": 0.3939540864725847, "grad_norm": 0.691046953201294, "learning_rate": 4.8261188669227645e-05, "loss": 0.3878, "step": 9344 }, { "epoch": 0.39399624765478425, "grad_norm": 0.9854079484939575, "learning_rate": 4.826040788281958e-05, "loss": 0.501, "step": 9345 }, { "epoch": 0.39403840883698377, "grad_norm": 1.1173902750015259, "learning_rate": 4.825962709641151e-05, "loss": 0.9338, "step": 9346 }, { "epoch": 0.39408057001918334, "grad_norm": 0.9083266854286194, "learning_rate": 4.825884631000344e-05, "loss": 0.5321, "step": 9347 }, { "epoch": 0.3941227312013829, "grad_norm": 0.6240510940551758, "learning_rate": 4.825806552359537e-05, "loss": 0.0841, "step": 9348 }, { "epoch": 0.39416489238358243, "grad_norm": 0.6958338618278503, "learning_rate": 4.8257284737187295e-05, "loss": 0.3627, "step": 9349 }, { "epoch": 0.394207053565782, "grad_norm": 0.8661128282546997, "learning_rate": 4.8256503950779226e-05, "loss": 0.3395, "step": 9350 }, { "epoch": 0.3942492147479815, "grad_norm": 110.83943176269531, "learning_rate": 4.825572316437116e-05, "loss": 0.2305, "step": 9351 }, { "epoch": 0.3942913759301811, "grad_norm": 0.39414480328559875, "learning_rate": 4.825494237796309e-05, "loss": 0.1061, "step": 9352 }, { "epoch": 0.3943335371123806, "grad_norm": 0.7501365542411804, "learning_rate": 4.825416159155501e-05, "loss": 0.1083, "step": 9353 }, { "epoch": 0.3943756982945802, "grad_norm": 1.476394772529602, "learning_rate": 4.8253380805146944e-05, "loss": 1.2755, "step": 9354 }, { "epoch": 0.3944178594767797, "grad_norm": 0.29202204942703247, "learning_rate": 4.8252600018738875e-05, "loss": 0.1296, "step": 9355 }, { "epoch": 0.3944600206589793, "grad_norm": 0.7677525281906128, "learning_rate": 4.825181923233081e-05, "loss": 0.2183, "step": 9356 }, { "epoch": 0.3945021818411788, "grad_norm": 0.5728973150253296, "learning_rate": 4.825103844592274e-05, "loss": 0.128, "step": 9357 }, { "epoch": 0.3945443430233784, "grad_norm": 3.5535476207733154, "learning_rate": 4.825025765951466e-05, "loss": 0.5492, "step": 9358 }, { "epoch": 0.39458650420557795, "grad_norm": 0.6958374381065369, "learning_rate": 4.82494768731066e-05, "loss": 0.0798, "step": 9359 }, { "epoch": 0.39462866538777747, "grad_norm": 0.8929688334465027, "learning_rate": 4.8248696086698525e-05, "loss": 0.2781, "step": 9360 }, { "epoch": 0.39467082656997704, "grad_norm": 0.513333261013031, "learning_rate": 4.824791530029045e-05, "loss": 0.1817, "step": 9361 }, { "epoch": 0.39471298775217656, "grad_norm": 0.8354249000549316, "learning_rate": 4.824713451388239e-05, "loss": 0.4173, "step": 9362 }, { "epoch": 0.39475514893437613, "grad_norm": 0.5538308024406433, "learning_rate": 4.824635372747431e-05, "loss": 0.1448, "step": 9363 }, { "epoch": 0.39479731011657565, "grad_norm": 0.51270991563797, "learning_rate": 4.824557294106624e-05, "loss": 0.2115, "step": 9364 }, { "epoch": 0.3948394712987752, "grad_norm": 1.1154741048812866, "learning_rate": 4.8244792154658174e-05, "loss": 0.5245, "step": 9365 }, { "epoch": 0.39488163248097474, "grad_norm": 0.3867226243019104, "learning_rate": 4.8244011368250106e-05, "loss": 0.1179, "step": 9366 }, { "epoch": 0.3949237936631743, "grad_norm": 0.3016873002052307, "learning_rate": 4.824323058184204e-05, "loss": 0.1315, "step": 9367 }, { "epoch": 0.3949659548453739, "grad_norm": 2.6176295280456543, "learning_rate": 4.824244979543396e-05, "loss": 0.1349, "step": 9368 }, { "epoch": 0.3950081160275734, "grad_norm": 0.5250353217124939, "learning_rate": 4.824166900902589e-05, "loss": 0.2337, "step": 9369 }, { "epoch": 0.395050277209773, "grad_norm": 12.418532371520996, "learning_rate": 4.8240888222617824e-05, "loss": 0.3297, "step": 9370 }, { "epoch": 0.3950924383919725, "grad_norm": 0.7261799573898315, "learning_rate": 4.8240107436209755e-05, "loss": 0.4901, "step": 9371 }, { "epoch": 0.3951345995741721, "grad_norm": 0.8290916085243225, "learning_rate": 4.823932664980168e-05, "loss": 0.2528, "step": 9372 }, { "epoch": 0.3951767607563716, "grad_norm": 1.6710363626480103, "learning_rate": 4.823854586339361e-05, "loss": 0.7177, "step": 9373 }, { "epoch": 0.39521892193857117, "grad_norm": 0.3896801471710205, "learning_rate": 4.823776507698554e-05, "loss": 0.1658, "step": 9374 }, { "epoch": 0.3952610831207707, "grad_norm": 0.5457261800765991, "learning_rate": 4.823698429057747e-05, "loss": 0.0859, "step": 9375 }, { "epoch": 0.39530324430297026, "grad_norm": 0.9807615876197815, "learning_rate": 4.8236203504169405e-05, "loss": 0.1134, "step": 9376 }, { "epoch": 0.39534540548516983, "grad_norm": 0.7818842530250549, "learning_rate": 4.823542271776133e-05, "loss": 1.1059, "step": 9377 }, { "epoch": 0.39538756666736935, "grad_norm": 0.3593439757823944, "learning_rate": 4.823464193135326e-05, "loss": 0.0986, "step": 9378 }, { "epoch": 0.3954297278495689, "grad_norm": 2.057858943939209, "learning_rate": 4.823386114494519e-05, "loss": 1.0651, "step": 9379 }, { "epoch": 0.39547188903176844, "grad_norm": 0.6578262448310852, "learning_rate": 4.8233080358537116e-05, "loss": 0.4736, "step": 9380 }, { "epoch": 0.395514050213968, "grad_norm": 0.699281632900238, "learning_rate": 4.8232299572129054e-05, "loss": 0.0891, "step": 9381 }, { "epoch": 0.39555621139616753, "grad_norm": 0.6790111064910889, "learning_rate": 4.823151878572098e-05, "loss": 0.154, "step": 9382 }, { "epoch": 0.3955983725783671, "grad_norm": 0.3031339943408966, "learning_rate": 4.823073799931291e-05, "loss": 0.0689, "step": 9383 }, { "epoch": 0.3956405337605666, "grad_norm": 0.49663591384887695, "learning_rate": 4.822995721290484e-05, "loss": 0.3842, "step": 9384 }, { "epoch": 0.3956826949427662, "grad_norm": 0.7060630321502686, "learning_rate": 4.822917642649677e-05, "loss": 0.1636, "step": 9385 }, { "epoch": 0.3957248561249657, "grad_norm": 0.8404473066329956, "learning_rate": 4.82283956400887e-05, "loss": 0.2008, "step": 9386 }, { "epoch": 0.3957670173071653, "grad_norm": 0.45513811707496643, "learning_rate": 4.822761485368063e-05, "loss": 0.0659, "step": 9387 }, { "epoch": 0.39580917848936487, "grad_norm": 0.5469754338264465, "learning_rate": 4.822683406727256e-05, "loss": 0.1639, "step": 9388 }, { "epoch": 0.3958513396715644, "grad_norm": 0.5443307757377625, "learning_rate": 4.822605328086449e-05, "loss": 0.1095, "step": 9389 }, { "epoch": 0.39589350085376396, "grad_norm": 0.453572541475296, "learning_rate": 4.822527249445642e-05, "loss": 0.1564, "step": 9390 }, { "epoch": 0.3959356620359635, "grad_norm": 4.989023208618164, "learning_rate": 4.8224491708048347e-05, "loss": 0.9359, "step": 9391 }, { "epoch": 0.39597782321816305, "grad_norm": 2.4832348823547363, "learning_rate": 4.822371092164028e-05, "loss": 1.2109, "step": 9392 }, { "epoch": 0.39601998440036257, "grad_norm": 0.7968556880950928, "learning_rate": 4.822293013523221e-05, "loss": 0.2692, "step": 9393 }, { "epoch": 0.39606214558256214, "grad_norm": 0.44472211599349976, "learning_rate": 4.8222149348824134e-05, "loss": 0.292, "step": 9394 }, { "epoch": 0.39610430676476166, "grad_norm": 3.721928119659424, "learning_rate": 4.822136856241607e-05, "loss": 0.7649, "step": 9395 }, { "epoch": 0.39614646794696123, "grad_norm": 1.0640884637832642, "learning_rate": 4.8220587776007996e-05, "loss": 0.7414, "step": 9396 }, { "epoch": 0.3961886291291608, "grad_norm": 0.6843023300170898, "learning_rate": 4.821980698959993e-05, "loss": 0.1795, "step": 9397 }, { "epoch": 0.3962307903113603, "grad_norm": 0.46038174629211426, "learning_rate": 4.821902620319186e-05, "loss": 0.2214, "step": 9398 }, { "epoch": 0.3962729514935599, "grad_norm": 0.3656585216522217, "learning_rate": 4.821824541678378e-05, "loss": 0.0817, "step": 9399 }, { "epoch": 0.3963151126757594, "grad_norm": 0.3775768578052521, "learning_rate": 4.821746463037572e-05, "loss": 0.1158, "step": 9400 }, { "epoch": 0.396357273857959, "grad_norm": 1.266330361366272, "learning_rate": 4.8216683843967646e-05, "loss": 0.1886, "step": 9401 }, { "epoch": 0.3963994350401585, "grad_norm": 1.7437106370925903, "learning_rate": 4.821590305755958e-05, "loss": 0.9741, "step": 9402 }, { "epoch": 0.3964415962223581, "grad_norm": 4.405589580535889, "learning_rate": 4.821512227115151e-05, "loss": 0.7075, "step": 9403 }, { "epoch": 0.3964837574045576, "grad_norm": 1.3233743906021118, "learning_rate": 4.821434148474344e-05, "loss": 0.1291, "step": 9404 }, { "epoch": 0.3965259185867572, "grad_norm": 3.3852460384368896, "learning_rate": 4.8213560698335364e-05, "loss": 0.7927, "step": 9405 }, { "epoch": 0.3965680797689567, "grad_norm": 0.6176903247833252, "learning_rate": 4.8212779911927295e-05, "loss": 0.1616, "step": 9406 }, { "epoch": 0.39661024095115627, "grad_norm": 1.577389121055603, "learning_rate": 4.8211999125519226e-05, "loss": 0.2784, "step": 9407 }, { "epoch": 0.39665240213335584, "grad_norm": 0.5072506666183472, "learning_rate": 4.821121833911115e-05, "loss": 0.2192, "step": 9408 }, { "epoch": 0.39669456331555536, "grad_norm": 0.5759578347206116, "learning_rate": 4.821043755270309e-05, "loss": 0.6339, "step": 9409 }, { "epoch": 0.39673672449775493, "grad_norm": 0.4602046608924866, "learning_rate": 4.8209656766295013e-05, "loss": 0.1389, "step": 9410 }, { "epoch": 0.39677888567995445, "grad_norm": 1.340159296989441, "learning_rate": 4.8208875979886945e-05, "loss": 0.4529, "step": 9411 }, { "epoch": 0.396821046862154, "grad_norm": 19.748355865478516, "learning_rate": 4.8208095193478876e-05, "loss": 0.9976, "step": 9412 }, { "epoch": 0.39686320804435354, "grad_norm": 0.5184465646743774, "learning_rate": 4.82073144070708e-05, "loss": 0.065, "step": 9413 }, { "epoch": 0.3969053692265531, "grad_norm": 0.41645917296409607, "learning_rate": 4.820653362066274e-05, "loss": 0.1495, "step": 9414 }, { "epoch": 0.39694753040875264, "grad_norm": 0.9402076005935669, "learning_rate": 4.820575283425466e-05, "loss": 0.1743, "step": 9415 }, { "epoch": 0.3969896915909522, "grad_norm": 0.31735852360725403, "learning_rate": 4.8204972047846594e-05, "loss": 0.1175, "step": 9416 }, { "epoch": 0.3970318527731518, "grad_norm": 0.38382506370544434, "learning_rate": 4.8204191261438525e-05, "loss": 0.2328, "step": 9417 }, { "epoch": 0.3970740139553513, "grad_norm": 2.9549293518066406, "learning_rate": 4.820341047503045e-05, "loss": 0.3849, "step": 9418 }, { "epoch": 0.3971161751375509, "grad_norm": 1.6670340299606323, "learning_rate": 4.820262968862238e-05, "loss": 0.2483, "step": 9419 }, { "epoch": 0.3971583363197504, "grad_norm": 0.6226763129234314, "learning_rate": 4.820184890221431e-05, "loss": 0.205, "step": 9420 }, { "epoch": 0.39720049750194997, "grad_norm": 0.9889428615570068, "learning_rate": 4.8201068115806244e-05, "loss": 0.3476, "step": 9421 }, { "epoch": 0.3972426586841495, "grad_norm": 0.3965533375740051, "learning_rate": 4.820028732939817e-05, "loss": 0.1127, "step": 9422 }, { "epoch": 0.39728481986634906, "grad_norm": 0.838901162147522, "learning_rate": 4.81995065429901e-05, "loss": 0.5115, "step": 9423 }, { "epoch": 0.3973269810485486, "grad_norm": 1.0073918104171753, "learning_rate": 4.819872575658203e-05, "loss": 0.152, "step": 9424 }, { "epoch": 0.39736914223074815, "grad_norm": 1.0956552028656006, "learning_rate": 4.819794497017396e-05, "loss": 0.9398, "step": 9425 }, { "epoch": 0.3974113034129477, "grad_norm": 0.6266634464263916, "learning_rate": 4.819716418376589e-05, "loss": 0.6562, "step": 9426 }, { "epoch": 0.39745346459514724, "grad_norm": 0.6868720054626465, "learning_rate": 4.819638339735782e-05, "loss": 0.1399, "step": 9427 }, { "epoch": 0.3974956257773468, "grad_norm": 0.959772527217865, "learning_rate": 4.8195602610949756e-05, "loss": 0.5944, "step": 9428 }, { "epoch": 0.39753778695954634, "grad_norm": 0.6914798021316528, "learning_rate": 4.819482182454168e-05, "loss": 0.1905, "step": 9429 }, { "epoch": 0.3975799481417459, "grad_norm": 0.724215567111969, "learning_rate": 4.819404103813361e-05, "loss": 0.2243, "step": 9430 }, { "epoch": 0.3976221093239454, "grad_norm": 2.54982590675354, "learning_rate": 4.819326025172554e-05, "loss": 0.3676, "step": 9431 }, { "epoch": 0.397664270506145, "grad_norm": 34.0362548828125, "learning_rate": 4.819247946531747e-05, "loss": 0.6959, "step": 9432 }, { "epoch": 0.3977064316883445, "grad_norm": 0.9010624885559082, "learning_rate": 4.81916986789094e-05, "loss": 0.3556, "step": 9433 }, { "epoch": 0.3977485928705441, "grad_norm": 2.690521478652954, "learning_rate": 4.819091789250133e-05, "loss": 0.3736, "step": 9434 }, { "epoch": 0.3977907540527436, "grad_norm": 0.8644861578941345, "learning_rate": 4.819013710609326e-05, "loss": 0.3054, "step": 9435 }, { "epoch": 0.3978329152349432, "grad_norm": 1.2255111932754517, "learning_rate": 4.818935631968519e-05, "loss": 1.0209, "step": 9436 }, { "epoch": 0.39787507641714276, "grad_norm": 3.3390607833862305, "learning_rate": 4.818857553327712e-05, "loss": 0.282, "step": 9437 }, { "epoch": 0.3979172375993423, "grad_norm": 0.5077532529830933, "learning_rate": 4.818779474686905e-05, "loss": 0.1159, "step": 9438 }, { "epoch": 0.39795939878154185, "grad_norm": 0.8184794783592224, "learning_rate": 4.818701396046098e-05, "loss": 0.1394, "step": 9439 }, { "epoch": 0.39800155996374137, "grad_norm": 0.7484124898910522, "learning_rate": 4.818623317405291e-05, "loss": 0.26, "step": 9440 }, { "epoch": 0.39804372114594094, "grad_norm": 0.3853057026863098, "learning_rate": 4.8185452387644835e-05, "loss": 0.1741, "step": 9441 }, { "epoch": 0.39808588232814046, "grad_norm": 1.5204541683197021, "learning_rate": 4.8184671601236766e-05, "loss": 0.2292, "step": 9442 }, { "epoch": 0.39812804351034004, "grad_norm": 0.38601577281951904, "learning_rate": 4.81838908148287e-05, "loss": 0.0901, "step": 9443 }, { "epoch": 0.39817020469253955, "grad_norm": 1.593610405921936, "learning_rate": 4.818311002842062e-05, "loss": 1.1955, "step": 9444 }, { "epoch": 0.3982123658747391, "grad_norm": 1.055254340171814, "learning_rate": 4.818232924201256e-05, "loss": 0.4837, "step": 9445 }, { "epoch": 0.3982545270569387, "grad_norm": 0.7713969945907593, "learning_rate": 4.8181548455604485e-05, "loss": 0.2254, "step": 9446 }, { "epoch": 0.3982966882391382, "grad_norm": 653.6014404296875, "learning_rate": 4.818076766919642e-05, "loss": 0.1007, "step": 9447 }, { "epoch": 0.3983388494213378, "grad_norm": 0.448076456785202, "learning_rate": 4.817998688278835e-05, "loss": 0.0864, "step": 9448 }, { "epoch": 0.3983810106035373, "grad_norm": 0.707794725894928, "learning_rate": 4.817920609638027e-05, "loss": 0.2665, "step": 9449 }, { "epoch": 0.3984231717857369, "grad_norm": 0.5709816217422485, "learning_rate": 4.817842530997221e-05, "loss": 0.2268, "step": 9450 }, { "epoch": 0.3984653329679364, "grad_norm": 0.8606244921684265, "learning_rate": 4.8177644523564134e-05, "loss": 0.4871, "step": 9451 }, { "epoch": 0.398507494150136, "grad_norm": 0.2884119153022766, "learning_rate": 4.8176863737156065e-05, "loss": 0.0846, "step": 9452 }, { "epoch": 0.3985496553323355, "grad_norm": 0.3105485439300537, "learning_rate": 4.8176082950748e-05, "loss": 0.1382, "step": 9453 }, { "epoch": 0.39859181651453507, "grad_norm": 1.1973536014556885, "learning_rate": 4.817530216433993e-05, "loss": 0.183, "step": 9454 }, { "epoch": 0.39863397769673464, "grad_norm": 0.8940235376358032, "learning_rate": 4.817452137793185e-05, "loss": 0.2975, "step": 9455 }, { "epoch": 0.39867613887893416, "grad_norm": 1.063742995262146, "learning_rate": 4.8173740591523784e-05, "loss": 0.1372, "step": 9456 }, { "epoch": 0.39871830006113373, "grad_norm": 0.7611694931983948, "learning_rate": 4.8172959805115715e-05, "loss": 0.5486, "step": 9457 }, { "epoch": 0.39876046124333325, "grad_norm": 1.862575888633728, "learning_rate": 4.8172179018707646e-05, "loss": 0.23, "step": 9458 }, { "epoch": 0.3988026224255328, "grad_norm": 2.4189674854278564, "learning_rate": 4.817139823229958e-05, "loss": 0.8804, "step": 9459 }, { "epoch": 0.39884478360773234, "grad_norm": 0.5715959668159485, "learning_rate": 4.81706174458915e-05, "loss": 0.2318, "step": 9460 }, { "epoch": 0.3988869447899319, "grad_norm": 0.9462904334068298, "learning_rate": 4.816983665948343e-05, "loss": 0.3413, "step": 9461 }, { "epoch": 0.39892910597213144, "grad_norm": 0.6528865098953247, "learning_rate": 4.8169055873075364e-05, "loss": 0.2659, "step": 9462 }, { "epoch": 0.398971267154331, "grad_norm": 5.393374919891357, "learning_rate": 4.816827508666729e-05, "loss": 1.0672, "step": 9463 }, { "epoch": 0.39901342833653053, "grad_norm": 0.6268011331558228, "learning_rate": 4.816749430025923e-05, "loss": 0.3883, "step": 9464 }, { "epoch": 0.3990555895187301, "grad_norm": 5.477489948272705, "learning_rate": 4.816671351385115e-05, "loss": 0.8584, "step": 9465 }, { "epoch": 0.3990977507009297, "grad_norm": 0.2747894823551178, "learning_rate": 4.816593272744308e-05, "loss": 0.0748, "step": 9466 }, { "epoch": 0.3991399118831292, "grad_norm": 0.7901437878608704, "learning_rate": 4.8165151941035014e-05, "loss": 0.8135, "step": 9467 }, { "epoch": 0.39918207306532877, "grad_norm": 0.4885178208351135, "learning_rate": 4.816437115462694e-05, "loss": 0.2653, "step": 9468 }, { "epoch": 0.3992242342475283, "grad_norm": 0.49804002046585083, "learning_rate": 4.8163590368218876e-05, "loss": 0.1527, "step": 9469 }, { "epoch": 0.39926639542972786, "grad_norm": 0.5510052442550659, "learning_rate": 4.81628095818108e-05, "loss": 0.3153, "step": 9470 }, { "epoch": 0.3993085566119274, "grad_norm": 0.5118377208709717, "learning_rate": 4.816202879540273e-05, "loss": 0.1424, "step": 9471 }, { "epoch": 0.39935071779412695, "grad_norm": 21.454530715942383, "learning_rate": 4.8161248008994663e-05, "loss": 0.5474, "step": 9472 }, { "epoch": 0.39939287897632647, "grad_norm": 0.5506883859634399, "learning_rate": 4.8160467222586595e-05, "loss": 0.2299, "step": 9473 }, { "epoch": 0.39943504015852604, "grad_norm": 0.749614953994751, "learning_rate": 4.815968643617852e-05, "loss": 0.5535, "step": 9474 }, { "epoch": 0.3994772013407256, "grad_norm": 0.9579703211784363, "learning_rate": 4.815890564977045e-05, "loss": 0.2498, "step": 9475 }, { "epoch": 0.39951936252292514, "grad_norm": 1.382319450378418, "learning_rate": 4.815812486336238e-05, "loss": 0.241, "step": 9476 }, { "epoch": 0.3995615237051247, "grad_norm": 1.7334542274475098, "learning_rate": 4.8157344076954306e-05, "loss": 0.4932, "step": 9477 }, { "epoch": 0.39960368488732423, "grad_norm": 0.6514452695846558, "learning_rate": 4.8156563290546244e-05, "loss": 0.5735, "step": 9478 }, { "epoch": 0.3996458460695238, "grad_norm": 1.0953325033187866, "learning_rate": 4.815578250413817e-05, "loss": 0.3901, "step": 9479 }, { "epoch": 0.3996880072517233, "grad_norm": 1.7742149829864502, "learning_rate": 4.81550017177301e-05, "loss": 0.5077, "step": 9480 }, { "epoch": 0.3997301684339229, "grad_norm": 0.4379443824291229, "learning_rate": 4.815422093132203e-05, "loss": 0.1336, "step": 9481 }, { "epoch": 0.3997723296161224, "grad_norm": 0.7469119429588318, "learning_rate": 4.8153440144913956e-05, "loss": 0.208, "step": 9482 }, { "epoch": 0.399814490798322, "grad_norm": 0.3889063894748688, "learning_rate": 4.8152659358505894e-05, "loss": 0.1377, "step": 9483 }, { "epoch": 0.39985665198052156, "grad_norm": 0.3655491769313812, "learning_rate": 4.815187857209782e-05, "loss": 0.0778, "step": 9484 }, { "epoch": 0.3998988131627211, "grad_norm": 1.1110295057296753, "learning_rate": 4.815109778568975e-05, "loss": 0.1192, "step": 9485 }, { "epoch": 0.39994097434492065, "grad_norm": 1.8511775732040405, "learning_rate": 4.815031699928168e-05, "loss": 0.226, "step": 9486 }, { "epoch": 0.39998313552712017, "grad_norm": 0.802451491355896, "learning_rate": 4.8149536212873605e-05, "loss": 0.6109, "step": 9487 }, { "epoch": 0.40002529670931974, "grad_norm": 0.5374400615692139, "learning_rate": 4.8148755426465537e-05, "loss": 0.1712, "step": 9488 }, { "epoch": 0.40006745789151926, "grad_norm": 0.5850756168365479, "learning_rate": 4.814797464005747e-05, "loss": 0.1066, "step": 9489 }, { "epoch": 0.40010961907371884, "grad_norm": 1.020277738571167, "learning_rate": 4.81471938536494e-05, "loss": 0.1129, "step": 9490 }, { "epoch": 0.40015178025591835, "grad_norm": 0.5035251379013062, "learning_rate": 4.814641306724133e-05, "loss": 0.2394, "step": 9491 }, { "epoch": 0.40019394143811793, "grad_norm": 0.7609970569610596, "learning_rate": 4.814563228083326e-05, "loss": 0.5552, "step": 9492 }, { "epoch": 0.40023610262031745, "grad_norm": 0.9454950094223022, "learning_rate": 4.8144851494425186e-05, "loss": 0.3067, "step": 9493 }, { "epoch": 0.400278263802517, "grad_norm": 1.9908368587493896, "learning_rate": 4.814407070801712e-05, "loss": 0.1811, "step": 9494 }, { "epoch": 0.4003204249847166, "grad_norm": 0.7369356155395508, "learning_rate": 4.814328992160905e-05, "loss": 0.2357, "step": 9495 }, { "epoch": 0.4003625861669161, "grad_norm": 1.148496389389038, "learning_rate": 4.814250913520097e-05, "loss": 0.5045, "step": 9496 }, { "epoch": 0.4004047473491157, "grad_norm": 0.7758697867393494, "learning_rate": 4.814172834879291e-05, "loss": 0.2104, "step": 9497 }, { "epoch": 0.4004469085313152, "grad_norm": 0.7472009658813477, "learning_rate": 4.8140947562384836e-05, "loss": 0.4328, "step": 9498 }, { "epoch": 0.4004890697135148, "grad_norm": 0.8406028151512146, "learning_rate": 4.814016677597677e-05, "loss": 0.131, "step": 9499 }, { "epoch": 0.4005312308957143, "grad_norm": 0.9084563255310059, "learning_rate": 4.81393859895687e-05, "loss": 0.131, "step": 9500 }, { "epoch": 0.40057339207791387, "grad_norm": 1.0520985126495361, "learning_rate": 4.813860520316062e-05, "loss": 0.2221, "step": 9501 }, { "epoch": 0.4006155532601134, "grad_norm": 2.2883143424987793, "learning_rate": 4.8137824416752554e-05, "loss": 0.5566, "step": 9502 }, { "epoch": 0.40065771444231296, "grad_norm": 0.5834916234016418, "learning_rate": 4.8137043630344485e-05, "loss": 0.1692, "step": 9503 }, { "epoch": 0.40069987562451254, "grad_norm": 0.8456748127937317, "learning_rate": 4.8136262843936416e-05, "loss": 0.3728, "step": 9504 }, { "epoch": 0.40074203680671205, "grad_norm": 1.0660789012908936, "learning_rate": 4.813548205752835e-05, "loss": 0.1025, "step": 9505 }, { "epoch": 0.40078419798891163, "grad_norm": 0.5670413970947266, "learning_rate": 4.813470127112027e-05, "loss": 0.1671, "step": 9506 }, { "epoch": 0.40082635917111115, "grad_norm": 0.739517867565155, "learning_rate": 4.81339204847122e-05, "loss": 0.2276, "step": 9507 }, { "epoch": 0.4008685203533107, "grad_norm": 0.6336997747421265, "learning_rate": 4.8133139698304135e-05, "loss": 0.1455, "step": 9508 }, { "epoch": 0.40091068153551024, "grad_norm": 0.7070066928863525, "learning_rate": 4.8132358911896066e-05, "loss": 0.2031, "step": 9509 }, { "epoch": 0.4009528427177098, "grad_norm": 0.26404690742492676, "learning_rate": 4.813157812548799e-05, "loss": 0.0795, "step": 9510 }, { "epoch": 0.40099500389990933, "grad_norm": 0.6345298290252686, "learning_rate": 4.813079733907992e-05, "loss": 0.0776, "step": 9511 }, { "epoch": 0.4010371650821089, "grad_norm": 0.47702473402023315, "learning_rate": 4.813001655267185e-05, "loss": 0.2308, "step": 9512 }, { "epoch": 0.4010793262643085, "grad_norm": 1.2948943376541138, "learning_rate": 4.812923576626378e-05, "loss": 0.1722, "step": 9513 }, { "epoch": 0.401121487446508, "grad_norm": 0.5366660952568054, "learning_rate": 4.8128454979855715e-05, "loss": 0.2235, "step": 9514 }, { "epoch": 0.40116364862870757, "grad_norm": 0.7642658948898315, "learning_rate": 4.812767419344764e-05, "loss": 0.1943, "step": 9515 }, { "epoch": 0.4012058098109071, "grad_norm": 0.5886754989624023, "learning_rate": 4.812689340703958e-05, "loss": 0.4778, "step": 9516 }, { "epoch": 0.40124797099310666, "grad_norm": 0.9561139345169067, "learning_rate": 4.81261126206315e-05, "loss": 0.3343, "step": 9517 }, { "epoch": 0.4012901321753062, "grad_norm": 0.7341185212135315, "learning_rate": 4.8125331834223434e-05, "loss": 0.5282, "step": 9518 }, { "epoch": 0.40133229335750575, "grad_norm": 2.178635597229004, "learning_rate": 4.8124551047815365e-05, "loss": 0.194, "step": 9519 }, { "epoch": 0.40137445453970527, "grad_norm": 0.32265251874923706, "learning_rate": 4.812377026140729e-05, "loss": 0.0744, "step": 9520 }, { "epoch": 0.40141661572190485, "grad_norm": 1.4591790437698364, "learning_rate": 4.812298947499922e-05, "loss": 0.5577, "step": 9521 }, { "epoch": 0.40145877690410436, "grad_norm": 0.67377108335495, "learning_rate": 4.812220868859115e-05, "loss": 0.6105, "step": 9522 }, { "epoch": 0.40150093808630394, "grad_norm": 4.538821220397949, "learning_rate": 4.812142790218308e-05, "loss": 0.5321, "step": 9523 }, { "epoch": 0.4015430992685035, "grad_norm": 1.0742977857589722, "learning_rate": 4.812064711577501e-05, "loss": 0.6643, "step": 9524 }, { "epoch": 0.40158526045070303, "grad_norm": 37.28339385986328, "learning_rate": 4.811986632936694e-05, "loss": 0.2303, "step": 9525 }, { "epoch": 0.4016274216329026, "grad_norm": 0.8794848322868347, "learning_rate": 4.811908554295887e-05, "loss": 0.3865, "step": 9526 }, { "epoch": 0.4016695828151021, "grad_norm": 0.3710425794124603, "learning_rate": 4.81183047565508e-05, "loss": 0.1165, "step": 9527 }, { "epoch": 0.4017117439973017, "grad_norm": 2.9326727390289307, "learning_rate": 4.811752397014273e-05, "loss": 0.5043, "step": 9528 }, { "epoch": 0.4017539051795012, "grad_norm": 2.5172922611236572, "learning_rate": 4.811674318373466e-05, "loss": 0.4036, "step": 9529 }, { "epoch": 0.4017960663617008, "grad_norm": 2.8263230323791504, "learning_rate": 4.811596239732659e-05, "loss": 0.6419, "step": 9530 }, { "epoch": 0.4018382275439003, "grad_norm": 1.625917911529541, "learning_rate": 4.811518161091852e-05, "loss": 0.5243, "step": 9531 }, { "epoch": 0.4018803887260999, "grad_norm": 0.8702558279037476, "learning_rate": 4.8114400824510444e-05, "loss": 0.1749, "step": 9532 }, { "epoch": 0.40192254990829945, "grad_norm": 0.7633737325668335, "learning_rate": 4.811362003810238e-05, "loss": 0.6641, "step": 9533 }, { "epoch": 0.40196471109049897, "grad_norm": 2.089421033859253, "learning_rate": 4.811283925169431e-05, "loss": 0.2668, "step": 9534 }, { "epoch": 0.40200687227269855, "grad_norm": 2.0088062286376953, "learning_rate": 4.811205846528624e-05, "loss": 0.0825, "step": 9535 }, { "epoch": 0.40204903345489806, "grad_norm": 1.0039052963256836, "learning_rate": 4.811127767887817e-05, "loss": 1.1044, "step": 9536 }, { "epoch": 0.40209119463709764, "grad_norm": 0.7507470846176147, "learning_rate": 4.81104968924701e-05, "loss": 0.1521, "step": 9537 }, { "epoch": 0.40213335581929716, "grad_norm": 0.8047015070915222, "learning_rate": 4.810971610606203e-05, "loss": 0.2541, "step": 9538 }, { "epoch": 0.40217551700149673, "grad_norm": 0.672088623046875, "learning_rate": 4.8108935319653956e-05, "loss": 0.1604, "step": 9539 }, { "epoch": 0.40221767818369625, "grad_norm": 2.24576997756958, "learning_rate": 4.810815453324589e-05, "loss": 0.4317, "step": 9540 }, { "epoch": 0.4022598393658958, "grad_norm": 0.749694287776947, "learning_rate": 4.810737374683782e-05, "loss": 0.0744, "step": 9541 }, { "epoch": 0.4023020005480954, "grad_norm": 0.8986707925796509, "learning_rate": 4.810659296042975e-05, "loss": 0.3854, "step": 9542 }, { "epoch": 0.4023441617302949, "grad_norm": 0.49266088008880615, "learning_rate": 4.8105812174021674e-05, "loss": 0.1508, "step": 9543 }, { "epoch": 0.4023863229124945, "grad_norm": 0.7314556837081909, "learning_rate": 4.8105031387613606e-05, "loss": 0.2152, "step": 9544 }, { "epoch": 0.402428484094694, "grad_norm": 0.6666662096977234, "learning_rate": 4.810425060120554e-05, "loss": 0.4444, "step": 9545 }, { "epoch": 0.4024706452768936, "grad_norm": 0.7531851530075073, "learning_rate": 4.810346981479746e-05, "loss": 0.3595, "step": 9546 }, { "epoch": 0.4025128064590931, "grad_norm": 1.143862009048462, "learning_rate": 4.81026890283894e-05, "loss": 0.5634, "step": 9547 }, { "epoch": 0.40255496764129267, "grad_norm": 2.5183615684509277, "learning_rate": 4.8101908241981324e-05, "loss": 0.1824, "step": 9548 }, { "epoch": 0.4025971288234922, "grad_norm": 1.042327880859375, "learning_rate": 4.8101127455573255e-05, "loss": 0.6395, "step": 9549 }, { "epoch": 0.40263929000569176, "grad_norm": 0.6553325057029724, "learning_rate": 4.8100346669165187e-05, "loss": 0.212, "step": 9550 }, { "epoch": 0.4026814511878913, "grad_norm": 4.990105628967285, "learning_rate": 4.809956588275711e-05, "loss": 0.8792, "step": 9551 }, { "epoch": 0.40272361237009086, "grad_norm": 1.0884939432144165, "learning_rate": 4.809878509634905e-05, "loss": 0.2335, "step": 9552 }, { "epoch": 0.40276577355229043, "grad_norm": 0.5347087979316711, "learning_rate": 4.8098004309940974e-05, "loss": 0.1718, "step": 9553 }, { "epoch": 0.40280793473448995, "grad_norm": 0.6582790613174438, "learning_rate": 4.8097223523532905e-05, "loss": 0.2379, "step": 9554 }, { "epoch": 0.4028500959166895, "grad_norm": 30.500362396240234, "learning_rate": 4.8096442737124836e-05, "loss": 0.2125, "step": 9555 }, { "epoch": 0.40289225709888904, "grad_norm": 0.7499064803123474, "learning_rate": 4.809566195071676e-05, "loss": 0.147, "step": 9556 }, { "epoch": 0.4029344182810886, "grad_norm": 7.969039440155029, "learning_rate": 4.809488116430869e-05, "loss": 0.2362, "step": 9557 }, { "epoch": 0.40297657946328813, "grad_norm": 12.07973861694336, "learning_rate": 4.809410037790062e-05, "loss": 0.1225, "step": 9558 }, { "epoch": 0.4030187406454877, "grad_norm": 0.9920541048049927, "learning_rate": 4.8093319591492554e-05, "loss": 0.2641, "step": 9559 }, { "epoch": 0.4030609018276872, "grad_norm": 3.4550442695617676, "learning_rate": 4.8092538805084486e-05, "loss": 0.9883, "step": 9560 }, { "epoch": 0.4031030630098868, "grad_norm": 0.9925726652145386, "learning_rate": 4.809175801867642e-05, "loss": 0.1079, "step": 9561 }, { "epoch": 0.40314522419208637, "grad_norm": 0.5060190558433533, "learning_rate": 4.809097723226834e-05, "loss": 0.1169, "step": 9562 }, { "epoch": 0.4031873853742859, "grad_norm": 0.6120166182518005, "learning_rate": 4.809019644586027e-05, "loss": 0.2413, "step": 9563 }, { "epoch": 0.40322954655648546, "grad_norm": 0.8488950133323669, "learning_rate": 4.8089415659452204e-05, "loss": 0.2524, "step": 9564 }, { "epoch": 0.403271707738685, "grad_norm": 3.0110342502593994, "learning_rate": 4.808863487304413e-05, "loss": 0.1167, "step": 9565 }, { "epoch": 0.40331386892088456, "grad_norm": 0.8818640112876892, "learning_rate": 4.8087854086636066e-05, "loss": 0.1628, "step": 9566 }, { "epoch": 0.4033560301030841, "grad_norm": 0.8225048780441284, "learning_rate": 4.808707330022799e-05, "loss": 0.5195, "step": 9567 }, { "epoch": 0.40339819128528365, "grad_norm": 0.4408281743526459, "learning_rate": 4.808629251381992e-05, "loss": 0.0781, "step": 9568 }, { "epoch": 0.40344035246748317, "grad_norm": 1.2742754220962524, "learning_rate": 4.808551172741185e-05, "loss": 0.6645, "step": 9569 }, { "epoch": 0.40348251364968274, "grad_norm": 0.7975082397460938, "learning_rate": 4.808473094100378e-05, "loss": 0.1508, "step": 9570 }, { "epoch": 0.4035246748318823, "grad_norm": 1.3579870462417603, "learning_rate": 4.8083950154595716e-05, "loss": 0.4382, "step": 9571 }, { "epoch": 0.40356683601408183, "grad_norm": 0.9131385087966919, "learning_rate": 4.808316936818764e-05, "loss": 0.3469, "step": 9572 }, { "epoch": 0.4036089971962814, "grad_norm": 4.3262152671813965, "learning_rate": 4.808238858177957e-05, "loss": 0.25, "step": 9573 }, { "epoch": 0.4036511583784809, "grad_norm": 1.0872970819473267, "learning_rate": 4.80816077953715e-05, "loss": 0.4571, "step": 9574 }, { "epoch": 0.4036933195606805, "grad_norm": 1.1629948616027832, "learning_rate": 4.808082700896343e-05, "loss": 0.6183, "step": 9575 }, { "epoch": 0.40373548074288, "grad_norm": 6.926462173461914, "learning_rate": 4.808004622255536e-05, "loss": 0.6081, "step": 9576 }, { "epoch": 0.4037776419250796, "grad_norm": 32.5369987487793, "learning_rate": 4.807926543614729e-05, "loss": 0.5917, "step": 9577 }, { "epoch": 0.4038198031072791, "grad_norm": 6.16141939163208, "learning_rate": 4.807848464973922e-05, "loss": 1.5362, "step": 9578 }, { "epoch": 0.4038619642894787, "grad_norm": 9.974523544311523, "learning_rate": 4.8077703863331146e-05, "loss": 1.1826, "step": 9579 }, { "epoch": 0.4039041254716782, "grad_norm": 21.211727142333984, "learning_rate": 4.8076923076923084e-05, "loss": 1.0284, "step": 9580 }, { "epoch": 0.4039462866538778, "grad_norm": 1.8124698400497437, "learning_rate": 4.807614229051501e-05, "loss": 1.4652, "step": 9581 }, { "epoch": 0.40398844783607735, "grad_norm": 46.45866775512695, "learning_rate": 4.807536150410694e-05, "loss": 1.0297, "step": 9582 }, { "epoch": 0.40403060901827687, "grad_norm": 3500.801025390625, "learning_rate": 4.807458071769887e-05, "loss": 0.9378, "step": 9583 }, { "epoch": 0.40407277020047644, "grad_norm": 8.987638473510742, "learning_rate": 4.8073799931290795e-05, "loss": 1.1287, "step": 9584 }, { "epoch": 0.40411493138267596, "grad_norm": 6.243098258972168, "learning_rate": 4.807301914488273e-05, "loss": 0.8795, "step": 9585 }, { "epoch": 0.40415709256487553, "grad_norm": 6.859901428222656, "learning_rate": 4.807223835847466e-05, "loss": 1.0529, "step": 9586 }, { "epoch": 0.40419925374707505, "grad_norm": 1.5262739658355713, "learning_rate": 4.807145757206659e-05, "loss": 0.6817, "step": 9587 }, { "epoch": 0.4042414149292746, "grad_norm": 5.862540245056152, "learning_rate": 4.807067678565852e-05, "loss": 0.8929, "step": 9588 }, { "epoch": 0.40428357611147414, "grad_norm": 146.24378967285156, "learning_rate": 4.8069895999250445e-05, "loss": 0.9172, "step": 9589 }, { "epoch": 0.4043257372936737, "grad_norm": 2.068098545074463, "learning_rate": 4.8069115212842376e-05, "loss": 0.5178, "step": 9590 }, { "epoch": 0.4043678984758733, "grad_norm": 2.31109356880188, "learning_rate": 4.806833442643431e-05, "loss": 0.7264, "step": 9591 }, { "epoch": 0.4044100596580728, "grad_norm": 1.2637745141983032, "learning_rate": 4.806755364002624e-05, "loss": 0.3539, "step": 9592 }, { "epoch": 0.4044522208402724, "grad_norm": 1061.7257080078125, "learning_rate": 4.806677285361816e-05, "loss": 0.6005, "step": 9593 }, { "epoch": 0.4044943820224719, "grad_norm": 27.26010513305664, "learning_rate": 4.8065992067210094e-05, "loss": 0.2219, "step": 9594 }, { "epoch": 0.4045365432046715, "grad_norm": 0.8157386183738708, "learning_rate": 4.8065211280802025e-05, "loss": 0.3974, "step": 9595 }, { "epoch": 0.404578704386871, "grad_norm": 0.5625200271606445, "learning_rate": 4.806443049439396e-05, "loss": 0.1943, "step": 9596 }, { "epoch": 0.40462086556907056, "grad_norm": 0.6016032695770264, "learning_rate": 4.806364970798589e-05, "loss": 0.1739, "step": 9597 }, { "epoch": 0.4046630267512701, "grad_norm": 0.6443092823028564, "learning_rate": 4.806286892157781e-05, "loss": 0.113, "step": 9598 }, { "epoch": 0.40470518793346966, "grad_norm": 0.6577837467193604, "learning_rate": 4.806208813516975e-05, "loss": 0.2154, "step": 9599 }, { "epoch": 0.40474734911566923, "grad_norm": 0.491753488779068, "learning_rate": 4.8061307348761675e-05, "loss": 0.1251, "step": 9600 }, { "epoch": 0.40478951029786875, "grad_norm": 0.7068836092948914, "learning_rate": 4.80605265623536e-05, "loss": 0.5085, "step": 9601 }, { "epoch": 0.4048316714800683, "grad_norm": 0.6977285146713257, "learning_rate": 4.805974577594554e-05, "loss": 0.228, "step": 9602 }, { "epoch": 0.40487383266226784, "grad_norm": 1.0998955965042114, "learning_rate": 4.805896498953746e-05, "loss": 0.1648, "step": 9603 }, { "epoch": 0.4049159938444674, "grad_norm": 1.8766067028045654, "learning_rate": 4.805818420312939e-05, "loss": 0.2745, "step": 9604 }, { "epoch": 0.40495815502666693, "grad_norm": 1.0903137922286987, "learning_rate": 4.8057403416721324e-05, "loss": 0.5065, "step": 9605 }, { "epoch": 0.4050003162088665, "grad_norm": 0.7125757932662964, "learning_rate": 4.8056622630313256e-05, "loss": 0.5462, "step": 9606 }, { "epoch": 0.405042477391066, "grad_norm": 1.8432658910751343, "learning_rate": 4.805584184390519e-05, "loss": 1.0683, "step": 9607 }, { "epoch": 0.4050846385732656, "grad_norm": 6.530701637268066, "learning_rate": 4.805506105749711e-05, "loss": 1.0018, "step": 9608 }, { "epoch": 0.4051267997554651, "grad_norm": 1.3547265529632568, "learning_rate": 4.805428027108904e-05, "loss": 0.46, "step": 9609 }, { "epoch": 0.4051689609376647, "grad_norm": 3.5658514499664307, "learning_rate": 4.8053499484680974e-05, "loss": 0.5932, "step": 9610 }, { "epoch": 0.40521112211986426, "grad_norm": 1.4673794507980347, "learning_rate": 4.8052718698272905e-05, "loss": 0.9871, "step": 9611 }, { "epoch": 0.4052532833020638, "grad_norm": 0.666028618812561, "learning_rate": 4.805193791186483e-05, "loss": 0.701, "step": 9612 }, { "epoch": 0.40529544448426336, "grad_norm": 0.4767850339412689, "learning_rate": 4.805115712545676e-05, "loss": 0.3224, "step": 9613 }, { "epoch": 0.4053376056664629, "grad_norm": 1.0751824378967285, "learning_rate": 4.805037633904869e-05, "loss": 0.5025, "step": 9614 }, { "epoch": 0.40537976684866245, "grad_norm": 0.8776291608810425, "learning_rate": 4.804959555264062e-05, "loss": 0.6398, "step": 9615 }, { "epoch": 0.40542192803086197, "grad_norm": 0.5052894949913025, "learning_rate": 4.8048814766232555e-05, "loss": 0.2075, "step": 9616 }, { "epoch": 0.40546408921306154, "grad_norm": 0.42771220207214355, "learning_rate": 4.804803397982448e-05, "loss": 0.137, "step": 9617 }, { "epoch": 0.40550625039526106, "grad_norm": 0.6993048787117004, "learning_rate": 4.804725319341641e-05, "loss": 0.1973, "step": 9618 }, { "epoch": 0.40554841157746063, "grad_norm": 1.791885495185852, "learning_rate": 4.804647240700834e-05, "loss": 0.3256, "step": 9619 }, { "epoch": 0.4055905727596602, "grad_norm": 0.705007016658783, "learning_rate": 4.8045691620600266e-05, "loss": 0.2148, "step": 9620 }, { "epoch": 0.4056327339418597, "grad_norm": 2.7886526584625244, "learning_rate": 4.8044910834192204e-05, "loss": 1.0879, "step": 9621 }, { "epoch": 0.4056748951240593, "grad_norm": 4.814601421356201, "learning_rate": 4.804413004778413e-05, "loss": 0.7194, "step": 9622 }, { "epoch": 0.4057170563062588, "grad_norm": 1.0887658596038818, "learning_rate": 4.804334926137606e-05, "loss": 0.3998, "step": 9623 }, { "epoch": 0.4057592174884584, "grad_norm": 2.5226328372955322, "learning_rate": 4.804256847496799e-05, "loss": 0.1614, "step": 9624 }, { "epoch": 0.4058013786706579, "grad_norm": 0.8876458406448364, "learning_rate": 4.804178768855992e-05, "loss": 0.1978, "step": 9625 }, { "epoch": 0.4058435398528575, "grad_norm": 0.8994358777999878, "learning_rate": 4.804100690215185e-05, "loss": 0.2173, "step": 9626 }, { "epoch": 0.405885701035057, "grad_norm": 0.3131943643093109, "learning_rate": 4.804022611574378e-05, "loss": 0.0947, "step": 9627 }, { "epoch": 0.4059278622172566, "grad_norm": 0.3674412667751312, "learning_rate": 4.803944532933571e-05, "loss": 0.1379, "step": 9628 }, { "epoch": 0.40597002339945615, "grad_norm": 1.057267427444458, "learning_rate": 4.803866454292764e-05, "loss": 0.1501, "step": 9629 }, { "epoch": 0.40601218458165567, "grad_norm": 1.9513344764709473, "learning_rate": 4.803788375651957e-05, "loss": 0.5839, "step": 9630 }, { "epoch": 0.40605434576385524, "grad_norm": 0.7800675630569458, "learning_rate": 4.8037102970111497e-05, "loss": 0.3447, "step": 9631 }, { "epoch": 0.40609650694605476, "grad_norm": 0.6375932097434998, "learning_rate": 4.803632218370343e-05, "loss": 0.1865, "step": 9632 }, { "epoch": 0.40613866812825433, "grad_norm": 0.40911099314689636, "learning_rate": 4.803554139729536e-05, "loss": 0.2, "step": 9633 }, { "epoch": 0.40618082931045385, "grad_norm": 0.36327213048934937, "learning_rate": 4.8034760610887284e-05, "loss": 0.1306, "step": 9634 }, { "epoch": 0.4062229904926534, "grad_norm": 0.9461007714271545, "learning_rate": 4.803397982447922e-05, "loss": 0.1849, "step": 9635 }, { "epoch": 0.40626515167485294, "grad_norm": 0.6015740633010864, "learning_rate": 4.8033199038071146e-05, "loss": 0.5898, "step": 9636 }, { "epoch": 0.4063073128570525, "grad_norm": 1.0990159511566162, "learning_rate": 4.803241825166308e-05, "loss": 0.2775, "step": 9637 }, { "epoch": 0.40634947403925203, "grad_norm": 0.778044581413269, "learning_rate": 4.803163746525501e-05, "loss": 0.4517, "step": 9638 }, { "epoch": 0.4063916352214516, "grad_norm": 4.805224418640137, "learning_rate": 4.803085667884693e-05, "loss": 0.8429, "step": 9639 }, { "epoch": 0.4064337964036512, "grad_norm": 0.6649098992347717, "learning_rate": 4.803007589243887e-05, "loss": 0.307, "step": 9640 }, { "epoch": 0.4064759575858507, "grad_norm": 0.406084805727005, "learning_rate": 4.8029295106030796e-05, "loss": 0.118, "step": 9641 }, { "epoch": 0.4065181187680503, "grad_norm": 0.6891419291496277, "learning_rate": 4.802851431962273e-05, "loss": 0.2406, "step": 9642 }, { "epoch": 0.4065602799502498, "grad_norm": 1.110208511352539, "learning_rate": 4.802773353321466e-05, "loss": 0.4612, "step": 9643 }, { "epoch": 0.40660244113244937, "grad_norm": 0.9818741083145142, "learning_rate": 4.802695274680659e-05, "loss": 0.6476, "step": 9644 }, { "epoch": 0.4066446023146489, "grad_norm": 1.304090976715088, "learning_rate": 4.8026171960398514e-05, "loss": 0.2661, "step": 9645 }, { "epoch": 0.40668676349684846, "grad_norm": 0.9243986010551453, "learning_rate": 4.8025391173990445e-05, "loss": 0.1185, "step": 9646 }, { "epoch": 0.406728924679048, "grad_norm": 4.239387035369873, "learning_rate": 4.8024610387582376e-05, "loss": 0.1595, "step": 9647 }, { "epoch": 0.40677108586124755, "grad_norm": 0.7884100675582886, "learning_rate": 4.80238296011743e-05, "loss": 0.4978, "step": 9648 }, { "epoch": 0.4068132470434471, "grad_norm": 0.386406272649765, "learning_rate": 4.802304881476624e-05, "loss": 0.1153, "step": 9649 }, { "epoch": 0.40685540822564664, "grad_norm": 1.2968616485595703, "learning_rate": 4.8022268028358163e-05, "loss": 0.2802, "step": 9650 }, { "epoch": 0.4068975694078462, "grad_norm": 0.865830659866333, "learning_rate": 4.8021487241950095e-05, "loss": 0.365, "step": 9651 }, { "epoch": 0.40693973059004573, "grad_norm": 0.8131710886955261, "learning_rate": 4.8020706455542026e-05, "loss": 0.1275, "step": 9652 }, { "epoch": 0.4069818917722453, "grad_norm": 1.0522860288619995, "learning_rate": 4.801992566913395e-05, "loss": 0.1078, "step": 9653 }, { "epoch": 0.4070240529544448, "grad_norm": 1.4668934345245361, "learning_rate": 4.801914488272589e-05, "loss": 0.5102, "step": 9654 }, { "epoch": 0.4070662141366444, "grad_norm": 4.4325761795043945, "learning_rate": 4.801836409631781e-05, "loss": 0.5854, "step": 9655 }, { "epoch": 0.4071083753188439, "grad_norm": 0.8516280055046082, "learning_rate": 4.8017583309909744e-05, "loss": 0.1402, "step": 9656 }, { "epoch": 0.4071505365010435, "grad_norm": 0.5621041655540466, "learning_rate": 4.8016802523501675e-05, "loss": 0.1229, "step": 9657 }, { "epoch": 0.40719269768324307, "grad_norm": 0.664210855960846, "learning_rate": 4.80160217370936e-05, "loss": 0.1557, "step": 9658 }, { "epoch": 0.4072348588654426, "grad_norm": 0.5135124921798706, "learning_rate": 4.801524095068553e-05, "loss": 0.2065, "step": 9659 }, { "epoch": 0.40727702004764216, "grad_norm": 0.5183254480361938, "learning_rate": 4.801446016427746e-05, "loss": 0.0903, "step": 9660 }, { "epoch": 0.4073191812298417, "grad_norm": 1.6895235776901245, "learning_rate": 4.8013679377869394e-05, "loss": 0.8719, "step": 9661 }, { "epoch": 0.40736134241204125, "grad_norm": 0.5002369284629822, "learning_rate": 4.801289859146132e-05, "loss": 0.0802, "step": 9662 }, { "epoch": 0.40740350359424077, "grad_norm": 0.880911111831665, "learning_rate": 4.801211780505325e-05, "loss": 0.3219, "step": 9663 }, { "epoch": 0.40744566477644034, "grad_norm": 0.8815792798995972, "learning_rate": 4.801133701864518e-05, "loss": 0.2194, "step": 9664 }, { "epoch": 0.40748782595863986, "grad_norm": 0.6042813062667847, "learning_rate": 4.801055623223711e-05, "loss": 0.1968, "step": 9665 }, { "epoch": 0.40752998714083943, "grad_norm": 0.7919464707374573, "learning_rate": 4.800977544582904e-05, "loss": 0.1836, "step": 9666 }, { "epoch": 0.40757214832303895, "grad_norm": 2.801159620285034, "learning_rate": 4.800899465942097e-05, "loss": 0.8467, "step": 9667 }, { "epoch": 0.4076143095052385, "grad_norm": 4.371845722198486, "learning_rate": 4.8008213873012906e-05, "loss": 0.5884, "step": 9668 }, { "epoch": 0.4076564706874381, "grad_norm": 1.3708466291427612, "learning_rate": 4.800743308660483e-05, "loss": 1.0215, "step": 9669 }, { "epoch": 0.4076986318696376, "grad_norm": 0.7830452919006348, "learning_rate": 4.800665230019676e-05, "loss": 0.2751, "step": 9670 }, { "epoch": 0.4077407930518372, "grad_norm": 2.5647239685058594, "learning_rate": 4.800587151378869e-05, "loss": 0.9772, "step": 9671 }, { "epoch": 0.4077829542340367, "grad_norm": 0.3654167950153351, "learning_rate": 4.800509072738062e-05, "loss": 0.074, "step": 9672 }, { "epoch": 0.4078251154162363, "grad_norm": 0.9283328652381897, "learning_rate": 4.800430994097255e-05, "loss": 0.6578, "step": 9673 }, { "epoch": 0.4078672765984358, "grad_norm": 0.4110657572746277, "learning_rate": 4.800352915456448e-05, "loss": 0.153, "step": 9674 }, { "epoch": 0.4079094377806354, "grad_norm": 0.5505207777023315, "learning_rate": 4.800274836815641e-05, "loss": 0.2361, "step": 9675 }, { "epoch": 0.4079515989628349, "grad_norm": 1.5961692333221436, "learning_rate": 4.800196758174834e-05, "loss": 0.0985, "step": 9676 }, { "epoch": 0.40799376014503447, "grad_norm": 0.7045090794563293, "learning_rate": 4.800118679534027e-05, "loss": 0.1622, "step": 9677 }, { "epoch": 0.40803592132723404, "grad_norm": 0.33554786443710327, "learning_rate": 4.80004060089322e-05, "loss": 0.0893, "step": 9678 }, { "epoch": 0.40807808250943356, "grad_norm": 0.6949686408042908, "learning_rate": 4.799962522252413e-05, "loss": 0.5903, "step": 9679 }, { "epoch": 0.40812024369163313, "grad_norm": 5704.3310546875, "learning_rate": 4.799884443611606e-05, "loss": 0.526, "step": 9680 }, { "epoch": 0.40816240487383265, "grad_norm": 26.452911376953125, "learning_rate": 4.7998063649707985e-05, "loss": 0.6655, "step": 9681 }, { "epoch": 0.4082045660560322, "grad_norm": 0.5927451848983765, "learning_rate": 4.7997282863299916e-05, "loss": 0.4932, "step": 9682 }, { "epoch": 0.40824672723823174, "grad_norm": 0.8386524319648743, "learning_rate": 4.799650207689185e-05, "loss": 0.4626, "step": 9683 }, { "epoch": 0.4082888884204313, "grad_norm": 0.46815603971481323, "learning_rate": 4.799572129048377e-05, "loss": 0.1684, "step": 9684 }, { "epoch": 0.40833104960263084, "grad_norm": 0.6472854018211365, "learning_rate": 4.799494050407571e-05, "loss": 0.124, "step": 9685 }, { "epoch": 0.4083732107848304, "grad_norm": 1.722398281097412, "learning_rate": 4.7994159717667635e-05, "loss": 0.404, "step": 9686 }, { "epoch": 0.40841537196703, "grad_norm": 3.7371904850006104, "learning_rate": 4.799337893125957e-05, "loss": 1.0976, "step": 9687 }, { "epoch": 0.4084575331492295, "grad_norm": 0.4436758756637573, "learning_rate": 4.79925981448515e-05, "loss": 0.1079, "step": 9688 }, { "epoch": 0.4084996943314291, "grad_norm": 0.5121629238128662, "learning_rate": 4.799181735844343e-05, "loss": 0.0897, "step": 9689 }, { "epoch": 0.4085418555136286, "grad_norm": 1.6606945991516113, "learning_rate": 4.799103657203536e-05, "loss": 0.7102, "step": 9690 }, { "epoch": 0.40858401669582817, "grad_norm": 0.891097366809845, "learning_rate": 4.7990255785627284e-05, "loss": 0.5876, "step": 9691 }, { "epoch": 0.4086261778780277, "grad_norm": 0.6437088251113892, "learning_rate": 4.7989474999219215e-05, "loss": 0.1547, "step": 9692 }, { "epoch": 0.40866833906022726, "grad_norm": 0.5405550003051758, "learning_rate": 4.798869421281115e-05, "loss": 0.217, "step": 9693 }, { "epoch": 0.4087105002424268, "grad_norm": 1.964049220085144, "learning_rate": 4.798791342640308e-05, "loss": 0.9027, "step": 9694 }, { "epoch": 0.40875266142462635, "grad_norm": 0.5178130269050598, "learning_rate": 4.7987132639995e-05, "loss": 0.3463, "step": 9695 }, { "epoch": 0.40879482260682587, "grad_norm": 1.3556772470474243, "learning_rate": 4.7986351853586934e-05, "loss": 0.3849, "step": 9696 }, { "epoch": 0.40883698378902544, "grad_norm": 0.491337388753891, "learning_rate": 4.7985571067178865e-05, "loss": 0.1088, "step": 9697 }, { "epoch": 0.408879144971225, "grad_norm": 1.464639663696289, "learning_rate": 4.7984790280770796e-05, "loss": 0.6961, "step": 9698 }, { "epoch": 0.40892130615342454, "grad_norm": 1.2751076221466064, "learning_rate": 4.798400949436273e-05, "loss": 0.9079, "step": 9699 }, { "epoch": 0.4089634673356241, "grad_norm": 1.3416621685028076, "learning_rate": 4.798322870795465e-05, "loss": 0.5358, "step": 9700 }, { "epoch": 0.4090056285178236, "grad_norm": 1.35970139503479, "learning_rate": 4.798244792154658e-05, "loss": 0.4203, "step": 9701 }, { "epoch": 0.4090477897000232, "grad_norm": 2.1967761516571045, "learning_rate": 4.7981667135138514e-05, "loss": 0.0853, "step": 9702 }, { "epoch": 0.4090899508822227, "grad_norm": 0.9502485394477844, "learning_rate": 4.798088634873044e-05, "loss": 0.25, "step": 9703 }, { "epoch": 0.4091321120644223, "grad_norm": 1.2272932529449463, "learning_rate": 4.798010556232238e-05, "loss": 0.1571, "step": 9704 }, { "epoch": 0.4091742732466218, "grad_norm": 0.6869809031486511, "learning_rate": 4.79793247759143e-05, "loss": 0.3321, "step": 9705 }, { "epoch": 0.4092164344288214, "grad_norm": 0.6594030857086182, "learning_rate": 4.797854398950623e-05, "loss": 0.1117, "step": 9706 }, { "epoch": 0.40925859561102096, "grad_norm": 0.9093565940856934, "learning_rate": 4.7977763203098164e-05, "loss": 0.597, "step": 9707 }, { "epoch": 0.4093007567932205, "grad_norm": 3.4469990730285645, "learning_rate": 4.797698241669009e-05, "loss": 0.3044, "step": 9708 }, { "epoch": 0.40934291797542005, "grad_norm": 0.4942173957824707, "learning_rate": 4.7976201630282026e-05, "loss": 0.0967, "step": 9709 }, { "epoch": 0.40938507915761957, "grad_norm": 1.3022291660308838, "learning_rate": 4.797542084387395e-05, "loss": 0.1708, "step": 9710 }, { "epoch": 0.40942724033981914, "grad_norm": 0.5361248254776001, "learning_rate": 4.797464005746588e-05, "loss": 0.063, "step": 9711 }, { "epoch": 0.40946940152201866, "grad_norm": 0.731417179107666, "learning_rate": 4.7973859271057813e-05, "loss": 0.1714, "step": 9712 }, { "epoch": 0.40951156270421823, "grad_norm": 0.6507936120033264, "learning_rate": 4.7973078484649745e-05, "loss": 0.1814, "step": 9713 }, { "epoch": 0.40955372388641775, "grad_norm": 0.42998364567756653, "learning_rate": 4.797229769824167e-05, "loss": 0.1307, "step": 9714 }, { "epoch": 0.4095958850686173, "grad_norm": 1.4059405326843262, "learning_rate": 4.79715169118336e-05, "loss": 0.324, "step": 9715 }, { "epoch": 0.40963804625081685, "grad_norm": 1.8461097478866577, "learning_rate": 4.797073612542553e-05, "loss": 0.4732, "step": 9716 }, { "epoch": 0.4096802074330164, "grad_norm": 1.7738450765609741, "learning_rate": 4.7969955339017456e-05, "loss": 0.3869, "step": 9717 }, { "epoch": 0.409722368615216, "grad_norm": 2.2332730293273926, "learning_rate": 4.7969174552609394e-05, "loss": 0.9991, "step": 9718 }, { "epoch": 0.4097645297974155, "grad_norm": 0.6278132200241089, "learning_rate": 4.796839376620132e-05, "loss": 0.1774, "step": 9719 }, { "epoch": 0.4098066909796151, "grad_norm": 1.228725790977478, "learning_rate": 4.796761297979325e-05, "loss": 0.1216, "step": 9720 }, { "epoch": 0.4098488521618146, "grad_norm": 1.9197371006011963, "learning_rate": 4.796683219338518e-05, "loss": 0.1435, "step": 9721 }, { "epoch": 0.4098910133440142, "grad_norm": 0.6015625, "learning_rate": 4.7966051406977106e-05, "loss": 0.0596, "step": 9722 }, { "epoch": 0.4099331745262137, "grad_norm": 0.43293097615242004, "learning_rate": 4.7965270620569044e-05, "loss": 0.2135, "step": 9723 }, { "epoch": 0.40997533570841327, "grad_norm": 1.3658093214035034, "learning_rate": 4.796448983416097e-05, "loss": 0.3039, "step": 9724 }, { "epoch": 0.4100174968906128, "grad_norm": 0.9354677796363831, "learning_rate": 4.79637090477529e-05, "loss": 0.2727, "step": 9725 }, { "epoch": 0.41005965807281236, "grad_norm": 1.5728206634521484, "learning_rate": 4.796292826134483e-05, "loss": 0.1799, "step": 9726 }, { "epoch": 0.41010181925501193, "grad_norm": 0.6067072153091431, "learning_rate": 4.7962147474936755e-05, "loss": 0.0829, "step": 9727 }, { "epoch": 0.41014398043721145, "grad_norm": 0.4110657274723053, "learning_rate": 4.7961366688528686e-05, "loss": 0.1034, "step": 9728 }, { "epoch": 0.410186141619411, "grad_norm": 0.5871211886405945, "learning_rate": 4.796058590212062e-05, "loss": 0.18, "step": 9729 }, { "epoch": 0.41022830280161054, "grad_norm": 2.873384714126587, "learning_rate": 4.795980511571255e-05, "loss": 0.2042, "step": 9730 }, { "epoch": 0.4102704639838101, "grad_norm": 1.4765183925628662, "learning_rate": 4.795902432930448e-05, "loss": 1.0446, "step": 9731 }, { "epoch": 0.41031262516600964, "grad_norm": 0.5753685832023621, "learning_rate": 4.795824354289641e-05, "loss": 0.1176, "step": 9732 }, { "epoch": 0.4103547863482092, "grad_norm": 0.6459425091743469, "learning_rate": 4.7957462756488336e-05, "loss": 0.4786, "step": 9733 }, { "epoch": 0.41039694753040873, "grad_norm": 1.0130245685577393, "learning_rate": 4.795668197008027e-05, "loss": 0.59, "step": 9734 }, { "epoch": 0.4104391087126083, "grad_norm": 1.4912008047103882, "learning_rate": 4.79559011836722e-05, "loss": 0.9572, "step": 9735 }, { "epoch": 0.4104812698948079, "grad_norm": 0.7595909833908081, "learning_rate": 4.795512039726412e-05, "loss": 0.1421, "step": 9736 }, { "epoch": 0.4105234310770074, "grad_norm": 0.5045429468154907, "learning_rate": 4.795433961085606e-05, "loss": 0.1071, "step": 9737 }, { "epoch": 0.41056559225920697, "grad_norm": 0.5412925481796265, "learning_rate": 4.7953558824447986e-05, "loss": 0.1686, "step": 9738 }, { "epoch": 0.4106077534414065, "grad_norm": 0.934349000453949, "learning_rate": 4.795277803803992e-05, "loss": 0.2048, "step": 9739 }, { "epoch": 0.41064991462360606, "grad_norm": 0.9096880555152893, "learning_rate": 4.795199725163185e-05, "loss": 0.1178, "step": 9740 }, { "epoch": 0.4106920758058056, "grad_norm": 0.8801180720329285, "learning_rate": 4.795121646522377e-05, "loss": 0.1614, "step": 9741 }, { "epoch": 0.41073423698800515, "grad_norm": 1.2443504333496094, "learning_rate": 4.7950435678815704e-05, "loss": 0.1289, "step": 9742 }, { "epoch": 0.41077639817020467, "grad_norm": 1.9042141437530518, "learning_rate": 4.7949654892407635e-05, "loss": 0.5104, "step": 9743 }, { "epoch": 0.41081855935240424, "grad_norm": 0.38544946908950806, "learning_rate": 4.7948874105999566e-05, "loss": 0.1099, "step": 9744 }, { "epoch": 0.41086072053460376, "grad_norm": 0.4107016921043396, "learning_rate": 4.79480933195915e-05, "loss": 0.0864, "step": 9745 }, { "epoch": 0.41090288171680334, "grad_norm": 3.1073501110076904, "learning_rate": 4.794731253318342e-05, "loss": 0.5405, "step": 9746 }, { "epoch": 0.4109450428990029, "grad_norm": 1.2714985609054565, "learning_rate": 4.794653174677535e-05, "loss": 0.2319, "step": 9747 }, { "epoch": 0.41098720408120243, "grad_norm": 1.3130255937576294, "learning_rate": 4.7945750960367285e-05, "loss": 0.1979, "step": 9748 }, { "epoch": 0.411029365263402, "grad_norm": 12.0272216796875, "learning_rate": 4.7944970173959216e-05, "loss": 0.1193, "step": 9749 }, { "epoch": 0.4110715264456015, "grad_norm": 19.587997436523438, "learning_rate": 4.794418938755114e-05, "loss": 0.1079, "step": 9750 }, { "epoch": 0.4111136876278011, "grad_norm": 1.1064503192901611, "learning_rate": 4.794340860114307e-05, "loss": 0.1747, "step": 9751 }, { "epoch": 0.4111558488100006, "grad_norm": 0.654007613658905, "learning_rate": 4.7942627814735e-05, "loss": 0.2041, "step": 9752 }, { "epoch": 0.4111980099922002, "grad_norm": 1.4056086540222168, "learning_rate": 4.794184702832693e-05, "loss": 0.5502, "step": 9753 }, { "epoch": 0.4112401711743997, "grad_norm": 3.327951431274414, "learning_rate": 4.7941066241918865e-05, "loss": 0.4331, "step": 9754 }, { "epoch": 0.4112823323565993, "grad_norm": 0.5132291913032532, "learning_rate": 4.794028545551079e-05, "loss": 0.3787, "step": 9755 }, { "epoch": 0.41132449353879885, "grad_norm": 2.323904037475586, "learning_rate": 4.793950466910273e-05, "loss": 0.5204, "step": 9756 }, { "epoch": 0.41136665472099837, "grad_norm": 37.430484771728516, "learning_rate": 4.793872388269465e-05, "loss": 0.1313, "step": 9757 }, { "epoch": 0.41140881590319794, "grad_norm": 84.01014709472656, "learning_rate": 4.7937943096286584e-05, "loss": 0.4321, "step": 9758 }, { "epoch": 0.41145097708539746, "grad_norm": 1.1064376831054688, "learning_rate": 4.7937162309878515e-05, "loss": 0.1308, "step": 9759 }, { "epoch": 0.41149313826759704, "grad_norm": 1.036325454711914, "learning_rate": 4.793638152347044e-05, "loss": 0.0855, "step": 9760 }, { "epoch": 0.41153529944979655, "grad_norm": 1.8743270635604858, "learning_rate": 4.793560073706237e-05, "loss": 0.3898, "step": 9761 }, { "epoch": 0.41157746063199613, "grad_norm": 1.4652734994888306, "learning_rate": 4.79348199506543e-05, "loss": 0.5393, "step": 9762 }, { "epoch": 0.41161962181419565, "grad_norm": 0.3611540198326111, "learning_rate": 4.793403916424623e-05, "loss": 0.0635, "step": 9763 }, { "epoch": 0.4116617829963952, "grad_norm": 0.7030071020126343, "learning_rate": 4.793325837783816e-05, "loss": 0.5858, "step": 9764 }, { "epoch": 0.4117039441785948, "grad_norm": 11.881775856018066, "learning_rate": 4.793247759143009e-05, "loss": 0.0961, "step": 9765 }, { "epoch": 0.4117461053607943, "grad_norm": 1.2688360214233398, "learning_rate": 4.793169680502202e-05, "loss": 0.1204, "step": 9766 }, { "epoch": 0.4117882665429939, "grad_norm": 0.6656021475791931, "learning_rate": 4.793091601861395e-05, "loss": 0.1684, "step": 9767 }, { "epoch": 0.4118304277251934, "grad_norm": 4.34965181350708, "learning_rate": 4.793013523220588e-05, "loss": 1.1974, "step": 9768 }, { "epoch": 0.411872588907393, "grad_norm": 0.9802863597869873, "learning_rate": 4.792935444579781e-05, "loss": 0.1564, "step": 9769 }, { "epoch": 0.4119147500895925, "grad_norm": 0.9000479578971863, "learning_rate": 4.792857365938974e-05, "loss": 0.1886, "step": 9770 }, { "epoch": 0.41195691127179207, "grad_norm": 0.9735577702522278, "learning_rate": 4.792779287298167e-05, "loss": 0.3708, "step": 9771 }, { "epoch": 0.4119990724539916, "grad_norm": 1.4899210929870605, "learning_rate": 4.7927012086573594e-05, "loss": 0.2047, "step": 9772 }, { "epoch": 0.41204123363619116, "grad_norm": 1.5412381887435913, "learning_rate": 4.792623130016553e-05, "loss": 0.938, "step": 9773 }, { "epoch": 0.4120833948183907, "grad_norm": 4.6321635246276855, "learning_rate": 4.792545051375746e-05, "loss": 1.1322, "step": 9774 }, { "epoch": 0.41212555600059025, "grad_norm": 1.249016523361206, "learning_rate": 4.792466972734939e-05, "loss": 0.6091, "step": 9775 }, { "epoch": 0.41216771718278983, "grad_norm": 0.5222232937812805, "learning_rate": 4.792388894094132e-05, "loss": 0.1129, "step": 9776 }, { "epoch": 0.41220987836498935, "grad_norm": 1.8432080745697021, "learning_rate": 4.792310815453325e-05, "loss": 0.6405, "step": 9777 }, { "epoch": 0.4122520395471889, "grad_norm": 0.6739734411239624, "learning_rate": 4.792232736812518e-05, "loss": 0.1667, "step": 9778 }, { "epoch": 0.41229420072938844, "grad_norm": 0.6698874831199646, "learning_rate": 4.7921546581717106e-05, "loss": 0.1332, "step": 9779 }, { "epoch": 0.412336361911588, "grad_norm": 1.086025595664978, "learning_rate": 4.792076579530904e-05, "loss": 0.6577, "step": 9780 }, { "epoch": 0.41237852309378753, "grad_norm": 2.151482105255127, "learning_rate": 4.791998500890097e-05, "loss": 0.527, "step": 9781 }, { "epoch": 0.4124206842759871, "grad_norm": 2.5475826263427734, "learning_rate": 4.79192042224929e-05, "loss": 0.7251, "step": 9782 }, { "epoch": 0.4124628454581866, "grad_norm": 0.7702701687812805, "learning_rate": 4.7918423436084824e-05, "loss": 0.4871, "step": 9783 }, { "epoch": 0.4125050066403862, "grad_norm": 1.8516435623168945, "learning_rate": 4.7917642649676756e-05, "loss": 0.3371, "step": 9784 }, { "epoch": 0.41254716782258577, "grad_norm": 1.0776859521865845, "learning_rate": 4.791686186326869e-05, "loss": 0.58, "step": 9785 }, { "epoch": 0.4125893290047853, "grad_norm": 0.9580838680267334, "learning_rate": 4.791608107686061e-05, "loss": 0.1737, "step": 9786 }, { "epoch": 0.41263149018698486, "grad_norm": 0.2882930040359497, "learning_rate": 4.791530029045255e-05, "loss": 0.0856, "step": 9787 }, { "epoch": 0.4126736513691844, "grad_norm": 2.7199864387512207, "learning_rate": 4.7914519504044474e-05, "loss": 0.9523, "step": 9788 }, { "epoch": 0.41271581255138395, "grad_norm": 0.8092933893203735, "learning_rate": 4.7913738717636405e-05, "loss": 0.3321, "step": 9789 }, { "epoch": 0.41275797373358347, "grad_norm": 0.7677580118179321, "learning_rate": 4.7912957931228337e-05, "loss": 0.5964, "step": 9790 }, { "epoch": 0.41280013491578305, "grad_norm": 1.2186179161071777, "learning_rate": 4.791217714482026e-05, "loss": 0.2127, "step": 9791 }, { "epoch": 0.41284229609798256, "grad_norm": 0.8396876454353333, "learning_rate": 4.79113963584122e-05, "loss": 0.3802, "step": 9792 }, { "epoch": 0.41288445728018214, "grad_norm": 0.5354171991348267, "learning_rate": 4.7910615572004124e-05, "loss": 0.1359, "step": 9793 }, { "epoch": 0.4129266184623817, "grad_norm": 0.9057239890098572, "learning_rate": 4.7909834785596055e-05, "loss": 0.2685, "step": 9794 }, { "epoch": 0.41296877964458123, "grad_norm": 3.394230365753174, "learning_rate": 4.7909053999187986e-05, "loss": 0.5889, "step": 9795 }, { "epoch": 0.4130109408267808, "grad_norm": 1.080898642539978, "learning_rate": 4.790827321277991e-05, "loss": 0.6728, "step": 9796 }, { "epoch": 0.4130531020089803, "grad_norm": 0.9214401245117188, "learning_rate": 4.790749242637184e-05, "loss": 0.0976, "step": 9797 }, { "epoch": 0.4130952631911799, "grad_norm": 0.660959005355835, "learning_rate": 4.790671163996377e-05, "loss": 0.2466, "step": 9798 }, { "epoch": 0.4131374243733794, "grad_norm": 2.407799243927002, "learning_rate": 4.7905930853555704e-05, "loss": 0.5272, "step": 9799 }, { "epoch": 0.413179585555579, "grad_norm": 1.4507533311843872, "learning_rate": 4.7905150067147636e-05, "loss": 0.2784, "step": 9800 }, { "epoch": 0.4132217467377785, "grad_norm": 0.7274466753005981, "learning_rate": 4.790436928073957e-05, "loss": 0.1528, "step": 9801 }, { "epoch": 0.4132639079199781, "grad_norm": 0.8027195334434509, "learning_rate": 4.790358849433149e-05, "loss": 0.4495, "step": 9802 }, { "epoch": 0.4133060691021776, "grad_norm": 0.5311644077301025, "learning_rate": 4.790280770792342e-05, "loss": 0.1988, "step": 9803 }, { "epoch": 0.41334823028437717, "grad_norm": 0.8651569485664368, "learning_rate": 4.7902026921515354e-05, "loss": 0.6608, "step": 9804 }, { "epoch": 0.41339039146657675, "grad_norm": 0.7960143089294434, "learning_rate": 4.790124613510728e-05, "loss": 0.4386, "step": 9805 }, { "epoch": 0.41343255264877626, "grad_norm": 0.4448806643486023, "learning_rate": 4.7900465348699216e-05, "loss": 0.0662, "step": 9806 }, { "epoch": 0.41347471383097584, "grad_norm": 0.8535088300704956, "learning_rate": 4.789968456229114e-05, "loss": 0.1443, "step": 9807 }, { "epoch": 0.41351687501317536, "grad_norm": 0.4394434690475464, "learning_rate": 4.789890377588307e-05, "loss": 0.1254, "step": 9808 }, { "epoch": 0.41355903619537493, "grad_norm": 0.9134693741798401, "learning_rate": 4.7898122989475e-05, "loss": 0.5426, "step": 9809 }, { "epoch": 0.41360119737757445, "grad_norm": 2.2537660598754883, "learning_rate": 4.789734220306693e-05, "loss": 0.6511, "step": 9810 }, { "epoch": 0.413643358559774, "grad_norm": 0.6141820549964905, "learning_rate": 4.7896561416658866e-05, "loss": 0.091, "step": 9811 }, { "epoch": 0.41368551974197354, "grad_norm": 3.9131152629852295, "learning_rate": 4.789578063025079e-05, "loss": 0.575, "step": 9812 }, { "epoch": 0.4137276809241731, "grad_norm": 3.614152669906616, "learning_rate": 4.789499984384272e-05, "loss": 0.3814, "step": 9813 }, { "epoch": 0.4137698421063727, "grad_norm": 3.5813984870910645, "learning_rate": 4.789421905743465e-05, "loss": 0.6383, "step": 9814 }, { "epoch": 0.4138120032885722, "grad_norm": 2.592475414276123, "learning_rate": 4.789343827102658e-05, "loss": 0.5736, "step": 9815 }, { "epoch": 0.4138541644707718, "grad_norm": 0.9771522879600525, "learning_rate": 4.789265748461851e-05, "loss": 0.2054, "step": 9816 }, { "epoch": 0.4138963256529713, "grad_norm": 1.1031135320663452, "learning_rate": 4.789187669821044e-05, "loss": 0.1645, "step": 9817 }, { "epoch": 0.41393848683517087, "grad_norm": 1.1520544290542603, "learning_rate": 4.789109591180237e-05, "loss": 0.0899, "step": 9818 }, { "epoch": 0.4139806480173704, "grad_norm": 0.6584717631340027, "learning_rate": 4.7890315125394296e-05, "loss": 0.1979, "step": 9819 }, { "epoch": 0.41402280919956996, "grad_norm": 1.0231354236602783, "learning_rate": 4.7889534338986234e-05, "loss": 0.2298, "step": 9820 }, { "epoch": 0.4140649703817695, "grad_norm": 0.8590418696403503, "learning_rate": 4.788875355257816e-05, "loss": 0.1641, "step": 9821 }, { "epoch": 0.41410713156396906, "grad_norm": 0.4922982454299927, "learning_rate": 4.788797276617009e-05, "loss": 0.0942, "step": 9822 }, { "epoch": 0.41414929274616863, "grad_norm": 1.5737296342849731, "learning_rate": 4.788719197976202e-05, "loss": 0.9312, "step": 9823 }, { "epoch": 0.41419145392836815, "grad_norm": 0.4567042887210846, "learning_rate": 4.7886411193353945e-05, "loss": 0.1526, "step": 9824 }, { "epoch": 0.4142336151105677, "grad_norm": 2.055823802947998, "learning_rate": 4.788563040694588e-05, "loss": 0.6198, "step": 9825 }, { "epoch": 0.41427577629276724, "grad_norm": 0.7863750457763672, "learning_rate": 4.788484962053781e-05, "loss": 0.1879, "step": 9826 }, { "epoch": 0.4143179374749668, "grad_norm": 0.8553062677383423, "learning_rate": 4.788406883412974e-05, "loss": 0.1697, "step": 9827 }, { "epoch": 0.41436009865716633, "grad_norm": 0.8110292553901672, "learning_rate": 4.788328804772167e-05, "loss": 0.3115, "step": 9828 }, { "epoch": 0.4144022598393659, "grad_norm": 2.5213215351104736, "learning_rate": 4.7882507261313595e-05, "loss": 0.7169, "step": 9829 }, { "epoch": 0.4144444210215654, "grad_norm": 0.46613919734954834, "learning_rate": 4.7881726474905526e-05, "loss": 0.1171, "step": 9830 }, { "epoch": 0.414486582203765, "grad_norm": 1.402218222618103, "learning_rate": 4.788094568849746e-05, "loss": 0.2666, "step": 9831 }, { "epoch": 0.4145287433859645, "grad_norm": 0.43151065707206726, "learning_rate": 4.788016490208939e-05, "loss": 0.087, "step": 9832 }, { "epoch": 0.4145709045681641, "grad_norm": 1.232485294342041, "learning_rate": 4.787938411568131e-05, "loss": 0.3312, "step": 9833 }, { "epoch": 0.41461306575036366, "grad_norm": 2.0204832553863525, "learning_rate": 4.7878603329273244e-05, "loss": 0.5672, "step": 9834 }, { "epoch": 0.4146552269325632, "grad_norm": 0.5375451445579529, "learning_rate": 4.7877822542865175e-05, "loss": 0.1769, "step": 9835 }, { "epoch": 0.41469738811476275, "grad_norm": 1.239392876625061, "learning_rate": 4.787704175645711e-05, "loss": 0.7495, "step": 9836 }, { "epoch": 0.4147395492969623, "grad_norm": 0.7279128432273865, "learning_rate": 4.787626097004904e-05, "loss": 0.1462, "step": 9837 }, { "epoch": 0.41478171047916185, "grad_norm": 0.876617968082428, "learning_rate": 4.787548018364096e-05, "loss": 0.5562, "step": 9838 }, { "epoch": 0.41482387166136137, "grad_norm": 0.6977629661560059, "learning_rate": 4.78746993972329e-05, "loss": 0.1739, "step": 9839 }, { "epoch": 0.41486603284356094, "grad_norm": 0.8395140171051025, "learning_rate": 4.7873918610824825e-05, "loss": 0.1627, "step": 9840 }, { "epoch": 0.41490819402576046, "grad_norm": 0.7518528699874878, "learning_rate": 4.787313782441675e-05, "loss": 0.2705, "step": 9841 }, { "epoch": 0.41495035520796003, "grad_norm": 72372.9296875, "learning_rate": 4.787235703800869e-05, "loss": 0.1636, "step": 9842 }, { "epoch": 0.4149925163901596, "grad_norm": 1.1028035879135132, "learning_rate": 4.787157625160061e-05, "loss": 0.6988, "step": 9843 }, { "epoch": 0.4150346775723591, "grad_norm": 2.131398916244507, "learning_rate": 4.787079546519254e-05, "loss": 0.6963, "step": 9844 }, { "epoch": 0.4150768387545587, "grad_norm": 3.3718550205230713, "learning_rate": 4.7870014678784474e-05, "loss": 0.9932, "step": 9845 }, { "epoch": 0.4151189999367582, "grad_norm": 20.09994125366211, "learning_rate": 4.7869233892376406e-05, "loss": 0.2372, "step": 9846 }, { "epoch": 0.4151611611189578, "grad_norm": 1.7791107892990112, "learning_rate": 4.786845310596834e-05, "loss": 0.6582, "step": 9847 }, { "epoch": 0.4152033223011573, "grad_norm": 4.502616882324219, "learning_rate": 4.786767231956026e-05, "loss": 0.9751, "step": 9848 }, { "epoch": 0.4152454834833569, "grad_norm": 0.9930121898651123, "learning_rate": 4.786689153315219e-05, "loss": 0.4851, "step": 9849 }, { "epoch": 0.4152876446655564, "grad_norm": 2.54815936088562, "learning_rate": 4.7866110746744124e-05, "loss": 0.3332, "step": 9850 }, { "epoch": 0.415329805847756, "grad_norm": 0.6265565156936646, "learning_rate": 4.7865329960336055e-05, "loss": 0.2447, "step": 9851 }, { "epoch": 0.41537196702995555, "grad_norm": 0.7092563509941101, "learning_rate": 4.786454917392798e-05, "loss": 0.1183, "step": 9852 }, { "epoch": 0.41541412821215506, "grad_norm": 0.9871340990066528, "learning_rate": 4.786376838751991e-05, "loss": 0.2784, "step": 9853 }, { "epoch": 0.41545628939435464, "grad_norm": 0.4155878722667694, "learning_rate": 4.786298760111184e-05, "loss": 0.1175, "step": 9854 }, { "epoch": 0.41549845057655416, "grad_norm": 0.6965272426605225, "learning_rate": 4.786220681470377e-05, "loss": 0.0839, "step": 9855 }, { "epoch": 0.41554061175875373, "grad_norm": 0.8389132022857666, "learning_rate": 4.7861426028295705e-05, "loss": 0.2089, "step": 9856 }, { "epoch": 0.41558277294095325, "grad_norm": 1.4799022674560547, "learning_rate": 4.786064524188763e-05, "loss": 0.2391, "step": 9857 }, { "epoch": 0.4156249341231528, "grad_norm": 0.5132981538772583, "learning_rate": 4.785986445547956e-05, "loss": 0.1207, "step": 9858 }, { "epoch": 0.41566709530535234, "grad_norm": 2.367867946624756, "learning_rate": 4.785908366907149e-05, "loss": 1.0817, "step": 9859 }, { "epoch": 0.4157092564875519, "grad_norm": 2.4131968021392822, "learning_rate": 4.7858302882663416e-05, "loss": 0.4287, "step": 9860 }, { "epoch": 0.41575141766975143, "grad_norm": 2.6823244094848633, "learning_rate": 4.7857522096255354e-05, "loss": 0.8218, "step": 9861 }, { "epoch": 0.415793578851951, "grad_norm": 2.4931528568267822, "learning_rate": 4.785674130984728e-05, "loss": 0.2489, "step": 9862 }, { "epoch": 0.4158357400341506, "grad_norm": 1.1916406154632568, "learning_rate": 4.785596052343921e-05, "loss": 0.1003, "step": 9863 }, { "epoch": 0.4158779012163501, "grad_norm": 0.8549274802207947, "learning_rate": 4.785517973703114e-05, "loss": 1.0536, "step": 9864 }, { "epoch": 0.4159200623985497, "grad_norm": 1.1450767517089844, "learning_rate": 4.785439895062307e-05, "loss": 0.7742, "step": 9865 }, { "epoch": 0.4159622235807492, "grad_norm": 0.821527898311615, "learning_rate": 4.7853618164215e-05, "loss": 0.5754, "step": 9866 }, { "epoch": 0.41600438476294876, "grad_norm": 1.14577317237854, "learning_rate": 4.785283737780693e-05, "loss": 0.6213, "step": 9867 }, { "epoch": 0.4160465459451483, "grad_norm": 0.9263027906417847, "learning_rate": 4.785205659139886e-05, "loss": 0.5062, "step": 9868 }, { "epoch": 0.41608870712734786, "grad_norm": 30.7006893157959, "learning_rate": 4.785127580499079e-05, "loss": 0.6433, "step": 9869 }, { "epoch": 0.4161308683095474, "grad_norm": 1.1800743341445923, "learning_rate": 4.785049501858272e-05, "loss": 0.3032, "step": 9870 }, { "epoch": 0.41617302949174695, "grad_norm": 1.555530309677124, "learning_rate": 4.7849714232174647e-05, "loss": 0.6284, "step": 9871 }, { "epoch": 0.4162151906739465, "grad_norm": 0.6342523097991943, "learning_rate": 4.784893344576658e-05, "loss": 0.3084, "step": 9872 }, { "epoch": 0.41625735185614604, "grad_norm": 1.27448308467865, "learning_rate": 4.784815265935851e-05, "loss": 0.1207, "step": 9873 }, { "epoch": 0.4162995130383456, "grad_norm": 0.5976374745368958, "learning_rate": 4.7847371872950434e-05, "loss": 0.3795, "step": 9874 }, { "epoch": 0.41634167422054513, "grad_norm": 0.8663005828857422, "learning_rate": 4.784659108654237e-05, "loss": 0.1159, "step": 9875 }, { "epoch": 0.4163838354027447, "grad_norm": 0.6488385796546936, "learning_rate": 4.7845810300134296e-05, "loss": 0.0833, "step": 9876 }, { "epoch": 0.4164259965849442, "grad_norm": 1.058720350265503, "learning_rate": 4.784502951372623e-05, "loss": 0.379, "step": 9877 }, { "epoch": 0.4164681577671438, "grad_norm": 0.8697371482849121, "learning_rate": 4.784424872731816e-05, "loss": 0.1907, "step": 9878 }, { "epoch": 0.4165103189493433, "grad_norm": 0.6506315469741821, "learning_rate": 4.784346794091008e-05, "loss": 0.6208, "step": 9879 }, { "epoch": 0.4165524801315429, "grad_norm": 0.9593769907951355, "learning_rate": 4.784268715450202e-05, "loss": 0.3461, "step": 9880 }, { "epoch": 0.41659464131374246, "grad_norm": 1.7167764902114868, "learning_rate": 4.7841906368093946e-05, "loss": 0.4648, "step": 9881 }, { "epoch": 0.416636802495942, "grad_norm": 2.3306679725646973, "learning_rate": 4.784112558168588e-05, "loss": 0.1378, "step": 9882 }, { "epoch": 0.41667896367814156, "grad_norm": 0.5346845388412476, "learning_rate": 4.784034479527781e-05, "loss": 0.0881, "step": 9883 }, { "epoch": 0.4167211248603411, "grad_norm": 1.1896002292633057, "learning_rate": 4.783956400886974e-05, "loss": 0.3596, "step": 9884 }, { "epoch": 0.41676328604254065, "grad_norm": 25.54786491394043, "learning_rate": 4.7838783222461664e-05, "loss": 0.2195, "step": 9885 }, { "epoch": 0.41680544722474017, "grad_norm": 0.3880596160888672, "learning_rate": 4.7838002436053595e-05, "loss": 0.1304, "step": 9886 }, { "epoch": 0.41684760840693974, "grad_norm": 2.2393364906311035, "learning_rate": 4.7837221649645526e-05, "loss": 0.5128, "step": 9887 }, { "epoch": 0.41688976958913926, "grad_norm": 1.3939906358718872, "learning_rate": 4.783644086323745e-05, "loss": 0.9448, "step": 9888 }, { "epoch": 0.41693193077133883, "grad_norm": 0.985470175743103, "learning_rate": 4.783566007682939e-05, "loss": 0.6756, "step": 9889 }, { "epoch": 0.41697409195353835, "grad_norm": 21.936071395874023, "learning_rate": 4.7834879290421313e-05, "loss": 0.1119, "step": 9890 }, { "epoch": 0.4170162531357379, "grad_norm": 1.2725023031234741, "learning_rate": 4.7834098504013245e-05, "loss": 0.1848, "step": 9891 }, { "epoch": 0.4170584143179375, "grad_norm": 2.0165562629699707, "learning_rate": 4.7833317717605176e-05, "loss": 0.2149, "step": 9892 }, { "epoch": 0.417100575500137, "grad_norm": 1.4525264501571655, "learning_rate": 4.78325369311971e-05, "loss": 0.1908, "step": 9893 }, { "epoch": 0.4171427366823366, "grad_norm": 3.833364248275757, "learning_rate": 4.783175614478904e-05, "loss": 0.4996, "step": 9894 }, { "epoch": 0.4171848978645361, "grad_norm": 1.030060887336731, "learning_rate": 4.783097535838096e-05, "loss": 0.2077, "step": 9895 }, { "epoch": 0.4172270590467357, "grad_norm": 0.4734576642513275, "learning_rate": 4.7830194571972894e-05, "loss": 0.0682, "step": 9896 }, { "epoch": 0.4172692202289352, "grad_norm": 1.0336263179779053, "learning_rate": 4.7829413785564825e-05, "loss": 0.3524, "step": 9897 }, { "epoch": 0.4173113814111348, "grad_norm": 0.6203151345252991, "learning_rate": 4.782863299915675e-05, "loss": 0.1999, "step": 9898 }, { "epoch": 0.4173535425933343, "grad_norm": 3.3091793060302734, "learning_rate": 4.782785221274868e-05, "loss": 0.7402, "step": 9899 }, { "epoch": 0.41739570377553387, "grad_norm": 0.5703010559082031, "learning_rate": 4.782707142634061e-05, "loss": 0.2633, "step": 9900 }, { "epoch": 0.41743786495773344, "grad_norm": 0.30545127391815186, "learning_rate": 4.7826290639932544e-05, "loss": 0.0985, "step": 9901 }, { "epoch": 0.41748002613993296, "grad_norm": 1.0732722282409668, "learning_rate": 4.782550985352447e-05, "loss": 0.3599, "step": 9902 }, { "epoch": 0.41752218732213253, "grad_norm": 1.364159345626831, "learning_rate": 4.78247290671164e-05, "loss": 0.3703, "step": 9903 }, { "epoch": 0.41756434850433205, "grad_norm": 0.4388540983200073, "learning_rate": 4.782394828070833e-05, "loss": 0.1108, "step": 9904 }, { "epoch": 0.4176065096865316, "grad_norm": 3.0471324920654297, "learning_rate": 4.782316749430026e-05, "loss": 0.6584, "step": 9905 }, { "epoch": 0.41764867086873114, "grad_norm": 2.2497236728668213, "learning_rate": 4.782238670789219e-05, "loss": 0.9335, "step": 9906 }, { "epoch": 0.4176908320509307, "grad_norm": 3.68237566947937, "learning_rate": 4.782160592148412e-05, "loss": 0.5208, "step": 9907 }, { "epoch": 0.41773299323313023, "grad_norm": 0.4978249967098236, "learning_rate": 4.7820825135076056e-05, "loss": 0.1092, "step": 9908 }, { "epoch": 0.4177751544153298, "grad_norm": 0.5042338967323303, "learning_rate": 4.782004434866798e-05, "loss": 0.088, "step": 9909 }, { "epoch": 0.4178173155975294, "grad_norm": 0.7128061056137085, "learning_rate": 4.781926356225991e-05, "loss": 0.1268, "step": 9910 }, { "epoch": 0.4178594767797289, "grad_norm": 2.0960822105407715, "learning_rate": 4.781848277585184e-05, "loss": 0.427, "step": 9911 }, { "epoch": 0.4179016379619285, "grad_norm": 1.5232584476470947, "learning_rate": 4.781770198944377e-05, "loss": 0.2629, "step": 9912 }, { "epoch": 0.417943799144128, "grad_norm": 0.5985131859779358, "learning_rate": 4.78169212030357e-05, "loss": 0.1056, "step": 9913 }, { "epoch": 0.41798596032632757, "grad_norm": 2.783202886581421, "learning_rate": 4.781614041662763e-05, "loss": 0.869, "step": 9914 }, { "epoch": 0.4180281215085271, "grad_norm": 0.9830760359764099, "learning_rate": 4.781535963021956e-05, "loss": 0.6427, "step": 9915 }, { "epoch": 0.41807028269072666, "grad_norm": 12.883888244628906, "learning_rate": 4.781457884381149e-05, "loss": 0.149, "step": 9916 }, { "epoch": 0.4181124438729262, "grad_norm": 2.377202272415161, "learning_rate": 4.781379805740342e-05, "loss": 0.6144, "step": 9917 }, { "epoch": 0.41815460505512575, "grad_norm": 3.164109945297241, "learning_rate": 4.781301727099535e-05, "loss": 0.7074, "step": 9918 }, { "epoch": 0.41819676623732527, "grad_norm": 1.9895387887954712, "learning_rate": 4.781223648458728e-05, "loss": 0.4693, "step": 9919 }, { "epoch": 0.41823892741952484, "grad_norm": 0.9431159496307373, "learning_rate": 4.781145569817921e-05, "loss": 0.2092, "step": 9920 }, { "epoch": 0.4182810886017244, "grad_norm": 1.3936402797698975, "learning_rate": 4.7810674911771135e-05, "loss": 0.3302, "step": 9921 }, { "epoch": 0.41832324978392393, "grad_norm": 0.25022730231285095, "learning_rate": 4.7809894125363066e-05, "loss": 0.0512, "step": 9922 }, { "epoch": 0.4183654109661235, "grad_norm": 2.2044613361358643, "learning_rate": 4.7809113338955e-05, "loss": 0.4791, "step": 9923 }, { "epoch": 0.418407572148323, "grad_norm": 0.7857416868209839, "learning_rate": 4.780833255254692e-05, "loss": 0.1782, "step": 9924 }, { "epoch": 0.4184497333305226, "grad_norm": 1.0052942037582397, "learning_rate": 4.780755176613886e-05, "loss": 0.1033, "step": 9925 }, { "epoch": 0.4184918945127221, "grad_norm": 3.4263226985931396, "learning_rate": 4.7806770979730785e-05, "loss": 0.9623, "step": 9926 }, { "epoch": 0.4185340556949217, "grad_norm": 0.7299858927726746, "learning_rate": 4.780599019332272e-05, "loss": 0.1831, "step": 9927 }, { "epoch": 0.4185762168771212, "grad_norm": 0.5473616719245911, "learning_rate": 4.780520940691465e-05, "loss": 0.3298, "step": 9928 }, { "epoch": 0.4186183780593208, "grad_norm": 0.9890840649604797, "learning_rate": 4.780442862050658e-05, "loss": 0.4619, "step": 9929 }, { "epoch": 0.41866053924152036, "grad_norm": 0.8666648864746094, "learning_rate": 4.780364783409851e-05, "loss": 0.1083, "step": 9930 }, { "epoch": 0.4187027004237199, "grad_norm": 0.7569690346717834, "learning_rate": 4.7802867047690434e-05, "loss": 0.17, "step": 9931 }, { "epoch": 0.41874486160591945, "grad_norm": 0.4667508900165558, "learning_rate": 4.7802086261282365e-05, "loss": 0.0651, "step": 9932 }, { "epoch": 0.41878702278811897, "grad_norm": 5.088338375091553, "learning_rate": 4.78013054748743e-05, "loss": 1.1221, "step": 9933 }, { "epoch": 0.41882918397031854, "grad_norm": 1.1242461204528809, "learning_rate": 4.780052468846623e-05, "loss": 0.9937, "step": 9934 }, { "epoch": 0.41887134515251806, "grad_norm": 1.5696732997894287, "learning_rate": 4.779974390205815e-05, "loss": 0.2255, "step": 9935 }, { "epoch": 0.41891350633471763, "grad_norm": 1.169700026512146, "learning_rate": 4.7798963115650084e-05, "loss": 0.1498, "step": 9936 }, { "epoch": 0.41895566751691715, "grad_norm": 0.4543772339820862, "learning_rate": 4.7798182329242015e-05, "loss": 0.1218, "step": 9937 }, { "epoch": 0.4189978286991167, "grad_norm": 0.8940675258636475, "learning_rate": 4.7797401542833946e-05, "loss": 0.1608, "step": 9938 }, { "epoch": 0.4190399898813163, "grad_norm": 0.9774407744407654, "learning_rate": 4.779662075642588e-05, "loss": 0.3694, "step": 9939 }, { "epoch": 0.4190821510635158, "grad_norm": 0.7427220940589905, "learning_rate": 4.77958399700178e-05, "loss": 0.1282, "step": 9940 }, { "epoch": 0.4191243122457154, "grad_norm": 0.612892746925354, "learning_rate": 4.779505918360973e-05, "loss": 0.1094, "step": 9941 }, { "epoch": 0.4191664734279149, "grad_norm": 1.002885341644287, "learning_rate": 4.7794278397201664e-05, "loss": 0.1167, "step": 9942 }, { "epoch": 0.4192086346101145, "grad_norm": 0.5945339798927307, "learning_rate": 4.779349761079359e-05, "loss": 0.1398, "step": 9943 }, { "epoch": 0.419250795792314, "grad_norm": 3.1477420330047607, "learning_rate": 4.779271682438553e-05, "loss": 0.9255, "step": 9944 }, { "epoch": 0.4192929569745136, "grad_norm": 1.2777692079544067, "learning_rate": 4.779193603797745e-05, "loss": 0.1861, "step": 9945 }, { "epoch": 0.4193351181567131, "grad_norm": 0.6481615900993347, "learning_rate": 4.779115525156938e-05, "loss": 0.1642, "step": 9946 }, { "epoch": 0.41937727933891267, "grad_norm": 0.9040662050247192, "learning_rate": 4.7790374465161314e-05, "loss": 0.1802, "step": 9947 }, { "epoch": 0.4194194405211122, "grad_norm": 1.2667738199234009, "learning_rate": 4.778959367875324e-05, "loss": 0.1189, "step": 9948 }, { "epoch": 0.41946160170331176, "grad_norm": 0.6390776038169861, "learning_rate": 4.7788812892345176e-05, "loss": 0.2377, "step": 9949 }, { "epoch": 0.41950376288551133, "grad_norm": 0.805095374584198, "learning_rate": 4.77880321059371e-05, "loss": 0.1525, "step": 9950 }, { "epoch": 0.41954592406771085, "grad_norm": 0.38297751545906067, "learning_rate": 4.778725131952903e-05, "loss": 0.0872, "step": 9951 }, { "epoch": 0.4195880852499104, "grad_norm": 0.8033165335655212, "learning_rate": 4.7786470533120963e-05, "loss": 0.1702, "step": 9952 }, { "epoch": 0.41963024643210994, "grad_norm": 0.43385228514671326, "learning_rate": 4.7785689746712895e-05, "loss": 0.0641, "step": 9953 }, { "epoch": 0.4196724076143095, "grad_norm": 1.9681248664855957, "learning_rate": 4.778490896030482e-05, "loss": 0.4618, "step": 9954 }, { "epoch": 0.41971456879650904, "grad_norm": 0.5745772123336792, "learning_rate": 4.778412817389675e-05, "loss": 0.1789, "step": 9955 }, { "epoch": 0.4197567299787086, "grad_norm": 1.7549208402633667, "learning_rate": 4.778334738748868e-05, "loss": 0.2339, "step": 9956 }, { "epoch": 0.4197988911609081, "grad_norm": 0.46877923607826233, "learning_rate": 4.7782566601080606e-05, "loss": 0.2491, "step": 9957 }, { "epoch": 0.4198410523431077, "grad_norm": 1.1925581693649292, "learning_rate": 4.7781785814672544e-05, "loss": 0.1287, "step": 9958 }, { "epoch": 0.4198832135253073, "grad_norm": 2.164257287979126, "learning_rate": 4.778100502826447e-05, "loss": 0.7525, "step": 9959 }, { "epoch": 0.4199253747075068, "grad_norm": 0.5603511333465576, "learning_rate": 4.77802242418564e-05, "loss": 0.1461, "step": 9960 }, { "epoch": 0.41996753588970637, "grad_norm": 15.72542667388916, "learning_rate": 4.777944345544833e-05, "loss": 0.0964, "step": 9961 }, { "epoch": 0.4200096970719059, "grad_norm": 0.7154683470726013, "learning_rate": 4.7778662669040256e-05, "loss": 0.1635, "step": 9962 }, { "epoch": 0.42005185825410546, "grad_norm": 1.1002233028411865, "learning_rate": 4.7777881882632194e-05, "loss": 0.2261, "step": 9963 }, { "epoch": 0.420094019436305, "grad_norm": 0.8016431927680969, "learning_rate": 4.777710109622412e-05, "loss": 0.1953, "step": 9964 }, { "epoch": 0.42013618061850455, "grad_norm": 5.0830512046813965, "learning_rate": 4.777632030981605e-05, "loss": 0.5664, "step": 9965 }, { "epoch": 0.42017834180070407, "grad_norm": 1.8152375221252441, "learning_rate": 4.777553952340798e-05, "loss": 0.4877, "step": 9966 }, { "epoch": 0.42022050298290364, "grad_norm": 0.848619282245636, "learning_rate": 4.7774758736999905e-05, "loss": 0.2268, "step": 9967 }, { "epoch": 0.4202626641651032, "grad_norm": 0.8938298225402832, "learning_rate": 4.7773977950591836e-05, "loss": 0.1885, "step": 9968 }, { "epoch": 0.42030482534730274, "grad_norm": 1.1670303344726562, "learning_rate": 4.777319716418377e-05, "loss": 0.3691, "step": 9969 }, { "epoch": 0.4203469865295023, "grad_norm": 2.0215282440185547, "learning_rate": 4.77724163777757e-05, "loss": 0.3439, "step": 9970 }, { "epoch": 0.4203891477117018, "grad_norm": 46.020660400390625, "learning_rate": 4.7771635591367624e-05, "loss": 0.2122, "step": 9971 }, { "epoch": 0.4204313088939014, "grad_norm": 0.7373962998390198, "learning_rate": 4.777085480495956e-05, "loss": 0.3408, "step": 9972 }, { "epoch": 0.4204734700761009, "grad_norm": 0.6826852560043335, "learning_rate": 4.7770074018551486e-05, "loss": 0.2104, "step": 9973 }, { "epoch": 0.4205156312583005, "grad_norm": 181.13442993164062, "learning_rate": 4.776929323214342e-05, "loss": 0.2513, "step": 9974 }, { "epoch": 0.4205577924405, "grad_norm": 224.92095947265625, "learning_rate": 4.776851244573535e-05, "loss": 0.4364, "step": 9975 }, { "epoch": 0.4205999536226996, "grad_norm": 0.9695190787315369, "learning_rate": 4.776773165932727e-05, "loss": 0.5721, "step": 9976 }, { "epoch": 0.4206421148048991, "grad_norm": 41.058658599853516, "learning_rate": 4.776695087291921e-05, "loss": 0.7394, "step": 9977 }, { "epoch": 0.4206842759870987, "grad_norm": 14305.5576171875, "learning_rate": 4.7766170086511136e-05, "loss": 0.5765, "step": 9978 }, { "epoch": 0.42072643716929825, "grad_norm": 113.54920196533203, "learning_rate": 4.776538930010307e-05, "loss": 0.1302, "step": 9979 }, { "epoch": 0.42076859835149777, "grad_norm": 49.94491958618164, "learning_rate": 4.7764608513695e-05, "loss": 0.616, "step": 9980 }, { "epoch": 0.42081075953369734, "grad_norm": 7.061598300933838, "learning_rate": 4.776382772728692e-05, "loss": 0.8013, "step": 9981 }, { "epoch": 0.42085292071589686, "grad_norm": 3.080820322036743, "learning_rate": 4.7763046940878854e-05, "loss": 0.4783, "step": 9982 }, { "epoch": 0.42089508189809643, "grad_norm": 1.0431352853775024, "learning_rate": 4.7762266154470785e-05, "loss": 0.2003, "step": 9983 }, { "epoch": 0.42093724308029595, "grad_norm": 0.8350643515586853, "learning_rate": 4.7761485368062716e-05, "loss": 0.24, "step": 9984 }, { "epoch": 0.4209794042624955, "grad_norm": 0.9631856083869934, "learning_rate": 4.776070458165465e-05, "loss": 0.3077, "step": 9985 }, { "epoch": 0.42102156544469505, "grad_norm": 0.8078908920288086, "learning_rate": 4.775992379524657e-05, "loss": 0.2221, "step": 9986 }, { "epoch": 0.4210637266268946, "grad_norm": 7.431057929992676, "learning_rate": 4.77591430088385e-05, "loss": 1.1133, "step": 9987 }, { "epoch": 0.4211058878090942, "grad_norm": 85.78736114501953, "learning_rate": 4.7758362222430435e-05, "loss": 0.3276, "step": 9988 }, { "epoch": 0.4211480489912937, "grad_norm": 1.194682240486145, "learning_rate": 4.7757581436022366e-05, "loss": 0.1546, "step": 9989 }, { "epoch": 0.4211902101734933, "grad_norm": 22.414752960205078, "learning_rate": 4.775680064961429e-05, "loss": 0.237, "step": 9990 }, { "epoch": 0.4212323713556928, "grad_norm": 2.77844500541687, "learning_rate": 4.775601986320623e-05, "loss": 0.2066, "step": 9991 }, { "epoch": 0.4212745325378924, "grad_norm": 3.0519022941589355, "learning_rate": 4.775523907679815e-05, "loss": 0.5188, "step": 9992 }, { "epoch": 0.4213166937200919, "grad_norm": 1.4892551898956299, "learning_rate": 4.775445829039008e-05, "loss": 0.2615, "step": 9993 }, { "epoch": 0.42135885490229147, "grad_norm": 2.0688741207122803, "learning_rate": 4.7753677503982015e-05, "loss": 0.1898, "step": 9994 }, { "epoch": 0.421401016084491, "grad_norm": 8.242162704467773, "learning_rate": 4.775289671757394e-05, "loss": 0.2163, "step": 9995 }, { "epoch": 0.42144317726669056, "grad_norm": 6.201334476470947, "learning_rate": 4.775211593116588e-05, "loss": 0.3306, "step": 9996 }, { "epoch": 0.42148533844889013, "grad_norm": 3.955198287963867, "learning_rate": 4.77513351447578e-05, "loss": 0.8242, "step": 9997 }, { "epoch": 0.42152749963108965, "grad_norm": 20.59930992126465, "learning_rate": 4.7750554358349734e-05, "loss": 0.2693, "step": 9998 }, { "epoch": 0.4215696608132892, "grad_norm": 2.8041861057281494, "learning_rate": 4.7749773571941665e-05, "loss": 0.2076, "step": 9999 }, { "epoch": 0.42161182199548874, "grad_norm": 427.5036315917969, "learning_rate": 4.774899278553359e-05, "loss": 0.5076, "step": 10000 }, { "epoch": 0.4216539831776883, "grad_norm": 589.0046997070312, "learning_rate": 4.774821199912552e-05, "loss": 0.6224, "step": 10001 }, { "epoch": 0.42169614435988784, "grad_norm": 6.289798259735107, "learning_rate": 4.774743121271745e-05, "loss": 0.5342, "step": 10002 }, { "epoch": 0.4217383055420874, "grad_norm": 2.76222562789917, "learning_rate": 4.774665042630938e-05, "loss": 0.8492, "step": 10003 }, { "epoch": 0.42178046672428693, "grad_norm": 1019.194580078125, "learning_rate": 4.774586963990131e-05, "loss": 0.6556, "step": 10004 }, { "epoch": 0.4218226279064865, "grad_norm": 998.25634765625, "learning_rate": 4.774508885349324e-05, "loss": 1.6758, "step": 10005 }, { "epoch": 0.421864789088686, "grad_norm": 3357.427490234375, "learning_rate": 4.774430806708517e-05, "loss": 2.3493, "step": 10006 }, { "epoch": 0.4219069502708856, "grad_norm": 23590.48046875, "learning_rate": 4.77435272806771e-05, "loss": 4.0304, "step": 10007 }, { "epoch": 0.42194911145308517, "grad_norm": 12557.62109375, "learning_rate": 4.774274649426903e-05, "loss": 3.9319, "step": 10008 }, { "epoch": 0.4219912726352847, "grad_norm": 21247.7109375, "learning_rate": 4.774196570786096e-05, "loss": 9.2661, "step": 10009 }, { "epoch": 0.42203343381748426, "grad_norm": 47043.80859375, "learning_rate": 4.774118492145289e-05, "loss": 10.8597, "step": 10010 }, { "epoch": 0.4220755949996838, "grad_norm": 118947.40625, "learning_rate": 4.774040413504482e-05, "loss": 14.0784, "step": 10011 }, { "epoch": 0.42211775618188335, "grad_norm": 60279.21484375, "learning_rate": 4.7739623348636744e-05, "loss": 18.3129, "step": 10012 }, { "epoch": 0.42215991736408287, "grad_norm": 267567.75, "learning_rate": 4.773884256222868e-05, "loss": 17.9794, "step": 10013 }, { "epoch": 0.42220207854628244, "grad_norm": 60032.6875, "learning_rate": 4.773806177582061e-05, "loss": 12.5141, "step": 10014 }, { "epoch": 0.42224423972848196, "grad_norm": 55754.10546875, "learning_rate": 4.773728098941254e-05, "loss": 18.3309, "step": 10015 }, { "epoch": 0.42228640091068154, "grad_norm": 32153.296875, "learning_rate": 4.773650020300447e-05, "loss": 11.7545, "step": 10016 }, { "epoch": 0.4223285620928811, "grad_norm": 445708.90625, "learning_rate": 4.77357194165964e-05, "loss": 17.1479, "step": 10017 }, { "epoch": 0.42237072327508063, "grad_norm": 47513.3203125, "learning_rate": 4.773493863018833e-05, "loss": 12.324, "step": 10018 }, { "epoch": 0.4224128844572802, "grad_norm": 66883.5625, "learning_rate": 4.7734157843780256e-05, "loss": 16.9652, "step": 10019 }, { "epoch": 0.4224550456394797, "grad_norm": 250981.78125, "learning_rate": 4.773337705737219e-05, "loss": 28.8838, "step": 10020 }, { "epoch": 0.4224972068216793, "grad_norm": 172909.0, "learning_rate": 4.773259627096412e-05, "loss": 48.3671, "step": 10021 }, { "epoch": 0.4225393680038788, "grad_norm": 260680.71875, "learning_rate": 4.773181548455605e-05, "loss": 51.9218, "step": 10022 }, { "epoch": 0.4225815291860784, "grad_norm": 84043.7265625, "learning_rate": 4.7731034698147974e-05, "loss": 65.8233, "step": 10023 }, { "epoch": 0.4226236903682779, "grad_norm": 411764.875, "learning_rate": 4.7730253911739906e-05, "loss": 66.9569, "step": 10024 }, { "epoch": 0.4226658515504775, "grad_norm": 126655.671875, "learning_rate": 4.772947312533184e-05, "loss": 68.1746, "step": 10025 }, { "epoch": 0.42270801273267705, "grad_norm": 236812.828125, "learning_rate": 4.772869233892376e-05, "loss": 70.2916, "step": 10026 }, { "epoch": 0.42275017391487657, "grad_norm": 256996.046875, "learning_rate": 4.77279115525157e-05, "loss": 69.4757, "step": 10027 }, { "epoch": 0.42279233509707614, "grad_norm": 31088.21484375, "learning_rate": 4.7727130766107624e-05, "loss": 70.8114, "step": 10028 }, { "epoch": 0.42283449627927566, "grad_norm": 11727.8857421875, "learning_rate": 4.7726349979699555e-05, "loss": 67.732, "step": 10029 }, { "epoch": 0.42287665746147524, "grad_norm": 5388.697265625, "learning_rate": 4.7725569193291487e-05, "loss": 69.2659, "step": 10030 }, { "epoch": 0.42291881864367475, "grad_norm": 5386.365234375, "learning_rate": 4.772478840688341e-05, "loss": 67.8617, "step": 10031 }, { "epoch": 0.42296097982587433, "grad_norm": 34300.67578125, "learning_rate": 4.772400762047535e-05, "loss": 67.4985, "step": 10032 }, { "epoch": 0.42300314100807385, "grad_norm": 4950.98876953125, "learning_rate": 4.7723226834067274e-05, "loss": 68.6711, "step": 10033 }, { "epoch": 0.4230453021902734, "grad_norm": 7092.1064453125, "learning_rate": 4.7722446047659205e-05, "loss": 66.01, "step": 10034 }, { "epoch": 0.42308746337247294, "grad_norm": 8366.708984375, "learning_rate": 4.7721665261251136e-05, "loss": 67.7563, "step": 10035 }, { "epoch": 0.4231296245546725, "grad_norm": 9720.3525390625, "learning_rate": 4.772088447484306e-05, "loss": 66.6639, "step": 10036 }, { "epoch": 0.4231717857368721, "grad_norm": 8356.564453125, "learning_rate": 4.772010368843499e-05, "loss": 72.0249, "step": 10037 }, { "epoch": 0.4232139469190716, "grad_norm": 6382.8125, "learning_rate": 4.771932290202692e-05, "loss": 68.8608, "step": 10038 }, { "epoch": 0.4232561081012712, "grad_norm": 10546.32421875, "learning_rate": 4.7718542115618854e-05, "loss": 69.0061, "step": 10039 }, { "epoch": 0.4232982692834707, "grad_norm": 4976.9189453125, "learning_rate": 4.7717761329210786e-05, "loss": 67.2658, "step": 10040 }, { "epoch": 0.42334043046567027, "grad_norm": 13324.9228515625, "learning_rate": 4.771698054280272e-05, "loss": 69.7517, "step": 10041 }, { "epoch": 0.4233825916478698, "grad_norm": 49868.98828125, "learning_rate": 4.771619975639464e-05, "loss": 69.4085, "step": 10042 }, { "epoch": 0.42342475283006936, "grad_norm": 92861.3125, "learning_rate": 4.771541896998657e-05, "loss": 68.1215, "step": 10043 }, { "epoch": 0.4234669140122689, "grad_norm": 111681.140625, "learning_rate": 4.7714638183578504e-05, "loss": 67.2749, "step": 10044 }, { "epoch": 0.42350907519446845, "grad_norm": 308970.71875, "learning_rate": 4.771385739717043e-05, "loss": 69.1446, "step": 10045 }, { "epoch": 0.423551236376668, "grad_norm": 286784.6875, "learning_rate": 4.7713076610762366e-05, "loss": 71.8477, "step": 10046 }, { "epoch": 0.42359339755886755, "grad_norm": 829745.9375, "learning_rate": 4.771229582435429e-05, "loss": 66.6355, "step": 10047 }, { "epoch": 0.4236355587410671, "grad_norm": 297911.46875, "learning_rate": 4.771151503794622e-05, "loss": 66.3397, "step": 10048 }, { "epoch": 0.42367771992326664, "grad_norm": 211370.5, "learning_rate": 4.771073425153815e-05, "loss": 33.4133, "step": 10049 }, { "epoch": 0.4237198811054662, "grad_norm": 347499.4375, "learning_rate": 4.770995346513008e-05, "loss": 34.2751, "step": 10050 }, { "epoch": 0.42376204228766573, "grad_norm": 120906.65625, "learning_rate": 4.770917267872201e-05, "loss": 54.7504, "step": 10051 }, { "epoch": 0.4238042034698653, "grad_norm": 247013.390625, "learning_rate": 4.770839189231394e-05, "loss": 63.7762, "step": 10052 }, { "epoch": 0.4238463646520648, "grad_norm": 478528.0, "learning_rate": 4.770761110590587e-05, "loss": 65.2973, "step": 10053 }, { "epoch": 0.4238885258342644, "grad_norm": 80069.1953125, "learning_rate": 4.77068303194978e-05, "loss": 71.2126, "step": 10054 }, { "epoch": 0.4239306870164639, "grad_norm": 47202.9296875, "learning_rate": 4.770604953308973e-05, "loss": 66.7326, "step": 10055 }, { "epoch": 0.4239728481986635, "grad_norm": 18240.69140625, "learning_rate": 4.770526874668166e-05, "loss": 72.2916, "step": 10056 }, { "epoch": 0.42401500938086306, "grad_norm": 6073.29150390625, "learning_rate": 4.770448796027359e-05, "loss": 65.7302, "step": 10057 }, { "epoch": 0.4240571705630626, "grad_norm": 8672.5439453125, "learning_rate": 4.770370717386552e-05, "loss": 70.9152, "step": 10058 }, { "epoch": 0.42409933174526215, "grad_norm": 13566.166015625, "learning_rate": 4.7702926387457446e-05, "loss": 69.4, "step": 10059 }, { "epoch": 0.42414149292746167, "grad_norm": 15204.876953125, "learning_rate": 4.7702145601049384e-05, "loss": 66.3095, "step": 10060 }, { "epoch": 0.42418365410966125, "grad_norm": 11747.0654296875, "learning_rate": 4.770136481464131e-05, "loss": 69.8144, "step": 10061 }, { "epoch": 0.42422581529186076, "grad_norm": 8771.28125, "learning_rate": 4.770058402823324e-05, "loss": 69.8851, "step": 10062 }, { "epoch": 0.42426797647406034, "grad_norm": 39472.921875, "learning_rate": 4.769980324182517e-05, "loss": 68.2277, "step": 10063 }, { "epoch": 0.42431013765625986, "grad_norm": 7174.34130859375, "learning_rate": 4.7699022455417095e-05, "loss": 71.2862, "step": 10064 }, { "epoch": 0.42435229883845943, "grad_norm": 14266.298828125, "learning_rate": 4.769824166900903e-05, "loss": 68.9601, "step": 10065 }, { "epoch": 0.424394460020659, "grad_norm": 6822.66455078125, "learning_rate": 4.769746088260096e-05, "loss": 66.736, "step": 10066 }, { "epoch": 0.4244366212028585, "grad_norm": 14277.546875, "learning_rate": 4.769668009619289e-05, "loss": 66.8814, "step": 10067 }, { "epoch": 0.4244787823850581, "grad_norm": 21478.84375, "learning_rate": 4.769589930978482e-05, "loss": 65.6748, "step": 10068 }, { "epoch": 0.4245209435672576, "grad_norm": 491732.34375, "learning_rate": 4.7695118523376745e-05, "loss": 66.4764, "step": 10069 }, { "epoch": 0.4245631047494572, "grad_norm": 20565.822265625, "learning_rate": 4.7694337736968676e-05, "loss": 66.4692, "step": 10070 }, { "epoch": 0.4246052659316567, "grad_norm": 77681.0625, "learning_rate": 4.769355695056061e-05, "loss": 64.481, "step": 10071 }, { "epoch": 0.4246474271138563, "grad_norm": 72526.4921875, "learning_rate": 4.769277616415254e-05, "loss": 66.6042, "step": 10072 }, { "epoch": 0.4246895882960558, "grad_norm": 19670.083984375, "learning_rate": 4.769199537774446e-05, "loss": 68.8692, "step": 10073 }, { "epoch": 0.42473174947825537, "grad_norm": 12690.666015625, "learning_rate": 4.7691214591336394e-05, "loss": 68.8589, "step": 10074 }, { "epoch": 0.42477391066045495, "grad_norm": 40290.953125, "learning_rate": 4.7690433804928325e-05, "loss": 64.9038, "step": 10075 }, { "epoch": 0.42481607184265446, "grad_norm": 25275.154296875, "learning_rate": 4.768965301852026e-05, "loss": 67.1022, "step": 10076 }, { "epoch": 0.42485823302485404, "grad_norm": 12098.796875, "learning_rate": 4.768887223211219e-05, "loss": 68.0863, "step": 10077 }, { "epoch": 0.42490039420705356, "grad_norm": 11776.7451171875, "learning_rate": 4.768809144570411e-05, "loss": 66.4426, "step": 10078 }, { "epoch": 0.42494255538925313, "grad_norm": 47145.40625, "learning_rate": 4.768731065929605e-05, "loss": 65.768, "step": 10079 }, { "epoch": 0.42498471657145265, "grad_norm": 18026.556640625, "learning_rate": 4.7686529872887975e-05, "loss": 66.5927, "step": 10080 }, { "epoch": 0.4250268777536522, "grad_norm": 9172.57421875, "learning_rate": 4.76857490864799e-05, "loss": 62.2683, "step": 10081 }, { "epoch": 0.42506903893585174, "grad_norm": 7500.2041015625, "learning_rate": 4.768496830007184e-05, "loss": 70.7768, "step": 10082 }, { "epoch": 0.4251112001180513, "grad_norm": 12325.12890625, "learning_rate": 4.768418751366376e-05, "loss": 65.2979, "step": 10083 }, { "epoch": 0.42515336130025083, "grad_norm": 8236.8798828125, "learning_rate": 4.768340672725569e-05, "loss": 68.9547, "step": 10084 }, { "epoch": 0.4251955224824504, "grad_norm": 6587.6669921875, "learning_rate": 4.7682625940847624e-05, "loss": 69.577, "step": 10085 }, { "epoch": 0.42523768366465, "grad_norm": 8584.4384765625, "learning_rate": 4.7681845154439556e-05, "loss": 66.6783, "step": 10086 }, { "epoch": 0.4252798448468495, "grad_norm": 5239.00390625, "learning_rate": 4.768106436803149e-05, "loss": 65.0782, "step": 10087 }, { "epoch": 0.42532200602904907, "grad_norm": 11229.990234375, "learning_rate": 4.768028358162341e-05, "loss": 64.5568, "step": 10088 }, { "epoch": 0.4253641672112486, "grad_norm": 39369.09375, "learning_rate": 4.767950279521534e-05, "loss": 67.0693, "step": 10089 }, { "epoch": 0.42540632839344816, "grad_norm": 4976.49609375, "learning_rate": 4.7678722008807274e-05, "loss": 67.855, "step": 10090 }, { "epoch": 0.4254484895756477, "grad_norm": 7945.51611328125, "learning_rate": 4.7677941222399205e-05, "loss": 62.412, "step": 10091 }, { "epoch": 0.42549065075784726, "grad_norm": 4235.91552734375, "learning_rate": 4.767716043599113e-05, "loss": 62.129, "step": 10092 }, { "epoch": 0.4255328119400468, "grad_norm": 10218.841796875, "learning_rate": 4.767637964958306e-05, "loss": 67.4283, "step": 10093 }, { "epoch": 0.42557497312224635, "grad_norm": 7094.76318359375, "learning_rate": 4.767559886317499e-05, "loss": 67.3287, "step": 10094 }, { "epoch": 0.4256171343044459, "grad_norm": 16952.353515625, "learning_rate": 4.767481807676692e-05, "loss": 68.3945, "step": 10095 }, { "epoch": 0.42565929548664544, "grad_norm": 15254.6552734375, "learning_rate": 4.7674037290358855e-05, "loss": 62.9376, "step": 10096 }, { "epoch": 0.425701456668845, "grad_norm": 13539.8388671875, "learning_rate": 4.767325650395078e-05, "loss": 67.4841, "step": 10097 }, { "epoch": 0.42574361785104453, "grad_norm": 5596.31982421875, "learning_rate": 4.767247571754271e-05, "loss": 65.1657, "step": 10098 }, { "epoch": 0.4257857790332441, "grad_norm": 10259.1201171875, "learning_rate": 4.767169493113464e-05, "loss": 64.5058, "step": 10099 }, { "epoch": 0.4258279402154436, "grad_norm": 7518.6005859375, "learning_rate": 4.7670914144726566e-05, "loss": 64.9715, "step": 10100 }, { "epoch": 0.4258701013976432, "grad_norm": 3297.217041015625, "learning_rate": 4.7670133358318504e-05, "loss": 63.0359, "step": 10101 }, { "epoch": 0.4259122625798427, "grad_norm": 14524.7734375, "learning_rate": 4.766935257191043e-05, "loss": 63.4503, "step": 10102 }, { "epoch": 0.4259544237620423, "grad_norm": 10072.4716796875, "learning_rate": 4.766857178550236e-05, "loss": 66.7687, "step": 10103 }, { "epoch": 0.42599658494424186, "grad_norm": 7457.05859375, "learning_rate": 4.766779099909429e-05, "loss": 61.9885, "step": 10104 }, { "epoch": 0.4260387461264414, "grad_norm": 27283.455078125, "learning_rate": 4.766701021268622e-05, "loss": 64.6593, "step": 10105 }, { "epoch": 0.42608090730864095, "grad_norm": 9480.990234375, "learning_rate": 4.766622942627815e-05, "loss": 61.8157, "step": 10106 }, { "epoch": 0.4261230684908405, "grad_norm": 5627.2216796875, "learning_rate": 4.766544863987008e-05, "loss": 70.835, "step": 10107 }, { "epoch": 0.42616522967304005, "grad_norm": 17670.619140625, "learning_rate": 4.766466785346201e-05, "loss": 62.8384, "step": 10108 }, { "epoch": 0.42620739085523957, "grad_norm": 7542.9169921875, "learning_rate": 4.766388706705394e-05, "loss": 62.0065, "step": 10109 }, { "epoch": 0.42624955203743914, "grad_norm": 14103.560546875, "learning_rate": 4.766310628064587e-05, "loss": 64.5705, "step": 10110 }, { "epoch": 0.42629171321963866, "grad_norm": 11680.3896484375, "learning_rate": 4.7662325494237797e-05, "loss": 61.623, "step": 10111 }, { "epoch": 0.42633387440183823, "grad_norm": 10925.8544921875, "learning_rate": 4.766154470782973e-05, "loss": 61.1342, "step": 10112 }, { "epoch": 0.42637603558403775, "grad_norm": 8299.0341796875, "learning_rate": 4.766076392142166e-05, "loss": 60.0231, "step": 10113 }, { "epoch": 0.4264181967662373, "grad_norm": 8355.3955078125, "learning_rate": 4.7659983135013584e-05, "loss": 62.0102, "step": 10114 }, { "epoch": 0.4264603579484369, "grad_norm": 3654.53466796875, "learning_rate": 4.765920234860552e-05, "loss": 62.5257, "step": 10115 }, { "epoch": 0.4265025191306364, "grad_norm": 11288.0556640625, "learning_rate": 4.7658421562197446e-05, "loss": 60.0181, "step": 10116 }, { "epoch": 0.426544680312836, "grad_norm": 6758.662109375, "learning_rate": 4.765764077578938e-05, "loss": 61.0914, "step": 10117 }, { "epoch": 0.4265868414950355, "grad_norm": 2389.762451171875, "learning_rate": 4.765685998938131e-05, "loss": 61.2913, "step": 10118 }, { "epoch": 0.4266290026772351, "grad_norm": 12086.525390625, "learning_rate": 4.765607920297323e-05, "loss": 59.0893, "step": 10119 }, { "epoch": 0.4266711638594346, "grad_norm": 15714.7783203125, "learning_rate": 4.765529841656517e-05, "loss": 58.6725, "step": 10120 }, { "epoch": 0.4267133250416342, "grad_norm": 9114.61328125, "learning_rate": 4.7654517630157096e-05, "loss": 61.0173, "step": 10121 }, { "epoch": 0.4267554862238337, "grad_norm": 5620.28759765625, "learning_rate": 4.765373684374903e-05, "loss": 59.3549, "step": 10122 }, { "epoch": 0.42679764740603326, "grad_norm": 10204.8525390625, "learning_rate": 4.765295605734096e-05, "loss": 58.9204, "step": 10123 }, { "epoch": 0.42683980858823284, "grad_norm": 13906.43359375, "learning_rate": 4.765217527093289e-05, "loss": 59.2397, "step": 10124 }, { "epoch": 0.42688196977043236, "grad_norm": 13745.3935546875, "learning_rate": 4.7651394484524814e-05, "loss": 57.6061, "step": 10125 }, { "epoch": 0.42692413095263193, "grad_norm": 10674.34375, "learning_rate": 4.7650613698116745e-05, "loss": 59.8953, "step": 10126 }, { "epoch": 0.42696629213483145, "grad_norm": 10337.919921875, "learning_rate": 4.7649832911708676e-05, "loss": 59.0664, "step": 10127 }, { "epoch": 0.427008453317031, "grad_norm": 19116.412109375, "learning_rate": 4.76490521253006e-05, "loss": 56.5656, "step": 10128 }, { "epoch": 0.42705061449923054, "grad_norm": 6950.611328125, "learning_rate": 4.764827133889254e-05, "loss": 53.3069, "step": 10129 }, { "epoch": 0.4270927756814301, "grad_norm": 8136.85400390625, "learning_rate": 4.7647490552484463e-05, "loss": 53.3009, "step": 10130 }, { "epoch": 0.42713493686362963, "grad_norm": 7361.89697265625, "learning_rate": 4.7646709766076395e-05, "loss": 58.3268, "step": 10131 }, { "epoch": 0.4271770980458292, "grad_norm": 6015.4306640625, "learning_rate": 4.7645928979668326e-05, "loss": 54.291, "step": 10132 }, { "epoch": 0.4272192592280288, "grad_norm": 2373.1845703125, "learning_rate": 4.764514819326025e-05, "loss": 52.3831, "step": 10133 }, { "epoch": 0.4272614204102283, "grad_norm": 11122.46875, "learning_rate": 4.764436740685219e-05, "loss": 50.4712, "step": 10134 }, { "epoch": 0.4273035815924279, "grad_norm": 8186.388671875, "learning_rate": 4.764358662044411e-05, "loss": 53.1024, "step": 10135 }, { "epoch": 0.4273457427746274, "grad_norm": 6003.2392578125, "learning_rate": 4.7642805834036044e-05, "loss": 50.339, "step": 10136 }, { "epoch": 0.42738790395682696, "grad_norm": 6736.001953125, "learning_rate": 4.7642025047627975e-05, "loss": 49.9922, "step": 10137 }, { "epoch": 0.4274300651390265, "grad_norm": 2198.9521484375, "learning_rate": 4.76412442612199e-05, "loss": 52.5514, "step": 10138 }, { "epoch": 0.42747222632122606, "grad_norm": 1449.0343017578125, "learning_rate": 4.764046347481183e-05, "loss": 51.0899, "step": 10139 }, { "epoch": 0.4275143875034256, "grad_norm": 1324.5360107421875, "learning_rate": 4.763968268840376e-05, "loss": 50.8385, "step": 10140 }, { "epoch": 0.42755654868562515, "grad_norm": 2748.839111328125, "learning_rate": 4.7638901901995694e-05, "loss": 51.779, "step": 10141 }, { "epoch": 0.42759870986782467, "grad_norm": 3197.28564453125, "learning_rate": 4.763812111558762e-05, "loss": 46.6887, "step": 10142 }, { "epoch": 0.42764087105002424, "grad_norm": 4185.9169921875, "learning_rate": 4.763734032917955e-05, "loss": 48.8548, "step": 10143 }, { "epoch": 0.4276830322322238, "grad_norm": 3593.5087890625, "learning_rate": 4.763655954277148e-05, "loss": 46.1159, "step": 10144 }, { "epoch": 0.42772519341442333, "grad_norm": 7183.3876953125, "learning_rate": 4.763577875636341e-05, "loss": 43.138, "step": 10145 }, { "epoch": 0.4277673545966229, "grad_norm": 4596.869140625, "learning_rate": 4.763499796995534e-05, "loss": 44.3002, "step": 10146 }, { "epoch": 0.4278095157788224, "grad_norm": 2177.603271484375, "learning_rate": 4.763421718354727e-05, "loss": 41.4949, "step": 10147 }, { "epoch": 0.427851676961022, "grad_norm": 3446.170166015625, "learning_rate": 4.7633436397139206e-05, "loss": 40.9202, "step": 10148 }, { "epoch": 0.4278938381432215, "grad_norm": 3065.308349609375, "learning_rate": 4.763265561073113e-05, "loss": 40.2062, "step": 10149 }, { "epoch": 0.4279359993254211, "grad_norm": 3469.37451171875, "learning_rate": 4.763187482432306e-05, "loss": 38.8967, "step": 10150 }, { "epoch": 0.4279781605076206, "grad_norm": 2302.900390625, "learning_rate": 4.763109403791499e-05, "loss": 41.5114, "step": 10151 }, { "epoch": 0.4280203216898202, "grad_norm": 1935.5626220703125, "learning_rate": 4.763031325150692e-05, "loss": 40.3123, "step": 10152 }, { "epoch": 0.42806248287201976, "grad_norm": 2484.43115234375, "learning_rate": 4.762953246509885e-05, "loss": 38.7819, "step": 10153 }, { "epoch": 0.4281046440542193, "grad_norm": 3659.916748046875, "learning_rate": 4.762875167869078e-05, "loss": 39.513, "step": 10154 }, { "epoch": 0.42814680523641885, "grad_norm": 2125.703125, "learning_rate": 4.762797089228271e-05, "loss": 37.7244, "step": 10155 }, { "epoch": 0.42818896641861837, "grad_norm": 1088.8426513671875, "learning_rate": 4.762719010587464e-05, "loss": 37.1098, "step": 10156 }, { "epoch": 0.42823112760081794, "grad_norm": 1669.493896484375, "learning_rate": 4.762640931946657e-05, "loss": 39.0956, "step": 10157 }, { "epoch": 0.42827328878301746, "grad_norm": 3815.37646484375, "learning_rate": 4.76256285330585e-05, "loss": 34.4281, "step": 10158 }, { "epoch": 0.42831544996521703, "grad_norm": 7349.890625, "learning_rate": 4.762484774665043e-05, "loss": 33.1009, "step": 10159 }, { "epoch": 0.42835761114741655, "grad_norm": 1763.857666015625, "learning_rate": 4.762406696024236e-05, "loss": 35.2726, "step": 10160 }, { "epoch": 0.4283997723296161, "grad_norm": 1312.18310546875, "learning_rate": 4.7623286173834285e-05, "loss": 32.1147, "step": 10161 }, { "epoch": 0.4284419335118157, "grad_norm": 1439.00439453125, "learning_rate": 4.7622505387426216e-05, "loss": 33.9907, "step": 10162 }, { "epoch": 0.4284840946940152, "grad_norm": 4244.56640625, "learning_rate": 4.762172460101815e-05, "loss": 31.2102, "step": 10163 }, { "epoch": 0.4285262558762148, "grad_norm": 5754.05126953125, "learning_rate": 4.762094381461007e-05, "loss": 31.4545, "step": 10164 }, { "epoch": 0.4285684170584143, "grad_norm": 1690.093505859375, "learning_rate": 4.762016302820201e-05, "loss": 30.1398, "step": 10165 }, { "epoch": 0.4286105782406139, "grad_norm": 1231.3338623046875, "learning_rate": 4.7619382241793935e-05, "loss": 28.5812, "step": 10166 }, { "epoch": 0.4286527394228134, "grad_norm": 2373.91943359375, "learning_rate": 4.761860145538587e-05, "loss": 27.9184, "step": 10167 }, { "epoch": 0.428694900605013, "grad_norm": 1523.23828125, "learning_rate": 4.76178206689778e-05, "loss": 28.5697, "step": 10168 }, { "epoch": 0.4287370617872125, "grad_norm": 2617.455078125, "learning_rate": 4.761703988256973e-05, "loss": 25.2239, "step": 10169 }, { "epoch": 0.42877922296941207, "grad_norm": 1170.484619140625, "learning_rate": 4.761625909616166e-05, "loss": 25.3517, "step": 10170 }, { "epoch": 0.4288213841516116, "grad_norm": 1507.993408203125, "learning_rate": 4.7615478309753584e-05, "loss": 24.8609, "step": 10171 }, { "epoch": 0.42886354533381116, "grad_norm": 996.34814453125, "learning_rate": 4.7614697523345515e-05, "loss": 25.4142, "step": 10172 }, { "epoch": 0.42890570651601073, "grad_norm": 2596.402099609375, "learning_rate": 4.7613916736937447e-05, "loss": 24.0695, "step": 10173 }, { "epoch": 0.42894786769821025, "grad_norm": 1608.2313232421875, "learning_rate": 4.761313595052938e-05, "loss": 23.8089, "step": 10174 }, { "epoch": 0.4289900288804098, "grad_norm": 910.6967163085938, "learning_rate": 4.76123551641213e-05, "loss": 21.5697, "step": 10175 }, { "epoch": 0.42903219006260934, "grad_norm": 941.3162841796875, "learning_rate": 4.7611574377713234e-05, "loss": 19.8581, "step": 10176 }, { "epoch": 0.4290743512448089, "grad_norm": 601.8199462890625, "learning_rate": 4.7610793591305165e-05, "loss": 19.6143, "step": 10177 }, { "epoch": 0.42911651242700843, "grad_norm": 703.913330078125, "learning_rate": 4.7610012804897096e-05, "loss": 20.1465, "step": 10178 }, { "epoch": 0.429158673609208, "grad_norm": 681.7413940429688, "learning_rate": 4.760923201848903e-05, "loss": 19.4634, "step": 10179 }, { "epoch": 0.4292008347914075, "grad_norm": 4615.03369140625, "learning_rate": 4.760845123208095e-05, "loss": 18.5782, "step": 10180 }, { "epoch": 0.4292429959736071, "grad_norm": 10657.9443359375, "learning_rate": 4.760767044567288e-05, "loss": 17.9437, "step": 10181 }, { "epoch": 0.4292851571558067, "grad_norm": 7533.37451171875, "learning_rate": 4.7606889659264814e-05, "loss": 17.3171, "step": 10182 }, { "epoch": 0.4293273183380062, "grad_norm": 63601.15625, "learning_rate": 4.760610887285674e-05, "loss": 17.0462, "step": 10183 }, { "epoch": 0.42936947952020577, "grad_norm": 5019.82373046875, "learning_rate": 4.760532808644868e-05, "loss": 18.3633, "step": 10184 }, { "epoch": 0.4294116407024053, "grad_norm": 14343.212890625, "learning_rate": 4.76045473000406e-05, "loss": 16.4419, "step": 10185 }, { "epoch": 0.42945380188460486, "grad_norm": 4876.4599609375, "learning_rate": 4.760376651363253e-05, "loss": 15.9367, "step": 10186 }, { "epoch": 0.4294959630668044, "grad_norm": 1164.3404541015625, "learning_rate": 4.7602985727224464e-05, "loss": 17.4644, "step": 10187 }, { "epoch": 0.42953812424900395, "grad_norm": 4659.478515625, "learning_rate": 4.760220494081639e-05, "loss": 17.2422, "step": 10188 }, { "epoch": 0.42958028543120347, "grad_norm": 6242.853515625, "learning_rate": 4.7601424154408326e-05, "loss": 15.8491, "step": 10189 }, { "epoch": 0.42962244661340304, "grad_norm": 2041.811279296875, "learning_rate": 4.760064336800025e-05, "loss": 15.6957, "step": 10190 }, { "epoch": 0.4296646077956026, "grad_norm": 29302.810546875, "learning_rate": 4.759986258159218e-05, "loss": 15.1662, "step": 10191 }, { "epoch": 0.42970676897780213, "grad_norm": 118100064.0, "learning_rate": 4.7599081795184113e-05, "loss": 15.7942, "step": 10192 }, { "epoch": 0.4297489301600017, "grad_norm": 1647756.5, "learning_rate": 4.7598301008776045e-05, "loss": 15.1802, "step": 10193 }, { "epoch": 0.4297910913422012, "grad_norm": 167740112.0, "learning_rate": 4.759752022236797e-05, "loss": 15.4515, "step": 10194 }, { "epoch": 0.4298332525244008, "grad_norm": 75245.6796875, "learning_rate": 4.75967394359599e-05, "loss": 14.8497, "step": 10195 }, { "epoch": 0.4298754137066003, "grad_norm": 21194626.0, "learning_rate": 4.759595864955183e-05, "loss": 16.9536, "step": 10196 }, { "epoch": 0.4299175748887999, "grad_norm": 158676.296875, "learning_rate": 4.7595177863143756e-05, "loss": 14.5938, "step": 10197 }, { "epoch": 0.4299597360709994, "grad_norm": 265914.96875, "learning_rate": 4.7594397076735694e-05, "loss": 13.9363, "step": 10198 }, { "epoch": 0.430001897253199, "grad_norm": 57586.0859375, "learning_rate": 4.759361629032762e-05, "loss": 14.2762, "step": 10199 }, { "epoch": 0.4300440584353985, "grad_norm": 19010084.0, "learning_rate": 4.759283550391955e-05, "loss": 14.3877, "step": 10200 }, { "epoch": 0.4300862196175981, "grad_norm": 4246973.5, "learning_rate": 4.759205471751148e-05, "loss": 15.4319, "step": 10201 }, { "epoch": 0.43012838079979765, "grad_norm": 639089.625, "learning_rate": 4.7591273931103406e-05, "loss": 14.9621, "step": 10202 }, { "epoch": 0.43017054198199717, "grad_norm": 231966.390625, "learning_rate": 4.7590493144695344e-05, "loss": 16.1201, "step": 10203 }, { "epoch": 0.43021270316419674, "grad_norm": 3397113.25, "learning_rate": 4.758971235828727e-05, "loss": 15.2117, "step": 10204 }, { "epoch": 0.43025486434639626, "grad_norm": 36617.01953125, "learning_rate": 4.75889315718792e-05, "loss": 14.7804, "step": 10205 }, { "epoch": 0.43029702552859583, "grad_norm": 22517.029296875, "learning_rate": 4.758815078547113e-05, "loss": 13.9737, "step": 10206 }, { "epoch": 0.43033918671079535, "grad_norm": 266507.75, "learning_rate": 4.7587369999063055e-05, "loss": 14.4298, "step": 10207 }, { "epoch": 0.4303813478929949, "grad_norm": 1151210.625, "learning_rate": 4.7586589212654986e-05, "loss": 14.9027, "step": 10208 }, { "epoch": 0.43042350907519444, "grad_norm": 31816020.0, "learning_rate": 4.758580842624692e-05, "loss": 14.1864, "step": 10209 }, { "epoch": 0.430465670257394, "grad_norm": 498298.84375, "learning_rate": 4.758502763983885e-05, "loss": 14.9269, "step": 10210 }, { "epoch": 0.4305078314395936, "grad_norm": 2751.8212890625, "learning_rate": 4.7584246853430773e-05, "loss": 16.1333, "step": 10211 }, { "epoch": 0.4305499926217931, "grad_norm": 795.2531127929688, "learning_rate": 4.758346606702271e-05, "loss": 14.0143, "step": 10212 }, { "epoch": 0.4305921538039927, "grad_norm": 3936.546142578125, "learning_rate": 4.7582685280614636e-05, "loss": 14.6694, "step": 10213 }, { "epoch": 0.4306343149861922, "grad_norm": 1093.0765380859375, "learning_rate": 4.758190449420657e-05, "loss": 15.4157, "step": 10214 }, { "epoch": 0.4306764761683918, "grad_norm": 6854.83984375, "learning_rate": 4.75811237077985e-05, "loss": 13.6612, "step": 10215 }, { "epoch": 0.4307186373505913, "grad_norm": 4247.802734375, "learning_rate": 4.758034292139042e-05, "loss": 14.587, "step": 10216 }, { "epoch": 0.43076079853279087, "grad_norm": 1065.014892578125, "learning_rate": 4.757956213498236e-05, "loss": 14.5326, "step": 10217 }, { "epoch": 0.4308029597149904, "grad_norm": 1093.89990234375, "learning_rate": 4.7578781348574286e-05, "loss": 13.4195, "step": 10218 }, { "epoch": 0.43084512089718996, "grad_norm": 1461.466552734375, "learning_rate": 4.757800056216622e-05, "loss": 13.6837, "step": 10219 }, { "epoch": 0.43088728207938953, "grad_norm": 1279.3370361328125, "learning_rate": 4.757721977575815e-05, "loss": 13.2282, "step": 10220 }, { "epoch": 0.43092944326158905, "grad_norm": 3532.130859375, "learning_rate": 4.757643898935007e-05, "loss": 13.7406, "step": 10221 }, { "epoch": 0.4309716044437886, "grad_norm": 1555.776123046875, "learning_rate": 4.7575658202942004e-05, "loss": 13.6713, "step": 10222 }, { "epoch": 0.43101376562598814, "grad_norm": 1493.32080078125, "learning_rate": 4.7574877416533935e-05, "loss": 14.0109, "step": 10223 }, { "epoch": 0.4310559268081877, "grad_norm": 710.7109375, "learning_rate": 4.7574096630125866e-05, "loss": 13.9796, "step": 10224 }, { "epoch": 0.43109808799038724, "grad_norm": 672.1221313476562, "learning_rate": 4.75733158437178e-05, "loss": 12.7262, "step": 10225 }, { "epoch": 0.4311402491725868, "grad_norm": 2383.953369140625, "learning_rate": 4.757253505730972e-05, "loss": 12.1749, "step": 10226 }, { "epoch": 0.4311824103547863, "grad_norm": 10554.775390625, "learning_rate": 4.757175427090165e-05, "loss": 12.2796, "step": 10227 }, { "epoch": 0.4312245715369859, "grad_norm": 3979745.75, "learning_rate": 4.7570973484493585e-05, "loss": 12.6116, "step": 10228 }, { "epoch": 0.4312667327191854, "grad_norm": 3158.099365234375, "learning_rate": 4.7570192698085516e-05, "loss": 12.4936, "step": 10229 }, { "epoch": 0.431308893901385, "grad_norm": 17170.9375, "learning_rate": 4.756941191167744e-05, "loss": 11.6203, "step": 10230 }, { "epoch": 0.43135105508358457, "grad_norm": 1783.2000732421875, "learning_rate": 4.756863112526938e-05, "loss": 12.6525, "step": 10231 }, { "epoch": 0.4313932162657841, "grad_norm": 1700.5654296875, "learning_rate": 4.75678503388613e-05, "loss": 12.747, "step": 10232 }, { "epoch": 0.43143537744798366, "grad_norm": 831.0177001953125, "learning_rate": 4.756706955245323e-05, "loss": 11.0697, "step": 10233 }, { "epoch": 0.4314775386301832, "grad_norm": 743.50537109375, "learning_rate": 4.7566288766045165e-05, "loss": 11.4912, "step": 10234 }, { "epoch": 0.43151969981238275, "grad_norm": 1835.6297607421875, "learning_rate": 4.756550797963709e-05, "loss": 11.9665, "step": 10235 }, { "epoch": 0.43156186099458227, "grad_norm": 401.2403564453125, "learning_rate": 4.756472719322903e-05, "loss": 11.1975, "step": 10236 }, { "epoch": 0.43160402217678184, "grad_norm": 705.229248046875, "learning_rate": 4.756394640682095e-05, "loss": 10.5633, "step": 10237 }, { "epoch": 0.43164618335898136, "grad_norm": 6632.77392578125, "learning_rate": 4.7563165620412884e-05, "loss": 11.5758, "step": 10238 }, { "epoch": 0.43168834454118094, "grad_norm": 841.5562133789062, "learning_rate": 4.7562384834004815e-05, "loss": 11.6993, "step": 10239 }, { "epoch": 0.4317305057233805, "grad_norm": 346.5860595703125, "learning_rate": 4.756160404759674e-05, "loss": 10.4825, "step": 10240 }, { "epoch": 0.43177266690558, "grad_norm": 3185.840087890625, "learning_rate": 4.756082326118867e-05, "loss": 11.1557, "step": 10241 }, { "epoch": 0.4318148280877796, "grad_norm": 233.34365844726562, "learning_rate": 4.75600424747806e-05, "loss": 10.5529, "step": 10242 }, { "epoch": 0.4318569892699791, "grad_norm": 628.1873168945312, "learning_rate": 4.755926168837253e-05, "loss": 10.865, "step": 10243 }, { "epoch": 0.4318991504521787, "grad_norm": 329.017333984375, "learning_rate": 4.755848090196446e-05, "loss": 10.292, "step": 10244 }, { "epoch": 0.4319413116343782, "grad_norm": 257.5701599121094, "learning_rate": 4.755770011555639e-05, "loss": 10.1663, "step": 10245 }, { "epoch": 0.4319834728165778, "grad_norm": 750.7822265625, "learning_rate": 4.755691932914832e-05, "loss": 9.9256, "step": 10246 }, { "epoch": 0.4320256339987773, "grad_norm": 223.57272338867188, "learning_rate": 4.755613854274025e-05, "loss": 11.6797, "step": 10247 }, { "epoch": 0.4320677951809769, "grad_norm": 298.2427673339844, "learning_rate": 4.755535775633218e-05, "loss": 10.5997, "step": 10248 }, { "epoch": 0.43210995636317645, "grad_norm": 616.9548950195312, "learning_rate": 4.755457696992411e-05, "loss": 15.7303, "step": 10249 }, { "epoch": 0.43215211754537597, "grad_norm": 605.4960327148438, "learning_rate": 4.755379618351604e-05, "loss": 11.5698, "step": 10250 }, { "epoch": 0.43219427872757554, "grad_norm": 266.3812255859375, "learning_rate": 4.755301539710797e-05, "loss": 9.7515, "step": 10251 }, { "epoch": 0.43223643990977506, "grad_norm": 230.576171875, "learning_rate": 4.7552234610699894e-05, "loss": 11.2674, "step": 10252 }, { "epoch": 0.43227860109197463, "grad_norm": 165.11566162109375, "learning_rate": 4.755145382429183e-05, "loss": 9.8246, "step": 10253 }, { "epoch": 0.43232076227417415, "grad_norm": 161.2083282470703, "learning_rate": 4.755067303788376e-05, "loss": 9.9218, "step": 10254 }, { "epoch": 0.4323629234563737, "grad_norm": 195.58018493652344, "learning_rate": 4.754989225147569e-05, "loss": 11.1497, "step": 10255 }, { "epoch": 0.43240508463857324, "grad_norm": 281.8096008300781, "learning_rate": 4.754911146506762e-05, "loss": 9.5775, "step": 10256 }, { "epoch": 0.4324472458207728, "grad_norm": 201.15953063964844, "learning_rate": 4.754833067865955e-05, "loss": 10.1987, "step": 10257 }, { "epoch": 0.43248940700297234, "grad_norm": 299.98712158203125, "learning_rate": 4.754754989225148e-05, "loss": 9.7764, "step": 10258 }, { "epoch": 0.4325315681851719, "grad_norm": 603.2242431640625, "learning_rate": 4.7546769105843406e-05, "loss": 10.0475, "step": 10259 }, { "epoch": 0.4325737293673715, "grad_norm": 263.93182373046875, "learning_rate": 4.754598831943534e-05, "loss": 9.8713, "step": 10260 }, { "epoch": 0.432615890549571, "grad_norm": 342.45166015625, "learning_rate": 4.754520753302727e-05, "loss": 10.7782, "step": 10261 }, { "epoch": 0.4326580517317706, "grad_norm": 1219.7353515625, "learning_rate": 4.75444267466192e-05, "loss": 9.4033, "step": 10262 }, { "epoch": 0.4327002129139701, "grad_norm": 196.73883056640625, "learning_rate": 4.7543645960211124e-05, "loss": 9.9856, "step": 10263 }, { "epoch": 0.43274237409616967, "grad_norm": 254.07382202148438, "learning_rate": 4.7542865173803056e-05, "loss": 9.5427, "step": 10264 }, { "epoch": 0.4327845352783692, "grad_norm": 801.1953735351562, "learning_rate": 4.754208438739499e-05, "loss": 10.2832, "step": 10265 }, { "epoch": 0.43282669646056876, "grad_norm": 443.3643798828125, "learning_rate": 4.754130360098691e-05, "loss": 9.6455, "step": 10266 }, { "epoch": 0.4328688576427683, "grad_norm": 362.6656188964844, "learning_rate": 4.754052281457885e-05, "loss": 9.396, "step": 10267 }, { "epoch": 0.43291101882496785, "grad_norm": 580.8248901367188, "learning_rate": 4.7539742028170774e-05, "loss": 11.4348, "step": 10268 }, { "epoch": 0.4329531800071674, "grad_norm": 3593.475830078125, "learning_rate": 4.7538961241762705e-05, "loss": 11.0381, "step": 10269 }, { "epoch": 0.43299534118936694, "grad_norm": 320.1392517089844, "learning_rate": 4.7538180455354637e-05, "loss": 10.967, "step": 10270 }, { "epoch": 0.4330375023715665, "grad_norm": 417.7598571777344, "learning_rate": 4.753739966894656e-05, "loss": 9.6756, "step": 10271 }, { "epoch": 0.43307966355376604, "grad_norm": 685.04296875, "learning_rate": 4.75366188825385e-05, "loss": 9.0617, "step": 10272 }, { "epoch": 0.4331218247359656, "grad_norm": 776.7842407226562, "learning_rate": 4.7535838096130424e-05, "loss": 9.9307, "step": 10273 }, { "epoch": 0.43316398591816513, "grad_norm": 707.9800415039062, "learning_rate": 4.7535057309722355e-05, "loss": 10.0948, "step": 10274 }, { "epoch": 0.4332061471003647, "grad_norm": 286.51300048828125, "learning_rate": 4.7534276523314286e-05, "loss": 10.1719, "step": 10275 }, { "epoch": 0.4332483082825642, "grad_norm": 424.1825256347656, "learning_rate": 4.753349573690622e-05, "loss": 9.4355, "step": 10276 }, { "epoch": 0.4332904694647638, "grad_norm": 617.6268310546875, "learning_rate": 4.753271495049814e-05, "loss": 9.0518, "step": 10277 }, { "epoch": 0.43333263064696337, "grad_norm": 657.70947265625, "learning_rate": 4.753193416409007e-05, "loss": 9.3514, "step": 10278 }, { "epoch": 0.4333747918291629, "grad_norm": 774.4436645507812, "learning_rate": 4.7531153377682004e-05, "loss": 9.0214, "step": 10279 }, { "epoch": 0.43341695301136246, "grad_norm": 164.0991668701172, "learning_rate": 4.753037259127393e-05, "loss": 9.4165, "step": 10280 }, { "epoch": 0.433459114193562, "grad_norm": 2219.024658203125, "learning_rate": 4.752959180486587e-05, "loss": 9.4495, "step": 10281 }, { "epoch": 0.43350127537576155, "grad_norm": 1271.97900390625, "learning_rate": 4.752881101845779e-05, "loss": 10.5446, "step": 10282 }, { "epoch": 0.43354343655796107, "grad_norm": 1564.7144775390625, "learning_rate": 4.752803023204972e-05, "loss": 9.6991, "step": 10283 }, { "epoch": 0.43358559774016064, "grad_norm": 932.0067749023438, "learning_rate": 4.7527249445641654e-05, "loss": 9.0221, "step": 10284 }, { "epoch": 0.43362775892236016, "grad_norm": 423.169189453125, "learning_rate": 4.752646865923358e-05, "loss": 9.6282, "step": 10285 }, { "epoch": 0.43366992010455974, "grad_norm": 1622.802490234375, "learning_rate": 4.7525687872825516e-05, "loss": 8.8127, "step": 10286 }, { "epoch": 0.43371208128675925, "grad_norm": 1065.7794189453125, "learning_rate": 4.752490708641744e-05, "loss": 10.2107, "step": 10287 }, { "epoch": 0.43375424246895883, "grad_norm": 1367.462890625, "learning_rate": 4.752412630000937e-05, "loss": 9.6744, "step": 10288 }, { "epoch": 0.4337964036511584, "grad_norm": 775.7685546875, "learning_rate": 4.75233455136013e-05, "loss": 9.384, "step": 10289 }, { "epoch": 0.4338385648333579, "grad_norm": 623.3073120117188, "learning_rate": 4.752256472719323e-05, "loss": 9.3812, "step": 10290 }, { "epoch": 0.4338807260155575, "grad_norm": 254.523193359375, "learning_rate": 4.752178394078516e-05, "loss": 9.4499, "step": 10291 }, { "epoch": 0.433922887197757, "grad_norm": 201.78652954101562, "learning_rate": 4.752100315437709e-05, "loss": 11.1781, "step": 10292 }, { "epoch": 0.4339650483799566, "grad_norm": 256.28656005859375, "learning_rate": 4.752022236796902e-05, "loss": 8.9834, "step": 10293 }, { "epoch": 0.4340072095621561, "grad_norm": 1197.4842529296875, "learning_rate": 4.751944158156095e-05, "loss": 9.6562, "step": 10294 }, { "epoch": 0.4340493707443557, "grad_norm": 983.9232177734375, "learning_rate": 4.751866079515288e-05, "loss": 10.6857, "step": 10295 }, { "epoch": 0.4340915319265552, "grad_norm": 1431.883544921875, "learning_rate": 4.751788000874481e-05, "loss": 8.7619, "step": 10296 }, { "epoch": 0.43413369310875477, "grad_norm": 459.0527038574219, "learning_rate": 4.751709922233674e-05, "loss": 9.3839, "step": 10297 }, { "epoch": 0.43417585429095434, "grad_norm": 343.290771484375, "learning_rate": 4.751631843592867e-05, "loss": 8.7878, "step": 10298 }, { "epoch": 0.43421801547315386, "grad_norm": 1068.6077880859375, "learning_rate": 4.7515537649520596e-05, "loss": 10.0501, "step": 10299 }, { "epoch": 0.43426017665535344, "grad_norm": 241.31466674804688, "learning_rate": 4.7514756863112534e-05, "loss": 9.2885, "step": 10300 }, { "epoch": 0.43430233783755295, "grad_norm": 248.01356506347656, "learning_rate": 4.751397607670446e-05, "loss": 8.891, "step": 10301 }, { "epoch": 0.43434449901975253, "grad_norm": 245.06106567382812, "learning_rate": 4.751319529029639e-05, "loss": 9.7469, "step": 10302 }, { "epoch": 0.43438666020195205, "grad_norm": 697.8612670898438, "learning_rate": 4.751241450388832e-05, "loss": 9.0695, "step": 10303 }, { "epoch": 0.4344288213841516, "grad_norm": 360.40234375, "learning_rate": 4.7511633717480245e-05, "loss": 8.951, "step": 10304 }, { "epoch": 0.43447098256635114, "grad_norm": 413.2932434082031, "learning_rate": 4.751085293107218e-05, "loss": 9.2178, "step": 10305 }, { "epoch": 0.4345131437485507, "grad_norm": 470.88470458984375, "learning_rate": 4.751007214466411e-05, "loss": 9.1133, "step": 10306 }, { "epoch": 0.4345553049307503, "grad_norm": 464.35589599609375, "learning_rate": 4.750929135825604e-05, "loss": 9.0855, "step": 10307 }, { "epoch": 0.4345974661129498, "grad_norm": 410.014892578125, "learning_rate": 4.750851057184797e-05, "loss": 8.7084, "step": 10308 }, { "epoch": 0.4346396272951494, "grad_norm": 364.9948425292969, "learning_rate": 4.7507729785439895e-05, "loss": 8.9409, "step": 10309 }, { "epoch": 0.4346817884773489, "grad_norm": 1009.6121215820312, "learning_rate": 4.7506948999031826e-05, "loss": 9.5561, "step": 10310 }, { "epoch": 0.43472394965954847, "grad_norm": 1005.8134155273438, "learning_rate": 4.750616821262376e-05, "loss": 8.7976, "step": 10311 }, { "epoch": 0.434766110841748, "grad_norm": 542.3357543945312, "learning_rate": 4.750538742621569e-05, "loss": 8.9482, "step": 10312 }, { "epoch": 0.43480827202394756, "grad_norm": 924.07666015625, "learning_rate": 4.750460663980761e-05, "loss": 9.2214, "step": 10313 }, { "epoch": 0.4348504332061471, "grad_norm": 462.64312744140625, "learning_rate": 4.7503825853399544e-05, "loss": 9.2252, "step": 10314 }, { "epoch": 0.43489259438834665, "grad_norm": 273.68499755859375, "learning_rate": 4.7503045066991475e-05, "loss": 8.5561, "step": 10315 }, { "epoch": 0.43493475557054617, "grad_norm": 184.00741577148438, "learning_rate": 4.750226428058341e-05, "loss": 9.4194, "step": 10316 }, { "epoch": 0.43497691675274575, "grad_norm": 1850.230224609375, "learning_rate": 4.750148349417534e-05, "loss": 8.4173, "step": 10317 }, { "epoch": 0.4350190779349453, "grad_norm": 633.5872192382812, "learning_rate": 4.750070270776726e-05, "loss": 10.3845, "step": 10318 }, { "epoch": 0.43506123911714484, "grad_norm": 238.06869506835938, "learning_rate": 4.74999219213592e-05, "loss": 9.3216, "step": 10319 }, { "epoch": 0.4351034002993444, "grad_norm": 278.6080627441406, "learning_rate": 4.7499141134951125e-05, "loss": 8.6171, "step": 10320 }, { "epoch": 0.43514556148154393, "grad_norm": 210.6923828125, "learning_rate": 4.749836034854305e-05, "loss": 8.9398, "step": 10321 }, { "epoch": 0.4351877226637435, "grad_norm": 208.8592987060547, "learning_rate": 4.749757956213499e-05, "loss": 10.0024, "step": 10322 }, { "epoch": 0.435229883845943, "grad_norm": 199.252197265625, "learning_rate": 4.749679877572691e-05, "loss": 9.2247, "step": 10323 }, { "epoch": 0.4352720450281426, "grad_norm": 825.9542846679688, "learning_rate": 4.749601798931884e-05, "loss": 9.0533, "step": 10324 }, { "epoch": 0.4353142062103421, "grad_norm": 871.6721801757812, "learning_rate": 4.7495237202910774e-05, "loss": 9.5943, "step": 10325 }, { "epoch": 0.4353563673925417, "grad_norm": 247.07504272460938, "learning_rate": 4.7494456416502706e-05, "loss": 8.6959, "step": 10326 }, { "epoch": 0.43539852857474126, "grad_norm": 2069.571533203125, "learning_rate": 4.749367563009464e-05, "loss": 9.6358, "step": 10327 }, { "epoch": 0.4354406897569408, "grad_norm": 415.1199645996094, "learning_rate": 4.749289484368656e-05, "loss": 9.3754, "step": 10328 }, { "epoch": 0.43548285093914035, "grad_norm": 594.0272216796875, "learning_rate": 4.749211405727849e-05, "loss": 8.7879, "step": 10329 }, { "epoch": 0.43552501212133987, "grad_norm": 1458.8089599609375, "learning_rate": 4.7491333270870424e-05, "loss": 9.8552, "step": 10330 }, { "epoch": 0.43556717330353945, "grad_norm": 338.7960205078125, "learning_rate": 4.7490552484462355e-05, "loss": 9.494, "step": 10331 }, { "epoch": 0.43560933448573896, "grad_norm": 1301.489013671875, "learning_rate": 4.748977169805428e-05, "loss": 8.8701, "step": 10332 }, { "epoch": 0.43565149566793854, "grad_norm": 301.1697998046875, "learning_rate": 4.748899091164621e-05, "loss": 10.0743, "step": 10333 }, { "epoch": 0.43569365685013806, "grad_norm": 4800.4404296875, "learning_rate": 4.748821012523814e-05, "loss": 9.8594, "step": 10334 }, { "epoch": 0.43573581803233763, "grad_norm": 857.95556640625, "learning_rate": 4.748742933883007e-05, "loss": 9.7964, "step": 10335 }, { "epoch": 0.4357779792145372, "grad_norm": 1651.671630859375, "learning_rate": 4.7486648552422005e-05, "loss": 8.7268, "step": 10336 }, { "epoch": 0.4358201403967367, "grad_norm": 2112.322998046875, "learning_rate": 4.748586776601393e-05, "loss": 9.3395, "step": 10337 }, { "epoch": 0.4358623015789363, "grad_norm": 496.0100402832031, "learning_rate": 4.748508697960587e-05, "loss": 9.58, "step": 10338 }, { "epoch": 0.4359044627611358, "grad_norm": 695.3055419921875, "learning_rate": 4.748430619319779e-05, "loss": 8.9972, "step": 10339 }, { "epoch": 0.4359466239433354, "grad_norm": 1924.7230224609375, "learning_rate": 4.7483525406789716e-05, "loss": 8.5217, "step": 10340 }, { "epoch": 0.4359887851255349, "grad_norm": 814.2390747070312, "learning_rate": 4.7482744620381654e-05, "loss": 9.1257, "step": 10341 }, { "epoch": 0.4360309463077345, "grad_norm": 679.7388916015625, "learning_rate": 4.748196383397358e-05, "loss": 8.5982, "step": 10342 }, { "epoch": 0.436073107489934, "grad_norm": 496.43280029296875, "learning_rate": 4.748118304756551e-05, "loss": 8.5006, "step": 10343 }, { "epoch": 0.43611526867213357, "grad_norm": 1413.7445068359375, "learning_rate": 4.748040226115744e-05, "loss": 8.7692, "step": 10344 }, { "epoch": 0.4361574298543331, "grad_norm": 1126.43798828125, "learning_rate": 4.747962147474937e-05, "loss": 8.6657, "step": 10345 }, { "epoch": 0.43619959103653266, "grad_norm": 1727.9306640625, "learning_rate": 4.74788406883413e-05, "loss": 9.4422, "step": 10346 }, { "epoch": 0.43624175221873224, "grad_norm": 968.0904541015625, "learning_rate": 4.747805990193323e-05, "loss": 10.3277, "step": 10347 }, { "epoch": 0.43628391340093176, "grad_norm": 1034.2298583984375, "learning_rate": 4.747727911552516e-05, "loss": 9.5775, "step": 10348 }, { "epoch": 0.43632607458313133, "grad_norm": 590.4878540039062, "learning_rate": 4.747649832911709e-05, "loss": 8.7036, "step": 10349 }, { "epoch": 0.43636823576533085, "grad_norm": 939.2133178710938, "learning_rate": 4.747571754270902e-05, "loss": 9.9862, "step": 10350 }, { "epoch": 0.4364103969475304, "grad_norm": 1743.87939453125, "learning_rate": 4.7474936756300947e-05, "loss": 8.7969, "step": 10351 }, { "epoch": 0.43645255812972994, "grad_norm": 2098.372802734375, "learning_rate": 4.747415596989288e-05, "loss": 8.9461, "step": 10352 }, { "epoch": 0.4364947193119295, "grad_norm": 3119.538818359375, "learning_rate": 4.747337518348481e-05, "loss": 10.2296, "step": 10353 }, { "epoch": 0.43653688049412903, "grad_norm": 1711.58154296875, "learning_rate": 4.7472594397076734e-05, "loss": 10.6996, "step": 10354 }, { "epoch": 0.4365790416763286, "grad_norm": 9085.8662109375, "learning_rate": 4.747181361066867e-05, "loss": 9.3133, "step": 10355 }, { "epoch": 0.4366212028585282, "grad_norm": 720.9588012695312, "learning_rate": 4.7471032824260596e-05, "loss": 10.153, "step": 10356 }, { "epoch": 0.4366633640407277, "grad_norm": 1272.4185791015625, "learning_rate": 4.747025203785253e-05, "loss": 9.2246, "step": 10357 }, { "epoch": 0.43670552522292727, "grad_norm": 830.6892700195312, "learning_rate": 4.746947125144446e-05, "loss": 9.6724, "step": 10358 }, { "epoch": 0.4367476864051268, "grad_norm": 1444.650634765625, "learning_rate": 4.746869046503638e-05, "loss": 9.1611, "step": 10359 }, { "epoch": 0.43678984758732636, "grad_norm": 546.4075317382812, "learning_rate": 4.7467909678628314e-05, "loss": 9.469, "step": 10360 }, { "epoch": 0.4368320087695259, "grad_norm": 1577.1751708984375, "learning_rate": 4.7467128892220246e-05, "loss": 9.0659, "step": 10361 }, { "epoch": 0.43687416995172546, "grad_norm": 2751.12353515625, "learning_rate": 4.746634810581218e-05, "loss": 10.1774, "step": 10362 }, { "epoch": 0.436916331133925, "grad_norm": 3470.3134765625, "learning_rate": 4.746556731940411e-05, "loss": 8.1292, "step": 10363 }, { "epoch": 0.43695849231612455, "grad_norm": 933.936279296875, "learning_rate": 4.746478653299604e-05, "loss": 8.9163, "step": 10364 }, { "epoch": 0.43700065349832407, "grad_norm": 561.461669921875, "learning_rate": 4.7464005746587964e-05, "loss": 9.0038, "step": 10365 }, { "epoch": 0.43704281468052364, "grad_norm": 1395.526611328125, "learning_rate": 4.7463224960179895e-05, "loss": 9.1677, "step": 10366 }, { "epoch": 0.4370849758627232, "grad_norm": 2358.153076171875, "learning_rate": 4.7462444173771826e-05, "loss": 8.8505, "step": 10367 }, { "epoch": 0.43712713704492273, "grad_norm": 1011.3880004882812, "learning_rate": 4.746166338736375e-05, "loss": 8.6759, "step": 10368 }, { "epoch": 0.4371692982271223, "grad_norm": 1345.79296875, "learning_rate": 4.746088260095569e-05, "loss": 8.9335, "step": 10369 }, { "epoch": 0.4372114594093218, "grad_norm": 1467.4215087890625, "learning_rate": 4.7460101814547613e-05, "loss": 9.2325, "step": 10370 }, { "epoch": 0.4372536205915214, "grad_norm": 955.3640747070312, "learning_rate": 4.7459321028139545e-05, "loss": 9.1847, "step": 10371 }, { "epoch": 0.4372957817737209, "grad_norm": 2451.774169921875, "learning_rate": 4.7458540241731476e-05, "loss": 10.0528, "step": 10372 }, { "epoch": 0.4373379429559205, "grad_norm": 1937.0760498046875, "learning_rate": 4.74577594553234e-05, "loss": 8.8397, "step": 10373 }, { "epoch": 0.43738010413812, "grad_norm": 2062.95068359375, "learning_rate": 4.745697866891534e-05, "loss": 9.24, "step": 10374 }, { "epoch": 0.4374222653203196, "grad_norm": 2301.542724609375, "learning_rate": 4.745619788250726e-05, "loss": 9.7997, "step": 10375 }, { "epoch": 0.43746442650251915, "grad_norm": 1157.78955078125, "learning_rate": 4.7455417096099194e-05, "loss": 9.7403, "step": 10376 }, { "epoch": 0.4375065876847187, "grad_norm": 770.3562622070312, "learning_rate": 4.7454636309691125e-05, "loss": 9.2872, "step": 10377 }, { "epoch": 0.43754874886691825, "grad_norm": 1132.8349609375, "learning_rate": 4.745385552328305e-05, "loss": 9.3356, "step": 10378 }, { "epoch": 0.43759091004911777, "grad_norm": 5208.66748046875, "learning_rate": 4.745307473687498e-05, "loss": 9.2681, "step": 10379 }, { "epoch": 0.43763307123131734, "grad_norm": 1430.3087158203125, "learning_rate": 4.745229395046691e-05, "loss": 9.541, "step": 10380 }, { "epoch": 0.43767523241351686, "grad_norm": 8137.513671875, "learning_rate": 4.7451513164058844e-05, "loss": 8.4233, "step": 10381 }, { "epoch": 0.43771739359571643, "grad_norm": 2986.139404296875, "learning_rate": 4.745073237765077e-05, "loss": 8.7394, "step": 10382 }, { "epoch": 0.43775955477791595, "grad_norm": 7593.7802734375, "learning_rate": 4.74499515912427e-05, "loss": 8.6331, "step": 10383 }, { "epoch": 0.4378017159601155, "grad_norm": 603.1563110351562, "learning_rate": 4.744917080483463e-05, "loss": 8.6048, "step": 10384 }, { "epoch": 0.4378438771423151, "grad_norm": 3124.45458984375, "learning_rate": 4.744839001842656e-05, "loss": 9.0053, "step": 10385 }, { "epoch": 0.4378860383245146, "grad_norm": 1563.40869140625, "learning_rate": 4.744760923201849e-05, "loss": 8.4075, "step": 10386 }, { "epoch": 0.4379281995067142, "grad_norm": 1163.6669921875, "learning_rate": 4.744682844561042e-05, "loss": 8.7822, "step": 10387 }, { "epoch": 0.4379703606889137, "grad_norm": 2880.73828125, "learning_rate": 4.7446047659202356e-05, "loss": 8.3665, "step": 10388 }, { "epoch": 0.4380125218711133, "grad_norm": 1326.4681396484375, "learning_rate": 4.744526687279428e-05, "loss": 8.627, "step": 10389 }, { "epoch": 0.4380546830533128, "grad_norm": 1231.14990234375, "learning_rate": 4.744448608638621e-05, "loss": 8.7345, "step": 10390 }, { "epoch": 0.4380968442355124, "grad_norm": 2130.390380859375, "learning_rate": 4.744370529997814e-05, "loss": 9.4812, "step": 10391 }, { "epoch": 0.4381390054177119, "grad_norm": 3163.9619140625, "learning_rate": 4.744292451357007e-05, "loss": 8.4638, "step": 10392 }, { "epoch": 0.43818116659991146, "grad_norm": 761.31787109375, "learning_rate": 4.7442143727162e-05, "loss": 8.5325, "step": 10393 }, { "epoch": 0.438223327782111, "grad_norm": 4495.82763671875, "learning_rate": 4.744136294075393e-05, "loss": 8.7784, "step": 10394 }, { "epoch": 0.43826548896431056, "grad_norm": 19506.505859375, "learning_rate": 4.744058215434586e-05, "loss": 9.3842, "step": 10395 }, { "epoch": 0.43830765014651013, "grad_norm": 740.1141967773438, "learning_rate": 4.743980136793779e-05, "loss": 8.1994, "step": 10396 }, { "epoch": 0.43834981132870965, "grad_norm": 788.564697265625, "learning_rate": 4.743902058152972e-05, "loss": 8.5759, "step": 10397 }, { "epoch": 0.4383919725109092, "grad_norm": 1232.9141845703125, "learning_rate": 4.743823979512165e-05, "loss": 8.5312, "step": 10398 }, { "epoch": 0.43843413369310874, "grad_norm": 586.9892578125, "learning_rate": 4.743745900871358e-05, "loss": 8.51, "step": 10399 }, { "epoch": 0.4384762948753083, "grad_norm": 1063.0277099609375, "learning_rate": 4.743667822230551e-05, "loss": 8.4896, "step": 10400 }, { "epoch": 0.43851845605750783, "grad_norm": 684.0381469726562, "learning_rate": 4.7435897435897435e-05, "loss": 8.4687, "step": 10401 }, { "epoch": 0.4385606172397074, "grad_norm": 4856.2109375, "learning_rate": 4.7435116649489366e-05, "loss": 9.0692, "step": 10402 }, { "epoch": 0.4386027784219069, "grad_norm": 8430.65234375, "learning_rate": 4.74343358630813e-05, "loss": 9.9831, "step": 10403 }, { "epoch": 0.4386449396041065, "grad_norm": 19785.10546875, "learning_rate": 4.743355507667322e-05, "loss": 8.9992, "step": 10404 }, { "epoch": 0.4386871007863061, "grad_norm": 981.3108520507812, "learning_rate": 4.743277429026516e-05, "loss": 9.2976, "step": 10405 }, { "epoch": 0.4387292619685056, "grad_norm": 2507.413330078125, "learning_rate": 4.7431993503857085e-05, "loss": 9.5752, "step": 10406 }, { "epoch": 0.43877142315070516, "grad_norm": 784.8739624023438, "learning_rate": 4.743121271744902e-05, "loss": 8.9724, "step": 10407 }, { "epoch": 0.4388135843329047, "grad_norm": 2414.4580078125, "learning_rate": 4.743043193104095e-05, "loss": 8.8856, "step": 10408 }, { "epoch": 0.43885574551510426, "grad_norm": 1087.3055419921875, "learning_rate": 4.742965114463288e-05, "loss": 9.4609, "step": 10409 }, { "epoch": 0.4388979066973038, "grad_norm": 4343.99169921875, "learning_rate": 4.742887035822481e-05, "loss": 10.3719, "step": 10410 }, { "epoch": 0.43894006787950335, "grad_norm": 5961.1787109375, "learning_rate": 4.7428089571816734e-05, "loss": 8.2467, "step": 10411 }, { "epoch": 0.43898222906170287, "grad_norm": 1601.587890625, "learning_rate": 4.7427308785408665e-05, "loss": 9.057, "step": 10412 }, { "epoch": 0.43902439024390244, "grad_norm": 2213.686279296875, "learning_rate": 4.7426527999000597e-05, "loss": 8.2564, "step": 10413 }, { "epoch": 0.439066551426102, "grad_norm": 2520.924560546875, "learning_rate": 4.742574721259253e-05, "loss": 9.4436, "step": 10414 }, { "epoch": 0.43910871260830153, "grad_norm": 2605.146728515625, "learning_rate": 4.742496642618445e-05, "loss": 9.3362, "step": 10415 }, { "epoch": 0.4391508737905011, "grad_norm": 2624.96826171875, "learning_rate": 4.7424185639776384e-05, "loss": 8.4457, "step": 10416 }, { "epoch": 0.4391930349727006, "grad_norm": 50579.95703125, "learning_rate": 4.7423404853368315e-05, "loss": 9.745, "step": 10417 }, { "epoch": 0.4392351961549002, "grad_norm": 2950.571533203125, "learning_rate": 4.7422624066960246e-05, "loss": 10.422, "step": 10418 }, { "epoch": 0.4392773573370997, "grad_norm": 7335.0205078125, "learning_rate": 4.742184328055218e-05, "loss": 8.0564, "step": 10419 }, { "epoch": 0.4393195185192993, "grad_norm": 2816.480224609375, "learning_rate": 4.74210624941441e-05, "loss": 8.5207, "step": 10420 }, { "epoch": 0.4393616797014988, "grad_norm": 902.3738403320312, "learning_rate": 4.742028170773603e-05, "loss": 8.7643, "step": 10421 }, { "epoch": 0.4394038408836984, "grad_norm": 2797.57275390625, "learning_rate": 4.7419500921327964e-05, "loss": 8.3701, "step": 10422 }, { "epoch": 0.4394460020658979, "grad_norm": 1494.1524658203125, "learning_rate": 4.741872013491989e-05, "loss": 8.8168, "step": 10423 }, { "epoch": 0.4394881632480975, "grad_norm": 3817.789794921875, "learning_rate": 4.741793934851183e-05, "loss": 9.1297, "step": 10424 }, { "epoch": 0.43953032443029705, "grad_norm": 821.2020874023438, "learning_rate": 4.741715856210375e-05, "loss": 8.6309, "step": 10425 }, { "epoch": 0.43957248561249657, "grad_norm": 1082.26318359375, "learning_rate": 4.741637777569568e-05, "loss": 8.3906, "step": 10426 }, { "epoch": 0.43961464679469614, "grad_norm": 3033.400634765625, "learning_rate": 4.7415596989287614e-05, "loss": 9.3194, "step": 10427 }, { "epoch": 0.43965680797689566, "grad_norm": 1198.6885986328125, "learning_rate": 4.741481620287954e-05, "loss": 8.5383, "step": 10428 }, { "epoch": 0.43969896915909523, "grad_norm": 1283.0994873046875, "learning_rate": 4.7414035416471476e-05, "loss": 10.4853, "step": 10429 }, { "epoch": 0.43974113034129475, "grad_norm": 1942.6336669921875, "learning_rate": 4.74132546300634e-05, "loss": 8.3068, "step": 10430 }, { "epoch": 0.4397832915234943, "grad_norm": 1033.434326171875, "learning_rate": 4.741247384365533e-05, "loss": 8.3698, "step": 10431 }, { "epoch": 0.43982545270569384, "grad_norm": 1624.687255859375, "learning_rate": 4.7411693057247263e-05, "loss": 9.4695, "step": 10432 }, { "epoch": 0.4398676138878934, "grad_norm": 1366.885986328125, "learning_rate": 4.7410912270839195e-05, "loss": 8.6939, "step": 10433 }, { "epoch": 0.439909775070093, "grad_norm": 3422.559814453125, "learning_rate": 4.741013148443112e-05, "loss": 8.6325, "step": 10434 }, { "epoch": 0.4399519362522925, "grad_norm": 969.994873046875, "learning_rate": 4.740935069802305e-05, "loss": 8.453, "step": 10435 }, { "epoch": 0.4399940974344921, "grad_norm": 830.0562133789062, "learning_rate": 4.740856991161498e-05, "loss": 9.2588, "step": 10436 }, { "epoch": 0.4400362586166916, "grad_norm": 8629.46484375, "learning_rate": 4.7407789125206906e-05, "loss": 8.7896, "step": 10437 }, { "epoch": 0.4400784197988912, "grad_norm": 10283.0283203125, "learning_rate": 4.7407008338798844e-05, "loss": 8.5958, "step": 10438 }, { "epoch": 0.4401205809810907, "grad_norm": 1747.310791015625, "learning_rate": 4.740622755239077e-05, "loss": 10.035, "step": 10439 }, { "epoch": 0.44016274216329027, "grad_norm": 2295.426513671875, "learning_rate": 4.74054467659827e-05, "loss": 9.5379, "step": 10440 }, { "epoch": 0.4402049033454898, "grad_norm": 3459.834228515625, "learning_rate": 4.740466597957463e-05, "loss": 9.0252, "step": 10441 }, { "epoch": 0.44024706452768936, "grad_norm": 11180.6005859375, "learning_rate": 4.7403885193166556e-05, "loss": 9.1382, "step": 10442 }, { "epoch": 0.44028922570988893, "grad_norm": 1048.44482421875, "learning_rate": 4.7403104406758494e-05, "loss": 9.0042, "step": 10443 }, { "epoch": 0.44033138689208845, "grad_norm": 1049.4632568359375, "learning_rate": 4.740232362035042e-05, "loss": 8.9955, "step": 10444 }, { "epoch": 0.440373548074288, "grad_norm": 5143.53125, "learning_rate": 4.740154283394235e-05, "loss": 9.1408, "step": 10445 }, { "epoch": 0.44041570925648754, "grad_norm": 1064.6771240234375, "learning_rate": 4.740076204753428e-05, "loss": 9.4899, "step": 10446 }, { "epoch": 0.4404578704386871, "grad_norm": 22696.103515625, "learning_rate": 4.7399981261126205e-05, "loss": 9.0625, "step": 10447 }, { "epoch": 0.44050003162088663, "grad_norm": 4278.7255859375, "learning_rate": 4.7399200474718136e-05, "loss": 10.5461, "step": 10448 }, { "epoch": 0.4405421928030862, "grad_norm": 25058.2890625, "learning_rate": 4.739841968831007e-05, "loss": 8.9862, "step": 10449 }, { "epoch": 0.4405843539852857, "grad_norm": 1194.0892333984375, "learning_rate": 4.7397638901902e-05, "loss": 10.038, "step": 10450 }, { "epoch": 0.4406265151674853, "grad_norm": 840.5604248046875, "learning_rate": 4.7396858115493923e-05, "loss": 10.0933, "step": 10451 }, { "epoch": 0.4406686763496848, "grad_norm": 10148.919921875, "learning_rate": 4.739607732908586e-05, "loss": 10.2236, "step": 10452 }, { "epoch": 0.4407108375318844, "grad_norm": 4261.79150390625, "learning_rate": 4.7395296542677786e-05, "loss": 9.6341, "step": 10453 }, { "epoch": 0.44075299871408397, "grad_norm": 836.96630859375, "learning_rate": 4.739451575626972e-05, "loss": 10.0002, "step": 10454 }, { "epoch": 0.4407951598962835, "grad_norm": 789.9462890625, "learning_rate": 4.739373496986165e-05, "loss": 9.8772, "step": 10455 }, { "epoch": 0.44083732107848306, "grad_norm": 7491.56396484375, "learning_rate": 4.739295418345357e-05, "loss": 9.9334, "step": 10456 }, { "epoch": 0.4408794822606826, "grad_norm": 1062.476318359375, "learning_rate": 4.739217339704551e-05, "loss": 9.5563, "step": 10457 }, { "epoch": 0.44092164344288215, "grad_norm": 1391.989990234375, "learning_rate": 4.7391392610637436e-05, "loss": 9.5649, "step": 10458 }, { "epoch": 0.44096380462508167, "grad_norm": 2721.86328125, "learning_rate": 4.739061182422937e-05, "loss": 9.4674, "step": 10459 }, { "epoch": 0.44100596580728124, "grad_norm": 52580.66015625, "learning_rate": 4.73898310378213e-05, "loss": 10.373, "step": 10460 }, { "epoch": 0.44104812698948076, "grad_norm": 3093.583740234375, "learning_rate": 4.738905025141322e-05, "loss": 9.0627, "step": 10461 }, { "epoch": 0.44109028817168033, "grad_norm": 784.5543823242188, "learning_rate": 4.7388269465005154e-05, "loss": 9.1412, "step": 10462 }, { "epoch": 0.4411324493538799, "grad_norm": 1354.7406005859375, "learning_rate": 4.7387488678597085e-05, "loss": 9.7491, "step": 10463 }, { "epoch": 0.4411746105360794, "grad_norm": 1241.0174560546875, "learning_rate": 4.7386707892189016e-05, "loss": 9.615, "step": 10464 }, { "epoch": 0.441216771718279, "grad_norm": 702.9310913085938, "learning_rate": 4.738592710578095e-05, "loss": 9.5333, "step": 10465 }, { "epoch": 0.4412589329004785, "grad_norm": 470.5857849121094, "learning_rate": 4.738514631937287e-05, "loss": 9.9507, "step": 10466 }, { "epoch": 0.4413010940826781, "grad_norm": 9135.3662109375, "learning_rate": 4.73843655329648e-05, "loss": 8.5598, "step": 10467 }, { "epoch": 0.4413432552648776, "grad_norm": 1157.892578125, "learning_rate": 4.7383584746556735e-05, "loss": 8.5416, "step": 10468 }, { "epoch": 0.4413854164470772, "grad_norm": 4811.02734375, "learning_rate": 4.7382803960148666e-05, "loss": 9.4308, "step": 10469 }, { "epoch": 0.4414275776292767, "grad_norm": 5840.2373046875, "learning_rate": 4.738202317374059e-05, "loss": 9.5319, "step": 10470 }, { "epoch": 0.4414697388114763, "grad_norm": 11902.6123046875, "learning_rate": 4.738124238733253e-05, "loss": 8.279, "step": 10471 }, { "epoch": 0.44151189999367585, "grad_norm": 2057.767333984375, "learning_rate": 4.738046160092445e-05, "loss": 8.3339, "step": 10472 }, { "epoch": 0.44155406117587537, "grad_norm": 664.1489868164062, "learning_rate": 4.737968081451638e-05, "loss": 9.3924, "step": 10473 }, { "epoch": 0.44159622235807494, "grad_norm": 3921.63427734375, "learning_rate": 4.7378900028108315e-05, "loss": 8.9426, "step": 10474 }, { "epoch": 0.44163838354027446, "grad_norm": 3895.028076171875, "learning_rate": 4.737811924170024e-05, "loss": 9.864, "step": 10475 }, { "epoch": 0.44168054472247403, "grad_norm": 11006.3232421875, "learning_rate": 4.737733845529218e-05, "loss": 8.3014, "step": 10476 }, { "epoch": 0.44172270590467355, "grad_norm": 3115.930419921875, "learning_rate": 4.73765576688841e-05, "loss": 7.9863, "step": 10477 }, { "epoch": 0.4417648670868731, "grad_norm": 3232.9345703125, "learning_rate": 4.7375776882476034e-05, "loss": 8.4479, "step": 10478 }, { "epoch": 0.44180702826907264, "grad_norm": 8208.994140625, "learning_rate": 4.7374996096067965e-05, "loss": 8.2994, "step": 10479 }, { "epoch": 0.4418491894512722, "grad_norm": 6772.3828125, "learning_rate": 4.737421530965989e-05, "loss": 9.2267, "step": 10480 }, { "epoch": 0.44189135063347174, "grad_norm": 2025.588134765625, "learning_rate": 4.737343452325182e-05, "loss": 9.5697, "step": 10481 }, { "epoch": 0.4419335118156713, "grad_norm": 3599.557373046875, "learning_rate": 4.737265373684375e-05, "loss": 9.893, "step": 10482 }, { "epoch": 0.4419756729978709, "grad_norm": 46141.93359375, "learning_rate": 4.737187295043568e-05, "loss": 8.9274, "step": 10483 }, { "epoch": 0.4420178341800704, "grad_norm": 6525.080078125, "learning_rate": 4.737109216402761e-05, "loss": 9.0484, "step": 10484 }, { "epoch": 0.44205999536227, "grad_norm": 4944.876953125, "learning_rate": 4.737031137761954e-05, "loss": 8.8111, "step": 10485 }, { "epoch": 0.4421021565444695, "grad_norm": 11376.4541015625, "learning_rate": 4.736953059121147e-05, "loss": 8.5685, "step": 10486 }, { "epoch": 0.44214431772666907, "grad_norm": 3499.97802734375, "learning_rate": 4.73687498048034e-05, "loss": 10.1467, "step": 10487 }, { "epoch": 0.4421864789088686, "grad_norm": 7228.89599609375, "learning_rate": 4.736796901839533e-05, "loss": 8.5597, "step": 10488 }, { "epoch": 0.44222864009106816, "grad_norm": 6836.912109375, "learning_rate": 4.736718823198726e-05, "loss": 8.2454, "step": 10489 }, { "epoch": 0.4422708012732677, "grad_norm": 4526.74267578125, "learning_rate": 4.736640744557919e-05, "loss": 8.7137, "step": 10490 }, { "epoch": 0.44231296245546725, "grad_norm": 9857.9052734375, "learning_rate": 4.736562665917112e-05, "loss": 8.6732, "step": 10491 }, { "epoch": 0.4423551236376668, "grad_norm": 7841.14697265625, "learning_rate": 4.7364845872763044e-05, "loss": 8.3598, "step": 10492 }, { "epoch": 0.44239728481986634, "grad_norm": 3737.905517578125, "learning_rate": 4.736406508635498e-05, "loss": 9.3273, "step": 10493 }, { "epoch": 0.4424394460020659, "grad_norm": 3865.007568359375, "learning_rate": 4.736328429994691e-05, "loss": 9.0083, "step": 10494 }, { "epoch": 0.44248160718426544, "grad_norm": 13034.4501953125, "learning_rate": 4.736250351353884e-05, "loss": 8.2102, "step": 10495 }, { "epoch": 0.442523768366465, "grad_norm": 3733.680908203125, "learning_rate": 4.736172272713077e-05, "loss": 9.194, "step": 10496 }, { "epoch": 0.4425659295486645, "grad_norm": 4109.17919921875, "learning_rate": 4.73609419407227e-05, "loss": 9.1419, "step": 10497 }, { "epoch": 0.4426080907308641, "grad_norm": 5372.54541015625, "learning_rate": 4.736016115431463e-05, "loss": 8.9248, "step": 10498 }, { "epoch": 0.4426502519130636, "grad_norm": 4334.6064453125, "learning_rate": 4.7359380367906556e-05, "loss": 8.5613, "step": 10499 }, { "epoch": 0.4426924130952632, "grad_norm": 2571.486328125, "learning_rate": 4.735859958149849e-05, "loss": 9.4371, "step": 10500 }, { "epoch": 0.44273457427746277, "grad_norm": 1585.9100341796875, "learning_rate": 4.735781879509042e-05, "loss": 9.1205, "step": 10501 }, { "epoch": 0.4427767354596623, "grad_norm": 2942.968505859375, "learning_rate": 4.735703800868235e-05, "loss": 9.9101, "step": 10502 }, { "epoch": 0.44281889664186186, "grad_norm": 3173.44970703125, "learning_rate": 4.7356257222274274e-05, "loss": 8.378, "step": 10503 }, { "epoch": 0.4428610578240614, "grad_norm": 17070.451171875, "learning_rate": 4.7355476435866206e-05, "loss": 8.4779, "step": 10504 }, { "epoch": 0.44290321900626095, "grad_norm": 10633.396484375, "learning_rate": 4.735469564945814e-05, "loss": 13.567, "step": 10505 }, { "epoch": 0.44294538018846047, "grad_norm": 3229.012939453125, "learning_rate": 4.735391486305006e-05, "loss": 8.6582, "step": 10506 }, { "epoch": 0.44298754137066004, "grad_norm": 847.8447875976562, "learning_rate": 4.7353134076642e-05, "loss": 8.6296, "step": 10507 }, { "epoch": 0.44302970255285956, "grad_norm": 4239.4970703125, "learning_rate": 4.7352353290233924e-05, "loss": 8.6608, "step": 10508 }, { "epoch": 0.44307186373505913, "grad_norm": 8773.8798828125, "learning_rate": 4.7351572503825855e-05, "loss": 8.8491, "step": 10509 }, { "epoch": 0.44311402491725865, "grad_norm": 55100.83203125, "learning_rate": 4.7350791717417787e-05, "loss": 9.428, "step": 10510 }, { "epoch": 0.4431561860994582, "grad_norm": 3567.185546875, "learning_rate": 4.735001093100971e-05, "loss": 9.4046, "step": 10511 }, { "epoch": 0.4431983472816578, "grad_norm": 7529.8466796875, "learning_rate": 4.734923014460165e-05, "loss": 8.3031, "step": 10512 }, { "epoch": 0.4432405084638573, "grad_norm": 6023.3271484375, "learning_rate": 4.7348449358193574e-05, "loss": 8.2356, "step": 10513 }, { "epoch": 0.4432826696460569, "grad_norm": 1963.764404296875, "learning_rate": 4.7347668571785505e-05, "loss": 9.6321, "step": 10514 }, { "epoch": 0.4433248308282564, "grad_norm": 13571.9736328125, "learning_rate": 4.7346887785377436e-05, "loss": 9.3426, "step": 10515 }, { "epoch": 0.443366992010456, "grad_norm": 7114.953125, "learning_rate": 4.734610699896937e-05, "loss": 8.5139, "step": 10516 }, { "epoch": 0.4434091531926555, "grad_norm": 19509.240234375, "learning_rate": 4.734532621256129e-05, "loss": 10.5138, "step": 10517 }, { "epoch": 0.4434513143748551, "grad_norm": 26078.228515625, "learning_rate": 4.734454542615322e-05, "loss": 8.9636, "step": 10518 }, { "epoch": 0.4434934755570546, "grad_norm": 23331.310546875, "learning_rate": 4.7343764639745154e-05, "loss": 9.0098, "step": 10519 }, { "epoch": 0.44353563673925417, "grad_norm": 55615.71484375, "learning_rate": 4.734298385333708e-05, "loss": 8.9373, "step": 10520 }, { "epoch": 0.44357779792145374, "grad_norm": 17584.955078125, "learning_rate": 4.734220306692902e-05, "loss": 8.6337, "step": 10521 }, { "epoch": 0.44361995910365326, "grad_norm": 19383.958984375, "learning_rate": 4.734142228052094e-05, "loss": 11.0745, "step": 10522 }, { "epoch": 0.44366212028585283, "grad_norm": 16950.45703125, "learning_rate": 4.734064149411287e-05, "loss": 9.0041, "step": 10523 }, { "epoch": 0.44370428146805235, "grad_norm": 12402.9990234375, "learning_rate": 4.7339860707704804e-05, "loss": 10.1802, "step": 10524 }, { "epoch": 0.4437464426502519, "grad_norm": 15738.72265625, "learning_rate": 4.733907992129673e-05, "loss": 8.7698, "step": 10525 }, { "epoch": 0.44378860383245144, "grad_norm": 8312.83203125, "learning_rate": 4.7338299134888666e-05, "loss": 9.8561, "step": 10526 }, { "epoch": 0.443830765014651, "grad_norm": 7802.30224609375, "learning_rate": 4.733751834848059e-05, "loss": 8.6316, "step": 10527 }, { "epoch": 0.44387292619685054, "grad_norm": 1962.473876953125, "learning_rate": 4.733673756207252e-05, "loss": 9.2129, "step": 10528 }, { "epoch": 0.4439150873790501, "grad_norm": 33822.1796875, "learning_rate": 4.733595677566445e-05, "loss": 9.1863, "step": 10529 }, { "epoch": 0.4439572485612497, "grad_norm": 14094.2392578125, "learning_rate": 4.733517598925638e-05, "loss": 10.2024, "step": 10530 }, { "epoch": 0.4439994097434492, "grad_norm": 5565.990234375, "learning_rate": 4.733439520284831e-05, "loss": 8.6875, "step": 10531 }, { "epoch": 0.4440415709256488, "grad_norm": 772.3516845703125, "learning_rate": 4.733361441644024e-05, "loss": 8.7247, "step": 10532 }, { "epoch": 0.4440837321078483, "grad_norm": 11167.5029296875, "learning_rate": 4.733283363003217e-05, "loss": 9.0933, "step": 10533 }, { "epoch": 0.44412589329004787, "grad_norm": 10726.8955078125, "learning_rate": 4.73320528436241e-05, "loss": 9.883, "step": 10534 }, { "epoch": 0.4441680544722474, "grad_norm": 1903.0081787109375, "learning_rate": 4.733127205721603e-05, "loss": 9.3439, "step": 10535 }, { "epoch": 0.44421021565444696, "grad_norm": 20721.455078125, "learning_rate": 4.733049127080796e-05, "loss": 9.1829, "step": 10536 }, { "epoch": 0.4442523768366465, "grad_norm": 7098.666015625, "learning_rate": 4.732971048439989e-05, "loss": 9.0846, "step": 10537 }, { "epoch": 0.44429453801884605, "grad_norm": 1876.95947265625, "learning_rate": 4.732892969799182e-05, "loss": 9.0809, "step": 10538 }, { "epoch": 0.44433669920104557, "grad_norm": 2771.97021484375, "learning_rate": 4.7328148911583746e-05, "loss": 9.6648, "step": 10539 }, { "epoch": 0.44437886038324514, "grad_norm": 12950.9931640625, "learning_rate": 4.7327368125175684e-05, "loss": 8.7664, "step": 10540 }, { "epoch": 0.4444210215654447, "grad_norm": 714.35546875, "learning_rate": 4.732658733876761e-05, "loss": 8.6374, "step": 10541 }, { "epoch": 0.44446318274764424, "grad_norm": 7903.68798828125, "learning_rate": 4.732580655235954e-05, "loss": 9.9144, "step": 10542 }, { "epoch": 0.4445053439298438, "grad_norm": 2361.375732421875, "learning_rate": 4.732502576595147e-05, "loss": 9.6729, "step": 10543 }, { "epoch": 0.44454750511204333, "grad_norm": 13413.0615234375, "learning_rate": 4.7324244979543395e-05, "loss": 8.7448, "step": 10544 }, { "epoch": 0.4445896662942429, "grad_norm": 3431.4248046875, "learning_rate": 4.732346419313533e-05, "loss": 9.6724, "step": 10545 }, { "epoch": 0.4446318274764424, "grad_norm": 1351.148681640625, "learning_rate": 4.732268340672726e-05, "loss": 8.7031, "step": 10546 }, { "epoch": 0.444673988658642, "grad_norm": 16252.3505859375, "learning_rate": 4.732190262031919e-05, "loss": 8.6706, "step": 10547 }, { "epoch": 0.4447161498408415, "grad_norm": 2196.260986328125, "learning_rate": 4.732112183391112e-05, "loss": 8.4845, "step": 10548 }, { "epoch": 0.4447583110230411, "grad_norm": 22940.015625, "learning_rate": 4.7320341047503045e-05, "loss": 8.4302, "step": 10549 }, { "epoch": 0.44480047220524066, "grad_norm": 36947.05078125, "learning_rate": 4.7319560261094976e-05, "loss": 8.7543, "step": 10550 }, { "epoch": 0.4448426333874402, "grad_norm": 8273.2509765625, "learning_rate": 4.731877947468691e-05, "loss": 8.8122, "step": 10551 }, { "epoch": 0.44488479456963975, "grad_norm": 14900.3984375, "learning_rate": 4.731799868827884e-05, "loss": 9.8493, "step": 10552 }, { "epoch": 0.44492695575183927, "grad_norm": 3589.6044921875, "learning_rate": 4.731721790187076e-05, "loss": 8.4196, "step": 10553 }, { "epoch": 0.44496911693403884, "grad_norm": 6236.486328125, "learning_rate": 4.7316437115462694e-05, "loss": 8.6289, "step": 10554 }, { "epoch": 0.44501127811623836, "grad_norm": 209813.734375, "learning_rate": 4.7315656329054625e-05, "loss": 9.4777, "step": 10555 }, { "epoch": 0.44505343929843794, "grad_norm": 9642.7568359375, "learning_rate": 4.731487554264656e-05, "loss": 9.9512, "step": 10556 }, { "epoch": 0.44509560048063745, "grad_norm": 3100.9716796875, "learning_rate": 4.731409475623849e-05, "loss": 8.6079, "step": 10557 }, { "epoch": 0.44513776166283703, "grad_norm": 18455.357421875, "learning_rate": 4.731331396983041e-05, "loss": 10.522, "step": 10558 }, { "epoch": 0.4451799228450366, "grad_norm": 4544.892578125, "learning_rate": 4.731253318342235e-05, "loss": 8.8666, "step": 10559 }, { "epoch": 0.4452220840272361, "grad_norm": 947.9598999023438, "learning_rate": 4.7311752397014275e-05, "loss": 9.2624, "step": 10560 }, { "epoch": 0.4452642452094357, "grad_norm": 39006.3515625, "learning_rate": 4.7310971610606206e-05, "loss": 8.3931, "step": 10561 }, { "epoch": 0.4453064063916352, "grad_norm": 7627.2138671875, "learning_rate": 4.731019082419814e-05, "loss": 10.1737, "step": 10562 }, { "epoch": 0.4453485675738348, "grad_norm": 2299.39453125, "learning_rate": 4.730941003779006e-05, "loss": 8.615, "step": 10563 }, { "epoch": 0.4453907287560343, "grad_norm": 1627.49951171875, "learning_rate": 4.730862925138199e-05, "loss": 8.4215, "step": 10564 }, { "epoch": 0.4454328899382339, "grad_norm": 4609251.5, "learning_rate": 4.7307848464973924e-05, "loss": 8.9368, "step": 10565 }, { "epoch": 0.4454750511204334, "grad_norm": 9131.0390625, "learning_rate": 4.7307067678565856e-05, "loss": 8.9475, "step": 10566 }, { "epoch": 0.44551721230263297, "grad_norm": 12232.33203125, "learning_rate": 4.730628689215779e-05, "loss": 9.6143, "step": 10567 }, { "epoch": 0.4455593734848325, "grad_norm": 42101.578125, "learning_rate": 4.730550610574971e-05, "loss": 10.5627, "step": 10568 }, { "epoch": 0.44560153466703206, "grad_norm": 5742.853515625, "learning_rate": 4.730472531934164e-05, "loss": 9.8906, "step": 10569 }, { "epoch": 0.44564369584923164, "grad_norm": 2369.3369140625, "learning_rate": 4.7303944532933574e-05, "loss": 9.85, "step": 10570 }, { "epoch": 0.44568585703143115, "grad_norm": 6266.07421875, "learning_rate": 4.7303163746525505e-05, "loss": 8.6562, "step": 10571 }, { "epoch": 0.44572801821363073, "grad_norm": 1280.5059814453125, "learning_rate": 4.730238296011743e-05, "loss": 8.7959, "step": 10572 }, { "epoch": 0.44577017939583025, "grad_norm": 10186.478515625, "learning_rate": 4.730160217370936e-05, "loss": 9.892, "step": 10573 }, { "epoch": 0.4458123405780298, "grad_norm": 42282.328125, "learning_rate": 4.730082138730129e-05, "loss": 8.7605, "step": 10574 }, { "epoch": 0.44585450176022934, "grad_norm": 1318556.25, "learning_rate": 4.730004060089322e-05, "loss": 9.2944, "step": 10575 }, { "epoch": 0.4458966629424289, "grad_norm": 13465.228515625, "learning_rate": 4.7299259814485155e-05, "loss": 8.9053, "step": 10576 }, { "epoch": 0.44593882412462843, "grad_norm": 5301.048828125, "learning_rate": 4.729847902807708e-05, "loss": 8.8259, "step": 10577 }, { "epoch": 0.445980985306828, "grad_norm": 9526.97265625, "learning_rate": 4.729769824166902e-05, "loss": 8.7782, "step": 10578 }, { "epoch": 0.4460231464890276, "grad_norm": 29572.837890625, "learning_rate": 4.729691745526094e-05, "loss": 8.8299, "step": 10579 }, { "epoch": 0.4460653076712271, "grad_norm": 2374.33154296875, "learning_rate": 4.7296136668852866e-05, "loss": 9.9895, "step": 10580 }, { "epoch": 0.44610746885342667, "grad_norm": 4514.4443359375, "learning_rate": 4.7295355882444804e-05, "loss": 9.0948, "step": 10581 }, { "epoch": 0.4461496300356262, "grad_norm": 6324.2412109375, "learning_rate": 4.729457509603673e-05, "loss": 9.6438, "step": 10582 }, { "epoch": 0.44619179121782576, "grad_norm": 34235.3984375, "learning_rate": 4.729379430962866e-05, "loss": 9.1316, "step": 10583 }, { "epoch": 0.4462339524000253, "grad_norm": 135991.8125, "learning_rate": 4.729301352322059e-05, "loss": 10.4308, "step": 10584 }, { "epoch": 0.44627611358222485, "grad_norm": 4964.494140625, "learning_rate": 4.729223273681252e-05, "loss": 9.5848, "step": 10585 }, { "epoch": 0.44631827476442437, "grad_norm": 15956.2880859375, "learning_rate": 4.729145195040445e-05, "loss": 8.7357, "step": 10586 }, { "epoch": 0.44636043594662395, "grad_norm": 8664.22265625, "learning_rate": 4.729067116399638e-05, "loss": 9.2968, "step": 10587 }, { "epoch": 0.4464025971288235, "grad_norm": 17483.1875, "learning_rate": 4.728989037758831e-05, "loss": 8.8135, "step": 10588 }, { "epoch": 0.44644475831102304, "grad_norm": 1718717.25, "learning_rate": 4.728910959118024e-05, "loss": 9.9731, "step": 10589 }, { "epoch": 0.4464869194932226, "grad_norm": 100223.5078125, "learning_rate": 4.728832880477217e-05, "loss": 8.8477, "step": 10590 }, { "epoch": 0.44652908067542213, "grad_norm": 67804.40625, "learning_rate": 4.7287548018364097e-05, "loss": 8.6615, "step": 10591 }, { "epoch": 0.4465712418576217, "grad_norm": 23663.634765625, "learning_rate": 4.728676723195603e-05, "loss": 9.3817, "step": 10592 }, { "epoch": 0.4466134030398212, "grad_norm": 29814.658203125, "learning_rate": 4.728598644554796e-05, "loss": 9.1874, "step": 10593 }, { "epoch": 0.4466555642220208, "grad_norm": 209709.03125, "learning_rate": 4.7285205659139884e-05, "loss": 9.3036, "step": 10594 }, { "epoch": 0.4466977254042203, "grad_norm": 86786.25, "learning_rate": 4.728442487273182e-05, "loss": 8.7169, "step": 10595 }, { "epoch": 0.4467398865864199, "grad_norm": 14691.546875, "learning_rate": 4.7283644086323746e-05, "loss": 9.6128, "step": 10596 }, { "epoch": 0.4467820477686194, "grad_norm": 7115.640625, "learning_rate": 4.728286329991568e-05, "loss": 9.0783, "step": 10597 }, { "epoch": 0.446824208950819, "grad_norm": 36078.79296875, "learning_rate": 4.728208251350761e-05, "loss": 9.4795, "step": 10598 }, { "epoch": 0.44686637013301855, "grad_norm": 8334.6474609375, "learning_rate": 4.728130172709953e-05, "loss": 9.4261, "step": 10599 }, { "epoch": 0.44690853131521807, "grad_norm": 11122.974609375, "learning_rate": 4.7280520940691464e-05, "loss": 10.2991, "step": 10600 }, { "epoch": 0.44695069249741765, "grad_norm": 13311.8994140625, "learning_rate": 4.7279740154283396e-05, "loss": 9.0848, "step": 10601 }, { "epoch": 0.44699285367961716, "grad_norm": 16699.875, "learning_rate": 4.727895936787533e-05, "loss": 11.202, "step": 10602 }, { "epoch": 0.44703501486181674, "grad_norm": 6244.12451171875, "learning_rate": 4.727817858146726e-05, "loss": 8.8261, "step": 10603 }, { "epoch": 0.44707717604401626, "grad_norm": 17498.845703125, "learning_rate": 4.727739779505919e-05, "loss": 9.0578, "step": 10604 }, { "epoch": 0.44711933722621583, "grad_norm": 26204.23828125, "learning_rate": 4.7276617008651114e-05, "loss": 10.1727, "step": 10605 }, { "epoch": 0.44716149840841535, "grad_norm": 48996.60546875, "learning_rate": 4.7275836222243045e-05, "loss": 9.5986, "step": 10606 }, { "epoch": 0.4472036595906149, "grad_norm": 18466.759765625, "learning_rate": 4.7275055435834976e-05, "loss": 9.2093, "step": 10607 }, { "epoch": 0.4472458207728145, "grad_norm": 15728.8828125, "learning_rate": 4.72742746494269e-05, "loss": 9.0084, "step": 10608 }, { "epoch": 0.447287981955014, "grad_norm": 9154.83984375, "learning_rate": 4.727349386301884e-05, "loss": 9.8954, "step": 10609 }, { "epoch": 0.4473301431372136, "grad_norm": 3014.171875, "learning_rate": 4.7272713076610763e-05, "loss": 9.986, "step": 10610 }, { "epoch": 0.4473723043194131, "grad_norm": 23979.263671875, "learning_rate": 4.7271932290202695e-05, "loss": 9.4732, "step": 10611 }, { "epoch": 0.4474144655016127, "grad_norm": 5248.86279296875, "learning_rate": 4.7271151503794626e-05, "loss": 10.1294, "step": 10612 }, { "epoch": 0.4474566266838122, "grad_norm": 10775.806640625, "learning_rate": 4.727037071738655e-05, "loss": 8.7178, "step": 10613 }, { "epoch": 0.44749878786601177, "grad_norm": 217733.453125, "learning_rate": 4.726958993097849e-05, "loss": 8.6204, "step": 10614 }, { "epoch": 0.4475409490482113, "grad_norm": 1395.4798583984375, "learning_rate": 4.726880914457041e-05, "loss": 8.5113, "step": 10615 }, { "epoch": 0.44758311023041086, "grad_norm": 9178.2841796875, "learning_rate": 4.7268028358162344e-05, "loss": 10.0945, "step": 10616 }, { "epoch": 0.44762527141261044, "grad_norm": 442963.09375, "learning_rate": 4.7267247571754275e-05, "loss": 8.5256, "step": 10617 }, { "epoch": 0.44766743259480996, "grad_norm": 17902.236328125, "learning_rate": 4.72664667853462e-05, "loss": 8.9589, "step": 10618 }, { "epoch": 0.44770959377700953, "grad_norm": 22536.939453125, "learning_rate": 4.726568599893813e-05, "loss": 8.6542, "step": 10619 }, { "epoch": 0.44775175495920905, "grad_norm": 990.9450073242188, "learning_rate": 4.726490521253006e-05, "loss": 9.3801, "step": 10620 }, { "epoch": 0.4477939161414086, "grad_norm": 2089.07861328125, "learning_rate": 4.7264124426121994e-05, "loss": 8.6044, "step": 10621 }, { "epoch": 0.44783607732360814, "grad_norm": 22889.05078125, "learning_rate": 4.726334363971392e-05, "loss": 9.6791, "step": 10622 }, { "epoch": 0.4478782385058077, "grad_norm": 19439.23046875, "learning_rate": 4.7262562853305856e-05, "loss": 9.6233, "step": 10623 }, { "epoch": 0.44792039968800723, "grad_norm": 109850.203125, "learning_rate": 4.726178206689778e-05, "loss": 9.1275, "step": 10624 }, { "epoch": 0.4479625608702068, "grad_norm": 39545.60546875, "learning_rate": 4.726100128048971e-05, "loss": 9.3992, "step": 10625 }, { "epoch": 0.4480047220524063, "grad_norm": 58229.5625, "learning_rate": 4.726022049408164e-05, "loss": 10.1354, "step": 10626 }, { "epoch": 0.4480468832346059, "grad_norm": 33117.046875, "learning_rate": 4.725943970767357e-05, "loss": 9.9381, "step": 10627 }, { "epoch": 0.44808904441680547, "grad_norm": 13755.724609375, "learning_rate": 4.7258658921265506e-05, "loss": 9.386, "step": 10628 }, { "epoch": 0.448131205599005, "grad_norm": 24690.865234375, "learning_rate": 4.725787813485743e-05, "loss": 10.287, "step": 10629 }, { "epoch": 0.44817336678120456, "grad_norm": 1919.12841796875, "learning_rate": 4.725709734844936e-05, "loss": 9.2608, "step": 10630 }, { "epoch": 0.4482155279634041, "grad_norm": 64900.9140625, "learning_rate": 4.725631656204129e-05, "loss": 9.1845, "step": 10631 }, { "epoch": 0.44825768914560365, "grad_norm": 1362.92041015625, "learning_rate": 4.725553577563322e-05, "loss": 10.0716, "step": 10632 }, { "epoch": 0.4482998503278032, "grad_norm": 4637.53173828125, "learning_rate": 4.725475498922515e-05, "loss": 9.3561, "step": 10633 }, { "epoch": 0.44834201151000275, "grad_norm": 555082.0625, "learning_rate": 4.725397420281708e-05, "loss": 9.6319, "step": 10634 }, { "epoch": 0.44838417269220227, "grad_norm": 2414.658935546875, "learning_rate": 4.725319341640901e-05, "loss": 9.9041, "step": 10635 }, { "epoch": 0.44842633387440184, "grad_norm": 7262.0693359375, "learning_rate": 4.725241263000094e-05, "loss": 11.6595, "step": 10636 }, { "epoch": 0.4484684950566014, "grad_norm": 1746.0673828125, "learning_rate": 4.725163184359287e-05, "loss": 11.4847, "step": 10637 }, { "epoch": 0.44851065623880093, "grad_norm": 2954.736328125, "learning_rate": 4.72508510571848e-05, "loss": 9.4994, "step": 10638 }, { "epoch": 0.4485528174210005, "grad_norm": 38441.01953125, "learning_rate": 4.725007027077673e-05, "loss": 9.3452, "step": 10639 }, { "epoch": 0.4485949786032, "grad_norm": 12475.5107421875, "learning_rate": 4.724928948436866e-05, "loss": 9.6282, "step": 10640 }, { "epoch": 0.4486371397853996, "grad_norm": 36664.8671875, "learning_rate": 4.7248508697960585e-05, "loss": 9.6928, "step": 10641 }, { "epoch": 0.4486793009675991, "grad_norm": 10396.517578125, "learning_rate": 4.7247727911552516e-05, "loss": 9.4333, "step": 10642 }, { "epoch": 0.4487214621497987, "grad_norm": 2735.427001953125, "learning_rate": 4.724694712514445e-05, "loss": 9.2677, "step": 10643 }, { "epoch": 0.4487636233319982, "grad_norm": 55167.515625, "learning_rate": 4.724616633873637e-05, "loss": 10.6579, "step": 10644 }, { "epoch": 0.4488057845141978, "grad_norm": 4148.31787109375, "learning_rate": 4.724538555232831e-05, "loss": 10.7007, "step": 10645 }, { "epoch": 0.44884794569639735, "grad_norm": 715.0469970703125, "learning_rate": 4.7244604765920235e-05, "loss": 8.9454, "step": 10646 }, { "epoch": 0.4488901068785969, "grad_norm": 776.0418701171875, "learning_rate": 4.724382397951217e-05, "loss": 10.2717, "step": 10647 }, { "epoch": 0.44893226806079645, "grad_norm": 15490.6015625, "learning_rate": 4.72430431931041e-05, "loss": 9.29, "step": 10648 }, { "epoch": 0.44897442924299596, "grad_norm": 3672.82373046875, "learning_rate": 4.724226240669603e-05, "loss": 9.0412, "step": 10649 }, { "epoch": 0.44901659042519554, "grad_norm": 790.3472900390625, "learning_rate": 4.724148162028796e-05, "loss": 9.0669, "step": 10650 }, { "epoch": 0.44905875160739506, "grad_norm": 1896.9600830078125, "learning_rate": 4.7240700833879884e-05, "loss": 9.0353, "step": 10651 }, { "epoch": 0.44910091278959463, "grad_norm": 4888.619140625, "learning_rate": 4.7239920047471815e-05, "loss": 8.8106, "step": 10652 }, { "epoch": 0.44914307397179415, "grad_norm": 17961.94140625, "learning_rate": 4.7239139261063747e-05, "loss": 9.0448, "step": 10653 }, { "epoch": 0.4491852351539937, "grad_norm": 672.5504760742188, "learning_rate": 4.723835847465568e-05, "loss": 8.8068, "step": 10654 }, { "epoch": 0.44922739633619324, "grad_norm": 3560.441162109375, "learning_rate": 4.72375776882476e-05, "loss": 8.0973, "step": 10655 }, { "epoch": 0.4492695575183928, "grad_norm": 6241.51953125, "learning_rate": 4.7236796901839534e-05, "loss": 8.4542, "step": 10656 }, { "epoch": 0.4493117187005924, "grad_norm": 5909.08203125, "learning_rate": 4.7236016115431465e-05, "loss": 8.1756, "step": 10657 }, { "epoch": 0.4493538798827919, "grad_norm": 8008.2705078125, "learning_rate": 4.7235235329023396e-05, "loss": 8.4978, "step": 10658 }, { "epoch": 0.4493960410649915, "grad_norm": 24578.408203125, "learning_rate": 4.723445454261533e-05, "loss": 9.2385, "step": 10659 }, { "epoch": 0.449438202247191, "grad_norm": 8946.177734375, "learning_rate": 4.723367375620725e-05, "loss": 7.8409, "step": 10660 }, { "epoch": 0.4494803634293906, "grad_norm": 8548.0986328125, "learning_rate": 4.723289296979918e-05, "loss": 8.4744, "step": 10661 }, { "epoch": 0.4495225246115901, "grad_norm": 275063.3125, "learning_rate": 4.7232112183391114e-05, "loss": 8.4099, "step": 10662 }, { "epoch": 0.44956468579378966, "grad_norm": 4659.79052734375, "learning_rate": 4.723133139698304e-05, "loss": 8.5162, "step": 10663 }, { "epoch": 0.4496068469759892, "grad_norm": 2797.260009765625, "learning_rate": 4.723055061057498e-05, "loss": 9.037, "step": 10664 }, { "epoch": 0.44964900815818876, "grad_norm": 408956.65625, "learning_rate": 4.72297698241669e-05, "loss": 9.2755, "step": 10665 }, { "epoch": 0.44969116934038833, "grad_norm": 56255.546875, "learning_rate": 4.722898903775883e-05, "loss": 8.3383, "step": 10666 }, { "epoch": 0.44973333052258785, "grad_norm": 3802.891845703125, "learning_rate": 4.7228208251350764e-05, "loss": 9.0661, "step": 10667 }, { "epoch": 0.4497754917047874, "grad_norm": 42302.57421875, "learning_rate": 4.722742746494269e-05, "loss": 8.1866, "step": 10668 }, { "epoch": 0.44981765288698694, "grad_norm": 32230.392578125, "learning_rate": 4.7226646678534626e-05, "loss": 8.8294, "step": 10669 }, { "epoch": 0.4498598140691865, "grad_norm": 2828.070556640625, "learning_rate": 4.722586589212655e-05, "loss": 9.1834, "step": 10670 }, { "epoch": 0.44990197525138603, "grad_norm": 245346.78125, "learning_rate": 4.722508510571848e-05, "loss": 9.244, "step": 10671 }, { "epoch": 0.4499441364335856, "grad_norm": 3372.307861328125, "learning_rate": 4.7224304319310413e-05, "loss": 9.4287, "step": 10672 }, { "epoch": 0.4499862976157851, "grad_norm": 3012.452880859375, "learning_rate": 4.7223523532902345e-05, "loss": 10.105, "step": 10673 }, { "epoch": 0.4500284587979847, "grad_norm": 5478.04052734375, "learning_rate": 4.722274274649427e-05, "loss": 8.3446, "step": 10674 }, { "epoch": 0.4500706199801842, "grad_norm": 98846.84375, "learning_rate": 4.72219619600862e-05, "loss": 10.1479, "step": 10675 }, { "epoch": 0.4501127811623838, "grad_norm": 26665.486328125, "learning_rate": 4.722118117367813e-05, "loss": 8.2345, "step": 10676 }, { "epoch": 0.45015494234458336, "grad_norm": 9034.8330078125, "learning_rate": 4.7220400387270056e-05, "loss": 9.3486, "step": 10677 }, { "epoch": 0.4501971035267829, "grad_norm": 14859.4970703125, "learning_rate": 4.7219619600861994e-05, "loss": 9.3275, "step": 10678 }, { "epoch": 0.45023926470898246, "grad_norm": 20934802.0, "learning_rate": 4.721883881445392e-05, "loss": 9.5933, "step": 10679 }, { "epoch": 0.450281425891182, "grad_norm": 2341.81689453125, "learning_rate": 4.721805802804585e-05, "loss": 11.5124, "step": 10680 }, { "epoch": 0.45032358707338155, "grad_norm": 38487.75390625, "learning_rate": 4.721727724163778e-05, "loss": 8.8577, "step": 10681 }, { "epoch": 0.45036574825558107, "grad_norm": 5400.28076171875, "learning_rate": 4.7216496455229706e-05, "loss": 8.1161, "step": 10682 }, { "epoch": 0.45040790943778064, "grad_norm": 3585.52587890625, "learning_rate": 4.7215715668821644e-05, "loss": 9.5322, "step": 10683 }, { "epoch": 0.45045007061998016, "grad_norm": 15880.1708984375, "learning_rate": 4.721493488241357e-05, "loss": 8.4543, "step": 10684 }, { "epoch": 0.45049223180217973, "grad_norm": 5282.9130859375, "learning_rate": 4.72141540960055e-05, "loss": 8.8622, "step": 10685 }, { "epoch": 0.4505343929843793, "grad_norm": 4029.440185546875, "learning_rate": 4.721337330959743e-05, "loss": 9.3681, "step": 10686 }, { "epoch": 0.4505765541665788, "grad_norm": 5646.58056640625, "learning_rate": 4.7212592523189355e-05, "loss": 8.5427, "step": 10687 }, { "epoch": 0.4506187153487784, "grad_norm": 5374.275390625, "learning_rate": 4.7211811736781286e-05, "loss": 10.1551, "step": 10688 }, { "epoch": 0.4506608765309779, "grad_norm": 16547.90625, "learning_rate": 4.721103095037322e-05, "loss": 9.2565, "step": 10689 }, { "epoch": 0.4507030377131775, "grad_norm": 2513.5341796875, "learning_rate": 4.721025016396515e-05, "loss": 8.4211, "step": 10690 }, { "epoch": 0.450745198895377, "grad_norm": 520.1073608398438, "learning_rate": 4.7209469377557073e-05, "loss": 8.2807, "step": 10691 }, { "epoch": 0.4507873600775766, "grad_norm": 1478.412353515625, "learning_rate": 4.720868859114901e-05, "loss": 8.7632, "step": 10692 }, { "epoch": 0.4508295212597761, "grad_norm": 156222.375, "learning_rate": 4.7207907804740936e-05, "loss": 10.2776, "step": 10693 }, { "epoch": 0.4508716824419757, "grad_norm": 22177.3046875, "learning_rate": 4.720712701833287e-05, "loss": 8.418, "step": 10694 }, { "epoch": 0.45091384362417525, "grad_norm": 740.6395263671875, "learning_rate": 4.72063462319248e-05, "loss": 10.5518, "step": 10695 }, { "epoch": 0.45095600480637477, "grad_norm": 43783.55078125, "learning_rate": 4.720556544551672e-05, "loss": 8.1815, "step": 10696 }, { "epoch": 0.45099816598857434, "grad_norm": 1441.464111328125, "learning_rate": 4.720478465910866e-05, "loss": 8.3164, "step": 10697 }, { "epoch": 0.45104032717077386, "grad_norm": 41572.3671875, "learning_rate": 4.7204003872700586e-05, "loss": 8.9152, "step": 10698 }, { "epoch": 0.45108248835297343, "grad_norm": 2629.708251953125, "learning_rate": 4.720322308629252e-05, "loss": 9.5994, "step": 10699 }, { "epoch": 0.45112464953517295, "grad_norm": 81198.4140625, "learning_rate": 4.720244229988445e-05, "loss": 9.2225, "step": 10700 }, { "epoch": 0.4511668107173725, "grad_norm": 7256.07861328125, "learning_rate": 4.720166151347637e-05, "loss": 9.7016, "step": 10701 }, { "epoch": 0.45120897189957204, "grad_norm": 17647.05078125, "learning_rate": 4.7200880727068304e-05, "loss": 8.9113, "step": 10702 }, { "epoch": 0.4512511330817716, "grad_norm": 10958.4921875, "learning_rate": 4.7200099940660235e-05, "loss": 9.5167, "step": 10703 }, { "epoch": 0.45129329426397113, "grad_norm": 4875261.5, "learning_rate": 4.7199319154252166e-05, "loss": 8.6472, "step": 10704 }, { "epoch": 0.4513354554461707, "grad_norm": 51116.7734375, "learning_rate": 4.71985383678441e-05, "loss": 9.1894, "step": 10705 }, { "epoch": 0.4513776166283703, "grad_norm": 14349.986328125, "learning_rate": 4.719775758143602e-05, "loss": 10.3827, "step": 10706 }, { "epoch": 0.4514197778105698, "grad_norm": 17248.51953125, "learning_rate": 4.719697679502795e-05, "loss": 8.6655, "step": 10707 }, { "epoch": 0.4514619389927694, "grad_norm": 657.2952270507812, "learning_rate": 4.7196196008619885e-05, "loss": 9.7299, "step": 10708 }, { "epoch": 0.4515041001749689, "grad_norm": 4032.85400390625, "learning_rate": 4.7195415222211816e-05, "loss": 8.8239, "step": 10709 }, { "epoch": 0.45154626135716847, "grad_norm": 6990.6279296875, "learning_rate": 4.719463443580374e-05, "loss": 8.8701, "step": 10710 }, { "epoch": 0.451588422539368, "grad_norm": 15472.1787109375, "learning_rate": 4.719385364939568e-05, "loss": 12.0389, "step": 10711 }, { "epoch": 0.45163058372156756, "grad_norm": 59532.6328125, "learning_rate": 4.71930728629876e-05, "loss": 9.5191, "step": 10712 }, { "epoch": 0.4516727449037671, "grad_norm": 38414.95703125, "learning_rate": 4.719229207657953e-05, "loss": 9.736, "step": 10713 }, { "epoch": 0.45171490608596665, "grad_norm": 5957404.0, "learning_rate": 4.7191511290171465e-05, "loss": 9.7326, "step": 10714 }, { "epoch": 0.4517570672681662, "grad_norm": 6468.38134765625, "learning_rate": 4.719073050376339e-05, "loss": 9.3521, "step": 10715 }, { "epoch": 0.45179922845036574, "grad_norm": 8966.5966796875, "learning_rate": 4.718994971735533e-05, "loss": 9.758, "step": 10716 }, { "epoch": 0.4518413896325653, "grad_norm": 3250.064697265625, "learning_rate": 4.718916893094725e-05, "loss": 9.0946, "step": 10717 }, { "epoch": 0.45188355081476483, "grad_norm": 70482.96875, "learning_rate": 4.7188388144539184e-05, "loss": 10.9535, "step": 10718 }, { "epoch": 0.4519257119969644, "grad_norm": 9025.474609375, "learning_rate": 4.7187607358131115e-05, "loss": 9.2901, "step": 10719 }, { "epoch": 0.4519678731791639, "grad_norm": 8999.5185546875, "learning_rate": 4.718682657172304e-05, "loss": 8.7322, "step": 10720 }, { "epoch": 0.4520100343613635, "grad_norm": 1878.4874267578125, "learning_rate": 4.718604578531497e-05, "loss": 8.6564, "step": 10721 }, { "epoch": 0.452052195543563, "grad_norm": 4215.6025390625, "learning_rate": 4.71852649989069e-05, "loss": 8.575, "step": 10722 }, { "epoch": 0.4520943567257626, "grad_norm": 23057.580078125, "learning_rate": 4.718448421249883e-05, "loss": 10.6654, "step": 10723 }, { "epoch": 0.45213651790796217, "grad_norm": 34122.6015625, "learning_rate": 4.718370342609076e-05, "loss": 9.1257, "step": 10724 }, { "epoch": 0.4521786790901617, "grad_norm": 13619.39453125, "learning_rate": 4.718292263968269e-05, "loss": 9.4519, "step": 10725 }, { "epoch": 0.45222084027236126, "grad_norm": 9952.9580078125, "learning_rate": 4.718214185327462e-05, "loss": 9.2895, "step": 10726 }, { "epoch": 0.4522630014545608, "grad_norm": 7771.87939453125, "learning_rate": 4.718136106686655e-05, "loss": 9.0057, "step": 10727 }, { "epoch": 0.45230516263676035, "grad_norm": 15940.0634765625, "learning_rate": 4.718058028045848e-05, "loss": 8.9433, "step": 10728 }, { "epoch": 0.45234732381895987, "grad_norm": 4270.27880859375, "learning_rate": 4.717979949405041e-05, "loss": 10.1258, "step": 10729 }, { "epoch": 0.45238948500115944, "grad_norm": 6363.0751953125, "learning_rate": 4.717901870764234e-05, "loss": 10.0041, "step": 10730 }, { "epoch": 0.45243164618335896, "grad_norm": 52221.1640625, "learning_rate": 4.717823792123427e-05, "loss": 9.2868, "step": 10731 }, { "epoch": 0.45247380736555853, "grad_norm": 2614315.5, "learning_rate": 4.7177457134826194e-05, "loss": 10.1849, "step": 10732 }, { "epoch": 0.45251596854775805, "grad_norm": 142708.53125, "learning_rate": 4.717667634841813e-05, "loss": 9.277, "step": 10733 }, { "epoch": 0.4525581297299576, "grad_norm": 23684.73046875, "learning_rate": 4.717589556201006e-05, "loss": 9.3312, "step": 10734 }, { "epoch": 0.4526002909121572, "grad_norm": 9187.458984375, "learning_rate": 4.717511477560199e-05, "loss": 9.6759, "step": 10735 }, { "epoch": 0.4526424520943567, "grad_norm": 5140.78955078125, "learning_rate": 4.717433398919392e-05, "loss": 9.8284, "step": 10736 }, { "epoch": 0.4526846132765563, "grad_norm": 4414.150390625, "learning_rate": 4.717355320278585e-05, "loss": 10.2366, "step": 10737 }, { "epoch": 0.4527267744587558, "grad_norm": 6103.2744140625, "learning_rate": 4.717277241637778e-05, "loss": 9.5001, "step": 10738 }, { "epoch": 0.4527689356409554, "grad_norm": 27903.8984375, "learning_rate": 4.7171991629969706e-05, "loss": 9.1698, "step": 10739 }, { "epoch": 0.4528110968231549, "grad_norm": 79479.46875, "learning_rate": 4.717121084356164e-05, "loss": 9.3141, "step": 10740 }, { "epoch": 0.4528532580053545, "grad_norm": 8512.0126953125, "learning_rate": 4.717043005715357e-05, "loss": 9.8672, "step": 10741 }, { "epoch": 0.452895419187554, "grad_norm": 19120.865234375, "learning_rate": 4.71696492707455e-05, "loss": 8.9246, "step": 10742 }, { "epoch": 0.45293758036975357, "grad_norm": 75434.8671875, "learning_rate": 4.7168868484337424e-05, "loss": 9.2639, "step": 10743 }, { "epoch": 0.45297974155195314, "grad_norm": 54920.5703125, "learning_rate": 4.7168087697929356e-05, "loss": 9.8663, "step": 10744 }, { "epoch": 0.45302190273415266, "grad_norm": 15115.8662109375, "learning_rate": 4.716730691152129e-05, "loss": 9.9279, "step": 10745 }, { "epoch": 0.45306406391635223, "grad_norm": 70034.53125, "learning_rate": 4.716652612511321e-05, "loss": 9.6146, "step": 10746 }, { "epoch": 0.45310622509855175, "grad_norm": 77938.6171875, "learning_rate": 4.716574533870515e-05, "loss": 11.7356, "step": 10747 }, { "epoch": 0.4531483862807513, "grad_norm": 20036.134765625, "learning_rate": 4.7164964552297074e-05, "loss": 9.5435, "step": 10748 }, { "epoch": 0.45319054746295084, "grad_norm": 3426.5947265625, "learning_rate": 4.7164183765889005e-05, "loss": 10.8695, "step": 10749 }, { "epoch": 0.4532327086451504, "grad_norm": 85786.7890625, "learning_rate": 4.7163402979480937e-05, "loss": 10.2273, "step": 10750 }, { "epoch": 0.45327486982734994, "grad_norm": 170494.234375, "learning_rate": 4.716262219307286e-05, "loss": 10.6781, "step": 10751 }, { "epoch": 0.4533170310095495, "grad_norm": 45024.3046875, "learning_rate": 4.71618414066648e-05, "loss": 9.5659, "step": 10752 }, { "epoch": 0.4533591921917491, "grad_norm": 10742.125, "learning_rate": 4.7161060620256724e-05, "loss": 11.7756, "step": 10753 }, { "epoch": 0.4534013533739486, "grad_norm": 6889.99267578125, "learning_rate": 4.7160279833848655e-05, "loss": 9.5276, "step": 10754 }, { "epoch": 0.4534435145561482, "grad_norm": 46481.57421875, "learning_rate": 4.7159499047440586e-05, "loss": 9.0947, "step": 10755 }, { "epoch": 0.4534856757383477, "grad_norm": 30573.00390625, "learning_rate": 4.715871826103252e-05, "loss": 10.7003, "step": 10756 }, { "epoch": 0.45352783692054727, "grad_norm": 29914.611328125, "learning_rate": 4.715793747462444e-05, "loss": 9.3639, "step": 10757 }, { "epoch": 0.4535699981027468, "grad_norm": 5778.9921875, "learning_rate": 4.715715668821637e-05, "loss": 9.8213, "step": 10758 }, { "epoch": 0.45361215928494636, "grad_norm": 2756.505615234375, "learning_rate": 4.7156375901808304e-05, "loss": 11.2715, "step": 10759 }, { "epoch": 0.4536543204671459, "grad_norm": 54767.76953125, "learning_rate": 4.715559511540023e-05, "loss": 11.3441, "step": 10760 }, { "epoch": 0.45369648164934545, "grad_norm": 163471.734375, "learning_rate": 4.715481432899217e-05, "loss": 10.0267, "step": 10761 }, { "epoch": 0.45373864283154497, "grad_norm": 24761.578125, "learning_rate": 4.715403354258409e-05, "loss": 8.9735, "step": 10762 }, { "epoch": 0.45378080401374454, "grad_norm": 1061547.375, "learning_rate": 4.715325275617602e-05, "loss": 9.2484, "step": 10763 }, { "epoch": 0.4538229651959441, "grad_norm": 13510.8681640625, "learning_rate": 4.7152471969767954e-05, "loss": 9.3378, "step": 10764 }, { "epoch": 0.45386512637814364, "grad_norm": 14351.5400390625, "learning_rate": 4.715169118335988e-05, "loss": 9.4173, "step": 10765 }, { "epoch": 0.4539072875603432, "grad_norm": 222664.421875, "learning_rate": 4.7150910396951816e-05, "loss": 9.6707, "step": 10766 }, { "epoch": 0.4539494487425427, "grad_norm": 183806.078125, "learning_rate": 4.715012961054374e-05, "loss": 8.8047, "step": 10767 }, { "epoch": 0.4539916099247423, "grad_norm": 1035811.375, "learning_rate": 4.714934882413567e-05, "loss": 10.1576, "step": 10768 }, { "epoch": 0.4540337711069418, "grad_norm": 27473.36328125, "learning_rate": 4.71485680377276e-05, "loss": 9.1457, "step": 10769 }, { "epoch": 0.4540759322891414, "grad_norm": 71108.3046875, "learning_rate": 4.714778725131953e-05, "loss": 8.8656, "step": 10770 }, { "epoch": 0.4541180934713409, "grad_norm": 79023.4609375, "learning_rate": 4.714700646491146e-05, "loss": 10.0167, "step": 10771 }, { "epoch": 0.4541602546535405, "grad_norm": 8776916.0, "learning_rate": 4.714622567850339e-05, "loss": 9.8244, "step": 10772 }, { "epoch": 0.45420241583574006, "grad_norm": 1090718.125, "learning_rate": 4.714544489209532e-05, "loss": 9.1659, "step": 10773 }, { "epoch": 0.4542445770179396, "grad_norm": 142166.953125, "learning_rate": 4.714466410568725e-05, "loss": 9.6002, "step": 10774 }, { "epoch": 0.45428673820013915, "grad_norm": 294803.0625, "learning_rate": 4.714388331927918e-05, "loss": 9.154, "step": 10775 }, { "epoch": 0.45432889938233867, "grad_norm": 936634.5, "learning_rate": 4.714310253287111e-05, "loss": 9.8962, "step": 10776 }, { "epoch": 0.45437106056453824, "grad_norm": 995500.0, "learning_rate": 4.714232174646304e-05, "loss": 9.9483, "step": 10777 }, { "epoch": 0.45441322174673776, "grad_norm": 230544.90625, "learning_rate": 4.714154096005497e-05, "loss": 11.6958, "step": 10778 }, { "epoch": 0.45445538292893733, "grad_norm": 85654.65625, "learning_rate": 4.7140760173646896e-05, "loss": 9.6289, "step": 10779 }, { "epoch": 0.45449754411113685, "grad_norm": 2664639.5, "learning_rate": 4.7139979387238834e-05, "loss": 9.511, "step": 10780 }, { "epoch": 0.4545397052933364, "grad_norm": 746649.375, "learning_rate": 4.713919860083076e-05, "loss": 9.9473, "step": 10781 }, { "epoch": 0.454581866475536, "grad_norm": 8430168.0, "learning_rate": 4.713841781442269e-05, "loss": 9.6135, "step": 10782 }, { "epoch": 0.4546240276577355, "grad_norm": 745275.25, "learning_rate": 4.713763702801462e-05, "loss": 9.6219, "step": 10783 }, { "epoch": 0.4546661888399351, "grad_norm": 43593.58984375, "learning_rate": 4.7136856241606545e-05, "loss": 9.1779, "step": 10784 }, { "epoch": 0.4547083500221346, "grad_norm": 84935.5390625, "learning_rate": 4.713607545519848e-05, "loss": 10.4976, "step": 10785 }, { "epoch": 0.4547505112043342, "grad_norm": 12209.0458984375, "learning_rate": 4.713529466879041e-05, "loss": 10.1167, "step": 10786 }, { "epoch": 0.4547926723865337, "grad_norm": 22195.958984375, "learning_rate": 4.713451388238234e-05, "loss": 9.3761, "step": 10787 }, { "epoch": 0.4548348335687333, "grad_norm": 100348.0625, "learning_rate": 4.713373309597427e-05, "loss": 9.399, "step": 10788 }, { "epoch": 0.4548769947509328, "grad_norm": 52383.72265625, "learning_rate": 4.7132952309566195e-05, "loss": 9.5237, "step": 10789 }, { "epoch": 0.45491915593313237, "grad_norm": 49512.01171875, "learning_rate": 4.7132171523158126e-05, "loss": 9.4302, "step": 10790 }, { "epoch": 0.4549613171153319, "grad_norm": 56217.5390625, "learning_rate": 4.713139073675006e-05, "loss": 11.3213, "step": 10791 }, { "epoch": 0.45500347829753146, "grad_norm": 80172.921875, "learning_rate": 4.713060995034199e-05, "loss": 9.0656, "step": 10792 }, { "epoch": 0.45504563947973103, "grad_norm": 315249.8125, "learning_rate": 4.712982916393391e-05, "loss": 9.3325, "step": 10793 }, { "epoch": 0.45508780066193055, "grad_norm": 8523.96484375, "learning_rate": 4.7129048377525844e-05, "loss": 9.5861, "step": 10794 }, { "epoch": 0.4551299618441301, "grad_norm": 54440.18359375, "learning_rate": 4.7128267591117775e-05, "loss": 9.6861, "step": 10795 }, { "epoch": 0.45517212302632964, "grad_norm": 6536.47998046875, "learning_rate": 4.712748680470971e-05, "loss": 10.207, "step": 10796 }, { "epoch": 0.4552142842085292, "grad_norm": 166722.453125, "learning_rate": 4.712670601830164e-05, "loss": 9.0438, "step": 10797 }, { "epoch": 0.45525644539072874, "grad_norm": 38379.44140625, "learning_rate": 4.712592523189356e-05, "loss": 9.2027, "step": 10798 }, { "epoch": 0.4552986065729283, "grad_norm": 9759.6533203125, "learning_rate": 4.71251444454855e-05, "loss": 9.6413, "step": 10799 }, { "epoch": 0.45534076775512783, "grad_norm": 33526.4140625, "learning_rate": 4.7124363659077425e-05, "loss": 9.3563, "step": 10800 }, { "epoch": 0.4553829289373274, "grad_norm": 24355.015625, "learning_rate": 4.7123582872669356e-05, "loss": 9.6783, "step": 10801 }, { "epoch": 0.455425090119527, "grad_norm": 1301.4093017578125, "learning_rate": 4.712280208626129e-05, "loss": 10.821, "step": 10802 }, { "epoch": 0.4554672513017265, "grad_norm": 20501.962890625, "learning_rate": 4.712202129985321e-05, "loss": 10.0669, "step": 10803 }, { "epoch": 0.45550941248392607, "grad_norm": 1321.84423828125, "learning_rate": 4.712124051344514e-05, "loss": 10.0408, "step": 10804 }, { "epoch": 0.4555515736661256, "grad_norm": 400632.625, "learning_rate": 4.7120459727037074e-05, "loss": 9.4382, "step": 10805 }, { "epoch": 0.45559373484832516, "grad_norm": 61097.6171875, "learning_rate": 4.7119678940629006e-05, "loss": 8.9945, "step": 10806 }, { "epoch": 0.4556358960305247, "grad_norm": 1983.6051025390625, "learning_rate": 4.711889815422094e-05, "loss": 10.2242, "step": 10807 }, { "epoch": 0.45567805721272425, "grad_norm": 7065.22900390625, "learning_rate": 4.711811736781286e-05, "loss": 9.9249, "step": 10808 }, { "epoch": 0.45572021839492377, "grad_norm": 74956.7109375, "learning_rate": 4.711733658140479e-05, "loss": 8.8935, "step": 10809 }, { "epoch": 0.45576237957712334, "grad_norm": 36585.1875, "learning_rate": 4.7116555794996724e-05, "loss": 8.9507, "step": 10810 }, { "epoch": 0.4558045407593229, "grad_norm": 70884.7734375, "learning_rate": 4.7115775008588655e-05, "loss": 8.5911, "step": 10811 }, { "epoch": 0.45584670194152244, "grad_norm": 4802.00390625, "learning_rate": 4.711499422218058e-05, "loss": 9.4939, "step": 10812 }, { "epoch": 0.455888863123722, "grad_norm": 688016.4375, "learning_rate": 4.711421343577251e-05, "loss": 9.39, "step": 10813 }, { "epoch": 0.45593102430592153, "grad_norm": 161574.953125, "learning_rate": 4.711343264936444e-05, "loss": 10.7149, "step": 10814 }, { "epoch": 0.4559731854881211, "grad_norm": 8728.1171875, "learning_rate": 4.711265186295637e-05, "loss": 9.4316, "step": 10815 }, { "epoch": 0.4560153466703206, "grad_norm": 72121.171875, "learning_rate": 4.7111871076548305e-05, "loss": 10.2785, "step": 10816 }, { "epoch": 0.4560575078525202, "grad_norm": 126994.8203125, "learning_rate": 4.711109029014023e-05, "loss": 10.407, "step": 10817 }, { "epoch": 0.4560996690347197, "grad_norm": 17056.638671875, "learning_rate": 4.711030950373217e-05, "loss": 10.6139, "step": 10818 }, { "epoch": 0.4561418302169193, "grad_norm": 222541.65625, "learning_rate": 4.710952871732409e-05, "loss": 9.0124, "step": 10819 }, { "epoch": 0.4561839913991188, "grad_norm": 358844.75, "learning_rate": 4.7108747930916016e-05, "loss": 9.2515, "step": 10820 }, { "epoch": 0.4562261525813184, "grad_norm": 1989950.5, "learning_rate": 4.7107967144507954e-05, "loss": 10.1856, "step": 10821 }, { "epoch": 0.45626831376351795, "grad_norm": 3347819.25, "learning_rate": 4.710718635809988e-05, "loss": 9.9495, "step": 10822 }, { "epoch": 0.45631047494571747, "grad_norm": 12107188.0, "learning_rate": 4.710640557169181e-05, "loss": 8.7657, "step": 10823 }, { "epoch": 0.45635263612791704, "grad_norm": 120437.328125, "learning_rate": 4.710562478528374e-05, "loss": 9.7698, "step": 10824 }, { "epoch": 0.45639479731011656, "grad_norm": 7554089.5, "learning_rate": 4.710484399887567e-05, "loss": 9.6622, "step": 10825 }, { "epoch": 0.45643695849231614, "grad_norm": 7486.791015625, "learning_rate": 4.71040632124676e-05, "loss": 9.1418, "step": 10826 }, { "epoch": 0.45647911967451565, "grad_norm": 16376.3974609375, "learning_rate": 4.710328242605953e-05, "loss": 9.0015, "step": 10827 }, { "epoch": 0.45652128085671523, "grad_norm": 192533.84375, "learning_rate": 4.710250163965146e-05, "loss": 9.423, "step": 10828 }, { "epoch": 0.45656344203891475, "grad_norm": 71230.9453125, "learning_rate": 4.7101720853243384e-05, "loss": 9.3132, "step": 10829 }, { "epoch": 0.4566056032211143, "grad_norm": 9468.62109375, "learning_rate": 4.710094006683532e-05, "loss": 9.5584, "step": 10830 }, { "epoch": 0.4566477644033139, "grad_norm": 24008.39453125, "learning_rate": 4.7100159280427247e-05, "loss": 8.8581, "step": 10831 }, { "epoch": 0.4566899255855134, "grad_norm": 438515.0625, "learning_rate": 4.709937849401918e-05, "loss": 9.1089, "step": 10832 }, { "epoch": 0.456732086767713, "grad_norm": 42521220.0, "learning_rate": 4.709859770761111e-05, "loss": 11.1394, "step": 10833 }, { "epoch": 0.4567742479499125, "grad_norm": 175512.734375, "learning_rate": 4.7097816921203034e-05, "loss": 9.9234, "step": 10834 }, { "epoch": 0.4568164091321121, "grad_norm": 37892.07421875, "learning_rate": 4.709703613479497e-05, "loss": 10.0513, "step": 10835 }, { "epoch": 0.4568585703143116, "grad_norm": 136033.78125, "learning_rate": 4.7096255348386896e-05, "loss": 9.3722, "step": 10836 }, { "epoch": 0.45690073149651117, "grad_norm": 14868.9443359375, "learning_rate": 4.709547456197883e-05, "loss": 9.4423, "step": 10837 }, { "epoch": 0.4569428926787107, "grad_norm": 8801.259765625, "learning_rate": 4.709469377557076e-05, "loss": 8.9727, "step": 10838 }, { "epoch": 0.45698505386091026, "grad_norm": 140772.0, "learning_rate": 4.709391298916268e-05, "loss": 9.0992, "step": 10839 }, { "epoch": 0.45702721504310984, "grad_norm": 5887.71630859375, "learning_rate": 4.7093132202754614e-05, "loss": 9.5217, "step": 10840 }, { "epoch": 0.45706937622530935, "grad_norm": 9074.9150390625, "learning_rate": 4.7092351416346546e-05, "loss": 9.0177, "step": 10841 }, { "epoch": 0.4571115374075089, "grad_norm": 7809.1552734375, "learning_rate": 4.709157062993848e-05, "loss": 9.9319, "step": 10842 }, { "epoch": 0.45715369858970845, "grad_norm": 2112.581787109375, "learning_rate": 4.709078984353041e-05, "loss": 9.6142, "step": 10843 }, { "epoch": 0.457195859771908, "grad_norm": 19070.54296875, "learning_rate": 4.709000905712234e-05, "loss": 9.7885, "step": 10844 }, { "epoch": 0.45723802095410754, "grad_norm": 21061.962890625, "learning_rate": 4.7089228270714264e-05, "loss": 8.9529, "step": 10845 }, { "epoch": 0.4572801821363071, "grad_norm": 282324.375, "learning_rate": 4.7088447484306195e-05, "loss": 9.2486, "step": 10846 }, { "epoch": 0.45732234331850663, "grad_norm": 164198.015625, "learning_rate": 4.7087666697898126e-05, "loss": 9.256, "step": 10847 }, { "epoch": 0.4573645045007062, "grad_norm": 121884.828125, "learning_rate": 4.708688591149005e-05, "loss": 9.0262, "step": 10848 }, { "epoch": 0.4574066656829057, "grad_norm": 9362.9111328125, "learning_rate": 4.708610512508199e-05, "loss": 9.7644, "step": 10849 }, { "epoch": 0.4574488268651053, "grad_norm": 2049.02099609375, "learning_rate": 4.7085324338673913e-05, "loss": 10.2848, "step": 10850 }, { "epoch": 0.45749098804730487, "grad_norm": 20890.220703125, "learning_rate": 4.7084543552265845e-05, "loss": 9.7814, "step": 10851 }, { "epoch": 0.4575331492295044, "grad_norm": 73828.671875, "learning_rate": 4.7083762765857776e-05, "loss": 8.8718, "step": 10852 }, { "epoch": 0.45757531041170396, "grad_norm": 11598.59765625, "learning_rate": 4.70829819794497e-05, "loss": 9.9064, "step": 10853 }, { "epoch": 0.4576174715939035, "grad_norm": 19464.6484375, "learning_rate": 4.708220119304164e-05, "loss": 9.4947, "step": 10854 }, { "epoch": 0.45765963277610305, "grad_norm": 80075.6171875, "learning_rate": 4.708142040663356e-05, "loss": 11.3726, "step": 10855 }, { "epoch": 0.45770179395830257, "grad_norm": 141064.09375, "learning_rate": 4.7080639620225494e-05, "loss": 8.8789, "step": 10856 }, { "epoch": 0.45774395514050215, "grad_norm": 5586563.5, "learning_rate": 4.7079858833817425e-05, "loss": 8.7938, "step": 10857 }, { "epoch": 0.45778611632270166, "grad_norm": 30670.77734375, "learning_rate": 4.707907804740935e-05, "loss": 10.3626, "step": 10858 }, { "epoch": 0.45782827750490124, "grad_norm": 31515.365234375, "learning_rate": 4.707829726100128e-05, "loss": 11.314, "step": 10859 }, { "epoch": 0.4578704386871008, "grad_norm": 9504274.0, "learning_rate": 4.707751647459321e-05, "loss": 9.7219, "step": 10860 }, { "epoch": 0.45791259986930033, "grad_norm": 51220.1875, "learning_rate": 4.7076735688185144e-05, "loss": 9.2759, "step": 10861 }, { "epoch": 0.4579547610514999, "grad_norm": 7869.6513671875, "learning_rate": 4.707595490177707e-05, "loss": 10.0759, "step": 10862 }, { "epoch": 0.4579969222336994, "grad_norm": 324817.6875, "learning_rate": 4.7075174115369006e-05, "loss": 10.3107, "step": 10863 }, { "epoch": 0.458039083415899, "grad_norm": 3766525.5, "learning_rate": 4.707439332896093e-05, "loss": 8.9139, "step": 10864 }, { "epoch": 0.4580812445980985, "grad_norm": 65832.5625, "learning_rate": 4.707361254255286e-05, "loss": 11.1806, "step": 10865 }, { "epoch": 0.4581234057802981, "grad_norm": 2653645.25, "learning_rate": 4.707283175614479e-05, "loss": 9.0259, "step": 10866 }, { "epoch": 0.4581655669624976, "grad_norm": 24971.751953125, "learning_rate": 4.707205096973672e-05, "loss": 9.7826, "step": 10867 }, { "epoch": 0.4582077281446972, "grad_norm": 6609.1865234375, "learning_rate": 4.7071270183328656e-05, "loss": 8.8851, "step": 10868 }, { "epoch": 0.45824988932689675, "grad_norm": 177417.078125, "learning_rate": 4.707048939692058e-05, "loss": 8.8209, "step": 10869 }, { "epoch": 0.45829205050909627, "grad_norm": 1344.855712890625, "learning_rate": 4.706970861051251e-05, "loss": 10.8752, "step": 10870 }, { "epoch": 0.45833421169129585, "grad_norm": 11749.47265625, "learning_rate": 4.706892782410444e-05, "loss": 8.8014, "step": 10871 }, { "epoch": 0.45837637287349536, "grad_norm": 1127903.0, "learning_rate": 4.706814703769637e-05, "loss": 9.9355, "step": 10872 }, { "epoch": 0.45841853405569494, "grad_norm": 379580.4375, "learning_rate": 4.70673662512883e-05, "loss": 8.6179, "step": 10873 }, { "epoch": 0.45846069523789446, "grad_norm": 80095.9609375, "learning_rate": 4.706658546488023e-05, "loss": 8.5464, "step": 10874 }, { "epoch": 0.45850285642009403, "grad_norm": 213084.34375, "learning_rate": 4.706580467847216e-05, "loss": 10.0846, "step": 10875 }, { "epoch": 0.45854501760229355, "grad_norm": 4951.97412109375, "learning_rate": 4.706502389206409e-05, "loss": 9.5039, "step": 10876 }, { "epoch": 0.4585871787844931, "grad_norm": 32180.984375, "learning_rate": 4.706424310565602e-05, "loss": 9.408, "step": 10877 }, { "epoch": 0.45862933996669264, "grad_norm": 59099.98046875, "learning_rate": 4.706346231924795e-05, "loss": 8.9717, "step": 10878 }, { "epoch": 0.4586715011488922, "grad_norm": 45916.58203125, "learning_rate": 4.706268153283988e-05, "loss": 9.6874, "step": 10879 }, { "epoch": 0.4587136623310918, "grad_norm": 17270.427734375, "learning_rate": 4.706190074643181e-05, "loss": 9.5635, "step": 10880 }, { "epoch": 0.4587558235132913, "grad_norm": 6589.5166015625, "learning_rate": 4.7061119960023735e-05, "loss": 10.4596, "step": 10881 }, { "epoch": 0.4587979846954909, "grad_norm": 1520.8828125, "learning_rate": 4.7060339173615666e-05, "loss": 9.2513, "step": 10882 }, { "epoch": 0.4588401458776904, "grad_norm": 460323.875, "learning_rate": 4.70595583872076e-05, "loss": 9.6005, "step": 10883 }, { "epoch": 0.45888230705988997, "grad_norm": 13204.03125, "learning_rate": 4.705877760079952e-05, "loss": 9.9761, "step": 10884 }, { "epoch": 0.4589244682420895, "grad_norm": 11326.693359375, "learning_rate": 4.705799681439146e-05, "loss": 10.4779, "step": 10885 }, { "epoch": 0.45896662942428906, "grad_norm": 15833.1650390625, "learning_rate": 4.7057216027983385e-05, "loss": 11.0898, "step": 10886 }, { "epoch": 0.4590087906064886, "grad_norm": 3498.96630859375, "learning_rate": 4.705643524157532e-05, "loss": 9.6128, "step": 10887 }, { "epoch": 0.45905095178868816, "grad_norm": 4463.119140625, "learning_rate": 4.705565445516725e-05, "loss": 10.5179, "step": 10888 }, { "epoch": 0.45909311297088773, "grad_norm": 30384.39453125, "learning_rate": 4.705487366875918e-05, "loss": 9.6244, "step": 10889 }, { "epoch": 0.45913527415308725, "grad_norm": 46729.2265625, "learning_rate": 4.705409288235111e-05, "loss": 8.832, "step": 10890 }, { "epoch": 0.4591774353352868, "grad_norm": 3303.549560546875, "learning_rate": 4.7053312095943034e-05, "loss": 10.2532, "step": 10891 }, { "epoch": 0.45921959651748634, "grad_norm": 14785.318359375, "learning_rate": 4.7052531309534965e-05, "loss": 9.1469, "step": 10892 }, { "epoch": 0.4592617576996859, "grad_norm": 280121.5625, "learning_rate": 4.7051750523126897e-05, "loss": 9.7875, "step": 10893 }, { "epoch": 0.45930391888188543, "grad_norm": 12033.23828125, "learning_rate": 4.705096973671883e-05, "loss": 9.4229, "step": 10894 }, { "epoch": 0.459346080064085, "grad_norm": 75441.3671875, "learning_rate": 4.705018895031075e-05, "loss": 10.5826, "step": 10895 }, { "epoch": 0.4593882412462845, "grad_norm": 10983.15625, "learning_rate": 4.7049408163902684e-05, "loss": 9.5464, "step": 10896 }, { "epoch": 0.4594304024284841, "grad_norm": 74839.7265625, "learning_rate": 4.7048627377494615e-05, "loss": 9.9849, "step": 10897 }, { "epoch": 0.45947256361068367, "grad_norm": 19212.349609375, "learning_rate": 4.7047846591086546e-05, "loss": 8.7089, "step": 10898 }, { "epoch": 0.4595147247928832, "grad_norm": 15888.15234375, "learning_rate": 4.704706580467848e-05, "loss": 9.0589, "step": 10899 }, { "epoch": 0.45955688597508276, "grad_norm": 38912.07421875, "learning_rate": 4.70462850182704e-05, "loss": 9.6088, "step": 10900 }, { "epoch": 0.4595990471572823, "grad_norm": 5720.3876953125, "learning_rate": 4.704550423186233e-05, "loss": 8.8093, "step": 10901 }, { "epoch": 0.45964120833948185, "grad_norm": 66935.3125, "learning_rate": 4.7044723445454264e-05, "loss": 9.2481, "step": 10902 }, { "epoch": 0.4596833695216814, "grad_norm": 34162.80078125, "learning_rate": 4.704394265904619e-05, "loss": 9.4182, "step": 10903 }, { "epoch": 0.45972553070388095, "grad_norm": 18999.578125, "learning_rate": 4.704316187263813e-05, "loss": 10.8963, "step": 10904 }, { "epoch": 0.45976769188608047, "grad_norm": 16410.048828125, "learning_rate": 4.704238108623005e-05, "loss": 9.5091, "step": 10905 }, { "epoch": 0.45980985306828004, "grad_norm": 290825.5625, "learning_rate": 4.704160029982198e-05, "loss": 10.3918, "step": 10906 }, { "epoch": 0.45985201425047956, "grad_norm": 36644.375, "learning_rate": 4.7040819513413914e-05, "loss": 9.9021, "step": 10907 }, { "epoch": 0.45989417543267913, "grad_norm": 16716.525390625, "learning_rate": 4.7040038727005845e-05, "loss": 9.5836, "step": 10908 }, { "epoch": 0.4599363366148787, "grad_norm": 86497528.0, "learning_rate": 4.703925794059777e-05, "loss": 9.9291, "step": 10909 }, { "epoch": 0.4599784977970782, "grad_norm": 77522.1328125, "learning_rate": 4.70384771541897e-05, "loss": 9.9273, "step": 10910 }, { "epoch": 0.4600206589792778, "grad_norm": 2066862208.0, "learning_rate": 4.703769636778163e-05, "loss": 11.7295, "step": 10911 }, { "epoch": 0.4600628201614773, "grad_norm": 130086.6015625, "learning_rate": 4.7036915581373563e-05, "loss": 10.3935, "step": 10912 }, { "epoch": 0.4601049813436769, "grad_norm": 141039.21875, "learning_rate": 4.7036134794965495e-05, "loss": 9.7935, "step": 10913 }, { "epoch": 0.4601471425258764, "grad_norm": 424760.90625, "learning_rate": 4.703535400855742e-05, "loss": 9.516, "step": 10914 }, { "epoch": 0.460189303708076, "grad_norm": 100157.6875, "learning_rate": 4.703457322214935e-05, "loss": 10.573, "step": 10915 }, { "epoch": 0.4602314648902755, "grad_norm": 177793.09375, "learning_rate": 4.703379243574128e-05, "loss": 11.2602, "step": 10916 }, { "epoch": 0.4602736260724751, "grad_norm": 605389.625, "learning_rate": 4.7033011649333206e-05, "loss": 9.7774, "step": 10917 }, { "epoch": 0.46031578725467465, "grad_norm": 221667.359375, "learning_rate": 4.7032230862925144e-05, "loss": 10.6852, "step": 10918 }, { "epoch": 0.46035794843687416, "grad_norm": 58863.96484375, "learning_rate": 4.703145007651707e-05, "loss": 10.6895, "step": 10919 }, { "epoch": 0.46040010961907374, "grad_norm": 37409.9453125, "learning_rate": 4.7030669290109e-05, "loss": 10.6838, "step": 10920 }, { "epoch": 0.46044227080127326, "grad_norm": 58063.53515625, "learning_rate": 4.702988850370093e-05, "loss": 10.6143, "step": 10921 }, { "epoch": 0.46048443198347283, "grad_norm": 93143.9765625, "learning_rate": 4.7029107717292856e-05, "loss": 10.2701, "step": 10922 }, { "epoch": 0.46052659316567235, "grad_norm": 263705.84375, "learning_rate": 4.7028326930884794e-05, "loss": 10.3067, "step": 10923 }, { "epoch": 0.4605687543478719, "grad_norm": 150520.625, "learning_rate": 4.702754614447672e-05, "loss": 10.5108, "step": 10924 }, { "epoch": 0.46061091553007144, "grad_norm": 157274.796875, "learning_rate": 4.702676535806865e-05, "loss": 9.7227, "step": 10925 }, { "epoch": 0.460653076712271, "grad_norm": 35340.78515625, "learning_rate": 4.702598457166058e-05, "loss": 11.3731, "step": 10926 }, { "epoch": 0.4606952378944706, "grad_norm": 445232.5625, "learning_rate": 4.7025203785252505e-05, "loss": 9.85, "step": 10927 }, { "epoch": 0.4607373990766701, "grad_norm": 1332000.125, "learning_rate": 4.7024422998844436e-05, "loss": 9.3737, "step": 10928 }, { "epoch": 0.4607795602588697, "grad_norm": 67413.046875, "learning_rate": 4.702364221243637e-05, "loss": 10.181, "step": 10929 }, { "epoch": 0.4608217214410692, "grad_norm": 78999.046875, "learning_rate": 4.70228614260283e-05, "loss": 9.4488, "step": 10930 }, { "epoch": 0.4608638826232688, "grad_norm": 304262.59375, "learning_rate": 4.7022080639620223e-05, "loss": 10.4681, "step": 10931 }, { "epoch": 0.4609060438054683, "grad_norm": 25156.4609375, "learning_rate": 4.702129985321216e-05, "loss": 10.7208, "step": 10932 }, { "epoch": 0.46094820498766786, "grad_norm": 143881.25, "learning_rate": 4.7020519066804086e-05, "loss": 9.5218, "step": 10933 }, { "epoch": 0.4609903661698674, "grad_norm": 86007.4765625, "learning_rate": 4.701973828039602e-05, "loss": 9.2186, "step": 10934 }, { "epoch": 0.46103252735206696, "grad_norm": 20866.986328125, "learning_rate": 4.701895749398795e-05, "loss": 11.4932, "step": 10935 }, { "epoch": 0.4610746885342665, "grad_norm": 10127.61328125, "learning_rate": 4.701817670757987e-05, "loss": 9.9982, "step": 10936 }, { "epoch": 0.46111684971646605, "grad_norm": 30931.8984375, "learning_rate": 4.701739592117181e-05, "loss": 9.4509, "step": 10937 }, { "epoch": 0.4611590108986656, "grad_norm": 129278.96875, "learning_rate": 4.7016615134763736e-05, "loss": 10.8992, "step": 10938 }, { "epoch": 0.46120117208086514, "grad_norm": 59003.7578125, "learning_rate": 4.701583434835567e-05, "loss": 11.0467, "step": 10939 }, { "epoch": 0.4612433332630647, "grad_norm": 2908.8759765625, "learning_rate": 4.70150535619476e-05, "loss": 9.1776, "step": 10940 }, { "epoch": 0.46128549444526423, "grad_norm": 1494209.75, "learning_rate": 4.701427277553952e-05, "loss": 8.9561, "step": 10941 }, { "epoch": 0.4613276556274638, "grad_norm": 73404.0234375, "learning_rate": 4.7013491989131454e-05, "loss": 9.6759, "step": 10942 }, { "epoch": 0.4613698168096633, "grad_norm": 65861.1953125, "learning_rate": 4.7012711202723385e-05, "loss": 9.7007, "step": 10943 }, { "epoch": 0.4614119779918629, "grad_norm": 112989.8046875, "learning_rate": 4.7011930416315316e-05, "loss": 8.9843, "step": 10944 }, { "epoch": 0.4614541391740624, "grad_norm": 265144.71875, "learning_rate": 4.701114962990725e-05, "loss": 8.8608, "step": 10945 }, { "epoch": 0.461496300356262, "grad_norm": 27342.728515625, "learning_rate": 4.701036884349917e-05, "loss": 10.0759, "step": 10946 }, { "epoch": 0.46153846153846156, "grad_norm": 64222.42578125, "learning_rate": 4.70095880570911e-05, "loss": 10.327, "step": 10947 }, { "epoch": 0.4615806227206611, "grad_norm": 6490.8232421875, "learning_rate": 4.7008807270683035e-05, "loss": 10.7585, "step": 10948 }, { "epoch": 0.46162278390286066, "grad_norm": 8334.919921875, "learning_rate": 4.7008026484274966e-05, "loss": 9.1949, "step": 10949 }, { "epoch": 0.4616649450850602, "grad_norm": 571.3961791992188, "learning_rate": 4.700724569786689e-05, "loss": 8.8976, "step": 10950 }, { "epoch": 0.46170710626725975, "grad_norm": 3678.38720703125, "learning_rate": 4.700646491145883e-05, "loss": 9.2506, "step": 10951 }, { "epoch": 0.46174926744945927, "grad_norm": 1819.3135986328125, "learning_rate": 4.700568412505075e-05, "loss": 9.6116, "step": 10952 }, { "epoch": 0.46179142863165884, "grad_norm": 4110.22607421875, "learning_rate": 4.700490333864268e-05, "loss": 9.3158, "step": 10953 }, { "epoch": 0.46183358981385836, "grad_norm": 7381.80126953125, "learning_rate": 4.7004122552234615e-05, "loss": 8.8453, "step": 10954 }, { "epoch": 0.46187575099605793, "grad_norm": 1122.496826171875, "learning_rate": 4.700334176582654e-05, "loss": 8.9637, "step": 10955 }, { "epoch": 0.4619179121782575, "grad_norm": 1347.5367431640625, "learning_rate": 4.700256097941848e-05, "loss": 9.9472, "step": 10956 }, { "epoch": 0.461960073360457, "grad_norm": 5347.97509765625, "learning_rate": 4.70017801930104e-05, "loss": 8.982, "step": 10957 }, { "epoch": 0.4620022345426566, "grad_norm": 572.5010375976562, "learning_rate": 4.7000999406602334e-05, "loss": 10.6181, "step": 10958 }, { "epoch": 0.4620443957248561, "grad_norm": 14952.451171875, "learning_rate": 4.7000218620194265e-05, "loss": 9.1277, "step": 10959 }, { "epoch": 0.4620865569070557, "grad_norm": 567.2510986328125, "learning_rate": 4.699943783378619e-05, "loss": 8.977, "step": 10960 }, { "epoch": 0.4621287180892552, "grad_norm": 457.8420104980469, "learning_rate": 4.699865704737812e-05, "loss": 9.0861, "step": 10961 }, { "epoch": 0.4621708792714548, "grad_norm": 5778.20458984375, "learning_rate": 4.699787626097005e-05, "loss": 9.1302, "step": 10962 }, { "epoch": 0.4622130404536543, "grad_norm": 546.1326904296875, "learning_rate": 4.699709547456198e-05, "loss": 8.3555, "step": 10963 }, { "epoch": 0.4622552016358539, "grad_norm": 303.76214599609375, "learning_rate": 4.699631468815391e-05, "loss": 8.6437, "step": 10964 }, { "epoch": 0.4622973628180534, "grad_norm": 752.8685302734375, "learning_rate": 4.699553390174584e-05, "loss": 8.6452, "step": 10965 }, { "epoch": 0.46233952400025297, "grad_norm": 249.5378875732422, "learning_rate": 4.699475311533777e-05, "loss": 9.919, "step": 10966 }, { "epoch": 0.46238168518245254, "grad_norm": 112.53152465820312, "learning_rate": 4.69939723289297e-05, "loss": 9.5043, "step": 10967 }, { "epoch": 0.46242384636465206, "grad_norm": 417.1719970703125, "learning_rate": 4.699319154252163e-05, "loss": 9.0181, "step": 10968 }, { "epoch": 0.46246600754685163, "grad_norm": 1164.9749755859375, "learning_rate": 4.699241075611356e-05, "loss": 9.0123, "step": 10969 }, { "epoch": 0.46250816872905115, "grad_norm": 798.54736328125, "learning_rate": 4.6991629969705495e-05, "loss": 9.0201, "step": 10970 }, { "epoch": 0.4625503299112507, "grad_norm": 1361.1409912109375, "learning_rate": 4.699084918329742e-05, "loss": 8.942, "step": 10971 }, { "epoch": 0.46259249109345024, "grad_norm": 221.46539306640625, "learning_rate": 4.6990068396889344e-05, "loss": 9.7679, "step": 10972 }, { "epoch": 0.4626346522756498, "grad_norm": 250.75540161132812, "learning_rate": 4.698928761048128e-05, "loss": 10.0064, "step": 10973 }, { "epoch": 0.46267681345784933, "grad_norm": 116.1337890625, "learning_rate": 4.698850682407321e-05, "loss": 9.3221, "step": 10974 }, { "epoch": 0.4627189746400489, "grad_norm": 135.5090789794922, "learning_rate": 4.698772603766514e-05, "loss": 10.2272, "step": 10975 }, { "epoch": 0.4627611358222485, "grad_norm": 167.09780883789062, "learning_rate": 4.698694525125707e-05, "loss": 8.8246, "step": 10976 }, { "epoch": 0.462803297004448, "grad_norm": 292.5644836425781, "learning_rate": 4.6986164464849e-05, "loss": 8.7362, "step": 10977 }, { "epoch": 0.4628454581866476, "grad_norm": 125.25008392333984, "learning_rate": 4.698538367844093e-05, "loss": 8.3242, "step": 10978 }, { "epoch": 0.4628876193688471, "grad_norm": 146.37696838378906, "learning_rate": 4.6984602892032856e-05, "loss": 10.6709, "step": 10979 }, { "epoch": 0.46292978055104667, "grad_norm": 193.60504150390625, "learning_rate": 4.698382210562479e-05, "loss": 8.4761, "step": 10980 }, { "epoch": 0.4629719417332462, "grad_norm": 195.0832061767578, "learning_rate": 4.698304131921672e-05, "loss": 9.493, "step": 10981 }, { "epoch": 0.46301410291544576, "grad_norm": 2589.987548828125, "learning_rate": 4.698226053280865e-05, "loss": 9.1369, "step": 10982 }, { "epoch": 0.4630562640976453, "grad_norm": 311.57672119140625, "learning_rate": 4.6981479746400574e-05, "loss": 9.6423, "step": 10983 }, { "epoch": 0.46309842527984485, "grad_norm": 334.953857421875, "learning_rate": 4.6980698959992506e-05, "loss": 8.4842, "step": 10984 }, { "epoch": 0.46314058646204437, "grad_norm": 754.6959838867188, "learning_rate": 4.697991817358444e-05, "loss": 9.266, "step": 10985 }, { "epoch": 0.46318274764424394, "grad_norm": 157.3220977783203, "learning_rate": 4.697913738717636e-05, "loss": 8.4113, "step": 10986 }, { "epoch": 0.4632249088264435, "grad_norm": 279.0370178222656, "learning_rate": 4.69783566007683e-05, "loss": 9.5127, "step": 10987 }, { "epoch": 0.46326707000864303, "grad_norm": 1703.780029296875, "learning_rate": 4.6977575814360224e-05, "loss": 9.2055, "step": 10988 }, { "epoch": 0.4633092311908426, "grad_norm": 163.05093383789062, "learning_rate": 4.6976795027952155e-05, "loss": 8.9652, "step": 10989 }, { "epoch": 0.4633513923730421, "grad_norm": 227.36582946777344, "learning_rate": 4.6976014241544087e-05, "loss": 8.3658, "step": 10990 }, { "epoch": 0.4633935535552417, "grad_norm": 249.0227813720703, "learning_rate": 4.697523345513601e-05, "loss": 8.4083, "step": 10991 }, { "epoch": 0.4634357147374412, "grad_norm": 191.29104614257812, "learning_rate": 4.697445266872795e-05, "loss": 9.0902, "step": 10992 }, { "epoch": 0.4634778759196408, "grad_norm": 184.6138458251953, "learning_rate": 4.6973671882319874e-05, "loss": 9.1582, "step": 10993 }, { "epoch": 0.4635200371018403, "grad_norm": 179.01380920410156, "learning_rate": 4.6972891095911805e-05, "loss": 8.9933, "step": 10994 }, { "epoch": 0.4635621982840399, "grad_norm": 175.60238647460938, "learning_rate": 4.6972110309503736e-05, "loss": 9.9583, "step": 10995 }, { "epoch": 0.46360435946623946, "grad_norm": 173.25633239746094, "learning_rate": 4.697132952309567e-05, "loss": 8.4967, "step": 10996 }, { "epoch": 0.463646520648439, "grad_norm": 1141.7735595703125, "learning_rate": 4.697054873668759e-05, "loss": 9.9901, "step": 10997 }, { "epoch": 0.46368868183063855, "grad_norm": 314004.375, "learning_rate": 4.696976795027952e-05, "loss": 10.1741, "step": 10998 }, { "epoch": 0.46373084301283807, "grad_norm": 628268.4375, "learning_rate": 4.6968987163871454e-05, "loss": 9.514, "step": 10999 }, { "epoch": 0.46377300419503764, "grad_norm": 65542.7734375, "learning_rate": 4.696820637746338e-05, "loss": 10.4339, "step": 11000 }, { "epoch": 0.46381516537723716, "grad_norm": 512625.53125, "learning_rate": 4.696742559105532e-05, "loss": 11.5662, "step": 11001 }, { "epoch": 0.46385732655943673, "grad_norm": 46030.1796875, "learning_rate": 4.696664480464724e-05, "loss": 10.147, "step": 11002 }, { "epoch": 0.46389948774163625, "grad_norm": 5681.48828125, "learning_rate": 4.696586401823917e-05, "loss": 10.1081, "step": 11003 }, { "epoch": 0.4639416489238358, "grad_norm": 108442.765625, "learning_rate": 4.6965083231831104e-05, "loss": 10.1921, "step": 11004 }, { "epoch": 0.4639838101060354, "grad_norm": 8548870.0, "learning_rate": 4.696430244542303e-05, "loss": 9.9039, "step": 11005 }, { "epoch": 0.4640259712882349, "grad_norm": 79412.2265625, "learning_rate": 4.6963521659014966e-05, "loss": 9.0776, "step": 11006 }, { "epoch": 0.4640681324704345, "grad_norm": 20086.443359375, "learning_rate": 4.696274087260689e-05, "loss": 9.0284, "step": 11007 }, { "epoch": 0.464110293652634, "grad_norm": 431.630615234375, "learning_rate": 4.696196008619882e-05, "loss": 8.7093, "step": 11008 }, { "epoch": 0.4641524548348336, "grad_norm": 340.8463439941406, "learning_rate": 4.696117929979075e-05, "loss": 9.5446, "step": 11009 }, { "epoch": 0.4641946160170331, "grad_norm": 7718.79931640625, "learning_rate": 4.696039851338268e-05, "loss": 8.8949, "step": 11010 }, { "epoch": 0.4642367771992327, "grad_norm": 9349.853515625, "learning_rate": 4.695961772697461e-05, "loss": 8.8363, "step": 11011 }, { "epoch": 0.4642789383814322, "grad_norm": 320.663818359375, "learning_rate": 4.695883694056654e-05, "loss": 8.8603, "step": 11012 }, { "epoch": 0.46432109956363177, "grad_norm": 491.5821533203125, "learning_rate": 4.695805615415847e-05, "loss": 8.5755, "step": 11013 }, { "epoch": 0.4643632607458313, "grad_norm": 18541.05078125, "learning_rate": 4.69572753677504e-05, "loss": 9.707, "step": 11014 }, { "epoch": 0.46440542192803086, "grad_norm": 8556.0068359375, "learning_rate": 4.695649458134233e-05, "loss": 9.439, "step": 11015 }, { "epoch": 0.46444758311023043, "grad_norm": 5042.68310546875, "learning_rate": 4.695571379493426e-05, "loss": 9.0191, "step": 11016 }, { "epoch": 0.46448974429242995, "grad_norm": 2842.91748046875, "learning_rate": 4.695493300852619e-05, "loss": 9.0497, "step": 11017 }, { "epoch": 0.4645319054746295, "grad_norm": 62717.453125, "learning_rate": 4.695415222211812e-05, "loss": 10.2573, "step": 11018 }, { "epoch": 0.46457406665682904, "grad_norm": 3346.609619140625, "learning_rate": 4.6953371435710046e-05, "loss": 10.1036, "step": 11019 }, { "epoch": 0.4646162278390286, "grad_norm": 1031993.3125, "learning_rate": 4.6952590649301984e-05, "loss": 9.2149, "step": 11020 }, { "epoch": 0.46465838902122814, "grad_norm": 122796.921875, "learning_rate": 4.695180986289391e-05, "loss": 11.5411, "step": 11021 }, { "epoch": 0.4647005502034277, "grad_norm": 270896.21875, "learning_rate": 4.695102907648584e-05, "loss": 9.8138, "step": 11022 }, { "epoch": 0.4647427113856272, "grad_norm": 5453.75537109375, "learning_rate": 4.695024829007777e-05, "loss": 9.9016, "step": 11023 }, { "epoch": 0.4647848725678268, "grad_norm": 1257174.375, "learning_rate": 4.6949467503669695e-05, "loss": 9.7324, "step": 11024 }, { "epoch": 0.4648270337500264, "grad_norm": 157800.1875, "learning_rate": 4.694868671726163e-05, "loss": 9.986, "step": 11025 }, { "epoch": 0.4648691949322259, "grad_norm": 7432.26513671875, "learning_rate": 4.694790593085356e-05, "loss": 9.7632, "step": 11026 }, { "epoch": 0.46491135611442547, "grad_norm": 163512944.0, "learning_rate": 4.694712514444549e-05, "loss": 9.7289, "step": 11027 }, { "epoch": 0.464953517296625, "grad_norm": 67297928.0, "learning_rate": 4.694634435803742e-05, "loss": 10.9423, "step": 11028 }, { "epoch": 0.46499567847882456, "grad_norm": 1118.2535400390625, "learning_rate": 4.6945563571629345e-05, "loss": 9.9931, "step": 11029 }, { "epoch": 0.4650378396610241, "grad_norm": 6313.58642578125, "learning_rate": 4.6944782785221276e-05, "loss": 9.5142, "step": 11030 }, { "epoch": 0.46508000084322365, "grad_norm": 12315.498046875, "learning_rate": 4.694400199881321e-05, "loss": 9.608, "step": 11031 }, { "epoch": 0.46512216202542317, "grad_norm": 1026.4053955078125, "learning_rate": 4.694322121240514e-05, "loss": 10.0217, "step": 11032 }, { "epoch": 0.46516432320762274, "grad_norm": 1254908.875, "learning_rate": 4.694244042599706e-05, "loss": 9.4706, "step": 11033 }, { "epoch": 0.4652064843898223, "grad_norm": 1270.8583984375, "learning_rate": 4.6941659639588994e-05, "loss": 10.7036, "step": 11034 }, { "epoch": 0.46524864557202184, "grad_norm": 1176.7677001953125, "learning_rate": 4.6940878853180925e-05, "loss": 9.8173, "step": 11035 }, { "epoch": 0.4652908067542214, "grad_norm": 599.1016845703125, "learning_rate": 4.694009806677286e-05, "loss": 9.4928, "step": 11036 }, { "epoch": 0.4653329679364209, "grad_norm": 4887.49755859375, "learning_rate": 4.693931728036479e-05, "loss": 10.0603, "step": 11037 }, { "epoch": 0.4653751291186205, "grad_norm": 885.7584838867188, "learning_rate": 4.693853649395671e-05, "loss": 10.9655, "step": 11038 }, { "epoch": 0.46541729030082, "grad_norm": 707.4408569335938, "learning_rate": 4.693775570754865e-05, "loss": 10.0583, "step": 11039 }, { "epoch": 0.4654594514830196, "grad_norm": 1137.6715087890625, "learning_rate": 4.6936974921140575e-05, "loss": 10.7218, "step": 11040 }, { "epoch": 0.4655016126652191, "grad_norm": 775.5728759765625, "learning_rate": 4.6936194134732506e-05, "loss": 10.937, "step": 11041 }, { "epoch": 0.4655437738474187, "grad_norm": 474.1878967285156, "learning_rate": 4.693541334832444e-05, "loss": 11.2533, "step": 11042 }, { "epoch": 0.4655859350296182, "grad_norm": 8541.591796875, "learning_rate": 4.693463256191636e-05, "loss": 11.7818, "step": 11043 }, { "epoch": 0.4656280962118178, "grad_norm": 1109.8018798828125, "learning_rate": 4.693385177550829e-05, "loss": 11.4063, "step": 11044 }, { "epoch": 0.46567025739401735, "grad_norm": 2071.698486328125, "learning_rate": 4.6933070989100224e-05, "loss": 10.9479, "step": 11045 }, { "epoch": 0.46571241857621687, "grad_norm": 781.3609008789062, "learning_rate": 4.6932290202692156e-05, "loss": 13.239, "step": 11046 }, { "epoch": 0.46575457975841644, "grad_norm": 372.2491149902344, "learning_rate": 4.693150941628409e-05, "loss": 12.2032, "step": 11047 }, { "epoch": 0.46579674094061596, "grad_norm": 583.09033203125, "learning_rate": 4.693072862987601e-05, "loss": 11.9658, "step": 11048 }, { "epoch": 0.46583890212281553, "grad_norm": 8827411.0, "learning_rate": 4.692994784346794e-05, "loss": 11.7041, "step": 11049 }, { "epoch": 0.46588106330501505, "grad_norm": 2200.689697265625, "learning_rate": 4.6929167057059874e-05, "loss": 12.6531, "step": 11050 }, { "epoch": 0.4659232244872146, "grad_norm": 1241.549072265625, "learning_rate": 4.6928386270651805e-05, "loss": 12.5556, "step": 11051 }, { "epoch": 0.46596538566941414, "grad_norm": 802.6312255859375, "learning_rate": 4.692760548424373e-05, "loss": 13.0255, "step": 11052 }, { "epoch": 0.4660075468516137, "grad_norm": 41468.40625, "learning_rate": 4.692682469783566e-05, "loss": 12.9921, "step": 11053 }, { "epoch": 0.4660497080338133, "grad_norm": 406.5533752441406, "learning_rate": 4.692604391142759e-05, "loss": 13.0737, "step": 11054 }, { "epoch": 0.4660918692160128, "grad_norm": 618.1888427734375, "learning_rate": 4.692526312501952e-05, "loss": 12.1337, "step": 11055 }, { "epoch": 0.4661340303982124, "grad_norm": 3701956.25, "learning_rate": 4.6924482338611455e-05, "loss": 13.0747, "step": 11056 }, { "epoch": 0.4661761915804119, "grad_norm": 1922106.875, "learning_rate": 4.692370155220338e-05, "loss": 13.1769, "step": 11057 }, { "epoch": 0.4662183527626115, "grad_norm": 277232.40625, "learning_rate": 4.692292076579532e-05, "loss": 13.0609, "step": 11058 }, { "epoch": 0.466260513944811, "grad_norm": 7188223.0, "learning_rate": 4.692213997938724e-05, "loss": 13.0388, "step": 11059 }, { "epoch": 0.46630267512701057, "grad_norm": 152391.375, "learning_rate": 4.6921359192979166e-05, "loss": 13.5437, "step": 11060 }, { "epoch": 0.4663448363092101, "grad_norm": 32041.767578125, "learning_rate": 4.6920578406571104e-05, "loss": 12.7587, "step": 11061 }, { "epoch": 0.46638699749140966, "grad_norm": 147106.765625, "learning_rate": 4.691979762016303e-05, "loss": 13.3758, "step": 11062 }, { "epoch": 0.46642915867360923, "grad_norm": 538673.25, "learning_rate": 4.691901683375496e-05, "loss": 14.6164, "step": 11063 }, { "epoch": 0.46647131985580875, "grad_norm": 611386.125, "learning_rate": 4.691823604734689e-05, "loss": 14.2509, "step": 11064 }, { "epoch": 0.4665134810380083, "grad_norm": 3519.082763671875, "learning_rate": 4.691745526093882e-05, "loss": 13.8909, "step": 11065 }, { "epoch": 0.46655564222020784, "grad_norm": 9995434.0, "learning_rate": 4.691667447453075e-05, "loss": 14.6921, "step": 11066 }, { "epoch": 0.4665978034024074, "grad_norm": 3265.625, "learning_rate": 4.691589368812268e-05, "loss": 15.6579, "step": 11067 }, { "epoch": 0.46663996458460694, "grad_norm": 4241.49658203125, "learning_rate": 4.691511290171461e-05, "loss": 16.0429, "step": 11068 }, { "epoch": 0.4666821257668065, "grad_norm": 4191.99072265625, "learning_rate": 4.6914332115306534e-05, "loss": 16.2933, "step": 11069 }, { "epoch": 0.46672428694900603, "grad_norm": 5117.65283203125, "learning_rate": 4.691355132889847e-05, "loss": 17.4, "step": 11070 }, { "epoch": 0.4667664481312056, "grad_norm": 4756.88623046875, "learning_rate": 4.6912770542490397e-05, "loss": 21.0796, "step": 11071 }, { "epoch": 0.4668086093134051, "grad_norm": 4521.96826171875, "learning_rate": 4.691198975608233e-05, "loss": 20.6703, "step": 11072 }, { "epoch": 0.4668507704956047, "grad_norm": 6120.7392578125, "learning_rate": 4.691120896967426e-05, "loss": 22.0828, "step": 11073 }, { "epoch": 0.46689293167780427, "grad_norm": 46134.5625, "learning_rate": 4.6910428183266184e-05, "loss": 24.3195, "step": 11074 }, { "epoch": 0.4669350928600038, "grad_norm": 6190.86181640625, "learning_rate": 4.690964739685812e-05, "loss": 25.3103, "step": 11075 }, { "epoch": 0.46697725404220336, "grad_norm": 3572.714599609375, "learning_rate": 4.6908866610450046e-05, "loss": 27.8554, "step": 11076 }, { "epoch": 0.4670194152244029, "grad_norm": 6568.50537109375, "learning_rate": 4.690808582404198e-05, "loss": 26.7992, "step": 11077 }, { "epoch": 0.46706157640660245, "grad_norm": 8873.8154296875, "learning_rate": 4.690730503763391e-05, "loss": 29.7976, "step": 11078 }, { "epoch": 0.46710373758880197, "grad_norm": 4975.60400390625, "learning_rate": 4.690652425122583e-05, "loss": 32.0404, "step": 11079 }, { "epoch": 0.46714589877100154, "grad_norm": 4248.97802734375, "learning_rate": 4.6905743464817764e-05, "loss": 31.4406, "step": 11080 }, { "epoch": 0.46718805995320106, "grad_norm": 5168.70947265625, "learning_rate": 4.6904962678409696e-05, "loss": 34.16, "step": 11081 }, { "epoch": 0.46723022113540064, "grad_norm": 5002.54931640625, "learning_rate": 4.690418189200163e-05, "loss": 34.2649, "step": 11082 }, { "epoch": 0.4672723823176002, "grad_norm": 6827.99169921875, "learning_rate": 4.690340110559356e-05, "loss": 35.8466, "step": 11083 }, { "epoch": 0.46731454349979973, "grad_norm": 5028.7490234375, "learning_rate": 4.690262031918549e-05, "loss": 36.5574, "step": 11084 }, { "epoch": 0.4673567046819993, "grad_norm": 3920.96728515625, "learning_rate": 4.6901839532777414e-05, "loss": 37.3769, "step": 11085 }, { "epoch": 0.4673988658641988, "grad_norm": 2830.944580078125, "learning_rate": 4.6901058746369345e-05, "loss": 34.9251, "step": 11086 }, { "epoch": 0.4674410270463984, "grad_norm": 6395.29541015625, "learning_rate": 4.6900277959961276e-05, "loss": 38.6008, "step": 11087 }, { "epoch": 0.4674831882285979, "grad_norm": 5082.52197265625, "learning_rate": 4.68994971735532e-05, "loss": 37.2331, "step": 11088 }, { "epoch": 0.4675253494107975, "grad_norm": 8880.4501953125, "learning_rate": 4.689871638714514e-05, "loss": 38.8791, "step": 11089 }, { "epoch": 0.467567510592997, "grad_norm": 10316.5517578125, "learning_rate": 4.6897935600737063e-05, "loss": 37.9116, "step": 11090 }, { "epoch": 0.4676096717751966, "grad_norm": 7134.92041015625, "learning_rate": 4.6897154814328995e-05, "loss": 39.2534, "step": 11091 }, { "epoch": 0.46765183295739615, "grad_norm": 14107.9326171875, "learning_rate": 4.6896374027920926e-05, "loss": 37.1278, "step": 11092 }, { "epoch": 0.46769399413959567, "grad_norm": 22593.271484375, "learning_rate": 4.689559324151285e-05, "loss": 36.0174, "step": 11093 }, { "epoch": 0.46773615532179524, "grad_norm": 313356.25, "learning_rate": 4.689481245510479e-05, "loss": 20.0794, "step": 11094 }, { "epoch": 0.46777831650399476, "grad_norm": 6835.3837890625, "learning_rate": 4.689403166869671e-05, "loss": 18.0849, "step": 11095 }, { "epoch": 0.46782047768619434, "grad_norm": 12777.6796875, "learning_rate": 4.6893250882288644e-05, "loss": 19.206, "step": 11096 }, { "epoch": 0.46786263886839385, "grad_norm": 16228.7177734375, "learning_rate": 4.6892470095880575e-05, "loss": 19.1226, "step": 11097 }, { "epoch": 0.46790480005059343, "grad_norm": 7201.40771484375, "learning_rate": 4.68916893094725e-05, "loss": 17.7065, "step": 11098 }, { "epoch": 0.46794696123279295, "grad_norm": 9929.001953125, "learning_rate": 4.689090852306443e-05, "loss": 17.4198, "step": 11099 }, { "epoch": 0.4679891224149925, "grad_norm": 8408.6044921875, "learning_rate": 4.689012773665636e-05, "loss": 17.3866, "step": 11100 }, { "epoch": 0.46803128359719204, "grad_norm": 7549.783203125, "learning_rate": 4.6889346950248294e-05, "loss": 16.787, "step": 11101 }, { "epoch": 0.4680734447793916, "grad_norm": 7600.82861328125, "learning_rate": 4.688856616384022e-05, "loss": 18.1563, "step": 11102 }, { "epoch": 0.4681156059615912, "grad_norm": 9381.05078125, "learning_rate": 4.6887785377432156e-05, "loss": 18.0213, "step": 11103 }, { "epoch": 0.4681577671437907, "grad_norm": 4243.75830078125, "learning_rate": 4.688700459102408e-05, "loss": 19.3545, "step": 11104 }, { "epoch": 0.4681999283259903, "grad_norm": 7709.759765625, "learning_rate": 4.688622380461601e-05, "loss": 17.5721, "step": 11105 }, { "epoch": 0.4682420895081898, "grad_norm": 8611.056640625, "learning_rate": 4.688544301820794e-05, "loss": 17.1344, "step": 11106 }, { "epoch": 0.46828425069038937, "grad_norm": 4449.8388671875, "learning_rate": 4.688466223179987e-05, "loss": 16.5762, "step": 11107 }, { "epoch": 0.4683264118725889, "grad_norm": 3673.2763671875, "learning_rate": 4.6883881445391806e-05, "loss": 16.9986, "step": 11108 }, { "epoch": 0.46836857305478846, "grad_norm": 4728.734375, "learning_rate": 4.688310065898373e-05, "loss": 16.0375, "step": 11109 }, { "epoch": 0.468410734236988, "grad_norm": 4994.7685546875, "learning_rate": 4.688231987257566e-05, "loss": 17.1159, "step": 11110 }, { "epoch": 0.46845289541918755, "grad_norm": 3340.146240234375, "learning_rate": 4.688153908616759e-05, "loss": 16.9704, "step": 11111 }, { "epoch": 0.4684950566013871, "grad_norm": 4595.6484375, "learning_rate": 4.688075829975952e-05, "loss": 15.1651, "step": 11112 }, { "epoch": 0.46853721778358665, "grad_norm": 1498.8560791015625, "learning_rate": 4.687997751335145e-05, "loss": 16.6952, "step": 11113 }, { "epoch": 0.4685793789657862, "grad_norm": 2363.8095703125, "learning_rate": 4.687919672694338e-05, "loss": 18.07, "step": 11114 }, { "epoch": 0.46862154014798574, "grad_norm": 2631.298095703125, "learning_rate": 4.687841594053531e-05, "loss": 16.5378, "step": 11115 }, { "epoch": 0.4686637013301853, "grad_norm": 2603.931884765625, "learning_rate": 4.687763515412724e-05, "loss": 16.3994, "step": 11116 }, { "epoch": 0.46870586251238483, "grad_norm": 1858.4190673828125, "learning_rate": 4.687685436771917e-05, "loss": 16.7007, "step": 11117 }, { "epoch": 0.4687480236945844, "grad_norm": 326.532470703125, "learning_rate": 4.68760735813111e-05, "loss": 17.3275, "step": 11118 }, { "epoch": 0.4687901848767839, "grad_norm": 997.8203735351562, "learning_rate": 4.687529279490303e-05, "loss": 17.3319, "step": 11119 }, { "epoch": 0.4688323460589835, "grad_norm": 705.1913452148438, "learning_rate": 4.687451200849496e-05, "loss": 16.032, "step": 11120 }, { "epoch": 0.46887450724118307, "grad_norm": 283.3387756347656, "learning_rate": 4.6873731222086885e-05, "loss": 17.8244, "step": 11121 }, { "epoch": 0.4689166684233826, "grad_norm": 539.6858520507812, "learning_rate": 4.6872950435678816e-05, "loss": 18.0461, "step": 11122 }, { "epoch": 0.46895882960558216, "grad_norm": 310.7389221191406, "learning_rate": 4.687216964927075e-05, "loss": 17.3144, "step": 11123 }, { "epoch": 0.4690009907877817, "grad_norm": 302.8934020996094, "learning_rate": 4.687138886286267e-05, "loss": 17.6152, "step": 11124 }, { "epoch": 0.46904315196998125, "grad_norm": 339.80999755859375, "learning_rate": 4.687060807645461e-05, "loss": 17.3133, "step": 11125 }, { "epoch": 0.46908531315218077, "grad_norm": 340.9446716308594, "learning_rate": 4.6869827290046535e-05, "loss": 17.489, "step": 11126 }, { "epoch": 0.46912747433438035, "grad_norm": 345.9542541503906, "learning_rate": 4.686904650363847e-05, "loss": 17.8868, "step": 11127 }, { "epoch": 0.46916963551657986, "grad_norm": 308.3093566894531, "learning_rate": 4.68682657172304e-05, "loss": 18.6104, "step": 11128 }, { "epoch": 0.46921179669877944, "grad_norm": 292.3078308105469, "learning_rate": 4.686748493082233e-05, "loss": 18.3603, "step": 11129 }, { "epoch": 0.46925395788097896, "grad_norm": 327.6274719238281, "learning_rate": 4.686670414441426e-05, "loss": 17.0527, "step": 11130 }, { "epoch": 0.46929611906317853, "grad_norm": 309.4680480957031, "learning_rate": 4.6865923358006184e-05, "loss": 17.8455, "step": 11131 }, { "epoch": 0.4693382802453781, "grad_norm": 294.6850280761719, "learning_rate": 4.6865142571598115e-05, "loss": 19.2993, "step": 11132 }, { "epoch": 0.4693804414275776, "grad_norm": 485.8830871582031, "learning_rate": 4.6864361785190047e-05, "loss": 18.7843, "step": 11133 }, { "epoch": 0.4694226026097772, "grad_norm": 211.62916564941406, "learning_rate": 4.686358099878198e-05, "loss": 19.7111, "step": 11134 }, { "epoch": 0.4694647637919767, "grad_norm": 374.09954833984375, "learning_rate": 4.68628002123739e-05, "loss": 19.0711, "step": 11135 }, { "epoch": 0.4695069249741763, "grad_norm": 711.07763671875, "learning_rate": 4.6862019425965834e-05, "loss": 18.5429, "step": 11136 }, { "epoch": 0.4695490861563758, "grad_norm": 610.7547607421875, "learning_rate": 4.6861238639557765e-05, "loss": 18.8725, "step": 11137 }, { "epoch": 0.4695912473385754, "grad_norm": 470.1511535644531, "learning_rate": 4.6860457853149696e-05, "loss": 20.216, "step": 11138 }, { "epoch": 0.4696334085207749, "grad_norm": 1315.36181640625, "learning_rate": 4.685967706674163e-05, "loss": 18.6669, "step": 11139 }, { "epoch": 0.46967556970297447, "grad_norm": 298.33441162109375, "learning_rate": 4.685889628033355e-05, "loss": 19.6817, "step": 11140 }, { "epoch": 0.46971773088517405, "grad_norm": 444.86578369140625, "learning_rate": 4.685811549392548e-05, "loss": 19.6009, "step": 11141 }, { "epoch": 0.46975989206737356, "grad_norm": 525.1151123046875, "learning_rate": 4.6857334707517414e-05, "loss": 17.8938, "step": 11142 }, { "epoch": 0.46980205324957314, "grad_norm": 1307.1846923828125, "learning_rate": 4.685655392110934e-05, "loss": 20.7481, "step": 11143 }, { "epoch": 0.46984421443177266, "grad_norm": 346.8219299316406, "learning_rate": 4.685577313470128e-05, "loss": 18.8753, "step": 11144 }, { "epoch": 0.46988637561397223, "grad_norm": 589.4281616210938, "learning_rate": 4.68549923482932e-05, "loss": 19.1675, "step": 11145 }, { "epoch": 0.46992853679617175, "grad_norm": 188.65664672851562, "learning_rate": 4.685421156188513e-05, "loss": 20.3262, "step": 11146 }, { "epoch": 0.4699706979783713, "grad_norm": 534.1218872070312, "learning_rate": 4.6853430775477064e-05, "loss": 19.5541, "step": 11147 }, { "epoch": 0.47001285916057084, "grad_norm": 622.2230834960938, "learning_rate": 4.6852649989068995e-05, "loss": 18.2752, "step": 11148 }, { "epoch": 0.4700550203427704, "grad_norm": 869.4801635742188, "learning_rate": 4.685186920266092e-05, "loss": 19.828, "step": 11149 }, { "epoch": 0.47009718152497, "grad_norm": 2456.24853515625, "learning_rate": 4.685108841625285e-05, "loss": 18.0721, "step": 11150 }, { "epoch": 0.4701393427071695, "grad_norm": 4379.55322265625, "learning_rate": 4.685030762984478e-05, "loss": 18.9587, "step": 11151 }, { "epoch": 0.4701815038893691, "grad_norm": 4942.2724609375, "learning_rate": 4.6849526843436713e-05, "loss": 17.9051, "step": 11152 }, { "epoch": 0.4702236650715686, "grad_norm": 1327.6717529296875, "learning_rate": 4.6848746057028645e-05, "loss": 18.6099, "step": 11153 }, { "epoch": 0.47026582625376817, "grad_norm": 821.2469482421875, "learning_rate": 4.684796527062057e-05, "loss": 18.7922, "step": 11154 }, { "epoch": 0.4703079874359677, "grad_norm": 2101.958251953125, "learning_rate": 4.68471844842125e-05, "loss": 18.3519, "step": 11155 }, { "epoch": 0.47035014861816726, "grad_norm": 852.6859741210938, "learning_rate": 4.684640369780443e-05, "loss": 19.078, "step": 11156 }, { "epoch": 0.4703923098003668, "grad_norm": 876.678466796875, "learning_rate": 4.6845622911396356e-05, "loss": 16.8847, "step": 11157 }, { "epoch": 0.47043447098256636, "grad_norm": 1371.1710205078125, "learning_rate": 4.6844842124988294e-05, "loss": 18.6404, "step": 11158 }, { "epoch": 0.4704766321647659, "grad_norm": 838.667724609375, "learning_rate": 4.684406133858022e-05, "loss": 18.2667, "step": 11159 }, { "epoch": 0.47051879334696545, "grad_norm": 7335.49658203125, "learning_rate": 4.684328055217215e-05, "loss": 18.3245, "step": 11160 }, { "epoch": 0.470560954529165, "grad_norm": 3072.5458984375, "learning_rate": 4.684249976576408e-05, "loss": 18.3197, "step": 11161 }, { "epoch": 0.47060311571136454, "grad_norm": 1449.6966552734375, "learning_rate": 4.6841718979356006e-05, "loss": 18.1069, "step": 11162 }, { "epoch": 0.4706452768935641, "grad_norm": 1343.6717529296875, "learning_rate": 4.6840938192947944e-05, "loss": 18.2322, "step": 11163 }, { "epoch": 0.47068743807576363, "grad_norm": 1792.9747314453125, "learning_rate": 4.684015740653987e-05, "loss": 17.1972, "step": 11164 }, { "epoch": 0.4707295992579632, "grad_norm": 3360.72216796875, "learning_rate": 4.68393766201318e-05, "loss": 18.4669, "step": 11165 }, { "epoch": 0.4707717604401627, "grad_norm": 332.1051025390625, "learning_rate": 4.683859583372373e-05, "loss": 20.0878, "step": 11166 }, { "epoch": 0.4708139216223623, "grad_norm": 490.8712463378906, "learning_rate": 4.6837815047315655e-05, "loss": 19.5158, "step": 11167 }, { "epoch": 0.4708560828045618, "grad_norm": 2696.515625, "learning_rate": 4.6837034260907586e-05, "loss": 17.9781, "step": 11168 }, { "epoch": 0.4708982439867614, "grad_norm": 2338.98779296875, "learning_rate": 4.683625347449952e-05, "loss": 19.9849, "step": 11169 }, { "epoch": 0.47094040516896096, "grad_norm": 1010.6285400390625, "learning_rate": 4.683547268809145e-05, "loss": 19.0039, "step": 11170 }, { "epoch": 0.4709825663511605, "grad_norm": 1457.6666259765625, "learning_rate": 4.6834691901683373e-05, "loss": 17.8822, "step": 11171 }, { "epoch": 0.47102472753336005, "grad_norm": 2578.351318359375, "learning_rate": 4.683391111527531e-05, "loss": 17.8584, "step": 11172 }, { "epoch": 0.4710668887155596, "grad_norm": 120615.0859375, "learning_rate": 4.6833130328867236e-05, "loss": 17.1896, "step": 11173 }, { "epoch": 0.47110904989775915, "grad_norm": 1376.897216796875, "learning_rate": 4.683234954245917e-05, "loss": 17.4364, "step": 11174 }, { "epoch": 0.47115121107995867, "grad_norm": 4436.79150390625, "learning_rate": 4.68315687560511e-05, "loss": 15.8229, "step": 11175 }, { "epoch": 0.47119337226215824, "grad_norm": 1684.0428466796875, "learning_rate": 4.683078796964302e-05, "loss": 15.9433, "step": 11176 }, { "epoch": 0.47123553344435776, "grad_norm": 1187.1123046875, "learning_rate": 4.683000718323496e-05, "loss": 18.7274, "step": 11177 }, { "epoch": 0.47127769462655733, "grad_norm": 7616.19189453125, "learning_rate": 4.6829226396826886e-05, "loss": 16.443, "step": 11178 }, { "epoch": 0.4713198558087569, "grad_norm": 823.0576171875, "learning_rate": 4.682844561041882e-05, "loss": 16.5357, "step": 11179 }, { "epoch": 0.4713620169909564, "grad_norm": 2063.25439453125, "learning_rate": 4.682766482401075e-05, "loss": 17.1604, "step": 11180 }, { "epoch": 0.471404178173156, "grad_norm": 6578.6650390625, "learning_rate": 4.682688403760267e-05, "loss": 15.5843, "step": 11181 }, { "epoch": 0.4714463393553555, "grad_norm": 660.839599609375, "learning_rate": 4.6826103251194604e-05, "loss": 15.8305, "step": 11182 }, { "epoch": 0.4714885005375551, "grad_norm": 1237.6494140625, "learning_rate": 4.6825322464786535e-05, "loss": 16.4696, "step": 11183 }, { "epoch": 0.4715306617197546, "grad_norm": 12435.80078125, "learning_rate": 4.6824541678378466e-05, "loss": 14.6666, "step": 11184 }, { "epoch": 0.4715728229019542, "grad_norm": 2622.70068359375, "learning_rate": 4.68237608919704e-05, "loss": 15.1205, "step": 11185 }, { "epoch": 0.4716149840841537, "grad_norm": 577.2345581054688, "learning_rate": 4.682298010556232e-05, "loss": 16.8783, "step": 11186 }, { "epoch": 0.4716571452663533, "grad_norm": 2756.143310546875, "learning_rate": 4.682219931915425e-05, "loss": 16.4608, "step": 11187 }, { "epoch": 0.4716993064485528, "grad_norm": 809.8614501953125, "learning_rate": 4.6821418532746185e-05, "loss": 15.6423, "step": 11188 }, { "epoch": 0.47174146763075236, "grad_norm": 1457.6610107421875, "learning_rate": 4.6820637746338116e-05, "loss": 16.7466, "step": 11189 }, { "epoch": 0.47178362881295194, "grad_norm": 3403.962646484375, "learning_rate": 4.681985695993004e-05, "loss": 15.6766, "step": 11190 }, { "epoch": 0.47182578999515146, "grad_norm": 1062.0048828125, "learning_rate": 4.681907617352198e-05, "loss": 17.4051, "step": 11191 }, { "epoch": 0.47186795117735103, "grad_norm": 835.6171264648438, "learning_rate": 4.68182953871139e-05, "loss": 16.6956, "step": 11192 }, { "epoch": 0.47191011235955055, "grad_norm": 1053.70556640625, "learning_rate": 4.6817514600705834e-05, "loss": 16.3527, "step": 11193 }, { "epoch": 0.4719522735417501, "grad_norm": 517.3336181640625, "learning_rate": 4.6816733814297765e-05, "loss": 15.8241, "step": 11194 }, { "epoch": 0.47199443472394964, "grad_norm": 732.8933715820312, "learning_rate": 4.681595302788969e-05, "loss": 15.7209, "step": 11195 }, { "epoch": 0.4720365959061492, "grad_norm": 1849.02734375, "learning_rate": 4.681517224148163e-05, "loss": 16.1066, "step": 11196 }, { "epoch": 0.47207875708834873, "grad_norm": 118.44274139404297, "learning_rate": 4.681439145507355e-05, "loss": 16.3472, "step": 11197 }, { "epoch": 0.4721209182705483, "grad_norm": 736.5899658203125, "learning_rate": 4.6813610668665484e-05, "loss": 15.826, "step": 11198 }, { "epoch": 0.4721630794527479, "grad_norm": 23589.36328125, "learning_rate": 4.6812829882257415e-05, "loss": 16.72, "step": 11199 }, { "epoch": 0.4722052406349474, "grad_norm": 232.028564453125, "learning_rate": 4.681204909584934e-05, "loss": 16.6155, "step": 11200 }, { "epoch": 0.472247401817147, "grad_norm": 521.478759765625, "learning_rate": 4.681126830944127e-05, "loss": 16.4402, "step": 11201 }, { "epoch": 0.4722895629993465, "grad_norm": 259.2904052734375, "learning_rate": 4.68104875230332e-05, "loss": 17.5018, "step": 11202 }, { "epoch": 0.47233172418154606, "grad_norm": 428.8916015625, "learning_rate": 4.680970673662513e-05, "loss": 18.6687, "step": 11203 }, { "epoch": 0.4723738853637456, "grad_norm": 1348.779296875, "learning_rate": 4.680892595021706e-05, "loss": 15.9663, "step": 11204 }, { "epoch": 0.47241604654594516, "grad_norm": 1235.2349853515625, "learning_rate": 4.680814516380899e-05, "loss": 16.913, "step": 11205 }, { "epoch": 0.4724582077281447, "grad_norm": 1422.8447265625, "learning_rate": 4.680736437740092e-05, "loss": 16.8278, "step": 11206 }, { "epoch": 0.47250036891034425, "grad_norm": 116.66036987304688, "learning_rate": 4.680658359099285e-05, "loss": 16.7195, "step": 11207 }, { "epoch": 0.4725425300925438, "grad_norm": 563.7160034179688, "learning_rate": 4.680580280458478e-05, "loss": 16.8271, "step": 11208 }, { "epoch": 0.47258469127474334, "grad_norm": 284.60626220703125, "learning_rate": 4.680502201817671e-05, "loss": 15.6192, "step": 11209 }, { "epoch": 0.4726268524569429, "grad_norm": 187.8636016845703, "learning_rate": 4.6804241231768645e-05, "loss": 18.4795, "step": 11210 }, { "epoch": 0.47266901363914243, "grad_norm": 564.5892333984375, "learning_rate": 4.680346044536057e-05, "loss": 15.5757, "step": 11211 }, { "epoch": 0.472711174821342, "grad_norm": 347.3932800292969, "learning_rate": 4.6802679658952494e-05, "loss": 18.1975, "step": 11212 }, { "epoch": 0.4727533360035415, "grad_norm": 891.235595703125, "learning_rate": 4.680189887254443e-05, "loss": 16.8243, "step": 11213 }, { "epoch": 0.4727954971857411, "grad_norm": 1257.776123046875, "learning_rate": 4.680111808613636e-05, "loss": 17.0609, "step": 11214 }, { "epoch": 0.4728376583679406, "grad_norm": 1235.759765625, "learning_rate": 4.680033729972829e-05, "loss": 19.5101, "step": 11215 }, { "epoch": 0.4728798195501402, "grad_norm": 261.8477478027344, "learning_rate": 4.679955651332022e-05, "loss": 15.6967, "step": 11216 }, { "epoch": 0.4729219807323397, "grad_norm": 2765.8896484375, "learning_rate": 4.679877572691215e-05, "loss": 17.0392, "step": 11217 }, { "epoch": 0.4729641419145393, "grad_norm": 2072.561279296875, "learning_rate": 4.6797994940504075e-05, "loss": 16.5689, "step": 11218 }, { "epoch": 0.47300630309673886, "grad_norm": 1778.49462890625, "learning_rate": 4.6797214154096006e-05, "loss": 16.3282, "step": 11219 }, { "epoch": 0.4730484642789384, "grad_norm": 1148.42724609375, "learning_rate": 4.679643336768794e-05, "loss": 16.521, "step": 11220 }, { "epoch": 0.47309062546113795, "grad_norm": 2872.9814453125, "learning_rate": 4.679565258127987e-05, "loss": 18.5534, "step": 11221 }, { "epoch": 0.47313278664333747, "grad_norm": 753.5897827148438, "learning_rate": 4.67948717948718e-05, "loss": 17.1067, "step": 11222 }, { "epoch": 0.47317494782553704, "grad_norm": 816.4202880859375, "learning_rate": 4.6794091008463724e-05, "loss": 16.3276, "step": 11223 }, { "epoch": 0.47321710900773656, "grad_norm": 2992.51708984375, "learning_rate": 4.6793310222055656e-05, "loss": 17.1432, "step": 11224 }, { "epoch": 0.47325927018993613, "grad_norm": 895.5631713867188, "learning_rate": 4.679252943564759e-05, "loss": 17.0672, "step": 11225 }, { "epoch": 0.47330143137213565, "grad_norm": 3591.96044921875, "learning_rate": 4.679174864923951e-05, "loss": 17.1205, "step": 11226 }, { "epoch": 0.4733435925543352, "grad_norm": 1931.5775146484375, "learning_rate": 4.679096786283145e-05, "loss": 15.0848, "step": 11227 }, { "epoch": 0.4733857537365348, "grad_norm": 2729.794677734375, "learning_rate": 4.6790187076423374e-05, "loss": 17.6772, "step": 11228 }, { "epoch": 0.4734279149187343, "grad_norm": 7811.90185546875, "learning_rate": 4.6789406290015305e-05, "loss": 17.8838, "step": 11229 }, { "epoch": 0.4734700761009339, "grad_norm": 7010.60205078125, "learning_rate": 4.6788625503607237e-05, "loss": 16.123, "step": 11230 }, { "epoch": 0.4735122372831334, "grad_norm": 6054.47705078125, "learning_rate": 4.678784471719916e-05, "loss": 17.0219, "step": 11231 }, { "epoch": 0.473554398465333, "grad_norm": 3639.57568359375, "learning_rate": 4.67870639307911e-05, "loss": 18.3676, "step": 11232 }, { "epoch": 0.4735965596475325, "grad_norm": 4118.77197265625, "learning_rate": 4.6786283144383024e-05, "loss": 15.7285, "step": 11233 }, { "epoch": 0.4736387208297321, "grad_norm": 1343.260498046875, "learning_rate": 4.6785502357974955e-05, "loss": 17.4058, "step": 11234 }, { "epoch": 0.4736808820119316, "grad_norm": 669.6691284179688, "learning_rate": 4.6784721571566886e-05, "loss": 14.0404, "step": 11235 }, { "epoch": 0.47372304319413117, "grad_norm": 1069.41455078125, "learning_rate": 4.678394078515882e-05, "loss": 14.8992, "step": 11236 }, { "epoch": 0.47376520437633074, "grad_norm": 489.5892028808594, "learning_rate": 4.678315999875074e-05, "loss": 14.3763, "step": 11237 }, { "epoch": 0.47380736555853026, "grad_norm": 10537.6630859375, "learning_rate": 4.678237921234267e-05, "loss": 14.936, "step": 11238 }, { "epoch": 0.47384952674072983, "grad_norm": 9672.7734375, "learning_rate": 4.6781598425934604e-05, "loss": 15.0232, "step": 11239 }, { "epoch": 0.47389168792292935, "grad_norm": 6798.890625, "learning_rate": 4.678081763952653e-05, "loss": 14.8965, "step": 11240 }, { "epoch": 0.4739338491051289, "grad_norm": 87618.9765625, "learning_rate": 4.678003685311847e-05, "loss": 14.9314, "step": 11241 }, { "epoch": 0.47397601028732844, "grad_norm": 539.922119140625, "learning_rate": 4.677925606671039e-05, "loss": 15.5087, "step": 11242 }, { "epoch": 0.474018171469528, "grad_norm": 3704.78173828125, "learning_rate": 4.677847528030232e-05, "loss": 16.4749, "step": 11243 }, { "epoch": 0.47406033265172753, "grad_norm": 9762.392578125, "learning_rate": 4.6777694493894254e-05, "loss": 16.1475, "step": 11244 }, { "epoch": 0.4741024938339271, "grad_norm": 2705.045166015625, "learning_rate": 4.677691370748618e-05, "loss": 15.6045, "step": 11245 }, { "epoch": 0.4741446550161266, "grad_norm": 7600.30322265625, "learning_rate": 4.6776132921078116e-05, "loss": 14.6605, "step": 11246 }, { "epoch": 0.4741868161983262, "grad_norm": 13232.8310546875, "learning_rate": 4.677535213467004e-05, "loss": 15.3981, "step": 11247 }, { "epoch": 0.4742289773805258, "grad_norm": 1442.178466796875, "learning_rate": 4.677457134826197e-05, "loss": 14.2296, "step": 11248 }, { "epoch": 0.4742711385627253, "grad_norm": 9722.3603515625, "learning_rate": 4.67737905618539e-05, "loss": 14.9641, "step": 11249 }, { "epoch": 0.47431329974492487, "grad_norm": 2831.3857421875, "learning_rate": 4.677300977544583e-05, "loss": 16.0217, "step": 11250 }, { "epoch": 0.4743554609271244, "grad_norm": 13459.6376953125, "learning_rate": 4.677222898903776e-05, "loss": 16.9141, "step": 11251 }, { "epoch": 0.47439762210932396, "grad_norm": 20339.119140625, "learning_rate": 4.677144820262969e-05, "loss": 16.4938, "step": 11252 }, { "epoch": 0.4744397832915235, "grad_norm": 744.0917358398438, "learning_rate": 4.677066741622162e-05, "loss": 16.2044, "step": 11253 }, { "epoch": 0.47448194447372305, "grad_norm": 4760.6923828125, "learning_rate": 4.676988662981355e-05, "loss": 15.9527, "step": 11254 }, { "epoch": 0.47452410565592257, "grad_norm": 13452.4404296875, "learning_rate": 4.6769105843405484e-05, "loss": 15.4872, "step": 11255 }, { "epoch": 0.47456626683812214, "grad_norm": 672.3013916015625, "learning_rate": 4.676832505699741e-05, "loss": 15.8444, "step": 11256 }, { "epoch": 0.4746084280203217, "grad_norm": 907.3564453125, "learning_rate": 4.676754427058934e-05, "loss": 14.532, "step": 11257 }, { "epoch": 0.47465058920252123, "grad_norm": 1499.4012451171875, "learning_rate": 4.676676348418127e-05, "loss": 13.796, "step": 11258 }, { "epoch": 0.4746927503847208, "grad_norm": 767.2735595703125, "learning_rate": 4.6765982697773196e-05, "loss": 14.8702, "step": 11259 }, { "epoch": 0.4747349115669203, "grad_norm": 1131.2818603515625, "learning_rate": 4.6765201911365134e-05, "loss": 15.5241, "step": 11260 }, { "epoch": 0.4747770727491199, "grad_norm": 305.8817138671875, "learning_rate": 4.676442112495706e-05, "loss": 15.717, "step": 11261 }, { "epoch": 0.4748192339313194, "grad_norm": 3250.846923828125, "learning_rate": 4.676364033854899e-05, "loss": 14.6041, "step": 11262 }, { "epoch": 0.474861395113519, "grad_norm": 420.33441162109375, "learning_rate": 4.676285955214092e-05, "loss": 14.8042, "step": 11263 }, { "epoch": 0.4749035562957185, "grad_norm": 1809.1055908203125, "learning_rate": 4.6762078765732845e-05, "loss": 13.7726, "step": 11264 }, { "epoch": 0.4749457174779181, "grad_norm": 1857.34130859375, "learning_rate": 4.676129797932478e-05, "loss": 15.4534, "step": 11265 }, { "epoch": 0.47498787866011766, "grad_norm": 1517.337890625, "learning_rate": 4.676051719291671e-05, "loss": 14.842, "step": 11266 }, { "epoch": 0.4750300398423172, "grad_norm": 2506.469482421875, "learning_rate": 4.675973640650864e-05, "loss": 14.4763, "step": 11267 }, { "epoch": 0.47507220102451675, "grad_norm": 1277.7012939453125, "learning_rate": 4.675895562010057e-05, "loss": 14.81, "step": 11268 }, { "epoch": 0.47511436220671627, "grad_norm": 1465.3980712890625, "learning_rate": 4.6758174833692495e-05, "loss": 16.9165, "step": 11269 }, { "epoch": 0.47515652338891584, "grad_norm": 2976.990478515625, "learning_rate": 4.6757394047284426e-05, "loss": 16.4962, "step": 11270 }, { "epoch": 0.47519868457111536, "grad_norm": 4129.02294921875, "learning_rate": 4.675661326087636e-05, "loss": 15.3707, "step": 11271 }, { "epoch": 0.47524084575331493, "grad_norm": 26404.5703125, "learning_rate": 4.675583247446829e-05, "loss": 14.5325, "step": 11272 }, { "epoch": 0.47528300693551445, "grad_norm": 8740.138671875, "learning_rate": 4.675505168806021e-05, "loss": 15.5446, "step": 11273 }, { "epoch": 0.475325168117714, "grad_norm": 6425.57763671875, "learning_rate": 4.6754270901652144e-05, "loss": 15.4216, "step": 11274 }, { "epoch": 0.47536732929991354, "grad_norm": 1239.2413330078125, "learning_rate": 4.6753490115244075e-05, "loss": 14.034, "step": 11275 }, { "epoch": 0.4754094904821131, "grad_norm": 1367.381591796875, "learning_rate": 4.675270932883601e-05, "loss": 14.7477, "step": 11276 }, { "epoch": 0.4754516516643127, "grad_norm": 831.896240234375, "learning_rate": 4.675192854242794e-05, "loss": 15.0859, "step": 11277 }, { "epoch": 0.4754938128465122, "grad_norm": 1780.51806640625, "learning_rate": 4.675114775601986e-05, "loss": 14.4433, "step": 11278 }, { "epoch": 0.4755359740287118, "grad_norm": 2236.10546875, "learning_rate": 4.67503669696118e-05, "loss": 16.2546, "step": 11279 }, { "epoch": 0.4755781352109113, "grad_norm": 2090.323486328125, "learning_rate": 4.6749586183203725e-05, "loss": 15.172, "step": 11280 }, { "epoch": 0.4756202963931109, "grad_norm": 1684.7083740234375, "learning_rate": 4.6748805396795656e-05, "loss": 14.1051, "step": 11281 }, { "epoch": 0.4756624575753104, "grad_norm": 1852.519775390625, "learning_rate": 4.674802461038759e-05, "loss": 15.0353, "step": 11282 }, { "epoch": 0.47570461875750997, "grad_norm": 1145.01025390625, "learning_rate": 4.674724382397951e-05, "loss": 13.9543, "step": 11283 }, { "epoch": 0.4757467799397095, "grad_norm": 2106.82861328125, "learning_rate": 4.674646303757144e-05, "loss": 13.7285, "step": 11284 }, { "epoch": 0.47578894112190906, "grad_norm": 3130.158203125, "learning_rate": 4.6745682251163374e-05, "loss": 14.2206, "step": 11285 }, { "epoch": 0.47583110230410863, "grad_norm": 698.6174926757812, "learning_rate": 4.6744901464755306e-05, "loss": 18.2021, "step": 11286 }, { "epoch": 0.47587326348630815, "grad_norm": 508.8934631347656, "learning_rate": 4.674412067834724e-05, "loss": 15.9707, "step": 11287 }, { "epoch": 0.4759154246685077, "grad_norm": 2268.046875, "learning_rate": 4.674333989193916e-05, "loss": 14.725, "step": 11288 }, { "epoch": 0.47595758585070724, "grad_norm": 1888.6650390625, "learning_rate": 4.674255910553109e-05, "loss": 14.6978, "step": 11289 }, { "epoch": 0.4759997470329068, "grad_norm": 4089.103759765625, "learning_rate": 4.6741778319123024e-05, "loss": 15.3518, "step": 11290 }, { "epoch": 0.47604190821510634, "grad_norm": 1557.2799072265625, "learning_rate": 4.6740997532714955e-05, "loss": 14.9855, "step": 11291 }, { "epoch": 0.4760840693973059, "grad_norm": 1595.2081298828125, "learning_rate": 4.674021674630688e-05, "loss": 14.1043, "step": 11292 }, { "epoch": 0.4761262305795054, "grad_norm": 1010.2977905273438, "learning_rate": 4.673943595989881e-05, "loss": 14.5378, "step": 11293 }, { "epoch": 0.476168391761705, "grad_norm": 1180.4136962890625, "learning_rate": 4.673865517349074e-05, "loss": 13.7647, "step": 11294 }, { "epoch": 0.4762105529439046, "grad_norm": 623.5261840820312, "learning_rate": 4.673787438708267e-05, "loss": 13.9331, "step": 11295 }, { "epoch": 0.4762527141261041, "grad_norm": 617.9439086914062, "learning_rate": 4.6737093600674605e-05, "loss": 14.8169, "step": 11296 }, { "epoch": 0.47629487530830367, "grad_norm": 361.6736755371094, "learning_rate": 4.673631281426653e-05, "loss": 18.3333, "step": 11297 }, { "epoch": 0.4763370364905032, "grad_norm": 311.9393005371094, "learning_rate": 4.673553202785846e-05, "loss": 16.9678, "step": 11298 }, { "epoch": 0.47637919767270276, "grad_norm": 130.24615478515625, "learning_rate": 4.673475124145039e-05, "loss": 15.975, "step": 11299 }, { "epoch": 0.4764213588549023, "grad_norm": 332.3328857421875, "learning_rate": 4.6733970455042316e-05, "loss": 16.0413, "step": 11300 }, { "epoch": 0.47646352003710185, "grad_norm": 594.8782958984375, "learning_rate": 4.6733189668634254e-05, "loss": 15.9317, "step": 11301 }, { "epoch": 0.47650568121930137, "grad_norm": 384.921142578125, "learning_rate": 4.673240888222618e-05, "loss": 16.1115, "step": 11302 }, { "epoch": 0.47654784240150094, "grad_norm": 183.04110717773438, "learning_rate": 4.673162809581811e-05, "loss": 16.8068, "step": 11303 }, { "epoch": 0.47659000358370046, "grad_norm": 303.8658142089844, "learning_rate": 4.673084730941004e-05, "loss": 16.5695, "step": 11304 }, { "epoch": 0.47663216476590003, "grad_norm": 471.2748107910156, "learning_rate": 4.673006652300197e-05, "loss": 15.0795, "step": 11305 }, { "epoch": 0.4766743259480996, "grad_norm": 210.53076171875, "learning_rate": 4.67292857365939e-05, "loss": 15.668, "step": 11306 }, { "epoch": 0.4767164871302991, "grad_norm": 451.3153076171875, "learning_rate": 4.672850495018583e-05, "loss": 15.4519, "step": 11307 }, { "epoch": 0.4767586483124987, "grad_norm": 168.0139617919922, "learning_rate": 4.672772416377776e-05, "loss": 17.3081, "step": 11308 }, { "epoch": 0.4768008094946982, "grad_norm": 576.9119873046875, "learning_rate": 4.6726943377369684e-05, "loss": 16.4414, "step": 11309 }, { "epoch": 0.4768429706768978, "grad_norm": 371.51806640625, "learning_rate": 4.672616259096162e-05, "loss": 15.8819, "step": 11310 }, { "epoch": 0.4768851318590973, "grad_norm": 109.46621704101562, "learning_rate": 4.6725381804553547e-05, "loss": 16.2133, "step": 11311 }, { "epoch": 0.4769272930412969, "grad_norm": 113.96572875976562, "learning_rate": 4.672460101814548e-05, "loss": 15.1683, "step": 11312 }, { "epoch": 0.4769694542234964, "grad_norm": 464.9334716796875, "learning_rate": 4.672382023173741e-05, "loss": 17.1475, "step": 11313 }, { "epoch": 0.477011615405696, "grad_norm": 529.3947143554688, "learning_rate": 4.6723039445329334e-05, "loss": 19.5534, "step": 11314 }, { "epoch": 0.47705377658789555, "grad_norm": 1784.6484375, "learning_rate": 4.672225865892127e-05, "loss": 16.7812, "step": 11315 }, { "epoch": 0.47709593777009507, "grad_norm": 396.82452392578125, "learning_rate": 4.6721477872513196e-05, "loss": 17.3087, "step": 11316 }, { "epoch": 0.47713809895229464, "grad_norm": 965.7152709960938, "learning_rate": 4.672069708610513e-05, "loss": 16.4144, "step": 11317 }, { "epoch": 0.47718026013449416, "grad_norm": 448.09326171875, "learning_rate": 4.671991629969706e-05, "loss": 16.8168, "step": 11318 }, { "epoch": 0.47722242131669373, "grad_norm": 3405.547119140625, "learning_rate": 4.671913551328898e-05, "loss": 16.903, "step": 11319 }, { "epoch": 0.47726458249889325, "grad_norm": 8029.84375, "learning_rate": 4.6718354726880914e-05, "loss": 17.2925, "step": 11320 }, { "epoch": 0.4773067436810928, "grad_norm": 1034.3848876953125, "learning_rate": 4.6717573940472846e-05, "loss": 16.9029, "step": 11321 }, { "epoch": 0.47734890486329234, "grad_norm": 4065.587890625, "learning_rate": 4.671679315406478e-05, "loss": 15.9448, "step": 11322 }, { "epoch": 0.4773910660454919, "grad_norm": 3214.179443359375, "learning_rate": 4.671601236765671e-05, "loss": 16.395, "step": 11323 }, { "epoch": 0.47743322722769144, "grad_norm": 1662.7633056640625, "learning_rate": 4.671523158124864e-05, "loss": 17.3471, "step": 11324 }, { "epoch": 0.477475388409891, "grad_norm": 2678.116943359375, "learning_rate": 4.6714450794840564e-05, "loss": 15.8708, "step": 11325 }, { "epoch": 0.4775175495920906, "grad_norm": 1258.6409912109375, "learning_rate": 4.6713670008432495e-05, "loss": 18.2126, "step": 11326 }, { "epoch": 0.4775597107742901, "grad_norm": 475.0738830566406, "learning_rate": 4.6712889222024426e-05, "loss": 17.1161, "step": 11327 }, { "epoch": 0.4776018719564897, "grad_norm": 13468.2265625, "learning_rate": 4.671210843561635e-05, "loss": 16.2464, "step": 11328 }, { "epoch": 0.4776440331386892, "grad_norm": 1942.40966796875, "learning_rate": 4.671132764920829e-05, "loss": 19.1743, "step": 11329 }, { "epoch": 0.47768619432088877, "grad_norm": 12931.7880859375, "learning_rate": 4.6710546862800213e-05, "loss": 17.746, "step": 11330 }, { "epoch": 0.4777283555030883, "grad_norm": 3827.50146484375, "learning_rate": 4.6709766076392145e-05, "loss": 16.8367, "step": 11331 }, { "epoch": 0.47777051668528786, "grad_norm": 14803.705078125, "learning_rate": 4.6708985289984076e-05, "loss": 15.5003, "step": 11332 }, { "epoch": 0.4778126778674874, "grad_norm": 5192.68994140625, "learning_rate": 4.6708204503576e-05, "loss": 16.7257, "step": 11333 }, { "epoch": 0.47785483904968695, "grad_norm": 3721.08740234375, "learning_rate": 4.670742371716794e-05, "loss": 16.1934, "step": 11334 }, { "epoch": 0.4778970002318865, "grad_norm": 1196.3001708984375, "learning_rate": 4.670664293075986e-05, "loss": 16.644, "step": 11335 }, { "epoch": 0.47793916141408604, "grad_norm": 11368.380859375, "learning_rate": 4.6705862144351794e-05, "loss": 15.2526, "step": 11336 }, { "epoch": 0.4779813225962856, "grad_norm": 54203.54296875, "learning_rate": 4.6705081357943725e-05, "loss": 16.1907, "step": 11337 }, { "epoch": 0.47802348377848514, "grad_norm": 13638.3232421875, "learning_rate": 4.670430057153565e-05, "loss": 18.2447, "step": 11338 }, { "epoch": 0.4780656449606847, "grad_norm": 5045.509765625, "learning_rate": 4.670351978512758e-05, "loss": 15.9791, "step": 11339 }, { "epoch": 0.47810780614288423, "grad_norm": 724.9221801757812, "learning_rate": 4.670273899871951e-05, "loss": 15.84, "step": 11340 }, { "epoch": 0.4781499673250838, "grad_norm": 1821.834716796875, "learning_rate": 4.6701958212311444e-05, "loss": 15.7227, "step": 11341 }, { "epoch": 0.4781921285072833, "grad_norm": 63481.9609375, "learning_rate": 4.670117742590337e-05, "loss": 16.8729, "step": 11342 }, { "epoch": 0.4782342896894829, "grad_norm": 231.51052856445312, "learning_rate": 4.6700396639495306e-05, "loss": 14.4643, "step": 11343 }, { "epoch": 0.47827645087168247, "grad_norm": 277.19122314453125, "learning_rate": 4.669961585308723e-05, "loss": 14.6252, "step": 11344 }, { "epoch": 0.478318612053882, "grad_norm": 1855.251220703125, "learning_rate": 4.669883506667916e-05, "loss": 14.8206, "step": 11345 }, { "epoch": 0.47836077323608156, "grad_norm": 521.650146484375, "learning_rate": 4.669805428027109e-05, "loss": 16.8407, "step": 11346 }, { "epoch": 0.4784029344182811, "grad_norm": 690.765625, "learning_rate": 4.669727349386302e-05, "loss": 14.8011, "step": 11347 }, { "epoch": 0.47844509560048065, "grad_norm": 1209.2308349609375, "learning_rate": 4.6696492707454956e-05, "loss": 14.6259, "step": 11348 }, { "epoch": 0.47848725678268017, "grad_norm": 2157.29150390625, "learning_rate": 4.669571192104688e-05, "loss": 15.6487, "step": 11349 }, { "epoch": 0.47852941796487974, "grad_norm": 2227.27490234375, "learning_rate": 4.669493113463881e-05, "loss": 14.7252, "step": 11350 }, { "epoch": 0.47857157914707926, "grad_norm": 1822.112548828125, "learning_rate": 4.669415034823074e-05, "loss": 15.5345, "step": 11351 }, { "epoch": 0.47861374032927884, "grad_norm": 215.47767639160156, "learning_rate": 4.669336956182267e-05, "loss": 14.8924, "step": 11352 }, { "epoch": 0.47865590151147835, "grad_norm": 1000.676513671875, "learning_rate": 4.66925887754146e-05, "loss": 14.1733, "step": 11353 }, { "epoch": 0.47869806269367793, "grad_norm": 5989.55810546875, "learning_rate": 4.669180798900653e-05, "loss": 16.3138, "step": 11354 }, { "epoch": 0.4787402238758775, "grad_norm": 1739.113525390625, "learning_rate": 4.669102720259846e-05, "loss": 17.2527, "step": 11355 }, { "epoch": 0.478782385058077, "grad_norm": 678.484130859375, "learning_rate": 4.669024641619039e-05, "loss": 16.0197, "step": 11356 }, { "epoch": 0.4788245462402766, "grad_norm": 979.837158203125, "learning_rate": 4.668946562978232e-05, "loss": 15.5808, "step": 11357 }, { "epoch": 0.4788667074224761, "grad_norm": 2642.97314453125, "learning_rate": 4.668868484337425e-05, "loss": 14.2847, "step": 11358 }, { "epoch": 0.4789088686046757, "grad_norm": 548.1511840820312, "learning_rate": 4.668790405696618e-05, "loss": 17.7853, "step": 11359 }, { "epoch": 0.4789510297868752, "grad_norm": 1552.909912109375, "learning_rate": 4.668712327055811e-05, "loss": 16.5979, "step": 11360 }, { "epoch": 0.4789931909690748, "grad_norm": 166.24087524414062, "learning_rate": 4.6686342484150035e-05, "loss": 18.4855, "step": 11361 }, { "epoch": 0.4790353521512743, "grad_norm": 452.9913024902344, "learning_rate": 4.6685561697741966e-05, "loss": 16.4067, "step": 11362 }, { "epoch": 0.47907751333347387, "grad_norm": 912.6624755859375, "learning_rate": 4.66847809113339e-05, "loss": 16.474, "step": 11363 }, { "epoch": 0.47911967451567344, "grad_norm": 647.1862182617188, "learning_rate": 4.668400012492582e-05, "loss": 16.7874, "step": 11364 }, { "epoch": 0.47916183569787296, "grad_norm": 1164.68359375, "learning_rate": 4.668321933851776e-05, "loss": 15.109, "step": 11365 }, { "epoch": 0.47920399688007254, "grad_norm": 453.4937744140625, "learning_rate": 4.6682438552109685e-05, "loss": 15.8424, "step": 11366 }, { "epoch": 0.47924615806227205, "grad_norm": 293.740478515625, "learning_rate": 4.668165776570162e-05, "loss": 17.5009, "step": 11367 }, { "epoch": 0.47928831924447163, "grad_norm": 200.1683807373047, "learning_rate": 4.668087697929355e-05, "loss": 17.4399, "step": 11368 }, { "epoch": 0.47933048042667115, "grad_norm": 95.69328308105469, "learning_rate": 4.668009619288548e-05, "loss": 18.9852, "step": 11369 }, { "epoch": 0.4793726416088707, "grad_norm": 118.06135559082031, "learning_rate": 4.667931540647741e-05, "loss": 15.5851, "step": 11370 }, { "epoch": 0.47941480279107024, "grad_norm": 667.72705078125, "learning_rate": 4.6678534620069334e-05, "loss": 16.3102, "step": 11371 }, { "epoch": 0.4794569639732698, "grad_norm": 635.0054931640625, "learning_rate": 4.6677753833661265e-05, "loss": 16.7729, "step": 11372 }, { "epoch": 0.4794991251554694, "grad_norm": 289.4246826171875, "learning_rate": 4.6676973047253197e-05, "loss": 17.6241, "step": 11373 }, { "epoch": 0.4795412863376689, "grad_norm": 129.4275665283203, "learning_rate": 4.667619226084513e-05, "loss": 15.9794, "step": 11374 }, { "epoch": 0.4795834475198685, "grad_norm": 7579.015625, "learning_rate": 4.667541147443705e-05, "loss": 17.1595, "step": 11375 }, { "epoch": 0.479625608702068, "grad_norm": 115.94728088378906, "learning_rate": 4.6674630688028984e-05, "loss": 17.3902, "step": 11376 }, { "epoch": 0.47966776988426757, "grad_norm": 183.32801818847656, "learning_rate": 4.6673849901620915e-05, "loss": 16.1875, "step": 11377 }, { "epoch": 0.4797099310664671, "grad_norm": 158.64695739746094, "learning_rate": 4.667306911521284e-05, "loss": 19.1461, "step": 11378 }, { "epoch": 0.47975209224866666, "grad_norm": 144.0396270751953, "learning_rate": 4.667228832880478e-05, "loss": 16.8593, "step": 11379 }, { "epoch": 0.4797942534308662, "grad_norm": 115.05393981933594, "learning_rate": 4.66715075423967e-05, "loss": 17.5936, "step": 11380 }, { "epoch": 0.47983641461306575, "grad_norm": 91.04911041259766, "learning_rate": 4.667072675598863e-05, "loss": 18.0488, "step": 11381 }, { "epoch": 0.47987857579526527, "grad_norm": 227.69332885742188, "learning_rate": 4.6669945969580564e-05, "loss": 16.4573, "step": 11382 }, { "epoch": 0.47992073697746485, "grad_norm": 128.48533630371094, "learning_rate": 4.666916518317249e-05, "loss": 18.2942, "step": 11383 }, { "epoch": 0.4799628981596644, "grad_norm": 167.33543395996094, "learning_rate": 4.666838439676443e-05, "loss": 17.016, "step": 11384 }, { "epoch": 0.48000505934186394, "grad_norm": 212.7919464111328, "learning_rate": 4.666760361035635e-05, "loss": 17.4567, "step": 11385 }, { "epoch": 0.4800472205240635, "grad_norm": 189.53662109375, "learning_rate": 4.666682282394828e-05, "loss": 15.8677, "step": 11386 }, { "epoch": 0.48008938170626303, "grad_norm": 156.82252502441406, "learning_rate": 4.6666042037540214e-05, "loss": 17.8873, "step": 11387 }, { "epoch": 0.4801315428884626, "grad_norm": 169.59609985351562, "learning_rate": 4.6665261251132145e-05, "loss": 16.9343, "step": 11388 }, { "epoch": 0.4801737040706621, "grad_norm": 192.3872833251953, "learning_rate": 4.666448046472407e-05, "loss": 16.3775, "step": 11389 }, { "epoch": 0.4802158652528617, "grad_norm": 6208.70361328125, "learning_rate": 4.6663699678316e-05, "loss": 17.832, "step": 11390 }, { "epoch": 0.4802580264350612, "grad_norm": 133.46368408203125, "learning_rate": 4.666291889190793e-05, "loss": 16.6384, "step": 11391 }, { "epoch": 0.4803001876172608, "grad_norm": 183.1693878173828, "learning_rate": 4.6662138105499863e-05, "loss": 17.2922, "step": 11392 }, { "epoch": 0.48034234879946036, "grad_norm": 156.56866455078125, "learning_rate": 4.6661357319091795e-05, "loss": 17.7496, "step": 11393 }, { "epoch": 0.4803845099816599, "grad_norm": 309.05267333984375, "learning_rate": 4.666057653268372e-05, "loss": 15.8097, "step": 11394 }, { "epoch": 0.48042667116385945, "grad_norm": 398.86065673828125, "learning_rate": 4.665979574627565e-05, "loss": 15.7913, "step": 11395 }, { "epoch": 0.48046883234605897, "grad_norm": 169.21405029296875, "learning_rate": 4.665901495986758e-05, "loss": 15.5053, "step": 11396 }, { "epoch": 0.48051099352825855, "grad_norm": 141.9420166015625, "learning_rate": 4.6658234173459506e-05, "loss": 18.1809, "step": 11397 }, { "epoch": 0.48055315471045806, "grad_norm": 256.3729553222656, "learning_rate": 4.6657453387051444e-05, "loss": 16.3608, "step": 11398 }, { "epoch": 0.48059531589265764, "grad_norm": 334.8798828125, "learning_rate": 4.665667260064337e-05, "loss": 15.0988, "step": 11399 }, { "epoch": 0.48063747707485716, "grad_norm": 334.6314697265625, "learning_rate": 4.66558918142353e-05, "loss": 17.1021, "step": 11400 }, { "epoch": 0.48067963825705673, "grad_norm": 352.6166687011719, "learning_rate": 4.665511102782723e-05, "loss": 15.9505, "step": 11401 }, { "epoch": 0.4807217994392563, "grad_norm": 135.58297729492188, "learning_rate": 4.6654330241419156e-05, "loss": 15.2682, "step": 11402 }, { "epoch": 0.4807639606214558, "grad_norm": 775.6846313476562, "learning_rate": 4.6653549455011094e-05, "loss": 14.0951, "step": 11403 }, { "epoch": 0.4808061218036554, "grad_norm": 206.30953979492188, "learning_rate": 4.665276866860302e-05, "loss": 14.4978, "step": 11404 }, { "epoch": 0.4808482829858549, "grad_norm": 803.7129516601562, "learning_rate": 4.665198788219495e-05, "loss": 16.1029, "step": 11405 }, { "epoch": 0.4808904441680545, "grad_norm": 3313.85205078125, "learning_rate": 4.665120709578688e-05, "loss": 14.6844, "step": 11406 }, { "epoch": 0.480932605350254, "grad_norm": 2200.5009765625, "learning_rate": 4.6650426309378805e-05, "loss": 14.9776, "step": 11407 }, { "epoch": 0.4809747665324536, "grad_norm": 3096.226806640625, "learning_rate": 4.6649645522970736e-05, "loss": 16.097, "step": 11408 }, { "epoch": 0.4810169277146531, "grad_norm": 113.40570831298828, "learning_rate": 4.664886473656267e-05, "loss": 15.2432, "step": 11409 }, { "epoch": 0.48105908889685267, "grad_norm": 2630.849609375, "learning_rate": 4.66480839501546e-05, "loss": 13.5869, "step": 11410 }, { "epoch": 0.4811012500790522, "grad_norm": 140.50723266601562, "learning_rate": 4.6647303163746523e-05, "loss": 13.1944, "step": 11411 }, { "epoch": 0.48114341126125176, "grad_norm": 22400.8671875, "learning_rate": 4.664652237733846e-05, "loss": 13.2823, "step": 11412 }, { "epoch": 0.48118557244345134, "grad_norm": 431.11517333984375, "learning_rate": 4.6645741590930386e-05, "loss": 13.5537, "step": 11413 }, { "epoch": 0.48122773362565086, "grad_norm": 367.5897521972656, "learning_rate": 4.664496080452232e-05, "loss": 13.4692, "step": 11414 }, { "epoch": 0.48126989480785043, "grad_norm": 4081.40478515625, "learning_rate": 4.664418001811425e-05, "loss": 17.6165, "step": 11415 }, { "epoch": 0.48131205599004995, "grad_norm": 3439.246337890625, "learning_rate": 4.664339923170617e-05, "loss": 16.1517, "step": 11416 }, { "epoch": 0.4813542171722495, "grad_norm": 307.4464111328125, "learning_rate": 4.664261844529811e-05, "loss": 15.6474, "step": 11417 }, { "epoch": 0.48139637835444904, "grad_norm": 2562.093505859375, "learning_rate": 4.6641837658890036e-05, "loss": 15.2766, "step": 11418 }, { "epoch": 0.4814385395366486, "grad_norm": 2598.38525390625, "learning_rate": 4.664105687248197e-05, "loss": 13.7794, "step": 11419 }, { "epoch": 0.48148070071884813, "grad_norm": 4417.37744140625, "learning_rate": 4.66402760860739e-05, "loss": 14.3261, "step": 11420 }, { "epoch": 0.4815228619010477, "grad_norm": 845.7589721679688, "learning_rate": 4.663949529966582e-05, "loss": 13.952, "step": 11421 }, { "epoch": 0.4815650230832473, "grad_norm": 2044.9156494140625, "learning_rate": 4.6638714513257754e-05, "loss": 13.9967, "step": 11422 }, { "epoch": 0.4816071842654468, "grad_norm": 1442.4122314453125, "learning_rate": 4.6637933726849685e-05, "loss": 13.8279, "step": 11423 }, { "epoch": 0.48164934544764637, "grad_norm": 1151.9049072265625, "learning_rate": 4.6637152940441616e-05, "loss": 13.246, "step": 11424 }, { "epoch": 0.4816915066298459, "grad_norm": 839.5711059570312, "learning_rate": 4.663637215403355e-05, "loss": 15.7596, "step": 11425 }, { "epoch": 0.48173366781204546, "grad_norm": 872.7534790039062, "learning_rate": 4.663559136762547e-05, "loss": 13.5033, "step": 11426 }, { "epoch": 0.481775828994245, "grad_norm": 436.8988037109375, "learning_rate": 4.66348105812174e-05, "loss": 14.2212, "step": 11427 }, { "epoch": 0.48181799017644455, "grad_norm": 775.60498046875, "learning_rate": 4.6634029794809335e-05, "loss": 14.0924, "step": 11428 }, { "epoch": 0.4818601513586441, "grad_norm": 1006.62548828125, "learning_rate": 4.6633249008401266e-05, "loss": 13.7186, "step": 11429 }, { "epoch": 0.48190231254084365, "grad_norm": 525.4096069335938, "learning_rate": 4.663246822199319e-05, "loss": 13.6202, "step": 11430 }, { "epoch": 0.4819444737230432, "grad_norm": 372.13934326171875, "learning_rate": 4.663168743558513e-05, "loss": 13.2282, "step": 11431 }, { "epoch": 0.48198663490524274, "grad_norm": 522.9902954101562, "learning_rate": 4.663090664917705e-05, "loss": 16.4789, "step": 11432 }, { "epoch": 0.4820287960874423, "grad_norm": 205.05206298828125, "learning_rate": 4.6630125862768984e-05, "loss": 15.4147, "step": 11433 }, { "epoch": 0.48207095726964183, "grad_norm": 370.6871643066406, "learning_rate": 4.6629345076360915e-05, "loss": 13.6599, "step": 11434 }, { "epoch": 0.4821131184518414, "grad_norm": 155.74009704589844, "learning_rate": 4.662856428995284e-05, "loss": 14.0935, "step": 11435 }, { "epoch": 0.4821552796340409, "grad_norm": 173.1523895263672, "learning_rate": 4.662778350354478e-05, "loss": 16.0142, "step": 11436 }, { "epoch": 0.4821974408162405, "grad_norm": 408.5744323730469, "learning_rate": 4.66270027171367e-05, "loss": 13.5448, "step": 11437 }, { "epoch": 0.48223960199844, "grad_norm": 510.431640625, "learning_rate": 4.6626221930728634e-05, "loss": 14.0198, "step": 11438 }, { "epoch": 0.4822817631806396, "grad_norm": 98.48478698730469, "learning_rate": 4.6625441144320565e-05, "loss": 16.7425, "step": 11439 }, { "epoch": 0.4823239243628391, "grad_norm": 114.57971954345703, "learning_rate": 4.662466035791249e-05, "loss": 15.4523, "step": 11440 }, { "epoch": 0.4823660855450387, "grad_norm": 148.11911010742188, "learning_rate": 4.662387957150442e-05, "loss": 13.4306, "step": 11441 }, { "epoch": 0.48240824672723825, "grad_norm": 206.7402801513672, "learning_rate": 4.662309878509635e-05, "loss": 17.2277, "step": 11442 }, { "epoch": 0.4824504079094378, "grad_norm": 150.1053924560547, "learning_rate": 4.662231799868828e-05, "loss": 15.4673, "step": 11443 }, { "epoch": 0.48249256909163735, "grad_norm": 114.66951751708984, "learning_rate": 4.662153721228021e-05, "loss": 18.2363, "step": 11444 }, { "epoch": 0.48253473027383686, "grad_norm": 117.25394439697266, "learning_rate": 4.662075642587214e-05, "loss": 15.4401, "step": 11445 }, { "epoch": 0.48257689145603644, "grad_norm": 128.6331787109375, "learning_rate": 4.661997563946407e-05, "loss": 15.8777, "step": 11446 }, { "epoch": 0.48261905263823596, "grad_norm": 116.289306640625, "learning_rate": 4.6619194853056e-05, "loss": 14.7634, "step": 11447 }, { "epoch": 0.48266121382043553, "grad_norm": 132.4709930419922, "learning_rate": 4.661841406664793e-05, "loss": 17.711, "step": 11448 }, { "epoch": 0.48270337500263505, "grad_norm": 122.66789245605469, "learning_rate": 4.661763328023986e-05, "loss": 17.5706, "step": 11449 }, { "epoch": 0.4827455361848346, "grad_norm": 107.39714050292969, "learning_rate": 4.6616852493831795e-05, "loss": 17.9873, "step": 11450 }, { "epoch": 0.4827876973670342, "grad_norm": 131.0175323486328, "learning_rate": 4.661607170742372e-05, "loss": 16.877, "step": 11451 }, { "epoch": 0.4828298585492337, "grad_norm": 151.4274444580078, "learning_rate": 4.6615290921015644e-05, "loss": 16.8052, "step": 11452 }, { "epoch": 0.4828720197314333, "grad_norm": 134.48373413085938, "learning_rate": 4.661451013460758e-05, "loss": 16.7308, "step": 11453 }, { "epoch": 0.4829141809136328, "grad_norm": 165.2495880126953, "learning_rate": 4.661372934819951e-05, "loss": 15.6022, "step": 11454 }, { "epoch": 0.4829563420958324, "grad_norm": 122.45941925048828, "learning_rate": 4.661294856179144e-05, "loss": 16.7674, "step": 11455 }, { "epoch": 0.4829985032780319, "grad_norm": 174.2340545654297, "learning_rate": 4.661216777538337e-05, "loss": 19.0725, "step": 11456 }, { "epoch": 0.4830406644602315, "grad_norm": 138.06930541992188, "learning_rate": 4.66113869889753e-05, "loss": 18.5697, "step": 11457 }, { "epoch": 0.483082825642431, "grad_norm": 131.04612731933594, "learning_rate": 4.6610606202567225e-05, "loss": 18.5875, "step": 11458 }, { "epoch": 0.48312498682463056, "grad_norm": 132.73764038085938, "learning_rate": 4.6609825416159156e-05, "loss": 18.2492, "step": 11459 }, { "epoch": 0.48316714800683014, "grad_norm": 147.83648681640625, "learning_rate": 4.660904462975109e-05, "loss": 18.6392, "step": 11460 }, { "epoch": 0.48320930918902966, "grad_norm": 140.41139221191406, "learning_rate": 4.660826384334302e-05, "loss": 18.719, "step": 11461 }, { "epoch": 0.48325147037122923, "grad_norm": 158.7552032470703, "learning_rate": 4.660748305693495e-05, "loss": 19.6538, "step": 11462 }, { "epoch": 0.48329363155342875, "grad_norm": 156.67991638183594, "learning_rate": 4.6606702270526874e-05, "loss": 17.807, "step": 11463 }, { "epoch": 0.4833357927356283, "grad_norm": 145.99652099609375, "learning_rate": 4.6605921484118806e-05, "loss": 19.874, "step": 11464 }, { "epoch": 0.48337795391782784, "grad_norm": 154.47341918945312, "learning_rate": 4.660514069771074e-05, "loss": 20.249, "step": 11465 }, { "epoch": 0.4834201151000274, "grad_norm": 128.16102600097656, "learning_rate": 4.660435991130266e-05, "loss": 20.9496, "step": 11466 }, { "epoch": 0.48346227628222693, "grad_norm": 147.70472717285156, "learning_rate": 4.66035791248946e-05, "loss": 19.759, "step": 11467 }, { "epoch": 0.4835044374644265, "grad_norm": 144.2873077392578, "learning_rate": 4.6602798338486524e-05, "loss": 19.4391, "step": 11468 }, { "epoch": 0.483546598646626, "grad_norm": 136.25039672851562, "learning_rate": 4.6602017552078455e-05, "loss": 19.9811, "step": 11469 }, { "epoch": 0.4835887598288256, "grad_norm": 142.18154907226562, "learning_rate": 4.6601236765670387e-05, "loss": 20.1421, "step": 11470 }, { "epoch": 0.48363092101102517, "grad_norm": 139.26881408691406, "learning_rate": 4.660045597926231e-05, "loss": 20.0303, "step": 11471 }, { "epoch": 0.4836730821932247, "grad_norm": 130.986572265625, "learning_rate": 4.659967519285425e-05, "loss": 21.0007, "step": 11472 }, { "epoch": 0.48371524337542426, "grad_norm": 125.75054931640625, "learning_rate": 4.6598894406446174e-05, "loss": 23.4341, "step": 11473 }, { "epoch": 0.4837574045576238, "grad_norm": 134.32147216796875, "learning_rate": 4.6598113620038105e-05, "loss": 20.3663, "step": 11474 }, { "epoch": 0.48379956573982336, "grad_norm": 114.921875, "learning_rate": 4.6597332833630036e-05, "loss": 21.5538, "step": 11475 }, { "epoch": 0.4838417269220229, "grad_norm": 131.0596160888672, "learning_rate": 4.659655204722197e-05, "loss": 23.3924, "step": 11476 }, { "epoch": 0.48388388810422245, "grad_norm": 164.5839080810547, "learning_rate": 4.659577126081389e-05, "loss": 19.7923, "step": 11477 }, { "epoch": 0.48392604928642197, "grad_norm": 125.76180267333984, "learning_rate": 4.659499047440582e-05, "loss": 21.6363, "step": 11478 }, { "epoch": 0.48396821046862154, "grad_norm": 116.47372436523438, "learning_rate": 4.6594209687997754e-05, "loss": 21.3836, "step": 11479 }, { "epoch": 0.4840103716508211, "grad_norm": 92.84940338134766, "learning_rate": 4.659342890158968e-05, "loss": 22.1802, "step": 11480 }, { "epoch": 0.48405253283302063, "grad_norm": 113.4803695678711, "learning_rate": 4.659264811518162e-05, "loss": 21.3534, "step": 11481 }, { "epoch": 0.4840946940152202, "grad_norm": 106.67500305175781, "learning_rate": 4.659186732877354e-05, "loss": 21.0352, "step": 11482 }, { "epoch": 0.4841368551974197, "grad_norm": 96.73863220214844, "learning_rate": 4.659108654236547e-05, "loss": 20.7305, "step": 11483 }, { "epoch": 0.4841790163796193, "grad_norm": 97.53919219970703, "learning_rate": 4.6590305755957404e-05, "loss": 20.7821, "step": 11484 }, { "epoch": 0.4842211775618188, "grad_norm": 109.54879760742188, "learning_rate": 4.658952496954933e-05, "loss": 21.7128, "step": 11485 }, { "epoch": 0.4842633387440184, "grad_norm": 108.97203826904297, "learning_rate": 4.6588744183141266e-05, "loss": 21.1492, "step": 11486 }, { "epoch": 0.4843054999262179, "grad_norm": 102.52863311767578, "learning_rate": 4.658796339673319e-05, "loss": 19.8507, "step": 11487 }, { "epoch": 0.4843476611084175, "grad_norm": 91.68067932128906, "learning_rate": 4.658718261032512e-05, "loss": 20.7368, "step": 11488 }, { "epoch": 0.48438982229061706, "grad_norm": 80.24995422363281, "learning_rate": 4.658640182391705e-05, "loss": 20.9922, "step": 11489 }, { "epoch": 0.4844319834728166, "grad_norm": 78.41148376464844, "learning_rate": 4.658562103750898e-05, "loss": 21.5907, "step": 11490 }, { "epoch": 0.48447414465501615, "grad_norm": 94.62510681152344, "learning_rate": 4.658484025110091e-05, "loss": 21.7683, "step": 11491 }, { "epoch": 0.48451630583721567, "grad_norm": 69.13249969482422, "learning_rate": 4.658405946469284e-05, "loss": 21.4779, "step": 11492 }, { "epoch": 0.48455846701941524, "grad_norm": 76.14418029785156, "learning_rate": 4.658327867828477e-05, "loss": 21.1598, "step": 11493 }, { "epoch": 0.48460062820161476, "grad_norm": 91.21678924560547, "learning_rate": 4.65824978918767e-05, "loss": 20.5038, "step": 11494 }, { "epoch": 0.48464278938381433, "grad_norm": 73.0735855102539, "learning_rate": 4.6581717105468634e-05, "loss": 22.8035, "step": 11495 }, { "epoch": 0.48468495056601385, "grad_norm": 81.96640014648438, "learning_rate": 4.658093631906056e-05, "loss": 20.8617, "step": 11496 }, { "epoch": 0.4847271117482134, "grad_norm": 71.64869689941406, "learning_rate": 4.658015553265249e-05, "loss": 22.5608, "step": 11497 }, { "epoch": 0.48476927293041294, "grad_norm": 101.56840515136719, "learning_rate": 4.657937474624442e-05, "loss": 19.7416, "step": 11498 }, { "epoch": 0.4848114341126125, "grad_norm": 95.08061218261719, "learning_rate": 4.6578593959836346e-05, "loss": 20.8956, "step": 11499 }, { "epoch": 0.4848535952948121, "grad_norm": 64.3941879272461, "learning_rate": 4.6577813173428284e-05, "loss": 21.1102, "step": 11500 }, { "epoch": 0.4848957564770116, "grad_norm": 67.90065002441406, "learning_rate": 4.657703238702021e-05, "loss": 20.044, "step": 11501 }, { "epoch": 0.4849379176592112, "grad_norm": 74.97453308105469, "learning_rate": 4.657625160061214e-05, "loss": 20.3212, "step": 11502 }, { "epoch": 0.4849800788414107, "grad_norm": 87.21442413330078, "learning_rate": 4.657547081420407e-05, "loss": 20.3824, "step": 11503 }, { "epoch": 0.4850222400236103, "grad_norm": 148.96536254882812, "learning_rate": 4.6574690027795995e-05, "loss": 19.5357, "step": 11504 }, { "epoch": 0.4850644012058098, "grad_norm": 70.64389038085938, "learning_rate": 4.657390924138793e-05, "loss": 20.5791, "step": 11505 }, { "epoch": 0.48510656238800937, "grad_norm": 65.28844451904297, "learning_rate": 4.657312845497986e-05, "loss": 19.5744, "step": 11506 }, { "epoch": 0.4851487235702089, "grad_norm": 62.21462631225586, "learning_rate": 4.657234766857179e-05, "loss": 19.301, "step": 11507 }, { "epoch": 0.48519088475240846, "grad_norm": 71.94905853271484, "learning_rate": 4.657156688216372e-05, "loss": 20.5399, "step": 11508 }, { "epoch": 0.48523304593460803, "grad_norm": 72.90213775634766, "learning_rate": 4.6570786095755645e-05, "loss": 20.5179, "step": 11509 }, { "epoch": 0.48527520711680755, "grad_norm": 89.30763244628906, "learning_rate": 4.6570005309347576e-05, "loss": 20.5066, "step": 11510 }, { "epoch": 0.4853173682990071, "grad_norm": 59.540096282958984, "learning_rate": 4.656922452293951e-05, "loss": 19.7993, "step": 11511 }, { "epoch": 0.48535952948120664, "grad_norm": 82.76805114746094, "learning_rate": 4.656844373653144e-05, "loss": 21.2422, "step": 11512 }, { "epoch": 0.4854016906634062, "grad_norm": 77.57274627685547, "learning_rate": 4.656766295012336e-05, "loss": 18.0053, "step": 11513 }, { "epoch": 0.48544385184560573, "grad_norm": 48.60126876831055, "learning_rate": 4.6566882163715294e-05, "loss": 18.5605, "step": 11514 }, { "epoch": 0.4854860130278053, "grad_norm": 65.8672103881836, "learning_rate": 4.6566101377307225e-05, "loss": 17.7122, "step": 11515 }, { "epoch": 0.4855281742100048, "grad_norm": 59.01713180541992, "learning_rate": 4.656532059089916e-05, "loss": 17.9959, "step": 11516 }, { "epoch": 0.4855703353922044, "grad_norm": 66.7761459350586, "learning_rate": 4.656453980449109e-05, "loss": 19.8087, "step": 11517 }, { "epoch": 0.485612496574404, "grad_norm": 71.81974792480469, "learning_rate": 4.656375901808301e-05, "loss": 16.6854, "step": 11518 }, { "epoch": 0.4856546577566035, "grad_norm": 56.27787780761719, "learning_rate": 4.656297823167495e-05, "loss": 17.8484, "step": 11519 }, { "epoch": 0.48569681893880307, "grad_norm": 76.94019317626953, "learning_rate": 4.6562197445266875e-05, "loss": 17.6305, "step": 11520 }, { "epoch": 0.4857389801210026, "grad_norm": 46.50360107421875, "learning_rate": 4.6561416658858806e-05, "loss": 16.383, "step": 11521 }, { "epoch": 0.48578114130320216, "grad_norm": 39.852657318115234, "learning_rate": 4.656063587245074e-05, "loss": 16.9572, "step": 11522 }, { "epoch": 0.4858233024854017, "grad_norm": 60.82897186279297, "learning_rate": 4.655985508604266e-05, "loss": 18.8304, "step": 11523 }, { "epoch": 0.48586546366760125, "grad_norm": 146.71241760253906, "learning_rate": 4.655907429963459e-05, "loss": 16.0412, "step": 11524 }, { "epoch": 0.48590762484980077, "grad_norm": 68.20579528808594, "learning_rate": 4.6558293513226524e-05, "loss": 16.9191, "step": 11525 }, { "epoch": 0.48594978603200034, "grad_norm": 74.75284576416016, "learning_rate": 4.6557512726818456e-05, "loss": 16.3472, "step": 11526 }, { "epoch": 0.48599194721419986, "grad_norm": 59.72254943847656, "learning_rate": 4.655673194041039e-05, "loss": 16.0897, "step": 11527 }, { "epoch": 0.48603410839639943, "grad_norm": 60.217288970947266, "learning_rate": 4.655595115400231e-05, "loss": 15.7632, "step": 11528 }, { "epoch": 0.486076269578599, "grad_norm": 58.7186164855957, "learning_rate": 4.655517036759424e-05, "loss": 16.065, "step": 11529 }, { "epoch": 0.4861184307607985, "grad_norm": 67.95716094970703, "learning_rate": 4.6554389581186174e-05, "loss": 15.8966, "step": 11530 }, { "epoch": 0.4861605919429981, "grad_norm": 105.35977935791016, "learning_rate": 4.6553608794778105e-05, "loss": 17.5597, "step": 11531 }, { "epoch": 0.4862027531251976, "grad_norm": 69.880859375, "learning_rate": 4.655282800837003e-05, "loss": 14.4109, "step": 11532 }, { "epoch": 0.4862449143073972, "grad_norm": 74.0153579711914, "learning_rate": 4.655204722196196e-05, "loss": 15.8117, "step": 11533 }, { "epoch": 0.4862870754895967, "grad_norm": 52.44272232055664, "learning_rate": 4.655126643555389e-05, "loss": 15.8302, "step": 11534 }, { "epoch": 0.4863292366717963, "grad_norm": 58.33414840698242, "learning_rate": 4.655048564914582e-05, "loss": 16.4909, "step": 11535 }, { "epoch": 0.4863713978539958, "grad_norm": 78.6836929321289, "learning_rate": 4.6549704862737755e-05, "loss": 14.6595, "step": 11536 }, { "epoch": 0.4864135590361954, "grad_norm": 98.04617309570312, "learning_rate": 4.654892407632968e-05, "loss": 17.4214, "step": 11537 }, { "epoch": 0.48645572021839495, "grad_norm": 44.97675323486328, "learning_rate": 4.654814328992161e-05, "loss": 16.0593, "step": 11538 }, { "epoch": 0.48649788140059447, "grad_norm": 48.00479507446289, "learning_rate": 4.654736250351354e-05, "loss": 16.0293, "step": 11539 }, { "epoch": 0.48654004258279404, "grad_norm": 50.47493362426758, "learning_rate": 4.654658171710547e-05, "loss": 15.6933, "step": 11540 }, { "epoch": 0.48658220376499356, "grad_norm": 46.88778305053711, "learning_rate": 4.6545800930697404e-05, "loss": 16.8719, "step": 11541 }, { "epoch": 0.48662436494719313, "grad_norm": 45.21567916870117, "learning_rate": 4.654502014428933e-05, "loss": 16.6121, "step": 11542 }, { "epoch": 0.48666652612939265, "grad_norm": 46.466346740722656, "learning_rate": 4.654423935788126e-05, "loss": 17.3256, "step": 11543 }, { "epoch": 0.4867086873115922, "grad_norm": 71.78176879882812, "learning_rate": 4.654345857147319e-05, "loss": 16.6042, "step": 11544 }, { "epoch": 0.48675084849379174, "grad_norm": 43.45033264160156, "learning_rate": 4.654267778506512e-05, "loss": 18.0179, "step": 11545 }, { "epoch": 0.4867930096759913, "grad_norm": 49.86886978149414, "learning_rate": 4.654189699865705e-05, "loss": 15.2934, "step": 11546 }, { "epoch": 0.4868351708581909, "grad_norm": 46.0860595703125, "learning_rate": 4.654111621224898e-05, "loss": 16.3486, "step": 11547 }, { "epoch": 0.4868773320403904, "grad_norm": 49.1130485534668, "learning_rate": 4.654033542584091e-05, "loss": 14.4448, "step": 11548 }, { "epoch": 0.48691949322259, "grad_norm": 41.50354766845703, "learning_rate": 4.6539554639432834e-05, "loss": 16.3599, "step": 11549 }, { "epoch": 0.4869616544047895, "grad_norm": 48.608543395996094, "learning_rate": 4.653877385302477e-05, "loss": 15.0254, "step": 11550 }, { "epoch": 0.4870038155869891, "grad_norm": 58.85116195678711, "learning_rate": 4.6537993066616697e-05, "loss": 15.5825, "step": 11551 }, { "epoch": 0.4870459767691886, "grad_norm": 72.71319580078125, "learning_rate": 4.653721228020863e-05, "loss": 15.0634, "step": 11552 }, { "epoch": 0.48708813795138817, "grad_norm": 57.72809600830078, "learning_rate": 4.653643149380056e-05, "loss": 14.8961, "step": 11553 }, { "epoch": 0.4871302991335877, "grad_norm": 63.96424865722656, "learning_rate": 4.6535650707392484e-05, "loss": 17.6506, "step": 11554 }, { "epoch": 0.48717246031578726, "grad_norm": 71.14749908447266, "learning_rate": 4.653486992098442e-05, "loss": 14.5787, "step": 11555 }, { "epoch": 0.4872146214979868, "grad_norm": 70.08602142333984, "learning_rate": 4.6534089134576346e-05, "loss": 16.017, "step": 11556 }, { "epoch": 0.48725678268018635, "grad_norm": 47.090614318847656, "learning_rate": 4.653330834816828e-05, "loss": 16.2895, "step": 11557 }, { "epoch": 0.4872989438623859, "grad_norm": 56.70907974243164, "learning_rate": 4.653252756176021e-05, "loss": 15.4027, "step": 11558 }, { "epoch": 0.48734110504458544, "grad_norm": 58.549903869628906, "learning_rate": 4.653174677535213e-05, "loss": 15.4806, "step": 11559 }, { "epoch": 0.487383266226785, "grad_norm": 57.728694915771484, "learning_rate": 4.6530965988944064e-05, "loss": 18.6986, "step": 11560 }, { "epoch": 0.48742542740898454, "grad_norm": 114.923095703125, "learning_rate": 4.6530185202535996e-05, "loss": 18.176, "step": 11561 }, { "epoch": 0.4874675885911841, "grad_norm": 315.7654724121094, "learning_rate": 4.652940441612793e-05, "loss": 16.3213, "step": 11562 }, { "epoch": 0.4875097497733836, "grad_norm": 78.53807067871094, "learning_rate": 4.652862362971986e-05, "loss": 16.8543, "step": 11563 }, { "epoch": 0.4875519109555832, "grad_norm": 167.45156860351562, "learning_rate": 4.652784284331179e-05, "loss": 16.2159, "step": 11564 }, { "epoch": 0.4875940721377827, "grad_norm": 201.8092041015625, "learning_rate": 4.6527062056903714e-05, "loss": 16.5265, "step": 11565 }, { "epoch": 0.4876362333199823, "grad_norm": 114.40388488769531, "learning_rate": 4.6526281270495645e-05, "loss": 16.2476, "step": 11566 }, { "epoch": 0.48767839450218187, "grad_norm": 482.20782470703125, "learning_rate": 4.6525500484087576e-05, "loss": 16.5295, "step": 11567 }, { "epoch": 0.4877205556843814, "grad_norm": 629.2255249023438, "learning_rate": 4.65247196976795e-05, "loss": 16.1361, "step": 11568 }, { "epoch": 0.48776271686658096, "grad_norm": 1809.75732421875, "learning_rate": 4.652393891127144e-05, "loss": 16.774, "step": 11569 }, { "epoch": 0.4878048780487805, "grad_norm": 2796.029052734375, "learning_rate": 4.6523158124863363e-05, "loss": 15.5779, "step": 11570 }, { "epoch": 0.48784703923098005, "grad_norm": 1226.10009765625, "learning_rate": 4.6522377338455295e-05, "loss": 15.5467, "step": 11571 }, { "epoch": 0.48788920041317957, "grad_norm": 4961.03271484375, "learning_rate": 4.6521596552047226e-05, "loss": 15.443, "step": 11572 }, { "epoch": 0.48793136159537914, "grad_norm": 3893.264404296875, "learning_rate": 4.652081576563915e-05, "loss": 17.5137, "step": 11573 }, { "epoch": 0.48797352277757866, "grad_norm": 24299.365234375, "learning_rate": 4.652003497923109e-05, "loss": 15.9716, "step": 11574 }, { "epoch": 0.48801568395977823, "grad_norm": 874.4743041992188, "learning_rate": 4.651925419282301e-05, "loss": 17.1106, "step": 11575 }, { "epoch": 0.4880578451419778, "grad_norm": 361.0589904785156, "learning_rate": 4.6518473406414944e-05, "loss": 18.0432, "step": 11576 }, { "epoch": 0.4881000063241773, "grad_norm": 2002.2938232421875, "learning_rate": 4.6517692620006875e-05, "loss": 15.687, "step": 11577 }, { "epoch": 0.4881421675063769, "grad_norm": 584.703369140625, "learning_rate": 4.65169118335988e-05, "loss": 16.4032, "step": 11578 }, { "epoch": 0.4881843286885764, "grad_norm": 130.056640625, "learning_rate": 4.651613104719073e-05, "loss": 15.1608, "step": 11579 }, { "epoch": 0.488226489870776, "grad_norm": 114.81829071044922, "learning_rate": 4.651535026078266e-05, "loss": 17.0051, "step": 11580 }, { "epoch": 0.4882686510529755, "grad_norm": 720.8120727539062, "learning_rate": 4.6514569474374594e-05, "loss": 17.169, "step": 11581 }, { "epoch": 0.4883108122351751, "grad_norm": 105.91358184814453, "learning_rate": 4.651378868796652e-05, "loss": 17.6099, "step": 11582 }, { "epoch": 0.4883529734173746, "grad_norm": 140.95693969726562, "learning_rate": 4.6513007901558456e-05, "loss": 16.2009, "step": 11583 }, { "epoch": 0.4883951345995742, "grad_norm": 220.93663024902344, "learning_rate": 4.651222711515038e-05, "loss": 16.8433, "step": 11584 }, { "epoch": 0.4884372957817737, "grad_norm": 114.59618377685547, "learning_rate": 4.651144632874231e-05, "loss": 18.9549, "step": 11585 }, { "epoch": 0.48847945696397327, "grad_norm": 116.63944244384766, "learning_rate": 4.651066554233424e-05, "loss": 17.7897, "step": 11586 }, { "epoch": 0.48852161814617284, "grad_norm": 99.01984405517578, "learning_rate": 4.650988475592617e-05, "loss": 19.3447, "step": 11587 }, { "epoch": 0.48856377932837236, "grad_norm": 125.92293548583984, "learning_rate": 4.6509103969518106e-05, "loss": 16.6887, "step": 11588 }, { "epoch": 0.48860594051057193, "grad_norm": 156.318603515625, "learning_rate": 4.650832318311003e-05, "loss": 17.6319, "step": 11589 }, { "epoch": 0.48864810169277145, "grad_norm": 115.81696319580078, "learning_rate": 4.650754239670196e-05, "loss": 17.0376, "step": 11590 }, { "epoch": 0.488690262874971, "grad_norm": 87.05288696289062, "learning_rate": 4.650676161029389e-05, "loss": 17.5728, "step": 11591 }, { "epoch": 0.48873242405717054, "grad_norm": 90.96098327636719, "learning_rate": 4.650598082388582e-05, "loss": 17.3146, "step": 11592 }, { "epoch": 0.4887745852393701, "grad_norm": 102.29825592041016, "learning_rate": 4.650520003747775e-05, "loss": 17.1015, "step": 11593 }, { "epoch": 0.48881674642156964, "grad_norm": 96.69892120361328, "learning_rate": 4.650441925106968e-05, "loss": 18.561, "step": 11594 }, { "epoch": 0.4888589076037692, "grad_norm": 92.42157745361328, "learning_rate": 4.650363846466161e-05, "loss": 17.3399, "step": 11595 }, { "epoch": 0.4889010687859688, "grad_norm": 69.73120880126953, "learning_rate": 4.650285767825354e-05, "loss": 18.7037, "step": 11596 }, { "epoch": 0.4889432299681683, "grad_norm": 186.13829040527344, "learning_rate": 4.650207689184547e-05, "loss": 19.0824, "step": 11597 }, { "epoch": 0.4889853911503679, "grad_norm": 75.61886596679688, "learning_rate": 4.65012961054374e-05, "loss": 22.4087, "step": 11598 }, { "epoch": 0.4890275523325674, "grad_norm": 36.83750534057617, "learning_rate": 4.650051531902933e-05, "loss": 17.8425, "step": 11599 }, { "epoch": 0.48906971351476697, "grad_norm": 128.70223999023438, "learning_rate": 4.649973453262126e-05, "loss": 17.8372, "step": 11600 }, { "epoch": 0.4891118746969665, "grad_norm": 38.78228759765625, "learning_rate": 4.6498953746213185e-05, "loss": 18.504, "step": 11601 }, { "epoch": 0.48915403587916606, "grad_norm": 36.980628967285156, "learning_rate": 4.6498172959805116e-05, "loss": 16.7803, "step": 11602 }, { "epoch": 0.4891961970613656, "grad_norm": 50.820594787597656, "learning_rate": 4.649739217339705e-05, "loss": 18.243, "step": 11603 }, { "epoch": 0.48923835824356515, "grad_norm": 43.307804107666016, "learning_rate": 4.649661138698897e-05, "loss": 16.293, "step": 11604 }, { "epoch": 0.4892805194257647, "grad_norm": 50.877655029296875, "learning_rate": 4.649583060058091e-05, "loss": 17.3733, "step": 11605 }, { "epoch": 0.48932268060796424, "grad_norm": 43.30990982055664, "learning_rate": 4.6495049814172835e-05, "loss": 17.6236, "step": 11606 }, { "epoch": 0.4893648417901638, "grad_norm": 39.5788459777832, "learning_rate": 4.649426902776477e-05, "loss": 17.0305, "step": 11607 }, { "epoch": 0.48940700297236334, "grad_norm": 29.864892959594727, "learning_rate": 4.64934882413567e-05, "loss": 16.5545, "step": 11608 }, { "epoch": 0.4894491641545629, "grad_norm": 74.22017669677734, "learning_rate": 4.649270745494863e-05, "loss": 15.377, "step": 11609 }, { "epoch": 0.48949132533676243, "grad_norm": 39.574073791503906, "learning_rate": 4.649192666854056e-05, "loss": 15.7688, "step": 11610 }, { "epoch": 0.489533486518962, "grad_norm": 34.20168685913086, "learning_rate": 4.6491145882132484e-05, "loss": 15.4764, "step": 11611 }, { "epoch": 0.4895756477011615, "grad_norm": 26.9871883392334, "learning_rate": 4.6490365095724415e-05, "loss": 14.0764, "step": 11612 }, { "epoch": 0.4896178088833611, "grad_norm": 50.61260986328125, "learning_rate": 4.6489584309316347e-05, "loss": 13.5565, "step": 11613 }, { "epoch": 0.4896599700655606, "grad_norm": 34.898101806640625, "learning_rate": 4.648880352290828e-05, "loss": 15.8536, "step": 11614 }, { "epoch": 0.4897021312477602, "grad_norm": 32.06472396850586, "learning_rate": 4.64880227365002e-05, "loss": 15.8173, "step": 11615 }, { "epoch": 0.48974429242995976, "grad_norm": 77.42221069335938, "learning_rate": 4.6487241950092134e-05, "loss": 13.7345, "step": 11616 }, { "epoch": 0.4897864536121593, "grad_norm": 41.072383880615234, "learning_rate": 4.6486461163684065e-05, "loss": 13.5657, "step": 11617 }, { "epoch": 0.48982861479435885, "grad_norm": 89.2135238647461, "learning_rate": 4.648568037727599e-05, "loss": 13.7481, "step": 11618 }, { "epoch": 0.48987077597655837, "grad_norm": 50.30002975463867, "learning_rate": 4.648489959086793e-05, "loss": 13.9503, "step": 11619 }, { "epoch": 0.48991293715875794, "grad_norm": 38.475223541259766, "learning_rate": 4.648411880445985e-05, "loss": 13.9759, "step": 11620 }, { "epoch": 0.48995509834095746, "grad_norm": 37.71576690673828, "learning_rate": 4.648333801805178e-05, "loss": 12.8769, "step": 11621 }, { "epoch": 0.48999725952315704, "grad_norm": 30.875505447387695, "learning_rate": 4.6482557231643714e-05, "loss": 13.3117, "step": 11622 }, { "epoch": 0.49003942070535655, "grad_norm": 33.00849533081055, "learning_rate": 4.648177644523564e-05, "loss": 14.0561, "step": 11623 }, { "epoch": 0.49008158188755613, "grad_norm": 32.0596809387207, "learning_rate": 4.648099565882758e-05, "loss": 13.0343, "step": 11624 }, { "epoch": 0.4901237430697557, "grad_norm": 32.832523345947266, "learning_rate": 4.64802148724195e-05, "loss": 12.2105, "step": 11625 }, { "epoch": 0.4901659042519552, "grad_norm": 27.33806037902832, "learning_rate": 4.647943408601143e-05, "loss": 12.9338, "step": 11626 }, { "epoch": 0.4902080654341548, "grad_norm": 46.47328567504883, "learning_rate": 4.6478653299603364e-05, "loss": 12.9689, "step": 11627 }, { "epoch": 0.4902502266163543, "grad_norm": 89.3916015625, "learning_rate": 4.6477872513195295e-05, "loss": 12.9406, "step": 11628 }, { "epoch": 0.4902923877985539, "grad_norm": 111.11249542236328, "learning_rate": 4.647709172678722e-05, "loss": 13.0347, "step": 11629 }, { "epoch": 0.4903345489807534, "grad_norm": 177.9041290283203, "learning_rate": 4.647631094037915e-05, "loss": 15.453, "step": 11630 }, { "epoch": 0.490376710162953, "grad_norm": 140.34571838378906, "learning_rate": 4.647553015397108e-05, "loss": 14.3905, "step": 11631 }, { "epoch": 0.4904188713451525, "grad_norm": 177.89402770996094, "learning_rate": 4.6474749367563013e-05, "loss": 12.7238, "step": 11632 }, { "epoch": 0.49046103252735207, "grad_norm": 530.5958862304688, "learning_rate": 4.6473968581154945e-05, "loss": 13.8911, "step": 11633 }, { "epoch": 0.4905031937095516, "grad_norm": 387.1665954589844, "learning_rate": 4.647318779474687e-05, "loss": 12.4106, "step": 11634 }, { "epoch": 0.49054535489175116, "grad_norm": 40328.484375, "learning_rate": 4.64724070083388e-05, "loss": 13.242, "step": 11635 }, { "epoch": 0.49058751607395074, "grad_norm": 10069.4990234375, "learning_rate": 4.647162622193073e-05, "loss": 13.3537, "step": 11636 }, { "epoch": 0.49062967725615025, "grad_norm": 3981.75830078125, "learning_rate": 4.6470845435522656e-05, "loss": 12.446, "step": 11637 }, { "epoch": 0.4906718384383498, "grad_norm": 3822.79833984375, "learning_rate": 4.6470064649114594e-05, "loss": 13.7972, "step": 11638 }, { "epoch": 0.49071399962054935, "grad_norm": 1917.267822265625, "learning_rate": 4.646928386270652e-05, "loss": 13.5728, "step": 11639 }, { "epoch": 0.4907561608027489, "grad_norm": 2590.20458984375, "learning_rate": 4.646850307629845e-05, "loss": 13.9425, "step": 11640 }, { "epoch": 0.49079832198494844, "grad_norm": 11023.435546875, "learning_rate": 4.646772228989038e-05, "loss": 12.872, "step": 11641 }, { "epoch": 0.490840483167148, "grad_norm": 12372.52734375, "learning_rate": 4.6466941503482306e-05, "loss": 12.6509, "step": 11642 }, { "epoch": 0.49088264434934753, "grad_norm": 79.6281509399414, "learning_rate": 4.6466160717074244e-05, "loss": 13.7832, "step": 11643 }, { "epoch": 0.4909248055315471, "grad_norm": 3370.673828125, "learning_rate": 4.646537993066617e-05, "loss": 13.8265, "step": 11644 }, { "epoch": 0.4909669667137467, "grad_norm": 633.7952270507812, "learning_rate": 4.64645991442581e-05, "loss": 12.3018, "step": 11645 }, { "epoch": 0.4910091278959462, "grad_norm": 316.3112487792969, "learning_rate": 4.646381835785003e-05, "loss": 13.1194, "step": 11646 }, { "epoch": 0.49105128907814577, "grad_norm": 229.2689971923828, "learning_rate": 4.6463037571441955e-05, "loss": 14.0881, "step": 11647 }, { "epoch": 0.4910934502603453, "grad_norm": 97.95829010009766, "learning_rate": 4.6462256785033886e-05, "loss": 13.031, "step": 11648 }, { "epoch": 0.49113561144254486, "grad_norm": 134.8800811767578, "learning_rate": 4.646147599862582e-05, "loss": 13.1774, "step": 11649 }, { "epoch": 0.4911777726247444, "grad_norm": 122.37042999267578, "learning_rate": 4.646069521221775e-05, "loss": 14.1251, "step": 11650 }, { "epoch": 0.49121993380694395, "grad_norm": 87.90876770019531, "learning_rate": 4.6459914425809673e-05, "loss": 13.1624, "step": 11651 }, { "epoch": 0.49126209498914347, "grad_norm": 2177.668212890625, "learning_rate": 4.645913363940161e-05, "loss": 14.3119, "step": 11652 }, { "epoch": 0.49130425617134305, "grad_norm": 116.15025329589844, "learning_rate": 4.6458352852993536e-05, "loss": 13.4663, "step": 11653 }, { "epoch": 0.4913464173535426, "grad_norm": 120.79115295410156, "learning_rate": 4.645757206658547e-05, "loss": 14.3778, "step": 11654 }, { "epoch": 0.49138857853574214, "grad_norm": 1034.9129638671875, "learning_rate": 4.64567912801774e-05, "loss": 14.5041, "step": 11655 }, { "epoch": 0.4914307397179417, "grad_norm": 96.8763198852539, "learning_rate": 4.645601049376932e-05, "loss": 13.3189, "step": 11656 }, { "epoch": 0.49147290090014123, "grad_norm": 105.66354370117188, "learning_rate": 4.645522970736126e-05, "loss": 14.7197, "step": 11657 }, { "epoch": 0.4915150620823408, "grad_norm": 4444.56640625, "learning_rate": 4.6454448920953186e-05, "loss": 15.9167, "step": 11658 }, { "epoch": 0.4915572232645403, "grad_norm": 72.27545928955078, "learning_rate": 4.645366813454512e-05, "loss": 15.0297, "step": 11659 }, { "epoch": 0.4915993844467399, "grad_norm": 76.88407897949219, "learning_rate": 4.645288734813705e-05, "loss": 14.6817, "step": 11660 }, { "epoch": 0.4916415456289394, "grad_norm": 58.85978698730469, "learning_rate": 4.645210656172897e-05, "loss": 14.2466, "step": 11661 }, { "epoch": 0.491683706811139, "grad_norm": 250.99188232421875, "learning_rate": 4.6451325775320904e-05, "loss": 14.031, "step": 11662 }, { "epoch": 0.4917258679933385, "grad_norm": 134.32057189941406, "learning_rate": 4.6450544988912835e-05, "loss": 16.276, "step": 11663 }, { "epoch": 0.4917680291755381, "grad_norm": 70.76810455322266, "learning_rate": 4.6449764202504766e-05, "loss": 15.2331, "step": 11664 }, { "epoch": 0.49181019035773765, "grad_norm": 68.56050109863281, "learning_rate": 4.64489834160967e-05, "loss": 13.5419, "step": 11665 }, { "epoch": 0.49185235153993717, "grad_norm": 37.66680145263672, "learning_rate": 4.644820262968862e-05, "loss": 15.5909, "step": 11666 }, { "epoch": 0.49189451272213675, "grad_norm": 44.33855056762695, "learning_rate": 4.644742184328055e-05, "loss": 15.3686, "step": 11667 }, { "epoch": 0.49193667390433626, "grad_norm": 51.52583312988281, "learning_rate": 4.6446641056872485e-05, "loss": 15.4389, "step": 11668 }, { "epoch": 0.49197883508653584, "grad_norm": 33.23393630981445, "learning_rate": 4.6445860270464416e-05, "loss": 14.6367, "step": 11669 }, { "epoch": 0.49202099626873536, "grad_norm": 42.684288024902344, "learning_rate": 4.644507948405634e-05, "loss": 15.8128, "step": 11670 }, { "epoch": 0.49206315745093493, "grad_norm": 37.36015701293945, "learning_rate": 4.644429869764828e-05, "loss": 16.4942, "step": 11671 }, { "epoch": 0.49210531863313445, "grad_norm": 44.38057327270508, "learning_rate": 4.64435179112402e-05, "loss": 15.7783, "step": 11672 }, { "epoch": 0.492147479815334, "grad_norm": 51.42408752441406, "learning_rate": 4.6442737124832134e-05, "loss": 16.0717, "step": 11673 }, { "epoch": 0.4921896409975336, "grad_norm": 36.5336799621582, "learning_rate": 4.6441956338424065e-05, "loss": 14.7189, "step": 11674 }, { "epoch": 0.4922318021797331, "grad_norm": 32.07858657836914, "learning_rate": 4.644117555201599e-05, "loss": 15.8112, "step": 11675 }, { "epoch": 0.4922739633619327, "grad_norm": 180.74948120117188, "learning_rate": 4.644039476560793e-05, "loss": 15.1191, "step": 11676 }, { "epoch": 0.4923161245441322, "grad_norm": 32.49269104003906, "learning_rate": 4.643961397919985e-05, "loss": 15.6814, "step": 11677 }, { "epoch": 0.4923582857263318, "grad_norm": 138.49432373046875, "learning_rate": 4.6438833192791784e-05, "loss": 14.1834, "step": 11678 }, { "epoch": 0.4924004469085313, "grad_norm": 27.55551528930664, "learning_rate": 4.6438052406383715e-05, "loss": 13.9408, "step": 11679 }, { "epoch": 0.49244260809073087, "grad_norm": 33.18275451660156, "learning_rate": 4.643727161997564e-05, "loss": 14.1139, "step": 11680 }, { "epoch": 0.4924847692729304, "grad_norm": 32.29576873779297, "learning_rate": 4.643649083356757e-05, "loss": 14.2926, "step": 11681 }, { "epoch": 0.49252693045512996, "grad_norm": 30.096120834350586, "learning_rate": 4.64357100471595e-05, "loss": 12.7037, "step": 11682 }, { "epoch": 0.49256909163732954, "grad_norm": 27.94110107421875, "learning_rate": 4.643492926075143e-05, "loss": 13.4735, "step": 11683 }, { "epoch": 0.49261125281952906, "grad_norm": 21.533037185668945, "learning_rate": 4.643414847434336e-05, "loss": 14.2034, "step": 11684 }, { "epoch": 0.49265341400172863, "grad_norm": 22.502063751220703, "learning_rate": 4.643336768793529e-05, "loss": 13.9591, "step": 11685 }, { "epoch": 0.49269557518392815, "grad_norm": 20.771751403808594, "learning_rate": 4.643258690152722e-05, "loss": 13.1919, "step": 11686 }, { "epoch": 0.4927377363661277, "grad_norm": 37.044891357421875, "learning_rate": 4.6431806115119145e-05, "loss": 15.6581, "step": 11687 }, { "epoch": 0.49277989754832724, "grad_norm": 20.81553077697754, "learning_rate": 4.643102532871108e-05, "loss": 13.317, "step": 11688 }, { "epoch": 0.4928220587305268, "grad_norm": 21.38714027404785, "learning_rate": 4.643024454230301e-05, "loss": 13.3463, "step": 11689 }, { "epoch": 0.49286421991272633, "grad_norm": 20.39994239807129, "learning_rate": 4.6429463755894945e-05, "loss": 12.1423, "step": 11690 }, { "epoch": 0.4929063810949259, "grad_norm": 35.592376708984375, "learning_rate": 4.642868296948687e-05, "loss": 13.1104, "step": 11691 }, { "epoch": 0.4929485422771254, "grad_norm": 25.706317901611328, "learning_rate": 4.6427902183078794e-05, "loss": 13.191, "step": 11692 }, { "epoch": 0.492990703459325, "grad_norm": 57.56497573852539, "learning_rate": 4.642712139667073e-05, "loss": 14.0472, "step": 11693 }, { "epoch": 0.49303286464152457, "grad_norm": 22.6177978515625, "learning_rate": 4.642634061026266e-05, "loss": 11.4515, "step": 11694 }, { "epoch": 0.4930750258237241, "grad_norm": 31.432004928588867, "learning_rate": 4.642555982385459e-05, "loss": 11.001, "step": 11695 }, { "epoch": 0.49311718700592366, "grad_norm": 26.44585418701172, "learning_rate": 4.642477903744652e-05, "loss": 11.9282, "step": 11696 }, { "epoch": 0.4931593481881232, "grad_norm": 24.58942413330078, "learning_rate": 4.642399825103845e-05, "loss": 11.5158, "step": 11697 }, { "epoch": 0.49320150937032275, "grad_norm": 18.39488410949707, "learning_rate": 4.6423217464630375e-05, "loss": 11.4954, "step": 11698 }, { "epoch": 0.4932436705525223, "grad_norm": 18.212596893310547, "learning_rate": 4.6422436678222306e-05, "loss": 11.3507, "step": 11699 }, { "epoch": 0.49328583173472185, "grad_norm": 23.809188842773438, "learning_rate": 4.642165589181424e-05, "loss": 10.4557, "step": 11700 }, { "epoch": 0.49332799291692137, "grad_norm": 27.49920654296875, "learning_rate": 4.642087510540617e-05, "loss": 11.7832, "step": 11701 }, { "epoch": 0.49337015409912094, "grad_norm": 18.213544845581055, "learning_rate": 4.64200943189981e-05, "loss": 11.8065, "step": 11702 }, { "epoch": 0.4934123152813205, "grad_norm": 18.736360549926758, "learning_rate": 4.6419313532590024e-05, "loss": 11.7651, "step": 11703 }, { "epoch": 0.49345447646352003, "grad_norm": 17.64032554626465, "learning_rate": 4.6418532746181956e-05, "loss": 11.3501, "step": 11704 }, { "epoch": 0.4934966376457196, "grad_norm": 18.980199813842773, "learning_rate": 4.641775195977389e-05, "loss": 11.2032, "step": 11705 }, { "epoch": 0.4935387988279191, "grad_norm": 16.664031982421875, "learning_rate": 4.641697117336581e-05, "loss": 10.9448, "step": 11706 }, { "epoch": 0.4935809600101187, "grad_norm": 15.874828338623047, "learning_rate": 4.641619038695775e-05, "loss": 11.0661, "step": 11707 }, { "epoch": 0.4936231211923182, "grad_norm": 36.53801727294922, "learning_rate": 4.6415409600549674e-05, "loss": 11.0595, "step": 11708 }, { "epoch": 0.4936652823745178, "grad_norm": 21.678709030151367, "learning_rate": 4.6414628814141605e-05, "loss": 12.9079, "step": 11709 }, { "epoch": 0.4937074435567173, "grad_norm": 24.242233276367188, "learning_rate": 4.6413848027733537e-05, "loss": 12.9779, "step": 11710 }, { "epoch": 0.4937496047389169, "grad_norm": 20.72310447692871, "learning_rate": 4.641306724132546e-05, "loss": 11.5678, "step": 11711 }, { "epoch": 0.49379176592111645, "grad_norm": 17.707847595214844, "learning_rate": 4.64122864549174e-05, "loss": 10.483, "step": 11712 }, { "epoch": 0.493833927103316, "grad_norm": 22.728208541870117, "learning_rate": 4.6411505668509324e-05, "loss": 10.9293, "step": 11713 }, { "epoch": 0.49387608828551555, "grad_norm": 28.11190414428711, "learning_rate": 4.6410724882101255e-05, "loss": 11.7672, "step": 11714 }, { "epoch": 0.49391824946771506, "grad_norm": 25.653017044067383, "learning_rate": 4.6409944095693186e-05, "loss": 12.2651, "step": 11715 }, { "epoch": 0.49396041064991464, "grad_norm": 23.819372177124023, "learning_rate": 4.640916330928512e-05, "loss": 10.6299, "step": 11716 }, { "epoch": 0.49400257183211416, "grad_norm": 27.7428035736084, "learning_rate": 4.640838252287704e-05, "loss": 11.9071, "step": 11717 }, { "epoch": 0.49404473301431373, "grad_norm": 20.417909622192383, "learning_rate": 4.640760173646897e-05, "loss": 9.7021, "step": 11718 }, { "epoch": 0.49408689419651325, "grad_norm": 18.175695419311523, "learning_rate": 4.6406820950060904e-05, "loss": 11.4084, "step": 11719 }, { "epoch": 0.4941290553787128, "grad_norm": 19.79462432861328, "learning_rate": 4.640604016365283e-05, "loss": 11.5028, "step": 11720 }, { "epoch": 0.49417121656091234, "grad_norm": 20.930171966552734, "learning_rate": 4.640525937724477e-05, "loss": 10.5617, "step": 11721 }, { "epoch": 0.4942133777431119, "grad_norm": 23.635353088378906, "learning_rate": 4.640447859083669e-05, "loss": 10.3709, "step": 11722 }, { "epoch": 0.4942555389253115, "grad_norm": 21.63753890991211, "learning_rate": 4.640369780442862e-05, "loss": 10.6151, "step": 11723 }, { "epoch": 0.494297700107511, "grad_norm": 20.788909912109375, "learning_rate": 4.6402917018020554e-05, "loss": 11.0206, "step": 11724 }, { "epoch": 0.4943398612897106, "grad_norm": 41.65485382080078, "learning_rate": 4.640213623161248e-05, "loss": 12.1107, "step": 11725 }, { "epoch": 0.4943820224719101, "grad_norm": 66.59357452392578, "learning_rate": 4.6401355445204416e-05, "loss": 10.3582, "step": 11726 }, { "epoch": 0.4944241836541097, "grad_norm": 53.244285583496094, "learning_rate": 4.640057465879634e-05, "loss": 11.8571, "step": 11727 }, { "epoch": 0.4944663448363092, "grad_norm": 317.21636962890625, "learning_rate": 4.639979387238827e-05, "loss": 12.5611, "step": 11728 }, { "epoch": 0.49450850601850876, "grad_norm": 20.730504989624023, "learning_rate": 4.63990130859802e-05, "loss": 12.058, "step": 11729 }, { "epoch": 0.4945506672007083, "grad_norm": 898.16162109375, "learning_rate": 4.639823229957213e-05, "loss": 10.8304, "step": 11730 }, { "epoch": 0.49459282838290786, "grad_norm": 839.0836791992188, "learning_rate": 4.639745151316406e-05, "loss": 10.7137, "step": 11731 }, { "epoch": 0.49463498956510743, "grad_norm": 1505.7489013671875, "learning_rate": 4.639667072675599e-05, "loss": 12.4334, "step": 11732 }, { "epoch": 0.49467715074730695, "grad_norm": 9494.3251953125, "learning_rate": 4.639588994034792e-05, "loss": 10.8504, "step": 11733 }, { "epoch": 0.4947193119295065, "grad_norm": 6417.4208984375, "learning_rate": 4.639510915393985e-05, "loss": 10.4816, "step": 11734 }, { "epoch": 0.49476147311170604, "grad_norm": 116.38665771484375, "learning_rate": 4.6394328367531784e-05, "loss": 11.5724, "step": 11735 }, { "epoch": 0.4948036342939056, "grad_norm": 9961.1337890625, "learning_rate": 4.639354758112371e-05, "loss": 11.3749, "step": 11736 }, { "epoch": 0.49484579547610513, "grad_norm": 56.06163024902344, "learning_rate": 4.639276679471564e-05, "loss": 10.7714, "step": 11737 }, { "epoch": 0.4948879566583047, "grad_norm": 9697.8056640625, "learning_rate": 4.639198600830757e-05, "loss": 10.7779, "step": 11738 }, { "epoch": 0.4949301178405042, "grad_norm": 1237.0074462890625, "learning_rate": 4.6391205221899496e-05, "loss": 10.7915, "step": 11739 }, { "epoch": 0.4949722790227038, "grad_norm": 360.1095275878906, "learning_rate": 4.6390424435491434e-05, "loss": 11.4939, "step": 11740 }, { "epoch": 0.49501444020490337, "grad_norm": 2249.405029296875, "learning_rate": 4.638964364908336e-05, "loss": 10.8081, "step": 11741 }, { "epoch": 0.4950566013871029, "grad_norm": 1716.950927734375, "learning_rate": 4.638886286267529e-05, "loss": 10.0892, "step": 11742 }, { "epoch": 0.49509876256930246, "grad_norm": 1158.6356201171875, "learning_rate": 4.638808207626722e-05, "loss": 13.4509, "step": 11743 }, { "epoch": 0.495140923751502, "grad_norm": 218.45242309570312, "learning_rate": 4.6387301289859145e-05, "loss": 11.2801, "step": 11744 }, { "epoch": 0.49518308493370156, "grad_norm": 198.01319885253906, "learning_rate": 4.638652050345108e-05, "loss": 10.4408, "step": 11745 }, { "epoch": 0.4952252461159011, "grad_norm": 58.595916748046875, "learning_rate": 4.638573971704301e-05, "loss": 10.7255, "step": 11746 }, { "epoch": 0.49526740729810065, "grad_norm": 113.93099212646484, "learning_rate": 4.638495893063494e-05, "loss": 10.6216, "step": 11747 }, { "epoch": 0.49530956848030017, "grad_norm": 308.0628967285156, "learning_rate": 4.638417814422687e-05, "loss": 10.7385, "step": 11748 }, { "epoch": 0.49535172966249974, "grad_norm": 88.2845687866211, "learning_rate": 4.6383397357818795e-05, "loss": 10.7899, "step": 11749 }, { "epoch": 0.49539389084469926, "grad_norm": 69.01258087158203, "learning_rate": 4.6382616571410726e-05, "loss": 10.459, "step": 11750 }, { "epoch": 0.49543605202689883, "grad_norm": 103.79247283935547, "learning_rate": 4.638183578500266e-05, "loss": 10.2327, "step": 11751 }, { "epoch": 0.4954782132090984, "grad_norm": 49.28877639770508, "learning_rate": 4.638105499859459e-05, "loss": 11.2766, "step": 11752 }, { "epoch": 0.4955203743912979, "grad_norm": 55.6826057434082, "learning_rate": 4.638027421218651e-05, "loss": 10.5226, "step": 11753 }, { "epoch": 0.4955625355734975, "grad_norm": 46.18689727783203, "learning_rate": 4.6379493425778444e-05, "loss": 11.7704, "step": 11754 }, { "epoch": 0.495604696755697, "grad_norm": 89.17078399658203, "learning_rate": 4.6378712639370375e-05, "loss": 11.8878, "step": 11755 }, { "epoch": 0.4956468579378966, "grad_norm": 33.99412155151367, "learning_rate": 4.637793185296231e-05, "loss": 11.396, "step": 11756 }, { "epoch": 0.4956890191200961, "grad_norm": 49.45012664794922, "learning_rate": 4.637715106655424e-05, "loss": 11.8096, "step": 11757 }, { "epoch": 0.4957311803022957, "grad_norm": 40.07582473754883, "learning_rate": 4.637637028014616e-05, "loss": 12.2069, "step": 11758 }, { "epoch": 0.4957733414844952, "grad_norm": 49.8058967590332, "learning_rate": 4.63755894937381e-05, "loss": 11.3112, "step": 11759 }, { "epoch": 0.4958155026666948, "grad_norm": 28.90351104736328, "learning_rate": 4.6374808707330025e-05, "loss": 11.3563, "step": 11760 }, { "epoch": 0.49585766384889435, "grad_norm": 30.55677032470703, "learning_rate": 4.6374027920921956e-05, "loss": 11.6316, "step": 11761 }, { "epoch": 0.49589982503109387, "grad_norm": 18.60417938232422, "learning_rate": 4.637324713451389e-05, "loss": 12.457, "step": 11762 }, { "epoch": 0.49594198621329344, "grad_norm": 23.835359573364258, "learning_rate": 4.637246634810581e-05, "loss": 12.1013, "step": 11763 }, { "epoch": 0.49598414739549296, "grad_norm": 15.85185718536377, "learning_rate": 4.637168556169774e-05, "loss": 12.3138, "step": 11764 }, { "epoch": 0.49602630857769253, "grad_norm": 21.828250885009766, "learning_rate": 4.6370904775289674e-05, "loss": 12.6633, "step": 11765 }, { "epoch": 0.49606846975989205, "grad_norm": 29.266963958740234, "learning_rate": 4.6370123988881606e-05, "loss": 11.2754, "step": 11766 }, { "epoch": 0.4961106309420916, "grad_norm": 44.727134704589844, "learning_rate": 4.636934320247353e-05, "loss": 11.6267, "step": 11767 }, { "epoch": 0.49615279212429114, "grad_norm": 15.271617889404297, "learning_rate": 4.636856241606546e-05, "loss": 11.9304, "step": 11768 }, { "epoch": 0.4961949533064907, "grad_norm": 37.22346115112305, "learning_rate": 4.636778162965739e-05, "loss": 11.7799, "step": 11769 }, { "epoch": 0.4962371144886903, "grad_norm": 31.83363914489746, "learning_rate": 4.6367000843249324e-05, "loss": 11.1161, "step": 11770 }, { "epoch": 0.4962792756708898, "grad_norm": 80.13459777832031, "learning_rate": 4.6366220056841255e-05, "loss": 10.9339, "step": 11771 }, { "epoch": 0.4963214368530894, "grad_norm": 59.288299560546875, "learning_rate": 4.636543927043318e-05, "loss": 10.6321, "step": 11772 }, { "epoch": 0.4963635980352889, "grad_norm": 101.68877410888672, "learning_rate": 4.636465848402511e-05, "loss": 11.0805, "step": 11773 }, { "epoch": 0.4964057592174885, "grad_norm": 130.5559844970703, "learning_rate": 4.636387769761704e-05, "loss": 10.9747, "step": 11774 }, { "epoch": 0.496447920399688, "grad_norm": 658.432373046875, "learning_rate": 4.636309691120897e-05, "loss": 10.9456, "step": 11775 }, { "epoch": 0.49649008158188757, "grad_norm": 1787.67724609375, "learning_rate": 4.6362316124800905e-05, "loss": 11.3953, "step": 11776 }, { "epoch": 0.4965322427640871, "grad_norm": 32.4023323059082, "learning_rate": 4.636153533839283e-05, "loss": 11.0596, "step": 11777 }, { "epoch": 0.49657440394628666, "grad_norm": 4100.35986328125, "learning_rate": 4.636075455198476e-05, "loss": 10.8146, "step": 11778 }, { "epoch": 0.4966165651284862, "grad_norm": 3090.1513671875, "learning_rate": 4.635997376557669e-05, "loss": 10.5767, "step": 11779 }, { "epoch": 0.49665872631068575, "grad_norm": 4029.85791015625, "learning_rate": 4.635919297916862e-05, "loss": 12.2262, "step": 11780 }, { "epoch": 0.4967008874928853, "grad_norm": 2215.20263671875, "learning_rate": 4.6358412192760554e-05, "loss": 11.2995, "step": 11781 }, { "epoch": 0.49674304867508484, "grad_norm": 2190.91943359375, "learning_rate": 4.635763140635248e-05, "loss": 10.8543, "step": 11782 }, { "epoch": 0.4967852098572844, "grad_norm": 12479.33984375, "learning_rate": 4.635685061994441e-05, "loss": 11.4072, "step": 11783 }, { "epoch": 0.49682737103948393, "grad_norm": 9129.318359375, "learning_rate": 4.635606983353634e-05, "loss": 10.6649, "step": 11784 }, { "epoch": 0.4968695322216835, "grad_norm": 5869.36669921875, "learning_rate": 4.635528904712827e-05, "loss": 12.9479, "step": 11785 }, { "epoch": 0.496911693403883, "grad_norm": 642.4754638671875, "learning_rate": 4.63545082607202e-05, "loss": 12.9174, "step": 11786 }, { "epoch": 0.4969538545860826, "grad_norm": 8062.15185546875, "learning_rate": 4.635372747431213e-05, "loss": 11.3689, "step": 11787 }, { "epoch": 0.4969960157682821, "grad_norm": 8650.0341796875, "learning_rate": 4.635294668790406e-05, "loss": 11.6929, "step": 11788 }, { "epoch": 0.4970381769504817, "grad_norm": 6178.013671875, "learning_rate": 4.6352165901495984e-05, "loss": 11.0783, "step": 11789 }, { "epoch": 0.49708033813268127, "grad_norm": 44165.734375, "learning_rate": 4.635138511508792e-05, "loss": 10.5509, "step": 11790 }, { "epoch": 0.4971224993148808, "grad_norm": 6290.81689453125, "learning_rate": 4.6350604328679847e-05, "loss": 11.0096, "step": 11791 }, { "epoch": 0.49716466049708036, "grad_norm": 1269.30078125, "learning_rate": 4.634982354227178e-05, "loss": 10.7813, "step": 11792 }, { "epoch": 0.4972068216792799, "grad_norm": 964.113525390625, "learning_rate": 4.634904275586371e-05, "loss": 11.0796, "step": 11793 }, { "epoch": 0.49724898286147945, "grad_norm": 1559.6514892578125, "learning_rate": 4.6348261969455634e-05, "loss": 10.418, "step": 11794 }, { "epoch": 0.49729114404367897, "grad_norm": 106.29056549072266, "learning_rate": 4.634748118304757e-05, "loss": 11.0938, "step": 11795 }, { "epoch": 0.49733330522587854, "grad_norm": 38.49861145019531, "learning_rate": 4.6346700396639496e-05, "loss": 10.5632, "step": 11796 }, { "epoch": 0.49737546640807806, "grad_norm": 64.45005798339844, "learning_rate": 4.634591961023143e-05, "loss": 10.9425, "step": 11797 }, { "epoch": 0.49741762759027763, "grad_norm": 47.96970748901367, "learning_rate": 4.634513882382336e-05, "loss": 11.0746, "step": 11798 }, { "epoch": 0.4974597887724772, "grad_norm": 74.06410217285156, "learning_rate": 4.634435803741528e-05, "loss": 10.4793, "step": 11799 }, { "epoch": 0.4975019499546767, "grad_norm": 61.8734130859375, "learning_rate": 4.6343577251007214e-05, "loss": 11.4522, "step": 11800 }, { "epoch": 0.4975441111368763, "grad_norm": 61.82181167602539, "learning_rate": 4.6342796464599146e-05, "loss": 12.5768, "step": 11801 }, { "epoch": 0.4975862723190758, "grad_norm": 56.17780303955078, "learning_rate": 4.634201567819108e-05, "loss": 10.6356, "step": 11802 }, { "epoch": 0.4976284335012754, "grad_norm": 38.36943054199219, "learning_rate": 4.634123489178301e-05, "loss": 11.1874, "step": 11803 }, { "epoch": 0.4976705946834749, "grad_norm": 34.43208312988281, "learning_rate": 4.634045410537494e-05, "loss": 11.7929, "step": 11804 }, { "epoch": 0.4977127558656745, "grad_norm": 43.359493255615234, "learning_rate": 4.6339673318966864e-05, "loss": 11.7748, "step": 11805 }, { "epoch": 0.497754917047874, "grad_norm": 47.63283920288086, "learning_rate": 4.6338892532558795e-05, "loss": 11.5532, "step": 11806 }, { "epoch": 0.4977970782300736, "grad_norm": 42.2318115234375, "learning_rate": 4.6338111746150726e-05, "loss": 11.6421, "step": 11807 }, { "epoch": 0.4978392394122731, "grad_norm": 34.91887283325195, "learning_rate": 4.633733095974265e-05, "loss": 11.6509, "step": 11808 }, { "epoch": 0.49788140059447267, "grad_norm": 412.4742736816406, "learning_rate": 4.633655017333459e-05, "loss": 11.628, "step": 11809 }, { "epoch": 0.49792356177667224, "grad_norm": 65.88456726074219, "learning_rate": 4.6335769386926513e-05, "loss": 11.3062, "step": 11810 }, { "epoch": 0.49796572295887176, "grad_norm": 56.39039611816406, "learning_rate": 4.6334988600518445e-05, "loss": 14.484, "step": 11811 }, { "epoch": 0.49800788414107133, "grad_norm": 19.61841583251953, "learning_rate": 4.6334207814110376e-05, "loss": 11.3258, "step": 11812 }, { "epoch": 0.49805004532327085, "grad_norm": 57.636295318603516, "learning_rate": 4.63334270277023e-05, "loss": 12.7332, "step": 11813 }, { "epoch": 0.4980922065054704, "grad_norm": 26.88374137878418, "learning_rate": 4.633264624129424e-05, "loss": 11.2288, "step": 11814 }, { "epoch": 0.49813436768766994, "grad_norm": 26.282564163208008, "learning_rate": 4.633186545488616e-05, "loss": 12.528, "step": 11815 }, { "epoch": 0.4981765288698695, "grad_norm": 56.40142059326172, "learning_rate": 4.6331084668478094e-05, "loss": 11.2217, "step": 11816 }, { "epoch": 0.49821869005206904, "grad_norm": 237.49874877929688, "learning_rate": 4.6330303882070025e-05, "loss": 11.3397, "step": 11817 }, { "epoch": 0.4982608512342686, "grad_norm": 153.0418701171875, "learning_rate": 4.632952309566195e-05, "loss": 12.334, "step": 11818 }, { "epoch": 0.4983030124164682, "grad_norm": 349.5934143066406, "learning_rate": 4.632874230925388e-05, "loss": 13.2059, "step": 11819 }, { "epoch": 0.4983451735986677, "grad_norm": 272.8626403808594, "learning_rate": 4.632796152284581e-05, "loss": 12.6761, "step": 11820 }, { "epoch": 0.4983873347808673, "grad_norm": 360.1849670410156, "learning_rate": 4.6327180736437744e-05, "loss": 12.1439, "step": 11821 }, { "epoch": 0.4984294959630668, "grad_norm": 154.73587036132812, "learning_rate": 4.632639995002967e-05, "loss": 13.3814, "step": 11822 }, { "epoch": 0.49847165714526637, "grad_norm": 788.208984375, "learning_rate": 4.6325619163621606e-05, "loss": 11.4621, "step": 11823 }, { "epoch": 0.4985138183274659, "grad_norm": 31.516889572143555, "learning_rate": 4.632483837721353e-05, "loss": 14.1387, "step": 11824 }, { "epoch": 0.49855597950966546, "grad_norm": 15.586057662963867, "learning_rate": 4.632405759080546e-05, "loss": 12.0454, "step": 11825 }, { "epoch": 0.498598140691865, "grad_norm": 82.36544036865234, "learning_rate": 4.632327680439739e-05, "loss": 10.627, "step": 11826 }, { "epoch": 0.49864030187406455, "grad_norm": 30.14101219177246, "learning_rate": 4.632249601798932e-05, "loss": 12.5346, "step": 11827 }, { "epoch": 0.4986824630562641, "grad_norm": 295.4556884765625, "learning_rate": 4.6321715231581256e-05, "loss": 11.2392, "step": 11828 }, { "epoch": 0.49872462423846364, "grad_norm": 41.59814453125, "learning_rate": 4.632093444517318e-05, "loss": 12.8866, "step": 11829 }, { "epoch": 0.4987667854206632, "grad_norm": 86.2600326538086, "learning_rate": 4.632015365876511e-05, "loss": 11.3378, "step": 11830 }, { "epoch": 0.49880894660286274, "grad_norm": 218.45462036132812, "learning_rate": 4.631937287235704e-05, "loss": 12.4718, "step": 11831 }, { "epoch": 0.4988511077850623, "grad_norm": 195.3949737548828, "learning_rate": 4.631859208594897e-05, "loss": 12.9589, "step": 11832 }, { "epoch": 0.4988932689672618, "grad_norm": 211.1729278564453, "learning_rate": 4.63178112995409e-05, "loss": 11.4575, "step": 11833 }, { "epoch": 0.4989354301494614, "grad_norm": 113.2733154296875, "learning_rate": 4.631703051313283e-05, "loss": 12.1165, "step": 11834 }, { "epoch": 0.4989775913316609, "grad_norm": 235.37222290039062, "learning_rate": 4.631624972672476e-05, "loss": 11.3074, "step": 11835 }, { "epoch": 0.4990197525138605, "grad_norm": 48.56479263305664, "learning_rate": 4.631546894031669e-05, "loss": 11.2757, "step": 11836 }, { "epoch": 0.49906191369606, "grad_norm": 122.03820037841797, "learning_rate": 4.631468815390862e-05, "loss": 11.258, "step": 11837 }, { "epoch": 0.4991040748782596, "grad_norm": 96.09656524658203, "learning_rate": 4.631390736750055e-05, "loss": 11.2899, "step": 11838 }, { "epoch": 0.49914623606045916, "grad_norm": 43.700740814208984, "learning_rate": 4.631312658109248e-05, "loss": 12.0504, "step": 11839 }, { "epoch": 0.4991883972426587, "grad_norm": 34.14548110961914, "learning_rate": 4.631234579468441e-05, "loss": 11.6775, "step": 11840 }, { "epoch": 0.49923055842485825, "grad_norm": 119.45032501220703, "learning_rate": 4.6311565008276335e-05, "loss": 11.5483, "step": 11841 }, { "epoch": 0.49927271960705777, "grad_norm": 39.901336669921875, "learning_rate": 4.631078422186827e-05, "loss": 11.1984, "step": 11842 }, { "epoch": 0.49931488078925734, "grad_norm": 46.09782791137695, "learning_rate": 4.63100034354602e-05, "loss": 11.9405, "step": 11843 }, { "epoch": 0.49935704197145686, "grad_norm": 45.62253189086914, "learning_rate": 4.630922264905212e-05, "loss": 10.7955, "step": 11844 }, { "epoch": 0.49939920315365643, "grad_norm": 53.9639778137207, "learning_rate": 4.630844186264406e-05, "loss": 12.8404, "step": 11845 }, { "epoch": 0.49944136433585595, "grad_norm": 45.28071975708008, "learning_rate": 4.6307661076235985e-05, "loss": 12.6942, "step": 11846 }, { "epoch": 0.4994835255180555, "grad_norm": 24.248762130737305, "learning_rate": 4.6306880289827916e-05, "loss": 13.0958, "step": 11847 }, { "epoch": 0.4995256867002551, "grad_norm": 44.702354431152344, "learning_rate": 4.630609950341985e-05, "loss": 11.5879, "step": 11848 }, { "epoch": 0.4995678478824546, "grad_norm": 38.3474006652832, "learning_rate": 4.630531871701178e-05, "loss": 11.8632, "step": 11849 }, { "epoch": 0.4996100090646542, "grad_norm": 29.66574478149414, "learning_rate": 4.630453793060371e-05, "loss": 11.5847, "step": 11850 }, { "epoch": 0.4996521702468537, "grad_norm": 21.44057846069336, "learning_rate": 4.6303757144195634e-05, "loss": 11.2526, "step": 11851 }, { "epoch": 0.4996943314290533, "grad_norm": 27.497568130493164, "learning_rate": 4.6302976357787565e-05, "loss": 11.9163, "step": 11852 }, { "epoch": 0.4997364926112528, "grad_norm": 21.6080265045166, "learning_rate": 4.6302195571379497e-05, "loss": 11.0182, "step": 11853 }, { "epoch": 0.4997786537934524, "grad_norm": 15.331035614013672, "learning_rate": 4.630141478497143e-05, "loss": 11.762, "step": 11854 }, { "epoch": 0.4998208149756519, "grad_norm": 24.747751235961914, "learning_rate": 4.630063399856335e-05, "loss": 11.5001, "step": 11855 }, { "epoch": 0.49986297615785147, "grad_norm": 17.384389877319336, "learning_rate": 4.6299853212155284e-05, "loss": 12.9765, "step": 11856 }, { "epoch": 0.49990513734005104, "grad_norm": 14.807623863220215, "learning_rate": 4.6299072425747215e-05, "loss": 11.018, "step": 11857 }, { "epoch": 0.49994729852225056, "grad_norm": 15.741525650024414, "learning_rate": 4.629829163933914e-05, "loss": 12.8453, "step": 11858 }, { "epoch": 0.49998945970445013, "grad_norm": 23.356935501098633, "learning_rate": 4.629751085293108e-05, "loss": 11.1067, "step": 11859 }, { "epoch": 0.5000316208866497, "grad_norm": 14.27096939086914, "learning_rate": 4.6296730066523e-05, "loss": 11.6576, "step": 11860 }, { "epoch": 0.5000737820688492, "grad_norm": 24.06934928894043, "learning_rate": 4.629594928011493e-05, "loss": 12.6179, "step": 11861 }, { "epoch": 0.5001159432510488, "grad_norm": 16.073225021362305, "learning_rate": 4.6295168493706864e-05, "loss": 11.032, "step": 11862 }, { "epoch": 0.5001581044332483, "grad_norm": 13.981496810913086, "learning_rate": 4.629438770729879e-05, "loss": 11.103, "step": 11863 }, { "epoch": 0.5002002656154478, "grad_norm": 46.456871032714844, "learning_rate": 4.629360692089073e-05, "loss": 11.6213, "step": 11864 }, { "epoch": 0.5002424267976474, "grad_norm": 14.303961753845215, "learning_rate": 4.629282613448265e-05, "loss": 11.1868, "step": 11865 }, { "epoch": 0.500284587979847, "grad_norm": 13.636601448059082, "learning_rate": 4.629204534807458e-05, "loss": 11.3735, "step": 11866 }, { "epoch": 0.5003267491620466, "grad_norm": 16.427433013916016, "learning_rate": 4.6291264561666514e-05, "loss": 11.2552, "step": 11867 }, { "epoch": 0.500368910344246, "grad_norm": 16.288543701171875, "learning_rate": 4.6290483775258445e-05, "loss": 10.9488, "step": 11868 }, { "epoch": 0.5004110715264456, "grad_norm": 15.884294509887695, "learning_rate": 4.628970298885037e-05, "loss": 10.5919, "step": 11869 }, { "epoch": 0.5004532327086452, "grad_norm": 22.188817977905273, "learning_rate": 4.62889222024423e-05, "loss": 11.0003, "step": 11870 }, { "epoch": 0.5004953938908447, "grad_norm": 19.68981170654297, "learning_rate": 4.628814141603423e-05, "loss": 11.1903, "step": 11871 }, { "epoch": 0.5005375550730442, "grad_norm": 15.733154296875, "learning_rate": 4.6287360629626163e-05, "loss": 11.11, "step": 11872 }, { "epoch": 0.5005797162552438, "grad_norm": 29.82745361328125, "learning_rate": 4.6286579843218095e-05, "loss": 10.2709, "step": 11873 }, { "epoch": 0.5006218774374434, "grad_norm": 41.85045623779297, "learning_rate": 4.628579905681002e-05, "loss": 10.9992, "step": 11874 }, { "epoch": 0.5006640386196429, "grad_norm": 15.823761940002441, "learning_rate": 4.628501827040195e-05, "loss": 10.9868, "step": 11875 }, { "epoch": 0.5007061998018424, "grad_norm": 81.28671264648438, "learning_rate": 4.628423748399388e-05, "loss": 10.5252, "step": 11876 }, { "epoch": 0.500748360984042, "grad_norm": 27.25835609436035, "learning_rate": 4.6283456697585806e-05, "loss": 11.2997, "step": 11877 }, { "epoch": 0.5007905221662415, "grad_norm": 43.833316802978516, "learning_rate": 4.6282675911177744e-05, "loss": 10.0758, "step": 11878 }, { "epoch": 0.5008326833484411, "grad_norm": 207.97988891601562, "learning_rate": 4.628189512476967e-05, "loss": 10.428, "step": 11879 }, { "epoch": 0.5008748445306407, "grad_norm": 13.054567337036133, "learning_rate": 4.62811143383616e-05, "loss": 12.1237, "step": 11880 }, { "epoch": 0.5009170057128401, "grad_norm": 262.2446594238281, "learning_rate": 4.628033355195353e-05, "loss": 9.5974, "step": 11881 }, { "epoch": 0.5009591668950397, "grad_norm": 13.783349990844727, "learning_rate": 4.6279552765545456e-05, "loss": 12.4477, "step": 11882 }, { "epoch": 0.5010013280772393, "grad_norm": 468.5789794921875, "learning_rate": 4.6278771979137394e-05, "loss": 9.9528, "step": 11883 }, { "epoch": 0.5010434892594389, "grad_norm": 353.61199951171875, "learning_rate": 4.627799119272932e-05, "loss": 10.4346, "step": 11884 }, { "epoch": 0.5010856504416383, "grad_norm": 589.0536499023438, "learning_rate": 4.627721040632125e-05, "loss": 11.0367, "step": 11885 }, { "epoch": 0.5011278116238379, "grad_norm": 651.36181640625, "learning_rate": 4.627642961991318e-05, "loss": 11.2093, "step": 11886 }, { "epoch": 0.5011699728060375, "grad_norm": 2160.36962890625, "learning_rate": 4.6275648833505105e-05, "loss": 9.9647, "step": 11887 }, { "epoch": 0.501212133988237, "grad_norm": 975.1152954101562, "learning_rate": 4.6274868047097036e-05, "loss": 10.5552, "step": 11888 }, { "epoch": 0.5012542951704366, "grad_norm": 10753.1923828125, "learning_rate": 4.627408726068897e-05, "loss": 9.8983, "step": 11889 }, { "epoch": 0.5012964563526361, "grad_norm": 556.7330932617188, "learning_rate": 4.62733064742809e-05, "loss": 10.2021, "step": 11890 }, { "epoch": 0.5013386175348357, "grad_norm": 2600.1123046875, "learning_rate": 4.6272525687872823e-05, "loss": 11.0993, "step": 11891 }, { "epoch": 0.5013807787170352, "grad_norm": 475.35272216796875, "learning_rate": 4.627174490146476e-05, "loss": 10.3103, "step": 11892 }, { "epoch": 0.5014229398992348, "grad_norm": 1439.195556640625, "learning_rate": 4.6270964115056686e-05, "loss": 10.853, "step": 11893 }, { "epoch": 0.5014651010814343, "grad_norm": 544.5853271484375, "learning_rate": 4.627018332864862e-05, "loss": 10.8168, "step": 11894 }, { "epoch": 0.5015072622636338, "grad_norm": 449.4422912597656, "learning_rate": 4.626940254224055e-05, "loss": 11.1443, "step": 11895 }, { "epoch": 0.5015494234458334, "grad_norm": 21.516965866088867, "learning_rate": 4.626862175583247e-05, "loss": 10.412, "step": 11896 }, { "epoch": 0.501591584628033, "grad_norm": 15.266242027282715, "learning_rate": 4.626784096942441e-05, "loss": 10.2585, "step": 11897 }, { "epoch": 0.5016337458102326, "grad_norm": 96.46782684326172, "learning_rate": 4.6267060183016336e-05, "loss": 10.4677, "step": 11898 }, { "epoch": 0.501675906992432, "grad_norm": 58.743778228759766, "learning_rate": 4.626627939660827e-05, "loss": 9.8189, "step": 11899 }, { "epoch": 0.5017180681746316, "grad_norm": 32.37036895751953, "learning_rate": 4.62654986102002e-05, "loss": 10.8077, "step": 11900 }, { "epoch": 0.5017602293568312, "grad_norm": 36.477081298828125, "learning_rate": 4.626471782379212e-05, "loss": 10.0596, "step": 11901 }, { "epoch": 0.5018023905390308, "grad_norm": 33.96795654296875, "learning_rate": 4.6263937037384054e-05, "loss": 11.4526, "step": 11902 }, { "epoch": 0.5018445517212302, "grad_norm": 18.7662353515625, "learning_rate": 4.6263156250975985e-05, "loss": 10.9208, "step": 11903 }, { "epoch": 0.5018867129034298, "grad_norm": 27.115108489990234, "learning_rate": 4.6262375464567916e-05, "loss": 10.6912, "step": 11904 }, { "epoch": 0.5019288740856294, "grad_norm": 29.550308227539062, "learning_rate": 4.626159467815985e-05, "loss": 11.2839, "step": 11905 }, { "epoch": 0.5019710352678289, "grad_norm": 66.4714584350586, "learning_rate": 4.626081389175177e-05, "loss": 10.2366, "step": 11906 }, { "epoch": 0.5020131964500285, "grad_norm": 9.400750160217285, "learning_rate": 4.62600331053437e-05, "loss": 10.5006, "step": 11907 }, { "epoch": 0.502055357632228, "grad_norm": 25.450420379638672, "learning_rate": 4.6259252318935635e-05, "loss": 10.069, "step": 11908 }, { "epoch": 0.5020975188144275, "grad_norm": 9.328808784484863, "learning_rate": 4.6258471532527566e-05, "loss": 10.1864, "step": 11909 }, { "epoch": 0.5021396799966271, "grad_norm": 18.214488983154297, "learning_rate": 4.625769074611949e-05, "loss": 10.3852, "step": 11910 }, { "epoch": 0.5021818411788267, "grad_norm": 24.18954849243164, "learning_rate": 4.625690995971143e-05, "loss": 9.6131, "step": 11911 }, { "epoch": 0.5022240023610262, "grad_norm": 21.48043441772461, "learning_rate": 4.625612917330335e-05, "loss": 10.4913, "step": 11912 }, { "epoch": 0.5022661635432257, "grad_norm": 128.86744689941406, "learning_rate": 4.6255348386895284e-05, "loss": 11.304, "step": 11913 }, { "epoch": 0.5023083247254253, "grad_norm": 11.39963150024414, "learning_rate": 4.6254567600487215e-05, "loss": 11.2365, "step": 11914 }, { "epoch": 0.5023504859076249, "grad_norm": 65.69625854492188, "learning_rate": 4.625378681407914e-05, "loss": 10.3307, "step": 11915 }, { "epoch": 0.5023926470898245, "grad_norm": 11.688799858093262, "learning_rate": 4.625300602767108e-05, "loss": 10.3583, "step": 11916 }, { "epoch": 0.5024348082720239, "grad_norm": 15.319676399230957, "learning_rate": 4.6252225241263e-05, "loss": 9.6048, "step": 11917 }, { "epoch": 0.5024769694542235, "grad_norm": 32.630924224853516, "learning_rate": 4.6251444454854934e-05, "loss": 10.212, "step": 11918 }, { "epoch": 0.5025191306364231, "grad_norm": 13.679361343383789, "learning_rate": 4.6250663668446865e-05, "loss": 12.4684, "step": 11919 }, { "epoch": 0.5025612918186226, "grad_norm": 13.36674976348877, "learning_rate": 4.624988288203879e-05, "loss": 10.0415, "step": 11920 }, { "epoch": 0.5026034530008221, "grad_norm": 18.178462982177734, "learning_rate": 4.624910209563072e-05, "loss": 10.1493, "step": 11921 }, { "epoch": 0.5026456141830217, "grad_norm": 29.256114959716797, "learning_rate": 4.624832130922265e-05, "loss": 10.9207, "step": 11922 }, { "epoch": 0.5026877753652212, "grad_norm": 14.819157600402832, "learning_rate": 4.624754052281458e-05, "loss": 9.6016, "step": 11923 }, { "epoch": 0.5027299365474208, "grad_norm": 23.007749557495117, "learning_rate": 4.624675973640651e-05, "loss": 10.086, "step": 11924 }, { "epoch": 0.5027720977296204, "grad_norm": 89.05233764648438, "learning_rate": 4.624597894999844e-05, "loss": 10.4278, "step": 11925 }, { "epoch": 0.5028142589118199, "grad_norm": 43.8101806640625, "learning_rate": 4.624519816359037e-05, "loss": 10.2488, "step": 11926 }, { "epoch": 0.5028564200940194, "grad_norm": 104.99779510498047, "learning_rate": 4.6244417377182295e-05, "loss": 10.4712, "step": 11927 }, { "epoch": 0.502898581276219, "grad_norm": 29.478694915771484, "learning_rate": 4.624363659077423e-05, "loss": 10.6925, "step": 11928 }, { "epoch": 0.5029407424584186, "grad_norm": 215.86325073242188, "learning_rate": 4.624285580436616e-05, "loss": 10.1607, "step": 11929 }, { "epoch": 0.502982903640618, "grad_norm": 356.03466796875, "learning_rate": 4.6242075017958095e-05, "loss": 9.2598, "step": 11930 }, { "epoch": 0.5030250648228176, "grad_norm": 291.3038024902344, "learning_rate": 4.624129423155002e-05, "loss": 10.3925, "step": 11931 }, { "epoch": 0.5030672260050172, "grad_norm": 107.81902313232422, "learning_rate": 4.6240513445141944e-05, "loss": 9.8587, "step": 11932 }, { "epoch": 0.5031093871872168, "grad_norm": 45.87864303588867, "learning_rate": 4.623973265873388e-05, "loss": 10.215, "step": 11933 }, { "epoch": 0.5031515483694162, "grad_norm": 682.0436401367188, "learning_rate": 4.623895187232581e-05, "loss": 12.0105, "step": 11934 }, { "epoch": 0.5031937095516158, "grad_norm": 945.872802734375, "learning_rate": 4.623817108591774e-05, "loss": 10.5514, "step": 11935 }, { "epoch": 0.5032358707338154, "grad_norm": 976.9561767578125, "learning_rate": 4.623739029950967e-05, "loss": 11.0833, "step": 11936 }, { "epoch": 0.503278031916015, "grad_norm": 482.18365478515625, "learning_rate": 4.62366095131016e-05, "loss": 10.0372, "step": 11937 }, { "epoch": 0.5033201930982145, "grad_norm": 249.5332794189453, "learning_rate": 4.6235828726693525e-05, "loss": 10.2904, "step": 11938 }, { "epoch": 0.503362354280414, "grad_norm": 354.28668212890625, "learning_rate": 4.6235047940285456e-05, "loss": 10.2644, "step": 11939 }, { "epoch": 0.5034045154626136, "grad_norm": 48.41199493408203, "learning_rate": 4.623426715387739e-05, "loss": 10.8211, "step": 11940 }, { "epoch": 0.5034466766448131, "grad_norm": 2854.035400390625, "learning_rate": 4.623348636746932e-05, "loss": 9.6301, "step": 11941 }, { "epoch": 0.5034888378270127, "grad_norm": 61.67398452758789, "learning_rate": 4.623270558106125e-05, "loss": 10.304, "step": 11942 }, { "epoch": 0.5035309990092122, "grad_norm": 45.86015319824219, "learning_rate": 4.6231924794653174e-05, "loss": 10.1658, "step": 11943 }, { "epoch": 0.5035731601914117, "grad_norm": 35.62799072265625, "learning_rate": 4.6231144008245106e-05, "loss": 9.7752, "step": 11944 }, { "epoch": 0.5036153213736113, "grad_norm": 29.41254425048828, "learning_rate": 4.623036322183704e-05, "loss": 10.5231, "step": 11945 }, { "epoch": 0.5036574825558109, "grad_norm": 19.701635360717773, "learning_rate": 4.622958243542896e-05, "loss": 10.9469, "step": 11946 }, { "epoch": 0.5036996437380105, "grad_norm": 24.66425132751465, "learning_rate": 4.62288016490209e-05, "loss": 10.2384, "step": 11947 }, { "epoch": 0.5037418049202099, "grad_norm": 42.937564849853516, "learning_rate": 4.6228020862612824e-05, "loss": 9.8566, "step": 11948 }, { "epoch": 0.5037839661024095, "grad_norm": 69.09300231933594, "learning_rate": 4.6227240076204755e-05, "loss": 9.8133, "step": 11949 }, { "epoch": 0.5038261272846091, "grad_norm": 12.86917781829834, "learning_rate": 4.6226459289796687e-05, "loss": 9.9775, "step": 11950 }, { "epoch": 0.5038682884668086, "grad_norm": 26.677997589111328, "learning_rate": 4.622567850338861e-05, "loss": 10.3355, "step": 11951 }, { "epoch": 0.5039104496490081, "grad_norm": 34.20306396484375, "learning_rate": 4.622489771698055e-05, "loss": 9.985, "step": 11952 }, { "epoch": 0.5039526108312077, "grad_norm": 35.54084014892578, "learning_rate": 4.6224116930572474e-05, "loss": 11.1217, "step": 11953 }, { "epoch": 0.5039947720134073, "grad_norm": 26.321380615234375, "learning_rate": 4.6223336144164405e-05, "loss": 10.0303, "step": 11954 }, { "epoch": 0.5040369331956068, "grad_norm": 21.903114318847656, "learning_rate": 4.6222555357756336e-05, "loss": 9.9962, "step": 11955 }, { "epoch": 0.5040790943778064, "grad_norm": 17.142013549804688, "learning_rate": 4.622177457134827e-05, "loss": 10.5031, "step": 11956 }, { "epoch": 0.5041212555600059, "grad_norm": 30.541345596313477, "learning_rate": 4.622099378494019e-05, "loss": 10.3938, "step": 11957 }, { "epoch": 0.5041634167422054, "grad_norm": 22.508970260620117, "learning_rate": 4.622021299853212e-05, "loss": 10.3892, "step": 11958 }, { "epoch": 0.504205577924405, "grad_norm": 14.139625549316406, "learning_rate": 4.6219432212124054e-05, "loss": 10.3622, "step": 11959 }, { "epoch": 0.5042477391066046, "grad_norm": 37.36812973022461, "learning_rate": 4.621865142571598e-05, "loss": 10.2037, "step": 11960 }, { "epoch": 0.504289900288804, "grad_norm": 20.323762893676758, "learning_rate": 4.621787063930792e-05, "loss": 10.2438, "step": 11961 }, { "epoch": 0.5043320614710036, "grad_norm": 18.171005249023438, "learning_rate": 4.621708985289984e-05, "loss": 10.7121, "step": 11962 }, { "epoch": 0.5043742226532032, "grad_norm": 18.357341766357422, "learning_rate": 4.621630906649177e-05, "loss": 10.2833, "step": 11963 }, { "epoch": 0.5044163838354028, "grad_norm": 27.926483154296875, "learning_rate": 4.6215528280083704e-05, "loss": 10.5538, "step": 11964 }, { "epoch": 0.5044585450176023, "grad_norm": 33.88220977783203, "learning_rate": 4.621474749367563e-05, "loss": 9.6295, "step": 11965 }, { "epoch": 0.5045007061998018, "grad_norm": 9.200841903686523, "learning_rate": 4.6213966707267566e-05, "loss": 9.8613, "step": 11966 }, { "epoch": 0.5045428673820014, "grad_norm": 17.125051498413086, "learning_rate": 4.621318592085949e-05, "loss": 10.2747, "step": 11967 }, { "epoch": 0.504585028564201, "grad_norm": 13.025556564331055, "learning_rate": 4.621240513445142e-05, "loss": 10.002, "step": 11968 }, { "epoch": 0.5046271897464005, "grad_norm": 14.961664199829102, "learning_rate": 4.621162434804335e-05, "loss": 10.3321, "step": 11969 }, { "epoch": 0.5046693509286, "grad_norm": 33.25154495239258, "learning_rate": 4.621084356163528e-05, "loss": 10.6079, "step": 11970 }, { "epoch": 0.5047115121107996, "grad_norm": 295.79058837890625, "learning_rate": 4.621006277522721e-05, "loss": 9.6541, "step": 11971 }, { "epoch": 0.5047536732929991, "grad_norm": 307.6005859375, "learning_rate": 4.620928198881914e-05, "loss": 10.0357, "step": 11972 }, { "epoch": 0.5047958344751987, "grad_norm": 177.26927185058594, "learning_rate": 4.620850120241107e-05, "loss": 9.9935, "step": 11973 }, { "epoch": 0.5048379956573983, "grad_norm": 108.12248992919922, "learning_rate": 4.6207720416003e-05, "loss": 10.6084, "step": 11974 }, { "epoch": 0.5048801568395977, "grad_norm": 152.33421325683594, "learning_rate": 4.6206939629594934e-05, "loss": 10.4319, "step": 11975 }, { "epoch": 0.5049223180217973, "grad_norm": 420.73486328125, "learning_rate": 4.620615884318686e-05, "loss": 9.1559, "step": 11976 }, { "epoch": 0.5049644792039969, "grad_norm": 44.75786590576172, "learning_rate": 4.620537805677879e-05, "loss": 9.4677, "step": 11977 }, { "epoch": 0.5050066403861965, "grad_norm": 34.34197235107422, "learning_rate": 4.620459727037072e-05, "loss": 10.2055, "step": 11978 }, { "epoch": 0.5050488015683959, "grad_norm": 12.086565971374512, "learning_rate": 4.6203816483962646e-05, "loss": 10.1025, "step": 11979 }, { "epoch": 0.5050909627505955, "grad_norm": 50.25096893310547, "learning_rate": 4.6203035697554584e-05, "loss": 9.6461, "step": 11980 }, { "epoch": 0.5051331239327951, "grad_norm": 25.609132766723633, "learning_rate": 4.620225491114651e-05, "loss": 11.0758, "step": 11981 }, { "epoch": 0.5051752851149947, "grad_norm": 13.039712905883789, "learning_rate": 4.620147412473844e-05, "loss": 10.7566, "step": 11982 }, { "epoch": 0.5052174462971942, "grad_norm": 23.76917266845703, "learning_rate": 4.620069333833037e-05, "loss": 9.126, "step": 11983 }, { "epoch": 0.5052596074793937, "grad_norm": 20.046003341674805, "learning_rate": 4.6199912551922295e-05, "loss": 8.8912, "step": 11984 }, { "epoch": 0.5053017686615933, "grad_norm": 52.68342590332031, "learning_rate": 4.619913176551423e-05, "loss": 9.3455, "step": 11985 }, { "epoch": 0.5053439298437928, "grad_norm": 20.763504028320312, "learning_rate": 4.619835097910616e-05, "loss": 9.2026, "step": 11986 }, { "epoch": 0.5053860910259924, "grad_norm": 75.51754760742188, "learning_rate": 4.619757019269809e-05, "loss": 10.2513, "step": 11987 }, { "epoch": 0.5054282522081919, "grad_norm": 44.23623275756836, "learning_rate": 4.619678940629002e-05, "loss": 9.8757, "step": 11988 }, { "epoch": 0.5054704133903914, "grad_norm": 47.369876861572266, "learning_rate": 4.6196008619881945e-05, "loss": 9.7418, "step": 11989 }, { "epoch": 0.505512574572591, "grad_norm": 26.74108123779297, "learning_rate": 4.6195227833473876e-05, "loss": 9.5446, "step": 11990 }, { "epoch": 0.5055547357547906, "grad_norm": 13.92699909210205, "learning_rate": 4.619444704706581e-05, "loss": 9.8134, "step": 11991 }, { "epoch": 0.5055968969369901, "grad_norm": 13.641356468200684, "learning_rate": 4.619366626065774e-05, "loss": 9.9089, "step": 11992 }, { "epoch": 0.5056390581191896, "grad_norm": 8.331847190856934, "learning_rate": 4.619288547424966e-05, "loss": 10.8251, "step": 11993 }, { "epoch": 0.5056812193013892, "grad_norm": 11.9741792678833, "learning_rate": 4.6192104687841594e-05, "loss": 9.3094, "step": 11994 }, { "epoch": 0.5057233804835888, "grad_norm": 115.49524688720703, "learning_rate": 4.6191323901433525e-05, "loss": 9.6363, "step": 11995 }, { "epoch": 0.5057655416657884, "grad_norm": 11.967494010925293, "learning_rate": 4.619054311502546e-05, "loss": 10.2629, "step": 11996 }, { "epoch": 0.5058077028479878, "grad_norm": 19.562358856201172, "learning_rate": 4.618976232861739e-05, "loss": 8.8552, "step": 11997 }, { "epoch": 0.5058498640301874, "grad_norm": 12.940190315246582, "learning_rate": 4.618898154220931e-05, "loss": 10.1421, "step": 11998 }, { "epoch": 0.505892025212387, "grad_norm": 9.647469520568848, "learning_rate": 4.618820075580125e-05, "loss": 10.2806, "step": 11999 }, { "epoch": 0.5059341863945865, "grad_norm": 10.703936576843262, "learning_rate": 4.6187419969393175e-05, "loss": 11.3255, "step": 12000 }, { "epoch": 0.505976347576786, "grad_norm": 15.436309814453125, "learning_rate": 4.6186639182985106e-05, "loss": 9.8569, "step": 12001 }, { "epoch": 0.5060185087589856, "grad_norm": 11.603365898132324, "learning_rate": 4.618585839657704e-05, "loss": 9.7978, "step": 12002 }, { "epoch": 0.5060606699411851, "grad_norm": 16.63055992126465, "learning_rate": 4.618507761016896e-05, "loss": 9.3272, "step": 12003 }, { "epoch": 0.5061028311233847, "grad_norm": 19.383752822875977, "learning_rate": 4.618429682376089e-05, "loss": 9.5584, "step": 12004 }, { "epoch": 0.5061449923055843, "grad_norm": 44.89236068725586, "learning_rate": 4.6183516037352824e-05, "loss": 9.3771, "step": 12005 }, { "epoch": 0.5061871534877838, "grad_norm": 32.124542236328125, "learning_rate": 4.6182735250944756e-05, "loss": 9.0982, "step": 12006 }, { "epoch": 0.5062293146699833, "grad_norm": 16.692773818969727, "learning_rate": 4.618195446453668e-05, "loss": 10.1156, "step": 12007 }, { "epoch": 0.5062714758521829, "grad_norm": 40.48482131958008, "learning_rate": 4.618117367812861e-05, "loss": 9.7429, "step": 12008 }, { "epoch": 0.5063136370343825, "grad_norm": 23.67383575439453, "learning_rate": 4.618039289172054e-05, "loss": 9.6353, "step": 12009 }, { "epoch": 0.5063557982165819, "grad_norm": 24.38941192626953, "learning_rate": 4.6179612105312474e-05, "loss": 10.0248, "step": 12010 }, { "epoch": 0.5063979593987815, "grad_norm": 26.233240127563477, "learning_rate": 4.6178831318904405e-05, "loss": 9.1002, "step": 12011 }, { "epoch": 0.5064401205809811, "grad_norm": 75.48461151123047, "learning_rate": 4.617805053249633e-05, "loss": 10.9839, "step": 12012 }, { "epoch": 0.5064822817631807, "grad_norm": 34.2131462097168, "learning_rate": 4.617726974608826e-05, "loss": 10.0273, "step": 12013 }, { "epoch": 0.5065244429453802, "grad_norm": 69.73307037353516, "learning_rate": 4.617648895968019e-05, "loss": 8.9988, "step": 12014 }, { "epoch": 0.5065666041275797, "grad_norm": 126.03901672363281, "learning_rate": 4.617570817327212e-05, "loss": 8.9513, "step": 12015 }, { "epoch": 0.5066087653097793, "grad_norm": 56.38081741333008, "learning_rate": 4.6174927386864055e-05, "loss": 9.823, "step": 12016 }, { "epoch": 0.5066509264919788, "grad_norm": 66.16698455810547, "learning_rate": 4.617414660045598e-05, "loss": 11.3806, "step": 12017 }, { "epoch": 0.5066930876741784, "grad_norm": 100.94764709472656, "learning_rate": 4.617336581404791e-05, "loss": 9.7386, "step": 12018 }, { "epoch": 0.5067352488563779, "grad_norm": 28.11998176574707, "learning_rate": 4.617258502763984e-05, "loss": 8.6045, "step": 12019 }, { "epoch": 0.5067774100385775, "grad_norm": 30.46010971069336, "learning_rate": 4.617180424123177e-05, "loss": 8.8126, "step": 12020 }, { "epoch": 0.506819571220777, "grad_norm": 24.697526931762695, "learning_rate": 4.6171023454823704e-05, "loss": 9.2204, "step": 12021 }, { "epoch": 0.5068617324029766, "grad_norm": 38.0969123840332, "learning_rate": 4.617024266841563e-05, "loss": 9.5249, "step": 12022 }, { "epoch": 0.5069038935851762, "grad_norm": 16.684383392333984, "learning_rate": 4.616946188200756e-05, "loss": 9.6168, "step": 12023 }, { "epoch": 0.5069460547673756, "grad_norm": 31.24737548828125, "learning_rate": 4.616868109559949e-05, "loss": 9.5313, "step": 12024 }, { "epoch": 0.5069882159495752, "grad_norm": 15.40954875946045, "learning_rate": 4.616790030919142e-05, "loss": 9.2019, "step": 12025 }, { "epoch": 0.5070303771317748, "grad_norm": 33.184635162353516, "learning_rate": 4.616711952278335e-05, "loss": 8.9875, "step": 12026 }, { "epoch": 0.5070725383139744, "grad_norm": 12.357611656188965, "learning_rate": 4.616633873637528e-05, "loss": 9.7088, "step": 12027 }, { "epoch": 0.5071146994961738, "grad_norm": 13.589360237121582, "learning_rate": 4.616555794996721e-05, "loss": 9.4782, "step": 12028 }, { "epoch": 0.5071568606783734, "grad_norm": 175.33602905273438, "learning_rate": 4.6164777163559134e-05, "loss": 9.9063, "step": 12029 }, { "epoch": 0.507199021860573, "grad_norm": 41.52195739746094, "learning_rate": 4.616399637715107e-05, "loss": 10.4585, "step": 12030 }, { "epoch": 0.5072411830427725, "grad_norm": 24.08175277709961, "learning_rate": 4.6163215590742997e-05, "loss": 10.1684, "step": 12031 }, { "epoch": 0.5072833442249721, "grad_norm": 18.32440948486328, "learning_rate": 4.616243480433493e-05, "loss": 10.7133, "step": 12032 }, { "epoch": 0.5073255054071716, "grad_norm": 37.86302185058594, "learning_rate": 4.616165401792686e-05, "loss": 9.4039, "step": 12033 }, { "epoch": 0.5073676665893712, "grad_norm": 23.67854881286621, "learning_rate": 4.6160873231518784e-05, "loss": 9.4284, "step": 12034 }, { "epoch": 0.5074098277715707, "grad_norm": 21.176666259765625, "learning_rate": 4.616009244511072e-05, "loss": 9.1564, "step": 12035 }, { "epoch": 0.5074519889537703, "grad_norm": 19.885276794433594, "learning_rate": 4.6159311658702646e-05, "loss": 8.9907, "step": 12036 }, { "epoch": 0.5074941501359698, "grad_norm": 10.189878463745117, "learning_rate": 4.615853087229458e-05, "loss": 9.6541, "step": 12037 }, { "epoch": 0.5075363113181693, "grad_norm": 8.385865211486816, "learning_rate": 4.615775008588651e-05, "loss": 8.9906, "step": 12038 }, { "epoch": 0.5075784725003689, "grad_norm": 18.11686134338379, "learning_rate": 4.615696929947843e-05, "loss": 9.6835, "step": 12039 }, { "epoch": 0.5076206336825685, "grad_norm": 18.2094783782959, "learning_rate": 4.6156188513070364e-05, "loss": 8.8363, "step": 12040 }, { "epoch": 0.5076627948647681, "grad_norm": 38.63359451293945, "learning_rate": 4.6155407726662296e-05, "loss": 9.4649, "step": 12041 }, { "epoch": 0.5077049560469675, "grad_norm": 18.588489532470703, "learning_rate": 4.615462694025423e-05, "loss": 9.6967, "step": 12042 }, { "epoch": 0.5077471172291671, "grad_norm": 17.154245376586914, "learning_rate": 4.615384615384616e-05, "loss": 8.5328, "step": 12043 }, { "epoch": 0.5077892784113667, "grad_norm": 17.72423553466797, "learning_rate": 4.615306536743809e-05, "loss": 9.3888, "step": 12044 }, { "epoch": 0.5078314395935662, "grad_norm": 12.705223083496094, "learning_rate": 4.6152284581030014e-05, "loss": 9.0151, "step": 12045 }, { "epoch": 0.5078736007757657, "grad_norm": 19.345067977905273, "learning_rate": 4.6151503794621945e-05, "loss": 10.5652, "step": 12046 }, { "epoch": 0.5079157619579653, "grad_norm": 14.980422973632812, "learning_rate": 4.6150723008213876e-05, "loss": 8.5089, "step": 12047 }, { "epoch": 0.5079579231401649, "grad_norm": 20.10231590270996, "learning_rate": 4.61499422218058e-05, "loss": 9.6529, "step": 12048 }, { "epoch": 0.5080000843223644, "grad_norm": 10.536091804504395, "learning_rate": 4.614916143539774e-05, "loss": 10.0746, "step": 12049 }, { "epoch": 0.5080422455045639, "grad_norm": 15.932591438293457, "learning_rate": 4.6148380648989663e-05, "loss": 8.6999, "step": 12050 }, { "epoch": 0.5080844066867635, "grad_norm": 11.49415397644043, "learning_rate": 4.6147599862581595e-05, "loss": 9.9473, "step": 12051 }, { "epoch": 0.508126567868963, "grad_norm": 16.061487197875977, "learning_rate": 4.6146819076173526e-05, "loss": 10.8901, "step": 12052 }, { "epoch": 0.5081687290511626, "grad_norm": 13.19579029083252, "learning_rate": 4.614603828976545e-05, "loss": 9.1598, "step": 12053 }, { "epoch": 0.5082108902333622, "grad_norm": 10.881561279296875, "learning_rate": 4.614525750335739e-05, "loss": 8.9445, "step": 12054 }, { "epoch": 0.5082530514155617, "grad_norm": 10.033308029174805, "learning_rate": 4.614447671694931e-05, "loss": 8.8128, "step": 12055 }, { "epoch": 0.5082952125977612, "grad_norm": 15.845145225524902, "learning_rate": 4.6143695930541244e-05, "loss": 10.1958, "step": 12056 }, { "epoch": 0.5083373737799608, "grad_norm": 25.85428810119629, "learning_rate": 4.6142915144133175e-05, "loss": 9.1273, "step": 12057 }, { "epoch": 0.5083795349621604, "grad_norm": 28.815608978271484, "learning_rate": 4.61421343577251e-05, "loss": 8.7817, "step": 12058 }, { "epoch": 0.5084216961443598, "grad_norm": 6.1443023681640625, "learning_rate": 4.614135357131703e-05, "loss": 9.3801, "step": 12059 }, { "epoch": 0.5084638573265594, "grad_norm": 15.544561386108398, "learning_rate": 4.614057278490896e-05, "loss": 8.7257, "step": 12060 }, { "epoch": 0.508506018508759, "grad_norm": 41.56686782836914, "learning_rate": 4.6139791998500894e-05, "loss": 9.7112, "step": 12061 }, { "epoch": 0.5085481796909586, "grad_norm": 18.566205978393555, "learning_rate": 4.613901121209282e-05, "loss": 8.6116, "step": 12062 }, { "epoch": 0.5085903408731581, "grad_norm": 16.907371520996094, "learning_rate": 4.6138230425684756e-05, "loss": 10.0492, "step": 12063 }, { "epoch": 0.5086325020553576, "grad_norm": 51.190757751464844, "learning_rate": 4.613744963927668e-05, "loss": 8.9344, "step": 12064 }, { "epoch": 0.5086746632375572, "grad_norm": 139.91476440429688, "learning_rate": 4.613666885286861e-05, "loss": 9.0532, "step": 12065 }, { "epoch": 0.5087168244197567, "grad_norm": 19.602251052856445, "learning_rate": 4.613588806646054e-05, "loss": 9.1818, "step": 12066 }, { "epoch": 0.5087589856019563, "grad_norm": 18.146778106689453, "learning_rate": 4.613510728005247e-05, "loss": 8.6521, "step": 12067 }, { "epoch": 0.5088011467841558, "grad_norm": 32.661895751953125, "learning_rate": 4.6134326493644406e-05, "loss": 8.9024, "step": 12068 }, { "epoch": 0.5088433079663554, "grad_norm": 39.630889892578125, "learning_rate": 4.613354570723633e-05, "loss": 9.202, "step": 12069 }, { "epoch": 0.5088854691485549, "grad_norm": 15.297717094421387, "learning_rate": 4.613276492082826e-05, "loss": 10.1656, "step": 12070 }, { "epoch": 0.5089276303307545, "grad_norm": 12.914249420166016, "learning_rate": 4.613198413442019e-05, "loss": 8.7927, "step": 12071 }, { "epoch": 0.5089697915129541, "grad_norm": 13.944409370422363, "learning_rate": 4.613120334801212e-05, "loss": 8.8655, "step": 12072 }, { "epoch": 0.5090119526951535, "grad_norm": 38.053470611572266, "learning_rate": 4.613042256160405e-05, "loss": 9.6256, "step": 12073 }, { "epoch": 0.5090541138773531, "grad_norm": 20.218923568725586, "learning_rate": 4.612964177519598e-05, "loss": 8.6557, "step": 12074 }, { "epoch": 0.5090962750595527, "grad_norm": 29.17159652709961, "learning_rate": 4.612886098878791e-05, "loss": 9.6359, "step": 12075 }, { "epoch": 0.5091384362417523, "grad_norm": 19.155370712280273, "learning_rate": 4.612808020237984e-05, "loss": 9.555, "step": 12076 }, { "epoch": 0.5091805974239517, "grad_norm": 22.052688598632812, "learning_rate": 4.612729941597177e-05, "loss": 8.5907, "step": 12077 }, { "epoch": 0.5092227586061513, "grad_norm": 17.628782272338867, "learning_rate": 4.61265186295637e-05, "loss": 8.7736, "step": 12078 }, { "epoch": 0.5092649197883509, "grad_norm": 23.757740020751953, "learning_rate": 4.612573784315563e-05, "loss": 8.5488, "step": 12079 }, { "epoch": 0.5093070809705504, "grad_norm": 20.716556549072266, "learning_rate": 4.612495705674756e-05, "loss": 8.8362, "step": 12080 }, { "epoch": 0.50934924215275, "grad_norm": 33.435752868652344, "learning_rate": 4.6124176270339485e-05, "loss": 10.169, "step": 12081 }, { "epoch": 0.5093914033349495, "grad_norm": 12.210338592529297, "learning_rate": 4.612339548393142e-05, "loss": 9.3411, "step": 12082 }, { "epoch": 0.509433564517149, "grad_norm": 16.310638427734375, "learning_rate": 4.612261469752335e-05, "loss": 9.0676, "step": 12083 }, { "epoch": 0.5094757256993486, "grad_norm": 7.041306495666504, "learning_rate": 4.612183391111527e-05, "loss": 9.1129, "step": 12084 }, { "epoch": 0.5095178868815482, "grad_norm": 11.271323204040527, "learning_rate": 4.612105312470721e-05, "loss": 8.871, "step": 12085 }, { "epoch": 0.5095600480637477, "grad_norm": 13.4918212890625, "learning_rate": 4.6120272338299135e-05, "loss": 9.4802, "step": 12086 }, { "epoch": 0.5096022092459472, "grad_norm": 35.40632629394531, "learning_rate": 4.6119491551891066e-05, "loss": 9.0097, "step": 12087 }, { "epoch": 0.5096443704281468, "grad_norm": 11.402470588684082, "learning_rate": 4.6118710765483e-05, "loss": 8.6543, "step": 12088 }, { "epoch": 0.5096865316103464, "grad_norm": 12.695907592773438, "learning_rate": 4.611792997907493e-05, "loss": 8.8133, "step": 12089 }, { "epoch": 0.509728692792546, "grad_norm": 125.58489227294922, "learning_rate": 4.611714919266686e-05, "loss": 8.6984, "step": 12090 }, { "epoch": 0.5097708539747454, "grad_norm": 13.621066093444824, "learning_rate": 4.6116368406258784e-05, "loss": 9.4974, "step": 12091 }, { "epoch": 0.509813015156945, "grad_norm": 12.833900451660156, "learning_rate": 4.6115587619850715e-05, "loss": 9.5053, "step": 12092 }, { "epoch": 0.5098551763391446, "grad_norm": 8.0645112991333, "learning_rate": 4.6114806833442647e-05, "loss": 8.8796, "step": 12093 }, { "epoch": 0.5098973375213441, "grad_norm": 18.655582427978516, "learning_rate": 4.611402604703458e-05, "loss": 10.3243, "step": 12094 }, { "epoch": 0.5099394987035436, "grad_norm": 11.904275894165039, "learning_rate": 4.61132452606265e-05, "loss": 8.9142, "step": 12095 }, { "epoch": 0.5099816598857432, "grad_norm": 13.538291931152344, "learning_rate": 4.6112464474218434e-05, "loss": 9.1091, "step": 12096 }, { "epoch": 0.5100238210679428, "grad_norm": 15.381031036376953, "learning_rate": 4.6111683687810365e-05, "loss": 8.4407, "step": 12097 }, { "epoch": 0.5100659822501423, "grad_norm": 34.88471984863281, "learning_rate": 4.611090290140229e-05, "loss": 8.7523, "step": 12098 }, { "epoch": 0.5101081434323418, "grad_norm": 14.655206680297852, "learning_rate": 4.611012211499423e-05, "loss": 8.8962, "step": 12099 }, { "epoch": 0.5101503046145414, "grad_norm": 45.453067779541016, "learning_rate": 4.610934132858615e-05, "loss": 9.864, "step": 12100 }, { "epoch": 0.5101924657967409, "grad_norm": 12.370018005371094, "learning_rate": 4.610856054217808e-05, "loss": 9.8479, "step": 12101 }, { "epoch": 0.5102346269789405, "grad_norm": 8.371062278747559, "learning_rate": 4.6107779755770014e-05, "loss": 9.2771, "step": 12102 }, { "epoch": 0.5102767881611401, "grad_norm": 17.16974449157715, "learning_rate": 4.610699896936194e-05, "loss": 8.9837, "step": 12103 }, { "epoch": 0.5103189493433395, "grad_norm": 17.208515167236328, "learning_rate": 4.610621818295388e-05, "loss": 8.5921, "step": 12104 }, { "epoch": 0.5103611105255391, "grad_norm": 6.4156174659729, "learning_rate": 4.61054373965458e-05, "loss": 8.2447, "step": 12105 }, { "epoch": 0.5104032717077387, "grad_norm": 8.124101638793945, "learning_rate": 4.610465661013773e-05, "loss": 10.1141, "step": 12106 }, { "epoch": 0.5104454328899383, "grad_norm": 11.083891868591309, "learning_rate": 4.6103875823729664e-05, "loss": 9.9065, "step": 12107 }, { "epoch": 0.5104875940721377, "grad_norm": 24.247175216674805, "learning_rate": 4.6103095037321595e-05, "loss": 8.2105, "step": 12108 }, { "epoch": 0.5105297552543373, "grad_norm": 15.143781661987305, "learning_rate": 4.610231425091352e-05, "loss": 9.5012, "step": 12109 }, { "epoch": 0.5105719164365369, "grad_norm": 29.389543533325195, "learning_rate": 4.610153346450545e-05, "loss": 8.2219, "step": 12110 }, { "epoch": 0.5106140776187365, "grad_norm": 20.82186508178711, "learning_rate": 4.610075267809738e-05, "loss": 9.0745, "step": 12111 }, { "epoch": 0.510656238800936, "grad_norm": 28.90283203125, "learning_rate": 4.6099971891689313e-05, "loss": 8.4602, "step": 12112 }, { "epoch": 0.5106983999831355, "grad_norm": 12.456192016601562, "learning_rate": 4.6099191105281245e-05, "loss": 9.0725, "step": 12113 }, { "epoch": 0.5107405611653351, "grad_norm": 7.793822288513184, "learning_rate": 4.609841031887317e-05, "loss": 9.7428, "step": 12114 }, { "epoch": 0.5107827223475346, "grad_norm": 5.4553751945495605, "learning_rate": 4.60976295324651e-05, "loss": 8.6233, "step": 12115 }, { "epoch": 0.5108248835297342, "grad_norm": 12.485526084899902, "learning_rate": 4.609684874605703e-05, "loss": 8.4846, "step": 12116 }, { "epoch": 0.5108670447119337, "grad_norm": 34.035152435302734, "learning_rate": 4.6096067959648956e-05, "loss": 8.4122, "step": 12117 }, { "epoch": 0.5109092058941332, "grad_norm": 19.459468841552734, "learning_rate": 4.6095287173240894e-05, "loss": 8.1866, "step": 12118 }, { "epoch": 0.5109513670763328, "grad_norm": 49.720333099365234, "learning_rate": 4.609450638683282e-05, "loss": 8.4758, "step": 12119 }, { "epoch": 0.5109935282585324, "grad_norm": 10.1669921875, "learning_rate": 4.609372560042475e-05, "loss": 8.3968, "step": 12120 }, { "epoch": 0.511035689440732, "grad_norm": 20.658802032470703, "learning_rate": 4.609294481401668e-05, "loss": 9.3311, "step": 12121 }, { "epoch": 0.5110778506229314, "grad_norm": 14.454079627990723, "learning_rate": 4.6092164027608606e-05, "loss": 8.3633, "step": 12122 }, { "epoch": 0.511120011805131, "grad_norm": 7.898335933685303, "learning_rate": 4.6091383241200544e-05, "loss": 8.4531, "step": 12123 }, { "epoch": 0.5111621729873306, "grad_norm": 5.959489345550537, "learning_rate": 4.609060245479247e-05, "loss": 8.332, "step": 12124 }, { "epoch": 0.5112043341695302, "grad_norm": 18.719032287597656, "learning_rate": 4.60898216683844e-05, "loss": 8.5342, "step": 12125 }, { "epoch": 0.5112464953517296, "grad_norm": 20.947660446166992, "learning_rate": 4.608904088197633e-05, "loss": 8.1037, "step": 12126 }, { "epoch": 0.5112886565339292, "grad_norm": 12.3997220993042, "learning_rate": 4.608826009556826e-05, "loss": 8.9423, "step": 12127 }, { "epoch": 0.5113308177161288, "grad_norm": 6.791267395019531, "learning_rate": 4.6087479309160186e-05, "loss": 8.6535, "step": 12128 }, { "epoch": 0.5113729788983283, "grad_norm": 11.640296936035156, "learning_rate": 4.608669852275212e-05, "loss": 9.2066, "step": 12129 }, { "epoch": 0.5114151400805279, "grad_norm": 8.932181358337402, "learning_rate": 4.608591773634405e-05, "loss": 8.807, "step": 12130 }, { "epoch": 0.5114573012627274, "grad_norm": 16.047176361083984, "learning_rate": 4.6085136949935973e-05, "loss": 7.89, "step": 12131 }, { "epoch": 0.5114994624449269, "grad_norm": 9.72645092010498, "learning_rate": 4.608435616352791e-05, "loss": 8.5362, "step": 12132 }, { "epoch": 0.5115416236271265, "grad_norm": 5.4165120124816895, "learning_rate": 4.6083575377119836e-05, "loss": 8.2559, "step": 12133 }, { "epoch": 0.5115837848093261, "grad_norm": 12.540751457214355, "learning_rate": 4.608279459071177e-05, "loss": 8.5432, "step": 12134 }, { "epoch": 0.5116259459915256, "grad_norm": 19.659324645996094, "learning_rate": 4.60820138043037e-05, "loss": 8.3976, "step": 12135 }, { "epoch": 0.5116681071737251, "grad_norm": 7.647658824920654, "learning_rate": 4.608123301789562e-05, "loss": 8.2765, "step": 12136 }, { "epoch": 0.5117102683559247, "grad_norm": 9.796798706054688, "learning_rate": 4.608045223148756e-05, "loss": 9.8858, "step": 12137 }, { "epoch": 0.5117524295381243, "grad_norm": 8.224590301513672, "learning_rate": 4.6079671445079486e-05, "loss": 8.9272, "step": 12138 }, { "epoch": 0.5117945907203238, "grad_norm": 8.57607364654541, "learning_rate": 4.607889065867142e-05, "loss": 8.767, "step": 12139 }, { "epoch": 0.5118367519025233, "grad_norm": 10.272390365600586, "learning_rate": 4.607810987226335e-05, "loss": 8.8049, "step": 12140 }, { "epoch": 0.5118789130847229, "grad_norm": 10.915438652038574, "learning_rate": 4.607732908585527e-05, "loss": 11.4006, "step": 12141 }, { "epoch": 0.5119210742669225, "grad_norm": 9.698220252990723, "learning_rate": 4.6076548299447204e-05, "loss": 9.2016, "step": 12142 }, { "epoch": 0.511963235449122, "grad_norm": 5.248912811279297, "learning_rate": 4.6075767513039135e-05, "loss": 8.9311, "step": 12143 }, { "epoch": 0.5120053966313215, "grad_norm": 8.401043891906738, "learning_rate": 4.6074986726631066e-05, "loss": 7.6391, "step": 12144 }, { "epoch": 0.5120475578135211, "grad_norm": 7.920811653137207, "learning_rate": 4.6074205940223e-05, "loss": 9.1784, "step": 12145 }, { "epoch": 0.5120897189957206, "grad_norm": 5.383946418762207, "learning_rate": 4.607342515381492e-05, "loss": 9.0073, "step": 12146 }, { "epoch": 0.5121318801779202, "grad_norm": 5.025763511657715, "learning_rate": 4.607264436740685e-05, "loss": 7.9653, "step": 12147 }, { "epoch": 0.5121740413601198, "grad_norm": 3.9825832843780518, "learning_rate": 4.6071863580998785e-05, "loss": 9.0978, "step": 12148 }, { "epoch": 0.5122162025423193, "grad_norm": 6.087832927703857, "learning_rate": 4.6071082794590716e-05, "loss": 7.9977, "step": 12149 }, { "epoch": 0.5122583637245188, "grad_norm": 9.727324485778809, "learning_rate": 4.607030200818264e-05, "loss": 9.9851, "step": 12150 }, { "epoch": 0.5123005249067184, "grad_norm": 7.689426898956299, "learning_rate": 4.606952122177458e-05, "loss": 8.0505, "step": 12151 }, { "epoch": 0.512342686088918, "grad_norm": 4.41148567199707, "learning_rate": 4.60687404353665e-05, "loss": 8.7354, "step": 12152 }, { "epoch": 0.5123848472711174, "grad_norm": 7.332434177398682, "learning_rate": 4.6067959648958434e-05, "loss": 7.7646, "step": 12153 }, { "epoch": 0.512427008453317, "grad_norm": 5.995861053466797, "learning_rate": 4.6067178862550365e-05, "loss": 9.6642, "step": 12154 }, { "epoch": 0.5124691696355166, "grad_norm": 6.409979343414307, "learning_rate": 4.606639807614229e-05, "loss": 9.5658, "step": 12155 }, { "epoch": 0.5125113308177162, "grad_norm": 4.037621021270752, "learning_rate": 4.606561728973422e-05, "loss": 8.0194, "step": 12156 }, { "epoch": 0.5125534919999156, "grad_norm": 3.584681272506714, "learning_rate": 4.606483650332615e-05, "loss": 7.8355, "step": 12157 }, { "epoch": 0.5125956531821152, "grad_norm": 5.552277565002441, "learning_rate": 4.6064055716918084e-05, "loss": 8.6764, "step": 12158 }, { "epoch": 0.5126378143643148, "grad_norm": 8.070819854736328, "learning_rate": 4.6063274930510015e-05, "loss": 7.7002, "step": 12159 }, { "epoch": 0.5126799755465143, "grad_norm": 4.86931848526001, "learning_rate": 4.606249414410194e-05, "loss": 7.8545, "step": 12160 }, { "epoch": 0.5127221367287139, "grad_norm": 4.546297550201416, "learning_rate": 4.606171335769387e-05, "loss": 8.2012, "step": 12161 }, { "epoch": 0.5127642979109134, "grad_norm": 6.025990962982178, "learning_rate": 4.60609325712858e-05, "loss": 7.4777, "step": 12162 }, { "epoch": 0.512806459093113, "grad_norm": 5.53622579574585, "learning_rate": 4.606015178487773e-05, "loss": 8.4556, "step": 12163 }, { "epoch": 0.5128486202753125, "grad_norm": 6.668311595916748, "learning_rate": 4.605937099846966e-05, "loss": 9.0201, "step": 12164 }, { "epoch": 0.5128907814575121, "grad_norm": 5.595378398895264, "learning_rate": 4.605859021206159e-05, "loss": 7.4941, "step": 12165 }, { "epoch": 0.5129329426397116, "grad_norm": 5.891815662384033, "learning_rate": 4.605780942565352e-05, "loss": 9.5181, "step": 12166 }, { "epoch": 0.5129751038219111, "grad_norm": 3.5774459838867188, "learning_rate": 4.6057028639245445e-05, "loss": 8.5609, "step": 12167 }, { "epoch": 0.5130172650041107, "grad_norm": 3.695953607559204, "learning_rate": 4.605624785283738e-05, "loss": 7.9423, "step": 12168 }, { "epoch": 0.5130594261863103, "grad_norm": 4.011291980743408, "learning_rate": 4.605546706642931e-05, "loss": 8.0779, "step": 12169 }, { "epoch": 0.5131015873685099, "grad_norm": 4.795922756195068, "learning_rate": 4.6054686280021245e-05, "loss": 7.9202, "step": 12170 }, { "epoch": 0.5131437485507093, "grad_norm": 3.72735857963562, "learning_rate": 4.605390549361317e-05, "loss": 7.9749, "step": 12171 }, { "epoch": 0.5131859097329089, "grad_norm": 5.143901348114014, "learning_rate": 4.6053124707205094e-05, "loss": 7.6646, "step": 12172 }, { "epoch": 0.5132280709151085, "grad_norm": 4.7907562255859375, "learning_rate": 4.605234392079703e-05, "loss": 7.7419, "step": 12173 }, { "epoch": 0.513270232097308, "grad_norm": 4.01146936416626, "learning_rate": 4.605156313438896e-05, "loss": 7.7316, "step": 12174 }, { "epoch": 0.5133123932795075, "grad_norm": 4.561883926391602, "learning_rate": 4.605078234798089e-05, "loss": 7.9623, "step": 12175 }, { "epoch": 0.5133545544617071, "grad_norm": 3.6055519580841064, "learning_rate": 4.605000156157282e-05, "loss": 8.1277, "step": 12176 }, { "epoch": 0.5133967156439067, "grad_norm": 8.034016609191895, "learning_rate": 4.604922077516475e-05, "loss": 9.3416, "step": 12177 }, { "epoch": 0.5134388768261062, "grad_norm": 5.243134498596191, "learning_rate": 4.6048439988756675e-05, "loss": 8.3853, "step": 12178 }, { "epoch": 0.5134810380083058, "grad_norm": 5.129549503326416, "learning_rate": 4.6047659202348606e-05, "loss": 8.7664, "step": 12179 }, { "epoch": 0.5135231991905053, "grad_norm": 5.102325916290283, "learning_rate": 4.604687841594054e-05, "loss": 8.3036, "step": 12180 }, { "epoch": 0.5135653603727048, "grad_norm": 5.313324928283691, "learning_rate": 4.604609762953247e-05, "loss": 7.8624, "step": 12181 }, { "epoch": 0.5136075215549044, "grad_norm": 4.779802322387695, "learning_rate": 4.60453168431244e-05, "loss": 7.8307, "step": 12182 }, { "epoch": 0.513649682737104, "grad_norm": 4.578166484832764, "learning_rate": 4.6044536056716324e-05, "loss": 8.0587, "step": 12183 }, { "epoch": 0.5136918439193034, "grad_norm": 3.905885696411133, "learning_rate": 4.6043755270308256e-05, "loss": 8.4022, "step": 12184 }, { "epoch": 0.513734005101503, "grad_norm": 3.111802816390991, "learning_rate": 4.604297448390019e-05, "loss": 7.9029, "step": 12185 }, { "epoch": 0.5137761662837026, "grad_norm": 3.6796700954437256, "learning_rate": 4.604219369749211e-05, "loss": 8.1933, "step": 12186 }, { "epoch": 0.5138183274659022, "grad_norm": 3.252028226852417, "learning_rate": 4.604141291108405e-05, "loss": 8.2262, "step": 12187 }, { "epoch": 0.5138604886481017, "grad_norm": 6.001636505126953, "learning_rate": 4.6040632124675974e-05, "loss": 9.223, "step": 12188 }, { "epoch": 0.5139026498303012, "grad_norm": 3.859893798828125, "learning_rate": 4.6039851338267905e-05, "loss": 8.2324, "step": 12189 }, { "epoch": 0.5139448110125008, "grad_norm": 3.716911554336548, "learning_rate": 4.6039070551859837e-05, "loss": 7.3577, "step": 12190 }, { "epoch": 0.5139869721947004, "grad_norm": 4.803835868835449, "learning_rate": 4.603828976545176e-05, "loss": 7.5536, "step": 12191 }, { "epoch": 0.5140291333768999, "grad_norm": 3.792598009109497, "learning_rate": 4.60375089790437e-05, "loss": 7.6963, "step": 12192 }, { "epoch": 0.5140712945590994, "grad_norm": 7.588057518005371, "learning_rate": 4.6036728192635624e-05, "loss": 7.6235, "step": 12193 }, { "epoch": 0.514113455741299, "grad_norm": 4.725024700164795, "learning_rate": 4.6035947406227555e-05, "loss": 8.7753, "step": 12194 }, { "epoch": 0.5141556169234985, "grad_norm": 3.747875452041626, "learning_rate": 4.6035166619819486e-05, "loss": 7.3977, "step": 12195 }, { "epoch": 0.5141977781056981, "grad_norm": 5.336648464202881, "learning_rate": 4.603438583341142e-05, "loss": 8.2422, "step": 12196 }, { "epoch": 0.5142399392878977, "grad_norm": 3.3609132766723633, "learning_rate": 4.603360504700334e-05, "loss": 7.621, "step": 12197 }, { "epoch": 0.5142821004700971, "grad_norm": 3.4106013774871826, "learning_rate": 4.603282426059527e-05, "loss": 7.4795, "step": 12198 }, { "epoch": 0.5143242616522967, "grad_norm": 3.6285316944122314, "learning_rate": 4.6032043474187204e-05, "loss": 8.1939, "step": 12199 }, { "epoch": 0.5143664228344963, "grad_norm": 5.539993762969971, "learning_rate": 4.603126268777913e-05, "loss": 8.2012, "step": 12200 }, { "epoch": 0.5144085840166959, "grad_norm": 2.725660562515259, "learning_rate": 4.603048190137107e-05, "loss": 7.7036, "step": 12201 }, { "epoch": 0.5144507451988953, "grad_norm": 6.125186443328857, "learning_rate": 4.602970111496299e-05, "loss": 8.4994, "step": 12202 }, { "epoch": 0.5144929063810949, "grad_norm": 6.21065092086792, "learning_rate": 4.602892032855492e-05, "loss": 8.5589, "step": 12203 }, { "epoch": 0.5145350675632945, "grad_norm": 10.446921348571777, "learning_rate": 4.6028139542146854e-05, "loss": 10.0481, "step": 12204 }, { "epoch": 0.514577228745494, "grad_norm": 3.319240093231201, "learning_rate": 4.602735875573878e-05, "loss": 7.3383, "step": 12205 }, { "epoch": 0.5146193899276936, "grad_norm": 3.4099667072296143, "learning_rate": 4.6026577969330716e-05, "loss": 7.5312, "step": 12206 }, { "epoch": 0.5146615511098931, "grad_norm": 3.1566202640533447, "learning_rate": 4.602579718292264e-05, "loss": 7.2688, "step": 12207 }, { "epoch": 0.5147037122920927, "grad_norm": 3.484617233276367, "learning_rate": 4.602501639651457e-05, "loss": 8.0376, "step": 12208 }, { "epoch": 0.5147458734742922, "grad_norm": 3.8131253719329834, "learning_rate": 4.60242356101065e-05, "loss": 8.1101, "step": 12209 }, { "epoch": 0.5147880346564918, "grad_norm": 5.1643171310424805, "learning_rate": 4.602345482369843e-05, "loss": 8.6917, "step": 12210 }, { "epoch": 0.5148301958386913, "grad_norm": 3.992384672164917, "learning_rate": 4.602267403729036e-05, "loss": 7.6262, "step": 12211 }, { "epoch": 0.5148723570208908, "grad_norm": 2.4785664081573486, "learning_rate": 4.602189325088229e-05, "loss": 7.4028, "step": 12212 }, { "epoch": 0.5149145182030904, "grad_norm": 3.50683331489563, "learning_rate": 4.602111246447422e-05, "loss": 7.4568, "step": 12213 }, { "epoch": 0.51495667938529, "grad_norm": 3.4245293140411377, "learning_rate": 4.602033167806615e-05, "loss": 7.5167, "step": 12214 }, { "epoch": 0.5149988405674895, "grad_norm": 4.267003536224365, "learning_rate": 4.6019550891658084e-05, "loss": 7.3631, "step": 12215 }, { "epoch": 0.515041001749689, "grad_norm": 4.980077266693115, "learning_rate": 4.601877010525001e-05, "loss": 7.8349, "step": 12216 }, { "epoch": 0.5150831629318886, "grad_norm": 3.9409122467041016, "learning_rate": 4.601798931884194e-05, "loss": 7.2527, "step": 12217 }, { "epoch": 0.5151253241140882, "grad_norm": 9.081912994384766, "learning_rate": 4.601720853243387e-05, "loss": 9.1114, "step": 12218 }, { "epoch": 0.5151674852962878, "grad_norm": 3.8371191024780273, "learning_rate": 4.6016427746025796e-05, "loss": 7.0449, "step": 12219 }, { "epoch": 0.5152096464784872, "grad_norm": 2.956045389175415, "learning_rate": 4.6015646959617734e-05, "loss": 7.4338, "step": 12220 }, { "epoch": 0.5152518076606868, "grad_norm": 2.6118602752685547, "learning_rate": 4.601486617320966e-05, "loss": 7.8849, "step": 12221 }, { "epoch": 0.5152939688428864, "grad_norm": 2.8927419185638428, "learning_rate": 4.601408538680159e-05, "loss": 7.1505, "step": 12222 }, { "epoch": 0.5153361300250859, "grad_norm": 3.3106420040130615, "learning_rate": 4.601330460039352e-05, "loss": 7.1308, "step": 12223 }, { "epoch": 0.5153782912072854, "grad_norm": 5.919384956359863, "learning_rate": 4.6012523813985445e-05, "loss": 8.3059, "step": 12224 }, { "epoch": 0.515420452389485, "grad_norm": 2.9762091636657715, "learning_rate": 4.601174302757738e-05, "loss": 7.247, "step": 12225 }, { "epoch": 0.5154626135716845, "grad_norm": 3.222151517868042, "learning_rate": 4.601096224116931e-05, "loss": 7.2565, "step": 12226 }, { "epoch": 0.5155047747538841, "grad_norm": 3.0749850273132324, "learning_rate": 4.601018145476124e-05, "loss": 7.8541, "step": 12227 }, { "epoch": 0.5155469359360837, "grad_norm": 3.2397661209106445, "learning_rate": 4.600940066835317e-05, "loss": 7.3567, "step": 12228 }, { "epoch": 0.5155890971182832, "grad_norm": 2.3673255443573, "learning_rate": 4.6008619881945095e-05, "loss": 7.3314, "step": 12229 }, { "epoch": 0.5156312583004827, "grad_norm": 4.584691047668457, "learning_rate": 4.6007839095537026e-05, "loss": 8.5146, "step": 12230 }, { "epoch": 0.5156734194826823, "grad_norm": 2.47729754447937, "learning_rate": 4.600705830912896e-05, "loss": 7.3965, "step": 12231 }, { "epoch": 0.5157155806648819, "grad_norm": 8.708706855773926, "learning_rate": 4.600627752272089e-05, "loss": 9.4713, "step": 12232 }, { "epoch": 0.5157577418470813, "grad_norm": 2.250269651412964, "learning_rate": 4.600549673631281e-05, "loss": 7.356, "step": 12233 }, { "epoch": 0.5157999030292809, "grad_norm": 2.450016736984253, "learning_rate": 4.6004715949904744e-05, "loss": 7.2277, "step": 12234 }, { "epoch": 0.5158420642114805, "grad_norm": 2.183210611343384, "learning_rate": 4.6003935163496675e-05, "loss": 7.2064, "step": 12235 }, { "epoch": 0.5158842253936801, "grad_norm": 4.045843124389648, "learning_rate": 4.60031543770886e-05, "loss": 7.9862, "step": 12236 }, { "epoch": 0.5159263865758796, "grad_norm": 1.9662046432495117, "learning_rate": 4.600237359068054e-05, "loss": 7.21, "step": 12237 }, { "epoch": 0.5159685477580791, "grad_norm": 2.404005527496338, "learning_rate": 4.600159280427246e-05, "loss": 7.6199, "step": 12238 }, { "epoch": 0.5160107089402787, "grad_norm": 7.388726234436035, "learning_rate": 4.60008120178644e-05, "loss": 8.1148, "step": 12239 }, { "epoch": 0.5160528701224782, "grad_norm": 4.641256332397461, "learning_rate": 4.6000031231456325e-05, "loss": 8.8443, "step": 12240 }, { "epoch": 0.5160950313046778, "grad_norm": 5.781316757202148, "learning_rate": 4.5999250445048256e-05, "loss": 8.4153, "step": 12241 }, { "epoch": 0.5161371924868773, "grad_norm": 2.676616668701172, "learning_rate": 4.599846965864019e-05, "loss": 7.7374, "step": 12242 }, { "epoch": 0.5161793536690769, "grad_norm": 5.190017223358154, "learning_rate": 4.599768887223211e-05, "loss": 8.6459, "step": 12243 }, { "epoch": 0.5162215148512764, "grad_norm": 5.271056652069092, "learning_rate": 4.599690808582404e-05, "loss": 7.2684, "step": 12244 }, { "epoch": 0.516263676033476, "grad_norm": 3.954702615737915, "learning_rate": 4.5996127299415974e-05, "loss": 7.0883, "step": 12245 }, { "epoch": 0.5163058372156756, "grad_norm": 2.272444486618042, "learning_rate": 4.5995346513007906e-05, "loss": 7.2998, "step": 12246 }, { "epoch": 0.516347998397875, "grad_norm": 3.492208957672119, "learning_rate": 4.599456572659983e-05, "loss": 7.2308, "step": 12247 }, { "epoch": 0.5163901595800746, "grad_norm": 3.644787311553955, "learning_rate": 4.599378494019176e-05, "loss": 7.4569, "step": 12248 }, { "epoch": 0.5164323207622742, "grad_norm": 4.179746627807617, "learning_rate": 4.599300415378369e-05, "loss": 7.736, "step": 12249 }, { "epoch": 0.5164744819444738, "grad_norm": 2.620241403579712, "learning_rate": 4.5992223367375624e-05, "loss": 7.4356, "step": 12250 }, { "epoch": 0.5165166431266732, "grad_norm": 3.834944248199463, "learning_rate": 4.5991442580967555e-05, "loss": 8.6711, "step": 12251 }, { "epoch": 0.5165588043088728, "grad_norm": 2.2233104705810547, "learning_rate": 4.599066179455948e-05, "loss": 7.2461, "step": 12252 }, { "epoch": 0.5166009654910724, "grad_norm": 4.64019250869751, "learning_rate": 4.598988100815141e-05, "loss": 8.0324, "step": 12253 }, { "epoch": 0.516643126673272, "grad_norm": 6.682190895080566, "learning_rate": 4.598910022174334e-05, "loss": 9.3788, "step": 12254 }, { "epoch": 0.5166852878554715, "grad_norm": 2.694272994995117, "learning_rate": 4.598831943533527e-05, "loss": 7.7687, "step": 12255 }, { "epoch": 0.516727449037671, "grad_norm": 4.774729251861572, "learning_rate": 4.5987538648927205e-05, "loss": 7.9684, "step": 12256 }, { "epoch": 0.5167696102198706, "grad_norm": 5.211630344390869, "learning_rate": 4.598675786251913e-05, "loss": 8.8719, "step": 12257 }, { "epoch": 0.5168117714020701, "grad_norm": 4.430159568786621, "learning_rate": 4.598597707611106e-05, "loss": 7.8309, "step": 12258 }, { "epoch": 0.5168539325842697, "grad_norm": 3.9097952842712402, "learning_rate": 4.598519628970299e-05, "loss": 6.9419, "step": 12259 }, { "epoch": 0.5168960937664692, "grad_norm": 3.8118667602539062, "learning_rate": 4.598441550329492e-05, "loss": 7.4689, "step": 12260 }, { "epoch": 0.5169382549486687, "grad_norm": 3.0551271438598633, "learning_rate": 4.5983634716886854e-05, "loss": 7.941, "step": 12261 }, { "epoch": 0.5169804161308683, "grad_norm": 3.062378406524658, "learning_rate": 4.598285393047878e-05, "loss": 7.8566, "step": 12262 }, { "epoch": 0.5170225773130679, "grad_norm": 2.955275774002075, "learning_rate": 4.598207314407071e-05, "loss": 7.1275, "step": 12263 }, { "epoch": 0.5170647384952675, "grad_norm": 3.396113634109497, "learning_rate": 4.598129235766264e-05, "loss": 7.9515, "step": 12264 }, { "epoch": 0.5171068996774669, "grad_norm": 4.322360992431641, "learning_rate": 4.598051157125457e-05, "loss": 7.7597, "step": 12265 }, { "epoch": 0.5171490608596665, "grad_norm": 2.888270378112793, "learning_rate": 4.59797307848465e-05, "loss": 7.5286, "step": 12266 }, { "epoch": 0.5171912220418661, "grad_norm": 2.914425849914551, "learning_rate": 4.597894999843843e-05, "loss": 7.914, "step": 12267 }, { "epoch": 0.5172333832240656, "grad_norm": 3.3079330921173096, "learning_rate": 4.597816921203036e-05, "loss": 7.8593, "step": 12268 }, { "epoch": 0.5172755444062651, "grad_norm": 2.3549299240112305, "learning_rate": 4.5977388425622284e-05, "loss": 7.2396, "step": 12269 }, { "epoch": 0.5173177055884647, "grad_norm": 2.6681535243988037, "learning_rate": 4.597660763921422e-05, "loss": 7.0005, "step": 12270 }, { "epoch": 0.5173598667706643, "grad_norm": 3.3569223880767822, "learning_rate": 4.5975826852806147e-05, "loss": 8.6223, "step": 12271 }, { "epoch": 0.5174020279528638, "grad_norm": 3.48812198638916, "learning_rate": 4.597504606639808e-05, "loss": 8.5686, "step": 12272 }, { "epoch": 0.5174441891350633, "grad_norm": 5.931090831756592, "learning_rate": 4.597426527999001e-05, "loss": 7.3268, "step": 12273 }, { "epoch": 0.5174863503172629, "grad_norm": 3.196235418319702, "learning_rate": 4.5973484493581934e-05, "loss": 7.1149, "step": 12274 }, { "epoch": 0.5175285114994624, "grad_norm": 4.4809746742248535, "learning_rate": 4.597270370717387e-05, "loss": 7.8118, "step": 12275 }, { "epoch": 0.517570672681662, "grad_norm": 3.7758748531341553, "learning_rate": 4.5971922920765796e-05, "loss": 8.1802, "step": 12276 }, { "epoch": 0.5176128338638616, "grad_norm": 2.3593108654022217, "learning_rate": 4.597114213435773e-05, "loss": 7.5321, "step": 12277 }, { "epoch": 0.517654995046061, "grad_norm": 3.6703264713287354, "learning_rate": 4.597036134794966e-05, "loss": 8.6682, "step": 12278 }, { "epoch": 0.5176971562282606, "grad_norm": 2.626842498779297, "learning_rate": 4.596958056154158e-05, "loss": 7.8102, "step": 12279 }, { "epoch": 0.5177393174104602, "grad_norm": 2.80073618888855, "learning_rate": 4.5968799775133514e-05, "loss": 7.1641, "step": 12280 }, { "epoch": 0.5177814785926598, "grad_norm": 9.986783981323242, "learning_rate": 4.5968018988725446e-05, "loss": 8.6408, "step": 12281 }, { "epoch": 0.5178236397748592, "grad_norm": 5.208540439605713, "learning_rate": 4.596723820231738e-05, "loss": 8.5394, "step": 12282 }, { "epoch": 0.5178658009570588, "grad_norm": 5.356685161590576, "learning_rate": 4.596645741590931e-05, "loss": 8.0516, "step": 12283 }, { "epoch": 0.5179079621392584, "grad_norm": 7.409824371337891, "learning_rate": 4.596567662950124e-05, "loss": 8.3981, "step": 12284 }, { "epoch": 0.517950123321458, "grad_norm": 2.4379351139068604, "learning_rate": 4.5964895843093164e-05, "loss": 7.0544, "step": 12285 }, { "epoch": 0.5179922845036575, "grad_norm": 4.8663649559021, "learning_rate": 4.5964115056685095e-05, "loss": 7.368, "step": 12286 }, { "epoch": 0.518034445685857, "grad_norm": 2.835602045059204, "learning_rate": 4.5963334270277026e-05, "loss": 6.9979, "step": 12287 }, { "epoch": 0.5180766068680566, "grad_norm": 5.301723003387451, "learning_rate": 4.596255348386895e-05, "loss": 7.3934, "step": 12288 }, { "epoch": 0.5181187680502561, "grad_norm": 5.657270431518555, "learning_rate": 4.596177269746089e-05, "loss": 8.5942, "step": 12289 }, { "epoch": 0.5181609292324557, "grad_norm": 4.371915817260742, "learning_rate": 4.5960991911052813e-05, "loss": 7.0397, "step": 12290 }, { "epoch": 0.5182030904146552, "grad_norm": 2.5445244312286377, "learning_rate": 4.5960211124644745e-05, "loss": 7.1592, "step": 12291 }, { "epoch": 0.5182452515968548, "grad_norm": 2.5387930870056152, "learning_rate": 4.5959430338236676e-05, "loss": 7.7779, "step": 12292 }, { "epoch": 0.5182874127790543, "grad_norm": 2.7323899269104004, "learning_rate": 4.59586495518286e-05, "loss": 7.6686, "step": 12293 }, { "epoch": 0.5183295739612539, "grad_norm": 6.244013786315918, "learning_rate": 4.595786876542054e-05, "loss": 8.364, "step": 12294 }, { "epoch": 0.5183717351434535, "grad_norm": 2.793750524520874, "learning_rate": 4.595708797901246e-05, "loss": 7.1238, "step": 12295 }, { "epoch": 0.5184138963256529, "grad_norm": 3.455676794052124, "learning_rate": 4.5956307192604394e-05, "loss": 7.8712, "step": 12296 }, { "epoch": 0.5184560575078525, "grad_norm": 2.4673309326171875, "learning_rate": 4.5955526406196325e-05, "loss": 6.8605, "step": 12297 }, { "epoch": 0.5184982186900521, "grad_norm": 3.0767390727996826, "learning_rate": 4.595474561978825e-05, "loss": 7.6564, "step": 12298 }, { "epoch": 0.5185403798722517, "grad_norm": 138.37095642089844, "learning_rate": 4.595396483338018e-05, "loss": 7.938, "step": 12299 }, { "epoch": 0.5185825410544511, "grad_norm": 3.098792552947998, "learning_rate": 4.595318404697211e-05, "loss": 6.7501, "step": 12300 }, { "epoch": 0.5186247022366507, "grad_norm": 3.181535482406616, "learning_rate": 4.5952403260564044e-05, "loss": 7.2566, "step": 12301 }, { "epoch": 0.5186668634188503, "grad_norm": 3.6710641384124756, "learning_rate": 4.595162247415597e-05, "loss": 8.1474, "step": 12302 }, { "epoch": 0.5187090246010498, "grad_norm": 1.9619979858398438, "learning_rate": 4.5950841687747906e-05, "loss": 7.0649, "step": 12303 }, { "epoch": 0.5187511857832494, "grad_norm": 2.0433385372161865, "learning_rate": 4.595006090133983e-05, "loss": 7.3162, "step": 12304 }, { "epoch": 0.5187933469654489, "grad_norm": 2.705108642578125, "learning_rate": 4.594928011493176e-05, "loss": 7.4171, "step": 12305 }, { "epoch": 0.5188355081476484, "grad_norm": 2.9685750007629395, "learning_rate": 4.594849932852369e-05, "loss": 7.5603, "step": 12306 }, { "epoch": 0.518877669329848, "grad_norm": 2.782378911972046, "learning_rate": 4.594771854211562e-05, "loss": 7.1269, "step": 12307 }, { "epoch": 0.5189198305120476, "grad_norm": 4.432992458343506, "learning_rate": 4.5946937755707556e-05, "loss": 8.3044, "step": 12308 }, { "epoch": 0.5189619916942471, "grad_norm": 3.6333370208740234, "learning_rate": 4.594615696929948e-05, "loss": 8.2724, "step": 12309 }, { "epoch": 0.5190041528764466, "grad_norm": 6.58012580871582, "learning_rate": 4.594537618289141e-05, "loss": 9.7543, "step": 12310 }, { "epoch": 0.5190463140586462, "grad_norm": 1.7235990762710571, "learning_rate": 4.594459539648334e-05, "loss": 7.178, "step": 12311 }, { "epoch": 0.5190884752408458, "grad_norm": 2.6457571983337402, "learning_rate": 4.594381461007527e-05, "loss": 7.186, "step": 12312 }, { "epoch": 0.5191306364230454, "grad_norm": 2.425861358642578, "learning_rate": 4.59430338236672e-05, "loss": 7.7032, "step": 12313 }, { "epoch": 0.5191727976052448, "grad_norm": 1.8992258310317993, "learning_rate": 4.594225303725913e-05, "loss": 7.0312, "step": 12314 }, { "epoch": 0.5192149587874444, "grad_norm": 4.716669082641602, "learning_rate": 4.594147225085106e-05, "loss": 8.1999, "step": 12315 }, { "epoch": 0.519257119969644, "grad_norm": 3.413982629776001, "learning_rate": 4.5940691464442986e-05, "loss": 7.6402, "step": 12316 }, { "epoch": 0.5192992811518435, "grad_norm": 6.786108493804932, "learning_rate": 4.593991067803492e-05, "loss": 8.1052, "step": 12317 }, { "epoch": 0.519341442334043, "grad_norm": 2.2876319885253906, "learning_rate": 4.593912989162685e-05, "loss": 7.3953, "step": 12318 }, { "epoch": 0.5193836035162426, "grad_norm": 2.5274291038513184, "learning_rate": 4.593834910521878e-05, "loss": 7.3981, "step": 12319 }, { "epoch": 0.5194257646984421, "grad_norm": 5.062121868133545, "learning_rate": 4.593756831881071e-05, "loss": 8.1135, "step": 12320 }, { "epoch": 0.5194679258806417, "grad_norm": 1.9221084117889404, "learning_rate": 4.5936787532402635e-05, "loss": 7.1674, "step": 12321 }, { "epoch": 0.5195100870628413, "grad_norm": 3.4246180057525635, "learning_rate": 4.593600674599457e-05, "loss": 7.2587, "step": 12322 }, { "epoch": 0.5195522482450408, "grad_norm": 3.1804237365722656, "learning_rate": 4.59352259595865e-05, "loss": 6.9337, "step": 12323 }, { "epoch": 0.5195944094272403, "grad_norm": 4.9272332191467285, "learning_rate": 4.593444517317842e-05, "loss": 8.7216, "step": 12324 }, { "epoch": 0.5196365706094399, "grad_norm": 2.1803805828094482, "learning_rate": 4.593366438677036e-05, "loss": 7.4944, "step": 12325 }, { "epoch": 0.5196787317916395, "grad_norm": 2.1051297187805176, "learning_rate": 4.5932883600362285e-05, "loss": 7.0491, "step": 12326 }, { "epoch": 0.5197208929738389, "grad_norm": 4.571244716644287, "learning_rate": 4.5932102813954216e-05, "loss": 8.1807, "step": 12327 }, { "epoch": 0.5197630541560385, "grad_norm": 2.1397581100463867, "learning_rate": 4.593132202754615e-05, "loss": 6.9829, "step": 12328 }, { "epoch": 0.5198052153382381, "grad_norm": 2.2853000164031982, "learning_rate": 4.593054124113808e-05, "loss": 6.992, "step": 12329 }, { "epoch": 0.5198473765204377, "grad_norm": 5.279094219207764, "learning_rate": 4.592976045473001e-05, "loss": 9.2497, "step": 12330 }, { "epoch": 0.5198895377026371, "grad_norm": 2.0233969688415527, "learning_rate": 4.5928979668321934e-05, "loss": 7.0511, "step": 12331 }, { "epoch": 0.5199316988848367, "grad_norm": 2.6447157859802246, "learning_rate": 4.5928198881913865e-05, "loss": 7.2131, "step": 12332 }, { "epoch": 0.5199738600670363, "grad_norm": 2.947948455810547, "learning_rate": 4.5927418095505797e-05, "loss": 6.8142, "step": 12333 }, { "epoch": 0.5200160212492358, "grad_norm": 2.003561496734619, "learning_rate": 4.592663730909773e-05, "loss": 7.0521, "step": 12334 }, { "epoch": 0.5200581824314354, "grad_norm": 2.806199312210083, "learning_rate": 4.592585652268965e-05, "loss": 6.6655, "step": 12335 }, { "epoch": 0.5201003436136349, "grad_norm": 1.7838190793991089, "learning_rate": 4.5925075736281584e-05, "loss": 7.0973, "step": 12336 }, { "epoch": 0.5201425047958345, "grad_norm": 6.530656814575195, "learning_rate": 4.5924294949873515e-05, "loss": 8.311, "step": 12337 }, { "epoch": 0.520184665978034, "grad_norm": 2.538778781890869, "learning_rate": 4.592351416346544e-05, "loss": 6.9741, "step": 12338 }, { "epoch": 0.5202268271602336, "grad_norm": 1.8583226203918457, "learning_rate": 4.592273337705738e-05, "loss": 6.9527, "step": 12339 }, { "epoch": 0.5202689883424331, "grad_norm": 2.741438627243042, "learning_rate": 4.59219525906493e-05, "loss": 7.3129, "step": 12340 }, { "epoch": 0.5203111495246326, "grad_norm": 4.395619869232178, "learning_rate": 4.592117180424123e-05, "loss": 8.2048, "step": 12341 }, { "epoch": 0.5203533107068322, "grad_norm": 5.500894546508789, "learning_rate": 4.5920391017833164e-05, "loss": 8.2277, "step": 12342 }, { "epoch": 0.5203954718890318, "grad_norm": 1.748170256614685, "learning_rate": 4.591961023142509e-05, "loss": 6.9062, "step": 12343 }, { "epoch": 0.5204376330712314, "grad_norm": 3.8211400508880615, "learning_rate": 4.591882944501703e-05, "loss": 7.6766, "step": 12344 }, { "epoch": 0.5204797942534308, "grad_norm": 3.73982310295105, "learning_rate": 4.591804865860895e-05, "loss": 7.5555, "step": 12345 }, { "epoch": 0.5205219554356304, "grad_norm": 113.39151000976562, "learning_rate": 4.591726787220088e-05, "loss": 6.8694, "step": 12346 }, { "epoch": 0.52056411661783, "grad_norm": 2.2387309074401855, "learning_rate": 4.5916487085792814e-05, "loss": 6.8357, "step": 12347 }, { "epoch": 0.5206062778000295, "grad_norm": 228.5186309814453, "learning_rate": 4.5915706299384745e-05, "loss": 7.9158, "step": 12348 }, { "epoch": 0.520648438982229, "grad_norm": 10.200868606567383, "learning_rate": 4.591492551297667e-05, "loss": 7.1914, "step": 12349 }, { "epoch": 0.5206906001644286, "grad_norm": 18.488107681274414, "learning_rate": 4.59141447265686e-05, "loss": 7.1046, "step": 12350 }, { "epoch": 0.5207327613466282, "grad_norm": 4.64100980758667, "learning_rate": 4.591336394016053e-05, "loss": 7.8841, "step": 12351 }, { "epoch": 0.5207749225288277, "grad_norm": 12.439457893371582, "learning_rate": 4.5912583153752463e-05, "loss": 7.021, "step": 12352 }, { "epoch": 0.5208170837110273, "grad_norm": 212.0615234375, "learning_rate": 4.5911802367344395e-05, "loss": 6.5816, "step": 12353 }, { "epoch": 0.5208592448932268, "grad_norm": 2.186765193939209, "learning_rate": 4.591102158093632e-05, "loss": 7.067, "step": 12354 }, { "epoch": 0.5209014060754263, "grad_norm": 2.172942638397217, "learning_rate": 4.591024079452825e-05, "loss": 7.4339, "step": 12355 }, { "epoch": 0.5209435672576259, "grad_norm": 30.175825119018555, "learning_rate": 4.590946000812018e-05, "loss": 7.6522, "step": 12356 }, { "epoch": 0.5209857284398255, "grad_norm": 2.7731244564056396, "learning_rate": 4.5908679221712106e-05, "loss": 7.2067, "step": 12357 }, { "epoch": 0.521027889622025, "grad_norm": 4.544091701507568, "learning_rate": 4.5907898435304044e-05, "loss": 8.643, "step": 12358 }, { "epoch": 0.5210700508042245, "grad_norm": 3.4213180541992188, "learning_rate": 4.590711764889597e-05, "loss": 8.6464, "step": 12359 }, { "epoch": 0.5211122119864241, "grad_norm": 3.925144672393799, "learning_rate": 4.59063368624879e-05, "loss": 6.6239, "step": 12360 }, { "epoch": 0.5211543731686237, "grad_norm": 3.225945472717285, "learning_rate": 4.590555607607983e-05, "loss": 8.0228, "step": 12361 }, { "epoch": 0.5211965343508232, "grad_norm": 3.7803499698638916, "learning_rate": 4.5904775289671756e-05, "loss": 6.6055, "step": 12362 }, { "epoch": 0.5212386955330227, "grad_norm": 4.906617164611816, "learning_rate": 4.5903994503263694e-05, "loss": 6.9258, "step": 12363 }, { "epoch": 0.5212808567152223, "grad_norm": 2.0624420642852783, "learning_rate": 4.590321371685562e-05, "loss": 6.9509, "step": 12364 }, { "epoch": 0.5213230178974219, "grad_norm": 2.169560194015503, "learning_rate": 4.590243293044755e-05, "loss": 7.3893, "step": 12365 }, { "epoch": 0.5213651790796214, "grad_norm": 2.5241169929504395, "learning_rate": 4.590165214403948e-05, "loss": 7.6755, "step": 12366 }, { "epoch": 0.5214073402618209, "grad_norm": 3.0072524547576904, "learning_rate": 4.590087135763141e-05, "loss": 6.7726, "step": 12367 }, { "epoch": 0.5214495014440205, "grad_norm": 2.0021002292633057, "learning_rate": 4.5900090571223336e-05, "loss": 6.7549, "step": 12368 }, { "epoch": 0.52149166262622, "grad_norm": 4.30246639251709, "learning_rate": 4.589930978481527e-05, "loss": 8.3274, "step": 12369 }, { "epoch": 0.5215338238084196, "grad_norm": 2.554678440093994, "learning_rate": 4.58985289984072e-05, "loss": 6.8698, "step": 12370 }, { "epoch": 0.5215759849906192, "grad_norm": 2.713620901107788, "learning_rate": 4.5897748211999123e-05, "loss": 7.8224, "step": 12371 }, { "epoch": 0.5216181461728187, "grad_norm": 4.58193826675415, "learning_rate": 4.589696742559106e-05, "loss": 8.5822, "step": 12372 }, { "epoch": 0.5216603073550182, "grad_norm": 2.778233051300049, "learning_rate": 4.5896186639182986e-05, "loss": 7.0713, "step": 12373 }, { "epoch": 0.5217024685372178, "grad_norm": 2.875190496444702, "learning_rate": 4.589540585277492e-05, "loss": 7.6647, "step": 12374 }, { "epoch": 0.5217446297194174, "grad_norm": 4.125247478485107, "learning_rate": 4.589462506636685e-05, "loss": 7.0459, "step": 12375 }, { "epoch": 0.5217867909016168, "grad_norm": 2.4661009311676025, "learning_rate": 4.589384427995877e-05, "loss": 7.5156, "step": 12376 }, { "epoch": 0.5218289520838164, "grad_norm": 2.105247974395752, "learning_rate": 4.589306349355071e-05, "loss": 6.8114, "step": 12377 }, { "epoch": 0.521871113266016, "grad_norm": 3.869281053543091, "learning_rate": 4.5892282707142636e-05, "loss": 8.2648, "step": 12378 }, { "epoch": 0.5219132744482156, "grad_norm": 2.2903594970703125, "learning_rate": 4.589150192073457e-05, "loss": 6.6123, "step": 12379 }, { "epoch": 0.5219554356304151, "grad_norm": 1.9446157217025757, "learning_rate": 4.58907211343265e-05, "loss": 6.726, "step": 12380 }, { "epoch": 0.5219975968126146, "grad_norm": 3.7891197204589844, "learning_rate": 4.588994034791842e-05, "loss": 7.38, "step": 12381 }, { "epoch": 0.5220397579948142, "grad_norm": 2.8044519424438477, "learning_rate": 4.5889159561510354e-05, "loss": 7.5965, "step": 12382 }, { "epoch": 0.5220819191770137, "grad_norm": 2.7018890380859375, "learning_rate": 4.5888378775102285e-05, "loss": 6.7707, "step": 12383 }, { "epoch": 0.5221240803592133, "grad_norm": 2.0921921730041504, "learning_rate": 4.5887597988694216e-05, "loss": 6.8148, "step": 12384 }, { "epoch": 0.5221662415414128, "grad_norm": 1.8915417194366455, "learning_rate": 4.588681720228615e-05, "loss": 6.7295, "step": 12385 }, { "epoch": 0.5222084027236124, "grad_norm": 2.228078603744507, "learning_rate": 4.588603641587807e-05, "loss": 6.9022, "step": 12386 }, { "epoch": 0.5222505639058119, "grad_norm": 2.1769673824310303, "learning_rate": 4.588525562947e-05, "loss": 6.7887, "step": 12387 }, { "epoch": 0.5222927250880115, "grad_norm": 2.8782877922058105, "learning_rate": 4.5884474843061935e-05, "loss": 7.3511, "step": 12388 }, { "epoch": 0.522334886270211, "grad_norm": 4.2827043533325195, "learning_rate": 4.5883694056653866e-05, "loss": 7.7474, "step": 12389 }, { "epoch": 0.5223770474524105, "grad_norm": 1.7836893796920776, "learning_rate": 4.588291327024579e-05, "loss": 7.1859, "step": 12390 }, { "epoch": 0.5224192086346101, "grad_norm": 1.755415916442871, "learning_rate": 4.588213248383773e-05, "loss": 6.9191, "step": 12391 }, { "epoch": 0.5224613698168097, "grad_norm": 2.6383297443389893, "learning_rate": 4.588135169742965e-05, "loss": 7.165, "step": 12392 }, { "epoch": 0.5225035309990093, "grad_norm": 3.0432238578796387, "learning_rate": 4.5880570911021584e-05, "loss": 7.178, "step": 12393 }, { "epoch": 0.5225456921812087, "grad_norm": 2.549534559249878, "learning_rate": 4.5879790124613515e-05, "loss": 7.778, "step": 12394 }, { "epoch": 0.5225878533634083, "grad_norm": 1.9481956958770752, "learning_rate": 4.587900933820544e-05, "loss": 7.2994, "step": 12395 }, { "epoch": 0.5226300145456079, "grad_norm": 1.6589808464050293, "learning_rate": 4.587822855179737e-05, "loss": 6.7482, "step": 12396 }, { "epoch": 0.5226721757278074, "grad_norm": 2.198133945465088, "learning_rate": 4.58774477653893e-05, "loss": 6.5928, "step": 12397 }, { "epoch": 0.5227143369100069, "grad_norm": 4.03613805770874, "learning_rate": 4.5876666978981234e-05, "loss": 7.7356, "step": 12398 }, { "epoch": 0.5227564980922065, "grad_norm": 4.373698711395264, "learning_rate": 4.5875886192573165e-05, "loss": 7.5339, "step": 12399 }, { "epoch": 0.522798659274406, "grad_norm": 2.0694198608398438, "learning_rate": 4.587510540616509e-05, "loss": 6.5758, "step": 12400 }, { "epoch": 0.5228408204566056, "grad_norm": 2.6006178855895996, "learning_rate": 4.587432461975702e-05, "loss": 6.6609, "step": 12401 }, { "epoch": 0.5228829816388052, "grad_norm": 2.5940566062927246, "learning_rate": 4.587354383334895e-05, "loss": 7.1158, "step": 12402 }, { "epoch": 0.5229251428210047, "grad_norm": 4.468811988830566, "learning_rate": 4.587276304694088e-05, "loss": 8.1139, "step": 12403 }, { "epoch": 0.5229673040032042, "grad_norm": 1.5585246086120605, "learning_rate": 4.587198226053281e-05, "loss": 7.0292, "step": 12404 }, { "epoch": 0.5230094651854038, "grad_norm": 2.659111738204956, "learning_rate": 4.587120147412474e-05, "loss": 6.5962, "step": 12405 }, { "epoch": 0.5230516263676034, "grad_norm": 1.4181146621704102, "learning_rate": 4.587042068771667e-05, "loss": 6.8789, "step": 12406 }, { "epoch": 0.5230937875498028, "grad_norm": 2.285013437271118, "learning_rate": 4.5869639901308595e-05, "loss": 6.9276, "step": 12407 }, { "epoch": 0.5231359487320024, "grad_norm": 2.1435935497283936, "learning_rate": 4.586885911490053e-05, "loss": 6.6427, "step": 12408 }, { "epoch": 0.523178109914202, "grad_norm": 3.584942102432251, "learning_rate": 4.586807832849246e-05, "loss": 6.9323, "step": 12409 }, { "epoch": 0.5232202710964016, "grad_norm": 2.254668951034546, "learning_rate": 4.5867297542084395e-05, "loss": 6.9435, "step": 12410 }, { "epoch": 0.5232624322786011, "grad_norm": 2.38918399810791, "learning_rate": 4.586651675567632e-05, "loss": 7.4783, "step": 12411 }, { "epoch": 0.5233045934608006, "grad_norm": 4.299013137817383, "learning_rate": 4.586573596926825e-05, "loss": 7.8734, "step": 12412 }, { "epoch": 0.5233467546430002, "grad_norm": 2.3672895431518555, "learning_rate": 4.586495518286018e-05, "loss": 6.7167, "step": 12413 }, { "epoch": 0.5233889158251998, "grad_norm": 2.596904993057251, "learning_rate": 4.586417439645211e-05, "loss": 6.6556, "step": 12414 }, { "epoch": 0.5234310770073993, "grad_norm": 1.8913229703903198, "learning_rate": 4.586339361004404e-05, "loss": 6.8569, "step": 12415 }, { "epoch": 0.5234732381895988, "grad_norm": 1.588692307472229, "learning_rate": 4.586261282363597e-05, "loss": 6.6991, "step": 12416 }, { "epoch": 0.5235153993717984, "grad_norm": 2.1158230304718018, "learning_rate": 4.58618320372279e-05, "loss": 6.6361, "step": 12417 }, { "epoch": 0.5235575605539979, "grad_norm": 1.7970794439315796, "learning_rate": 4.5861051250819825e-05, "loss": 6.8276, "step": 12418 }, { "epoch": 0.5235997217361975, "grad_norm": 1.7953263521194458, "learning_rate": 4.5860270464411756e-05, "loss": 6.885, "step": 12419 }, { "epoch": 0.5236418829183971, "grad_norm": 5.7883806228637695, "learning_rate": 4.585948967800369e-05, "loss": 8.4807, "step": 12420 }, { "epoch": 0.5236840441005965, "grad_norm": 1.8910548686981201, "learning_rate": 4.585870889159562e-05, "loss": 6.8685, "step": 12421 }, { "epoch": 0.5237262052827961, "grad_norm": 4.829803466796875, "learning_rate": 4.585792810518755e-05, "loss": 7.2391, "step": 12422 }, { "epoch": 0.5237683664649957, "grad_norm": 1.6459717750549316, "learning_rate": 4.5857147318779474e-05, "loss": 6.8526, "step": 12423 }, { "epoch": 0.5238105276471953, "grad_norm": 1.4473848342895508, "learning_rate": 4.5856366532371406e-05, "loss": 6.8543, "step": 12424 }, { "epoch": 0.5238526888293947, "grad_norm": 1.4961107969284058, "learning_rate": 4.585558574596334e-05, "loss": 6.9554, "step": 12425 }, { "epoch": 0.5238948500115943, "grad_norm": 2.2511796951293945, "learning_rate": 4.585480495955526e-05, "loss": 7.1464, "step": 12426 }, { "epoch": 0.5239370111937939, "grad_norm": 1.4912266731262207, "learning_rate": 4.58540241731472e-05, "loss": 6.716, "step": 12427 }, { "epoch": 0.5239791723759935, "grad_norm": 2.3180429935455322, "learning_rate": 4.5853243386739124e-05, "loss": 7.8578, "step": 12428 }, { "epoch": 0.524021333558193, "grad_norm": 2.509671688079834, "learning_rate": 4.5852462600331055e-05, "loss": 6.7794, "step": 12429 }, { "epoch": 0.5240634947403925, "grad_norm": 4.803190231323242, "learning_rate": 4.5851681813922986e-05, "loss": 8.362, "step": 12430 }, { "epoch": 0.5241056559225921, "grad_norm": 2.883044958114624, "learning_rate": 4.585090102751491e-05, "loss": 7.9112, "step": 12431 }, { "epoch": 0.5241478171047916, "grad_norm": 2.2765250205993652, "learning_rate": 4.585012024110685e-05, "loss": 7.3746, "step": 12432 }, { "epoch": 0.5241899782869912, "grad_norm": 1.7968400716781616, "learning_rate": 4.5849339454698774e-05, "loss": 6.8297, "step": 12433 }, { "epoch": 0.5242321394691907, "grad_norm": 2.432936668395996, "learning_rate": 4.5848558668290705e-05, "loss": 7.3743, "step": 12434 }, { "epoch": 0.5242743006513902, "grad_norm": 1.6570924520492554, "learning_rate": 4.5847777881882636e-05, "loss": 6.7312, "step": 12435 }, { "epoch": 0.5243164618335898, "grad_norm": 1.935652256011963, "learning_rate": 4.584699709547457e-05, "loss": 6.948, "step": 12436 }, { "epoch": 0.5243586230157894, "grad_norm": 4.336211681365967, "learning_rate": 4.584621630906649e-05, "loss": 8.3635, "step": 12437 }, { "epoch": 0.5244007841979889, "grad_norm": 1.7518513202667236, "learning_rate": 4.584543552265842e-05, "loss": 6.9742, "step": 12438 }, { "epoch": 0.5244429453801884, "grad_norm": 4.2811479568481445, "learning_rate": 4.5844654736250354e-05, "loss": 8.4625, "step": 12439 }, { "epoch": 0.524485106562388, "grad_norm": 3.181285858154297, "learning_rate": 4.584387394984228e-05, "loss": 7.3761, "step": 12440 }, { "epoch": 0.5245272677445876, "grad_norm": 3.041386365890503, "learning_rate": 4.584309316343422e-05, "loss": 7.1855, "step": 12441 }, { "epoch": 0.5245694289267872, "grad_norm": 1.8524999618530273, "learning_rate": 4.584231237702614e-05, "loss": 6.6399, "step": 12442 }, { "epoch": 0.5246115901089866, "grad_norm": 2.1356937885284424, "learning_rate": 4.584153159061807e-05, "loss": 7.5984, "step": 12443 }, { "epoch": 0.5246537512911862, "grad_norm": 2.656149387359619, "learning_rate": 4.5840750804210004e-05, "loss": 7.6432, "step": 12444 }, { "epoch": 0.5246959124733858, "grad_norm": 1.9057443141937256, "learning_rate": 4.583997001780193e-05, "loss": 7.0148, "step": 12445 }, { "epoch": 0.5247380736555853, "grad_norm": 2.050994634628296, "learning_rate": 4.5839189231393866e-05, "loss": 7.3609, "step": 12446 }, { "epoch": 0.5247802348377848, "grad_norm": 3.41493558883667, "learning_rate": 4.583840844498579e-05, "loss": 8.24, "step": 12447 }, { "epoch": 0.5248223960199844, "grad_norm": 1.9859222173690796, "learning_rate": 4.583762765857772e-05, "loss": 6.853, "step": 12448 }, { "epoch": 0.524864557202184, "grad_norm": 1.4667017459869385, "learning_rate": 4.583684687216965e-05, "loss": 6.7781, "step": 12449 }, { "epoch": 0.5249067183843835, "grad_norm": 1.4961434602737427, "learning_rate": 4.583606608576158e-05, "loss": 7.1781, "step": 12450 }, { "epoch": 0.5249488795665831, "grad_norm": 2.341634511947632, "learning_rate": 4.583528529935351e-05, "loss": 6.576, "step": 12451 }, { "epoch": 0.5249910407487826, "grad_norm": 1.6209228038787842, "learning_rate": 4.583450451294544e-05, "loss": 7.3631, "step": 12452 }, { "epoch": 0.5250332019309821, "grad_norm": 2.4482994079589844, "learning_rate": 4.583372372653737e-05, "loss": 7.4303, "step": 12453 }, { "epoch": 0.5250753631131817, "grad_norm": 1.878141164779663, "learning_rate": 4.58329429401293e-05, "loss": 7.3392, "step": 12454 }, { "epoch": 0.5251175242953813, "grad_norm": 1.48190176486969, "learning_rate": 4.5832162153721234e-05, "loss": 6.7465, "step": 12455 }, { "epoch": 0.5251596854775807, "grad_norm": 2.2252228260040283, "learning_rate": 4.583138136731316e-05, "loss": 6.6752, "step": 12456 }, { "epoch": 0.5252018466597803, "grad_norm": 1.6102144718170166, "learning_rate": 4.583060058090509e-05, "loss": 6.8337, "step": 12457 }, { "epoch": 0.5252440078419799, "grad_norm": 1.614827275276184, "learning_rate": 4.582981979449702e-05, "loss": 6.738, "step": 12458 }, { "epoch": 0.5252861690241795, "grad_norm": 1.6711679697036743, "learning_rate": 4.5829039008088946e-05, "loss": 7.4543, "step": 12459 }, { "epoch": 0.525328330206379, "grad_norm": 2.585860252380371, "learning_rate": 4.5828258221680884e-05, "loss": 7.1936, "step": 12460 }, { "epoch": 0.5253704913885785, "grad_norm": 1.9486219882965088, "learning_rate": 4.582747743527281e-05, "loss": 6.7254, "step": 12461 }, { "epoch": 0.5254126525707781, "grad_norm": 1.6925307512283325, "learning_rate": 4.582669664886474e-05, "loss": 6.6264, "step": 12462 }, { "epoch": 0.5254548137529776, "grad_norm": 2.494527816772461, "learning_rate": 4.582591586245667e-05, "loss": 7.1932, "step": 12463 }, { "epoch": 0.5254969749351772, "grad_norm": 2.247387409210205, "learning_rate": 4.5825135076048595e-05, "loss": 7.8334, "step": 12464 }, { "epoch": 0.5255391361173767, "grad_norm": 1.64693021774292, "learning_rate": 4.582435428964053e-05, "loss": 7.5518, "step": 12465 }, { "epoch": 0.5255812972995763, "grad_norm": 5.358400821685791, "learning_rate": 4.582357350323246e-05, "loss": 8.33, "step": 12466 }, { "epoch": 0.5256234584817758, "grad_norm": 1.7973634004592896, "learning_rate": 4.582279271682439e-05, "loss": 7.2956, "step": 12467 }, { "epoch": 0.5256656196639754, "grad_norm": 1.3614895343780518, "learning_rate": 4.582201193041632e-05, "loss": 6.7827, "step": 12468 }, { "epoch": 0.525707780846175, "grad_norm": 5.293837070465088, "learning_rate": 4.5821231144008245e-05, "loss": 7.874, "step": 12469 }, { "epoch": 0.5257499420283744, "grad_norm": 3.4265880584716797, "learning_rate": 4.5820450357600176e-05, "loss": 8.5761, "step": 12470 }, { "epoch": 0.525792103210574, "grad_norm": 3.0211544036865234, "learning_rate": 4.581966957119211e-05, "loss": 7.3043, "step": 12471 }, { "epoch": 0.5258342643927736, "grad_norm": 1.9426250457763672, "learning_rate": 4.581888878478404e-05, "loss": 6.7616, "step": 12472 }, { "epoch": 0.5258764255749732, "grad_norm": 4.787106037139893, "learning_rate": 4.581810799837596e-05, "loss": 8.4269, "step": 12473 }, { "epoch": 0.5259185867571726, "grad_norm": 4.103071212768555, "learning_rate": 4.58173272119679e-05, "loss": 8.159, "step": 12474 }, { "epoch": 0.5259607479393722, "grad_norm": 2.403067111968994, "learning_rate": 4.5816546425559825e-05, "loss": 7.8092, "step": 12475 }, { "epoch": 0.5260029091215718, "grad_norm": 4.00134801864624, "learning_rate": 4.581576563915175e-05, "loss": 8.6562, "step": 12476 }, { "epoch": 0.5260450703037713, "grad_norm": 2.0367119312286377, "learning_rate": 4.581498485274369e-05, "loss": 6.7563, "step": 12477 }, { "epoch": 0.5260872314859709, "grad_norm": 3.790113687515259, "learning_rate": 4.581420406633561e-05, "loss": 6.6377, "step": 12478 }, { "epoch": 0.5261293926681704, "grad_norm": 1.9004589319229126, "learning_rate": 4.581342327992755e-05, "loss": 6.8119, "step": 12479 }, { "epoch": 0.52617155385037, "grad_norm": 3.201080322265625, "learning_rate": 4.5812642493519475e-05, "loss": 6.8456, "step": 12480 }, { "epoch": 0.5262137150325695, "grad_norm": 1.7123596668243408, "learning_rate": 4.5811861707111406e-05, "loss": 6.7328, "step": 12481 }, { "epoch": 0.5262558762147691, "grad_norm": 3.0583202838897705, "learning_rate": 4.581108092070334e-05, "loss": 7.366, "step": 12482 }, { "epoch": 0.5262980373969686, "grad_norm": 3.3679039478302, "learning_rate": 4.581030013429526e-05, "loss": 8.2176, "step": 12483 }, { "epoch": 0.5263401985791681, "grad_norm": 2.3244857788085938, "learning_rate": 4.580951934788719e-05, "loss": 7.4821, "step": 12484 }, { "epoch": 0.5263823597613677, "grad_norm": 2.6173343658447266, "learning_rate": 4.5808738561479124e-05, "loss": 7.4109, "step": 12485 }, { "epoch": 0.5264245209435673, "grad_norm": 1.9788591861724854, "learning_rate": 4.5807957775071056e-05, "loss": 7.0331, "step": 12486 }, { "epoch": 0.5264666821257669, "grad_norm": 4.200056076049805, "learning_rate": 4.580717698866298e-05, "loss": 7.8483, "step": 12487 }, { "epoch": 0.5265088433079663, "grad_norm": 2.162050724029541, "learning_rate": 4.580639620225491e-05, "loss": 7.0593, "step": 12488 }, { "epoch": 0.5265510044901659, "grad_norm": 1.8463398218154907, "learning_rate": 4.580561541584684e-05, "loss": 6.723, "step": 12489 }, { "epoch": 0.5265931656723655, "grad_norm": 1.5561271905899048, "learning_rate": 4.5804834629438774e-05, "loss": 6.7445, "step": 12490 }, { "epoch": 0.526635326854565, "grad_norm": 1.9803155660629272, "learning_rate": 4.5804053843030705e-05, "loss": 7.281, "step": 12491 }, { "epoch": 0.5266774880367645, "grad_norm": 3.77066969871521, "learning_rate": 4.580327305662263e-05, "loss": 6.4732, "step": 12492 }, { "epoch": 0.5267196492189641, "grad_norm": 1.8144437074661255, "learning_rate": 4.580249227021456e-05, "loss": 6.5723, "step": 12493 }, { "epoch": 0.5267618104011637, "grad_norm": 3.003434658050537, "learning_rate": 4.580171148380649e-05, "loss": 7.3858, "step": 12494 }, { "epoch": 0.5268039715833632, "grad_norm": 3.9954230785369873, "learning_rate": 4.580093069739842e-05, "loss": 8.6925, "step": 12495 }, { "epoch": 0.5268461327655627, "grad_norm": 2.1548256874084473, "learning_rate": 4.5800149910990355e-05, "loss": 7.8738, "step": 12496 }, { "epoch": 0.5268882939477623, "grad_norm": 2.240844249725342, "learning_rate": 4.579936912458228e-05, "loss": 7.3502, "step": 12497 }, { "epoch": 0.5269304551299618, "grad_norm": 2.838529109954834, "learning_rate": 4.579858833817421e-05, "loss": 6.5453, "step": 12498 }, { "epoch": 0.5269726163121614, "grad_norm": 4.2376708984375, "learning_rate": 4.579780755176614e-05, "loss": 6.644, "step": 12499 }, { "epoch": 0.527014777494361, "grad_norm": 3.4799039363861084, "learning_rate": 4.579702676535807e-05, "loss": 7.2288, "step": 12500 }, { "epoch": 0.5270569386765604, "grad_norm": 1.7085498571395874, "learning_rate": 4.5796245978950004e-05, "loss": 7.2449, "step": 12501 }, { "epoch": 0.52709909985876, "grad_norm": 2.30313777923584, "learning_rate": 4.579546519254193e-05, "loss": 7.6586, "step": 12502 }, { "epoch": 0.5271412610409596, "grad_norm": 1.9010156393051147, "learning_rate": 4.579468440613386e-05, "loss": 7.538, "step": 12503 }, { "epoch": 0.5271834222231592, "grad_norm": 1.7512383460998535, "learning_rate": 4.579390361972579e-05, "loss": 6.667, "step": 12504 }, { "epoch": 0.5272255834053586, "grad_norm": 3.3609375953674316, "learning_rate": 4.579312283331772e-05, "loss": 7.539, "step": 12505 }, { "epoch": 0.5272677445875582, "grad_norm": 5.310394287109375, "learning_rate": 4.579234204690965e-05, "loss": 8.1561, "step": 12506 }, { "epoch": 0.5273099057697578, "grad_norm": 2.7965805530548096, "learning_rate": 4.579156126050158e-05, "loss": 7.4708, "step": 12507 }, { "epoch": 0.5273520669519574, "grad_norm": 1.994994878768921, "learning_rate": 4.579078047409351e-05, "loss": 7.2679, "step": 12508 }, { "epoch": 0.5273942281341569, "grad_norm": 2.050410032272339, "learning_rate": 4.5789999687685434e-05, "loss": 6.6791, "step": 12509 }, { "epoch": 0.5274363893163564, "grad_norm": 2.038637399673462, "learning_rate": 4.578921890127737e-05, "loss": 6.7934, "step": 12510 }, { "epoch": 0.527478550498556, "grad_norm": 2.868067741394043, "learning_rate": 4.5788438114869297e-05, "loss": 8.4067, "step": 12511 }, { "epoch": 0.5275207116807555, "grad_norm": 2.5709874629974365, "learning_rate": 4.578765732846123e-05, "loss": 6.7477, "step": 12512 }, { "epoch": 0.5275628728629551, "grad_norm": 4.6477274894714355, "learning_rate": 4.578687654205316e-05, "loss": 8.1587, "step": 12513 }, { "epoch": 0.5276050340451546, "grad_norm": 2.7289419174194336, "learning_rate": 4.5786095755645084e-05, "loss": 7.3069, "step": 12514 }, { "epoch": 0.5276471952273541, "grad_norm": 1.9583349227905273, "learning_rate": 4.578531496923702e-05, "loss": 6.8125, "step": 12515 }, { "epoch": 0.5276893564095537, "grad_norm": 3.67097544670105, "learning_rate": 4.5784534182828946e-05, "loss": 7.7045, "step": 12516 }, { "epoch": 0.5277315175917533, "grad_norm": 2.221867561340332, "learning_rate": 4.578375339642088e-05, "loss": 7.2779, "step": 12517 }, { "epoch": 0.5277736787739529, "grad_norm": 2.4014933109283447, "learning_rate": 4.578297261001281e-05, "loss": 7.4807, "step": 12518 }, { "epoch": 0.5278158399561523, "grad_norm": 3.096708059310913, "learning_rate": 4.578219182360473e-05, "loss": 7.5899, "step": 12519 }, { "epoch": 0.5278580011383519, "grad_norm": 3.086843967437744, "learning_rate": 4.5781411037196664e-05, "loss": 7.8574, "step": 12520 }, { "epoch": 0.5279001623205515, "grad_norm": 2.786468505859375, "learning_rate": 4.5780630250788596e-05, "loss": 6.5479, "step": 12521 }, { "epoch": 0.527942323502751, "grad_norm": 4.300774574279785, "learning_rate": 4.577984946438053e-05, "loss": 7.5873, "step": 12522 }, { "epoch": 0.5279844846849505, "grad_norm": 1.838172197341919, "learning_rate": 4.577906867797246e-05, "loss": 6.9811, "step": 12523 }, { "epoch": 0.5280266458671501, "grad_norm": 3.9296560287475586, "learning_rate": 4.577828789156439e-05, "loss": 6.8499, "step": 12524 }, { "epoch": 0.5280688070493497, "grad_norm": 1.731421947479248, "learning_rate": 4.5777507105156314e-05, "loss": 7.0905, "step": 12525 }, { "epoch": 0.5281109682315492, "grad_norm": 1.9902894496917725, "learning_rate": 4.5776726318748245e-05, "loss": 6.8335, "step": 12526 }, { "epoch": 0.5281531294137488, "grad_norm": 2.473843812942505, "learning_rate": 4.5775945532340176e-05, "loss": 6.8114, "step": 12527 }, { "epoch": 0.5281952905959483, "grad_norm": 2.2777209281921387, "learning_rate": 4.57751647459321e-05, "loss": 7.0416, "step": 12528 }, { "epoch": 0.5282374517781478, "grad_norm": 1.7376371622085571, "learning_rate": 4.577438395952404e-05, "loss": 7.3783, "step": 12529 }, { "epoch": 0.5282796129603474, "grad_norm": 3.560011863708496, "learning_rate": 4.5773603173115963e-05, "loss": 8.2788, "step": 12530 }, { "epoch": 0.528321774142547, "grad_norm": 1.6800005435943604, "learning_rate": 4.5772822386707895e-05, "loss": 6.9213, "step": 12531 }, { "epoch": 0.5283639353247465, "grad_norm": 1.6324739456176758, "learning_rate": 4.5772041600299826e-05, "loss": 6.7162, "step": 12532 }, { "epoch": 0.528406096506946, "grad_norm": 3.850006103515625, "learning_rate": 4.577126081389175e-05, "loss": 8.228, "step": 12533 }, { "epoch": 0.5284482576891456, "grad_norm": 3.478520393371582, "learning_rate": 4.577048002748369e-05, "loss": 8.1889, "step": 12534 }, { "epoch": 0.5284904188713452, "grad_norm": 1.4100027084350586, "learning_rate": 4.576969924107561e-05, "loss": 6.9334, "step": 12535 }, { "epoch": 0.5285325800535448, "grad_norm": 3.9217658042907715, "learning_rate": 4.5768918454667544e-05, "loss": 7.7189, "step": 12536 }, { "epoch": 0.5285747412357442, "grad_norm": 2.3710575103759766, "learning_rate": 4.5768137668259475e-05, "loss": 8.1225, "step": 12537 }, { "epoch": 0.5286169024179438, "grad_norm": 2.682436466217041, "learning_rate": 4.57673568818514e-05, "loss": 6.9915, "step": 12538 }, { "epoch": 0.5286590636001434, "grad_norm": 2.161180257797241, "learning_rate": 4.576657609544333e-05, "loss": 6.5455, "step": 12539 }, { "epoch": 0.5287012247823429, "grad_norm": 1.4637583494186401, "learning_rate": 4.576579530903526e-05, "loss": 7.1947, "step": 12540 }, { "epoch": 0.5287433859645424, "grad_norm": 1.4836726188659668, "learning_rate": 4.5765014522627194e-05, "loss": 6.9427, "step": 12541 }, { "epoch": 0.528785547146742, "grad_norm": 2.464833974838257, "learning_rate": 4.576423373621912e-05, "loss": 6.7502, "step": 12542 }, { "epoch": 0.5288277083289415, "grad_norm": 2.024191379547119, "learning_rate": 4.5763452949811056e-05, "loss": 7.5035, "step": 12543 }, { "epoch": 0.5288698695111411, "grad_norm": 2.2723326683044434, "learning_rate": 4.576267216340298e-05, "loss": 6.5222, "step": 12544 }, { "epoch": 0.5289120306933407, "grad_norm": 1.4743767976760864, "learning_rate": 4.576189137699491e-05, "loss": 7.0134, "step": 12545 }, { "epoch": 0.5289541918755402, "grad_norm": 1.8998173475265503, "learning_rate": 4.576111059058684e-05, "loss": 6.5302, "step": 12546 }, { "epoch": 0.5289963530577397, "grad_norm": 1.3540877103805542, "learning_rate": 4.576032980417877e-05, "loss": 6.9728, "step": 12547 }, { "epoch": 0.5290385142399393, "grad_norm": 1.8551379442214966, "learning_rate": 4.5759549017770706e-05, "loss": 6.5791, "step": 12548 }, { "epoch": 0.5290806754221389, "grad_norm": 2.471578359603882, "learning_rate": 4.575876823136263e-05, "loss": 7.3733, "step": 12549 }, { "epoch": 0.5291228366043383, "grad_norm": 2.015681505203247, "learning_rate": 4.575798744495456e-05, "loss": 6.8745, "step": 12550 }, { "epoch": 0.5291649977865379, "grad_norm": 1.8993326425552368, "learning_rate": 4.575720665854649e-05, "loss": 7.7043, "step": 12551 }, { "epoch": 0.5292071589687375, "grad_norm": 2.517782211303711, "learning_rate": 4.575642587213842e-05, "loss": 7.3778, "step": 12552 }, { "epoch": 0.5292493201509371, "grad_norm": 1.3289196491241455, "learning_rate": 4.575564508573035e-05, "loss": 6.6916, "step": 12553 }, { "epoch": 0.5292914813331365, "grad_norm": 1.5829650163650513, "learning_rate": 4.575486429932228e-05, "loss": 6.7697, "step": 12554 }, { "epoch": 0.5293336425153361, "grad_norm": 3.390202283859253, "learning_rate": 4.575408351291421e-05, "loss": 6.8405, "step": 12555 }, { "epoch": 0.5293758036975357, "grad_norm": 2.323796510696411, "learning_rate": 4.5753302726506136e-05, "loss": 7.7753, "step": 12556 }, { "epoch": 0.5294179648797352, "grad_norm": 4.143942832946777, "learning_rate": 4.575252194009807e-05, "loss": 7.6701, "step": 12557 }, { "epoch": 0.5294601260619348, "grad_norm": 2.966872215270996, "learning_rate": 4.575174115369e-05, "loss": 7.6532, "step": 12558 }, { "epoch": 0.5295022872441343, "grad_norm": 2.0027921199798584, "learning_rate": 4.575096036728193e-05, "loss": 7.3297, "step": 12559 }, { "epoch": 0.5295444484263339, "grad_norm": 2.043234348297119, "learning_rate": 4.575017958087386e-05, "loss": 7.1057, "step": 12560 }, { "epoch": 0.5295866096085334, "grad_norm": 4.329424858093262, "learning_rate": 4.5749398794465785e-05, "loss": 6.8456, "step": 12561 }, { "epoch": 0.529628770790733, "grad_norm": 1.7813379764556885, "learning_rate": 4.574861800805772e-05, "loss": 6.7463, "step": 12562 }, { "epoch": 0.5296709319729325, "grad_norm": 2.223895788192749, "learning_rate": 4.574783722164965e-05, "loss": 6.6953, "step": 12563 }, { "epoch": 0.529713093155132, "grad_norm": 1.6243863105773926, "learning_rate": 4.574705643524157e-05, "loss": 7.2165, "step": 12564 }, { "epoch": 0.5297552543373316, "grad_norm": 1.4853092432022095, "learning_rate": 4.574627564883351e-05, "loss": 7.0814, "step": 12565 }, { "epoch": 0.5297974155195312, "grad_norm": 3.183694839477539, "learning_rate": 4.5745494862425435e-05, "loss": 7.1297, "step": 12566 }, { "epoch": 0.5298395767017308, "grad_norm": 1.523935317993164, "learning_rate": 4.5744714076017366e-05, "loss": 6.7104, "step": 12567 }, { "epoch": 0.5298817378839302, "grad_norm": 1.3247888088226318, "learning_rate": 4.57439332896093e-05, "loss": 6.9591, "step": 12568 }, { "epoch": 0.5299238990661298, "grad_norm": 3.359981060028076, "learning_rate": 4.574315250320123e-05, "loss": 7.8886, "step": 12569 }, { "epoch": 0.5299660602483294, "grad_norm": 2.4962639808654785, "learning_rate": 4.574237171679316e-05, "loss": 7.3724, "step": 12570 }, { "epoch": 0.530008221430529, "grad_norm": 1.6495004892349243, "learning_rate": 4.5741590930385084e-05, "loss": 6.6432, "step": 12571 }, { "epoch": 0.5300503826127284, "grad_norm": 3.955177068710327, "learning_rate": 4.5740810143977015e-05, "loss": 9.1822, "step": 12572 }, { "epoch": 0.530092543794928, "grad_norm": 1.687997817993164, "learning_rate": 4.5740029357568947e-05, "loss": 6.8188, "step": 12573 }, { "epoch": 0.5301347049771276, "grad_norm": 2.513073682785034, "learning_rate": 4.573924857116088e-05, "loss": 7.3284, "step": 12574 }, { "epoch": 0.5301768661593271, "grad_norm": 1.596295714378357, "learning_rate": 4.57384677847528e-05, "loss": 6.6463, "step": 12575 }, { "epoch": 0.5302190273415267, "grad_norm": 2.1324515342712402, "learning_rate": 4.5737686998344734e-05, "loss": 6.4527, "step": 12576 }, { "epoch": 0.5302611885237262, "grad_norm": 2.865936040878296, "learning_rate": 4.5736906211936665e-05, "loss": 8.3298, "step": 12577 }, { "epoch": 0.5303033497059257, "grad_norm": 1.6918193101882935, "learning_rate": 4.573612542552859e-05, "loss": 6.9604, "step": 12578 }, { "epoch": 0.5303455108881253, "grad_norm": 2.1467347145080566, "learning_rate": 4.573534463912053e-05, "loss": 6.6368, "step": 12579 }, { "epoch": 0.5303876720703249, "grad_norm": 2.153493881225586, "learning_rate": 4.573456385271245e-05, "loss": 7.1667, "step": 12580 }, { "epoch": 0.5304298332525244, "grad_norm": 2.125767230987549, "learning_rate": 4.573378306630438e-05, "loss": 6.813, "step": 12581 }, { "epoch": 0.5304719944347239, "grad_norm": 1.7441219091415405, "learning_rate": 4.5733002279896314e-05, "loss": 7.3249, "step": 12582 }, { "epoch": 0.5305141556169235, "grad_norm": 3.0151360034942627, "learning_rate": 4.573222149348824e-05, "loss": 7.4261, "step": 12583 }, { "epoch": 0.5305563167991231, "grad_norm": 2.449349880218506, "learning_rate": 4.573144070708018e-05, "loss": 6.6675, "step": 12584 }, { "epoch": 0.5305984779813226, "grad_norm": 2.360668182373047, "learning_rate": 4.57306599206721e-05, "loss": 6.4636, "step": 12585 }, { "epoch": 0.5306406391635221, "grad_norm": 3.8863229751586914, "learning_rate": 4.572987913426403e-05, "loss": 7.8397, "step": 12586 }, { "epoch": 0.5306828003457217, "grad_norm": 1.338729739189148, "learning_rate": 4.5729098347855964e-05, "loss": 6.8504, "step": 12587 }, { "epoch": 0.5307249615279213, "grad_norm": 2.171914577484131, "learning_rate": 4.5728317561447895e-05, "loss": 7.3724, "step": 12588 }, { "epoch": 0.5307671227101208, "grad_norm": 3.084362268447876, "learning_rate": 4.572753677503982e-05, "loss": 6.7324, "step": 12589 }, { "epoch": 0.5308092838923203, "grad_norm": 2.2756032943725586, "learning_rate": 4.572675598863175e-05, "loss": 7.55, "step": 12590 }, { "epoch": 0.5308514450745199, "grad_norm": 1.8747533559799194, "learning_rate": 4.572597520222368e-05, "loss": 7.3919, "step": 12591 }, { "epoch": 0.5308936062567194, "grad_norm": 2.915889024734497, "learning_rate": 4.5725194415815613e-05, "loss": 8.0799, "step": 12592 }, { "epoch": 0.530935767438919, "grad_norm": 2.221287965774536, "learning_rate": 4.5724413629407545e-05, "loss": 6.681, "step": 12593 }, { "epoch": 0.5309779286211186, "grad_norm": 2.256834030151367, "learning_rate": 4.572363284299947e-05, "loss": 7.5457, "step": 12594 }, { "epoch": 0.531020089803318, "grad_norm": 2.3618059158325195, "learning_rate": 4.57228520565914e-05, "loss": 6.5807, "step": 12595 }, { "epoch": 0.5310622509855176, "grad_norm": 2.958030939102173, "learning_rate": 4.572207127018333e-05, "loss": 7.9106, "step": 12596 }, { "epoch": 0.5311044121677172, "grad_norm": 2.220996141433716, "learning_rate": 4.5721290483775256e-05, "loss": 6.5486, "step": 12597 }, { "epoch": 0.5311465733499168, "grad_norm": 2.5035207271575928, "learning_rate": 4.5720509697367194e-05, "loss": 6.844, "step": 12598 }, { "epoch": 0.5311887345321162, "grad_norm": 4.285445213317871, "learning_rate": 4.571972891095912e-05, "loss": 8.2981, "step": 12599 }, { "epoch": 0.5312308957143158, "grad_norm": 1.945897102355957, "learning_rate": 4.571894812455105e-05, "loss": 6.8705, "step": 12600 }, { "epoch": 0.5312730568965154, "grad_norm": 3.9997878074645996, "learning_rate": 4.571816733814298e-05, "loss": 8.0958, "step": 12601 }, { "epoch": 0.531315218078715, "grad_norm": 3.26725697517395, "learning_rate": 4.5717386551734906e-05, "loss": 8.6104, "step": 12602 }, { "epoch": 0.5313573792609145, "grad_norm": 1.7071890830993652, "learning_rate": 4.5716605765326844e-05, "loss": 6.9958, "step": 12603 }, { "epoch": 0.531399540443114, "grad_norm": 1.6230134963989258, "learning_rate": 4.571582497891877e-05, "loss": 6.7285, "step": 12604 }, { "epoch": 0.5314417016253136, "grad_norm": 1.7508032321929932, "learning_rate": 4.57150441925107e-05, "loss": 6.8408, "step": 12605 }, { "epoch": 0.5314838628075131, "grad_norm": 2.45528507232666, "learning_rate": 4.571426340610263e-05, "loss": 6.986, "step": 12606 }, { "epoch": 0.5315260239897127, "grad_norm": 1.7583128213882446, "learning_rate": 4.571348261969456e-05, "loss": 6.8011, "step": 12607 }, { "epoch": 0.5315681851719122, "grad_norm": 1.4450408220291138, "learning_rate": 4.5712701833286486e-05, "loss": 6.7745, "step": 12608 }, { "epoch": 0.5316103463541118, "grad_norm": 1.6303932666778564, "learning_rate": 4.571192104687842e-05, "loss": 6.9886, "step": 12609 }, { "epoch": 0.5316525075363113, "grad_norm": 2.648022413253784, "learning_rate": 4.571114026047035e-05, "loss": 6.5443, "step": 12610 }, { "epoch": 0.5316946687185109, "grad_norm": 2.200122594833374, "learning_rate": 4.5710359474062273e-05, "loss": 6.6487, "step": 12611 }, { "epoch": 0.5317368299007104, "grad_norm": 1.2574762105941772, "learning_rate": 4.570957868765421e-05, "loss": 6.5833, "step": 12612 }, { "epoch": 0.5317789910829099, "grad_norm": 2.3847362995147705, "learning_rate": 4.5708797901246136e-05, "loss": 7.9377, "step": 12613 }, { "epoch": 0.5318211522651095, "grad_norm": 2.0697765350341797, "learning_rate": 4.570801711483807e-05, "loss": 7.607, "step": 12614 }, { "epoch": 0.5318633134473091, "grad_norm": 1.706430196762085, "learning_rate": 4.570723632843e-05, "loss": 6.8441, "step": 12615 }, { "epoch": 0.5319054746295087, "grad_norm": 1.8945775032043457, "learning_rate": 4.570645554202192e-05, "loss": 6.7385, "step": 12616 }, { "epoch": 0.5319476358117081, "grad_norm": 3.127519369125366, "learning_rate": 4.570567475561386e-05, "loss": 7.8005, "step": 12617 }, { "epoch": 0.5319897969939077, "grad_norm": 3.1960198879241943, "learning_rate": 4.5704893969205786e-05, "loss": 6.9088, "step": 12618 }, { "epoch": 0.5320319581761073, "grad_norm": 2.330491065979004, "learning_rate": 4.570411318279772e-05, "loss": 6.6315, "step": 12619 }, { "epoch": 0.5320741193583068, "grad_norm": 2.0397098064422607, "learning_rate": 4.570333239638965e-05, "loss": 6.6219, "step": 12620 }, { "epoch": 0.5321162805405063, "grad_norm": 1.6303554773330688, "learning_rate": 4.570255160998157e-05, "loss": 6.6677, "step": 12621 }, { "epoch": 0.5321584417227059, "grad_norm": 1.2841464281082153, "learning_rate": 4.5701770823573504e-05, "loss": 6.7795, "step": 12622 }, { "epoch": 0.5322006029049055, "grad_norm": 2.1165571212768555, "learning_rate": 4.5700990037165435e-05, "loss": 6.7379, "step": 12623 }, { "epoch": 0.532242764087105, "grad_norm": 4.637545108795166, "learning_rate": 4.5700209250757366e-05, "loss": 7.9556, "step": 12624 }, { "epoch": 0.5322849252693046, "grad_norm": 3.4789059162139893, "learning_rate": 4.569942846434929e-05, "loss": 7.2904, "step": 12625 }, { "epoch": 0.5323270864515041, "grad_norm": 3.120460033416748, "learning_rate": 4.569864767794122e-05, "loss": 6.6498, "step": 12626 }, { "epoch": 0.5323692476337036, "grad_norm": 4.424717426300049, "learning_rate": 4.569786689153315e-05, "loss": 8.2714, "step": 12627 }, { "epoch": 0.5324114088159032, "grad_norm": 4.1733622550964355, "learning_rate": 4.5697086105125085e-05, "loss": 8.3172, "step": 12628 }, { "epoch": 0.5324535699981028, "grad_norm": 1.791094183921814, "learning_rate": 4.5696305318717016e-05, "loss": 7.1842, "step": 12629 }, { "epoch": 0.5324957311803022, "grad_norm": 3.2529489994049072, "learning_rate": 4.569552453230894e-05, "loss": 6.4883, "step": 12630 }, { "epoch": 0.5325378923625018, "grad_norm": 2.6006762981414795, "learning_rate": 4.569474374590088e-05, "loss": 7.3138, "step": 12631 }, { "epoch": 0.5325800535447014, "grad_norm": 1.9597162008285522, "learning_rate": 4.56939629594928e-05, "loss": 7.1058, "step": 12632 }, { "epoch": 0.532622214726901, "grad_norm": 2.8103933334350586, "learning_rate": 4.5693182173084734e-05, "loss": 8.1817, "step": 12633 }, { "epoch": 0.5326643759091005, "grad_norm": 2.463200807571411, "learning_rate": 4.5692401386676665e-05, "loss": 6.9375, "step": 12634 }, { "epoch": 0.5327065370913, "grad_norm": 3.623046875, "learning_rate": 4.569162060026859e-05, "loss": 8.0807, "step": 12635 }, { "epoch": 0.5327486982734996, "grad_norm": 1.9804880619049072, "learning_rate": 4.569083981386052e-05, "loss": 7.5995, "step": 12636 }, { "epoch": 0.5327908594556992, "grad_norm": 3.042428493499756, "learning_rate": 4.569005902745245e-05, "loss": 7.826, "step": 12637 }, { "epoch": 0.5328330206378987, "grad_norm": 1.5530128479003906, "learning_rate": 4.5689278241044384e-05, "loss": 6.7491, "step": 12638 }, { "epoch": 0.5328751818200982, "grad_norm": 3.1633670330047607, "learning_rate": 4.5688497454636315e-05, "loss": 8.3498, "step": 12639 }, { "epoch": 0.5329173430022978, "grad_norm": 1.3815386295318604, "learning_rate": 4.568771666822824e-05, "loss": 6.6497, "step": 12640 }, { "epoch": 0.5329595041844973, "grad_norm": 1.5454119443893433, "learning_rate": 4.568693588182017e-05, "loss": 6.6114, "step": 12641 }, { "epoch": 0.5330016653666969, "grad_norm": 3.2613353729248047, "learning_rate": 4.56861550954121e-05, "loss": 6.7033, "step": 12642 }, { "epoch": 0.5330438265488965, "grad_norm": 1.8475942611694336, "learning_rate": 4.568537430900403e-05, "loss": 7.3533, "step": 12643 }, { "epoch": 0.533085987731096, "grad_norm": 1.9281872510910034, "learning_rate": 4.568459352259596e-05, "loss": 6.557, "step": 12644 }, { "epoch": 0.5331281489132955, "grad_norm": 3.3941850662231445, "learning_rate": 4.568381273618789e-05, "loss": 6.3572, "step": 12645 }, { "epoch": 0.5331703100954951, "grad_norm": 4.0813188552856445, "learning_rate": 4.568303194977982e-05, "loss": 7.7664, "step": 12646 }, { "epoch": 0.5332124712776947, "grad_norm": 2.0423004627227783, "learning_rate": 4.5682251163371745e-05, "loss": 7.3367, "step": 12647 }, { "epoch": 0.5332546324598941, "grad_norm": 1.294850468635559, "learning_rate": 4.568147037696368e-05, "loss": 6.616, "step": 12648 }, { "epoch": 0.5332967936420937, "grad_norm": 2.243626832962036, "learning_rate": 4.568068959055561e-05, "loss": 7.489, "step": 12649 }, { "epoch": 0.5333389548242933, "grad_norm": 1.3561756610870361, "learning_rate": 4.5679908804147545e-05, "loss": 6.5944, "step": 12650 }, { "epoch": 0.5333811160064929, "grad_norm": 1.1633566617965698, "learning_rate": 4.567912801773947e-05, "loss": 7.1037, "step": 12651 }, { "epoch": 0.5334232771886924, "grad_norm": 4.064992904663086, "learning_rate": 4.56783472313314e-05, "loss": 7.3127, "step": 12652 }, { "epoch": 0.5334654383708919, "grad_norm": 3.2689919471740723, "learning_rate": 4.567756644492333e-05, "loss": 8.366, "step": 12653 }, { "epoch": 0.5335075995530915, "grad_norm": 2.952019453048706, "learning_rate": 4.567678565851526e-05, "loss": 6.3915, "step": 12654 }, { "epoch": 0.533549760735291, "grad_norm": 1.3680169582366943, "learning_rate": 4.567600487210719e-05, "loss": 6.548, "step": 12655 }, { "epoch": 0.5335919219174906, "grad_norm": 2.612180233001709, "learning_rate": 4.567522408569912e-05, "loss": 6.981, "step": 12656 }, { "epoch": 0.5336340830996901, "grad_norm": 1.4174646139144897, "learning_rate": 4.567444329929105e-05, "loss": 6.7392, "step": 12657 }, { "epoch": 0.5336762442818896, "grad_norm": 2.0252110958099365, "learning_rate": 4.5673662512882975e-05, "loss": 7.5769, "step": 12658 }, { "epoch": 0.5337184054640892, "grad_norm": 1.760716438293457, "learning_rate": 4.5672881726474906e-05, "loss": 6.8494, "step": 12659 }, { "epoch": 0.5337605666462888, "grad_norm": 1.3739384412765503, "learning_rate": 4.567210094006684e-05, "loss": 6.5129, "step": 12660 }, { "epoch": 0.5338027278284884, "grad_norm": 1.38373601436615, "learning_rate": 4.567132015365877e-05, "loss": 6.8394, "step": 12661 }, { "epoch": 0.5338448890106878, "grad_norm": 1.2723714113235474, "learning_rate": 4.56705393672507e-05, "loss": 6.5025, "step": 12662 }, { "epoch": 0.5338870501928874, "grad_norm": 1.8330940008163452, "learning_rate": 4.5669758580842624e-05, "loss": 7.3853, "step": 12663 }, { "epoch": 0.533929211375087, "grad_norm": 1.679729700088501, "learning_rate": 4.5668977794434556e-05, "loss": 7.1487, "step": 12664 }, { "epoch": 0.5339713725572865, "grad_norm": 2.1569972038269043, "learning_rate": 4.566819700802649e-05, "loss": 6.7755, "step": 12665 }, { "epoch": 0.534013533739486, "grad_norm": 1.7613478899002075, "learning_rate": 4.566741622161841e-05, "loss": 6.9026, "step": 12666 }, { "epoch": 0.5340556949216856, "grad_norm": 4.781427383422852, "learning_rate": 4.566663543521035e-05, "loss": 9.5967, "step": 12667 }, { "epoch": 0.5340978561038852, "grad_norm": 3.99355149269104, "learning_rate": 4.5665854648802274e-05, "loss": 8.401, "step": 12668 }, { "epoch": 0.5341400172860847, "grad_norm": 2.927069664001465, "learning_rate": 4.5665073862394205e-05, "loss": 6.4691, "step": 12669 }, { "epoch": 0.5341821784682842, "grad_norm": 1.7145694494247437, "learning_rate": 4.5664293075986136e-05, "loss": 7.3888, "step": 12670 }, { "epoch": 0.5342243396504838, "grad_norm": 1.8067660331726074, "learning_rate": 4.566351228957806e-05, "loss": 6.7071, "step": 12671 }, { "epoch": 0.5342665008326833, "grad_norm": 1.5279606580734253, "learning_rate": 4.566273150317e-05, "loss": 6.7747, "step": 12672 }, { "epoch": 0.5343086620148829, "grad_norm": 3.0815372467041016, "learning_rate": 4.5661950716761924e-05, "loss": 7.3009, "step": 12673 }, { "epoch": 0.5343508231970825, "grad_norm": 1.4210681915283203, "learning_rate": 4.5661169930353855e-05, "loss": 6.7455, "step": 12674 }, { "epoch": 0.534392984379282, "grad_norm": 3.290497303009033, "learning_rate": 4.5660389143945786e-05, "loss": 8.3175, "step": 12675 }, { "epoch": 0.5344351455614815, "grad_norm": 1.3427079916000366, "learning_rate": 4.565960835753772e-05, "loss": 6.589, "step": 12676 }, { "epoch": 0.5344773067436811, "grad_norm": 3.6798062324523926, "learning_rate": 4.565882757112964e-05, "loss": 7.5676, "step": 12677 }, { "epoch": 0.5345194679258807, "grad_norm": 1.7551778554916382, "learning_rate": 4.565804678472157e-05, "loss": 7.2153, "step": 12678 }, { "epoch": 0.5345616291080801, "grad_norm": 2.02756404876709, "learning_rate": 4.5657265998313504e-05, "loss": 7.4214, "step": 12679 }, { "epoch": 0.5346037902902797, "grad_norm": 1.2881652116775513, "learning_rate": 4.565648521190543e-05, "loss": 7.279, "step": 12680 }, { "epoch": 0.5346459514724793, "grad_norm": 2.715160846710205, "learning_rate": 4.565570442549737e-05, "loss": 7.6506, "step": 12681 }, { "epoch": 0.5346881126546789, "grad_norm": 1.6410000324249268, "learning_rate": 4.565492363908929e-05, "loss": 7.2939, "step": 12682 }, { "epoch": 0.5347302738368784, "grad_norm": 2.2095725536346436, "learning_rate": 4.565414285268122e-05, "loss": 6.6286, "step": 12683 }, { "epoch": 0.5347724350190779, "grad_norm": 2.2411367893218994, "learning_rate": 4.5653362066273154e-05, "loss": 8.0249, "step": 12684 }, { "epoch": 0.5348145962012775, "grad_norm": 2.7640795707702637, "learning_rate": 4.565258127986508e-05, "loss": 6.8382, "step": 12685 }, { "epoch": 0.534856757383477, "grad_norm": 1.301888108253479, "learning_rate": 4.5651800493457016e-05, "loss": 6.6452, "step": 12686 }, { "epoch": 0.5348989185656766, "grad_norm": 1.7215428352355957, "learning_rate": 4.565101970704894e-05, "loss": 6.8092, "step": 12687 }, { "epoch": 0.5349410797478761, "grad_norm": 3.415761947631836, "learning_rate": 4.565023892064087e-05, "loss": 8.052, "step": 12688 }, { "epoch": 0.5349832409300757, "grad_norm": 3.3500733375549316, "learning_rate": 4.56494581342328e-05, "loss": 7.4915, "step": 12689 }, { "epoch": 0.5350254021122752, "grad_norm": 1.536209225654602, "learning_rate": 4.564867734782473e-05, "loss": 6.8694, "step": 12690 }, { "epoch": 0.5350675632944748, "grad_norm": 2.715296745300293, "learning_rate": 4.564789656141666e-05, "loss": 6.5815, "step": 12691 }, { "epoch": 0.5351097244766744, "grad_norm": 1.5895875692367554, "learning_rate": 4.564711577500859e-05, "loss": 7.4136, "step": 12692 }, { "epoch": 0.5351518856588738, "grad_norm": 1.6525424718856812, "learning_rate": 4.564633498860052e-05, "loss": 7.3226, "step": 12693 }, { "epoch": 0.5351940468410734, "grad_norm": 2.564760684967041, "learning_rate": 4.564555420219245e-05, "loss": 7.0467, "step": 12694 }, { "epoch": 0.535236208023273, "grad_norm": 2.9153051376342773, "learning_rate": 4.5644773415784384e-05, "loss": 8.2527, "step": 12695 }, { "epoch": 0.5352783692054726, "grad_norm": 1.870459794998169, "learning_rate": 4.564399262937631e-05, "loss": 7.0849, "step": 12696 }, { "epoch": 0.535320530387672, "grad_norm": 2.287079095840454, "learning_rate": 4.564321184296824e-05, "loss": 7.0852, "step": 12697 }, { "epoch": 0.5353626915698716, "grad_norm": 1.447357416152954, "learning_rate": 4.564243105656017e-05, "loss": 6.566, "step": 12698 }, { "epoch": 0.5354048527520712, "grad_norm": 4.38064432144165, "learning_rate": 4.5641650270152096e-05, "loss": 8.5358, "step": 12699 }, { "epoch": 0.5354470139342707, "grad_norm": 1.405680775642395, "learning_rate": 4.5640869483744034e-05, "loss": 7.1111, "step": 12700 }, { "epoch": 0.5354891751164703, "grad_norm": 2.761842966079712, "learning_rate": 4.564008869733596e-05, "loss": 7.1484, "step": 12701 }, { "epoch": 0.5355313362986698, "grad_norm": 4.3199687004089355, "learning_rate": 4.563930791092789e-05, "loss": 7.7209, "step": 12702 }, { "epoch": 0.5355734974808694, "grad_norm": 2.5147511959075928, "learning_rate": 4.563852712451982e-05, "loss": 7.3173, "step": 12703 }, { "epoch": 0.5356156586630689, "grad_norm": 1.2428560256958008, "learning_rate": 4.5637746338111745e-05, "loss": 6.559, "step": 12704 }, { "epoch": 0.5356578198452685, "grad_norm": 1.515159010887146, "learning_rate": 4.5636965551703676e-05, "loss": 6.7374, "step": 12705 }, { "epoch": 0.535699981027468, "grad_norm": 1.5458203554153442, "learning_rate": 4.563618476529561e-05, "loss": 6.6173, "step": 12706 }, { "epoch": 0.5357421422096675, "grad_norm": 2.679152727127075, "learning_rate": 4.563540397888754e-05, "loss": 7.3724, "step": 12707 }, { "epoch": 0.5357843033918671, "grad_norm": 3.05523419380188, "learning_rate": 4.563462319247947e-05, "loss": 7.8545, "step": 12708 }, { "epoch": 0.5358264645740667, "grad_norm": 2.397425889968872, "learning_rate": 4.5633842406071395e-05, "loss": 7.3252, "step": 12709 }, { "epoch": 0.5358686257562663, "grad_norm": 1.779801845550537, "learning_rate": 4.5633061619663326e-05, "loss": 6.8665, "step": 12710 }, { "epoch": 0.5359107869384657, "grad_norm": 3.4015889167785645, "learning_rate": 4.563228083325526e-05, "loss": 7.2879, "step": 12711 }, { "epoch": 0.5359529481206653, "grad_norm": 1.5806254148483276, "learning_rate": 4.563150004684719e-05, "loss": 7.6347, "step": 12712 }, { "epoch": 0.5359951093028649, "grad_norm": 1.6706628799438477, "learning_rate": 4.563071926043911e-05, "loss": 7.3745, "step": 12713 }, { "epoch": 0.5360372704850644, "grad_norm": 2.3377625942230225, "learning_rate": 4.562993847403105e-05, "loss": 6.7263, "step": 12714 }, { "epoch": 0.5360794316672639, "grad_norm": 3.261129140853882, "learning_rate": 4.5629157687622975e-05, "loss": 8.146, "step": 12715 }, { "epoch": 0.5361215928494635, "grad_norm": 3.140709638595581, "learning_rate": 4.56283769012149e-05, "loss": 8.2258, "step": 12716 }, { "epoch": 0.536163754031663, "grad_norm": 1.647789478302002, "learning_rate": 4.562759611480684e-05, "loss": 6.5529, "step": 12717 }, { "epoch": 0.5362059152138626, "grad_norm": 1.8184115886688232, "learning_rate": 4.562681532839876e-05, "loss": 6.3883, "step": 12718 }, { "epoch": 0.5362480763960622, "grad_norm": 1.9573538303375244, "learning_rate": 4.56260345419907e-05, "loss": 6.8074, "step": 12719 }, { "epoch": 0.5362902375782617, "grad_norm": 1.6844240427017212, "learning_rate": 4.5625253755582625e-05, "loss": 7.6347, "step": 12720 }, { "epoch": 0.5363323987604612, "grad_norm": 3.4000768661499023, "learning_rate": 4.5624472969174556e-05, "loss": 8.2475, "step": 12721 }, { "epoch": 0.5363745599426608, "grad_norm": 1.4187803268432617, "learning_rate": 4.562369218276649e-05, "loss": 6.6405, "step": 12722 }, { "epoch": 0.5364167211248604, "grad_norm": 2.662637948989868, "learning_rate": 4.562291139635841e-05, "loss": 7.7535, "step": 12723 }, { "epoch": 0.5364588823070598, "grad_norm": 1.9813737869262695, "learning_rate": 4.562213060995034e-05, "loss": 6.6256, "step": 12724 }, { "epoch": 0.5365010434892594, "grad_norm": 1.6457535028457642, "learning_rate": 4.5621349823542274e-05, "loss": 6.699, "step": 12725 }, { "epoch": 0.536543204671459, "grad_norm": 1.4724550247192383, "learning_rate": 4.5620569037134206e-05, "loss": 7.2486, "step": 12726 }, { "epoch": 0.5365853658536586, "grad_norm": 1.5572599172592163, "learning_rate": 4.561978825072613e-05, "loss": 6.7655, "step": 12727 }, { "epoch": 0.536627527035858, "grad_norm": 4.001200199127197, "learning_rate": 4.561900746431806e-05, "loss": 7.4541, "step": 12728 }, { "epoch": 0.5366696882180576, "grad_norm": 1.339192509651184, "learning_rate": 4.561822667790999e-05, "loss": 6.7487, "step": 12729 }, { "epoch": 0.5367118494002572, "grad_norm": 1.2795064449310303, "learning_rate": 4.5617445891501924e-05, "loss": 7.0856, "step": 12730 }, { "epoch": 0.5367540105824568, "grad_norm": 3.672968864440918, "learning_rate": 4.5616665105093855e-05, "loss": 8.3532, "step": 12731 }, { "epoch": 0.5367961717646563, "grad_norm": 1.101537823677063, "learning_rate": 4.561588431868578e-05, "loss": 6.914, "step": 12732 }, { "epoch": 0.5368383329468558, "grad_norm": 2.935532331466675, "learning_rate": 4.561510353227771e-05, "loss": 8.5345, "step": 12733 }, { "epoch": 0.5368804941290554, "grad_norm": 2.520258665084839, "learning_rate": 4.561432274586964e-05, "loss": 7.5383, "step": 12734 }, { "epoch": 0.5369226553112549, "grad_norm": 3.1217684745788574, "learning_rate": 4.561354195946157e-05, "loss": 6.6459, "step": 12735 }, { "epoch": 0.5369648164934545, "grad_norm": 2.805572986602783, "learning_rate": 4.5612761173053505e-05, "loss": 7.5059, "step": 12736 }, { "epoch": 0.537006977675654, "grad_norm": 1.1417911052703857, "learning_rate": 4.561198038664543e-05, "loss": 6.7761, "step": 12737 }, { "epoch": 0.5370491388578535, "grad_norm": 3.2227368354797363, "learning_rate": 4.561119960023736e-05, "loss": 6.4689, "step": 12738 }, { "epoch": 0.5370913000400531, "grad_norm": 1.5445632934570312, "learning_rate": 4.561041881382929e-05, "loss": 7.3981, "step": 12739 }, { "epoch": 0.5371334612222527, "grad_norm": 1.8707839250564575, "learning_rate": 4.560963802742122e-05, "loss": 6.6648, "step": 12740 }, { "epoch": 0.5371756224044523, "grad_norm": 1.4590983390808105, "learning_rate": 4.5608857241013154e-05, "loss": 6.7785, "step": 12741 }, { "epoch": 0.5372177835866517, "grad_norm": 1.088279128074646, "learning_rate": 4.560807645460508e-05, "loss": 6.7708, "step": 12742 }, { "epoch": 0.5372599447688513, "grad_norm": 3.3475499153137207, "learning_rate": 4.560729566819701e-05, "loss": 6.5494, "step": 12743 }, { "epoch": 0.5373021059510509, "grad_norm": 1.9107500314712524, "learning_rate": 4.560651488178894e-05, "loss": 7.0823, "step": 12744 }, { "epoch": 0.5373442671332505, "grad_norm": 3.1544251441955566, "learning_rate": 4.560573409538087e-05, "loss": 6.4224, "step": 12745 }, { "epoch": 0.5373864283154499, "grad_norm": 1.3226304054260254, "learning_rate": 4.56049533089728e-05, "loss": 6.7886, "step": 12746 }, { "epoch": 0.5374285894976495, "grad_norm": 1.343850016593933, "learning_rate": 4.560417252256473e-05, "loss": 6.9966, "step": 12747 }, { "epoch": 0.5374707506798491, "grad_norm": 1.4092289209365845, "learning_rate": 4.560339173615666e-05, "loss": 6.7782, "step": 12748 }, { "epoch": 0.5375129118620486, "grad_norm": 1.580850601196289, "learning_rate": 4.5602610949748584e-05, "loss": 7.1943, "step": 12749 }, { "epoch": 0.5375550730442482, "grad_norm": 1.3384153842926025, "learning_rate": 4.560183016334052e-05, "loss": 6.5534, "step": 12750 }, { "epoch": 0.5375972342264477, "grad_norm": 2.7683563232421875, "learning_rate": 4.5601049376932447e-05, "loss": 8.0461, "step": 12751 }, { "epoch": 0.5376393954086472, "grad_norm": 1.1023225784301758, "learning_rate": 4.560026859052438e-05, "loss": 6.8052, "step": 12752 }, { "epoch": 0.5376815565908468, "grad_norm": 2.119622230529785, "learning_rate": 4.559948780411631e-05, "loss": 7.2863, "step": 12753 }, { "epoch": 0.5377237177730464, "grad_norm": 2.8361001014709473, "learning_rate": 4.5598707017708234e-05, "loss": 7.9444, "step": 12754 }, { "epoch": 0.5377658789552459, "grad_norm": 1.9220244884490967, "learning_rate": 4.559792623130017e-05, "loss": 7.5444, "step": 12755 }, { "epoch": 0.5378080401374454, "grad_norm": 2.3567004203796387, "learning_rate": 4.5597145444892096e-05, "loss": 7.4289, "step": 12756 }, { "epoch": 0.537850201319645, "grad_norm": 2.3256685733795166, "learning_rate": 4.559636465848403e-05, "loss": 6.7195, "step": 12757 }, { "epoch": 0.5378923625018446, "grad_norm": 1.962311029434204, "learning_rate": 4.559558387207596e-05, "loss": 7.212, "step": 12758 }, { "epoch": 0.5379345236840442, "grad_norm": 1.2780284881591797, "learning_rate": 4.559480308566789e-05, "loss": 6.616, "step": 12759 }, { "epoch": 0.5379766848662436, "grad_norm": 1.1619504690170288, "learning_rate": 4.5594022299259814e-05, "loss": 6.5653, "step": 12760 }, { "epoch": 0.5380188460484432, "grad_norm": 1.0255186557769775, "learning_rate": 4.5593241512851746e-05, "loss": 6.6172, "step": 12761 }, { "epoch": 0.5380610072306428, "grad_norm": 3.3672690391540527, "learning_rate": 4.559246072644368e-05, "loss": 7.1964, "step": 12762 }, { "epoch": 0.5381031684128423, "grad_norm": 1.4797351360321045, "learning_rate": 4.559167994003561e-05, "loss": 6.5562, "step": 12763 }, { "epoch": 0.5381453295950418, "grad_norm": 1.3087875843048096, "learning_rate": 4.559089915362754e-05, "loss": 7.2022, "step": 12764 }, { "epoch": 0.5381874907772414, "grad_norm": 3.942814350128174, "learning_rate": 4.5590118367219464e-05, "loss": 7.5336, "step": 12765 }, { "epoch": 0.538229651959441, "grad_norm": 1.3996143341064453, "learning_rate": 4.5589337580811395e-05, "loss": 6.5575, "step": 12766 }, { "epoch": 0.5382718131416405, "grad_norm": 1.6603434085845947, "learning_rate": 4.5588556794403326e-05, "loss": 6.3756, "step": 12767 }, { "epoch": 0.5383139743238401, "grad_norm": 2.565981864929199, "learning_rate": 4.558777600799525e-05, "loss": 6.5235, "step": 12768 }, { "epoch": 0.5383561355060396, "grad_norm": 3.884143352508545, "learning_rate": 4.558699522158719e-05, "loss": 7.9396, "step": 12769 }, { "epoch": 0.5383982966882391, "grad_norm": 1.2319751977920532, "learning_rate": 4.5586214435179113e-05, "loss": 6.7651, "step": 12770 }, { "epoch": 0.5384404578704387, "grad_norm": 1.320574402809143, "learning_rate": 4.5585433648771045e-05, "loss": 6.815, "step": 12771 }, { "epoch": 0.5384826190526383, "grad_norm": 2.297614097595215, "learning_rate": 4.5584652862362976e-05, "loss": 6.3773, "step": 12772 }, { "epoch": 0.5385247802348377, "grad_norm": 2.5152225494384766, "learning_rate": 4.55838720759549e-05, "loss": 7.9075, "step": 12773 }, { "epoch": 0.5385669414170373, "grad_norm": 2.477468490600586, "learning_rate": 4.558309128954684e-05, "loss": 7.2033, "step": 12774 }, { "epoch": 0.5386091025992369, "grad_norm": 1.9567430019378662, "learning_rate": 4.558231050313876e-05, "loss": 7.3947, "step": 12775 }, { "epoch": 0.5386512637814365, "grad_norm": 3.317661762237549, "learning_rate": 4.5581529716730694e-05, "loss": 7.2934, "step": 12776 }, { "epoch": 0.5386934249636359, "grad_norm": 1.607178807258606, "learning_rate": 4.5580748930322625e-05, "loss": 7.0473, "step": 12777 }, { "epoch": 0.5387355861458355, "grad_norm": 2.086013078689575, "learning_rate": 4.557996814391455e-05, "loss": 6.3849, "step": 12778 }, { "epoch": 0.5387777473280351, "grad_norm": 1.6330269575119019, "learning_rate": 4.557918735750648e-05, "loss": 7.3743, "step": 12779 }, { "epoch": 0.5388199085102346, "grad_norm": 2.243652820587158, "learning_rate": 4.557840657109841e-05, "loss": 6.7982, "step": 12780 }, { "epoch": 0.5388620696924342, "grad_norm": 1.6173886060714722, "learning_rate": 4.5577625784690344e-05, "loss": 6.877, "step": 12781 }, { "epoch": 0.5389042308746337, "grad_norm": 1.964572548866272, "learning_rate": 4.557684499828227e-05, "loss": 7.0211, "step": 12782 }, { "epoch": 0.5389463920568333, "grad_norm": 2.3726773262023926, "learning_rate": 4.5576064211874206e-05, "loss": 7.263, "step": 12783 }, { "epoch": 0.5389885532390328, "grad_norm": 1.404255747795105, "learning_rate": 4.557528342546613e-05, "loss": 7.0897, "step": 12784 }, { "epoch": 0.5390307144212324, "grad_norm": 1.3235670328140259, "learning_rate": 4.557450263905806e-05, "loss": 6.3415, "step": 12785 }, { "epoch": 0.5390728756034319, "grad_norm": 1.34659743309021, "learning_rate": 4.557372185264999e-05, "loss": 7.3174, "step": 12786 }, { "epoch": 0.5391150367856314, "grad_norm": 4.811240196228027, "learning_rate": 4.557294106624192e-05, "loss": 10.1381, "step": 12787 }, { "epoch": 0.539157197967831, "grad_norm": 3.076237678527832, "learning_rate": 4.5572160279833856e-05, "loss": 8.5234, "step": 12788 }, { "epoch": 0.5391993591500306, "grad_norm": 1.3862292766571045, "learning_rate": 4.557137949342578e-05, "loss": 7.1647, "step": 12789 }, { "epoch": 0.5392415203322302, "grad_norm": 4.497127056121826, "learning_rate": 4.557059870701771e-05, "loss": 8.0335, "step": 12790 }, { "epoch": 0.5392836815144296, "grad_norm": 1.4413156509399414, "learning_rate": 4.556981792060964e-05, "loss": 6.8178, "step": 12791 }, { "epoch": 0.5393258426966292, "grad_norm": 3.275440216064453, "learning_rate": 4.556903713420157e-05, "loss": 7.7807, "step": 12792 }, { "epoch": 0.5393680038788288, "grad_norm": 4.03137731552124, "learning_rate": 4.55682563477935e-05, "loss": 9.6823, "step": 12793 }, { "epoch": 0.5394101650610283, "grad_norm": 2.9392807483673096, "learning_rate": 4.556747556138543e-05, "loss": 6.3578, "step": 12794 }, { "epoch": 0.5394523262432278, "grad_norm": 3.5440726280212402, "learning_rate": 4.556669477497736e-05, "loss": 8.2053, "step": 12795 }, { "epoch": 0.5394944874254274, "grad_norm": 3.104414224624634, "learning_rate": 4.5565913988569286e-05, "loss": 7.9133, "step": 12796 }, { "epoch": 0.539536648607627, "grad_norm": 1.2949386835098267, "learning_rate": 4.556513320216122e-05, "loss": 6.5934, "step": 12797 }, { "epoch": 0.5395788097898265, "grad_norm": 3.3855371475219727, "learning_rate": 4.556435241575315e-05, "loss": 8.0276, "step": 12798 }, { "epoch": 0.5396209709720261, "grad_norm": 1.7067283391952515, "learning_rate": 4.556357162934508e-05, "loss": 6.8818, "step": 12799 }, { "epoch": 0.5396631321542256, "grad_norm": 2.6601288318634033, "learning_rate": 4.556279084293701e-05, "loss": 6.753, "step": 12800 }, { "epoch": 0.5397052933364251, "grad_norm": 1.7281415462493896, "learning_rate": 4.5562010056528935e-05, "loss": 6.5468, "step": 12801 }, { "epoch": 0.5397474545186247, "grad_norm": 1.5209587812423706, "learning_rate": 4.556122927012087e-05, "loss": 6.9478, "step": 12802 }, { "epoch": 0.5397896157008243, "grad_norm": 1.7146395444869995, "learning_rate": 4.55604484837128e-05, "loss": 7.1769, "step": 12803 }, { "epoch": 0.5398317768830238, "grad_norm": 1.714707612991333, "learning_rate": 4.555966769730472e-05, "loss": 6.7273, "step": 12804 }, { "epoch": 0.5398739380652233, "grad_norm": 1.6576576232910156, "learning_rate": 4.555888691089666e-05, "loss": 6.7376, "step": 12805 }, { "epoch": 0.5399160992474229, "grad_norm": 3.7221078872680664, "learning_rate": 4.5558106124488585e-05, "loss": 7.6691, "step": 12806 }, { "epoch": 0.5399582604296225, "grad_norm": 4.782291889190674, "learning_rate": 4.5557325338080516e-05, "loss": 9.8147, "step": 12807 }, { "epoch": 0.540000421611822, "grad_norm": 2.122278928756714, "learning_rate": 4.555654455167245e-05, "loss": 7.8485, "step": 12808 }, { "epoch": 0.5400425827940215, "grad_norm": 2.4185068607330322, "learning_rate": 4.555576376526438e-05, "loss": 7.8877, "step": 12809 }, { "epoch": 0.5400847439762211, "grad_norm": 1.246107816696167, "learning_rate": 4.555498297885631e-05, "loss": 6.604, "step": 12810 }, { "epoch": 0.5401269051584207, "grad_norm": 2.771148920059204, "learning_rate": 4.5554202192448234e-05, "loss": 7.1498, "step": 12811 }, { "epoch": 0.5401690663406202, "grad_norm": 2.452979803085327, "learning_rate": 4.5553421406040165e-05, "loss": 6.9068, "step": 12812 }, { "epoch": 0.5402112275228197, "grad_norm": 1.5697319507598877, "learning_rate": 4.5552640619632097e-05, "loss": 7.2034, "step": 12813 }, { "epoch": 0.5402533887050193, "grad_norm": 1.5194453001022339, "learning_rate": 4.555185983322403e-05, "loss": 6.4676, "step": 12814 }, { "epoch": 0.5402955498872188, "grad_norm": 3.1535069942474365, "learning_rate": 4.555107904681595e-05, "loss": 8.191, "step": 12815 }, { "epoch": 0.5403377110694184, "grad_norm": 3.7132155895233154, "learning_rate": 4.5550298260407884e-05, "loss": 7.919, "step": 12816 }, { "epoch": 0.540379872251618, "grad_norm": 1.2198048830032349, "learning_rate": 4.5549517473999815e-05, "loss": 6.8148, "step": 12817 }, { "epoch": 0.5404220334338175, "grad_norm": 5.039782524108887, "learning_rate": 4.554873668759174e-05, "loss": 7.3469, "step": 12818 }, { "epoch": 0.540464194616017, "grad_norm": 1.5633183717727661, "learning_rate": 4.554795590118368e-05, "loss": 7.178, "step": 12819 }, { "epoch": 0.5405063557982166, "grad_norm": 1.5533522367477417, "learning_rate": 4.55471751147756e-05, "loss": 6.7583, "step": 12820 }, { "epoch": 0.5405485169804162, "grad_norm": 1.485514521598816, "learning_rate": 4.554639432836753e-05, "loss": 6.4213, "step": 12821 }, { "epoch": 0.5405906781626156, "grad_norm": 1.1987528800964355, "learning_rate": 4.5545613541959464e-05, "loss": 6.8855, "step": 12822 }, { "epoch": 0.5406328393448152, "grad_norm": 2.979158878326416, "learning_rate": 4.554483275555139e-05, "loss": 7.8431, "step": 12823 }, { "epoch": 0.5406750005270148, "grad_norm": 1.308321475982666, "learning_rate": 4.554405196914333e-05, "loss": 6.8909, "step": 12824 }, { "epoch": 0.5407171617092144, "grad_norm": 1.18715500831604, "learning_rate": 4.554327118273525e-05, "loss": 6.6454, "step": 12825 }, { "epoch": 0.5407593228914139, "grad_norm": 1.3708516359329224, "learning_rate": 4.554249039632718e-05, "loss": 7.1545, "step": 12826 }, { "epoch": 0.5408014840736134, "grad_norm": 2.1633808612823486, "learning_rate": 4.5541709609919114e-05, "loss": 6.6386, "step": 12827 }, { "epoch": 0.540843645255813, "grad_norm": 1.3163563013076782, "learning_rate": 4.5540928823511045e-05, "loss": 7.2345, "step": 12828 }, { "epoch": 0.5408858064380125, "grad_norm": 1.754185676574707, "learning_rate": 4.554014803710297e-05, "loss": 7.2698, "step": 12829 }, { "epoch": 0.5409279676202121, "grad_norm": 1.8270374536514282, "learning_rate": 4.55393672506949e-05, "loss": 6.5195, "step": 12830 }, { "epoch": 0.5409701288024116, "grad_norm": 1.2116749286651611, "learning_rate": 4.553858646428683e-05, "loss": 6.7083, "step": 12831 }, { "epoch": 0.5410122899846111, "grad_norm": 1.321959376335144, "learning_rate": 4.5537805677878763e-05, "loss": 6.6738, "step": 12832 }, { "epoch": 0.5410544511668107, "grad_norm": 1.0349578857421875, "learning_rate": 4.5537024891470695e-05, "loss": 6.7343, "step": 12833 }, { "epoch": 0.5410966123490103, "grad_norm": 2.395522117614746, "learning_rate": 4.553624410506262e-05, "loss": 7.155, "step": 12834 }, { "epoch": 0.5411387735312098, "grad_norm": 1.5821760892868042, "learning_rate": 4.553546331865455e-05, "loss": 6.7201, "step": 12835 }, { "epoch": 0.5411809347134093, "grad_norm": 0.9143993854522705, "learning_rate": 4.553468253224648e-05, "loss": 6.4653, "step": 12836 }, { "epoch": 0.5412230958956089, "grad_norm": 3.319615602493286, "learning_rate": 4.5533901745838406e-05, "loss": 8.2273, "step": 12837 }, { "epoch": 0.5412652570778085, "grad_norm": 1.9470531940460205, "learning_rate": 4.5533120959430344e-05, "loss": 7.2473, "step": 12838 }, { "epoch": 0.541307418260008, "grad_norm": 4.187618732452393, "learning_rate": 4.553234017302227e-05, "loss": 7.8246, "step": 12839 }, { "epoch": 0.5413495794422075, "grad_norm": 1.5944833755493164, "learning_rate": 4.55315593866142e-05, "loss": 7.2592, "step": 12840 }, { "epoch": 0.5413917406244071, "grad_norm": 5.790596961975098, "learning_rate": 4.553077860020613e-05, "loss": 7.3051, "step": 12841 }, { "epoch": 0.5414339018066067, "grad_norm": 2.0593721866607666, "learning_rate": 4.5529997813798056e-05, "loss": 6.8444, "step": 12842 }, { "epoch": 0.5414760629888062, "grad_norm": 1.5364034175872803, "learning_rate": 4.5529217027389994e-05, "loss": 7.0222, "step": 12843 }, { "epoch": 0.5415182241710057, "grad_norm": 1.6476751565933228, "learning_rate": 4.552843624098192e-05, "loss": 6.6575, "step": 12844 }, { "epoch": 0.5415603853532053, "grad_norm": 2.276567220687866, "learning_rate": 4.552765545457385e-05, "loss": 7.2896, "step": 12845 }, { "epoch": 0.5416025465354048, "grad_norm": 1.1006042957305908, "learning_rate": 4.552687466816578e-05, "loss": 6.5139, "step": 12846 }, { "epoch": 0.5416447077176044, "grad_norm": 2.255824327468872, "learning_rate": 4.552609388175771e-05, "loss": 6.3267, "step": 12847 }, { "epoch": 0.541686868899804, "grad_norm": 2.55664324760437, "learning_rate": 4.5525313095349636e-05, "loss": 7.3399, "step": 12848 }, { "epoch": 0.5417290300820035, "grad_norm": 1.3101427555084229, "learning_rate": 4.552453230894157e-05, "loss": 6.7611, "step": 12849 }, { "epoch": 0.541771191264203, "grad_norm": 2.119941234588623, "learning_rate": 4.55237515225335e-05, "loss": 7.3619, "step": 12850 }, { "epoch": 0.5418133524464026, "grad_norm": 1.1159557104110718, "learning_rate": 4.5522970736125423e-05, "loss": 6.8162, "step": 12851 }, { "epoch": 0.5418555136286022, "grad_norm": 2.4594645500183105, "learning_rate": 4.552218994971736e-05, "loss": 7.6274, "step": 12852 }, { "epoch": 0.5418976748108016, "grad_norm": 1.456046462059021, "learning_rate": 4.5521409163309286e-05, "loss": 7.0045, "step": 12853 }, { "epoch": 0.5419398359930012, "grad_norm": 2.458799362182617, "learning_rate": 4.552062837690122e-05, "loss": 7.651, "step": 12854 }, { "epoch": 0.5419819971752008, "grad_norm": 1.648191213607788, "learning_rate": 4.551984759049315e-05, "loss": 6.5476, "step": 12855 }, { "epoch": 0.5420241583574004, "grad_norm": 1.9647759199142456, "learning_rate": 4.551906680408507e-05, "loss": 7.5181, "step": 12856 }, { "epoch": 0.5420663195395999, "grad_norm": 1.5987557172775269, "learning_rate": 4.551828601767701e-05, "loss": 6.5026, "step": 12857 }, { "epoch": 0.5421084807217994, "grad_norm": 2.4267361164093018, "learning_rate": 4.5517505231268936e-05, "loss": 6.9643, "step": 12858 }, { "epoch": 0.542150641903999, "grad_norm": 2.134894609451294, "learning_rate": 4.551672444486087e-05, "loss": 7.2792, "step": 12859 }, { "epoch": 0.5421928030861985, "grad_norm": 1.7555230855941772, "learning_rate": 4.55159436584528e-05, "loss": 6.432, "step": 12860 }, { "epoch": 0.5422349642683981, "grad_norm": 3.422499418258667, "learning_rate": 4.551516287204472e-05, "loss": 8.0532, "step": 12861 }, { "epoch": 0.5422771254505976, "grad_norm": 1.2374330759048462, "learning_rate": 4.5514382085636654e-05, "loss": 6.4786, "step": 12862 }, { "epoch": 0.5423192866327972, "grad_norm": 1.3482600450515747, "learning_rate": 4.5513601299228585e-05, "loss": 6.6478, "step": 12863 }, { "epoch": 0.5423614478149967, "grad_norm": 1.903304100036621, "learning_rate": 4.5512820512820516e-05, "loss": 7.0458, "step": 12864 }, { "epoch": 0.5424036089971963, "grad_norm": 3.0403199195861816, "learning_rate": 4.551203972641244e-05, "loss": 7.905, "step": 12865 }, { "epoch": 0.5424457701793959, "grad_norm": 0.9780241847038269, "learning_rate": 4.551125894000437e-05, "loss": 6.5465, "step": 12866 }, { "epoch": 0.5424879313615953, "grad_norm": 1.484257459640503, "learning_rate": 4.55104781535963e-05, "loss": 7.3991, "step": 12867 }, { "epoch": 0.5425300925437949, "grad_norm": 1.7956792116165161, "learning_rate": 4.5509697367188235e-05, "loss": 7.1473, "step": 12868 }, { "epoch": 0.5425722537259945, "grad_norm": 2.4448564052581787, "learning_rate": 4.5508916580780166e-05, "loss": 8.0433, "step": 12869 }, { "epoch": 0.5426144149081941, "grad_norm": 1.7946252822875977, "learning_rate": 4.550813579437209e-05, "loss": 7.1582, "step": 12870 }, { "epoch": 0.5426565760903935, "grad_norm": 3.3004038333892822, "learning_rate": 4.550735500796403e-05, "loss": 7.7692, "step": 12871 }, { "epoch": 0.5426987372725931, "grad_norm": 1.580535888671875, "learning_rate": 4.550657422155595e-05, "loss": 6.5591, "step": 12872 }, { "epoch": 0.5427408984547927, "grad_norm": 1.4554964303970337, "learning_rate": 4.5505793435147884e-05, "loss": 7.1315, "step": 12873 }, { "epoch": 0.5427830596369922, "grad_norm": 1.3363608121871948, "learning_rate": 4.5505012648739815e-05, "loss": 7.0008, "step": 12874 }, { "epoch": 0.5428252208191918, "grad_norm": 1.5878689289093018, "learning_rate": 4.550423186233174e-05, "loss": 7.2813, "step": 12875 }, { "epoch": 0.5428673820013913, "grad_norm": 2.2091009616851807, "learning_rate": 4.550345107592367e-05, "loss": 7.4395, "step": 12876 }, { "epoch": 0.5429095431835909, "grad_norm": 2.024409294128418, "learning_rate": 4.55026702895156e-05, "loss": 6.946, "step": 12877 }, { "epoch": 0.5429517043657904, "grad_norm": 4.161984443664551, "learning_rate": 4.5501889503107534e-05, "loss": 8.1256, "step": 12878 }, { "epoch": 0.54299386554799, "grad_norm": 2.530761480331421, "learning_rate": 4.5501108716699465e-05, "loss": 7.7475, "step": 12879 }, { "epoch": 0.5430360267301895, "grad_norm": 1.3102918863296509, "learning_rate": 4.550032793029139e-05, "loss": 6.935, "step": 12880 }, { "epoch": 0.543078187912389, "grad_norm": 2.0254640579223633, "learning_rate": 4.549954714388332e-05, "loss": 6.5501, "step": 12881 }, { "epoch": 0.5431203490945886, "grad_norm": 1.6472656726837158, "learning_rate": 4.549876635747525e-05, "loss": 6.6427, "step": 12882 }, { "epoch": 0.5431625102767882, "grad_norm": 3.068763494491577, "learning_rate": 4.549798557106718e-05, "loss": 6.3978, "step": 12883 }, { "epoch": 0.5432046714589878, "grad_norm": 1.6427992582321167, "learning_rate": 4.549720478465911e-05, "loss": 6.5921, "step": 12884 }, { "epoch": 0.5432468326411872, "grad_norm": 1.3650474548339844, "learning_rate": 4.549642399825104e-05, "loss": 7.1638, "step": 12885 }, { "epoch": 0.5432889938233868, "grad_norm": 1.2959667444229126, "learning_rate": 4.549564321184297e-05, "loss": 6.8695, "step": 12886 }, { "epoch": 0.5433311550055864, "grad_norm": 1.5901833772659302, "learning_rate": 4.5494862425434895e-05, "loss": 6.5092, "step": 12887 }, { "epoch": 0.543373316187786, "grad_norm": 2.5626347064971924, "learning_rate": 4.549408163902683e-05, "loss": 6.635, "step": 12888 }, { "epoch": 0.5434154773699854, "grad_norm": 0.9600765109062195, "learning_rate": 4.549330085261876e-05, "loss": 6.7996, "step": 12889 }, { "epoch": 0.543457638552185, "grad_norm": 1.225318193435669, "learning_rate": 4.5492520066210695e-05, "loss": 6.7262, "step": 12890 }, { "epoch": 0.5434997997343846, "grad_norm": 3.09504771232605, "learning_rate": 4.549173927980262e-05, "loss": 8.469, "step": 12891 }, { "epoch": 0.5435419609165841, "grad_norm": 0.8227521777153015, "learning_rate": 4.549095849339455e-05, "loss": 6.5524, "step": 12892 }, { "epoch": 0.5435841220987836, "grad_norm": 1.646539568901062, "learning_rate": 4.549017770698648e-05, "loss": 6.3411, "step": 12893 }, { "epoch": 0.5436262832809832, "grad_norm": 1.9110575914382935, "learning_rate": 4.548939692057841e-05, "loss": 6.8181, "step": 12894 }, { "epoch": 0.5436684444631827, "grad_norm": 1.4973760843276978, "learning_rate": 4.548861613417034e-05, "loss": 7.1448, "step": 12895 }, { "epoch": 0.5437106056453823, "grad_norm": 1.8120766878128052, "learning_rate": 4.548783534776227e-05, "loss": 6.4548, "step": 12896 }, { "epoch": 0.5437527668275819, "grad_norm": 1.7806040048599243, "learning_rate": 4.54870545613542e-05, "loss": 6.6088, "step": 12897 }, { "epoch": 0.5437949280097814, "grad_norm": 4.08356237411499, "learning_rate": 4.5486273774946125e-05, "loss": 7.7075, "step": 12898 }, { "epoch": 0.5438370891919809, "grad_norm": 2.152538537979126, "learning_rate": 4.5485492988538056e-05, "loss": 6.982, "step": 12899 }, { "epoch": 0.5438792503741805, "grad_norm": 3.7563328742980957, "learning_rate": 4.548471220212999e-05, "loss": 7.5583, "step": 12900 }, { "epoch": 0.5439214115563801, "grad_norm": 1.4365476369857788, "learning_rate": 4.548393141572192e-05, "loss": 6.9053, "step": 12901 }, { "epoch": 0.5439635727385795, "grad_norm": 3.5852210521698, "learning_rate": 4.548315062931385e-05, "loss": 7.1073, "step": 12902 }, { "epoch": 0.5440057339207791, "grad_norm": 2.9450440406799316, "learning_rate": 4.5482369842905774e-05, "loss": 7.5629, "step": 12903 }, { "epoch": 0.5440478951029787, "grad_norm": 1.2741563320159912, "learning_rate": 4.5481589056497706e-05, "loss": 6.4592, "step": 12904 }, { "epoch": 0.5440900562851783, "grad_norm": 3.6164817810058594, "learning_rate": 4.548080827008964e-05, "loss": 8.1554, "step": 12905 }, { "epoch": 0.5441322174673778, "grad_norm": 4.0912065505981445, "learning_rate": 4.548002748368156e-05, "loss": 7.6412, "step": 12906 }, { "epoch": 0.5441743786495773, "grad_norm": 1.922993540763855, "learning_rate": 4.54792466972735e-05, "loss": 7.4228, "step": 12907 }, { "epoch": 0.5442165398317769, "grad_norm": 2.1384973526000977, "learning_rate": 4.5478465910865424e-05, "loss": 6.9904, "step": 12908 }, { "epoch": 0.5442587010139764, "grad_norm": 2.308502435684204, "learning_rate": 4.5477685124457355e-05, "loss": 7.2925, "step": 12909 }, { "epoch": 0.544300862196176, "grad_norm": 3.5212526321411133, "learning_rate": 4.5476904338049286e-05, "loss": 6.9343, "step": 12910 }, { "epoch": 0.5443430233783755, "grad_norm": 1.706459879875183, "learning_rate": 4.547612355164121e-05, "loss": 7.0256, "step": 12911 }, { "epoch": 0.544385184560575, "grad_norm": 5.1503167152404785, "learning_rate": 4.547534276523315e-05, "loss": 7.4523, "step": 12912 }, { "epoch": 0.5444273457427746, "grad_norm": 2.082460403442383, "learning_rate": 4.5474561978825073e-05, "loss": 6.8242, "step": 12913 }, { "epoch": 0.5444695069249742, "grad_norm": 1.6478416919708252, "learning_rate": 4.5473781192417005e-05, "loss": 7.3931, "step": 12914 }, { "epoch": 0.5445116681071738, "grad_norm": 3.5553736686706543, "learning_rate": 4.5473000406008936e-05, "loss": 7.1731, "step": 12915 }, { "epoch": 0.5445538292893732, "grad_norm": 1.9907528162002563, "learning_rate": 4.547221961960087e-05, "loss": 6.4517, "step": 12916 }, { "epoch": 0.5445959904715728, "grad_norm": 1.442659854888916, "learning_rate": 4.547143883319279e-05, "loss": 6.5581, "step": 12917 }, { "epoch": 0.5446381516537724, "grad_norm": 2.192948579788208, "learning_rate": 4.547065804678472e-05, "loss": 7.1531, "step": 12918 }, { "epoch": 0.544680312835972, "grad_norm": 3.0581142902374268, "learning_rate": 4.5469877260376654e-05, "loss": 7.5481, "step": 12919 }, { "epoch": 0.5447224740181714, "grad_norm": 2.66974139213562, "learning_rate": 4.546909647396858e-05, "loss": 7.9631, "step": 12920 }, { "epoch": 0.544764635200371, "grad_norm": 2.310854911804199, "learning_rate": 4.546831568756052e-05, "loss": 6.6719, "step": 12921 }, { "epoch": 0.5448067963825706, "grad_norm": 1.7395738363265991, "learning_rate": 4.546753490115244e-05, "loss": 6.9887, "step": 12922 }, { "epoch": 0.5448489575647701, "grad_norm": 1.936892032623291, "learning_rate": 4.546675411474437e-05, "loss": 7.4512, "step": 12923 }, { "epoch": 0.5448911187469697, "grad_norm": 1.996708869934082, "learning_rate": 4.5465973328336304e-05, "loss": 6.5168, "step": 12924 }, { "epoch": 0.5449332799291692, "grad_norm": 1.6753616333007812, "learning_rate": 4.546519254192823e-05, "loss": 6.8558, "step": 12925 }, { "epoch": 0.5449754411113688, "grad_norm": 2.1477644443511963, "learning_rate": 4.5464411755520166e-05, "loss": 8.0118, "step": 12926 }, { "epoch": 0.5450176022935683, "grad_norm": 1.3091281652450562, "learning_rate": 4.546363096911209e-05, "loss": 6.3453, "step": 12927 }, { "epoch": 0.5450597634757679, "grad_norm": 1.85468590259552, "learning_rate": 4.546285018270402e-05, "loss": 7.3212, "step": 12928 }, { "epoch": 0.5451019246579674, "grad_norm": 2.0488715171813965, "learning_rate": 4.546206939629595e-05, "loss": 7.345, "step": 12929 }, { "epoch": 0.5451440858401669, "grad_norm": 0.9753944873809814, "learning_rate": 4.546128860988788e-05, "loss": 6.6259, "step": 12930 }, { "epoch": 0.5451862470223665, "grad_norm": 2.4154603481292725, "learning_rate": 4.546050782347981e-05, "loss": 6.4674, "step": 12931 }, { "epoch": 0.5452284082045661, "grad_norm": 1.1812024116516113, "learning_rate": 4.545972703707174e-05, "loss": 7.142, "step": 12932 }, { "epoch": 0.5452705693867657, "grad_norm": 1.83653724193573, "learning_rate": 4.545894625066367e-05, "loss": 6.9309, "step": 12933 }, { "epoch": 0.5453127305689651, "grad_norm": 1.222940444946289, "learning_rate": 4.54581654642556e-05, "loss": 6.5809, "step": 12934 }, { "epoch": 0.5453548917511647, "grad_norm": 1.273835301399231, "learning_rate": 4.5457384677847534e-05, "loss": 6.7261, "step": 12935 }, { "epoch": 0.5453970529333643, "grad_norm": 0.8834477663040161, "learning_rate": 4.545660389143946e-05, "loss": 6.8604, "step": 12936 }, { "epoch": 0.5454392141155638, "grad_norm": 1.1649470329284668, "learning_rate": 4.545582310503139e-05, "loss": 6.5561, "step": 12937 }, { "epoch": 0.5454813752977633, "grad_norm": 4.2700042724609375, "learning_rate": 4.545504231862332e-05, "loss": 8.1994, "step": 12938 }, { "epoch": 0.5455235364799629, "grad_norm": 1.1062374114990234, "learning_rate": 4.5454261532215246e-05, "loss": 6.614, "step": 12939 }, { "epoch": 0.5455656976621625, "grad_norm": 2.0611777305603027, "learning_rate": 4.5453480745807184e-05, "loss": 7.5084, "step": 12940 }, { "epoch": 0.545607858844362, "grad_norm": 2.194356918334961, "learning_rate": 4.545269995939911e-05, "loss": 7.2051, "step": 12941 }, { "epoch": 0.5456500200265616, "grad_norm": 2.578749895095825, "learning_rate": 4.545191917299104e-05, "loss": 6.8815, "step": 12942 }, { "epoch": 0.5456921812087611, "grad_norm": 1.9386550188064575, "learning_rate": 4.545113838658297e-05, "loss": 6.9736, "step": 12943 }, { "epoch": 0.5457343423909606, "grad_norm": 2.3607065677642822, "learning_rate": 4.5450357600174895e-05, "loss": 6.3763, "step": 12944 }, { "epoch": 0.5457765035731602, "grad_norm": 2.837766170501709, "learning_rate": 4.5449576813766826e-05, "loss": 6.6421, "step": 12945 }, { "epoch": 0.5458186647553598, "grad_norm": 2.3484582901000977, "learning_rate": 4.544879602735876e-05, "loss": 7.2351, "step": 12946 }, { "epoch": 0.5458608259375592, "grad_norm": 1.3979990482330322, "learning_rate": 4.544801524095069e-05, "loss": 6.7671, "step": 12947 }, { "epoch": 0.5459029871197588, "grad_norm": 0.9187667369842529, "learning_rate": 4.544723445454262e-05, "loss": 6.5393, "step": 12948 }, { "epoch": 0.5459451483019584, "grad_norm": 2.622039556503296, "learning_rate": 4.5446453668134545e-05, "loss": 8.1903, "step": 12949 }, { "epoch": 0.545987309484158, "grad_norm": 2.583904266357422, "learning_rate": 4.5445672881726476e-05, "loss": 7.1285, "step": 12950 }, { "epoch": 0.5460294706663574, "grad_norm": 2.3251748085021973, "learning_rate": 4.544489209531841e-05, "loss": 6.4193, "step": 12951 }, { "epoch": 0.546071631848557, "grad_norm": 1.9049652814865112, "learning_rate": 4.544411130891034e-05, "loss": 6.7249, "step": 12952 }, { "epoch": 0.5461137930307566, "grad_norm": 1.316055417060852, "learning_rate": 4.544333052250226e-05, "loss": 6.5518, "step": 12953 }, { "epoch": 0.5461559542129562, "grad_norm": 1.5139250755310059, "learning_rate": 4.54425497360942e-05, "loss": 7.4801, "step": 12954 }, { "epoch": 0.5461981153951557, "grad_norm": 2.3484911918640137, "learning_rate": 4.5441768949686125e-05, "loss": 7.3221, "step": 12955 }, { "epoch": 0.5462402765773552, "grad_norm": 1.9047523736953735, "learning_rate": 4.544098816327805e-05, "loss": 6.3221, "step": 12956 }, { "epoch": 0.5462824377595548, "grad_norm": 1.9286442995071411, "learning_rate": 4.544020737686999e-05, "loss": 7.3315, "step": 12957 }, { "epoch": 0.5463245989417543, "grad_norm": 2.874717950820923, "learning_rate": 4.543942659046191e-05, "loss": 7.2485, "step": 12958 }, { "epoch": 0.5463667601239539, "grad_norm": 2.122282028198242, "learning_rate": 4.543864580405385e-05, "loss": 7.4843, "step": 12959 }, { "epoch": 0.5464089213061534, "grad_norm": 1.7173296213150024, "learning_rate": 4.5437865017645775e-05, "loss": 6.4534, "step": 12960 }, { "epoch": 0.546451082488353, "grad_norm": 1.6301319599151611, "learning_rate": 4.5437084231237706e-05, "loss": 6.9929, "step": 12961 }, { "epoch": 0.5464932436705525, "grad_norm": 1.274715781211853, "learning_rate": 4.543630344482964e-05, "loss": 7.3916, "step": 12962 }, { "epoch": 0.5465354048527521, "grad_norm": 1.0548498630523682, "learning_rate": 4.543552265842156e-05, "loss": 6.4654, "step": 12963 }, { "epoch": 0.5465775660349517, "grad_norm": 1.1774471998214722, "learning_rate": 4.543474187201349e-05, "loss": 6.9073, "step": 12964 }, { "epoch": 0.5466197272171511, "grad_norm": 2.1578328609466553, "learning_rate": 4.5433961085605424e-05, "loss": 7.3339, "step": 12965 }, { "epoch": 0.5466618883993507, "grad_norm": 1.6240766048431396, "learning_rate": 4.5433180299197356e-05, "loss": 6.4613, "step": 12966 }, { "epoch": 0.5467040495815503, "grad_norm": 1.263226866722107, "learning_rate": 4.543239951278928e-05, "loss": 6.6031, "step": 12967 }, { "epoch": 0.5467462107637499, "grad_norm": 2.579573631286621, "learning_rate": 4.543161872638121e-05, "loss": 7.967, "step": 12968 }, { "epoch": 0.5467883719459493, "grad_norm": 2.2176337242126465, "learning_rate": 4.543083793997314e-05, "loss": 6.542, "step": 12969 }, { "epoch": 0.5468305331281489, "grad_norm": 2.7677862644195557, "learning_rate": 4.5430057153565074e-05, "loss": 7.3805, "step": 12970 }, { "epoch": 0.5468726943103485, "grad_norm": 2.2745442390441895, "learning_rate": 4.5429276367157005e-05, "loss": 7.8216, "step": 12971 }, { "epoch": 0.546914855492548, "grad_norm": 3.349886894226074, "learning_rate": 4.542849558074893e-05, "loss": 6.3929, "step": 12972 }, { "epoch": 0.5469570166747476, "grad_norm": 1.3585398197174072, "learning_rate": 4.542771479434086e-05, "loss": 7.1816, "step": 12973 }, { "epoch": 0.5469991778569471, "grad_norm": 2.5066304206848145, "learning_rate": 4.542693400793279e-05, "loss": 7.7925, "step": 12974 }, { "epoch": 0.5470413390391466, "grad_norm": 1.5959625244140625, "learning_rate": 4.542615322152472e-05, "loss": 6.35, "step": 12975 }, { "epoch": 0.5470835002213462, "grad_norm": 1.1388219594955444, "learning_rate": 4.5425372435116655e-05, "loss": 6.6974, "step": 12976 }, { "epoch": 0.5471256614035458, "grad_norm": 1.216471791267395, "learning_rate": 4.542459164870858e-05, "loss": 6.5765, "step": 12977 }, { "epoch": 0.5471678225857453, "grad_norm": 1.4410189390182495, "learning_rate": 4.542381086230051e-05, "loss": 6.6497, "step": 12978 }, { "epoch": 0.5472099837679448, "grad_norm": 2.400024175643921, "learning_rate": 4.542303007589244e-05, "loss": 7.7859, "step": 12979 }, { "epoch": 0.5472521449501444, "grad_norm": 2.746631622314453, "learning_rate": 4.542224928948437e-05, "loss": 7.0981, "step": 12980 }, { "epoch": 0.547294306132344, "grad_norm": 1.2286450862884521, "learning_rate": 4.5421468503076304e-05, "loss": 6.8906, "step": 12981 }, { "epoch": 0.5473364673145436, "grad_norm": 2.44197940826416, "learning_rate": 4.542068771666823e-05, "loss": 6.5712, "step": 12982 }, { "epoch": 0.547378628496743, "grad_norm": 3.2295985221862793, "learning_rate": 4.541990693026016e-05, "loss": 8.2711, "step": 12983 }, { "epoch": 0.5474207896789426, "grad_norm": 1.6344375610351562, "learning_rate": 4.541912614385209e-05, "loss": 7.2608, "step": 12984 }, { "epoch": 0.5474629508611422, "grad_norm": 1.7586359977722168, "learning_rate": 4.541834535744402e-05, "loss": 6.4897, "step": 12985 }, { "epoch": 0.5475051120433417, "grad_norm": 1.7034412622451782, "learning_rate": 4.541756457103595e-05, "loss": 7.1178, "step": 12986 }, { "epoch": 0.5475472732255412, "grad_norm": 1.334101676940918, "learning_rate": 4.541678378462788e-05, "loss": 6.541, "step": 12987 }, { "epoch": 0.5475894344077408, "grad_norm": 1.697593331336975, "learning_rate": 4.541600299821981e-05, "loss": 7.3234, "step": 12988 }, { "epoch": 0.5476315955899403, "grad_norm": 1.0102953910827637, "learning_rate": 4.5415222211811734e-05, "loss": 6.6638, "step": 12989 }, { "epoch": 0.5476737567721399, "grad_norm": 2.128260374069214, "learning_rate": 4.541444142540367e-05, "loss": 7.1515, "step": 12990 }, { "epoch": 0.5477159179543395, "grad_norm": 1.1550148725509644, "learning_rate": 4.5413660638995597e-05, "loss": 6.4726, "step": 12991 }, { "epoch": 0.547758079136539, "grad_norm": 1.5907871723175049, "learning_rate": 4.541287985258753e-05, "loss": 6.4208, "step": 12992 }, { "epoch": 0.5478002403187385, "grad_norm": 1.041866660118103, "learning_rate": 4.541209906617946e-05, "loss": 7.0818, "step": 12993 }, { "epoch": 0.5478424015009381, "grad_norm": 1.571272850036621, "learning_rate": 4.5411318279771384e-05, "loss": 6.6325, "step": 12994 }, { "epoch": 0.5478845626831377, "grad_norm": 4.173155307769775, "learning_rate": 4.541053749336332e-05, "loss": 7.5546, "step": 12995 }, { "epoch": 0.5479267238653371, "grad_norm": 1.8885313272476196, "learning_rate": 4.5409756706955246e-05, "loss": 6.9551, "step": 12996 }, { "epoch": 0.5479688850475367, "grad_norm": 1.7026392221450806, "learning_rate": 4.540897592054718e-05, "loss": 7.118, "step": 12997 }, { "epoch": 0.5480110462297363, "grad_norm": 1.1078277826309204, "learning_rate": 4.540819513413911e-05, "loss": 7.0424, "step": 12998 }, { "epoch": 0.5480532074119359, "grad_norm": 1.528984785079956, "learning_rate": 4.540741434773104e-05, "loss": 6.7848, "step": 12999 }, { "epoch": 0.5480953685941354, "grad_norm": 1.3253852128982544, "learning_rate": 4.5406633561322964e-05, "loss": 7.0418, "step": 13000 }, { "epoch": 0.5481375297763349, "grad_norm": 2.488105535507202, "learning_rate": 4.5405852774914896e-05, "loss": 7.3663, "step": 13001 }, { "epoch": 0.5481796909585345, "grad_norm": 1.6205180883407593, "learning_rate": 4.540507198850683e-05, "loss": 7.2198, "step": 13002 }, { "epoch": 0.548221852140734, "grad_norm": 1.1509168148040771, "learning_rate": 4.540429120209876e-05, "loss": 6.8154, "step": 13003 }, { "epoch": 0.5482640133229336, "grad_norm": 2.706571340560913, "learning_rate": 4.540351041569069e-05, "loss": 6.295, "step": 13004 }, { "epoch": 0.5483061745051331, "grad_norm": 1.3640130758285522, "learning_rate": 4.5402729629282614e-05, "loss": 7.1653, "step": 13005 }, { "epoch": 0.5483483356873327, "grad_norm": 1.4903299808502197, "learning_rate": 4.5401948842874545e-05, "loss": 6.4631, "step": 13006 }, { "epoch": 0.5483904968695322, "grad_norm": 2.587414503097534, "learning_rate": 4.5401168056466476e-05, "loss": 7.7952, "step": 13007 }, { "epoch": 0.5484326580517318, "grad_norm": 3.292342185974121, "learning_rate": 4.54003872700584e-05, "loss": 7.6695, "step": 13008 }, { "epoch": 0.5484748192339313, "grad_norm": 1.5840860605239868, "learning_rate": 4.539960648365034e-05, "loss": 7.3949, "step": 13009 }, { "epoch": 0.5485169804161308, "grad_norm": 1.103418231010437, "learning_rate": 4.5398825697242263e-05, "loss": 6.5356, "step": 13010 }, { "epoch": 0.5485591415983304, "grad_norm": 1.2967708110809326, "learning_rate": 4.5398044910834195e-05, "loss": 7.0301, "step": 13011 }, { "epoch": 0.54860130278053, "grad_norm": 1.5158675909042358, "learning_rate": 4.5397264124426126e-05, "loss": 6.2326, "step": 13012 }, { "epoch": 0.5486434639627296, "grad_norm": 2.9651622772216797, "learning_rate": 4.539648333801805e-05, "loss": 8.04, "step": 13013 }, { "epoch": 0.548685625144929, "grad_norm": 1.1651312112808228, "learning_rate": 4.539570255160998e-05, "loss": 6.531, "step": 13014 }, { "epoch": 0.5487277863271286, "grad_norm": 1.2102452516555786, "learning_rate": 4.539492176520191e-05, "loss": 6.8227, "step": 13015 }, { "epoch": 0.5487699475093282, "grad_norm": 4.63826322555542, "learning_rate": 4.5394140978793844e-05, "loss": 7.3573, "step": 13016 }, { "epoch": 0.5488121086915277, "grad_norm": 1.371631145477295, "learning_rate": 4.5393360192385775e-05, "loss": 6.4511, "step": 13017 }, { "epoch": 0.5488542698737272, "grad_norm": 1.4727275371551514, "learning_rate": 4.53925794059777e-05, "loss": 6.4015, "step": 13018 }, { "epoch": 0.5488964310559268, "grad_norm": 3.223235607147217, "learning_rate": 4.539179861956963e-05, "loss": 8.0941, "step": 13019 }, { "epoch": 0.5489385922381264, "grad_norm": 1.873768925666809, "learning_rate": 4.539101783316156e-05, "loss": 7.1778, "step": 13020 }, { "epoch": 0.5489807534203259, "grad_norm": 1.0121424198150635, "learning_rate": 4.5390237046753494e-05, "loss": 6.7715, "step": 13021 }, { "epoch": 0.5490229146025255, "grad_norm": 3.2204864025115967, "learning_rate": 4.538945626034542e-05, "loss": 7.7651, "step": 13022 }, { "epoch": 0.549065075784725, "grad_norm": 1.9348220825195312, "learning_rate": 4.5388675473937356e-05, "loss": 6.6076, "step": 13023 }, { "epoch": 0.5491072369669245, "grad_norm": 1.0573421716690063, "learning_rate": 4.538789468752928e-05, "loss": 6.418, "step": 13024 }, { "epoch": 0.5491493981491241, "grad_norm": 1.7575838565826416, "learning_rate": 4.538711390112121e-05, "loss": 7.0778, "step": 13025 }, { "epoch": 0.5491915593313237, "grad_norm": 2.6611804962158203, "learning_rate": 4.538633311471314e-05, "loss": 7.8079, "step": 13026 }, { "epoch": 0.5492337205135231, "grad_norm": 2.6626837253570557, "learning_rate": 4.538555232830507e-05, "loss": 7.2837, "step": 13027 }, { "epoch": 0.5492758816957227, "grad_norm": 1.2049400806427002, "learning_rate": 4.5384771541897006e-05, "loss": 7.1016, "step": 13028 }, { "epoch": 0.5493180428779223, "grad_norm": 1.6595304012298584, "learning_rate": 4.538399075548893e-05, "loss": 6.5294, "step": 13029 }, { "epoch": 0.5493602040601219, "grad_norm": 2.6392197608947754, "learning_rate": 4.538320996908086e-05, "loss": 7.6626, "step": 13030 }, { "epoch": 0.5494023652423214, "grad_norm": 0.8864319920539856, "learning_rate": 4.538242918267279e-05, "loss": 6.564, "step": 13031 }, { "epoch": 0.5494445264245209, "grad_norm": 1.4230425357818604, "learning_rate": 4.538164839626472e-05, "loss": 6.8276, "step": 13032 }, { "epoch": 0.5494866876067205, "grad_norm": 1.1782350540161133, "learning_rate": 4.538086760985665e-05, "loss": 6.9503, "step": 13033 }, { "epoch": 0.54952884878892, "grad_norm": 1.478691816329956, "learning_rate": 4.538008682344858e-05, "loss": 7.0269, "step": 13034 }, { "epoch": 0.5495710099711196, "grad_norm": 1.2854444980621338, "learning_rate": 4.537930603704051e-05, "loss": 7.2707, "step": 13035 }, { "epoch": 0.5496131711533191, "grad_norm": 1.4517101049423218, "learning_rate": 4.5378525250632436e-05, "loss": 7.2451, "step": 13036 }, { "epoch": 0.5496553323355187, "grad_norm": 1.1688637733459473, "learning_rate": 4.537774446422437e-05, "loss": 6.6221, "step": 13037 }, { "epoch": 0.5496974935177182, "grad_norm": 1.6906046867370605, "learning_rate": 4.53769636778163e-05, "loss": 6.5357, "step": 13038 }, { "epoch": 0.5497396546999178, "grad_norm": 1.3072584867477417, "learning_rate": 4.537618289140823e-05, "loss": 6.7073, "step": 13039 }, { "epoch": 0.5497818158821174, "grad_norm": 2.8709049224853516, "learning_rate": 4.537540210500016e-05, "loss": 6.3541, "step": 13040 }, { "epoch": 0.5498239770643168, "grad_norm": 2.573089599609375, "learning_rate": 4.5374621318592085e-05, "loss": 7.57, "step": 13041 }, { "epoch": 0.5498661382465164, "grad_norm": 1.4714356660842896, "learning_rate": 4.537384053218402e-05, "loss": 6.3962, "step": 13042 }, { "epoch": 0.549908299428716, "grad_norm": 2.9770917892456055, "learning_rate": 4.537305974577595e-05, "loss": 7.3839, "step": 13043 }, { "epoch": 0.5499504606109156, "grad_norm": 1.2050156593322754, "learning_rate": 4.537227895936788e-05, "loss": 6.5776, "step": 13044 }, { "epoch": 0.549992621793115, "grad_norm": 1.6720333099365234, "learning_rate": 4.537149817295981e-05, "loss": 7.2969, "step": 13045 }, { "epoch": 0.5500347829753146, "grad_norm": 3.605968713760376, "learning_rate": 4.5370717386551735e-05, "loss": 7.8372, "step": 13046 }, { "epoch": 0.5500769441575142, "grad_norm": 3.1982004642486572, "learning_rate": 4.5369936600143666e-05, "loss": 8.2221, "step": 13047 }, { "epoch": 0.5501191053397138, "grad_norm": 1.4572569131851196, "learning_rate": 4.53691558137356e-05, "loss": 6.4186, "step": 13048 }, { "epoch": 0.5501612665219133, "grad_norm": 2.2109529972076416, "learning_rate": 4.536837502732753e-05, "loss": 6.6675, "step": 13049 }, { "epoch": 0.5502034277041128, "grad_norm": 1.7416049242019653, "learning_rate": 4.536759424091946e-05, "loss": 6.3646, "step": 13050 }, { "epoch": 0.5502455888863124, "grad_norm": 1.5008302927017212, "learning_rate": 4.5366813454511384e-05, "loss": 6.6687, "step": 13051 }, { "epoch": 0.5502877500685119, "grad_norm": 1.4525542259216309, "learning_rate": 4.5366032668103315e-05, "loss": 6.3755, "step": 13052 }, { "epoch": 0.5503299112507115, "grad_norm": 2.3321259021759033, "learning_rate": 4.5365251881695247e-05, "loss": 6.2795, "step": 13053 }, { "epoch": 0.550372072432911, "grad_norm": 0.9882248044013977, "learning_rate": 4.536447109528718e-05, "loss": 6.5477, "step": 13054 }, { "epoch": 0.5504142336151105, "grad_norm": 2.483978509902954, "learning_rate": 4.53636903088791e-05, "loss": 6.4493, "step": 13055 }, { "epoch": 0.5504563947973101, "grad_norm": 1.398753046989441, "learning_rate": 4.5362909522471034e-05, "loss": 6.5915, "step": 13056 }, { "epoch": 0.5504985559795097, "grad_norm": 1.0981855392456055, "learning_rate": 4.5362128736062965e-05, "loss": 6.4243, "step": 13057 }, { "epoch": 0.5505407171617092, "grad_norm": 2.8652892112731934, "learning_rate": 4.536134794965489e-05, "loss": 6.3303, "step": 13058 }, { "epoch": 0.5505828783439087, "grad_norm": 3.8636794090270996, "learning_rate": 4.536056716324683e-05, "loss": 7.6745, "step": 13059 }, { "epoch": 0.5506250395261083, "grad_norm": 1.2919045686721802, "learning_rate": 4.535978637683875e-05, "loss": 7.3581, "step": 13060 }, { "epoch": 0.5506672007083079, "grad_norm": 1.531520128250122, "learning_rate": 4.535900559043069e-05, "loss": 6.7742, "step": 13061 }, { "epoch": 0.5507093618905075, "grad_norm": 2.4151570796966553, "learning_rate": 4.5358224804022614e-05, "loss": 6.2083, "step": 13062 }, { "epoch": 0.5507515230727069, "grad_norm": 2.1755967140197754, "learning_rate": 4.535744401761454e-05, "loss": 6.5828, "step": 13063 }, { "epoch": 0.5507936842549065, "grad_norm": 2.1405439376831055, "learning_rate": 4.535666323120648e-05, "loss": 6.736, "step": 13064 }, { "epoch": 0.5508358454371061, "grad_norm": 1.8202033042907715, "learning_rate": 4.53558824447984e-05, "loss": 7.2556, "step": 13065 }, { "epoch": 0.5508780066193056, "grad_norm": 3.4071102142333984, "learning_rate": 4.535510165839033e-05, "loss": 8.0304, "step": 13066 }, { "epoch": 0.5509201678015051, "grad_norm": 1.584818959236145, "learning_rate": 4.5354320871982264e-05, "loss": 6.749, "step": 13067 }, { "epoch": 0.5509623289837047, "grad_norm": 1.3633863925933838, "learning_rate": 4.5353540085574195e-05, "loss": 6.619, "step": 13068 }, { "epoch": 0.5510044901659042, "grad_norm": 1.9064579010009766, "learning_rate": 4.535275929916612e-05, "loss": 6.6945, "step": 13069 }, { "epoch": 0.5510466513481038, "grad_norm": 1.2740802764892578, "learning_rate": 4.535197851275805e-05, "loss": 6.6095, "step": 13070 }, { "epoch": 0.5510888125303034, "grad_norm": 1.3082644939422607, "learning_rate": 4.535119772634998e-05, "loss": 6.7052, "step": 13071 }, { "epoch": 0.5511309737125029, "grad_norm": 1.1514142751693726, "learning_rate": 4.5350416939941913e-05, "loss": 6.7102, "step": 13072 }, { "epoch": 0.5511731348947024, "grad_norm": 2.0211877822875977, "learning_rate": 4.5349636153533845e-05, "loss": 6.7151, "step": 13073 }, { "epoch": 0.551215296076902, "grad_norm": 1.3708044290542603, "learning_rate": 4.534885536712577e-05, "loss": 7.1366, "step": 13074 }, { "epoch": 0.5512574572591016, "grad_norm": 2.8284146785736084, "learning_rate": 4.53480745807177e-05, "loss": 7.974, "step": 13075 }, { "epoch": 0.551299618441301, "grad_norm": 1.9429510831832886, "learning_rate": 4.534729379430963e-05, "loss": 6.3104, "step": 13076 }, { "epoch": 0.5513417796235006, "grad_norm": 2.2702200412750244, "learning_rate": 4.5346513007901556e-05, "loss": 7.9233, "step": 13077 }, { "epoch": 0.5513839408057002, "grad_norm": 1.9134601354599, "learning_rate": 4.5345732221493494e-05, "loss": 8.0884, "step": 13078 }, { "epoch": 0.5514261019878998, "grad_norm": 1.0433422327041626, "learning_rate": 4.534495143508542e-05, "loss": 6.765, "step": 13079 }, { "epoch": 0.5514682631700993, "grad_norm": 0.9672181010246277, "learning_rate": 4.534417064867735e-05, "loss": 6.659, "step": 13080 }, { "epoch": 0.5515104243522988, "grad_norm": 2.1051297187805176, "learning_rate": 4.534338986226928e-05, "loss": 7.5497, "step": 13081 }, { "epoch": 0.5515525855344984, "grad_norm": 1.6162598133087158, "learning_rate": 4.5342609075861206e-05, "loss": 6.3214, "step": 13082 }, { "epoch": 0.551594746716698, "grad_norm": 1.2039133310317993, "learning_rate": 4.5341828289453144e-05, "loss": 6.9297, "step": 13083 }, { "epoch": 0.5516369078988975, "grad_norm": 2.4800732135772705, "learning_rate": 4.534104750304507e-05, "loss": 6.6258, "step": 13084 }, { "epoch": 0.551679069081097, "grad_norm": 1.6865925788879395, "learning_rate": 4.5340266716637e-05, "loss": 7.0939, "step": 13085 }, { "epoch": 0.5517212302632966, "grad_norm": 1.4662723541259766, "learning_rate": 4.533948593022893e-05, "loss": 6.5104, "step": 13086 }, { "epoch": 0.5517633914454961, "grad_norm": 3.8206491470336914, "learning_rate": 4.533870514382086e-05, "loss": 7.5789, "step": 13087 }, { "epoch": 0.5518055526276957, "grad_norm": 1.7204267978668213, "learning_rate": 4.5337924357412786e-05, "loss": 6.6578, "step": 13088 }, { "epoch": 0.5518477138098953, "grad_norm": 1.5114139318466187, "learning_rate": 4.533714357100472e-05, "loss": 7.1484, "step": 13089 }, { "epoch": 0.5518898749920947, "grad_norm": 2.617357015609741, "learning_rate": 4.533636278459665e-05, "loss": 7.9559, "step": 13090 }, { "epoch": 0.5519320361742943, "grad_norm": 1.61314857006073, "learning_rate": 4.5335581998188573e-05, "loss": 6.3612, "step": 13091 }, { "epoch": 0.5519741973564939, "grad_norm": 1.704121708869934, "learning_rate": 4.533480121178051e-05, "loss": 7.2902, "step": 13092 }, { "epoch": 0.5520163585386935, "grad_norm": 2.6597278118133545, "learning_rate": 4.5334020425372436e-05, "loss": 8.1983, "step": 13093 }, { "epoch": 0.5520585197208929, "grad_norm": 2.0871989727020264, "learning_rate": 4.533323963896437e-05, "loss": 6.8203, "step": 13094 }, { "epoch": 0.5521006809030925, "grad_norm": 1.9435421228408813, "learning_rate": 4.53324588525563e-05, "loss": 6.5347, "step": 13095 }, { "epoch": 0.5521428420852921, "grad_norm": 3.5500876903533936, "learning_rate": 4.533167806614822e-05, "loss": 8.3723, "step": 13096 }, { "epoch": 0.5521850032674916, "grad_norm": 2.262545585632324, "learning_rate": 4.533089727974016e-05, "loss": 7.3059, "step": 13097 }, { "epoch": 0.5522271644496912, "grad_norm": 2.251528739929199, "learning_rate": 4.5330116493332086e-05, "loss": 7.9114, "step": 13098 }, { "epoch": 0.5522693256318907, "grad_norm": 1.5578795671463013, "learning_rate": 4.532933570692402e-05, "loss": 6.7028, "step": 13099 }, { "epoch": 0.5523114868140903, "grad_norm": 1.8410433530807495, "learning_rate": 4.532855492051595e-05, "loss": 6.426, "step": 13100 }, { "epoch": 0.5523536479962898, "grad_norm": 1.3134058713912964, "learning_rate": 4.532777413410787e-05, "loss": 6.483, "step": 13101 }, { "epoch": 0.5523958091784894, "grad_norm": 1.054704189300537, "learning_rate": 4.5326993347699804e-05, "loss": 6.6642, "step": 13102 }, { "epoch": 0.5524379703606889, "grad_norm": 1.2499994039535522, "learning_rate": 4.5326212561291735e-05, "loss": 6.9603, "step": 13103 }, { "epoch": 0.5524801315428884, "grad_norm": 1.5646628141403198, "learning_rate": 4.5325431774883666e-05, "loss": 6.51, "step": 13104 }, { "epoch": 0.552522292725088, "grad_norm": 1.5340051651000977, "learning_rate": 4.532465098847559e-05, "loss": 6.5618, "step": 13105 }, { "epoch": 0.5525644539072876, "grad_norm": 1.0208097696304321, "learning_rate": 4.532387020206752e-05, "loss": 6.3973, "step": 13106 }, { "epoch": 0.5526066150894872, "grad_norm": 1.4247522354125977, "learning_rate": 4.532308941565945e-05, "loss": 6.4314, "step": 13107 }, { "epoch": 0.5526487762716866, "grad_norm": 4.521799087524414, "learning_rate": 4.5322308629251385e-05, "loss": 7.5487, "step": 13108 }, { "epoch": 0.5526909374538862, "grad_norm": 2.16448712348938, "learning_rate": 4.5321527842843316e-05, "loss": 7.2409, "step": 13109 }, { "epoch": 0.5527330986360858, "grad_norm": 3.78086519241333, "learning_rate": 4.532074705643524e-05, "loss": 9.4423, "step": 13110 }, { "epoch": 0.5527752598182853, "grad_norm": 1.4920812845230103, "learning_rate": 4.531996627002718e-05, "loss": 6.3525, "step": 13111 }, { "epoch": 0.5528174210004848, "grad_norm": 1.8266454935073853, "learning_rate": 4.53191854836191e-05, "loss": 8.0017, "step": 13112 }, { "epoch": 0.5528595821826844, "grad_norm": 1.462897777557373, "learning_rate": 4.5318404697211034e-05, "loss": 6.5358, "step": 13113 }, { "epoch": 0.552901743364884, "grad_norm": 1.7399955987930298, "learning_rate": 4.5317623910802965e-05, "loss": 6.6606, "step": 13114 }, { "epoch": 0.5529439045470835, "grad_norm": 2.1960012912750244, "learning_rate": 4.531684312439489e-05, "loss": 6.8704, "step": 13115 }, { "epoch": 0.552986065729283, "grad_norm": 1.1041191816329956, "learning_rate": 4.531606233798682e-05, "loss": 7.0356, "step": 13116 }, { "epoch": 0.5530282269114826, "grad_norm": 1.7156779766082764, "learning_rate": 4.531528155157875e-05, "loss": 7.0976, "step": 13117 }, { "epoch": 0.5530703880936821, "grad_norm": 1.658776044845581, "learning_rate": 4.5314500765170684e-05, "loss": 7.0983, "step": 13118 }, { "epoch": 0.5531125492758817, "grad_norm": 1.5128886699676514, "learning_rate": 4.5313719978762615e-05, "loss": 6.5065, "step": 13119 }, { "epoch": 0.5531547104580813, "grad_norm": 2.0729684829711914, "learning_rate": 4.531293919235454e-05, "loss": 6.2948, "step": 13120 }, { "epoch": 0.5531968716402808, "grad_norm": 2.3889193534851074, "learning_rate": 4.531215840594647e-05, "loss": 7.4131, "step": 13121 }, { "epoch": 0.5532390328224803, "grad_norm": 1.5360852479934692, "learning_rate": 4.53113776195384e-05, "loss": 7.439, "step": 13122 }, { "epoch": 0.5532811940046799, "grad_norm": 1.1548032760620117, "learning_rate": 4.531059683313033e-05, "loss": 6.5897, "step": 13123 }, { "epoch": 0.5533233551868795, "grad_norm": 2.2210779190063477, "learning_rate": 4.530981604672226e-05, "loss": 6.4296, "step": 13124 }, { "epoch": 0.5533655163690789, "grad_norm": 1.539104700088501, "learning_rate": 4.530903526031419e-05, "loss": 6.736, "step": 13125 }, { "epoch": 0.5534076775512785, "grad_norm": 3.2116105556488037, "learning_rate": 4.530825447390612e-05, "loss": 9.2732, "step": 13126 }, { "epoch": 0.5534498387334781, "grad_norm": 1.5371607542037964, "learning_rate": 4.5307473687498045e-05, "loss": 7.1977, "step": 13127 }, { "epoch": 0.5534919999156777, "grad_norm": 2.962292432785034, "learning_rate": 4.530669290108998e-05, "loss": 7.6475, "step": 13128 }, { "epoch": 0.5535341610978772, "grad_norm": 1.7834935188293457, "learning_rate": 4.530591211468191e-05, "loss": 6.8613, "step": 13129 }, { "epoch": 0.5535763222800767, "grad_norm": 1.3844506740570068, "learning_rate": 4.5305131328273845e-05, "loss": 6.4617, "step": 13130 }, { "epoch": 0.5536184834622763, "grad_norm": 0.9689217805862427, "learning_rate": 4.530435054186577e-05, "loss": 6.5737, "step": 13131 }, { "epoch": 0.5536606446444758, "grad_norm": 2.7966177463531494, "learning_rate": 4.53035697554577e-05, "loss": 7.3178, "step": 13132 }, { "epoch": 0.5537028058266754, "grad_norm": 3.340667247772217, "learning_rate": 4.530278896904963e-05, "loss": 7.4409, "step": 13133 }, { "epoch": 0.5537449670088749, "grad_norm": 3.5323526859283447, "learning_rate": 4.530200818264156e-05, "loss": 7.2285, "step": 13134 }, { "epoch": 0.5537871281910745, "grad_norm": 2.728087902069092, "learning_rate": 4.530122739623349e-05, "loss": 7.4128, "step": 13135 }, { "epoch": 0.553829289373274, "grad_norm": 1.1287577152252197, "learning_rate": 4.530044660982542e-05, "loss": 6.5593, "step": 13136 }, { "epoch": 0.5538714505554736, "grad_norm": 0.9964160919189453, "learning_rate": 4.529966582341735e-05, "loss": 6.7522, "step": 13137 }, { "epoch": 0.5539136117376732, "grad_norm": 1.1458086967468262, "learning_rate": 4.5298885037009275e-05, "loss": 6.577, "step": 13138 }, { "epoch": 0.5539557729198726, "grad_norm": 4.357144355773926, "learning_rate": 4.5298104250601206e-05, "loss": 6.3063, "step": 13139 }, { "epoch": 0.5539979341020722, "grad_norm": 1.1708118915557861, "learning_rate": 4.529732346419314e-05, "loss": 6.4692, "step": 13140 }, { "epoch": 0.5540400952842718, "grad_norm": 1.872987151145935, "learning_rate": 4.529654267778507e-05, "loss": 7.2482, "step": 13141 }, { "epoch": 0.5540822564664714, "grad_norm": 2.1958038806915283, "learning_rate": 4.5295761891377e-05, "loss": 8.4551, "step": 13142 }, { "epoch": 0.5541244176486708, "grad_norm": 2.6654510498046875, "learning_rate": 4.5294981104968924e-05, "loss": 7.4527, "step": 13143 }, { "epoch": 0.5541665788308704, "grad_norm": 1.1649963855743408, "learning_rate": 4.5294200318560856e-05, "loss": 6.7897, "step": 13144 }, { "epoch": 0.55420874001307, "grad_norm": 1.7368444204330444, "learning_rate": 4.529341953215279e-05, "loss": 6.8139, "step": 13145 }, { "epoch": 0.5542509011952695, "grad_norm": 2.332308530807495, "learning_rate": 4.529263874574471e-05, "loss": 7.5015, "step": 13146 }, { "epoch": 0.5542930623774691, "grad_norm": 2.0498812198638916, "learning_rate": 4.529185795933665e-05, "loss": 6.8688, "step": 13147 }, { "epoch": 0.5543352235596686, "grad_norm": 2.200507640838623, "learning_rate": 4.5291077172928574e-05, "loss": 7.652, "step": 13148 }, { "epoch": 0.5543773847418682, "grad_norm": 2.440171480178833, "learning_rate": 4.5290296386520505e-05, "loss": 7.7177, "step": 13149 }, { "epoch": 0.5544195459240677, "grad_norm": 1.1233720779418945, "learning_rate": 4.5289515600112436e-05, "loss": 7.2161, "step": 13150 }, { "epoch": 0.5544617071062673, "grad_norm": 2.2616701126098633, "learning_rate": 4.528873481370436e-05, "loss": 7.1359, "step": 13151 }, { "epoch": 0.5545038682884668, "grad_norm": 2.7731635570526123, "learning_rate": 4.52879540272963e-05, "loss": 6.2836, "step": 13152 }, { "epoch": 0.5545460294706663, "grad_norm": 2.5789833068847656, "learning_rate": 4.5287173240888223e-05, "loss": 7.6586, "step": 13153 }, { "epoch": 0.5545881906528659, "grad_norm": 1.1855900287628174, "learning_rate": 4.5286392454480155e-05, "loss": 6.697, "step": 13154 }, { "epoch": 0.5546303518350655, "grad_norm": 2.6985092163085938, "learning_rate": 4.5285611668072086e-05, "loss": 7.1577, "step": 13155 }, { "epoch": 0.5546725130172651, "grad_norm": 1.5803471803665161, "learning_rate": 4.528483088166402e-05, "loss": 7.1223, "step": 13156 }, { "epoch": 0.5547146741994645, "grad_norm": 1.3368089199066162, "learning_rate": 4.528405009525594e-05, "loss": 7.2978, "step": 13157 }, { "epoch": 0.5547568353816641, "grad_norm": 2.0206656455993652, "learning_rate": 4.528326930884787e-05, "loss": 6.6234, "step": 13158 }, { "epoch": 0.5547989965638637, "grad_norm": 2.2933566570281982, "learning_rate": 4.5282488522439804e-05, "loss": 8.0224, "step": 13159 }, { "epoch": 0.5548411577460632, "grad_norm": 2.086806058883667, "learning_rate": 4.528170773603173e-05, "loss": 6.4747, "step": 13160 }, { "epoch": 0.5548833189282627, "grad_norm": 2.3193187713623047, "learning_rate": 4.528092694962367e-05, "loss": 6.5587, "step": 13161 }, { "epoch": 0.5549254801104623, "grad_norm": 1.890268325805664, "learning_rate": 4.528014616321559e-05, "loss": 7.3265, "step": 13162 }, { "epoch": 0.5549676412926619, "grad_norm": 1.2000354528427124, "learning_rate": 4.527936537680752e-05, "loss": 6.6166, "step": 13163 }, { "epoch": 0.5550098024748614, "grad_norm": 1.7751811742782593, "learning_rate": 4.5278584590399454e-05, "loss": 7.3923, "step": 13164 }, { "epoch": 0.555051963657061, "grad_norm": 1.6551482677459717, "learning_rate": 4.527780380399138e-05, "loss": 6.4956, "step": 13165 }, { "epoch": 0.5550941248392605, "grad_norm": 2.951328754425049, "learning_rate": 4.5277023017583316e-05, "loss": 7.9048, "step": 13166 }, { "epoch": 0.55513628602146, "grad_norm": 1.4331352710723877, "learning_rate": 4.527624223117524e-05, "loss": 6.6779, "step": 13167 }, { "epoch": 0.5551784472036596, "grad_norm": 1.5177055597305298, "learning_rate": 4.527546144476717e-05, "loss": 6.4676, "step": 13168 }, { "epoch": 0.5552206083858592, "grad_norm": 2.5382795333862305, "learning_rate": 4.52746806583591e-05, "loss": 6.1784, "step": 13169 }, { "epoch": 0.5552627695680586, "grad_norm": 1.4531067609786987, "learning_rate": 4.527389987195103e-05, "loss": 6.6803, "step": 13170 }, { "epoch": 0.5553049307502582, "grad_norm": 1.4860808849334717, "learning_rate": 4.527311908554296e-05, "loss": 6.7741, "step": 13171 }, { "epoch": 0.5553470919324578, "grad_norm": 3.2930116653442383, "learning_rate": 4.527233829913489e-05, "loss": 7.8994, "step": 13172 }, { "epoch": 0.5553892531146574, "grad_norm": 1.161247968673706, "learning_rate": 4.527155751272682e-05, "loss": 7.0491, "step": 13173 }, { "epoch": 0.5554314142968568, "grad_norm": 1.5903366804122925, "learning_rate": 4.5270776726318746e-05, "loss": 6.9335, "step": 13174 }, { "epoch": 0.5554735754790564, "grad_norm": 4.548093795776367, "learning_rate": 4.5269995939910684e-05, "loss": 7.1554, "step": 13175 }, { "epoch": 0.555515736661256, "grad_norm": 1.2392191886901855, "learning_rate": 4.526921515350261e-05, "loss": 6.5035, "step": 13176 }, { "epoch": 0.5555578978434556, "grad_norm": 1.3805526494979858, "learning_rate": 4.526843436709454e-05, "loss": 7.1046, "step": 13177 }, { "epoch": 0.5556000590256551, "grad_norm": 2.4355623722076416, "learning_rate": 4.526765358068647e-05, "loss": 7.9762, "step": 13178 }, { "epoch": 0.5556422202078546, "grad_norm": 1.2376066446304321, "learning_rate": 4.5266872794278396e-05, "loss": 7.2453, "step": 13179 }, { "epoch": 0.5556843813900542, "grad_norm": 1.3583619594573975, "learning_rate": 4.5266092007870334e-05, "loss": 6.7054, "step": 13180 }, { "epoch": 0.5557265425722537, "grad_norm": 2.6556005477905273, "learning_rate": 4.526531122146226e-05, "loss": 8.052, "step": 13181 }, { "epoch": 0.5557687037544533, "grad_norm": 1.5024300813674927, "learning_rate": 4.526453043505419e-05, "loss": 6.6592, "step": 13182 }, { "epoch": 0.5558108649366528, "grad_norm": 1.5449460744857788, "learning_rate": 4.526374964864612e-05, "loss": 6.357, "step": 13183 }, { "epoch": 0.5558530261188523, "grad_norm": 1.4975463151931763, "learning_rate": 4.5262968862238045e-05, "loss": 7.3796, "step": 13184 }, { "epoch": 0.5558951873010519, "grad_norm": 3.2579851150512695, "learning_rate": 4.5262188075829976e-05, "loss": 7.2878, "step": 13185 }, { "epoch": 0.5559373484832515, "grad_norm": 1.9862128496170044, "learning_rate": 4.526140728942191e-05, "loss": 6.7982, "step": 13186 }, { "epoch": 0.5559795096654511, "grad_norm": 2.1789252758026123, "learning_rate": 4.526062650301384e-05, "loss": 8.4021, "step": 13187 }, { "epoch": 0.5560216708476505, "grad_norm": 1.8854161500930786, "learning_rate": 4.525984571660577e-05, "loss": 6.3605, "step": 13188 }, { "epoch": 0.5560638320298501, "grad_norm": 1.5082002878189087, "learning_rate": 4.5259064930197695e-05, "loss": 6.467, "step": 13189 }, { "epoch": 0.5561059932120497, "grad_norm": 1.302392840385437, "learning_rate": 4.5258284143789626e-05, "loss": 6.4426, "step": 13190 }, { "epoch": 0.5561481543942493, "grad_norm": 1.418898582458496, "learning_rate": 4.525750335738156e-05, "loss": 7.4894, "step": 13191 }, { "epoch": 0.5561903155764487, "grad_norm": 7.0623064041137695, "learning_rate": 4.525672257097349e-05, "loss": 12.1535, "step": 13192 }, { "epoch": 0.5562324767586483, "grad_norm": 3.898681402206421, "learning_rate": 4.525594178456541e-05, "loss": 7.8546, "step": 13193 }, { "epoch": 0.5562746379408479, "grad_norm": 1.3877102136611938, "learning_rate": 4.525516099815735e-05, "loss": 6.8635, "step": 13194 }, { "epoch": 0.5563167991230474, "grad_norm": 1.4364662170410156, "learning_rate": 4.5254380211749275e-05, "loss": 6.7498, "step": 13195 }, { "epoch": 0.556358960305247, "grad_norm": 1.1681251525878906, "learning_rate": 4.52535994253412e-05, "loss": 6.5211, "step": 13196 }, { "epoch": 0.5564011214874465, "grad_norm": 2.7492730617523193, "learning_rate": 4.525281863893314e-05, "loss": 7.6174, "step": 13197 }, { "epoch": 0.556443282669646, "grad_norm": 6.811948776245117, "learning_rate": 4.525203785252506e-05, "loss": 11.8107, "step": 13198 }, { "epoch": 0.5564854438518456, "grad_norm": 2.5504016876220703, "learning_rate": 4.5251257066117e-05, "loss": 7.9099, "step": 13199 }, { "epoch": 0.5565276050340452, "grad_norm": 1.7291994094848633, "learning_rate": 4.5250476279708925e-05, "loss": 6.53, "step": 13200 }, { "epoch": 0.5565697662162447, "grad_norm": 1.4041074514389038, "learning_rate": 4.5249695493300856e-05, "loss": 7.0498, "step": 13201 }, { "epoch": 0.5566119273984442, "grad_norm": 2.488172769546509, "learning_rate": 4.524891470689279e-05, "loss": 7.5976, "step": 13202 }, { "epoch": 0.5566540885806438, "grad_norm": 0.9445634484291077, "learning_rate": 4.524813392048471e-05, "loss": 6.4985, "step": 13203 }, { "epoch": 0.5566962497628434, "grad_norm": 2.1915438175201416, "learning_rate": 4.524735313407664e-05, "loss": 7.1911, "step": 13204 }, { "epoch": 0.556738410945043, "grad_norm": 2.9725050926208496, "learning_rate": 4.5246572347668574e-05, "loss": 8.0437, "step": 13205 }, { "epoch": 0.5567805721272424, "grad_norm": 1.5460965633392334, "learning_rate": 4.5245791561260506e-05, "loss": 6.4274, "step": 13206 }, { "epoch": 0.556822733309442, "grad_norm": 1.9409804344177246, "learning_rate": 4.524501077485243e-05, "loss": 6.8642, "step": 13207 }, { "epoch": 0.5568648944916416, "grad_norm": 1.540337085723877, "learning_rate": 4.524422998844436e-05, "loss": 7.2534, "step": 13208 }, { "epoch": 0.5569070556738411, "grad_norm": 1.5821614265441895, "learning_rate": 4.524344920203629e-05, "loss": 6.372, "step": 13209 }, { "epoch": 0.5569492168560406, "grad_norm": 1.0413237810134888, "learning_rate": 4.5242668415628224e-05, "loss": 6.7184, "step": 13210 }, { "epoch": 0.5569913780382402, "grad_norm": 1.1901971101760864, "learning_rate": 4.5241887629220155e-05, "loss": 7.1092, "step": 13211 }, { "epoch": 0.5570335392204397, "grad_norm": 3.5504868030548096, "learning_rate": 4.524110684281208e-05, "loss": 7.2402, "step": 13212 }, { "epoch": 0.5570757004026393, "grad_norm": 2.767184257507324, "learning_rate": 4.524032605640401e-05, "loss": 7.934, "step": 13213 }, { "epoch": 0.5571178615848389, "grad_norm": 1.2437565326690674, "learning_rate": 4.523954526999594e-05, "loss": 6.6036, "step": 13214 }, { "epoch": 0.5571600227670384, "grad_norm": 1.3256222009658813, "learning_rate": 4.523876448358787e-05, "loss": 6.4528, "step": 13215 }, { "epoch": 0.5572021839492379, "grad_norm": 1.4806184768676758, "learning_rate": 4.5237983697179805e-05, "loss": 6.4136, "step": 13216 }, { "epoch": 0.5572443451314375, "grad_norm": 1.292913794517517, "learning_rate": 4.523720291077173e-05, "loss": 6.5516, "step": 13217 }, { "epoch": 0.5572865063136371, "grad_norm": 1.1370785236358643, "learning_rate": 4.523642212436366e-05, "loss": 6.7751, "step": 13218 }, { "epoch": 0.5573286674958365, "grad_norm": 2.948833703994751, "learning_rate": 4.523564133795559e-05, "loss": 8.0326, "step": 13219 }, { "epoch": 0.5573708286780361, "grad_norm": 1.1186684370040894, "learning_rate": 4.523486055154752e-05, "loss": 6.6717, "step": 13220 }, { "epoch": 0.5574129898602357, "grad_norm": 1.1570156812667847, "learning_rate": 4.5234079765139454e-05, "loss": 6.6196, "step": 13221 }, { "epoch": 0.5574551510424353, "grad_norm": 1.409475326538086, "learning_rate": 4.523329897873138e-05, "loss": 6.8017, "step": 13222 }, { "epoch": 0.5574973122246348, "grad_norm": 1.2223119735717773, "learning_rate": 4.523251819232331e-05, "loss": 6.8915, "step": 13223 }, { "epoch": 0.5575394734068343, "grad_norm": 3.5134406089782715, "learning_rate": 4.523173740591524e-05, "loss": 6.595, "step": 13224 }, { "epoch": 0.5575816345890339, "grad_norm": 1.3067537546157837, "learning_rate": 4.523095661950717e-05, "loss": 6.5202, "step": 13225 }, { "epoch": 0.5576237957712334, "grad_norm": 1.9094160795211792, "learning_rate": 4.52301758330991e-05, "loss": 7.0752, "step": 13226 }, { "epoch": 0.557665956953433, "grad_norm": 1.0903071165084839, "learning_rate": 4.522939504669103e-05, "loss": 6.5498, "step": 13227 }, { "epoch": 0.5577081181356325, "grad_norm": 1.1950817108154297, "learning_rate": 4.522861426028296e-05, "loss": 6.6038, "step": 13228 }, { "epoch": 0.557750279317832, "grad_norm": 2.341718912124634, "learning_rate": 4.5227833473874884e-05, "loss": 6.4494, "step": 13229 }, { "epoch": 0.5577924405000316, "grad_norm": 1.6041542291641235, "learning_rate": 4.522705268746682e-05, "loss": 6.5252, "step": 13230 }, { "epoch": 0.5578346016822312, "grad_norm": 1.4619687795639038, "learning_rate": 4.5226271901058747e-05, "loss": 6.3503, "step": 13231 }, { "epoch": 0.5578767628644307, "grad_norm": 1.5001541376113892, "learning_rate": 4.522549111465068e-05, "loss": 7.2534, "step": 13232 }, { "epoch": 0.5579189240466302, "grad_norm": 0.8670036792755127, "learning_rate": 4.522471032824261e-05, "loss": 6.9605, "step": 13233 }, { "epoch": 0.5579610852288298, "grad_norm": 3.207486391067505, "learning_rate": 4.5223929541834534e-05, "loss": 7.7809, "step": 13234 }, { "epoch": 0.5580032464110294, "grad_norm": 1.5423839092254639, "learning_rate": 4.522314875542647e-05, "loss": 6.3767, "step": 13235 }, { "epoch": 0.558045407593229, "grad_norm": 3.438922166824341, "learning_rate": 4.5222367969018396e-05, "loss": 7.7235, "step": 13236 }, { "epoch": 0.5580875687754284, "grad_norm": 1.184915542602539, "learning_rate": 4.522158718261033e-05, "loss": 6.6563, "step": 13237 }, { "epoch": 0.558129729957628, "grad_norm": 1.1347421407699585, "learning_rate": 4.522080639620226e-05, "loss": 6.9351, "step": 13238 }, { "epoch": 0.5581718911398276, "grad_norm": 1.4302321672439575, "learning_rate": 4.522002560979419e-05, "loss": 6.6381, "step": 13239 }, { "epoch": 0.5582140523220271, "grad_norm": 2.2857601642608643, "learning_rate": 4.5219244823386114e-05, "loss": 7.1973, "step": 13240 }, { "epoch": 0.5582562135042266, "grad_norm": 4.27992582321167, "learning_rate": 4.5218464036978046e-05, "loss": 9.3689, "step": 13241 }, { "epoch": 0.5582983746864262, "grad_norm": 1.5623148679733276, "learning_rate": 4.521768325056998e-05, "loss": 7.8245, "step": 13242 }, { "epoch": 0.5583405358686258, "grad_norm": 1.126867413520813, "learning_rate": 4.521690246416191e-05, "loss": 6.4302, "step": 13243 }, { "epoch": 0.5583826970508253, "grad_norm": 1.6888539791107178, "learning_rate": 4.521612167775384e-05, "loss": 7.3649, "step": 13244 }, { "epoch": 0.5584248582330249, "grad_norm": 1.0357089042663574, "learning_rate": 4.5215340891345764e-05, "loss": 6.5034, "step": 13245 }, { "epoch": 0.5584670194152244, "grad_norm": 0.9402845501899719, "learning_rate": 4.5214560104937695e-05, "loss": 6.3579, "step": 13246 }, { "epoch": 0.5585091805974239, "grad_norm": 1.136767029762268, "learning_rate": 4.5213779318529626e-05, "loss": 6.7522, "step": 13247 }, { "epoch": 0.5585513417796235, "grad_norm": 2.9195244312286377, "learning_rate": 4.521299853212155e-05, "loss": 7.8247, "step": 13248 }, { "epoch": 0.5585935029618231, "grad_norm": 1.1317615509033203, "learning_rate": 4.521221774571349e-05, "loss": 6.7283, "step": 13249 }, { "epoch": 0.5586356641440225, "grad_norm": 1.100024938583374, "learning_rate": 4.5211436959305413e-05, "loss": 6.6027, "step": 13250 }, { "epoch": 0.5586778253262221, "grad_norm": 1.259565830230713, "learning_rate": 4.5210656172897345e-05, "loss": 6.9394, "step": 13251 }, { "epoch": 0.5587199865084217, "grad_norm": 1.4606568813323975, "learning_rate": 4.5209875386489276e-05, "loss": 6.2782, "step": 13252 }, { "epoch": 0.5587621476906213, "grad_norm": 2.852919578552246, "learning_rate": 4.52090946000812e-05, "loss": 7.5145, "step": 13253 }, { "epoch": 0.5588043088728208, "grad_norm": 1.7384014129638672, "learning_rate": 4.520831381367313e-05, "loss": 7.2125, "step": 13254 }, { "epoch": 0.5588464700550203, "grad_norm": 1.012221336364746, "learning_rate": 4.520753302726506e-05, "loss": 6.7681, "step": 13255 }, { "epoch": 0.5588886312372199, "grad_norm": 1.7337466478347778, "learning_rate": 4.5206752240856994e-05, "loss": 7.2139, "step": 13256 }, { "epoch": 0.5589307924194195, "grad_norm": 0.9812993407249451, "learning_rate": 4.5205971454448925e-05, "loss": 6.6268, "step": 13257 }, { "epoch": 0.558972953601619, "grad_norm": 1.0605827569961548, "learning_rate": 4.520519066804085e-05, "loss": 6.7661, "step": 13258 }, { "epoch": 0.5590151147838185, "grad_norm": 1.6552516222000122, "learning_rate": 4.520440988163278e-05, "loss": 7.4074, "step": 13259 }, { "epoch": 0.5590572759660181, "grad_norm": 2.4067671298980713, "learning_rate": 4.520362909522471e-05, "loss": 7.0639, "step": 13260 }, { "epoch": 0.5590994371482176, "grad_norm": 2.750629186630249, "learning_rate": 4.5202848308816644e-05, "loss": 7.8693, "step": 13261 }, { "epoch": 0.5591415983304172, "grad_norm": 2.2708466053009033, "learning_rate": 4.520206752240857e-05, "loss": 7.8342, "step": 13262 }, { "epoch": 0.5591837595126168, "grad_norm": 1.5194460153579712, "learning_rate": 4.5201286736000506e-05, "loss": 7.0895, "step": 13263 }, { "epoch": 0.5592259206948162, "grad_norm": 1.4329464435577393, "learning_rate": 4.520050594959243e-05, "loss": 6.5102, "step": 13264 }, { "epoch": 0.5592680818770158, "grad_norm": 1.4610397815704346, "learning_rate": 4.519972516318436e-05, "loss": 6.7564, "step": 13265 }, { "epoch": 0.5593102430592154, "grad_norm": 1.520101547241211, "learning_rate": 4.519894437677629e-05, "loss": 6.8189, "step": 13266 }, { "epoch": 0.559352404241415, "grad_norm": 1.066478967666626, "learning_rate": 4.519816359036822e-05, "loss": 6.841, "step": 13267 }, { "epoch": 0.5593945654236144, "grad_norm": 2.2384989261627197, "learning_rate": 4.5197382803960156e-05, "loss": 6.4433, "step": 13268 }, { "epoch": 0.559436726605814, "grad_norm": 2.47196364402771, "learning_rate": 4.519660201755208e-05, "loss": 7.8261, "step": 13269 }, { "epoch": 0.5594788877880136, "grad_norm": 1.6807515621185303, "learning_rate": 4.519582123114401e-05, "loss": 7.5392, "step": 13270 }, { "epoch": 0.5595210489702132, "grad_norm": 1.2925273180007935, "learning_rate": 4.519504044473594e-05, "loss": 6.5417, "step": 13271 }, { "epoch": 0.5595632101524127, "grad_norm": 2.402176856994629, "learning_rate": 4.519425965832787e-05, "loss": 6.8055, "step": 13272 }, { "epoch": 0.5596053713346122, "grad_norm": 2.4929397106170654, "learning_rate": 4.51934788719198e-05, "loss": 7.9213, "step": 13273 }, { "epoch": 0.5596475325168118, "grad_norm": 2.629682779312134, "learning_rate": 4.519269808551173e-05, "loss": 8.0611, "step": 13274 }, { "epoch": 0.5596896936990113, "grad_norm": 4.078824996948242, "learning_rate": 4.519191729910366e-05, "loss": 7.325, "step": 13275 }, { "epoch": 0.5597318548812109, "grad_norm": 1.2103952169418335, "learning_rate": 4.5191136512695585e-05, "loss": 7.0963, "step": 13276 }, { "epoch": 0.5597740160634104, "grad_norm": 1.1531291007995605, "learning_rate": 4.519035572628752e-05, "loss": 6.7255, "step": 13277 }, { "epoch": 0.55981617724561, "grad_norm": 1.8387175798416138, "learning_rate": 4.518957493987945e-05, "loss": 6.8637, "step": 13278 }, { "epoch": 0.5598583384278095, "grad_norm": 1.6569567918777466, "learning_rate": 4.518879415347138e-05, "loss": 7.3439, "step": 13279 }, { "epoch": 0.5599004996100091, "grad_norm": 1.4985039234161377, "learning_rate": 4.518801336706331e-05, "loss": 6.6606, "step": 13280 }, { "epoch": 0.5599426607922087, "grad_norm": 1.5527327060699463, "learning_rate": 4.5187232580655235e-05, "loss": 6.3715, "step": 13281 }, { "epoch": 0.5599848219744081, "grad_norm": 3.2388134002685547, "learning_rate": 4.518645179424717e-05, "loss": 7.7388, "step": 13282 }, { "epoch": 0.5600269831566077, "grad_norm": 2.949760913848877, "learning_rate": 4.51856710078391e-05, "loss": 8.046, "step": 13283 }, { "epoch": 0.5600691443388073, "grad_norm": 1.1016486883163452, "learning_rate": 4.518489022143103e-05, "loss": 7.0277, "step": 13284 }, { "epoch": 0.5601113055210069, "grad_norm": 1.5292316675186157, "learning_rate": 4.518410943502296e-05, "loss": 7.1236, "step": 13285 }, { "epoch": 0.5601534667032063, "grad_norm": 2.7529895305633545, "learning_rate": 4.5183328648614885e-05, "loss": 7.8694, "step": 13286 }, { "epoch": 0.5601956278854059, "grad_norm": 1.633165955543518, "learning_rate": 4.5182547862206816e-05, "loss": 6.7389, "step": 13287 }, { "epoch": 0.5602377890676055, "grad_norm": 1.2148851156234741, "learning_rate": 4.518176707579875e-05, "loss": 6.4615, "step": 13288 }, { "epoch": 0.560279950249805, "grad_norm": 2.1235899925231934, "learning_rate": 4.518098628939068e-05, "loss": 6.9206, "step": 13289 }, { "epoch": 0.5603221114320045, "grad_norm": 2.2530083656311035, "learning_rate": 4.518020550298261e-05, "loss": 6.5354, "step": 13290 }, { "epoch": 0.5603642726142041, "grad_norm": 1.6692278385162354, "learning_rate": 4.5179424716574534e-05, "loss": 6.3645, "step": 13291 }, { "epoch": 0.5604064337964036, "grad_norm": 1.30752432346344, "learning_rate": 4.5178643930166465e-05, "loss": 7.1226, "step": 13292 }, { "epoch": 0.5604485949786032, "grad_norm": 2.906014919281006, "learning_rate": 4.5177863143758397e-05, "loss": 7.9322, "step": 13293 }, { "epoch": 0.5604907561608028, "grad_norm": 3.3406741619110107, "learning_rate": 4.517708235735033e-05, "loss": 7.1008, "step": 13294 }, { "epoch": 0.5605329173430023, "grad_norm": 2.04718279838562, "learning_rate": 4.517630157094225e-05, "loss": 6.8756, "step": 13295 }, { "epoch": 0.5605750785252018, "grad_norm": 1.1925164461135864, "learning_rate": 4.5175520784534184e-05, "loss": 6.6548, "step": 13296 }, { "epoch": 0.5606172397074014, "grad_norm": 1.2823673486709595, "learning_rate": 4.5174739998126115e-05, "loss": 6.5639, "step": 13297 }, { "epoch": 0.560659400889601, "grad_norm": 1.545029878616333, "learning_rate": 4.517395921171804e-05, "loss": 6.5405, "step": 13298 }, { "epoch": 0.5607015620718004, "grad_norm": 1.0270456075668335, "learning_rate": 4.517317842530998e-05, "loss": 6.5572, "step": 13299 }, { "epoch": 0.560743723254, "grad_norm": 1.3010653257369995, "learning_rate": 4.51723976389019e-05, "loss": 6.3576, "step": 13300 }, { "epoch": 0.5607858844361996, "grad_norm": 3.655696392059326, "learning_rate": 4.517161685249384e-05, "loss": 7.3066, "step": 13301 }, { "epoch": 0.5608280456183992, "grad_norm": 2.1905016899108887, "learning_rate": 4.5170836066085764e-05, "loss": 6.6304, "step": 13302 }, { "epoch": 0.5608702068005987, "grad_norm": 1.2259362936019897, "learning_rate": 4.517005527967769e-05, "loss": 6.5865, "step": 13303 }, { "epoch": 0.5609123679827982, "grad_norm": 2.966287851333618, "learning_rate": 4.516927449326963e-05, "loss": 7.2292, "step": 13304 }, { "epoch": 0.5609545291649978, "grad_norm": 1.1560205221176147, "learning_rate": 4.516849370686155e-05, "loss": 6.8248, "step": 13305 }, { "epoch": 0.5609966903471973, "grad_norm": 1.2432664632797241, "learning_rate": 4.516771292045348e-05, "loss": 6.6774, "step": 13306 }, { "epoch": 0.5610388515293969, "grad_norm": 1.4249625205993652, "learning_rate": 4.5166932134045414e-05, "loss": 6.565, "step": 13307 }, { "epoch": 0.5610810127115964, "grad_norm": 2.6708195209503174, "learning_rate": 4.5166151347637345e-05, "loss": 7.3628, "step": 13308 }, { "epoch": 0.561123173893796, "grad_norm": 1.2402876615524292, "learning_rate": 4.516537056122927e-05, "loss": 6.5104, "step": 13309 }, { "epoch": 0.5611653350759955, "grad_norm": 1.894021987915039, "learning_rate": 4.51645897748212e-05, "loss": 7.2205, "step": 13310 }, { "epoch": 0.5612074962581951, "grad_norm": 2.6045546531677246, "learning_rate": 4.516380898841313e-05, "loss": 7.5601, "step": 13311 }, { "epoch": 0.5612496574403947, "grad_norm": 3.318960428237915, "learning_rate": 4.5163028202005063e-05, "loss": 8.0471, "step": 13312 }, { "epoch": 0.5612918186225941, "grad_norm": 1.5075191259384155, "learning_rate": 4.5162247415596995e-05, "loss": 6.4847, "step": 13313 }, { "epoch": 0.5613339798047937, "grad_norm": 2.431527853012085, "learning_rate": 4.516146662918892e-05, "loss": 7.0676, "step": 13314 }, { "epoch": 0.5613761409869933, "grad_norm": 2.0493273735046387, "learning_rate": 4.516068584278085e-05, "loss": 7.7153, "step": 13315 }, { "epoch": 0.5614183021691929, "grad_norm": 1.591222882270813, "learning_rate": 4.515990505637278e-05, "loss": 8.1196, "step": 13316 }, { "epoch": 0.5614604633513923, "grad_norm": 1.281996726989746, "learning_rate": 4.5159124269964706e-05, "loss": 7.0841, "step": 13317 }, { "epoch": 0.5615026245335919, "grad_norm": 3.021158218383789, "learning_rate": 4.5158343483556644e-05, "loss": 7.8249, "step": 13318 }, { "epoch": 0.5615447857157915, "grad_norm": 1.4298903942108154, "learning_rate": 4.515756269714857e-05, "loss": 7.4056, "step": 13319 }, { "epoch": 0.561586946897991, "grad_norm": 1.0196946859359741, "learning_rate": 4.51567819107405e-05, "loss": 6.5893, "step": 13320 }, { "epoch": 0.5616291080801906, "grad_norm": 1.5693453550338745, "learning_rate": 4.515600112433243e-05, "loss": 6.541, "step": 13321 }, { "epoch": 0.5616712692623901, "grad_norm": 1.1406497955322266, "learning_rate": 4.5155220337924356e-05, "loss": 6.4735, "step": 13322 }, { "epoch": 0.5617134304445897, "grad_norm": 3.018946647644043, "learning_rate": 4.5154439551516294e-05, "loss": 6.3009, "step": 13323 }, { "epoch": 0.5617555916267892, "grad_norm": 1.451829195022583, "learning_rate": 4.515365876510822e-05, "loss": 7.5477, "step": 13324 }, { "epoch": 0.5617977528089888, "grad_norm": 2.109248399734497, "learning_rate": 4.515287797870015e-05, "loss": 7.5306, "step": 13325 }, { "epoch": 0.5618399139911883, "grad_norm": 1.3836123943328857, "learning_rate": 4.515209719229208e-05, "loss": 6.569, "step": 13326 }, { "epoch": 0.5618820751733878, "grad_norm": 2.48825740814209, "learning_rate": 4.515131640588401e-05, "loss": 6.5363, "step": 13327 }, { "epoch": 0.5619242363555874, "grad_norm": 1.1385784149169922, "learning_rate": 4.5150535619475936e-05, "loss": 6.957, "step": 13328 }, { "epoch": 0.561966397537787, "grad_norm": 2.693791151046753, "learning_rate": 4.514975483306787e-05, "loss": 7.3158, "step": 13329 }, { "epoch": 0.5620085587199866, "grad_norm": 0.9961920380592346, "learning_rate": 4.51489740466598e-05, "loss": 6.4196, "step": 13330 }, { "epoch": 0.562050719902186, "grad_norm": 2.1114420890808105, "learning_rate": 4.5148193260251723e-05, "loss": 8.1718, "step": 13331 }, { "epoch": 0.5620928810843856, "grad_norm": 1.2189222574234009, "learning_rate": 4.514741247384366e-05, "loss": 6.835, "step": 13332 }, { "epoch": 0.5621350422665852, "grad_norm": 1.8838833570480347, "learning_rate": 4.5146631687435586e-05, "loss": 6.888, "step": 13333 }, { "epoch": 0.5621772034487847, "grad_norm": 2.330296277999878, "learning_rate": 4.514585090102752e-05, "loss": 7.6183, "step": 13334 }, { "epoch": 0.5622193646309842, "grad_norm": 2.18811297416687, "learning_rate": 4.514507011461945e-05, "loss": 8.2885, "step": 13335 }, { "epoch": 0.5622615258131838, "grad_norm": 1.4686275720596313, "learning_rate": 4.514428932821137e-05, "loss": 6.3702, "step": 13336 }, { "epoch": 0.5623036869953834, "grad_norm": 2.7523021697998047, "learning_rate": 4.514350854180331e-05, "loss": 7.6684, "step": 13337 }, { "epoch": 0.5623458481775829, "grad_norm": 3.3216848373413086, "learning_rate": 4.5142727755395236e-05, "loss": 6.8302, "step": 13338 }, { "epoch": 0.5623880093597825, "grad_norm": 1.7137216329574585, "learning_rate": 4.514194696898717e-05, "loss": 6.4378, "step": 13339 }, { "epoch": 0.562430170541982, "grad_norm": 1.0556524991989136, "learning_rate": 4.51411661825791e-05, "loss": 6.5876, "step": 13340 }, { "epoch": 0.5624723317241815, "grad_norm": 0.9088811278343201, "learning_rate": 4.514038539617102e-05, "loss": 6.5505, "step": 13341 }, { "epoch": 0.5625144929063811, "grad_norm": 3.043483257293701, "learning_rate": 4.5139604609762954e-05, "loss": 7.0127, "step": 13342 }, { "epoch": 0.5625566540885807, "grad_norm": 1.370318055152893, "learning_rate": 4.5138823823354885e-05, "loss": 6.9502, "step": 13343 }, { "epoch": 0.5625988152707802, "grad_norm": 2.105245590209961, "learning_rate": 4.5138043036946816e-05, "loss": 6.4479, "step": 13344 }, { "epoch": 0.5626409764529797, "grad_norm": 3.8665497303009033, "learning_rate": 4.513726225053874e-05, "loss": 8.1439, "step": 13345 }, { "epoch": 0.5626831376351793, "grad_norm": 1.6649307012557983, "learning_rate": 4.513648146413068e-05, "loss": 6.369, "step": 13346 }, { "epoch": 0.5627252988173789, "grad_norm": 1.0518440008163452, "learning_rate": 4.51357006777226e-05, "loss": 6.353, "step": 13347 }, { "epoch": 0.5627674599995783, "grad_norm": 0.9397681951522827, "learning_rate": 4.5134919891314535e-05, "loss": 6.5622, "step": 13348 }, { "epoch": 0.5628096211817779, "grad_norm": 1.3439466953277588, "learning_rate": 4.5134139104906466e-05, "loss": 6.3991, "step": 13349 }, { "epoch": 0.5628517823639775, "grad_norm": 1.433050274848938, "learning_rate": 4.513335831849839e-05, "loss": 7.0325, "step": 13350 }, { "epoch": 0.5628939435461771, "grad_norm": 2.9344778060913086, "learning_rate": 4.513257753209033e-05, "loss": 8.7711, "step": 13351 }, { "epoch": 0.5629361047283766, "grad_norm": 2.8679511547088623, "learning_rate": 4.513179674568225e-05, "loss": 7.1102, "step": 13352 }, { "epoch": 0.5629782659105761, "grad_norm": 1.3740140199661255, "learning_rate": 4.5131015959274184e-05, "loss": 6.3573, "step": 13353 }, { "epoch": 0.5630204270927757, "grad_norm": 1.0320672988891602, "learning_rate": 4.5130235172866115e-05, "loss": 6.6031, "step": 13354 }, { "epoch": 0.5630625882749752, "grad_norm": 1.355630874633789, "learning_rate": 4.512945438645804e-05, "loss": 7.1476, "step": 13355 }, { "epoch": 0.5631047494571748, "grad_norm": 2.2975993156433105, "learning_rate": 4.512867360004997e-05, "loss": 7.2133, "step": 13356 }, { "epoch": 0.5631469106393743, "grad_norm": 2.6690330505371094, "learning_rate": 4.51278928136419e-05, "loss": 6.2395, "step": 13357 }, { "epoch": 0.5631890718215739, "grad_norm": 1.580995798110962, "learning_rate": 4.5127112027233834e-05, "loss": 7.2163, "step": 13358 }, { "epoch": 0.5632312330037734, "grad_norm": 1.2730110883712769, "learning_rate": 4.5126331240825765e-05, "loss": 6.583, "step": 13359 }, { "epoch": 0.563273394185973, "grad_norm": 2.7038333415985107, "learning_rate": 4.512555045441769e-05, "loss": 7.6752, "step": 13360 }, { "epoch": 0.5633155553681726, "grad_norm": 1.0660128593444824, "learning_rate": 4.512476966800962e-05, "loss": 6.5513, "step": 13361 }, { "epoch": 0.563357716550372, "grad_norm": 1.2705059051513672, "learning_rate": 4.512398888160155e-05, "loss": 7.1225, "step": 13362 }, { "epoch": 0.5633998777325716, "grad_norm": 1.1657828092575073, "learning_rate": 4.512320809519348e-05, "loss": 6.3279, "step": 13363 }, { "epoch": 0.5634420389147712, "grad_norm": 2.7199513912200928, "learning_rate": 4.512242730878541e-05, "loss": 7.8516, "step": 13364 }, { "epoch": 0.5634842000969708, "grad_norm": 1.0161290168762207, "learning_rate": 4.512164652237734e-05, "loss": 6.8629, "step": 13365 }, { "epoch": 0.5635263612791702, "grad_norm": 2.284733295440674, "learning_rate": 4.512086573596927e-05, "loss": 7.6347, "step": 13366 }, { "epoch": 0.5635685224613698, "grad_norm": 4.048526763916016, "learning_rate": 4.5120084949561195e-05, "loss": 7.2595, "step": 13367 }, { "epoch": 0.5636106836435694, "grad_norm": 1.6626406908035278, "learning_rate": 4.511930416315313e-05, "loss": 6.93, "step": 13368 }, { "epoch": 0.5636528448257689, "grad_norm": 1.1956770420074463, "learning_rate": 4.511852337674506e-05, "loss": 6.6463, "step": 13369 }, { "epoch": 0.5636950060079685, "grad_norm": 1.4298113584518433, "learning_rate": 4.5117742590336995e-05, "loss": 6.8409, "step": 13370 }, { "epoch": 0.563737167190168, "grad_norm": 0.8987817168235779, "learning_rate": 4.511696180392892e-05, "loss": 6.7069, "step": 13371 }, { "epoch": 0.5637793283723675, "grad_norm": 0.8327142596244812, "learning_rate": 4.511618101752085e-05, "loss": 6.7615, "step": 13372 }, { "epoch": 0.5638214895545671, "grad_norm": 2.3774983882904053, "learning_rate": 4.511540023111278e-05, "loss": 7.2479, "step": 13373 }, { "epoch": 0.5638636507367667, "grad_norm": 1.2809009552001953, "learning_rate": 4.511461944470471e-05, "loss": 6.3575, "step": 13374 }, { "epoch": 0.5639058119189662, "grad_norm": 2.6007344722747803, "learning_rate": 4.511383865829664e-05, "loss": 8.3259, "step": 13375 }, { "epoch": 0.5639479731011657, "grad_norm": 1.6257505416870117, "learning_rate": 4.511305787188857e-05, "loss": 6.5238, "step": 13376 }, { "epoch": 0.5639901342833653, "grad_norm": 2.645977735519409, "learning_rate": 4.51122770854805e-05, "loss": 7.6376, "step": 13377 }, { "epoch": 0.5640322954655649, "grad_norm": 2.497539520263672, "learning_rate": 4.5111496299072425e-05, "loss": 7.4279, "step": 13378 }, { "epoch": 0.5640744566477645, "grad_norm": 1.7429523468017578, "learning_rate": 4.5110715512664356e-05, "loss": 6.8067, "step": 13379 }, { "epoch": 0.5641166178299639, "grad_norm": 1.9385818243026733, "learning_rate": 4.510993472625629e-05, "loss": 6.289, "step": 13380 }, { "epoch": 0.5641587790121635, "grad_norm": 1.3502089977264404, "learning_rate": 4.510915393984822e-05, "loss": 6.7925, "step": 13381 }, { "epoch": 0.5642009401943631, "grad_norm": 1.5632556676864624, "learning_rate": 4.510837315344015e-05, "loss": 7.5632, "step": 13382 }, { "epoch": 0.5642431013765626, "grad_norm": 1.7057665586471558, "learning_rate": 4.5107592367032074e-05, "loss": 7.1243, "step": 13383 }, { "epoch": 0.5642852625587621, "grad_norm": 1.266159176826477, "learning_rate": 4.5106811580624006e-05, "loss": 6.4108, "step": 13384 }, { "epoch": 0.5643274237409617, "grad_norm": 2.538116216659546, "learning_rate": 4.510603079421594e-05, "loss": 7.8175, "step": 13385 }, { "epoch": 0.5643695849231612, "grad_norm": 1.335556983947754, "learning_rate": 4.510525000780786e-05, "loss": 7.0176, "step": 13386 }, { "epoch": 0.5644117461053608, "grad_norm": 1.4547349214553833, "learning_rate": 4.51044692213998e-05, "loss": 7.3365, "step": 13387 }, { "epoch": 0.5644539072875604, "grad_norm": 0.9198090434074402, "learning_rate": 4.5103688434991724e-05, "loss": 6.5808, "step": 13388 }, { "epoch": 0.5644960684697599, "grad_norm": 2.16497540473938, "learning_rate": 4.5102907648583655e-05, "loss": 6.2296, "step": 13389 }, { "epoch": 0.5645382296519594, "grad_norm": 1.7940824031829834, "learning_rate": 4.5102126862175586e-05, "loss": 6.4378, "step": 13390 }, { "epoch": 0.564580390834159, "grad_norm": 2.702634811401367, "learning_rate": 4.510134607576751e-05, "loss": 7.9659, "step": 13391 }, { "epoch": 0.5646225520163586, "grad_norm": 3.1075007915496826, "learning_rate": 4.510056528935945e-05, "loss": 7.2598, "step": 13392 }, { "epoch": 0.564664713198558, "grad_norm": 1.9173766374588013, "learning_rate": 4.5099784502951373e-05, "loss": 7.2889, "step": 13393 }, { "epoch": 0.5647068743807576, "grad_norm": 1.3001470565795898, "learning_rate": 4.5099003716543305e-05, "loss": 6.5322, "step": 13394 }, { "epoch": 0.5647490355629572, "grad_norm": 1.8481321334838867, "learning_rate": 4.5098222930135236e-05, "loss": 6.0724, "step": 13395 }, { "epoch": 0.5647911967451568, "grad_norm": 1.4527024030685425, "learning_rate": 4.509744214372717e-05, "loss": 6.7793, "step": 13396 }, { "epoch": 0.5648333579273562, "grad_norm": 3.8563783168792725, "learning_rate": 4.509666135731909e-05, "loss": 7.6214, "step": 13397 }, { "epoch": 0.5648755191095558, "grad_norm": 1.5954402685165405, "learning_rate": 4.509588057091102e-05, "loss": 7.0001, "step": 13398 }, { "epoch": 0.5649176802917554, "grad_norm": 3.317701816558838, "learning_rate": 4.5095099784502954e-05, "loss": 7.582, "step": 13399 }, { "epoch": 0.564959841473955, "grad_norm": 2.528629779815674, "learning_rate": 4.509431899809488e-05, "loss": 7.9426, "step": 13400 }, { "epoch": 0.5650020026561545, "grad_norm": 1.3834178447723389, "learning_rate": 4.509353821168682e-05, "loss": 6.3898, "step": 13401 }, { "epoch": 0.565044163838354, "grad_norm": 1.373300313949585, "learning_rate": 4.509275742527874e-05, "loss": 7.1305, "step": 13402 }, { "epoch": 0.5650863250205536, "grad_norm": 1.4988149404525757, "learning_rate": 4.509197663887067e-05, "loss": 7.1566, "step": 13403 }, { "epoch": 0.5651284862027531, "grad_norm": 2.590001344680786, "learning_rate": 4.5091195852462604e-05, "loss": 7.5084, "step": 13404 }, { "epoch": 0.5651706473849527, "grad_norm": 1.232589602470398, "learning_rate": 4.509041506605453e-05, "loss": 7.7347, "step": 13405 }, { "epoch": 0.5652128085671522, "grad_norm": 1.2677291631698608, "learning_rate": 4.5089634279646466e-05, "loss": 6.9805, "step": 13406 }, { "epoch": 0.5652549697493517, "grad_norm": 1.4602104425430298, "learning_rate": 4.508885349323839e-05, "loss": 7.8016, "step": 13407 }, { "epoch": 0.5652971309315513, "grad_norm": 2.05892276763916, "learning_rate": 4.508807270683032e-05, "loss": 8.4103, "step": 13408 }, { "epoch": 0.5653392921137509, "grad_norm": 1.5611438751220703, "learning_rate": 4.508729192042225e-05, "loss": 6.399, "step": 13409 }, { "epoch": 0.5653814532959505, "grad_norm": 1.9152692556381226, "learning_rate": 4.508651113401418e-05, "loss": 6.5281, "step": 13410 }, { "epoch": 0.5654236144781499, "grad_norm": 1.2356679439544678, "learning_rate": 4.508573034760611e-05, "loss": 6.9542, "step": 13411 }, { "epoch": 0.5654657756603495, "grad_norm": 1.179765224456787, "learning_rate": 4.508494956119804e-05, "loss": 6.7319, "step": 13412 }, { "epoch": 0.5655079368425491, "grad_norm": 1.399836540222168, "learning_rate": 4.508416877478997e-05, "loss": 7.149, "step": 13413 }, { "epoch": 0.5655500980247486, "grad_norm": 2.0812811851501465, "learning_rate": 4.5083387988381896e-05, "loss": 6.4599, "step": 13414 }, { "epoch": 0.5655922592069481, "grad_norm": 1.7112038135528564, "learning_rate": 4.5082607201973834e-05, "loss": 6.336, "step": 13415 }, { "epoch": 0.5656344203891477, "grad_norm": 0.8200690150260925, "learning_rate": 4.508182641556576e-05, "loss": 6.6077, "step": 13416 }, { "epoch": 0.5656765815713473, "grad_norm": 0.9628126621246338, "learning_rate": 4.508104562915769e-05, "loss": 6.4879, "step": 13417 }, { "epoch": 0.5657187427535468, "grad_norm": 0.858880877494812, "learning_rate": 4.508026484274962e-05, "loss": 6.4639, "step": 13418 }, { "epoch": 0.5657609039357464, "grad_norm": 1.2362732887268066, "learning_rate": 4.5079484056341546e-05, "loss": 6.2737, "step": 13419 }, { "epoch": 0.5658030651179459, "grad_norm": 1.2682247161865234, "learning_rate": 4.5078703269933484e-05, "loss": 6.3665, "step": 13420 }, { "epoch": 0.5658452263001454, "grad_norm": 1.5047876834869385, "learning_rate": 4.507792248352541e-05, "loss": 6.7312, "step": 13421 }, { "epoch": 0.565887387482345, "grad_norm": 1.7772338390350342, "learning_rate": 4.507714169711734e-05, "loss": 7.6083, "step": 13422 }, { "epoch": 0.5659295486645446, "grad_norm": 1.3867778778076172, "learning_rate": 4.507636091070927e-05, "loss": 6.3765, "step": 13423 }, { "epoch": 0.565971709846744, "grad_norm": 0.7804117798805237, "learning_rate": 4.5075580124301195e-05, "loss": 6.3593, "step": 13424 }, { "epoch": 0.5660138710289436, "grad_norm": 1.8667391538619995, "learning_rate": 4.5074799337893126e-05, "loss": 7.2212, "step": 13425 }, { "epoch": 0.5660560322111432, "grad_norm": 1.8776072263717651, "learning_rate": 4.507401855148506e-05, "loss": 6.5016, "step": 13426 }, { "epoch": 0.5660981933933428, "grad_norm": 0.8921732902526855, "learning_rate": 4.507323776507699e-05, "loss": 6.4158, "step": 13427 }, { "epoch": 0.5661403545755423, "grad_norm": 1.617301106452942, "learning_rate": 4.507245697866892e-05, "loss": 7.1702, "step": 13428 }, { "epoch": 0.5661825157577418, "grad_norm": 0.9385973811149597, "learning_rate": 4.5071676192260845e-05, "loss": 6.4361, "step": 13429 }, { "epoch": 0.5662246769399414, "grad_norm": 2.456639289855957, "learning_rate": 4.5070895405852776e-05, "loss": 6.4782, "step": 13430 }, { "epoch": 0.566266838122141, "grad_norm": 3.774221181869507, "learning_rate": 4.507011461944471e-05, "loss": 7.9795, "step": 13431 }, { "epoch": 0.5663089993043405, "grad_norm": 1.091423511505127, "learning_rate": 4.506933383303664e-05, "loss": 6.9341, "step": 13432 }, { "epoch": 0.56635116048654, "grad_norm": 0.9916476607322693, "learning_rate": 4.506855304662856e-05, "loss": 6.4761, "step": 13433 }, { "epoch": 0.5663933216687396, "grad_norm": 2.505058765411377, "learning_rate": 4.50677722602205e-05, "loss": 7.8954, "step": 13434 }, { "epoch": 0.5664354828509391, "grad_norm": 0.9000751972198486, "learning_rate": 4.5066991473812425e-05, "loss": 6.7231, "step": 13435 }, { "epoch": 0.5664776440331387, "grad_norm": 0.9484280347824097, "learning_rate": 4.506621068740435e-05, "loss": 6.9044, "step": 13436 }, { "epoch": 0.5665198052153383, "grad_norm": 0.8284943103790283, "learning_rate": 4.506542990099629e-05, "loss": 6.4122, "step": 13437 }, { "epoch": 0.5665619663975378, "grad_norm": 2.833327531814575, "learning_rate": 4.506464911458821e-05, "loss": 6.8613, "step": 13438 }, { "epoch": 0.5666041275797373, "grad_norm": 4.029247283935547, "learning_rate": 4.506386832818015e-05, "loss": 7.7079, "step": 13439 }, { "epoch": 0.5666462887619369, "grad_norm": 2.14313006401062, "learning_rate": 4.5063087541772075e-05, "loss": 6.902, "step": 13440 }, { "epoch": 0.5666884499441365, "grad_norm": 0.9759231805801392, "learning_rate": 4.5062306755364006e-05, "loss": 6.3081, "step": 13441 }, { "epoch": 0.5667306111263359, "grad_norm": 1.0420252084732056, "learning_rate": 4.506152596895594e-05, "loss": 6.7816, "step": 13442 }, { "epoch": 0.5667727723085355, "grad_norm": 1.5209951400756836, "learning_rate": 4.506074518254786e-05, "loss": 7.0755, "step": 13443 }, { "epoch": 0.5668149334907351, "grad_norm": 1.105394721031189, "learning_rate": 4.505996439613979e-05, "loss": 6.3203, "step": 13444 }, { "epoch": 0.5668570946729347, "grad_norm": 1.265617847442627, "learning_rate": 4.5059183609731724e-05, "loss": 6.5206, "step": 13445 }, { "epoch": 0.5668992558551342, "grad_norm": 1.2999776601791382, "learning_rate": 4.5058402823323656e-05, "loss": 6.3568, "step": 13446 }, { "epoch": 0.5669414170373337, "grad_norm": 1.28837251663208, "learning_rate": 4.505762203691558e-05, "loss": 6.4723, "step": 13447 }, { "epoch": 0.5669835782195333, "grad_norm": 2.2993648052215576, "learning_rate": 4.505684125050751e-05, "loss": 7.9174, "step": 13448 }, { "epoch": 0.5670257394017328, "grad_norm": 1.8359522819519043, "learning_rate": 4.505606046409944e-05, "loss": 6.6723, "step": 13449 }, { "epoch": 0.5670679005839324, "grad_norm": 3.600276231765747, "learning_rate": 4.5055279677691374e-05, "loss": 7.5268, "step": 13450 }, { "epoch": 0.5671100617661319, "grad_norm": 2.5661158561706543, "learning_rate": 4.5054498891283305e-05, "loss": 7.163, "step": 13451 }, { "epoch": 0.5671522229483315, "grad_norm": 1.1330400705337524, "learning_rate": 4.505371810487523e-05, "loss": 6.77, "step": 13452 }, { "epoch": 0.567194384130531, "grad_norm": 1.109900712966919, "learning_rate": 4.505293731846716e-05, "loss": 6.8757, "step": 13453 }, { "epoch": 0.5672365453127306, "grad_norm": 4.422662258148193, "learning_rate": 4.505215653205909e-05, "loss": 7.541, "step": 13454 }, { "epoch": 0.5672787064949301, "grad_norm": 2.4993131160736084, "learning_rate": 4.505137574565102e-05, "loss": 6.5824, "step": 13455 }, { "epoch": 0.5673208676771296, "grad_norm": 0.9901347160339355, "learning_rate": 4.5050594959242955e-05, "loss": 6.4493, "step": 13456 }, { "epoch": 0.5673630288593292, "grad_norm": 2.7195472717285156, "learning_rate": 4.504981417283488e-05, "loss": 7.9248, "step": 13457 }, { "epoch": 0.5674051900415288, "grad_norm": 2.1514737606048584, "learning_rate": 4.504903338642681e-05, "loss": 7.1616, "step": 13458 }, { "epoch": 0.5674473512237284, "grad_norm": 1.1828083992004395, "learning_rate": 4.504825260001874e-05, "loss": 7.1201, "step": 13459 }, { "epoch": 0.5674895124059278, "grad_norm": 1.5695804357528687, "learning_rate": 4.504747181361067e-05, "loss": 7.1388, "step": 13460 }, { "epoch": 0.5675316735881274, "grad_norm": 1.9242043495178223, "learning_rate": 4.5046691027202604e-05, "loss": 7.2119, "step": 13461 }, { "epoch": 0.567573834770327, "grad_norm": 1.315948486328125, "learning_rate": 4.504591024079453e-05, "loss": 6.745, "step": 13462 }, { "epoch": 0.5676159959525265, "grad_norm": 1.361410140991211, "learning_rate": 4.504512945438646e-05, "loss": 6.3092, "step": 13463 }, { "epoch": 0.567658157134726, "grad_norm": 1.9863874912261963, "learning_rate": 4.504434866797839e-05, "loss": 7.3319, "step": 13464 }, { "epoch": 0.5677003183169256, "grad_norm": 1.2732256650924683, "learning_rate": 4.504356788157032e-05, "loss": 6.7458, "step": 13465 }, { "epoch": 0.5677424794991252, "grad_norm": 2.8305211067199707, "learning_rate": 4.504278709516225e-05, "loss": 7.9169, "step": 13466 }, { "epoch": 0.5677846406813247, "grad_norm": 1.3809000253677368, "learning_rate": 4.504200630875418e-05, "loss": 6.4227, "step": 13467 }, { "epoch": 0.5678268018635243, "grad_norm": 1.192696213722229, "learning_rate": 4.504122552234611e-05, "loss": 6.4631, "step": 13468 }, { "epoch": 0.5678689630457238, "grad_norm": 2.9089975357055664, "learning_rate": 4.5040444735938034e-05, "loss": 7.7109, "step": 13469 }, { "epoch": 0.5679111242279233, "grad_norm": 1.921802282333374, "learning_rate": 4.503966394952997e-05, "loss": 6.203, "step": 13470 }, { "epoch": 0.5679532854101229, "grad_norm": 1.3727601766586304, "learning_rate": 4.5038883163121897e-05, "loss": 6.6034, "step": 13471 }, { "epoch": 0.5679954465923225, "grad_norm": 1.044851541519165, "learning_rate": 4.503810237671383e-05, "loss": 6.8761, "step": 13472 }, { "epoch": 0.568037607774522, "grad_norm": 1.1647545099258423, "learning_rate": 4.503732159030576e-05, "loss": 6.3946, "step": 13473 }, { "epoch": 0.5680797689567215, "grad_norm": 1.2650247812271118, "learning_rate": 4.5036540803897684e-05, "loss": 6.3561, "step": 13474 }, { "epoch": 0.5681219301389211, "grad_norm": 1.0211279392242432, "learning_rate": 4.503576001748962e-05, "loss": 6.3117, "step": 13475 }, { "epoch": 0.5681640913211207, "grad_norm": 1.7913058996200562, "learning_rate": 4.5034979231081546e-05, "loss": 6.6807, "step": 13476 }, { "epoch": 0.5682062525033202, "grad_norm": 1.5699485540390015, "learning_rate": 4.503419844467348e-05, "loss": 6.357, "step": 13477 }, { "epoch": 0.5682484136855197, "grad_norm": 1.217486023902893, "learning_rate": 4.503341765826541e-05, "loss": 6.5712, "step": 13478 }, { "epoch": 0.5682905748677193, "grad_norm": 1.8564077615737915, "learning_rate": 4.503263687185734e-05, "loss": 6.9826, "step": 13479 }, { "epoch": 0.5683327360499189, "grad_norm": 1.3230091333389282, "learning_rate": 4.5031856085449264e-05, "loss": 7.2845, "step": 13480 }, { "epoch": 0.5683748972321184, "grad_norm": 2.029956340789795, "learning_rate": 4.5031075299041196e-05, "loss": 6.5422, "step": 13481 }, { "epoch": 0.5684170584143179, "grad_norm": 2.6067545413970947, "learning_rate": 4.503029451263313e-05, "loss": 8.5382, "step": 13482 }, { "epoch": 0.5684592195965175, "grad_norm": 2.6535048484802246, "learning_rate": 4.502951372622505e-05, "loss": 7.703, "step": 13483 }, { "epoch": 0.568501380778717, "grad_norm": 4.2398810386657715, "learning_rate": 4.502873293981699e-05, "loss": 7.232, "step": 13484 }, { "epoch": 0.5685435419609166, "grad_norm": 1.6220935583114624, "learning_rate": 4.5027952153408914e-05, "loss": 7.023, "step": 13485 }, { "epoch": 0.5685857031431162, "grad_norm": 1.2620375156402588, "learning_rate": 4.5027171367000845e-05, "loss": 6.5292, "step": 13486 }, { "epoch": 0.5686278643253156, "grad_norm": 2.1696789264678955, "learning_rate": 4.5026390580592776e-05, "loss": 6.8194, "step": 13487 }, { "epoch": 0.5686700255075152, "grad_norm": 3.580955743789673, "learning_rate": 4.50256097941847e-05, "loss": 8.0121, "step": 13488 }, { "epoch": 0.5687121866897148, "grad_norm": 1.7884259223937988, "learning_rate": 4.502482900777664e-05, "loss": 6.8359, "step": 13489 }, { "epoch": 0.5687543478719144, "grad_norm": 3.187516212463379, "learning_rate": 4.5024048221368563e-05, "loss": 6.2271, "step": 13490 }, { "epoch": 0.5687965090541138, "grad_norm": 1.250959038734436, "learning_rate": 4.5023267434960495e-05, "loss": 6.3947, "step": 13491 }, { "epoch": 0.5688386702363134, "grad_norm": 1.1059024333953857, "learning_rate": 4.5022486648552426e-05, "loss": 6.5826, "step": 13492 }, { "epoch": 0.568880831418513, "grad_norm": 1.0577312707901, "learning_rate": 4.502170586214435e-05, "loss": 6.4623, "step": 13493 }, { "epoch": 0.5689229926007126, "grad_norm": 2.540928363800049, "learning_rate": 4.502092507573628e-05, "loss": 6.26, "step": 13494 }, { "epoch": 0.5689651537829121, "grad_norm": 2.7487552165985107, "learning_rate": 4.502014428932821e-05, "loss": 8.002, "step": 13495 }, { "epoch": 0.5690073149651116, "grad_norm": 1.85516357421875, "learning_rate": 4.5019363502920144e-05, "loss": 7.1131, "step": 13496 }, { "epoch": 0.5690494761473112, "grad_norm": 4.093142986297607, "learning_rate": 4.5018582716512075e-05, "loss": 7.5946, "step": 13497 }, { "epoch": 0.5690916373295107, "grad_norm": 1.839554786682129, "learning_rate": 4.5017801930104e-05, "loss": 6.9907, "step": 13498 }, { "epoch": 0.5691337985117103, "grad_norm": 1.6684242486953735, "learning_rate": 4.501702114369593e-05, "loss": 6.936, "step": 13499 }, { "epoch": 0.5691759596939098, "grad_norm": 1.4999738931655884, "learning_rate": 4.501624035728786e-05, "loss": 6.4201, "step": 13500 }, { "epoch": 0.5692181208761093, "grad_norm": 1.963043212890625, "learning_rate": 4.5015459570879794e-05, "loss": 7.1394, "step": 13501 }, { "epoch": 0.5692602820583089, "grad_norm": 3.5730607509613037, "learning_rate": 4.501467878447172e-05, "loss": 8.1429, "step": 13502 }, { "epoch": 0.5693024432405085, "grad_norm": 2.117896795272827, "learning_rate": 4.5013897998063656e-05, "loss": 6.6268, "step": 13503 }, { "epoch": 0.5693446044227081, "grad_norm": 1.1733037233352661, "learning_rate": 4.501311721165558e-05, "loss": 6.8578, "step": 13504 }, { "epoch": 0.5693867656049075, "grad_norm": 1.3927268981933594, "learning_rate": 4.501233642524751e-05, "loss": 6.4266, "step": 13505 }, { "epoch": 0.5694289267871071, "grad_norm": 2.639714479446411, "learning_rate": 4.501155563883944e-05, "loss": 7.1888, "step": 13506 }, { "epoch": 0.5694710879693067, "grad_norm": 2.0477089881896973, "learning_rate": 4.501077485243137e-05, "loss": 6.411, "step": 13507 }, { "epoch": 0.5695132491515063, "grad_norm": 2.2394301891326904, "learning_rate": 4.5009994066023306e-05, "loss": 6.6593, "step": 13508 }, { "epoch": 0.5695554103337057, "grad_norm": 2.408064603805542, "learning_rate": 4.500921327961523e-05, "loss": 6.3341, "step": 13509 }, { "epoch": 0.5695975715159053, "grad_norm": 2.213592290878296, "learning_rate": 4.500843249320716e-05, "loss": 7.7367, "step": 13510 }, { "epoch": 0.5696397326981049, "grad_norm": 1.548404335975647, "learning_rate": 4.500765170679909e-05, "loss": 6.9476, "step": 13511 }, { "epoch": 0.5696818938803044, "grad_norm": 0.8146890997886658, "learning_rate": 4.500687092039102e-05, "loss": 6.5262, "step": 13512 }, { "epoch": 0.5697240550625039, "grad_norm": 1.5164976119995117, "learning_rate": 4.500609013398295e-05, "loss": 7.0607, "step": 13513 }, { "epoch": 0.5697662162447035, "grad_norm": 1.9258960485458374, "learning_rate": 4.500530934757488e-05, "loss": 6.5763, "step": 13514 }, { "epoch": 0.569808377426903, "grad_norm": 2.2381510734558105, "learning_rate": 4.500452856116681e-05, "loss": 6.686, "step": 13515 }, { "epoch": 0.5698505386091026, "grad_norm": 3.8020801544189453, "learning_rate": 4.5003747774758735e-05, "loss": 7.8952, "step": 13516 }, { "epoch": 0.5698926997913022, "grad_norm": 1.132673978805542, "learning_rate": 4.500296698835067e-05, "loss": 6.5003, "step": 13517 }, { "epoch": 0.5699348609735017, "grad_norm": 0.9905626773834229, "learning_rate": 4.50021862019426e-05, "loss": 6.5205, "step": 13518 }, { "epoch": 0.5699770221557012, "grad_norm": 1.352869987487793, "learning_rate": 4.500140541553453e-05, "loss": 6.8444, "step": 13519 }, { "epoch": 0.5700191833379008, "grad_norm": 1.935511589050293, "learning_rate": 4.500062462912646e-05, "loss": 6.4868, "step": 13520 }, { "epoch": 0.5700613445201004, "grad_norm": 3.1022887229919434, "learning_rate": 4.4999843842718385e-05, "loss": 6.6795, "step": 13521 }, { "epoch": 0.5701035057022998, "grad_norm": 1.9209858179092407, "learning_rate": 4.499906305631032e-05, "loss": 6.4519, "step": 13522 }, { "epoch": 0.5701456668844994, "grad_norm": 2.6326489448547363, "learning_rate": 4.499828226990225e-05, "loss": 6.17, "step": 13523 }, { "epoch": 0.570187828066699, "grad_norm": 1.9479601383209229, "learning_rate": 4.499750148349418e-05, "loss": 6.4077, "step": 13524 }, { "epoch": 0.5702299892488986, "grad_norm": 2.0742239952087402, "learning_rate": 4.499672069708611e-05, "loss": 7.2257, "step": 13525 }, { "epoch": 0.5702721504310981, "grad_norm": 0.9092978239059448, "learning_rate": 4.4995939910678035e-05, "loss": 6.7213, "step": 13526 }, { "epoch": 0.5703143116132976, "grad_norm": 1.2484586238861084, "learning_rate": 4.4995159124269966e-05, "loss": 6.3431, "step": 13527 }, { "epoch": 0.5703564727954972, "grad_norm": 1.218145489692688, "learning_rate": 4.49943783378619e-05, "loss": 7.2962, "step": 13528 }, { "epoch": 0.5703986339776967, "grad_norm": 1.1986665725708008, "learning_rate": 4.499359755145383e-05, "loss": 6.5293, "step": 13529 }, { "epoch": 0.5704407951598963, "grad_norm": 1.251435399055481, "learning_rate": 4.499281676504576e-05, "loss": 6.6975, "step": 13530 }, { "epoch": 0.5704829563420958, "grad_norm": 1.4004476070404053, "learning_rate": 4.4992035978637684e-05, "loss": 6.9619, "step": 13531 }, { "epoch": 0.5705251175242954, "grad_norm": 0.996062159538269, "learning_rate": 4.4991255192229615e-05, "loss": 6.4387, "step": 13532 }, { "epoch": 0.5705672787064949, "grad_norm": 1.0363718271255493, "learning_rate": 4.4990474405821547e-05, "loss": 6.4035, "step": 13533 }, { "epoch": 0.5706094398886945, "grad_norm": 2.818478584289551, "learning_rate": 4.498969361941348e-05, "loss": 6.2872, "step": 13534 }, { "epoch": 0.5706516010708941, "grad_norm": 0.920201301574707, "learning_rate": 4.49889128330054e-05, "loss": 6.7538, "step": 13535 }, { "epoch": 0.5706937622530935, "grad_norm": 0.7694056034088135, "learning_rate": 4.4988132046597334e-05, "loss": 6.4503, "step": 13536 }, { "epoch": 0.5707359234352931, "grad_norm": 1.536399483680725, "learning_rate": 4.4987351260189265e-05, "loss": 7.1116, "step": 13537 }, { "epoch": 0.5707780846174927, "grad_norm": 0.8520563840866089, "learning_rate": 4.498657047378119e-05, "loss": 7.015, "step": 13538 }, { "epoch": 0.5708202457996923, "grad_norm": 1.0669324398040771, "learning_rate": 4.498578968737313e-05, "loss": 6.5412, "step": 13539 }, { "epoch": 0.5708624069818917, "grad_norm": 1.111201286315918, "learning_rate": 4.498500890096505e-05, "loss": 7.0556, "step": 13540 }, { "epoch": 0.5709045681640913, "grad_norm": 1.6502583026885986, "learning_rate": 4.498422811455699e-05, "loss": 7.0628, "step": 13541 }, { "epoch": 0.5709467293462909, "grad_norm": 1.4876646995544434, "learning_rate": 4.4983447328148914e-05, "loss": 7.2426, "step": 13542 }, { "epoch": 0.5709888905284904, "grad_norm": 0.9199610948562622, "learning_rate": 4.498266654174084e-05, "loss": 6.3811, "step": 13543 }, { "epoch": 0.57103105171069, "grad_norm": 1.4499053955078125, "learning_rate": 4.498188575533278e-05, "loss": 6.4296, "step": 13544 }, { "epoch": 0.5710732128928895, "grad_norm": 1.5531400442123413, "learning_rate": 4.49811049689247e-05, "loss": 6.9898, "step": 13545 }, { "epoch": 0.571115374075089, "grad_norm": 3.9780290126800537, "learning_rate": 4.498032418251663e-05, "loss": 8.077, "step": 13546 }, { "epoch": 0.5711575352572886, "grad_norm": 2.0630621910095215, "learning_rate": 4.4979543396108564e-05, "loss": 7.5066, "step": 13547 }, { "epoch": 0.5711996964394882, "grad_norm": 1.4676494598388672, "learning_rate": 4.4978762609700495e-05, "loss": 6.4908, "step": 13548 }, { "epoch": 0.5712418576216877, "grad_norm": 1.786860466003418, "learning_rate": 4.497798182329242e-05, "loss": 6.8582, "step": 13549 }, { "epoch": 0.5712840188038872, "grad_norm": 1.6790043115615845, "learning_rate": 4.497720103688435e-05, "loss": 6.3269, "step": 13550 }, { "epoch": 0.5713261799860868, "grad_norm": 1.2404237985610962, "learning_rate": 4.497642025047628e-05, "loss": 6.5199, "step": 13551 }, { "epoch": 0.5713683411682864, "grad_norm": 1.5736435651779175, "learning_rate": 4.4975639464068213e-05, "loss": 6.6598, "step": 13552 }, { "epoch": 0.571410502350486, "grad_norm": 2.310213088989258, "learning_rate": 4.4974858677660145e-05, "loss": 7.2566, "step": 13553 }, { "epoch": 0.5714526635326854, "grad_norm": 2.53426456451416, "learning_rate": 4.497407789125207e-05, "loss": 7.2942, "step": 13554 }, { "epoch": 0.571494824714885, "grad_norm": 1.6087180376052856, "learning_rate": 4.4973297104844e-05, "loss": 7.1486, "step": 13555 }, { "epoch": 0.5715369858970846, "grad_norm": 2.08013916015625, "learning_rate": 4.497251631843593e-05, "loss": 7.0404, "step": 13556 }, { "epoch": 0.5715791470792841, "grad_norm": 1.4991121292114258, "learning_rate": 4.4971735532027856e-05, "loss": 7.3422, "step": 13557 }, { "epoch": 0.5716213082614836, "grad_norm": 1.4284123182296753, "learning_rate": 4.4970954745619794e-05, "loss": 6.3876, "step": 13558 }, { "epoch": 0.5716634694436832, "grad_norm": 1.5649994611740112, "learning_rate": 4.497017395921172e-05, "loss": 6.742, "step": 13559 }, { "epoch": 0.5717056306258828, "grad_norm": 2.033874988555908, "learning_rate": 4.496939317280365e-05, "loss": 7.7374, "step": 13560 }, { "epoch": 0.5717477918080823, "grad_norm": 2.043388605117798, "learning_rate": 4.496861238639558e-05, "loss": 6.4984, "step": 13561 }, { "epoch": 0.5717899529902819, "grad_norm": 1.281991958618164, "learning_rate": 4.4967831599987506e-05, "loss": 6.4527, "step": 13562 }, { "epoch": 0.5718321141724814, "grad_norm": 2.8483245372772217, "learning_rate": 4.496705081357944e-05, "loss": 6.5418, "step": 13563 }, { "epoch": 0.5718742753546809, "grad_norm": 1.9633415937423706, "learning_rate": 4.496627002717137e-05, "loss": 6.3729, "step": 13564 }, { "epoch": 0.5719164365368805, "grad_norm": 1.071616530418396, "learning_rate": 4.49654892407633e-05, "loss": 6.4963, "step": 13565 }, { "epoch": 0.5719585977190801, "grad_norm": 2.547057628631592, "learning_rate": 4.496470845435523e-05, "loss": 7.3038, "step": 13566 }, { "epoch": 0.5720007589012795, "grad_norm": 1.740390658378601, "learning_rate": 4.496392766794716e-05, "loss": 6.3058, "step": 13567 }, { "epoch": 0.5720429200834791, "grad_norm": 0.9674397110939026, "learning_rate": 4.4963146881539086e-05, "loss": 6.3901, "step": 13568 }, { "epoch": 0.5720850812656787, "grad_norm": 1.8700865507125854, "learning_rate": 4.496236609513102e-05, "loss": 6.869, "step": 13569 }, { "epoch": 0.5721272424478783, "grad_norm": 1.4355100393295288, "learning_rate": 4.496158530872295e-05, "loss": 6.1859, "step": 13570 }, { "epoch": 0.5721694036300777, "grad_norm": 0.7326136827468872, "learning_rate": 4.4960804522314873e-05, "loss": 6.693, "step": 13571 }, { "epoch": 0.5722115648122773, "grad_norm": 3.7618091106414795, "learning_rate": 4.496002373590681e-05, "loss": 6.9197, "step": 13572 }, { "epoch": 0.5722537259944769, "grad_norm": 2.1045634746551514, "learning_rate": 4.4959242949498736e-05, "loss": 7.5608, "step": 13573 }, { "epoch": 0.5722958871766765, "grad_norm": 1.3530727624893188, "learning_rate": 4.495846216309067e-05, "loss": 6.4302, "step": 13574 }, { "epoch": 0.572338048358876, "grad_norm": 1.0726128816604614, "learning_rate": 4.49576813766826e-05, "loss": 6.3196, "step": 13575 }, { "epoch": 0.5723802095410755, "grad_norm": 1.1734306812286377, "learning_rate": 4.495690059027452e-05, "loss": 6.894, "step": 13576 }, { "epoch": 0.5724223707232751, "grad_norm": 6.109511375427246, "learning_rate": 4.495611980386646e-05, "loss": 10.9985, "step": 13577 }, { "epoch": 0.5724645319054746, "grad_norm": 2.285747766494751, "learning_rate": 4.4955339017458386e-05, "loss": 6.8535, "step": 13578 }, { "epoch": 0.5725066930876742, "grad_norm": 1.7099034786224365, "learning_rate": 4.495455823105032e-05, "loss": 6.7525, "step": 13579 }, { "epoch": 0.5725488542698737, "grad_norm": 1.813086748123169, "learning_rate": 4.495377744464225e-05, "loss": 6.4114, "step": 13580 }, { "epoch": 0.5725910154520732, "grad_norm": 1.794350504875183, "learning_rate": 4.495299665823417e-05, "loss": 7.2048, "step": 13581 }, { "epoch": 0.5726331766342728, "grad_norm": 1.8707362413406372, "learning_rate": 4.4952215871826104e-05, "loss": 7.0745, "step": 13582 }, { "epoch": 0.5726753378164724, "grad_norm": 0.8780477643013, "learning_rate": 4.4951435085418035e-05, "loss": 6.5561, "step": 13583 }, { "epoch": 0.572717498998672, "grad_norm": 1.0122464895248413, "learning_rate": 4.4950654299009966e-05, "loss": 6.3709, "step": 13584 }, { "epoch": 0.5727596601808714, "grad_norm": 0.9458169937133789, "learning_rate": 4.494987351260189e-05, "loss": 6.6943, "step": 13585 }, { "epoch": 0.572801821363071, "grad_norm": 1.0749082565307617, "learning_rate": 4.494909272619383e-05, "loss": 6.5302, "step": 13586 }, { "epoch": 0.5728439825452706, "grad_norm": 1.5716041326522827, "learning_rate": 4.494831193978575e-05, "loss": 6.9074, "step": 13587 }, { "epoch": 0.5728861437274702, "grad_norm": 3.2736270427703857, "learning_rate": 4.4947531153377685e-05, "loss": 7.6341, "step": 13588 }, { "epoch": 0.5729283049096696, "grad_norm": 1.0912504196166992, "learning_rate": 4.4946750366969616e-05, "loss": 6.4743, "step": 13589 }, { "epoch": 0.5729704660918692, "grad_norm": 1.457697868347168, "learning_rate": 4.494596958056154e-05, "loss": 6.7458, "step": 13590 }, { "epoch": 0.5730126272740688, "grad_norm": 1.091570258140564, "learning_rate": 4.494518879415348e-05, "loss": 6.532, "step": 13591 }, { "epoch": 0.5730547884562683, "grad_norm": 3.1708953380584717, "learning_rate": 4.49444080077454e-05, "loss": 8.0158, "step": 13592 }, { "epoch": 0.5730969496384679, "grad_norm": 2.503098249435425, "learning_rate": 4.4943627221337334e-05, "loss": 7.2319, "step": 13593 }, { "epoch": 0.5731391108206674, "grad_norm": 1.4173812866210938, "learning_rate": 4.4942846434929265e-05, "loss": 7.5556, "step": 13594 }, { "epoch": 0.573181272002867, "grad_norm": 1.2059779167175293, "learning_rate": 4.494206564852119e-05, "loss": 6.6902, "step": 13595 }, { "epoch": 0.5732234331850665, "grad_norm": 1.9323842525482178, "learning_rate": 4.494128486211312e-05, "loss": 6.6811, "step": 13596 }, { "epoch": 0.5732655943672661, "grad_norm": 1.497712254524231, "learning_rate": 4.494050407570505e-05, "loss": 7.224, "step": 13597 }, { "epoch": 0.5733077555494656, "grad_norm": 2.07682466506958, "learning_rate": 4.4939723289296984e-05, "loss": 6.5359, "step": 13598 }, { "epoch": 0.5733499167316651, "grad_norm": 2.2950472831726074, "learning_rate": 4.4938942502888915e-05, "loss": 6.2989, "step": 13599 }, { "epoch": 0.5733920779138647, "grad_norm": 1.74684476852417, "learning_rate": 4.493816171648084e-05, "loss": 7.4812, "step": 13600 }, { "epoch": 0.5734342390960643, "grad_norm": 1.0947047472000122, "learning_rate": 4.493738093007277e-05, "loss": 6.9979, "step": 13601 }, { "epoch": 0.5734764002782639, "grad_norm": 2.065993070602417, "learning_rate": 4.49366001436647e-05, "loss": 7.5298, "step": 13602 }, { "epoch": 0.5735185614604633, "grad_norm": 1.8860024213790894, "learning_rate": 4.493581935725663e-05, "loss": 8.028, "step": 13603 }, { "epoch": 0.5735607226426629, "grad_norm": 1.876175045967102, "learning_rate": 4.493503857084856e-05, "loss": 6.465, "step": 13604 }, { "epoch": 0.5736028838248625, "grad_norm": 2.4398183822631836, "learning_rate": 4.493425778444049e-05, "loss": 7.6168, "step": 13605 }, { "epoch": 0.573645045007062, "grad_norm": 1.5643527507781982, "learning_rate": 4.493347699803242e-05, "loss": 6.2861, "step": 13606 }, { "epoch": 0.5736872061892615, "grad_norm": 1.6808722019195557, "learning_rate": 4.4932696211624345e-05, "loss": 6.3761, "step": 13607 }, { "epoch": 0.5737293673714611, "grad_norm": 2.6723124980926514, "learning_rate": 4.493191542521628e-05, "loss": 7.6459, "step": 13608 }, { "epoch": 0.5737715285536606, "grad_norm": 1.650338888168335, "learning_rate": 4.493113463880821e-05, "loss": 6.7759, "step": 13609 }, { "epoch": 0.5738136897358602, "grad_norm": 2.8506317138671875, "learning_rate": 4.4930353852400145e-05, "loss": 6.2181, "step": 13610 }, { "epoch": 0.5738558509180598, "grad_norm": 1.3682626485824585, "learning_rate": 4.492957306599207e-05, "loss": 6.4693, "step": 13611 }, { "epoch": 0.5738980121002593, "grad_norm": 3.3633010387420654, "learning_rate": 4.4928792279584e-05, "loss": 8.0233, "step": 13612 }, { "epoch": 0.5739401732824588, "grad_norm": 2.5874974727630615, "learning_rate": 4.492801149317593e-05, "loss": 7.6138, "step": 13613 }, { "epoch": 0.5739823344646584, "grad_norm": 1.9768586158752441, "learning_rate": 4.492723070676786e-05, "loss": 6.5754, "step": 13614 }, { "epoch": 0.574024495646858, "grad_norm": 1.6851526498794556, "learning_rate": 4.492644992035979e-05, "loss": 6.3801, "step": 13615 }, { "epoch": 0.5740666568290574, "grad_norm": 1.7493754625320435, "learning_rate": 4.492566913395172e-05, "loss": 6.524, "step": 13616 }, { "epoch": 0.574108818011257, "grad_norm": 1.7000367641448975, "learning_rate": 4.492488834754365e-05, "loss": 6.4945, "step": 13617 }, { "epoch": 0.5741509791934566, "grad_norm": 1.1833239793777466, "learning_rate": 4.4924107561135575e-05, "loss": 6.8928, "step": 13618 }, { "epoch": 0.5741931403756562, "grad_norm": 1.2305387258529663, "learning_rate": 4.4923326774727506e-05, "loss": 6.7329, "step": 13619 }, { "epoch": 0.5742353015578557, "grad_norm": 1.1523349285125732, "learning_rate": 4.492254598831944e-05, "loss": 6.5185, "step": 13620 }, { "epoch": 0.5742774627400552, "grad_norm": 1.7887811660766602, "learning_rate": 4.492176520191137e-05, "loss": 7.2136, "step": 13621 }, { "epoch": 0.5743196239222548, "grad_norm": 1.9188430309295654, "learning_rate": 4.49209844155033e-05, "loss": 7.2325, "step": 13622 }, { "epoch": 0.5743617851044543, "grad_norm": 1.16733980178833, "learning_rate": 4.4920203629095224e-05, "loss": 7.0556, "step": 13623 }, { "epoch": 0.5744039462866539, "grad_norm": 1.4810417890548706, "learning_rate": 4.4919422842687156e-05, "loss": 6.3477, "step": 13624 }, { "epoch": 0.5744461074688534, "grad_norm": 0.9892929792404175, "learning_rate": 4.491864205627909e-05, "loss": 6.4896, "step": 13625 }, { "epoch": 0.574488268651053, "grad_norm": 2.091874122619629, "learning_rate": 4.491786126987101e-05, "loss": 6.4665, "step": 13626 }, { "epoch": 0.5745304298332525, "grad_norm": 2.3756046295166016, "learning_rate": 4.491708048346295e-05, "loss": 6.5154, "step": 13627 }, { "epoch": 0.5745725910154521, "grad_norm": 1.3425066471099854, "learning_rate": 4.4916299697054874e-05, "loss": 6.6159, "step": 13628 }, { "epoch": 0.5746147521976516, "grad_norm": 3.9646642208099365, "learning_rate": 4.4915518910646805e-05, "loss": 8.3841, "step": 13629 }, { "epoch": 0.5746569133798511, "grad_norm": 2.501201629638672, "learning_rate": 4.4914738124238736e-05, "loss": 7.1225, "step": 13630 }, { "epoch": 0.5746990745620507, "grad_norm": 0.8542658090591431, "learning_rate": 4.491395733783067e-05, "loss": 6.4904, "step": 13631 }, { "epoch": 0.5747412357442503, "grad_norm": 2.317403793334961, "learning_rate": 4.49131765514226e-05, "loss": 6.5989, "step": 13632 }, { "epoch": 0.5747833969264499, "grad_norm": 1.5917558670043945, "learning_rate": 4.4912395765014523e-05, "loss": 6.1716, "step": 13633 }, { "epoch": 0.5748255581086493, "grad_norm": 1.2014238834381104, "learning_rate": 4.4911614978606455e-05, "loss": 6.6932, "step": 13634 }, { "epoch": 0.5748677192908489, "grad_norm": 0.7989298105239868, "learning_rate": 4.4910834192198386e-05, "loss": 6.3172, "step": 13635 }, { "epoch": 0.5749098804730485, "grad_norm": 1.7017900943756104, "learning_rate": 4.491005340579032e-05, "loss": 6.4598, "step": 13636 }, { "epoch": 0.574952041655248, "grad_norm": 2.676021099090576, "learning_rate": 4.490927261938224e-05, "loss": 7.0047, "step": 13637 }, { "epoch": 0.5749942028374475, "grad_norm": 2.818178415298462, "learning_rate": 4.490849183297417e-05, "loss": 6.3825, "step": 13638 }, { "epoch": 0.5750363640196471, "grad_norm": 1.6180896759033203, "learning_rate": 4.4907711046566104e-05, "loss": 6.845, "step": 13639 }, { "epoch": 0.5750785252018467, "grad_norm": 1.2070393562316895, "learning_rate": 4.490693026015803e-05, "loss": 6.407, "step": 13640 }, { "epoch": 0.5751206863840462, "grad_norm": 2.002486228942871, "learning_rate": 4.490614947374997e-05, "loss": 7.4266, "step": 13641 }, { "epoch": 0.5751628475662458, "grad_norm": 3.3263580799102783, "learning_rate": 4.490536868734189e-05, "loss": 8.1796, "step": 13642 }, { "epoch": 0.5752050087484453, "grad_norm": 1.891072392463684, "learning_rate": 4.490458790093382e-05, "loss": 7.5219, "step": 13643 }, { "epoch": 0.5752471699306448, "grad_norm": 3.1680634021759033, "learning_rate": 4.4903807114525754e-05, "loss": 7.9887, "step": 13644 }, { "epoch": 0.5752893311128444, "grad_norm": 2.509181022644043, "learning_rate": 4.490302632811768e-05, "loss": 6.7852, "step": 13645 }, { "epoch": 0.575331492295044, "grad_norm": 2.5477492809295654, "learning_rate": 4.4902245541709616e-05, "loss": 6.3253, "step": 13646 }, { "epoch": 0.5753736534772435, "grad_norm": 1.4395166635513306, "learning_rate": 4.490146475530154e-05, "loss": 6.6118, "step": 13647 }, { "epoch": 0.575415814659443, "grad_norm": 1.3322341442108154, "learning_rate": 4.490068396889347e-05, "loss": 6.6544, "step": 13648 }, { "epoch": 0.5754579758416426, "grad_norm": 4.206340789794922, "learning_rate": 4.48999031824854e-05, "loss": 8.0754, "step": 13649 }, { "epoch": 0.5755001370238422, "grad_norm": 2.0934267044067383, "learning_rate": 4.489912239607733e-05, "loss": 6.6559, "step": 13650 }, { "epoch": 0.5755422982060417, "grad_norm": 2.1018097400665283, "learning_rate": 4.489834160966926e-05, "loss": 6.5255, "step": 13651 }, { "epoch": 0.5755844593882412, "grad_norm": 3.0528688430786133, "learning_rate": 4.489756082326119e-05, "loss": 7.2702, "step": 13652 }, { "epoch": 0.5756266205704408, "grad_norm": 2.8753178119659424, "learning_rate": 4.489678003685312e-05, "loss": 7.9389, "step": 13653 }, { "epoch": 0.5756687817526404, "grad_norm": 1.3927706480026245, "learning_rate": 4.4895999250445046e-05, "loss": 6.6863, "step": 13654 }, { "epoch": 0.5757109429348399, "grad_norm": 2.2378647327423096, "learning_rate": 4.4895218464036984e-05, "loss": 7.9625, "step": 13655 }, { "epoch": 0.5757531041170394, "grad_norm": 2.378317356109619, "learning_rate": 4.489443767762891e-05, "loss": 6.2655, "step": 13656 }, { "epoch": 0.575795265299239, "grad_norm": 2.0355474948883057, "learning_rate": 4.489365689122084e-05, "loss": 6.4818, "step": 13657 }, { "epoch": 0.5758374264814385, "grad_norm": 2.270759105682373, "learning_rate": 4.489287610481277e-05, "loss": 7.8613, "step": 13658 }, { "epoch": 0.5758795876636381, "grad_norm": 1.4660162925720215, "learning_rate": 4.4892095318404696e-05, "loss": 6.4066, "step": 13659 }, { "epoch": 0.5759217488458377, "grad_norm": 1.3784679174423218, "learning_rate": 4.4891314531996634e-05, "loss": 6.7332, "step": 13660 }, { "epoch": 0.5759639100280372, "grad_norm": 1.0208148956298828, "learning_rate": 4.489053374558856e-05, "loss": 6.4707, "step": 13661 }, { "epoch": 0.5760060712102367, "grad_norm": 1.9133763313293457, "learning_rate": 4.488975295918049e-05, "loss": 6.2541, "step": 13662 }, { "epoch": 0.5760482323924363, "grad_norm": 1.4983645677566528, "learning_rate": 4.488897217277242e-05, "loss": 6.6161, "step": 13663 }, { "epoch": 0.5760903935746359, "grad_norm": 2.956688642501831, "learning_rate": 4.4888191386364345e-05, "loss": 8.352, "step": 13664 }, { "epoch": 0.5761325547568353, "grad_norm": 1.9888088703155518, "learning_rate": 4.4887410599956276e-05, "loss": 6.9196, "step": 13665 }, { "epoch": 0.5761747159390349, "grad_norm": 2.1198551654815674, "learning_rate": 4.488662981354821e-05, "loss": 7.278, "step": 13666 }, { "epoch": 0.5762168771212345, "grad_norm": 2.4157204627990723, "learning_rate": 4.488584902714014e-05, "loss": 7.7716, "step": 13667 }, { "epoch": 0.5762590383034341, "grad_norm": 1.7747207880020142, "learning_rate": 4.488506824073207e-05, "loss": 6.3564, "step": 13668 }, { "epoch": 0.5763011994856336, "grad_norm": 1.4126640558242798, "learning_rate": 4.4884287454323995e-05, "loss": 6.4439, "step": 13669 }, { "epoch": 0.5763433606678331, "grad_norm": 3.1825976371765137, "learning_rate": 4.4883506667915926e-05, "loss": 7.4054, "step": 13670 }, { "epoch": 0.5763855218500327, "grad_norm": 1.8174196481704712, "learning_rate": 4.488272588150786e-05, "loss": 6.3658, "step": 13671 }, { "epoch": 0.5764276830322322, "grad_norm": 2.871389389038086, "learning_rate": 4.488194509509979e-05, "loss": 7.6058, "step": 13672 }, { "epoch": 0.5764698442144318, "grad_norm": 1.601519227027893, "learning_rate": 4.488116430869171e-05, "loss": 6.7359, "step": 13673 }, { "epoch": 0.5765120053966313, "grad_norm": 1.830225944519043, "learning_rate": 4.488038352228365e-05, "loss": 6.8421, "step": 13674 }, { "epoch": 0.5765541665788309, "grad_norm": 1.214154839515686, "learning_rate": 4.4879602735875575e-05, "loss": 6.7357, "step": 13675 }, { "epoch": 0.5765963277610304, "grad_norm": 1.7401000261306763, "learning_rate": 4.48788219494675e-05, "loss": 6.7128, "step": 13676 }, { "epoch": 0.57663848894323, "grad_norm": 1.7296074628829956, "learning_rate": 4.487804116305944e-05, "loss": 7.1631, "step": 13677 }, { "epoch": 0.5766806501254296, "grad_norm": 2.7880752086639404, "learning_rate": 4.487726037665136e-05, "loss": 8.0335, "step": 13678 }, { "epoch": 0.576722811307629, "grad_norm": 2.1730611324310303, "learning_rate": 4.48764795902433e-05, "loss": 6.9999, "step": 13679 }, { "epoch": 0.5767649724898286, "grad_norm": 2.1458547115325928, "learning_rate": 4.4875698803835225e-05, "loss": 7.3427, "step": 13680 }, { "epoch": 0.5768071336720282, "grad_norm": 2.679666757583618, "learning_rate": 4.4874918017427156e-05, "loss": 7.9133, "step": 13681 }, { "epoch": 0.5768492948542278, "grad_norm": 2.3568758964538574, "learning_rate": 4.487413723101909e-05, "loss": 6.5231, "step": 13682 }, { "epoch": 0.5768914560364272, "grad_norm": 2.797396659851074, "learning_rate": 4.487335644461101e-05, "loss": 7.7288, "step": 13683 }, { "epoch": 0.5769336172186268, "grad_norm": 2.5129334926605225, "learning_rate": 4.487257565820294e-05, "loss": 6.2944, "step": 13684 }, { "epoch": 0.5769757784008264, "grad_norm": 1.6277588605880737, "learning_rate": 4.4871794871794874e-05, "loss": 6.7454, "step": 13685 }, { "epoch": 0.5770179395830259, "grad_norm": 2.5589563846588135, "learning_rate": 4.4871014085386806e-05, "loss": 7.0951, "step": 13686 }, { "epoch": 0.5770601007652254, "grad_norm": 3.0889458656311035, "learning_rate": 4.487023329897873e-05, "loss": 7.0335, "step": 13687 }, { "epoch": 0.577102261947425, "grad_norm": 2.6805930137634277, "learning_rate": 4.486945251257066e-05, "loss": 6.7218, "step": 13688 }, { "epoch": 0.5771444231296246, "grad_norm": 2.2781453132629395, "learning_rate": 4.486867172616259e-05, "loss": 6.4846, "step": 13689 }, { "epoch": 0.5771865843118241, "grad_norm": 3.7096621990203857, "learning_rate": 4.4867890939754524e-05, "loss": 7.8721, "step": 13690 }, { "epoch": 0.5772287454940237, "grad_norm": 1.882346749305725, "learning_rate": 4.4867110153346455e-05, "loss": 6.6284, "step": 13691 }, { "epoch": 0.5772709066762232, "grad_norm": 1.6030703783035278, "learning_rate": 4.486632936693838e-05, "loss": 6.4189, "step": 13692 }, { "epoch": 0.5773130678584227, "grad_norm": 2.0175602436065674, "learning_rate": 4.486554858053032e-05, "loss": 6.507, "step": 13693 }, { "epoch": 0.5773552290406223, "grad_norm": 2.0369904041290283, "learning_rate": 4.486476779412224e-05, "loss": 7.1327, "step": 13694 }, { "epoch": 0.5773973902228219, "grad_norm": 1.610300064086914, "learning_rate": 4.486398700771417e-05, "loss": 6.8305, "step": 13695 }, { "epoch": 0.5774395514050213, "grad_norm": 1.4535588026046753, "learning_rate": 4.4863206221306105e-05, "loss": 7.156, "step": 13696 }, { "epoch": 0.5774817125872209, "grad_norm": 1.2618701457977295, "learning_rate": 4.486242543489803e-05, "loss": 6.8422, "step": 13697 }, { "epoch": 0.5775238737694205, "grad_norm": 3.554558515548706, "learning_rate": 4.486164464848996e-05, "loss": 9.1815, "step": 13698 }, { "epoch": 0.5775660349516201, "grad_norm": 1.045352578163147, "learning_rate": 4.486086386208189e-05, "loss": 6.7554, "step": 13699 }, { "epoch": 0.5776081961338196, "grad_norm": 0.9691663384437561, "learning_rate": 4.486008307567382e-05, "loss": 7.0989, "step": 13700 }, { "epoch": 0.5776503573160191, "grad_norm": 1.0983525514602661, "learning_rate": 4.4859302289265754e-05, "loss": 6.7527, "step": 13701 }, { "epoch": 0.5776925184982187, "grad_norm": 1.3115733861923218, "learning_rate": 4.485852150285768e-05, "loss": 6.5829, "step": 13702 }, { "epoch": 0.5777346796804183, "grad_norm": 1.0547071695327759, "learning_rate": 4.485774071644961e-05, "loss": 6.6371, "step": 13703 }, { "epoch": 0.5777768408626178, "grad_norm": 1.7627217769622803, "learning_rate": 4.485695993004154e-05, "loss": 7.1769, "step": 13704 }, { "epoch": 0.5778190020448173, "grad_norm": 1.8489248752593994, "learning_rate": 4.485617914363347e-05, "loss": 7.1776, "step": 13705 }, { "epoch": 0.5778611632270169, "grad_norm": 2.0056264400482178, "learning_rate": 4.48553983572254e-05, "loss": 6.5451, "step": 13706 }, { "epoch": 0.5779033244092164, "grad_norm": 1.5045158863067627, "learning_rate": 4.485461757081733e-05, "loss": 7.045, "step": 13707 }, { "epoch": 0.577945485591416, "grad_norm": 1.192564845085144, "learning_rate": 4.485383678440926e-05, "loss": 6.6294, "step": 13708 }, { "epoch": 0.5779876467736156, "grad_norm": 2.9499948024749756, "learning_rate": 4.4853055998001184e-05, "loss": 7.8737, "step": 13709 }, { "epoch": 0.578029807955815, "grad_norm": 1.4835097789764404, "learning_rate": 4.485227521159312e-05, "loss": 7.0989, "step": 13710 }, { "epoch": 0.5780719691380146, "grad_norm": 2.8439855575561523, "learning_rate": 4.4851494425185047e-05, "loss": 8.386, "step": 13711 }, { "epoch": 0.5781141303202142, "grad_norm": 3.215637445449829, "learning_rate": 4.485071363877698e-05, "loss": 7.9674, "step": 13712 }, { "epoch": 0.5781562915024138, "grad_norm": 1.3301594257354736, "learning_rate": 4.484993285236891e-05, "loss": 7.1374, "step": 13713 }, { "epoch": 0.5781984526846132, "grad_norm": 2.107614755630493, "learning_rate": 4.4849152065960834e-05, "loss": 8.166, "step": 13714 }, { "epoch": 0.5782406138668128, "grad_norm": 2.023637294769287, "learning_rate": 4.484837127955277e-05, "loss": 8.0279, "step": 13715 }, { "epoch": 0.5782827750490124, "grad_norm": 2.0386569499969482, "learning_rate": 4.4847590493144696e-05, "loss": 6.5845, "step": 13716 }, { "epoch": 0.578324936231212, "grad_norm": 1.1451067924499512, "learning_rate": 4.484680970673663e-05, "loss": 6.8609, "step": 13717 }, { "epoch": 0.5783670974134115, "grad_norm": 1.2211952209472656, "learning_rate": 4.484602892032856e-05, "loss": 7.1181, "step": 13718 }, { "epoch": 0.578409258595611, "grad_norm": 1.6445422172546387, "learning_rate": 4.484524813392049e-05, "loss": 7.1648, "step": 13719 }, { "epoch": 0.5784514197778106, "grad_norm": 1.3172070980072021, "learning_rate": 4.4844467347512414e-05, "loss": 6.3297, "step": 13720 }, { "epoch": 0.5784935809600101, "grad_norm": 2.11063551902771, "learning_rate": 4.4843686561104346e-05, "loss": 7.7567, "step": 13721 }, { "epoch": 0.5785357421422097, "grad_norm": 0.9705243110656738, "learning_rate": 4.484290577469628e-05, "loss": 7.0558, "step": 13722 }, { "epoch": 0.5785779033244092, "grad_norm": 2.5527865886688232, "learning_rate": 4.48421249882882e-05, "loss": 6.5599, "step": 13723 }, { "epoch": 0.5786200645066087, "grad_norm": 1.4609941244125366, "learning_rate": 4.484134420188014e-05, "loss": 7.0383, "step": 13724 }, { "epoch": 0.5786622256888083, "grad_norm": 4.016117095947266, "learning_rate": 4.4840563415472064e-05, "loss": 7.0326, "step": 13725 }, { "epoch": 0.5787043868710079, "grad_norm": 2.301434278488159, "learning_rate": 4.4839782629063995e-05, "loss": 7.247, "step": 13726 }, { "epoch": 0.5787465480532075, "grad_norm": 3.038440704345703, "learning_rate": 4.4839001842655926e-05, "loss": 6.4197, "step": 13727 }, { "epoch": 0.5787887092354069, "grad_norm": 3.0200862884521484, "learning_rate": 4.483822105624785e-05, "loss": 6.5646, "step": 13728 }, { "epoch": 0.5788308704176065, "grad_norm": 2.7420618534088135, "learning_rate": 4.483744026983979e-05, "loss": 7.0167, "step": 13729 }, { "epoch": 0.5788730315998061, "grad_norm": 3.651186943054199, "learning_rate": 4.4836659483431713e-05, "loss": 7.2745, "step": 13730 }, { "epoch": 0.5789151927820056, "grad_norm": 1.638249158859253, "learning_rate": 4.4835878697023645e-05, "loss": 7.2, "step": 13731 }, { "epoch": 0.5789573539642051, "grad_norm": 1.2783281803131104, "learning_rate": 4.4835097910615576e-05, "loss": 6.5014, "step": 13732 }, { "epoch": 0.5789995151464047, "grad_norm": 1.1916415691375732, "learning_rate": 4.48343171242075e-05, "loss": 6.6439, "step": 13733 }, { "epoch": 0.5790416763286043, "grad_norm": 3.341123342514038, "learning_rate": 4.483353633779943e-05, "loss": 8.7801, "step": 13734 }, { "epoch": 0.5790838375108038, "grad_norm": 1.4377373456954956, "learning_rate": 4.483275555139136e-05, "loss": 6.4701, "step": 13735 }, { "epoch": 0.5791259986930033, "grad_norm": 1.327973484992981, "learning_rate": 4.4831974764983294e-05, "loss": 6.6988, "step": 13736 }, { "epoch": 0.5791681598752029, "grad_norm": 1.566627860069275, "learning_rate": 4.4831193978575225e-05, "loss": 6.9126, "step": 13737 }, { "epoch": 0.5792103210574024, "grad_norm": 1.1527516841888428, "learning_rate": 4.483041319216715e-05, "loss": 6.9694, "step": 13738 }, { "epoch": 0.579252482239602, "grad_norm": 1.5469920635223389, "learning_rate": 4.482963240575908e-05, "loss": 6.7139, "step": 13739 }, { "epoch": 0.5792946434218016, "grad_norm": 0.9846861958503723, "learning_rate": 4.482885161935101e-05, "loss": 6.6712, "step": 13740 }, { "epoch": 0.579336804604001, "grad_norm": 2.5465123653411865, "learning_rate": 4.4828070832942944e-05, "loss": 7.9242, "step": 13741 }, { "epoch": 0.5793789657862006, "grad_norm": 2.5962154865264893, "learning_rate": 4.482729004653487e-05, "loss": 6.2711, "step": 13742 }, { "epoch": 0.5794211269684002, "grad_norm": 1.281256914138794, "learning_rate": 4.4826509260126806e-05, "loss": 6.4613, "step": 13743 }, { "epoch": 0.5794632881505998, "grad_norm": 3.535985231399536, "learning_rate": 4.482572847371873e-05, "loss": 8.2847, "step": 13744 }, { "epoch": 0.5795054493327992, "grad_norm": 2.228827953338623, "learning_rate": 4.482494768731066e-05, "loss": 6.748, "step": 13745 }, { "epoch": 0.5795476105149988, "grad_norm": 1.4526722431182861, "learning_rate": 4.482416690090259e-05, "loss": 6.4145, "step": 13746 }, { "epoch": 0.5795897716971984, "grad_norm": 1.2541669607162476, "learning_rate": 4.482338611449452e-05, "loss": 6.5755, "step": 13747 }, { "epoch": 0.579631932879398, "grad_norm": 3.3449199199676514, "learning_rate": 4.4822605328086456e-05, "loss": 7.83, "step": 13748 }, { "epoch": 0.5796740940615975, "grad_norm": 1.337026834487915, "learning_rate": 4.482182454167838e-05, "loss": 6.7286, "step": 13749 }, { "epoch": 0.579716255243797, "grad_norm": 1.2440075874328613, "learning_rate": 4.482104375527031e-05, "loss": 6.4435, "step": 13750 }, { "epoch": 0.5797584164259966, "grad_norm": 2.8845770359039307, "learning_rate": 4.482026296886224e-05, "loss": 7.5887, "step": 13751 }, { "epoch": 0.5798005776081961, "grad_norm": 2.9998586177825928, "learning_rate": 4.481948218245417e-05, "loss": 6.3732, "step": 13752 }, { "epoch": 0.5798427387903957, "grad_norm": 1.2052898406982422, "learning_rate": 4.48187013960461e-05, "loss": 6.6784, "step": 13753 }, { "epoch": 0.5798848999725952, "grad_norm": 2.7515504360198975, "learning_rate": 4.481792060963803e-05, "loss": 7.8262, "step": 13754 }, { "epoch": 0.5799270611547948, "grad_norm": 3.084890842437744, "learning_rate": 4.481713982322996e-05, "loss": 7.8206, "step": 13755 }, { "epoch": 0.5799692223369943, "grad_norm": 1.9840461015701294, "learning_rate": 4.4816359036821885e-05, "loss": 6.5355, "step": 13756 }, { "epoch": 0.5800113835191939, "grad_norm": 3.016044855117798, "learning_rate": 4.481557825041382e-05, "loss": 6.3512, "step": 13757 }, { "epoch": 0.5800535447013935, "grad_norm": 2.6722893714904785, "learning_rate": 4.481479746400575e-05, "loss": 6.3644, "step": 13758 }, { "epoch": 0.5800957058835929, "grad_norm": 1.25296151638031, "learning_rate": 4.481401667759768e-05, "loss": 6.8818, "step": 13759 }, { "epoch": 0.5801378670657925, "grad_norm": 1.4254112243652344, "learning_rate": 4.481323589118961e-05, "loss": 6.4314, "step": 13760 }, { "epoch": 0.5801800282479921, "grad_norm": 1.5980372428894043, "learning_rate": 4.4812455104781535e-05, "loss": 6.4372, "step": 13761 }, { "epoch": 0.5802221894301917, "grad_norm": 1.6377357244491577, "learning_rate": 4.481167431837347e-05, "loss": 6.9513, "step": 13762 }, { "epoch": 0.5802643506123911, "grad_norm": 1.7305947542190552, "learning_rate": 4.48108935319654e-05, "loss": 6.211, "step": 13763 }, { "epoch": 0.5803065117945907, "grad_norm": 1.5921095609664917, "learning_rate": 4.481011274555733e-05, "loss": 6.2943, "step": 13764 }, { "epoch": 0.5803486729767903, "grad_norm": 1.280982255935669, "learning_rate": 4.480933195914926e-05, "loss": 6.2802, "step": 13765 }, { "epoch": 0.5803908341589898, "grad_norm": 2.4496707916259766, "learning_rate": 4.4808551172741185e-05, "loss": 8.2862, "step": 13766 }, { "epoch": 0.5804329953411894, "grad_norm": 2.6996145248413086, "learning_rate": 4.4807770386333116e-05, "loss": 7.3458, "step": 13767 }, { "epoch": 0.5804751565233889, "grad_norm": 1.1100242137908936, "learning_rate": 4.480698959992505e-05, "loss": 6.3278, "step": 13768 }, { "epoch": 0.5805173177055885, "grad_norm": 1.0915230512619019, "learning_rate": 4.480620881351698e-05, "loss": 6.3348, "step": 13769 }, { "epoch": 0.580559478887788, "grad_norm": 2.291794776916504, "learning_rate": 4.480542802710891e-05, "loss": 7.6499, "step": 13770 }, { "epoch": 0.5806016400699876, "grad_norm": 3.439110040664673, "learning_rate": 4.4804647240700834e-05, "loss": 8.1087, "step": 13771 }, { "epoch": 0.5806438012521871, "grad_norm": 1.336005449295044, "learning_rate": 4.4803866454292765e-05, "loss": 6.5779, "step": 13772 }, { "epoch": 0.5806859624343866, "grad_norm": 2.4807872772216797, "learning_rate": 4.4803085667884697e-05, "loss": 7.709, "step": 13773 }, { "epoch": 0.5807281236165862, "grad_norm": 1.3410345315933228, "learning_rate": 4.480230488147663e-05, "loss": 6.9532, "step": 13774 }, { "epoch": 0.5807702847987858, "grad_norm": 1.569804072380066, "learning_rate": 4.480152409506855e-05, "loss": 7.1321, "step": 13775 }, { "epoch": 0.5808124459809854, "grad_norm": 2.111537456512451, "learning_rate": 4.4800743308660484e-05, "loss": 6.3521, "step": 13776 }, { "epoch": 0.5808546071631848, "grad_norm": 1.5345333814620972, "learning_rate": 4.4799962522252415e-05, "loss": 7.1587, "step": 13777 }, { "epoch": 0.5808967683453844, "grad_norm": 1.1341450214385986, "learning_rate": 4.479918173584434e-05, "loss": 6.8912, "step": 13778 }, { "epoch": 0.580938929527584, "grad_norm": 3.3452858924865723, "learning_rate": 4.479840094943628e-05, "loss": 8.1094, "step": 13779 }, { "epoch": 0.5809810907097835, "grad_norm": 1.5377049446105957, "learning_rate": 4.47976201630282e-05, "loss": 6.7862, "step": 13780 }, { "epoch": 0.581023251891983, "grad_norm": 2.94929838180542, "learning_rate": 4.479683937662014e-05, "loss": 6.2405, "step": 13781 }, { "epoch": 0.5810654130741826, "grad_norm": 1.7785649299621582, "learning_rate": 4.4796058590212064e-05, "loss": 7.6232, "step": 13782 }, { "epoch": 0.5811075742563822, "grad_norm": 2.6606178283691406, "learning_rate": 4.479527780380399e-05, "loss": 7.5189, "step": 13783 }, { "epoch": 0.5811497354385817, "grad_norm": 1.50804603099823, "learning_rate": 4.479449701739593e-05, "loss": 7.2297, "step": 13784 }, { "epoch": 0.5811918966207813, "grad_norm": 1.5391144752502441, "learning_rate": 4.479371623098785e-05, "loss": 7.02, "step": 13785 }, { "epoch": 0.5812340578029808, "grad_norm": 1.345658540725708, "learning_rate": 4.479293544457978e-05, "loss": 6.9623, "step": 13786 }, { "epoch": 0.5812762189851803, "grad_norm": 2.182152032852173, "learning_rate": 4.4792154658171714e-05, "loss": 7.1539, "step": 13787 }, { "epoch": 0.5813183801673799, "grad_norm": 2.4866139888763428, "learning_rate": 4.4791373871763645e-05, "loss": 6.6838, "step": 13788 }, { "epoch": 0.5813605413495795, "grad_norm": 1.301599144935608, "learning_rate": 4.479059308535557e-05, "loss": 7.1035, "step": 13789 }, { "epoch": 0.581402702531779, "grad_norm": 1.563852071762085, "learning_rate": 4.47898122989475e-05, "loss": 6.453, "step": 13790 }, { "epoch": 0.5814448637139785, "grad_norm": 2.471482038497925, "learning_rate": 4.478903151253943e-05, "loss": 7.5639, "step": 13791 }, { "epoch": 0.5814870248961781, "grad_norm": 1.0074132680892944, "learning_rate": 4.4788250726131363e-05, "loss": 6.9316, "step": 13792 }, { "epoch": 0.5815291860783777, "grad_norm": 1.379726529121399, "learning_rate": 4.4787469939723295e-05, "loss": 7.0181, "step": 13793 }, { "epoch": 0.5815713472605771, "grad_norm": 1.965004324913025, "learning_rate": 4.478668915331522e-05, "loss": 7.0886, "step": 13794 }, { "epoch": 0.5816135084427767, "grad_norm": 1.235705852508545, "learning_rate": 4.478590836690715e-05, "loss": 6.5206, "step": 13795 }, { "epoch": 0.5816556696249763, "grad_norm": 1.2212022542953491, "learning_rate": 4.478512758049908e-05, "loss": 6.5473, "step": 13796 }, { "epoch": 0.5816978308071759, "grad_norm": 1.32829749584198, "learning_rate": 4.4784346794091006e-05, "loss": 6.4785, "step": 13797 }, { "epoch": 0.5817399919893754, "grad_norm": 2.223757743835449, "learning_rate": 4.4783566007682944e-05, "loss": 6.6839, "step": 13798 }, { "epoch": 0.5817821531715749, "grad_norm": 2.6811575889587402, "learning_rate": 4.478278522127487e-05, "loss": 7.2207, "step": 13799 }, { "epoch": 0.5818243143537745, "grad_norm": 2.579051971435547, "learning_rate": 4.47820044348668e-05, "loss": 6.3296, "step": 13800 }, { "epoch": 0.581866475535974, "grad_norm": 1.1996521949768066, "learning_rate": 4.478122364845873e-05, "loss": 6.422, "step": 13801 }, { "epoch": 0.5819086367181736, "grad_norm": 1.2327845096588135, "learning_rate": 4.4780442862050656e-05, "loss": 6.4682, "step": 13802 }, { "epoch": 0.5819507979003731, "grad_norm": 3.0913848876953125, "learning_rate": 4.477966207564259e-05, "loss": 7.9346, "step": 13803 }, { "epoch": 0.5819929590825726, "grad_norm": 1.2233703136444092, "learning_rate": 4.477888128923452e-05, "loss": 6.4897, "step": 13804 }, { "epoch": 0.5820351202647722, "grad_norm": 1.7196835279464722, "learning_rate": 4.477810050282645e-05, "loss": 6.6314, "step": 13805 }, { "epoch": 0.5820772814469718, "grad_norm": 1.5239685773849487, "learning_rate": 4.477731971641838e-05, "loss": 7.0207, "step": 13806 }, { "epoch": 0.5821194426291714, "grad_norm": 1.6509674787521362, "learning_rate": 4.477653893001031e-05, "loss": 7.1959, "step": 13807 }, { "epoch": 0.5821616038113708, "grad_norm": 1.8186509609222412, "learning_rate": 4.4775758143602236e-05, "loss": 6.6512, "step": 13808 }, { "epoch": 0.5822037649935704, "grad_norm": 1.734758973121643, "learning_rate": 4.477497735719417e-05, "loss": 6.2813, "step": 13809 }, { "epoch": 0.58224592617577, "grad_norm": 1.6205146312713623, "learning_rate": 4.47741965707861e-05, "loss": 6.8907, "step": 13810 }, { "epoch": 0.5822880873579696, "grad_norm": 2.3308839797973633, "learning_rate": 4.4773415784378023e-05, "loss": 7.1139, "step": 13811 }, { "epoch": 0.582330248540169, "grad_norm": 1.552493929862976, "learning_rate": 4.477263499796996e-05, "loss": 6.718, "step": 13812 }, { "epoch": 0.5823724097223686, "grad_norm": 2.326491117477417, "learning_rate": 4.4771854211561886e-05, "loss": 7.8452, "step": 13813 }, { "epoch": 0.5824145709045682, "grad_norm": 1.2481226921081543, "learning_rate": 4.477107342515382e-05, "loss": 6.4714, "step": 13814 }, { "epoch": 0.5824567320867677, "grad_norm": 1.3886418342590332, "learning_rate": 4.477029263874575e-05, "loss": 6.8374, "step": 13815 }, { "epoch": 0.5824988932689673, "grad_norm": 2.1439361572265625, "learning_rate": 4.476951185233767e-05, "loss": 6.4052, "step": 13816 }, { "epoch": 0.5825410544511668, "grad_norm": 3.1645431518554688, "learning_rate": 4.476873106592961e-05, "loss": 7.1212, "step": 13817 }, { "epoch": 0.5825832156333663, "grad_norm": 1.6885377168655396, "learning_rate": 4.4767950279521536e-05, "loss": 7.4532, "step": 13818 }, { "epoch": 0.5826253768155659, "grad_norm": 1.6454689502716064, "learning_rate": 4.476716949311347e-05, "loss": 6.3364, "step": 13819 }, { "epoch": 0.5826675379977655, "grad_norm": 2.8393514156341553, "learning_rate": 4.47663887067054e-05, "loss": 8.1425, "step": 13820 }, { "epoch": 0.582709699179965, "grad_norm": 2.3754734992980957, "learning_rate": 4.476560792029732e-05, "loss": 6.6078, "step": 13821 }, { "epoch": 0.5827518603621645, "grad_norm": 2.246664047241211, "learning_rate": 4.4764827133889254e-05, "loss": 6.5202, "step": 13822 }, { "epoch": 0.5827940215443641, "grad_norm": 1.4886882305145264, "learning_rate": 4.4764046347481185e-05, "loss": 7.2476, "step": 13823 }, { "epoch": 0.5828361827265637, "grad_norm": 1.4148544073104858, "learning_rate": 4.4763265561073116e-05, "loss": 6.8231, "step": 13824 }, { "epoch": 0.5828783439087633, "grad_norm": 2.461620807647705, "learning_rate": 4.476248477466504e-05, "loss": 7.4668, "step": 13825 }, { "epoch": 0.5829205050909627, "grad_norm": 2.2650609016418457, "learning_rate": 4.476170398825698e-05, "loss": 7.5466, "step": 13826 }, { "epoch": 0.5829626662731623, "grad_norm": 0.8744463324546814, "learning_rate": 4.47609232018489e-05, "loss": 6.689, "step": 13827 }, { "epoch": 0.5830048274553619, "grad_norm": 1.2489519119262695, "learning_rate": 4.4760142415440835e-05, "loss": 7.0913, "step": 13828 }, { "epoch": 0.5830469886375614, "grad_norm": 1.61585533618927, "learning_rate": 4.4759361629032766e-05, "loss": 6.2992, "step": 13829 }, { "epoch": 0.5830891498197609, "grad_norm": 2.786807060241699, "learning_rate": 4.475858084262469e-05, "loss": 7.8428, "step": 13830 }, { "epoch": 0.5831313110019605, "grad_norm": 1.1056160926818848, "learning_rate": 4.475780005621663e-05, "loss": 7.1507, "step": 13831 }, { "epoch": 0.58317347218416, "grad_norm": 1.7905759811401367, "learning_rate": 4.475701926980855e-05, "loss": 6.6436, "step": 13832 }, { "epoch": 0.5832156333663596, "grad_norm": 1.776021122932434, "learning_rate": 4.4756238483400484e-05, "loss": 6.3653, "step": 13833 }, { "epoch": 0.5832577945485592, "grad_norm": 1.6947381496429443, "learning_rate": 4.4755457696992415e-05, "loss": 6.7044, "step": 13834 }, { "epoch": 0.5832999557307587, "grad_norm": 1.3582782745361328, "learning_rate": 4.475467691058434e-05, "loss": 6.609, "step": 13835 }, { "epoch": 0.5833421169129582, "grad_norm": 2.7540946006774902, "learning_rate": 4.475389612417627e-05, "loss": 7.2019, "step": 13836 }, { "epoch": 0.5833842780951578, "grad_norm": 1.4158155918121338, "learning_rate": 4.47531153377682e-05, "loss": 6.1375, "step": 13837 }, { "epoch": 0.5834264392773574, "grad_norm": 1.9007028341293335, "learning_rate": 4.4752334551360134e-05, "loss": 6.7694, "step": 13838 }, { "epoch": 0.5834686004595568, "grad_norm": 4.021967887878418, "learning_rate": 4.4751553764952065e-05, "loss": 7.2842, "step": 13839 }, { "epoch": 0.5835107616417564, "grad_norm": 2.836864948272705, "learning_rate": 4.475077297854399e-05, "loss": 7.8327, "step": 13840 }, { "epoch": 0.583552922823956, "grad_norm": 3.3126728534698486, "learning_rate": 4.474999219213592e-05, "loss": 7.631, "step": 13841 }, { "epoch": 0.5835950840061556, "grad_norm": 1.9339197874069214, "learning_rate": 4.474921140572785e-05, "loss": 6.8547, "step": 13842 }, { "epoch": 0.5836372451883551, "grad_norm": 2.5424036979675293, "learning_rate": 4.474843061931978e-05, "loss": 7.0109, "step": 13843 }, { "epoch": 0.5836794063705546, "grad_norm": 1.4905571937561035, "learning_rate": 4.474764983291171e-05, "loss": 7.1296, "step": 13844 }, { "epoch": 0.5837215675527542, "grad_norm": 2.196164608001709, "learning_rate": 4.474686904650364e-05, "loss": 6.7194, "step": 13845 }, { "epoch": 0.5837637287349537, "grad_norm": 1.0264005661010742, "learning_rate": 4.474608826009557e-05, "loss": 6.8722, "step": 13846 }, { "epoch": 0.5838058899171533, "grad_norm": 1.441480040550232, "learning_rate": 4.4745307473687495e-05, "loss": 6.9753, "step": 13847 }, { "epoch": 0.5838480510993528, "grad_norm": 1.2330257892608643, "learning_rate": 4.474452668727943e-05, "loss": 6.2925, "step": 13848 }, { "epoch": 0.5838902122815524, "grad_norm": 1.196685552597046, "learning_rate": 4.474374590087136e-05, "loss": 6.4837, "step": 13849 }, { "epoch": 0.5839323734637519, "grad_norm": 1.8321110010147095, "learning_rate": 4.4742965114463295e-05, "loss": 7.152, "step": 13850 }, { "epoch": 0.5839745346459515, "grad_norm": 1.1036313772201538, "learning_rate": 4.474218432805522e-05, "loss": 6.2786, "step": 13851 }, { "epoch": 0.584016695828151, "grad_norm": 1.168187141418457, "learning_rate": 4.474140354164715e-05, "loss": 7.5214, "step": 13852 }, { "epoch": 0.5840588570103505, "grad_norm": 2.572704553604126, "learning_rate": 4.474062275523908e-05, "loss": 7.3011, "step": 13853 }, { "epoch": 0.5841010181925501, "grad_norm": 2.2424700260162354, "learning_rate": 4.473984196883101e-05, "loss": 6.318, "step": 13854 }, { "epoch": 0.5841431793747497, "grad_norm": 1.4859968423843384, "learning_rate": 4.473906118242294e-05, "loss": 6.571, "step": 13855 }, { "epoch": 0.5841853405569493, "grad_norm": 1.4715449810028076, "learning_rate": 4.473828039601487e-05, "loss": 6.5207, "step": 13856 }, { "epoch": 0.5842275017391487, "grad_norm": 1.700851559638977, "learning_rate": 4.47374996096068e-05, "loss": 6.9529, "step": 13857 }, { "epoch": 0.5842696629213483, "grad_norm": 1.3151240348815918, "learning_rate": 4.4736718823198725e-05, "loss": 6.3998, "step": 13858 }, { "epoch": 0.5843118241035479, "grad_norm": 1.7975436449050903, "learning_rate": 4.4735938036790656e-05, "loss": 6.4093, "step": 13859 }, { "epoch": 0.5843539852857474, "grad_norm": 1.5125577449798584, "learning_rate": 4.473515725038259e-05, "loss": 6.5315, "step": 13860 }, { "epoch": 0.5843961464679469, "grad_norm": 2.8475708961486816, "learning_rate": 4.473437646397452e-05, "loss": 7.4042, "step": 13861 }, { "epoch": 0.5844383076501465, "grad_norm": 1.8928327560424805, "learning_rate": 4.473359567756645e-05, "loss": 7.1384, "step": 13862 }, { "epoch": 0.5844804688323461, "grad_norm": 1.5961867570877075, "learning_rate": 4.4732814891158374e-05, "loss": 7.1095, "step": 13863 }, { "epoch": 0.5845226300145456, "grad_norm": 2.5102741718292236, "learning_rate": 4.4732034104750306e-05, "loss": 6.8617, "step": 13864 }, { "epoch": 0.5845647911967452, "grad_norm": 2.749145269393921, "learning_rate": 4.473125331834224e-05, "loss": 7.722, "step": 13865 }, { "epoch": 0.5846069523789447, "grad_norm": 1.0922654867172241, "learning_rate": 4.473047253193416e-05, "loss": 6.6411, "step": 13866 }, { "epoch": 0.5846491135611442, "grad_norm": 0.8985841274261475, "learning_rate": 4.47296917455261e-05, "loss": 6.4466, "step": 13867 }, { "epoch": 0.5846912747433438, "grad_norm": 1.2337812185287476, "learning_rate": 4.4728910959118024e-05, "loss": 6.415, "step": 13868 }, { "epoch": 0.5847334359255434, "grad_norm": 1.1209633350372314, "learning_rate": 4.4728130172709955e-05, "loss": 6.5633, "step": 13869 }, { "epoch": 0.5847755971077429, "grad_norm": 1.1355143785476685, "learning_rate": 4.4727349386301886e-05, "loss": 6.6813, "step": 13870 }, { "epoch": 0.5848177582899424, "grad_norm": 1.0630362033843994, "learning_rate": 4.472656859989382e-05, "loss": 6.7912, "step": 13871 }, { "epoch": 0.584859919472142, "grad_norm": 1.2998902797698975, "learning_rate": 4.472578781348575e-05, "loss": 7.0086, "step": 13872 }, { "epoch": 0.5849020806543416, "grad_norm": 2.1209726333618164, "learning_rate": 4.4725007027077673e-05, "loss": 8.2274, "step": 13873 }, { "epoch": 0.5849442418365411, "grad_norm": 1.3526870012283325, "learning_rate": 4.4724226240669605e-05, "loss": 7.0743, "step": 13874 }, { "epoch": 0.5849864030187406, "grad_norm": 1.6676385402679443, "learning_rate": 4.4723445454261536e-05, "loss": 6.6307, "step": 13875 }, { "epoch": 0.5850285642009402, "grad_norm": 1.3412435054779053, "learning_rate": 4.472266466785347e-05, "loss": 7.1882, "step": 13876 }, { "epoch": 0.5850707253831398, "grad_norm": 2.6916890144348145, "learning_rate": 4.472188388144539e-05, "loss": 7.7857, "step": 13877 }, { "epoch": 0.5851128865653393, "grad_norm": 1.735807180404663, "learning_rate": 4.472110309503732e-05, "loss": 6.2658, "step": 13878 }, { "epoch": 0.5851550477475388, "grad_norm": 1.2753031253814697, "learning_rate": 4.4720322308629254e-05, "loss": 6.5019, "step": 13879 }, { "epoch": 0.5851972089297384, "grad_norm": 1.4097273349761963, "learning_rate": 4.471954152222118e-05, "loss": 6.441, "step": 13880 }, { "epoch": 0.5852393701119379, "grad_norm": 3.7559194564819336, "learning_rate": 4.471876073581312e-05, "loss": 8.3089, "step": 13881 }, { "epoch": 0.5852815312941375, "grad_norm": 1.136826753616333, "learning_rate": 4.471797994940504e-05, "loss": 6.3057, "step": 13882 }, { "epoch": 0.5853236924763371, "grad_norm": 2.152168035507202, "learning_rate": 4.471719916299697e-05, "loss": 6.8331, "step": 13883 }, { "epoch": 0.5853658536585366, "grad_norm": 2.2242255210876465, "learning_rate": 4.4716418376588904e-05, "loss": 7.6994, "step": 13884 }, { "epoch": 0.5854080148407361, "grad_norm": 3.5126075744628906, "learning_rate": 4.471563759018083e-05, "loss": 7.439, "step": 13885 }, { "epoch": 0.5854501760229357, "grad_norm": 1.0635448694229126, "learning_rate": 4.4714856803772766e-05, "loss": 6.5541, "step": 13886 }, { "epoch": 0.5854923372051353, "grad_norm": 2.2226626873016357, "learning_rate": 4.471407601736469e-05, "loss": 7.2176, "step": 13887 }, { "epoch": 0.5855344983873347, "grad_norm": 0.9845512509346008, "learning_rate": 4.471329523095662e-05, "loss": 6.7129, "step": 13888 }, { "epoch": 0.5855766595695343, "grad_norm": 1.330505132675171, "learning_rate": 4.471251444454855e-05, "loss": 6.4872, "step": 13889 }, { "epoch": 0.5856188207517339, "grad_norm": 0.8917661309242249, "learning_rate": 4.471173365814048e-05, "loss": 6.3633, "step": 13890 }, { "epoch": 0.5856609819339335, "grad_norm": 1.3860633373260498, "learning_rate": 4.471095287173241e-05, "loss": 6.211, "step": 13891 }, { "epoch": 0.585703143116133, "grad_norm": 1.165671467781067, "learning_rate": 4.471017208532434e-05, "loss": 6.5278, "step": 13892 }, { "epoch": 0.5857453042983325, "grad_norm": 3.4585721492767334, "learning_rate": 4.470939129891627e-05, "loss": 6.952, "step": 13893 }, { "epoch": 0.5857874654805321, "grad_norm": 3.86576771736145, "learning_rate": 4.4708610512508196e-05, "loss": 7.8416, "step": 13894 }, { "epoch": 0.5858296266627316, "grad_norm": 4.36515998840332, "learning_rate": 4.4707829726100134e-05, "loss": 7.8068, "step": 13895 }, { "epoch": 0.5858717878449312, "grad_norm": 1.243367075920105, "learning_rate": 4.470704893969206e-05, "loss": 6.669, "step": 13896 }, { "epoch": 0.5859139490271307, "grad_norm": 1.1914691925048828, "learning_rate": 4.470626815328399e-05, "loss": 6.3885, "step": 13897 }, { "epoch": 0.5859561102093302, "grad_norm": 1.1438790559768677, "learning_rate": 4.470548736687592e-05, "loss": 6.7605, "step": 13898 }, { "epoch": 0.5859982713915298, "grad_norm": 1.508188247680664, "learning_rate": 4.4704706580467846e-05, "loss": 6.6886, "step": 13899 }, { "epoch": 0.5860404325737294, "grad_norm": 3.3179771900177, "learning_rate": 4.4703925794059784e-05, "loss": 7.1354, "step": 13900 }, { "epoch": 0.586082593755929, "grad_norm": 2.3360350131988525, "learning_rate": 4.470314500765171e-05, "loss": 7.7533, "step": 13901 }, { "epoch": 0.5861247549381284, "grad_norm": 1.7664059400558472, "learning_rate": 4.470236422124364e-05, "loss": 7.4524, "step": 13902 }, { "epoch": 0.586166916120328, "grad_norm": 1.0743566751480103, "learning_rate": 4.470158343483557e-05, "loss": 6.4321, "step": 13903 }, { "epoch": 0.5862090773025276, "grad_norm": 2.347470283508301, "learning_rate": 4.4700802648427495e-05, "loss": 6.8791, "step": 13904 }, { "epoch": 0.5862512384847272, "grad_norm": 0.8737768530845642, "learning_rate": 4.4700021862019426e-05, "loss": 6.3714, "step": 13905 }, { "epoch": 0.5862933996669266, "grad_norm": 2.7050464153289795, "learning_rate": 4.469924107561136e-05, "loss": 7.1318, "step": 13906 }, { "epoch": 0.5863355608491262, "grad_norm": 1.2046005725860596, "learning_rate": 4.469846028920329e-05, "loss": 6.6521, "step": 13907 }, { "epoch": 0.5863777220313258, "grad_norm": 1.3456376791000366, "learning_rate": 4.469767950279522e-05, "loss": 7.0926, "step": 13908 }, { "epoch": 0.5864198832135253, "grad_norm": 2.4805808067321777, "learning_rate": 4.4696898716387145e-05, "loss": 7.6611, "step": 13909 }, { "epoch": 0.5864620443957248, "grad_norm": 1.5049721002578735, "learning_rate": 4.4696117929979076e-05, "loss": 7.0224, "step": 13910 }, { "epoch": 0.5865042055779244, "grad_norm": 1.94036865234375, "learning_rate": 4.469533714357101e-05, "loss": 7.0362, "step": 13911 }, { "epoch": 0.586546366760124, "grad_norm": 1.375795841217041, "learning_rate": 4.469455635716294e-05, "loss": 7.0859, "step": 13912 }, { "epoch": 0.5865885279423235, "grad_norm": 1.309747576713562, "learning_rate": 4.469377557075486e-05, "loss": 6.4485, "step": 13913 }, { "epoch": 0.5866306891245231, "grad_norm": 2.6134462356567383, "learning_rate": 4.46929947843468e-05, "loss": 7.3252, "step": 13914 }, { "epoch": 0.5866728503067226, "grad_norm": 1.9536746740341187, "learning_rate": 4.4692213997938725e-05, "loss": 6.8302, "step": 13915 }, { "epoch": 0.5867150114889221, "grad_norm": 1.8437989950180054, "learning_rate": 4.469143321153066e-05, "loss": 6.9279, "step": 13916 }, { "epoch": 0.5867571726711217, "grad_norm": 2.1123223304748535, "learning_rate": 4.469065242512259e-05, "loss": 6.4509, "step": 13917 }, { "epoch": 0.5867993338533213, "grad_norm": 1.2033106088638306, "learning_rate": 4.468987163871451e-05, "loss": 6.315, "step": 13918 }, { "epoch": 0.5868414950355207, "grad_norm": 0.9335532784461975, "learning_rate": 4.468909085230645e-05, "loss": 6.387, "step": 13919 }, { "epoch": 0.5868836562177203, "grad_norm": 2.0246894359588623, "learning_rate": 4.4688310065898375e-05, "loss": 6.8696, "step": 13920 }, { "epoch": 0.5869258173999199, "grad_norm": 1.162024736404419, "learning_rate": 4.4687529279490306e-05, "loss": 6.7492, "step": 13921 }, { "epoch": 0.5869679785821195, "grad_norm": 0.8984385132789612, "learning_rate": 4.468674849308224e-05, "loss": 6.6671, "step": 13922 }, { "epoch": 0.587010139764319, "grad_norm": 3.3537440299987793, "learning_rate": 4.468596770667416e-05, "loss": 7.4538, "step": 13923 }, { "epoch": 0.5870523009465185, "grad_norm": 2.287943124771118, "learning_rate": 4.468518692026609e-05, "loss": 7.1133, "step": 13924 }, { "epoch": 0.5870944621287181, "grad_norm": 2.6121084690093994, "learning_rate": 4.4684406133858024e-05, "loss": 7.6245, "step": 13925 }, { "epoch": 0.5871366233109176, "grad_norm": 1.8819496631622314, "learning_rate": 4.4683625347449956e-05, "loss": 6.4435, "step": 13926 }, { "epoch": 0.5871787844931172, "grad_norm": 1.8303841352462769, "learning_rate": 4.468284456104188e-05, "loss": 6.4905, "step": 13927 }, { "epoch": 0.5872209456753167, "grad_norm": 1.4588334560394287, "learning_rate": 4.468206377463381e-05, "loss": 6.3723, "step": 13928 }, { "epoch": 0.5872631068575163, "grad_norm": 2.4239261150360107, "learning_rate": 4.468128298822574e-05, "loss": 6.3649, "step": 13929 }, { "epoch": 0.5873052680397158, "grad_norm": 1.1566020250320435, "learning_rate": 4.4680502201817674e-05, "loss": 6.3247, "step": 13930 }, { "epoch": 0.5873474292219154, "grad_norm": 1.0876039266586304, "learning_rate": 4.4679721415409605e-05, "loss": 6.7065, "step": 13931 }, { "epoch": 0.587389590404115, "grad_norm": 1.7010674476623535, "learning_rate": 4.467894062900153e-05, "loss": 7.0576, "step": 13932 }, { "epoch": 0.5874317515863144, "grad_norm": 1.9061890840530396, "learning_rate": 4.467815984259347e-05, "loss": 7.6435, "step": 13933 }, { "epoch": 0.587473912768514, "grad_norm": 1.4089499711990356, "learning_rate": 4.467737905618539e-05, "loss": 7.0143, "step": 13934 }, { "epoch": 0.5875160739507136, "grad_norm": 1.7226535081863403, "learning_rate": 4.467659826977732e-05, "loss": 6.7309, "step": 13935 }, { "epoch": 0.5875582351329132, "grad_norm": 1.498606562614441, "learning_rate": 4.4675817483369255e-05, "loss": 6.687, "step": 13936 }, { "epoch": 0.5876003963151126, "grad_norm": 2.6658565998077393, "learning_rate": 4.467503669696118e-05, "loss": 7.7607, "step": 13937 }, { "epoch": 0.5876425574973122, "grad_norm": 1.0643432140350342, "learning_rate": 4.467425591055311e-05, "loss": 6.2901, "step": 13938 }, { "epoch": 0.5876847186795118, "grad_norm": 2.318241596221924, "learning_rate": 4.467347512414504e-05, "loss": 7.6256, "step": 13939 }, { "epoch": 0.5877268798617113, "grad_norm": 2.2395310401916504, "learning_rate": 4.467269433773697e-05, "loss": 6.9196, "step": 13940 }, { "epoch": 0.5877690410439109, "grad_norm": 2.5805816650390625, "learning_rate": 4.4671913551328904e-05, "loss": 7.2698, "step": 13941 }, { "epoch": 0.5878112022261104, "grad_norm": 1.072243332862854, "learning_rate": 4.467113276492083e-05, "loss": 6.7102, "step": 13942 }, { "epoch": 0.58785336340831, "grad_norm": 0.9464010000228882, "learning_rate": 4.467035197851276e-05, "loss": 6.8084, "step": 13943 }, { "epoch": 0.5878955245905095, "grad_norm": 2.898822546005249, "learning_rate": 4.466957119210469e-05, "loss": 7.9261, "step": 13944 }, { "epoch": 0.5879376857727091, "grad_norm": 2.5643086433410645, "learning_rate": 4.466879040569662e-05, "loss": 6.3378, "step": 13945 }, { "epoch": 0.5879798469549086, "grad_norm": 1.5006991624832153, "learning_rate": 4.466800961928855e-05, "loss": 7.2575, "step": 13946 }, { "epoch": 0.5880220081371081, "grad_norm": 1.7708251476287842, "learning_rate": 4.466722883288048e-05, "loss": 6.3441, "step": 13947 }, { "epoch": 0.5880641693193077, "grad_norm": 1.9335237741470337, "learning_rate": 4.466644804647241e-05, "loss": 7.3085, "step": 13948 }, { "epoch": 0.5881063305015073, "grad_norm": 1.1024062633514404, "learning_rate": 4.4665667260064334e-05, "loss": 6.3806, "step": 13949 }, { "epoch": 0.5881484916837069, "grad_norm": 2.82039737701416, "learning_rate": 4.466488647365627e-05, "loss": 7.6384, "step": 13950 }, { "epoch": 0.5881906528659063, "grad_norm": 2.9226696491241455, "learning_rate": 4.4664105687248197e-05, "loss": 7.7962, "step": 13951 }, { "epoch": 0.5882328140481059, "grad_norm": 1.6985563039779663, "learning_rate": 4.466332490084013e-05, "loss": 6.5011, "step": 13952 }, { "epoch": 0.5882749752303055, "grad_norm": 1.6037960052490234, "learning_rate": 4.466254411443206e-05, "loss": 6.7204, "step": 13953 }, { "epoch": 0.588317136412505, "grad_norm": 1.024959921836853, "learning_rate": 4.4661763328023984e-05, "loss": 6.5735, "step": 13954 }, { "epoch": 0.5883592975947045, "grad_norm": 1.3905223608016968, "learning_rate": 4.466098254161592e-05, "loss": 6.3242, "step": 13955 }, { "epoch": 0.5884014587769041, "grad_norm": 2.3808035850524902, "learning_rate": 4.4660201755207846e-05, "loss": 7.9472, "step": 13956 }, { "epoch": 0.5884436199591037, "grad_norm": 1.4228875637054443, "learning_rate": 4.465942096879978e-05, "loss": 6.4092, "step": 13957 }, { "epoch": 0.5884857811413032, "grad_norm": 0.9460673928260803, "learning_rate": 4.465864018239171e-05, "loss": 6.7865, "step": 13958 }, { "epoch": 0.5885279423235028, "grad_norm": 3.2058398723602295, "learning_rate": 4.465785939598364e-05, "loss": 7.4262, "step": 13959 }, { "epoch": 0.5885701035057023, "grad_norm": 0.8732950687408447, "learning_rate": 4.4657078609575564e-05, "loss": 6.5858, "step": 13960 }, { "epoch": 0.5886122646879018, "grad_norm": 2.127702474594116, "learning_rate": 4.4656297823167496e-05, "loss": 7.3421, "step": 13961 }, { "epoch": 0.5886544258701014, "grad_norm": 0.9624263644218445, "learning_rate": 4.465551703675943e-05, "loss": 6.4399, "step": 13962 }, { "epoch": 0.588696587052301, "grad_norm": 0.8069973587989807, "learning_rate": 4.465473625035135e-05, "loss": 6.577, "step": 13963 }, { "epoch": 0.5887387482345005, "grad_norm": 4.129042148590088, "learning_rate": 4.465395546394329e-05, "loss": 8.4526, "step": 13964 }, { "epoch": 0.5887809094167, "grad_norm": 0.7077462673187256, "learning_rate": 4.4653174677535214e-05, "loss": 6.4782, "step": 13965 }, { "epoch": 0.5888230705988996, "grad_norm": 0.9108072519302368, "learning_rate": 4.4652393891127145e-05, "loss": 6.2951, "step": 13966 }, { "epoch": 0.5888652317810992, "grad_norm": 1.3591084480285645, "learning_rate": 4.4651613104719076e-05, "loss": 6.4481, "step": 13967 }, { "epoch": 0.5889073929632986, "grad_norm": 1.106311321258545, "learning_rate": 4.4650832318311e-05, "loss": 6.5167, "step": 13968 }, { "epoch": 0.5889495541454982, "grad_norm": 2.530719757080078, "learning_rate": 4.465005153190294e-05, "loss": 6.4728, "step": 13969 }, { "epoch": 0.5889917153276978, "grad_norm": 2.3549962043762207, "learning_rate": 4.4649270745494863e-05, "loss": 7.2934, "step": 13970 }, { "epoch": 0.5890338765098974, "grad_norm": 1.1033304929733276, "learning_rate": 4.4648489959086795e-05, "loss": 6.3908, "step": 13971 }, { "epoch": 0.5890760376920969, "grad_norm": 2.8838260173797607, "learning_rate": 4.4647709172678726e-05, "loss": 6.988, "step": 13972 }, { "epoch": 0.5891181988742964, "grad_norm": 1.8256739377975464, "learning_rate": 4.464692838627065e-05, "loss": 6.6159, "step": 13973 }, { "epoch": 0.589160360056496, "grad_norm": 1.257542610168457, "learning_rate": 4.464614759986258e-05, "loss": 6.5823, "step": 13974 }, { "epoch": 0.5892025212386955, "grad_norm": 0.9837090969085693, "learning_rate": 4.464536681345451e-05, "loss": 6.3053, "step": 13975 }, { "epoch": 0.5892446824208951, "grad_norm": 1.157023310661316, "learning_rate": 4.4644586027046444e-05, "loss": 6.9936, "step": 13976 }, { "epoch": 0.5892868436030946, "grad_norm": 4.5710883140563965, "learning_rate": 4.4643805240638375e-05, "loss": 8.6253, "step": 13977 }, { "epoch": 0.5893290047852942, "grad_norm": 1.6642855405807495, "learning_rate": 4.464302445423031e-05, "loss": 6.8052, "step": 13978 }, { "epoch": 0.5893711659674937, "grad_norm": 1.6056650876998901, "learning_rate": 4.464224366782223e-05, "loss": 6.182, "step": 13979 }, { "epoch": 0.5894133271496933, "grad_norm": 2.77848219871521, "learning_rate": 4.464146288141416e-05, "loss": 7.9814, "step": 13980 }, { "epoch": 0.5894554883318929, "grad_norm": 1.3236446380615234, "learning_rate": 4.4640682095006094e-05, "loss": 6.6283, "step": 13981 }, { "epoch": 0.5894976495140923, "grad_norm": 2.3395450115203857, "learning_rate": 4.463990130859802e-05, "loss": 7.6989, "step": 13982 }, { "epoch": 0.5895398106962919, "grad_norm": 1.9872808456420898, "learning_rate": 4.4639120522189956e-05, "loss": 6.5664, "step": 13983 }, { "epoch": 0.5895819718784915, "grad_norm": 1.7803936004638672, "learning_rate": 4.463833973578188e-05, "loss": 6.54, "step": 13984 }, { "epoch": 0.5896241330606911, "grad_norm": 3.2140486240386963, "learning_rate": 4.463755894937381e-05, "loss": 8.1748, "step": 13985 }, { "epoch": 0.5896662942428905, "grad_norm": 1.3831828832626343, "learning_rate": 4.463677816296574e-05, "loss": 7.0387, "step": 13986 }, { "epoch": 0.5897084554250901, "grad_norm": 2.7984089851379395, "learning_rate": 4.463599737655767e-05, "loss": 7.8017, "step": 13987 }, { "epoch": 0.5897506166072897, "grad_norm": 2.966161012649536, "learning_rate": 4.4635216590149606e-05, "loss": 7.9476, "step": 13988 }, { "epoch": 0.5897927777894892, "grad_norm": 1.6540682315826416, "learning_rate": 4.463443580374153e-05, "loss": 6.5601, "step": 13989 }, { "epoch": 0.5898349389716888, "grad_norm": 3.168431520462036, "learning_rate": 4.463365501733346e-05, "loss": 8.0069, "step": 13990 }, { "epoch": 0.5898771001538883, "grad_norm": 1.4053680896759033, "learning_rate": 4.463287423092539e-05, "loss": 6.9908, "step": 13991 }, { "epoch": 0.5899192613360879, "grad_norm": 2.1449742317199707, "learning_rate": 4.463209344451732e-05, "loss": 6.3688, "step": 13992 }, { "epoch": 0.5899614225182874, "grad_norm": 3.872237205505371, "learning_rate": 4.463131265810925e-05, "loss": 7.0763, "step": 13993 }, { "epoch": 0.590003583700487, "grad_norm": 1.9734090566635132, "learning_rate": 4.463053187170118e-05, "loss": 6.4548, "step": 13994 }, { "epoch": 0.5900457448826865, "grad_norm": 6.814635753631592, "learning_rate": 4.462975108529311e-05, "loss": 10.8172, "step": 13995 }, { "epoch": 0.590087906064886, "grad_norm": 2.8451473712921143, "learning_rate": 4.4628970298885035e-05, "loss": 7.238, "step": 13996 }, { "epoch": 0.5901300672470856, "grad_norm": 2.689201593399048, "learning_rate": 4.462818951247697e-05, "loss": 6.6899, "step": 13997 }, { "epoch": 0.5901722284292852, "grad_norm": 4.160082817077637, "learning_rate": 4.46274087260689e-05, "loss": 7.7237, "step": 13998 }, { "epoch": 0.5902143896114848, "grad_norm": 2.070896625518799, "learning_rate": 4.462662793966083e-05, "loss": 6.2329, "step": 13999 }, { "epoch": 0.5902565507936842, "grad_norm": 1.4106271266937256, "learning_rate": 4.462584715325276e-05, "loss": 7.0107, "step": 14000 }, { "epoch": 0.5902987119758838, "grad_norm": 1.983460545539856, "learning_rate": 4.4625066366844685e-05, "loss": 6.8161, "step": 14001 }, { "epoch": 0.5903408731580834, "grad_norm": 1.526984453201294, "learning_rate": 4.462428558043662e-05, "loss": 7.1456, "step": 14002 }, { "epoch": 0.5903830343402829, "grad_norm": 2.224158763885498, "learning_rate": 4.462350479402855e-05, "loss": 6.5801, "step": 14003 }, { "epoch": 0.5904251955224824, "grad_norm": 4.043676376342773, "learning_rate": 4.462272400762048e-05, "loss": 8.0435, "step": 14004 }, { "epoch": 0.590467356704682, "grad_norm": 1.5814915895462036, "learning_rate": 4.462194322121241e-05, "loss": 7.3155, "step": 14005 }, { "epoch": 0.5905095178868816, "grad_norm": 1.3085196018218994, "learning_rate": 4.4621162434804335e-05, "loss": 6.4927, "step": 14006 }, { "epoch": 0.5905516790690811, "grad_norm": 2.1622073650360107, "learning_rate": 4.4620381648396266e-05, "loss": 6.5085, "step": 14007 }, { "epoch": 0.5905938402512807, "grad_norm": 1.173667550086975, "learning_rate": 4.46196008619882e-05, "loss": 6.5203, "step": 14008 }, { "epoch": 0.5906360014334802, "grad_norm": 3.448031187057495, "learning_rate": 4.461882007558013e-05, "loss": 7.65, "step": 14009 }, { "epoch": 0.5906781626156797, "grad_norm": 1.4642215967178345, "learning_rate": 4.461803928917206e-05, "loss": 6.3825, "step": 14010 }, { "epoch": 0.5907203237978793, "grad_norm": 1.5379283428192139, "learning_rate": 4.4617258502763984e-05, "loss": 6.7096, "step": 14011 }, { "epoch": 0.5907624849800789, "grad_norm": 1.996936559677124, "learning_rate": 4.4616477716355915e-05, "loss": 7.7244, "step": 14012 }, { "epoch": 0.5908046461622783, "grad_norm": 1.2874388694763184, "learning_rate": 4.4615696929947847e-05, "loss": 7.013, "step": 14013 }, { "epoch": 0.5908468073444779, "grad_norm": 1.9292982816696167, "learning_rate": 4.461491614353978e-05, "loss": 6.6238, "step": 14014 }, { "epoch": 0.5908889685266775, "grad_norm": 1.2885169982910156, "learning_rate": 4.46141353571317e-05, "loss": 7.0933, "step": 14015 }, { "epoch": 0.5909311297088771, "grad_norm": 1.317014455795288, "learning_rate": 4.4613354570723634e-05, "loss": 7.1813, "step": 14016 }, { "epoch": 0.5909732908910765, "grad_norm": 1.4501713514328003, "learning_rate": 4.4612573784315565e-05, "loss": 6.5396, "step": 14017 }, { "epoch": 0.5910154520732761, "grad_norm": 1.3610481023788452, "learning_rate": 4.461179299790749e-05, "loss": 6.2874, "step": 14018 }, { "epoch": 0.5910576132554757, "grad_norm": 2.503798723220825, "learning_rate": 4.461101221149943e-05, "loss": 7.1703, "step": 14019 }, { "epoch": 0.5910997744376753, "grad_norm": 2.1938276290893555, "learning_rate": 4.461023142509135e-05, "loss": 6.7544, "step": 14020 }, { "epoch": 0.5911419356198748, "grad_norm": 2.5274393558502197, "learning_rate": 4.460945063868329e-05, "loss": 7.9257, "step": 14021 }, { "epoch": 0.5911840968020743, "grad_norm": 2.234285354614258, "learning_rate": 4.4608669852275214e-05, "loss": 6.615, "step": 14022 }, { "epoch": 0.5912262579842739, "grad_norm": 2.5412087440490723, "learning_rate": 4.460788906586714e-05, "loss": 6.792, "step": 14023 }, { "epoch": 0.5912684191664734, "grad_norm": 1.3437331914901733, "learning_rate": 4.460710827945908e-05, "loss": 6.5247, "step": 14024 }, { "epoch": 0.591310580348673, "grad_norm": 1.4595500230789185, "learning_rate": 4.4606327493051e-05, "loss": 7.0783, "step": 14025 }, { "epoch": 0.5913527415308725, "grad_norm": 3.429594039916992, "learning_rate": 4.460554670664293e-05, "loss": 8.337, "step": 14026 }, { "epoch": 0.591394902713072, "grad_norm": 1.1587306261062622, "learning_rate": 4.4604765920234864e-05, "loss": 6.4528, "step": 14027 }, { "epoch": 0.5914370638952716, "grad_norm": 2.3622817993164062, "learning_rate": 4.4603985133826795e-05, "loss": 6.6636, "step": 14028 }, { "epoch": 0.5914792250774712, "grad_norm": 1.0822571516036987, "learning_rate": 4.460320434741872e-05, "loss": 6.4681, "step": 14029 }, { "epoch": 0.5915213862596708, "grad_norm": 1.728186845779419, "learning_rate": 4.460242356101065e-05, "loss": 6.7002, "step": 14030 }, { "epoch": 0.5915635474418702, "grad_norm": 1.409063696861267, "learning_rate": 4.460164277460258e-05, "loss": 6.4053, "step": 14031 }, { "epoch": 0.5916057086240698, "grad_norm": 0.9296857714653015, "learning_rate": 4.460086198819451e-05, "loss": 6.4787, "step": 14032 }, { "epoch": 0.5916478698062694, "grad_norm": 1.8829116821289062, "learning_rate": 4.4600081201786445e-05, "loss": 6.4258, "step": 14033 }, { "epoch": 0.591690030988469, "grad_norm": 3.8640730381011963, "learning_rate": 4.459930041537837e-05, "loss": 7.8311, "step": 14034 }, { "epoch": 0.5917321921706684, "grad_norm": 1.0986411571502686, "learning_rate": 4.45985196289703e-05, "loss": 6.6394, "step": 14035 }, { "epoch": 0.591774353352868, "grad_norm": 1.199453353881836, "learning_rate": 4.459773884256223e-05, "loss": 6.2593, "step": 14036 }, { "epoch": 0.5918165145350676, "grad_norm": 4.754847526550293, "learning_rate": 4.4596958056154156e-05, "loss": 7.8288, "step": 14037 }, { "epoch": 0.5918586757172671, "grad_norm": 1.7607976198196411, "learning_rate": 4.4596177269746094e-05, "loss": 6.3487, "step": 14038 }, { "epoch": 0.5919008368994667, "grad_norm": 3.5390541553497314, "learning_rate": 4.459539648333802e-05, "loss": 7.2414, "step": 14039 }, { "epoch": 0.5919429980816662, "grad_norm": 1.078083872795105, "learning_rate": 4.459461569692995e-05, "loss": 6.4988, "step": 14040 }, { "epoch": 0.5919851592638657, "grad_norm": 0.8455734848976135, "learning_rate": 4.459383491052188e-05, "loss": 6.5203, "step": 14041 }, { "epoch": 0.5920273204460653, "grad_norm": 1.088749885559082, "learning_rate": 4.4593054124113806e-05, "loss": 6.621, "step": 14042 }, { "epoch": 0.5920694816282649, "grad_norm": 2.792543649673462, "learning_rate": 4.459227333770574e-05, "loss": 7.1301, "step": 14043 }, { "epoch": 0.5921116428104644, "grad_norm": 1.7035713195800781, "learning_rate": 4.459149255129767e-05, "loss": 6.6569, "step": 14044 }, { "epoch": 0.5921538039926639, "grad_norm": 1.2787028551101685, "learning_rate": 4.45907117648896e-05, "loss": 6.4287, "step": 14045 }, { "epoch": 0.5921959651748635, "grad_norm": 2.4040379524230957, "learning_rate": 4.458993097848153e-05, "loss": 6.4799, "step": 14046 }, { "epoch": 0.5922381263570631, "grad_norm": 1.2212024927139282, "learning_rate": 4.458915019207346e-05, "loss": 6.264, "step": 14047 }, { "epoch": 0.5922802875392627, "grad_norm": 2.8255350589752197, "learning_rate": 4.4588369405665386e-05, "loss": 8.0733, "step": 14048 }, { "epoch": 0.5923224487214621, "grad_norm": 2.563560724258423, "learning_rate": 4.458758861925732e-05, "loss": 8.0009, "step": 14049 }, { "epoch": 0.5923646099036617, "grad_norm": 1.3057845830917358, "learning_rate": 4.458680783284925e-05, "loss": 6.9062, "step": 14050 }, { "epoch": 0.5924067710858613, "grad_norm": 2.5537526607513428, "learning_rate": 4.4586027046441173e-05, "loss": 6.3278, "step": 14051 }, { "epoch": 0.5924489322680608, "grad_norm": 1.354975938796997, "learning_rate": 4.458524626003311e-05, "loss": 6.956, "step": 14052 }, { "epoch": 0.5924910934502603, "grad_norm": 2.5373106002807617, "learning_rate": 4.4584465473625036e-05, "loss": 7.3542, "step": 14053 }, { "epoch": 0.5925332546324599, "grad_norm": 3.4757461547851562, "learning_rate": 4.458368468721697e-05, "loss": 8.0425, "step": 14054 }, { "epoch": 0.5925754158146594, "grad_norm": 1.8352261781692505, "learning_rate": 4.45829039008089e-05, "loss": 6.889, "step": 14055 }, { "epoch": 0.592617576996859, "grad_norm": 2.4171409606933594, "learning_rate": 4.458212311440082e-05, "loss": 6.461, "step": 14056 }, { "epoch": 0.5926597381790586, "grad_norm": 2.610211133956909, "learning_rate": 4.458134232799276e-05, "loss": 6.8406, "step": 14057 }, { "epoch": 0.5927018993612581, "grad_norm": 1.3066672086715698, "learning_rate": 4.4580561541584686e-05, "loss": 7.0623, "step": 14058 }, { "epoch": 0.5927440605434576, "grad_norm": 3.1563618183135986, "learning_rate": 4.457978075517662e-05, "loss": 7.7665, "step": 14059 }, { "epoch": 0.5927862217256572, "grad_norm": 1.4058928489685059, "learning_rate": 4.457899996876855e-05, "loss": 6.5864, "step": 14060 }, { "epoch": 0.5928283829078568, "grad_norm": 3.513138771057129, "learning_rate": 4.457821918236047e-05, "loss": 7.6225, "step": 14061 }, { "epoch": 0.5928705440900562, "grad_norm": 1.9861575365066528, "learning_rate": 4.4577438395952404e-05, "loss": 6.5497, "step": 14062 }, { "epoch": 0.5929127052722558, "grad_norm": 2.9009547233581543, "learning_rate": 4.4576657609544335e-05, "loss": 6.8511, "step": 14063 }, { "epoch": 0.5929548664544554, "grad_norm": 0.9967166781425476, "learning_rate": 4.4575876823136266e-05, "loss": 6.9435, "step": 14064 }, { "epoch": 0.592997027636655, "grad_norm": 1.576357364654541, "learning_rate": 4.457509603672819e-05, "loss": 7.0804, "step": 14065 }, { "epoch": 0.5930391888188545, "grad_norm": 2.0886693000793457, "learning_rate": 4.457431525032013e-05, "loss": 6.5469, "step": 14066 }, { "epoch": 0.593081350001054, "grad_norm": 1.595973253250122, "learning_rate": 4.457353446391205e-05, "loss": 7.3946, "step": 14067 }, { "epoch": 0.5931235111832536, "grad_norm": 1.5794353485107422, "learning_rate": 4.4572753677503985e-05, "loss": 7.1055, "step": 14068 }, { "epoch": 0.5931656723654531, "grad_norm": 1.2036747932434082, "learning_rate": 4.4571972891095916e-05, "loss": 6.5198, "step": 14069 }, { "epoch": 0.5932078335476527, "grad_norm": 1.8329355716705322, "learning_rate": 4.457119210468784e-05, "loss": 6.9083, "step": 14070 }, { "epoch": 0.5932499947298522, "grad_norm": 1.5196070671081543, "learning_rate": 4.457041131827978e-05, "loss": 6.2957, "step": 14071 }, { "epoch": 0.5932921559120518, "grad_norm": 1.5409762859344482, "learning_rate": 4.45696305318717e-05, "loss": 6.6648, "step": 14072 }, { "epoch": 0.5933343170942513, "grad_norm": 0.9576740264892578, "learning_rate": 4.4568849745463634e-05, "loss": 6.5246, "step": 14073 }, { "epoch": 0.5933764782764509, "grad_norm": 1.9633477926254272, "learning_rate": 4.4568068959055565e-05, "loss": 7.2046, "step": 14074 }, { "epoch": 0.5934186394586504, "grad_norm": 5.043374061584473, "learning_rate": 4.456728817264749e-05, "loss": 10.4796, "step": 14075 }, { "epoch": 0.5934608006408499, "grad_norm": 3.3662900924682617, "learning_rate": 4.456650738623942e-05, "loss": 7.1435, "step": 14076 }, { "epoch": 0.5935029618230495, "grad_norm": 2.6060938835144043, "learning_rate": 4.456572659983135e-05, "loss": 7.2506, "step": 14077 }, { "epoch": 0.5935451230052491, "grad_norm": 2.4544410705566406, "learning_rate": 4.4564945813423284e-05, "loss": 7.5494, "step": 14078 }, { "epoch": 0.5935872841874487, "grad_norm": 1.8678494691848755, "learning_rate": 4.4564165027015215e-05, "loss": 6.4127, "step": 14079 }, { "epoch": 0.5936294453696481, "grad_norm": 2.8229053020477295, "learning_rate": 4.456338424060714e-05, "loss": 7.5634, "step": 14080 }, { "epoch": 0.5936716065518477, "grad_norm": 1.1552332639694214, "learning_rate": 4.456260345419907e-05, "loss": 6.9213, "step": 14081 }, { "epoch": 0.5937137677340473, "grad_norm": 1.3241032361984253, "learning_rate": 4.4561822667791e-05, "loss": 7.017, "step": 14082 }, { "epoch": 0.5937559289162468, "grad_norm": 1.8997061252593994, "learning_rate": 4.456104188138293e-05, "loss": 7.0503, "step": 14083 }, { "epoch": 0.5937980900984463, "grad_norm": 1.3345890045166016, "learning_rate": 4.456026109497486e-05, "loss": 6.9233, "step": 14084 }, { "epoch": 0.5938402512806459, "grad_norm": 1.3234935998916626, "learning_rate": 4.455948030856679e-05, "loss": 6.4961, "step": 14085 }, { "epoch": 0.5938824124628455, "grad_norm": 1.686074137687683, "learning_rate": 4.455869952215872e-05, "loss": 6.8632, "step": 14086 }, { "epoch": 0.593924573645045, "grad_norm": 2.3205015659332275, "learning_rate": 4.4557918735750645e-05, "loss": 7.2036, "step": 14087 }, { "epoch": 0.5939667348272446, "grad_norm": 2.7206804752349854, "learning_rate": 4.455713794934258e-05, "loss": 6.3708, "step": 14088 }, { "epoch": 0.5940088960094441, "grad_norm": 1.6682610511779785, "learning_rate": 4.455635716293451e-05, "loss": 6.356, "step": 14089 }, { "epoch": 0.5940510571916436, "grad_norm": 1.7673588991165161, "learning_rate": 4.4555576376526445e-05, "loss": 7.0924, "step": 14090 }, { "epoch": 0.5940932183738432, "grad_norm": 1.1864784955978394, "learning_rate": 4.455479559011837e-05, "loss": 7.4188, "step": 14091 }, { "epoch": 0.5941353795560428, "grad_norm": 1.4114627838134766, "learning_rate": 4.45540148037103e-05, "loss": 6.5988, "step": 14092 }, { "epoch": 0.5941775407382422, "grad_norm": 1.6463359594345093, "learning_rate": 4.455323401730223e-05, "loss": 6.7137, "step": 14093 }, { "epoch": 0.5942197019204418, "grad_norm": 2.306243419647217, "learning_rate": 4.455245323089416e-05, "loss": 6.4811, "step": 14094 }, { "epoch": 0.5942618631026414, "grad_norm": 1.180167317390442, "learning_rate": 4.455167244448609e-05, "loss": 6.921, "step": 14095 }, { "epoch": 0.594304024284841, "grad_norm": 1.7917765378952026, "learning_rate": 4.455089165807802e-05, "loss": 6.5979, "step": 14096 }, { "epoch": 0.5943461854670405, "grad_norm": 1.4492541551589966, "learning_rate": 4.455011087166995e-05, "loss": 6.6184, "step": 14097 }, { "epoch": 0.59438834664924, "grad_norm": 1.7905668020248413, "learning_rate": 4.4549330085261875e-05, "loss": 6.4434, "step": 14098 }, { "epoch": 0.5944305078314396, "grad_norm": 1.3919579982757568, "learning_rate": 4.4548549298853806e-05, "loss": 6.2887, "step": 14099 }, { "epoch": 0.5944726690136392, "grad_norm": 1.469177007675171, "learning_rate": 4.454776851244574e-05, "loss": 7.0226, "step": 14100 }, { "epoch": 0.5945148301958387, "grad_norm": 2.1703224182128906, "learning_rate": 4.454698772603767e-05, "loss": 6.9067, "step": 14101 }, { "epoch": 0.5945569913780382, "grad_norm": 7.406850814819336, "learning_rate": 4.45462069396296e-05, "loss": 11.6393, "step": 14102 }, { "epoch": 0.5945991525602378, "grad_norm": 1.6777101755142212, "learning_rate": 4.4545426153221524e-05, "loss": 6.8941, "step": 14103 }, { "epoch": 0.5946413137424373, "grad_norm": 2.2248599529266357, "learning_rate": 4.4544645366813456e-05, "loss": 7.131, "step": 14104 }, { "epoch": 0.5946834749246369, "grad_norm": 1.0719155073165894, "learning_rate": 4.454386458040539e-05, "loss": 6.4052, "step": 14105 }, { "epoch": 0.5947256361068365, "grad_norm": 1.2791104316711426, "learning_rate": 4.454308379399731e-05, "loss": 6.7015, "step": 14106 }, { "epoch": 0.594767797289036, "grad_norm": 1.2220985889434814, "learning_rate": 4.454230300758925e-05, "loss": 6.591, "step": 14107 }, { "epoch": 0.5948099584712355, "grad_norm": 1.2895333766937256, "learning_rate": 4.4541522221181174e-05, "loss": 7.0046, "step": 14108 }, { "epoch": 0.5948521196534351, "grad_norm": 1.3423709869384766, "learning_rate": 4.4540741434773105e-05, "loss": 6.3321, "step": 14109 }, { "epoch": 0.5948942808356347, "grad_norm": 3.550168991088867, "learning_rate": 4.4539960648365036e-05, "loss": 7.7042, "step": 14110 }, { "epoch": 0.5949364420178341, "grad_norm": 1.8211073875427246, "learning_rate": 4.453917986195697e-05, "loss": 7.0602, "step": 14111 }, { "epoch": 0.5949786032000337, "grad_norm": 1.5202414989471436, "learning_rate": 4.453839907554889e-05, "loss": 6.7299, "step": 14112 }, { "epoch": 0.5950207643822333, "grad_norm": 4.220954895019531, "learning_rate": 4.4537618289140823e-05, "loss": 6.6625, "step": 14113 }, { "epoch": 0.5950629255644329, "grad_norm": 1.5897929668426514, "learning_rate": 4.4536837502732755e-05, "loss": 6.4095, "step": 14114 }, { "epoch": 0.5951050867466324, "grad_norm": 1.0625684261322021, "learning_rate": 4.4536056716324686e-05, "loss": 7.0603, "step": 14115 }, { "epoch": 0.5951472479288319, "grad_norm": 0.9554828405380249, "learning_rate": 4.453527592991662e-05, "loss": 6.3374, "step": 14116 }, { "epoch": 0.5951894091110315, "grad_norm": 0.9574723243713379, "learning_rate": 4.453449514350854e-05, "loss": 6.5632, "step": 14117 }, { "epoch": 0.595231570293231, "grad_norm": 3.0522072315216064, "learning_rate": 4.453371435710047e-05, "loss": 8.6593, "step": 14118 }, { "epoch": 0.5952737314754306, "grad_norm": 1.2629607915878296, "learning_rate": 4.4532933570692404e-05, "loss": 6.7415, "step": 14119 }, { "epoch": 0.5953158926576301, "grad_norm": 1.2780567407608032, "learning_rate": 4.453215278428433e-05, "loss": 7.166, "step": 14120 }, { "epoch": 0.5953580538398296, "grad_norm": 1.470270037651062, "learning_rate": 4.453137199787627e-05, "loss": 6.4316, "step": 14121 }, { "epoch": 0.5954002150220292, "grad_norm": 1.3378387689590454, "learning_rate": 4.453059121146819e-05, "loss": 6.4764, "step": 14122 }, { "epoch": 0.5954423762042288, "grad_norm": 1.4175752401351929, "learning_rate": 4.452981042506012e-05, "loss": 6.3508, "step": 14123 }, { "epoch": 0.5954845373864284, "grad_norm": 1.5126830339431763, "learning_rate": 4.4529029638652054e-05, "loss": 6.9138, "step": 14124 }, { "epoch": 0.5955266985686278, "grad_norm": 2.3506898880004883, "learning_rate": 4.452824885224398e-05, "loss": 7.147, "step": 14125 }, { "epoch": 0.5955688597508274, "grad_norm": 2.047938108444214, "learning_rate": 4.4527468065835916e-05, "loss": 7.5909, "step": 14126 }, { "epoch": 0.595611020933027, "grad_norm": 1.4695087671279907, "learning_rate": 4.452668727942784e-05, "loss": 6.2789, "step": 14127 }, { "epoch": 0.5956531821152266, "grad_norm": 0.9904171228408813, "learning_rate": 4.452590649301977e-05, "loss": 6.7392, "step": 14128 }, { "epoch": 0.595695343297426, "grad_norm": 1.1027346849441528, "learning_rate": 4.45251257066117e-05, "loss": 6.2444, "step": 14129 }, { "epoch": 0.5957375044796256, "grad_norm": 0.9132255911827087, "learning_rate": 4.452434492020363e-05, "loss": 6.6667, "step": 14130 }, { "epoch": 0.5957796656618252, "grad_norm": 0.9710085988044739, "learning_rate": 4.452356413379556e-05, "loss": 6.388, "step": 14131 }, { "epoch": 0.5958218268440247, "grad_norm": 3.4142396450042725, "learning_rate": 4.452278334738749e-05, "loss": 6.8984, "step": 14132 }, { "epoch": 0.5958639880262242, "grad_norm": 2.9720842838287354, "learning_rate": 4.452200256097942e-05, "loss": 6.4132, "step": 14133 }, { "epoch": 0.5959061492084238, "grad_norm": 3.7844908237457275, "learning_rate": 4.4521221774571346e-05, "loss": 7.3889, "step": 14134 }, { "epoch": 0.5959483103906233, "grad_norm": 0.9128275513648987, "learning_rate": 4.4520440988163284e-05, "loss": 6.3986, "step": 14135 }, { "epoch": 0.5959904715728229, "grad_norm": 1.589590311050415, "learning_rate": 4.451966020175521e-05, "loss": 7.1522, "step": 14136 }, { "epoch": 0.5960326327550225, "grad_norm": 1.2542555332183838, "learning_rate": 4.451887941534714e-05, "loss": 6.3284, "step": 14137 }, { "epoch": 0.596074793937222, "grad_norm": 1.585548758506775, "learning_rate": 4.451809862893907e-05, "loss": 7.1779, "step": 14138 }, { "epoch": 0.5961169551194215, "grad_norm": 2.9317667484283447, "learning_rate": 4.4517317842530996e-05, "loss": 7.1528, "step": 14139 }, { "epoch": 0.5961591163016211, "grad_norm": 2.8728389739990234, "learning_rate": 4.4516537056122934e-05, "loss": 7.2359, "step": 14140 }, { "epoch": 0.5962012774838207, "grad_norm": 0.8231099843978882, "learning_rate": 4.451575626971486e-05, "loss": 6.9649, "step": 14141 }, { "epoch": 0.5962434386660201, "grad_norm": 1.3448033332824707, "learning_rate": 4.451497548330679e-05, "loss": 6.2323, "step": 14142 }, { "epoch": 0.5962855998482197, "grad_norm": 2.1749556064605713, "learning_rate": 4.451419469689872e-05, "loss": 6.395, "step": 14143 }, { "epoch": 0.5963277610304193, "grad_norm": 0.7764455676078796, "learning_rate": 4.4513413910490645e-05, "loss": 6.5026, "step": 14144 }, { "epoch": 0.5963699222126189, "grad_norm": 2.032039165496826, "learning_rate": 4.4512633124082576e-05, "loss": 7.5845, "step": 14145 }, { "epoch": 0.5964120833948184, "grad_norm": 2.521052360534668, "learning_rate": 4.451185233767451e-05, "loss": 6.3243, "step": 14146 }, { "epoch": 0.5964542445770179, "grad_norm": 1.2743579149246216, "learning_rate": 4.451107155126644e-05, "loss": 6.4467, "step": 14147 }, { "epoch": 0.5964964057592175, "grad_norm": 3.13873028755188, "learning_rate": 4.451029076485837e-05, "loss": 7.9203, "step": 14148 }, { "epoch": 0.596538566941417, "grad_norm": 2.066962480545044, "learning_rate": 4.4509509978450295e-05, "loss": 6.7146, "step": 14149 }, { "epoch": 0.5965807281236166, "grad_norm": 1.2858086824417114, "learning_rate": 4.4508729192042226e-05, "loss": 6.4111, "step": 14150 }, { "epoch": 0.5966228893058161, "grad_norm": 2.805765390396118, "learning_rate": 4.450794840563416e-05, "loss": 7.7619, "step": 14151 }, { "epoch": 0.5966650504880157, "grad_norm": 1.1388722658157349, "learning_rate": 4.450716761922609e-05, "loss": 6.2697, "step": 14152 }, { "epoch": 0.5967072116702152, "grad_norm": 3.734570264816284, "learning_rate": 4.450638683281801e-05, "loss": 7.8651, "step": 14153 }, { "epoch": 0.5967493728524148, "grad_norm": 1.304555058479309, "learning_rate": 4.450560604640995e-05, "loss": 6.6335, "step": 14154 }, { "epoch": 0.5967915340346144, "grad_norm": 3.595659017562866, "learning_rate": 4.4504825260001875e-05, "loss": 9.2009, "step": 14155 }, { "epoch": 0.5968336952168138, "grad_norm": 2.1788954734802246, "learning_rate": 4.450404447359381e-05, "loss": 6.4519, "step": 14156 }, { "epoch": 0.5968758563990134, "grad_norm": 1.4262357950210571, "learning_rate": 4.450326368718574e-05, "loss": 7.3185, "step": 14157 }, { "epoch": 0.596918017581213, "grad_norm": 2.23126482963562, "learning_rate": 4.450248290077766e-05, "loss": 6.7554, "step": 14158 }, { "epoch": 0.5969601787634126, "grad_norm": 1.6482995748519897, "learning_rate": 4.45017021143696e-05, "loss": 6.4282, "step": 14159 }, { "epoch": 0.597002339945612, "grad_norm": 1.5118448734283447, "learning_rate": 4.4500921327961525e-05, "loss": 6.3634, "step": 14160 }, { "epoch": 0.5970445011278116, "grad_norm": 2.4736006259918213, "learning_rate": 4.4500140541553456e-05, "loss": 6.8825, "step": 14161 }, { "epoch": 0.5970866623100112, "grad_norm": 3.4721133708953857, "learning_rate": 4.449935975514539e-05, "loss": 7.7868, "step": 14162 }, { "epoch": 0.5971288234922107, "grad_norm": 2.346623182296753, "learning_rate": 4.449857896873731e-05, "loss": 6.3385, "step": 14163 }, { "epoch": 0.5971709846744103, "grad_norm": 3.785050630569458, "learning_rate": 4.449779818232924e-05, "loss": 6.8737, "step": 14164 }, { "epoch": 0.5972131458566098, "grad_norm": 1.748659610748291, "learning_rate": 4.4497017395921174e-05, "loss": 6.2535, "step": 14165 }, { "epoch": 0.5972553070388094, "grad_norm": 0.9904860854148865, "learning_rate": 4.4496236609513106e-05, "loss": 6.8103, "step": 14166 }, { "epoch": 0.5972974682210089, "grad_norm": 2.4474010467529297, "learning_rate": 4.449545582310503e-05, "loss": 6.4131, "step": 14167 }, { "epoch": 0.5973396294032085, "grad_norm": 2.704127073287964, "learning_rate": 4.449467503669696e-05, "loss": 6.9044, "step": 14168 }, { "epoch": 0.597381790585408, "grad_norm": 1.7051680088043213, "learning_rate": 4.449389425028889e-05, "loss": 7.0957, "step": 14169 }, { "epoch": 0.5974239517676075, "grad_norm": 3.452526569366455, "learning_rate": 4.4493113463880824e-05, "loss": 6.7652, "step": 14170 }, { "epoch": 0.5974661129498071, "grad_norm": 3.94891095161438, "learning_rate": 4.4492332677472755e-05, "loss": 7.2792, "step": 14171 }, { "epoch": 0.5975082741320067, "grad_norm": 1.6702425479888916, "learning_rate": 4.449155189106468e-05, "loss": 6.4241, "step": 14172 }, { "epoch": 0.5975504353142063, "grad_norm": 3.666985273361206, "learning_rate": 4.449077110465662e-05, "loss": 7.5691, "step": 14173 }, { "epoch": 0.5975925964964057, "grad_norm": 4.198912143707275, "learning_rate": 4.448999031824854e-05, "loss": 9.3962, "step": 14174 }, { "epoch": 0.5976347576786053, "grad_norm": 1.690913438796997, "learning_rate": 4.448920953184047e-05, "loss": 6.3701, "step": 14175 }, { "epoch": 0.5976769188608049, "grad_norm": 1.4868801832199097, "learning_rate": 4.4488428745432405e-05, "loss": 6.5976, "step": 14176 }, { "epoch": 0.5977190800430044, "grad_norm": 1.7435699701309204, "learning_rate": 4.448764795902433e-05, "loss": 6.2255, "step": 14177 }, { "epoch": 0.5977612412252039, "grad_norm": 0.8442075252532959, "learning_rate": 4.448686717261626e-05, "loss": 6.8278, "step": 14178 }, { "epoch": 0.5978034024074035, "grad_norm": 2.035477638244629, "learning_rate": 4.448608638620819e-05, "loss": 7.1453, "step": 14179 }, { "epoch": 0.5978455635896031, "grad_norm": 0.8626003265380859, "learning_rate": 4.448530559980012e-05, "loss": 6.3686, "step": 14180 }, { "epoch": 0.5978877247718026, "grad_norm": 1.0732120275497437, "learning_rate": 4.4484524813392054e-05, "loss": 6.6655, "step": 14181 }, { "epoch": 0.5979298859540022, "grad_norm": 1.097367525100708, "learning_rate": 4.448374402698398e-05, "loss": 6.2151, "step": 14182 }, { "epoch": 0.5979720471362017, "grad_norm": 2.9272639751434326, "learning_rate": 4.448296324057591e-05, "loss": 6.3172, "step": 14183 }, { "epoch": 0.5980142083184012, "grad_norm": 1.4307748079299927, "learning_rate": 4.448218245416784e-05, "loss": 6.2259, "step": 14184 }, { "epoch": 0.5980563695006008, "grad_norm": 3.3005456924438477, "learning_rate": 4.448140166775977e-05, "loss": 8.0006, "step": 14185 }, { "epoch": 0.5980985306828004, "grad_norm": 2.021275758743286, "learning_rate": 4.44806208813517e-05, "loss": 6.4865, "step": 14186 }, { "epoch": 0.5981406918649999, "grad_norm": 1.4324839115142822, "learning_rate": 4.447984009494363e-05, "loss": 6.7195, "step": 14187 }, { "epoch": 0.5981828530471994, "grad_norm": 3.3595354557037354, "learning_rate": 4.447905930853556e-05, "loss": 7.9061, "step": 14188 }, { "epoch": 0.598225014229399, "grad_norm": 2.9206366539001465, "learning_rate": 4.4478278522127484e-05, "loss": 6.266, "step": 14189 }, { "epoch": 0.5982671754115986, "grad_norm": 2.208287477493286, "learning_rate": 4.447749773571942e-05, "loss": 6.8929, "step": 14190 }, { "epoch": 0.598309336593798, "grad_norm": 1.2804092168807983, "learning_rate": 4.4476716949311347e-05, "loss": 6.3509, "step": 14191 }, { "epoch": 0.5983514977759976, "grad_norm": 1.5996402502059937, "learning_rate": 4.447593616290328e-05, "loss": 7.1304, "step": 14192 }, { "epoch": 0.5983936589581972, "grad_norm": 1.4587005376815796, "learning_rate": 4.447515537649521e-05, "loss": 6.8232, "step": 14193 }, { "epoch": 0.5984358201403968, "grad_norm": 1.3081368207931519, "learning_rate": 4.4474374590087134e-05, "loss": 6.3202, "step": 14194 }, { "epoch": 0.5984779813225963, "grad_norm": 1.5722146034240723, "learning_rate": 4.447359380367907e-05, "loss": 6.418, "step": 14195 }, { "epoch": 0.5985201425047958, "grad_norm": 1.8107935190200806, "learning_rate": 4.4472813017270996e-05, "loss": 7.0765, "step": 14196 }, { "epoch": 0.5985623036869954, "grad_norm": 1.530106544494629, "learning_rate": 4.447203223086293e-05, "loss": 6.5842, "step": 14197 }, { "epoch": 0.5986044648691949, "grad_norm": 1.8326694965362549, "learning_rate": 4.447125144445486e-05, "loss": 7.2253, "step": 14198 }, { "epoch": 0.5986466260513945, "grad_norm": 2.489968776702881, "learning_rate": 4.447047065804679e-05, "loss": 7.2705, "step": 14199 }, { "epoch": 0.598688787233594, "grad_norm": 1.4407018423080444, "learning_rate": 4.4469689871638714e-05, "loss": 6.9979, "step": 14200 }, { "epoch": 0.5987309484157936, "grad_norm": 2.627288341522217, "learning_rate": 4.4468909085230646e-05, "loss": 6.1254, "step": 14201 }, { "epoch": 0.5987731095979931, "grad_norm": 2.2064807415008545, "learning_rate": 4.446812829882258e-05, "loss": 7.0455, "step": 14202 }, { "epoch": 0.5988152707801927, "grad_norm": 1.7522929906845093, "learning_rate": 4.44673475124145e-05, "loss": 6.6417, "step": 14203 }, { "epoch": 0.5988574319623923, "grad_norm": 2.1529831886291504, "learning_rate": 4.446656672600644e-05, "loss": 6.7571, "step": 14204 }, { "epoch": 0.5988995931445917, "grad_norm": 1.2277288436889648, "learning_rate": 4.4465785939598364e-05, "loss": 6.7964, "step": 14205 }, { "epoch": 0.5989417543267913, "grad_norm": 1.6902052164077759, "learning_rate": 4.4465005153190295e-05, "loss": 6.2567, "step": 14206 }, { "epoch": 0.5989839155089909, "grad_norm": 3.7528326511383057, "learning_rate": 4.4464224366782226e-05, "loss": 6.1476, "step": 14207 }, { "epoch": 0.5990260766911905, "grad_norm": 2.270810842514038, "learning_rate": 4.446344358037415e-05, "loss": 6.6105, "step": 14208 }, { "epoch": 0.5990682378733899, "grad_norm": 1.472590684890747, "learning_rate": 4.446266279396609e-05, "loss": 6.8842, "step": 14209 }, { "epoch": 0.5991103990555895, "grad_norm": 2.3760626316070557, "learning_rate": 4.4461882007558013e-05, "loss": 7.5481, "step": 14210 }, { "epoch": 0.5991525602377891, "grad_norm": 1.563822865486145, "learning_rate": 4.4461101221149945e-05, "loss": 6.3302, "step": 14211 }, { "epoch": 0.5991947214199886, "grad_norm": 2.3199968338012695, "learning_rate": 4.4460320434741876e-05, "loss": 6.7245, "step": 14212 }, { "epoch": 0.5992368826021882, "grad_norm": 1.5247852802276611, "learning_rate": 4.44595396483338e-05, "loss": 6.7926, "step": 14213 }, { "epoch": 0.5992790437843877, "grad_norm": 1.1358227729797363, "learning_rate": 4.445875886192573e-05, "loss": 7.0026, "step": 14214 }, { "epoch": 0.5993212049665873, "grad_norm": 1.218875527381897, "learning_rate": 4.445797807551766e-05, "loss": 6.4574, "step": 14215 }, { "epoch": 0.5993633661487868, "grad_norm": 1.579237461090088, "learning_rate": 4.4457197289109594e-05, "loss": 6.6403, "step": 14216 }, { "epoch": 0.5994055273309864, "grad_norm": 1.6060059070587158, "learning_rate": 4.4456416502701525e-05, "loss": 6.7809, "step": 14217 }, { "epoch": 0.5994476885131859, "grad_norm": 1.4617854356765747, "learning_rate": 4.445563571629346e-05, "loss": 6.874, "step": 14218 }, { "epoch": 0.5994898496953854, "grad_norm": 2.369457721710205, "learning_rate": 4.445485492988538e-05, "loss": 7.698, "step": 14219 }, { "epoch": 0.599532010877585, "grad_norm": 2.6411423683166504, "learning_rate": 4.445407414347731e-05, "loss": 7.1183, "step": 14220 }, { "epoch": 0.5995741720597846, "grad_norm": 2.9503068923950195, "learning_rate": 4.4453293357069244e-05, "loss": 7.7684, "step": 14221 }, { "epoch": 0.5996163332419842, "grad_norm": 2.8522675037384033, "learning_rate": 4.445251257066117e-05, "loss": 6.3911, "step": 14222 }, { "epoch": 0.5996584944241836, "grad_norm": 2.5130462646484375, "learning_rate": 4.4451731784253106e-05, "loss": 7.6546, "step": 14223 }, { "epoch": 0.5997006556063832, "grad_norm": 2.3408875465393066, "learning_rate": 4.445095099784503e-05, "loss": 6.4668, "step": 14224 }, { "epoch": 0.5997428167885828, "grad_norm": 2.548814296722412, "learning_rate": 4.445017021143696e-05, "loss": 7.2272, "step": 14225 }, { "epoch": 0.5997849779707823, "grad_norm": 3.1940743923187256, "learning_rate": 4.444938942502889e-05, "loss": 6.5119, "step": 14226 }, { "epoch": 0.5998271391529818, "grad_norm": 2.134584665298462, "learning_rate": 4.444860863862082e-05, "loss": 6.486, "step": 14227 }, { "epoch": 0.5998693003351814, "grad_norm": 2.1066644191741943, "learning_rate": 4.4447827852212756e-05, "loss": 7.1485, "step": 14228 }, { "epoch": 0.599911461517381, "grad_norm": 2.0518569946289062, "learning_rate": 4.444704706580468e-05, "loss": 6.357, "step": 14229 }, { "epoch": 0.5999536226995805, "grad_norm": 1.517566442489624, "learning_rate": 4.444626627939661e-05, "loss": 6.5529, "step": 14230 }, { "epoch": 0.5999957838817801, "grad_norm": 1.7323204278945923, "learning_rate": 4.444548549298854e-05, "loss": 6.7732, "step": 14231 }, { "epoch": 0.6000379450639796, "grad_norm": 1.5963008403778076, "learning_rate": 4.444470470658047e-05, "loss": 6.8421, "step": 14232 }, { "epoch": 0.6000801062461791, "grad_norm": 3.946855068206787, "learning_rate": 4.44439239201724e-05, "loss": 7.5434, "step": 14233 }, { "epoch": 0.6001222674283787, "grad_norm": 1.2687907218933105, "learning_rate": 4.444314313376433e-05, "loss": 6.6066, "step": 14234 }, { "epoch": 0.6001644286105783, "grad_norm": 2.1312618255615234, "learning_rate": 4.444236234735626e-05, "loss": 6.6124, "step": 14235 }, { "epoch": 0.6002065897927777, "grad_norm": 1.1382440328598022, "learning_rate": 4.4441581560948185e-05, "loss": 6.451, "step": 14236 }, { "epoch": 0.6002487509749773, "grad_norm": 2.3304100036621094, "learning_rate": 4.444080077454012e-05, "loss": 6.7216, "step": 14237 }, { "epoch": 0.6002909121571769, "grad_norm": 3.3259851932525635, "learning_rate": 4.444001998813205e-05, "loss": 6.5086, "step": 14238 }, { "epoch": 0.6003330733393765, "grad_norm": 2.9138143062591553, "learning_rate": 4.443923920172398e-05, "loss": 7.1082, "step": 14239 }, { "epoch": 0.600375234521576, "grad_norm": 2.056023120880127, "learning_rate": 4.443845841531591e-05, "loss": 7.0267, "step": 14240 }, { "epoch": 0.6004173957037755, "grad_norm": 6.092745304107666, "learning_rate": 4.4437677628907835e-05, "loss": 7.2868, "step": 14241 }, { "epoch": 0.6004595568859751, "grad_norm": 1.7452038526535034, "learning_rate": 4.443689684249977e-05, "loss": 6.7157, "step": 14242 }, { "epoch": 0.6005017180681747, "grad_norm": 3.5262298583984375, "learning_rate": 4.44361160560917e-05, "loss": 7.723, "step": 14243 }, { "epoch": 0.6005438792503742, "grad_norm": 4.194153785705566, "learning_rate": 4.443533526968363e-05, "loss": 6.7927, "step": 14244 }, { "epoch": 0.6005860404325737, "grad_norm": 4.956404685974121, "learning_rate": 4.443455448327556e-05, "loss": 6.5121, "step": 14245 }, { "epoch": 0.6006282016147733, "grad_norm": 5.378892421722412, "learning_rate": 4.4433773696867485e-05, "loss": 6.3695, "step": 14246 }, { "epoch": 0.6006703627969728, "grad_norm": 2.8871259689331055, "learning_rate": 4.4432992910459416e-05, "loss": 6.242, "step": 14247 }, { "epoch": 0.6007125239791724, "grad_norm": 1.6739442348480225, "learning_rate": 4.443221212405135e-05, "loss": 6.9279, "step": 14248 }, { "epoch": 0.6007546851613719, "grad_norm": 1.977475881576538, "learning_rate": 4.443143133764328e-05, "loss": 7.1399, "step": 14249 }, { "epoch": 0.6007968463435714, "grad_norm": 3.2486560344696045, "learning_rate": 4.443065055123521e-05, "loss": 6.6356, "step": 14250 }, { "epoch": 0.600839007525771, "grad_norm": 1.787045955657959, "learning_rate": 4.4429869764827134e-05, "loss": 6.9601, "step": 14251 }, { "epoch": 0.6008811687079706, "grad_norm": 2.7378697395324707, "learning_rate": 4.4429088978419065e-05, "loss": 7.2312, "step": 14252 }, { "epoch": 0.6009233298901702, "grad_norm": 2.3755154609680176, "learning_rate": 4.4428308192010997e-05, "loss": 6.346, "step": 14253 }, { "epoch": 0.6009654910723696, "grad_norm": 1.9057649374008179, "learning_rate": 4.442752740560293e-05, "loss": 6.5606, "step": 14254 }, { "epoch": 0.6010076522545692, "grad_norm": 2.0287044048309326, "learning_rate": 4.442674661919485e-05, "loss": 6.4812, "step": 14255 }, { "epoch": 0.6010498134367688, "grad_norm": 2.3094024658203125, "learning_rate": 4.4425965832786784e-05, "loss": 6.9308, "step": 14256 }, { "epoch": 0.6010919746189683, "grad_norm": 2.6914570331573486, "learning_rate": 4.4425185046378715e-05, "loss": 6.839, "step": 14257 }, { "epoch": 0.6011341358011678, "grad_norm": 2.2312097549438477, "learning_rate": 4.442440425997064e-05, "loss": 6.3746, "step": 14258 }, { "epoch": 0.6011762969833674, "grad_norm": 1.8109912872314453, "learning_rate": 4.442362347356258e-05, "loss": 6.9392, "step": 14259 }, { "epoch": 0.601218458165567, "grad_norm": 1.8678889274597168, "learning_rate": 4.44228426871545e-05, "loss": 6.8838, "step": 14260 }, { "epoch": 0.6012606193477665, "grad_norm": 1.3250449895858765, "learning_rate": 4.442206190074644e-05, "loss": 6.7125, "step": 14261 }, { "epoch": 0.6013027805299661, "grad_norm": 2.2733521461486816, "learning_rate": 4.4421281114338364e-05, "loss": 6.9424, "step": 14262 }, { "epoch": 0.6013449417121656, "grad_norm": 2.1787846088409424, "learning_rate": 4.4420500327930296e-05, "loss": 7.3922, "step": 14263 }, { "epoch": 0.6013871028943651, "grad_norm": 2.640442132949829, "learning_rate": 4.441971954152223e-05, "loss": 6.5196, "step": 14264 }, { "epoch": 0.6014292640765647, "grad_norm": 1.9032280445098877, "learning_rate": 4.441893875511415e-05, "loss": 7.0964, "step": 14265 }, { "epoch": 0.6014714252587643, "grad_norm": 2.142728567123413, "learning_rate": 4.441815796870608e-05, "loss": 7.3065, "step": 14266 }, { "epoch": 0.6015135864409638, "grad_norm": 1.5006974935531616, "learning_rate": 4.4417377182298014e-05, "loss": 6.4221, "step": 14267 }, { "epoch": 0.6015557476231633, "grad_norm": 3.2898294925689697, "learning_rate": 4.4416596395889945e-05, "loss": 7.2849, "step": 14268 }, { "epoch": 0.6015979088053629, "grad_norm": 3.2042579650878906, "learning_rate": 4.441581560948187e-05, "loss": 6.2883, "step": 14269 }, { "epoch": 0.6016400699875625, "grad_norm": 2.990086793899536, "learning_rate": 4.44150348230738e-05, "loss": 7.4391, "step": 14270 }, { "epoch": 0.601682231169762, "grad_norm": 2.4001760482788086, "learning_rate": 4.441425403666573e-05, "loss": 6.8265, "step": 14271 }, { "epoch": 0.6017243923519615, "grad_norm": 1.388145089149475, "learning_rate": 4.441347325025766e-05, "loss": 6.4162, "step": 14272 }, { "epoch": 0.6017665535341611, "grad_norm": 1.2848553657531738, "learning_rate": 4.4412692463849595e-05, "loss": 6.4041, "step": 14273 }, { "epoch": 0.6018087147163607, "grad_norm": 1.9684944152832031, "learning_rate": 4.441191167744152e-05, "loss": 7.0035, "step": 14274 }, { "epoch": 0.6018508758985602, "grad_norm": 2.7521650791168213, "learning_rate": 4.441113089103345e-05, "loss": 7.4939, "step": 14275 }, { "epoch": 0.6018930370807597, "grad_norm": 2.2937161922454834, "learning_rate": 4.441035010462538e-05, "loss": 6.7293, "step": 14276 }, { "epoch": 0.6019351982629593, "grad_norm": 2.654836654663086, "learning_rate": 4.4409569318217306e-05, "loss": 6.3521, "step": 14277 }, { "epoch": 0.6019773594451588, "grad_norm": 2.925091028213501, "learning_rate": 4.4408788531809244e-05, "loss": 7.3593, "step": 14278 }, { "epoch": 0.6020195206273584, "grad_norm": 1.701098918914795, "learning_rate": 4.440800774540117e-05, "loss": 6.6289, "step": 14279 }, { "epoch": 0.602061681809558, "grad_norm": 3.5961380004882812, "learning_rate": 4.44072269589931e-05, "loss": 7.0283, "step": 14280 }, { "epoch": 0.6021038429917575, "grad_norm": 4.469405651092529, "learning_rate": 4.440644617258503e-05, "loss": 7.8013, "step": 14281 }, { "epoch": 0.602146004173957, "grad_norm": 3.7581865787506104, "learning_rate": 4.4405665386176956e-05, "loss": 7.0664, "step": 14282 }, { "epoch": 0.6021881653561566, "grad_norm": 3.0523226261138916, "learning_rate": 4.440488459976889e-05, "loss": 7.1182, "step": 14283 }, { "epoch": 0.6022303265383562, "grad_norm": 2.8735668659210205, "learning_rate": 4.440410381336082e-05, "loss": 7.5586, "step": 14284 }, { "epoch": 0.6022724877205556, "grad_norm": 1.7161606550216675, "learning_rate": 4.440332302695275e-05, "loss": 6.8362, "step": 14285 }, { "epoch": 0.6023146489027552, "grad_norm": 2.4689619541168213, "learning_rate": 4.440254224054468e-05, "loss": 6.3099, "step": 14286 }, { "epoch": 0.6023568100849548, "grad_norm": 1.673492193222046, "learning_rate": 4.440176145413661e-05, "loss": 6.9391, "step": 14287 }, { "epoch": 0.6023989712671544, "grad_norm": 2.638343334197998, "learning_rate": 4.4400980667728536e-05, "loss": 6.3508, "step": 14288 }, { "epoch": 0.6024411324493539, "grad_norm": 3.0410804748535156, "learning_rate": 4.440019988132047e-05, "loss": 6.6568, "step": 14289 }, { "epoch": 0.6024832936315534, "grad_norm": 1.5749152898788452, "learning_rate": 4.43994190949124e-05, "loss": 6.7837, "step": 14290 }, { "epoch": 0.602525454813753, "grad_norm": 2.123924732208252, "learning_rate": 4.4398638308504323e-05, "loss": 6.237, "step": 14291 }, { "epoch": 0.6025676159959525, "grad_norm": 3.748147487640381, "learning_rate": 4.439785752209626e-05, "loss": 6.2773, "step": 14292 }, { "epoch": 0.6026097771781521, "grad_norm": 1.5801037549972534, "learning_rate": 4.4397076735688186e-05, "loss": 6.4171, "step": 14293 }, { "epoch": 0.6026519383603516, "grad_norm": 2.4545273780822754, "learning_rate": 4.439629594928012e-05, "loss": 6.4926, "step": 14294 }, { "epoch": 0.6026940995425512, "grad_norm": 1.4260581731796265, "learning_rate": 4.439551516287205e-05, "loss": 6.4017, "step": 14295 }, { "epoch": 0.6027362607247507, "grad_norm": 1.3939296007156372, "learning_rate": 4.439473437646397e-05, "loss": 6.5958, "step": 14296 }, { "epoch": 0.6027784219069503, "grad_norm": 3.103201389312744, "learning_rate": 4.439395359005591e-05, "loss": 7.924, "step": 14297 }, { "epoch": 0.6028205830891499, "grad_norm": 1.9393081665039062, "learning_rate": 4.4393172803647836e-05, "loss": 6.4304, "step": 14298 }, { "epoch": 0.6028627442713493, "grad_norm": 2.638597249984741, "learning_rate": 4.439239201723977e-05, "loss": 6.9256, "step": 14299 }, { "epoch": 0.6029049054535489, "grad_norm": 2.004638195037842, "learning_rate": 4.43916112308317e-05, "loss": 6.9751, "step": 14300 }, { "epoch": 0.6029470666357485, "grad_norm": 2.5884556770324707, "learning_rate": 4.439083044442362e-05, "loss": 7.0655, "step": 14301 }, { "epoch": 0.6029892278179481, "grad_norm": 4.222589492797852, "learning_rate": 4.4390049658015554e-05, "loss": 7.8524, "step": 14302 }, { "epoch": 0.6030313890001475, "grad_norm": 1.3503841161727905, "learning_rate": 4.4389268871607485e-05, "loss": 6.8302, "step": 14303 }, { "epoch": 0.6030735501823471, "grad_norm": 3.3620362281799316, "learning_rate": 4.4388488085199416e-05, "loss": 6.6657, "step": 14304 }, { "epoch": 0.6031157113645467, "grad_norm": 2.918897867202759, "learning_rate": 4.438770729879134e-05, "loss": 6.3988, "step": 14305 }, { "epoch": 0.6031578725467462, "grad_norm": 3.6028494834899902, "learning_rate": 4.438692651238328e-05, "loss": 6.4097, "step": 14306 }, { "epoch": 0.6032000337289457, "grad_norm": 1.4033513069152832, "learning_rate": 4.43861457259752e-05, "loss": 6.894, "step": 14307 }, { "epoch": 0.6032421949111453, "grad_norm": 1.3816914558410645, "learning_rate": 4.4385364939567135e-05, "loss": 6.5369, "step": 14308 }, { "epoch": 0.6032843560933449, "grad_norm": 1.1360455751419067, "learning_rate": 4.4384584153159066e-05, "loss": 6.675, "step": 14309 }, { "epoch": 0.6033265172755444, "grad_norm": 2.4174997806549072, "learning_rate": 4.438380336675099e-05, "loss": 6.2963, "step": 14310 }, { "epoch": 0.603368678457744, "grad_norm": 1.3112927675247192, "learning_rate": 4.438302258034293e-05, "loss": 6.424, "step": 14311 }, { "epoch": 0.6034108396399435, "grad_norm": 1.1873201131820679, "learning_rate": 4.438224179393485e-05, "loss": 6.4244, "step": 14312 }, { "epoch": 0.603453000822143, "grad_norm": 4.879882335662842, "learning_rate": 4.4381461007526784e-05, "loss": 7.5442, "step": 14313 }, { "epoch": 0.6034951620043426, "grad_norm": 1.9602971076965332, "learning_rate": 4.4380680221118715e-05, "loss": 6.976, "step": 14314 }, { "epoch": 0.6035373231865422, "grad_norm": 2.3073644638061523, "learning_rate": 4.437989943471064e-05, "loss": 6.4616, "step": 14315 }, { "epoch": 0.6035794843687416, "grad_norm": 2.3131604194641113, "learning_rate": 4.437911864830257e-05, "loss": 7.0285, "step": 14316 }, { "epoch": 0.6036216455509412, "grad_norm": 2.897500514984131, "learning_rate": 4.43783378618945e-05, "loss": 7.3481, "step": 14317 }, { "epoch": 0.6036638067331408, "grad_norm": 3.7201719284057617, "learning_rate": 4.4377557075486434e-05, "loss": 8.0482, "step": 14318 }, { "epoch": 0.6037059679153404, "grad_norm": 2.4812910556793213, "learning_rate": 4.4376776289078365e-05, "loss": 6.8581, "step": 14319 }, { "epoch": 0.6037481290975399, "grad_norm": 1.8969206809997559, "learning_rate": 4.437599550267029e-05, "loss": 6.9638, "step": 14320 }, { "epoch": 0.6037902902797394, "grad_norm": 2.763227701187134, "learning_rate": 4.437521471626222e-05, "loss": 6.2017, "step": 14321 }, { "epoch": 0.603832451461939, "grad_norm": 2.9148194789886475, "learning_rate": 4.437443392985415e-05, "loss": 7.6148, "step": 14322 }, { "epoch": 0.6038746126441386, "grad_norm": 2.5246644020080566, "learning_rate": 4.437365314344608e-05, "loss": 6.9286, "step": 14323 }, { "epoch": 0.6039167738263381, "grad_norm": 1.887856125831604, "learning_rate": 4.437287235703801e-05, "loss": 6.5292, "step": 14324 }, { "epoch": 0.6039589350085376, "grad_norm": 1.5445466041564941, "learning_rate": 4.4372091570629946e-05, "loss": 7.0239, "step": 14325 }, { "epoch": 0.6040010961907372, "grad_norm": 3.133730411529541, "learning_rate": 4.437131078422187e-05, "loss": 6.9759, "step": 14326 }, { "epoch": 0.6040432573729367, "grad_norm": 4.367567539215088, "learning_rate": 4.4370529997813795e-05, "loss": 8.11, "step": 14327 }, { "epoch": 0.6040854185551363, "grad_norm": 1.5149708986282349, "learning_rate": 4.436974921140573e-05, "loss": 6.9585, "step": 14328 }, { "epoch": 0.6041275797373359, "grad_norm": 3.6163008213043213, "learning_rate": 4.436896842499766e-05, "loss": 7.5479, "step": 14329 }, { "epoch": 0.6041697409195353, "grad_norm": 2.0071427822113037, "learning_rate": 4.4368187638589595e-05, "loss": 6.5842, "step": 14330 }, { "epoch": 0.6042119021017349, "grad_norm": 1.5663527250289917, "learning_rate": 4.436740685218152e-05, "loss": 6.6414, "step": 14331 }, { "epoch": 0.6042540632839345, "grad_norm": 4.25115966796875, "learning_rate": 4.436662606577345e-05, "loss": 6.9807, "step": 14332 }, { "epoch": 0.6042962244661341, "grad_norm": 2.273139238357544, "learning_rate": 4.436584527936538e-05, "loss": 6.5507, "step": 14333 }, { "epoch": 0.6043383856483335, "grad_norm": 3.052927017211914, "learning_rate": 4.436506449295731e-05, "loss": 7.788, "step": 14334 }, { "epoch": 0.6043805468305331, "grad_norm": 1.941480040550232, "learning_rate": 4.436428370654924e-05, "loss": 6.5771, "step": 14335 }, { "epoch": 0.6044227080127327, "grad_norm": 1.919568657875061, "learning_rate": 4.436350292014117e-05, "loss": 6.33, "step": 14336 }, { "epoch": 0.6044648691949323, "grad_norm": 2.28092098236084, "learning_rate": 4.43627221337331e-05, "loss": 7.5147, "step": 14337 }, { "epoch": 0.6045070303771318, "grad_norm": 3.2776050567626953, "learning_rate": 4.4361941347325025e-05, "loss": 7.0087, "step": 14338 }, { "epoch": 0.6045491915593313, "grad_norm": 1.4754599332809448, "learning_rate": 4.4361160560916956e-05, "loss": 6.7595, "step": 14339 }, { "epoch": 0.6045913527415309, "grad_norm": 2.321302652359009, "learning_rate": 4.436037977450889e-05, "loss": 6.987, "step": 14340 }, { "epoch": 0.6046335139237304, "grad_norm": 4.1086225509643555, "learning_rate": 4.435959898810081e-05, "loss": 6.4619, "step": 14341 }, { "epoch": 0.60467567510593, "grad_norm": 2.8338475227355957, "learning_rate": 4.435881820169275e-05, "loss": 7.029, "step": 14342 }, { "epoch": 0.6047178362881295, "grad_norm": 2.2192764282226562, "learning_rate": 4.4358037415284674e-05, "loss": 7.3439, "step": 14343 }, { "epoch": 0.604759997470329, "grad_norm": 2.469897985458374, "learning_rate": 4.4357256628876606e-05, "loss": 6.9813, "step": 14344 }, { "epoch": 0.6048021586525286, "grad_norm": 2.608792781829834, "learning_rate": 4.435647584246854e-05, "loss": 7.123, "step": 14345 }, { "epoch": 0.6048443198347282, "grad_norm": 1.4495905637741089, "learning_rate": 4.435569505606046e-05, "loss": 6.8837, "step": 14346 }, { "epoch": 0.6048864810169278, "grad_norm": 3.1432995796203613, "learning_rate": 4.43549142696524e-05, "loss": 6.7295, "step": 14347 }, { "epoch": 0.6049286421991272, "grad_norm": 1.3274396657943726, "learning_rate": 4.4354133483244324e-05, "loss": 6.6652, "step": 14348 }, { "epoch": 0.6049708033813268, "grad_norm": 2.474153757095337, "learning_rate": 4.4353352696836255e-05, "loss": 6.4936, "step": 14349 }, { "epoch": 0.6050129645635264, "grad_norm": 1.4461392164230347, "learning_rate": 4.4352571910428186e-05, "loss": 6.5241, "step": 14350 }, { "epoch": 0.605055125745726, "grad_norm": 1.801153540611267, "learning_rate": 4.435179112402012e-05, "loss": 6.9594, "step": 14351 }, { "epoch": 0.6050972869279254, "grad_norm": 0.9991348385810852, "learning_rate": 4.435101033761204e-05, "loss": 6.5594, "step": 14352 }, { "epoch": 0.605139448110125, "grad_norm": 2.1194493770599365, "learning_rate": 4.4350229551203973e-05, "loss": 6.5207, "step": 14353 }, { "epoch": 0.6051816092923246, "grad_norm": 2.5754103660583496, "learning_rate": 4.4349448764795905e-05, "loss": 6.6852, "step": 14354 }, { "epoch": 0.6052237704745241, "grad_norm": 2.8479349613189697, "learning_rate": 4.4348667978387836e-05, "loss": 6.9457, "step": 14355 }, { "epoch": 0.6052659316567236, "grad_norm": 3.4548044204711914, "learning_rate": 4.434788719197977e-05, "loss": 6.2667, "step": 14356 }, { "epoch": 0.6053080928389232, "grad_norm": 1.9126861095428467, "learning_rate": 4.434710640557169e-05, "loss": 6.4011, "step": 14357 }, { "epoch": 0.6053502540211227, "grad_norm": 1.8868699073791504, "learning_rate": 4.434632561916362e-05, "loss": 7.0886, "step": 14358 }, { "epoch": 0.6053924152033223, "grad_norm": 2.314289093017578, "learning_rate": 4.4345544832755554e-05, "loss": 7.0007, "step": 14359 }, { "epoch": 0.6054345763855219, "grad_norm": 4.148879528045654, "learning_rate": 4.434476404634748e-05, "loss": 7.0027, "step": 14360 }, { "epoch": 0.6054767375677214, "grad_norm": 1.8168613910675049, "learning_rate": 4.434398325993942e-05, "loss": 6.7257, "step": 14361 }, { "epoch": 0.6055188987499209, "grad_norm": 2.0292809009552, "learning_rate": 4.434320247353134e-05, "loss": 6.5747, "step": 14362 }, { "epoch": 0.6055610599321205, "grad_norm": 2.075969934463501, "learning_rate": 4.434242168712327e-05, "loss": 6.9222, "step": 14363 }, { "epoch": 0.6056032211143201, "grad_norm": 1.3473364114761353, "learning_rate": 4.4341640900715204e-05, "loss": 6.4881, "step": 14364 }, { "epoch": 0.6056453822965195, "grad_norm": 1.5760772228240967, "learning_rate": 4.434086011430713e-05, "loss": 6.8045, "step": 14365 }, { "epoch": 0.6056875434787191, "grad_norm": 2.2803051471710205, "learning_rate": 4.4340079327899066e-05, "loss": 6.4093, "step": 14366 }, { "epoch": 0.6057297046609187, "grad_norm": 1.8616331815719604, "learning_rate": 4.433929854149099e-05, "loss": 6.9418, "step": 14367 }, { "epoch": 0.6057718658431183, "grad_norm": 1.5847818851470947, "learning_rate": 4.433851775508292e-05, "loss": 6.3331, "step": 14368 }, { "epoch": 0.6058140270253178, "grad_norm": 2.055356979370117, "learning_rate": 4.433773696867485e-05, "loss": 6.4365, "step": 14369 }, { "epoch": 0.6058561882075173, "grad_norm": 1.4233895540237427, "learning_rate": 4.433695618226678e-05, "loss": 6.9555, "step": 14370 }, { "epoch": 0.6058983493897169, "grad_norm": 3.1234846115112305, "learning_rate": 4.433617539585871e-05, "loss": 7.2265, "step": 14371 }, { "epoch": 0.6059405105719164, "grad_norm": 1.1951414346694946, "learning_rate": 4.433539460945064e-05, "loss": 6.8655, "step": 14372 }, { "epoch": 0.605982671754116, "grad_norm": 2.6271450519561768, "learning_rate": 4.433461382304257e-05, "loss": 6.7191, "step": 14373 }, { "epoch": 0.6060248329363155, "grad_norm": 1.324628233909607, "learning_rate": 4.4333833036634496e-05, "loss": 6.4853, "step": 14374 }, { "epoch": 0.6060669941185151, "grad_norm": 4.288486480712891, "learning_rate": 4.4333052250226434e-05, "loss": 7.5847, "step": 14375 }, { "epoch": 0.6061091553007146, "grad_norm": 1.8499795198440552, "learning_rate": 4.433227146381836e-05, "loss": 6.5826, "step": 14376 }, { "epoch": 0.6061513164829142, "grad_norm": 3.3478968143463135, "learning_rate": 4.433149067741029e-05, "loss": 7.0998, "step": 14377 }, { "epoch": 0.6061934776651138, "grad_norm": 1.9568315744400024, "learning_rate": 4.433070989100222e-05, "loss": 7.0333, "step": 14378 }, { "epoch": 0.6062356388473132, "grad_norm": 3.3707292079925537, "learning_rate": 4.4329929104594146e-05, "loss": 9.1061, "step": 14379 }, { "epoch": 0.6062778000295128, "grad_norm": 3.5162367820739746, "learning_rate": 4.4329148318186084e-05, "loss": 6.468, "step": 14380 }, { "epoch": 0.6063199612117124, "grad_norm": 2.0960822105407715, "learning_rate": 4.432836753177801e-05, "loss": 6.593, "step": 14381 }, { "epoch": 0.606362122393912, "grad_norm": 2.5438170433044434, "learning_rate": 4.432758674536994e-05, "loss": 6.7473, "step": 14382 }, { "epoch": 0.6064042835761114, "grad_norm": 1.9132015705108643, "learning_rate": 4.432680595896187e-05, "loss": 6.9995, "step": 14383 }, { "epoch": 0.606446444758311, "grad_norm": 2.177061080932617, "learning_rate": 4.4326025172553795e-05, "loss": 6.6795, "step": 14384 }, { "epoch": 0.6064886059405106, "grad_norm": 2.084348440170288, "learning_rate": 4.4325244386145726e-05, "loss": 7.3738, "step": 14385 }, { "epoch": 0.6065307671227101, "grad_norm": 3.0360403060913086, "learning_rate": 4.432446359973766e-05, "loss": 6.223, "step": 14386 }, { "epoch": 0.6065729283049097, "grad_norm": 1.6041549444198608, "learning_rate": 4.432368281332959e-05, "loss": 6.599, "step": 14387 }, { "epoch": 0.6066150894871092, "grad_norm": 1.6880296468734741, "learning_rate": 4.432290202692152e-05, "loss": 6.4964, "step": 14388 }, { "epoch": 0.6066572506693088, "grad_norm": 2.3921799659729004, "learning_rate": 4.4322121240513445e-05, "loss": 6.8945, "step": 14389 }, { "epoch": 0.6066994118515083, "grad_norm": 6.948581695556641, "learning_rate": 4.4321340454105376e-05, "loss": 11.5279, "step": 14390 }, { "epoch": 0.6067415730337079, "grad_norm": 2.9354002475738525, "learning_rate": 4.432055966769731e-05, "loss": 7.4962, "step": 14391 }, { "epoch": 0.6067837342159074, "grad_norm": 3.582908868789673, "learning_rate": 4.431977888128924e-05, "loss": 7.9789, "step": 14392 }, { "epoch": 0.6068258953981069, "grad_norm": 4.0034708976745605, "learning_rate": 4.431899809488116e-05, "loss": 6.758, "step": 14393 }, { "epoch": 0.6068680565803065, "grad_norm": 3.6031110286712646, "learning_rate": 4.43182173084731e-05, "loss": 6.6893, "step": 14394 }, { "epoch": 0.6069102177625061, "grad_norm": 2.906944513320923, "learning_rate": 4.4317436522065025e-05, "loss": 6.4979, "step": 14395 }, { "epoch": 0.6069523789447057, "grad_norm": 2.8226685523986816, "learning_rate": 4.431665573565696e-05, "loss": 6.4257, "step": 14396 }, { "epoch": 0.6069945401269051, "grad_norm": 1.855062484741211, "learning_rate": 4.431587494924889e-05, "loss": 6.6818, "step": 14397 }, { "epoch": 0.6070367013091047, "grad_norm": 3.734006881713867, "learning_rate": 4.431509416284081e-05, "loss": 6.8461, "step": 14398 }, { "epoch": 0.6070788624913043, "grad_norm": 3.437420129776001, "learning_rate": 4.431431337643275e-05, "loss": 6.3543, "step": 14399 }, { "epoch": 0.6071210236735038, "grad_norm": 4.091054916381836, "learning_rate": 4.4313532590024675e-05, "loss": 6.2561, "step": 14400 }, { "epoch": 0.6071631848557033, "grad_norm": 2.1960301399230957, "learning_rate": 4.4312751803616606e-05, "loss": 6.4063, "step": 14401 }, { "epoch": 0.6072053460379029, "grad_norm": 2.1184475421905518, "learning_rate": 4.431197101720854e-05, "loss": 6.435, "step": 14402 }, { "epoch": 0.6072475072201025, "grad_norm": 2.6617696285247803, "learning_rate": 4.431119023080046e-05, "loss": 6.5311, "step": 14403 }, { "epoch": 0.607289668402302, "grad_norm": 1.9762516021728516, "learning_rate": 4.431040944439239e-05, "loss": 6.7893, "step": 14404 }, { "epoch": 0.6073318295845016, "grad_norm": 2.453964948654175, "learning_rate": 4.4309628657984324e-05, "loss": 6.5737, "step": 14405 }, { "epoch": 0.6073739907667011, "grad_norm": 2.070899724960327, "learning_rate": 4.4308847871576256e-05, "loss": 7.3413, "step": 14406 }, { "epoch": 0.6074161519489006, "grad_norm": 2.27917218208313, "learning_rate": 4.430806708516818e-05, "loss": 6.3205, "step": 14407 }, { "epoch": 0.6074583131311002, "grad_norm": 2.4715442657470703, "learning_rate": 4.430728629876011e-05, "loss": 7.8659, "step": 14408 }, { "epoch": 0.6075004743132998, "grad_norm": 4.866781234741211, "learning_rate": 4.430650551235204e-05, "loss": 8.9435, "step": 14409 }, { "epoch": 0.6075426354954993, "grad_norm": 1.798460841178894, "learning_rate": 4.4305724725943974e-05, "loss": 7.1266, "step": 14410 }, { "epoch": 0.6075847966776988, "grad_norm": 4.443849086761475, "learning_rate": 4.4304943939535905e-05, "loss": 10.0196, "step": 14411 }, { "epoch": 0.6076269578598984, "grad_norm": 2.1934587955474854, "learning_rate": 4.430416315312783e-05, "loss": 6.5458, "step": 14412 }, { "epoch": 0.607669119042098, "grad_norm": 2.489053964614868, "learning_rate": 4.430338236671977e-05, "loss": 7.6889, "step": 14413 }, { "epoch": 0.6077112802242974, "grad_norm": 3.178114891052246, "learning_rate": 4.430260158031169e-05, "loss": 6.974, "step": 14414 }, { "epoch": 0.607753441406497, "grad_norm": 3.7241857051849365, "learning_rate": 4.430182079390362e-05, "loss": 7.4582, "step": 14415 }, { "epoch": 0.6077956025886966, "grad_norm": 2.9099204540252686, "learning_rate": 4.4301040007495555e-05, "loss": 6.3485, "step": 14416 }, { "epoch": 0.6078377637708962, "grad_norm": 4.676935195922852, "learning_rate": 4.430025922108748e-05, "loss": 7.7751, "step": 14417 }, { "epoch": 0.6078799249530957, "grad_norm": 2.4017810821533203, "learning_rate": 4.429947843467941e-05, "loss": 6.6707, "step": 14418 }, { "epoch": 0.6079220861352952, "grad_norm": 2.4603755474090576, "learning_rate": 4.429869764827134e-05, "loss": 7.753, "step": 14419 }, { "epoch": 0.6079642473174948, "grad_norm": 2.132885694503784, "learning_rate": 4.429791686186327e-05, "loss": 6.6162, "step": 14420 }, { "epoch": 0.6080064084996943, "grad_norm": 2.0809502601623535, "learning_rate": 4.42971360754552e-05, "loss": 7.0005, "step": 14421 }, { "epoch": 0.6080485696818939, "grad_norm": 2.0085391998291016, "learning_rate": 4.429635528904713e-05, "loss": 6.3685, "step": 14422 }, { "epoch": 0.6080907308640934, "grad_norm": 2.440229654312134, "learning_rate": 4.429557450263906e-05, "loss": 6.846, "step": 14423 }, { "epoch": 0.608132892046293, "grad_norm": 2.6242165565490723, "learning_rate": 4.429479371623099e-05, "loss": 7.3485, "step": 14424 }, { "epoch": 0.6081750532284925, "grad_norm": 2.0466725826263428, "learning_rate": 4.429401292982292e-05, "loss": 6.8856, "step": 14425 }, { "epoch": 0.6082172144106921, "grad_norm": 2.515185594558716, "learning_rate": 4.429323214341485e-05, "loss": 6.3771, "step": 14426 }, { "epoch": 0.6082593755928917, "grad_norm": 1.7493523359298706, "learning_rate": 4.429245135700678e-05, "loss": 7.0857, "step": 14427 }, { "epoch": 0.6083015367750911, "grad_norm": 2.715397596359253, "learning_rate": 4.429167057059871e-05, "loss": 7.1533, "step": 14428 }, { "epoch": 0.6083436979572907, "grad_norm": 3.3839528560638428, "learning_rate": 4.4290889784190634e-05, "loss": 6.4175, "step": 14429 }, { "epoch": 0.6083858591394903, "grad_norm": 3.1711363792419434, "learning_rate": 4.429010899778257e-05, "loss": 7.7931, "step": 14430 }, { "epoch": 0.6084280203216899, "grad_norm": 4.11336088180542, "learning_rate": 4.4289328211374497e-05, "loss": 7.4484, "step": 14431 }, { "epoch": 0.6084701815038893, "grad_norm": 2.7129721641540527, "learning_rate": 4.428854742496643e-05, "loss": 6.3246, "step": 14432 }, { "epoch": 0.6085123426860889, "grad_norm": 3.9015634059906006, "learning_rate": 4.428776663855836e-05, "loss": 7.6839, "step": 14433 }, { "epoch": 0.6085545038682885, "grad_norm": 4.0535664558410645, "learning_rate": 4.4286985852150284e-05, "loss": 6.3151, "step": 14434 }, { "epoch": 0.608596665050488, "grad_norm": 1.8931281566619873, "learning_rate": 4.428620506574222e-05, "loss": 7.0635, "step": 14435 }, { "epoch": 0.6086388262326876, "grad_norm": 4.411761283874512, "learning_rate": 4.4285424279334146e-05, "loss": 6.4873, "step": 14436 }, { "epoch": 0.6086809874148871, "grad_norm": 2.7109789848327637, "learning_rate": 4.428464349292608e-05, "loss": 6.4225, "step": 14437 }, { "epoch": 0.6087231485970866, "grad_norm": 2.2562596797943115, "learning_rate": 4.428386270651801e-05, "loss": 6.6768, "step": 14438 }, { "epoch": 0.6087653097792862, "grad_norm": 2.9393234252929688, "learning_rate": 4.428308192010994e-05, "loss": 6.7338, "step": 14439 }, { "epoch": 0.6088074709614858, "grad_norm": 3.0619149208068848, "learning_rate": 4.4282301133701864e-05, "loss": 7.6734, "step": 14440 }, { "epoch": 0.6088496321436853, "grad_norm": 1.749031662940979, "learning_rate": 4.4281520347293796e-05, "loss": 6.3283, "step": 14441 }, { "epoch": 0.6088917933258848, "grad_norm": 3.165637969970703, "learning_rate": 4.428073956088573e-05, "loss": 7.2373, "step": 14442 }, { "epoch": 0.6089339545080844, "grad_norm": 7.04353666305542, "learning_rate": 4.427995877447765e-05, "loss": 7.7342, "step": 14443 }, { "epoch": 0.608976115690284, "grad_norm": 2.5205650329589844, "learning_rate": 4.427917798806959e-05, "loss": 7.1463, "step": 14444 }, { "epoch": 0.6090182768724836, "grad_norm": 2.778205156326294, "learning_rate": 4.4278397201661514e-05, "loss": 6.8098, "step": 14445 }, { "epoch": 0.609060438054683, "grad_norm": 3.2483606338500977, "learning_rate": 4.4277616415253445e-05, "loss": 7.2312, "step": 14446 }, { "epoch": 0.6091025992368826, "grad_norm": 2.204766035079956, "learning_rate": 4.4276835628845376e-05, "loss": 6.6944, "step": 14447 }, { "epoch": 0.6091447604190822, "grad_norm": 2.4832770824432373, "learning_rate": 4.42760548424373e-05, "loss": 7.8517, "step": 14448 }, { "epoch": 0.6091869216012817, "grad_norm": 4.311534881591797, "learning_rate": 4.427527405602924e-05, "loss": 6.878, "step": 14449 }, { "epoch": 0.6092290827834812, "grad_norm": 2.6244475841522217, "learning_rate": 4.4274493269621163e-05, "loss": 6.9411, "step": 14450 }, { "epoch": 0.6092712439656808, "grad_norm": 2.929722547531128, "learning_rate": 4.4273712483213095e-05, "loss": 7.0358, "step": 14451 }, { "epoch": 0.6093134051478803, "grad_norm": 2.0130937099456787, "learning_rate": 4.4272931696805026e-05, "loss": 6.3557, "step": 14452 }, { "epoch": 0.6093555663300799, "grad_norm": 1.9360054731369019, "learning_rate": 4.427215091039695e-05, "loss": 7.0127, "step": 14453 }, { "epoch": 0.6093977275122795, "grad_norm": 1.5879703760147095, "learning_rate": 4.427137012398888e-05, "loss": 6.7073, "step": 14454 }, { "epoch": 0.609439888694479, "grad_norm": 2.8128890991210938, "learning_rate": 4.427058933758081e-05, "loss": 6.5809, "step": 14455 }, { "epoch": 0.6094820498766785, "grad_norm": 2.4536571502685547, "learning_rate": 4.4269808551172744e-05, "loss": 6.6927, "step": 14456 }, { "epoch": 0.6095242110588781, "grad_norm": 2.2263875007629395, "learning_rate": 4.4269027764764675e-05, "loss": 6.7807, "step": 14457 }, { "epoch": 0.6095663722410777, "grad_norm": 2.1155576705932617, "learning_rate": 4.426824697835661e-05, "loss": 7.0868, "step": 14458 }, { "epoch": 0.6096085334232771, "grad_norm": 3.316866874694824, "learning_rate": 4.426746619194853e-05, "loss": 6.7852, "step": 14459 }, { "epoch": 0.6096506946054767, "grad_norm": 2.612905502319336, "learning_rate": 4.426668540554046e-05, "loss": 7.4294, "step": 14460 }, { "epoch": 0.6096928557876763, "grad_norm": 2.457742214202881, "learning_rate": 4.4265904619132394e-05, "loss": 7.2701, "step": 14461 }, { "epoch": 0.6097350169698759, "grad_norm": 2.5114259719848633, "learning_rate": 4.426512383272432e-05, "loss": 6.6874, "step": 14462 }, { "epoch": 0.6097771781520754, "grad_norm": 1.7912369966506958, "learning_rate": 4.4264343046316256e-05, "loss": 6.5823, "step": 14463 }, { "epoch": 0.6098193393342749, "grad_norm": 1.866686463356018, "learning_rate": 4.426356225990818e-05, "loss": 6.7402, "step": 14464 }, { "epoch": 0.6098615005164745, "grad_norm": 2.741024971008301, "learning_rate": 4.426278147350011e-05, "loss": 7.1029, "step": 14465 }, { "epoch": 0.609903661698674, "grad_norm": 1.6102430820465088, "learning_rate": 4.426200068709204e-05, "loss": 7.171, "step": 14466 }, { "epoch": 0.6099458228808736, "grad_norm": 3.186927556991577, "learning_rate": 4.426121990068397e-05, "loss": 6.3746, "step": 14467 }, { "epoch": 0.6099879840630731, "grad_norm": 2.1709821224212646, "learning_rate": 4.4260439114275906e-05, "loss": 6.677, "step": 14468 }, { "epoch": 0.6100301452452727, "grad_norm": 2.350003242492676, "learning_rate": 4.425965832786783e-05, "loss": 7.3368, "step": 14469 }, { "epoch": 0.6100723064274722, "grad_norm": 2.0487186908721924, "learning_rate": 4.425887754145976e-05, "loss": 6.6137, "step": 14470 }, { "epoch": 0.6101144676096718, "grad_norm": 2.0708770751953125, "learning_rate": 4.425809675505169e-05, "loss": 6.8191, "step": 14471 }, { "epoch": 0.6101566287918713, "grad_norm": 2.536283493041992, "learning_rate": 4.425731596864362e-05, "loss": 7.055, "step": 14472 }, { "epoch": 0.6101987899740708, "grad_norm": 1.5107523202896118, "learning_rate": 4.425653518223555e-05, "loss": 6.7741, "step": 14473 }, { "epoch": 0.6102409511562704, "grad_norm": 1.8258488178253174, "learning_rate": 4.425575439582748e-05, "loss": 6.7462, "step": 14474 }, { "epoch": 0.61028311233847, "grad_norm": 2.7999303340911865, "learning_rate": 4.425497360941941e-05, "loss": 6.5705, "step": 14475 }, { "epoch": 0.6103252735206696, "grad_norm": 2.181588888168335, "learning_rate": 4.4254192823011335e-05, "loss": 6.2785, "step": 14476 }, { "epoch": 0.610367434702869, "grad_norm": 2.64686918258667, "learning_rate": 4.425341203660327e-05, "loss": 7.0878, "step": 14477 }, { "epoch": 0.6104095958850686, "grad_norm": 2.3137872219085693, "learning_rate": 4.42526312501952e-05, "loss": 6.4841, "step": 14478 }, { "epoch": 0.6104517570672682, "grad_norm": 1.643322229385376, "learning_rate": 4.425185046378713e-05, "loss": 6.7087, "step": 14479 }, { "epoch": 0.6104939182494677, "grad_norm": 6.6615166664123535, "learning_rate": 4.425106967737906e-05, "loss": 10.795, "step": 14480 }, { "epoch": 0.6105360794316672, "grad_norm": 2.5003535747528076, "learning_rate": 4.4250288890970985e-05, "loss": 7.1813, "step": 14481 }, { "epoch": 0.6105782406138668, "grad_norm": 1.9486879110336304, "learning_rate": 4.424950810456292e-05, "loss": 6.598, "step": 14482 }, { "epoch": 0.6106204017960664, "grad_norm": 1.4576739072799683, "learning_rate": 4.424872731815485e-05, "loss": 6.3689, "step": 14483 }, { "epoch": 0.6106625629782659, "grad_norm": 4.148116111755371, "learning_rate": 4.424794653174678e-05, "loss": 7.7816, "step": 14484 }, { "epoch": 0.6107047241604655, "grad_norm": 3.634396553039551, "learning_rate": 4.424716574533871e-05, "loss": 7.3567, "step": 14485 }, { "epoch": 0.610746885342665, "grad_norm": 2.2762773036956787, "learning_rate": 4.4246384958930635e-05, "loss": 6.999, "step": 14486 }, { "epoch": 0.6107890465248645, "grad_norm": 2.6795384883880615, "learning_rate": 4.4245604172522566e-05, "loss": 7.5936, "step": 14487 }, { "epoch": 0.6108312077070641, "grad_norm": 3.4180994033813477, "learning_rate": 4.42448233861145e-05, "loss": 7.7602, "step": 14488 }, { "epoch": 0.6108733688892637, "grad_norm": 2.3989439010620117, "learning_rate": 4.424404259970643e-05, "loss": 7.0184, "step": 14489 }, { "epoch": 0.6109155300714632, "grad_norm": 2.031175374984741, "learning_rate": 4.424326181329836e-05, "loss": 7.1012, "step": 14490 }, { "epoch": 0.6109576912536627, "grad_norm": 2.1166675090789795, "learning_rate": 4.4242481026890284e-05, "loss": 6.4657, "step": 14491 }, { "epoch": 0.6109998524358623, "grad_norm": 3.230293035507202, "learning_rate": 4.4241700240482215e-05, "loss": 7.9114, "step": 14492 }, { "epoch": 0.6110420136180619, "grad_norm": 1.55039381980896, "learning_rate": 4.4240919454074147e-05, "loss": 6.6861, "step": 14493 }, { "epoch": 0.6110841748002614, "grad_norm": 1.9067909717559814, "learning_rate": 4.424013866766608e-05, "loss": 6.5161, "step": 14494 }, { "epoch": 0.6111263359824609, "grad_norm": 3.809453248977661, "learning_rate": 4.4239357881258e-05, "loss": 6.7903, "step": 14495 }, { "epoch": 0.6111684971646605, "grad_norm": 2.2714719772338867, "learning_rate": 4.4238577094849934e-05, "loss": 6.4972, "step": 14496 }, { "epoch": 0.6112106583468601, "grad_norm": 2.5179784297943115, "learning_rate": 4.4237796308441865e-05, "loss": 6.6727, "step": 14497 }, { "epoch": 0.6112528195290596, "grad_norm": 2.9806065559387207, "learning_rate": 4.423701552203379e-05, "loss": 6.6844, "step": 14498 }, { "epoch": 0.6112949807112591, "grad_norm": 1.2054635286331177, "learning_rate": 4.423623473562573e-05, "loss": 6.5157, "step": 14499 }, { "epoch": 0.6113371418934587, "grad_norm": 1.5130590200424194, "learning_rate": 4.423545394921765e-05, "loss": 6.4951, "step": 14500 }, { "epoch": 0.6113793030756582, "grad_norm": 2.5705244541168213, "learning_rate": 4.423467316280958e-05, "loss": 7.3878, "step": 14501 }, { "epoch": 0.6114214642578578, "grad_norm": 2.4960360527038574, "learning_rate": 4.4233892376401514e-05, "loss": 6.8303, "step": 14502 }, { "epoch": 0.6114636254400574, "grad_norm": 2.1475579738616943, "learning_rate": 4.4233111589993446e-05, "loss": 6.4308, "step": 14503 }, { "epoch": 0.6115057866222569, "grad_norm": 3.382937431335449, "learning_rate": 4.423233080358538e-05, "loss": 8.0847, "step": 14504 }, { "epoch": 0.6115479478044564, "grad_norm": 3.0582525730133057, "learning_rate": 4.42315500171773e-05, "loss": 7.4988, "step": 14505 }, { "epoch": 0.611590108986656, "grad_norm": 1.4275658130645752, "learning_rate": 4.423076923076923e-05, "loss": 6.2877, "step": 14506 }, { "epoch": 0.6116322701688556, "grad_norm": 3.5586774349212646, "learning_rate": 4.4229988444361164e-05, "loss": 6.9516, "step": 14507 }, { "epoch": 0.611674431351055, "grad_norm": 2.523793935775757, "learning_rate": 4.4229207657953095e-05, "loss": 7.3454, "step": 14508 }, { "epoch": 0.6117165925332546, "grad_norm": 2.1396052837371826, "learning_rate": 4.422842687154502e-05, "loss": 6.7826, "step": 14509 }, { "epoch": 0.6117587537154542, "grad_norm": 1.2059739828109741, "learning_rate": 4.422764608513695e-05, "loss": 6.4109, "step": 14510 }, { "epoch": 0.6118009148976538, "grad_norm": 1.027330756187439, "learning_rate": 4.422686529872888e-05, "loss": 6.7996, "step": 14511 }, { "epoch": 0.6118430760798533, "grad_norm": 1.1831080913543701, "learning_rate": 4.422608451232081e-05, "loss": 6.7161, "step": 14512 }, { "epoch": 0.6118852372620528, "grad_norm": 1.1783795356750488, "learning_rate": 4.4225303725912745e-05, "loss": 6.3891, "step": 14513 }, { "epoch": 0.6119273984442524, "grad_norm": 1.6543208360671997, "learning_rate": 4.422452293950467e-05, "loss": 6.2428, "step": 14514 }, { "epoch": 0.6119695596264519, "grad_norm": 1.2355401515960693, "learning_rate": 4.42237421530966e-05, "loss": 6.9143, "step": 14515 }, { "epoch": 0.6120117208086515, "grad_norm": 1.7725661993026733, "learning_rate": 4.422296136668853e-05, "loss": 6.3804, "step": 14516 }, { "epoch": 0.612053881990851, "grad_norm": 1.976455569267273, "learning_rate": 4.4222180580280456e-05, "loss": 6.3678, "step": 14517 }, { "epoch": 0.6120960431730506, "grad_norm": 1.7644598484039307, "learning_rate": 4.4221399793872394e-05, "loss": 7.2782, "step": 14518 }, { "epoch": 0.6121382043552501, "grad_norm": 3.629286527633667, "learning_rate": 4.422061900746432e-05, "loss": 7.887, "step": 14519 }, { "epoch": 0.6121803655374497, "grad_norm": 2.1948132514953613, "learning_rate": 4.421983822105625e-05, "loss": 6.5737, "step": 14520 }, { "epoch": 0.6122225267196493, "grad_norm": 4.934746265411377, "learning_rate": 4.421905743464818e-05, "loss": 10.7186, "step": 14521 }, { "epoch": 0.6122646879018487, "grad_norm": 2.9110138416290283, "learning_rate": 4.4218276648240106e-05, "loss": 6.8915, "step": 14522 }, { "epoch": 0.6123068490840483, "grad_norm": 3.362577438354492, "learning_rate": 4.421749586183204e-05, "loss": 7.8949, "step": 14523 }, { "epoch": 0.6123490102662479, "grad_norm": 1.776516079902649, "learning_rate": 4.421671507542397e-05, "loss": 6.9701, "step": 14524 }, { "epoch": 0.6123911714484475, "grad_norm": 1.1609834432601929, "learning_rate": 4.42159342890159e-05, "loss": 6.6545, "step": 14525 }, { "epoch": 0.6124333326306469, "grad_norm": 1.8907865285873413, "learning_rate": 4.421515350260783e-05, "loss": 7.06, "step": 14526 }, { "epoch": 0.6124754938128465, "grad_norm": 2.2207539081573486, "learning_rate": 4.421437271619976e-05, "loss": 6.2577, "step": 14527 }, { "epoch": 0.6125176549950461, "grad_norm": 1.7581919431686401, "learning_rate": 4.4213591929791686e-05, "loss": 7.0365, "step": 14528 }, { "epoch": 0.6125598161772456, "grad_norm": 1.7874764204025269, "learning_rate": 4.421281114338362e-05, "loss": 6.345, "step": 14529 }, { "epoch": 0.6126019773594451, "grad_norm": 1.9518373012542725, "learning_rate": 4.421203035697555e-05, "loss": 7.3942, "step": 14530 }, { "epoch": 0.6126441385416447, "grad_norm": 3.298889636993408, "learning_rate": 4.4211249570567473e-05, "loss": 6.6018, "step": 14531 }, { "epoch": 0.6126862997238443, "grad_norm": 2.3000638484954834, "learning_rate": 4.421046878415941e-05, "loss": 6.3671, "step": 14532 }, { "epoch": 0.6127284609060438, "grad_norm": 1.5162690877914429, "learning_rate": 4.4209687997751336e-05, "loss": 7.1365, "step": 14533 }, { "epoch": 0.6127706220882434, "grad_norm": 1.5939152240753174, "learning_rate": 4.420890721134327e-05, "loss": 6.7312, "step": 14534 }, { "epoch": 0.6128127832704429, "grad_norm": 1.5992413759231567, "learning_rate": 4.42081264249352e-05, "loss": 6.4789, "step": 14535 }, { "epoch": 0.6128549444526424, "grad_norm": 1.8310751914978027, "learning_rate": 4.420734563852712e-05, "loss": 6.3023, "step": 14536 }, { "epoch": 0.612897105634842, "grad_norm": 3.744701623916626, "learning_rate": 4.420656485211906e-05, "loss": 7.808, "step": 14537 }, { "epoch": 0.6129392668170416, "grad_norm": 3.0574817657470703, "learning_rate": 4.4205784065710986e-05, "loss": 7.3438, "step": 14538 }, { "epoch": 0.612981427999241, "grad_norm": 1.2646079063415527, "learning_rate": 4.420500327930292e-05, "loss": 6.9732, "step": 14539 }, { "epoch": 0.6130235891814406, "grad_norm": 1.203420877456665, "learning_rate": 4.420422249289485e-05, "loss": 7.0212, "step": 14540 }, { "epoch": 0.6130657503636402, "grad_norm": 2.799722909927368, "learning_rate": 4.420344170648677e-05, "loss": 6.2924, "step": 14541 }, { "epoch": 0.6131079115458398, "grad_norm": 2.6762845516204834, "learning_rate": 4.4202660920078704e-05, "loss": 6.5005, "step": 14542 }, { "epoch": 0.6131500727280393, "grad_norm": 2.873596429824829, "learning_rate": 4.4201880133670635e-05, "loss": 7.3587, "step": 14543 }, { "epoch": 0.6131922339102388, "grad_norm": 3.2313966751098633, "learning_rate": 4.4201099347262566e-05, "loss": 7.5491, "step": 14544 }, { "epoch": 0.6132343950924384, "grad_norm": 1.5948901176452637, "learning_rate": 4.420031856085449e-05, "loss": 6.5495, "step": 14545 }, { "epoch": 0.613276556274638, "grad_norm": 3.2596733570098877, "learning_rate": 4.419953777444643e-05, "loss": 7.7465, "step": 14546 }, { "epoch": 0.6133187174568375, "grad_norm": 2.2580275535583496, "learning_rate": 4.419875698803835e-05, "loss": 6.5057, "step": 14547 }, { "epoch": 0.613360878639037, "grad_norm": 2.6783175468444824, "learning_rate": 4.4197976201630285e-05, "loss": 6.4276, "step": 14548 }, { "epoch": 0.6134030398212366, "grad_norm": 1.4887720346450806, "learning_rate": 4.4197195415222216e-05, "loss": 6.5733, "step": 14549 }, { "epoch": 0.6134452010034361, "grad_norm": 1.2361714839935303, "learning_rate": 4.419641462881414e-05, "loss": 6.3526, "step": 14550 }, { "epoch": 0.6134873621856357, "grad_norm": 1.400648593902588, "learning_rate": 4.419563384240608e-05, "loss": 6.7787, "step": 14551 }, { "epoch": 0.6135295233678353, "grad_norm": 2.019428253173828, "learning_rate": 4.4194853055998e-05, "loss": 7.106, "step": 14552 }, { "epoch": 0.6135716845500347, "grad_norm": 1.8145930767059326, "learning_rate": 4.4194072269589934e-05, "loss": 6.979, "step": 14553 }, { "epoch": 0.6136138457322343, "grad_norm": 2.3013594150543213, "learning_rate": 4.4193291483181865e-05, "loss": 7.2165, "step": 14554 }, { "epoch": 0.6136560069144339, "grad_norm": 1.6377718448638916, "learning_rate": 4.419251069677379e-05, "loss": 6.9239, "step": 14555 }, { "epoch": 0.6136981680966335, "grad_norm": 1.314504623413086, "learning_rate": 4.419172991036572e-05, "loss": 7.2681, "step": 14556 }, { "epoch": 0.6137403292788329, "grad_norm": 2.8089706897735596, "learning_rate": 4.419094912395765e-05, "loss": 7.7668, "step": 14557 }, { "epoch": 0.6137824904610325, "grad_norm": 1.2961078882217407, "learning_rate": 4.4190168337549584e-05, "loss": 6.5398, "step": 14558 }, { "epoch": 0.6138246516432321, "grad_norm": 2.8289384841918945, "learning_rate": 4.4189387551141515e-05, "loss": 6.5295, "step": 14559 }, { "epoch": 0.6138668128254317, "grad_norm": 1.743035912513733, "learning_rate": 4.418860676473344e-05, "loss": 6.8258, "step": 14560 }, { "epoch": 0.6139089740076312, "grad_norm": 2.6906604766845703, "learning_rate": 4.418782597832537e-05, "loss": 7.0405, "step": 14561 }, { "epoch": 0.6139511351898307, "grad_norm": 2.8143372535705566, "learning_rate": 4.41870451919173e-05, "loss": 7.822, "step": 14562 }, { "epoch": 0.6139932963720303, "grad_norm": 2.7192325592041016, "learning_rate": 4.418626440550923e-05, "loss": 7.3385, "step": 14563 }, { "epoch": 0.6140354575542298, "grad_norm": 2.8230836391448975, "learning_rate": 4.418548361910116e-05, "loss": 7.6405, "step": 14564 }, { "epoch": 0.6140776187364294, "grad_norm": 3.10113787651062, "learning_rate": 4.4184702832693096e-05, "loss": 6.2152, "step": 14565 }, { "epoch": 0.6141197799186289, "grad_norm": 1.9515854120254517, "learning_rate": 4.418392204628502e-05, "loss": 6.7483, "step": 14566 }, { "epoch": 0.6141619411008284, "grad_norm": 3.0915822982788086, "learning_rate": 4.4183141259876945e-05, "loss": 7.521, "step": 14567 }, { "epoch": 0.614204102283028, "grad_norm": 1.7386996746063232, "learning_rate": 4.418236047346888e-05, "loss": 6.9556, "step": 14568 }, { "epoch": 0.6142462634652276, "grad_norm": 1.5258649587631226, "learning_rate": 4.418157968706081e-05, "loss": 6.7873, "step": 14569 }, { "epoch": 0.6142884246474272, "grad_norm": 4.56411600112915, "learning_rate": 4.4180798900652745e-05, "loss": 7.8862, "step": 14570 }, { "epoch": 0.6143305858296266, "grad_norm": 1.5169568061828613, "learning_rate": 4.418001811424467e-05, "loss": 6.3957, "step": 14571 }, { "epoch": 0.6143727470118262, "grad_norm": 3.334132432937622, "learning_rate": 4.41792373278366e-05, "loss": 7.3358, "step": 14572 }, { "epoch": 0.6144149081940258, "grad_norm": 4.392404079437256, "learning_rate": 4.417845654142853e-05, "loss": 7.6172, "step": 14573 }, { "epoch": 0.6144570693762254, "grad_norm": 2.389110565185547, "learning_rate": 4.417767575502046e-05, "loss": 6.5872, "step": 14574 }, { "epoch": 0.6144992305584248, "grad_norm": 1.7892495393753052, "learning_rate": 4.417689496861239e-05, "loss": 6.8351, "step": 14575 }, { "epoch": 0.6145413917406244, "grad_norm": 1.7416281700134277, "learning_rate": 4.417611418220432e-05, "loss": 6.7998, "step": 14576 }, { "epoch": 0.614583552922824, "grad_norm": 2.9373676776885986, "learning_rate": 4.417533339579625e-05, "loss": 6.2174, "step": 14577 }, { "epoch": 0.6146257141050235, "grad_norm": 2.8076252937316895, "learning_rate": 4.4174552609388175e-05, "loss": 6.7945, "step": 14578 }, { "epoch": 0.6146678752872231, "grad_norm": 2.1280484199523926, "learning_rate": 4.4173771822980106e-05, "loss": 6.8213, "step": 14579 }, { "epoch": 0.6147100364694226, "grad_norm": 2.19170880317688, "learning_rate": 4.417299103657204e-05, "loss": 6.3382, "step": 14580 }, { "epoch": 0.6147521976516221, "grad_norm": 3.4390552043914795, "learning_rate": 4.417221025016396e-05, "loss": 7.8114, "step": 14581 }, { "epoch": 0.6147943588338217, "grad_norm": 3.560744524002075, "learning_rate": 4.41714294637559e-05, "loss": 8.0167, "step": 14582 }, { "epoch": 0.6148365200160213, "grad_norm": 2.6607167720794678, "learning_rate": 4.4170648677347824e-05, "loss": 7.4108, "step": 14583 }, { "epoch": 0.6148786811982208, "grad_norm": 1.774306297302246, "learning_rate": 4.4169867890939756e-05, "loss": 6.8645, "step": 14584 }, { "epoch": 0.6149208423804203, "grad_norm": 1.0829273462295532, "learning_rate": 4.416908710453169e-05, "loss": 6.525, "step": 14585 }, { "epoch": 0.6149630035626199, "grad_norm": 3.5675129890441895, "learning_rate": 4.416830631812361e-05, "loss": 7.8253, "step": 14586 }, { "epoch": 0.6150051647448195, "grad_norm": 2.069246292114258, "learning_rate": 4.416752553171555e-05, "loss": 6.7368, "step": 14587 }, { "epoch": 0.6150473259270189, "grad_norm": 2.8631186485290527, "learning_rate": 4.4166744745307474e-05, "loss": 7.2175, "step": 14588 }, { "epoch": 0.6150894871092185, "grad_norm": 2.4457435607910156, "learning_rate": 4.4165963958899405e-05, "loss": 7.0663, "step": 14589 }, { "epoch": 0.6151316482914181, "grad_norm": 2.6841821670532227, "learning_rate": 4.4165183172491336e-05, "loss": 6.2417, "step": 14590 }, { "epoch": 0.6151738094736177, "grad_norm": 1.4144413471221924, "learning_rate": 4.416440238608327e-05, "loss": 6.4038, "step": 14591 }, { "epoch": 0.6152159706558172, "grad_norm": 1.601274847984314, "learning_rate": 4.416362159967519e-05, "loss": 6.8144, "step": 14592 }, { "epoch": 0.6152581318380167, "grad_norm": 1.6316014528274536, "learning_rate": 4.4162840813267123e-05, "loss": 6.4851, "step": 14593 }, { "epoch": 0.6153002930202163, "grad_norm": 1.805262804031372, "learning_rate": 4.4162060026859055e-05, "loss": 6.6018, "step": 14594 }, { "epoch": 0.6153424542024158, "grad_norm": 1.196092963218689, "learning_rate": 4.4161279240450986e-05, "loss": 6.647, "step": 14595 }, { "epoch": 0.6153846153846154, "grad_norm": 1.2949343919754028, "learning_rate": 4.416049845404292e-05, "loss": 6.5084, "step": 14596 }, { "epoch": 0.6154267765668149, "grad_norm": 1.2564321756362915, "learning_rate": 4.415971766763484e-05, "loss": 6.3586, "step": 14597 }, { "epoch": 0.6154689377490145, "grad_norm": 2.183652400970459, "learning_rate": 4.415893688122677e-05, "loss": 6.95, "step": 14598 }, { "epoch": 0.615511098931214, "grad_norm": 2.576082468032837, "learning_rate": 4.4158156094818704e-05, "loss": 6.7089, "step": 14599 }, { "epoch": 0.6155532601134136, "grad_norm": 1.5122528076171875, "learning_rate": 4.415737530841063e-05, "loss": 6.2868, "step": 14600 }, { "epoch": 0.6155954212956132, "grad_norm": 2.1786563396453857, "learning_rate": 4.415659452200257e-05, "loss": 6.786, "step": 14601 }, { "epoch": 0.6156375824778126, "grad_norm": 1.8686250448226929, "learning_rate": 4.415581373559449e-05, "loss": 6.3604, "step": 14602 }, { "epoch": 0.6156797436600122, "grad_norm": 1.8312641382217407, "learning_rate": 4.415503294918642e-05, "loss": 6.7856, "step": 14603 }, { "epoch": 0.6157219048422118, "grad_norm": 3.497271776199341, "learning_rate": 4.4154252162778354e-05, "loss": 7.1282, "step": 14604 }, { "epoch": 0.6157640660244114, "grad_norm": 1.5694019794464111, "learning_rate": 4.415347137637028e-05, "loss": 6.7501, "step": 14605 }, { "epoch": 0.6158062272066108, "grad_norm": 3.094151258468628, "learning_rate": 4.4152690589962216e-05, "loss": 6.9825, "step": 14606 }, { "epoch": 0.6158483883888104, "grad_norm": 1.1803710460662842, "learning_rate": 4.415190980355414e-05, "loss": 6.439, "step": 14607 }, { "epoch": 0.61589054957101, "grad_norm": 2.3547677993774414, "learning_rate": 4.415112901714607e-05, "loss": 7.3893, "step": 14608 }, { "epoch": 0.6159327107532095, "grad_norm": 1.7947113513946533, "learning_rate": 4.4150348230738e-05, "loss": 7.0825, "step": 14609 }, { "epoch": 0.6159748719354091, "grad_norm": 2.826143741607666, "learning_rate": 4.4149567444329935e-05, "loss": 7.1829, "step": 14610 }, { "epoch": 0.6160170331176086, "grad_norm": 2.402226686477661, "learning_rate": 4.414878665792186e-05, "loss": 6.98, "step": 14611 }, { "epoch": 0.6160591942998082, "grad_norm": 3.139407157897949, "learning_rate": 4.414800587151379e-05, "loss": 6.8111, "step": 14612 }, { "epoch": 0.6161013554820077, "grad_norm": 2.7721610069274902, "learning_rate": 4.414722508510572e-05, "loss": 6.3987, "step": 14613 }, { "epoch": 0.6161435166642073, "grad_norm": 2.6178882122039795, "learning_rate": 4.4146444298697646e-05, "loss": 7.4784, "step": 14614 }, { "epoch": 0.6161856778464068, "grad_norm": 3.359980821609497, "learning_rate": 4.4145663512289584e-05, "loss": 7.2673, "step": 14615 }, { "epoch": 0.6162278390286063, "grad_norm": 3.5751936435699463, "learning_rate": 4.414488272588151e-05, "loss": 7.2315, "step": 14616 }, { "epoch": 0.6162700002108059, "grad_norm": 3.787745952606201, "learning_rate": 4.414410193947344e-05, "loss": 6.2937, "step": 14617 }, { "epoch": 0.6163121613930055, "grad_norm": 3.0243020057678223, "learning_rate": 4.414332115306537e-05, "loss": 6.9669, "step": 14618 }, { "epoch": 0.6163543225752051, "grad_norm": 1.6181191205978394, "learning_rate": 4.4142540366657296e-05, "loss": 6.5177, "step": 14619 }, { "epoch": 0.6163964837574045, "grad_norm": 2.8391427993774414, "learning_rate": 4.4141759580249234e-05, "loss": 7.4688, "step": 14620 }, { "epoch": 0.6164386449396041, "grad_norm": 2.301659345626831, "learning_rate": 4.414097879384116e-05, "loss": 6.414, "step": 14621 }, { "epoch": 0.6164808061218037, "grad_norm": 1.996127724647522, "learning_rate": 4.414019800743309e-05, "loss": 6.9623, "step": 14622 }, { "epoch": 0.6165229673040032, "grad_norm": 1.4858368635177612, "learning_rate": 4.413941722102502e-05, "loss": 6.802, "step": 14623 }, { "epoch": 0.6165651284862027, "grad_norm": 2.6317710876464844, "learning_rate": 4.4138636434616945e-05, "loss": 7.6016, "step": 14624 }, { "epoch": 0.6166072896684023, "grad_norm": 2.022381544113159, "learning_rate": 4.4137855648208876e-05, "loss": 6.4726, "step": 14625 }, { "epoch": 0.6166494508506019, "grad_norm": 1.9379607439041138, "learning_rate": 4.413707486180081e-05, "loss": 7.3749, "step": 14626 }, { "epoch": 0.6166916120328014, "grad_norm": 2.157050609588623, "learning_rate": 4.413629407539274e-05, "loss": 7.4393, "step": 14627 }, { "epoch": 0.616733773215001, "grad_norm": 2.3828213214874268, "learning_rate": 4.413551328898467e-05, "loss": 7.507, "step": 14628 }, { "epoch": 0.6167759343972005, "grad_norm": 2.1497087478637695, "learning_rate": 4.4134732502576595e-05, "loss": 8.2783, "step": 14629 }, { "epoch": 0.6168180955794, "grad_norm": 2.9960062503814697, "learning_rate": 4.4133951716168526e-05, "loss": 7.2895, "step": 14630 }, { "epoch": 0.6168602567615996, "grad_norm": 2.903599262237549, "learning_rate": 4.413317092976046e-05, "loss": 6.7085, "step": 14631 }, { "epoch": 0.6169024179437992, "grad_norm": 3.0574686527252197, "learning_rate": 4.413239014335239e-05, "loss": 7.1315, "step": 14632 }, { "epoch": 0.6169445791259986, "grad_norm": 2.7274937629699707, "learning_rate": 4.413160935694431e-05, "loss": 6.5005, "step": 14633 }, { "epoch": 0.6169867403081982, "grad_norm": 3.708065986633301, "learning_rate": 4.413082857053625e-05, "loss": 7.3044, "step": 14634 }, { "epoch": 0.6170289014903978, "grad_norm": 1.551579236984253, "learning_rate": 4.4130047784128175e-05, "loss": 6.6612, "step": 14635 }, { "epoch": 0.6170710626725974, "grad_norm": 2.610623836517334, "learning_rate": 4.412926699772011e-05, "loss": 6.4116, "step": 14636 }, { "epoch": 0.6171132238547969, "grad_norm": 1.7720590829849243, "learning_rate": 4.412848621131204e-05, "loss": 6.634, "step": 14637 }, { "epoch": 0.6171553850369964, "grad_norm": 1.243456482887268, "learning_rate": 4.412770542490396e-05, "loss": 6.9784, "step": 14638 }, { "epoch": 0.617197546219196, "grad_norm": 1.0093914270401, "learning_rate": 4.41269246384959e-05, "loss": 6.3935, "step": 14639 }, { "epoch": 0.6172397074013956, "grad_norm": 1.1822363138198853, "learning_rate": 4.4126143852087825e-05, "loss": 6.6882, "step": 14640 }, { "epoch": 0.6172818685835951, "grad_norm": 1.22621488571167, "learning_rate": 4.4125363065679756e-05, "loss": 6.5323, "step": 14641 }, { "epoch": 0.6173240297657946, "grad_norm": 1.9111822843551636, "learning_rate": 4.412458227927169e-05, "loss": 6.9668, "step": 14642 }, { "epoch": 0.6173661909479942, "grad_norm": 1.403119683265686, "learning_rate": 4.412380149286361e-05, "loss": 6.8868, "step": 14643 }, { "epoch": 0.6174083521301937, "grad_norm": 1.4158140420913696, "learning_rate": 4.412302070645554e-05, "loss": 6.397, "step": 14644 }, { "epoch": 0.6174505133123933, "grad_norm": 3.3216869831085205, "learning_rate": 4.4122239920047474e-05, "loss": 7.4722, "step": 14645 }, { "epoch": 0.6174926744945928, "grad_norm": 1.4351078271865845, "learning_rate": 4.4121459133639406e-05, "loss": 6.7739, "step": 14646 }, { "epoch": 0.6175348356767923, "grad_norm": 2.134237051010132, "learning_rate": 4.412067834723133e-05, "loss": 6.3689, "step": 14647 }, { "epoch": 0.6175769968589919, "grad_norm": 2.8151209354400635, "learning_rate": 4.411989756082326e-05, "loss": 6.5679, "step": 14648 }, { "epoch": 0.6176191580411915, "grad_norm": 2.200016975402832, "learning_rate": 4.411911677441519e-05, "loss": 6.93, "step": 14649 }, { "epoch": 0.6176613192233911, "grad_norm": 2.0216562747955322, "learning_rate": 4.4118335988007124e-05, "loss": 7.1041, "step": 14650 }, { "epoch": 0.6177034804055905, "grad_norm": 2.613142490386963, "learning_rate": 4.4117555201599055e-05, "loss": 6.4628, "step": 14651 }, { "epoch": 0.6177456415877901, "grad_norm": 2.144252061843872, "learning_rate": 4.411677441519098e-05, "loss": 7.1448, "step": 14652 }, { "epoch": 0.6177878027699897, "grad_norm": 2.6656928062438965, "learning_rate": 4.411599362878292e-05, "loss": 8.1526, "step": 14653 }, { "epoch": 0.6178299639521893, "grad_norm": 3.0152857303619385, "learning_rate": 4.411521284237484e-05, "loss": 7.89, "step": 14654 }, { "epoch": 0.6178721251343887, "grad_norm": 1.4612468481063843, "learning_rate": 4.411443205596677e-05, "loss": 7.1216, "step": 14655 }, { "epoch": 0.6179142863165883, "grad_norm": 2.8222029209136963, "learning_rate": 4.4113651269558705e-05, "loss": 7.6936, "step": 14656 }, { "epoch": 0.6179564474987879, "grad_norm": 2.7648532390594482, "learning_rate": 4.411287048315063e-05, "loss": 6.4639, "step": 14657 }, { "epoch": 0.6179986086809874, "grad_norm": 4.1264262199401855, "learning_rate": 4.411208969674256e-05, "loss": 7.4718, "step": 14658 }, { "epoch": 0.618040769863187, "grad_norm": 2.3920974731445312, "learning_rate": 4.411130891033449e-05, "loss": 6.6463, "step": 14659 }, { "epoch": 0.6180829310453865, "grad_norm": 3.023495674133301, "learning_rate": 4.411052812392642e-05, "loss": 7.4474, "step": 14660 }, { "epoch": 0.618125092227586, "grad_norm": 3.102322578430176, "learning_rate": 4.410974733751835e-05, "loss": 6.2563, "step": 14661 }, { "epoch": 0.6181672534097856, "grad_norm": 5.230820655822754, "learning_rate": 4.410896655111028e-05, "loss": 8.3334, "step": 14662 }, { "epoch": 0.6182094145919852, "grad_norm": 1.984695315361023, "learning_rate": 4.410818576470221e-05, "loss": 6.7474, "step": 14663 }, { "epoch": 0.6182515757741847, "grad_norm": 1.856621503829956, "learning_rate": 4.410740497829414e-05, "loss": 6.3337, "step": 14664 }, { "epoch": 0.6182937369563842, "grad_norm": 1.5084309577941895, "learning_rate": 4.410662419188607e-05, "loss": 6.4737, "step": 14665 }, { "epoch": 0.6183358981385838, "grad_norm": 2.6744720935821533, "learning_rate": 4.4105843405478e-05, "loss": 6.202, "step": 14666 }, { "epoch": 0.6183780593207834, "grad_norm": 1.4506663084030151, "learning_rate": 4.410506261906993e-05, "loss": 6.2936, "step": 14667 }, { "epoch": 0.618420220502983, "grad_norm": 2.289092540740967, "learning_rate": 4.410428183266186e-05, "loss": 6.9981, "step": 14668 }, { "epoch": 0.6184623816851824, "grad_norm": 1.9245160818099976, "learning_rate": 4.4103501046253784e-05, "loss": 7.0049, "step": 14669 }, { "epoch": 0.618504542867382, "grad_norm": 1.7769132852554321, "learning_rate": 4.410272025984572e-05, "loss": 7.0721, "step": 14670 }, { "epoch": 0.6185467040495816, "grad_norm": 2.449777841567993, "learning_rate": 4.4101939473437647e-05, "loss": 7.0663, "step": 14671 }, { "epoch": 0.6185888652317811, "grad_norm": 1.9027528762817383, "learning_rate": 4.410115868702958e-05, "loss": 6.9141, "step": 14672 }, { "epoch": 0.6186310264139806, "grad_norm": 1.7520866394042969, "learning_rate": 4.410037790062151e-05, "loss": 6.3925, "step": 14673 }, { "epoch": 0.6186731875961802, "grad_norm": 3.0696308612823486, "learning_rate": 4.4099597114213434e-05, "loss": 7.6228, "step": 14674 }, { "epoch": 0.6187153487783797, "grad_norm": 1.957228183746338, "learning_rate": 4.409881632780537e-05, "loss": 7.5185, "step": 14675 }, { "epoch": 0.6187575099605793, "grad_norm": 1.9069358110427856, "learning_rate": 4.4098035541397296e-05, "loss": 7.0906, "step": 14676 }, { "epoch": 0.6187996711427789, "grad_norm": 1.8366113901138306, "learning_rate": 4.409725475498923e-05, "loss": 7.0924, "step": 14677 }, { "epoch": 0.6188418323249784, "grad_norm": 1.7758415937423706, "learning_rate": 4.409647396858116e-05, "loss": 6.5022, "step": 14678 }, { "epoch": 0.6188839935071779, "grad_norm": 3.27341628074646, "learning_rate": 4.409569318217309e-05, "loss": 7.0689, "step": 14679 }, { "epoch": 0.6189261546893775, "grad_norm": 2.301439046859741, "learning_rate": 4.4094912395765014e-05, "loss": 6.6172, "step": 14680 }, { "epoch": 0.6189683158715771, "grad_norm": 1.827834963798523, "learning_rate": 4.4094131609356946e-05, "loss": 6.3972, "step": 14681 }, { "epoch": 0.6190104770537765, "grad_norm": 2.2514584064483643, "learning_rate": 4.409335082294888e-05, "loss": 6.9302, "step": 14682 }, { "epoch": 0.6190526382359761, "grad_norm": 4.184304714202881, "learning_rate": 4.40925700365408e-05, "loss": 7.2395, "step": 14683 }, { "epoch": 0.6190947994181757, "grad_norm": 1.3267379999160767, "learning_rate": 4.409178925013274e-05, "loss": 6.5673, "step": 14684 }, { "epoch": 0.6191369606003753, "grad_norm": 1.6150431632995605, "learning_rate": 4.4091008463724664e-05, "loss": 6.616, "step": 14685 }, { "epoch": 0.6191791217825748, "grad_norm": 1.4265565872192383, "learning_rate": 4.4090227677316595e-05, "loss": 6.3352, "step": 14686 }, { "epoch": 0.6192212829647743, "grad_norm": 2.7413294315338135, "learning_rate": 4.4089446890908526e-05, "loss": 7.8103, "step": 14687 }, { "epoch": 0.6192634441469739, "grad_norm": 1.270912766456604, "learning_rate": 4.408866610450045e-05, "loss": 6.3468, "step": 14688 }, { "epoch": 0.6193056053291734, "grad_norm": 1.7336629629135132, "learning_rate": 4.408788531809239e-05, "loss": 7.0017, "step": 14689 }, { "epoch": 0.619347766511373, "grad_norm": 3.3289825916290283, "learning_rate": 4.4087104531684313e-05, "loss": 7.2222, "step": 14690 }, { "epoch": 0.6193899276935725, "grad_norm": 2.563450813293457, "learning_rate": 4.4086323745276245e-05, "loss": 6.8112, "step": 14691 }, { "epoch": 0.6194320888757721, "grad_norm": 3.262242078781128, "learning_rate": 4.4085542958868176e-05, "loss": 7.2698, "step": 14692 }, { "epoch": 0.6194742500579716, "grad_norm": 2.1081254482269287, "learning_rate": 4.40847621724601e-05, "loss": 7.046, "step": 14693 }, { "epoch": 0.6195164112401712, "grad_norm": 2.5051631927490234, "learning_rate": 4.408398138605203e-05, "loss": 6.4031, "step": 14694 }, { "epoch": 0.6195585724223707, "grad_norm": 2.1411561965942383, "learning_rate": 4.408320059964396e-05, "loss": 7.0571, "step": 14695 }, { "epoch": 0.6196007336045702, "grad_norm": 2.0863256454467773, "learning_rate": 4.4082419813235894e-05, "loss": 7.2282, "step": 14696 }, { "epoch": 0.6196428947867698, "grad_norm": 1.7972302436828613, "learning_rate": 4.4081639026827825e-05, "loss": 6.6632, "step": 14697 }, { "epoch": 0.6196850559689694, "grad_norm": 2.453711748123169, "learning_rate": 4.408085824041976e-05, "loss": 7.602, "step": 14698 }, { "epoch": 0.619727217151169, "grad_norm": 2.1357948780059814, "learning_rate": 4.408007745401168e-05, "loss": 6.331, "step": 14699 }, { "epoch": 0.6197693783333684, "grad_norm": 1.5441476106643677, "learning_rate": 4.407929666760361e-05, "loss": 6.4235, "step": 14700 }, { "epoch": 0.619811539515568, "grad_norm": 1.3054711818695068, "learning_rate": 4.4078515881195544e-05, "loss": 6.3116, "step": 14701 }, { "epoch": 0.6198537006977676, "grad_norm": 1.6497666835784912, "learning_rate": 4.407773509478747e-05, "loss": 6.5187, "step": 14702 }, { "epoch": 0.6198958618799671, "grad_norm": 1.7809218168258667, "learning_rate": 4.4076954308379406e-05, "loss": 6.9916, "step": 14703 }, { "epoch": 0.6199380230621666, "grad_norm": 1.3958088159561157, "learning_rate": 4.407617352197133e-05, "loss": 6.6409, "step": 14704 }, { "epoch": 0.6199801842443662, "grad_norm": 1.3129081726074219, "learning_rate": 4.407539273556326e-05, "loss": 6.3516, "step": 14705 }, { "epoch": 0.6200223454265658, "grad_norm": 2.184102773666382, "learning_rate": 4.407461194915519e-05, "loss": 7.5917, "step": 14706 }, { "epoch": 0.6200645066087653, "grad_norm": 1.7954095602035522, "learning_rate": 4.407383116274712e-05, "loss": 6.9304, "step": 14707 }, { "epoch": 0.6201066677909649, "grad_norm": 1.0830525159835815, "learning_rate": 4.4073050376339056e-05, "loss": 6.4306, "step": 14708 }, { "epoch": 0.6201488289731644, "grad_norm": 3.67193341255188, "learning_rate": 4.407226958993098e-05, "loss": 8.5708, "step": 14709 }, { "epoch": 0.6201909901553639, "grad_norm": 2.6456332206726074, "learning_rate": 4.407148880352291e-05, "loss": 7.5573, "step": 14710 }, { "epoch": 0.6202331513375635, "grad_norm": 3.0739388465881348, "learning_rate": 4.407070801711484e-05, "loss": 7.7088, "step": 14711 }, { "epoch": 0.6202753125197631, "grad_norm": 1.6064187288284302, "learning_rate": 4.406992723070677e-05, "loss": 6.4666, "step": 14712 }, { "epoch": 0.6203174737019626, "grad_norm": 1.0569576025009155, "learning_rate": 4.40691464442987e-05, "loss": 6.5506, "step": 14713 }, { "epoch": 0.6203596348841621, "grad_norm": 3.1555471420288086, "learning_rate": 4.406836565789063e-05, "loss": 7.7757, "step": 14714 }, { "epoch": 0.6204017960663617, "grad_norm": 1.0744608640670776, "learning_rate": 4.406758487148256e-05, "loss": 6.6643, "step": 14715 }, { "epoch": 0.6204439572485613, "grad_norm": 1.8236771821975708, "learning_rate": 4.4066804085074485e-05, "loss": 6.5533, "step": 14716 }, { "epoch": 0.6204861184307608, "grad_norm": 1.5669593811035156, "learning_rate": 4.406602329866642e-05, "loss": 6.3997, "step": 14717 }, { "epoch": 0.6205282796129603, "grad_norm": 1.8541209697723389, "learning_rate": 4.406524251225835e-05, "loss": 7.0086, "step": 14718 }, { "epoch": 0.6205704407951599, "grad_norm": 1.21272873878479, "learning_rate": 4.406446172585028e-05, "loss": 6.6787, "step": 14719 }, { "epoch": 0.6206126019773595, "grad_norm": 3.086760997772217, "learning_rate": 4.406368093944221e-05, "loss": 6.4406, "step": 14720 }, { "epoch": 0.620654763159559, "grad_norm": 1.5425918102264404, "learning_rate": 4.4062900153034135e-05, "loss": 6.6578, "step": 14721 }, { "epoch": 0.6206969243417585, "grad_norm": 2.8022775650024414, "learning_rate": 4.406211936662607e-05, "loss": 6.6256, "step": 14722 }, { "epoch": 0.6207390855239581, "grad_norm": 1.0673000812530518, "learning_rate": 4.4061338580218e-05, "loss": 6.8882, "step": 14723 }, { "epoch": 0.6207812467061576, "grad_norm": 1.314584493637085, "learning_rate": 4.406055779380993e-05, "loss": 6.5575, "step": 14724 }, { "epoch": 0.6208234078883572, "grad_norm": 3.2140040397644043, "learning_rate": 4.405977700740186e-05, "loss": 7.3101, "step": 14725 }, { "epoch": 0.6208655690705568, "grad_norm": 1.3064782619476318, "learning_rate": 4.4058996220993785e-05, "loss": 6.6422, "step": 14726 }, { "epoch": 0.6209077302527563, "grad_norm": 2.817082643508911, "learning_rate": 4.4058215434585716e-05, "loss": 7.7587, "step": 14727 }, { "epoch": 0.6209498914349558, "grad_norm": 1.700236439704895, "learning_rate": 4.405743464817765e-05, "loss": 6.9567, "step": 14728 }, { "epoch": 0.6209920526171554, "grad_norm": 2.469801664352417, "learning_rate": 4.405665386176958e-05, "loss": 7.7134, "step": 14729 }, { "epoch": 0.621034213799355, "grad_norm": 2.475825548171997, "learning_rate": 4.405587307536151e-05, "loss": 6.5056, "step": 14730 }, { "epoch": 0.6210763749815544, "grad_norm": 3.1934123039245605, "learning_rate": 4.4055092288953434e-05, "loss": 7.8003, "step": 14731 }, { "epoch": 0.621118536163754, "grad_norm": 1.4691678285598755, "learning_rate": 4.4054311502545365e-05, "loss": 6.9612, "step": 14732 }, { "epoch": 0.6211606973459536, "grad_norm": 1.6015609502792358, "learning_rate": 4.4053530716137297e-05, "loss": 6.5506, "step": 14733 }, { "epoch": 0.6212028585281532, "grad_norm": 1.339146375656128, "learning_rate": 4.405274992972923e-05, "loss": 6.9585, "step": 14734 }, { "epoch": 0.6212450197103527, "grad_norm": 2.3878180980682373, "learning_rate": 4.405196914332115e-05, "loss": 7.2819, "step": 14735 }, { "epoch": 0.6212871808925522, "grad_norm": 1.4657412767410278, "learning_rate": 4.4051188356913084e-05, "loss": 6.4245, "step": 14736 }, { "epoch": 0.6213293420747518, "grad_norm": 1.9707063436508179, "learning_rate": 4.4050407570505015e-05, "loss": 7.0996, "step": 14737 }, { "epoch": 0.6213715032569513, "grad_norm": 1.201013207435608, "learning_rate": 4.404962678409694e-05, "loss": 6.6594, "step": 14738 }, { "epoch": 0.6214136644391509, "grad_norm": 1.6700570583343506, "learning_rate": 4.404884599768888e-05, "loss": 7.1406, "step": 14739 }, { "epoch": 0.6214558256213504, "grad_norm": 2.128984212875366, "learning_rate": 4.40480652112808e-05, "loss": 6.7803, "step": 14740 }, { "epoch": 0.62149798680355, "grad_norm": 4.452347755432129, "learning_rate": 4.404728442487273e-05, "loss": 8.8913, "step": 14741 }, { "epoch": 0.6215401479857495, "grad_norm": 1.7026416063308716, "learning_rate": 4.4046503638464664e-05, "loss": 6.3724, "step": 14742 }, { "epoch": 0.6215823091679491, "grad_norm": 1.2447483539581299, "learning_rate": 4.4045722852056596e-05, "loss": 6.438, "step": 14743 }, { "epoch": 0.6216244703501487, "grad_norm": 1.0424505472183228, "learning_rate": 4.404494206564853e-05, "loss": 6.5132, "step": 14744 }, { "epoch": 0.6216666315323481, "grad_norm": 3.9988291263580322, "learning_rate": 4.404416127924045e-05, "loss": 7.8124, "step": 14745 }, { "epoch": 0.6217087927145477, "grad_norm": 4.145760536193848, "learning_rate": 4.404338049283238e-05, "loss": 6.9451, "step": 14746 }, { "epoch": 0.6217509538967473, "grad_norm": 1.6666617393493652, "learning_rate": 4.4042599706424314e-05, "loss": 6.3718, "step": 14747 }, { "epoch": 0.6217931150789469, "grad_norm": 1.1732574701309204, "learning_rate": 4.4041818920016245e-05, "loss": 7.313, "step": 14748 }, { "epoch": 0.6218352762611463, "grad_norm": 1.9916924238204956, "learning_rate": 4.404103813360817e-05, "loss": 6.4406, "step": 14749 }, { "epoch": 0.6218774374433459, "grad_norm": 1.5198798179626465, "learning_rate": 4.40402573472001e-05, "loss": 6.5449, "step": 14750 }, { "epoch": 0.6219195986255455, "grad_norm": 1.5497097969055176, "learning_rate": 4.403947656079203e-05, "loss": 6.5446, "step": 14751 }, { "epoch": 0.621961759807745, "grad_norm": 1.1918338537216187, "learning_rate": 4.403869577438396e-05, "loss": 6.5214, "step": 14752 }, { "epoch": 0.6220039209899445, "grad_norm": 1.781853199005127, "learning_rate": 4.4037914987975895e-05, "loss": 6.5077, "step": 14753 }, { "epoch": 0.6220460821721441, "grad_norm": 1.0370705127716064, "learning_rate": 4.403713420156782e-05, "loss": 6.5069, "step": 14754 }, { "epoch": 0.6220882433543437, "grad_norm": 2.8425817489624023, "learning_rate": 4.403635341515975e-05, "loss": 7.0179, "step": 14755 }, { "epoch": 0.6221304045365432, "grad_norm": 1.291132926940918, "learning_rate": 4.403557262875168e-05, "loss": 6.4154, "step": 14756 }, { "epoch": 0.6221725657187428, "grad_norm": 1.4406094551086426, "learning_rate": 4.4034791842343606e-05, "loss": 6.8168, "step": 14757 }, { "epoch": 0.6222147269009423, "grad_norm": 1.1714956760406494, "learning_rate": 4.4034011055935544e-05, "loss": 6.673, "step": 14758 }, { "epoch": 0.6222568880831418, "grad_norm": 2.7442586421966553, "learning_rate": 4.403323026952747e-05, "loss": 7.984, "step": 14759 }, { "epoch": 0.6222990492653414, "grad_norm": 1.4755514860153198, "learning_rate": 4.40324494831194e-05, "loss": 6.2801, "step": 14760 }, { "epoch": 0.622341210447541, "grad_norm": 1.7608531713485718, "learning_rate": 4.403166869671133e-05, "loss": 6.2853, "step": 14761 }, { "epoch": 0.6223833716297404, "grad_norm": 2.8216676712036133, "learning_rate": 4.4030887910303256e-05, "loss": 7.737, "step": 14762 }, { "epoch": 0.62242553281194, "grad_norm": 1.10542631149292, "learning_rate": 4.403010712389519e-05, "loss": 6.4505, "step": 14763 }, { "epoch": 0.6224676939941396, "grad_norm": 2.443789005279541, "learning_rate": 4.402932633748712e-05, "loss": 7.371, "step": 14764 }, { "epoch": 0.6225098551763392, "grad_norm": 2.3113036155700684, "learning_rate": 4.402854555107905e-05, "loss": 7.0115, "step": 14765 }, { "epoch": 0.6225520163585387, "grad_norm": 1.4976437091827393, "learning_rate": 4.402776476467098e-05, "loss": 6.737, "step": 14766 }, { "epoch": 0.6225941775407382, "grad_norm": 1.9138273000717163, "learning_rate": 4.402698397826291e-05, "loss": 7.1172, "step": 14767 }, { "epoch": 0.6226363387229378, "grad_norm": 3.6536078453063965, "learning_rate": 4.4026203191854836e-05, "loss": 7.8662, "step": 14768 }, { "epoch": 0.6226784999051374, "grad_norm": 3.560483455657959, "learning_rate": 4.402542240544677e-05, "loss": 7.5258, "step": 14769 }, { "epoch": 0.6227206610873369, "grad_norm": 2.0579142570495605, "learning_rate": 4.40246416190387e-05, "loss": 6.3651, "step": 14770 }, { "epoch": 0.6227628222695364, "grad_norm": 3.0545244216918945, "learning_rate": 4.4023860832630623e-05, "loss": 6.5368, "step": 14771 }, { "epoch": 0.622804983451736, "grad_norm": 2.0414886474609375, "learning_rate": 4.402308004622256e-05, "loss": 7.5295, "step": 14772 }, { "epoch": 0.6228471446339355, "grad_norm": 2.961219072341919, "learning_rate": 4.4022299259814486e-05, "loss": 7.9195, "step": 14773 }, { "epoch": 0.6228893058161351, "grad_norm": 1.864362120628357, "learning_rate": 4.402151847340642e-05, "loss": 6.9021, "step": 14774 }, { "epoch": 0.6229314669983347, "grad_norm": 2.154705047607422, "learning_rate": 4.402073768699835e-05, "loss": 6.8867, "step": 14775 }, { "epoch": 0.6229736281805341, "grad_norm": 2.5087473392486572, "learning_rate": 4.401995690059027e-05, "loss": 6.3356, "step": 14776 }, { "epoch": 0.6230157893627337, "grad_norm": 1.553270697593689, "learning_rate": 4.401917611418221e-05, "loss": 6.5188, "step": 14777 }, { "epoch": 0.6230579505449333, "grad_norm": 1.297860026359558, "learning_rate": 4.4018395327774136e-05, "loss": 6.7676, "step": 14778 }, { "epoch": 0.6231001117271329, "grad_norm": 2.4366915225982666, "learning_rate": 4.401761454136607e-05, "loss": 7.4536, "step": 14779 }, { "epoch": 0.6231422729093323, "grad_norm": 2.7916581630706787, "learning_rate": 4.4016833754958e-05, "loss": 6.4491, "step": 14780 }, { "epoch": 0.6231844340915319, "grad_norm": 3.0172882080078125, "learning_rate": 4.401605296854992e-05, "loss": 6.52, "step": 14781 }, { "epoch": 0.6232265952737315, "grad_norm": 2.0588197708129883, "learning_rate": 4.4015272182141854e-05, "loss": 6.8407, "step": 14782 }, { "epoch": 0.623268756455931, "grad_norm": 1.3535228967666626, "learning_rate": 4.4014491395733785e-05, "loss": 6.5817, "step": 14783 }, { "epoch": 0.6233109176381306, "grad_norm": 3.3807687759399414, "learning_rate": 4.4013710609325716e-05, "loss": 6.3697, "step": 14784 }, { "epoch": 0.6233530788203301, "grad_norm": 1.5462559461593628, "learning_rate": 4.401292982291764e-05, "loss": 6.5409, "step": 14785 }, { "epoch": 0.6233952400025297, "grad_norm": 2.651466131210327, "learning_rate": 4.401214903650958e-05, "loss": 6.4742, "step": 14786 }, { "epoch": 0.6234374011847292, "grad_norm": 0.845417857170105, "learning_rate": 4.40113682501015e-05, "loss": 6.4837, "step": 14787 }, { "epoch": 0.6234795623669288, "grad_norm": 2.3568708896636963, "learning_rate": 4.4010587463693435e-05, "loss": 6.9572, "step": 14788 }, { "epoch": 0.6235217235491283, "grad_norm": 1.8150227069854736, "learning_rate": 4.4009806677285366e-05, "loss": 7.6041, "step": 14789 }, { "epoch": 0.6235638847313278, "grad_norm": 2.3454091548919678, "learning_rate": 4.400902589087729e-05, "loss": 7.1396, "step": 14790 }, { "epoch": 0.6236060459135274, "grad_norm": 2.904860734939575, "learning_rate": 4.400824510446923e-05, "loss": 6.8492, "step": 14791 }, { "epoch": 0.623648207095727, "grad_norm": 2.3231260776519775, "learning_rate": 4.400746431806115e-05, "loss": 7.231, "step": 14792 }, { "epoch": 0.6236903682779266, "grad_norm": 1.2096996307373047, "learning_rate": 4.4006683531653084e-05, "loss": 6.9164, "step": 14793 }, { "epoch": 0.623732529460126, "grad_norm": 2.110059976577759, "learning_rate": 4.4005902745245015e-05, "loss": 6.768, "step": 14794 }, { "epoch": 0.6237746906423256, "grad_norm": 4.783871173858643, "learning_rate": 4.400512195883694e-05, "loss": 7.9162, "step": 14795 }, { "epoch": 0.6238168518245252, "grad_norm": 1.5147488117218018, "learning_rate": 4.400434117242887e-05, "loss": 6.5631, "step": 14796 }, { "epoch": 0.6238590130067247, "grad_norm": 1.267677903175354, "learning_rate": 4.40035603860208e-05, "loss": 6.6733, "step": 14797 }, { "epoch": 0.6239011741889242, "grad_norm": 2.03251576423645, "learning_rate": 4.4002779599612734e-05, "loss": 7.4981, "step": 14798 }, { "epoch": 0.6239433353711238, "grad_norm": 2.1451992988586426, "learning_rate": 4.4001998813204665e-05, "loss": 6.5186, "step": 14799 }, { "epoch": 0.6239854965533234, "grad_norm": 1.5808730125427246, "learning_rate": 4.400121802679659e-05, "loss": 6.7231, "step": 14800 }, { "epoch": 0.6240276577355229, "grad_norm": 2.309903144836426, "learning_rate": 4.400043724038852e-05, "loss": 6.9938, "step": 14801 }, { "epoch": 0.6240698189177225, "grad_norm": 2.251509189605713, "learning_rate": 4.399965645398045e-05, "loss": 7.1837, "step": 14802 }, { "epoch": 0.624111980099922, "grad_norm": 1.219325065612793, "learning_rate": 4.399887566757238e-05, "loss": 6.7377, "step": 14803 }, { "epoch": 0.6241541412821215, "grad_norm": 1.7502875328063965, "learning_rate": 4.399809488116431e-05, "loss": 6.3643, "step": 14804 }, { "epoch": 0.6241963024643211, "grad_norm": 1.7625257968902588, "learning_rate": 4.3997314094756246e-05, "loss": 6.7467, "step": 14805 }, { "epoch": 0.6242384636465207, "grad_norm": 2.7066586017608643, "learning_rate": 4.399653330834817e-05, "loss": 7.6719, "step": 14806 }, { "epoch": 0.6242806248287202, "grad_norm": 1.5565450191497803, "learning_rate": 4.3995752521940095e-05, "loss": 6.9864, "step": 14807 }, { "epoch": 0.6243227860109197, "grad_norm": 2.366001605987549, "learning_rate": 4.399497173553203e-05, "loss": 7.1887, "step": 14808 }, { "epoch": 0.6243649471931193, "grad_norm": 1.169626235961914, "learning_rate": 4.399419094912396e-05, "loss": 6.6859, "step": 14809 }, { "epoch": 0.6244071083753189, "grad_norm": 2.3875887393951416, "learning_rate": 4.399341016271589e-05, "loss": 7.2332, "step": 14810 }, { "epoch": 0.6244492695575183, "grad_norm": 2.7482504844665527, "learning_rate": 4.399262937630782e-05, "loss": 7.637, "step": 14811 }, { "epoch": 0.6244914307397179, "grad_norm": 1.3151476383209229, "learning_rate": 4.399184858989975e-05, "loss": 7.036, "step": 14812 }, { "epoch": 0.6245335919219175, "grad_norm": 3.1395316123962402, "learning_rate": 4.399106780349168e-05, "loss": 7.1419, "step": 14813 }, { "epoch": 0.6245757531041171, "grad_norm": 2.414569616317749, "learning_rate": 4.399028701708361e-05, "loss": 7.6934, "step": 14814 }, { "epoch": 0.6246179142863166, "grad_norm": 3.444793939590454, "learning_rate": 4.398950623067554e-05, "loss": 7.6192, "step": 14815 }, { "epoch": 0.6246600754685161, "grad_norm": 2.3905155658721924, "learning_rate": 4.398872544426747e-05, "loss": 7.408, "step": 14816 }, { "epoch": 0.6247022366507157, "grad_norm": 1.2111477851867676, "learning_rate": 4.39879446578594e-05, "loss": 6.8041, "step": 14817 }, { "epoch": 0.6247443978329152, "grad_norm": 2.128556251525879, "learning_rate": 4.3987163871451325e-05, "loss": 6.3838, "step": 14818 }, { "epoch": 0.6247865590151148, "grad_norm": 1.6494396924972534, "learning_rate": 4.3986383085043256e-05, "loss": 6.7964, "step": 14819 }, { "epoch": 0.6248287201973143, "grad_norm": 1.7547621726989746, "learning_rate": 4.398560229863519e-05, "loss": 6.6046, "step": 14820 }, { "epoch": 0.6248708813795139, "grad_norm": 3.0400350093841553, "learning_rate": 4.398482151222711e-05, "loss": 7.5967, "step": 14821 }, { "epoch": 0.6249130425617134, "grad_norm": 2.2259788513183594, "learning_rate": 4.398404072581905e-05, "loss": 6.4198, "step": 14822 }, { "epoch": 0.624955203743913, "grad_norm": 1.097306728363037, "learning_rate": 4.3983259939410974e-05, "loss": 7.0034, "step": 14823 }, { "epoch": 0.6249973649261126, "grad_norm": 2.6543822288513184, "learning_rate": 4.3982479153002906e-05, "loss": 6.7401, "step": 14824 }, { "epoch": 0.625039526108312, "grad_norm": 3.1367807388305664, "learning_rate": 4.398169836659484e-05, "loss": 6.393, "step": 14825 }, { "epoch": 0.6250816872905116, "grad_norm": 3.0946619510650635, "learning_rate": 4.398091758018676e-05, "loss": 6.2808, "step": 14826 }, { "epoch": 0.6251238484727112, "grad_norm": 1.957250714302063, "learning_rate": 4.39801367937787e-05, "loss": 6.2985, "step": 14827 }, { "epoch": 0.6251660096549108, "grad_norm": 3.693192958831787, "learning_rate": 4.3979356007370624e-05, "loss": 7.791, "step": 14828 }, { "epoch": 0.6252081708371102, "grad_norm": 1.5540448427200317, "learning_rate": 4.3978575220962555e-05, "loss": 6.644, "step": 14829 }, { "epoch": 0.6252503320193098, "grad_norm": 1.3820979595184326, "learning_rate": 4.3977794434554486e-05, "loss": 6.3306, "step": 14830 }, { "epoch": 0.6252924932015094, "grad_norm": 3.734605073928833, "learning_rate": 4.397701364814642e-05, "loss": 7.5891, "step": 14831 }, { "epoch": 0.6253346543837089, "grad_norm": 1.4399070739746094, "learning_rate": 4.397623286173834e-05, "loss": 6.9203, "step": 14832 }, { "epoch": 0.6253768155659085, "grad_norm": 1.759442687034607, "learning_rate": 4.3975452075330273e-05, "loss": 7.1247, "step": 14833 }, { "epoch": 0.625418976748108, "grad_norm": 1.1741373538970947, "learning_rate": 4.3974671288922205e-05, "loss": 6.7025, "step": 14834 }, { "epoch": 0.6254611379303076, "grad_norm": 1.2199918031692505, "learning_rate": 4.3973890502514136e-05, "loss": 6.4238, "step": 14835 }, { "epoch": 0.6255032991125071, "grad_norm": 1.8610212802886963, "learning_rate": 4.397310971610607e-05, "loss": 6.8166, "step": 14836 }, { "epoch": 0.6255454602947067, "grad_norm": 1.3440041542053223, "learning_rate": 4.397232892969799e-05, "loss": 6.3036, "step": 14837 }, { "epoch": 0.6255876214769062, "grad_norm": 1.565452218055725, "learning_rate": 4.397154814328992e-05, "loss": 6.9582, "step": 14838 }, { "epoch": 0.6256297826591057, "grad_norm": 0.9770897626876831, "learning_rate": 4.3970767356881854e-05, "loss": 6.402, "step": 14839 }, { "epoch": 0.6256719438413053, "grad_norm": 2.48355770111084, "learning_rate": 4.396998657047378e-05, "loss": 7.0887, "step": 14840 }, { "epoch": 0.6257141050235049, "grad_norm": 1.143252968788147, "learning_rate": 4.396920578406572e-05, "loss": 6.7826, "step": 14841 }, { "epoch": 0.6257562662057045, "grad_norm": 1.7077943086624146, "learning_rate": 4.396842499765764e-05, "loss": 6.3808, "step": 14842 }, { "epoch": 0.6257984273879039, "grad_norm": 1.5146969556808472, "learning_rate": 4.396764421124957e-05, "loss": 6.7536, "step": 14843 }, { "epoch": 0.6258405885701035, "grad_norm": 1.7369352579116821, "learning_rate": 4.3966863424841504e-05, "loss": 6.9091, "step": 14844 }, { "epoch": 0.6258827497523031, "grad_norm": 2.2697558403015137, "learning_rate": 4.396608263843343e-05, "loss": 7.0397, "step": 14845 }, { "epoch": 0.6259249109345026, "grad_norm": 2.5727336406707764, "learning_rate": 4.3965301852025366e-05, "loss": 6.5107, "step": 14846 }, { "epoch": 0.6259670721167021, "grad_norm": 1.506582260131836, "learning_rate": 4.396452106561729e-05, "loss": 6.6035, "step": 14847 }, { "epoch": 0.6260092332989017, "grad_norm": 1.657724380493164, "learning_rate": 4.396374027920922e-05, "loss": 7.041, "step": 14848 }, { "epoch": 0.6260513944811013, "grad_norm": 2.7802884578704834, "learning_rate": 4.396295949280115e-05, "loss": 7.3871, "step": 14849 }, { "epoch": 0.6260935556633008, "grad_norm": 1.1709835529327393, "learning_rate": 4.3962178706393085e-05, "loss": 6.7677, "step": 14850 }, { "epoch": 0.6261357168455004, "grad_norm": 3.9704957008361816, "learning_rate": 4.396139791998501e-05, "loss": 7.5377, "step": 14851 }, { "epoch": 0.6261778780276999, "grad_norm": 1.3906729221343994, "learning_rate": 4.396061713357694e-05, "loss": 6.7559, "step": 14852 }, { "epoch": 0.6262200392098994, "grad_norm": 3.036259651184082, "learning_rate": 4.395983634716887e-05, "loss": 7.5637, "step": 14853 }, { "epoch": 0.626262200392099, "grad_norm": 1.4066414833068848, "learning_rate": 4.3959055560760796e-05, "loss": 7.1322, "step": 14854 }, { "epoch": 0.6263043615742986, "grad_norm": 1.3847929239273071, "learning_rate": 4.3958274774352734e-05, "loss": 6.9393, "step": 14855 }, { "epoch": 0.626346522756498, "grad_norm": 2.281398296356201, "learning_rate": 4.395749398794466e-05, "loss": 6.6934, "step": 14856 }, { "epoch": 0.6263886839386976, "grad_norm": 1.9139572381973267, "learning_rate": 4.395671320153659e-05, "loss": 6.43, "step": 14857 }, { "epoch": 0.6264308451208972, "grad_norm": 1.4051302671432495, "learning_rate": 4.395593241512852e-05, "loss": 6.4705, "step": 14858 }, { "epoch": 0.6264730063030968, "grad_norm": 1.8703151941299438, "learning_rate": 4.3955151628720446e-05, "loss": 6.5778, "step": 14859 }, { "epoch": 0.6265151674852963, "grad_norm": 3.0095901489257812, "learning_rate": 4.3954370842312384e-05, "loss": 7.0912, "step": 14860 }, { "epoch": 0.6265573286674958, "grad_norm": 2.327373743057251, "learning_rate": 4.395359005590431e-05, "loss": 7.0169, "step": 14861 }, { "epoch": 0.6265994898496954, "grad_norm": 3.265977382659912, "learning_rate": 4.395280926949624e-05, "loss": 7.6882, "step": 14862 }, { "epoch": 0.626641651031895, "grad_norm": 1.7139486074447632, "learning_rate": 4.395202848308817e-05, "loss": 7.0645, "step": 14863 }, { "epoch": 0.6266838122140945, "grad_norm": 2.5433924198150635, "learning_rate": 4.3951247696680095e-05, "loss": 6.6574, "step": 14864 }, { "epoch": 0.626725973396294, "grad_norm": 3.2414872646331787, "learning_rate": 4.3950466910272026e-05, "loss": 6.5235, "step": 14865 }, { "epoch": 0.6267681345784936, "grad_norm": 2.198948383331299, "learning_rate": 4.394968612386396e-05, "loss": 6.6293, "step": 14866 }, { "epoch": 0.6268102957606931, "grad_norm": 2.8980228900909424, "learning_rate": 4.394890533745589e-05, "loss": 8.1341, "step": 14867 }, { "epoch": 0.6268524569428927, "grad_norm": 1.8318400382995605, "learning_rate": 4.394812455104782e-05, "loss": 7.0716, "step": 14868 }, { "epoch": 0.6268946181250922, "grad_norm": 2.896989107131958, "learning_rate": 4.3947343764639745e-05, "loss": 7.6423, "step": 14869 }, { "epoch": 0.6269367793072917, "grad_norm": 1.696601390838623, "learning_rate": 4.3946562978231676e-05, "loss": 7.0119, "step": 14870 }, { "epoch": 0.6269789404894913, "grad_norm": 1.00990629196167, "learning_rate": 4.394578219182361e-05, "loss": 7.0095, "step": 14871 }, { "epoch": 0.6270211016716909, "grad_norm": 2.394456386566162, "learning_rate": 4.394500140541554e-05, "loss": 6.4959, "step": 14872 }, { "epoch": 0.6270632628538905, "grad_norm": 2.1398937702178955, "learning_rate": 4.394422061900746e-05, "loss": 6.8277, "step": 14873 }, { "epoch": 0.6271054240360899, "grad_norm": 1.8451306819915771, "learning_rate": 4.39434398325994e-05, "loss": 6.6107, "step": 14874 }, { "epoch": 0.6271475852182895, "grad_norm": 3.116056203842163, "learning_rate": 4.3942659046191325e-05, "loss": 7.0541, "step": 14875 }, { "epoch": 0.6271897464004891, "grad_norm": 1.5408843755722046, "learning_rate": 4.394187825978326e-05, "loss": 6.8843, "step": 14876 }, { "epoch": 0.6272319075826887, "grad_norm": 2.25500750541687, "learning_rate": 4.394109747337519e-05, "loss": 6.8717, "step": 14877 }, { "epoch": 0.6272740687648881, "grad_norm": 1.1869724988937378, "learning_rate": 4.394031668696711e-05, "loss": 6.5985, "step": 14878 }, { "epoch": 0.6273162299470877, "grad_norm": 1.5482068061828613, "learning_rate": 4.393953590055905e-05, "loss": 6.3462, "step": 14879 }, { "epoch": 0.6273583911292873, "grad_norm": 1.4803348779678345, "learning_rate": 4.3938755114150975e-05, "loss": 6.3272, "step": 14880 }, { "epoch": 0.6274005523114868, "grad_norm": 3.147203207015991, "learning_rate": 4.3937974327742906e-05, "loss": 7.5573, "step": 14881 }, { "epoch": 0.6274427134936864, "grad_norm": 1.2991877794265747, "learning_rate": 4.393719354133484e-05, "loss": 6.3609, "step": 14882 }, { "epoch": 0.6274848746758859, "grad_norm": 1.5365240573883057, "learning_rate": 4.393641275492676e-05, "loss": 6.5119, "step": 14883 }, { "epoch": 0.6275270358580854, "grad_norm": 1.6949905157089233, "learning_rate": 4.393563196851869e-05, "loss": 7.0445, "step": 14884 }, { "epoch": 0.627569197040285, "grad_norm": 4.016100883483887, "learning_rate": 4.3934851182110624e-05, "loss": 7.7437, "step": 14885 }, { "epoch": 0.6276113582224846, "grad_norm": 1.6009536981582642, "learning_rate": 4.3934070395702556e-05, "loss": 6.3889, "step": 14886 }, { "epoch": 0.6276535194046841, "grad_norm": 3.1648430824279785, "learning_rate": 4.393328960929448e-05, "loss": 7.8656, "step": 14887 }, { "epoch": 0.6276956805868836, "grad_norm": 1.7472466230392456, "learning_rate": 4.393250882288641e-05, "loss": 6.6472, "step": 14888 }, { "epoch": 0.6277378417690832, "grad_norm": 1.2504911422729492, "learning_rate": 4.393172803647834e-05, "loss": 6.5479, "step": 14889 }, { "epoch": 0.6277800029512828, "grad_norm": 1.818038821220398, "learning_rate": 4.393094725007027e-05, "loss": 7.1081, "step": 14890 }, { "epoch": 0.6278221641334824, "grad_norm": 3.026667594909668, "learning_rate": 4.3930166463662205e-05, "loss": 8.0609, "step": 14891 }, { "epoch": 0.6278643253156818, "grad_norm": 1.5765467882156372, "learning_rate": 4.392938567725413e-05, "loss": 6.491, "step": 14892 }, { "epoch": 0.6279064864978814, "grad_norm": 1.4363738298416138, "learning_rate": 4.392860489084607e-05, "loss": 6.5828, "step": 14893 }, { "epoch": 0.627948647680081, "grad_norm": 1.4749903678894043, "learning_rate": 4.392782410443799e-05, "loss": 7.0205, "step": 14894 }, { "epoch": 0.6279908088622805, "grad_norm": 1.1101205348968506, "learning_rate": 4.3927043318029924e-05, "loss": 6.4795, "step": 14895 }, { "epoch": 0.62803297004448, "grad_norm": 0.9218024015426636, "learning_rate": 4.3926262531621855e-05, "loss": 6.3523, "step": 14896 }, { "epoch": 0.6280751312266796, "grad_norm": 2.43190860748291, "learning_rate": 4.392548174521378e-05, "loss": 6.9607, "step": 14897 }, { "epoch": 0.6281172924088791, "grad_norm": 1.5619794130325317, "learning_rate": 4.392470095880571e-05, "loss": 6.5567, "step": 14898 }, { "epoch": 0.6281594535910787, "grad_norm": 2.0378711223602295, "learning_rate": 4.392392017239764e-05, "loss": 6.3311, "step": 14899 }, { "epoch": 0.6282016147732783, "grad_norm": 2.4393351078033447, "learning_rate": 4.392313938598957e-05, "loss": 7.0808, "step": 14900 }, { "epoch": 0.6282437759554778, "grad_norm": 1.1961877346038818, "learning_rate": 4.39223585995815e-05, "loss": 6.5196, "step": 14901 }, { "epoch": 0.6282859371376773, "grad_norm": 3.6038601398468018, "learning_rate": 4.392157781317343e-05, "loss": 7.8357, "step": 14902 }, { "epoch": 0.6283280983198769, "grad_norm": 3.0934271812438965, "learning_rate": 4.392079702676536e-05, "loss": 6.474, "step": 14903 }, { "epoch": 0.6283702595020765, "grad_norm": 2.2577743530273438, "learning_rate": 4.392001624035729e-05, "loss": 6.9498, "step": 14904 }, { "epoch": 0.6284124206842759, "grad_norm": 2.1879563331604004, "learning_rate": 4.391923545394922e-05, "loss": 6.8111, "step": 14905 }, { "epoch": 0.6284545818664755, "grad_norm": 3.180204153060913, "learning_rate": 4.391845466754115e-05, "loss": 7.1665, "step": 14906 }, { "epoch": 0.6284967430486751, "grad_norm": 2.0675129890441895, "learning_rate": 4.391767388113308e-05, "loss": 6.753, "step": 14907 }, { "epoch": 0.6285389042308747, "grad_norm": 3.6668453216552734, "learning_rate": 4.391689309472501e-05, "loss": 8.3791, "step": 14908 }, { "epoch": 0.6285810654130742, "grad_norm": 1.598602533340454, "learning_rate": 4.3916112308316934e-05, "loss": 6.528, "step": 14909 }, { "epoch": 0.6286232265952737, "grad_norm": 1.4490060806274414, "learning_rate": 4.391533152190887e-05, "loss": 6.5255, "step": 14910 }, { "epoch": 0.6286653877774733, "grad_norm": 2.2613651752471924, "learning_rate": 4.3914550735500797e-05, "loss": 6.667, "step": 14911 }, { "epoch": 0.6287075489596728, "grad_norm": 2.9956376552581787, "learning_rate": 4.391376994909273e-05, "loss": 7.778, "step": 14912 }, { "epoch": 0.6287497101418724, "grad_norm": 1.767479658126831, "learning_rate": 4.391298916268466e-05, "loss": 6.8968, "step": 14913 }, { "epoch": 0.6287918713240719, "grad_norm": 1.5503535270690918, "learning_rate": 4.3912208376276584e-05, "loss": 6.4293, "step": 14914 }, { "epoch": 0.6288340325062715, "grad_norm": 2.6401724815368652, "learning_rate": 4.391142758986852e-05, "loss": 7.3588, "step": 14915 }, { "epoch": 0.628876193688471, "grad_norm": 3.445467710494995, "learning_rate": 4.3910646803460446e-05, "loss": 8.5172, "step": 14916 }, { "epoch": 0.6289183548706706, "grad_norm": 1.5061603784561157, "learning_rate": 4.390986601705238e-05, "loss": 7.1844, "step": 14917 }, { "epoch": 0.6289605160528702, "grad_norm": 1.4366871118545532, "learning_rate": 4.390908523064431e-05, "loss": 6.5504, "step": 14918 }, { "epoch": 0.6290026772350696, "grad_norm": 3.4669137001037598, "learning_rate": 4.390830444423624e-05, "loss": 7.5806, "step": 14919 }, { "epoch": 0.6290448384172692, "grad_norm": 1.3801944255828857, "learning_rate": 4.3907523657828164e-05, "loss": 6.7427, "step": 14920 }, { "epoch": 0.6290869995994688, "grad_norm": 1.3551561832427979, "learning_rate": 4.3906742871420096e-05, "loss": 6.6372, "step": 14921 }, { "epoch": 0.6291291607816684, "grad_norm": 2.3420987129211426, "learning_rate": 4.390596208501203e-05, "loss": 6.4922, "step": 14922 }, { "epoch": 0.6291713219638678, "grad_norm": 2.8946871757507324, "learning_rate": 4.390518129860395e-05, "loss": 7.9368, "step": 14923 }, { "epoch": 0.6292134831460674, "grad_norm": 1.5068156719207764, "learning_rate": 4.390440051219589e-05, "loss": 6.3167, "step": 14924 }, { "epoch": 0.629255644328267, "grad_norm": 1.895263910293579, "learning_rate": 4.3903619725787814e-05, "loss": 6.778, "step": 14925 }, { "epoch": 0.6292978055104665, "grad_norm": 1.0429123640060425, "learning_rate": 4.3902838939379745e-05, "loss": 6.5187, "step": 14926 }, { "epoch": 0.629339966692666, "grad_norm": 2.501739025115967, "learning_rate": 4.3902058152971676e-05, "loss": 7.1578, "step": 14927 }, { "epoch": 0.6293821278748656, "grad_norm": 1.1541281938552856, "learning_rate": 4.39012773665636e-05, "loss": 6.391, "step": 14928 }, { "epoch": 0.6294242890570652, "grad_norm": 2.4077823162078857, "learning_rate": 4.390049658015554e-05, "loss": 6.8556, "step": 14929 }, { "epoch": 0.6294664502392647, "grad_norm": 2.3788836002349854, "learning_rate": 4.3899715793747463e-05, "loss": 6.9456, "step": 14930 }, { "epoch": 0.6295086114214643, "grad_norm": 2.6916534900665283, "learning_rate": 4.3898935007339395e-05, "loss": 8.6979, "step": 14931 }, { "epoch": 0.6295507726036638, "grad_norm": 2.21232008934021, "learning_rate": 4.3898154220931326e-05, "loss": 6.4873, "step": 14932 }, { "epoch": 0.6295929337858633, "grad_norm": 2.1428816318511963, "learning_rate": 4.389737343452325e-05, "loss": 6.6673, "step": 14933 }, { "epoch": 0.6296350949680629, "grad_norm": 1.2967052459716797, "learning_rate": 4.389659264811518e-05, "loss": 6.472, "step": 14934 }, { "epoch": 0.6296772561502625, "grad_norm": 1.847038745880127, "learning_rate": 4.389581186170711e-05, "loss": 6.5005, "step": 14935 }, { "epoch": 0.629719417332462, "grad_norm": 2.182673215866089, "learning_rate": 4.3895031075299044e-05, "loss": 6.6027, "step": 14936 }, { "epoch": 0.6297615785146615, "grad_norm": 3.635226011276245, "learning_rate": 4.3894250288890975e-05, "loss": 6.3685, "step": 14937 }, { "epoch": 0.6298037396968611, "grad_norm": 2.7535204887390137, "learning_rate": 4.389346950248291e-05, "loss": 7.1716, "step": 14938 }, { "epoch": 0.6298459008790607, "grad_norm": 2.6123409271240234, "learning_rate": 4.389268871607483e-05, "loss": 6.6383, "step": 14939 }, { "epoch": 0.6298880620612602, "grad_norm": 1.3538126945495605, "learning_rate": 4.389190792966676e-05, "loss": 6.3227, "step": 14940 }, { "epoch": 0.6299302232434597, "grad_norm": 1.6540768146514893, "learning_rate": 4.3891127143258694e-05, "loss": 6.3548, "step": 14941 }, { "epoch": 0.6299723844256593, "grad_norm": 2.415487051010132, "learning_rate": 4.389034635685062e-05, "loss": 6.4253, "step": 14942 }, { "epoch": 0.6300145456078589, "grad_norm": 2.523954153060913, "learning_rate": 4.3889565570442556e-05, "loss": 7.792, "step": 14943 }, { "epoch": 0.6300567067900584, "grad_norm": 2.127329111099243, "learning_rate": 4.388878478403448e-05, "loss": 7.0635, "step": 14944 }, { "epoch": 0.6300988679722579, "grad_norm": 1.1995326280593872, "learning_rate": 4.388800399762641e-05, "loss": 6.4411, "step": 14945 }, { "epoch": 0.6301410291544575, "grad_norm": 1.7455461025238037, "learning_rate": 4.388722321121834e-05, "loss": 6.9, "step": 14946 }, { "epoch": 0.630183190336657, "grad_norm": 2.948223114013672, "learning_rate": 4.388644242481027e-05, "loss": 8.1399, "step": 14947 }, { "epoch": 0.6302253515188566, "grad_norm": 1.415719985961914, "learning_rate": 4.3885661638402206e-05, "loss": 6.5637, "step": 14948 }, { "epoch": 0.6302675127010562, "grad_norm": 2.0143673419952393, "learning_rate": 4.388488085199413e-05, "loss": 6.9317, "step": 14949 }, { "epoch": 0.6303096738832557, "grad_norm": 1.4306560754776, "learning_rate": 4.388410006558606e-05, "loss": 6.4531, "step": 14950 }, { "epoch": 0.6303518350654552, "grad_norm": 1.7967281341552734, "learning_rate": 4.388331927917799e-05, "loss": 6.4169, "step": 14951 }, { "epoch": 0.6303939962476548, "grad_norm": 2.96051287651062, "learning_rate": 4.388253849276992e-05, "loss": 7.1818, "step": 14952 }, { "epoch": 0.6304361574298544, "grad_norm": 1.2424070835113525, "learning_rate": 4.388175770636185e-05, "loss": 6.314, "step": 14953 }, { "epoch": 0.6304783186120538, "grad_norm": 2.841442108154297, "learning_rate": 4.388097691995378e-05, "loss": 6.4432, "step": 14954 }, { "epoch": 0.6305204797942534, "grad_norm": 2.4116580486297607, "learning_rate": 4.388019613354571e-05, "loss": 6.2848, "step": 14955 }, { "epoch": 0.630562640976453, "grad_norm": 3.6466691493988037, "learning_rate": 4.3879415347137635e-05, "loss": 7.8252, "step": 14956 }, { "epoch": 0.6306048021586526, "grad_norm": 1.5010676383972168, "learning_rate": 4.387863456072957e-05, "loss": 6.4045, "step": 14957 }, { "epoch": 0.6306469633408521, "grad_norm": 1.6742311716079712, "learning_rate": 4.38778537743215e-05, "loss": 6.7967, "step": 14958 }, { "epoch": 0.6306891245230516, "grad_norm": 0.9705008268356323, "learning_rate": 4.387707298791343e-05, "loss": 6.4893, "step": 14959 }, { "epoch": 0.6307312857052512, "grad_norm": 1.248133659362793, "learning_rate": 4.387629220150536e-05, "loss": 6.4464, "step": 14960 }, { "epoch": 0.6307734468874507, "grad_norm": 1.0728012323379517, "learning_rate": 4.3875511415097285e-05, "loss": 6.3842, "step": 14961 }, { "epoch": 0.6308156080696503, "grad_norm": 1.511568546295166, "learning_rate": 4.387473062868922e-05, "loss": 6.5428, "step": 14962 }, { "epoch": 0.6308577692518498, "grad_norm": 3.629086971282959, "learning_rate": 4.387394984228115e-05, "loss": 7.5945, "step": 14963 }, { "epoch": 0.6308999304340493, "grad_norm": 1.7022852897644043, "learning_rate": 4.387316905587308e-05, "loss": 6.5056, "step": 14964 }, { "epoch": 0.6309420916162489, "grad_norm": 1.3962527513504028, "learning_rate": 4.387238826946501e-05, "loss": 6.845, "step": 14965 }, { "epoch": 0.6309842527984485, "grad_norm": 1.4899829626083374, "learning_rate": 4.3871607483056935e-05, "loss": 7.1553, "step": 14966 }, { "epoch": 0.6310264139806481, "grad_norm": 1.9352080821990967, "learning_rate": 4.3870826696648866e-05, "loss": 6.2775, "step": 14967 }, { "epoch": 0.6310685751628475, "grad_norm": 1.213633418083191, "learning_rate": 4.38700459102408e-05, "loss": 6.5035, "step": 14968 }, { "epoch": 0.6311107363450471, "grad_norm": 1.370010495185852, "learning_rate": 4.386926512383273e-05, "loss": 6.3467, "step": 14969 }, { "epoch": 0.6311528975272467, "grad_norm": 2.3575117588043213, "learning_rate": 4.386848433742465e-05, "loss": 6.3767, "step": 14970 }, { "epoch": 0.6311950587094463, "grad_norm": 1.0172158479690552, "learning_rate": 4.3867703551016584e-05, "loss": 6.6993, "step": 14971 }, { "epoch": 0.6312372198916457, "grad_norm": 1.690371036529541, "learning_rate": 4.3866922764608515e-05, "loss": 6.8656, "step": 14972 }, { "epoch": 0.6312793810738453, "grad_norm": 2.911720037460327, "learning_rate": 4.3866141978200447e-05, "loss": 7.963, "step": 14973 }, { "epoch": 0.6313215422560449, "grad_norm": 1.2581969499588013, "learning_rate": 4.386536119179238e-05, "loss": 6.9355, "step": 14974 }, { "epoch": 0.6313637034382444, "grad_norm": 2.204442262649536, "learning_rate": 4.38645804053843e-05, "loss": 7.0, "step": 14975 }, { "epoch": 0.6314058646204439, "grad_norm": 3.615227460861206, "learning_rate": 4.3863799618976234e-05, "loss": 8.2609, "step": 14976 }, { "epoch": 0.6314480258026435, "grad_norm": 1.370405673980713, "learning_rate": 4.3863018832568165e-05, "loss": 6.4598, "step": 14977 }, { "epoch": 0.631490186984843, "grad_norm": 1.0989574193954468, "learning_rate": 4.386223804616009e-05, "loss": 6.3528, "step": 14978 }, { "epoch": 0.6315323481670426, "grad_norm": 2.313253879547119, "learning_rate": 4.386145725975203e-05, "loss": 7.0872, "step": 14979 }, { "epoch": 0.6315745093492422, "grad_norm": 3.155224561691284, "learning_rate": 4.386067647334395e-05, "loss": 7.692, "step": 14980 }, { "epoch": 0.6316166705314417, "grad_norm": 2.4644973278045654, "learning_rate": 4.385989568693588e-05, "loss": 6.5469, "step": 14981 }, { "epoch": 0.6316588317136412, "grad_norm": 1.8738070726394653, "learning_rate": 4.3859114900527814e-05, "loss": 6.8227, "step": 14982 }, { "epoch": 0.6317009928958408, "grad_norm": 1.7581126689910889, "learning_rate": 4.3858334114119746e-05, "loss": 6.4014, "step": 14983 }, { "epoch": 0.6317431540780404, "grad_norm": 1.357163906097412, "learning_rate": 4.385755332771168e-05, "loss": 6.8429, "step": 14984 }, { "epoch": 0.6317853152602398, "grad_norm": 2.735334634780884, "learning_rate": 4.38567725413036e-05, "loss": 7.2467, "step": 14985 }, { "epoch": 0.6318274764424394, "grad_norm": 4.228466033935547, "learning_rate": 4.385599175489553e-05, "loss": 7.7188, "step": 14986 }, { "epoch": 0.631869637624639, "grad_norm": 1.2237539291381836, "learning_rate": 4.3855210968487464e-05, "loss": 6.5469, "step": 14987 }, { "epoch": 0.6319117988068386, "grad_norm": 1.672922134399414, "learning_rate": 4.3854430182079395e-05, "loss": 6.9828, "step": 14988 }, { "epoch": 0.6319539599890381, "grad_norm": 2.828150510787964, "learning_rate": 4.385364939567132e-05, "loss": 7.0154, "step": 14989 }, { "epoch": 0.6319961211712376, "grad_norm": 3.2265028953552246, "learning_rate": 4.385286860926325e-05, "loss": 7.3923, "step": 14990 }, { "epoch": 0.6320382823534372, "grad_norm": 3.153975248336792, "learning_rate": 4.385208782285518e-05, "loss": 7.7339, "step": 14991 }, { "epoch": 0.6320804435356367, "grad_norm": 1.1945327520370483, "learning_rate": 4.385130703644711e-05, "loss": 6.4744, "step": 14992 }, { "epoch": 0.6321226047178363, "grad_norm": 1.8268648386001587, "learning_rate": 4.3850526250039045e-05, "loss": 6.9701, "step": 14993 }, { "epoch": 0.6321647659000358, "grad_norm": 1.1306571960449219, "learning_rate": 4.384974546363097e-05, "loss": 6.3861, "step": 14994 }, { "epoch": 0.6322069270822354, "grad_norm": 1.546478033065796, "learning_rate": 4.38489646772229e-05, "loss": 6.2198, "step": 14995 }, { "epoch": 0.6322490882644349, "grad_norm": 1.2767932415008545, "learning_rate": 4.384818389081483e-05, "loss": 6.8924, "step": 14996 }, { "epoch": 0.6322912494466345, "grad_norm": 2.1585328578948975, "learning_rate": 4.3847403104406756e-05, "loss": 7.4193, "step": 14997 }, { "epoch": 0.6323334106288341, "grad_norm": 1.6157007217407227, "learning_rate": 4.3846622317998694e-05, "loss": 6.9459, "step": 14998 }, { "epoch": 0.6323755718110335, "grad_norm": 1.599568486213684, "learning_rate": 4.384584153159062e-05, "loss": 6.6877, "step": 14999 }, { "epoch": 0.6324177329932331, "grad_norm": 1.634877324104309, "learning_rate": 4.384506074518255e-05, "loss": 7.5681, "step": 15000 }, { "epoch": 0.6324598941754327, "grad_norm": 1.7489155530929565, "learning_rate": 4.384427995877448e-05, "loss": 6.6192, "step": 15001 }, { "epoch": 0.6325020553576323, "grad_norm": 1.760226845741272, "learning_rate": 4.3843499172366406e-05, "loss": 7.2646, "step": 15002 }, { "epoch": 0.6325442165398317, "grad_norm": 1.5988563299179077, "learning_rate": 4.384271838595834e-05, "loss": 6.2746, "step": 15003 }, { "epoch": 0.6325863777220313, "grad_norm": 2.069939374923706, "learning_rate": 4.384193759955027e-05, "loss": 6.4112, "step": 15004 }, { "epoch": 0.6326285389042309, "grad_norm": 2.7322473526000977, "learning_rate": 4.38411568131422e-05, "loss": 7.0869, "step": 15005 }, { "epoch": 0.6326707000864304, "grad_norm": 1.5865050554275513, "learning_rate": 4.384037602673413e-05, "loss": 6.4212, "step": 15006 }, { "epoch": 0.63271286126863, "grad_norm": 1.3228285312652588, "learning_rate": 4.383959524032606e-05, "loss": 6.7097, "step": 15007 }, { "epoch": 0.6327550224508295, "grad_norm": 1.1740508079528809, "learning_rate": 4.3838814453917986e-05, "loss": 6.8125, "step": 15008 }, { "epoch": 0.6327971836330291, "grad_norm": 1.2140960693359375, "learning_rate": 4.383803366750992e-05, "loss": 6.6636, "step": 15009 }, { "epoch": 0.6328393448152286, "grad_norm": 2.5420596599578857, "learning_rate": 4.383725288110185e-05, "loss": 7.0802, "step": 15010 }, { "epoch": 0.6328815059974282, "grad_norm": 1.2453678846359253, "learning_rate": 4.3836472094693773e-05, "loss": 6.6694, "step": 15011 }, { "epoch": 0.6329236671796277, "grad_norm": 1.2099393606185913, "learning_rate": 4.383569130828571e-05, "loss": 6.6086, "step": 15012 }, { "epoch": 0.6329658283618272, "grad_norm": 2.8937723636627197, "learning_rate": 4.3834910521877636e-05, "loss": 7.0547, "step": 15013 }, { "epoch": 0.6330079895440268, "grad_norm": 1.276033639907837, "learning_rate": 4.383412973546957e-05, "loss": 6.4377, "step": 15014 }, { "epoch": 0.6330501507262264, "grad_norm": 1.161119818687439, "learning_rate": 4.38333489490615e-05, "loss": 6.623, "step": 15015 }, { "epoch": 0.633092311908426, "grad_norm": 3.627967119216919, "learning_rate": 4.383256816265342e-05, "loss": 7.5212, "step": 15016 }, { "epoch": 0.6331344730906254, "grad_norm": 2.6888532638549805, "learning_rate": 4.383178737624536e-05, "loss": 7.5439, "step": 15017 }, { "epoch": 0.633176634272825, "grad_norm": 3.032101631164551, "learning_rate": 4.3831006589837286e-05, "loss": 7.7799, "step": 15018 }, { "epoch": 0.6332187954550246, "grad_norm": 1.5955651998519897, "learning_rate": 4.383022580342922e-05, "loss": 6.5587, "step": 15019 }, { "epoch": 0.6332609566372241, "grad_norm": 1.0155023336410522, "learning_rate": 4.382944501702115e-05, "loss": 6.9087, "step": 15020 }, { "epoch": 0.6333031178194236, "grad_norm": 3.804412603378296, "learning_rate": 4.382866423061307e-05, "loss": 7.576, "step": 15021 }, { "epoch": 0.6333452790016232, "grad_norm": 1.5225015878677368, "learning_rate": 4.3827883444205004e-05, "loss": 6.9389, "step": 15022 }, { "epoch": 0.6333874401838228, "grad_norm": 1.597866177558899, "learning_rate": 4.3827102657796935e-05, "loss": 7.1302, "step": 15023 }, { "epoch": 0.6334296013660223, "grad_norm": 2.6689796447753906, "learning_rate": 4.3826321871388866e-05, "loss": 6.4972, "step": 15024 }, { "epoch": 0.6334717625482219, "grad_norm": 2.923898458480835, "learning_rate": 4.382554108498079e-05, "loss": 7.7218, "step": 15025 }, { "epoch": 0.6335139237304214, "grad_norm": 1.5623509883880615, "learning_rate": 4.382476029857273e-05, "loss": 6.6094, "step": 15026 }, { "epoch": 0.6335560849126209, "grad_norm": 2.2782416343688965, "learning_rate": 4.382397951216465e-05, "loss": 6.9681, "step": 15027 }, { "epoch": 0.6335982460948205, "grad_norm": 1.8127940893173218, "learning_rate": 4.3823198725756585e-05, "loss": 6.3871, "step": 15028 }, { "epoch": 0.6336404072770201, "grad_norm": 1.3281970024108887, "learning_rate": 4.3822417939348516e-05, "loss": 6.7694, "step": 15029 }, { "epoch": 0.6336825684592196, "grad_norm": 1.181469202041626, "learning_rate": 4.382163715294044e-05, "loss": 6.5455, "step": 15030 }, { "epoch": 0.6337247296414191, "grad_norm": 2.7946488857269287, "learning_rate": 4.382085636653238e-05, "loss": 6.8358, "step": 15031 }, { "epoch": 0.6337668908236187, "grad_norm": 3.438542366027832, "learning_rate": 4.38200755801243e-05, "loss": 7.5775, "step": 15032 }, { "epoch": 0.6338090520058183, "grad_norm": 2.2698020935058594, "learning_rate": 4.3819294793716234e-05, "loss": 6.3763, "step": 15033 }, { "epoch": 0.6338512131880177, "grad_norm": 1.237310767173767, "learning_rate": 4.3818514007308165e-05, "loss": 6.4641, "step": 15034 }, { "epoch": 0.6338933743702173, "grad_norm": 1.0163891315460205, "learning_rate": 4.381773322090009e-05, "loss": 6.9778, "step": 15035 }, { "epoch": 0.6339355355524169, "grad_norm": 3.44677472114563, "learning_rate": 4.381695243449202e-05, "loss": 8.2388, "step": 15036 }, { "epoch": 0.6339776967346165, "grad_norm": 1.6806859970092773, "learning_rate": 4.381617164808395e-05, "loss": 6.9, "step": 15037 }, { "epoch": 0.634019857916816, "grad_norm": 3.687272071838379, "learning_rate": 4.3815390861675884e-05, "loss": 6.9552, "step": 15038 }, { "epoch": 0.6340620190990155, "grad_norm": 3.048863410949707, "learning_rate": 4.3814610075267815e-05, "loss": 7.6013, "step": 15039 }, { "epoch": 0.6341041802812151, "grad_norm": 2.0967600345611572, "learning_rate": 4.381382928885974e-05, "loss": 7.6458, "step": 15040 }, { "epoch": 0.6341463414634146, "grad_norm": 1.638623595237732, "learning_rate": 4.381304850245167e-05, "loss": 7.2471, "step": 15041 }, { "epoch": 0.6341885026456142, "grad_norm": 2.4965429306030273, "learning_rate": 4.38122677160436e-05, "loss": 6.758, "step": 15042 }, { "epoch": 0.6342306638278137, "grad_norm": 1.1444426774978638, "learning_rate": 4.381148692963553e-05, "loss": 6.4928, "step": 15043 }, { "epoch": 0.6342728250100133, "grad_norm": 1.6098624467849731, "learning_rate": 4.381070614322746e-05, "loss": 6.2605, "step": 15044 }, { "epoch": 0.6343149861922128, "grad_norm": 2.1092092990875244, "learning_rate": 4.3809925356819396e-05, "loss": 6.9343, "step": 15045 }, { "epoch": 0.6343571473744124, "grad_norm": 1.1621079444885254, "learning_rate": 4.380914457041132e-05, "loss": 6.4977, "step": 15046 }, { "epoch": 0.634399308556612, "grad_norm": 0.8976702690124512, "learning_rate": 4.3808363784003245e-05, "loss": 6.5706, "step": 15047 }, { "epoch": 0.6344414697388114, "grad_norm": 1.6095777750015259, "learning_rate": 4.380758299759518e-05, "loss": 6.7005, "step": 15048 }, { "epoch": 0.634483630921011, "grad_norm": 1.7576532363891602, "learning_rate": 4.380680221118711e-05, "loss": 6.6392, "step": 15049 }, { "epoch": 0.6345257921032106, "grad_norm": 1.4317476749420166, "learning_rate": 4.380602142477904e-05, "loss": 6.8539, "step": 15050 }, { "epoch": 0.6345679532854102, "grad_norm": 1.2439944744110107, "learning_rate": 4.380524063837097e-05, "loss": 6.6297, "step": 15051 }, { "epoch": 0.6346101144676096, "grad_norm": 1.5435587167739868, "learning_rate": 4.38044598519629e-05, "loss": 7.0855, "step": 15052 }, { "epoch": 0.6346522756498092, "grad_norm": 1.9637376070022583, "learning_rate": 4.380367906555483e-05, "loss": 6.7887, "step": 15053 }, { "epoch": 0.6346944368320088, "grad_norm": 1.0731773376464844, "learning_rate": 4.380289827914676e-05, "loss": 6.5343, "step": 15054 }, { "epoch": 0.6347365980142083, "grad_norm": 3.6378278732299805, "learning_rate": 4.380211749273869e-05, "loss": 7.7811, "step": 15055 }, { "epoch": 0.6347787591964079, "grad_norm": 1.2570315599441528, "learning_rate": 4.380133670633062e-05, "loss": 6.8051, "step": 15056 }, { "epoch": 0.6348209203786074, "grad_norm": 1.1479074954986572, "learning_rate": 4.380055591992255e-05, "loss": 6.5365, "step": 15057 }, { "epoch": 0.634863081560807, "grad_norm": 1.3668562173843384, "learning_rate": 4.3799775133514475e-05, "loss": 6.493, "step": 15058 }, { "epoch": 0.6349052427430065, "grad_norm": 3.575230836868286, "learning_rate": 4.3798994347106406e-05, "loss": 7.2788, "step": 15059 }, { "epoch": 0.6349474039252061, "grad_norm": 1.3984851837158203, "learning_rate": 4.379821356069834e-05, "loss": 7.1136, "step": 15060 }, { "epoch": 0.6349895651074056, "grad_norm": 1.724631905555725, "learning_rate": 4.379743277429026e-05, "loss": 6.4463, "step": 15061 }, { "epoch": 0.6350317262896051, "grad_norm": 1.4874955415725708, "learning_rate": 4.37966519878822e-05, "loss": 6.4183, "step": 15062 }, { "epoch": 0.6350738874718047, "grad_norm": 3.1566851139068604, "learning_rate": 4.3795871201474124e-05, "loss": 6.2898, "step": 15063 }, { "epoch": 0.6351160486540043, "grad_norm": 1.211057186126709, "learning_rate": 4.3795090415066056e-05, "loss": 6.3664, "step": 15064 }, { "epoch": 0.6351582098362039, "grad_norm": 3.1652259826660156, "learning_rate": 4.379430962865799e-05, "loss": 7.563, "step": 15065 }, { "epoch": 0.6352003710184033, "grad_norm": 2.4662837982177734, "learning_rate": 4.379352884224991e-05, "loss": 6.7671, "step": 15066 }, { "epoch": 0.6352425322006029, "grad_norm": 0.9524003863334656, "learning_rate": 4.379274805584185e-05, "loss": 6.7007, "step": 15067 }, { "epoch": 0.6352846933828025, "grad_norm": 1.5055413246154785, "learning_rate": 4.3791967269433774e-05, "loss": 6.5836, "step": 15068 }, { "epoch": 0.635326854565002, "grad_norm": 1.5465377569198608, "learning_rate": 4.3791186483025705e-05, "loss": 6.6675, "step": 15069 }, { "epoch": 0.6353690157472015, "grad_norm": 1.123049020767212, "learning_rate": 4.3790405696617636e-05, "loss": 6.6492, "step": 15070 }, { "epoch": 0.6354111769294011, "grad_norm": 1.4708447456359863, "learning_rate": 4.378962491020957e-05, "loss": 6.6323, "step": 15071 }, { "epoch": 0.6354533381116007, "grad_norm": 3.763943910598755, "learning_rate": 4.378884412380149e-05, "loss": 7.3226, "step": 15072 }, { "epoch": 0.6354954992938002, "grad_norm": 3.6636579036712646, "learning_rate": 4.3788063337393423e-05, "loss": 7.4304, "step": 15073 }, { "epoch": 0.6355376604759998, "grad_norm": 3.8621666431427, "learning_rate": 4.3787282550985355e-05, "loss": 6.9707, "step": 15074 }, { "epoch": 0.6355798216581993, "grad_norm": 1.4950495958328247, "learning_rate": 4.3786501764577286e-05, "loss": 6.5061, "step": 15075 }, { "epoch": 0.6356219828403988, "grad_norm": 3.5450119972229004, "learning_rate": 4.378572097816922e-05, "loss": 6.6068, "step": 15076 }, { "epoch": 0.6356641440225984, "grad_norm": 1.6534490585327148, "learning_rate": 4.378494019176114e-05, "loss": 6.2508, "step": 15077 }, { "epoch": 0.635706305204798, "grad_norm": 2.5483815670013428, "learning_rate": 4.378415940535307e-05, "loss": 7.1022, "step": 15078 }, { "epoch": 0.6357484663869974, "grad_norm": 2.1456215381622314, "learning_rate": 4.3783378618945004e-05, "loss": 6.6633, "step": 15079 }, { "epoch": 0.635790627569197, "grad_norm": 2.4367239475250244, "learning_rate": 4.378259783253693e-05, "loss": 6.4159, "step": 15080 }, { "epoch": 0.6358327887513966, "grad_norm": 2.102268934249878, "learning_rate": 4.378181704612887e-05, "loss": 6.3195, "step": 15081 }, { "epoch": 0.6358749499335962, "grad_norm": 1.2372907400131226, "learning_rate": 4.378103625972079e-05, "loss": 6.3735, "step": 15082 }, { "epoch": 0.6359171111157957, "grad_norm": 1.269911527633667, "learning_rate": 4.378025547331272e-05, "loss": 6.3015, "step": 15083 }, { "epoch": 0.6359592722979952, "grad_norm": 2.833221912384033, "learning_rate": 4.3779474686904654e-05, "loss": 6.3925, "step": 15084 }, { "epoch": 0.6360014334801948, "grad_norm": 3.753905773162842, "learning_rate": 4.377869390049658e-05, "loss": 6.2812, "step": 15085 }, { "epoch": 0.6360435946623944, "grad_norm": 2.826988697052002, "learning_rate": 4.3777913114088516e-05, "loss": 7.0906, "step": 15086 }, { "epoch": 0.6360857558445939, "grad_norm": 1.7715654373168945, "learning_rate": 4.377713232768044e-05, "loss": 7.0594, "step": 15087 }, { "epoch": 0.6361279170267934, "grad_norm": 2.8581104278564453, "learning_rate": 4.377635154127237e-05, "loss": 6.5043, "step": 15088 }, { "epoch": 0.636170078208993, "grad_norm": 2.8980865478515625, "learning_rate": 4.37755707548643e-05, "loss": 7.8492, "step": 15089 }, { "epoch": 0.6362122393911925, "grad_norm": 2.1472110748291016, "learning_rate": 4.3774789968456235e-05, "loss": 7.0138, "step": 15090 }, { "epoch": 0.6362544005733921, "grad_norm": 3.8092849254608154, "learning_rate": 4.377400918204816e-05, "loss": 7.8795, "step": 15091 }, { "epoch": 0.6362965617555916, "grad_norm": 1.5257253646850586, "learning_rate": 4.377322839564009e-05, "loss": 7.0217, "step": 15092 }, { "epoch": 0.6363387229377911, "grad_norm": 2.4326701164245605, "learning_rate": 4.377244760923202e-05, "loss": 7.0755, "step": 15093 }, { "epoch": 0.6363808841199907, "grad_norm": 2.1155221462249756, "learning_rate": 4.3771666822823946e-05, "loss": 7.0292, "step": 15094 }, { "epoch": 0.6364230453021903, "grad_norm": 0.9337542057037354, "learning_rate": 4.3770886036415884e-05, "loss": 6.4431, "step": 15095 }, { "epoch": 0.6364652064843899, "grad_norm": 1.0563932657241821, "learning_rate": 4.377010525000781e-05, "loss": 6.5329, "step": 15096 }, { "epoch": 0.6365073676665893, "grad_norm": 2.9686803817749023, "learning_rate": 4.376932446359974e-05, "loss": 6.1732, "step": 15097 }, { "epoch": 0.6365495288487889, "grad_norm": 1.7532548904418945, "learning_rate": 4.376854367719167e-05, "loss": 6.649, "step": 15098 }, { "epoch": 0.6365916900309885, "grad_norm": 1.6060335636138916, "learning_rate": 4.3767762890783596e-05, "loss": 6.8703, "step": 15099 }, { "epoch": 0.636633851213188, "grad_norm": 2.0546905994415283, "learning_rate": 4.3766982104375534e-05, "loss": 7.6408, "step": 15100 }, { "epoch": 0.6366760123953875, "grad_norm": 4.30914831161499, "learning_rate": 4.376620131796746e-05, "loss": 7.9091, "step": 15101 }, { "epoch": 0.6367181735775871, "grad_norm": 2.1161208152770996, "learning_rate": 4.376542053155939e-05, "loss": 7.3417, "step": 15102 }, { "epoch": 0.6367603347597867, "grad_norm": 3.0726659297943115, "learning_rate": 4.376463974515132e-05, "loss": 7.7231, "step": 15103 }, { "epoch": 0.6368024959419862, "grad_norm": 4.644534587860107, "learning_rate": 4.3763858958743245e-05, "loss": 7.7771, "step": 15104 }, { "epoch": 0.6368446571241858, "grad_norm": 2.29807186126709, "learning_rate": 4.3763078172335176e-05, "loss": 7.033, "step": 15105 }, { "epoch": 0.6368868183063853, "grad_norm": 2.337718963623047, "learning_rate": 4.376229738592711e-05, "loss": 6.846, "step": 15106 }, { "epoch": 0.6369289794885848, "grad_norm": 1.5077444314956665, "learning_rate": 4.376151659951904e-05, "loss": 6.5871, "step": 15107 }, { "epoch": 0.6369711406707844, "grad_norm": 2.730005979537964, "learning_rate": 4.376073581311097e-05, "loss": 6.9028, "step": 15108 }, { "epoch": 0.637013301852984, "grad_norm": 3.1766486167907715, "learning_rate": 4.3759955026702895e-05, "loss": 6.6467, "step": 15109 }, { "epoch": 0.6370554630351835, "grad_norm": 3.335899591445923, "learning_rate": 4.3759174240294826e-05, "loss": 7.201, "step": 15110 }, { "epoch": 0.637097624217383, "grad_norm": 1.776434302330017, "learning_rate": 4.375839345388676e-05, "loss": 6.6578, "step": 15111 }, { "epoch": 0.6371397853995826, "grad_norm": 1.508104920387268, "learning_rate": 4.375761266747869e-05, "loss": 6.2283, "step": 15112 }, { "epoch": 0.6371819465817822, "grad_norm": 3.006518840789795, "learning_rate": 4.375683188107061e-05, "loss": 7.5949, "step": 15113 }, { "epoch": 0.6372241077639818, "grad_norm": 2.9652926921844482, "learning_rate": 4.375605109466255e-05, "loss": 7.2515, "step": 15114 }, { "epoch": 0.6372662689461812, "grad_norm": 1.581114649772644, "learning_rate": 4.3755270308254475e-05, "loss": 6.4421, "step": 15115 }, { "epoch": 0.6373084301283808, "grad_norm": 1.1569278240203857, "learning_rate": 4.375448952184641e-05, "loss": 6.6437, "step": 15116 }, { "epoch": 0.6373505913105804, "grad_norm": 1.3944696187973022, "learning_rate": 4.375370873543834e-05, "loss": 7.0179, "step": 15117 }, { "epoch": 0.6373927524927799, "grad_norm": 1.6396214962005615, "learning_rate": 4.375292794903026e-05, "loss": 6.4824, "step": 15118 }, { "epoch": 0.6374349136749794, "grad_norm": 3.2216126918792725, "learning_rate": 4.37521471626222e-05, "loss": 7.2286, "step": 15119 }, { "epoch": 0.637477074857179, "grad_norm": 1.0757408142089844, "learning_rate": 4.3751366376214125e-05, "loss": 6.4891, "step": 15120 }, { "epoch": 0.6375192360393785, "grad_norm": 0.8083873391151428, "learning_rate": 4.3750585589806056e-05, "loss": 6.4041, "step": 15121 }, { "epoch": 0.6375613972215781, "grad_norm": 2.0952913761138916, "learning_rate": 4.374980480339799e-05, "loss": 6.8808, "step": 15122 }, { "epoch": 0.6376035584037777, "grad_norm": 1.7179282903671265, "learning_rate": 4.374902401698991e-05, "loss": 6.3586, "step": 15123 }, { "epoch": 0.6376457195859772, "grad_norm": 3.810441732406616, "learning_rate": 4.374824323058184e-05, "loss": 7.4806, "step": 15124 }, { "epoch": 0.6376878807681767, "grad_norm": 1.991894006729126, "learning_rate": 4.3747462444173774e-05, "loss": 6.5294, "step": 15125 }, { "epoch": 0.6377300419503763, "grad_norm": 1.4170269966125488, "learning_rate": 4.3746681657765706e-05, "loss": 6.88, "step": 15126 }, { "epoch": 0.6377722031325759, "grad_norm": 1.1390857696533203, "learning_rate": 4.374590087135763e-05, "loss": 6.5991, "step": 15127 }, { "epoch": 0.6378143643147753, "grad_norm": 0.905189573764801, "learning_rate": 4.374512008494956e-05, "loss": 6.5328, "step": 15128 }, { "epoch": 0.6378565254969749, "grad_norm": 3.337628126144409, "learning_rate": 4.374433929854149e-05, "loss": 7.3576, "step": 15129 }, { "epoch": 0.6378986866791745, "grad_norm": 0.7552186846733093, "learning_rate": 4.374355851213342e-05, "loss": 6.3583, "step": 15130 }, { "epoch": 0.6379408478613741, "grad_norm": 0.898926854133606, "learning_rate": 4.3742777725725355e-05, "loss": 6.4191, "step": 15131 }, { "epoch": 0.6379830090435736, "grad_norm": 0.9891159534454346, "learning_rate": 4.374199693931728e-05, "loss": 6.6267, "step": 15132 }, { "epoch": 0.6380251702257731, "grad_norm": 2.48343825340271, "learning_rate": 4.374121615290922e-05, "loss": 6.5016, "step": 15133 }, { "epoch": 0.6380673314079727, "grad_norm": 3.6491496562957764, "learning_rate": 4.374043536650114e-05, "loss": 8.8083, "step": 15134 }, { "epoch": 0.6381094925901722, "grad_norm": 1.7414931058883667, "learning_rate": 4.3739654580093074e-05, "loss": 6.9891, "step": 15135 }, { "epoch": 0.6381516537723718, "grad_norm": 3.622474193572998, "learning_rate": 4.3738873793685005e-05, "loss": 7.5792, "step": 15136 }, { "epoch": 0.6381938149545713, "grad_norm": 3.2155368328094482, "learning_rate": 4.373809300727693e-05, "loss": 7.6112, "step": 15137 }, { "epoch": 0.6382359761367709, "grad_norm": 2.0697197914123535, "learning_rate": 4.373731222086886e-05, "loss": 6.8868, "step": 15138 }, { "epoch": 0.6382781373189704, "grad_norm": 1.489532232284546, "learning_rate": 4.373653143446079e-05, "loss": 6.6135, "step": 15139 }, { "epoch": 0.63832029850117, "grad_norm": 4.447201251983643, "learning_rate": 4.373575064805272e-05, "loss": 7.8616, "step": 15140 }, { "epoch": 0.6383624596833696, "grad_norm": 4.698927402496338, "learning_rate": 4.373496986164465e-05, "loss": 7.7263, "step": 15141 }, { "epoch": 0.638404620865569, "grad_norm": 2.9777417182922363, "learning_rate": 4.373418907523658e-05, "loss": 6.5806, "step": 15142 }, { "epoch": 0.6384467820477686, "grad_norm": 2.0482113361358643, "learning_rate": 4.373340828882851e-05, "loss": 6.6083, "step": 15143 }, { "epoch": 0.6384889432299682, "grad_norm": 2.5264689922332764, "learning_rate": 4.373262750242044e-05, "loss": 6.9266, "step": 15144 }, { "epoch": 0.6385311044121678, "grad_norm": 2.4905714988708496, "learning_rate": 4.373184671601237e-05, "loss": 6.5076, "step": 15145 }, { "epoch": 0.6385732655943672, "grad_norm": 3.380115509033203, "learning_rate": 4.37310659296043e-05, "loss": 6.7754, "step": 15146 }, { "epoch": 0.6386154267765668, "grad_norm": 3.847080707550049, "learning_rate": 4.373028514319623e-05, "loss": 6.3419, "step": 15147 }, { "epoch": 0.6386575879587664, "grad_norm": 3.457444906234741, "learning_rate": 4.372950435678816e-05, "loss": 7.5737, "step": 15148 }, { "epoch": 0.6386997491409659, "grad_norm": 1.633480191230774, "learning_rate": 4.3728723570380084e-05, "loss": 7.8004, "step": 15149 }, { "epoch": 0.6387419103231654, "grad_norm": 2.463343620300293, "learning_rate": 4.372794278397202e-05, "loss": 7.0151, "step": 15150 }, { "epoch": 0.638784071505365, "grad_norm": 1.2767943143844604, "learning_rate": 4.3727161997563947e-05, "loss": 6.5612, "step": 15151 }, { "epoch": 0.6388262326875646, "grad_norm": 4.149812698364258, "learning_rate": 4.372638121115588e-05, "loss": 7.8051, "step": 15152 }, { "epoch": 0.6388683938697641, "grad_norm": 3.8361260890960693, "learning_rate": 4.372560042474781e-05, "loss": 7.7882, "step": 15153 }, { "epoch": 0.6389105550519637, "grad_norm": 3.8628194332122803, "learning_rate": 4.3724819638339734e-05, "loss": 6.4778, "step": 15154 }, { "epoch": 0.6389527162341632, "grad_norm": 1.3443565368652344, "learning_rate": 4.372403885193167e-05, "loss": 6.629, "step": 15155 }, { "epoch": 0.6389948774163627, "grad_norm": 2.273632049560547, "learning_rate": 4.3723258065523596e-05, "loss": 6.4909, "step": 15156 }, { "epoch": 0.6390370385985623, "grad_norm": 3.80607008934021, "learning_rate": 4.372247727911553e-05, "loss": 8.2617, "step": 15157 }, { "epoch": 0.6390791997807619, "grad_norm": 2.791698694229126, "learning_rate": 4.372169649270746e-05, "loss": 7.1018, "step": 15158 }, { "epoch": 0.6391213609629613, "grad_norm": 4.271052360534668, "learning_rate": 4.372091570629939e-05, "loss": 7.4389, "step": 15159 }, { "epoch": 0.6391635221451609, "grad_norm": 2.075140953063965, "learning_rate": 4.3720134919891314e-05, "loss": 7.3496, "step": 15160 }, { "epoch": 0.6392056833273605, "grad_norm": 1.7338981628417969, "learning_rate": 4.3719354133483246e-05, "loss": 7.1104, "step": 15161 }, { "epoch": 0.6392478445095601, "grad_norm": 1.8697055578231812, "learning_rate": 4.371857334707518e-05, "loss": 7.3189, "step": 15162 }, { "epoch": 0.6392900056917596, "grad_norm": 1.8452084064483643, "learning_rate": 4.37177925606671e-05, "loss": 6.3043, "step": 15163 }, { "epoch": 0.6393321668739591, "grad_norm": 1.5139310359954834, "learning_rate": 4.371701177425904e-05, "loss": 6.6659, "step": 15164 }, { "epoch": 0.6393743280561587, "grad_norm": 1.6129993200302124, "learning_rate": 4.3716230987850964e-05, "loss": 6.9709, "step": 15165 }, { "epoch": 0.6394164892383583, "grad_norm": 1.7025896310806274, "learning_rate": 4.3715450201442895e-05, "loss": 6.2659, "step": 15166 }, { "epoch": 0.6394586504205578, "grad_norm": 1.7958040237426758, "learning_rate": 4.3714669415034826e-05, "loss": 6.9543, "step": 15167 }, { "epoch": 0.6395008116027573, "grad_norm": 1.4797592163085938, "learning_rate": 4.371388862862675e-05, "loss": 6.8008, "step": 15168 }, { "epoch": 0.6395429727849569, "grad_norm": 1.7319905757904053, "learning_rate": 4.371310784221869e-05, "loss": 6.8881, "step": 15169 }, { "epoch": 0.6395851339671564, "grad_norm": 2.4274544715881348, "learning_rate": 4.371232705581061e-05, "loss": 7.5122, "step": 15170 }, { "epoch": 0.639627295149356, "grad_norm": 3.4714813232421875, "learning_rate": 4.3711546269402545e-05, "loss": 7.243, "step": 15171 }, { "epoch": 0.6396694563315556, "grad_norm": 1.9681682586669922, "learning_rate": 4.3710765482994476e-05, "loss": 7.0272, "step": 15172 }, { "epoch": 0.639711617513755, "grad_norm": 2.0092999935150146, "learning_rate": 4.37099846965864e-05, "loss": 6.3503, "step": 15173 }, { "epoch": 0.6397537786959546, "grad_norm": 2.556276798248291, "learning_rate": 4.370920391017833e-05, "loss": 7.449, "step": 15174 }, { "epoch": 0.6397959398781542, "grad_norm": 1.0937330722808838, "learning_rate": 4.370842312377026e-05, "loss": 6.6796, "step": 15175 }, { "epoch": 0.6398381010603538, "grad_norm": 1.5695732831954956, "learning_rate": 4.3707642337362194e-05, "loss": 6.5076, "step": 15176 }, { "epoch": 0.6398802622425532, "grad_norm": 1.9681036472320557, "learning_rate": 4.3706861550954125e-05, "loss": 6.4141, "step": 15177 }, { "epoch": 0.6399224234247528, "grad_norm": 1.4832335710525513, "learning_rate": 4.370608076454606e-05, "loss": 7.321, "step": 15178 }, { "epoch": 0.6399645846069524, "grad_norm": 4.194911479949951, "learning_rate": 4.370529997813798e-05, "loss": 7.3048, "step": 15179 }, { "epoch": 0.640006745789152, "grad_norm": 2.573683738708496, "learning_rate": 4.370451919172991e-05, "loss": 6.9961, "step": 15180 }, { "epoch": 0.6400489069713515, "grad_norm": 4.188765048980713, "learning_rate": 4.3703738405321844e-05, "loss": 7.7417, "step": 15181 }, { "epoch": 0.640091068153551, "grad_norm": 3.4882309436798096, "learning_rate": 4.370295761891377e-05, "loss": 7.6427, "step": 15182 }, { "epoch": 0.6401332293357506, "grad_norm": 1.692472219467163, "learning_rate": 4.3702176832505706e-05, "loss": 6.7329, "step": 15183 }, { "epoch": 0.6401753905179501, "grad_norm": 2.4055233001708984, "learning_rate": 4.370139604609763e-05, "loss": 6.5208, "step": 15184 }, { "epoch": 0.6402175517001497, "grad_norm": 2.5239474773406982, "learning_rate": 4.370061525968956e-05, "loss": 6.9012, "step": 15185 }, { "epoch": 0.6402597128823492, "grad_norm": 3.944985866546631, "learning_rate": 4.369983447328149e-05, "loss": 8.8601, "step": 15186 }, { "epoch": 0.6403018740645487, "grad_norm": 2.508141040802002, "learning_rate": 4.369905368687342e-05, "loss": 6.4616, "step": 15187 }, { "epoch": 0.6403440352467483, "grad_norm": 1.6524769067764282, "learning_rate": 4.3698272900465356e-05, "loss": 6.9573, "step": 15188 }, { "epoch": 0.6403861964289479, "grad_norm": 2.1987788677215576, "learning_rate": 4.369749211405728e-05, "loss": 6.9192, "step": 15189 }, { "epoch": 0.6404283576111475, "grad_norm": 1.4928239583969116, "learning_rate": 4.369671132764921e-05, "loss": 6.3882, "step": 15190 }, { "epoch": 0.6404705187933469, "grad_norm": 2.900256633758545, "learning_rate": 4.369593054124114e-05, "loss": 6.3323, "step": 15191 }, { "epoch": 0.6405126799755465, "grad_norm": 3.2945241928100586, "learning_rate": 4.369514975483307e-05, "loss": 7.1338, "step": 15192 }, { "epoch": 0.6405548411577461, "grad_norm": 2.298624277114868, "learning_rate": 4.3694368968425e-05, "loss": 6.5958, "step": 15193 }, { "epoch": 0.6405970023399457, "grad_norm": 1.0667897462844849, "learning_rate": 4.369358818201693e-05, "loss": 6.4374, "step": 15194 }, { "epoch": 0.6406391635221451, "grad_norm": 3.14420485496521, "learning_rate": 4.369280739560886e-05, "loss": 7.5375, "step": 15195 }, { "epoch": 0.6406813247043447, "grad_norm": 1.2130897045135498, "learning_rate": 4.3692026609200785e-05, "loss": 6.1806, "step": 15196 }, { "epoch": 0.6407234858865443, "grad_norm": 2.409615993499756, "learning_rate": 4.3691245822792724e-05, "loss": 6.9902, "step": 15197 }, { "epoch": 0.6407656470687438, "grad_norm": 1.965262532234192, "learning_rate": 4.369046503638465e-05, "loss": 6.5788, "step": 15198 }, { "epoch": 0.6408078082509434, "grad_norm": 1.8066776990890503, "learning_rate": 4.368968424997658e-05, "loss": 6.677, "step": 15199 }, { "epoch": 0.6408499694331429, "grad_norm": 1.8129100799560547, "learning_rate": 4.368890346356851e-05, "loss": 7.0007, "step": 15200 }, { "epoch": 0.6408921306153424, "grad_norm": 2.32943058013916, "learning_rate": 4.3688122677160435e-05, "loss": 6.4195, "step": 15201 }, { "epoch": 0.640934291797542, "grad_norm": 2.094187021255493, "learning_rate": 4.368734189075237e-05, "loss": 7.0966, "step": 15202 }, { "epoch": 0.6409764529797416, "grad_norm": 2.2708969116210938, "learning_rate": 4.36865611043443e-05, "loss": 7.1076, "step": 15203 }, { "epoch": 0.6410186141619411, "grad_norm": 2.630117177963257, "learning_rate": 4.368578031793623e-05, "loss": 6.1881, "step": 15204 }, { "epoch": 0.6410607753441406, "grad_norm": 3.2165281772613525, "learning_rate": 4.368499953152816e-05, "loss": 7.4227, "step": 15205 }, { "epoch": 0.6411029365263402, "grad_norm": 3.962264060974121, "learning_rate": 4.3684218745120085e-05, "loss": 7.8206, "step": 15206 }, { "epoch": 0.6411450977085398, "grad_norm": 1.6480073928833008, "learning_rate": 4.3683437958712016e-05, "loss": 6.717, "step": 15207 }, { "epoch": 0.6411872588907392, "grad_norm": 1.5457478761672974, "learning_rate": 4.368265717230395e-05, "loss": 6.402, "step": 15208 }, { "epoch": 0.6412294200729388, "grad_norm": 3.0954065322875977, "learning_rate": 4.368187638589588e-05, "loss": 6.2069, "step": 15209 }, { "epoch": 0.6412715812551384, "grad_norm": 1.9514203071594238, "learning_rate": 4.36810955994878e-05, "loss": 6.2776, "step": 15210 }, { "epoch": 0.641313742437338, "grad_norm": 1.5848859548568726, "learning_rate": 4.3680314813079734e-05, "loss": 6.9087, "step": 15211 }, { "epoch": 0.6413559036195375, "grad_norm": 3.30220365524292, "learning_rate": 4.3679534026671665e-05, "loss": 7.4769, "step": 15212 }, { "epoch": 0.641398064801737, "grad_norm": 3.229856491088867, "learning_rate": 4.3678753240263597e-05, "loss": 7.5991, "step": 15213 }, { "epoch": 0.6414402259839366, "grad_norm": 1.8817059993743896, "learning_rate": 4.367797245385553e-05, "loss": 6.3872, "step": 15214 }, { "epoch": 0.6414823871661361, "grad_norm": 1.9378528594970703, "learning_rate": 4.367719166744745e-05, "loss": 6.7641, "step": 15215 }, { "epoch": 0.6415245483483357, "grad_norm": 1.3380506038665771, "learning_rate": 4.3676410881039384e-05, "loss": 6.927, "step": 15216 }, { "epoch": 0.6415667095305352, "grad_norm": 3.3360073566436768, "learning_rate": 4.3675630094631315e-05, "loss": 7.3225, "step": 15217 }, { "epoch": 0.6416088707127348, "grad_norm": 1.060240626335144, "learning_rate": 4.367484930822324e-05, "loss": 6.3584, "step": 15218 }, { "epoch": 0.6416510318949343, "grad_norm": 1.5517688989639282, "learning_rate": 4.367406852181518e-05, "loss": 6.6397, "step": 15219 }, { "epoch": 0.6416931930771339, "grad_norm": 4.018189430236816, "learning_rate": 4.36732877354071e-05, "loss": 7.6717, "step": 15220 }, { "epoch": 0.6417353542593335, "grad_norm": 1.4253654479980469, "learning_rate": 4.367250694899903e-05, "loss": 6.4926, "step": 15221 }, { "epoch": 0.6417775154415329, "grad_norm": 2.657585859298706, "learning_rate": 4.3671726162590964e-05, "loss": 8.139, "step": 15222 }, { "epoch": 0.6418196766237325, "grad_norm": 2.292694330215454, "learning_rate": 4.3670945376182896e-05, "loss": 6.7853, "step": 15223 }, { "epoch": 0.6418618378059321, "grad_norm": 2.061821222305298, "learning_rate": 4.367016458977483e-05, "loss": 6.9987, "step": 15224 }, { "epoch": 0.6419039989881317, "grad_norm": 2.243967056274414, "learning_rate": 4.366938380336675e-05, "loss": 6.2648, "step": 15225 }, { "epoch": 0.6419461601703311, "grad_norm": 1.169089436531067, "learning_rate": 4.366860301695868e-05, "loss": 6.8304, "step": 15226 }, { "epoch": 0.6419883213525307, "grad_norm": 1.550149917602539, "learning_rate": 4.3667822230550614e-05, "loss": 6.3258, "step": 15227 }, { "epoch": 0.6420304825347303, "grad_norm": 1.8822859525680542, "learning_rate": 4.3667041444142545e-05, "loss": 6.6924, "step": 15228 }, { "epoch": 0.6420726437169298, "grad_norm": 3.5705418586730957, "learning_rate": 4.366626065773447e-05, "loss": 6.6726, "step": 15229 }, { "epoch": 0.6421148048991294, "grad_norm": 1.8394206762313843, "learning_rate": 4.36654798713264e-05, "loss": 6.3412, "step": 15230 }, { "epoch": 0.6421569660813289, "grad_norm": 2.5399374961853027, "learning_rate": 4.366469908491833e-05, "loss": 6.7789, "step": 15231 }, { "epoch": 0.6421991272635285, "grad_norm": 1.4729828834533691, "learning_rate": 4.366391829851026e-05, "loss": 7.0794, "step": 15232 }, { "epoch": 0.642241288445728, "grad_norm": 1.6596345901489258, "learning_rate": 4.3663137512102195e-05, "loss": 6.577, "step": 15233 }, { "epoch": 0.6422834496279276, "grad_norm": 3.2580528259277344, "learning_rate": 4.366235672569412e-05, "loss": 7.5679, "step": 15234 }, { "epoch": 0.6423256108101271, "grad_norm": 2.6257033348083496, "learning_rate": 4.366157593928605e-05, "loss": 6.4154, "step": 15235 }, { "epoch": 0.6423677719923266, "grad_norm": 1.4336612224578857, "learning_rate": 4.366079515287798e-05, "loss": 6.89, "step": 15236 }, { "epoch": 0.6424099331745262, "grad_norm": 1.7141729593276978, "learning_rate": 4.3660014366469906e-05, "loss": 6.6551, "step": 15237 }, { "epoch": 0.6424520943567258, "grad_norm": 4.600090026855469, "learning_rate": 4.3659233580061844e-05, "loss": 7.6647, "step": 15238 }, { "epoch": 0.6424942555389254, "grad_norm": 2.4500861167907715, "learning_rate": 4.365845279365377e-05, "loss": 6.7956, "step": 15239 }, { "epoch": 0.6425364167211248, "grad_norm": 3.1412136554718018, "learning_rate": 4.36576720072457e-05, "loss": 7.7766, "step": 15240 }, { "epoch": 0.6425785779033244, "grad_norm": 3.028226852416992, "learning_rate": 4.365689122083763e-05, "loss": 7.3581, "step": 15241 }, { "epoch": 0.642620739085524, "grad_norm": 2.4386820793151855, "learning_rate": 4.3656110434429556e-05, "loss": 6.6351, "step": 15242 }, { "epoch": 0.6426629002677235, "grad_norm": 2.5629117488861084, "learning_rate": 4.365532964802149e-05, "loss": 6.6892, "step": 15243 }, { "epoch": 0.642705061449923, "grad_norm": 2.248072624206543, "learning_rate": 4.365454886161342e-05, "loss": 7.5826, "step": 15244 }, { "epoch": 0.6427472226321226, "grad_norm": 1.9477955102920532, "learning_rate": 4.365376807520535e-05, "loss": 6.3977, "step": 15245 }, { "epoch": 0.6427893838143222, "grad_norm": 2.5336811542510986, "learning_rate": 4.365298728879728e-05, "loss": 7.1115, "step": 15246 }, { "epoch": 0.6428315449965217, "grad_norm": 1.7461122274398804, "learning_rate": 4.365220650238921e-05, "loss": 6.487, "step": 15247 }, { "epoch": 0.6428737061787213, "grad_norm": 1.801558494567871, "learning_rate": 4.3651425715981136e-05, "loss": 6.7609, "step": 15248 }, { "epoch": 0.6429158673609208, "grad_norm": 1.540289282798767, "learning_rate": 4.365064492957307e-05, "loss": 6.2994, "step": 15249 }, { "epoch": 0.6429580285431203, "grad_norm": 2.0399513244628906, "learning_rate": 4.3649864143165e-05, "loss": 6.355, "step": 15250 }, { "epoch": 0.6430001897253199, "grad_norm": 2.353053569793701, "learning_rate": 4.3649083356756923e-05, "loss": 6.9096, "step": 15251 }, { "epoch": 0.6430423509075195, "grad_norm": 1.710946798324585, "learning_rate": 4.364830257034886e-05, "loss": 7.1882, "step": 15252 }, { "epoch": 0.643084512089719, "grad_norm": 2.0751380920410156, "learning_rate": 4.3647521783940786e-05, "loss": 6.5014, "step": 15253 }, { "epoch": 0.6431266732719185, "grad_norm": 2.0627474784851074, "learning_rate": 4.364674099753272e-05, "loss": 6.8383, "step": 15254 }, { "epoch": 0.6431688344541181, "grad_norm": 2.6827750205993652, "learning_rate": 4.364596021112465e-05, "loss": 6.4786, "step": 15255 }, { "epoch": 0.6432109956363177, "grad_norm": 1.9287610054016113, "learning_rate": 4.364517942471657e-05, "loss": 6.778, "step": 15256 }, { "epoch": 0.6432531568185172, "grad_norm": 2.9212636947631836, "learning_rate": 4.364439863830851e-05, "loss": 7.2137, "step": 15257 }, { "epoch": 0.6432953180007167, "grad_norm": 2.233257293701172, "learning_rate": 4.3643617851900436e-05, "loss": 7.178, "step": 15258 }, { "epoch": 0.6433374791829163, "grad_norm": 1.6236704587936401, "learning_rate": 4.364283706549237e-05, "loss": 6.296, "step": 15259 }, { "epoch": 0.6433796403651159, "grad_norm": 1.6719783544540405, "learning_rate": 4.36420562790843e-05, "loss": 6.7765, "step": 15260 }, { "epoch": 0.6434218015473154, "grad_norm": 3.135327100753784, "learning_rate": 4.364127549267622e-05, "loss": 8.9068, "step": 15261 }, { "epoch": 0.6434639627295149, "grad_norm": 3.107593536376953, "learning_rate": 4.3640494706268154e-05, "loss": 6.6234, "step": 15262 }, { "epoch": 0.6435061239117145, "grad_norm": 2.415285348892212, "learning_rate": 4.3639713919860085e-05, "loss": 7.6267, "step": 15263 }, { "epoch": 0.643548285093914, "grad_norm": 2.541337490081787, "learning_rate": 4.3638933133452016e-05, "loss": 6.637, "step": 15264 }, { "epoch": 0.6435904462761136, "grad_norm": 1.5673378705978394, "learning_rate": 4.363815234704394e-05, "loss": 7.1541, "step": 15265 }, { "epoch": 0.6436326074583131, "grad_norm": 2.21036434173584, "learning_rate": 4.363737156063588e-05, "loss": 6.3146, "step": 15266 }, { "epoch": 0.6436747686405127, "grad_norm": 1.8798669576644897, "learning_rate": 4.36365907742278e-05, "loss": 6.4617, "step": 15267 }, { "epoch": 0.6437169298227122, "grad_norm": 1.7764892578125, "learning_rate": 4.3635809987819735e-05, "loss": 6.3578, "step": 15268 }, { "epoch": 0.6437590910049118, "grad_norm": 2.6724510192871094, "learning_rate": 4.3635029201411666e-05, "loss": 6.989, "step": 15269 }, { "epoch": 0.6438012521871114, "grad_norm": 1.8050806522369385, "learning_rate": 4.363424841500359e-05, "loss": 6.946, "step": 15270 }, { "epoch": 0.6438434133693108, "grad_norm": 1.8515549898147583, "learning_rate": 4.363346762859553e-05, "loss": 6.4366, "step": 15271 }, { "epoch": 0.6438855745515104, "grad_norm": 2.8808634281158447, "learning_rate": 4.363268684218745e-05, "loss": 7.3619, "step": 15272 }, { "epoch": 0.64392773573371, "grad_norm": 2.0821962356567383, "learning_rate": 4.3631906055779384e-05, "loss": 6.4088, "step": 15273 }, { "epoch": 0.6439698969159096, "grad_norm": 3.0976321697235107, "learning_rate": 4.3631125269371315e-05, "loss": 6.2757, "step": 15274 }, { "epoch": 0.644012058098109, "grad_norm": 2.620816469192505, "learning_rate": 4.363034448296324e-05, "loss": 7.7527, "step": 15275 }, { "epoch": 0.6440542192803086, "grad_norm": 2.1289634704589844, "learning_rate": 4.362956369655517e-05, "loss": 7.0395, "step": 15276 }, { "epoch": 0.6440963804625082, "grad_norm": 0.8602120280265808, "learning_rate": 4.36287829101471e-05, "loss": 6.4023, "step": 15277 }, { "epoch": 0.6441385416447077, "grad_norm": 1.3767173290252686, "learning_rate": 4.3628002123739034e-05, "loss": 6.4398, "step": 15278 }, { "epoch": 0.6441807028269073, "grad_norm": 1.087644338607788, "learning_rate": 4.362722133733096e-05, "loss": 6.2869, "step": 15279 }, { "epoch": 0.6442228640091068, "grad_norm": 3.804081678390503, "learning_rate": 4.362644055092289e-05, "loss": 7.0552, "step": 15280 }, { "epoch": 0.6442650251913064, "grad_norm": 2.747004508972168, "learning_rate": 4.362565976451482e-05, "loss": 7.8162, "step": 15281 }, { "epoch": 0.6443071863735059, "grad_norm": 2.8360860347747803, "learning_rate": 4.362487897810675e-05, "loss": 6.2411, "step": 15282 }, { "epoch": 0.6443493475557055, "grad_norm": 2.2734320163726807, "learning_rate": 4.362409819169868e-05, "loss": 6.8427, "step": 15283 }, { "epoch": 0.644391508737905, "grad_norm": 2.4782004356384277, "learning_rate": 4.362331740529061e-05, "loss": 6.4683, "step": 15284 }, { "epoch": 0.6444336699201045, "grad_norm": 1.230851411819458, "learning_rate": 4.3622536618882546e-05, "loss": 6.8054, "step": 15285 }, { "epoch": 0.6444758311023041, "grad_norm": 1.4237315654754639, "learning_rate": 4.362175583247447e-05, "loss": 6.6078, "step": 15286 }, { "epoch": 0.6445179922845037, "grad_norm": 2.4303689002990723, "learning_rate": 4.3620975046066395e-05, "loss": 6.5971, "step": 15287 }, { "epoch": 0.6445601534667033, "grad_norm": 2.4109199047088623, "learning_rate": 4.362019425965833e-05, "loss": 7.1345, "step": 15288 }, { "epoch": 0.6446023146489027, "grad_norm": 2.4983959197998047, "learning_rate": 4.361941347325026e-05, "loss": 6.2326, "step": 15289 }, { "epoch": 0.6446444758311023, "grad_norm": 2.067411422729492, "learning_rate": 4.361863268684219e-05, "loss": 6.6145, "step": 15290 }, { "epoch": 0.6446866370133019, "grad_norm": 3.067551851272583, "learning_rate": 4.361785190043412e-05, "loss": 6.9566, "step": 15291 }, { "epoch": 0.6447287981955014, "grad_norm": 1.6110692024230957, "learning_rate": 4.361707111402605e-05, "loss": 6.4327, "step": 15292 }, { "epoch": 0.6447709593777009, "grad_norm": 1.772041916847229, "learning_rate": 4.361629032761798e-05, "loss": 6.8611, "step": 15293 }, { "epoch": 0.6448131205599005, "grad_norm": 1.5040457248687744, "learning_rate": 4.361550954120991e-05, "loss": 6.5099, "step": 15294 }, { "epoch": 0.6448552817421, "grad_norm": 1.8983019590377808, "learning_rate": 4.361472875480184e-05, "loss": 6.7049, "step": 15295 }, { "epoch": 0.6448974429242996, "grad_norm": 1.317762017250061, "learning_rate": 4.361394796839377e-05, "loss": 6.4128, "step": 15296 }, { "epoch": 0.6449396041064992, "grad_norm": 1.5670808553695679, "learning_rate": 4.36131671819857e-05, "loss": 6.3646, "step": 15297 }, { "epoch": 0.6449817652886987, "grad_norm": 2.0795512199401855, "learning_rate": 4.3612386395577625e-05, "loss": 7.1496, "step": 15298 }, { "epoch": 0.6450239264708982, "grad_norm": 2.2343246936798096, "learning_rate": 4.3611605609169556e-05, "loss": 7.0111, "step": 15299 }, { "epoch": 0.6450660876530978, "grad_norm": 2.500581741333008, "learning_rate": 4.361082482276149e-05, "loss": 6.4678, "step": 15300 }, { "epoch": 0.6451082488352974, "grad_norm": 2.469587564468384, "learning_rate": 4.361004403635341e-05, "loss": 6.8551, "step": 15301 }, { "epoch": 0.6451504100174968, "grad_norm": 1.7286385297775269, "learning_rate": 4.360926324994535e-05, "loss": 6.6728, "step": 15302 }, { "epoch": 0.6451925711996964, "grad_norm": 1.5371953248977661, "learning_rate": 4.3608482463537274e-05, "loss": 6.4069, "step": 15303 }, { "epoch": 0.645234732381896, "grad_norm": 1.2458442449569702, "learning_rate": 4.3607701677129206e-05, "loss": 6.3082, "step": 15304 }, { "epoch": 0.6452768935640956, "grad_norm": 2.8859505653381348, "learning_rate": 4.360692089072114e-05, "loss": 6.7125, "step": 15305 }, { "epoch": 0.6453190547462951, "grad_norm": 3.2216320037841797, "learning_rate": 4.360614010431306e-05, "loss": 7.0906, "step": 15306 }, { "epoch": 0.6453612159284946, "grad_norm": 1.3839792013168335, "learning_rate": 4.3605359317905e-05, "loss": 6.3894, "step": 15307 }, { "epoch": 0.6454033771106942, "grad_norm": 1.19708251953125, "learning_rate": 4.3604578531496924e-05, "loss": 6.3684, "step": 15308 }, { "epoch": 0.6454455382928938, "grad_norm": 5.12266206741333, "learning_rate": 4.3603797745088855e-05, "loss": 7.9129, "step": 15309 }, { "epoch": 0.6454876994750933, "grad_norm": 2.99569034576416, "learning_rate": 4.3603016958680786e-05, "loss": 7.5739, "step": 15310 }, { "epoch": 0.6455298606572928, "grad_norm": 1.9928398132324219, "learning_rate": 4.360223617227272e-05, "loss": 6.8146, "step": 15311 }, { "epoch": 0.6455720218394924, "grad_norm": 2.34425687789917, "learning_rate": 4.360145538586464e-05, "loss": 7.3323, "step": 15312 }, { "epoch": 0.6456141830216919, "grad_norm": 1.9201899766921997, "learning_rate": 4.3600674599456573e-05, "loss": 6.4059, "step": 15313 }, { "epoch": 0.6456563442038915, "grad_norm": 1.2986242771148682, "learning_rate": 4.3599893813048505e-05, "loss": 6.4291, "step": 15314 }, { "epoch": 0.645698505386091, "grad_norm": 2.6647696495056152, "learning_rate": 4.3599113026640436e-05, "loss": 7.4172, "step": 15315 }, { "epoch": 0.6457406665682905, "grad_norm": 1.7625287771224976, "learning_rate": 4.359833224023237e-05, "loss": 6.5906, "step": 15316 }, { "epoch": 0.6457828277504901, "grad_norm": 1.7294445037841797, "learning_rate": 4.359755145382429e-05, "loss": 6.411, "step": 15317 }, { "epoch": 0.6458249889326897, "grad_norm": 2.0960981845855713, "learning_rate": 4.359677066741622e-05, "loss": 6.3572, "step": 15318 }, { "epoch": 0.6458671501148893, "grad_norm": 2.571213483810425, "learning_rate": 4.3595989881008154e-05, "loss": 6.6331, "step": 15319 }, { "epoch": 0.6459093112970887, "grad_norm": 2.2023701667785645, "learning_rate": 4.359520909460008e-05, "loss": 6.9471, "step": 15320 }, { "epoch": 0.6459514724792883, "grad_norm": 1.5331685543060303, "learning_rate": 4.359442830819202e-05, "loss": 6.6106, "step": 15321 }, { "epoch": 0.6459936336614879, "grad_norm": 1.6719287633895874, "learning_rate": 4.359364752178394e-05, "loss": 6.4014, "step": 15322 }, { "epoch": 0.6460357948436874, "grad_norm": 1.843235731124878, "learning_rate": 4.359286673537587e-05, "loss": 6.5521, "step": 15323 }, { "epoch": 0.6460779560258869, "grad_norm": 3.0276761054992676, "learning_rate": 4.3592085948967804e-05, "loss": 7.9165, "step": 15324 }, { "epoch": 0.6461201172080865, "grad_norm": 2.0850703716278076, "learning_rate": 4.359130516255973e-05, "loss": 6.379, "step": 15325 }, { "epoch": 0.6461622783902861, "grad_norm": 1.9729143381118774, "learning_rate": 4.3590524376151666e-05, "loss": 6.901, "step": 15326 }, { "epoch": 0.6462044395724856, "grad_norm": 1.5193341970443726, "learning_rate": 4.358974358974359e-05, "loss": 6.9131, "step": 15327 }, { "epoch": 0.6462466007546852, "grad_norm": 2.3048207759857178, "learning_rate": 4.358896280333552e-05, "loss": 6.5812, "step": 15328 }, { "epoch": 0.6462887619368847, "grad_norm": 2.9537465572357178, "learning_rate": 4.358818201692745e-05, "loss": 7.6302, "step": 15329 }, { "epoch": 0.6463309231190842, "grad_norm": 3.2812931537628174, "learning_rate": 4.3587401230519385e-05, "loss": 6.93, "step": 15330 }, { "epoch": 0.6463730843012838, "grad_norm": 2.4581665992736816, "learning_rate": 4.358662044411131e-05, "loss": 7.6676, "step": 15331 }, { "epoch": 0.6464152454834834, "grad_norm": 2.7322185039520264, "learning_rate": 4.358583965770324e-05, "loss": 7.0261, "step": 15332 }, { "epoch": 0.6464574066656829, "grad_norm": 2.13273549079895, "learning_rate": 4.358505887129517e-05, "loss": 6.4842, "step": 15333 }, { "epoch": 0.6464995678478824, "grad_norm": 1.8932764530181885, "learning_rate": 4.3584278084887096e-05, "loss": 6.5922, "step": 15334 }, { "epoch": 0.646541729030082, "grad_norm": 1.314401388168335, "learning_rate": 4.3583497298479034e-05, "loss": 6.4821, "step": 15335 }, { "epoch": 0.6465838902122816, "grad_norm": 3.1714446544647217, "learning_rate": 4.358271651207096e-05, "loss": 6.7137, "step": 15336 }, { "epoch": 0.6466260513944811, "grad_norm": 1.8041077852249146, "learning_rate": 4.358193572566289e-05, "loss": 6.6353, "step": 15337 }, { "epoch": 0.6466682125766806, "grad_norm": 2.2964730262756348, "learning_rate": 4.358115493925482e-05, "loss": 7.216, "step": 15338 }, { "epoch": 0.6467103737588802, "grad_norm": 2.8495092391967773, "learning_rate": 4.3580374152846746e-05, "loss": 6.6943, "step": 15339 }, { "epoch": 0.6467525349410798, "grad_norm": 2.612222671508789, "learning_rate": 4.3579593366438684e-05, "loss": 6.7735, "step": 15340 }, { "epoch": 0.6467946961232793, "grad_norm": 3.57092022895813, "learning_rate": 4.357881258003061e-05, "loss": 6.3838, "step": 15341 }, { "epoch": 0.6468368573054788, "grad_norm": 3.2478678226470947, "learning_rate": 4.357803179362254e-05, "loss": 6.9537, "step": 15342 }, { "epoch": 0.6468790184876784, "grad_norm": 2.0277278423309326, "learning_rate": 4.357725100721447e-05, "loss": 6.4875, "step": 15343 }, { "epoch": 0.6469211796698779, "grad_norm": 1.3846794366836548, "learning_rate": 4.3576470220806395e-05, "loss": 6.4143, "step": 15344 }, { "epoch": 0.6469633408520775, "grad_norm": 3.3721141815185547, "learning_rate": 4.3575689434398326e-05, "loss": 7.1575, "step": 15345 }, { "epoch": 0.6470055020342771, "grad_norm": 5.92758846282959, "learning_rate": 4.357490864799026e-05, "loss": 7.9328, "step": 15346 }, { "epoch": 0.6470476632164766, "grad_norm": 3.4806015491485596, "learning_rate": 4.357412786158219e-05, "loss": 6.9779, "step": 15347 }, { "epoch": 0.6470898243986761, "grad_norm": 2.3881876468658447, "learning_rate": 4.357334707517412e-05, "loss": 6.6481, "step": 15348 }, { "epoch": 0.6471319855808757, "grad_norm": 2.9204232692718506, "learning_rate": 4.3572566288766045e-05, "loss": 6.602, "step": 15349 }, { "epoch": 0.6471741467630753, "grad_norm": 2.4757814407348633, "learning_rate": 4.3571785502357976e-05, "loss": 7.1137, "step": 15350 }, { "epoch": 0.6472163079452747, "grad_norm": 3.6463944911956787, "learning_rate": 4.357100471594991e-05, "loss": 6.9325, "step": 15351 }, { "epoch": 0.6472584691274743, "grad_norm": 2.9837257862091064, "learning_rate": 4.357022392954184e-05, "loss": 6.8712, "step": 15352 }, { "epoch": 0.6473006303096739, "grad_norm": 2.9975173473358154, "learning_rate": 4.356944314313376e-05, "loss": 7.6703, "step": 15353 }, { "epoch": 0.6473427914918735, "grad_norm": 2.2540788650512695, "learning_rate": 4.35686623567257e-05, "loss": 7.2789, "step": 15354 }, { "epoch": 0.647384952674073, "grad_norm": 3.581550121307373, "learning_rate": 4.3567881570317625e-05, "loss": 7.5908, "step": 15355 }, { "epoch": 0.6474271138562725, "grad_norm": 2.4142656326293945, "learning_rate": 4.356710078390956e-05, "loss": 6.5591, "step": 15356 }, { "epoch": 0.6474692750384721, "grad_norm": 2.16960072517395, "learning_rate": 4.356631999750149e-05, "loss": 6.4251, "step": 15357 }, { "epoch": 0.6475114362206716, "grad_norm": 1.508545160293579, "learning_rate": 4.356553921109341e-05, "loss": 6.4381, "step": 15358 }, { "epoch": 0.6475535974028712, "grad_norm": 2.203375816345215, "learning_rate": 4.3564758424685344e-05, "loss": 6.6044, "step": 15359 }, { "epoch": 0.6475957585850707, "grad_norm": 1.4899582862854004, "learning_rate": 4.3563977638277275e-05, "loss": 6.6664, "step": 15360 }, { "epoch": 0.6476379197672703, "grad_norm": 2.6064987182617188, "learning_rate": 4.3563196851869206e-05, "loss": 6.3917, "step": 15361 }, { "epoch": 0.6476800809494698, "grad_norm": 2.503092050552368, "learning_rate": 4.356241606546114e-05, "loss": 7.4266, "step": 15362 }, { "epoch": 0.6477222421316694, "grad_norm": 2.897869825363159, "learning_rate": 4.356163527905306e-05, "loss": 8.1924, "step": 15363 }, { "epoch": 0.647764403313869, "grad_norm": 1.1489934921264648, "learning_rate": 4.356085449264499e-05, "loss": 6.4645, "step": 15364 }, { "epoch": 0.6478065644960684, "grad_norm": 2.567406415939331, "learning_rate": 4.3560073706236924e-05, "loss": 7.1092, "step": 15365 }, { "epoch": 0.647848725678268, "grad_norm": 1.8745769262313843, "learning_rate": 4.3559292919828856e-05, "loss": 6.9452, "step": 15366 }, { "epoch": 0.6478908868604676, "grad_norm": 2.281900644302368, "learning_rate": 4.355851213342078e-05, "loss": 6.7859, "step": 15367 }, { "epoch": 0.6479330480426672, "grad_norm": 4.3924431800842285, "learning_rate": 4.355773134701271e-05, "loss": 8.2347, "step": 15368 }, { "epoch": 0.6479752092248666, "grad_norm": 2.427170515060425, "learning_rate": 4.355695056060464e-05, "loss": 7.498, "step": 15369 }, { "epoch": 0.6480173704070662, "grad_norm": 3.1728856563568115, "learning_rate": 4.355616977419657e-05, "loss": 6.2436, "step": 15370 }, { "epoch": 0.6480595315892658, "grad_norm": 2.055945873260498, "learning_rate": 4.3555388987788505e-05, "loss": 6.3707, "step": 15371 }, { "epoch": 0.6481016927714653, "grad_norm": 1.3474518060684204, "learning_rate": 4.355460820138043e-05, "loss": 7.1304, "step": 15372 }, { "epoch": 0.6481438539536648, "grad_norm": 1.2160654067993164, "learning_rate": 4.355382741497237e-05, "loss": 6.7458, "step": 15373 }, { "epoch": 0.6481860151358644, "grad_norm": 3.6196563243865967, "learning_rate": 4.355304662856429e-05, "loss": 7.0502, "step": 15374 }, { "epoch": 0.648228176318064, "grad_norm": 1.8160828351974487, "learning_rate": 4.3552265842156224e-05, "loss": 6.217, "step": 15375 }, { "epoch": 0.6482703375002635, "grad_norm": 1.8399369716644287, "learning_rate": 4.3551485055748155e-05, "loss": 6.3257, "step": 15376 }, { "epoch": 0.6483124986824631, "grad_norm": 1.425909161567688, "learning_rate": 4.355070426934008e-05, "loss": 6.6378, "step": 15377 }, { "epoch": 0.6483546598646626, "grad_norm": 2.059284210205078, "learning_rate": 4.354992348293201e-05, "loss": 7.0365, "step": 15378 }, { "epoch": 0.6483968210468621, "grad_norm": 5.629696846008301, "learning_rate": 4.354914269652394e-05, "loss": 9.5586, "step": 15379 }, { "epoch": 0.6484389822290617, "grad_norm": 1.8503713607788086, "learning_rate": 4.354836191011587e-05, "loss": 6.4339, "step": 15380 }, { "epoch": 0.6484811434112613, "grad_norm": 2.4530439376831055, "learning_rate": 4.35475811237078e-05, "loss": 7.1666, "step": 15381 }, { "epoch": 0.6485233045934607, "grad_norm": 1.6752591133117676, "learning_rate": 4.354680033729973e-05, "loss": 6.9044, "step": 15382 }, { "epoch": 0.6485654657756603, "grad_norm": 1.7234066724777222, "learning_rate": 4.354601955089166e-05, "loss": 6.6431, "step": 15383 }, { "epoch": 0.6486076269578599, "grad_norm": 2.7038631439208984, "learning_rate": 4.354523876448359e-05, "loss": 6.4713, "step": 15384 }, { "epoch": 0.6486497881400595, "grad_norm": 1.9267950057983398, "learning_rate": 4.354445797807552e-05, "loss": 6.6642, "step": 15385 }, { "epoch": 0.648691949322259, "grad_norm": 3.709887981414795, "learning_rate": 4.354367719166745e-05, "loss": 7.672, "step": 15386 }, { "epoch": 0.6487341105044585, "grad_norm": 1.7291972637176514, "learning_rate": 4.354289640525938e-05, "loss": 6.7328, "step": 15387 }, { "epoch": 0.6487762716866581, "grad_norm": 3.929076671600342, "learning_rate": 4.354211561885131e-05, "loss": 7.362, "step": 15388 }, { "epoch": 0.6488184328688577, "grad_norm": 3.7337028980255127, "learning_rate": 4.3541334832443234e-05, "loss": 7.099, "step": 15389 }, { "epoch": 0.6488605940510572, "grad_norm": 3.1772990226745605, "learning_rate": 4.354055404603517e-05, "loss": 6.9528, "step": 15390 }, { "epoch": 0.6489027552332567, "grad_norm": 5.3670268058776855, "learning_rate": 4.3539773259627097e-05, "loss": 6.6616, "step": 15391 }, { "epoch": 0.6489449164154563, "grad_norm": 2.4355125427246094, "learning_rate": 4.353899247321903e-05, "loss": 6.5355, "step": 15392 }, { "epoch": 0.6489870775976558, "grad_norm": 2.0221540927886963, "learning_rate": 4.353821168681096e-05, "loss": 7.0594, "step": 15393 }, { "epoch": 0.6490292387798554, "grad_norm": 3.692324638366699, "learning_rate": 4.3537430900402884e-05, "loss": 9.4209, "step": 15394 }, { "epoch": 0.649071399962055, "grad_norm": 3.170072317123413, "learning_rate": 4.353665011399482e-05, "loss": 6.581, "step": 15395 }, { "epoch": 0.6491135611442544, "grad_norm": 1.805403232574463, "learning_rate": 4.3535869327586746e-05, "loss": 6.8206, "step": 15396 }, { "epoch": 0.649155722326454, "grad_norm": 4.141464710235596, "learning_rate": 4.353508854117868e-05, "loss": 7.5516, "step": 15397 }, { "epoch": 0.6491978835086536, "grad_norm": 2.406641721725464, "learning_rate": 4.353430775477061e-05, "loss": 7.2732, "step": 15398 }, { "epoch": 0.6492400446908532, "grad_norm": 2.7973129749298096, "learning_rate": 4.353352696836254e-05, "loss": 7.4565, "step": 15399 }, { "epoch": 0.6492822058730526, "grad_norm": 2.322298765182495, "learning_rate": 4.3532746181954464e-05, "loss": 7.3994, "step": 15400 }, { "epoch": 0.6493243670552522, "grad_norm": 2.208508253097534, "learning_rate": 4.3531965395546396e-05, "loss": 6.6783, "step": 15401 }, { "epoch": 0.6493665282374518, "grad_norm": 2.310032844543457, "learning_rate": 4.353118460913833e-05, "loss": 6.7062, "step": 15402 }, { "epoch": 0.6494086894196514, "grad_norm": 2.373518705368042, "learning_rate": 4.353040382273025e-05, "loss": 6.7711, "step": 15403 }, { "epoch": 0.6494508506018509, "grad_norm": 4.213244915008545, "learning_rate": 4.352962303632219e-05, "loss": 7.9881, "step": 15404 }, { "epoch": 0.6494930117840504, "grad_norm": 4.047142028808594, "learning_rate": 4.3528842249914114e-05, "loss": 7.6433, "step": 15405 }, { "epoch": 0.64953517296625, "grad_norm": 2.7915892601013184, "learning_rate": 4.3528061463506045e-05, "loss": 7.3721, "step": 15406 }, { "epoch": 0.6495773341484495, "grad_norm": 1.4299936294555664, "learning_rate": 4.3527280677097976e-05, "loss": 6.682, "step": 15407 }, { "epoch": 0.6496194953306491, "grad_norm": 2.779193878173828, "learning_rate": 4.35264998906899e-05, "loss": 6.372, "step": 15408 }, { "epoch": 0.6496616565128486, "grad_norm": 1.544974446296692, "learning_rate": 4.352571910428184e-05, "loss": 6.5256, "step": 15409 }, { "epoch": 0.6497038176950481, "grad_norm": 1.874833345413208, "learning_rate": 4.352493831787376e-05, "loss": 6.5135, "step": 15410 }, { "epoch": 0.6497459788772477, "grad_norm": 2.029510259628296, "learning_rate": 4.3524157531465695e-05, "loss": 6.493, "step": 15411 }, { "epoch": 0.6497881400594473, "grad_norm": 1.5835009813308716, "learning_rate": 4.3523376745057626e-05, "loss": 6.3663, "step": 15412 }, { "epoch": 0.6498303012416469, "grad_norm": 2.5497195720672607, "learning_rate": 4.352259595864955e-05, "loss": 7.1388, "step": 15413 }, { "epoch": 0.6498724624238463, "grad_norm": 2.820366859436035, "learning_rate": 4.352181517224148e-05, "loss": 6.4798, "step": 15414 }, { "epoch": 0.6499146236060459, "grad_norm": 2.562696933746338, "learning_rate": 4.352103438583341e-05, "loss": 7.7596, "step": 15415 }, { "epoch": 0.6499567847882455, "grad_norm": 2.601912498474121, "learning_rate": 4.3520253599425344e-05, "loss": 6.5349, "step": 15416 }, { "epoch": 0.649998945970445, "grad_norm": 2.647214889526367, "learning_rate": 4.3519472813017275e-05, "loss": 7.3857, "step": 15417 }, { "epoch": 0.6500411071526445, "grad_norm": 3.12567400932312, "learning_rate": 4.351869202660921e-05, "loss": 7.2085, "step": 15418 }, { "epoch": 0.6500832683348441, "grad_norm": 1.8650139570236206, "learning_rate": 4.351791124020113e-05, "loss": 6.4898, "step": 15419 }, { "epoch": 0.6501254295170437, "grad_norm": 2.0178446769714355, "learning_rate": 4.351713045379306e-05, "loss": 7.1069, "step": 15420 }, { "epoch": 0.6501675906992432, "grad_norm": 1.8318350315093994, "learning_rate": 4.3516349667384994e-05, "loss": 6.7288, "step": 15421 }, { "epoch": 0.6502097518814428, "grad_norm": 1.3792141675949097, "learning_rate": 4.351556888097692e-05, "loss": 6.3466, "step": 15422 }, { "epoch": 0.6502519130636423, "grad_norm": 1.774482011795044, "learning_rate": 4.3514788094568856e-05, "loss": 6.3474, "step": 15423 }, { "epoch": 0.6502940742458418, "grad_norm": 1.7166587114334106, "learning_rate": 4.351400730816078e-05, "loss": 6.8276, "step": 15424 }, { "epoch": 0.6503362354280414, "grad_norm": 1.7695369720458984, "learning_rate": 4.351322652175271e-05, "loss": 6.2798, "step": 15425 }, { "epoch": 0.650378396610241, "grad_norm": 1.3373301029205322, "learning_rate": 4.351244573534464e-05, "loss": 6.6149, "step": 15426 }, { "epoch": 0.6504205577924405, "grad_norm": 2.095320224761963, "learning_rate": 4.351166494893657e-05, "loss": 6.3322, "step": 15427 }, { "epoch": 0.65046271897464, "grad_norm": 1.5732060670852661, "learning_rate": 4.3510884162528506e-05, "loss": 6.666, "step": 15428 }, { "epoch": 0.6505048801568396, "grad_norm": 1.9625848531723022, "learning_rate": 4.351010337612043e-05, "loss": 6.3987, "step": 15429 }, { "epoch": 0.6505470413390392, "grad_norm": 2.161745071411133, "learning_rate": 4.350932258971236e-05, "loss": 6.7847, "step": 15430 }, { "epoch": 0.6505892025212386, "grad_norm": 1.1858899593353271, "learning_rate": 4.350854180330429e-05, "loss": 6.3245, "step": 15431 }, { "epoch": 0.6506313637034382, "grad_norm": 2.3203787803649902, "learning_rate": 4.350776101689622e-05, "loss": 6.4499, "step": 15432 }, { "epoch": 0.6506735248856378, "grad_norm": 1.3873233795166016, "learning_rate": 4.350698023048815e-05, "loss": 6.3339, "step": 15433 }, { "epoch": 0.6507156860678374, "grad_norm": 1.0630416870117188, "learning_rate": 4.350619944408008e-05, "loss": 6.319, "step": 15434 }, { "epoch": 0.6507578472500369, "grad_norm": 3.9621522426605225, "learning_rate": 4.350541865767201e-05, "loss": 7.926, "step": 15435 }, { "epoch": 0.6508000084322364, "grad_norm": 1.4947292804718018, "learning_rate": 4.3504637871263935e-05, "loss": 6.7637, "step": 15436 }, { "epoch": 0.650842169614436, "grad_norm": 1.874883770942688, "learning_rate": 4.3503857084855874e-05, "loss": 6.6375, "step": 15437 }, { "epoch": 0.6508843307966355, "grad_norm": 1.7204163074493408, "learning_rate": 4.35030762984478e-05, "loss": 6.8567, "step": 15438 }, { "epoch": 0.6509264919788351, "grad_norm": 2.1921792030334473, "learning_rate": 4.350229551203972e-05, "loss": 7.2598, "step": 15439 }, { "epoch": 0.6509686531610346, "grad_norm": 1.5938314199447632, "learning_rate": 4.350151472563166e-05, "loss": 6.5254, "step": 15440 }, { "epoch": 0.6510108143432342, "grad_norm": 1.257832407951355, "learning_rate": 4.3500733939223585e-05, "loss": 6.5835, "step": 15441 }, { "epoch": 0.6510529755254337, "grad_norm": 1.1355159282684326, "learning_rate": 4.349995315281552e-05, "loss": 6.5706, "step": 15442 }, { "epoch": 0.6510951367076333, "grad_norm": 1.358863115310669, "learning_rate": 4.349917236640745e-05, "loss": 6.901, "step": 15443 }, { "epoch": 0.6511372978898329, "grad_norm": 1.6005035638809204, "learning_rate": 4.349839157999938e-05, "loss": 6.487, "step": 15444 }, { "epoch": 0.6511794590720323, "grad_norm": 1.7425174713134766, "learning_rate": 4.349761079359131e-05, "loss": 6.6487, "step": 15445 }, { "epoch": 0.6512216202542319, "grad_norm": 2.858915090560913, "learning_rate": 4.3496830007183235e-05, "loss": 7.1355, "step": 15446 }, { "epoch": 0.6512637814364315, "grad_norm": 2.401099443435669, "learning_rate": 4.3496049220775166e-05, "loss": 7.4616, "step": 15447 }, { "epoch": 0.6513059426186311, "grad_norm": 3.084052562713623, "learning_rate": 4.34952684343671e-05, "loss": 7.002, "step": 15448 }, { "epoch": 0.6513481038008305, "grad_norm": 3.0999245643615723, "learning_rate": 4.349448764795903e-05, "loss": 6.3303, "step": 15449 }, { "epoch": 0.6513902649830301, "grad_norm": 2.9761593341827393, "learning_rate": 4.349370686155095e-05, "loss": 6.3206, "step": 15450 }, { "epoch": 0.6514324261652297, "grad_norm": 1.6652374267578125, "learning_rate": 4.3492926075142884e-05, "loss": 6.6397, "step": 15451 }, { "epoch": 0.6514745873474292, "grad_norm": 2.28920841217041, "learning_rate": 4.3492145288734815e-05, "loss": 6.5108, "step": 15452 }, { "epoch": 0.6515167485296288, "grad_norm": 4.527610778808594, "learning_rate": 4.3491364502326747e-05, "loss": 7.0251, "step": 15453 }, { "epoch": 0.6515589097118283, "grad_norm": 2.7480711936950684, "learning_rate": 4.349058371591868e-05, "loss": 6.6289, "step": 15454 }, { "epoch": 0.6516010708940279, "grad_norm": 3.0268356800079346, "learning_rate": 4.34898029295106e-05, "loss": 6.494, "step": 15455 }, { "epoch": 0.6516432320762274, "grad_norm": 1.8232046365737915, "learning_rate": 4.3489022143102534e-05, "loss": 7.1356, "step": 15456 }, { "epoch": 0.651685393258427, "grad_norm": 2.992347002029419, "learning_rate": 4.3488241356694465e-05, "loss": 7.8975, "step": 15457 }, { "epoch": 0.6517275544406265, "grad_norm": 3.5934646129608154, "learning_rate": 4.348746057028639e-05, "loss": 7.459, "step": 15458 }, { "epoch": 0.651769715622826, "grad_norm": 3.5150415897369385, "learning_rate": 4.348667978387833e-05, "loss": 6.7287, "step": 15459 }, { "epoch": 0.6518118768050256, "grad_norm": 3.5267181396484375, "learning_rate": 4.348589899747025e-05, "loss": 7.7837, "step": 15460 }, { "epoch": 0.6518540379872252, "grad_norm": 2.511563777923584, "learning_rate": 4.348511821106218e-05, "loss": 6.356, "step": 15461 }, { "epoch": 0.6518961991694248, "grad_norm": 2.4082868099212646, "learning_rate": 4.3484337424654114e-05, "loss": 7.2402, "step": 15462 }, { "epoch": 0.6519383603516242, "grad_norm": 3.736609935760498, "learning_rate": 4.3483556638246046e-05, "loss": 6.3733, "step": 15463 }, { "epoch": 0.6519805215338238, "grad_norm": 2.4623167514801025, "learning_rate": 4.348277585183798e-05, "loss": 6.8228, "step": 15464 }, { "epoch": 0.6520226827160234, "grad_norm": 4.184895992279053, "learning_rate": 4.34819950654299e-05, "loss": 7.3427, "step": 15465 }, { "epoch": 0.652064843898223, "grad_norm": 3.7502357959747314, "learning_rate": 4.348121427902183e-05, "loss": 7.6205, "step": 15466 }, { "epoch": 0.6521070050804224, "grad_norm": 1.6233563423156738, "learning_rate": 4.3480433492613764e-05, "loss": 6.4397, "step": 15467 }, { "epoch": 0.652149166262622, "grad_norm": 3.9334709644317627, "learning_rate": 4.3479652706205695e-05, "loss": 7.5774, "step": 15468 }, { "epoch": 0.6521913274448216, "grad_norm": 3.4742722511291504, "learning_rate": 4.347887191979762e-05, "loss": 7.6631, "step": 15469 }, { "epoch": 0.6522334886270211, "grad_norm": 3.6802990436553955, "learning_rate": 4.347809113338955e-05, "loss": 6.4816, "step": 15470 }, { "epoch": 0.6522756498092207, "grad_norm": 2.2968461513519287, "learning_rate": 4.347731034698148e-05, "loss": 6.9167, "step": 15471 }, { "epoch": 0.6523178109914202, "grad_norm": 2.7863190174102783, "learning_rate": 4.347652956057341e-05, "loss": 6.2201, "step": 15472 }, { "epoch": 0.6523599721736197, "grad_norm": 1.9178966283798218, "learning_rate": 4.3475748774165345e-05, "loss": 6.7812, "step": 15473 }, { "epoch": 0.6524021333558193, "grad_norm": 3.8939156532287598, "learning_rate": 4.347496798775727e-05, "loss": 7.3803, "step": 15474 }, { "epoch": 0.6524442945380189, "grad_norm": 2.5315206050872803, "learning_rate": 4.34741872013492e-05, "loss": 6.9943, "step": 15475 }, { "epoch": 0.6524864557202184, "grad_norm": 1.4613670110702515, "learning_rate": 4.347340641494113e-05, "loss": 6.6344, "step": 15476 }, { "epoch": 0.6525286169024179, "grad_norm": 1.6350377798080444, "learning_rate": 4.3472625628533056e-05, "loss": 6.7362, "step": 15477 }, { "epoch": 0.6525707780846175, "grad_norm": 2.224743127822876, "learning_rate": 4.3471844842124994e-05, "loss": 6.7227, "step": 15478 }, { "epoch": 0.6526129392668171, "grad_norm": 1.7327638864517212, "learning_rate": 4.347106405571692e-05, "loss": 7.3009, "step": 15479 }, { "epoch": 0.6526551004490166, "grad_norm": 3.590970516204834, "learning_rate": 4.347028326930885e-05, "loss": 7.455, "step": 15480 }, { "epoch": 0.6526972616312161, "grad_norm": 2.9999163150787354, "learning_rate": 4.346950248290078e-05, "loss": 7.2461, "step": 15481 }, { "epoch": 0.6527394228134157, "grad_norm": 1.5550274848937988, "learning_rate": 4.346872169649271e-05, "loss": 6.7387, "step": 15482 }, { "epoch": 0.6527815839956153, "grad_norm": 3.2249033451080322, "learning_rate": 4.346794091008464e-05, "loss": 7.7607, "step": 15483 }, { "epoch": 0.6528237451778148, "grad_norm": 1.7728495597839355, "learning_rate": 4.346716012367657e-05, "loss": 7.5439, "step": 15484 }, { "epoch": 0.6528659063600143, "grad_norm": 2.138686418533325, "learning_rate": 4.34663793372685e-05, "loss": 6.4068, "step": 15485 }, { "epoch": 0.6529080675422139, "grad_norm": 1.6085866689682007, "learning_rate": 4.346559855086043e-05, "loss": 7.1415, "step": 15486 }, { "epoch": 0.6529502287244134, "grad_norm": 3.8548502922058105, "learning_rate": 4.346481776445236e-05, "loss": 7.7362, "step": 15487 }, { "epoch": 0.652992389906613, "grad_norm": 1.9388600587844849, "learning_rate": 4.3464036978044286e-05, "loss": 6.4403, "step": 15488 }, { "epoch": 0.6530345510888125, "grad_norm": 1.3793635368347168, "learning_rate": 4.346325619163622e-05, "loss": 6.8979, "step": 15489 }, { "epoch": 0.653076712271012, "grad_norm": 2.273233652114868, "learning_rate": 4.346247540522815e-05, "loss": 6.5027, "step": 15490 }, { "epoch": 0.6531188734532116, "grad_norm": 3.6147115230560303, "learning_rate": 4.3461694618820073e-05, "loss": 7.7485, "step": 15491 }, { "epoch": 0.6531610346354112, "grad_norm": 3.9782917499542236, "learning_rate": 4.346091383241201e-05, "loss": 8.6562, "step": 15492 }, { "epoch": 0.6532031958176108, "grad_norm": 2.8386099338531494, "learning_rate": 4.3460133046003936e-05, "loss": 7.6889, "step": 15493 }, { "epoch": 0.6532453569998102, "grad_norm": 3.416703224182129, "learning_rate": 4.345935225959587e-05, "loss": 6.513, "step": 15494 }, { "epoch": 0.6532875181820098, "grad_norm": 4.709090709686279, "learning_rate": 4.34585714731878e-05, "loss": 7.1442, "step": 15495 }, { "epoch": 0.6533296793642094, "grad_norm": 2.0441603660583496, "learning_rate": 4.345779068677972e-05, "loss": 6.5184, "step": 15496 }, { "epoch": 0.653371840546409, "grad_norm": 1.6010184288024902, "learning_rate": 4.345700990037166e-05, "loss": 6.8066, "step": 15497 }, { "epoch": 0.6534140017286084, "grad_norm": 3.009246587753296, "learning_rate": 4.3456229113963586e-05, "loss": 6.5057, "step": 15498 }, { "epoch": 0.653456162910808, "grad_norm": 2.817009210586548, "learning_rate": 4.345544832755552e-05, "loss": 6.4866, "step": 15499 }, { "epoch": 0.6534983240930076, "grad_norm": 2.6205735206604004, "learning_rate": 4.345466754114745e-05, "loss": 6.6014, "step": 15500 }, { "epoch": 0.6535404852752071, "grad_norm": 1.797174334526062, "learning_rate": 4.345388675473937e-05, "loss": 6.8929, "step": 15501 }, { "epoch": 0.6535826464574067, "grad_norm": 1.8076835870742798, "learning_rate": 4.3453105968331304e-05, "loss": 6.6492, "step": 15502 }, { "epoch": 0.6536248076396062, "grad_norm": 2.6314754486083984, "learning_rate": 4.3452325181923235e-05, "loss": 7.0325, "step": 15503 }, { "epoch": 0.6536669688218057, "grad_norm": 2.420888662338257, "learning_rate": 4.3451544395515166e-05, "loss": 6.7269, "step": 15504 }, { "epoch": 0.6537091300040053, "grad_norm": 1.7622178792953491, "learning_rate": 4.345076360910709e-05, "loss": 6.4829, "step": 15505 }, { "epoch": 0.6537512911862049, "grad_norm": 3.0791194438934326, "learning_rate": 4.344998282269903e-05, "loss": 7.8308, "step": 15506 }, { "epoch": 0.6537934523684044, "grad_norm": 1.84001624584198, "learning_rate": 4.344920203629095e-05, "loss": 6.4587, "step": 15507 }, { "epoch": 0.6538356135506039, "grad_norm": 3.4962074756622314, "learning_rate": 4.3448421249882885e-05, "loss": 6.924, "step": 15508 }, { "epoch": 0.6538777747328035, "grad_norm": 3.079608678817749, "learning_rate": 4.3447640463474816e-05, "loss": 7.0293, "step": 15509 }, { "epoch": 0.6539199359150031, "grad_norm": 2.7620339393615723, "learning_rate": 4.344685967706674e-05, "loss": 6.9471, "step": 15510 }, { "epoch": 0.6539620970972027, "grad_norm": 2.874678134918213, "learning_rate": 4.344607889065868e-05, "loss": 6.5475, "step": 15511 }, { "epoch": 0.6540042582794021, "grad_norm": 2.8487606048583984, "learning_rate": 4.34452981042506e-05, "loss": 6.622, "step": 15512 }, { "epoch": 0.6540464194616017, "grad_norm": 3.4975783824920654, "learning_rate": 4.3444517317842534e-05, "loss": 6.8164, "step": 15513 }, { "epoch": 0.6540885806438013, "grad_norm": 2.640233039855957, "learning_rate": 4.3443736531434465e-05, "loss": 7.5801, "step": 15514 }, { "epoch": 0.6541307418260008, "grad_norm": 1.8189036846160889, "learning_rate": 4.344295574502639e-05, "loss": 6.3812, "step": 15515 }, { "epoch": 0.6541729030082003, "grad_norm": 2.0797529220581055, "learning_rate": 4.344217495861832e-05, "loss": 6.583, "step": 15516 }, { "epoch": 0.6542150641903999, "grad_norm": 2.9267988204956055, "learning_rate": 4.344139417221025e-05, "loss": 6.6218, "step": 15517 }, { "epoch": 0.6542572253725994, "grad_norm": 3.6748464107513428, "learning_rate": 4.3440613385802184e-05, "loss": 7.1382, "step": 15518 }, { "epoch": 0.654299386554799, "grad_norm": 2.560218095779419, "learning_rate": 4.343983259939411e-05, "loss": 6.4624, "step": 15519 }, { "epoch": 0.6543415477369986, "grad_norm": 3.925846815109253, "learning_rate": 4.343905181298604e-05, "loss": 7.2447, "step": 15520 }, { "epoch": 0.6543837089191981, "grad_norm": 2.6202285289764404, "learning_rate": 4.343827102657797e-05, "loss": 6.5222, "step": 15521 }, { "epoch": 0.6544258701013976, "grad_norm": 3.828171730041504, "learning_rate": 4.34374902401699e-05, "loss": 7.5764, "step": 15522 }, { "epoch": 0.6544680312835972, "grad_norm": 3.505128860473633, "learning_rate": 4.343670945376183e-05, "loss": 7.7535, "step": 15523 }, { "epoch": 0.6545101924657968, "grad_norm": 3.702859401702881, "learning_rate": 4.343592866735376e-05, "loss": 6.3585, "step": 15524 }, { "epoch": 0.6545523536479962, "grad_norm": 2.7933542728424072, "learning_rate": 4.3435147880945696e-05, "loss": 6.8532, "step": 15525 }, { "epoch": 0.6545945148301958, "grad_norm": 3.697843551635742, "learning_rate": 4.343436709453762e-05, "loss": 6.5368, "step": 15526 }, { "epoch": 0.6546366760123954, "grad_norm": 1.4611191749572754, "learning_rate": 4.3433586308129545e-05, "loss": 6.3183, "step": 15527 }, { "epoch": 0.654678837194595, "grad_norm": 2.32053279876709, "learning_rate": 4.343280552172148e-05, "loss": 6.9957, "step": 15528 }, { "epoch": 0.6547209983767945, "grad_norm": 4.916295528411865, "learning_rate": 4.343202473531341e-05, "loss": 7.6448, "step": 15529 }, { "epoch": 0.654763159558994, "grad_norm": 3.634983777999878, "learning_rate": 4.343124394890534e-05, "loss": 7.0973, "step": 15530 }, { "epoch": 0.6548053207411936, "grad_norm": 2.049100637435913, "learning_rate": 4.343046316249727e-05, "loss": 6.3207, "step": 15531 }, { "epoch": 0.6548474819233931, "grad_norm": 1.9295843839645386, "learning_rate": 4.34296823760892e-05, "loss": 6.425, "step": 15532 }, { "epoch": 0.6548896431055927, "grad_norm": 1.8953588008880615, "learning_rate": 4.342890158968113e-05, "loss": 7.0905, "step": 15533 }, { "epoch": 0.6549318042877922, "grad_norm": 2.682159185409546, "learning_rate": 4.342812080327306e-05, "loss": 6.957, "step": 15534 }, { "epoch": 0.6549739654699918, "grad_norm": 2.641439437866211, "learning_rate": 4.342734001686499e-05, "loss": 6.356, "step": 15535 }, { "epoch": 0.6550161266521913, "grad_norm": 3.371253728866577, "learning_rate": 4.342655923045692e-05, "loss": 7.8225, "step": 15536 }, { "epoch": 0.6550582878343909, "grad_norm": 1.7614119052886963, "learning_rate": 4.342577844404885e-05, "loss": 6.2809, "step": 15537 }, { "epoch": 0.6551004490165905, "grad_norm": 2.495375156402588, "learning_rate": 4.3424997657640775e-05, "loss": 6.5921, "step": 15538 }, { "epoch": 0.6551426101987899, "grad_norm": 2.391050100326538, "learning_rate": 4.3424216871232706e-05, "loss": 6.6891, "step": 15539 }, { "epoch": 0.6551847713809895, "grad_norm": 2.0736918449401855, "learning_rate": 4.342343608482464e-05, "loss": 6.2247, "step": 15540 }, { "epoch": 0.6552269325631891, "grad_norm": 3.358520746231079, "learning_rate": 4.342265529841656e-05, "loss": 7.6189, "step": 15541 }, { "epoch": 0.6552690937453887, "grad_norm": 1.6873352527618408, "learning_rate": 4.34218745120085e-05, "loss": 6.3823, "step": 15542 }, { "epoch": 0.6553112549275881, "grad_norm": 1.3963514566421509, "learning_rate": 4.3421093725600424e-05, "loss": 6.6386, "step": 15543 }, { "epoch": 0.6553534161097877, "grad_norm": 1.3511556386947632, "learning_rate": 4.342031293919236e-05, "loss": 6.4784, "step": 15544 }, { "epoch": 0.6553955772919873, "grad_norm": 3.516331672668457, "learning_rate": 4.341953215278429e-05, "loss": 7.4331, "step": 15545 }, { "epoch": 0.6554377384741868, "grad_norm": 1.7316679954528809, "learning_rate": 4.341875136637621e-05, "loss": 6.596, "step": 15546 }, { "epoch": 0.6554798996563863, "grad_norm": 1.553259015083313, "learning_rate": 4.341797057996815e-05, "loss": 6.4291, "step": 15547 }, { "epoch": 0.6555220608385859, "grad_norm": 2.9407033920288086, "learning_rate": 4.3417189793560074e-05, "loss": 7.1841, "step": 15548 }, { "epoch": 0.6555642220207855, "grad_norm": 1.9108089208602905, "learning_rate": 4.3416409007152005e-05, "loss": 6.8696, "step": 15549 }, { "epoch": 0.655606383202985, "grad_norm": 3.5696890354156494, "learning_rate": 4.3415628220743936e-05, "loss": 7.6515, "step": 15550 }, { "epoch": 0.6556485443851846, "grad_norm": 2.011061429977417, "learning_rate": 4.341484743433587e-05, "loss": 6.8312, "step": 15551 }, { "epoch": 0.6556907055673841, "grad_norm": 3.1839396953582764, "learning_rate": 4.341406664792779e-05, "loss": 8.0389, "step": 15552 }, { "epoch": 0.6557328667495836, "grad_norm": 3.7158541679382324, "learning_rate": 4.3413285861519723e-05, "loss": 6.5164, "step": 15553 }, { "epoch": 0.6557750279317832, "grad_norm": 2.234605550765991, "learning_rate": 4.3412505075111655e-05, "loss": 7.0768, "step": 15554 }, { "epoch": 0.6558171891139828, "grad_norm": 1.9043371677398682, "learning_rate": 4.3411724288703586e-05, "loss": 6.1829, "step": 15555 }, { "epoch": 0.6558593502961823, "grad_norm": 1.2684091329574585, "learning_rate": 4.341094350229552e-05, "loss": 6.7599, "step": 15556 }, { "epoch": 0.6559015114783818, "grad_norm": 3.1988685131073, "learning_rate": 4.341016271588744e-05, "loss": 6.6852, "step": 15557 }, { "epoch": 0.6559436726605814, "grad_norm": 2.3756439685821533, "learning_rate": 4.340938192947937e-05, "loss": 6.3716, "step": 15558 }, { "epoch": 0.655985833842781, "grad_norm": 2.35723614692688, "learning_rate": 4.3408601143071304e-05, "loss": 7.1285, "step": 15559 }, { "epoch": 0.6560279950249805, "grad_norm": 1.898383378982544, "learning_rate": 4.340782035666323e-05, "loss": 6.641, "step": 15560 }, { "epoch": 0.65607015620718, "grad_norm": 2.9345364570617676, "learning_rate": 4.340703957025517e-05, "loss": 7.3106, "step": 15561 }, { "epoch": 0.6561123173893796, "grad_norm": 1.601750373840332, "learning_rate": 4.340625878384709e-05, "loss": 6.2949, "step": 15562 }, { "epoch": 0.6561544785715792, "grad_norm": 1.9440425634384155, "learning_rate": 4.340547799743902e-05, "loss": 6.3075, "step": 15563 }, { "epoch": 0.6561966397537787, "grad_norm": 3.097175121307373, "learning_rate": 4.3404697211030954e-05, "loss": 7.4301, "step": 15564 }, { "epoch": 0.6562388009359782, "grad_norm": 2.3751983642578125, "learning_rate": 4.340391642462288e-05, "loss": 6.9107, "step": 15565 }, { "epoch": 0.6562809621181778, "grad_norm": 3.1495676040649414, "learning_rate": 4.3403135638214816e-05, "loss": 7.8886, "step": 15566 }, { "epoch": 0.6563231233003773, "grad_norm": 1.5640676021575928, "learning_rate": 4.340235485180674e-05, "loss": 6.2026, "step": 15567 }, { "epoch": 0.6563652844825769, "grad_norm": 1.6106816530227661, "learning_rate": 4.340157406539867e-05, "loss": 6.5023, "step": 15568 }, { "epoch": 0.6564074456647765, "grad_norm": 1.6691957712173462, "learning_rate": 4.34007932789906e-05, "loss": 6.5584, "step": 15569 }, { "epoch": 0.656449606846976, "grad_norm": 1.2105664014816284, "learning_rate": 4.3400012492582535e-05, "loss": 6.3945, "step": 15570 }, { "epoch": 0.6564917680291755, "grad_norm": 1.3886364698410034, "learning_rate": 4.339923170617446e-05, "loss": 6.7579, "step": 15571 }, { "epoch": 0.6565339292113751, "grad_norm": 2.5420193672180176, "learning_rate": 4.339845091976639e-05, "loss": 6.3583, "step": 15572 }, { "epoch": 0.6565760903935747, "grad_norm": 2.630671262741089, "learning_rate": 4.339767013335832e-05, "loss": 6.2362, "step": 15573 }, { "epoch": 0.6566182515757741, "grad_norm": 2.056929588317871, "learning_rate": 4.3396889346950246e-05, "loss": 6.448, "step": 15574 }, { "epoch": 0.6566604127579737, "grad_norm": 2.553473711013794, "learning_rate": 4.3396108560542184e-05, "loss": 6.0652, "step": 15575 }, { "epoch": 0.6567025739401733, "grad_norm": 1.8894379138946533, "learning_rate": 4.339532777413411e-05, "loss": 7.0576, "step": 15576 }, { "epoch": 0.6567447351223729, "grad_norm": 1.987424373626709, "learning_rate": 4.339454698772604e-05, "loss": 6.5276, "step": 15577 }, { "epoch": 0.6567868963045724, "grad_norm": 2.840718984603882, "learning_rate": 4.339376620131797e-05, "loss": 7.1589, "step": 15578 }, { "epoch": 0.6568290574867719, "grad_norm": 3.209709882736206, "learning_rate": 4.3392985414909896e-05, "loss": 7.8043, "step": 15579 }, { "epoch": 0.6568712186689715, "grad_norm": 2.524831771850586, "learning_rate": 4.3392204628501834e-05, "loss": 7.1836, "step": 15580 }, { "epoch": 0.656913379851171, "grad_norm": 2.231058120727539, "learning_rate": 4.339142384209376e-05, "loss": 7.2132, "step": 15581 }, { "epoch": 0.6569555410333706, "grad_norm": 4.286967754364014, "learning_rate": 4.339064305568569e-05, "loss": 7.4177, "step": 15582 }, { "epoch": 0.6569977022155701, "grad_norm": 2.0687389373779297, "learning_rate": 4.338986226927762e-05, "loss": 7.1564, "step": 15583 }, { "epoch": 0.6570398633977697, "grad_norm": 2.1765451431274414, "learning_rate": 4.3389081482869545e-05, "loss": 6.3717, "step": 15584 }, { "epoch": 0.6570820245799692, "grad_norm": 2.7712018489837646, "learning_rate": 4.3388300696461476e-05, "loss": 7.1441, "step": 15585 }, { "epoch": 0.6571241857621688, "grad_norm": 2.420781135559082, "learning_rate": 4.338751991005341e-05, "loss": 6.4833, "step": 15586 }, { "epoch": 0.6571663469443684, "grad_norm": 1.9875327348709106, "learning_rate": 4.338673912364534e-05, "loss": 6.7925, "step": 15587 }, { "epoch": 0.6572085081265678, "grad_norm": 3.4947266578674316, "learning_rate": 4.338595833723727e-05, "loss": 6.5544, "step": 15588 }, { "epoch": 0.6572506693087674, "grad_norm": 2.9389851093292236, "learning_rate": 4.3385177550829195e-05, "loss": 6.3772, "step": 15589 }, { "epoch": 0.657292830490967, "grad_norm": 3.8705034255981445, "learning_rate": 4.3384396764421126e-05, "loss": 7.1698, "step": 15590 }, { "epoch": 0.6573349916731666, "grad_norm": 3.509648323059082, "learning_rate": 4.338361597801306e-05, "loss": 8.1134, "step": 15591 }, { "epoch": 0.657377152855366, "grad_norm": 2.684359550476074, "learning_rate": 4.338283519160499e-05, "loss": 7.824, "step": 15592 }, { "epoch": 0.6574193140375656, "grad_norm": 3.774608612060547, "learning_rate": 4.338205440519691e-05, "loss": 6.611, "step": 15593 }, { "epoch": 0.6574614752197652, "grad_norm": 1.4110764265060425, "learning_rate": 4.338127361878885e-05, "loss": 7.7488, "step": 15594 }, { "epoch": 0.6575036364019647, "grad_norm": 4.298306941986084, "learning_rate": 4.3380492832380775e-05, "loss": 8.4022, "step": 15595 }, { "epoch": 0.6575457975841643, "grad_norm": 2.6620733737945557, "learning_rate": 4.337971204597271e-05, "loss": 7.0311, "step": 15596 }, { "epoch": 0.6575879587663638, "grad_norm": 2.72552752494812, "learning_rate": 4.337893125956464e-05, "loss": 7.6656, "step": 15597 }, { "epoch": 0.6576301199485634, "grad_norm": 3.053722858428955, "learning_rate": 4.337815047315656e-05, "loss": 7.0343, "step": 15598 }, { "epoch": 0.6576722811307629, "grad_norm": 2.753455877304077, "learning_rate": 4.3377369686748494e-05, "loss": 6.6922, "step": 15599 }, { "epoch": 0.6577144423129625, "grad_norm": 2.3094215393066406, "learning_rate": 4.3376588900340425e-05, "loss": 7.2058, "step": 15600 }, { "epoch": 0.657756603495162, "grad_norm": 2.306082248687744, "learning_rate": 4.3375808113932356e-05, "loss": 6.9005, "step": 15601 }, { "epoch": 0.6577987646773615, "grad_norm": 2.7979812622070312, "learning_rate": 4.337502732752429e-05, "loss": 7.0601, "step": 15602 }, { "epoch": 0.6578409258595611, "grad_norm": 1.9154748916625977, "learning_rate": 4.337424654111621e-05, "loss": 6.6035, "step": 15603 }, { "epoch": 0.6578830870417607, "grad_norm": 1.7108845710754395, "learning_rate": 4.337346575470814e-05, "loss": 6.8148, "step": 15604 }, { "epoch": 0.6579252482239601, "grad_norm": 2.9330103397369385, "learning_rate": 4.3372684968300074e-05, "loss": 6.7586, "step": 15605 }, { "epoch": 0.6579674094061597, "grad_norm": 3.2198007106781006, "learning_rate": 4.3371904181892006e-05, "loss": 7.1337, "step": 15606 }, { "epoch": 0.6580095705883593, "grad_norm": 2.0293102264404297, "learning_rate": 4.337112339548393e-05, "loss": 6.9696, "step": 15607 }, { "epoch": 0.6580517317705589, "grad_norm": 2.2752034664154053, "learning_rate": 4.337034260907586e-05, "loss": 6.4588, "step": 15608 }, { "epoch": 0.6580938929527584, "grad_norm": 2.799661636352539, "learning_rate": 4.336956182266779e-05, "loss": 7.0399, "step": 15609 }, { "epoch": 0.6581360541349579, "grad_norm": 2.034209966659546, "learning_rate": 4.336878103625972e-05, "loss": 6.6407, "step": 15610 }, { "epoch": 0.6581782153171575, "grad_norm": 1.7867636680603027, "learning_rate": 4.3368000249851655e-05, "loss": 6.7448, "step": 15611 }, { "epoch": 0.658220376499357, "grad_norm": 1.715954065322876, "learning_rate": 4.336721946344358e-05, "loss": 6.5297, "step": 15612 }, { "epoch": 0.6582625376815566, "grad_norm": 1.8824334144592285, "learning_rate": 4.336643867703552e-05, "loss": 6.4651, "step": 15613 }, { "epoch": 0.6583046988637561, "grad_norm": 6.003273010253906, "learning_rate": 4.336565789062744e-05, "loss": 10.0352, "step": 15614 }, { "epoch": 0.6583468600459557, "grad_norm": 1.956108570098877, "learning_rate": 4.3364877104219373e-05, "loss": 6.5412, "step": 15615 }, { "epoch": 0.6583890212281552, "grad_norm": 1.722405195236206, "learning_rate": 4.3364096317811305e-05, "loss": 6.6544, "step": 15616 }, { "epoch": 0.6584311824103548, "grad_norm": 4.009115695953369, "learning_rate": 4.336331553140323e-05, "loss": 7.2362, "step": 15617 }, { "epoch": 0.6584733435925544, "grad_norm": 3.2024428844451904, "learning_rate": 4.336253474499516e-05, "loss": 7.524, "step": 15618 }, { "epoch": 0.6585155047747538, "grad_norm": 4.450326442718506, "learning_rate": 4.336175395858709e-05, "loss": 7.8308, "step": 15619 }, { "epoch": 0.6585576659569534, "grad_norm": 1.5057809352874756, "learning_rate": 4.336097317217902e-05, "loss": 6.8732, "step": 15620 }, { "epoch": 0.658599827139153, "grad_norm": 3.0022332668304443, "learning_rate": 4.336019238577095e-05, "loss": 7.7468, "step": 15621 }, { "epoch": 0.6586419883213526, "grad_norm": 1.94450843334198, "learning_rate": 4.335941159936288e-05, "loss": 6.4949, "step": 15622 }, { "epoch": 0.658684149503552, "grad_norm": 2.9711050987243652, "learning_rate": 4.335863081295481e-05, "loss": 6.2606, "step": 15623 }, { "epoch": 0.6587263106857516, "grad_norm": 3.3273673057556152, "learning_rate": 4.335785002654674e-05, "loss": 6.398, "step": 15624 }, { "epoch": 0.6587684718679512, "grad_norm": 1.4962445497512817, "learning_rate": 4.335706924013867e-05, "loss": 6.8433, "step": 15625 }, { "epoch": 0.6588106330501508, "grad_norm": 1.9721935987472534, "learning_rate": 4.33562884537306e-05, "loss": 6.7131, "step": 15626 }, { "epoch": 0.6588527942323503, "grad_norm": 2.8551063537597656, "learning_rate": 4.335550766732253e-05, "loss": 6.4482, "step": 15627 }, { "epoch": 0.6588949554145498, "grad_norm": 2.682543992996216, "learning_rate": 4.335472688091446e-05, "loss": 7.4181, "step": 15628 }, { "epoch": 0.6589371165967494, "grad_norm": 2.640915632247925, "learning_rate": 4.3353946094506384e-05, "loss": 6.8258, "step": 15629 }, { "epoch": 0.6589792777789489, "grad_norm": 2.1571269035339355, "learning_rate": 4.335316530809832e-05, "loss": 7.1923, "step": 15630 }, { "epoch": 0.6590214389611485, "grad_norm": 2.4539947509765625, "learning_rate": 4.3352384521690247e-05, "loss": 7.1067, "step": 15631 }, { "epoch": 0.659063600143348, "grad_norm": 2.2360363006591797, "learning_rate": 4.335160373528218e-05, "loss": 6.5133, "step": 15632 }, { "epoch": 0.6591057613255475, "grad_norm": 1.81088387966156, "learning_rate": 4.335082294887411e-05, "loss": 7.1861, "step": 15633 }, { "epoch": 0.6591479225077471, "grad_norm": 2.9590373039245605, "learning_rate": 4.3350042162466034e-05, "loss": 6.0992, "step": 15634 }, { "epoch": 0.6591900836899467, "grad_norm": 1.1354031562805176, "learning_rate": 4.334926137605797e-05, "loss": 6.4546, "step": 15635 }, { "epoch": 0.6592322448721463, "grad_norm": 1.7827281951904297, "learning_rate": 4.3348480589649896e-05, "loss": 6.4613, "step": 15636 }, { "epoch": 0.6592744060543457, "grad_norm": 2.207059621810913, "learning_rate": 4.334769980324183e-05, "loss": 7.1234, "step": 15637 }, { "epoch": 0.6593165672365453, "grad_norm": 2.0576252937316895, "learning_rate": 4.334691901683376e-05, "loss": 7.1145, "step": 15638 }, { "epoch": 0.6593587284187449, "grad_norm": 1.3589977025985718, "learning_rate": 4.334613823042569e-05, "loss": 6.7205, "step": 15639 }, { "epoch": 0.6594008896009445, "grad_norm": 1.4174113273620605, "learning_rate": 4.3345357444017614e-05, "loss": 7.0152, "step": 15640 }, { "epoch": 0.6594430507831439, "grad_norm": 1.5523465871810913, "learning_rate": 4.3344576657609546e-05, "loss": 6.8159, "step": 15641 }, { "epoch": 0.6594852119653435, "grad_norm": 3.536203384399414, "learning_rate": 4.334379587120148e-05, "loss": 6.3397, "step": 15642 }, { "epoch": 0.6595273731475431, "grad_norm": 1.6349817514419556, "learning_rate": 4.33430150847934e-05, "loss": 6.3111, "step": 15643 }, { "epoch": 0.6595695343297426, "grad_norm": 3.9344558715820312, "learning_rate": 4.334223429838534e-05, "loss": 7.478, "step": 15644 }, { "epoch": 0.6596116955119422, "grad_norm": 2.146566152572632, "learning_rate": 4.3341453511977264e-05, "loss": 6.8536, "step": 15645 }, { "epoch": 0.6596538566941417, "grad_norm": 4.294794082641602, "learning_rate": 4.3340672725569195e-05, "loss": 7.5936, "step": 15646 }, { "epoch": 0.6596960178763412, "grad_norm": 1.3130632638931274, "learning_rate": 4.3339891939161126e-05, "loss": 7.1217, "step": 15647 }, { "epoch": 0.6597381790585408, "grad_norm": 2.698448419570923, "learning_rate": 4.333911115275305e-05, "loss": 7.6954, "step": 15648 }, { "epoch": 0.6597803402407404, "grad_norm": 2.1725242137908936, "learning_rate": 4.333833036634499e-05, "loss": 6.551, "step": 15649 }, { "epoch": 0.6598225014229399, "grad_norm": 2.8717260360717773, "learning_rate": 4.333754957993691e-05, "loss": 7.5386, "step": 15650 }, { "epoch": 0.6598646626051394, "grad_norm": 2.153467893600464, "learning_rate": 4.3336768793528845e-05, "loss": 7.024, "step": 15651 }, { "epoch": 0.659906823787339, "grad_norm": 2.8753108978271484, "learning_rate": 4.3335988007120776e-05, "loss": 6.7121, "step": 15652 }, { "epoch": 0.6599489849695386, "grad_norm": 1.9641849994659424, "learning_rate": 4.33352072207127e-05, "loss": 6.3345, "step": 15653 }, { "epoch": 0.659991146151738, "grad_norm": 2.2380530834198, "learning_rate": 4.333442643430463e-05, "loss": 6.6272, "step": 15654 }, { "epoch": 0.6600333073339376, "grad_norm": 2.4931139945983887, "learning_rate": 4.333364564789656e-05, "loss": 6.3292, "step": 15655 }, { "epoch": 0.6600754685161372, "grad_norm": 1.3766030073165894, "learning_rate": 4.3332864861488494e-05, "loss": 6.374, "step": 15656 }, { "epoch": 0.6601176296983368, "grad_norm": 1.5446202754974365, "learning_rate": 4.3332084075080425e-05, "loss": 6.5178, "step": 15657 }, { "epoch": 0.6601597908805363, "grad_norm": 3.252394199371338, "learning_rate": 4.333130328867236e-05, "loss": 8.0635, "step": 15658 }, { "epoch": 0.6602019520627358, "grad_norm": 3.2017338275909424, "learning_rate": 4.333052250226428e-05, "loss": 7.6234, "step": 15659 }, { "epoch": 0.6602441132449354, "grad_norm": 2.3435540199279785, "learning_rate": 4.332974171585621e-05, "loss": 7.1688, "step": 15660 }, { "epoch": 0.660286274427135, "grad_norm": 2.2383100986480713, "learning_rate": 4.3328960929448144e-05, "loss": 6.6444, "step": 15661 }, { "epoch": 0.6603284356093345, "grad_norm": 1.821954607963562, "learning_rate": 4.332818014304007e-05, "loss": 6.6191, "step": 15662 }, { "epoch": 0.660370596791534, "grad_norm": 2.0952439308166504, "learning_rate": 4.3327399356632006e-05, "loss": 6.8331, "step": 15663 }, { "epoch": 0.6604127579737336, "grad_norm": 3.3767528533935547, "learning_rate": 4.332661857022393e-05, "loss": 6.4494, "step": 15664 }, { "epoch": 0.6604549191559331, "grad_norm": 1.748968482017517, "learning_rate": 4.332583778381586e-05, "loss": 6.4523, "step": 15665 }, { "epoch": 0.6604970803381327, "grad_norm": 1.6751792430877686, "learning_rate": 4.332505699740779e-05, "loss": 6.3702, "step": 15666 }, { "epoch": 0.6605392415203323, "grad_norm": 1.7425655126571655, "learning_rate": 4.332427621099972e-05, "loss": 7.0206, "step": 15667 }, { "epoch": 0.6605814027025317, "grad_norm": 1.5018688440322876, "learning_rate": 4.3323495424591656e-05, "loss": 6.5555, "step": 15668 }, { "epoch": 0.6606235638847313, "grad_norm": 3.3994951248168945, "learning_rate": 4.332271463818358e-05, "loss": 7.6346, "step": 15669 }, { "epoch": 0.6606657250669309, "grad_norm": 2.4183077812194824, "learning_rate": 4.332193385177551e-05, "loss": 6.3504, "step": 15670 }, { "epoch": 0.6607078862491305, "grad_norm": 2.231844902038574, "learning_rate": 4.332115306536744e-05, "loss": 6.9415, "step": 15671 }, { "epoch": 0.6607500474313299, "grad_norm": 3.4757635593414307, "learning_rate": 4.332037227895937e-05, "loss": 8.14, "step": 15672 }, { "epoch": 0.6607922086135295, "grad_norm": 4.7006516456604, "learning_rate": 4.33195914925513e-05, "loss": 7.459, "step": 15673 }, { "epoch": 0.6608343697957291, "grad_norm": 3.7546143531799316, "learning_rate": 4.331881070614323e-05, "loss": 7.7796, "step": 15674 }, { "epoch": 0.6608765309779286, "grad_norm": 1.8940997123718262, "learning_rate": 4.331802991973516e-05, "loss": 6.9257, "step": 15675 }, { "epoch": 0.6609186921601282, "grad_norm": 1.789474606513977, "learning_rate": 4.3317249133327085e-05, "loss": 6.7905, "step": 15676 }, { "epoch": 0.6609608533423277, "grad_norm": 2.883334159851074, "learning_rate": 4.3316468346919024e-05, "loss": 6.9494, "step": 15677 }, { "epoch": 0.6610030145245273, "grad_norm": 1.9580744504928589, "learning_rate": 4.331568756051095e-05, "loss": 7.1476, "step": 15678 }, { "epoch": 0.6610451757067268, "grad_norm": 1.6789827346801758, "learning_rate": 4.331490677410287e-05, "loss": 6.854, "step": 15679 }, { "epoch": 0.6610873368889264, "grad_norm": 2.357121467590332, "learning_rate": 4.331412598769481e-05, "loss": 6.4498, "step": 15680 }, { "epoch": 0.6611294980711259, "grad_norm": 2.9790990352630615, "learning_rate": 4.3313345201286735e-05, "loss": 6.4383, "step": 15681 }, { "epoch": 0.6611716592533254, "grad_norm": 3.0800976753234863, "learning_rate": 4.331256441487867e-05, "loss": 8.5724, "step": 15682 }, { "epoch": 0.661213820435525, "grad_norm": 2.0604031085968018, "learning_rate": 4.33117836284706e-05, "loss": 6.4763, "step": 15683 }, { "epoch": 0.6612559816177246, "grad_norm": 2.094054698944092, "learning_rate": 4.331100284206253e-05, "loss": 7.0044, "step": 15684 }, { "epoch": 0.6612981427999242, "grad_norm": 1.884493112564087, "learning_rate": 4.331022205565446e-05, "loss": 6.4115, "step": 15685 }, { "epoch": 0.6613403039821236, "grad_norm": 4.226258277893066, "learning_rate": 4.3309441269246385e-05, "loss": 7.5507, "step": 15686 }, { "epoch": 0.6613824651643232, "grad_norm": 1.7095493078231812, "learning_rate": 4.3308660482838316e-05, "loss": 6.4626, "step": 15687 }, { "epoch": 0.6614246263465228, "grad_norm": 1.798384189605713, "learning_rate": 4.330787969643025e-05, "loss": 7.0247, "step": 15688 }, { "epoch": 0.6614667875287223, "grad_norm": 1.5272613763809204, "learning_rate": 4.330709891002218e-05, "loss": 7.1452, "step": 15689 }, { "epoch": 0.6615089487109218, "grad_norm": 3.169301986694336, "learning_rate": 4.33063181236141e-05, "loss": 6.768, "step": 15690 }, { "epoch": 0.6615511098931214, "grad_norm": 4.401741981506348, "learning_rate": 4.3305537337206034e-05, "loss": 7.8285, "step": 15691 }, { "epoch": 0.661593271075321, "grad_norm": 2.4473958015441895, "learning_rate": 4.3304756550797965e-05, "loss": 7.4986, "step": 15692 }, { "epoch": 0.6616354322575205, "grad_norm": 3.1032252311706543, "learning_rate": 4.3303975764389897e-05, "loss": 6.647, "step": 15693 }, { "epoch": 0.6616775934397201, "grad_norm": 1.5353715419769287, "learning_rate": 4.330319497798183e-05, "loss": 6.4411, "step": 15694 }, { "epoch": 0.6617197546219196, "grad_norm": 1.3973182439804077, "learning_rate": 4.330241419157375e-05, "loss": 6.7766, "step": 15695 }, { "epoch": 0.6617619158041191, "grad_norm": 2.5588269233703613, "learning_rate": 4.3301633405165684e-05, "loss": 7.7269, "step": 15696 }, { "epoch": 0.6618040769863187, "grad_norm": 2.280851125717163, "learning_rate": 4.3300852618757615e-05, "loss": 6.617, "step": 15697 }, { "epoch": 0.6618462381685183, "grad_norm": 2.849693536758423, "learning_rate": 4.330007183234954e-05, "loss": 6.62, "step": 15698 }, { "epoch": 0.6618883993507177, "grad_norm": 1.8098527193069458, "learning_rate": 4.329929104594148e-05, "loss": 6.6322, "step": 15699 }, { "epoch": 0.6619305605329173, "grad_norm": 2.302644729614258, "learning_rate": 4.32985102595334e-05, "loss": 7.7817, "step": 15700 }, { "epoch": 0.6619727217151169, "grad_norm": 2.33622145652771, "learning_rate": 4.329772947312533e-05, "loss": 6.9592, "step": 15701 }, { "epoch": 0.6620148828973165, "grad_norm": 2.508537530899048, "learning_rate": 4.3296948686717264e-05, "loss": 6.3708, "step": 15702 }, { "epoch": 0.662057044079516, "grad_norm": 3.0668275356292725, "learning_rate": 4.3296167900309196e-05, "loss": 7.6544, "step": 15703 }, { "epoch": 0.6620992052617155, "grad_norm": 2.853733777999878, "learning_rate": 4.329538711390113e-05, "loss": 7.8067, "step": 15704 }, { "epoch": 0.6621413664439151, "grad_norm": 1.6173640489578247, "learning_rate": 4.329460632749305e-05, "loss": 6.2668, "step": 15705 }, { "epoch": 0.6621835276261147, "grad_norm": 2.9544355869293213, "learning_rate": 4.329382554108498e-05, "loss": 7.4902, "step": 15706 }, { "epoch": 0.6622256888083142, "grad_norm": 2.546233654022217, "learning_rate": 4.3293044754676914e-05, "loss": 6.9666, "step": 15707 }, { "epoch": 0.6622678499905137, "grad_norm": 1.978068470954895, "learning_rate": 4.3292263968268845e-05, "loss": 6.4463, "step": 15708 }, { "epoch": 0.6623100111727133, "grad_norm": 1.692202091217041, "learning_rate": 4.329148318186077e-05, "loss": 6.5231, "step": 15709 }, { "epoch": 0.6623521723549128, "grad_norm": 2.2566616535186768, "learning_rate": 4.32907023954527e-05, "loss": 6.4456, "step": 15710 }, { "epoch": 0.6623943335371124, "grad_norm": 1.7565979957580566, "learning_rate": 4.328992160904463e-05, "loss": 6.8952, "step": 15711 }, { "epoch": 0.6624364947193119, "grad_norm": 1.3658454418182373, "learning_rate": 4.328914082263656e-05, "loss": 6.3428, "step": 15712 }, { "epoch": 0.6624786559015114, "grad_norm": 1.9076423645019531, "learning_rate": 4.3288360036228495e-05, "loss": 6.4895, "step": 15713 }, { "epoch": 0.662520817083711, "grad_norm": 3.5780129432678223, "learning_rate": 4.328757924982042e-05, "loss": 7.7548, "step": 15714 }, { "epoch": 0.6625629782659106, "grad_norm": 1.991158366203308, "learning_rate": 4.328679846341235e-05, "loss": 6.9516, "step": 15715 }, { "epoch": 0.6626051394481102, "grad_norm": 1.645477056503296, "learning_rate": 4.328601767700428e-05, "loss": 6.8233, "step": 15716 }, { "epoch": 0.6626473006303096, "grad_norm": 3.056316375732422, "learning_rate": 4.3285236890596206e-05, "loss": 7.6409, "step": 15717 }, { "epoch": 0.6626894618125092, "grad_norm": 4.424330234527588, "learning_rate": 4.3284456104188144e-05, "loss": 7.2741, "step": 15718 }, { "epoch": 0.6627316229947088, "grad_norm": 2.4143171310424805, "learning_rate": 4.328367531778007e-05, "loss": 7.0512, "step": 15719 }, { "epoch": 0.6627737841769084, "grad_norm": 1.5086473226547241, "learning_rate": 4.3282894531372e-05, "loss": 6.944, "step": 15720 }, { "epoch": 0.6628159453591078, "grad_norm": 2.058896541595459, "learning_rate": 4.328211374496393e-05, "loss": 6.5646, "step": 15721 }, { "epoch": 0.6628581065413074, "grad_norm": 1.9551359415054321, "learning_rate": 4.328133295855586e-05, "loss": 7.0533, "step": 15722 }, { "epoch": 0.662900267723507, "grad_norm": 2.0082390308380127, "learning_rate": 4.328055217214779e-05, "loss": 7.0987, "step": 15723 }, { "epoch": 0.6629424289057065, "grad_norm": 1.768394112586975, "learning_rate": 4.327977138573972e-05, "loss": 6.9738, "step": 15724 }, { "epoch": 0.6629845900879061, "grad_norm": 2.234484910964966, "learning_rate": 4.327899059933165e-05, "loss": 6.8429, "step": 15725 }, { "epoch": 0.6630267512701056, "grad_norm": 3.206387758255005, "learning_rate": 4.327820981292358e-05, "loss": 6.4586, "step": 15726 }, { "epoch": 0.6630689124523051, "grad_norm": 4.412238597869873, "learning_rate": 4.327742902651551e-05, "loss": 7.3996, "step": 15727 }, { "epoch": 0.6631110736345047, "grad_norm": 1.7116049528121948, "learning_rate": 4.3276648240107436e-05, "loss": 7.0783, "step": 15728 }, { "epoch": 0.6631532348167043, "grad_norm": 1.6932095289230347, "learning_rate": 4.327586745369937e-05, "loss": 6.9035, "step": 15729 }, { "epoch": 0.6631953959989038, "grad_norm": 3.273101329803467, "learning_rate": 4.32750866672913e-05, "loss": 7.3172, "step": 15730 }, { "epoch": 0.6632375571811033, "grad_norm": 1.8638113737106323, "learning_rate": 4.3274305880883223e-05, "loss": 6.7399, "step": 15731 }, { "epoch": 0.6632797183633029, "grad_norm": 1.8149938583374023, "learning_rate": 4.327352509447516e-05, "loss": 7.0214, "step": 15732 }, { "epoch": 0.6633218795455025, "grad_norm": 2.491142988204956, "learning_rate": 4.3272744308067086e-05, "loss": 7.5568, "step": 15733 }, { "epoch": 0.663364040727702, "grad_norm": 1.7451465129852295, "learning_rate": 4.327196352165902e-05, "loss": 6.5395, "step": 15734 }, { "epoch": 0.6634062019099015, "grad_norm": 1.3239587545394897, "learning_rate": 4.327118273525095e-05, "loss": 6.8929, "step": 15735 }, { "epoch": 0.6634483630921011, "grad_norm": 1.9589093923568726, "learning_rate": 4.327040194884287e-05, "loss": 7.2227, "step": 15736 }, { "epoch": 0.6634905242743007, "grad_norm": 2.1631340980529785, "learning_rate": 4.326962116243481e-05, "loss": 6.4599, "step": 15737 }, { "epoch": 0.6635326854565002, "grad_norm": 3.08315110206604, "learning_rate": 4.3268840376026736e-05, "loss": 6.2817, "step": 15738 }, { "epoch": 0.6635748466386997, "grad_norm": 2.537184000015259, "learning_rate": 4.326805958961867e-05, "loss": 6.8331, "step": 15739 }, { "epoch": 0.6636170078208993, "grad_norm": 1.8437986373901367, "learning_rate": 4.32672788032106e-05, "loss": 6.5058, "step": 15740 }, { "epoch": 0.6636591690030988, "grad_norm": 1.229486107826233, "learning_rate": 4.326649801680252e-05, "loss": 6.5965, "step": 15741 }, { "epoch": 0.6637013301852984, "grad_norm": 1.5031812191009521, "learning_rate": 4.3265717230394454e-05, "loss": 6.3894, "step": 15742 }, { "epoch": 0.663743491367498, "grad_norm": 3.826519012451172, "learning_rate": 4.3264936443986385e-05, "loss": 7.6546, "step": 15743 }, { "epoch": 0.6637856525496975, "grad_norm": 2.118837594985962, "learning_rate": 4.3264155657578316e-05, "loss": 7.0325, "step": 15744 }, { "epoch": 0.663827813731897, "grad_norm": 2.286268711090088, "learning_rate": 4.326337487117024e-05, "loss": 6.7148, "step": 15745 }, { "epoch": 0.6638699749140966, "grad_norm": 2.7378406524658203, "learning_rate": 4.326259408476218e-05, "loss": 7.4387, "step": 15746 }, { "epoch": 0.6639121360962962, "grad_norm": 1.8353395462036133, "learning_rate": 4.32618132983541e-05, "loss": 6.5344, "step": 15747 }, { "epoch": 0.6639542972784956, "grad_norm": 2.971339225769043, "learning_rate": 4.3261032511946035e-05, "loss": 7.7354, "step": 15748 }, { "epoch": 0.6639964584606952, "grad_norm": 2.4793825149536133, "learning_rate": 4.3260251725537966e-05, "loss": 7.973, "step": 15749 }, { "epoch": 0.6640386196428948, "grad_norm": 2.1699330806732178, "learning_rate": 4.325947093912989e-05, "loss": 6.8116, "step": 15750 }, { "epoch": 0.6640807808250944, "grad_norm": 1.9965146780014038, "learning_rate": 4.325869015272183e-05, "loss": 6.9776, "step": 15751 }, { "epoch": 0.6641229420072939, "grad_norm": 1.7096155881881714, "learning_rate": 4.325790936631375e-05, "loss": 6.596, "step": 15752 }, { "epoch": 0.6641651031894934, "grad_norm": 3.4810547828674316, "learning_rate": 4.3257128579905684e-05, "loss": 6.4768, "step": 15753 }, { "epoch": 0.664207264371693, "grad_norm": 2.3066000938415527, "learning_rate": 4.3256347793497615e-05, "loss": 6.5385, "step": 15754 }, { "epoch": 0.6642494255538925, "grad_norm": 2.098691701889038, "learning_rate": 4.325556700708954e-05, "loss": 6.8995, "step": 15755 }, { "epoch": 0.6642915867360921, "grad_norm": 3.4739060401916504, "learning_rate": 4.325478622068147e-05, "loss": 7.4808, "step": 15756 }, { "epoch": 0.6643337479182916, "grad_norm": 1.6594959497451782, "learning_rate": 4.32540054342734e-05, "loss": 6.3467, "step": 15757 }, { "epoch": 0.6643759091004912, "grad_norm": 2.0372495651245117, "learning_rate": 4.3253224647865334e-05, "loss": 6.6137, "step": 15758 }, { "epoch": 0.6644180702826907, "grad_norm": 2.7777085304260254, "learning_rate": 4.325244386145726e-05, "loss": 6.4398, "step": 15759 }, { "epoch": 0.6644602314648903, "grad_norm": 3.169541597366333, "learning_rate": 4.325166307504919e-05, "loss": 7.5813, "step": 15760 }, { "epoch": 0.6645023926470899, "grad_norm": 3.715529680252075, "learning_rate": 4.325088228864112e-05, "loss": 7.9734, "step": 15761 }, { "epoch": 0.6645445538292893, "grad_norm": 1.8020585775375366, "learning_rate": 4.325010150223305e-05, "loss": 6.8296, "step": 15762 }, { "epoch": 0.6645867150114889, "grad_norm": 2.7819414138793945, "learning_rate": 4.324932071582498e-05, "loss": 6.8061, "step": 15763 }, { "epoch": 0.6646288761936885, "grad_norm": 2.696913480758667, "learning_rate": 4.324853992941691e-05, "loss": 7.6613, "step": 15764 }, { "epoch": 0.6646710373758881, "grad_norm": 2.2437877655029297, "learning_rate": 4.3247759143008846e-05, "loss": 6.5283, "step": 15765 }, { "epoch": 0.6647131985580875, "grad_norm": 1.5784045457839966, "learning_rate": 4.324697835660077e-05, "loss": 6.7118, "step": 15766 }, { "epoch": 0.6647553597402871, "grad_norm": 1.812295913696289, "learning_rate": 4.32461975701927e-05, "loss": 6.3843, "step": 15767 }, { "epoch": 0.6647975209224867, "grad_norm": 1.8976192474365234, "learning_rate": 4.324541678378463e-05, "loss": 7.5075, "step": 15768 }, { "epoch": 0.6648396821046862, "grad_norm": 1.798610806465149, "learning_rate": 4.324463599737656e-05, "loss": 6.3872, "step": 15769 }, { "epoch": 0.6648818432868857, "grad_norm": 3.426025867462158, "learning_rate": 4.324385521096849e-05, "loss": 6.7743, "step": 15770 }, { "epoch": 0.6649240044690853, "grad_norm": 1.3894243240356445, "learning_rate": 4.324307442456042e-05, "loss": 6.4009, "step": 15771 }, { "epoch": 0.6649661656512849, "grad_norm": 1.5536352396011353, "learning_rate": 4.324229363815235e-05, "loss": 6.5395, "step": 15772 }, { "epoch": 0.6650083268334844, "grad_norm": 3.969797134399414, "learning_rate": 4.324151285174428e-05, "loss": 7.1818, "step": 15773 }, { "epoch": 0.665050488015684, "grad_norm": 1.8311002254486084, "learning_rate": 4.324073206533621e-05, "loss": 6.6185, "step": 15774 }, { "epoch": 0.6650926491978835, "grad_norm": 2.1103920936584473, "learning_rate": 4.323995127892814e-05, "loss": 6.677, "step": 15775 }, { "epoch": 0.665134810380083, "grad_norm": 2.3970913887023926, "learning_rate": 4.323917049252007e-05, "loss": 6.6622, "step": 15776 }, { "epoch": 0.6651769715622826, "grad_norm": 2.9030354022979736, "learning_rate": 4.3238389706112e-05, "loss": 7.8645, "step": 15777 }, { "epoch": 0.6652191327444822, "grad_norm": 2.128412961959839, "learning_rate": 4.3237608919703925e-05, "loss": 6.4896, "step": 15778 }, { "epoch": 0.6652612939266817, "grad_norm": 1.584986925125122, "learning_rate": 4.3236828133295856e-05, "loss": 6.4698, "step": 15779 }, { "epoch": 0.6653034551088812, "grad_norm": 5.6571364402771, "learning_rate": 4.323604734688779e-05, "loss": 6.992, "step": 15780 }, { "epoch": 0.6653456162910808, "grad_norm": 2.454901695251465, "learning_rate": 4.323526656047971e-05, "loss": 7.002, "step": 15781 }, { "epoch": 0.6653877774732804, "grad_norm": 3.470964193344116, "learning_rate": 4.323448577407165e-05, "loss": 7.1752, "step": 15782 }, { "epoch": 0.66542993865548, "grad_norm": 3.0712718963623047, "learning_rate": 4.3233704987663574e-05, "loss": 7.7972, "step": 15783 }, { "epoch": 0.6654720998376794, "grad_norm": 2.120194911956787, "learning_rate": 4.323292420125551e-05, "loss": 7.0199, "step": 15784 }, { "epoch": 0.665514261019879, "grad_norm": 1.9793051481246948, "learning_rate": 4.323214341484744e-05, "loss": 6.8738, "step": 15785 }, { "epoch": 0.6655564222020786, "grad_norm": 1.5059415102005005, "learning_rate": 4.323136262843936e-05, "loss": 6.672, "step": 15786 }, { "epoch": 0.6655985833842781, "grad_norm": 1.5721501111984253, "learning_rate": 4.32305818420313e-05, "loss": 6.7574, "step": 15787 }, { "epoch": 0.6656407445664776, "grad_norm": 2.675729990005493, "learning_rate": 4.3229801055623224e-05, "loss": 6.4501, "step": 15788 }, { "epoch": 0.6656829057486772, "grad_norm": 1.5764155387878418, "learning_rate": 4.3229020269215155e-05, "loss": 6.3398, "step": 15789 }, { "epoch": 0.6657250669308767, "grad_norm": 1.244850993156433, "learning_rate": 4.3228239482807086e-05, "loss": 6.374, "step": 15790 }, { "epoch": 0.6657672281130763, "grad_norm": 1.9827637672424316, "learning_rate": 4.322745869639902e-05, "loss": 7.0822, "step": 15791 }, { "epoch": 0.6658093892952759, "grad_norm": 2.7568540573120117, "learning_rate": 4.322667790999094e-05, "loss": 7.3878, "step": 15792 }, { "epoch": 0.6658515504774754, "grad_norm": 2.924222946166992, "learning_rate": 4.3225897123582873e-05, "loss": 7.0418, "step": 15793 }, { "epoch": 0.6658937116596749, "grad_norm": 3.0272862911224365, "learning_rate": 4.3225116337174805e-05, "loss": 7.5734, "step": 15794 }, { "epoch": 0.6659358728418745, "grad_norm": 3.5258893966674805, "learning_rate": 4.3224335550766736e-05, "loss": 7.7621, "step": 15795 }, { "epoch": 0.6659780340240741, "grad_norm": 2.3305747509002686, "learning_rate": 4.322355476435867e-05, "loss": 6.6289, "step": 15796 }, { "epoch": 0.6660201952062735, "grad_norm": 2.7482292652130127, "learning_rate": 4.322277397795059e-05, "loss": 6.3842, "step": 15797 }, { "epoch": 0.6660623563884731, "grad_norm": 1.6254910230636597, "learning_rate": 4.322199319154252e-05, "loss": 6.7826, "step": 15798 }, { "epoch": 0.6661045175706727, "grad_norm": 3.3328635692596436, "learning_rate": 4.3221212405134454e-05, "loss": 7.9716, "step": 15799 }, { "epoch": 0.6661466787528723, "grad_norm": 1.6711958646774292, "learning_rate": 4.322043161872638e-05, "loss": 7.357, "step": 15800 }, { "epoch": 0.6661888399350718, "grad_norm": 2.137608766555786, "learning_rate": 4.321965083231832e-05, "loss": 7.0656, "step": 15801 }, { "epoch": 0.6662310011172713, "grad_norm": 1.6163698434829712, "learning_rate": 4.321887004591024e-05, "loss": 7.217, "step": 15802 }, { "epoch": 0.6662731622994709, "grad_norm": 1.8910523653030396, "learning_rate": 4.321808925950217e-05, "loss": 7.0261, "step": 15803 }, { "epoch": 0.6663153234816704, "grad_norm": 2.0454819202423096, "learning_rate": 4.3217308473094104e-05, "loss": 6.6727, "step": 15804 }, { "epoch": 0.66635748466387, "grad_norm": 1.919880747795105, "learning_rate": 4.321652768668603e-05, "loss": 6.3554, "step": 15805 }, { "epoch": 0.6663996458460695, "grad_norm": 1.617277979850769, "learning_rate": 4.3215746900277966e-05, "loss": 6.8869, "step": 15806 }, { "epoch": 0.666441807028269, "grad_norm": 1.9254052639007568, "learning_rate": 4.321496611386989e-05, "loss": 6.2091, "step": 15807 }, { "epoch": 0.6664839682104686, "grad_norm": 1.9834623336791992, "learning_rate": 4.321418532746182e-05, "loss": 6.3107, "step": 15808 }, { "epoch": 0.6665261293926682, "grad_norm": 1.3046802282333374, "learning_rate": 4.321340454105375e-05, "loss": 6.2914, "step": 15809 }, { "epoch": 0.6665682905748678, "grad_norm": 2.1867997646331787, "learning_rate": 4.3212623754645685e-05, "loss": 6.9118, "step": 15810 }, { "epoch": 0.6666104517570672, "grad_norm": 2.6792571544647217, "learning_rate": 4.321184296823761e-05, "loss": 6.5382, "step": 15811 }, { "epoch": 0.6666526129392668, "grad_norm": 1.5363342761993408, "learning_rate": 4.321106218182954e-05, "loss": 6.3342, "step": 15812 }, { "epoch": 0.6666947741214664, "grad_norm": 2.3662495613098145, "learning_rate": 4.321028139542147e-05, "loss": 7.0308, "step": 15813 }, { "epoch": 0.666736935303666, "grad_norm": 2.2691292762756348, "learning_rate": 4.3209500609013396e-05, "loss": 6.7747, "step": 15814 }, { "epoch": 0.6667790964858654, "grad_norm": 2.286756992340088, "learning_rate": 4.3208719822605334e-05, "loss": 6.6463, "step": 15815 }, { "epoch": 0.666821257668065, "grad_norm": 2.527787208557129, "learning_rate": 4.320793903619726e-05, "loss": 6.9867, "step": 15816 }, { "epoch": 0.6668634188502646, "grad_norm": 4.228785514831543, "learning_rate": 4.320715824978919e-05, "loss": 7.0672, "step": 15817 }, { "epoch": 0.6669055800324641, "grad_norm": 2.071269989013672, "learning_rate": 4.320637746338112e-05, "loss": 6.4128, "step": 15818 }, { "epoch": 0.6669477412146637, "grad_norm": 2.486510992050171, "learning_rate": 4.3205596676973046e-05, "loss": 6.6897, "step": 15819 }, { "epoch": 0.6669899023968632, "grad_norm": 1.842179536819458, "learning_rate": 4.3204815890564984e-05, "loss": 7.0266, "step": 15820 }, { "epoch": 0.6670320635790628, "grad_norm": 1.5298333168029785, "learning_rate": 4.320403510415691e-05, "loss": 6.3664, "step": 15821 }, { "epoch": 0.6670742247612623, "grad_norm": 3.971625566482544, "learning_rate": 4.320325431774884e-05, "loss": 7.5178, "step": 15822 }, { "epoch": 0.6671163859434619, "grad_norm": 1.833193302154541, "learning_rate": 4.320247353134077e-05, "loss": 7.0721, "step": 15823 }, { "epoch": 0.6671585471256614, "grad_norm": 3.128655433654785, "learning_rate": 4.3201692744932695e-05, "loss": 7.9138, "step": 15824 }, { "epoch": 0.6672007083078609, "grad_norm": 3.030895471572876, "learning_rate": 4.3200911958524626e-05, "loss": 6.4709, "step": 15825 }, { "epoch": 0.6672428694900605, "grad_norm": 1.653751254081726, "learning_rate": 4.320013117211656e-05, "loss": 6.6324, "step": 15826 }, { "epoch": 0.6672850306722601, "grad_norm": 1.8881756067276, "learning_rate": 4.319935038570849e-05, "loss": 6.5713, "step": 15827 }, { "epoch": 0.6673271918544595, "grad_norm": 2.3049967288970947, "learning_rate": 4.319856959930041e-05, "loss": 7.3543, "step": 15828 }, { "epoch": 0.6673693530366591, "grad_norm": 2.654784917831421, "learning_rate": 4.319778881289235e-05, "loss": 6.3946, "step": 15829 }, { "epoch": 0.6674115142188587, "grad_norm": 1.7840179204940796, "learning_rate": 4.3197008026484276e-05, "loss": 6.3351, "step": 15830 }, { "epoch": 0.6674536754010583, "grad_norm": 1.3474211692810059, "learning_rate": 4.319622724007621e-05, "loss": 6.813, "step": 15831 }, { "epoch": 0.6674958365832578, "grad_norm": 1.5823960304260254, "learning_rate": 4.319544645366814e-05, "loss": 7.0846, "step": 15832 }, { "epoch": 0.6675379977654573, "grad_norm": 2.190246105194092, "learning_rate": 4.319466566726006e-05, "loss": 6.8489, "step": 15833 }, { "epoch": 0.6675801589476569, "grad_norm": 1.5348149538040161, "learning_rate": 4.3193884880852e-05, "loss": 6.9738, "step": 15834 }, { "epoch": 0.6676223201298565, "grad_norm": 2.3222718238830566, "learning_rate": 4.3193104094443925e-05, "loss": 6.2604, "step": 15835 }, { "epoch": 0.667664481312056, "grad_norm": 1.1544262170791626, "learning_rate": 4.319232330803586e-05, "loss": 6.4982, "step": 15836 }, { "epoch": 0.6677066424942555, "grad_norm": 1.5322020053863525, "learning_rate": 4.319154252162779e-05, "loss": 6.3815, "step": 15837 }, { "epoch": 0.6677488036764551, "grad_norm": 1.6837865114212036, "learning_rate": 4.319076173521971e-05, "loss": 6.5127, "step": 15838 }, { "epoch": 0.6677909648586546, "grad_norm": 2.061927556991577, "learning_rate": 4.3189980948811644e-05, "loss": 6.4135, "step": 15839 }, { "epoch": 0.6678331260408542, "grad_norm": 1.7326747179031372, "learning_rate": 4.3189200162403575e-05, "loss": 6.8925, "step": 15840 }, { "epoch": 0.6678752872230538, "grad_norm": 1.2342312335968018, "learning_rate": 4.3188419375995506e-05, "loss": 6.3804, "step": 15841 }, { "epoch": 0.6679174484052532, "grad_norm": 1.9991050958633423, "learning_rate": 4.318763858958744e-05, "loss": 6.7163, "step": 15842 }, { "epoch": 0.6679596095874528, "grad_norm": 1.261755108833313, "learning_rate": 4.318685780317936e-05, "loss": 6.4325, "step": 15843 }, { "epoch": 0.6680017707696524, "grad_norm": 1.741598129272461, "learning_rate": 4.318607701677129e-05, "loss": 6.639, "step": 15844 }, { "epoch": 0.668043931951852, "grad_norm": 3.3344616889953613, "learning_rate": 4.3185296230363224e-05, "loss": 7.1488, "step": 15845 }, { "epoch": 0.6680860931340514, "grad_norm": 1.167807936668396, "learning_rate": 4.3184515443955156e-05, "loss": 6.2859, "step": 15846 }, { "epoch": 0.668128254316251, "grad_norm": 2.6374027729034424, "learning_rate": 4.318373465754708e-05, "loss": 6.7348, "step": 15847 }, { "epoch": 0.6681704154984506, "grad_norm": 1.2162079811096191, "learning_rate": 4.318295387113901e-05, "loss": 6.3804, "step": 15848 }, { "epoch": 0.6682125766806501, "grad_norm": 1.214010238647461, "learning_rate": 4.318217308473094e-05, "loss": 6.5025, "step": 15849 }, { "epoch": 0.6682547378628497, "grad_norm": 1.8441928625106812, "learning_rate": 4.318139229832287e-05, "loss": 6.9524, "step": 15850 }, { "epoch": 0.6682968990450492, "grad_norm": 2.0002081394195557, "learning_rate": 4.3180611511914805e-05, "loss": 6.3984, "step": 15851 }, { "epoch": 0.6683390602272488, "grad_norm": 2.9893529415130615, "learning_rate": 4.317983072550673e-05, "loss": 8.1511, "step": 15852 }, { "epoch": 0.6683812214094483, "grad_norm": 0.9973959922790527, "learning_rate": 4.317904993909867e-05, "loss": 6.381, "step": 15853 }, { "epoch": 0.6684233825916479, "grad_norm": 3.2148094177246094, "learning_rate": 4.317826915269059e-05, "loss": 7.8761, "step": 15854 }, { "epoch": 0.6684655437738474, "grad_norm": 1.5572117567062378, "learning_rate": 4.3177488366282523e-05, "loss": 6.9281, "step": 15855 }, { "epoch": 0.6685077049560469, "grad_norm": 1.252759337425232, "learning_rate": 4.3176707579874455e-05, "loss": 6.675, "step": 15856 }, { "epoch": 0.6685498661382465, "grad_norm": 1.1535965204238892, "learning_rate": 4.317592679346638e-05, "loss": 6.4197, "step": 15857 }, { "epoch": 0.6685920273204461, "grad_norm": 2.316662073135376, "learning_rate": 4.317514600705831e-05, "loss": 7.1224, "step": 15858 }, { "epoch": 0.6686341885026457, "grad_norm": 2.967278480529785, "learning_rate": 4.317436522065024e-05, "loss": 7.3495, "step": 15859 }, { "epoch": 0.6686763496848451, "grad_norm": 1.4076834917068481, "learning_rate": 4.317358443424217e-05, "loss": 6.5442, "step": 15860 }, { "epoch": 0.6687185108670447, "grad_norm": 1.4814293384552002, "learning_rate": 4.31728036478341e-05, "loss": 6.713, "step": 15861 }, { "epoch": 0.6687606720492443, "grad_norm": 2.4661126136779785, "learning_rate": 4.317202286142603e-05, "loss": 6.4676, "step": 15862 }, { "epoch": 0.6688028332314438, "grad_norm": 2.6905345916748047, "learning_rate": 4.317124207501796e-05, "loss": 6.993, "step": 15863 }, { "epoch": 0.6688449944136433, "grad_norm": 2.1020748615264893, "learning_rate": 4.317046128860989e-05, "loss": 6.4053, "step": 15864 }, { "epoch": 0.6688871555958429, "grad_norm": 2.5109801292419434, "learning_rate": 4.316968050220182e-05, "loss": 6.6185, "step": 15865 }, { "epoch": 0.6689293167780425, "grad_norm": 1.4101762771606445, "learning_rate": 4.316889971579375e-05, "loss": 6.6502, "step": 15866 }, { "epoch": 0.668971477960242, "grad_norm": 1.1368646621704102, "learning_rate": 4.316811892938568e-05, "loss": 6.656, "step": 15867 }, { "epoch": 0.6690136391424416, "grad_norm": 1.3379502296447754, "learning_rate": 4.316733814297761e-05, "loss": 6.3495, "step": 15868 }, { "epoch": 0.6690558003246411, "grad_norm": 2.667917251586914, "learning_rate": 4.3166557356569534e-05, "loss": 7.3198, "step": 15869 }, { "epoch": 0.6690979615068406, "grad_norm": 2.5877835750579834, "learning_rate": 4.316577657016147e-05, "loss": 6.4651, "step": 15870 }, { "epoch": 0.6691401226890402, "grad_norm": 3.4196619987487793, "learning_rate": 4.3164995783753397e-05, "loss": 7.9754, "step": 15871 }, { "epoch": 0.6691822838712398, "grad_norm": 1.4503518342971802, "learning_rate": 4.316421499734533e-05, "loss": 6.3781, "step": 15872 }, { "epoch": 0.6692244450534393, "grad_norm": 1.265715479850769, "learning_rate": 4.316343421093726e-05, "loss": 6.9896, "step": 15873 }, { "epoch": 0.6692666062356388, "grad_norm": 2.6813998222351074, "learning_rate": 4.3162653424529184e-05, "loss": 6.4163, "step": 15874 }, { "epoch": 0.6693087674178384, "grad_norm": 3.0788047313690186, "learning_rate": 4.316187263812112e-05, "loss": 7.6423, "step": 15875 }, { "epoch": 0.669350928600038, "grad_norm": 1.5454797744750977, "learning_rate": 4.3161091851713046e-05, "loss": 6.2195, "step": 15876 }, { "epoch": 0.6693930897822375, "grad_norm": 1.3258039951324463, "learning_rate": 4.316031106530498e-05, "loss": 6.9959, "step": 15877 }, { "epoch": 0.669435250964437, "grad_norm": 0.785789430141449, "learning_rate": 4.315953027889691e-05, "loss": 6.4064, "step": 15878 }, { "epoch": 0.6694774121466366, "grad_norm": 1.7220523357391357, "learning_rate": 4.315874949248884e-05, "loss": 6.4066, "step": 15879 }, { "epoch": 0.6695195733288362, "grad_norm": 4.097823143005371, "learning_rate": 4.3157968706080764e-05, "loss": 7.9282, "step": 15880 }, { "epoch": 0.6695617345110357, "grad_norm": 2.484166145324707, "learning_rate": 4.3157187919672696e-05, "loss": 6.842, "step": 15881 }, { "epoch": 0.6696038956932352, "grad_norm": 1.8728594779968262, "learning_rate": 4.315640713326463e-05, "loss": 6.8189, "step": 15882 }, { "epoch": 0.6696460568754348, "grad_norm": 1.2023539543151855, "learning_rate": 4.315562634685655e-05, "loss": 6.7861, "step": 15883 }, { "epoch": 0.6696882180576343, "grad_norm": 2.665156841278076, "learning_rate": 4.315484556044849e-05, "loss": 8.3197, "step": 15884 }, { "epoch": 0.6697303792398339, "grad_norm": 3.9044699668884277, "learning_rate": 4.3154064774040414e-05, "loss": 7.0107, "step": 15885 }, { "epoch": 0.6697725404220334, "grad_norm": 2.6820409297943115, "learning_rate": 4.3153283987632345e-05, "loss": 6.5294, "step": 15886 }, { "epoch": 0.669814701604233, "grad_norm": 2.025132417678833, "learning_rate": 4.3152503201224276e-05, "loss": 7.0204, "step": 15887 }, { "epoch": 0.6698568627864325, "grad_norm": 1.6299562454223633, "learning_rate": 4.31517224148162e-05, "loss": 6.669, "step": 15888 }, { "epoch": 0.6698990239686321, "grad_norm": 1.4309207201004028, "learning_rate": 4.315094162840814e-05, "loss": 6.7145, "step": 15889 }, { "epoch": 0.6699411851508317, "grad_norm": 4.249679088592529, "learning_rate": 4.315016084200006e-05, "loss": 8.0266, "step": 15890 }, { "epoch": 0.6699833463330311, "grad_norm": 2.196716070175171, "learning_rate": 4.3149380055591995e-05, "loss": 7.1097, "step": 15891 }, { "epoch": 0.6700255075152307, "grad_norm": 3.0184621810913086, "learning_rate": 4.3148599269183926e-05, "loss": 8.046, "step": 15892 }, { "epoch": 0.6700676686974303, "grad_norm": 1.9224975109100342, "learning_rate": 4.314781848277585e-05, "loss": 7.4188, "step": 15893 }, { "epoch": 0.6701098298796299, "grad_norm": 1.9345086812973022, "learning_rate": 4.314703769636778e-05, "loss": 7.0443, "step": 15894 }, { "epoch": 0.6701519910618293, "grad_norm": 2.222275733947754, "learning_rate": 4.314625690995971e-05, "loss": 7.988, "step": 15895 }, { "epoch": 0.6701941522440289, "grad_norm": 3.6189768314361572, "learning_rate": 4.3145476123551644e-05, "loss": 7.1956, "step": 15896 }, { "epoch": 0.6702363134262285, "grad_norm": 2.984890937805176, "learning_rate": 4.3144695337143575e-05, "loss": 7.4776, "step": 15897 }, { "epoch": 0.670278474608428, "grad_norm": 2.2049474716186523, "learning_rate": 4.314391455073551e-05, "loss": 6.8724, "step": 15898 }, { "epoch": 0.6703206357906276, "grad_norm": 2.362637758255005, "learning_rate": 4.314313376432743e-05, "loss": 7.1103, "step": 15899 }, { "epoch": 0.6703627969728271, "grad_norm": 1.9625619649887085, "learning_rate": 4.314235297791936e-05, "loss": 7.6415, "step": 15900 }, { "epoch": 0.6704049581550267, "grad_norm": 1.9950909614562988, "learning_rate": 4.3141572191511294e-05, "loss": 6.4809, "step": 15901 }, { "epoch": 0.6704471193372262, "grad_norm": 2.2765536308288574, "learning_rate": 4.314079140510322e-05, "loss": 6.7376, "step": 15902 }, { "epoch": 0.6704892805194258, "grad_norm": 2.8584675788879395, "learning_rate": 4.3140010618695156e-05, "loss": 6.9736, "step": 15903 }, { "epoch": 0.6705314417016253, "grad_norm": 1.7236249446868896, "learning_rate": 4.313922983228708e-05, "loss": 6.5563, "step": 15904 }, { "epoch": 0.6705736028838248, "grad_norm": 1.9531069993972778, "learning_rate": 4.313844904587901e-05, "loss": 6.357, "step": 15905 }, { "epoch": 0.6706157640660244, "grad_norm": 2.565650224685669, "learning_rate": 4.313766825947094e-05, "loss": 6.884, "step": 15906 }, { "epoch": 0.670657925248224, "grad_norm": 2.281052350997925, "learning_rate": 4.313688747306287e-05, "loss": 6.4994, "step": 15907 }, { "epoch": 0.6707000864304236, "grad_norm": 1.742991328239441, "learning_rate": 4.31361066866548e-05, "loss": 6.4871, "step": 15908 }, { "epoch": 0.670742247612623, "grad_norm": 3.2877047061920166, "learning_rate": 4.313532590024673e-05, "loss": 7.3808, "step": 15909 }, { "epoch": 0.6707844087948226, "grad_norm": 2.134474277496338, "learning_rate": 4.313454511383866e-05, "loss": 7.4121, "step": 15910 }, { "epoch": 0.6708265699770222, "grad_norm": 2.0818569660186768, "learning_rate": 4.313376432743059e-05, "loss": 6.7443, "step": 15911 }, { "epoch": 0.6708687311592217, "grad_norm": 8.183237075805664, "learning_rate": 4.313298354102252e-05, "loss": 7.5842, "step": 15912 }, { "epoch": 0.6709108923414212, "grad_norm": 3.807852029800415, "learning_rate": 4.313220275461445e-05, "loss": 7.7385, "step": 15913 }, { "epoch": 0.6709530535236208, "grad_norm": 2.4712305068969727, "learning_rate": 4.313142196820638e-05, "loss": 7.242, "step": 15914 }, { "epoch": 0.6709952147058204, "grad_norm": 2.6236259937286377, "learning_rate": 4.313064118179831e-05, "loss": 7.1028, "step": 15915 }, { "epoch": 0.6710373758880199, "grad_norm": 2.2204294204711914, "learning_rate": 4.3129860395390235e-05, "loss": 7.0978, "step": 15916 }, { "epoch": 0.6710795370702195, "grad_norm": 2.3522586822509766, "learning_rate": 4.3129079608982174e-05, "loss": 6.4829, "step": 15917 }, { "epoch": 0.671121698252419, "grad_norm": 2.0188469886779785, "learning_rate": 4.31282988225741e-05, "loss": 6.2321, "step": 15918 }, { "epoch": 0.6711638594346185, "grad_norm": 2.778987407684326, "learning_rate": 4.312751803616602e-05, "loss": 7.4445, "step": 15919 }, { "epoch": 0.6712060206168181, "grad_norm": 2.075134515762329, "learning_rate": 4.312673724975796e-05, "loss": 6.3798, "step": 15920 }, { "epoch": 0.6712481817990177, "grad_norm": 1.4816399812698364, "learning_rate": 4.3125956463349885e-05, "loss": 6.7575, "step": 15921 }, { "epoch": 0.6712903429812171, "grad_norm": 2.0692994594573975, "learning_rate": 4.312517567694182e-05, "loss": 7.4217, "step": 15922 }, { "epoch": 0.6713325041634167, "grad_norm": 2.1490862369537354, "learning_rate": 4.312439489053375e-05, "loss": 7.1716, "step": 15923 }, { "epoch": 0.6713746653456163, "grad_norm": 1.9982936382293701, "learning_rate": 4.312361410412568e-05, "loss": 6.715, "step": 15924 }, { "epoch": 0.6714168265278159, "grad_norm": 1.9928196668624878, "learning_rate": 4.312283331771761e-05, "loss": 6.435, "step": 15925 }, { "epoch": 0.6714589877100154, "grad_norm": 2.518481492996216, "learning_rate": 4.3122052531309535e-05, "loss": 6.5062, "step": 15926 }, { "epoch": 0.6715011488922149, "grad_norm": 2.42634654045105, "learning_rate": 4.3121271744901466e-05, "loss": 6.5918, "step": 15927 }, { "epoch": 0.6715433100744145, "grad_norm": 2.0847036838531494, "learning_rate": 4.31204909584934e-05, "loss": 7.1741, "step": 15928 }, { "epoch": 0.671585471256614, "grad_norm": 1.701890230178833, "learning_rate": 4.311971017208533e-05, "loss": 6.6936, "step": 15929 }, { "epoch": 0.6716276324388136, "grad_norm": 1.6238274574279785, "learning_rate": 4.311892938567725e-05, "loss": 6.4104, "step": 15930 }, { "epoch": 0.6716697936210131, "grad_norm": 0.928548276424408, "learning_rate": 4.3118148599269184e-05, "loss": 6.4336, "step": 15931 }, { "epoch": 0.6717119548032127, "grad_norm": 4.755917072296143, "learning_rate": 4.3117367812861115e-05, "loss": 7.2032, "step": 15932 }, { "epoch": 0.6717541159854122, "grad_norm": 1.9381362199783325, "learning_rate": 4.3116587026453047e-05, "loss": 6.7791, "step": 15933 }, { "epoch": 0.6717962771676118, "grad_norm": 3.714101552963257, "learning_rate": 4.311580624004498e-05, "loss": 7.7574, "step": 15934 }, { "epoch": 0.6718384383498113, "grad_norm": 1.7180700302124023, "learning_rate": 4.31150254536369e-05, "loss": 6.8728, "step": 15935 }, { "epoch": 0.6718805995320108, "grad_norm": 1.3355412483215332, "learning_rate": 4.3114244667228834e-05, "loss": 6.9013, "step": 15936 }, { "epoch": 0.6719227607142104, "grad_norm": 1.3006412982940674, "learning_rate": 4.3113463880820765e-05, "loss": 7.1838, "step": 15937 }, { "epoch": 0.67196492189641, "grad_norm": 3.5118203163146973, "learning_rate": 4.311268309441269e-05, "loss": 7.5082, "step": 15938 }, { "epoch": 0.6720070830786096, "grad_norm": 2.4952681064605713, "learning_rate": 4.311190230800463e-05, "loss": 6.4725, "step": 15939 }, { "epoch": 0.672049244260809, "grad_norm": 2.4718692302703857, "learning_rate": 4.311112152159655e-05, "loss": 7.028, "step": 15940 }, { "epoch": 0.6720914054430086, "grad_norm": 2.0508809089660645, "learning_rate": 4.311034073518848e-05, "loss": 6.4542, "step": 15941 }, { "epoch": 0.6721335666252082, "grad_norm": 2.3310844898223877, "learning_rate": 4.3109559948780414e-05, "loss": 6.9578, "step": 15942 }, { "epoch": 0.6721757278074078, "grad_norm": 4.301896095275879, "learning_rate": 4.3108779162372346e-05, "loss": 7.1643, "step": 15943 }, { "epoch": 0.6722178889896072, "grad_norm": 2.369197130203247, "learning_rate": 4.310799837596428e-05, "loss": 6.7407, "step": 15944 }, { "epoch": 0.6722600501718068, "grad_norm": 3.2283480167388916, "learning_rate": 4.31072175895562e-05, "loss": 7.3723, "step": 15945 }, { "epoch": 0.6723022113540064, "grad_norm": 2.3493406772613525, "learning_rate": 4.310643680314813e-05, "loss": 6.3355, "step": 15946 }, { "epoch": 0.6723443725362059, "grad_norm": 2.404693603515625, "learning_rate": 4.3105656016740064e-05, "loss": 7.8673, "step": 15947 }, { "epoch": 0.6723865337184055, "grad_norm": 2.080338478088379, "learning_rate": 4.3104875230331995e-05, "loss": 6.2616, "step": 15948 }, { "epoch": 0.672428694900605, "grad_norm": 3.5421531200408936, "learning_rate": 4.310409444392392e-05, "loss": 7.826, "step": 15949 }, { "epoch": 0.6724708560828045, "grad_norm": 2.390263319015503, "learning_rate": 4.310331365751585e-05, "loss": 7.2164, "step": 15950 }, { "epoch": 0.6725130172650041, "grad_norm": 3.023254632949829, "learning_rate": 4.310253287110778e-05, "loss": 6.3861, "step": 15951 }, { "epoch": 0.6725551784472037, "grad_norm": 3.1372175216674805, "learning_rate": 4.310175208469971e-05, "loss": 6.2428, "step": 15952 }, { "epoch": 0.6725973396294032, "grad_norm": 1.331396222114563, "learning_rate": 4.3100971298291645e-05, "loss": 6.9205, "step": 15953 }, { "epoch": 0.6726395008116027, "grad_norm": 2.599576473236084, "learning_rate": 4.310019051188357e-05, "loss": 7.3298, "step": 15954 }, { "epoch": 0.6726816619938023, "grad_norm": 1.4622806310653687, "learning_rate": 4.30994097254755e-05, "loss": 6.619, "step": 15955 }, { "epoch": 0.6727238231760019, "grad_norm": 2.5067546367645264, "learning_rate": 4.309862893906743e-05, "loss": 7.1554, "step": 15956 }, { "epoch": 0.6727659843582015, "grad_norm": 1.778982400894165, "learning_rate": 4.3097848152659356e-05, "loss": 6.8372, "step": 15957 }, { "epoch": 0.6728081455404009, "grad_norm": 1.6068588495254517, "learning_rate": 4.3097067366251294e-05, "loss": 7.1591, "step": 15958 }, { "epoch": 0.6728503067226005, "grad_norm": 2.032413959503174, "learning_rate": 4.309628657984322e-05, "loss": 7.3647, "step": 15959 }, { "epoch": 0.6728924679048001, "grad_norm": 1.9311503171920776, "learning_rate": 4.309550579343515e-05, "loss": 6.7549, "step": 15960 }, { "epoch": 0.6729346290869996, "grad_norm": 2.11334228515625, "learning_rate": 4.309472500702708e-05, "loss": 6.8311, "step": 15961 }, { "epoch": 0.6729767902691991, "grad_norm": 2.458552837371826, "learning_rate": 4.309394422061901e-05, "loss": 6.3912, "step": 15962 }, { "epoch": 0.6730189514513987, "grad_norm": 1.2516722679138184, "learning_rate": 4.309316343421094e-05, "loss": 6.8502, "step": 15963 }, { "epoch": 0.6730611126335982, "grad_norm": 3.4615044593811035, "learning_rate": 4.309238264780287e-05, "loss": 7.8821, "step": 15964 }, { "epoch": 0.6731032738157978, "grad_norm": 2.398535966873169, "learning_rate": 4.30916018613948e-05, "loss": 6.7159, "step": 15965 }, { "epoch": 0.6731454349979974, "grad_norm": 1.63722562789917, "learning_rate": 4.309082107498673e-05, "loss": 6.4246, "step": 15966 }, { "epoch": 0.6731875961801969, "grad_norm": 2.91601824760437, "learning_rate": 4.309004028857866e-05, "loss": 7.7855, "step": 15967 }, { "epoch": 0.6732297573623964, "grad_norm": 3.6439208984375, "learning_rate": 4.3089259502170586e-05, "loss": 7.6142, "step": 15968 }, { "epoch": 0.673271918544596, "grad_norm": 1.5903645753860474, "learning_rate": 4.308847871576252e-05, "loss": 6.7405, "step": 15969 }, { "epoch": 0.6733140797267956, "grad_norm": 3.3337762355804443, "learning_rate": 4.308769792935445e-05, "loss": 7.7603, "step": 15970 }, { "epoch": 0.673356240908995, "grad_norm": 2.3674557209014893, "learning_rate": 4.3086917142946373e-05, "loss": 6.4799, "step": 15971 }, { "epoch": 0.6733984020911946, "grad_norm": 2.1220552921295166, "learning_rate": 4.308613635653831e-05, "loss": 7.0507, "step": 15972 }, { "epoch": 0.6734405632733942, "grad_norm": 2.2395050525665283, "learning_rate": 4.3085355570130236e-05, "loss": 6.4262, "step": 15973 }, { "epoch": 0.6734827244555938, "grad_norm": 2.197486162185669, "learning_rate": 4.308457478372217e-05, "loss": 6.923, "step": 15974 }, { "epoch": 0.6735248856377933, "grad_norm": 2.441345691680908, "learning_rate": 4.30837939973141e-05, "loss": 6.9828, "step": 15975 }, { "epoch": 0.6735670468199928, "grad_norm": 2.4811925888061523, "learning_rate": 4.308301321090602e-05, "loss": 7.0271, "step": 15976 }, { "epoch": 0.6736092080021924, "grad_norm": 4.2999114990234375, "learning_rate": 4.308223242449796e-05, "loss": 7.2432, "step": 15977 }, { "epoch": 0.673651369184392, "grad_norm": 2.9410789012908936, "learning_rate": 4.3081451638089885e-05, "loss": 7.1524, "step": 15978 }, { "epoch": 0.6736935303665915, "grad_norm": 2.208284378051758, "learning_rate": 4.308067085168182e-05, "loss": 7.0133, "step": 15979 }, { "epoch": 0.673735691548791, "grad_norm": 1.900193452835083, "learning_rate": 4.307989006527375e-05, "loss": 6.6021, "step": 15980 }, { "epoch": 0.6737778527309906, "grad_norm": 1.4711452722549438, "learning_rate": 4.307910927886567e-05, "loss": 6.7187, "step": 15981 }, { "epoch": 0.6738200139131901, "grad_norm": 2.4437942504882812, "learning_rate": 4.3078328492457604e-05, "loss": 7.1462, "step": 15982 }, { "epoch": 0.6738621750953897, "grad_norm": 3.2191710472106934, "learning_rate": 4.3077547706049535e-05, "loss": 6.2749, "step": 15983 }, { "epoch": 0.6739043362775893, "grad_norm": 3.250155210494995, "learning_rate": 4.3076766919641466e-05, "loss": 6.3033, "step": 15984 }, { "epoch": 0.6739464974597887, "grad_norm": 3.0400142669677734, "learning_rate": 4.307598613323339e-05, "loss": 6.283, "step": 15985 }, { "epoch": 0.6739886586419883, "grad_norm": 1.7175294160842896, "learning_rate": 4.307520534682533e-05, "loss": 6.5096, "step": 15986 }, { "epoch": 0.6740308198241879, "grad_norm": 1.5101736783981323, "learning_rate": 4.307442456041725e-05, "loss": 6.3643, "step": 15987 }, { "epoch": 0.6740729810063875, "grad_norm": 1.8610180616378784, "learning_rate": 4.3073643774009185e-05, "loss": 6.8159, "step": 15988 }, { "epoch": 0.6741151421885869, "grad_norm": 1.69087815284729, "learning_rate": 4.3072862987601116e-05, "loss": 6.6419, "step": 15989 }, { "epoch": 0.6741573033707865, "grad_norm": 2.6427009105682373, "learning_rate": 4.307208220119304e-05, "loss": 7.1257, "step": 15990 }, { "epoch": 0.6741994645529861, "grad_norm": 1.4416148662567139, "learning_rate": 4.307130141478498e-05, "loss": 6.4655, "step": 15991 }, { "epoch": 0.6742416257351856, "grad_norm": 1.441084623336792, "learning_rate": 4.30705206283769e-05, "loss": 6.5788, "step": 15992 }, { "epoch": 0.6742837869173851, "grad_norm": 2.599818706512451, "learning_rate": 4.3069739841968834e-05, "loss": 7.3964, "step": 15993 }, { "epoch": 0.6743259480995847, "grad_norm": 2.012321710586548, "learning_rate": 4.3068959055560765e-05, "loss": 6.462, "step": 15994 }, { "epoch": 0.6743681092817843, "grad_norm": 2.5244815349578857, "learning_rate": 4.306817826915269e-05, "loss": 7.5592, "step": 15995 }, { "epoch": 0.6744102704639838, "grad_norm": 1.7050817012786865, "learning_rate": 4.306739748274462e-05, "loss": 6.404, "step": 15996 }, { "epoch": 0.6744524316461834, "grad_norm": 1.8060455322265625, "learning_rate": 4.306661669633655e-05, "loss": 7.0813, "step": 15997 }, { "epoch": 0.6744945928283829, "grad_norm": 4.092238426208496, "learning_rate": 4.3065835909928484e-05, "loss": 7.3293, "step": 15998 }, { "epoch": 0.6745367540105824, "grad_norm": 3.9297783374786377, "learning_rate": 4.306505512352041e-05, "loss": 7.7313, "step": 15999 }, { "epoch": 0.674578915192782, "grad_norm": 2.3943896293640137, "learning_rate": 4.306427433711234e-05, "loss": 6.4063, "step": 16000 }, { "epoch": 0.6746210763749816, "grad_norm": 2.376786470413208, "learning_rate": 4.306349355070427e-05, "loss": 6.7108, "step": 16001 }, { "epoch": 0.674663237557181, "grad_norm": 2.6084978580474854, "learning_rate": 4.30627127642962e-05, "loss": 7.4064, "step": 16002 }, { "epoch": 0.6747053987393806, "grad_norm": 2.152902603149414, "learning_rate": 4.306193197788813e-05, "loss": 6.751, "step": 16003 }, { "epoch": 0.6747475599215802, "grad_norm": 3.3691229820251465, "learning_rate": 4.306115119148006e-05, "loss": 6.6112, "step": 16004 }, { "epoch": 0.6747897211037798, "grad_norm": 3.4290010929107666, "learning_rate": 4.3060370405071996e-05, "loss": 6.3892, "step": 16005 }, { "epoch": 0.6748318822859793, "grad_norm": 3.5626184940338135, "learning_rate": 4.305958961866392e-05, "loss": 6.6708, "step": 16006 }, { "epoch": 0.6748740434681788, "grad_norm": 1.45986807346344, "learning_rate": 4.305880883225585e-05, "loss": 6.6053, "step": 16007 }, { "epoch": 0.6749162046503784, "grad_norm": 3.7826812267303467, "learning_rate": 4.305802804584778e-05, "loss": 7.4332, "step": 16008 }, { "epoch": 0.674958365832578, "grad_norm": 2.9798593521118164, "learning_rate": 4.305724725943971e-05, "loss": 7.2609, "step": 16009 }, { "epoch": 0.6750005270147775, "grad_norm": 2.193690776824951, "learning_rate": 4.305646647303164e-05, "loss": 6.5973, "step": 16010 }, { "epoch": 0.675042688196977, "grad_norm": 2.7881200313568115, "learning_rate": 4.305568568662357e-05, "loss": 7.1609, "step": 16011 }, { "epoch": 0.6750848493791766, "grad_norm": 1.8610621690750122, "learning_rate": 4.30549049002155e-05, "loss": 6.37, "step": 16012 }, { "epoch": 0.6751270105613761, "grad_norm": 1.8416537046432495, "learning_rate": 4.305412411380743e-05, "loss": 6.5888, "step": 16013 }, { "epoch": 0.6751691717435757, "grad_norm": 2.380474805831909, "learning_rate": 4.305334332739936e-05, "loss": 7.3055, "step": 16014 }, { "epoch": 0.6752113329257753, "grad_norm": 2.393444776535034, "learning_rate": 4.305256254099129e-05, "loss": 7.1648, "step": 16015 }, { "epoch": 0.6752534941079747, "grad_norm": 2.6273765563964844, "learning_rate": 4.305178175458322e-05, "loss": 6.9681, "step": 16016 }, { "epoch": 0.6752956552901743, "grad_norm": 2.2049813270568848, "learning_rate": 4.305100096817515e-05, "loss": 6.681, "step": 16017 }, { "epoch": 0.6753378164723739, "grad_norm": 3.512354612350464, "learning_rate": 4.3050220181767075e-05, "loss": 7.495, "step": 16018 }, { "epoch": 0.6753799776545735, "grad_norm": 4.132789134979248, "learning_rate": 4.3049439395359006e-05, "loss": 7.7413, "step": 16019 }, { "epoch": 0.6754221388367729, "grad_norm": 3.7764334678649902, "learning_rate": 4.304865860895094e-05, "loss": 6.9189, "step": 16020 }, { "epoch": 0.6754643000189725, "grad_norm": 2.7210299968719482, "learning_rate": 4.304787782254286e-05, "loss": 6.3557, "step": 16021 }, { "epoch": 0.6755064612011721, "grad_norm": 3.6220808029174805, "learning_rate": 4.30470970361348e-05, "loss": 7.7153, "step": 16022 }, { "epoch": 0.6755486223833717, "grad_norm": 2.166762113571167, "learning_rate": 4.3046316249726724e-05, "loss": 6.39, "step": 16023 }, { "epoch": 0.6755907835655712, "grad_norm": 2.1543571949005127, "learning_rate": 4.304553546331866e-05, "loss": 6.3923, "step": 16024 }, { "epoch": 0.6756329447477707, "grad_norm": 2.629348039627075, "learning_rate": 4.304475467691059e-05, "loss": 6.5401, "step": 16025 }, { "epoch": 0.6756751059299703, "grad_norm": 1.0951197147369385, "learning_rate": 4.304397389050251e-05, "loss": 7.2293, "step": 16026 }, { "epoch": 0.6757172671121698, "grad_norm": 4.493756294250488, "learning_rate": 4.304319310409445e-05, "loss": 9.0371, "step": 16027 }, { "epoch": 0.6757594282943694, "grad_norm": 1.9428108930587769, "learning_rate": 4.3042412317686374e-05, "loss": 6.8488, "step": 16028 }, { "epoch": 0.6758015894765689, "grad_norm": 3.9014382362365723, "learning_rate": 4.3041631531278305e-05, "loss": 7.1021, "step": 16029 }, { "epoch": 0.6758437506587684, "grad_norm": 1.6526378393173218, "learning_rate": 4.3040850744870236e-05, "loss": 6.2884, "step": 16030 }, { "epoch": 0.675885911840968, "grad_norm": 1.5796903371810913, "learning_rate": 4.304006995846217e-05, "loss": 6.3553, "step": 16031 }, { "epoch": 0.6759280730231676, "grad_norm": 1.924961805343628, "learning_rate": 4.303928917205409e-05, "loss": 7.1224, "step": 16032 }, { "epoch": 0.6759702342053672, "grad_norm": 2.213435411453247, "learning_rate": 4.3038508385646023e-05, "loss": 7.1931, "step": 16033 }, { "epoch": 0.6760123953875666, "grad_norm": 1.8598897457122803, "learning_rate": 4.3037727599237955e-05, "loss": 6.5425, "step": 16034 }, { "epoch": 0.6760545565697662, "grad_norm": 1.3900550603866577, "learning_rate": 4.3036946812829886e-05, "loss": 6.8428, "step": 16035 }, { "epoch": 0.6760967177519658, "grad_norm": 1.8851799964904785, "learning_rate": 4.303616602642182e-05, "loss": 6.7621, "step": 16036 }, { "epoch": 0.6761388789341654, "grad_norm": 1.290193796157837, "learning_rate": 4.303538524001374e-05, "loss": 6.5074, "step": 16037 }, { "epoch": 0.6761810401163648, "grad_norm": 1.5781583786010742, "learning_rate": 4.303460445360567e-05, "loss": 7.2992, "step": 16038 }, { "epoch": 0.6762232012985644, "grad_norm": 1.3132885694503784, "learning_rate": 4.3033823667197604e-05, "loss": 7.0356, "step": 16039 }, { "epoch": 0.676265362480764, "grad_norm": 1.980993390083313, "learning_rate": 4.303304288078953e-05, "loss": 6.436, "step": 16040 }, { "epoch": 0.6763075236629635, "grad_norm": 2.687058210372925, "learning_rate": 4.303226209438147e-05, "loss": 6.6245, "step": 16041 }, { "epoch": 0.6763496848451631, "grad_norm": 2.743544816970825, "learning_rate": 4.303148130797339e-05, "loss": 7.7852, "step": 16042 }, { "epoch": 0.6763918460273626, "grad_norm": 1.4685242176055908, "learning_rate": 4.303070052156532e-05, "loss": 6.4766, "step": 16043 }, { "epoch": 0.6764340072095621, "grad_norm": 1.9852068424224854, "learning_rate": 4.3029919735157254e-05, "loss": 6.7344, "step": 16044 }, { "epoch": 0.6764761683917617, "grad_norm": 2.2828238010406494, "learning_rate": 4.302913894874918e-05, "loss": 6.3217, "step": 16045 }, { "epoch": 0.6765183295739613, "grad_norm": 1.3908535242080688, "learning_rate": 4.3028358162341116e-05, "loss": 6.4982, "step": 16046 }, { "epoch": 0.6765604907561608, "grad_norm": 1.935227394104004, "learning_rate": 4.302757737593304e-05, "loss": 7.1088, "step": 16047 }, { "epoch": 0.6766026519383603, "grad_norm": 1.4508336782455444, "learning_rate": 4.302679658952497e-05, "loss": 6.8864, "step": 16048 }, { "epoch": 0.6766448131205599, "grad_norm": 2.9560728073120117, "learning_rate": 4.30260158031169e-05, "loss": 7.1378, "step": 16049 }, { "epoch": 0.6766869743027595, "grad_norm": 2.484212875366211, "learning_rate": 4.3025235016708835e-05, "loss": 7.8011, "step": 16050 }, { "epoch": 0.6767291354849589, "grad_norm": 2.950650453567505, "learning_rate": 4.302445423030076e-05, "loss": 6.5316, "step": 16051 }, { "epoch": 0.6767712966671585, "grad_norm": 2.260836362838745, "learning_rate": 4.302367344389269e-05, "loss": 6.5461, "step": 16052 }, { "epoch": 0.6768134578493581, "grad_norm": 1.7272380590438843, "learning_rate": 4.302289265748462e-05, "loss": 7.1476, "step": 16053 }, { "epoch": 0.6768556190315577, "grad_norm": 1.2218728065490723, "learning_rate": 4.3022111871076546e-05, "loss": 6.4623, "step": 16054 }, { "epoch": 0.6768977802137572, "grad_norm": 2.5265703201293945, "learning_rate": 4.3021331084668484e-05, "loss": 7.2304, "step": 16055 }, { "epoch": 0.6769399413959567, "grad_norm": 2.8225021362304688, "learning_rate": 4.302055029826041e-05, "loss": 6.5764, "step": 16056 }, { "epoch": 0.6769821025781563, "grad_norm": 2.5344455242156982, "learning_rate": 4.301976951185234e-05, "loss": 6.485, "step": 16057 }, { "epoch": 0.6770242637603558, "grad_norm": 1.967244267463684, "learning_rate": 4.301898872544427e-05, "loss": 7.1577, "step": 16058 }, { "epoch": 0.6770664249425554, "grad_norm": 2.15409779548645, "learning_rate": 4.3018207939036196e-05, "loss": 6.8454, "step": 16059 }, { "epoch": 0.6771085861247549, "grad_norm": 2.2594668865203857, "learning_rate": 4.3017427152628134e-05, "loss": 6.4603, "step": 16060 }, { "epoch": 0.6771507473069545, "grad_norm": 2.161661386489868, "learning_rate": 4.301664636622006e-05, "loss": 7.0029, "step": 16061 }, { "epoch": 0.677192908489154, "grad_norm": 3.314008951187134, "learning_rate": 4.301586557981199e-05, "loss": 7.6974, "step": 16062 }, { "epoch": 0.6772350696713536, "grad_norm": 3.6472220420837402, "learning_rate": 4.301508479340392e-05, "loss": 6.8598, "step": 16063 }, { "epoch": 0.6772772308535532, "grad_norm": 1.788238763809204, "learning_rate": 4.3014304006995845e-05, "loss": 6.5106, "step": 16064 }, { "epoch": 0.6773193920357526, "grad_norm": 1.7666972875595093, "learning_rate": 4.3013523220587776e-05, "loss": 7.2573, "step": 16065 }, { "epoch": 0.6773615532179522, "grad_norm": 2.8251121044158936, "learning_rate": 4.301274243417971e-05, "loss": 7.8016, "step": 16066 }, { "epoch": 0.6774037144001518, "grad_norm": 3.2318129539489746, "learning_rate": 4.301196164777164e-05, "loss": 7.6396, "step": 16067 }, { "epoch": 0.6774458755823514, "grad_norm": 2.4486920833587646, "learning_rate": 4.301118086136356e-05, "loss": 6.607, "step": 16068 }, { "epoch": 0.6774880367645508, "grad_norm": 1.9071253538131714, "learning_rate": 4.30104000749555e-05, "loss": 6.3638, "step": 16069 }, { "epoch": 0.6775301979467504, "grad_norm": 3.423356056213379, "learning_rate": 4.3009619288547426e-05, "loss": 7.6775, "step": 16070 }, { "epoch": 0.67757235912895, "grad_norm": 3.2404544353485107, "learning_rate": 4.300883850213936e-05, "loss": 7.2479, "step": 16071 }, { "epoch": 0.6776145203111495, "grad_norm": 1.746062994003296, "learning_rate": 4.300805771573129e-05, "loss": 6.3441, "step": 16072 }, { "epoch": 0.6776566814933491, "grad_norm": 2.461245059967041, "learning_rate": 4.300727692932321e-05, "loss": 7.0327, "step": 16073 }, { "epoch": 0.6776988426755486, "grad_norm": 1.5612411499023438, "learning_rate": 4.300649614291515e-05, "loss": 6.6978, "step": 16074 }, { "epoch": 0.6777410038577482, "grad_norm": 1.4571218490600586, "learning_rate": 4.3005715356507075e-05, "loss": 6.5072, "step": 16075 }, { "epoch": 0.6777831650399477, "grad_norm": 2.99249005317688, "learning_rate": 4.300493457009901e-05, "loss": 6.217, "step": 16076 }, { "epoch": 0.6778253262221473, "grad_norm": 2.8296430110931396, "learning_rate": 4.300415378369094e-05, "loss": 6.3155, "step": 16077 }, { "epoch": 0.6778674874043468, "grad_norm": 1.7041441202163696, "learning_rate": 4.300337299728286e-05, "loss": 6.759, "step": 16078 }, { "epoch": 0.6779096485865463, "grad_norm": 1.2848970890045166, "learning_rate": 4.3002592210874794e-05, "loss": 6.6122, "step": 16079 }, { "epoch": 0.6779518097687459, "grad_norm": 1.2337993383407593, "learning_rate": 4.3001811424466725e-05, "loss": 6.3543, "step": 16080 }, { "epoch": 0.6779939709509455, "grad_norm": 2.316312074661255, "learning_rate": 4.3001030638058656e-05, "loss": 7.0632, "step": 16081 }, { "epoch": 0.6780361321331451, "grad_norm": 3.5180208683013916, "learning_rate": 4.300024985165059e-05, "loss": 6.8878, "step": 16082 }, { "epoch": 0.6780782933153445, "grad_norm": 1.9879570007324219, "learning_rate": 4.299946906524251e-05, "loss": 6.4684, "step": 16083 }, { "epoch": 0.6781204544975441, "grad_norm": 1.6138131618499756, "learning_rate": 4.299868827883444e-05, "loss": 6.4017, "step": 16084 }, { "epoch": 0.6781626156797437, "grad_norm": 4.221836090087891, "learning_rate": 4.2997907492426374e-05, "loss": 8.4644, "step": 16085 }, { "epoch": 0.6782047768619432, "grad_norm": 2.0228028297424316, "learning_rate": 4.2997126706018306e-05, "loss": 7.0193, "step": 16086 }, { "epoch": 0.6782469380441427, "grad_norm": 1.560063123703003, "learning_rate": 4.299634591961023e-05, "loss": 6.2634, "step": 16087 }, { "epoch": 0.6782890992263423, "grad_norm": 2.1653637886047363, "learning_rate": 4.299556513320216e-05, "loss": 6.6522, "step": 16088 }, { "epoch": 0.6783312604085419, "grad_norm": 3.470832109451294, "learning_rate": 4.299478434679409e-05, "loss": 7.3837, "step": 16089 }, { "epoch": 0.6783734215907414, "grad_norm": 2.035569667816162, "learning_rate": 4.299400356038602e-05, "loss": 6.6105, "step": 16090 }, { "epoch": 0.678415582772941, "grad_norm": 2.328071355819702, "learning_rate": 4.2993222773977955e-05, "loss": 6.9237, "step": 16091 }, { "epoch": 0.6784577439551405, "grad_norm": 4.043030261993408, "learning_rate": 4.299244198756988e-05, "loss": 7.6319, "step": 16092 }, { "epoch": 0.67849990513734, "grad_norm": 1.881757378578186, "learning_rate": 4.299166120116182e-05, "loss": 7.0616, "step": 16093 }, { "epoch": 0.6785420663195396, "grad_norm": 1.4452943801879883, "learning_rate": 4.299088041475374e-05, "loss": 6.3412, "step": 16094 }, { "epoch": 0.6785842275017392, "grad_norm": 3.613877534866333, "learning_rate": 4.2990099628345673e-05, "loss": 7.4037, "step": 16095 }, { "epoch": 0.6786263886839387, "grad_norm": 2.9817254543304443, "learning_rate": 4.2989318841937605e-05, "loss": 7.3905, "step": 16096 }, { "epoch": 0.6786685498661382, "grad_norm": 1.4611198902130127, "learning_rate": 4.298853805552953e-05, "loss": 6.4159, "step": 16097 }, { "epoch": 0.6787107110483378, "grad_norm": 1.4968047142028809, "learning_rate": 4.298775726912146e-05, "loss": 6.3869, "step": 16098 }, { "epoch": 0.6787528722305374, "grad_norm": 1.3392033576965332, "learning_rate": 4.298697648271339e-05, "loss": 6.5614, "step": 16099 }, { "epoch": 0.678795033412737, "grad_norm": 4.180418491363525, "learning_rate": 4.298619569630532e-05, "loss": 6.3172, "step": 16100 }, { "epoch": 0.6788371945949364, "grad_norm": 1.8553603887557983, "learning_rate": 4.298541490989725e-05, "loss": 6.9661, "step": 16101 }, { "epoch": 0.678879355777136, "grad_norm": 2.72810435295105, "learning_rate": 4.298463412348918e-05, "loss": 6.7512, "step": 16102 }, { "epoch": 0.6789215169593356, "grad_norm": 1.7804793119430542, "learning_rate": 4.298385333708111e-05, "loss": 6.8996, "step": 16103 }, { "epoch": 0.6789636781415351, "grad_norm": 1.9523485898971558, "learning_rate": 4.298307255067304e-05, "loss": 6.3605, "step": 16104 }, { "epoch": 0.6790058393237346, "grad_norm": 2.2418482303619385, "learning_rate": 4.298229176426497e-05, "loss": 6.3846, "step": 16105 }, { "epoch": 0.6790480005059342, "grad_norm": 2.952068567276001, "learning_rate": 4.29815109778569e-05, "loss": 6.5059, "step": 16106 }, { "epoch": 0.6790901616881337, "grad_norm": 2.387432813644409, "learning_rate": 4.298073019144883e-05, "loss": 6.9117, "step": 16107 }, { "epoch": 0.6791323228703333, "grad_norm": 1.5478341579437256, "learning_rate": 4.297994940504076e-05, "loss": 6.3823, "step": 16108 }, { "epoch": 0.6791744840525328, "grad_norm": 2.945037841796875, "learning_rate": 4.2979168618632684e-05, "loss": 6.2284, "step": 16109 }, { "epoch": 0.6792166452347324, "grad_norm": 2.814147710800171, "learning_rate": 4.297838783222462e-05, "loss": 7.0183, "step": 16110 }, { "epoch": 0.6792588064169319, "grad_norm": 4.604671001434326, "learning_rate": 4.2977607045816547e-05, "loss": 7.7324, "step": 16111 }, { "epoch": 0.6793009675991315, "grad_norm": 2.1497395038604736, "learning_rate": 4.297682625940848e-05, "loss": 7.0311, "step": 16112 }, { "epoch": 0.6793431287813311, "grad_norm": 3.468919277191162, "learning_rate": 4.297604547300041e-05, "loss": 8.3666, "step": 16113 }, { "epoch": 0.6793852899635305, "grad_norm": 3.7013823986053467, "learning_rate": 4.297526468659234e-05, "loss": 7.4458, "step": 16114 }, { "epoch": 0.6794274511457301, "grad_norm": 5.428066730499268, "learning_rate": 4.297448390018427e-05, "loss": 6.9719, "step": 16115 }, { "epoch": 0.6794696123279297, "grad_norm": 3.261976957321167, "learning_rate": 4.2973703113776196e-05, "loss": 6.4696, "step": 16116 }, { "epoch": 0.6795117735101293, "grad_norm": 3.225158452987671, "learning_rate": 4.297292232736813e-05, "loss": 6.3023, "step": 16117 }, { "epoch": 0.6795539346923287, "grad_norm": 5.901594161987305, "learning_rate": 4.297214154096006e-05, "loss": 7.1701, "step": 16118 }, { "epoch": 0.6795960958745283, "grad_norm": 2.1398766040802, "learning_rate": 4.297136075455199e-05, "loss": 6.586, "step": 16119 }, { "epoch": 0.6796382570567279, "grad_norm": 3.0808498859405518, "learning_rate": 4.2970579968143914e-05, "loss": 6.9595, "step": 16120 }, { "epoch": 0.6796804182389274, "grad_norm": 5.680851936340332, "learning_rate": 4.2969799181735846e-05, "loss": 8.2976, "step": 16121 }, { "epoch": 0.679722579421127, "grad_norm": 2.4594366550445557, "learning_rate": 4.296901839532778e-05, "loss": 6.822, "step": 16122 }, { "epoch": 0.6797647406033265, "grad_norm": 1.9951773881912231, "learning_rate": 4.29682376089197e-05, "loss": 6.6452, "step": 16123 }, { "epoch": 0.679806901785526, "grad_norm": 1.4660987854003906, "learning_rate": 4.296745682251164e-05, "loss": 6.9486, "step": 16124 }, { "epoch": 0.6798490629677256, "grad_norm": 2.0293521881103516, "learning_rate": 4.2966676036103564e-05, "loss": 6.9255, "step": 16125 }, { "epoch": 0.6798912241499252, "grad_norm": 2.8192851543426514, "learning_rate": 4.2965895249695495e-05, "loss": 6.784, "step": 16126 }, { "epoch": 0.6799333853321247, "grad_norm": 3.174961566925049, "learning_rate": 4.2965114463287426e-05, "loss": 6.3543, "step": 16127 }, { "epoch": 0.6799755465143242, "grad_norm": 2.9837806224823, "learning_rate": 4.296433367687935e-05, "loss": 7.9272, "step": 16128 }, { "epoch": 0.6800177076965238, "grad_norm": 2.677187204360962, "learning_rate": 4.296355289047129e-05, "loss": 7.1965, "step": 16129 }, { "epoch": 0.6800598688787234, "grad_norm": 1.843481421470642, "learning_rate": 4.296277210406321e-05, "loss": 7.1237, "step": 16130 }, { "epoch": 0.680102030060923, "grad_norm": 1.4794455766677856, "learning_rate": 4.2961991317655145e-05, "loss": 6.5631, "step": 16131 }, { "epoch": 0.6801441912431224, "grad_norm": 1.6033152341842651, "learning_rate": 4.2961210531247076e-05, "loss": 6.9055, "step": 16132 }, { "epoch": 0.680186352425322, "grad_norm": 1.6127352714538574, "learning_rate": 4.2960429744839e-05, "loss": 6.4862, "step": 16133 }, { "epoch": 0.6802285136075216, "grad_norm": 1.8543592691421509, "learning_rate": 4.295964895843093e-05, "loss": 6.5009, "step": 16134 }, { "epoch": 0.6802706747897211, "grad_norm": 1.3134981393814087, "learning_rate": 4.295886817202286e-05, "loss": 6.5291, "step": 16135 }, { "epoch": 0.6803128359719206, "grad_norm": 4.21884298324585, "learning_rate": 4.2958087385614794e-05, "loss": 7.214, "step": 16136 }, { "epoch": 0.6803549971541202, "grad_norm": 1.396640658378601, "learning_rate": 4.295730659920672e-05, "loss": 6.6917, "step": 16137 }, { "epoch": 0.6803971583363198, "grad_norm": 2.3562240600585938, "learning_rate": 4.295652581279866e-05, "loss": 6.8007, "step": 16138 }, { "epoch": 0.6804393195185193, "grad_norm": 4.452334880828857, "learning_rate": 4.295574502639058e-05, "loss": 7.6756, "step": 16139 }, { "epoch": 0.6804814807007189, "grad_norm": 2.255464792251587, "learning_rate": 4.295496423998251e-05, "loss": 6.9212, "step": 16140 }, { "epoch": 0.6805236418829184, "grad_norm": 1.9532002210617065, "learning_rate": 4.2954183453574444e-05, "loss": 6.7696, "step": 16141 }, { "epoch": 0.6805658030651179, "grad_norm": 2.601491928100586, "learning_rate": 4.295340266716637e-05, "loss": 6.5268, "step": 16142 }, { "epoch": 0.6806079642473175, "grad_norm": 2.8711118698120117, "learning_rate": 4.2952621880758306e-05, "loss": 7.2374, "step": 16143 }, { "epoch": 0.6806501254295171, "grad_norm": 5.162904262542725, "learning_rate": 4.295184109435023e-05, "loss": 10.346, "step": 16144 }, { "epoch": 0.6806922866117165, "grad_norm": 1.7859512567520142, "learning_rate": 4.295106030794216e-05, "loss": 6.3551, "step": 16145 }, { "epoch": 0.6807344477939161, "grad_norm": 2.2777202129364014, "learning_rate": 4.295027952153409e-05, "loss": 6.818, "step": 16146 }, { "epoch": 0.6807766089761157, "grad_norm": 3.4728119373321533, "learning_rate": 4.294949873512602e-05, "loss": 6.3988, "step": 16147 }, { "epoch": 0.6808187701583153, "grad_norm": 2.6698737144470215, "learning_rate": 4.294871794871795e-05, "loss": 7.1335, "step": 16148 }, { "epoch": 0.6808609313405148, "grad_norm": 1.765748381614685, "learning_rate": 4.294793716230988e-05, "loss": 6.4932, "step": 16149 }, { "epoch": 0.6809030925227143, "grad_norm": 2.2606277465820312, "learning_rate": 4.294715637590181e-05, "loss": 6.3407, "step": 16150 }, { "epoch": 0.6809452537049139, "grad_norm": 2.190721035003662, "learning_rate": 4.294637558949374e-05, "loss": 6.4259, "step": 16151 }, { "epoch": 0.6809874148871135, "grad_norm": 4.7068634033203125, "learning_rate": 4.294559480308567e-05, "loss": 7.2244, "step": 16152 }, { "epoch": 0.681029576069313, "grad_norm": 2.737238645553589, "learning_rate": 4.29448140166776e-05, "loss": 7.2296, "step": 16153 }, { "epoch": 0.6810717372515125, "grad_norm": 2.694101572036743, "learning_rate": 4.294403323026953e-05, "loss": 6.9592, "step": 16154 }, { "epoch": 0.6811138984337121, "grad_norm": 2.7030696868896484, "learning_rate": 4.294325244386146e-05, "loss": 7.1147, "step": 16155 }, { "epoch": 0.6811560596159116, "grad_norm": 1.3954476118087769, "learning_rate": 4.2942471657453385e-05, "loss": 6.4417, "step": 16156 }, { "epoch": 0.6811982207981112, "grad_norm": 1.490713119506836, "learning_rate": 4.2941690871045324e-05, "loss": 6.7785, "step": 16157 }, { "epoch": 0.6812403819803108, "grad_norm": 2.397704601287842, "learning_rate": 4.294091008463725e-05, "loss": 6.4292, "step": 16158 }, { "epoch": 0.6812825431625102, "grad_norm": 2.5146853923797607, "learning_rate": 4.294012929822917e-05, "loss": 6.537, "step": 16159 }, { "epoch": 0.6813247043447098, "grad_norm": 2.38456392288208, "learning_rate": 4.293934851182111e-05, "loss": 6.4174, "step": 16160 }, { "epoch": 0.6813668655269094, "grad_norm": 2.859907865524292, "learning_rate": 4.2938567725413035e-05, "loss": 7.4404, "step": 16161 }, { "epoch": 0.681409026709109, "grad_norm": 1.910681128501892, "learning_rate": 4.293778693900497e-05, "loss": 6.8271, "step": 16162 }, { "epoch": 0.6814511878913084, "grad_norm": 3.175729513168335, "learning_rate": 4.29370061525969e-05, "loss": 7.0472, "step": 16163 }, { "epoch": 0.681493349073508, "grad_norm": 3.018700361251831, "learning_rate": 4.293622536618883e-05, "loss": 7.6722, "step": 16164 }, { "epoch": 0.6815355102557076, "grad_norm": 1.9471173286437988, "learning_rate": 4.293544457978076e-05, "loss": 6.9376, "step": 16165 }, { "epoch": 0.6815776714379072, "grad_norm": 1.5736297369003296, "learning_rate": 4.2934663793372685e-05, "loss": 7.0175, "step": 16166 }, { "epoch": 0.6816198326201066, "grad_norm": 1.878893256187439, "learning_rate": 4.2933883006964616e-05, "loss": 7.0431, "step": 16167 }, { "epoch": 0.6816619938023062, "grad_norm": 3.7129688262939453, "learning_rate": 4.293310222055655e-05, "loss": 7.0469, "step": 16168 }, { "epoch": 0.6817041549845058, "grad_norm": 3.5623114109039307, "learning_rate": 4.293232143414848e-05, "loss": 6.835, "step": 16169 }, { "epoch": 0.6817463161667053, "grad_norm": 2.8999483585357666, "learning_rate": 4.29315406477404e-05, "loss": 6.8888, "step": 16170 }, { "epoch": 0.6817884773489049, "grad_norm": 2.354902982711792, "learning_rate": 4.2930759861332334e-05, "loss": 6.7809, "step": 16171 }, { "epoch": 0.6818306385311044, "grad_norm": 1.709458827972412, "learning_rate": 4.2929979074924265e-05, "loss": 6.5914, "step": 16172 }, { "epoch": 0.681872799713304, "grad_norm": 2.3902082443237305, "learning_rate": 4.2929198288516197e-05, "loss": 6.6551, "step": 16173 }, { "epoch": 0.6819149608955035, "grad_norm": 2.7276546955108643, "learning_rate": 4.292841750210813e-05, "loss": 6.5442, "step": 16174 }, { "epoch": 0.6819571220777031, "grad_norm": 2.944406032562256, "learning_rate": 4.292763671570005e-05, "loss": 7.1955, "step": 16175 }, { "epoch": 0.6819992832599026, "grad_norm": 2.460294008255005, "learning_rate": 4.292685592929199e-05, "loss": 6.6658, "step": 16176 }, { "epoch": 0.6820414444421021, "grad_norm": 4.26214075088501, "learning_rate": 4.2926075142883915e-05, "loss": 7.4771, "step": 16177 }, { "epoch": 0.6820836056243017, "grad_norm": 2.1053197383880615, "learning_rate": 4.292529435647584e-05, "loss": 6.6914, "step": 16178 }, { "epoch": 0.6821257668065013, "grad_norm": 3.2470574378967285, "learning_rate": 4.292451357006778e-05, "loss": 7.6999, "step": 16179 }, { "epoch": 0.6821679279887009, "grad_norm": 2.6656494140625, "learning_rate": 4.29237327836597e-05, "loss": 6.3118, "step": 16180 }, { "epoch": 0.6822100891709003, "grad_norm": 2.063568353652954, "learning_rate": 4.292295199725163e-05, "loss": 6.4043, "step": 16181 }, { "epoch": 0.6822522503530999, "grad_norm": 1.602023959159851, "learning_rate": 4.2922171210843564e-05, "loss": 6.7929, "step": 16182 }, { "epoch": 0.6822944115352995, "grad_norm": 3.6612563133239746, "learning_rate": 4.2921390424435496e-05, "loss": 7.5266, "step": 16183 }, { "epoch": 0.682336572717499, "grad_norm": 2.8645033836364746, "learning_rate": 4.292060963802743e-05, "loss": 6.4546, "step": 16184 }, { "epoch": 0.6823787338996985, "grad_norm": 1.926192283630371, "learning_rate": 4.291982885161935e-05, "loss": 6.5172, "step": 16185 }, { "epoch": 0.6824208950818981, "grad_norm": 1.8290596008300781, "learning_rate": 4.291904806521128e-05, "loss": 6.3147, "step": 16186 }, { "epoch": 0.6824630562640976, "grad_norm": 3.4512784481048584, "learning_rate": 4.2918267278803214e-05, "loss": 7.379, "step": 16187 }, { "epoch": 0.6825052174462972, "grad_norm": 1.3941410779953003, "learning_rate": 4.2917486492395145e-05, "loss": 6.4691, "step": 16188 }, { "epoch": 0.6825473786284968, "grad_norm": 2.6298604011535645, "learning_rate": 4.291670570598707e-05, "loss": 7.3362, "step": 16189 }, { "epoch": 0.6825895398106963, "grad_norm": 1.7582916021347046, "learning_rate": 4.2915924919579e-05, "loss": 6.4798, "step": 16190 }, { "epoch": 0.6826317009928958, "grad_norm": 4.330686569213867, "learning_rate": 4.291514413317093e-05, "loss": 8.1065, "step": 16191 }, { "epoch": 0.6826738621750954, "grad_norm": 3.753995418548584, "learning_rate": 4.291436334676286e-05, "loss": 7.6334, "step": 16192 }, { "epoch": 0.682716023357295, "grad_norm": 2.1170246601104736, "learning_rate": 4.2913582560354795e-05, "loss": 6.4441, "step": 16193 }, { "epoch": 0.6827581845394944, "grad_norm": 1.825900673866272, "learning_rate": 4.291280177394672e-05, "loss": 6.4443, "step": 16194 }, { "epoch": 0.682800345721694, "grad_norm": 2.2503020763397217, "learning_rate": 4.291202098753865e-05, "loss": 6.8712, "step": 16195 }, { "epoch": 0.6828425069038936, "grad_norm": 1.421341061592102, "learning_rate": 4.291124020113058e-05, "loss": 6.3347, "step": 16196 }, { "epoch": 0.6828846680860932, "grad_norm": 3.659123420715332, "learning_rate": 4.2910459414722506e-05, "loss": 7.7209, "step": 16197 }, { "epoch": 0.6829268292682927, "grad_norm": 2.0967864990234375, "learning_rate": 4.2909678628314444e-05, "loss": 6.7283, "step": 16198 }, { "epoch": 0.6829689904504922, "grad_norm": 1.6069042682647705, "learning_rate": 4.290889784190637e-05, "loss": 6.4385, "step": 16199 }, { "epoch": 0.6830111516326918, "grad_norm": 1.4941428899765015, "learning_rate": 4.29081170554983e-05, "loss": 6.5853, "step": 16200 }, { "epoch": 0.6830533128148913, "grad_norm": 2.8839221000671387, "learning_rate": 4.290733626909023e-05, "loss": 6.8795, "step": 16201 }, { "epoch": 0.6830954739970909, "grad_norm": 1.665127158164978, "learning_rate": 4.290655548268216e-05, "loss": 6.7162, "step": 16202 }, { "epoch": 0.6831376351792904, "grad_norm": 2.6773571968078613, "learning_rate": 4.290577469627409e-05, "loss": 6.8241, "step": 16203 }, { "epoch": 0.68317979636149, "grad_norm": 1.6944077014923096, "learning_rate": 4.290499390986602e-05, "loss": 6.8155, "step": 16204 }, { "epoch": 0.6832219575436895, "grad_norm": 2.334541082382202, "learning_rate": 4.290421312345795e-05, "loss": 6.8224, "step": 16205 }, { "epoch": 0.6832641187258891, "grad_norm": 2.1133193969726562, "learning_rate": 4.290343233704988e-05, "loss": 7.0083, "step": 16206 }, { "epoch": 0.6833062799080887, "grad_norm": 2.176677703857422, "learning_rate": 4.290265155064181e-05, "loss": 7.1235, "step": 16207 }, { "epoch": 0.6833484410902881, "grad_norm": 3.7567765712738037, "learning_rate": 4.2901870764233736e-05, "loss": 7.6778, "step": 16208 }, { "epoch": 0.6833906022724877, "grad_norm": 2.5567026138305664, "learning_rate": 4.290108997782567e-05, "loss": 7.2058, "step": 16209 }, { "epoch": 0.6834327634546873, "grad_norm": 1.880383014678955, "learning_rate": 4.29003091914176e-05, "loss": 6.8707, "step": 16210 }, { "epoch": 0.6834749246368869, "grad_norm": 2.609956979751587, "learning_rate": 4.2899528405009523e-05, "loss": 6.297, "step": 16211 }, { "epoch": 0.6835170858190863, "grad_norm": 3.410520076751709, "learning_rate": 4.289874761860146e-05, "loss": 7.27, "step": 16212 }, { "epoch": 0.6835592470012859, "grad_norm": 2.3037819862365723, "learning_rate": 4.2897966832193386e-05, "loss": 6.7356, "step": 16213 }, { "epoch": 0.6836014081834855, "grad_norm": 3.2001094818115234, "learning_rate": 4.289718604578532e-05, "loss": 6.6917, "step": 16214 }, { "epoch": 0.683643569365685, "grad_norm": 1.8411836624145508, "learning_rate": 4.289640525937725e-05, "loss": 6.7466, "step": 16215 }, { "epoch": 0.6836857305478846, "grad_norm": 1.5562076568603516, "learning_rate": 4.289562447296917e-05, "loss": 6.5029, "step": 16216 }, { "epoch": 0.6837278917300841, "grad_norm": 1.7283517122268677, "learning_rate": 4.2894843686561104e-05, "loss": 6.4186, "step": 16217 }, { "epoch": 0.6837700529122837, "grad_norm": 3.586421489715576, "learning_rate": 4.2894062900153035e-05, "loss": 7.695, "step": 16218 }, { "epoch": 0.6838122140944832, "grad_norm": 2.028454303741455, "learning_rate": 4.289328211374497e-05, "loss": 6.3627, "step": 16219 }, { "epoch": 0.6838543752766828, "grad_norm": 1.4749778509140015, "learning_rate": 4.28925013273369e-05, "loss": 6.6922, "step": 16220 }, { "epoch": 0.6838965364588823, "grad_norm": 2.0112626552581787, "learning_rate": 4.289172054092882e-05, "loss": 6.6965, "step": 16221 }, { "epoch": 0.6839386976410818, "grad_norm": 2.667065382003784, "learning_rate": 4.2890939754520754e-05, "loss": 7.0792, "step": 16222 }, { "epoch": 0.6839808588232814, "grad_norm": 2.4205856323242188, "learning_rate": 4.2890158968112685e-05, "loss": 7.125, "step": 16223 }, { "epoch": 0.684023020005481, "grad_norm": 2.0866734981536865, "learning_rate": 4.2889378181704616e-05, "loss": 6.5837, "step": 16224 }, { "epoch": 0.6840651811876804, "grad_norm": 1.7248915433883667, "learning_rate": 4.288859739529654e-05, "loss": 6.4432, "step": 16225 }, { "epoch": 0.68410734236988, "grad_norm": 1.7785431146621704, "learning_rate": 4.288781660888848e-05, "loss": 6.6303, "step": 16226 }, { "epoch": 0.6841495035520796, "grad_norm": 1.6893094778060913, "learning_rate": 4.28870358224804e-05, "loss": 6.5439, "step": 16227 }, { "epoch": 0.6841916647342792, "grad_norm": 2.390064001083374, "learning_rate": 4.2886255036072335e-05, "loss": 6.7861, "step": 16228 }, { "epoch": 0.6842338259164787, "grad_norm": 3.2984778881073, "learning_rate": 4.2885474249664266e-05, "loss": 7.0318, "step": 16229 }, { "epoch": 0.6842759870986782, "grad_norm": 1.9257701635360718, "learning_rate": 4.288469346325619e-05, "loss": 6.67, "step": 16230 }, { "epoch": 0.6843181482808778, "grad_norm": 1.327471137046814, "learning_rate": 4.288391267684813e-05, "loss": 6.6545, "step": 16231 }, { "epoch": 0.6843603094630774, "grad_norm": 1.4194039106369019, "learning_rate": 4.288313189044005e-05, "loss": 6.3422, "step": 16232 }, { "epoch": 0.6844024706452769, "grad_norm": 1.8627870082855225, "learning_rate": 4.2882351104031984e-05, "loss": 6.976, "step": 16233 }, { "epoch": 0.6844446318274764, "grad_norm": 1.2995675802230835, "learning_rate": 4.2881570317623915e-05, "loss": 6.9587, "step": 16234 }, { "epoch": 0.684486793009676, "grad_norm": 1.8233965635299683, "learning_rate": 4.288078953121584e-05, "loss": 6.6893, "step": 16235 }, { "epoch": 0.6845289541918755, "grad_norm": 1.6167283058166504, "learning_rate": 4.288000874480777e-05, "loss": 6.5173, "step": 16236 }, { "epoch": 0.6845711153740751, "grad_norm": 1.0616307258605957, "learning_rate": 4.28792279583997e-05, "loss": 6.39, "step": 16237 }, { "epoch": 0.6846132765562747, "grad_norm": 3.1601741313934326, "learning_rate": 4.2878447171991634e-05, "loss": 7.6417, "step": 16238 }, { "epoch": 0.6846554377384741, "grad_norm": 1.6842596530914307, "learning_rate": 4.287766638558356e-05, "loss": 6.6644, "step": 16239 }, { "epoch": 0.6846975989206737, "grad_norm": 3.115532159805298, "learning_rate": 4.287688559917549e-05, "loss": 6.6213, "step": 16240 }, { "epoch": 0.6847397601028733, "grad_norm": 2.2176945209503174, "learning_rate": 4.287610481276742e-05, "loss": 6.7644, "step": 16241 }, { "epoch": 0.6847819212850729, "grad_norm": 1.8375743627548218, "learning_rate": 4.287532402635935e-05, "loss": 6.5685, "step": 16242 }, { "epoch": 0.6848240824672723, "grad_norm": 1.5279200077056885, "learning_rate": 4.287454323995128e-05, "loss": 7.0218, "step": 16243 }, { "epoch": 0.6848662436494719, "grad_norm": 2.8581013679504395, "learning_rate": 4.287376245354321e-05, "loss": 7.4532, "step": 16244 }, { "epoch": 0.6849084048316715, "grad_norm": 2.193990707397461, "learning_rate": 4.2872981667135146e-05, "loss": 6.6699, "step": 16245 }, { "epoch": 0.684950566013871, "grad_norm": 1.366359829902649, "learning_rate": 4.287220088072707e-05, "loss": 6.5292, "step": 16246 }, { "epoch": 0.6849927271960706, "grad_norm": 3.0225017070770264, "learning_rate": 4.2871420094319e-05, "loss": 7.1044, "step": 16247 }, { "epoch": 0.6850348883782701, "grad_norm": 1.731926679611206, "learning_rate": 4.287063930791093e-05, "loss": 6.5014, "step": 16248 }, { "epoch": 0.6850770495604697, "grad_norm": 2.1393136978149414, "learning_rate": 4.286985852150286e-05, "loss": 7.1829, "step": 16249 }, { "epoch": 0.6851192107426692, "grad_norm": 1.141806960105896, "learning_rate": 4.286907773509479e-05, "loss": 6.5805, "step": 16250 }, { "epoch": 0.6851613719248688, "grad_norm": 1.5857102870941162, "learning_rate": 4.286829694868672e-05, "loss": 6.4026, "step": 16251 }, { "epoch": 0.6852035331070683, "grad_norm": 2.380074977874756, "learning_rate": 4.286751616227865e-05, "loss": 6.392, "step": 16252 }, { "epoch": 0.6852456942892678, "grad_norm": 1.622052788734436, "learning_rate": 4.286673537587058e-05, "loss": 6.3804, "step": 16253 }, { "epoch": 0.6852878554714674, "grad_norm": 2.1618807315826416, "learning_rate": 4.286595458946251e-05, "loss": 7.1403, "step": 16254 }, { "epoch": 0.685330016653667, "grad_norm": 0.9738902449607849, "learning_rate": 4.286517380305444e-05, "loss": 6.4715, "step": 16255 }, { "epoch": 0.6853721778358666, "grad_norm": 1.0781604051589966, "learning_rate": 4.286439301664637e-05, "loss": 6.4454, "step": 16256 }, { "epoch": 0.685414339018066, "grad_norm": 1.3829312324523926, "learning_rate": 4.28636122302383e-05, "loss": 6.8678, "step": 16257 }, { "epoch": 0.6854565002002656, "grad_norm": 2.2849466800689697, "learning_rate": 4.2862831443830225e-05, "loss": 6.9191, "step": 16258 }, { "epoch": 0.6854986613824652, "grad_norm": 3.0020947456359863, "learning_rate": 4.2862050657422156e-05, "loss": 7.7829, "step": 16259 }, { "epoch": 0.6855408225646648, "grad_norm": 1.8570246696472168, "learning_rate": 4.286126987101409e-05, "loss": 6.364, "step": 16260 }, { "epoch": 0.6855829837468642, "grad_norm": 1.3815234899520874, "learning_rate": 4.286048908460601e-05, "loss": 6.6826, "step": 16261 }, { "epoch": 0.6856251449290638, "grad_norm": 1.3265687227249146, "learning_rate": 4.285970829819795e-05, "loss": 6.6521, "step": 16262 }, { "epoch": 0.6856673061112634, "grad_norm": 3.192777156829834, "learning_rate": 4.2858927511789874e-05, "loss": 7.745, "step": 16263 }, { "epoch": 0.6857094672934629, "grad_norm": 3.754852056503296, "learning_rate": 4.285814672538181e-05, "loss": 8.0759, "step": 16264 }, { "epoch": 0.6857516284756625, "grad_norm": 2.996232509613037, "learning_rate": 4.285736593897374e-05, "loss": 7.3139, "step": 16265 }, { "epoch": 0.685793789657862, "grad_norm": 2.439659357070923, "learning_rate": 4.285658515256566e-05, "loss": 6.5043, "step": 16266 }, { "epoch": 0.6858359508400615, "grad_norm": 1.6194210052490234, "learning_rate": 4.28558043661576e-05, "loss": 7.1685, "step": 16267 }, { "epoch": 0.6858781120222611, "grad_norm": 1.828212022781372, "learning_rate": 4.2855023579749524e-05, "loss": 6.8367, "step": 16268 }, { "epoch": 0.6859202732044607, "grad_norm": 2.8371872901916504, "learning_rate": 4.2854242793341455e-05, "loss": 7.7485, "step": 16269 }, { "epoch": 0.6859624343866602, "grad_norm": 2.123518705368042, "learning_rate": 4.2853462006933386e-05, "loss": 6.4601, "step": 16270 }, { "epoch": 0.6860045955688597, "grad_norm": 2.208359479904175, "learning_rate": 4.285268122052532e-05, "loss": 7.7442, "step": 16271 }, { "epoch": 0.6860467567510593, "grad_norm": 2.825126886367798, "learning_rate": 4.285190043411724e-05, "loss": 6.258, "step": 16272 }, { "epoch": 0.6860889179332589, "grad_norm": 1.7605822086334229, "learning_rate": 4.2851119647709173e-05, "loss": 7.1917, "step": 16273 }, { "epoch": 0.6861310791154583, "grad_norm": 1.4302185773849487, "learning_rate": 4.2850338861301105e-05, "loss": 6.8487, "step": 16274 }, { "epoch": 0.6861732402976579, "grad_norm": 1.875583291053772, "learning_rate": 4.2849558074893036e-05, "loss": 6.5452, "step": 16275 }, { "epoch": 0.6862154014798575, "grad_norm": 1.4226055145263672, "learning_rate": 4.284877728848497e-05, "loss": 6.4509, "step": 16276 }, { "epoch": 0.6862575626620571, "grad_norm": 2.4832375049591064, "learning_rate": 4.284799650207689e-05, "loss": 6.9943, "step": 16277 }, { "epoch": 0.6862997238442566, "grad_norm": 4.763381481170654, "learning_rate": 4.284721571566882e-05, "loss": 9.2714, "step": 16278 }, { "epoch": 0.6863418850264561, "grad_norm": 3.682652711868286, "learning_rate": 4.2846434929260754e-05, "loss": 7.7692, "step": 16279 }, { "epoch": 0.6863840462086557, "grad_norm": 3.0533323287963867, "learning_rate": 4.284565414285268e-05, "loss": 8.0798, "step": 16280 }, { "epoch": 0.6864262073908552, "grad_norm": 5.5742268562316895, "learning_rate": 4.284487335644462e-05, "loss": 7.9083, "step": 16281 }, { "epoch": 0.6864683685730548, "grad_norm": 3.038560152053833, "learning_rate": 4.284409257003654e-05, "loss": 7.3475, "step": 16282 }, { "epoch": 0.6865105297552543, "grad_norm": 3.3748528957366943, "learning_rate": 4.284331178362847e-05, "loss": 6.6014, "step": 16283 }, { "epoch": 0.6865526909374539, "grad_norm": 3.357238292694092, "learning_rate": 4.2842530997220404e-05, "loss": 7.1902, "step": 16284 }, { "epoch": 0.6865948521196534, "grad_norm": 3.670795440673828, "learning_rate": 4.284175021081233e-05, "loss": 7.5322, "step": 16285 }, { "epoch": 0.686637013301853, "grad_norm": 5.027807235717773, "learning_rate": 4.2840969424404266e-05, "loss": 8.3239, "step": 16286 }, { "epoch": 0.6866791744840526, "grad_norm": 2.3010663986206055, "learning_rate": 4.284018863799619e-05, "loss": 7.0776, "step": 16287 }, { "epoch": 0.686721335666252, "grad_norm": 1.7634098529815674, "learning_rate": 4.283940785158812e-05, "loss": 6.8896, "step": 16288 }, { "epoch": 0.6867634968484516, "grad_norm": 2.0404715538024902, "learning_rate": 4.283862706518005e-05, "loss": 7.0347, "step": 16289 }, { "epoch": 0.6868056580306512, "grad_norm": 2.244658946990967, "learning_rate": 4.2837846278771985e-05, "loss": 6.4696, "step": 16290 }, { "epoch": 0.6868478192128508, "grad_norm": 1.9864373207092285, "learning_rate": 4.283706549236391e-05, "loss": 7.0675, "step": 16291 }, { "epoch": 0.6868899803950502, "grad_norm": 3.512021064758301, "learning_rate": 4.283628470595584e-05, "loss": 6.6275, "step": 16292 }, { "epoch": 0.6869321415772498, "grad_norm": 2.845564126968384, "learning_rate": 4.283550391954777e-05, "loss": 6.6137, "step": 16293 }, { "epoch": 0.6869743027594494, "grad_norm": 1.5139542818069458, "learning_rate": 4.2834723133139696e-05, "loss": 6.886, "step": 16294 }, { "epoch": 0.687016463941649, "grad_norm": 4.332258701324463, "learning_rate": 4.2833942346731634e-05, "loss": 8.5903, "step": 16295 }, { "epoch": 0.6870586251238485, "grad_norm": 2.4043264389038086, "learning_rate": 4.283316156032356e-05, "loss": 6.7731, "step": 16296 }, { "epoch": 0.687100786306048, "grad_norm": 3.9922266006469727, "learning_rate": 4.283238077391549e-05, "loss": 7.4661, "step": 16297 }, { "epoch": 0.6871429474882476, "grad_norm": 1.6084994077682495, "learning_rate": 4.283159998750742e-05, "loss": 6.5532, "step": 16298 }, { "epoch": 0.6871851086704471, "grad_norm": 2.3865625858306885, "learning_rate": 4.2830819201099346e-05, "loss": 7.0062, "step": 16299 }, { "epoch": 0.6872272698526467, "grad_norm": 4.0642194747924805, "learning_rate": 4.2830038414691284e-05, "loss": 7.7732, "step": 16300 }, { "epoch": 0.6872694310348462, "grad_norm": 2.6619536876678467, "learning_rate": 4.282925762828321e-05, "loss": 7.0846, "step": 16301 }, { "epoch": 0.6873115922170457, "grad_norm": 1.685308814048767, "learning_rate": 4.282847684187514e-05, "loss": 6.8859, "step": 16302 }, { "epoch": 0.6873537533992453, "grad_norm": 1.6797778606414795, "learning_rate": 4.282769605546707e-05, "loss": 7.3322, "step": 16303 }, { "epoch": 0.6873959145814449, "grad_norm": 4.165064334869385, "learning_rate": 4.2826915269058995e-05, "loss": 6.8815, "step": 16304 }, { "epoch": 0.6874380757636445, "grad_norm": 3.524029016494751, "learning_rate": 4.2826134482650926e-05, "loss": 7.4261, "step": 16305 }, { "epoch": 0.6874802369458439, "grad_norm": 4.034078598022461, "learning_rate": 4.282535369624286e-05, "loss": 9.0906, "step": 16306 }, { "epoch": 0.6875223981280435, "grad_norm": 4.045433044433594, "learning_rate": 4.282457290983479e-05, "loss": 7.5882, "step": 16307 }, { "epoch": 0.6875645593102431, "grad_norm": 2.012985944747925, "learning_rate": 4.282379212342671e-05, "loss": 6.7008, "step": 16308 }, { "epoch": 0.6876067204924426, "grad_norm": 2.0692713260650635, "learning_rate": 4.282301133701865e-05, "loss": 6.8911, "step": 16309 }, { "epoch": 0.6876488816746421, "grad_norm": 2.381091356277466, "learning_rate": 4.2822230550610576e-05, "loss": 7.1206, "step": 16310 }, { "epoch": 0.6876910428568417, "grad_norm": 2.487621307373047, "learning_rate": 4.282144976420251e-05, "loss": 7.2923, "step": 16311 }, { "epoch": 0.6877332040390413, "grad_norm": 3.2907872200012207, "learning_rate": 4.282066897779444e-05, "loss": 6.3845, "step": 16312 }, { "epoch": 0.6877753652212408, "grad_norm": 2.4260082244873047, "learning_rate": 4.281988819138636e-05, "loss": 6.6813, "step": 16313 }, { "epoch": 0.6878175264034404, "grad_norm": 1.7194161415100098, "learning_rate": 4.28191074049783e-05, "loss": 6.7893, "step": 16314 }, { "epoch": 0.6878596875856399, "grad_norm": 1.6359962224960327, "learning_rate": 4.2818326618570225e-05, "loss": 6.8943, "step": 16315 }, { "epoch": 0.6879018487678394, "grad_norm": 1.3188607692718506, "learning_rate": 4.281754583216216e-05, "loss": 7.3667, "step": 16316 }, { "epoch": 0.687944009950039, "grad_norm": 1.3114382028579712, "learning_rate": 4.281676504575409e-05, "loss": 6.5504, "step": 16317 }, { "epoch": 0.6879861711322386, "grad_norm": 1.4145400524139404, "learning_rate": 4.281598425934601e-05, "loss": 6.6333, "step": 16318 }, { "epoch": 0.688028332314438, "grad_norm": 1.7768851518630981, "learning_rate": 4.2815203472937944e-05, "loss": 6.5116, "step": 16319 }, { "epoch": 0.6880704934966376, "grad_norm": 1.5698490142822266, "learning_rate": 4.2814422686529875e-05, "loss": 6.7235, "step": 16320 }, { "epoch": 0.6881126546788372, "grad_norm": 1.5579451322555542, "learning_rate": 4.2813641900121806e-05, "loss": 7.1705, "step": 16321 }, { "epoch": 0.6881548158610368, "grad_norm": 1.614162564277649, "learning_rate": 4.281286111371374e-05, "loss": 6.3826, "step": 16322 }, { "epoch": 0.6881969770432363, "grad_norm": 1.7830209732055664, "learning_rate": 4.281208032730566e-05, "loss": 6.5743, "step": 16323 }, { "epoch": 0.6882391382254358, "grad_norm": 3.131275177001953, "learning_rate": 4.281129954089759e-05, "loss": 7.5506, "step": 16324 }, { "epoch": 0.6882812994076354, "grad_norm": 1.4735336303710938, "learning_rate": 4.2810518754489524e-05, "loss": 7.0711, "step": 16325 }, { "epoch": 0.688323460589835, "grad_norm": 1.3587617874145508, "learning_rate": 4.2809737968081456e-05, "loss": 6.8299, "step": 16326 }, { "epoch": 0.6883656217720345, "grad_norm": 2.8858871459960938, "learning_rate": 4.280895718167338e-05, "loss": 6.5695, "step": 16327 }, { "epoch": 0.688407782954234, "grad_norm": 2.0214321613311768, "learning_rate": 4.280817639526531e-05, "loss": 6.9279, "step": 16328 }, { "epoch": 0.6884499441364336, "grad_norm": 1.317772388458252, "learning_rate": 4.280739560885724e-05, "loss": 6.5706, "step": 16329 }, { "epoch": 0.6884921053186331, "grad_norm": 4.113420486450195, "learning_rate": 4.280661482244917e-05, "loss": 7.4689, "step": 16330 }, { "epoch": 0.6885342665008327, "grad_norm": 2.0668742656707764, "learning_rate": 4.2805834036041105e-05, "loss": 7.792, "step": 16331 }, { "epoch": 0.6885764276830322, "grad_norm": 3.816666603088379, "learning_rate": 4.280505324963303e-05, "loss": 8.585, "step": 16332 }, { "epoch": 0.6886185888652318, "grad_norm": 2.123147964477539, "learning_rate": 4.280427246322497e-05, "loss": 6.4638, "step": 16333 }, { "epoch": 0.6886607500474313, "grad_norm": 1.2188353538513184, "learning_rate": 4.280349167681689e-05, "loss": 7.047, "step": 16334 }, { "epoch": 0.6887029112296309, "grad_norm": 1.414459466934204, "learning_rate": 4.2802710890408823e-05, "loss": 6.2866, "step": 16335 }, { "epoch": 0.6887450724118305, "grad_norm": 2.8876757621765137, "learning_rate": 4.2801930104000755e-05, "loss": 7.2738, "step": 16336 }, { "epoch": 0.6887872335940299, "grad_norm": 3.799020528793335, "learning_rate": 4.280114931759268e-05, "loss": 7.619, "step": 16337 }, { "epoch": 0.6888293947762295, "grad_norm": 2.3961548805236816, "learning_rate": 4.280036853118461e-05, "loss": 7.1919, "step": 16338 }, { "epoch": 0.6888715559584291, "grad_norm": 2.986945152282715, "learning_rate": 4.279958774477654e-05, "loss": 6.6444, "step": 16339 }, { "epoch": 0.6889137171406287, "grad_norm": 3.0242233276367188, "learning_rate": 4.279880695836847e-05, "loss": 6.5055, "step": 16340 }, { "epoch": 0.6889558783228281, "grad_norm": 3.276179075241089, "learning_rate": 4.27980261719604e-05, "loss": 6.392, "step": 16341 }, { "epoch": 0.6889980395050277, "grad_norm": 3.416727066040039, "learning_rate": 4.279724538555233e-05, "loss": 6.769, "step": 16342 }, { "epoch": 0.6890402006872273, "grad_norm": 2.388998031616211, "learning_rate": 4.279646459914426e-05, "loss": 6.818, "step": 16343 }, { "epoch": 0.6890823618694268, "grad_norm": 3.299672842025757, "learning_rate": 4.279568381273619e-05, "loss": 7.632, "step": 16344 }, { "epoch": 0.6891245230516264, "grad_norm": 1.238313913345337, "learning_rate": 4.279490302632812e-05, "loss": 6.4796, "step": 16345 }, { "epoch": 0.6891666842338259, "grad_norm": 1.7820426225662231, "learning_rate": 4.279412223992005e-05, "loss": 6.9587, "step": 16346 }, { "epoch": 0.6892088454160255, "grad_norm": 1.5354604721069336, "learning_rate": 4.279334145351198e-05, "loss": 6.5246, "step": 16347 }, { "epoch": 0.689251006598225, "grad_norm": 2.485426425933838, "learning_rate": 4.279256066710391e-05, "loss": 7.2267, "step": 16348 }, { "epoch": 0.6892931677804246, "grad_norm": 2.2728209495544434, "learning_rate": 4.2791779880695834e-05, "loss": 6.705, "step": 16349 }, { "epoch": 0.6893353289626241, "grad_norm": 2.8560688495635986, "learning_rate": 4.279099909428777e-05, "loss": 7.7075, "step": 16350 }, { "epoch": 0.6893774901448236, "grad_norm": 1.8076730966567993, "learning_rate": 4.2790218307879697e-05, "loss": 7.0138, "step": 16351 }, { "epoch": 0.6894196513270232, "grad_norm": 2.649661064147949, "learning_rate": 4.278943752147163e-05, "loss": 7.8215, "step": 16352 }, { "epoch": 0.6894618125092228, "grad_norm": 2.736483573913574, "learning_rate": 4.278865673506356e-05, "loss": 6.8266, "step": 16353 }, { "epoch": 0.6895039736914224, "grad_norm": 3.152785301208496, "learning_rate": 4.278787594865549e-05, "loss": 7.1141, "step": 16354 }, { "epoch": 0.6895461348736218, "grad_norm": 3.115562915802002, "learning_rate": 4.278709516224742e-05, "loss": 7.6084, "step": 16355 }, { "epoch": 0.6895882960558214, "grad_norm": 2.8436152935028076, "learning_rate": 4.2786314375839346e-05, "loss": 6.5952, "step": 16356 }, { "epoch": 0.689630457238021, "grad_norm": 1.4788967370986938, "learning_rate": 4.278553358943128e-05, "loss": 6.6232, "step": 16357 }, { "epoch": 0.6896726184202205, "grad_norm": 1.6620044708251953, "learning_rate": 4.278475280302321e-05, "loss": 6.6121, "step": 16358 }, { "epoch": 0.68971477960242, "grad_norm": 5.117774486541748, "learning_rate": 4.278397201661514e-05, "loss": 7.9324, "step": 16359 }, { "epoch": 0.6897569407846196, "grad_norm": 2.181739568710327, "learning_rate": 4.2783191230207064e-05, "loss": 6.5218, "step": 16360 }, { "epoch": 0.6897991019668192, "grad_norm": 1.9765642881393433, "learning_rate": 4.2782410443798996e-05, "loss": 6.4892, "step": 16361 }, { "epoch": 0.6898412631490187, "grad_norm": 1.5931129455566406, "learning_rate": 4.278162965739093e-05, "loss": 6.8861, "step": 16362 }, { "epoch": 0.6898834243312183, "grad_norm": 2.4476470947265625, "learning_rate": 4.278084887098285e-05, "loss": 6.5545, "step": 16363 }, { "epoch": 0.6899255855134178, "grad_norm": 3.0341176986694336, "learning_rate": 4.278006808457479e-05, "loss": 6.219, "step": 16364 }, { "epoch": 0.6899677466956173, "grad_norm": 2.6093835830688477, "learning_rate": 4.2779287298166714e-05, "loss": 6.3333, "step": 16365 }, { "epoch": 0.6900099078778169, "grad_norm": 2.8011016845703125, "learning_rate": 4.2778506511758645e-05, "loss": 6.6154, "step": 16366 }, { "epoch": 0.6900520690600165, "grad_norm": 2.1651241779327393, "learning_rate": 4.2777725725350576e-05, "loss": 6.702, "step": 16367 }, { "epoch": 0.690094230242216, "grad_norm": 1.5958436727523804, "learning_rate": 4.27769449389425e-05, "loss": 6.6794, "step": 16368 }, { "epoch": 0.6901363914244155, "grad_norm": 1.8050737380981445, "learning_rate": 4.277616415253444e-05, "loss": 6.6978, "step": 16369 }, { "epoch": 0.6901785526066151, "grad_norm": 4.3022356033325195, "learning_rate": 4.277538336612636e-05, "loss": 7.8381, "step": 16370 }, { "epoch": 0.6902207137888147, "grad_norm": 1.5618263483047485, "learning_rate": 4.2774602579718295e-05, "loss": 6.5417, "step": 16371 }, { "epoch": 0.6902628749710142, "grad_norm": 1.6527968645095825, "learning_rate": 4.2773821793310226e-05, "loss": 6.9485, "step": 16372 }, { "epoch": 0.6903050361532137, "grad_norm": 2.521162509918213, "learning_rate": 4.277304100690215e-05, "loss": 6.9635, "step": 16373 }, { "epoch": 0.6903471973354133, "grad_norm": 1.7390304803848267, "learning_rate": 4.277226022049408e-05, "loss": 7.071, "step": 16374 }, { "epoch": 0.6903893585176129, "grad_norm": 3.337937593460083, "learning_rate": 4.277147943408601e-05, "loss": 6.3629, "step": 16375 }, { "epoch": 0.6904315196998124, "grad_norm": 2.261120080947876, "learning_rate": 4.2770698647677944e-05, "loss": 6.9056, "step": 16376 }, { "epoch": 0.6904736808820119, "grad_norm": 3.962573766708374, "learning_rate": 4.276991786126987e-05, "loss": 8.0002, "step": 16377 }, { "epoch": 0.6905158420642115, "grad_norm": 1.4436229467391968, "learning_rate": 4.276913707486181e-05, "loss": 6.5623, "step": 16378 }, { "epoch": 0.690558003246411, "grad_norm": 1.8702383041381836, "learning_rate": 4.276835628845373e-05, "loss": 6.4847, "step": 16379 }, { "epoch": 0.6906001644286106, "grad_norm": 1.834950566291809, "learning_rate": 4.276757550204566e-05, "loss": 6.8819, "step": 16380 }, { "epoch": 0.6906423256108102, "grad_norm": 2.7610578536987305, "learning_rate": 4.2766794715637594e-05, "loss": 6.7642, "step": 16381 }, { "epoch": 0.6906844867930096, "grad_norm": 1.8843367099761963, "learning_rate": 4.276601392922952e-05, "loss": 6.6032, "step": 16382 }, { "epoch": 0.6907266479752092, "grad_norm": 3.248527765274048, "learning_rate": 4.2765233142821456e-05, "loss": 7.1466, "step": 16383 }, { "epoch": 0.6907688091574088, "grad_norm": 5.351835250854492, "learning_rate": 4.276445235641338e-05, "loss": 10.1857, "step": 16384 }, { "epoch": 0.6908109703396084, "grad_norm": 2.2925467491149902, "learning_rate": 4.276367157000531e-05, "loss": 7.0312, "step": 16385 }, { "epoch": 0.6908531315218078, "grad_norm": 2.136194944381714, "learning_rate": 4.276289078359724e-05, "loss": 7.1323, "step": 16386 }, { "epoch": 0.6908952927040074, "grad_norm": 3.2277746200561523, "learning_rate": 4.276210999718917e-05, "loss": 6.6558, "step": 16387 }, { "epoch": 0.690937453886207, "grad_norm": 3.192101240158081, "learning_rate": 4.27613292107811e-05, "loss": 6.4463, "step": 16388 }, { "epoch": 0.6909796150684065, "grad_norm": 2.3114330768585205, "learning_rate": 4.276054842437303e-05, "loss": 7.4281, "step": 16389 }, { "epoch": 0.691021776250606, "grad_norm": 1.6042072772979736, "learning_rate": 4.275976763796496e-05, "loss": 6.6985, "step": 16390 }, { "epoch": 0.6910639374328056, "grad_norm": 4.03806734085083, "learning_rate": 4.275898685155689e-05, "loss": 7.6617, "step": 16391 }, { "epoch": 0.6911060986150052, "grad_norm": 2.6733486652374268, "learning_rate": 4.275820606514882e-05, "loss": 7.1924, "step": 16392 }, { "epoch": 0.6911482597972047, "grad_norm": 3.8469908237457275, "learning_rate": 4.275742527874075e-05, "loss": 7.0097, "step": 16393 }, { "epoch": 0.6911904209794043, "grad_norm": 3.6118826866149902, "learning_rate": 4.275664449233268e-05, "loss": 7.8969, "step": 16394 }, { "epoch": 0.6912325821616038, "grad_norm": 2.9411966800689697, "learning_rate": 4.275586370592461e-05, "loss": 7.372, "step": 16395 }, { "epoch": 0.6912747433438033, "grad_norm": 1.7640832662582397, "learning_rate": 4.2755082919516535e-05, "loss": 6.8982, "step": 16396 }, { "epoch": 0.6913169045260029, "grad_norm": 2.95312237739563, "learning_rate": 4.2754302133108474e-05, "loss": 6.6228, "step": 16397 }, { "epoch": 0.6913590657082025, "grad_norm": 2.817328453063965, "learning_rate": 4.27535213467004e-05, "loss": 6.9813, "step": 16398 }, { "epoch": 0.691401226890402, "grad_norm": 2.381812334060669, "learning_rate": 4.275274056029233e-05, "loss": 6.6429, "step": 16399 }, { "epoch": 0.6914433880726015, "grad_norm": 3.5392796993255615, "learning_rate": 4.275195977388426e-05, "loss": 8.1038, "step": 16400 }, { "epoch": 0.6914855492548011, "grad_norm": 2.746096611022949, "learning_rate": 4.2751178987476185e-05, "loss": 6.9381, "step": 16401 }, { "epoch": 0.6915277104370007, "grad_norm": 2.7847957611083984, "learning_rate": 4.275039820106812e-05, "loss": 6.5068, "step": 16402 }, { "epoch": 0.6915698716192002, "grad_norm": 3.0745363235473633, "learning_rate": 4.274961741466005e-05, "loss": 7.0454, "step": 16403 }, { "epoch": 0.6916120328013997, "grad_norm": 2.0955617427825928, "learning_rate": 4.274883662825198e-05, "loss": 6.4937, "step": 16404 }, { "epoch": 0.6916541939835993, "grad_norm": 1.7706719636917114, "learning_rate": 4.274805584184391e-05, "loss": 6.5052, "step": 16405 }, { "epoch": 0.6916963551657989, "grad_norm": 1.637472152709961, "learning_rate": 4.2747275055435835e-05, "loss": 6.4833, "step": 16406 }, { "epoch": 0.6917385163479984, "grad_norm": 1.2665340900421143, "learning_rate": 4.2746494269027766e-05, "loss": 6.5442, "step": 16407 }, { "epoch": 0.6917806775301979, "grad_norm": 1.4699167013168335, "learning_rate": 4.27457134826197e-05, "loss": 6.5586, "step": 16408 }, { "epoch": 0.6918228387123975, "grad_norm": 2.3207638263702393, "learning_rate": 4.274493269621163e-05, "loss": 6.9805, "step": 16409 }, { "epoch": 0.691864999894597, "grad_norm": 1.5782870054244995, "learning_rate": 4.274415190980355e-05, "loss": 6.7397, "step": 16410 }, { "epoch": 0.6919071610767966, "grad_norm": 1.863631248474121, "learning_rate": 4.2743371123395484e-05, "loss": 6.6665, "step": 16411 }, { "epoch": 0.6919493222589962, "grad_norm": 2.0496346950531006, "learning_rate": 4.2742590336987415e-05, "loss": 6.9701, "step": 16412 }, { "epoch": 0.6919914834411957, "grad_norm": 3.314164876937866, "learning_rate": 4.2741809550579347e-05, "loss": 7.4067, "step": 16413 }, { "epoch": 0.6920336446233952, "grad_norm": 3.7922329902648926, "learning_rate": 4.274102876417128e-05, "loss": 7.2543, "step": 16414 }, { "epoch": 0.6920758058055948, "grad_norm": 3.145747423171997, "learning_rate": 4.27402479777632e-05, "loss": 7.5876, "step": 16415 }, { "epoch": 0.6921179669877944, "grad_norm": 2.5389902591705322, "learning_rate": 4.273946719135514e-05, "loss": 6.4482, "step": 16416 }, { "epoch": 0.6921601281699938, "grad_norm": 2.5769009590148926, "learning_rate": 4.2738686404947065e-05, "loss": 7.0182, "step": 16417 }, { "epoch": 0.6922022893521934, "grad_norm": 1.6599888801574707, "learning_rate": 4.273790561853899e-05, "loss": 7.0433, "step": 16418 }, { "epoch": 0.692244450534393, "grad_norm": 1.3092567920684814, "learning_rate": 4.273712483213093e-05, "loss": 6.7631, "step": 16419 }, { "epoch": 0.6922866117165926, "grad_norm": 2.2749600410461426, "learning_rate": 4.273634404572285e-05, "loss": 7.1679, "step": 16420 }, { "epoch": 0.6923287728987921, "grad_norm": 3.6489651203155518, "learning_rate": 4.273556325931478e-05, "loss": 6.4511, "step": 16421 }, { "epoch": 0.6923709340809916, "grad_norm": 3.432476282119751, "learning_rate": 4.2734782472906714e-05, "loss": 7.5569, "step": 16422 }, { "epoch": 0.6924130952631912, "grad_norm": 1.374456524848938, "learning_rate": 4.2734001686498646e-05, "loss": 6.4268, "step": 16423 }, { "epoch": 0.6924552564453907, "grad_norm": 2.285172700881958, "learning_rate": 4.273322090009058e-05, "loss": 7.1986, "step": 16424 }, { "epoch": 0.6924974176275903, "grad_norm": 4.155237674713135, "learning_rate": 4.27324401136825e-05, "loss": 9.9448, "step": 16425 }, { "epoch": 0.6925395788097898, "grad_norm": 2.0725739002227783, "learning_rate": 4.273165932727443e-05, "loss": 6.7783, "step": 16426 }, { "epoch": 0.6925817399919894, "grad_norm": 4.415273189544678, "learning_rate": 4.2730878540866364e-05, "loss": 7.1056, "step": 16427 }, { "epoch": 0.6926239011741889, "grad_norm": 2.1339831352233887, "learning_rate": 4.2730097754458295e-05, "loss": 6.6787, "step": 16428 }, { "epoch": 0.6926660623563885, "grad_norm": 2.1070637702941895, "learning_rate": 4.272931696805022e-05, "loss": 6.7319, "step": 16429 }, { "epoch": 0.6927082235385881, "grad_norm": 3.5434606075286865, "learning_rate": 4.272853618164215e-05, "loss": 6.985, "step": 16430 }, { "epoch": 0.6927503847207875, "grad_norm": 3.4803030490875244, "learning_rate": 4.272775539523408e-05, "loss": 7.639, "step": 16431 }, { "epoch": 0.6927925459029871, "grad_norm": 3.803590774536133, "learning_rate": 4.272697460882601e-05, "loss": 6.996, "step": 16432 }, { "epoch": 0.6928347070851867, "grad_norm": 3.0675466060638428, "learning_rate": 4.2726193822417945e-05, "loss": 6.5958, "step": 16433 }, { "epoch": 0.6928768682673863, "grad_norm": 1.7848495244979858, "learning_rate": 4.272541303600987e-05, "loss": 6.4339, "step": 16434 }, { "epoch": 0.6929190294495857, "grad_norm": 7.563145160675049, "learning_rate": 4.27246322496018e-05, "loss": 11.8885, "step": 16435 }, { "epoch": 0.6929611906317853, "grad_norm": 2.1191110610961914, "learning_rate": 4.272385146319373e-05, "loss": 6.4873, "step": 16436 }, { "epoch": 0.6930033518139849, "grad_norm": 3.0650603771209717, "learning_rate": 4.2723070676785656e-05, "loss": 6.371, "step": 16437 }, { "epoch": 0.6930455129961844, "grad_norm": 5.762583255767822, "learning_rate": 4.2722289890377594e-05, "loss": 7.3034, "step": 16438 }, { "epoch": 0.693087674178384, "grad_norm": 3.095747232437134, "learning_rate": 4.272150910396952e-05, "loss": 6.4601, "step": 16439 }, { "epoch": 0.6931298353605835, "grad_norm": 2.2898995876312256, "learning_rate": 4.272072831756145e-05, "loss": 6.8652, "step": 16440 }, { "epoch": 0.693171996542783, "grad_norm": 3.3633737564086914, "learning_rate": 4.271994753115338e-05, "loss": 6.3255, "step": 16441 }, { "epoch": 0.6932141577249826, "grad_norm": 3.437018871307373, "learning_rate": 4.271916674474531e-05, "loss": 6.8504, "step": 16442 }, { "epoch": 0.6932563189071822, "grad_norm": 2.690418243408203, "learning_rate": 4.271838595833724e-05, "loss": 6.6631, "step": 16443 }, { "epoch": 0.6932984800893817, "grad_norm": 2.4886255264282227, "learning_rate": 4.271760517192917e-05, "loss": 6.9277, "step": 16444 }, { "epoch": 0.6933406412715812, "grad_norm": 1.5068113803863525, "learning_rate": 4.27168243855211e-05, "loss": 6.5914, "step": 16445 }, { "epoch": 0.6933828024537808, "grad_norm": 1.706920862197876, "learning_rate": 4.271604359911303e-05, "loss": 6.7515, "step": 16446 }, { "epoch": 0.6934249636359804, "grad_norm": 2.769543409347534, "learning_rate": 4.271526281270496e-05, "loss": 6.2524, "step": 16447 }, { "epoch": 0.6934671248181798, "grad_norm": 3.7977826595306396, "learning_rate": 4.2714482026296886e-05, "loss": 7.8047, "step": 16448 }, { "epoch": 0.6935092860003794, "grad_norm": 1.6876431703567505, "learning_rate": 4.271370123988882e-05, "loss": 6.6422, "step": 16449 }, { "epoch": 0.693551447182579, "grad_norm": 3.8451364040374756, "learning_rate": 4.271292045348075e-05, "loss": 7.556, "step": 16450 }, { "epoch": 0.6935936083647786, "grad_norm": 2.0535690784454346, "learning_rate": 4.2712139667072673e-05, "loss": 7.0883, "step": 16451 }, { "epoch": 0.6936357695469781, "grad_norm": 2.7925806045532227, "learning_rate": 4.271135888066461e-05, "loss": 7.0596, "step": 16452 }, { "epoch": 0.6936779307291776, "grad_norm": 1.4047960042953491, "learning_rate": 4.2710578094256536e-05, "loss": 6.5541, "step": 16453 }, { "epoch": 0.6937200919113772, "grad_norm": 1.4361896514892578, "learning_rate": 4.270979730784847e-05, "loss": 6.6104, "step": 16454 }, { "epoch": 0.6937622530935768, "grad_norm": 3.242696523666382, "learning_rate": 4.27090165214404e-05, "loss": 7.4227, "step": 16455 }, { "epoch": 0.6938044142757763, "grad_norm": 2.8675737380981445, "learning_rate": 4.270823573503232e-05, "loss": 7.1123, "step": 16456 }, { "epoch": 0.6938465754579758, "grad_norm": 1.9737544059753418, "learning_rate": 4.2707454948624254e-05, "loss": 6.8118, "step": 16457 }, { "epoch": 0.6938887366401754, "grad_norm": 2.6651816368103027, "learning_rate": 4.2706674162216185e-05, "loss": 6.9831, "step": 16458 }, { "epoch": 0.6939308978223749, "grad_norm": 1.4559292793273926, "learning_rate": 4.270589337580812e-05, "loss": 7.1623, "step": 16459 }, { "epoch": 0.6939730590045745, "grad_norm": 2.153332233428955, "learning_rate": 4.270511258940005e-05, "loss": 6.62, "step": 16460 }, { "epoch": 0.6940152201867741, "grad_norm": 2.27052640914917, "learning_rate": 4.270433180299198e-05, "loss": 6.5588, "step": 16461 }, { "epoch": 0.6940573813689735, "grad_norm": 1.4350016117095947, "learning_rate": 4.2703551016583904e-05, "loss": 6.4186, "step": 16462 }, { "epoch": 0.6940995425511731, "grad_norm": 2.334774971008301, "learning_rate": 4.2702770230175835e-05, "loss": 7.6307, "step": 16463 }, { "epoch": 0.6941417037333727, "grad_norm": 3.406116485595703, "learning_rate": 4.2701989443767766e-05, "loss": 7.0824, "step": 16464 }, { "epoch": 0.6941838649155723, "grad_norm": 1.9707170724868774, "learning_rate": 4.270120865735969e-05, "loss": 6.4522, "step": 16465 }, { "epoch": 0.6942260260977717, "grad_norm": 1.9820733070373535, "learning_rate": 4.270042787095163e-05, "loss": 6.4996, "step": 16466 }, { "epoch": 0.6942681872799713, "grad_norm": 1.401857614517212, "learning_rate": 4.269964708454355e-05, "loss": 6.6262, "step": 16467 }, { "epoch": 0.6943103484621709, "grad_norm": 1.3859829902648926, "learning_rate": 4.2698866298135485e-05, "loss": 6.6261, "step": 16468 }, { "epoch": 0.6943525096443705, "grad_norm": 1.960455298423767, "learning_rate": 4.2698085511727416e-05, "loss": 6.3132, "step": 16469 }, { "epoch": 0.69439467082657, "grad_norm": 2.3617541790008545, "learning_rate": 4.269730472531934e-05, "loss": 7.0041, "step": 16470 }, { "epoch": 0.6944368320087695, "grad_norm": 2.914374589920044, "learning_rate": 4.269652393891128e-05, "loss": 6.3308, "step": 16471 }, { "epoch": 0.6944789931909691, "grad_norm": 1.449880838394165, "learning_rate": 4.26957431525032e-05, "loss": 6.4987, "step": 16472 }, { "epoch": 0.6945211543731686, "grad_norm": 1.1191767454147339, "learning_rate": 4.2694962366095134e-05, "loss": 6.4167, "step": 16473 }, { "epoch": 0.6945633155553682, "grad_norm": 1.5070691108703613, "learning_rate": 4.2694181579687065e-05, "loss": 6.3096, "step": 16474 }, { "epoch": 0.6946054767375677, "grad_norm": 2.41506028175354, "learning_rate": 4.269340079327899e-05, "loss": 6.9935, "step": 16475 }, { "epoch": 0.6946476379197672, "grad_norm": 4.118337154388428, "learning_rate": 4.269262000687092e-05, "loss": 7.2177, "step": 16476 }, { "epoch": 0.6946897991019668, "grad_norm": 3.622662305831909, "learning_rate": 4.269183922046285e-05, "loss": 7.6381, "step": 16477 }, { "epoch": 0.6947319602841664, "grad_norm": 2.0217885971069336, "learning_rate": 4.2691058434054784e-05, "loss": 6.9536, "step": 16478 }, { "epoch": 0.694774121466366, "grad_norm": 2.5895707607269287, "learning_rate": 4.269027764764671e-05, "loss": 7.7363, "step": 16479 }, { "epoch": 0.6948162826485654, "grad_norm": 3.3006834983825684, "learning_rate": 4.268949686123864e-05, "loss": 6.6944, "step": 16480 }, { "epoch": 0.694858443830765, "grad_norm": 3.185328722000122, "learning_rate": 4.268871607483057e-05, "loss": 6.5832, "step": 16481 }, { "epoch": 0.6949006050129646, "grad_norm": 2.816929817199707, "learning_rate": 4.26879352884225e-05, "loss": 7.0143, "step": 16482 }, { "epoch": 0.6949427661951642, "grad_norm": 2.1305134296417236, "learning_rate": 4.268715450201443e-05, "loss": 6.5538, "step": 16483 }, { "epoch": 0.6949849273773636, "grad_norm": 1.698906660079956, "learning_rate": 4.268637371560636e-05, "loss": 6.3626, "step": 16484 }, { "epoch": 0.6950270885595632, "grad_norm": 2.3963160514831543, "learning_rate": 4.2685592929198296e-05, "loss": 7.2137, "step": 16485 }, { "epoch": 0.6950692497417628, "grad_norm": 2.724736452102661, "learning_rate": 4.268481214279022e-05, "loss": 6.5623, "step": 16486 }, { "epoch": 0.6951114109239623, "grad_norm": 3.478158950805664, "learning_rate": 4.268403135638215e-05, "loss": 7.3473, "step": 16487 }, { "epoch": 0.6951535721061619, "grad_norm": 1.9807559251785278, "learning_rate": 4.268325056997408e-05, "loss": 6.7111, "step": 16488 }, { "epoch": 0.6951957332883614, "grad_norm": 1.125123381614685, "learning_rate": 4.268246978356601e-05, "loss": 6.4885, "step": 16489 }, { "epoch": 0.695237894470561, "grad_norm": 1.7999811172485352, "learning_rate": 4.268168899715794e-05, "loss": 6.9811, "step": 16490 }, { "epoch": 0.6952800556527605, "grad_norm": 1.7829697132110596, "learning_rate": 4.268090821074987e-05, "loss": 6.584, "step": 16491 }, { "epoch": 0.6953222168349601, "grad_norm": 1.713120937347412, "learning_rate": 4.26801274243418e-05, "loss": 6.9733, "step": 16492 }, { "epoch": 0.6953643780171596, "grad_norm": 3.918613910675049, "learning_rate": 4.267934663793373e-05, "loss": 7.9546, "step": 16493 }, { "epoch": 0.6954065391993591, "grad_norm": 2.736875534057617, "learning_rate": 4.267856585152566e-05, "loss": 7.0283, "step": 16494 }, { "epoch": 0.6954487003815587, "grad_norm": 2.014874219894409, "learning_rate": 4.267778506511759e-05, "loss": 6.3767, "step": 16495 }, { "epoch": 0.6954908615637583, "grad_norm": 4.935887336730957, "learning_rate": 4.267700427870952e-05, "loss": 9.0548, "step": 16496 }, { "epoch": 0.6955330227459579, "grad_norm": 2.002070188522339, "learning_rate": 4.267622349230145e-05, "loss": 6.3873, "step": 16497 }, { "epoch": 0.6955751839281573, "grad_norm": 2.337390184402466, "learning_rate": 4.2675442705893375e-05, "loss": 7.2697, "step": 16498 }, { "epoch": 0.6956173451103569, "grad_norm": 2.194613456726074, "learning_rate": 4.2674661919485306e-05, "loss": 6.5051, "step": 16499 }, { "epoch": 0.6956595062925565, "grad_norm": 1.6041691303253174, "learning_rate": 4.267388113307724e-05, "loss": 7.8562, "step": 16500 }, { "epoch": 0.695701667474756, "grad_norm": 2.7933156490325928, "learning_rate": 4.267310034666916e-05, "loss": 6.992, "step": 16501 }, { "epoch": 0.6957438286569555, "grad_norm": 3.480663299560547, "learning_rate": 4.26723195602611e-05, "loss": 6.6506, "step": 16502 }, { "epoch": 0.6957859898391551, "grad_norm": 2.308194160461426, "learning_rate": 4.2671538773853024e-05, "loss": 7.4299, "step": 16503 }, { "epoch": 0.6958281510213546, "grad_norm": 3.3743903636932373, "learning_rate": 4.267075798744496e-05, "loss": 6.3637, "step": 16504 }, { "epoch": 0.6958703122035542, "grad_norm": 1.3180172443389893, "learning_rate": 4.266997720103689e-05, "loss": 6.2742, "step": 16505 }, { "epoch": 0.6959124733857537, "grad_norm": 2.5190069675445557, "learning_rate": 4.266919641462881e-05, "loss": 7.518, "step": 16506 }, { "epoch": 0.6959546345679533, "grad_norm": 3.3058629035949707, "learning_rate": 4.266841562822075e-05, "loss": 7.2107, "step": 16507 }, { "epoch": 0.6959967957501528, "grad_norm": 1.547167420387268, "learning_rate": 4.2667634841812674e-05, "loss": 6.4282, "step": 16508 }, { "epoch": 0.6960389569323524, "grad_norm": 1.1812125444412231, "learning_rate": 4.2666854055404605e-05, "loss": 6.6164, "step": 16509 }, { "epoch": 0.696081118114552, "grad_norm": 2.7926137447357178, "learning_rate": 4.2666073268996536e-05, "loss": 8.3604, "step": 16510 }, { "epoch": 0.6961232792967514, "grad_norm": 2.0038700103759766, "learning_rate": 4.266529248258847e-05, "loss": 6.365, "step": 16511 }, { "epoch": 0.696165440478951, "grad_norm": 2.7662768363952637, "learning_rate": 4.266451169618039e-05, "loss": 7.091, "step": 16512 }, { "epoch": 0.6962076016611506, "grad_norm": 3.891404867172241, "learning_rate": 4.2663730909772323e-05, "loss": 7.8754, "step": 16513 }, { "epoch": 0.6962497628433502, "grad_norm": 3.329702615737915, "learning_rate": 4.2662950123364255e-05, "loss": 7.6938, "step": 16514 }, { "epoch": 0.6962919240255496, "grad_norm": 2.859431743621826, "learning_rate": 4.2662169336956186e-05, "loss": 6.4435, "step": 16515 }, { "epoch": 0.6963340852077492, "grad_norm": 1.800734281539917, "learning_rate": 4.266138855054812e-05, "loss": 7.062, "step": 16516 }, { "epoch": 0.6963762463899488, "grad_norm": 1.5422240495681763, "learning_rate": 4.266060776414004e-05, "loss": 6.6417, "step": 16517 }, { "epoch": 0.6964184075721483, "grad_norm": 2.4187822341918945, "learning_rate": 4.265982697773197e-05, "loss": 6.8712, "step": 16518 }, { "epoch": 0.6964605687543479, "grad_norm": 1.7763311862945557, "learning_rate": 4.2659046191323904e-05, "loss": 6.7681, "step": 16519 }, { "epoch": 0.6965027299365474, "grad_norm": 2.6765990257263184, "learning_rate": 4.265826540491583e-05, "loss": 7.0513, "step": 16520 }, { "epoch": 0.696544891118747, "grad_norm": 1.2751224040985107, "learning_rate": 4.265748461850777e-05, "loss": 6.6127, "step": 16521 }, { "epoch": 0.6965870523009465, "grad_norm": 1.3274116516113281, "learning_rate": 4.265670383209969e-05, "loss": 7.0264, "step": 16522 }, { "epoch": 0.6966292134831461, "grad_norm": 1.8645654916763306, "learning_rate": 4.265592304569162e-05, "loss": 6.9047, "step": 16523 }, { "epoch": 0.6966713746653456, "grad_norm": 2.3592801094055176, "learning_rate": 4.2655142259283554e-05, "loss": 6.9137, "step": 16524 }, { "epoch": 0.6967135358475451, "grad_norm": 1.428773283958435, "learning_rate": 4.265436147287548e-05, "loss": 6.8666, "step": 16525 }, { "epoch": 0.6967556970297447, "grad_norm": 1.7745128870010376, "learning_rate": 4.2653580686467416e-05, "loss": 6.312, "step": 16526 }, { "epoch": 0.6967978582119443, "grad_norm": 1.169000267982483, "learning_rate": 4.265279990005934e-05, "loss": 6.8257, "step": 16527 }, { "epoch": 0.6968400193941439, "grad_norm": 1.6796181201934814, "learning_rate": 4.265201911365127e-05, "loss": 6.6419, "step": 16528 }, { "epoch": 0.6968821805763433, "grad_norm": 1.2456469535827637, "learning_rate": 4.26512383272432e-05, "loss": 6.4357, "step": 16529 }, { "epoch": 0.6969243417585429, "grad_norm": 1.3519585132598877, "learning_rate": 4.2650457540835135e-05, "loss": 6.531, "step": 16530 }, { "epoch": 0.6969665029407425, "grad_norm": 1.6436043977737427, "learning_rate": 4.264967675442706e-05, "loss": 6.697, "step": 16531 }, { "epoch": 0.697008664122942, "grad_norm": 2.330099105834961, "learning_rate": 4.264889596801899e-05, "loss": 6.8644, "step": 16532 }, { "epoch": 0.6970508253051415, "grad_norm": 0.890482485294342, "learning_rate": 4.264811518161092e-05, "loss": 6.5196, "step": 16533 }, { "epoch": 0.6970929864873411, "grad_norm": 1.6944700479507446, "learning_rate": 4.2647334395202846e-05, "loss": 6.8308, "step": 16534 }, { "epoch": 0.6971351476695407, "grad_norm": 1.4741536378860474, "learning_rate": 4.2646553608794784e-05, "loss": 6.4998, "step": 16535 }, { "epoch": 0.6971773088517402, "grad_norm": 4.717737674713135, "learning_rate": 4.264577282238671e-05, "loss": 8.8048, "step": 16536 }, { "epoch": 0.6972194700339398, "grad_norm": 1.4185816049575806, "learning_rate": 4.264499203597864e-05, "loss": 6.7049, "step": 16537 }, { "epoch": 0.6972616312161393, "grad_norm": 3.3371756076812744, "learning_rate": 4.264421124957057e-05, "loss": 6.3223, "step": 16538 }, { "epoch": 0.6973037923983388, "grad_norm": 2.800626516342163, "learning_rate": 4.2643430463162496e-05, "loss": 6.6183, "step": 16539 }, { "epoch": 0.6973459535805384, "grad_norm": 1.8799898624420166, "learning_rate": 4.2642649676754434e-05, "loss": 6.3787, "step": 16540 }, { "epoch": 0.697388114762738, "grad_norm": 3.3843870162963867, "learning_rate": 4.264186889034636e-05, "loss": 7.895, "step": 16541 }, { "epoch": 0.6974302759449375, "grad_norm": 3.1180174350738525, "learning_rate": 4.264108810393829e-05, "loss": 7.2259, "step": 16542 }, { "epoch": 0.697472437127137, "grad_norm": 1.4659790992736816, "learning_rate": 4.264030731753022e-05, "loss": 6.8146, "step": 16543 }, { "epoch": 0.6975145983093366, "grad_norm": 1.3297570943832397, "learning_rate": 4.2639526531122145e-05, "loss": 6.8673, "step": 16544 }, { "epoch": 0.6975567594915362, "grad_norm": 1.5963701009750366, "learning_rate": 4.2638745744714076e-05, "loss": 7.062, "step": 16545 }, { "epoch": 0.6975989206737357, "grad_norm": 2.2461435794830322, "learning_rate": 4.263796495830601e-05, "loss": 6.9736, "step": 16546 }, { "epoch": 0.6976410818559352, "grad_norm": 1.604505181312561, "learning_rate": 4.263718417189794e-05, "loss": 6.3478, "step": 16547 }, { "epoch": 0.6976832430381348, "grad_norm": 2.254960536956787, "learning_rate": 4.263640338548986e-05, "loss": 7.2713, "step": 16548 }, { "epoch": 0.6977254042203344, "grad_norm": 3.273899555206299, "learning_rate": 4.26356225990818e-05, "loss": 7.15, "step": 16549 }, { "epoch": 0.6977675654025339, "grad_norm": 2.340123414993286, "learning_rate": 4.2634841812673726e-05, "loss": 7.1493, "step": 16550 }, { "epoch": 0.6978097265847334, "grad_norm": 4.838429927825928, "learning_rate": 4.263406102626566e-05, "loss": 10.5996, "step": 16551 }, { "epoch": 0.697851887766933, "grad_norm": 1.7904077768325806, "learning_rate": 4.263328023985759e-05, "loss": 7.4484, "step": 16552 }, { "epoch": 0.6978940489491325, "grad_norm": 2.371281147003174, "learning_rate": 4.263249945344951e-05, "loss": 6.4272, "step": 16553 }, { "epoch": 0.6979362101313321, "grad_norm": 1.3839505910873413, "learning_rate": 4.263171866704145e-05, "loss": 6.4063, "step": 16554 }, { "epoch": 0.6979783713135317, "grad_norm": 3.0165228843688965, "learning_rate": 4.2630937880633375e-05, "loss": 6.4582, "step": 16555 }, { "epoch": 0.6980205324957311, "grad_norm": 2.3264806270599365, "learning_rate": 4.263015709422531e-05, "loss": 6.612, "step": 16556 }, { "epoch": 0.6980626936779307, "grad_norm": 3.501661777496338, "learning_rate": 4.262937630781724e-05, "loss": 6.6071, "step": 16557 }, { "epoch": 0.6981048548601303, "grad_norm": 4.2351484298706055, "learning_rate": 4.262859552140916e-05, "loss": 6.9369, "step": 16558 }, { "epoch": 0.6981470160423299, "grad_norm": 4.083778381347656, "learning_rate": 4.2627814735001094e-05, "loss": 7.1386, "step": 16559 }, { "epoch": 0.6981891772245293, "grad_norm": 4.408188819885254, "learning_rate": 4.2627033948593025e-05, "loss": 7.5254, "step": 16560 }, { "epoch": 0.6982313384067289, "grad_norm": 3.4698832035064697, "learning_rate": 4.2626253162184956e-05, "loss": 7.0522, "step": 16561 }, { "epoch": 0.6982734995889285, "grad_norm": 2.0254547595977783, "learning_rate": 4.262547237577689e-05, "loss": 6.6092, "step": 16562 }, { "epoch": 0.698315660771128, "grad_norm": 2.726111888885498, "learning_rate": 4.262469158936881e-05, "loss": 6.7831, "step": 16563 }, { "epoch": 0.6983578219533275, "grad_norm": 2.8777875900268555, "learning_rate": 4.262391080296074e-05, "loss": 7.7873, "step": 16564 }, { "epoch": 0.6983999831355271, "grad_norm": 2.4848413467407227, "learning_rate": 4.2623130016552674e-05, "loss": 7.3, "step": 16565 }, { "epoch": 0.6984421443177267, "grad_norm": 2.7262094020843506, "learning_rate": 4.2622349230144606e-05, "loss": 8.692, "step": 16566 }, { "epoch": 0.6984843054999262, "grad_norm": 1.8011735677719116, "learning_rate": 4.262156844373653e-05, "loss": 6.4701, "step": 16567 }, { "epoch": 0.6985264666821258, "grad_norm": 1.4289740324020386, "learning_rate": 4.262078765732846e-05, "loss": 6.6586, "step": 16568 }, { "epoch": 0.6985686278643253, "grad_norm": 1.6933033466339111, "learning_rate": 4.262000687092039e-05, "loss": 6.362, "step": 16569 }, { "epoch": 0.6986107890465248, "grad_norm": 2.347341299057007, "learning_rate": 4.261922608451232e-05, "loss": 6.7774, "step": 16570 }, { "epoch": 0.6986529502287244, "grad_norm": 1.7693275213241577, "learning_rate": 4.2618445298104255e-05, "loss": 7.1409, "step": 16571 }, { "epoch": 0.698695111410924, "grad_norm": 1.8248943090438843, "learning_rate": 4.261766451169618e-05, "loss": 7.0944, "step": 16572 }, { "epoch": 0.6987372725931235, "grad_norm": 1.3728071451187134, "learning_rate": 4.261688372528812e-05, "loss": 7.4425, "step": 16573 }, { "epoch": 0.698779433775323, "grad_norm": 2.875731945037842, "learning_rate": 4.261610293888004e-05, "loss": 7.0803, "step": 16574 }, { "epoch": 0.6988215949575226, "grad_norm": 3.137580633163452, "learning_rate": 4.2615322152471973e-05, "loss": 7.2153, "step": 16575 }, { "epoch": 0.6988637561397222, "grad_norm": 2.6317286491394043, "learning_rate": 4.2614541366063905e-05, "loss": 6.4544, "step": 16576 }, { "epoch": 0.6989059173219218, "grad_norm": 3.291104555130005, "learning_rate": 4.261376057965583e-05, "loss": 7.51, "step": 16577 }, { "epoch": 0.6989480785041212, "grad_norm": 3.6493289470672607, "learning_rate": 4.261297979324776e-05, "loss": 6.43, "step": 16578 }, { "epoch": 0.6989902396863208, "grad_norm": 3.6314914226531982, "learning_rate": 4.261219900683969e-05, "loss": 7.0334, "step": 16579 }, { "epoch": 0.6990324008685204, "grad_norm": 5.168342590332031, "learning_rate": 4.261141822043162e-05, "loss": 7.8205, "step": 16580 }, { "epoch": 0.6990745620507199, "grad_norm": 2.140221118927002, "learning_rate": 4.261063743402355e-05, "loss": 7.0156, "step": 16581 }, { "epoch": 0.6991167232329194, "grad_norm": 1.2709888219833374, "learning_rate": 4.260985664761548e-05, "loss": 6.6222, "step": 16582 }, { "epoch": 0.699158884415119, "grad_norm": 1.7644217014312744, "learning_rate": 4.260907586120741e-05, "loss": 6.4105, "step": 16583 }, { "epoch": 0.6992010455973185, "grad_norm": 2.3257596492767334, "learning_rate": 4.260829507479934e-05, "loss": 6.8231, "step": 16584 }, { "epoch": 0.6992432067795181, "grad_norm": 2.950004816055298, "learning_rate": 4.260751428839127e-05, "loss": 6.8485, "step": 16585 }, { "epoch": 0.6992853679617177, "grad_norm": 1.8262368440628052, "learning_rate": 4.26067335019832e-05, "loss": 6.5405, "step": 16586 }, { "epoch": 0.6993275291439172, "grad_norm": 1.9699537754058838, "learning_rate": 4.260595271557513e-05, "loss": 7.0383, "step": 16587 }, { "epoch": 0.6993696903261167, "grad_norm": 3.6617844104766846, "learning_rate": 4.260517192916706e-05, "loss": 6.7837, "step": 16588 }, { "epoch": 0.6994118515083163, "grad_norm": 2.937398672103882, "learning_rate": 4.2604391142758984e-05, "loss": 7.0908, "step": 16589 }, { "epoch": 0.6994540126905159, "grad_norm": 3.166801691055298, "learning_rate": 4.260361035635092e-05, "loss": 6.8794, "step": 16590 }, { "epoch": 0.6994961738727153, "grad_norm": 5.239147186279297, "learning_rate": 4.2602829569942847e-05, "loss": 7.5824, "step": 16591 }, { "epoch": 0.6995383350549149, "grad_norm": 2.8740410804748535, "learning_rate": 4.260204878353478e-05, "loss": 6.4895, "step": 16592 }, { "epoch": 0.6995804962371145, "grad_norm": 1.9454628229141235, "learning_rate": 4.260126799712671e-05, "loss": 6.9112, "step": 16593 }, { "epoch": 0.6996226574193141, "grad_norm": 1.9526249170303345, "learning_rate": 4.260048721071864e-05, "loss": 7.1655, "step": 16594 }, { "epoch": 0.6996648186015136, "grad_norm": 2.839399814605713, "learning_rate": 4.259970642431057e-05, "loss": 6.5737, "step": 16595 }, { "epoch": 0.6997069797837131, "grad_norm": 3.6529366970062256, "learning_rate": 4.2598925637902496e-05, "loss": 6.2672, "step": 16596 }, { "epoch": 0.6997491409659127, "grad_norm": 2.7008886337280273, "learning_rate": 4.259814485149443e-05, "loss": 6.5144, "step": 16597 }, { "epoch": 0.6997913021481122, "grad_norm": 1.9263041019439697, "learning_rate": 4.259736406508636e-05, "loss": 7.0579, "step": 16598 }, { "epoch": 0.6998334633303118, "grad_norm": 4.0292768478393555, "learning_rate": 4.259658327867829e-05, "loss": 7.7125, "step": 16599 }, { "epoch": 0.6998756245125113, "grad_norm": 4.212540626525879, "learning_rate": 4.2595802492270214e-05, "loss": 7.7164, "step": 16600 }, { "epoch": 0.6999177856947109, "grad_norm": 3.5478057861328125, "learning_rate": 4.2595021705862146e-05, "loss": 7.5242, "step": 16601 }, { "epoch": 0.6999599468769104, "grad_norm": 1.7571792602539062, "learning_rate": 4.259424091945408e-05, "loss": 6.4954, "step": 16602 }, { "epoch": 0.70000210805911, "grad_norm": 3.8255958557128906, "learning_rate": 4.2593460133046e-05, "loss": 7.3487, "step": 16603 }, { "epoch": 0.7000442692413096, "grad_norm": 1.6735105514526367, "learning_rate": 4.259267934663794e-05, "loss": 6.8112, "step": 16604 }, { "epoch": 0.700086430423509, "grad_norm": 1.6285679340362549, "learning_rate": 4.2591898560229864e-05, "loss": 6.8315, "step": 16605 }, { "epoch": 0.7001285916057086, "grad_norm": 1.4594463109970093, "learning_rate": 4.2591117773821795e-05, "loss": 6.7972, "step": 16606 }, { "epoch": 0.7001707527879082, "grad_norm": 1.3429179191589355, "learning_rate": 4.2590336987413726e-05, "loss": 6.4758, "step": 16607 }, { "epoch": 0.7002129139701078, "grad_norm": 3.5608818531036377, "learning_rate": 4.258955620100565e-05, "loss": 7.7331, "step": 16608 }, { "epoch": 0.7002550751523072, "grad_norm": 2.8570127487182617, "learning_rate": 4.258877541459759e-05, "loss": 7.4745, "step": 16609 }, { "epoch": 0.7002972363345068, "grad_norm": 2.659680128097534, "learning_rate": 4.258799462818951e-05, "loss": 7.3854, "step": 16610 }, { "epoch": 0.7003393975167064, "grad_norm": 2.4578261375427246, "learning_rate": 4.2587213841781445e-05, "loss": 7.4292, "step": 16611 }, { "epoch": 0.700381558698906, "grad_norm": 3.344303607940674, "learning_rate": 4.2586433055373376e-05, "loss": 6.3841, "step": 16612 }, { "epoch": 0.7004237198811054, "grad_norm": 3.257112741470337, "learning_rate": 4.25856522689653e-05, "loss": 7.2413, "step": 16613 }, { "epoch": 0.700465881063305, "grad_norm": 2.5078258514404297, "learning_rate": 4.258487148255723e-05, "loss": 7.1804, "step": 16614 }, { "epoch": 0.7005080422455046, "grad_norm": 2.8555431365966797, "learning_rate": 4.258409069614916e-05, "loss": 6.835, "step": 16615 }, { "epoch": 0.7005502034277041, "grad_norm": 2.3961894512176514, "learning_rate": 4.2583309909741094e-05, "loss": 6.9784, "step": 16616 }, { "epoch": 0.7005923646099037, "grad_norm": 3.240766763687134, "learning_rate": 4.258252912333302e-05, "loss": 7.5454, "step": 16617 }, { "epoch": 0.7006345257921032, "grad_norm": 1.5601643323898315, "learning_rate": 4.258174833692496e-05, "loss": 6.8392, "step": 16618 }, { "epoch": 0.7006766869743027, "grad_norm": 1.920865774154663, "learning_rate": 4.258096755051688e-05, "loss": 6.5182, "step": 16619 }, { "epoch": 0.7007188481565023, "grad_norm": 2.9688854217529297, "learning_rate": 4.258018676410881e-05, "loss": 6.2378, "step": 16620 }, { "epoch": 0.7007610093387019, "grad_norm": 2.2846932411193848, "learning_rate": 4.2579405977700744e-05, "loss": 6.6576, "step": 16621 }, { "epoch": 0.7008031705209014, "grad_norm": 4.51241397857666, "learning_rate": 4.257862519129267e-05, "loss": 7.7102, "step": 16622 }, { "epoch": 0.7008453317031009, "grad_norm": 4.4824137687683105, "learning_rate": 4.2577844404884606e-05, "loss": 8.4951, "step": 16623 }, { "epoch": 0.7008874928853005, "grad_norm": 3.016371965408325, "learning_rate": 4.257706361847653e-05, "loss": 6.7466, "step": 16624 }, { "epoch": 0.7009296540675001, "grad_norm": 2.8445441722869873, "learning_rate": 4.257628283206846e-05, "loss": 6.6271, "step": 16625 }, { "epoch": 0.7009718152496996, "grad_norm": 2.4777092933654785, "learning_rate": 4.257550204566039e-05, "loss": 6.8954, "step": 16626 }, { "epoch": 0.7010139764318991, "grad_norm": 1.7419017553329468, "learning_rate": 4.257472125925232e-05, "loss": 6.7373, "step": 16627 }, { "epoch": 0.7010561376140987, "grad_norm": 1.7925021648406982, "learning_rate": 4.257394047284425e-05, "loss": 6.725, "step": 16628 }, { "epoch": 0.7010982987962983, "grad_norm": 3.6312320232391357, "learning_rate": 4.257315968643618e-05, "loss": 7.5394, "step": 16629 }, { "epoch": 0.7011404599784978, "grad_norm": 2.650590658187866, "learning_rate": 4.257237890002811e-05, "loss": 7.5346, "step": 16630 }, { "epoch": 0.7011826211606973, "grad_norm": 1.9699780941009521, "learning_rate": 4.257159811362004e-05, "loss": 6.9176, "step": 16631 }, { "epoch": 0.7012247823428969, "grad_norm": 2.4713776111602783, "learning_rate": 4.257081732721197e-05, "loss": 7.8857, "step": 16632 }, { "epoch": 0.7012669435250964, "grad_norm": 1.8768388032913208, "learning_rate": 4.25700365408039e-05, "loss": 6.4926, "step": 16633 }, { "epoch": 0.701309104707296, "grad_norm": 1.9740228652954102, "learning_rate": 4.256925575439583e-05, "loss": 6.5262, "step": 16634 }, { "epoch": 0.7013512658894956, "grad_norm": 2.223641872406006, "learning_rate": 4.256847496798776e-05, "loss": 7.0009, "step": 16635 }, { "epoch": 0.701393427071695, "grad_norm": 3.0574920177459717, "learning_rate": 4.2567694181579685e-05, "loss": 7.0639, "step": 16636 }, { "epoch": 0.7014355882538946, "grad_norm": 1.3909176588058472, "learning_rate": 4.2566913395171624e-05, "loss": 6.845, "step": 16637 }, { "epoch": 0.7014777494360942, "grad_norm": 1.7870203256607056, "learning_rate": 4.256613260876355e-05, "loss": 6.884, "step": 16638 }, { "epoch": 0.7015199106182938, "grad_norm": 2.019341230392456, "learning_rate": 4.256535182235548e-05, "loss": 7.2056, "step": 16639 }, { "epoch": 0.7015620718004932, "grad_norm": 1.7998592853546143, "learning_rate": 4.256457103594741e-05, "loss": 7.0801, "step": 16640 }, { "epoch": 0.7016042329826928, "grad_norm": 1.8319660425186157, "learning_rate": 4.2563790249539335e-05, "loss": 6.6247, "step": 16641 }, { "epoch": 0.7016463941648924, "grad_norm": 3.279297351837158, "learning_rate": 4.256300946313127e-05, "loss": 7.4415, "step": 16642 }, { "epoch": 0.701688555347092, "grad_norm": 2.02660870552063, "learning_rate": 4.25622286767232e-05, "loss": 6.898, "step": 16643 }, { "epoch": 0.7017307165292915, "grad_norm": 2.023139238357544, "learning_rate": 4.256144789031513e-05, "loss": 6.9419, "step": 16644 }, { "epoch": 0.701772877711491, "grad_norm": 2.201917886734009, "learning_rate": 4.256066710390706e-05, "loss": 6.499, "step": 16645 }, { "epoch": 0.7018150388936906, "grad_norm": 3.119976043701172, "learning_rate": 4.2559886317498985e-05, "loss": 7.6318, "step": 16646 }, { "epoch": 0.7018572000758901, "grad_norm": 1.8411036729812622, "learning_rate": 4.2559105531090916e-05, "loss": 7.1218, "step": 16647 }, { "epoch": 0.7018993612580897, "grad_norm": 1.766602635383606, "learning_rate": 4.255832474468285e-05, "loss": 6.6322, "step": 16648 }, { "epoch": 0.7019415224402892, "grad_norm": 1.473812222480774, "learning_rate": 4.255754395827478e-05, "loss": 6.5059, "step": 16649 }, { "epoch": 0.7019836836224888, "grad_norm": 5.79202127456665, "learning_rate": 4.25567631718667e-05, "loss": 6.5054, "step": 16650 }, { "epoch": 0.7020258448046883, "grad_norm": 2.2970845699310303, "learning_rate": 4.2555982385458634e-05, "loss": 6.7795, "step": 16651 }, { "epoch": 0.7020680059868879, "grad_norm": 3.058915138244629, "learning_rate": 4.2555201599050565e-05, "loss": 7.6668, "step": 16652 }, { "epoch": 0.7021101671690875, "grad_norm": 3.264558792114258, "learning_rate": 4.2554420812642497e-05, "loss": 6.6423, "step": 16653 }, { "epoch": 0.7021523283512869, "grad_norm": 3.820054531097412, "learning_rate": 4.255364002623443e-05, "loss": 7.408, "step": 16654 }, { "epoch": 0.7021944895334865, "grad_norm": 2.2685649394989014, "learning_rate": 4.255285923982635e-05, "loss": 6.419, "step": 16655 }, { "epoch": 0.7022366507156861, "grad_norm": 2.326711654663086, "learning_rate": 4.255207845341829e-05, "loss": 7.02, "step": 16656 }, { "epoch": 0.7022788118978857, "grad_norm": 2.3775200843811035, "learning_rate": 4.2551297667010215e-05, "loss": 6.7857, "step": 16657 }, { "epoch": 0.7023209730800851, "grad_norm": 2.685715675354004, "learning_rate": 4.255051688060214e-05, "loss": 6.5773, "step": 16658 }, { "epoch": 0.7023631342622847, "grad_norm": 2.1592087745666504, "learning_rate": 4.254973609419408e-05, "loss": 7.0053, "step": 16659 }, { "epoch": 0.7024052954444843, "grad_norm": 1.6083523035049438, "learning_rate": 4.2548955307786e-05, "loss": 6.4378, "step": 16660 }, { "epoch": 0.7024474566266838, "grad_norm": 2.0389199256896973, "learning_rate": 4.254817452137793e-05, "loss": 6.56, "step": 16661 }, { "epoch": 0.7024896178088834, "grad_norm": 2.174506425857544, "learning_rate": 4.2547393734969864e-05, "loss": 6.604, "step": 16662 }, { "epoch": 0.7025317789910829, "grad_norm": 3.9378907680511475, "learning_rate": 4.2546612948561796e-05, "loss": 7.4172, "step": 16663 }, { "epoch": 0.7025739401732825, "grad_norm": 1.4224023818969727, "learning_rate": 4.254583216215373e-05, "loss": 6.8753, "step": 16664 }, { "epoch": 0.702616101355482, "grad_norm": 1.8433295488357544, "learning_rate": 4.254505137574565e-05, "loss": 6.3955, "step": 16665 }, { "epoch": 0.7026582625376816, "grad_norm": 1.8139764070510864, "learning_rate": 4.254427058933758e-05, "loss": 6.8102, "step": 16666 }, { "epoch": 0.7027004237198811, "grad_norm": 2.3006019592285156, "learning_rate": 4.2543489802929514e-05, "loss": 6.4812, "step": 16667 }, { "epoch": 0.7027425849020806, "grad_norm": 4.640927791595459, "learning_rate": 4.2542709016521445e-05, "loss": 7.3412, "step": 16668 }, { "epoch": 0.7027847460842802, "grad_norm": 1.9861191511154175, "learning_rate": 4.254192823011337e-05, "loss": 6.6459, "step": 16669 }, { "epoch": 0.7028269072664798, "grad_norm": 2.0159428119659424, "learning_rate": 4.25411474437053e-05, "loss": 6.3758, "step": 16670 }, { "epoch": 0.7028690684486792, "grad_norm": 1.9333559274673462, "learning_rate": 4.254036665729723e-05, "loss": 6.8895, "step": 16671 }, { "epoch": 0.7029112296308788, "grad_norm": 1.341880202293396, "learning_rate": 4.253958587088916e-05, "loss": 6.6938, "step": 16672 }, { "epoch": 0.7029533908130784, "grad_norm": 2.0694241523742676, "learning_rate": 4.2538805084481095e-05, "loss": 6.3243, "step": 16673 }, { "epoch": 0.702995551995278, "grad_norm": 2.095822811126709, "learning_rate": 4.253802429807302e-05, "loss": 6.9971, "step": 16674 }, { "epoch": 0.7030377131774775, "grad_norm": 2.8000380992889404, "learning_rate": 4.253724351166495e-05, "loss": 6.86, "step": 16675 }, { "epoch": 0.703079874359677, "grad_norm": 2.840275287628174, "learning_rate": 4.253646272525688e-05, "loss": 6.203, "step": 16676 }, { "epoch": 0.7031220355418766, "grad_norm": 1.8496795892715454, "learning_rate": 4.2535681938848806e-05, "loss": 6.8822, "step": 16677 }, { "epoch": 0.7031641967240762, "grad_norm": 1.781411051750183, "learning_rate": 4.2534901152440744e-05, "loss": 6.8355, "step": 16678 }, { "epoch": 0.7032063579062757, "grad_norm": 1.520855188369751, "learning_rate": 4.253412036603267e-05, "loss": 6.2385, "step": 16679 }, { "epoch": 0.7032485190884752, "grad_norm": 1.5889513492584229, "learning_rate": 4.25333395796246e-05, "loss": 6.4358, "step": 16680 }, { "epoch": 0.7032906802706748, "grad_norm": 3.1838738918304443, "learning_rate": 4.253255879321653e-05, "loss": 7.1599, "step": 16681 }, { "epoch": 0.7033328414528743, "grad_norm": 1.492509126663208, "learning_rate": 4.253177800680846e-05, "loss": 6.3272, "step": 16682 }, { "epoch": 0.7033750026350739, "grad_norm": 1.799532413482666, "learning_rate": 4.253099722040039e-05, "loss": 6.9346, "step": 16683 }, { "epoch": 0.7034171638172735, "grad_norm": 4.588702201843262, "learning_rate": 4.253021643399232e-05, "loss": 9.8908, "step": 16684 }, { "epoch": 0.703459324999473, "grad_norm": 2.243150234222412, "learning_rate": 4.252943564758425e-05, "loss": 6.5869, "step": 16685 }, { "epoch": 0.7035014861816725, "grad_norm": 3.1884987354278564, "learning_rate": 4.2528654861176174e-05, "loss": 6.5352, "step": 16686 }, { "epoch": 0.7035436473638721, "grad_norm": 2.0636045932769775, "learning_rate": 4.252787407476811e-05, "loss": 6.3988, "step": 16687 }, { "epoch": 0.7035858085460717, "grad_norm": 2.2562477588653564, "learning_rate": 4.2527093288360036e-05, "loss": 7.7911, "step": 16688 }, { "epoch": 0.7036279697282711, "grad_norm": 3.3088369369506836, "learning_rate": 4.252631250195197e-05, "loss": 7.343, "step": 16689 }, { "epoch": 0.7036701309104707, "grad_norm": 2.5696208477020264, "learning_rate": 4.25255317155439e-05, "loss": 7.055, "step": 16690 }, { "epoch": 0.7037122920926703, "grad_norm": 1.9831044673919678, "learning_rate": 4.2524750929135823e-05, "loss": 6.5484, "step": 16691 }, { "epoch": 0.7037544532748699, "grad_norm": 1.9786150455474854, "learning_rate": 4.252397014272776e-05, "loss": 6.597, "step": 16692 }, { "epoch": 0.7037966144570694, "grad_norm": 2.111985683441162, "learning_rate": 4.2523189356319686e-05, "loss": 6.4603, "step": 16693 }, { "epoch": 0.7038387756392689, "grad_norm": 1.3249317407608032, "learning_rate": 4.252240856991162e-05, "loss": 6.3967, "step": 16694 }, { "epoch": 0.7038809368214685, "grad_norm": 2.248775005340576, "learning_rate": 4.252162778350355e-05, "loss": 7.2009, "step": 16695 }, { "epoch": 0.703923098003668, "grad_norm": 3.6351616382598877, "learning_rate": 4.252084699709547e-05, "loss": 7.6304, "step": 16696 }, { "epoch": 0.7039652591858676, "grad_norm": 1.9500595331192017, "learning_rate": 4.2520066210687404e-05, "loss": 7.0517, "step": 16697 }, { "epoch": 0.7040074203680671, "grad_norm": 1.6105393171310425, "learning_rate": 4.2519285424279335e-05, "loss": 6.9835, "step": 16698 }, { "epoch": 0.7040495815502666, "grad_norm": 2.408141851425171, "learning_rate": 4.251850463787127e-05, "loss": 7.5082, "step": 16699 }, { "epoch": 0.7040917427324662, "grad_norm": 1.0767905712127686, "learning_rate": 4.25177238514632e-05, "loss": 6.6745, "step": 16700 }, { "epoch": 0.7041339039146658, "grad_norm": 1.57651948928833, "learning_rate": 4.251694306505513e-05, "loss": 6.3291, "step": 16701 }, { "epoch": 0.7041760650968654, "grad_norm": 1.6068017482757568, "learning_rate": 4.2516162278647054e-05, "loss": 6.76, "step": 16702 }, { "epoch": 0.7042182262790648, "grad_norm": 2.550579071044922, "learning_rate": 4.2515381492238985e-05, "loss": 7.5795, "step": 16703 }, { "epoch": 0.7042603874612644, "grad_norm": 4.143132209777832, "learning_rate": 4.2514600705830916e-05, "loss": 7.6201, "step": 16704 }, { "epoch": 0.704302548643464, "grad_norm": 1.583651065826416, "learning_rate": 4.251381991942284e-05, "loss": 6.3826, "step": 16705 }, { "epoch": 0.7043447098256636, "grad_norm": 2.394789218902588, "learning_rate": 4.251303913301478e-05, "loss": 7.4773, "step": 16706 }, { "epoch": 0.704386871007863, "grad_norm": 1.855086088180542, "learning_rate": 4.25122583466067e-05, "loss": 6.8483, "step": 16707 }, { "epoch": 0.7044290321900626, "grad_norm": 1.9955487251281738, "learning_rate": 4.2511477560198635e-05, "loss": 6.5898, "step": 16708 }, { "epoch": 0.7044711933722622, "grad_norm": 1.536744236946106, "learning_rate": 4.2510696773790566e-05, "loss": 7.3199, "step": 16709 }, { "epoch": 0.7045133545544617, "grad_norm": 2.049651861190796, "learning_rate": 4.250991598738249e-05, "loss": 7.106, "step": 16710 }, { "epoch": 0.7045555157366613, "grad_norm": 1.788514256477356, "learning_rate": 4.250913520097443e-05, "loss": 6.3823, "step": 16711 }, { "epoch": 0.7045976769188608, "grad_norm": 4.0785369873046875, "learning_rate": 4.250835441456635e-05, "loss": 7.278, "step": 16712 }, { "epoch": 0.7046398381010603, "grad_norm": 1.7377246618270874, "learning_rate": 4.2507573628158284e-05, "loss": 6.6265, "step": 16713 }, { "epoch": 0.7046819992832599, "grad_norm": 3.768446445465088, "learning_rate": 4.2506792841750215e-05, "loss": 7.6914, "step": 16714 }, { "epoch": 0.7047241604654595, "grad_norm": 3.406583547592163, "learning_rate": 4.250601205534214e-05, "loss": 7.0073, "step": 16715 }, { "epoch": 0.704766321647659, "grad_norm": 3.1512439250946045, "learning_rate": 4.250523126893407e-05, "loss": 7.5332, "step": 16716 }, { "epoch": 0.7048084828298585, "grad_norm": 1.9395525455474854, "learning_rate": 4.2504450482526e-05, "loss": 6.6358, "step": 16717 }, { "epoch": 0.7048506440120581, "grad_norm": 2.450317859649658, "learning_rate": 4.2503669696117934e-05, "loss": 6.7833, "step": 16718 }, { "epoch": 0.7048928051942577, "grad_norm": 1.9994224309921265, "learning_rate": 4.250288890970986e-05, "loss": 6.6707, "step": 16719 }, { "epoch": 0.7049349663764573, "grad_norm": 3.1633527278900146, "learning_rate": 4.250210812330179e-05, "loss": 7.5922, "step": 16720 }, { "epoch": 0.7049771275586567, "grad_norm": 2.6264984607696533, "learning_rate": 4.250132733689372e-05, "loss": 7.3347, "step": 16721 }, { "epoch": 0.7050192887408563, "grad_norm": 1.9826740026474, "learning_rate": 4.250054655048565e-05, "loss": 6.8323, "step": 16722 }, { "epoch": 0.7050614499230559, "grad_norm": 1.5283123254776, "learning_rate": 4.249976576407758e-05, "loss": 6.3638, "step": 16723 }, { "epoch": 0.7051036111052554, "grad_norm": 1.6920974254608154, "learning_rate": 4.249898497766951e-05, "loss": 6.3483, "step": 16724 }, { "epoch": 0.7051457722874549, "grad_norm": 1.92130446434021, "learning_rate": 4.2498204191261446e-05, "loss": 6.8031, "step": 16725 }, { "epoch": 0.7051879334696545, "grad_norm": 1.932993769645691, "learning_rate": 4.249742340485337e-05, "loss": 6.9205, "step": 16726 }, { "epoch": 0.705230094651854, "grad_norm": 5.553319454193115, "learning_rate": 4.24966426184453e-05, "loss": 8.8, "step": 16727 }, { "epoch": 0.7052722558340536, "grad_norm": 1.4305039644241333, "learning_rate": 4.249586183203723e-05, "loss": 6.5441, "step": 16728 }, { "epoch": 0.7053144170162531, "grad_norm": 1.531320571899414, "learning_rate": 4.249508104562916e-05, "loss": 6.4959, "step": 16729 }, { "epoch": 0.7053565781984527, "grad_norm": 1.787736415863037, "learning_rate": 4.249430025922109e-05, "loss": 7.0425, "step": 16730 }, { "epoch": 0.7053987393806522, "grad_norm": 2.3210904598236084, "learning_rate": 4.249351947281302e-05, "loss": 6.5945, "step": 16731 }, { "epoch": 0.7054409005628518, "grad_norm": 2.3505468368530273, "learning_rate": 4.249273868640495e-05, "loss": 6.7207, "step": 16732 }, { "epoch": 0.7054830617450514, "grad_norm": 1.4447861909866333, "learning_rate": 4.249195789999688e-05, "loss": 6.4655, "step": 16733 }, { "epoch": 0.7055252229272508, "grad_norm": 2.7110447883605957, "learning_rate": 4.249117711358881e-05, "loss": 7.0807, "step": 16734 }, { "epoch": 0.7055673841094504, "grad_norm": 2.349126100540161, "learning_rate": 4.249039632718074e-05, "loss": 6.4808, "step": 16735 }, { "epoch": 0.70560954529165, "grad_norm": 2.908090114593506, "learning_rate": 4.248961554077267e-05, "loss": 7.5526, "step": 16736 }, { "epoch": 0.7056517064738496, "grad_norm": 2.910999298095703, "learning_rate": 4.24888347543646e-05, "loss": 6.8884, "step": 16737 }, { "epoch": 0.705693867656049, "grad_norm": 2.1801693439483643, "learning_rate": 4.2488053967956525e-05, "loss": 6.8169, "step": 16738 }, { "epoch": 0.7057360288382486, "grad_norm": 2.277474880218506, "learning_rate": 4.2487273181548456e-05, "loss": 6.2431, "step": 16739 }, { "epoch": 0.7057781900204482, "grad_norm": 1.8892512321472168, "learning_rate": 4.248649239514039e-05, "loss": 6.9123, "step": 16740 }, { "epoch": 0.7058203512026477, "grad_norm": 1.7481837272644043, "learning_rate": 4.248571160873231e-05, "loss": 6.6616, "step": 16741 }, { "epoch": 0.7058625123848473, "grad_norm": 1.4899568557739258, "learning_rate": 4.248493082232425e-05, "loss": 6.7875, "step": 16742 }, { "epoch": 0.7059046735670468, "grad_norm": 2.8943567276000977, "learning_rate": 4.2484150035916174e-05, "loss": 6.3886, "step": 16743 }, { "epoch": 0.7059468347492464, "grad_norm": 2.982520580291748, "learning_rate": 4.248336924950811e-05, "loss": 6.7086, "step": 16744 }, { "epoch": 0.7059889959314459, "grad_norm": 3.1399903297424316, "learning_rate": 4.248258846310004e-05, "loss": 7.3211, "step": 16745 }, { "epoch": 0.7060311571136455, "grad_norm": 4.553555965423584, "learning_rate": 4.248180767669197e-05, "loss": 7.7984, "step": 16746 }, { "epoch": 0.706073318295845, "grad_norm": 1.2278293371200562, "learning_rate": 4.24810268902839e-05, "loss": 6.5904, "step": 16747 }, { "epoch": 0.7061154794780445, "grad_norm": 3.0912539958953857, "learning_rate": 4.2480246103875824e-05, "loss": 7.3279, "step": 16748 }, { "epoch": 0.7061576406602441, "grad_norm": 1.7173926830291748, "learning_rate": 4.2479465317467755e-05, "loss": 6.5981, "step": 16749 }, { "epoch": 0.7061998018424437, "grad_norm": 1.7317978143692017, "learning_rate": 4.2478684531059686e-05, "loss": 7.8067, "step": 16750 }, { "epoch": 0.7062419630246433, "grad_norm": 3.127070903778076, "learning_rate": 4.247790374465162e-05, "loss": 6.2474, "step": 16751 }, { "epoch": 0.7062841242068427, "grad_norm": 1.8807882070541382, "learning_rate": 4.247712295824354e-05, "loss": 6.5055, "step": 16752 }, { "epoch": 0.7063262853890423, "grad_norm": 2.437505006790161, "learning_rate": 4.2476342171835473e-05, "loss": 6.7371, "step": 16753 }, { "epoch": 0.7063684465712419, "grad_norm": 1.7097529172897339, "learning_rate": 4.2475561385427405e-05, "loss": 6.8967, "step": 16754 }, { "epoch": 0.7064106077534414, "grad_norm": 1.2178070545196533, "learning_rate": 4.2474780599019336e-05, "loss": 6.5468, "step": 16755 }, { "epoch": 0.7064527689356409, "grad_norm": 2.8232011795043945, "learning_rate": 4.247399981261127e-05, "loss": 6.462, "step": 16756 }, { "epoch": 0.7064949301178405, "grad_norm": 1.5133957862854004, "learning_rate": 4.247321902620319e-05, "loss": 6.8214, "step": 16757 }, { "epoch": 0.70653709130004, "grad_norm": 3.730278253555298, "learning_rate": 4.247243823979512e-05, "loss": 7.207, "step": 16758 }, { "epoch": 0.7065792524822396, "grad_norm": 3.003363847732544, "learning_rate": 4.2471657453387054e-05, "loss": 7.3304, "step": 16759 }, { "epoch": 0.7066214136644392, "grad_norm": 2.169865369796753, "learning_rate": 4.247087666697898e-05, "loss": 7.2863, "step": 16760 }, { "epoch": 0.7066635748466387, "grad_norm": 3.0752248764038086, "learning_rate": 4.247009588057092e-05, "loss": 6.9119, "step": 16761 }, { "epoch": 0.7067057360288382, "grad_norm": 1.8852275609970093, "learning_rate": 4.246931509416284e-05, "loss": 6.6293, "step": 16762 }, { "epoch": 0.7067478972110378, "grad_norm": 4.056809902191162, "learning_rate": 4.246853430775477e-05, "loss": 7.7498, "step": 16763 }, { "epoch": 0.7067900583932374, "grad_norm": 1.5705105066299438, "learning_rate": 4.2467753521346704e-05, "loss": 6.9871, "step": 16764 }, { "epoch": 0.7068322195754368, "grad_norm": 2.3348312377929688, "learning_rate": 4.246697273493863e-05, "loss": 7.1927, "step": 16765 }, { "epoch": 0.7068743807576364, "grad_norm": 2.2558047771453857, "learning_rate": 4.246619194853056e-05, "loss": 7.7043, "step": 16766 }, { "epoch": 0.706916541939836, "grad_norm": 4.192974090576172, "learning_rate": 4.246541116212249e-05, "loss": 7.3914, "step": 16767 }, { "epoch": 0.7069587031220356, "grad_norm": 3.007800817489624, "learning_rate": 4.246463037571442e-05, "loss": 7.2979, "step": 16768 }, { "epoch": 0.7070008643042351, "grad_norm": 1.900254726409912, "learning_rate": 4.246384958930635e-05, "loss": 6.3936, "step": 16769 }, { "epoch": 0.7070430254864346, "grad_norm": 1.6212769746780396, "learning_rate": 4.2463068802898285e-05, "loss": 6.5242, "step": 16770 }, { "epoch": 0.7070851866686342, "grad_norm": 1.8474771976470947, "learning_rate": 4.246228801649021e-05, "loss": 6.672, "step": 16771 }, { "epoch": 0.7071273478508338, "grad_norm": 1.4911431074142456, "learning_rate": 4.246150723008214e-05, "loss": 6.6773, "step": 16772 }, { "epoch": 0.7071695090330333, "grad_norm": 1.5991482734680176, "learning_rate": 4.246072644367407e-05, "loss": 7.1826, "step": 16773 }, { "epoch": 0.7072116702152328, "grad_norm": 4.665774345397949, "learning_rate": 4.2459945657265996e-05, "loss": 7.5134, "step": 16774 }, { "epoch": 0.7072538313974324, "grad_norm": 2.2482187747955322, "learning_rate": 4.2459164870857934e-05, "loss": 7.0118, "step": 16775 }, { "epoch": 0.7072959925796319, "grad_norm": 1.9794089794158936, "learning_rate": 4.245838408444986e-05, "loss": 6.4164, "step": 16776 }, { "epoch": 0.7073381537618315, "grad_norm": 1.764245629310608, "learning_rate": 4.245760329804179e-05, "loss": 6.4177, "step": 16777 }, { "epoch": 0.7073803149440311, "grad_norm": 3.997870683670044, "learning_rate": 4.245682251163372e-05, "loss": 7.5456, "step": 16778 }, { "epoch": 0.7074224761262305, "grad_norm": 2.1958889961242676, "learning_rate": 4.2456041725225646e-05, "loss": 7.0213, "step": 16779 }, { "epoch": 0.7074646373084301, "grad_norm": 1.4143015146255493, "learning_rate": 4.2455260938817584e-05, "loss": 6.5909, "step": 16780 }, { "epoch": 0.7075067984906297, "grad_norm": 1.8296396732330322, "learning_rate": 4.245448015240951e-05, "loss": 6.308, "step": 16781 }, { "epoch": 0.7075489596728293, "grad_norm": 1.694947361946106, "learning_rate": 4.245369936600144e-05, "loss": 6.392, "step": 16782 }, { "epoch": 0.7075911208550287, "grad_norm": 1.4628809690475464, "learning_rate": 4.245291857959337e-05, "loss": 6.7663, "step": 16783 }, { "epoch": 0.7076332820372283, "grad_norm": 3.5699212551116943, "learning_rate": 4.2452137793185295e-05, "loss": 7.8253, "step": 16784 }, { "epoch": 0.7076754432194279, "grad_norm": 1.7558034658432007, "learning_rate": 4.2451357006777226e-05, "loss": 6.6453, "step": 16785 }, { "epoch": 0.7077176044016275, "grad_norm": 2.185244560241699, "learning_rate": 4.245057622036916e-05, "loss": 7.1631, "step": 16786 }, { "epoch": 0.7077597655838269, "grad_norm": 1.8099803924560547, "learning_rate": 4.244979543396109e-05, "loss": 6.409, "step": 16787 }, { "epoch": 0.7078019267660265, "grad_norm": 1.5892307758331299, "learning_rate": 4.244901464755301e-05, "loss": 6.4983, "step": 16788 }, { "epoch": 0.7078440879482261, "grad_norm": 1.2603598833084106, "learning_rate": 4.244823386114495e-05, "loss": 6.5265, "step": 16789 }, { "epoch": 0.7078862491304256, "grad_norm": 1.6457324028015137, "learning_rate": 4.2447453074736876e-05, "loss": 6.4168, "step": 16790 }, { "epoch": 0.7079284103126252, "grad_norm": 1.247078537940979, "learning_rate": 4.244667228832881e-05, "loss": 6.4743, "step": 16791 }, { "epoch": 0.7079705714948247, "grad_norm": 1.6874510049819946, "learning_rate": 4.244589150192074e-05, "loss": 6.9774, "step": 16792 }, { "epoch": 0.7080127326770242, "grad_norm": 3.7674245834350586, "learning_rate": 4.244511071551266e-05, "loss": 7.5965, "step": 16793 }, { "epoch": 0.7080548938592238, "grad_norm": 3.2267234325408936, "learning_rate": 4.24443299291046e-05, "loss": 6.9018, "step": 16794 }, { "epoch": 0.7080970550414234, "grad_norm": 1.9760401248931885, "learning_rate": 4.2443549142696525e-05, "loss": 7.2527, "step": 16795 }, { "epoch": 0.7081392162236229, "grad_norm": 1.7322840690612793, "learning_rate": 4.244276835628846e-05, "loss": 7.4317, "step": 16796 }, { "epoch": 0.7081813774058224, "grad_norm": 2.3413124084472656, "learning_rate": 4.244198756988039e-05, "loss": 6.4663, "step": 16797 }, { "epoch": 0.708223538588022, "grad_norm": 3.084442377090454, "learning_rate": 4.244120678347231e-05, "loss": 8.7332, "step": 16798 }, { "epoch": 0.7082656997702216, "grad_norm": 1.485307216644287, "learning_rate": 4.2440425997064244e-05, "loss": 7.0574, "step": 16799 }, { "epoch": 0.7083078609524212, "grad_norm": 1.3781706094741821, "learning_rate": 4.2439645210656175e-05, "loss": 6.4499, "step": 16800 }, { "epoch": 0.7083500221346206, "grad_norm": 1.1945899724960327, "learning_rate": 4.2438864424248106e-05, "loss": 6.5936, "step": 16801 }, { "epoch": 0.7083921833168202, "grad_norm": 2.4791927337646484, "learning_rate": 4.243808363784004e-05, "loss": 7.0924, "step": 16802 }, { "epoch": 0.7084343444990198, "grad_norm": 1.6732072830200195, "learning_rate": 4.243730285143196e-05, "loss": 6.4769, "step": 16803 }, { "epoch": 0.7084765056812193, "grad_norm": 1.1476025581359863, "learning_rate": 4.243652206502389e-05, "loss": 6.41, "step": 16804 }, { "epoch": 0.7085186668634188, "grad_norm": 1.5429304838180542, "learning_rate": 4.2435741278615824e-05, "loss": 6.851, "step": 16805 }, { "epoch": 0.7085608280456184, "grad_norm": 2.015913486480713, "learning_rate": 4.2434960492207756e-05, "loss": 6.8445, "step": 16806 }, { "epoch": 0.708602989227818, "grad_norm": 2.40995192527771, "learning_rate": 4.243417970579968e-05, "loss": 6.3948, "step": 16807 }, { "epoch": 0.7086451504100175, "grad_norm": 3.507235288619995, "learning_rate": 4.243339891939161e-05, "loss": 7.4323, "step": 16808 }, { "epoch": 0.7086873115922171, "grad_norm": 2.0047860145568848, "learning_rate": 4.243261813298354e-05, "loss": 6.6088, "step": 16809 }, { "epoch": 0.7087294727744166, "grad_norm": 1.3644016981124878, "learning_rate": 4.243183734657547e-05, "loss": 6.5064, "step": 16810 }, { "epoch": 0.7087716339566161, "grad_norm": 1.8844319581985474, "learning_rate": 4.2431056560167405e-05, "loss": 6.5648, "step": 16811 }, { "epoch": 0.7088137951388157, "grad_norm": 1.6382662057876587, "learning_rate": 4.243027577375933e-05, "loss": 6.8416, "step": 16812 }, { "epoch": 0.7088559563210153, "grad_norm": 1.9291702508926392, "learning_rate": 4.242949498735127e-05, "loss": 6.9391, "step": 16813 }, { "epoch": 0.7088981175032147, "grad_norm": 3.0149295330047607, "learning_rate": 4.242871420094319e-05, "loss": 7.6308, "step": 16814 }, { "epoch": 0.7089402786854143, "grad_norm": 1.5573234558105469, "learning_rate": 4.2427933414535123e-05, "loss": 6.6708, "step": 16815 }, { "epoch": 0.7089824398676139, "grad_norm": 2.387208938598633, "learning_rate": 4.2427152628127055e-05, "loss": 7.342, "step": 16816 }, { "epoch": 0.7090246010498135, "grad_norm": 2.2509024143218994, "learning_rate": 4.242637184171898e-05, "loss": 6.5202, "step": 16817 }, { "epoch": 0.709066762232013, "grad_norm": 1.393019676208496, "learning_rate": 4.242559105531091e-05, "loss": 6.8348, "step": 16818 }, { "epoch": 0.7091089234142125, "grad_norm": 1.403996467590332, "learning_rate": 4.242481026890284e-05, "loss": 6.4484, "step": 16819 }, { "epoch": 0.7091510845964121, "grad_norm": 2.3592138290405273, "learning_rate": 4.242402948249477e-05, "loss": 6.801, "step": 16820 }, { "epoch": 0.7091932457786116, "grad_norm": 3.5709869861602783, "learning_rate": 4.24232486960867e-05, "loss": 7.5195, "step": 16821 }, { "epoch": 0.7092354069608112, "grad_norm": 1.5999174118041992, "learning_rate": 4.242246790967863e-05, "loss": 6.8448, "step": 16822 }, { "epoch": 0.7092775681430107, "grad_norm": 3.0889809131622314, "learning_rate": 4.242168712327056e-05, "loss": 7.4606, "step": 16823 }, { "epoch": 0.7093197293252103, "grad_norm": 2.572741985321045, "learning_rate": 4.242090633686249e-05, "loss": 6.3485, "step": 16824 }, { "epoch": 0.7093618905074098, "grad_norm": 1.9415541887283325, "learning_rate": 4.242012555045442e-05, "loss": 6.9425, "step": 16825 }, { "epoch": 0.7094040516896094, "grad_norm": 1.3223705291748047, "learning_rate": 4.241934476404635e-05, "loss": 6.6213, "step": 16826 }, { "epoch": 0.709446212871809, "grad_norm": 1.0199501514434814, "learning_rate": 4.241856397763828e-05, "loss": 6.6826, "step": 16827 }, { "epoch": 0.7094883740540084, "grad_norm": 1.0668468475341797, "learning_rate": 4.241778319123021e-05, "loss": 6.6752, "step": 16828 }, { "epoch": 0.709530535236208, "grad_norm": 2.387211322784424, "learning_rate": 4.2417002404822134e-05, "loss": 7.1161, "step": 16829 }, { "epoch": 0.7095726964184076, "grad_norm": 1.2019668817520142, "learning_rate": 4.241622161841407e-05, "loss": 6.3527, "step": 16830 }, { "epoch": 0.7096148576006072, "grad_norm": 1.2450989484786987, "learning_rate": 4.2415440832005997e-05, "loss": 6.5913, "step": 16831 }, { "epoch": 0.7096570187828066, "grad_norm": 1.5650579929351807, "learning_rate": 4.241466004559793e-05, "loss": 6.4057, "step": 16832 }, { "epoch": 0.7096991799650062, "grad_norm": 1.593221664428711, "learning_rate": 4.241387925918986e-05, "loss": 7.1911, "step": 16833 }, { "epoch": 0.7097413411472058, "grad_norm": 1.1227823495864868, "learning_rate": 4.241309847278179e-05, "loss": 6.6816, "step": 16834 }, { "epoch": 0.7097835023294053, "grad_norm": 1.5843734741210938, "learning_rate": 4.241231768637372e-05, "loss": 6.4785, "step": 16835 }, { "epoch": 0.7098256635116049, "grad_norm": 1.4346469640731812, "learning_rate": 4.2411536899965646e-05, "loss": 6.5309, "step": 16836 }, { "epoch": 0.7098678246938044, "grad_norm": 1.3616058826446533, "learning_rate": 4.241075611355758e-05, "loss": 7.0764, "step": 16837 }, { "epoch": 0.709909985876004, "grad_norm": 2.22833514213562, "learning_rate": 4.240997532714951e-05, "loss": 6.9956, "step": 16838 }, { "epoch": 0.7099521470582035, "grad_norm": 1.484820008277893, "learning_rate": 4.240919454074144e-05, "loss": 6.9855, "step": 16839 }, { "epoch": 0.7099943082404031, "grad_norm": 1.3909891843795776, "learning_rate": 4.2408413754333364e-05, "loss": 6.7103, "step": 16840 }, { "epoch": 0.7100364694226026, "grad_norm": 1.2032368183135986, "learning_rate": 4.2407632967925296e-05, "loss": 7.0896, "step": 16841 }, { "epoch": 0.7100786306048021, "grad_norm": 1.3794384002685547, "learning_rate": 4.240685218151723e-05, "loss": 6.3614, "step": 16842 }, { "epoch": 0.7101207917870017, "grad_norm": 2.891996383666992, "learning_rate": 4.240607139510915e-05, "loss": 7.6804, "step": 16843 }, { "epoch": 0.7101629529692013, "grad_norm": 1.52028489112854, "learning_rate": 4.240529060870109e-05, "loss": 6.5773, "step": 16844 }, { "epoch": 0.7102051141514008, "grad_norm": 1.0561201572418213, "learning_rate": 4.2404509822293014e-05, "loss": 6.6866, "step": 16845 }, { "epoch": 0.7102472753336003, "grad_norm": 3.451154947280884, "learning_rate": 4.2403729035884945e-05, "loss": 7.7942, "step": 16846 }, { "epoch": 0.7102894365157999, "grad_norm": 2.5650060176849365, "learning_rate": 4.2402948249476876e-05, "loss": 6.4556, "step": 16847 }, { "epoch": 0.7103315976979995, "grad_norm": 1.3262968063354492, "learning_rate": 4.24021674630688e-05, "loss": 7.0879, "step": 16848 }, { "epoch": 0.710373758880199, "grad_norm": 2.1331992149353027, "learning_rate": 4.240138667666074e-05, "loss": 6.5628, "step": 16849 }, { "epoch": 0.7104159200623985, "grad_norm": 2.07696795463562, "learning_rate": 4.240060589025266e-05, "loss": 6.438, "step": 16850 }, { "epoch": 0.7104580812445981, "grad_norm": 1.6344629526138306, "learning_rate": 4.2399825103844595e-05, "loss": 6.5141, "step": 16851 }, { "epoch": 0.7105002424267977, "grad_norm": 4.718578338623047, "learning_rate": 4.2399044317436526e-05, "loss": 7.8108, "step": 16852 }, { "epoch": 0.7105424036089972, "grad_norm": 2.3901662826538086, "learning_rate": 4.239826353102845e-05, "loss": 7.2203, "step": 16853 }, { "epoch": 0.7105845647911967, "grad_norm": 2.8579671382904053, "learning_rate": 4.239748274462038e-05, "loss": 7.5002, "step": 16854 }, { "epoch": 0.7106267259733963, "grad_norm": 1.357926845550537, "learning_rate": 4.239670195821231e-05, "loss": 6.7634, "step": 16855 }, { "epoch": 0.7106688871555958, "grad_norm": 1.8251041173934937, "learning_rate": 4.2395921171804244e-05, "loss": 6.5538, "step": 16856 }, { "epoch": 0.7107110483377954, "grad_norm": 1.2378385066986084, "learning_rate": 4.239514038539617e-05, "loss": 6.4219, "step": 16857 }, { "epoch": 0.710753209519995, "grad_norm": 1.9329369068145752, "learning_rate": 4.239435959898811e-05, "loss": 6.5488, "step": 16858 }, { "epoch": 0.7107953707021945, "grad_norm": 1.0255039930343628, "learning_rate": 4.239357881258003e-05, "loss": 6.5291, "step": 16859 }, { "epoch": 0.710837531884394, "grad_norm": 2.2766706943511963, "learning_rate": 4.239279802617196e-05, "loss": 7.0505, "step": 16860 }, { "epoch": 0.7108796930665936, "grad_norm": 1.6690245866775513, "learning_rate": 4.2392017239763894e-05, "loss": 7.1207, "step": 16861 }, { "epoch": 0.7109218542487932, "grad_norm": 1.5213301181793213, "learning_rate": 4.239123645335582e-05, "loss": 7.0244, "step": 16862 }, { "epoch": 0.7109640154309926, "grad_norm": 3.112567663192749, "learning_rate": 4.2390455666947756e-05, "loss": 7.333, "step": 16863 }, { "epoch": 0.7110061766131922, "grad_norm": 1.5966509580612183, "learning_rate": 4.238967488053968e-05, "loss": 6.8853, "step": 16864 }, { "epoch": 0.7110483377953918, "grad_norm": 5.105603218078613, "learning_rate": 4.238889409413161e-05, "loss": 10.9286, "step": 16865 }, { "epoch": 0.7110904989775914, "grad_norm": 2.067169666290283, "learning_rate": 4.238811330772354e-05, "loss": 6.3936, "step": 16866 }, { "epoch": 0.7111326601597909, "grad_norm": 2.5057382583618164, "learning_rate": 4.238733252131547e-05, "loss": 6.5953, "step": 16867 }, { "epoch": 0.7111748213419904, "grad_norm": 1.5734535455703735, "learning_rate": 4.23865517349074e-05, "loss": 6.6637, "step": 16868 }, { "epoch": 0.71121698252419, "grad_norm": 3.593372106552124, "learning_rate": 4.238577094849933e-05, "loss": 6.892, "step": 16869 }, { "epoch": 0.7112591437063895, "grad_norm": 2.8733906745910645, "learning_rate": 4.238499016209126e-05, "loss": 7.0469, "step": 16870 }, { "epoch": 0.7113013048885891, "grad_norm": 3.8981122970581055, "learning_rate": 4.238420937568319e-05, "loss": 7.7422, "step": 16871 }, { "epoch": 0.7113434660707886, "grad_norm": 2.491339921951294, "learning_rate": 4.238342858927512e-05, "loss": 8.2436, "step": 16872 }, { "epoch": 0.7113856272529882, "grad_norm": 3.4989094734191895, "learning_rate": 4.238264780286705e-05, "loss": 7.2005, "step": 16873 }, { "epoch": 0.7114277884351877, "grad_norm": 2.1114373207092285, "learning_rate": 4.238186701645898e-05, "loss": 7.8389, "step": 16874 }, { "epoch": 0.7114699496173873, "grad_norm": 3.2380056381225586, "learning_rate": 4.238108623005091e-05, "loss": 7.6656, "step": 16875 }, { "epoch": 0.7115121107995869, "grad_norm": 2.101353645324707, "learning_rate": 4.2380305443642835e-05, "loss": 6.5808, "step": 16876 }, { "epoch": 0.7115542719817863, "grad_norm": 1.4490898847579956, "learning_rate": 4.2379524657234774e-05, "loss": 6.7294, "step": 16877 }, { "epoch": 0.7115964331639859, "grad_norm": 2.298797845840454, "learning_rate": 4.23787438708267e-05, "loss": 6.6062, "step": 16878 }, { "epoch": 0.7116385943461855, "grad_norm": 4.423199653625488, "learning_rate": 4.237796308441863e-05, "loss": 9.1604, "step": 16879 }, { "epoch": 0.7116807555283851, "grad_norm": 2.52896785736084, "learning_rate": 4.237718229801056e-05, "loss": 7.031, "step": 16880 }, { "epoch": 0.7117229167105845, "grad_norm": 1.5660520792007446, "learning_rate": 4.2376401511602485e-05, "loss": 6.459, "step": 16881 }, { "epoch": 0.7117650778927841, "grad_norm": 1.7342650890350342, "learning_rate": 4.237562072519442e-05, "loss": 6.9486, "step": 16882 }, { "epoch": 0.7118072390749837, "grad_norm": 2.834773540496826, "learning_rate": 4.237483993878635e-05, "loss": 7.5274, "step": 16883 }, { "epoch": 0.7118494002571832, "grad_norm": 4.004147529602051, "learning_rate": 4.237405915237828e-05, "loss": 7.2067, "step": 16884 }, { "epoch": 0.7118915614393828, "grad_norm": 2.4699249267578125, "learning_rate": 4.237327836597021e-05, "loss": 6.9259, "step": 16885 }, { "epoch": 0.7119337226215823, "grad_norm": 2.0674450397491455, "learning_rate": 4.2372497579562135e-05, "loss": 6.7446, "step": 16886 }, { "epoch": 0.7119758838037819, "grad_norm": 1.6562578678131104, "learning_rate": 4.2371716793154066e-05, "loss": 6.9299, "step": 16887 }, { "epoch": 0.7120180449859814, "grad_norm": 1.9274815320968628, "learning_rate": 4.2370936006746e-05, "loss": 6.7394, "step": 16888 }, { "epoch": 0.712060206168181, "grad_norm": 4.527523517608643, "learning_rate": 4.237015522033793e-05, "loss": 7.7485, "step": 16889 }, { "epoch": 0.7121023673503805, "grad_norm": 3.3810808658599854, "learning_rate": 4.236937443392985e-05, "loss": 7.4065, "step": 16890 }, { "epoch": 0.71214452853258, "grad_norm": 1.6764978170394897, "learning_rate": 4.2368593647521784e-05, "loss": 6.3951, "step": 16891 }, { "epoch": 0.7121866897147796, "grad_norm": 3.7120449542999268, "learning_rate": 4.2367812861113715e-05, "loss": 6.4849, "step": 16892 }, { "epoch": 0.7122288508969792, "grad_norm": 2.4784040451049805, "learning_rate": 4.2367032074705647e-05, "loss": 6.9989, "step": 16893 }, { "epoch": 0.7122710120791786, "grad_norm": 1.513633370399475, "learning_rate": 4.236625128829758e-05, "loss": 7.4913, "step": 16894 }, { "epoch": 0.7123131732613782, "grad_norm": 2.1045966148376465, "learning_rate": 4.23654705018895e-05, "loss": 6.6096, "step": 16895 }, { "epoch": 0.7123553344435778, "grad_norm": 2.597357749938965, "learning_rate": 4.236468971548144e-05, "loss": 6.8414, "step": 16896 }, { "epoch": 0.7123974956257774, "grad_norm": 1.8388099670410156, "learning_rate": 4.2363908929073365e-05, "loss": 7.2001, "step": 16897 }, { "epoch": 0.7124396568079769, "grad_norm": 3.412792921066284, "learning_rate": 4.236312814266529e-05, "loss": 7.694, "step": 16898 }, { "epoch": 0.7124818179901764, "grad_norm": 2.313133478164673, "learning_rate": 4.236234735625723e-05, "loss": 6.8807, "step": 16899 }, { "epoch": 0.712523979172376, "grad_norm": 3.331570863723755, "learning_rate": 4.236156656984915e-05, "loss": 7.729, "step": 16900 }, { "epoch": 0.7125661403545756, "grad_norm": 1.63130521774292, "learning_rate": 4.236078578344108e-05, "loss": 6.8212, "step": 16901 }, { "epoch": 0.7126083015367751, "grad_norm": 2.585109233856201, "learning_rate": 4.2360004997033014e-05, "loss": 7.4701, "step": 16902 }, { "epoch": 0.7126504627189746, "grad_norm": 2.4634604454040527, "learning_rate": 4.2359224210624946e-05, "loss": 8.6892, "step": 16903 }, { "epoch": 0.7126926239011742, "grad_norm": 3.2504684925079346, "learning_rate": 4.235844342421688e-05, "loss": 6.6017, "step": 16904 }, { "epoch": 0.7127347850833737, "grad_norm": 1.8254117965698242, "learning_rate": 4.23576626378088e-05, "loss": 7.061, "step": 16905 }, { "epoch": 0.7127769462655733, "grad_norm": 2.1941421031951904, "learning_rate": 4.235688185140073e-05, "loss": 6.4536, "step": 16906 }, { "epoch": 0.7128191074477729, "grad_norm": 2.8945741653442383, "learning_rate": 4.2356101064992664e-05, "loss": 6.8287, "step": 16907 }, { "epoch": 0.7128612686299723, "grad_norm": 2.6359784603118896, "learning_rate": 4.2355320278584595e-05, "loss": 7.3806, "step": 16908 }, { "epoch": 0.7129034298121719, "grad_norm": 4.133113384246826, "learning_rate": 4.235453949217652e-05, "loss": 7.2787, "step": 16909 }, { "epoch": 0.7129455909943715, "grad_norm": 2.189072370529175, "learning_rate": 4.235375870576845e-05, "loss": 6.9064, "step": 16910 }, { "epoch": 0.7129877521765711, "grad_norm": 1.627928376197815, "learning_rate": 4.235297791936038e-05, "loss": 6.4806, "step": 16911 }, { "epoch": 0.7130299133587705, "grad_norm": 1.3534921407699585, "learning_rate": 4.235219713295231e-05, "loss": 6.5509, "step": 16912 }, { "epoch": 0.7130720745409701, "grad_norm": 2.2244157791137695, "learning_rate": 4.2351416346544245e-05, "loss": 6.4329, "step": 16913 }, { "epoch": 0.7131142357231697, "grad_norm": 1.7091400623321533, "learning_rate": 4.235063556013617e-05, "loss": 6.5201, "step": 16914 }, { "epoch": 0.7131563969053692, "grad_norm": 1.7060129642486572, "learning_rate": 4.23498547737281e-05, "loss": 6.5235, "step": 16915 }, { "epoch": 0.7131985580875688, "grad_norm": 4.628494739532471, "learning_rate": 4.234907398732003e-05, "loss": 8.7866, "step": 16916 }, { "epoch": 0.7132407192697683, "grad_norm": 4.0883917808532715, "learning_rate": 4.2348293200911956e-05, "loss": 7.2738, "step": 16917 }, { "epoch": 0.7132828804519679, "grad_norm": 2.6904523372650146, "learning_rate": 4.2347512414503894e-05, "loss": 6.5874, "step": 16918 }, { "epoch": 0.7133250416341674, "grad_norm": 1.2330920696258545, "learning_rate": 4.234673162809582e-05, "loss": 6.3874, "step": 16919 }, { "epoch": 0.713367202816367, "grad_norm": 2.400514841079712, "learning_rate": 4.234595084168775e-05, "loss": 7.3177, "step": 16920 }, { "epoch": 0.7134093639985665, "grad_norm": 2.2573330402374268, "learning_rate": 4.234517005527968e-05, "loss": 7.0137, "step": 16921 }, { "epoch": 0.713451525180766, "grad_norm": 2.5203309059143066, "learning_rate": 4.234438926887161e-05, "loss": 6.9914, "step": 16922 }, { "epoch": 0.7134936863629656, "grad_norm": 3.633831024169922, "learning_rate": 4.234360848246354e-05, "loss": 7.7976, "step": 16923 }, { "epoch": 0.7135358475451652, "grad_norm": 1.8881173133850098, "learning_rate": 4.234282769605547e-05, "loss": 6.7565, "step": 16924 }, { "epoch": 0.7135780087273648, "grad_norm": 2.060248851776123, "learning_rate": 4.23420469096474e-05, "loss": 7.5603, "step": 16925 }, { "epoch": 0.7136201699095642, "grad_norm": 1.6803292036056519, "learning_rate": 4.2341266123239324e-05, "loss": 7.0312, "step": 16926 }, { "epoch": 0.7136623310917638, "grad_norm": 1.680856466293335, "learning_rate": 4.234048533683126e-05, "loss": 6.8888, "step": 16927 }, { "epoch": 0.7137044922739634, "grad_norm": 2.2046499252319336, "learning_rate": 4.2339704550423186e-05, "loss": 6.6099, "step": 16928 }, { "epoch": 0.713746653456163, "grad_norm": 1.729363203048706, "learning_rate": 4.233892376401512e-05, "loss": 6.513, "step": 16929 }, { "epoch": 0.7137888146383624, "grad_norm": 3.4681906700134277, "learning_rate": 4.233814297760705e-05, "loss": 7.8135, "step": 16930 }, { "epoch": 0.713830975820562, "grad_norm": 6.314182281494141, "learning_rate": 4.2337362191198973e-05, "loss": 10.5616, "step": 16931 }, { "epoch": 0.7138731370027616, "grad_norm": 2.3371193408966064, "learning_rate": 4.233658140479091e-05, "loss": 6.741, "step": 16932 }, { "epoch": 0.7139152981849611, "grad_norm": 3.146512031555176, "learning_rate": 4.2335800618382836e-05, "loss": 7.5794, "step": 16933 }, { "epoch": 0.7139574593671607, "grad_norm": 2.8709912300109863, "learning_rate": 4.233501983197477e-05, "loss": 9.1147, "step": 16934 }, { "epoch": 0.7139996205493602, "grad_norm": 3.112847089767456, "learning_rate": 4.23342390455667e-05, "loss": 7.3132, "step": 16935 }, { "epoch": 0.7140417817315597, "grad_norm": 2.738077402114868, "learning_rate": 4.233345825915862e-05, "loss": 6.6965, "step": 16936 }, { "epoch": 0.7140839429137593, "grad_norm": 1.9971545934677124, "learning_rate": 4.2332677472750554e-05, "loss": 6.5408, "step": 16937 }, { "epoch": 0.7141261040959589, "grad_norm": 1.2945835590362549, "learning_rate": 4.2331896686342485e-05, "loss": 6.5119, "step": 16938 }, { "epoch": 0.7141682652781584, "grad_norm": 3.019498825073242, "learning_rate": 4.233111589993442e-05, "loss": 6.8479, "step": 16939 }, { "epoch": 0.7142104264603579, "grad_norm": 3.115943193435669, "learning_rate": 4.233033511352635e-05, "loss": 7.0993, "step": 16940 }, { "epoch": 0.7142525876425575, "grad_norm": 2.6896936893463135, "learning_rate": 4.232955432711828e-05, "loss": 6.8538, "step": 16941 }, { "epoch": 0.7142947488247571, "grad_norm": 1.4337440729141235, "learning_rate": 4.2328773540710204e-05, "loss": 6.4253, "step": 16942 }, { "epoch": 0.7143369100069566, "grad_norm": 2.8350188732147217, "learning_rate": 4.2327992754302135e-05, "loss": 6.4382, "step": 16943 }, { "epoch": 0.7143790711891561, "grad_norm": 2.1340272426605225, "learning_rate": 4.2327211967894066e-05, "loss": 6.9864, "step": 16944 }, { "epoch": 0.7144212323713557, "grad_norm": 2.5040946006774902, "learning_rate": 4.232643118148599e-05, "loss": 7.097, "step": 16945 }, { "epoch": 0.7144633935535553, "grad_norm": 3.581042766571045, "learning_rate": 4.232565039507793e-05, "loss": 6.548, "step": 16946 }, { "epoch": 0.7145055547357548, "grad_norm": 2.9420390129089355, "learning_rate": 4.232486960866985e-05, "loss": 7.2281, "step": 16947 }, { "epoch": 0.7145477159179543, "grad_norm": 1.0948601961135864, "learning_rate": 4.2324088822261785e-05, "loss": 6.5323, "step": 16948 }, { "epoch": 0.7145898771001539, "grad_norm": 5.8350629806518555, "learning_rate": 4.2323308035853716e-05, "loss": 9.0008, "step": 16949 }, { "epoch": 0.7146320382823534, "grad_norm": 6.463706016540527, "learning_rate": 4.232252724944564e-05, "loss": 8.8507, "step": 16950 }, { "epoch": 0.714674199464553, "grad_norm": 3.9347362518310547, "learning_rate": 4.232174646303758e-05, "loss": 7.1051, "step": 16951 }, { "epoch": 0.7147163606467525, "grad_norm": 2.031294345855713, "learning_rate": 4.23209656766295e-05, "loss": 6.7285, "step": 16952 }, { "epoch": 0.714758521828952, "grad_norm": 1.6227620840072632, "learning_rate": 4.2320184890221434e-05, "loss": 6.7333, "step": 16953 }, { "epoch": 0.7148006830111516, "grad_norm": 1.63417387008667, "learning_rate": 4.2319404103813365e-05, "loss": 6.577, "step": 16954 }, { "epoch": 0.7148428441933512, "grad_norm": 2.0779025554656982, "learning_rate": 4.231862331740529e-05, "loss": 6.3407, "step": 16955 }, { "epoch": 0.7148850053755508, "grad_norm": 1.5487730503082275, "learning_rate": 4.231784253099722e-05, "loss": 6.606, "step": 16956 }, { "epoch": 0.7149271665577502, "grad_norm": 0.9227875471115112, "learning_rate": 4.231706174458915e-05, "loss": 6.7919, "step": 16957 }, { "epoch": 0.7149693277399498, "grad_norm": 3.3695528507232666, "learning_rate": 4.2316280958181084e-05, "loss": 7.5993, "step": 16958 }, { "epoch": 0.7150114889221494, "grad_norm": 4.47292947769165, "learning_rate": 4.231550017177301e-05, "loss": 7.6118, "step": 16959 }, { "epoch": 0.715053650104349, "grad_norm": 2.1739249229431152, "learning_rate": 4.231471938536494e-05, "loss": 6.9765, "step": 16960 }, { "epoch": 0.7150958112865484, "grad_norm": 3.699869155883789, "learning_rate": 4.231393859895687e-05, "loss": 7.5396, "step": 16961 }, { "epoch": 0.715137972468748, "grad_norm": 3.336395740509033, "learning_rate": 4.23131578125488e-05, "loss": 7.3305, "step": 16962 }, { "epoch": 0.7151801336509476, "grad_norm": 2.0245015621185303, "learning_rate": 4.231237702614073e-05, "loss": 7.1969, "step": 16963 }, { "epoch": 0.7152222948331471, "grad_norm": 3.0615317821502686, "learning_rate": 4.231159623973266e-05, "loss": 6.6012, "step": 16964 }, { "epoch": 0.7152644560153467, "grad_norm": 2.0021040439605713, "learning_rate": 4.2310815453324596e-05, "loss": 6.4242, "step": 16965 }, { "epoch": 0.7153066171975462, "grad_norm": 1.1747931241989136, "learning_rate": 4.231003466691652e-05, "loss": 6.7449, "step": 16966 }, { "epoch": 0.7153487783797458, "grad_norm": 1.424992561340332, "learning_rate": 4.230925388050845e-05, "loss": 6.4877, "step": 16967 }, { "epoch": 0.7153909395619453, "grad_norm": 3.284602403640747, "learning_rate": 4.230847309410038e-05, "loss": 7.1062, "step": 16968 }, { "epoch": 0.7154331007441449, "grad_norm": 2.6069533824920654, "learning_rate": 4.230769230769231e-05, "loss": 6.9958, "step": 16969 }, { "epoch": 0.7154752619263444, "grad_norm": 1.486657977104187, "learning_rate": 4.230691152128424e-05, "loss": 6.6574, "step": 16970 }, { "epoch": 0.7155174231085439, "grad_norm": 1.2666029930114746, "learning_rate": 4.230613073487617e-05, "loss": 6.7818, "step": 16971 }, { "epoch": 0.7155595842907435, "grad_norm": 2.6231062412261963, "learning_rate": 4.23053499484681e-05, "loss": 7.4832, "step": 16972 }, { "epoch": 0.7156017454729431, "grad_norm": 2.2192976474761963, "learning_rate": 4.230456916206003e-05, "loss": 6.8316, "step": 16973 }, { "epoch": 0.7156439066551427, "grad_norm": 2.969593048095703, "learning_rate": 4.230378837565196e-05, "loss": 7.4442, "step": 16974 }, { "epoch": 0.7156860678373421, "grad_norm": 2.4930732250213623, "learning_rate": 4.230300758924389e-05, "loss": 6.9086, "step": 16975 }, { "epoch": 0.7157282290195417, "grad_norm": 1.2750651836395264, "learning_rate": 4.230222680283582e-05, "loss": 6.5982, "step": 16976 }, { "epoch": 0.7157703902017413, "grad_norm": 1.7694085836410522, "learning_rate": 4.230144601642775e-05, "loss": 7.0174, "step": 16977 }, { "epoch": 0.7158125513839408, "grad_norm": 3.1679351329803467, "learning_rate": 4.2300665230019675e-05, "loss": 6.9461, "step": 16978 }, { "epoch": 0.7158547125661403, "grad_norm": 2.829911231994629, "learning_rate": 4.2299884443611606e-05, "loss": 7.3973, "step": 16979 }, { "epoch": 0.7158968737483399, "grad_norm": 1.436058521270752, "learning_rate": 4.229910365720354e-05, "loss": 6.6686, "step": 16980 }, { "epoch": 0.7159390349305395, "grad_norm": 2.4765453338623047, "learning_rate": 4.229832287079546e-05, "loss": 6.5194, "step": 16981 }, { "epoch": 0.715981196112739, "grad_norm": 2.9607913494110107, "learning_rate": 4.22975420843874e-05, "loss": 7.1722, "step": 16982 }, { "epoch": 0.7160233572949386, "grad_norm": 1.5986347198486328, "learning_rate": 4.2296761297979324e-05, "loss": 6.5505, "step": 16983 }, { "epoch": 0.7160655184771381, "grad_norm": 1.4335671663284302, "learning_rate": 4.229598051157126e-05, "loss": 6.6549, "step": 16984 }, { "epoch": 0.7161076796593376, "grad_norm": 2.0213797092437744, "learning_rate": 4.229519972516319e-05, "loss": 6.3973, "step": 16985 }, { "epoch": 0.7161498408415372, "grad_norm": 1.8798094987869263, "learning_rate": 4.229441893875512e-05, "loss": 6.9179, "step": 16986 }, { "epoch": 0.7161920020237368, "grad_norm": 3.0083837509155273, "learning_rate": 4.229363815234705e-05, "loss": 7.6277, "step": 16987 }, { "epoch": 0.7162341632059362, "grad_norm": 2.5932130813598633, "learning_rate": 4.2292857365938974e-05, "loss": 6.54, "step": 16988 }, { "epoch": 0.7162763243881358, "grad_norm": 1.912469744682312, "learning_rate": 4.2292076579530905e-05, "loss": 6.9865, "step": 16989 }, { "epoch": 0.7163184855703354, "grad_norm": 2.410824775695801, "learning_rate": 4.2291295793122836e-05, "loss": 6.538, "step": 16990 }, { "epoch": 0.716360646752535, "grad_norm": 2.694084882736206, "learning_rate": 4.229051500671477e-05, "loss": 7.6979, "step": 16991 }, { "epoch": 0.7164028079347345, "grad_norm": 1.5051612854003906, "learning_rate": 4.228973422030669e-05, "loss": 7.0733, "step": 16992 }, { "epoch": 0.716444969116934, "grad_norm": 3.0454792976379395, "learning_rate": 4.2288953433898623e-05, "loss": 6.7666, "step": 16993 }, { "epoch": 0.7164871302991336, "grad_norm": 1.5667716264724731, "learning_rate": 4.2288172647490555e-05, "loss": 6.373, "step": 16994 }, { "epoch": 0.7165292914813332, "grad_norm": 3.7486202716827393, "learning_rate": 4.2287391861082486e-05, "loss": 8.4277, "step": 16995 }, { "epoch": 0.7165714526635327, "grad_norm": 2.51597261428833, "learning_rate": 4.228661107467442e-05, "loss": 7.2811, "step": 16996 }, { "epoch": 0.7166136138457322, "grad_norm": 4.059803485870361, "learning_rate": 4.228583028826634e-05, "loss": 7.4441, "step": 16997 }, { "epoch": 0.7166557750279318, "grad_norm": 3.3881092071533203, "learning_rate": 4.228504950185827e-05, "loss": 7.6651, "step": 16998 }, { "epoch": 0.7166979362101313, "grad_norm": 1.5199322700500488, "learning_rate": 4.2284268715450204e-05, "loss": 6.5288, "step": 16999 }, { "epoch": 0.7167400973923309, "grad_norm": 1.4080737829208374, "learning_rate": 4.228348792904213e-05, "loss": 7.1848, "step": 17000 }, { "epoch": 0.7167822585745305, "grad_norm": 1.8521143198013306, "learning_rate": 4.228270714263407e-05, "loss": 6.6737, "step": 17001 }, { "epoch": 0.71682441975673, "grad_norm": 3.05427622795105, "learning_rate": 4.228192635622599e-05, "loss": 7.6891, "step": 17002 }, { "epoch": 0.7168665809389295, "grad_norm": 1.88894784450531, "learning_rate": 4.228114556981792e-05, "loss": 7.2405, "step": 17003 }, { "epoch": 0.7169087421211291, "grad_norm": 1.768316626548767, "learning_rate": 4.2280364783409854e-05, "loss": 6.4882, "step": 17004 }, { "epoch": 0.7169509033033287, "grad_norm": 1.5170793533325195, "learning_rate": 4.227958399700178e-05, "loss": 7.0241, "step": 17005 }, { "epoch": 0.7169930644855281, "grad_norm": 2.309762954711914, "learning_rate": 4.227880321059371e-05, "loss": 6.6827, "step": 17006 }, { "epoch": 0.7170352256677277, "grad_norm": 5.502964973449707, "learning_rate": 4.227802242418564e-05, "loss": 7.3625, "step": 17007 }, { "epoch": 0.7170773868499273, "grad_norm": 1.4020674228668213, "learning_rate": 4.227724163777757e-05, "loss": 6.9647, "step": 17008 }, { "epoch": 0.7171195480321269, "grad_norm": 2.819624423980713, "learning_rate": 4.22764608513695e-05, "loss": 7.3843, "step": 17009 }, { "epoch": 0.7171617092143263, "grad_norm": 1.9541716575622559, "learning_rate": 4.2275680064961435e-05, "loss": 6.5146, "step": 17010 }, { "epoch": 0.7172038703965259, "grad_norm": 3.371819019317627, "learning_rate": 4.227489927855336e-05, "loss": 7.0745, "step": 17011 }, { "epoch": 0.7172460315787255, "grad_norm": 3.104536533355713, "learning_rate": 4.227411849214529e-05, "loss": 7.4833, "step": 17012 }, { "epoch": 0.717288192760925, "grad_norm": 1.9693083763122559, "learning_rate": 4.227333770573722e-05, "loss": 6.7255, "step": 17013 }, { "epoch": 0.7173303539431246, "grad_norm": 1.8371102809906006, "learning_rate": 4.2272556919329146e-05, "loss": 6.9358, "step": 17014 }, { "epoch": 0.7173725151253241, "grad_norm": 2.444484233856201, "learning_rate": 4.2271776132921084e-05, "loss": 6.6508, "step": 17015 }, { "epoch": 0.7174146763075236, "grad_norm": 2.003699779510498, "learning_rate": 4.227099534651301e-05, "loss": 6.547, "step": 17016 }, { "epoch": 0.7174568374897232, "grad_norm": 4.532410621643066, "learning_rate": 4.227021456010494e-05, "loss": 7.0053, "step": 17017 }, { "epoch": 0.7174989986719228, "grad_norm": 2.098468542098999, "learning_rate": 4.226943377369687e-05, "loss": 6.9964, "step": 17018 }, { "epoch": 0.7175411598541223, "grad_norm": 1.0836073160171509, "learning_rate": 4.2268652987288796e-05, "loss": 6.5629, "step": 17019 }, { "epoch": 0.7175833210363218, "grad_norm": 1.2089070081710815, "learning_rate": 4.2267872200880734e-05, "loss": 6.5453, "step": 17020 }, { "epoch": 0.7176254822185214, "grad_norm": 5.610052585601807, "learning_rate": 4.226709141447266e-05, "loss": 9.6859, "step": 17021 }, { "epoch": 0.717667643400721, "grad_norm": 1.6341406106948853, "learning_rate": 4.226631062806459e-05, "loss": 6.4861, "step": 17022 }, { "epoch": 0.7177098045829206, "grad_norm": 2.8553810119628906, "learning_rate": 4.226552984165652e-05, "loss": 7.8349, "step": 17023 }, { "epoch": 0.71775196576512, "grad_norm": 1.7625949382781982, "learning_rate": 4.2264749055248445e-05, "loss": 6.9192, "step": 17024 }, { "epoch": 0.7177941269473196, "grad_norm": 1.4867514371871948, "learning_rate": 4.2263968268840376e-05, "loss": 6.6615, "step": 17025 }, { "epoch": 0.7178362881295192, "grad_norm": 1.958380103111267, "learning_rate": 4.226318748243231e-05, "loss": 6.8217, "step": 17026 }, { "epoch": 0.7178784493117187, "grad_norm": 2.8332884311676025, "learning_rate": 4.226240669602424e-05, "loss": 7.3513, "step": 17027 }, { "epoch": 0.7179206104939182, "grad_norm": 4.0766119956970215, "learning_rate": 4.226162590961616e-05, "loss": 6.6906, "step": 17028 }, { "epoch": 0.7179627716761178, "grad_norm": 2.432525634765625, "learning_rate": 4.22608451232081e-05, "loss": 6.9778, "step": 17029 }, { "epoch": 0.7180049328583173, "grad_norm": 1.760705828666687, "learning_rate": 4.2260064336800026e-05, "loss": 6.6439, "step": 17030 }, { "epoch": 0.7180470940405169, "grad_norm": 3.294400215148926, "learning_rate": 4.225928355039196e-05, "loss": 7.925, "step": 17031 }, { "epoch": 0.7180892552227165, "grad_norm": 3.9056711196899414, "learning_rate": 4.225850276398389e-05, "loss": 7.2023, "step": 17032 }, { "epoch": 0.718131416404916, "grad_norm": 2.697390556335449, "learning_rate": 4.225772197757581e-05, "loss": 6.6863, "step": 17033 }, { "epoch": 0.7181735775871155, "grad_norm": 4.925674915313721, "learning_rate": 4.225694119116775e-05, "loss": 7.8023, "step": 17034 }, { "epoch": 0.7182157387693151, "grad_norm": 2.0624282360076904, "learning_rate": 4.2256160404759675e-05, "loss": 6.8369, "step": 17035 }, { "epoch": 0.7182578999515147, "grad_norm": 2.416619300842285, "learning_rate": 4.225537961835161e-05, "loss": 6.7158, "step": 17036 }, { "epoch": 0.7183000611337141, "grad_norm": 2.2161340713500977, "learning_rate": 4.225459883194354e-05, "loss": 7.1788, "step": 17037 }, { "epoch": 0.7183422223159137, "grad_norm": 3.6291353702545166, "learning_rate": 4.225381804553546e-05, "loss": 7.6467, "step": 17038 }, { "epoch": 0.7183843834981133, "grad_norm": 2.647564649581909, "learning_rate": 4.2253037259127394e-05, "loss": 6.4833, "step": 17039 }, { "epoch": 0.7184265446803129, "grad_norm": 1.588670015335083, "learning_rate": 4.2252256472719325e-05, "loss": 6.991, "step": 17040 }, { "epoch": 0.7184687058625124, "grad_norm": 1.9225560426712036, "learning_rate": 4.2251475686311256e-05, "loss": 6.3935, "step": 17041 }, { "epoch": 0.7185108670447119, "grad_norm": 1.892966389656067, "learning_rate": 4.225069489990319e-05, "loss": 6.7163, "step": 17042 }, { "epoch": 0.7185530282269115, "grad_norm": 4.813758373260498, "learning_rate": 4.224991411349511e-05, "loss": 7.7842, "step": 17043 }, { "epoch": 0.718595189409111, "grad_norm": 4.418508052825928, "learning_rate": 4.224913332708704e-05, "loss": 7.7074, "step": 17044 }, { "epoch": 0.7186373505913106, "grad_norm": 1.6440199613571167, "learning_rate": 4.2248352540678974e-05, "loss": 6.495, "step": 17045 }, { "epoch": 0.7186795117735101, "grad_norm": 2.0229110717773438, "learning_rate": 4.2247571754270906e-05, "loss": 6.6839, "step": 17046 }, { "epoch": 0.7187216729557097, "grad_norm": 2.2637672424316406, "learning_rate": 4.224679096786283e-05, "loss": 6.8758, "step": 17047 }, { "epoch": 0.7187638341379092, "grad_norm": 1.3516024351119995, "learning_rate": 4.224601018145477e-05, "loss": 6.4655, "step": 17048 }, { "epoch": 0.7188059953201088, "grad_norm": 1.5361268520355225, "learning_rate": 4.224522939504669e-05, "loss": 6.8393, "step": 17049 }, { "epoch": 0.7188481565023084, "grad_norm": 2.112328052520752, "learning_rate": 4.224444860863862e-05, "loss": 6.9046, "step": 17050 }, { "epoch": 0.7188903176845078, "grad_norm": 1.301110029220581, "learning_rate": 4.2243667822230555e-05, "loss": 6.719, "step": 17051 }, { "epoch": 0.7189324788667074, "grad_norm": 2.4583277702331543, "learning_rate": 4.224288703582248e-05, "loss": 7.2483, "step": 17052 }, { "epoch": 0.718974640048907, "grad_norm": 1.9930404424667358, "learning_rate": 4.224210624941442e-05, "loss": 7.1867, "step": 17053 }, { "epoch": 0.7190168012311066, "grad_norm": 3.2787325382232666, "learning_rate": 4.224132546300634e-05, "loss": 7.2209, "step": 17054 }, { "epoch": 0.719058962413306, "grad_norm": 3.6504251956939697, "learning_rate": 4.2240544676598273e-05, "loss": 7.7312, "step": 17055 }, { "epoch": 0.7191011235955056, "grad_norm": 2.1368892192840576, "learning_rate": 4.2239763890190205e-05, "loss": 6.3462, "step": 17056 }, { "epoch": 0.7191432847777052, "grad_norm": 1.93477463722229, "learning_rate": 4.223898310378213e-05, "loss": 7.7819, "step": 17057 }, { "epoch": 0.7191854459599047, "grad_norm": 2.8009285926818848, "learning_rate": 4.223820231737406e-05, "loss": 6.584, "step": 17058 }, { "epoch": 0.7192276071421043, "grad_norm": 1.470191478729248, "learning_rate": 4.223742153096599e-05, "loss": 6.4904, "step": 17059 }, { "epoch": 0.7192697683243038, "grad_norm": 1.9327698945999146, "learning_rate": 4.223664074455792e-05, "loss": 6.8888, "step": 17060 }, { "epoch": 0.7193119295065034, "grad_norm": 1.295425534248352, "learning_rate": 4.223585995814985e-05, "loss": 6.5307, "step": 17061 }, { "epoch": 0.7193540906887029, "grad_norm": 3.9222733974456787, "learning_rate": 4.223507917174178e-05, "loss": 6.3714, "step": 17062 }, { "epoch": 0.7193962518709025, "grad_norm": 2.9114627838134766, "learning_rate": 4.223429838533371e-05, "loss": 6.3985, "step": 17063 }, { "epoch": 0.719438413053102, "grad_norm": 1.9789538383483887, "learning_rate": 4.223351759892564e-05, "loss": 6.823, "step": 17064 }, { "epoch": 0.7194805742353015, "grad_norm": 2.6739423274993896, "learning_rate": 4.223273681251757e-05, "loss": 7.6398, "step": 17065 }, { "epoch": 0.7195227354175011, "grad_norm": 3.766486406326294, "learning_rate": 4.22319560261095e-05, "loss": 7.8344, "step": 17066 }, { "epoch": 0.7195648965997007, "grad_norm": 2.098478317260742, "learning_rate": 4.223117523970143e-05, "loss": 6.895, "step": 17067 }, { "epoch": 0.7196070577819002, "grad_norm": 4.165891170501709, "learning_rate": 4.223039445329336e-05, "loss": 9.2004, "step": 17068 }, { "epoch": 0.7196492189640997, "grad_norm": 2.2466578483581543, "learning_rate": 4.2229613666885284e-05, "loss": 6.5167, "step": 17069 }, { "epoch": 0.7196913801462993, "grad_norm": 1.7482209205627441, "learning_rate": 4.222883288047722e-05, "loss": 7.5978, "step": 17070 }, { "epoch": 0.7197335413284989, "grad_norm": 1.7836058139801025, "learning_rate": 4.2228052094069147e-05, "loss": 6.4051, "step": 17071 }, { "epoch": 0.7197757025106984, "grad_norm": 1.1810340881347656, "learning_rate": 4.222727130766108e-05, "loss": 6.5461, "step": 17072 }, { "epoch": 0.7198178636928979, "grad_norm": 1.839820146560669, "learning_rate": 4.222649052125301e-05, "loss": 6.7795, "step": 17073 }, { "epoch": 0.7198600248750975, "grad_norm": 5.0599517822265625, "learning_rate": 4.222570973484494e-05, "loss": 7.4905, "step": 17074 }, { "epoch": 0.719902186057297, "grad_norm": 3.423957586288452, "learning_rate": 4.2224928948436865e-05, "loss": 7.1121, "step": 17075 }, { "epoch": 0.7199443472394966, "grad_norm": 2.9693281650543213, "learning_rate": 4.2224148162028796e-05, "loss": 7.0607, "step": 17076 }, { "epoch": 0.7199865084216961, "grad_norm": 2.55039381980896, "learning_rate": 4.222336737562073e-05, "loss": 6.6837, "step": 17077 }, { "epoch": 0.7200286696038957, "grad_norm": 2.246648073196411, "learning_rate": 4.222258658921266e-05, "loss": 6.9231, "step": 17078 }, { "epoch": 0.7200708307860952, "grad_norm": 3.2474143505096436, "learning_rate": 4.222180580280459e-05, "loss": 6.6771, "step": 17079 }, { "epoch": 0.7201129919682948, "grad_norm": 3.07460618019104, "learning_rate": 4.2221025016396514e-05, "loss": 9.6435, "step": 17080 }, { "epoch": 0.7201551531504944, "grad_norm": 3.5042662620544434, "learning_rate": 4.2220244229988446e-05, "loss": 6.6579, "step": 17081 }, { "epoch": 0.7201973143326938, "grad_norm": 2.299419641494751, "learning_rate": 4.221946344358038e-05, "loss": 6.4667, "step": 17082 }, { "epoch": 0.7202394755148934, "grad_norm": 3.765010118484497, "learning_rate": 4.22186826571723e-05, "loss": 7.5194, "step": 17083 }, { "epoch": 0.720281636697093, "grad_norm": 2.573129892349243, "learning_rate": 4.221790187076424e-05, "loss": 6.6367, "step": 17084 }, { "epoch": 0.7203237978792926, "grad_norm": 3.6745100021362305, "learning_rate": 4.2217121084356164e-05, "loss": 7.1176, "step": 17085 }, { "epoch": 0.720365959061492, "grad_norm": 3.0635650157928467, "learning_rate": 4.2216340297948095e-05, "loss": 6.6585, "step": 17086 }, { "epoch": 0.7204081202436916, "grad_norm": 2.796198844909668, "learning_rate": 4.2215559511540026e-05, "loss": 7.0523, "step": 17087 }, { "epoch": 0.7204502814258912, "grad_norm": 1.3554799556732178, "learning_rate": 4.221477872513195e-05, "loss": 7.0511, "step": 17088 }, { "epoch": 0.7204924426080908, "grad_norm": 2.033599376678467, "learning_rate": 4.221399793872389e-05, "loss": 6.6421, "step": 17089 }, { "epoch": 0.7205346037902903, "grad_norm": 2.5631868839263916, "learning_rate": 4.221321715231581e-05, "loss": 6.348, "step": 17090 }, { "epoch": 0.7205767649724898, "grad_norm": 2.853672981262207, "learning_rate": 4.2212436365907745e-05, "loss": 6.9571, "step": 17091 }, { "epoch": 0.7206189261546894, "grad_norm": 3.295408010482788, "learning_rate": 4.2211655579499676e-05, "loss": 7.7472, "step": 17092 }, { "epoch": 0.7206610873368889, "grad_norm": 1.703160047531128, "learning_rate": 4.22108747930916e-05, "loss": 7.1119, "step": 17093 }, { "epoch": 0.7207032485190885, "grad_norm": 1.379848599433899, "learning_rate": 4.221009400668353e-05, "loss": 6.4807, "step": 17094 }, { "epoch": 0.720745409701288, "grad_norm": 3.3444883823394775, "learning_rate": 4.220931322027546e-05, "loss": 7.4644, "step": 17095 }, { "epoch": 0.7207875708834875, "grad_norm": 3.1073949337005615, "learning_rate": 4.2208532433867394e-05, "loss": 6.5887, "step": 17096 }, { "epoch": 0.7208297320656871, "grad_norm": 1.7617518901824951, "learning_rate": 4.220775164745932e-05, "loss": 6.5483, "step": 17097 }, { "epoch": 0.7208718932478867, "grad_norm": 2.2700300216674805, "learning_rate": 4.220697086105126e-05, "loss": 6.8944, "step": 17098 }, { "epoch": 0.7209140544300863, "grad_norm": 3.576829195022583, "learning_rate": 4.220619007464318e-05, "loss": 6.5369, "step": 17099 }, { "epoch": 0.7209562156122857, "grad_norm": 3.309495449066162, "learning_rate": 4.220540928823511e-05, "loss": 8.5038, "step": 17100 }, { "epoch": 0.7209983767944853, "grad_norm": 2.3254597187042236, "learning_rate": 4.2204628501827044e-05, "loss": 6.3287, "step": 17101 }, { "epoch": 0.7210405379766849, "grad_norm": 1.9419997930526733, "learning_rate": 4.220384771541897e-05, "loss": 6.6704, "step": 17102 }, { "epoch": 0.7210826991588845, "grad_norm": 1.533848524093628, "learning_rate": 4.2203066929010906e-05, "loss": 6.6465, "step": 17103 }, { "epoch": 0.7211248603410839, "grad_norm": 1.6732224225997925, "learning_rate": 4.220228614260283e-05, "loss": 6.455, "step": 17104 }, { "epoch": 0.7211670215232835, "grad_norm": 2.4703800678253174, "learning_rate": 4.220150535619476e-05, "loss": 7.1284, "step": 17105 }, { "epoch": 0.7212091827054831, "grad_norm": 2.126574993133545, "learning_rate": 4.220072456978669e-05, "loss": 7.2277, "step": 17106 }, { "epoch": 0.7212513438876826, "grad_norm": 1.761273980140686, "learning_rate": 4.219994378337862e-05, "loss": 6.7334, "step": 17107 }, { "epoch": 0.7212935050698822, "grad_norm": 1.8876827955245972, "learning_rate": 4.219916299697055e-05, "loss": 7.0337, "step": 17108 }, { "epoch": 0.7213356662520817, "grad_norm": 2.2447619438171387, "learning_rate": 4.219838221056248e-05, "loss": 6.9901, "step": 17109 }, { "epoch": 0.7213778274342812, "grad_norm": 2.7107489109039307, "learning_rate": 4.219760142415441e-05, "loss": 6.8097, "step": 17110 }, { "epoch": 0.7214199886164808, "grad_norm": 2.3285107612609863, "learning_rate": 4.219682063774634e-05, "loss": 6.3457, "step": 17111 }, { "epoch": 0.7214621497986804, "grad_norm": 4.176926136016846, "learning_rate": 4.219603985133827e-05, "loss": 7.2095, "step": 17112 }, { "epoch": 0.7215043109808799, "grad_norm": 3.7684359550476074, "learning_rate": 4.21952590649302e-05, "loss": 7.7298, "step": 17113 }, { "epoch": 0.7215464721630794, "grad_norm": 2.9374117851257324, "learning_rate": 4.219447827852213e-05, "loss": 7.067, "step": 17114 }, { "epoch": 0.721588633345279, "grad_norm": 2.450817346572876, "learning_rate": 4.219369749211406e-05, "loss": 6.523, "step": 17115 }, { "epoch": 0.7216307945274786, "grad_norm": 2.151357650756836, "learning_rate": 4.2192916705705985e-05, "loss": 7.2367, "step": 17116 }, { "epoch": 0.7216729557096782, "grad_norm": 1.655324101448059, "learning_rate": 4.2192135919297924e-05, "loss": 6.9909, "step": 17117 }, { "epoch": 0.7217151168918776, "grad_norm": 2.5816800594329834, "learning_rate": 4.219135513288985e-05, "loss": 6.534, "step": 17118 }, { "epoch": 0.7217572780740772, "grad_norm": 2.589120388031006, "learning_rate": 4.219057434648178e-05, "loss": 7.5267, "step": 17119 }, { "epoch": 0.7217994392562768, "grad_norm": 3.2898712158203125, "learning_rate": 4.218979356007371e-05, "loss": 6.6107, "step": 17120 }, { "epoch": 0.7218416004384763, "grad_norm": 4.243585586547852, "learning_rate": 4.2189012773665635e-05, "loss": 7.5608, "step": 17121 }, { "epoch": 0.7218837616206758, "grad_norm": 2.3626708984375, "learning_rate": 4.218823198725757e-05, "loss": 6.4464, "step": 17122 }, { "epoch": 0.7219259228028754, "grad_norm": 1.7297472953796387, "learning_rate": 4.21874512008495e-05, "loss": 6.4514, "step": 17123 }, { "epoch": 0.721968083985075, "grad_norm": 3.7578511238098145, "learning_rate": 4.218667041444143e-05, "loss": 7.6296, "step": 17124 }, { "epoch": 0.7220102451672745, "grad_norm": 3.013533353805542, "learning_rate": 4.218588962803336e-05, "loss": 6.7879, "step": 17125 }, { "epoch": 0.722052406349474, "grad_norm": 2.4993724822998047, "learning_rate": 4.2185108841625285e-05, "loss": 6.6117, "step": 17126 }, { "epoch": 0.7220945675316736, "grad_norm": 1.5946590900421143, "learning_rate": 4.2184328055217216e-05, "loss": 6.9409, "step": 17127 }, { "epoch": 0.7221367287138731, "grad_norm": 1.1765276193618774, "learning_rate": 4.218354726880915e-05, "loss": 6.5669, "step": 17128 }, { "epoch": 0.7221788898960727, "grad_norm": 2.904343843460083, "learning_rate": 4.218276648240108e-05, "loss": 7.6599, "step": 17129 }, { "epoch": 0.7222210510782723, "grad_norm": 5.536872863769531, "learning_rate": 4.2181985695993e-05, "loss": 6.7764, "step": 17130 }, { "epoch": 0.7222632122604717, "grad_norm": 3.1136553287506104, "learning_rate": 4.2181204909584934e-05, "loss": 7.658, "step": 17131 }, { "epoch": 0.7223053734426713, "grad_norm": 1.8720543384552002, "learning_rate": 4.2180424123176865e-05, "loss": 6.7753, "step": 17132 }, { "epoch": 0.7223475346248709, "grad_norm": 2.005333662033081, "learning_rate": 4.2179643336768797e-05, "loss": 6.9001, "step": 17133 }, { "epoch": 0.7223896958070705, "grad_norm": 1.22640061378479, "learning_rate": 4.217886255036073e-05, "loss": 6.4877, "step": 17134 }, { "epoch": 0.7224318569892699, "grad_norm": 2.5446267127990723, "learning_rate": 4.217808176395265e-05, "loss": 6.6225, "step": 17135 }, { "epoch": 0.7224740181714695, "grad_norm": 3.484029769897461, "learning_rate": 4.217730097754459e-05, "loss": 7.1148, "step": 17136 }, { "epoch": 0.7225161793536691, "grad_norm": 4.394139289855957, "learning_rate": 4.2176520191136515e-05, "loss": 7.0616, "step": 17137 }, { "epoch": 0.7225583405358686, "grad_norm": 0.9556205868721008, "learning_rate": 4.217573940472844e-05, "loss": 6.6572, "step": 17138 }, { "epoch": 0.7226005017180682, "grad_norm": 3.3412251472473145, "learning_rate": 4.217495861832038e-05, "loss": 7.2599, "step": 17139 }, { "epoch": 0.7226426629002677, "grad_norm": 3.2021796703338623, "learning_rate": 4.21741778319123e-05, "loss": 7.6705, "step": 17140 }, { "epoch": 0.7226848240824673, "grad_norm": 2.0714712142944336, "learning_rate": 4.217339704550423e-05, "loss": 6.9741, "step": 17141 }, { "epoch": 0.7227269852646668, "grad_norm": 2.5528879165649414, "learning_rate": 4.2172616259096164e-05, "loss": 6.6663, "step": 17142 }, { "epoch": 0.7227691464468664, "grad_norm": 2.0347819328308105, "learning_rate": 4.2171835472688096e-05, "loss": 6.5052, "step": 17143 }, { "epoch": 0.7228113076290659, "grad_norm": 1.7339060306549072, "learning_rate": 4.217105468628003e-05, "loss": 7.0281, "step": 17144 }, { "epoch": 0.7228534688112654, "grad_norm": 2.128920793533325, "learning_rate": 4.217027389987195e-05, "loss": 6.9744, "step": 17145 }, { "epoch": 0.722895629993465, "grad_norm": 2.9085233211517334, "learning_rate": 4.216949311346388e-05, "loss": 6.5263, "step": 17146 }, { "epoch": 0.7229377911756646, "grad_norm": 1.2256511449813843, "learning_rate": 4.2168712327055814e-05, "loss": 6.5668, "step": 17147 }, { "epoch": 0.7229799523578642, "grad_norm": 2.74686336517334, "learning_rate": 4.2167931540647745e-05, "loss": 7.5678, "step": 17148 }, { "epoch": 0.7230221135400636, "grad_norm": 2.1285548210144043, "learning_rate": 4.216715075423967e-05, "loss": 6.8102, "step": 17149 }, { "epoch": 0.7230642747222632, "grad_norm": 2.862616539001465, "learning_rate": 4.21663699678316e-05, "loss": 7.1769, "step": 17150 }, { "epoch": 0.7231064359044628, "grad_norm": 2.57173752784729, "learning_rate": 4.216558918142353e-05, "loss": 6.6219, "step": 17151 }, { "epoch": 0.7231485970866623, "grad_norm": 1.3196089267730713, "learning_rate": 4.216480839501546e-05, "loss": 7.0041, "step": 17152 }, { "epoch": 0.7231907582688618, "grad_norm": 1.2413008213043213, "learning_rate": 4.2164027608607395e-05, "loss": 6.9267, "step": 17153 }, { "epoch": 0.7232329194510614, "grad_norm": 1.182984471321106, "learning_rate": 4.216324682219932e-05, "loss": 6.6231, "step": 17154 }, { "epoch": 0.723275080633261, "grad_norm": 3.6772263050079346, "learning_rate": 4.216246603579125e-05, "loss": 8.3222, "step": 17155 }, { "epoch": 0.7233172418154605, "grad_norm": 1.505933403968811, "learning_rate": 4.216168524938318e-05, "loss": 6.7363, "step": 17156 }, { "epoch": 0.7233594029976601, "grad_norm": 1.300161361694336, "learning_rate": 4.2160904462975106e-05, "loss": 6.7373, "step": 17157 }, { "epoch": 0.7234015641798596, "grad_norm": 4.094607830047607, "learning_rate": 4.2160123676567044e-05, "loss": 7.7139, "step": 17158 }, { "epoch": 0.7234437253620591, "grad_norm": 1.2531706094741821, "learning_rate": 4.215934289015897e-05, "loss": 6.6161, "step": 17159 }, { "epoch": 0.7234858865442587, "grad_norm": 2.2742691040039062, "learning_rate": 4.21585621037509e-05, "loss": 7.1064, "step": 17160 }, { "epoch": 0.7235280477264583, "grad_norm": 1.1236525774002075, "learning_rate": 4.215778131734283e-05, "loss": 6.4782, "step": 17161 }, { "epoch": 0.7235702089086578, "grad_norm": 1.09578275680542, "learning_rate": 4.215700053093476e-05, "loss": 6.9016, "step": 17162 }, { "epoch": 0.7236123700908573, "grad_norm": 1.6102631092071533, "learning_rate": 4.215621974452669e-05, "loss": 6.6342, "step": 17163 }, { "epoch": 0.7236545312730569, "grad_norm": 1.476288080215454, "learning_rate": 4.215543895811862e-05, "loss": 6.7431, "step": 17164 }, { "epoch": 0.7236966924552565, "grad_norm": 3.155388116836548, "learning_rate": 4.215465817171055e-05, "loss": 6.5924, "step": 17165 }, { "epoch": 0.723738853637456, "grad_norm": 1.6802617311477661, "learning_rate": 4.2153877385302474e-05, "loss": 6.9599, "step": 17166 }, { "epoch": 0.7237810148196555, "grad_norm": 1.411184549331665, "learning_rate": 4.215309659889441e-05, "loss": 6.6111, "step": 17167 }, { "epoch": 0.7238231760018551, "grad_norm": 1.274948000907898, "learning_rate": 4.2152315812486336e-05, "loss": 6.4045, "step": 17168 }, { "epoch": 0.7238653371840547, "grad_norm": 2.3103888034820557, "learning_rate": 4.215153502607827e-05, "loss": 6.475, "step": 17169 }, { "epoch": 0.7239074983662542, "grad_norm": 1.4092426300048828, "learning_rate": 4.21507542396702e-05, "loss": 6.8215, "step": 17170 }, { "epoch": 0.7239496595484537, "grad_norm": 3.9937164783477783, "learning_rate": 4.2149973453262123e-05, "loss": 7.2764, "step": 17171 }, { "epoch": 0.7239918207306533, "grad_norm": 1.9443330764770508, "learning_rate": 4.214919266685406e-05, "loss": 6.8656, "step": 17172 }, { "epoch": 0.7240339819128528, "grad_norm": 3.560769557952881, "learning_rate": 4.2148411880445986e-05, "loss": 7.4418, "step": 17173 }, { "epoch": 0.7240761430950524, "grad_norm": 1.9583505392074585, "learning_rate": 4.214763109403792e-05, "loss": 7.4029, "step": 17174 }, { "epoch": 0.724118304277252, "grad_norm": 3.626629114151001, "learning_rate": 4.214685030762985e-05, "loss": 6.7881, "step": 17175 }, { "epoch": 0.7241604654594515, "grad_norm": 2.4720609188079834, "learning_rate": 4.214606952122177e-05, "loss": 6.4576, "step": 17176 }, { "epoch": 0.724202626641651, "grad_norm": 2.1406586170196533, "learning_rate": 4.2145288734813704e-05, "loss": 6.8675, "step": 17177 }, { "epoch": 0.7242447878238506, "grad_norm": 1.4681702852249146, "learning_rate": 4.2144507948405635e-05, "loss": 6.7921, "step": 17178 }, { "epoch": 0.7242869490060502, "grad_norm": 4.012357711791992, "learning_rate": 4.214372716199757e-05, "loss": 7.6463, "step": 17179 }, { "epoch": 0.7243291101882496, "grad_norm": 1.6856539249420166, "learning_rate": 4.21429463755895e-05, "loss": 6.4739, "step": 17180 }, { "epoch": 0.7243712713704492, "grad_norm": 1.7969802618026733, "learning_rate": 4.214216558918143e-05, "loss": 6.8609, "step": 17181 }, { "epoch": 0.7244134325526488, "grad_norm": 1.7422829866409302, "learning_rate": 4.2141384802773354e-05, "loss": 6.4006, "step": 17182 }, { "epoch": 0.7244555937348484, "grad_norm": 1.7847455739974976, "learning_rate": 4.2140604016365285e-05, "loss": 6.4133, "step": 17183 }, { "epoch": 0.7244977549170478, "grad_norm": 1.4718784093856812, "learning_rate": 4.2139823229957216e-05, "loss": 6.711, "step": 17184 }, { "epoch": 0.7245399160992474, "grad_norm": 2.4233014583587646, "learning_rate": 4.213904244354914e-05, "loss": 7.2164, "step": 17185 }, { "epoch": 0.724582077281447, "grad_norm": 2.3071558475494385, "learning_rate": 4.213826165714108e-05, "loss": 6.6027, "step": 17186 }, { "epoch": 0.7246242384636465, "grad_norm": 3.0481224060058594, "learning_rate": 4.2137480870733e-05, "loss": 6.5509, "step": 17187 }, { "epoch": 0.7246663996458461, "grad_norm": 2.230292320251465, "learning_rate": 4.2136700084324935e-05, "loss": 6.8162, "step": 17188 }, { "epoch": 0.7247085608280456, "grad_norm": 1.4253839254379272, "learning_rate": 4.2135919297916866e-05, "loss": 6.9247, "step": 17189 }, { "epoch": 0.7247507220102452, "grad_norm": 1.144985556602478, "learning_rate": 4.213513851150879e-05, "loss": 6.3793, "step": 17190 }, { "epoch": 0.7247928831924447, "grad_norm": 1.4299306869506836, "learning_rate": 4.213435772510073e-05, "loss": 6.3682, "step": 17191 }, { "epoch": 0.7248350443746443, "grad_norm": 1.8002345561981201, "learning_rate": 4.213357693869265e-05, "loss": 6.7234, "step": 17192 }, { "epoch": 0.7248772055568438, "grad_norm": 2.306201696395874, "learning_rate": 4.2132796152284584e-05, "loss": 6.5833, "step": 17193 }, { "epoch": 0.7249193667390433, "grad_norm": 4.015716075897217, "learning_rate": 4.2132015365876515e-05, "loss": 7.4456, "step": 17194 }, { "epoch": 0.7249615279212429, "grad_norm": 1.9705137014389038, "learning_rate": 4.213123457946844e-05, "loss": 6.992, "step": 17195 }, { "epoch": 0.7250036891034425, "grad_norm": 2.0485591888427734, "learning_rate": 4.213045379306037e-05, "loss": 6.8454, "step": 17196 }, { "epoch": 0.7250458502856421, "grad_norm": 2.8677916526794434, "learning_rate": 4.21296730066523e-05, "loss": 8.1616, "step": 17197 }, { "epoch": 0.7250880114678415, "grad_norm": 1.262245774269104, "learning_rate": 4.2128892220244234e-05, "loss": 6.5724, "step": 17198 }, { "epoch": 0.7251301726500411, "grad_norm": 1.2913398742675781, "learning_rate": 4.212811143383616e-05, "loss": 7.1952, "step": 17199 }, { "epoch": 0.7251723338322407, "grad_norm": 1.6988533735275269, "learning_rate": 4.212733064742809e-05, "loss": 7.0476, "step": 17200 }, { "epoch": 0.7252144950144402, "grad_norm": 1.6679433584213257, "learning_rate": 4.212654986102002e-05, "loss": 6.5185, "step": 17201 }, { "epoch": 0.7252566561966397, "grad_norm": 1.6776758432388306, "learning_rate": 4.212576907461195e-05, "loss": 6.792, "step": 17202 }, { "epoch": 0.7252988173788393, "grad_norm": 1.0594267845153809, "learning_rate": 4.212498828820388e-05, "loss": 6.6083, "step": 17203 }, { "epoch": 0.7253409785610389, "grad_norm": 1.2063968181610107, "learning_rate": 4.212420750179581e-05, "loss": 6.322, "step": 17204 }, { "epoch": 0.7253831397432384, "grad_norm": 1.2371370792388916, "learning_rate": 4.2123426715387746e-05, "loss": 6.7752, "step": 17205 }, { "epoch": 0.725425300925438, "grad_norm": 3.8463850021362305, "learning_rate": 4.212264592897967e-05, "loss": 7.3867, "step": 17206 }, { "epoch": 0.7254674621076375, "grad_norm": 3.2301523685455322, "learning_rate": 4.21218651425716e-05, "loss": 7.2699, "step": 17207 }, { "epoch": 0.725509623289837, "grad_norm": 1.6056897640228271, "learning_rate": 4.212108435616353e-05, "loss": 6.5393, "step": 17208 }, { "epoch": 0.7255517844720366, "grad_norm": 1.6755741834640503, "learning_rate": 4.212030356975546e-05, "loss": 6.2734, "step": 17209 }, { "epoch": 0.7255939456542362, "grad_norm": 4.255266189575195, "learning_rate": 4.211952278334739e-05, "loss": 7.8372, "step": 17210 }, { "epoch": 0.7256361068364356, "grad_norm": 1.174630045890808, "learning_rate": 4.211874199693932e-05, "loss": 6.6759, "step": 17211 }, { "epoch": 0.7256782680186352, "grad_norm": 2.374527931213379, "learning_rate": 4.211796121053125e-05, "loss": 6.525, "step": 17212 }, { "epoch": 0.7257204292008348, "grad_norm": 1.929002046585083, "learning_rate": 4.211718042412318e-05, "loss": 6.4933, "step": 17213 }, { "epoch": 0.7257625903830344, "grad_norm": 2.941333532333374, "learning_rate": 4.211639963771511e-05, "loss": 7.1344, "step": 17214 }, { "epoch": 0.7258047515652339, "grad_norm": 2.623249053955078, "learning_rate": 4.211561885130704e-05, "loss": 6.3662, "step": 17215 }, { "epoch": 0.7258469127474334, "grad_norm": 1.2266653776168823, "learning_rate": 4.211483806489897e-05, "loss": 6.6264, "step": 17216 }, { "epoch": 0.725889073929633, "grad_norm": 2.7777230739593506, "learning_rate": 4.21140572784909e-05, "loss": 7.27, "step": 17217 }, { "epoch": 0.7259312351118326, "grad_norm": 3.6720685958862305, "learning_rate": 4.2113276492082825e-05, "loss": 7.9556, "step": 17218 }, { "epoch": 0.7259733962940321, "grad_norm": 3.597592353820801, "learning_rate": 4.2112495705674756e-05, "loss": 7.1832, "step": 17219 }, { "epoch": 0.7260155574762316, "grad_norm": 2.1399059295654297, "learning_rate": 4.211171491926669e-05, "loss": 6.6456, "step": 17220 }, { "epoch": 0.7260577186584312, "grad_norm": 1.5561436414718628, "learning_rate": 4.211093413285861e-05, "loss": 6.3427, "step": 17221 }, { "epoch": 0.7260998798406307, "grad_norm": 2.309269905090332, "learning_rate": 4.211015334645055e-05, "loss": 6.2854, "step": 17222 }, { "epoch": 0.7261420410228303, "grad_norm": 2.8660547733306885, "learning_rate": 4.2109372560042474e-05, "loss": 7.0496, "step": 17223 }, { "epoch": 0.7261842022050299, "grad_norm": 1.6891684532165527, "learning_rate": 4.210859177363441e-05, "loss": 6.5008, "step": 17224 }, { "epoch": 0.7262263633872293, "grad_norm": 1.3256447315216064, "learning_rate": 4.210781098722634e-05, "loss": 6.199, "step": 17225 }, { "epoch": 0.7262685245694289, "grad_norm": 2.0061895847320557, "learning_rate": 4.210703020081827e-05, "loss": 6.7757, "step": 17226 }, { "epoch": 0.7263106857516285, "grad_norm": 2.2902610301971436, "learning_rate": 4.21062494144102e-05, "loss": 6.9422, "step": 17227 }, { "epoch": 0.7263528469338281, "grad_norm": 2.562426805496216, "learning_rate": 4.2105468628002124e-05, "loss": 7.6928, "step": 17228 }, { "epoch": 0.7263950081160275, "grad_norm": 2.041637420654297, "learning_rate": 4.2104687841594055e-05, "loss": 6.599, "step": 17229 }, { "epoch": 0.7264371692982271, "grad_norm": 3.061650276184082, "learning_rate": 4.2103907055185986e-05, "loss": 6.9318, "step": 17230 }, { "epoch": 0.7264793304804267, "grad_norm": 2.3138885498046875, "learning_rate": 4.210312626877792e-05, "loss": 6.4863, "step": 17231 }, { "epoch": 0.7265214916626263, "grad_norm": 1.1340739727020264, "learning_rate": 4.210234548236984e-05, "loss": 6.3271, "step": 17232 }, { "epoch": 0.7265636528448257, "grad_norm": 1.204712986946106, "learning_rate": 4.2101564695961773e-05, "loss": 6.4552, "step": 17233 }, { "epoch": 0.7266058140270253, "grad_norm": 1.35707688331604, "learning_rate": 4.2100783909553705e-05, "loss": 6.3811, "step": 17234 }, { "epoch": 0.7266479752092249, "grad_norm": 3.6668879985809326, "learning_rate": 4.210000312314563e-05, "loss": 7.3601, "step": 17235 }, { "epoch": 0.7266901363914244, "grad_norm": 1.3338156938552856, "learning_rate": 4.209922233673757e-05, "loss": 6.4437, "step": 17236 }, { "epoch": 0.726732297573624, "grad_norm": 1.5221788883209229, "learning_rate": 4.209844155032949e-05, "loss": 6.8821, "step": 17237 }, { "epoch": 0.7267744587558235, "grad_norm": 2.7124383449554443, "learning_rate": 4.209766076392142e-05, "loss": 6.6125, "step": 17238 }, { "epoch": 0.726816619938023, "grad_norm": 4.369894981384277, "learning_rate": 4.2096879977513354e-05, "loss": 9.2214, "step": 17239 }, { "epoch": 0.7268587811202226, "grad_norm": 1.5140841007232666, "learning_rate": 4.209609919110528e-05, "loss": 6.6266, "step": 17240 }, { "epoch": 0.7269009423024222, "grad_norm": 1.6243575811386108, "learning_rate": 4.209531840469722e-05, "loss": 7.0775, "step": 17241 }, { "epoch": 0.7269431034846217, "grad_norm": 4.170954704284668, "learning_rate": 4.209453761828914e-05, "loss": 9.8127, "step": 17242 }, { "epoch": 0.7269852646668212, "grad_norm": 3.0491855144500732, "learning_rate": 4.209375683188107e-05, "loss": 6.3314, "step": 17243 }, { "epoch": 0.7270274258490208, "grad_norm": 1.5446540117263794, "learning_rate": 4.2092976045473004e-05, "loss": 6.4788, "step": 17244 }, { "epoch": 0.7270695870312204, "grad_norm": 1.7084473371505737, "learning_rate": 4.209219525906493e-05, "loss": 6.7101, "step": 17245 }, { "epoch": 0.72711174821342, "grad_norm": 1.4874436855316162, "learning_rate": 4.209141447265686e-05, "loss": 6.5007, "step": 17246 }, { "epoch": 0.7271539093956194, "grad_norm": 2.7720367908477783, "learning_rate": 4.209063368624879e-05, "loss": 6.4657, "step": 17247 }, { "epoch": 0.727196070577819, "grad_norm": 2.596198081970215, "learning_rate": 4.208985289984072e-05, "loss": 6.6096, "step": 17248 }, { "epoch": 0.7272382317600186, "grad_norm": 2.196664810180664, "learning_rate": 4.208907211343265e-05, "loss": 6.8544, "step": 17249 }, { "epoch": 0.7272803929422181, "grad_norm": 1.8926572799682617, "learning_rate": 4.2088291327024585e-05, "loss": 7.9621, "step": 17250 }, { "epoch": 0.7273225541244176, "grad_norm": 1.9835385084152222, "learning_rate": 4.208751054061651e-05, "loss": 6.3988, "step": 17251 }, { "epoch": 0.7273647153066172, "grad_norm": 1.6838018894195557, "learning_rate": 4.208672975420844e-05, "loss": 6.4491, "step": 17252 }, { "epoch": 0.7274068764888167, "grad_norm": 2.546402931213379, "learning_rate": 4.208594896780037e-05, "loss": 7.2196, "step": 17253 }, { "epoch": 0.7274490376710163, "grad_norm": 2.4346559047698975, "learning_rate": 4.2085168181392296e-05, "loss": 6.9473, "step": 17254 }, { "epoch": 0.7274911988532159, "grad_norm": 1.8362936973571777, "learning_rate": 4.2084387394984234e-05, "loss": 6.6662, "step": 17255 }, { "epoch": 0.7275333600354154, "grad_norm": 1.4054384231567383, "learning_rate": 4.208360660857616e-05, "loss": 6.475, "step": 17256 }, { "epoch": 0.7275755212176149, "grad_norm": 2.1592726707458496, "learning_rate": 4.208282582216809e-05, "loss": 6.8802, "step": 17257 }, { "epoch": 0.7276176823998145, "grad_norm": 1.8839912414550781, "learning_rate": 4.208204503576002e-05, "loss": 6.5721, "step": 17258 }, { "epoch": 0.7276598435820141, "grad_norm": 2.3931777477264404, "learning_rate": 4.2081264249351946e-05, "loss": 7.0024, "step": 17259 }, { "epoch": 0.7277020047642135, "grad_norm": 3.3550620079040527, "learning_rate": 4.2080483462943884e-05, "loss": 7.7619, "step": 17260 }, { "epoch": 0.7277441659464131, "grad_norm": 3.111591339111328, "learning_rate": 4.207970267653581e-05, "loss": 7.806, "step": 17261 }, { "epoch": 0.7277863271286127, "grad_norm": 2.7511353492736816, "learning_rate": 4.207892189012774e-05, "loss": 7.2024, "step": 17262 }, { "epoch": 0.7278284883108123, "grad_norm": 3.490800380706787, "learning_rate": 4.207814110371967e-05, "loss": 6.4559, "step": 17263 }, { "epoch": 0.7278706494930118, "grad_norm": 2.7726504802703857, "learning_rate": 4.2077360317311595e-05, "loss": 6.3134, "step": 17264 }, { "epoch": 0.7279128106752113, "grad_norm": 1.419228434562683, "learning_rate": 4.2076579530903526e-05, "loss": 6.4559, "step": 17265 }, { "epoch": 0.7279549718574109, "grad_norm": 1.838002324104309, "learning_rate": 4.207579874449546e-05, "loss": 6.6181, "step": 17266 }, { "epoch": 0.7279971330396104, "grad_norm": 2.601670742034912, "learning_rate": 4.207501795808739e-05, "loss": 6.6003, "step": 17267 }, { "epoch": 0.72803929422181, "grad_norm": 6.198034763336182, "learning_rate": 4.207423717167931e-05, "loss": 7.6839, "step": 17268 }, { "epoch": 0.7280814554040095, "grad_norm": 4.866711616516113, "learning_rate": 4.207345638527125e-05, "loss": 7.543, "step": 17269 }, { "epoch": 0.728123616586209, "grad_norm": 3.3356759548187256, "learning_rate": 4.2072675598863176e-05, "loss": 6.7535, "step": 17270 }, { "epoch": 0.7281657777684086, "grad_norm": 1.4299468994140625, "learning_rate": 4.207189481245511e-05, "loss": 6.5273, "step": 17271 }, { "epoch": 0.7282079389506082, "grad_norm": 2.0475778579711914, "learning_rate": 4.207111402604704e-05, "loss": 6.9531, "step": 17272 }, { "epoch": 0.7282501001328078, "grad_norm": 2.272672653198242, "learning_rate": 4.207033323963896e-05, "loss": 6.6012, "step": 17273 }, { "epoch": 0.7282922613150072, "grad_norm": 1.4342892169952393, "learning_rate": 4.20695524532309e-05, "loss": 6.3869, "step": 17274 }, { "epoch": 0.7283344224972068, "grad_norm": 1.2516193389892578, "learning_rate": 4.2068771666822825e-05, "loss": 6.5323, "step": 17275 }, { "epoch": 0.7283765836794064, "grad_norm": 3.879472494125366, "learning_rate": 4.206799088041476e-05, "loss": 7.7394, "step": 17276 }, { "epoch": 0.728418744861606, "grad_norm": 2.692685842514038, "learning_rate": 4.206721009400669e-05, "loss": 7.12, "step": 17277 }, { "epoch": 0.7284609060438054, "grad_norm": 1.03376305103302, "learning_rate": 4.206642930759861e-05, "loss": 6.4928, "step": 17278 }, { "epoch": 0.728503067226005, "grad_norm": 1.4863437414169312, "learning_rate": 4.2065648521190544e-05, "loss": 7.0438, "step": 17279 }, { "epoch": 0.7285452284082046, "grad_norm": 3.1433470249176025, "learning_rate": 4.2064867734782475e-05, "loss": 7.6235, "step": 17280 }, { "epoch": 0.7285873895904041, "grad_norm": 1.5099613666534424, "learning_rate": 4.2064086948374406e-05, "loss": 6.8926, "step": 17281 }, { "epoch": 0.7286295507726037, "grad_norm": 1.6734951734542847, "learning_rate": 4.206330616196634e-05, "loss": 7.0184, "step": 17282 }, { "epoch": 0.7286717119548032, "grad_norm": 1.171536922454834, "learning_rate": 4.206252537555826e-05, "loss": 6.678, "step": 17283 }, { "epoch": 0.7287138731370028, "grad_norm": 3.3971059322357178, "learning_rate": 4.206174458915019e-05, "loss": 7.5622, "step": 17284 }, { "epoch": 0.7287560343192023, "grad_norm": 2.5357778072357178, "learning_rate": 4.2060963802742124e-05, "loss": 7.5564, "step": 17285 }, { "epoch": 0.7287981955014019, "grad_norm": 1.486248254776001, "learning_rate": 4.2060183016334056e-05, "loss": 6.9579, "step": 17286 }, { "epoch": 0.7288403566836014, "grad_norm": 1.7093805074691772, "learning_rate": 4.205940222992598e-05, "loss": 6.7541, "step": 17287 }, { "epoch": 0.7288825178658009, "grad_norm": 1.1687774658203125, "learning_rate": 4.205862144351792e-05, "loss": 7.0117, "step": 17288 }, { "epoch": 0.7289246790480005, "grad_norm": 2.726574182510376, "learning_rate": 4.205784065710984e-05, "loss": 6.7048, "step": 17289 }, { "epoch": 0.7289668402302001, "grad_norm": 3.074979543685913, "learning_rate": 4.205705987070177e-05, "loss": 6.7216, "step": 17290 }, { "epoch": 0.7290090014123995, "grad_norm": 2.0921034812927246, "learning_rate": 4.2056279084293705e-05, "loss": 7.4668, "step": 17291 }, { "epoch": 0.7290511625945991, "grad_norm": 1.7283868789672852, "learning_rate": 4.205549829788563e-05, "loss": 6.6737, "step": 17292 }, { "epoch": 0.7290933237767987, "grad_norm": 1.2945011854171753, "learning_rate": 4.205471751147757e-05, "loss": 6.8144, "step": 17293 }, { "epoch": 0.7291354849589983, "grad_norm": 1.8567938804626465, "learning_rate": 4.205393672506949e-05, "loss": 6.6491, "step": 17294 }, { "epoch": 0.7291776461411978, "grad_norm": 1.6059213876724243, "learning_rate": 4.2053155938661423e-05, "loss": 6.7199, "step": 17295 }, { "epoch": 0.7292198073233973, "grad_norm": 1.4362750053405762, "learning_rate": 4.2052375152253355e-05, "loss": 6.4508, "step": 17296 }, { "epoch": 0.7292619685055969, "grad_norm": 2.208775043487549, "learning_rate": 4.205159436584528e-05, "loss": 6.9446, "step": 17297 }, { "epoch": 0.7293041296877965, "grad_norm": 4.036418437957764, "learning_rate": 4.205081357943721e-05, "loss": 6.3963, "step": 17298 }, { "epoch": 0.729346290869996, "grad_norm": 1.887190341949463, "learning_rate": 4.205003279302914e-05, "loss": 6.647, "step": 17299 }, { "epoch": 0.7293884520521955, "grad_norm": 1.8026127815246582, "learning_rate": 4.204925200662107e-05, "loss": 6.9863, "step": 17300 }, { "epoch": 0.7294306132343951, "grad_norm": 1.474281668663025, "learning_rate": 4.2048471220213e-05, "loss": 6.9775, "step": 17301 }, { "epoch": 0.7294727744165946, "grad_norm": 1.648328185081482, "learning_rate": 4.204769043380493e-05, "loss": 6.856, "step": 17302 }, { "epoch": 0.7295149355987942, "grad_norm": 0.9724710583686829, "learning_rate": 4.204690964739686e-05, "loss": 6.9417, "step": 17303 }, { "epoch": 0.7295570967809938, "grad_norm": 2.0438411235809326, "learning_rate": 4.204612886098879e-05, "loss": 6.8545, "step": 17304 }, { "epoch": 0.7295992579631932, "grad_norm": 4.175956726074219, "learning_rate": 4.204534807458072e-05, "loss": 7.8109, "step": 17305 }, { "epoch": 0.7296414191453928, "grad_norm": 3.5601327419281006, "learning_rate": 4.204456728817265e-05, "loss": 7.3872, "step": 17306 }, { "epoch": 0.7296835803275924, "grad_norm": 1.7516894340515137, "learning_rate": 4.204378650176458e-05, "loss": 6.5731, "step": 17307 }, { "epoch": 0.729725741509792, "grad_norm": 1.0158389806747437, "learning_rate": 4.204300571535651e-05, "loss": 6.8442, "step": 17308 }, { "epoch": 0.7297679026919914, "grad_norm": 4.058478832244873, "learning_rate": 4.2042224928948434e-05, "loss": 7.6215, "step": 17309 }, { "epoch": 0.729810063874191, "grad_norm": 1.3599867820739746, "learning_rate": 4.204144414254037e-05, "loss": 6.8888, "step": 17310 }, { "epoch": 0.7298522250563906, "grad_norm": 1.2340980768203735, "learning_rate": 4.2040663356132297e-05, "loss": 6.5702, "step": 17311 }, { "epoch": 0.7298943862385902, "grad_norm": 3.5923094749450684, "learning_rate": 4.203988256972423e-05, "loss": 7.3097, "step": 17312 }, { "epoch": 0.7299365474207897, "grad_norm": 3.1436431407928467, "learning_rate": 4.203910178331616e-05, "loss": 7.2036, "step": 17313 }, { "epoch": 0.7299787086029892, "grad_norm": 1.6466896533966064, "learning_rate": 4.203832099690809e-05, "loss": 6.3374, "step": 17314 }, { "epoch": 0.7300208697851888, "grad_norm": 1.5307984352111816, "learning_rate": 4.2037540210500015e-05, "loss": 6.5997, "step": 17315 }, { "epoch": 0.7300630309673883, "grad_norm": 2.726477861404419, "learning_rate": 4.2036759424091946e-05, "loss": 6.3085, "step": 17316 }, { "epoch": 0.7301051921495879, "grad_norm": 2.0369420051574707, "learning_rate": 4.203597863768388e-05, "loss": 6.5588, "step": 17317 }, { "epoch": 0.7301473533317874, "grad_norm": 2.321575164794922, "learning_rate": 4.203519785127581e-05, "loss": 6.6181, "step": 17318 }, { "epoch": 0.730189514513987, "grad_norm": 1.8661497831344604, "learning_rate": 4.203441706486774e-05, "loss": 7.1483, "step": 17319 }, { "epoch": 0.7302316756961865, "grad_norm": 2.6237502098083496, "learning_rate": 4.2033636278459664e-05, "loss": 6.9983, "step": 17320 }, { "epoch": 0.7302738368783861, "grad_norm": 2.4337406158447266, "learning_rate": 4.2032855492051596e-05, "loss": 7.7847, "step": 17321 }, { "epoch": 0.7303159980605857, "grad_norm": 2.9705348014831543, "learning_rate": 4.203207470564353e-05, "loss": 8.013, "step": 17322 }, { "epoch": 0.7303581592427851, "grad_norm": 3.01772141456604, "learning_rate": 4.203129391923545e-05, "loss": 6.728, "step": 17323 }, { "epoch": 0.7304003204249847, "grad_norm": 3.67283034324646, "learning_rate": 4.203051313282739e-05, "loss": 7.3195, "step": 17324 }, { "epoch": 0.7304424816071843, "grad_norm": 3.668975830078125, "learning_rate": 4.2029732346419314e-05, "loss": 7.275, "step": 17325 }, { "epoch": 0.7304846427893839, "grad_norm": 2.5866801738739014, "learning_rate": 4.2028951560011245e-05, "loss": 7.2933, "step": 17326 }, { "epoch": 0.7305268039715833, "grad_norm": 4.953227519989014, "learning_rate": 4.2028170773603176e-05, "loss": 8.5092, "step": 17327 }, { "epoch": 0.7305689651537829, "grad_norm": 2.108051300048828, "learning_rate": 4.20273899871951e-05, "loss": 6.8178, "step": 17328 }, { "epoch": 0.7306111263359825, "grad_norm": 2.0652594566345215, "learning_rate": 4.202660920078704e-05, "loss": 6.5982, "step": 17329 }, { "epoch": 0.730653287518182, "grad_norm": 2.487338066101074, "learning_rate": 4.202582841437896e-05, "loss": 7.2202, "step": 17330 }, { "epoch": 0.7306954487003816, "grad_norm": 2.993497848510742, "learning_rate": 4.2025047627970895e-05, "loss": 7.5503, "step": 17331 }, { "epoch": 0.7307376098825811, "grad_norm": 3.135133981704712, "learning_rate": 4.2024266841562826e-05, "loss": 6.2671, "step": 17332 }, { "epoch": 0.7307797710647806, "grad_norm": 2.4601449966430664, "learning_rate": 4.202348605515476e-05, "loss": 7.0018, "step": 17333 }, { "epoch": 0.7308219322469802, "grad_norm": 2.5007779598236084, "learning_rate": 4.202270526874668e-05, "loss": 6.6012, "step": 17334 }, { "epoch": 0.7308640934291798, "grad_norm": 1.5294992923736572, "learning_rate": 4.202192448233861e-05, "loss": 6.6431, "step": 17335 }, { "epoch": 0.7309062546113793, "grad_norm": 1.7167736291885376, "learning_rate": 4.2021143695930544e-05, "loss": 6.7877, "step": 17336 }, { "epoch": 0.7309484157935788, "grad_norm": 1.9226911067962646, "learning_rate": 4.202036290952247e-05, "loss": 6.4622, "step": 17337 }, { "epoch": 0.7309905769757784, "grad_norm": 1.4050495624542236, "learning_rate": 4.201958212311441e-05, "loss": 6.4323, "step": 17338 }, { "epoch": 0.731032738157978, "grad_norm": 2.9480996131896973, "learning_rate": 4.201880133670633e-05, "loss": 6.5483, "step": 17339 }, { "epoch": 0.7310748993401776, "grad_norm": 2.219590663909912, "learning_rate": 4.201802055029826e-05, "loss": 6.8849, "step": 17340 }, { "epoch": 0.731117060522377, "grad_norm": 1.8572900295257568, "learning_rate": 4.2017239763890194e-05, "loss": 7.0395, "step": 17341 }, { "epoch": 0.7311592217045766, "grad_norm": 3.341522455215454, "learning_rate": 4.201645897748212e-05, "loss": 6.8934, "step": 17342 }, { "epoch": 0.7312013828867762, "grad_norm": 1.6789604425430298, "learning_rate": 4.2015678191074056e-05, "loss": 6.3739, "step": 17343 }, { "epoch": 0.7312435440689757, "grad_norm": 1.2167195081710815, "learning_rate": 4.201489740466598e-05, "loss": 6.6415, "step": 17344 }, { "epoch": 0.7312857052511752, "grad_norm": 4.979882717132568, "learning_rate": 4.201411661825791e-05, "loss": 7.6778, "step": 17345 }, { "epoch": 0.7313278664333748, "grad_norm": 2.7057433128356934, "learning_rate": 4.201333583184984e-05, "loss": 7.1357, "step": 17346 }, { "epoch": 0.7313700276155743, "grad_norm": 1.370352864265442, "learning_rate": 4.201255504544177e-05, "loss": 6.374, "step": 17347 }, { "epoch": 0.7314121887977739, "grad_norm": 1.673128366470337, "learning_rate": 4.20117742590337e-05, "loss": 6.6433, "step": 17348 }, { "epoch": 0.7314543499799734, "grad_norm": 1.3595322370529175, "learning_rate": 4.201099347262563e-05, "loss": 6.6304, "step": 17349 }, { "epoch": 0.731496511162173, "grad_norm": 2.066668748855591, "learning_rate": 4.201021268621756e-05, "loss": 6.8797, "step": 17350 }, { "epoch": 0.7315386723443725, "grad_norm": 3.9016592502593994, "learning_rate": 4.200943189980949e-05, "loss": 7.7957, "step": 17351 }, { "epoch": 0.7315808335265721, "grad_norm": 1.1708571910858154, "learning_rate": 4.200865111340142e-05, "loss": 6.6217, "step": 17352 }, { "epoch": 0.7316229947087717, "grad_norm": 1.8760647773742676, "learning_rate": 4.200787032699335e-05, "loss": 6.6528, "step": 17353 }, { "epoch": 0.7316651558909711, "grad_norm": 2.7705817222595215, "learning_rate": 4.200708954058528e-05, "loss": 6.7713, "step": 17354 }, { "epoch": 0.7317073170731707, "grad_norm": 1.566423773765564, "learning_rate": 4.200630875417721e-05, "loss": 6.3926, "step": 17355 }, { "epoch": 0.7317494782553703, "grad_norm": 1.748746395111084, "learning_rate": 4.2005527967769135e-05, "loss": 6.4837, "step": 17356 }, { "epoch": 0.7317916394375699, "grad_norm": 1.8167184591293335, "learning_rate": 4.2004747181361074e-05, "loss": 7.1681, "step": 17357 }, { "epoch": 0.7318338006197693, "grad_norm": 1.2554140090942383, "learning_rate": 4.2003966394953e-05, "loss": 6.4772, "step": 17358 }, { "epoch": 0.7318759618019689, "grad_norm": 3.094719886779785, "learning_rate": 4.200318560854493e-05, "loss": 7.6114, "step": 17359 }, { "epoch": 0.7319181229841685, "grad_norm": 2.067809581756592, "learning_rate": 4.200240482213686e-05, "loss": 6.8433, "step": 17360 }, { "epoch": 0.731960284166368, "grad_norm": 1.3239188194274902, "learning_rate": 4.2001624035728785e-05, "loss": 6.5851, "step": 17361 }, { "epoch": 0.7320024453485676, "grad_norm": 2.7325336933135986, "learning_rate": 4.200084324932072e-05, "loss": 6.4315, "step": 17362 }, { "epoch": 0.7320446065307671, "grad_norm": 1.4060980081558228, "learning_rate": 4.200006246291265e-05, "loss": 6.6209, "step": 17363 }, { "epoch": 0.7320867677129667, "grad_norm": 1.737301230430603, "learning_rate": 4.199928167650458e-05, "loss": 6.9552, "step": 17364 }, { "epoch": 0.7321289288951662, "grad_norm": 2.9379208087921143, "learning_rate": 4.199850089009651e-05, "loss": 7.4996, "step": 17365 }, { "epoch": 0.7321710900773658, "grad_norm": 3.0261852741241455, "learning_rate": 4.1997720103688435e-05, "loss": 6.3538, "step": 17366 }, { "epoch": 0.7322132512595653, "grad_norm": 1.1939773559570312, "learning_rate": 4.1996939317280366e-05, "loss": 6.8896, "step": 17367 }, { "epoch": 0.7322554124417648, "grad_norm": 1.2049064636230469, "learning_rate": 4.19961585308723e-05, "loss": 6.4578, "step": 17368 }, { "epoch": 0.7322975736239644, "grad_norm": 1.4089806079864502, "learning_rate": 4.199537774446423e-05, "loss": 7.0222, "step": 17369 }, { "epoch": 0.732339734806164, "grad_norm": 1.9626415967941284, "learning_rate": 4.199459695805615e-05, "loss": 6.78, "step": 17370 }, { "epoch": 0.7323818959883636, "grad_norm": 1.7143990993499756, "learning_rate": 4.1993816171648084e-05, "loss": 6.9909, "step": 17371 }, { "epoch": 0.732424057170563, "grad_norm": 1.520140528678894, "learning_rate": 4.1993035385240015e-05, "loss": 6.4914, "step": 17372 }, { "epoch": 0.7324662183527626, "grad_norm": 2.4405384063720703, "learning_rate": 4.1992254598831947e-05, "loss": 7.0006, "step": 17373 }, { "epoch": 0.7325083795349622, "grad_norm": 2.204698085784912, "learning_rate": 4.199147381242388e-05, "loss": 6.4495, "step": 17374 }, { "epoch": 0.7325505407171617, "grad_norm": 2.040191888809204, "learning_rate": 4.19906930260158e-05, "loss": 6.3912, "step": 17375 }, { "epoch": 0.7325927018993612, "grad_norm": 2.517347812652588, "learning_rate": 4.198991223960774e-05, "loss": 7.434, "step": 17376 }, { "epoch": 0.7326348630815608, "grad_norm": 2.9624595642089844, "learning_rate": 4.1989131453199665e-05, "loss": 7.0696, "step": 17377 }, { "epoch": 0.7326770242637604, "grad_norm": 3.3382773399353027, "learning_rate": 4.198835066679159e-05, "loss": 6.9393, "step": 17378 }, { "epoch": 0.7327191854459599, "grad_norm": 3.706496477127075, "learning_rate": 4.198756988038353e-05, "loss": 6.9893, "step": 17379 }, { "epoch": 0.7327613466281595, "grad_norm": 1.2905174493789673, "learning_rate": 4.198678909397545e-05, "loss": 6.4561, "step": 17380 }, { "epoch": 0.732803507810359, "grad_norm": 1.7509692907333374, "learning_rate": 4.198600830756738e-05, "loss": 6.4576, "step": 17381 }, { "epoch": 0.7328456689925585, "grad_norm": 2.5797178745269775, "learning_rate": 4.1985227521159314e-05, "loss": 6.4446, "step": 17382 }, { "epoch": 0.7328878301747581, "grad_norm": 2.4694061279296875, "learning_rate": 4.1984446734751246e-05, "loss": 6.4534, "step": 17383 }, { "epoch": 0.7329299913569577, "grad_norm": 1.5395495891571045, "learning_rate": 4.198366594834318e-05, "loss": 6.6413, "step": 17384 }, { "epoch": 0.7329721525391572, "grad_norm": 4.30246114730835, "learning_rate": 4.19828851619351e-05, "loss": 7.5906, "step": 17385 }, { "epoch": 0.7330143137213567, "grad_norm": 4.8625264167785645, "learning_rate": 4.198210437552703e-05, "loss": 7.5789, "step": 17386 }, { "epoch": 0.7330564749035563, "grad_norm": 3.461702585220337, "learning_rate": 4.1981323589118964e-05, "loss": 7.7272, "step": 17387 }, { "epoch": 0.7330986360857559, "grad_norm": 1.5432568788528442, "learning_rate": 4.1980542802710895e-05, "loss": 6.6522, "step": 17388 }, { "epoch": 0.7331407972679554, "grad_norm": 1.6602216958999634, "learning_rate": 4.197976201630282e-05, "loss": 7.2939, "step": 17389 }, { "epoch": 0.7331829584501549, "grad_norm": 3.63208270072937, "learning_rate": 4.197898122989475e-05, "loss": 7.6232, "step": 17390 }, { "epoch": 0.7332251196323545, "grad_norm": 1.969191074371338, "learning_rate": 4.197820044348668e-05, "loss": 6.8831, "step": 17391 }, { "epoch": 0.7332672808145541, "grad_norm": 1.6784119606018066, "learning_rate": 4.197741965707861e-05, "loss": 6.8036, "step": 17392 }, { "epoch": 0.7333094419967536, "grad_norm": 3.0110199451446533, "learning_rate": 4.1976638870670545e-05, "loss": 7.2293, "step": 17393 }, { "epoch": 0.7333516031789531, "grad_norm": 1.4967765808105469, "learning_rate": 4.197585808426247e-05, "loss": 6.5303, "step": 17394 }, { "epoch": 0.7333937643611527, "grad_norm": 3.0787675380706787, "learning_rate": 4.19750772978544e-05, "loss": 6.9001, "step": 17395 }, { "epoch": 0.7334359255433522, "grad_norm": 3.10750675201416, "learning_rate": 4.197429651144633e-05, "loss": 6.5367, "step": 17396 }, { "epoch": 0.7334780867255518, "grad_norm": 2.4146552085876465, "learning_rate": 4.1973515725038256e-05, "loss": 6.7543, "step": 17397 }, { "epoch": 0.7335202479077514, "grad_norm": 2.1940319538116455, "learning_rate": 4.1972734938630194e-05, "loss": 6.8624, "step": 17398 }, { "epoch": 0.7335624090899509, "grad_norm": 2.2101941108703613, "learning_rate": 4.197195415222212e-05, "loss": 6.3725, "step": 17399 }, { "epoch": 0.7336045702721504, "grad_norm": 3.789477586746216, "learning_rate": 4.197117336581405e-05, "loss": 7.7505, "step": 17400 }, { "epoch": 0.73364673145435, "grad_norm": 3.0069661140441895, "learning_rate": 4.197039257940598e-05, "loss": 6.8282, "step": 17401 }, { "epoch": 0.7336888926365496, "grad_norm": 2.555986166000366, "learning_rate": 4.196961179299791e-05, "loss": 6.6808, "step": 17402 }, { "epoch": 0.733731053818749, "grad_norm": 2.968676805496216, "learning_rate": 4.196883100658984e-05, "loss": 6.3769, "step": 17403 }, { "epoch": 0.7337732150009486, "grad_norm": 2.2972726821899414, "learning_rate": 4.196805022018177e-05, "loss": 6.4427, "step": 17404 }, { "epoch": 0.7338153761831482, "grad_norm": 1.3593840599060059, "learning_rate": 4.19672694337737e-05, "loss": 6.7156, "step": 17405 }, { "epoch": 0.7338575373653478, "grad_norm": 3.9138503074645996, "learning_rate": 4.1966488647365624e-05, "loss": 7.2505, "step": 17406 }, { "epoch": 0.7338996985475472, "grad_norm": 1.821163296699524, "learning_rate": 4.196570786095756e-05, "loss": 6.3698, "step": 17407 }, { "epoch": 0.7339418597297468, "grad_norm": 2.586885690689087, "learning_rate": 4.1964927074549486e-05, "loss": 7.2515, "step": 17408 }, { "epoch": 0.7339840209119464, "grad_norm": 0.9876261949539185, "learning_rate": 4.196414628814142e-05, "loss": 6.2432, "step": 17409 }, { "epoch": 0.7340261820941459, "grad_norm": 1.8437285423278809, "learning_rate": 4.196336550173335e-05, "loss": 6.5322, "step": 17410 }, { "epoch": 0.7340683432763455, "grad_norm": 2.2939722537994385, "learning_rate": 4.1962584715325273e-05, "loss": 6.5775, "step": 17411 }, { "epoch": 0.734110504458545, "grad_norm": 1.790367603302002, "learning_rate": 4.196180392891721e-05, "loss": 6.3701, "step": 17412 }, { "epoch": 0.7341526656407446, "grad_norm": 2.702892780303955, "learning_rate": 4.1961023142509136e-05, "loss": 7.0863, "step": 17413 }, { "epoch": 0.7341948268229441, "grad_norm": 1.0026592016220093, "learning_rate": 4.196024235610107e-05, "loss": 6.6775, "step": 17414 }, { "epoch": 0.7342369880051437, "grad_norm": 1.958909511566162, "learning_rate": 4.1959461569693e-05, "loss": 7.039, "step": 17415 }, { "epoch": 0.7342791491873432, "grad_norm": 2.3054699897766113, "learning_rate": 4.195868078328492e-05, "loss": 7.1262, "step": 17416 }, { "epoch": 0.7343213103695427, "grad_norm": 2.6203298568725586, "learning_rate": 4.1957899996876854e-05, "loss": 8.6154, "step": 17417 }, { "epoch": 0.7343634715517423, "grad_norm": 2.092926025390625, "learning_rate": 4.1957119210468785e-05, "loss": 6.6864, "step": 17418 }, { "epoch": 0.7344056327339419, "grad_norm": 2.2806785106658936, "learning_rate": 4.195633842406072e-05, "loss": 6.9894, "step": 17419 }, { "epoch": 0.7344477939161415, "grad_norm": 3.902425527572632, "learning_rate": 4.195555763765265e-05, "loss": 7.2773, "step": 17420 }, { "epoch": 0.7344899550983409, "grad_norm": 1.7784690856933594, "learning_rate": 4.195477685124458e-05, "loss": 7.2847, "step": 17421 }, { "epoch": 0.7345321162805405, "grad_norm": 1.6702752113342285, "learning_rate": 4.1953996064836504e-05, "loss": 6.3529, "step": 17422 }, { "epoch": 0.7345742774627401, "grad_norm": 1.2197381258010864, "learning_rate": 4.1953215278428435e-05, "loss": 6.2064, "step": 17423 }, { "epoch": 0.7346164386449396, "grad_norm": 2.824897527694702, "learning_rate": 4.1952434492020366e-05, "loss": 7.6347, "step": 17424 }, { "epoch": 0.7346585998271391, "grad_norm": 2.480532169342041, "learning_rate": 4.195165370561229e-05, "loss": 7.0537, "step": 17425 }, { "epoch": 0.7347007610093387, "grad_norm": 2.0340826511383057, "learning_rate": 4.195087291920423e-05, "loss": 6.5539, "step": 17426 }, { "epoch": 0.7347429221915383, "grad_norm": 2.317720413208008, "learning_rate": 4.195009213279615e-05, "loss": 7.4376, "step": 17427 }, { "epoch": 0.7347850833737378, "grad_norm": 2.4901022911071777, "learning_rate": 4.1949311346388085e-05, "loss": 6.4619, "step": 17428 }, { "epoch": 0.7348272445559374, "grad_norm": 2.1866421699523926, "learning_rate": 4.1948530559980016e-05, "loss": 6.7831, "step": 17429 }, { "epoch": 0.7348694057381369, "grad_norm": 2.2878646850585938, "learning_rate": 4.194774977357194e-05, "loss": 6.7058, "step": 17430 }, { "epoch": 0.7349115669203364, "grad_norm": 1.020917296409607, "learning_rate": 4.194696898716388e-05, "loss": 6.5123, "step": 17431 }, { "epoch": 0.734953728102536, "grad_norm": 1.416793942451477, "learning_rate": 4.19461882007558e-05, "loss": 6.8967, "step": 17432 }, { "epoch": 0.7349958892847356, "grad_norm": 2.0127859115600586, "learning_rate": 4.1945407414347734e-05, "loss": 7.0135, "step": 17433 }, { "epoch": 0.735038050466935, "grad_norm": 1.3725184202194214, "learning_rate": 4.1944626627939665e-05, "loss": 6.8797, "step": 17434 }, { "epoch": 0.7350802116491346, "grad_norm": 1.7476485967636108, "learning_rate": 4.194384584153159e-05, "loss": 7.3054, "step": 17435 }, { "epoch": 0.7351223728313342, "grad_norm": 1.7576286792755127, "learning_rate": 4.194306505512352e-05, "loss": 6.3612, "step": 17436 }, { "epoch": 0.7351645340135338, "grad_norm": 3.051631212234497, "learning_rate": 4.194228426871545e-05, "loss": 6.3387, "step": 17437 }, { "epoch": 0.7352066951957333, "grad_norm": 2.6514461040496826, "learning_rate": 4.1941503482307384e-05, "loss": 6.6416, "step": 17438 }, { "epoch": 0.7352488563779328, "grad_norm": 1.4770803451538086, "learning_rate": 4.194072269589931e-05, "loss": 6.5786, "step": 17439 }, { "epoch": 0.7352910175601324, "grad_norm": 0.9852250218391418, "learning_rate": 4.193994190949124e-05, "loss": 6.3537, "step": 17440 }, { "epoch": 0.735333178742332, "grad_norm": 2.3899505138397217, "learning_rate": 4.193916112308317e-05, "loss": 6.9458, "step": 17441 }, { "epoch": 0.7353753399245315, "grad_norm": 1.9238011837005615, "learning_rate": 4.19383803366751e-05, "loss": 6.9869, "step": 17442 }, { "epoch": 0.735417501106731, "grad_norm": 1.6945894956588745, "learning_rate": 4.193759955026703e-05, "loss": 6.9656, "step": 17443 }, { "epoch": 0.7354596622889306, "grad_norm": 2.8905813694000244, "learning_rate": 4.193681876385896e-05, "loss": 6.8933, "step": 17444 }, { "epoch": 0.7355018234711301, "grad_norm": 1.9058773517608643, "learning_rate": 4.1936037977450896e-05, "loss": 6.7932, "step": 17445 }, { "epoch": 0.7355439846533297, "grad_norm": 1.7647976875305176, "learning_rate": 4.193525719104282e-05, "loss": 6.4784, "step": 17446 }, { "epoch": 0.7355861458355293, "grad_norm": 3.580303192138672, "learning_rate": 4.193447640463475e-05, "loss": 7.3555, "step": 17447 }, { "epoch": 0.7356283070177287, "grad_norm": 2.8149757385253906, "learning_rate": 4.193369561822668e-05, "loss": 6.555, "step": 17448 }, { "epoch": 0.7356704681999283, "grad_norm": 2.266272783279419, "learning_rate": 4.193291483181861e-05, "loss": 6.6994, "step": 17449 }, { "epoch": 0.7357126293821279, "grad_norm": 3.1142940521240234, "learning_rate": 4.193213404541054e-05, "loss": 9.0817, "step": 17450 }, { "epoch": 0.7357547905643275, "grad_norm": 1.6111029386520386, "learning_rate": 4.193135325900247e-05, "loss": 6.4623, "step": 17451 }, { "epoch": 0.7357969517465269, "grad_norm": 1.923100233078003, "learning_rate": 4.19305724725944e-05, "loss": 6.5633, "step": 17452 }, { "epoch": 0.7358391129287265, "grad_norm": 2.2794454097747803, "learning_rate": 4.192979168618633e-05, "loss": 7.5684, "step": 17453 }, { "epoch": 0.7358812741109261, "grad_norm": 3.277163505554199, "learning_rate": 4.192901089977826e-05, "loss": 7.4975, "step": 17454 }, { "epoch": 0.7359234352931256, "grad_norm": 2.2372493743896484, "learning_rate": 4.192823011337019e-05, "loss": 7.0777, "step": 17455 }, { "epoch": 0.7359655964753252, "grad_norm": 1.9005237817764282, "learning_rate": 4.192744932696212e-05, "loss": 6.8619, "step": 17456 }, { "epoch": 0.7360077576575247, "grad_norm": 1.5004198551177979, "learning_rate": 4.192666854055405e-05, "loss": 6.7456, "step": 17457 }, { "epoch": 0.7360499188397243, "grad_norm": 1.23131263256073, "learning_rate": 4.1925887754145975e-05, "loss": 6.4453, "step": 17458 }, { "epoch": 0.7360920800219238, "grad_norm": 2.7632148265838623, "learning_rate": 4.1925106967737906e-05, "loss": 7.3147, "step": 17459 }, { "epoch": 0.7361342412041234, "grad_norm": 1.9658052921295166, "learning_rate": 4.192432618132984e-05, "loss": 6.3118, "step": 17460 }, { "epoch": 0.7361764023863229, "grad_norm": 2.7749969959259033, "learning_rate": 4.192354539492176e-05, "loss": 6.4116, "step": 17461 }, { "epoch": 0.7362185635685224, "grad_norm": 2.175530433654785, "learning_rate": 4.19227646085137e-05, "loss": 6.6752, "step": 17462 }, { "epoch": 0.736260724750722, "grad_norm": 4.242486953735352, "learning_rate": 4.1921983822105624e-05, "loss": 7.5234, "step": 17463 }, { "epoch": 0.7363028859329216, "grad_norm": 2.9867196083068848, "learning_rate": 4.192120303569756e-05, "loss": 7.4552, "step": 17464 }, { "epoch": 0.736345047115121, "grad_norm": 1.4256736040115356, "learning_rate": 4.192042224928949e-05, "loss": 6.5359, "step": 17465 }, { "epoch": 0.7363872082973206, "grad_norm": 1.2603431940078735, "learning_rate": 4.191964146288142e-05, "loss": 6.567, "step": 17466 }, { "epoch": 0.7364293694795202, "grad_norm": 2.4650824069976807, "learning_rate": 4.191886067647335e-05, "loss": 7.2732, "step": 17467 }, { "epoch": 0.7364715306617198, "grad_norm": 1.7691296339035034, "learning_rate": 4.1918079890065274e-05, "loss": 7.0533, "step": 17468 }, { "epoch": 0.7365136918439193, "grad_norm": 1.293991208076477, "learning_rate": 4.1917299103657205e-05, "loss": 6.6904, "step": 17469 }, { "epoch": 0.7365558530261188, "grad_norm": 1.6028246879577637, "learning_rate": 4.1916518317249136e-05, "loss": 6.8225, "step": 17470 }, { "epoch": 0.7365980142083184, "grad_norm": 1.6319838762283325, "learning_rate": 4.191573753084107e-05, "loss": 6.3696, "step": 17471 }, { "epoch": 0.736640175390518, "grad_norm": 2.968709707260132, "learning_rate": 4.191495674443299e-05, "loss": 7.6445, "step": 17472 }, { "epoch": 0.7366823365727175, "grad_norm": 1.101697564125061, "learning_rate": 4.1914175958024923e-05, "loss": 6.7005, "step": 17473 }, { "epoch": 0.736724497754917, "grad_norm": 1.5437055826187134, "learning_rate": 4.1913395171616855e-05, "loss": 6.754, "step": 17474 }, { "epoch": 0.7367666589371166, "grad_norm": 1.091605305671692, "learning_rate": 4.191261438520878e-05, "loss": 6.6587, "step": 17475 }, { "epoch": 0.7368088201193161, "grad_norm": 1.6535097360610962, "learning_rate": 4.191183359880072e-05, "loss": 6.5412, "step": 17476 }, { "epoch": 0.7368509813015157, "grad_norm": 3.8198118209838867, "learning_rate": 4.191105281239264e-05, "loss": 7.4754, "step": 17477 }, { "epoch": 0.7368931424837153, "grad_norm": 3.753643751144409, "learning_rate": 4.191027202598457e-05, "loss": 7.9917, "step": 17478 }, { "epoch": 0.7369353036659148, "grad_norm": 2.032216787338257, "learning_rate": 4.1909491239576504e-05, "loss": 6.5305, "step": 17479 }, { "epoch": 0.7369774648481143, "grad_norm": 2.733511209487915, "learning_rate": 4.190871045316843e-05, "loss": 6.3571, "step": 17480 }, { "epoch": 0.7370196260303139, "grad_norm": 2.343329429626465, "learning_rate": 4.190792966676037e-05, "loss": 6.3317, "step": 17481 }, { "epoch": 0.7370617872125135, "grad_norm": 1.356054425239563, "learning_rate": 4.190714888035229e-05, "loss": 6.6282, "step": 17482 }, { "epoch": 0.7371039483947129, "grad_norm": 3.5746989250183105, "learning_rate": 4.190636809394422e-05, "loss": 7.3857, "step": 17483 }, { "epoch": 0.7371461095769125, "grad_norm": 3.820002794265747, "learning_rate": 4.1905587307536154e-05, "loss": 6.9033, "step": 17484 }, { "epoch": 0.7371882707591121, "grad_norm": 4.581761360168457, "learning_rate": 4.190480652112808e-05, "loss": 6.6631, "step": 17485 }, { "epoch": 0.7372304319413117, "grad_norm": 3.3267340660095215, "learning_rate": 4.190402573472001e-05, "loss": 7.2393, "step": 17486 }, { "epoch": 0.7372725931235112, "grad_norm": 5.815454483032227, "learning_rate": 4.190324494831194e-05, "loss": 8.3081, "step": 17487 }, { "epoch": 0.7373147543057107, "grad_norm": 1.3243038654327393, "learning_rate": 4.190246416190387e-05, "loss": 6.8902, "step": 17488 }, { "epoch": 0.7373569154879103, "grad_norm": 2.3667657375335693, "learning_rate": 4.19016833754958e-05, "loss": 6.9113, "step": 17489 }, { "epoch": 0.7373990766701098, "grad_norm": 2.8021905422210693, "learning_rate": 4.1900902589087735e-05, "loss": 6.9598, "step": 17490 }, { "epoch": 0.7374412378523094, "grad_norm": 2.7125039100646973, "learning_rate": 4.190012180267966e-05, "loss": 6.5368, "step": 17491 }, { "epoch": 0.7374833990345089, "grad_norm": 3.6200764179229736, "learning_rate": 4.189934101627159e-05, "loss": 6.7045, "step": 17492 }, { "epoch": 0.7375255602167085, "grad_norm": 1.5179107189178467, "learning_rate": 4.189856022986352e-05, "loss": 6.9017, "step": 17493 }, { "epoch": 0.737567721398908, "grad_norm": 4.1235432624816895, "learning_rate": 4.1897779443455446e-05, "loss": 7.9711, "step": 17494 }, { "epoch": 0.7376098825811076, "grad_norm": 3.572152853012085, "learning_rate": 4.1896998657047384e-05, "loss": 7.9964, "step": 17495 }, { "epoch": 0.7376520437633072, "grad_norm": 1.4822125434875488, "learning_rate": 4.189621787063931e-05, "loss": 6.489, "step": 17496 }, { "epoch": 0.7376942049455066, "grad_norm": 1.3136253356933594, "learning_rate": 4.189543708423124e-05, "loss": 6.4193, "step": 17497 }, { "epoch": 0.7377363661277062, "grad_norm": 1.3549418449401855, "learning_rate": 4.189465629782317e-05, "loss": 6.9815, "step": 17498 }, { "epoch": 0.7377785273099058, "grad_norm": 2.4773316383361816, "learning_rate": 4.1893875511415096e-05, "loss": 7.1012, "step": 17499 }, { "epoch": 0.7378206884921054, "grad_norm": 2.4192440509796143, "learning_rate": 4.1893094725007034e-05, "loss": 7.242, "step": 17500 }, { "epoch": 0.7378628496743048, "grad_norm": 2.4682788848876953, "learning_rate": 4.189231393859896e-05, "loss": 6.4808, "step": 17501 }, { "epoch": 0.7379050108565044, "grad_norm": 2.317174196243286, "learning_rate": 4.189153315219089e-05, "loss": 6.9144, "step": 17502 }, { "epoch": 0.737947172038704, "grad_norm": 1.4723944664001465, "learning_rate": 4.189075236578282e-05, "loss": 6.2494, "step": 17503 }, { "epoch": 0.7379893332209035, "grad_norm": 1.757182002067566, "learning_rate": 4.1889971579374745e-05, "loss": 6.4736, "step": 17504 }, { "epoch": 0.7380314944031031, "grad_norm": 3.3562936782836914, "learning_rate": 4.1889190792966676e-05, "loss": 6.9943, "step": 17505 }, { "epoch": 0.7380736555853026, "grad_norm": 1.4049103260040283, "learning_rate": 4.188841000655861e-05, "loss": 6.5251, "step": 17506 }, { "epoch": 0.7381158167675022, "grad_norm": 1.5402498245239258, "learning_rate": 4.188762922015054e-05, "loss": 6.9861, "step": 17507 }, { "epoch": 0.7381579779497017, "grad_norm": 1.436407446861267, "learning_rate": 4.188684843374246e-05, "loss": 7.5439, "step": 17508 }, { "epoch": 0.7382001391319013, "grad_norm": 2.602846145629883, "learning_rate": 4.18860676473344e-05, "loss": 6.4123, "step": 17509 }, { "epoch": 0.7382423003141008, "grad_norm": 3.345529556274414, "learning_rate": 4.1885286860926326e-05, "loss": 7.2913, "step": 17510 }, { "epoch": 0.7382844614963003, "grad_norm": 3.184399127960205, "learning_rate": 4.188450607451826e-05, "loss": 7.8107, "step": 17511 }, { "epoch": 0.7383266226784999, "grad_norm": 3.2409679889678955, "learning_rate": 4.188372528811019e-05, "loss": 6.6124, "step": 17512 }, { "epoch": 0.7383687838606995, "grad_norm": 1.5922073125839233, "learning_rate": 4.188294450170211e-05, "loss": 6.2494, "step": 17513 }, { "epoch": 0.738410945042899, "grad_norm": 1.1496456861495972, "learning_rate": 4.188216371529405e-05, "loss": 6.8894, "step": 17514 }, { "epoch": 0.7384531062250985, "grad_norm": 3.1145617961883545, "learning_rate": 4.1881382928885975e-05, "loss": 6.3819, "step": 17515 }, { "epoch": 0.7384952674072981, "grad_norm": 3.085052967071533, "learning_rate": 4.188060214247791e-05, "loss": 7.0654, "step": 17516 }, { "epoch": 0.7385374285894977, "grad_norm": 3.7509546279907227, "learning_rate": 4.187982135606984e-05, "loss": 7.1559, "step": 17517 }, { "epoch": 0.7385795897716972, "grad_norm": 3.486531972885132, "learning_rate": 4.187904056966176e-05, "loss": 7.789, "step": 17518 }, { "epoch": 0.7386217509538967, "grad_norm": 1.657425045967102, "learning_rate": 4.1878259783253694e-05, "loss": 6.9557, "step": 17519 }, { "epoch": 0.7386639121360963, "grad_norm": 2.0252652168273926, "learning_rate": 4.1877478996845625e-05, "loss": 6.9598, "step": 17520 }, { "epoch": 0.7387060733182959, "grad_norm": 2.2073416709899902, "learning_rate": 4.1876698210437556e-05, "loss": 6.6073, "step": 17521 }, { "epoch": 0.7387482345004954, "grad_norm": 1.491455078125, "learning_rate": 4.187591742402949e-05, "loss": 6.8887, "step": 17522 }, { "epoch": 0.7387903956826949, "grad_norm": 3.4127533435821533, "learning_rate": 4.187513663762141e-05, "loss": 7.729, "step": 17523 }, { "epoch": 0.7388325568648945, "grad_norm": 1.3255329132080078, "learning_rate": 4.187435585121334e-05, "loss": 6.4577, "step": 17524 }, { "epoch": 0.738874718047094, "grad_norm": 2.9092609882354736, "learning_rate": 4.1873575064805274e-05, "loss": 6.7347, "step": 17525 }, { "epoch": 0.7389168792292936, "grad_norm": 3.272271156311035, "learning_rate": 4.1872794278397206e-05, "loss": 7.1847, "step": 17526 }, { "epoch": 0.7389590404114932, "grad_norm": 2.157590627670288, "learning_rate": 4.187201349198913e-05, "loss": 6.6843, "step": 17527 }, { "epoch": 0.7390012015936926, "grad_norm": 2.504953384399414, "learning_rate": 4.187123270558107e-05, "loss": 6.5155, "step": 17528 }, { "epoch": 0.7390433627758922, "grad_norm": 2.1824495792388916, "learning_rate": 4.187045191917299e-05, "loss": 6.4984, "step": 17529 }, { "epoch": 0.7390855239580918, "grad_norm": 1.9729825258255005, "learning_rate": 4.186967113276492e-05, "loss": 6.3948, "step": 17530 }, { "epoch": 0.7391276851402914, "grad_norm": 2.4634110927581787, "learning_rate": 4.1868890346356855e-05, "loss": 6.985, "step": 17531 }, { "epoch": 0.7391698463224908, "grad_norm": 1.4504348039627075, "learning_rate": 4.186810955994878e-05, "loss": 6.3981, "step": 17532 }, { "epoch": 0.7392120075046904, "grad_norm": 1.5800501108169556, "learning_rate": 4.186732877354072e-05, "loss": 6.525, "step": 17533 }, { "epoch": 0.73925416868689, "grad_norm": 2.5743632316589355, "learning_rate": 4.186654798713264e-05, "loss": 6.6345, "step": 17534 }, { "epoch": 0.7392963298690896, "grad_norm": 2.6476473808288574, "learning_rate": 4.1865767200724573e-05, "loss": 6.8982, "step": 17535 }, { "epoch": 0.7393384910512891, "grad_norm": 0.8155523538589478, "learning_rate": 4.1864986414316505e-05, "loss": 6.6406, "step": 17536 }, { "epoch": 0.7393806522334886, "grad_norm": 2.804746150970459, "learning_rate": 4.186420562790843e-05, "loss": 6.7044, "step": 17537 }, { "epoch": 0.7394228134156882, "grad_norm": 2.4249489307403564, "learning_rate": 4.186342484150036e-05, "loss": 6.9387, "step": 17538 }, { "epoch": 0.7394649745978877, "grad_norm": 3.337648868560791, "learning_rate": 4.186264405509229e-05, "loss": 6.9819, "step": 17539 }, { "epoch": 0.7395071357800873, "grad_norm": 1.7265578508377075, "learning_rate": 4.186186326868422e-05, "loss": 6.7671, "step": 17540 }, { "epoch": 0.7395492969622868, "grad_norm": 1.6352217197418213, "learning_rate": 4.186108248227615e-05, "loss": 6.4542, "step": 17541 }, { "epoch": 0.7395914581444863, "grad_norm": 2.0501718521118164, "learning_rate": 4.186030169586808e-05, "loss": 6.5236, "step": 17542 }, { "epoch": 0.7396336193266859, "grad_norm": 4.015844345092773, "learning_rate": 4.185952090946001e-05, "loss": 7.7139, "step": 17543 }, { "epoch": 0.7396757805088855, "grad_norm": 2.7400999069213867, "learning_rate": 4.1858740123051934e-05, "loss": 7.4317, "step": 17544 }, { "epoch": 0.7397179416910851, "grad_norm": 1.5570775270462036, "learning_rate": 4.185795933664387e-05, "loss": 6.2373, "step": 17545 }, { "epoch": 0.7397601028732845, "grad_norm": 2.0909523963928223, "learning_rate": 4.18571785502358e-05, "loss": 7.0224, "step": 17546 }, { "epoch": 0.7398022640554841, "grad_norm": 1.7336788177490234, "learning_rate": 4.185639776382773e-05, "loss": 7.4981, "step": 17547 }, { "epoch": 0.7398444252376837, "grad_norm": 2.1992194652557373, "learning_rate": 4.185561697741966e-05, "loss": 7.3588, "step": 17548 }, { "epoch": 0.7398865864198833, "grad_norm": 3.107806921005249, "learning_rate": 4.1854836191011584e-05, "loss": 7.3655, "step": 17549 }, { "epoch": 0.7399287476020827, "grad_norm": 4.033840179443359, "learning_rate": 4.185405540460352e-05, "loss": 7.5639, "step": 17550 }, { "epoch": 0.7399709087842823, "grad_norm": 2.3858108520507812, "learning_rate": 4.1853274618195447e-05, "loss": 7.1282, "step": 17551 }, { "epoch": 0.7400130699664819, "grad_norm": 2.3947596549987793, "learning_rate": 4.185249383178738e-05, "loss": 6.4858, "step": 17552 }, { "epoch": 0.7400552311486814, "grad_norm": 1.7040636539459229, "learning_rate": 4.185171304537931e-05, "loss": 6.4182, "step": 17553 }, { "epoch": 0.740097392330881, "grad_norm": 1.6560227870941162, "learning_rate": 4.185093225897124e-05, "loss": 7.3816, "step": 17554 }, { "epoch": 0.7401395535130805, "grad_norm": 1.3430752754211426, "learning_rate": 4.1850151472563165e-05, "loss": 6.4969, "step": 17555 }, { "epoch": 0.74018171469528, "grad_norm": 1.3661742210388184, "learning_rate": 4.1849370686155096e-05, "loss": 6.4576, "step": 17556 }, { "epoch": 0.7402238758774796, "grad_norm": 1.6217122077941895, "learning_rate": 4.184858989974703e-05, "loss": 6.707, "step": 17557 }, { "epoch": 0.7402660370596792, "grad_norm": 1.8657188415527344, "learning_rate": 4.184780911333896e-05, "loss": 6.3562, "step": 17558 }, { "epoch": 0.7403081982418787, "grad_norm": 1.3729407787322998, "learning_rate": 4.184702832693089e-05, "loss": 6.5081, "step": 17559 }, { "epoch": 0.7403503594240782, "grad_norm": 2.262564182281494, "learning_rate": 4.1846247540522814e-05, "loss": 8.2574, "step": 17560 }, { "epoch": 0.7403925206062778, "grad_norm": 3.5572140216827393, "learning_rate": 4.1845466754114746e-05, "loss": 7.692, "step": 17561 }, { "epoch": 0.7404346817884774, "grad_norm": 2.1794166564941406, "learning_rate": 4.184468596770668e-05, "loss": 7.0762, "step": 17562 }, { "epoch": 0.740476842970677, "grad_norm": 2.4616477489471436, "learning_rate": 4.18439051812986e-05, "loss": 7.4789, "step": 17563 }, { "epoch": 0.7405190041528764, "grad_norm": 2.4664084911346436, "learning_rate": 4.184312439489054e-05, "loss": 6.5197, "step": 17564 }, { "epoch": 0.740561165335076, "grad_norm": 1.6167031526565552, "learning_rate": 4.1842343608482464e-05, "loss": 6.3848, "step": 17565 }, { "epoch": 0.7406033265172756, "grad_norm": 3.3323071002960205, "learning_rate": 4.1841562822074395e-05, "loss": 6.5864, "step": 17566 }, { "epoch": 0.7406454876994751, "grad_norm": 2.2976200580596924, "learning_rate": 4.1840782035666326e-05, "loss": 7.3615, "step": 17567 }, { "epoch": 0.7406876488816746, "grad_norm": 1.9197382926940918, "learning_rate": 4.184000124925825e-05, "loss": 6.5486, "step": 17568 }, { "epoch": 0.7407298100638742, "grad_norm": 2.851445198059082, "learning_rate": 4.183922046285019e-05, "loss": 7.6218, "step": 17569 }, { "epoch": 0.7407719712460737, "grad_norm": 1.241333246231079, "learning_rate": 4.183843967644211e-05, "loss": 6.7279, "step": 17570 }, { "epoch": 0.7408141324282733, "grad_norm": 1.3636789321899414, "learning_rate": 4.1837658890034045e-05, "loss": 6.7674, "step": 17571 }, { "epoch": 0.7408562936104728, "grad_norm": 2.6390013694763184, "learning_rate": 4.1836878103625976e-05, "loss": 7.8959, "step": 17572 }, { "epoch": 0.7408984547926724, "grad_norm": 2.31292986869812, "learning_rate": 4.183609731721791e-05, "loss": 6.7263, "step": 17573 }, { "epoch": 0.7409406159748719, "grad_norm": 1.6856787204742432, "learning_rate": 4.183531653080983e-05, "loss": 7.0992, "step": 17574 }, { "epoch": 0.7409827771570715, "grad_norm": 3.3327796459198, "learning_rate": 4.183453574440176e-05, "loss": 7.6992, "step": 17575 }, { "epoch": 0.7410249383392711, "grad_norm": 1.518061637878418, "learning_rate": 4.1833754957993694e-05, "loss": 7.1093, "step": 17576 }, { "epoch": 0.7410670995214705, "grad_norm": 1.191811203956604, "learning_rate": 4.183297417158562e-05, "loss": 6.667, "step": 17577 }, { "epoch": 0.7411092607036701, "grad_norm": 3.073939323425293, "learning_rate": 4.183219338517756e-05, "loss": 7.8346, "step": 17578 }, { "epoch": 0.7411514218858697, "grad_norm": 1.6747851371765137, "learning_rate": 4.183141259876948e-05, "loss": 6.7277, "step": 17579 }, { "epoch": 0.7411935830680693, "grad_norm": 2.8656370639801025, "learning_rate": 4.183063181236141e-05, "loss": 7.1418, "step": 17580 }, { "epoch": 0.7412357442502687, "grad_norm": 1.3030632734298706, "learning_rate": 4.1829851025953344e-05, "loss": 6.8154, "step": 17581 }, { "epoch": 0.7412779054324683, "grad_norm": 1.360188364982605, "learning_rate": 4.182907023954527e-05, "loss": 6.9005, "step": 17582 }, { "epoch": 0.7413200666146679, "grad_norm": 2.8979783058166504, "learning_rate": 4.1828289453137206e-05, "loss": 7.857, "step": 17583 }, { "epoch": 0.7413622277968674, "grad_norm": 1.4418338537216187, "learning_rate": 4.182750866672913e-05, "loss": 6.9466, "step": 17584 }, { "epoch": 0.741404388979067, "grad_norm": 1.3613797426223755, "learning_rate": 4.182672788032106e-05, "loss": 6.4193, "step": 17585 }, { "epoch": 0.7414465501612665, "grad_norm": 3.372814893722534, "learning_rate": 4.182594709391299e-05, "loss": 7.6589, "step": 17586 }, { "epoch": 0.7414887113434661, "grad_norm": 2.676386833190918, "learning_rate": 4.182516630750492e-05, "loss": 7.2125, "step": 17587 }, { "epoch": 0.7415308725256656, "grad_norm": 1.6857097148895264, "learning_rate": 4.182438552109685e-05, "loss": 7.3586, "step": 17588 }, { "epoch": 0.7415730337078652, "grad_norm": 2.4576544761657715, "learning_rate": 4.182360473468878e-05, "loss": 6.913, "step": 17589 }, { "epoch": 0.7416151948900647, "grad_norm": 1.9357125759124756, "learning_rate": 4.182282394828071e-05, "loss": 6.794, "step": 17590 }, { "epoch": 0.7416573560722642, "grad_norm": 1.9961094856262207, "learning_rate": 4.182204316187264e-05, "loss": 6.6222, "step": 17591 }, { "epoch": 0.7416995172544638, "grad_norm": 1.4807754755020142, "learning_rate": 4.182126237546457e-05, "loss": 6.561, "step": 17592 }, { "epoch": 0.7417416784366634, "grad_norm": 1.5246397256851196, "learning_rate": 4.18204815890565e-05, "loss": 6.6019, "step": 17593 }, { "epoch": 0.741783839618863, "grad_norm": 1.6433154344558716, "learning_rate": 4.181970080264843e-05, "loss": 6.5455, "step": 17594 }, { "epoch": 0.7418260008010624, "grad_norm": 1.3308610916137695, "learning_rate": 4.181892001624036e-05, "loss": 6.7728, "step": 17595 }, { "epoch": 0.741868161983262, "grad_norm": 1.2289708852767944, "learning_rate": 4.1818139229832285e-05, "loss": 7.1862, "step": 17596 }, { "epoch": 0.7419103231654616, "grad_norm": 2.0243563652038574, "learning_rate": 4.1817358443424224e-05, "loss": 6.7525, "step": 17597 }, { "epoch": 0.7419524843476611, "grad_norm": 3.731818199157715, "learning_rate": 4.181657765701615e-05, "loss": 7.8468, "step": 17598 }, { "epoch": 0.7419946455298606, "grad_norm": 2.747281789779663, "learning_rate": 4.181579687060808e-05, "loss": 6.5151, "step": 17599 }, { "epoch": 0.7420368067120602, "grad_norm": 1.8984917402267456, "learning_rate": 4.181501608420001e-05, "loss": 6.7604, "step": 17600 }, { "epoch": 0.7420789678942598, "grad_norm": 1.7280912399291992, "learning_rate": 4.1814235297791935e-05, "loss": 6.3793, "step": 17601 }, { "epoch": 0.7421211290764593, "grad_norm": 1.0519858598709106, "learning_rate": 4.181345451138387e-05, "loss": 6.5003, "step": 17602 }, { "epoch": 0.7421632902586589, "grad_norm": 1.9004936218261719, "learning_rate": 4.18126737249758e-05, "loss": 6.6054, "step": 17603 }, { "epoch": 0.7422054514408584, "grad_norm": 1.393117904663086, "learning_rate": 4.181189293856773e-05, "loss": 6.5233, "step": 17604 }, { "epoch": 0.7422476126230579, "grad_norm": 3.845381736755371, "learning_rate": 4.181111215215966e-05, "loss": 7.4152, "step": 17605 }, { "epoch": 0.7422897738052575, "grad_norm": 1.4225859642028809, "learning_rate": 4.1810331365751585e-05, "loss": 6.9074, "step": 17606 }, { "epoch": 0.7423319349874571, "grad_norm": 2.0457091331481934, "learning_rate": 4.1809550579343516e-05, "loss": 6.3904, "step": 17607 }, { "epoch": 0.7423740961696565, "grad_norm": 1.9938321113586426, "learning_rate": 4.180876979293545e-05, "loss": 6.4771, "step": 17608 }, { "epoch": 0.7424162573518561, "grad_norm": 1.3892027139663696, "learning_rate": 4.180798900652738e-05, "loss": 6.4193, "step": 17609 }, { "epoch": 0.7424584185340557, "grad_norm": 2.322742223739624, "learning_rate": 4.18072082201193e-05, "loss": 7.076, "step": 17610 }, { "epoch": 0.7425005797162553, "grad_norm": 2.6063754558563232, "learning_rate": 4.1806427433711234e-05, "loss": 6.9005, "step": 17611 }, { "epoch": 0.7425427408984548, "grad_norm": 1.8296663761138916, "learning_rate": 4.1805646647303165e-05, "loss": 6.7079, "step": 17612 }, { "epoch": 0.7425849020806543, "grad_norm": 1.161749243736267, "learning_rate": 4.1804865860895097e-05, "loss": 6.5516, "step": 17613 }, { "epoch": 0.7426270632628539, "grad_norm": 1.2858775854110718, "learning_rate": 4.180408507448703e-05, "loss": 6.4416, "step": 17614 }, { "epoch": 0.7426692244450535, "grad_norm": 1.4953380823135376, "learning_rate": 4.180330428807895e-05, "loss": 6.3903, "step": 17615 }, { "epoch": 0.742711385627253, "grad_norm": 1.3178954124450684, "learning_rate": 4.180252350167089e-05, "loss": 6.8252, "step": 17616 }, { "epoch": 0.7427535468094525, "grad_norm": 2.017002582550049, "learning_rate": 4.1801742715262815e-05, "loss": 6.9402, "step": 17617 }, { "epoch": 0.7427957079916521, "grad_norm": 1.3309216499328613, "learning_rate": 4.1800961928854746e-05, "loss": 6.5941, "step": 17618 }, { "epoch": 0.7428378691738516, "grad_norm": 1.9619745016098022, "learning_rate": 4.180018114244668e-05, "loss": 7.0322, "step": 17619 }, { "epoch": 0.7428800303560512, "grad_norm": 1.6047308444976807, "learning_rate": 4.17994003560386e-05, "loss": 6.8272, "step": 17620 }, { "epoch": 0.7429221915382508, "grad_norm": 4.316074848175049, "learning_rate": 4.179861956963053e-05, "loss": 7.4957, "step": 17621 }, { "epoch": 0.7429643527204502, "grad_norm": 2.9033403396606445, "learning_rate": 4.1797838783222464e-05, "loss": 7.1021, "step": 17622 }, { "epoch": 0.7430065139026498, "grad_norm": 2.669726848602295, "learning_rate": 4.1797057996814396e-05, "loss": 7.1945, "step": 17623 }, { "epoch": 0.7430486750848494, "grad_norm": 2.817643880844116, "learning_rate": 4.179627721040632e-05, "loss": 6.6241, "step": 17624 }, { "epoch": 0.743090836267049, "grad_norm": 2.541435480117798, "learning_rate": 4.179549642399825e-05, "loss": 6.506, "step": 17625 }, { "epoch": 0.7431329974492484, "grad_norm": 2.082087993621826, "learning_rate": 4.179471563759018e-05, "loss": 6.4276, "step": 17626 }, { "epoch": 0.743175158631448, "grad_norm": 1.8200525045394897, "learning_rate": 4.1793934851182114e-05, "loss": 7.1369, "step": 17627 }, { "epoch": 0.7432173198136476, "grad_norm": 3.0280909538269043, "learning_rate": 4.1793154064774045e-05, "loss": 7.0508, "step": 17628 }, { "epoch": 0.7432594809958472, "grad_norm": 3.1783242225646973, "learning_rate": 4.179237327836597e-05, "loss": 6.36, "step": 17629 }, { "epoch": 0.7433016421780466, "grad_norm": 2.2858619689941406, "learning_rate": 4.17915924919579e-05, "loss": 6.3921, "step": 17630 }, { "epoch": 0.7433438033602462, "grad_norm": 1.289075255393982, "learning_rate": 4.179081170554983e-05, "loss": 6.4548, "step": 17631 }, { "epoch": 0.7433859645424458, "grad_norm": 2.469686985015869, "learning_rate": 4.179003091914176e-05, "loss": 6.4381, "step": 17632 }, { "epoch": 0.7434281257246453, "grad_norm": 1.8405483961105347, "learning_rate": 4.1789250132733695e-05, "loss": 6.4559, "step": 17633 }, { "epoch": 0.7434702869068449, "grad_norm": 1.6978645324707031, "learning_rate": 4.178846934632562e-05, "loss": 7.358, "step": 17634 }, { "epoch": 0.7435124480890444, "grad_norm": 2.937851905822754, "learning_rate": 4.178768855991755e-05, "loss": 7.2909, "step": 17635 }, { "epoch": 0.743554609271244, "grad_norm": 1.9618947505950928, "learning_rate": 4.178690777350948e-05, "loss": 6.5723, "step": 17636 }, { "epoch": 0.7435967704534435, "grad_norm": 0.9754170775413513, "learning_rate": 4.1786126987101406e-05, "loss": 6.8397, "step": 17637 }, { "epoch": 0.7436389316356431, "grad_norm": 1.1531602144241333, "learning_rate": 4.1785346200693344e-05, "loss": 6.3381, "step": 17638 }, { "epoch": 0.7436810928178426, "grad_norm": 2.824758529663086, "learning_rate": 4.178456541428527e-05, "loss": 6.9134, "step": 17639 }, { "epoch": 0.7437232540000421, "grad_norm": 2.911043167114258, "learning_rate": 4.17837846278772e-05, "loss": 7.1111, "step": 17640 }, { "epoch": 0.7437654151822417, "grad_norm": 3.3916115760803223, "learning_rate": 4.178300384146913e-05, "loss": 7.3387, "step": 17641 }, { "epoch": 0.7438075763644413, "grad_norm": 1.7264536619186401, "learning_rate": 4.178222305506106e-05, "loss": 6.4943, "step": 17642 }, { "epoch": 0.7438497375466409, "grad_norm": 2.2587523460388184, "learning_rate": 4.178144226865299e-05, "loss": 6.6291, "step": 17643 }, { "epoch": 0.7438918987288403, "grad_norm": 2.8351657390594482, "learning_rate": 4.178066148224492e-05, "loss": 8.0009, "step": 17644 }, { "epoch": 0.7439340599110399, "grad_norm": 1.846794843673706, "learning_rate": 4.177988069583685e-05, "loss": 6.9796, "step": 17645 }, { "epoch": 0.7439762210932395, "grad_norm": 1.7121458053588867, "learning_rate": 4.1779099909428774e-05, "loss": 6.3661, "step": 17646 }, { "epoch": 0.744018382275439, "grad_norm": 1.2997163534164429, "learning_rate": 4.177831912302071e-05, "loss": 6.5238, "step": 17647 }, { "epoch": 0.7440605434576385, "grad_norm": 4.25673770904541, "learning_rate": 4.1777538336612636e-05, "loss": 7.0857, "step": 17648 }, { "epoch": 0.7441027046398381, "grad_norm": 3.7496564388275146, "learning_rate": 4.177675755020457e-05, "loss": 7.8393, "step": 17649 }, { "epoch": 0.7441448658220376, "grad_norm": 4.521410942077637, "learning_rate": 4.17759767637965e-05, "loss": 7.6127, "step": 17650 }, { "epoch": 0.7441870270042372, "grad_norm": 3.9952938556671143, "learning_rate": 4.1775195977388423e-05, "loss": 7.5162, "step": 17651 }, { "epoch": 0.7442291881864368, "grad_norm": 3.848778009414673, "learning_rate": 4.177441519098036e-05, "loss": 6.7854, "step": 17652 }, { "epoch": 0.7442713493686363, "grad_norm": 4.563476085662842, "learning_rate": 4.1773634404572286e-05, "loss": 7.0701, "step": 17653 }, { "epoch": 0.7443135105508358, "grad_norm": 1.4581713676452637, "learning_rate": 4.177285361816422e-05, "loss": 6.7356, "step": 17654 }, { "epoch": 0.7443556717330354, "grad_norm": 3.056447744369507, "learning_rate": 4.177207283175615e-05, "loss": 7.3402, "step": 17655 }, { "epoch": 0.744397832915235, "grad_norm": 2.4881751537323, "learning_rate": 4.177129204534807e-05, "loss": 6.9917, "step": 17656 }, { "epoch": 0.7444399940974344, "grad_norm": 3.1142804622650146, "learning_rate": 4.1770511258940004e-05, "loss": 7.8901, "step": 17657 }, { "epoch": 0.744482155279634, "grad_norm": 3.0465075969696045, "learning_rate": 4.1769730472531935e-05, "loss": 7.7189, "step": 17658 }, { "epoch": 0.7445243164618336, "grad_norm": 2.7386996746063232, "learning_rate": 4.176894968612387e-05, "loss": 7.6313, "step": 17659 }, { "epoch": 0.7445664776440332, "grad_norm": 2.0625991821289062, "learning_rate": 4.17681688997158e-05, "loss": 6.3982, "step": 17660 }, { "epoch": 0.7446086388262327, "grad_norm": 2.181342363357544, "learning_rate": 4.176738811330773e-05, "loss": 6.7266, "step": 17661 }, { "epoch": 0.7446508000084322, "grad_norm": 2.6892826557159424, "learning_rate": 4.1766607326899654e-05, "loss": 6.5586, "step": 17662 }, { "epoch": 0.7446929611906318, "grad_norm": 1.5902431011199951, "learning_rate": 4.1765826540491585e-05, "loss": 6.5883, "step": 17663 }, { "epoch": 0.7447351223728313, "grad_norm": 1.4508308172225952, "learning_rate": 4.1765045754083516e-05, "loss": 6.4446, "step": 17664 }, { "epoch": 0.7447772835550309, "grad_norm": 2.687927007675171, "learning_rate": 4.176426496767544e-05, "loss": 7.19, "step": 17665 }, { "epoch": 0.7448194447372304, "grad_norm": 3.861619234085083, "learning_rate": 4.176348418126738e-05, "loss": 7.5286, "step": 17666 }, { "epoch": 0.74486160591943, "grad_norm": 2.9268200397491455, "learning_rate": 4.17627033948593e-05, "loss": 7.3276, "step": 17667 }, { "epoch": 0.7449037671016295, "grad_norm": 3.583010673522949, "learning_rate": 4.1761922608451235e-05, "loss": 9.7736, "step": 17668 }, { "epoch": 0.7449459282838291, "grad_norm": 1.9791265726089478, "learning_rate": 4.1761141822043166e-05, "loss": 6.5952, "step": 17669 }, { "epoch": 0.7449880894660287, "grad_norm": 1.585341215133667, "learning_rate": 4.176036103563509e-05, "loss": 6.7531, "step": 17670 }, { "epoch": 0.7450302506482281, "grad_norm": 1.3270827531814575, "learning_rate": 4.175958024922703e-05, "loss": 6.8214, "step": 17671 }, { "epoch": 0.7450724118304277, "grad_norm": 1.6094402074813843, "learning_rate": 4.175879946281895e-05, "loss": 6.7145, "step": 17672 }, { "epoch": 0.7451145730126273, "grad_norm": 2.6861307621002197, "learning_rate": 4.1758018676410884e-05, "loss": 7.1947, "step": 17673 }, { "epoch": 0.7451567341948269, "grad_norm": 1.551544189453125, "learning_rate": 4.1757237890002815e-05, "loss": 6.8519, "step": 17674 }, { "epoch": 0.7451988953770263, "grad_norm": 1.2632970809936523, "learning_rate": 4.175645710359474e-05, "loss": 6.9705, "step": 17675 }, { "epoch": 0.7452410565592259, "grad_norm": 1.697112798690796, "learning_rate": 4.175567631718667e-05, "loss": 7.0905, "step": 17676 }, { "epoch": 0.7452832177414255, "grad_norm": 1.5528160333633423, "learning_rate": 4.17548955307786e-05, "loss": 7.0784, "step": 17677 }, { "epoch": 0.745325378923625, "grad_norm": 1.290324091911316, "learning_rate": 4.1754114744370534e-05, "loss": 6.6961, "step": 17678 }, { "epoch": 0.7453675401058246, "grad_norm": 2.231602907180786, "learning_rate": 4.175333395796246e-05, "loss": 6.9766, "step": 17679 }, { "epoch": 0.7454097012880241, "grad_norm": 1.2653254270553589, "learning_rate": 4.1752553171554396e-05, "loss": 6.8991, "step": 17680 }, { "epoch": 0.7454518624702237, "grad_norm": 1.7014517784118652, "learning_rate": 4.175177238514632e-05, "loss": 6.3793, "step": 17681 }, { "epoch": 0.7454940236524232, "grad_norm": 1.5038912296295166, "learning_rate": 4.175099159873825e-05, "loss": 6.9245, "step": 17682 }, { "epoch": 0.7455361848346228, "grad_norm": 3.09847354888916, "learning_rate": 4.175021081233018e-05, "loss": 7.6419, "step": 17683 }, { "epoch": 0.7455783460168223, "grad_norm": 2.869964599609375, "learning_rate": 4.174943002592211e-05, "loss": 8.2812, "step": 17684 }, { "epoch": 0.7456205071990218, "grad_norm": 1.6212847232818604, "learning_rate": 4.1748649239514046e-05, "loss": 6.5914, "step": 17685 }, { "epoch": 0.7456626683812214, "grad_norm": 2.135256290435791, "learning_rate": 4.174786845310597e-05, "loss": 6.6157, "step": 17686 }, { "epoch": 0.745704829563421, "grad_norm": 1.1752513647079468, "learning_rate": 4.17470876666979e-05, "loss": 6.6946, "step": 17687 }, { "epoch": 0.7457469907456205, "grad_norm": 4.317458152770996, "learning_rate": 4.174630688028983e-05, "loss": 9.6081, "step": 17688 }, { "epoch": 0.74578915192782, "grad_norm": 1.5950675010681152, "learning_rate": 4.174552609388176e-05, "loss": 6.9513, "step": 17689 }, { "epoch": 0.7458313131100196, "grad_norm": 2.333451986312866, "learning_rate": 4.174474530747369e-05, "loss": 6.4145, "step": 17690 }, { "epoch": 0.7458734742922192, "grad_norm": 3.5509653091430664, "learning_rate": 4.174396452106562e-05, "loss": 7.6193, "step": 17691 }, { "epoch": 0.7459156354744187, "grad_norm": 1.7828906774520874, "learning_rate": 4.174318373465755e-05, "loss": 6.572, "step": 17692 }, { "epoch": 0.7459577966566182, "grad_norm": 1.675929069519043, "learning_rate": 4.174240294824948e-05, "loss": 7.11, "step": 17693 }, { "epoch": 0.7459999578388178, "grad_norm": 2.7885825634002686, "learning_rate": 4.174162216184141e-05, "loss": 6.5384, "step": 17694 }, { "epoch": 0.7460421190210174, "grad_norm": 2.905078172683716, "learning_rate": 4.174084137543334e-05, "loss": 7.361, "step": 17695 }, { "epoch": 0.7460842802032169, "grad_norm": 2.389798164367676, "learning_rate": 4.174006058902527e-05, "loss": 7.069, "step": 17696 }, { "epoch": 0.7461264413854164, "grad_norm": 2.5012786388397217, "learning_rate": 4.17392798026172e-05, "loss": 7.5476, "step": 17697 }, { "epoch": 0.746168602567616, "grad_norm": 2.5083062648773193, "learning_rate": 4.1738499016209125e-05, "loss": 7.7662, "step": 17698 }, { "epoch": 0.7462107637498155, "grad_norm": 2.006136178970337, "learning_rate": 4.1737718229801056e-05, "loss": 6.6024, "step": 17699 }, { "epoch": 0.7462529249320151, "grad_norm": 1.6311898231506348, "learning_rate": 4.173693744339299e-05, "loss": 6.5297, "step": 17700 }, { "epoch": 0.7462950861142147, "grad_norm": 2.4277515411376953, "learning_rate": 4.173615665698491e-05, "loss": 6.8999, "step": 17701 }, { "epoch": 0.7463372472964142, "grad_norm": 1.8312331438064575, "learning_rate": 4.173537587057685e-05, "loss": 6.5586, "step": 17702 }, { "epoch": 0.7463794084786137, "grad_norm": 4.237653732299805, "learning_rate": 4.1734595084168774e-05, "loss": 8.4839, "step": 17703 }, { "epoch": 0.7464215696608133, "grad_norm": 1.2326247692108154, "learning_rate": 4.1733814297760706e-05, "loss": 6.9858, "step": 17704 }, { "epoch": 0.7464637308430129, "grad_norm": 1.6523401737213135, "learning_rate": 4.173303351135264e-05, "loss": 6.9414, "step": 17705 }, { "epoch": 0.7465058920252123, "grad_norm": 3.8495912551879883, "learning_rate": 4.173225272494457e-05, "loss": 7.2177, "step": 17706 }, { "epoch": 0.7465480532074119, "grad_norm": 2.87457275390625, "learning_rate": 4.17314719385365e-05, "loss": 7.7318, "step": 17707 }, { "epoch": 0.7465902143896115, "grad_norm": 5.236388683319092, "learning_rate": 4.1730691152128424e-05, "loss": 6.9183, "step": 17708 }, { "epoch": 0.7466323755718111, "grad_norm": 2.245422840118408, "learning_rate": 4.1729910365720355e-05, "loss": 7.7739, "step": 17709 }, { "epoch": 0.7466745367540106, "grad_norm": 1.403696894645691, "learning_rate": 4.1729129579312286e-05, "loss": 6.5185, "step": 17710 }, { "epoch": 0.7467166979362101, "grad_norm": 2.3565475940704346, "learning_rate": 4.172834879290422e-05, "loss": 6.7651, "step": 17711 }, { "epoch": 0.7467588591184097, "grad_norm": 1.5267494916915894, "learning_rate": 4.172756800649614e-05, "loss": 6.6912, "step": 17712 }, { "epoch": 0.7468010203006092, "grad_norm": 1.7518891096115112, "learning_rate": 4.1726787220088073e-05, "loss": 6.4577, "step": 17713 }, { "epoch": 0.7468431814828088, "grad_norm": 3.277216911315918, "learning_rate": 4.1726006433680005e-05, "loss": 7.7999, "step": 17714 }, { "epoch": 0.7468853426650083, "grad_norm": 5.038719177246094, "learning_rate": 4.172522564727193e-05, "loss": 8.5713, "step": 17715 }, { "epoch": 0.7469275038472079, "grad_norm": 2.5476505756378174, "learning_rate": 4.172444486086387e-05, "loss": 6.92, "step": 17716 }, { "epoch": 0.7469696650294074, "grad_norm": 1.9302068948745728, "learning_rate": 4.172366407445579e-05, "loss": 6.3708, "step": 17717 }, { "epoch": 0.747011826211607, "grad_norm": 1.7757352590560913, "learning_rate": 4.172288328804772e-05, "loss": 6.409, "step": 17718 }, { "epoch": 0.7470539873938066, "grad_norm": 1.4940580129623413, "learning_rate": 4.1722102501639654e-05, "loss": 7.1053, "step": 17719 }, { "epoch": 0.747096148576006, "grad_norm": 3.7274465560913086, "learning_rate": 4.172132171523158e-05, "loss": 9.6329, "step": 17720 }, { "epoch": 0.7471383097582056, "grad_norm": 1.3851464986801147, "learning_rate": 4.172054092882352e-05, "loss": 6.9459, "step": 17721 }, { "epoch": 0.7471804709404052, "grad_norm": 1.733206033706665, "learning_rate": 4.171976014241544e-05, "loss": 6.4993, "step": 17722 }, { "epoch": 0.7472226321226048, "grad_norm": 2.0553624629974365, "learning_rate": 4.171897935600737e-05, "loss": 6.9755, "step": 17723 }, { "epoch": 0.7472647933048042, "grad_norm": 2.4451217651367188, "learning_rate": 4.1718198569599304e-05, "loss": 6.7247, "step": 17724 }, { "epoch": 0.7473069544870038, "grad_norm": 2.0867321491241455, "learning_rate": 4.171741778319123e-05, "loss": 6.7581, "step": 17725 }, { "epoch": 0.7473491156692034, "grad_norm": 1.5596215724945068, "learning_rate": 4.171663699678316e-05, "loss": 6.8056, "step": 17726 }, { "epoch": 0.7473912768514029, "grad_norm": 1.1516832113265991, "learning_rate": 4.171585621037509e-05, "loss": 6.5562, "step": 17727 }, { "epoch": 0.7474334380336025, "grad_norm": 2.120664596557617, "learning_rate": 4.171507542396702e-05, "loss": 6.7167, "step": 17728 }, { "epoch": 0.747475599215802, "grad_norm": 1.4226230382919312, "learning_rate": 4.171429463755895e-05, "loss": 6.8399, "step": 17729 }, { "epoch": 0.7475177603980016, "grad_norm": 1.3626666069030762, "learning_rate": 4.1713513851150885e-05, "loss": 6.9369, "step": 17730 }, { "epoch": 0.7475599215802011, "grad_norm": 1.8205726146697998, "learning_rate": 4.171273306474281e-05, "loss": 6.841, "step": 17731 }, { "epoch": 0.7476020827624007, "grad_norm": 1.6834253072738647, "learning_rate": 4.171195227833474e-05, "loss": 6.5417, "step": 17732 }, { "epoch": 0.7476442439446002, "grad_norm": 2.3771018981933594, "learning_rate": 4.171117149192667e-05, "loss": 6.5046, "step": 17733 }, { "epoch": 0.7476864051267997, "grad_norm": 3.2598323822021484, "learning_rate": 4.1710390705518596e-05, "loss": 6.6048, "step": 17734 }, { "epoch": 0.7477285663089993, "grad_norm": 3.9830973148345947, "learning_rate": 4.1709609919110534e-05, "loss": 7.7214, "step": 17735 }, { "epoch": 0.7477707274911989, "grad_norm": 2.1143641471862793, "learning_rate": 4.170882913270246e-05, "loss": 6.8312, "step": 17736 }, { "epoch": 0.7478128886733985, "grad_norm": 1.3165323734283447, "learning_rate": 4.170804834629439e-05, "loss": 7.0732, "step": 17737 }, { "epoch": 0.7478550498555979, "grad_norm": 3.5339395999908447, "learning_rate": 4.170726755988632e-05, "loss": 6.9522, "step": 17738 }, { "epoch": 0.7478972110377975, "grad_norm": 1.3351134061813354, "learning_rate": 4.1706486773478246e-05, "loss": 6.7439, "step": 17739 }, { "epoch": 0.7479393722199971, "grad_norm": 2.9188709259033203, "learning_rate": 4.1705705987070184e-05, "loss": 8.115, "step": 17740 }, { "epoch": 0.7479815334021966, "grad_norm": 2.5748000144958496, "learning_rate": 4.170492520066211e-05, "loss": 7.4531, "step": 17741 }, { "epoch": 0.7480236945843961, "grad_norm": 2.961359977722168, "learning_rate": 4.170414441425404e-05, "loss": 7.1625, "step": 17742 }, { "epoch": 0.7480658557665957, "grad_norm": 1.8077677488327026, "learning_rate": 4.170336362784597e-05, "loss": 6.8532, "step": 17743 }, { "epoch": 0.7481080169487953, "grad_norm": 1.4703518152236938, "learning_rate": 4.1702582841437895e-05, "loss": 6.4813, "step": 17744 }, { "epoch": 0.7481501781309948, "grad_norm": 1.7946205139160156, "learning_rate": 4.1701802055029826e-05, "loss": 6.3595, "step": 17745 }, { "epoch": 0.7481923393131943, "grad_norm": 1.8343883752822876, "learning_rate": 4.170102126862176e-05, "loss": 6.7696, "step": 17746 }, { "epoch": 0.7482345004953939, "grad_norm": 0.9737592339515686, "learning_rate": 4.170024048221369e-05, "loss": 6.7192, "step": 17747 }, { "epoch": 0.7482766616775934, "grad_norm": 1.1697630882263184, "learning_rate": 4.169945969580561e-05, "loss": 6.7935, "step": 17748 }, { "epoch": 0.748318822859793, "grad_norm": 2.196650743484497, "learning_rate": 4.169867890939755e-05, "loss": 7.0721, "step": 17749 }, { "epoch": 0.7483609840419926, "grad_norm": 1.1426705121994019, "learning_rate": 4.1697898122989476e-05, "loss": 6.3706, "step": 17750 }, { "epoch": 0.748403145224192, "grad_norm": 1.6714842319488525, "learning_rate": 4.169711733658141e-05, "loss": 6.9418, "step": 17751 }, { "epoch": 0.7484453064063916, "grad_norm": 2.8336613178253174, "learning_rate": 4.169633655017334e-05, "loss": 7.676, "step": 17752 }, { "epoch": 0.7484874675885912, "grad_norm": 1.5621185302734375, "learning_rate": 4.169555576376526e-05, "loss": 6.5216, "step": 17753 }, { "epoch": 0.7485296287707908, "grad_norm": 1.086824655532837, "learning_rate": 4.16947749773572e-05, "loss": 6.4707, "step": 17754 }, { "epoch": 0.7485717899529902, "grad_norm": 3.595052480697632, "learning_rate": 4.1693994190949125e-05, "loss": 7.9088, "step": 17755 }, { "epoch": 0.7486139511351898, "grad_norm": 2.0606935024261475, "learning_rate": 4.169321340454106e-05, "loss": 6.9888, "step": 17756 }, { "epoch": 0.7486561123173894, "grad_norm": 1.9898409843444824, "learning_rate": 4.169243261813299e-05, "loss": 6.966, "step": 17757 }, { "epoch": 0.748698273499589, "grad_norm": 3.497246503829956, "learning_rate": 4.169165183172491e-05, "loss": 7.2108, "step": 17758 }, { "epoch": 0.7487404346817885, "grad_norm": 4.494208335876465, "learning_rate": 4.1690871045316844e-05, "loss": 7.51, "step": 17759 }, { "epoch": 0.748782595863988, "grad_norm": 2.2506422996520996, "learning_rate": 4.1690090258908775e-05, "loss": 6.4208, "step": 17760 }, { "epoch": 0.7488247570461876, "grad_norm": 1.6480162143707275, "learning_rate": 4.1689309472500706e-05, "loss": 6.5131, "step": 17761 }, { "epoch": 0.7488669182283871, "grad_norm": 1.1427855491638184, "learning_rate": 4.168852868609264e-05, "loss": 6.513, "step": 17762 }, { "epoch": 0.7489090794105867, "grad_norm": 1.4563913345336914, "learning_rate": 4.168774789968456e-05, "loss": 6.5459, "step": 17763 }, { "epoch": 0.7489512405927862, "grad_norm": 2.0184242725372314, "learning_rate": 4.168696711327649e-05, "loss": 7.0606, "step": 17764 }, { "epoch": 0.7489934017749857, "grad_norm": 1.7991623878479004, "learning_rate": 4.1686186326868424e-05, "loss": 6.5944, "step": 17765 }, { "epoch": 0.7490355629571853, "grad_norm": 1.4502588510513306, "learning_rate": 4.1685405540460356e-05, "loss": 6.9403, "step": 17766 }, { "epoch": 0.7490777241393849, "grad_norm": 3.7914083003997803, "learning_rate": 4.168462475405228e-05, "loss": 7.9373, "step": 17767 }, { "epoch": 0.7491198853215845, "grad_norm": 3.658060073852539, "learning_rate": 4.168384396764422e-05, "loss": 6.8072, "step": 17768 }, { "epoch": 0.7491620465037839, "grad_norm": 3.333937406539917, "learning_rate": 4.168306318123614e-05, "loss": 8.2374, "step": 17769 }, { "epoch": 0.7492042076859835, "grad_norm": 2.557610511779785, "learning_rate": 4.168228239482807e-05, "loss": 7.3272, "step": 17770 }, { "epoch": 0.7492463688681831, "grad_norm": 2.1847667694091797, "learning_rate": 4.1681501608420005e-05, "loss": 6.7872, "step": 17771 }, { "epoch": 0.7492885300503827, "grad_norm": 3.4996752738952637, "learning_rate": 4.168072082201193e-05, "loss": 7.8487, "step": 17772 }, { "epoch": 0.7493306912325821, "grad_norm": 2.5435516834259033, "learning_rate": 4.167994003560387e-05, "loss": 6.6249, "step": 17773 }, { "epoch": 0.7493728524147817, "grad_norm": 3.147279977798462, "learning_rate": 4.167915924919579e-05, "loss": 7.0879, "step": 17774 }, { "epoch": 0.7494150135969813, "grad_norm": 1.9723756313323975, "learning_rate": 4.1678378462787723e-05, "loss": 6.5849, "step": 17775 }, { "epoch": 0.7494571747791808, "grad_norm": 1.7170945405960083, "learning_rate": 4.1677597676379655e-05, "loss": 7.0859, "step": 17776 }, { "epoch": 0.7494993359613804, "grad_norm": 5.234770774841309, "learning_rate": 4.167681688997158e-05, "loss": 6.8967, "step": 17777 }, { "epoch": 0.7495414971435799, "grad_norm": 2.0227110385894775, "learning_rate": 4.167603610356351e-05, "loss": 8.2365, "step": 17778 }, { "epoch": 0.7495836583257794, "grad_norm": 3.188035249710083, "learning_rate": 4.167525531715544e-05, "loss": 7.4206, "step": 17779 }, { "epoch": 0.749625819507979, "grad_norm": 2.2027063369750977, "learning_rate": 4.167447453074737e-05, "loss": 6.444, "step": 17780 }, { "epoch": 0.7496679806901786, "grad_norm": 1.722116231918335, "learning_rate": 4.16736937443393e-05, "loss": 6.493, "step": 17781 }, { "epoch": 0.749710141872378, "grad_norm": 2.2443699836730957, "learning_rate": 4.167291295793123e-05, "loss": 6.8151, "step": 17782 }, { "epoch": 0.7497523030545776, "grad_norm": 1.844740629196167, "learning_rate": 4.167213217152316e-05, "loss": 6.6462, "step": 17783 }, { "epoch": 0.7497944642367772, "grad_norm": 1.5238624811172485, "learning_rate": 4.1671351385115084e-05, "loss": 6.6411, "step": 17784 }, { "epoch": 0.7498366254189768, "grad_norm": 1.3745864629745483, "learning_rate": 4.167057059870702e-05, "loss": 6.5006, "step": 17785 }, { "epoch": 0.7498787866011764, "grad_norm": 1.6618372201919556, "learning_rate": 4.166978981229895e-05, "loss": 6.3879, "step": 17786 }, { "epoch": 0.7499209477833758, "grad_norm": 1.24801504611969, "learning_rate": 4.166900902589088e-05, "loss": 6.6447, "step": 17787 }, { "epoch": 0.7499631089655754, "grad_norm": 1.2297348976135254, "learning_rate": 4.166822823948281e-05, "loss": 6.3839, "step": 17788 }, { "epoch": 0.750005270147775, "grad_norm": 2.2655487060546875, "learning_rate": 4.1667447453074734e-05, "loss": 7.0963, "step": 17789 }, { "epoch": 0.7500474313299745, "grad_norm": 1.8359384536743164, "learning_rate": 4.166666666666667e-05, "loss": 7.233, "step": 17790 }, { "epoch": 0.750089592512174, "grad_norm": 1.1635515689849854, "learning_rate": 4.1665885880258597e-05, "loss": 6.3867, "step": 17791 }, { "epoch": 0.7501317536943736, "grad_norm": 1.5997626781463623, "learning_rate": 4.166510509385053e-05, "loss": 6.8398, "step": 17792 }, { "epoch": 0.7501739148765731, "grad_norm": 0.9959156513214111, "learning_rate": 4.166432430744246e-05, "loss": 6.8025, "step": 17793 }, { "epoch": 0.7502160760587727, "grad_norm": 1.6194450855255127, "learning_rate": 4.166354352103439e-05, "loss": 6.7134, "step": 17794 }, { "epoch": 0.7502582372409723, "grad_norm": 1.0481358766555786, "learning_rate": 4.1662762734626315e-05, "loss": 6.4856, "step": 17795 }, { "epoch": 0.7503003984231718, "grad_norm": 3.314776659011841, "learning_rate": 4.1661981948218246e-05, "loss": 8.1422, "step": 17796 }, { "epoch": 0.7503425596053713, "grad_norm": 2.5319466590881348, "learning_rate": 4.166120116181018e-05, "loss": 7.6642, "step": 17797 }, { "epoch": 0.7503847207875709, "grad_norm": 3.1306633949279785, "learning_rate": 4.166042037540211e-05, "loss": 7.4506, "step": 17798 }, { "epoch": 0.7504268819697705, "grad_norm": 1.1719167232513428, "learning_rate": 4.165963958899404e-05, "loss": 6.4363, "step": 17799 }, { "epoch": 0.7504690431519699, "grad_norm": 0.8474745154380798, "learning_rate": 4.1658858802585964e-05, "loss": 6.5664, "step": 17800 }, { "epoch": 0.7505112043341695, "grad_norm": 1.7094279527664185, "learning_rate": 4.1658078016177896e-05, "loss": 6.8608, "step": 17801 }, { "epoch": 0.7505533655163691, "grad_norm": 0.8308771252632141, "learning_rate": 4.165729722976983e-05, "loss": 6.4836, "step": 17802 }, { "epoch": 0.7505955266985687, "grad_norm": 2.2855732440948486, "learning_rate": 4.165651644336175e-05, "loss": 6.8092, "step": 17803 }, { "epoch": 0.7506376878807681, "grad_norm": 1.2899922132492065, "learning_rate": 4.165573565695369e-05, "loss": 6.6526, "step": 17804 }, { "epoch": 0.7506798490629677, "grad_norm": 2.244094133377075, "learning_rate": 4.1654954870545614e-05, "loss": 6.6423, "step": 17805 }, { "epoch": 0.7507220102451673, "grad_norm": 1.293504238128662, "learning_rate": 4.1654174084137545e-05, "loss": 6.5568, "step": 17806 }, { "epoch": 0.7507641714273668, "grad_norm": 2.633765935897827, "learning_rate": 4.1653393297729476e-05, "loss": 7.1503, "step": 17807 }, { "epoch": 0.7508063326095664, "grad_norm": 3.162144184112549, "learning_rate": 4.16526125113214e-05, "loss": 7.1603, "step": 17808 }, { "epoch": 0.7508484937917659, "grad_norm": 2.71974778175354, "learning_rate": 4.165183172491334e-05, "loss": 7.0423, "step": 17809 }, { "epoch": 0.7508906549739655, "grad_norm": 1.880952000617981, "learning_rate": 4.165105093850526e-05, "loss": 7.1381, "step": 17810 }, { "epoch": 0.750932816156165, "grad_norm": 1.4688609838485718, "learning_rate": 4.1650270152097195e-05, "loss": 6.5501, "step": 17811 }, { "epoch": 0.7509749773383646, "grad_norm": 2.5759854316711426, "learning_rate": 4.1649489365689126e-05, "loss": 7.053, "step": 17812 }, { "epoch": 0.7510171385205641, "grad_norm": 3.774322509765625, "learning_rate": 4.164870857928106e-05, "loss": 7.6876, "step": 17813 }, { "epoch": 0.7510592997027636, "grad_norm": 2.332557201385498, "learning_rate": 4.164792779287298e-05, "loss": 6.6709, "step": 17814 }, { "epoch": 0.7511014608849632, "grad_norm": 4.4785943031311035, "learning_rate": 4.164714700646491e-05, "loss": 7.86, "step": 17815 }, { "epoch": 0.7511436220671628, "grad_norm": 1.3843084573745728, "learning_rate": 4.1646366220056844e-05, "loss": 6.8115, "step": 17816 }, { "epoch": 0.7511857832493624, "grad_norm": 2.0366439819335938, "learning_rate": 4.164558543364877e-05, "loss": 6.8166, "step": 17817 }, { "epoch": 0.7512279444315618, "grad_norm": 1.662521481513977, "learning_rate": 4.164480464724071e-05, "loss": 6.6028, "step": 17818 }, { "epoch": 0.7512701056137614, "grad_norm": 1.5393937826156616, "learning_rate": 4.164402386083263e-05, "loss": 6.6152, "step": 17819 }, { "epoch": 0.751312266795961, "grad_norm": 2.3238258361816406, "learning_rate": 4.164324307442456e-05, "loss": 7.0055, "step": 17820 }, { "epoch": 0.7513544279781605, "grad_norm": 1.9917032718658447, "learning_rate": 4.1642462288016494e-05, "loss": 7.2545, "step": 17821 }, { "epoch": 0.75139658916036, "grad_norm": 4.468733310699463, "learning_rate": 4.164168150160842e-05, "loss": 7.4628, "step": 17822 }, { "epoch": 0.7514387503425596, "grad_norm": 2.0673913955688477, "learning_rate": 4.1640900715200356e-05, "loss": 7.3241, "step": 17823 }, { "epoch": 0.7514809115247592, "grad_norm": 1.554716944694519, "learning_rate": 4.164011992879228e-05, "loss": 6.9749, "step": 17824 }, { "epoch": 0.7515230727069587, "grad_norm": 3.3685033321380615, "learning_rate": 4.163933914238421e-05, "loss": 7.2255, "step": 17825 }, { "epoch": 0.7515652338891583, "grad_norm": 3.1109237670898438, "learning_rate": 4.163855835597614e-05, "loss": 7.8099, "step": 17826 }, { "epoch": 0.7516073950713578, "grad_norm": 3.463616371154785, "learning_rate": 4.163777756956807e-05, "loss": 6.4464, "step": 17827 }, { "epoch": 0.7516495562535573, "grad_norm": 2.012714385986328, "learning_rate": 4.163699678316e-05, "loss": 7.1388, "step": 17828 }, { "epoch": 0.7516917174357569, "grad_norm": 1.8817811012268066, "learning_rate": 4.163621599675193e-05, "loss": 6.8389, "step": 17829 }, { "epoch": 0.7517338786179565, "grad_norm": 1.424257516860962, "learning_rate": 4.163543521034386e-05, "loss": 6.4862, "step": 17830 }, { "epoch": 0.751776039800156, "grad_norm": 1.2209073305130005, "learning_rate": 4.163465442393579e-05, "loss": 6.3462, "step": 17831 }, { "epoch": 0.7518182009823555, "grad_norm": 4.0962371826171875, "learning_rate": 4.163387363752772e-05, "loss": 7.64, "step": 17832 }, { "epoch": 0.7518603621645551, "grad_norm": 1.383428692817688, "learning_rate": 4.163309285111965e-05, "loss": 6.6643, "step": 17833 }, { "epoch": 0.7519025233467547, "grad_norm": 1.3555841445922852, "learning_rate": 4.163231206471158e-05, "loss": 6.6488, "step": 17834 }, { "epoch": 0.7519446845289542, "grad_norm": 1.069919466972351, "learning_rate": 4.163153127830351e-05, "loss": 6.866, "step": 17835 }, { "epoch": 0.7519868457111537, "grad_norm": 3.1754322052001953, "learning_rate": 4.1630750491895435e-05, "loss": 6.5498, "step": 17836 }, { "epoch": 0.7520290068933533, "grad_norm": 3.498833417892456, "learning_rate": 4.1629969705487374e-05, "loss": 9.9898, "step": 17837 }, { "epoch": 0.7520711680755529, "grad_norm": 2.664731025695801, "learning_rate": 4.16291889190793e-05, "loss": 6.6396, "step": 17838 }, { "epoch": 0.7521133292577524, "grad_norm": 2.1950788497924805, "learning_rate": 4.162840813267123e-05, "loss": 6.883, "step": 17839 }, { "epoch": 0.7521554904399519, "grad_norm": 1.7847297191619873, "learning_rate": 4.162762734626316e-05, "loss": 7.0674, "step": 17840 }, { "epoch": 0.7521976516221515, "grad_norm": 1.7433580160140991, "learning_rate": 4.1626846559855085e-05, "loss": 6.6849, "step": 17841 }, { "epoch": 0.752239812804351, "grad_norm": 3.8354811668395996, "learning_rate": 4.162606577344702e-05, "loss": 7.0931, "step": 17842 }, { "epoch": 0.7522819739865506, "grad_norm": 2.498180627822876, "learning_rate": 4.162528498703895e-05, "loss": 6.9254, "step": 17843 }, { "epoch": 0.7523241351687502, "grad_norm": 6.109010219573975, "learning_rate": 4.162450420063088e-05, "loss": 9.6506, "step": 17844 }, { "epoch": 0.7523662963509496, "grad_norm": 2.267056941986084, "learning_rate": 4.162372341422281e-05, "loss": 7.2238, "step": 17845 }, { "epoch": 0.7524084575331492, "grad_norm": 2.407315254211426, "learning_rate": 4.1622942627814735e-05, "loss": 7.4296, "step": 17846 }, { "epoch": 0.7524506187153488, "grad_norm": 1.9989863634109497, "learning_rate": 4.1622161841406666e-05, "loss": 6.3035, "step": 17847 }, { "epoch": 0.7524927798975484, "grad_norm": 2.356984853744507, "learning_rate": 4.16213810549986e-05, "loss": 6.5833, "step": 17848 }, { "epoch": 0.7525349410797478, "grad_norm": 2.364365816116333, "learning_rate": 4.162060026859053e-05, "loss": 6.7044, "step": 17849 }, { "epoch": 0.7525771022619474, "grad_norm": 3.5118448734283447, "learning_rate": 4.161981948218245e-05, "loss": 7.7741, "step": 17850 }, { "epoch": 0.752619263444147, "grad_norm": 1.7412465810775757, "learning_rate": 4.1619038695774384e-05, "loss": 6.3933, "step": 17851 }, { "epoch": 0.7526614246263466, "grad_norm": 1.5776292085647583, "learning_rate": 4.1618257909366315e-05, "loss": 6.5954, "step": 17852 }, { "epoch": 0.752703585808546, "grad_norm": 4.03812837600708, "learning_rate": 4.1617477122958247e-05, "loss": 8.4806, "step": 17853 }, { "epoch": 0.7527457469907456, "grad_norm": 2.300767660140991, "learning_rate": 4.161669633655018e-05, "loss": 6.6003, "step": 17854 }, { "epoch": 0.7527879081729452, "grad_norm": 2.3209176063537598, "learning_rate": 4.16159155501421e-05, "loss": 7.0987, "step": 17855 }, { "epoch": 0.7528300693551447, "grad_norm": 1.8301630020141602, "learning_rate": 4.161513476373404e-05, "loss": 6.9571, "step": 17856 }, { "epoch": 0.7528722305373443, "grad_norm": 2.4364635944366455, "learning_rate": 4.1614353977325965e-05, "loss": 6.353, "step": 17857 }, { "epoch": 0.7529143917195438, "grad_norm": 2.7777695655822754, "learning_rate": 4.1613573190917896e-05, "loss": 7.5674, "step": 17858 }, { "epoch": 0.7529565529017433, "grad_norm": 1.5078626871109009, "learning_rate": 4.161279240450983e-05, "loss": 6.4128, "step": 17859 }, { "epoch": 0.7529987140839429, "grad_norm": 1.9840439558029175, "learning_rate": 4.161201161810175e-05, "loss": 6.6806, "step": 17860 }, { "epoch": 0.7530408752661425, "grad_norm": 2.826813220977783, "learning_rate": 4.161123083169368e-05, "loss": 6.4508, "step": 17861 }, { "epoch": 0.753083036448342, "grad_norm": 2.542664051055908, "learning_rate": 4.1610450045285614e-05, "loss": 6.5363, "step": 17862 }, { "epoch": 0.7531251976305415, "grad_norm": 3.257194995880127, "learning_rate": 4.1609669258877546e-05, "loss": 6.5039, "step": 17863 }, { "epoch": 0.7531673588127411, "grad_norm": 1.4191843271255493, "learning_rate": 4.160888847246947e-05, "loss": 6.3066, "step": 17864 }, { "epoch": 0.7532095199949407, "grad_norm": 2.8871283531188965, "learning_rate": 4.16081076860614e-05, "loss": 6.8587, "step": 17865 }, { "epoch": 0.7532516811771403, "grad_norm": 1.6091139316558838, "learning_rate": 4.160732689965333e-05, "loss": 6.3367, "step": 17866 }, { "epoch": 0.7532938423593397, "grad_norm": 2.4094622135162354, "learning_rate": 4.1606546113245264e-05, "loss": 6.5492, "step": 17867 }, { "epoch": 0.7533360035415393, "grad_norm": 4.352652549743652, "learning_rate": 4.1605765326837195e-05, "loss": 7.9706, "step": 17868 }, { "epoch": 0.7533781647237389, "grad_norm": 2.5718588829040527, "learning_rate": 4.160498454042912e-05, "loss": 6.7522, "step": 17869 }, { "epoch": 0.7534203259059384, "grad_norm": 2.2106592655181885, "learning_rate": 4.160420375402105e-05, "loss": 6.7921, "step": 17870 }, { "epoch": 0.7534624870881379, "grad_norm": 3.8723487854003906, "learning_rate": 4.160342296761298e-05, "loss": 7.868, "step": 17871 }, { "epoch": 0.7535046482703375, "grad_norm": 0.744708776473999, "learning_rate": 4.1602642181204907e-05, "loss": 6.2423, "step": 17872 }, { "epoch": 0.753546809452537, "grad_norm": 1.6934185028076172, "learning_rate": 4.1601861394796845e-05, "loss": 6.6385, "step": 17873 }, { "epoch": 0.7535889706347366, "grad_norm": 1.8861300945281982, "learning_rate": 4.160108060838877e-05, "loss": 6.4509, "step": 17874 }, { "epoch": 0.7536311318169362, "grad_norm": 2.5492405891418457, "learning_rate": 4.16002998219807e-05, "loss": 8.0856, "step": 17875 }, { "epoch": 0.7536732929991357, "grad_norm": 3.3700742721557617, "learning_rate": 4.159951903557263e-05, "loss": 7.4947, "step": 17876 }, { "epoch": 0.7537154541813352, "grad_norm": 3.5180647373199463, "learning_rate": 4.1598738249164556e-05, "loss": 7.3372, "step": 17877 }, { "epoch": 0.7537576153635348, "grad_norm": 4.944811820983887, "learning_rate": 4.1597957462756494e-05, "loss": 11.1338, "step": 17878 }, { "epoch": 0.7537997765457344, "grad_norm": 2.7187881469726562, "learning_rate": 4.159717667634842e-05, "loss": 6.7129, "step": 17879 }, { "epoch": 0.7538419377279338, "grad_norm": 2.2028894424438477, "learning_rate": 4.159639588994035e-05, "loss": 6.8923, "step": 17880 }, { "epoch": 0.7538840989101334, "grad_norm": 1.498643398284912, "learning_rate": 4.159561510353228e-05, "loss": 6.8856, "step": 17881 }, { "epoch": 0.753926260092333, "grad_norm": 1.6773715019226074, "learning_rate": 4.159483431712421e-05, "loss": 7.1656, "step": 17882 }, { "epoch": 0.7539684212745326, "grad_norm": 2.75361704826355, "learning_rate": 4.159405353071614e-05, "loss": 7.6905, "step": 17883 }, { "epoch": 0.7540105824567321, "grad_norm": 1.4746577739715576, "learning_rate": 4.159327274430807e-05, "loss": 6.6682, "step": 17884 }, { "epoch": 0.7540527436389316, "grad_norm": 1.282378911972046, "learning_rate": 4.15924919579e-05, "loss": 6.7118, "step": 17885 }, { "epoch": 0.7540949048211312, "grad_norm": 1.6959971189498901, "learning_rate": 4.1591711171491924e-05, "loss": 6.4735, "step": 17886 }, { "epoch": 0.7541370660033307, "grad_norm": 1.280309796333313, "learning_rate": 4.159093038508386e-05, "loss": 6.9773, "step": 17887 }, { "epoch": 0.7541792271855303, "grad_norm": 1.234519362449646, "learning_rate": 4.1590149598675786e-05, "loss": 6.7565, "step": 17888 }, { "epoch": 0.7542213883677298, "grad_norm": 1.6346964836120605, "learning_rate": 4.158936881226772e-05, "loss": 6.5833, "step": 17889 }, { "epoch": 0.7542635495499294, "grad_norm": 3.482334613800049, "learning_rate": 4.158858802585965e-05, "loss": 7.6182, "step": 17890 }, { "epoch": 0.7543057107321289, "grad_norm": 2.4911696910858154, "learning_rate": 4.1587807239451573e-05, "loss": 8.1363, "step": 17891 }, { "epoch": 0.7543478719143285, "grad_norm": 2.538607597351074, "learning_rate": 4.158702645304351e-05, "loss": 6.8549, "step": 17892 }, { "epoch": 0.7543900330965281, "grad_norm": 1.4046772718429565, "learning_rate": 4.1586245666635436e-05, "loss": 6.7277, "step": 17893 }, { "epoch": 0.7544321942787275, "grad_norm": 1.2454465627670288, "learning_rate": 4.158546488022737e-05, "loss": 6.9734, "step": 17894 }, { "epoch": 0.7544743554609271, "grad_norm": 1.2441645860671997, "learning_rate": 4.15846840938193e-05, "loss": 6.5382, "step": 17895 }, { "epoch": 0.7545165166431267, "grad_norm": 1.3957630395889282, "learning_rate": 4.158390330741122e-05, "loss": 6.5897, "step": 17896 }, { "epoch": 0.7545586778253263, "grad_norm": 2.474936008453369, "learning_rate": 4.1583122521003154e-05, "loss": 7.5503, "step": 17897 }, { "epoch": 0.7546008390075257, "grad_norm": 1.7872414588928223, "learning_rate": 4.1582341734595085e-05, "loss": 7.0239, "step": 17898 }, { "epoch": 0.7546430001897253, "grad_norm": 3.929185628890991, "learning_rate": 4.158156094818702e-05, "loss": 9.0846, "step": 17899 }, { "epoch": 0.7546851613719249, "grad_norm": 1.5441060066223145, "learning_rate": 4.158078016177895e-05, "loss": 6.8918, "step": 17900 }, { "epoch": 0.7547273225541244, "grad_norm": 1.7367210388183594, "learning_rate": 4.157999937537088e-05, "loss": 7.1273, "step": 17901 }, { "epoch": 0.754769483736324, "grad_norm": 2.492997169494629, "learning_rate": 4.1579218588962804e-05, "loss": 7.6034, "step": 17902 }, { "epoch": 0.7548116449185235, "grad_norm": 1.5700485706329346, "learning_rate": 4.1578437802554735e-05, "loss": 6.5019, "step": 17903 }, { "epoch": 0.7548538061007231, "grad_norm": 1.3717235326766968, "learning_rate": 4.1577657016146666e-05, "loss": 6.6766, "step": 17904 }, { "epoch": 0.7548959672829226, "grad_norm": 1.4222323894500732, "learning_rate": 4.157687622973859e-05, "loss": 6.4196, "step": 17905 }, { "epoch": 0.7549381284651222, "grad_norm": 1.4547992944717407, "learning_rate": 4.157609544333053e-05, "loss": 6.9492, "step": 17906 }, { "epoch": 0.7549802896473217, "grad_norm": 1.3903841972351074, "learning_rate": 4.157531465692245e-05, "loss": 6.4236, "step": 17907 }, { "epoch": 0.7550224508295212, "grad_norm": 1.8920576572418213, "learning_rate": 4.1574533870514385e-05, "loss": 7.2087, "step": 17908 }, { "epoch": 0.7550646120117208, "grad_norm": 1.2833493947982788, "learning_rate": 4.1573753084106316e-05, "loss": 6.6245, "step": 17909 }, { "epoch": 0.7551067731939204, "grad_norm": 1.3976662158966064, "learning_rate": 4.157297229769824e-05, "loss": 7.014, "step": 17910 }, { "epoch": 0.7551489343761199, "grad_norm": 1.4480514526367188, "learning_rate": 4.157219151129018e-05, "loss": 6.7578, "step": 17911 }, { "epoch": 0.7551910955583194, "grad_norm": 0.9863097071647644, "learning_rate": 4.15714107248821e-05, "loss": 6.7909, "step": 17912 }, { "epoch": 0.755233256740519, "grad_norm": 3.156926393508911, "learning_rate": 4.1570629938474034e-05, "loss": 8.2442, "step": 17913 }, { "epoch": 0.7552754179227186, "grad_norm": 4.709752082824707, "learning_rate": 4.1569849152065965e-05, "loss": 8.0363, "step": 17914 }, { "epoch": 0.7553175791049181, "grad_norm": 2.572855234146118, "learning_rate": 4.156906836565789e-05, "loss": 7.2483, "step": 17915 }, { "epoch": 0.7553597402871176, "grad_norm": 2.3248226642608643, "learning_rate": 4.156828757924982e-05, "loss": 6.7374, "step": 17916 }, { "epoch": 0.7554019014693172, "grad_norm": 4.7265729904174805, "learning_rate": 4.156750679284175e-05, "loss": 8.1609, "step": 17917 }, { "epoch": 0.7554440626515168, "grad_norm": 1.3587380647659302, "learning_rate": 4.1566726006433684e-05, "loss": 6.8006, "step": 17918 }, { "epoch": 0.7554862238337163, "grad_norm": 2.109137773513794, "learning_rate": 4.156594522002561e-05, "loss": 7.0384, "step": 17919 }, { "epoch": 0.7555283850159158, "grad_norm": 3.640275001525879, "learning_rate": 4.1565164433617546e-05, "loss": 7.564, "step": 17920 }, { "epoch": 0.7555705461981154, "grad_norm": 1.7402688264846802, "learning_rate": 4.156438364720947e-05, "loss": 6.8698, "step": 17921 }, { "epoch": 0.7556127073803149, "grad_norm": 3.88082218170166, "learning_rate": 4.15636028608014e-05, "loss": 7.3551, "step": 17922 }, { "epoch": 0.7556548685625145, "grad_norm": 2.078521251678467, "learning_rate": 4.156282207439333e-05, "loss": 7.0809, "step": 17923 }, { "epoch": 0.7556970297447141, "grad_norm": 1.5752590894699097, "learning_rate": 4.156204128798526e-05, "loss": 6.9985, "step": 17924 }, { "epoch": 0.7557391909269136, "grad_norm": 1.7109357118606567, "learning_rate": 4.1561260501577196e-05, "loss": 7.016, "step": 17925 }, { "epoch": 0.7557813521091131, "grad_norm": 1.2996888160705566, "learning_rate": 4.156047971516912e-05, "loss": 6.4989, "step": 17926 }, { "epoch": 0.7558235132913127, "grad_norm": 1.8653067350387573, "learning_rate": 4.155969892876105e-05, "loss": 7.0511, "step": 17927 }, { "epoch": 0.7558656744735123, "grad_norm": 3.0972325801849365, "learning_rate": 4.155891814235298e-05, "loss": 7.1922, "step": 17928 }, { "epoch": 0.7559078356557117, "grad_norm": 1.8000611066818237, "learning_rate": 4.155813735594491e-05, "loss": 7.109, "step": 17929 }, { "epoch": 0.7559499968379113, "grad_norm": 1.4271389245986938, "learning_rate": 4.155735656953684e-05, "loss": 7.198, "step": 17930 }, { "epoch": 0.7559921580201109, "grad_norm": 3.965409517288208, "learning_rate": 4.155657578312877e-05, "loss": 7.3849, "step": 17931 }, { "epoch": 0.7560343192023105, "grad_norm": 1.9547104835510254, "learning_rate": 4.15557949967207e-05, "loss": 6.7024, "step": 17932 }, { "epoch": 0.75607648038451, "grad_norm": 2.6230785846710205, "learning_rate": 4.155501421031263e-05, "loss": 7.4637, "step": 17933 }, { "epoch": 0.7561186415667095, "grad_norm": 1.5298259258270264, "learning_rate": 4.155423342390456e-05, "loss": 7.5343, "step": 17934 }, { "epoch": 0.7561608027489091, "grad_norm": 2.059434175491333, "learning_rate": 4.155345263749649e-05, "loss": 7.6062, "step": 17935 }, { "epoch": 0.7562029639311086, "grad_norm": 2.163438320159912, "learning_rate": 4.155267185108842e-05, "loss": 6.882, "step": 17936 }, { "epoch": 0.7562451251133082, "grad_norm": 3.1464734077453613, "learning_rate": 4.155189106468035e-05, "loss": 8.4701, "step": 17937 }, { "epoch": 0.7562872862955077, "grad_norm": 3.1137444972991943, "learning_rate": 4.1551110278272275e-05, "loss": 7.2057, "step": 17938 }, { "epoch": 0.7563294474777073, "grad_norm": 1.687015414237976, "learning_rate": 4.1550329491864206e-05, "loss": 6.8651, "step": 17939 }, { "epoch": 0.7563716086599068, "grad_norm": 3.428175687789917, "learning_rate": 4.154954870545614e-05, "loss": 6.4562, "step": 17940 }, { "epoch": 0.7564137698421064, "grad_norm": 3.3268425464630127, "learning_rate": 4.154876791904806e-05, "loss": 6.4758, "step": 17941 }, { "epoch": 0.756455931024306, "grad_norm": 1.5741264820098877, "learning_rate": 4.154798713264e-05, "loss": 6.4086, "step": 17942 }, { "epoch": 0.7564980922065054, "grad_norm": 1.5806893110275269, "learning_rate": 4.1547206346231924e-05, "loss": 6.8118, "step": 17943 }, { "epoch": 0.756540253388705, "grad_norm": 1.9072272777557373, "learning_rate": 4.1546425559823856e-05, "loss": 6.6303, "step": 17944 }, { "epoch": 0.7565824145709046, "grad_norm": 4.08966588973999, "learning_rate": 4.154564477341579e-05, "loss": 7.5373, "step": 17945 }, { "epoch": 0.7566245757531042, "grad_norm": 4.290695667266846, "learning_rate": 4.154486398700772e-05, "loss": 7.8231, "step": 17946 }, { "epoch": 0.7566667369353036, "grad_norm": 1.9186123609542847, "learning_rate": 4.154408320059965e-05, "loss": 6.9852, "step": 17947 }, { "epoch": 0.7567088981175032, "grad_norm": 2.23957896232605, "learning_rate": 4.1543302414191574e-05, "loss": 7.1273, "step": 17948 }, { "epoch": 0.7567510592997028, "grad_norm": 2.1469905376434326, "learning_rate": 4.1542521627783505e-05, "loss": 6.6397, "step": 17949 }, { "epoch": 0.7567932204819023, "grad_norm": 3.0557379722595215, "learning_rate": 4.1541740841375436e-05, "loss": 7.686, "step": 17950 }, { "epoch": 0.7568353816641019, "grad_norm": 2.092785358428955, "learning_rate": 4.154096005496737e-05, "loss": 6.9083, "step": 17951 }, { "epoch": 0.7568775428463014, "grad_norm": 1.5127577781677246, "learning_rate": 4.154017926855929e-05, "loss": 7.0018, "step": 17952 }, { "epoch": 0.756919704028501, "grad_norm": 3.1358230113983154, "learning_rate": 4.1539398482151223e-05, "loss": 6.8011, "step": 17953 }, { "epoch": 0.7569618652107005, "grad_norm": 1.3364890813827515, "learning_rate": 4.1538617695743155e-05, "loss": 6.6209, "step": 17954 }, { "epoch": 0.7570040263929001, "grad_norm": 1.4368338584899902, "learning_rate": 4.153783690933508e-05, "loss": 6.7893, "step": 17955 }, { "epoch": 0.7570461875750996, "grad_norm": 1.870023488998413, "learning_rate": 4.153705612292702e-05, "loss": 6.4345, "step": 17956 }, { "epoch": 0.7570883487572991, "grad_norm": 1.4400386810302734, "learning_rate": 4.153627533651894e-05, "loss": 6.5843, "step": 17957 }, { "epoch": 0.7571305099394987, "grad_norm": 2.3492424488067627, "learning_rate": 4.153549455011087e-05, "loss": 6.7326, "step": 17958 }, { "epoch": 0.7571726711216983, "grad_norm": 3.066997766494751, "learning_rate": 4.1534713763702804e-05, "loss": 6.3823, "step": 17959 }, { "epoch": 0.7572148323038979, "grad_norm": 0.8841073513031006, "learning_rate": 4.153393297729473e-05, "loss": 6.4089, "step": 17960 }, { "epoch": 0.7572569934860973, "grad_norm": 1.4307833909988403, "learning_rate": 4.153315219088667e-05, "loss": 6.7602, "step": 17961 }, { "epoch": 0.7572991546682969, "grad_norm": 1.359289526939392, "learning_rate": 4.153237140447859e-05, "loss": 7.1754, "step": 17962 }, { "epoch": 0.7573413158504965, "grad_norm": 1.403520107269287, "learning_rate": 4.153159061807052e-05, "loss": 6.4072, "step": 17963 }, { "epoch": 0.757383477032696, "grad_norm": 3.58732271194458, "learning_rate": 4.1530809831662454e-05, "loss": 7.5309, "step": 17964 }, { "epoch": 0.7574256382148955, "grad_norm": 2.8208017349243164, "learning_rate": 4.1530029045254385e-05, "loss": 7.9737, "step": 17965 }, { "epoch": 0.7574677993970951, "grad_norm": 2.451296329498291, "learning_rate": 4.152924825884631e-05, "loss": 7.5537, "step": 17966 }, { "epoch": 0.7575099605792946, "grad_norm": 1.5641343593597412, "learning_rate": 4.152846747243824e-05, "loss": 6.8234, "step": 17967 }, { "epoch": 0.7575521217614942, "grad_norm": 5.175034999847412, "learning_rate": 4.152768668603017e-05, "loss": 11.0668, "step": 17968 }, { "epoch": 0.7575942829436937, "grad_norm": 4.261799335479736, "learning_rate": 4.15269058996221e-05, "loss": 10.1026, "step": 17969 }, { "epoch": 0.7576364441258933, "grad_norm": 1.9169646501541138, "learning_rate": 4.1526125113214035e-05, "loss": 6.989, "step": 17970 }, { "epoch": 0.7576786053080928, "grad_norm": 3.3053829669952393, "learning_rate": 4.152534432680596e-05, "loss": 7.7136, "step": 17971 }, { "epoch": 0.7577207664902924, "grad_norm": 2.5383119583129883, "learning_rate": 4.152456354039789e-05, "loss": 6.5441, "step": 17972 }, { "epoch": 0.757762927672492, "grad_norm": 2.262568235397339, "learning_rate": 4.152378275398982e-05, "loss": 6.6391, "step": 17973 }, { "epoch": 0.7578050888546914, "grad_norm": 1.8969112634658813, "learning_rate": 4.1523001967581746e-05, "loss": 6.4449, "step": 17974 }, { "epoch": 0.757847250036891, "grad_norm": 2.8501672744750977, "learning_rate": 4.1522221181173684e-05, "loss": 7.5652, "step": 17975 }, { "epoch": 0.7578894112190906, "grad_norm": 3.318937063217163, "learning_rate": 4.152144039476561e-05, "loss": 7.4996, "step": 17976 }, { "epoch": 0.7579315724012902, "grad_norm": 3.254805326461792, "learning_rate": 4.152065960835754e-05, "loss": 6.9705, "step": 17977 }, { "epoch": 0.7579737335834896, "grad_norm": 1.5862215757369995, "learning_rate": 4.151987882194947e-05, "loss": 6.8912, "step": 17978 }, { "epoch": 0.7580158947656892, "grad_norm": 2.6296725273132324, "learning_rate": 4.1519098035541396e-05, "loss": 7.4292, "step": 17979 }, { "epoch": 0.7580580559478888, "grad_norm": 4.33730936050415, "learning_rate": 4.1518317249133334e-05, "loss": 7.242, "step": 17980 }, { "epoch": 0.7581002171300883, "grad_norm": 3.6927480697631836, "learning_rate": 4.151753646272526e-05, "loss": 7.3529, "step": 17981 }, { "epoch": 0.7581423783122879, "grad_norm": 2.3378751277923584, "learning_rate": 4.151675567631719e-05, "loss": 6.9507, "step": 17982 }, { "epoch": 0.7581845394944874, "grad_norm": 2.9381356239318848, "learning_rate": 4.151597488990912e-05, "loss": 6.7492, "step": 17983 }, { "epoch": 0.758226700676687, "grad_norm": 1.6816562414169312, "learning_rate": 4.1515194103501045e-05, "loss": 6.4182, "step": 17984 }, { "epoch": 0.7582688618588865, "grad_norm": 3.233978748321533, "learning_rate": 4.1514413317092976e-05, "loss": 7.2583, "step": 17985 }, { "epoch": 0.7583110230410861, "grad_norm": 3.745414972305298, "learning_rate": 4.151363253068491e-05, "loss": 7.9403, "step": 17986 }, { "epoch": 0.7583531842232856, "grad_norm": 1.4321694374084473, "learning_rate": 4.151285174427684e-05, "loss": 6.5258, "step": 17987 }, { "epoch": 0.7583953454054851, "grad_norm": 1.8846912384033203, "learning_rate": 4.151207095786876e-05, "loss": 7.0955, "step": 17988 }, { "epoch": 0.7584375065876847, "grad_norm": 1.1224844455718994, "learning_rate": 4.15112901714607e-05, "loss": 6.7123, "step": 17989 }, { "epoch": 0.7584796677698843, "grad_norm": 2.4287784099578857, "learning_rate": 4.1510509385052626e-05, "loss": 6.4745, "step": 17990 }, { "epoch": 0.7585218289520839, "grad_norm": 3.0836501121520996, "learning_rate": 4.150972859864456e-05, "loss": 7.6577, "step": 17991 }, { "epoch": 0.7585639901342833, "grad_norm": 3.393378973007202, "learning_rate": 4.150894781223649e-05, "loss": 7.5225, "step": 17992 }, { "epoch": 0.7586061513164829, "grad_norm": 3.474778890609741, "learning_rate": 4.150816702582841e-05, "loss": 7.4341, "step": 17993 }, { "epoch": 0.7586483124986825, "grad_norm": 2.731351852416992, "learning_rate": 4.150738623942035e-05, "loss": 7.0569, "step": 17994 }, { "epoch": 0.758690473680882, "grad_norm": 2.1569464206695557, "learning_rate": 4.1506605453012275e-05, "loss": 6.5776, "step": 17995 }, { "epoch": 0.7587326348630815, "grad_norm": 3.1393465995788574, "learning_rate": 4.150582466660421e-05, "loss": 7.5054, "step": 17996 }, { "epoch": 0.7587747960452811, "grad_norm": 3.1283631324768066, "learning_rate": 4.150504388019614e-05, "loss": 7.0274, "step": 17997 }, { "epoch": 0.7588169572274807, "grad_norm": 5.871015548706055, "learning_rate": 4.150426309378806e-05, "loss": 7.6231, "step": 17998 }, { "epoch": 0.7588591184096802, "grad_norm": 4.51618766784668, "learning_rate": 4.1503482307379994e-05, "loss": 6.9542, "step": 17999 }, { "epoch": 0.7589012795918798, "grad_norm": 2.755932331085205, "learning_rate": 4.1502701520971925e-05, "loss": 6.9937, "step": 18000 }, { "epoch": 0.7589434407740793, "grad_norm": 4.033479690551758, "learning_rate": 4.1501920734563856e-05, "loss": 7.3164, "step": 18001 }, { "epoch": 0.7589856019562788, "grad_norm": 1.6415327787399292, "learning_rate": 4.150113994815579e-05, "loss": 6.7236, "step": 18002 }, { "epoch": 0.7590277631384784, "grad_norm": 2.567704439163208, "learning_rate": 4.150035916174771e-05, "loss": 6.5049, "step": 18003 }, { "epoch": 0.759069924320678, "grad_norm": 2.5613276958465576, "learning_rate": 4.149957837533964e-05, "loss": 6.7664, "step": 18004 }, { "epoch": 0.7591120855028775, "grad_norm": 3.139443874359131, "learning_rate": 4.1498797588931574e-05, "loss": 7.5631, "step": 18005 }, { "epoch": 0.759154246685077, "grad_norm": 1.8584342002868652, "learning_rate": 4.1498016802523506e-05, "loss": 6.892, "step": 18006 }, { "epoch": 0.7591964078672766, "grad_norm": 1.921966314315796, "learning_rate": 4.149723601611543e-05, "loss": 6.5136, "step": 18007 }, { "epoch": 0.7592385690494762, "grad_norm": 1.236875057220459, "learning_rate": 4.149645522970737e-05, "loss": 6.4602, "step": 18008 }, { "epoch": 0.7592807302316757, "grad_norm": 2.3546342849731445, "learning_rate": 4.149567444329929e-05, "loss": 6.6399, "step": 18009 }, { "epoch": 0.7593228914138752, "grad_norm": 2.7046961784362793, "learning_rate": 4.149489365689122e-05, "loss": 6.6059, "step": 18010 }, { "epoch": 0.7593650525960748, "grad_norm": 1.9672271013259888, "learning_rate": 4.1494112870483155e-05, "loss": 6.4644, "step": 18011 }, { "epoch": 0.7594072137782744, "grad_norm": 1.4123166799545288, "learning_rate": 4.149333208407508e-05, "loss": 6.4125, "step": 18012 }, { "epoch": 0.7594493749604739, "grad_norm": 1.1453006267547607, "learning_rate": 4.149255129766701e-05, "loss": 6.4797, "step": 18013 }, { "epoch": 0.7594915361426734, "grad_norm": 1.7934486865997314, "learning_rate": 4.149177051125894e-05, "loss": 6.5193, "step": 18014 }, { "epoch": 0.759533697324873, "grad_norm": 3.464460849761963, "learning_rate": 4.1490989724850873e-05, "loss": 6.9765, "step": 18015 }, { "epoch": 0.7595758585070725, "grad_norm": 2.789332151412964, "learning_rate": 4.1490208938442805e-05, "loss": 8.222, "step": 18016 }, { "epoch": 0.7596180196892721, "grad_norm": 1.8949549198150635, "learning_rate": 4.148942815203473e-05, "loss": 7.0535, "step": 18017 }, { "epoch": 0.7596601808714717, "grad_norm": 1.9373228549957275, "learning_rate": 4.148864736562666e-05, "loss": 6.8935, "step": 18018 }, { "epoch": 0.7597023420536712, "grad_norm": 4.046447277069092, "learning_rate": 4.148786657921859e-05, "loss": 7.1281, "step": 18019 }, { "epoch": 0.7597445032358707, "grad_norm": 2.7619519233703613, "learning_rate": 4.148708579281052e-05, "loss": 7.4158, "step": 18020 }, { "epoch": 0.7597866644180703, "grad_norm": 3.143298864364624, "learning_rate": 4.148630500640245e-05, "loss": 7.4731, "step": 18021 }, { "epoch": 0.7598288256002699, "grad_norm": 1.8679828643798828, "learning_rate": 4.148552421999438e-05, "loss": 6.4998, "step": 18022 }, { "epoch": 0.7598709867824693, "grad_norm": 3.926499843597412, "learning_rate": 4.148474343358631e-05, "loss": 8.0171, "step": 18023 }, { "epoch": 0.7599131479646689, "grad_norm": 3.488771677017212, "learning_rate": 4.1483962647178234e-05, "loss": 7.1371, "step": 18024 }, { "epoch": 0.7599553091468685, "grad_norm": 1.6913710832595825, "learning_rate": 4.148318186077017e-05, "loss": 6.4159, "step": 18025 }, { "epoch": 0.7599974703290681, "grad_norm": 2.0483529567718506, "learning_rate": 4.14824010743621e-05, "loss": 6.745, "step": 18026 }, { "epoch": 0.7600396315112675, "grad_norm": 2.4241487979888916, "learning_rate": 4.148162028795403e-05, "loss": 6.4493, "step": 18027 }, { "epoch": 0.7600817926934671, "grad_norm": 1.0904639959335327, "learning_rate": 4.148083950154596e-05, "loss": 6.5662, "step": 18028 }, { "epoch": 0.7601239538756667, "grad_norm": 2.3751542568206787, "learning_rate": 4.1480058715137884e-05, "loss": 6.3941, "step": 18029 }, { "epoch": 0.7601661150578662, "grad_norm": 2.829036235809326, "learning_rate": 4.147927792872982e-05, "loss": 6.3506, "step": 18030 }, { "epoch": 0.7602082762400658, "grad_norm": 3.610640048980713, "learning_rate": 4.1478497142321747e-05, "loss": 7.7552, "step": 18031 }, { "epoch": 0.7602504374222653, "grad_norm": 2.6027779579162598, "learning_rate": 4.147771635591368e-05, "loss": 6.4167, "step": 18032 }, { "epoch": 0.7602925986044649, "grad_norm": 4.681455135345459, "learning_rate": 4.147693556950561e-05, "loss": 7.2375, "step": 18033 }, { "epoch": 0.7603347597866644, "grad_norm": 3.6150176525115967, "learning_rate": 4.147615478309754e-05, "loss": 7.6631, "step": 18034 }, { "epoch": 0.760376920968864, "grad_norm": 1.4001970291137695, "learning_rate": 4.1475373996689465e-05, "loss": 6.77, "step": 18035 }, { "epoch": 0.7604190821510635, "grad_norm": 1.447845220565796, "learning_rate": 4.1474593210281396e-05, "loss": 6.354, "step": 18036 }, { "epoch": 0.760461243333263, "grad_norm": 2.3328349590301514, "learning_rate": 4.147381242387333e-05, "loss": 7.089, "step": 18037 }, { "epoch": 0.7605034045154626, "grad_norm": 1.549928069114685, "learning_rate": 4.147303163746526e-05, "loss": 6.7077, "step": 18038 }, { "epoch": 0.7605455656976622, "grad_norm": 3.552288293838501, "learning_rate": 4.147225085105719e-05, "loss": 7.7207, "step": 18039 }, { "epoch": 0.7605877268798618, "grad_norm": 3.9565536975860596, "learning_rate": 4.1471470064649114e-05, "loss": 7.7925, "step": 18040 }, { "epoch": 0.7606298880620612, "grad_norm": 2.1734273433685303, "learning_rate": 4.1470689278241046e-05, "loss": 7.0425, "step": 18041 }, { "epoch": 0.7606720492442608, "grad_norm": 1.7547763586044312, "learning_rate": 4.146990849183298e-05, "loss": 6.4719, "step": 18042 }, { "epoch": 0.7607142104264604, "grad_norm": 3.132856845855713, "learning_rate": 4.14691277054249e-05, "loss": 7.495, "step": 18043 }, { "epoch": 0.7607563716086599, "grad_norm": 2.2115516662597656, "learning_rate": 4.146834691901684e-05, "loss": 6.4098, "step": 18044 }, { "epoch": 0.7607985327908594, "grad_norm": 1.7832756042480469, "learning_rate": 4.1467566132608764e-05, "loss": 6.7883, "step": 18045 }, { "epoch": 0.760840693973059, "grad_norm": 1.6035550832748413, "learning_rate": 4.1466785346200695e-05, "loss": 7.3774, "step": 18046 }, { "epoch": 0.7608828551552586, "grad_norm": 1.197148084640503, "learning_rate": 4.1466004559792626e-05, "loss": 6.498, "step": 18047 }, { "epoch": 0.7609250163374581, "grad_norm": 3.741802930831909, "learning_rate": 4.146522377338455e-05, "loss": 7.8017, "step": 18048 }, { "epoch": 0.7609671775196577, "grad_norm": 2.3219094276428223, "learning_rate": 4.146444298697649e-05, "loss": 6.6404, "step": 18049 }, { "epoch": 0.7610093387018572, "grad_norm": 2.397067070007324, "learning_rate": 4.146366220056841e-05, "loss": 7.076, "step": 18050 }, { "epoch": 0.7610514998840567, "grad_norm": 1.693085789680481, "learning_rate": 4.1462881414160345e-05, "loss": 6.7891, "step": 18051 }, { "epoch": 0.7610936610662563, "grad_norm": 1.945673942565918, "learning_rate": 4.1462100627752276e-05, "loss": 6.7595, "step": 18052 }, { "epoch": 0.7611358222484559, "grad_norm": 3.3964085578918457, "learning_rate": 4.146131984134421e-05, "loss": 7.8898, "step": 18053 }, { "epoch": 0.7611779834306553, "grad_norm": 2.584839105606079, "learning_rate": 4.146053905493613e-05, "loss": 6.6209, "step": 18054 }, { "epoch": 0.7612201446128549, "grad_norm": 2.013734817504883, "learning_rate": 4.145975826852806e-05, "loss": 6.6132, "step": 18055 }, { "epoch": 0.7612623057950545, "grad_norm": 2.184601306915283, "learning_rate": 4.1458977482119994e-05, "loss": 7.2203, "step": 18056 }, { "epoch": 0.7613044669772541, "grad_norm": 2.468961000442505, "learning_rate": 4.145819669571192e-05, "loss": 6.7896, "step": 18057 }, { "epoch": 0.7613466281594536, "grad_norm": 2.910921096801758, "learning_rate": 4.145741590930386e-05, "loss": 6.2671, "step": 18058 }, { "epoch": 0.7613887893416531, "grad_norm": 2.653778553009033, "learning_rate": 4.145663512289578e-05, "loss": 6.9471, "step": 18059 }, { "epoch": 0.7614309505238527, "grad_norm": 2.4780211448669434, "learning_rate": 4.145585433648771e-05, "loss": 6.8643, "step": 18060 }, { "epoch": 0.7614731117060523, "grad_norm": 1.8444418907165527, "learning_rate": 4.1455073550079644e-05, "loss": 6.6107, "step": 18061 }, { "epoch": 0.7615152728882518, "grad_norm": 2.373824119567871, "learning_rate": 4.145429276367157e-05, "loss": 7.5326, "step": 18062 }, { "epoch": 0.7615574340704513, "grad_norm": 4.6713738441467285, "learning_rate": 4.1453511977263506e-05, "loss": 8.0133, "step": 18063 }, { "epoch": 0.7615995952526509, "grad_norm": 2.911898374557495, "learning_rate": 4.145273119085543e-05, "loss": 6.7937, "step": 18064 }, { "epoch": 0.7616417564348504, "grad_norm": 2.732522487640381, "learning_rate": 4.145195040444736e-05, "loss": 6.9239, "step": 18065 }, { "epoch": 0.76168391761705, "grad_norm": 1.4551516771316528, "learning_rate": 4.145116961803929e-05, "loss": 7.2319, "step": 18066 }, { "epoch": 0.7617260787992496, "grad_norm": 3.152907133102417, "learning_rate": 4.145038883163122e-05, "loss": 7.369, "step": 18067 }, { "epoch": 0.761768239981449, "grad_norm": 2.5474462509155273, "learning_rate": 4.144960804522315e-05, "loss": 7.2075, "step": 18068 }, { "epoch": 0.7618104011636486, "grad_norm": 1.7854398488998413, "learning_rate": 4.144882725881508e-05, "loss": 6.8913, "step": 18069 }, { "epoch": 0.7618525623458482, "grad_norm": 2.756864309310913, "learning_rate": 4.144804647240701e-05, "loss": 7.0504, "step": 18070 }, { "epoch": 0.7618947235280478, "grad_norm": 3.71681809425354, "learning_rate": 4.144726568599894e-05, "loss": 6.5091, "step": 18071 }, { "epoch": 0.7619368847102472, "grad_norm": 3.1983108520507812, "learning_rate": 4.144648489959087e-05, "loss": 6.3689, "step": 18072 }, { "epoch": 0.7619790458924468, "grad_norm": 4.1649298667907715, "learning_rate": 4.14457041131828e-05, "loss": 6.9567, "step": 18073 }, { "epoch": 0.7620212070746464, "grad_norm": 1.6283787488937378, "learning_rate": 4.144492332677473e-05, "loss": 7.8611, "step": 18074 }, { "epoch": 0.762063368256846, "grad_norm": 2.5072288513183594, "learning_rate": 4.144414254036666e-05, "loss": 6.5968, "step": 18075 }, { "epoch": 0.7621055294390455, "grad_norm": 4.53935432434082, "learning_rate": 4.1443361753958585e-05, "loss": 9.6582, "step": 18076 }, { "epoch": 0.762147690621245, "grad_norm": 1.7410420179367065, "learning_rate": 4.1442580967550524e-05, "loss": 6.725, "step": 18077 }, { "epoch": 0.7621898518034446, "grad_norm": 2.8013033866882324, "learning_rate": 4.144180018114245e-05, "loss": 8.1979, "step": 18078 }, { "epoch": 0.7622320129856441, "grad_norm": 2.5937509536743164, "learning_rate": 4.144101939473438e-05, "loss": 6.4912, "step": 18079 }, { "epoch": 0.7622741741678437, "grad_norm": 2.262057065963745, "learning_rate": 4.144023860832631e-05, "loss": 6.8148, "step": 18080 }, { "epoch": 0.7623163353500432, "grad_norm": 2.373847723007202, "learning_rate": 4.1439457821918235e-05, "loss": 7.3196, "step": 18081 }, { "epoch": 0.7623584965322427, "grad_norm": 2.054478406906128, "learning_rate": 4.143867703551017e-05, "loss": 6.4599, "step": 18082 }, { "epoch": 0.7624006577144423, "grad_norm": 2.8819918632507324, "learning_rate": 4.14378962491021e-05, "loss": 6.6368, "step": 18083 }, { "epoch": 0.7624428188966419, "grad_norm": 1.498081088066101, "learning_rate": 4.143711546269403e-05, "loss": 6.635, "step": 18084 }, { "epoch": 0.7624849800788414, "grad_norm": 1.3680249452590942, "learning_rate": 4.143633467628596e-05, "loss": 6.8669, "step": 18085 }, { "epoch": 0.7625271412610409, "grad_norm": 2.585271120071411, "learning_rate": 4.1435553889877885e-05, "loss": 6.4891, "step": 18086 }, { "epoch": 0.7625693024432405, "grad_norm": 2.5058629512786865, "learning_rate": 4.1434773103469816e-05, "loss": 6.9815, "step": 18087 }, { "epoch": 0.7626114636254401, "grad_norm": 2.150869607925415, "learning_rate": 4.143399231706175e-05, "loss": 7.282, "step": 18088 }, { "epoch": 0.7626536248076397, "grad_norm": 1.7390185594558716, "learning_rate": 4.143321153065368e-05, "loss": 6.4167, "step": 18089 }, { "epoch": 0.7626957859898391, "grad_norm": 1.3734310865402222, "learning_rate": 4.14324307442456e-05, "loss": 6.705, "step": 18090 }, { "epoch": 0.7627379471720387, "grad_norm": 2.585352897644043, "learning_rate": 4.1431649957837534e-05, "loss": 7.0576, "step": 18091 }, { "epoch": 0.7627801083542383, "grad_norm": 2.6132938861846924, "learning_rate": 4.1430869171429465e-05, "loss": 6.4721, "step": 18092 }, { "epoch": 0.7628222695364378, "grad_norm": 1.5295685529708862, "learning_rate": 4.143008838502139e-05, "loss": 6.5985, "step": 18093 }, { "epoch": 0.7628644307186373, "grad_norm": 2.2763190269470215, "learning_rate": 4.142930759861333e-05, "loss": 6.6094, "step": 18094 }, { "epoch": 0.7629065919008369, "grad_norm": 1.1999454498291016, "learning_rate": 4.142852681220525e-05, "loss": 7.1706, "step": 18095 }, { "epoch": 0.7629487530830364, "grad_norm": 1.867653250694275, "learning_rate": 4.142774602579719e-05, "loss": 6.9561, "step": 18096 }, { "epoch": 0.762990914265236, "grad_norm": 2.6378371715545654, "learning_rate": 4.1426965239389115e-05, "loss": 6.186, "step": 18097 }, { "epoch": 0.7630330754474356, "grad_norm": 4.525838375091553, "learning_rate": 4.1426184452981046e-05, "loss": 7.2402, "step": 18098 }, { "epoch": 0.7630752366296351, "grad_norm": 1.6674329042434692, "learning_rate": 4.142540366657298e-05, "loss": 6.8457, "step": 18099 }, { "epoch": 0.7631173978118346, "grad_norm": 4.2444353103637695, "learning_rate": 4.14246228801649e-05, "loss": 7.9012, "step": 18100 }, { "epoch": 0.7631595589940342, "grad_norm": 1.6039047241210938, "learning_rate": 4.142384209375683e-05, "loss": 6.5321, "step": 18101 }, { "epoch": 0.7632017201762338, "grad_norm": 2.169841766357422, "learning_rate": 4.1423061307348764e-05, "loss": 6.5072, "step": 18102 }, { "epoch": 0.7632438813584332, "grad_norm": 1.526963472366333, "learning_rate": 4.1422280520940696e-05, "loss": 6.568, "step": 18103 }, { "epoch": 0.7632860425406328, "grad_norm": 1.8113285303115845, "learning_rate": 4.142149973453262e-05, "loss": 6.4375, "step": 18104 }, { "epoch": 0.7633282037228324, "grad_norm": 1.9181166887283325, "learning_rate": 4.142071894812455e-05, "loss": 6.516, "step": 18105 }, { "epoch": 0.763370364905032, "grad_norm": 1.7988648414611816, "learning_rate": 4.141993816171648e-05, "loss": 6.6639, "step": 18106 }, { "epoch": 0.7634125260872315, "grad_norm": 2.006044864654541, "learning_rate": 4.1419157375308414e-05, "loss": 6.7484, "step": 18107 }, { "epoch": 0.763454687269431, "grad_norm": 2.1558728218078613, "learning_rate": 4.1418376588900345e-05, "loss": 6.843, "step": 18108 }, { "epoch": 0.7634968484516306, "grad_norm": 1.6971737146377563, "learning_rate": 4.141759580249227e-05, "loss": 6.4932, "step": 18109 }, { "epoch": 0.7635390096338301, "grad_norm": 1.6357669830322266, "learning_rate": 4.14168150160842e-05, "loss": 6.9602, "step": 18110 }, { "epoch": 0.7635811708160297, "grad_norm": 3.40804386138916, "learning_rate": 4.141603422967613e-05, "loss": 7.4008, "step": 18111 }, { "epoch": 0.7636233319982292, "grad_norm": 1.1639670133590698, "learning_rate": 4.1415253443268057e-05, "loss": 6.8381, "step": 18112 }, { "epoch": 0.7636654931804288, "grad_norm": 1.833106517791748, "learning_rate": 4.1414472656859995e-05, "loss": 6.6752, "step": 18113 }, { "epoch": 0.7637076543626283, "grad_norm": 0.991388201713562, "learning_rate": 4.141369187045192e-05, "loss": 6.6835, "step": 18114 }, { "epoch": 0.7637498155448279, "grad_norm": 1.1421594619750977, "learning_rate": 4.141291108404385e-05, "loss": 6.6565, "step": 18115 }, { "epoch": 0.7637919767270275, "grad_norm": 2.2673795223236084, "learning_rate": 4.141213029763578e-05, "loss": 7.2721, "step": 18116 }, { "epoch": 0.7638341379092269, "grad_norm": 0.9908216595649719, "learning_rate": 4.1411349511227706e-05, "loss": 6.4786, "step": 18117 }, { "epoch": 0.7638762990914265, "grad_norm": 3.749429941177368, "learning_rate": 4.1410568724819644e-05, "loss": 7.7241, "step": 18118 }, { "epoch": 0.7639184602736261, "grad_norm": 2.870145082473755, "learning_rate": 4.140978793841157e-05, "loss": 7.0875, "step": 18119 }, { "epoch": 0.7639606214558257, "grad_norm": 3.4936366081237793, "learning_rate": 4.14090071520035e-05, "loss": 7.7987, "step": 18120 }, { "epoch": 0.7640027826380251, "grad_norm": 5.33971643447876, "learning_rate": 4.140822636559543e-05, "loss": 7.9645, "step": 18121 }, { "epoch": 0.7640449438202247, "grad_norm": 1.541248083114624, "learning_rate": 4.140744557918736e-05, "loss": 6.913, "step": 18122 }, { "epoch": 0.7640871050024243, "grad_norm": 1.6627687215805054, "learning_rate": 4.140666479277929e-05, "loss": 6.4243, "step": 18123 }, { "epoch": 0.7641292661846238, "grad_norm": 2.9909074306488037, "learning_rate": 4.140588400637122e-05, "loss": 7.1205, "step": 18124 }, { "epoch": 0.7641714273668234, "grad_norm": 2.817349910736084, "learning_rate": 4.140510321996315e-05, "loss": 6.6123, "step": 18125 }, { "epoch": 0.7642135885490229, "grad_norm": 1.8481717109680176, "learning_rate": 4.1404322433555074e-05, "loss": 7.0354, "step": 18126 }, { "epoch": 0.7642557497312225, "grad_norm": 4.510000228881836, "learning_rate": 4.140354164714701e-05, "loss": 9.1519, "step": 18127 }, { "epoch": 0.764297910913422, "grad_norm": 1.3598978519439697, "learning_rate": 4.1402760860738936e-05, "loss": 6.414, "step": 18128 }, { "epoch": 0.7643400720956216, "grad_norm": 1.3973472118377686, "learning_rate": 4.140198007433087e-05, "loss": 6.6027, "step": 18129 }, { "epoch": 0.7643822332778211, "grad_norm": 1.182879090309143, "learning_rate": 4.14011992879228e-05, "loss": 6.7276, "step": 18130 }, { "epoch": 0.7644243944600206, "grad_norm": 2.7949743270874023, "learning_rate": 4.1400418501514723e-05, "loss": 7.0837, "step": 18131 }, { "epoch": 0.7644665556422202, "grad_norm": 1.553252935409546, "learning_rate": 4.139963771510666e-05, "loss": 6.4482, "step": 18132 }, { "epoch": 0.7645087168244198, "grad_norm": 2.5667531490325928, "learning_rate": 4.1398856928698586e-05, "loss": 7.3587, "step": 18133 }, { "epoch": 0.7645508780066194, "grad_norm": 1.6183892488479614, "learning_rate": 4.139807614229052e-05, "loss": 6.6318, "step": 18134 }, { "epoch": 0.7645930391888188, "grad_norm": 1.2358731031417847, "learning_rate": 4.139729535588245e-05, "loss": 6.817, "step": 18135 }, { "epoch": 0.7646352003710184, "grad_norm": 1.0261750221252441, "learning_rate": 4.139651456947437e-05, "loss": 6.4653, "step": 18136 }, { "epoch": 0.764677361553218, "grad_norm": 1.3195626735687256, "learning_rate": 4.1395733783066304e-05, "loss": 6.506, "step": 18137 }, { "epoch": 0.7647195227354175, "grad_norm": 1.2105516195297241, "learning_rate": 4.1394952996658235e-05, "loss": 6.3294, "step": 18138 }, { "epoch": 0.764761683917617, "grad_norm": 2.343496561050415, "learning_rate": 4.139417221025017e-05, "loss": 7.0, "step": 18139 }, { "epoch": 0.7648038450998166, "grad_norm": 1.5292656421661377, "learning_rate": 4.13933914238421e-05, "loss": 6.4491, "step": 18140 }, { "epoch": 0.7648460062820162, "grad_norm": 0.8965503573417664, "learning_rate": 4.139261063743403e-05, "loss": 6.6434, "step": 18141 }, { "epoch": 0.7648881674642157, "grad_norm": 2.042712450027466, "learning_rate": 4.1391829851025954e-05, "loss": 6.9604, "step": 18142 }, { "epoch": 0.7649303286464152, "grad_norm": 5.284828186035156, "learning_rate": 4.1391049064617885e-05, "loss": 8.5689, "step": 18143 }, { "epoch": 0.7649724898286148, "grad_norm": 2.479389190673828, "learning_rate": 4.1390268278209816e-05, "loss": 6.588, "step": 18144 }, { "epoch": 0.7650146510108143, "grad_norm": 1.782628059387207, "learning_rate": 4.138948749180174e-05, "loss": 6.8884, "step": 18145 }, { "epoch": 0.7650568121930139, "grad_norm": 1.2097032070159912, "learning_rate": 4.138870670539368e-05, "loss": 6.5912, "step": 18146 }, { "epoch": 0.7650989733752135, "grad_norm": 2.7101292610168457, "learning_rate": 4.13879259189856e-05, "loss": 7.9174, "step": 18147 }, { "epoch": 0.765141134557413, "grad_norm": 1.2254087924957275, "learning_rate": 4.1387145132577535e-05, "loss": 6.6239, "step": 18148 }, { "epoch": 0.7651832957396125, "grad_norm": 1.4054564237594604, "learning_rate": 4.1386364346169466e-05, "loss": 6.9624, "step": 18149 }, { "epoch": 0.7652254569218121, "grad_norm": 1.5793702602386475, "learning_rate": 4.138558355976139e-05, "loss": 6.9673, "step": 18150 }, { "epoch": 0.7652676181040117, "grad_norm": 1.4223707914352417, "learning_rate": 4.138480277335333e-05, "loss": 6.4646, "step": 18151 }, { "epoch": 0.7653097792862111, "grad_norm": 2.64178204536438, "learning_rate": 4.138402198694525e-05, "loss": 6.7548, "step": 18152 }, { "epoch": 0.7653519404684107, "grad_norm": 1.1387038230895996, "learning_rate": 4.1383241200537184e-05, "loss": 6.4965, "step": 18153 }, { "epoch": 0.7653941016506103, "grad_norm": 1.600762128829956, "learning_rate": 4.1382460414129115e-05, "loss": 6.4409, "step": 18154 }, { "epoch": 0.7654362628328099, "grad_norm": 2.3713176250457764, "learning_rate": 4.138167962772104e-05, "loss": 7.2114, "step": 18155 }, { "epoch": 0.7654784240150094, "grad_norm": 2.164029598236084, "learning_rate": 4.138089884131297e-05, "loss": 6.5308, "step": 18156 }, { "epoch": 0.7655205851972089, "grad_norm": 2.477937698364258, "learning_rate": 4.13801180549049e-05, "loss": 7.7669, "step": 18157 }, { "epoch": 0.7655627463794085, "grad_norm": 1.7666599750518799, "learning_rate": 4.1379337268496834e-05, "loss": 6.5574, "step": 18158 }, { "epoch": 0.765604907561608, "grad_norm": 1.8675652742385864, "learning_rate": 4.137855648208876e-05, "loss": 6.7937, "step": 18159 }, { "epoch": 0.7656470687438076, "grad_norm": 1.7150377035140991, "learning_rate": 4.1377775695680696e-05, "loss": 6.5144, "step": 18160 }, { "epoch": 0.7656892299260071, "grad_norm": 1.7271753549575806, "learning_rate": 4.137699490927262e-05, "loss": 6.6349, "step": 18161 }, { "epoch": 0.7657313911082066, "grad_norm": 3.668142795562744, "learning_rate": 4.137621412286455e-05, "loss": 7.3252, "step": 18162 }, { "epoch": 0.7657735522904062, "grad_norm": 3.0405917167663574, "learning_rate": 4.137543333645648e-05, "loss": 6.8183, "step": 18163 }, { "epoch": 0.7658157134726058, "grad_norm": 1.2565754652023315, "learning_rate": 4.137465255004841e-05, "loss": 6.6232, "step": 18164 }, { "epoch": 0.7658578746548054, "grad_norm": 1.0445129871368408, "learning_rate": 4.1373871763640346e-05, "loss": 6.4378, "step": 18165 }, { "epoch": 0.7659000358370048, "grad_norm": 1.758276104927063, "learning_rate": 4.137309097723227e-05, "loss": 7.191, "step": 18166 }, { "epoch": 0.7659421970192044, "grad_norm": 1.4191399812698364, "learning_rate": 4.13723101908242e-05, "loss": 6.7224, "step": 18167 }, { "epoch": 0.765984358201404, "grad_norm": 1.3361780643463135, "learning_rate": 4.137152940441613e-05, "loss": 6.5557, "step": 18168 }, { "epoch": 0.7660265193836036, "grad_norm": 3.373054265975952, "learning_rate": 4.137074861800806e-05, "loss": 7.7344, "step": 18169 }, { "epoch": 0.766068680565803, "grad_norm": 1.2110651731491089, "learning_rate": 4.136996783159999e-05, "loss": 6.6855, "step": 18170 }, { "epoch": 0.7661108417480026, "grad_norm": 1.6753907203674316, "learning_rate": 4.136918704519192e-05, "loss": 7.1056, "step": 18171 }, { "epoch": 0.7661530029302022, "grad_norm": 1.3923592567443848, "learning_rate": 4.136840625878385e-05, "loss": 7.1724, "step": 18172 }, { "epoch": 0.7661951641124017, "grad_norm": 1.4349862337112427, "learning_rate": 4.1367625472375775e-05, "loss": 7.1325, "step": 18173 }, { "epoch": 0.7662373252946013, "grad_norm": 2.4382741451263428, "learning_rate": 4.136684468596771e-05, "loss": 7.412, "step": 18174 }, { "epoch": 0.7662794864768008, "grad_norm": 2.038055419921875, "learning_rate": 4.136606389955964e-05, "loss": 6.5472, "step": 18175 }, { "epoch": 0.7663216476590003, "grad_norm": 2.8912858963012695, "learning_rate": 4.136528311315157e-05, "loss": 6.7363, "step": 18176 }, { "epoch": 0.7663638088411999, "grad_norm": 2.690446138381958, "learning_rate": 4.13645023267435e-05, "loss": 6.9853, "step": 18177 }, { "epoch": 0.7664059700233995, "grad_norm": 2.090938091278076, "learning_rate": 4.1363721540335425e-05, "loss": 6.5021, "step": 18178 }, { "epoch": 0.766448131205599, "grad_norm": 2.8435614109039307, "learning_rate": 4.1362940753927356e-05, "loss": 7.6778, "step": 18179 }, { "epoch": 0.7664902923877985, "grad_norm": 1.456408143043518, "learning_rate": 4.136215996751929e-05, "loss": 6.4899, "step": 18180 }, { "epoch": 0.7665324535699981, "grad_norm": 2.8268074989318848, "learning_rate": 4.136137918111121e-05, "loss": 6.7542, "step": 18181 }, { "epoch": 0.7665746147521977, "grad_norm": 5.259139537811279, "learning_rate": 4.136059839470315e-05, "loss": 7.5831, "step": 18182 }, { "epoch": 0.7666167759343973, "grad_norm": 4.778756618499756, "learning_rate": 4.1359817608295074e-05, "loss": 8.0439, "step": 18183 }, { "epoch": 0.7666589371165967, "grad_norm": 3.315631628036499, "learning_rate": 4.1359036821887006e-05, "loss": 6.3272, "step": 18184 }, { "epoch": 0.7667010982987963, "grad_norm": 1.6764248609542847, "learning_rate": 4.135825603547894e-05, "loss": 6.9378, "step": 18185 }, { "epoch": 0.7667432594809959, "grad_norm": 2.131523847579956, "learning_rate": 4.135747524907087e-05, "loss": 6.9217, "step": 18186 }, { "epoch": 0.7667854206631954, "grad_norm": 3.6821231842041016, "learning_rate": 4.13566944626628e-05, "loss": 6.59, "step": 18187 }, { "epoch": 0.7668275818453949, "grad_norm": 2.1276628971099854, "learning_rate": 4.1355913676254724e-05, "loss": 6.7906, "step": 18188 }, { "epoch": 0.7668697430275945, "grad_norm": 2.306759834289551, "learning_rate": 4.1355132889846655e-05, "loss": 7.0054, "step": 18189 }, { "epoch": 0.766911904209794, "grad_norm": 1.9740184545516968, "learning_rate": 4.1354352103438586e-05, "loss": 7.1265, "step": 18190 }, { "epoch": 0.7669540653919936, "grad_norm": 2.620802640914917, "learning_rate": 4.135357131703052e-05, "loss": 7.1354, "step": 18191 }, { "epoch": 0.7669962265741931, "grad_norm": 2.043433427810669, "learning_rate": 4.135279053062244e-05, "loss": 6.5406, "step": 18192 }, { "epoch": 0.7670383877563927, "grad_norm": 4.079569339752197, "learning_rate": 4.1352009744214373e-05, "loss": 7.0026, "step": 18193 }, { "epoch": 0.7670805489385922, "grad_norm": 2.8089327812194824, "learning_rate": 4.1351228957806305e-05, "loss": 7.1097, "step": 18194 }, { "epoch": 0.7671227101207918, "grad_norm": 2.878645658493042, "learning_rate": 4.135044817139823e-05, "loss": 6.2145, "step": 18195 }, { "epoch": 0.7671648713029914, "grad_norm": 1.6480183601379395, "learning_rate": 4.134966738499017e-05, "loss": 7.0863, "step": 18196 }, { "epoch": 0.7672070324851908, "grad_norm": 1.4715083837509155, "learning_rate": 4.134888659858209e-05, "loss": 6.7875, "step": 18197 }, { "epoch": 0.7672491936673904, "grad_norm": 3.2989518642425537, "learning_rate": 4.134810581217402e-05, "loss": 8.1052, "step": 18198 }, { "epoch": 0.76729135484959, "grad_norm": 3.049985408782959, "learning_rate": 4.1347325025765954e-05, "loss": 7.3576, "step": 18199 }, { "epoch": 0.7673335160317896, "grad_norm": 3.115116834640503, "learning_rate": 4.134654423935788e-05, "loss": 6.5236, "step": 18200 }, { "epoch": 0.767375677213989, "grad_norm": 3.199831485748291, "learning_rate": 4.134576345294982e-05, "loss": 7.1023, "step": 18201 }, { "epoch": 0.7674178383961886, "grad_norm": 2.848727226257324, "learning_rate": 4.134498266654174e-05, "loss": 6.2409, "step": 18202 }, { "epoch": 0.7674599995783882, "grad_norm": 1.438539743423462, "learning_rate": 4.134420188013367e-05, "loss": 6.5477, "step": 18203 }, { "epoch": 0.7675021607605877, "grad_norm": 3.655921459197998, "learning_rate": 4.1343421093725604e-05, "loss": 7.7463, "step": 18204 }, { "epoch": 0.7675443219427873, "grad_norm": 1.4640225172042847, "learning_rate": 4.1342640307317535e-05, "loss": 6.5487, "step": 18205 }, { "epoch": 0.7675864831249868, "grad_norm": 1.8119713068008423, "learning_rate": 4.134185952090946e-05, "loss": 7.1382, "step": 18206 }, { "epoch": 0.7676286443071864, "grad_norm": 2.1861720085144043, "learning_rate": 4.134107873450139e-05, "loss": 6.3656, "step": 18207 }, { "epoch": 0.7676708054893859, "grad_norm": 1.6459566354751587, "learning_rate": 4.134029794809332e-05, "loss": 7.0815, "step": 18208 }, { "epoch": 0.7677129666715855, "grad_norm": 2.491645336151123, "learning_rate": 4.133951716168525e-05, "loss": 7.3864, "step": 18209 }, { "epoch": 0.767755127853785, "grad_norm": 2.022404193878174, "learning_rate": 4.1338736375277185e-05, "loss": 6.4987, "step": 18210 }, { "epoch": 0.7677972890359845, "grad_norm": 1.4415316581726074, "learning_rate": 4.133795558886911e-05, "loss": 7.1246, "step": 18211 }, { "epoch": 0.7678394502181841, "grad_norm": 1.1878772974014282, "learning_rate": 4.133717480246104e-05, "loss": 6.6898, "step": 18212 }, { "epoch": 0.7678816114003837, "grad_norm": 1.5047420263290405, "learning_rate": 4.133639401605297e-05, "loss": 6.5557, "step": 18213 }, { "epoch": 0.7679237725825833, "grad_norm": 1.3762999773025513, "learning_rate": 4.1335613229644896e-05, "loss": 6.5661, "step": 18214 }, { "epoch": 0.7679659337647827, "grad_norm": 1.4511349201202393, "learning_rate": 4.1334832443236834e-05, "loss": 6.839, "step": 18215 }, { "epoch": 0.7680080949469823, "grad_norm": 1.9886759519577026, "learning_rate": 4.133405165682876e-05, "loss": 7.0007, "step": 18216 }, { "epoch": 0.7680502561291819, "grad_norm": 1.3262535333633423, "learning_rate": 4.133327087042069e-05, "loss": 6.6583, "step": 18217 }, { "epoch": 0.7680924173113814, "grad_norm": 2.66401743888855, "learning_rate": 4.133249008401262e-05, "loss": 7.3114, "step": 18218 }, { "epoch": 0.7681345784935809, "grad_norm": 2.796452522277832, "learning_rate": 4.1331709297604546e-05, "loss": 8.2877, "step": 18219 }, { "epoch": 0.7681767396757805, "grad_norm": 2.767345428466797, "learning_rate": 4.1330928511196484e-05, "loss": 8.1928, "step": 18220 }, { "epoch": 0.7682189008579801, "grad_norm": 2.9789795875549316, "learning_rate": 4.133014772478841e-05, "loss": 7.5916, "step": 18221 }, { "epoch": 0.7682610620401796, "grad_norm": 2.0764856338500977, "learning_rate": 4.132936693838034e-05, "loss": 6.4562, "step": 18222 }, { "epoch": 0.7683032232223792, "grad_norm": 2.1270875930786133, "learning_rate": 4.132858615197227e-05, "loss": 7.182, "step": 18223 }, { "epoch": 0.7683453844045787, "grad_norm": 2.691075563430786, "learning_rate": 4.1327805365564195e-05, "loss": 7.1802, "step": 18224 }, { "epoch": 0.7683875455867782, "grad_norm": 1.9660266637802124, "learning_rate": 4.1327024579156126e-05, "loss": 6.4178, "step": 18225 }, { "epoch": 0.7684297067689778, "grad_norm": 1.6079202890396118, "learning_rate": 4.132624379274806e-05, "loss": 6.9063, "step": 18226 }, { "epoch": 0.7684718679511774, "grad_norm": 1.6619398593902588, "learning_rate": 4.132546300633999e-05, "loss": 6.9036, "step": 18227 }, { "epoch": 0.7685140291333769, "grad_norm": 3.023271083831787, "learning_rate": 4.132468221993191e-05, "loss": 7.6087, "step": 18228 }, { "epoch": 0.7685561903155764, "grad_norm": 2.1486692428588867, "learning_rate": 4.132390143352385e-05, "loss": 6.8308, "step": 18229 }, { "epoch": 0.768598351497776, "grad_norm": 3.1567559242248535, "learning_rate": 4.1323120647115776e-05, "loss": 7.7363, "step": 18230 }, { "epoch": 0.7686405126799756, "grad_norm": 1.7892190217971802, "learning_rate": 4.132233986070771e-05, "loss": 7.2016, "step": 18231 }, { "epoch": 0.7686826738621751, "grad_norm": 1.5593342781066895, "learning_rate": 4.132155907429964e-05, "loss": 6.6867, "step": 18232 }, { "epoch": 0.7687248350443746, "grad_norm": 1.5985907316207886, "learning_rate": 4.132077828789156e-05, "loss": 6.901, "step": 18233 }, { "epoch": 0.7687669962265742, "grad_norm": 1.8351118564605713, "learning_rate": 4.13199975014835e-05, "loss": 6.733, "step": 18234 }, { "epoch": 0.7688091574087738, "grad_norm": 1.105616807937622, "learning_rate": 4.1319216715075425e-05, "loss": 6.4704, "step": 18235 }, { "epoch": 0.7688513185909733, "grad_norm": 1.320908784866333, "learning_rate": 4.131843592866736e-05, "loss": 6.8868, "step": 18236 }, { "epoch": 0.7688934797731728, "grad_norm": 2.461858034133911, "learning_rate": 4.131765514225929e-05, "loss": 6.9336, "step": 18237 }, { "epoch": 0.7689356409553724, "grad_norm": 4.165867328643799, "learning_rate": 4.131687435585121e-05, "loss": 7.4069, "step": 18238 }, { "epoch": 0.7689778021375719, "grad_norm": 2.7432639598846436, "learning_rate": 4.1316093569443144e-05, "loss": 7.6286, "step": 18239 }, { "epoch": 0.7690199633197715, "grad_norm": 1.4892752170562744, "learning_rate": 4.1315312783035075e-05, "loss": 6.5943, "step": 18240 }, { "epoch": 0.7690621245019711, "grad_norm": 2.936636209487915, "learning_rate": 4.1314531996627006e-05, "loss": 6.6968, "step": 18241 }, { "epoch": 0.7691042856841706, "grad_norm": 2.1084628105163574, "learning_rate": 4.131375121021894e-05, "loss": 6.5305, "step": 18242 }, { "epoch": 0.7691464468663701, "grad_norm": 2.8380775451660156, "learning_rate": 4.131297042381086e-05, "loss": 6.6262, "step": 18243 }, { "epoch": 0.7691886080485697, "grad_norm": 3.1516120433807373, "learning_rate": 4.131218963740279e-05, "loss": 6.4775, "step": 18244 }, { "epoch": 0.7692307692307693, "grad_norm": 2.417193651199341, "learning_rate": 4.1311408850994724e-05, "loss": 6.8043, "step": 18245 }, { "epoch": 0.7692729304129687, "grad_norm": 1.6995189189910889, "learning_rate": 4.1310628064586656e-05, "loss": 6.3251, "step": 18246 }, { "epoch": 0.7693150915951683, "grad_norm": 3.516958236694336, "learning_rate": 4.130984727817858e-05, "loss": 8.1943, "step": 18247 }, { "epoch": 0.7693572527773679, "grad_norm": 2.7016799449920654, "learning_rate": 4.130906649177052e-05, "loss": 7.2755, "step": 18248 }, { "epoch": 0.7693994139595675, "grad_norm": 4.930095672607422, "learning_rate": 4.130828570536244e-05, "loss": 8.0275, "step": 18249 }, { "epoch": 0.7694415751417669, "grad_norm": 2.4866340160369873, "learning_rate": 4.1307504918954374e-05, "loss": 6.9952, "step": 18250 }, { "epoch": 0.7694837363239665, "grad_norm": 1.864173412322998, "learning_rate": 4.1306724132546305e-05, "loss": 6.4927, "step": 18251 }, { "epoch": 0.7695258975061661, "grad_norm": 2.9509212970733643, "learning_rate": 4.130594334613823e-05, "loss": 7.6434, "step": 18252 }, { "epoch": 0.7695680586883656, "grad_norm": 2.348750352859497, "learning_rate": 4.130516255973016e-05, "loss": 6.581, "step": 18253 }, { "epoch": 0.7696102198705652, "grad_norm": 2.088670253753662, "learning_rate": 4.130438177332209e-05, "loss": 6.4314, "step": 18254 }, { "epoch": 0.7696523810527647, "grad_norm": 1.1813125610351562, "learning_rate": 4.1303600986914023e-05, "loss": 6.385, "step": 18255 }, { "epoch": 0.7696945422349643, "grad_norm": 1.8748981952667236, "learning_rate": 4.1302820200505955e-05, "loss": 6.9706, "step": 18256 }, { "epoch": 0.7697367034171638, "grad_norm": 1.2248073816299438, "learning_rate": 4.130203941409788e-05, "loss": 6.4727, "step": 18257 }, { "epoch": 0.7697788645993634, "grad_norm": 2.718182325363159, "learning_rate": 4.130125862768981e-05, "loss": 7.6469, "step": 18258 }, { "epoch": 0.7698210257815629, "grad_norm": 1.6870298385620117, "learning_rate": 4.130047784128174e-05, "loss": 6.4297, "step": 18259 }, { "epoch": 0.7698631869637624, "grad_norm": 1.8935835361480713, "learning_rate": 4.129969705487367e-05, "loss": 7.1292, "step": 18260 }, { "epoch": 0.769905348145962, "grad_norm": 2.286059856414795, "learning_rate": 4.12989162684656e-05, "loss": 7.1436, "step": 18261 }, { "epoch": 0.7699475093281616, "grad_norm": 1.4571361541748047, "learning_rate": 4.129813548205753e-05, "loss": 6.9675, "step": 18262 }, { "epoch": 0.7699896705103612, "grad_norm": 2.154925584793091, "learning_rate": 4.129735469564946e-05, "loss": 6.5103, "step": 18263 }, { "epoch": 0.7700318316925606, "grad_norm": 1.2221028804779053, "learning_rate": 4.1296573909241384e-05, "loss": 6.9846, "step": 18264 }, { "epoch": 0.7700739928747602, "grad_norm": 1.155838966369629, "learning_rate": 4.129579312283332e-05, "loss": 6.537, "step": 18265 }, { "epoch": 0.7701161540569598, "grad_norm": 2.4987752437591553, "learning_rate": 4.129501233642525e-05, "loss": 7.3509, "step": 18266 }, { "epoch": 0.7701583152391593, "grad_norm": 1.5661994218826294, "learning_rate": 4.1294231550017185e-05, "loss": 6.7272, "step": 18267 }, { "epoch": 0.7702004764213588, "grad_norm": 1.627873182296753, "learning_rate": 4.129345076360911e-05, "loss": 6.6662, "step": 18268 }, { "epoch": 0.7702426376035584, "grad_norm": 1.9665961265563965, "learning_rate": 4.1292669977201034e-05, "loss": 6.9659, "step": 18269 }, { "epoch": 0.770284798785758, "grad_norm": 5.288356304168701, "learning_rate": 4.129188919079297e-05, "loss": 8.9628, "step": 18270 }, { "epoch": 0.7703269599679575, "grad_norm": 2.9626388549804688, "learning_rate": 4.1291108404384897e-05, "loss": 6.7046, "step": 18271 }, { "epoch": 0.7703691211501571, "grad_norm": 1.5796923637390137, "learning_rate": 4.129032761797683e-05, "loss": 7.6957, "step": 18272 }, { "epoch": 0.7704112823323566, "grad_norm": 1.7717355489730835, "learning_rate": 4.128954683156876e-05, "loss": 6.4443, "step": 18273 }, { "epoch": 0.7704534435145561, "grad_norm": 2.360834836959839, "learning_rate": 4.128876604516069e-05, "loss": 7.1781, "step": 18274 }, { "epoch": 0.7704956046967557, "grad_norm": 3.882507085800171, "learning_rate": 4.1287985258752615e-05, "loss": 7.6619, "step": 18275 }, { "epoch": 0.7705377658789553, "grad_norm": 1.7497942447662354, "learning_rate": 4.1287204472344546e-05, "loss": 6.9192, "step": 18276 }, { "epoch": 0.7705799270611547, "grad_norm": 1.7517096996307373, "learning_rate": 4.128642368593648e-05, "loss": 6.7553, "step": 18277 }, { "epoch": 0.7706220882433543, "grad_norm": 1.6069645881652832, "learning_rate": 4.128564289952841e-05, "loss": 6.6652, "step": 18278 }, { "epoch": 0.7706642494255539, "grad_norm": 1.5533887147903442, "learning_rate": 4.128486211312034e-05, "loss": 7.0057, "step": 18279 }, { "epoch": 0.7707064106077535, "grad_norm": 1.346244215965271, "learning_rate": 4.1284081326712264e-05, "loss": 6.957, "step": 18280 }, { "epoch": 0.770748571789953, "grad_norm": 1.7811335325241089, "learning_rate": 4.1283300540304196e-05, "loss": 6.9413, "step": 18281 }, { "epoch": 0.7707907329721525, "grad_norm": 2.1161623001098633, "learning_rate": 4.128251975389613e-05, "loss": 7.0362, "step": 18282 }, { "epoch": 0.7708328941543521, "grad_norm": 1.236323356628418, "learning_rate": 4.128173896748805e-05, "loss": 6.9508, "step": 18283 }, { "epoch": 0.7708750553365517, "grad_norm": 1.2847074270248413, "learning_rate": 4.128095818107999e-05, "loss": 7.1197, "step": 18284 }, { "epoch": 0.7709172165187512, "grad_norm": 2.304211139678955, "learning_rate": 4.1280177394671914e-05, "loss": 6.7624, "step": 18285 }, { "epoch": 0.7709593777009507, "grad_norm": 2.7871859073638916, "learning_rate": 4.1279396608263845e-05, "loss": 7.7045, "step": 18286 }, { "epoch": 0.7710015388831503, "grad_norm": 1.766185998916626, "learning_rate": 4.1278615821855776e-05, "loss": 7.5242, "step": 18287 }, { "epoch": 0.7710437000653498, "grad_norm": 4.2961201667785645, "learning_rate": 4.12778350354477e-05, "loss": 8.6847, "step": 18288 }, { "epoch": 0.7710858612475494, "grad_norm": 2.6799423694610596, "learning_rate": 4.127705424903964e-05, "loss": 7.4069, "step": 18289 }, { "epoch": 0.771128022429749, "grad_norm": 2.1324896812438965, "learning_rate": 4.127627346263156e-05, "loss": 6.6019, "step": 18290 }, { "epoch": 0.7711701836119484, "grad_norm": 3.146329164505005, "learning_rate": 4.1275492676223495e-05, "loss": 7.0063, "step": 18291 }, { "epoch": 0.771212344794148, "grad_norm": 2.3399317264556885, "learning_rate": 4.1274711889815426e-05, "loss": 7.4923, "step": 18292 }, { "epoch": 0.7712545059763476, "grad_norm": 2.2229673862457275, "learning_rate": 4.127393110340736e-05, "loss": 7.5083, "step": 18293 }, { "epoch": 0.7712966671585472, "grad_norm": 1.592337727546692, "learning_rate": 4.127315031699928e-05, "loss": 6.6133, "step": 18294 }, { "epoch": 0.7713388283407466, "grad_norm": 1.6083731651306152, "learning_rate": 4.127236953059121e-05, "loss": 6.7379, "step": 18295 }, { "epoch": 0.7713809895229462, "grad_norm": 1.7355453968048096, "learning_rate": 4.1271588744183144e-05, "loss": 6.7856, "step": 18296 }, { "epoch": 0.7714231507051458, "grad_norm": 2.9756510257720947, "learning_rate": 4.127080795777507e-05, "loss": 8.3549, "step": 18297 }, { "epoch": 0.7714653118873454, "grad_norm": 1.268009901046753, "learning_rate": 4.127002717136701e-05, "loss": 6.6515, "step": 18298 }, { "epoch": 0.7715074730695449, "grad_norm": 3.0052454471588135, "learning_rate": 4.126924638495893e-05, "loss": 7.6451, "step": 18299 }, { "epoch": 0.7715496342517444, "grad_norm": 3.738297700881958, "learning_rate": 4.126846559855086e-05, "loss": 7.4799, "step": 18300 }, { "epoch": 0.771591795433944, "grad_norm": 2.953636407852173, "learning_rate": 4.1267684812142794e-05, "loss": 6.4555, "step": 18301 }, { "epoch": 0.7716339566161435, "grad_norm": 2.7866291999816895, "learning_rate": 4.126690402573472e-05, "loss": 6.7754, "step": 18302 }, { "epoch": 0.7716761177983431, "grad_norm": 3.1327714920043945, "learning_rate": 4.1266123239326656e-05, "loss": 7.2098, "step": 18303 }, { "epoch": 0.7717182789805426, "grad_norm": 1.5568122863769531, "learning_rate": 4.126534245291858e-05, "loss": 7.3504, "step": 18304 }, { "epoch": 0.7717604401627421, "grad_norm": 2.296029567718506, "learning_rate": 4.126456166651051e-05, "loss": 7.2915, "step": 18305 }, { "epoch": 0.7718026013449417, "grad_norm": 1.4984686374664307, "learning_rate": 4.126378088010244e-05, "loss": 7.0391, "step": 18306 }, { "epoch": 0.7718447625271413, "grad_norm": 1.5911914110183716, "learning_rate": 4.126300009369437e-05, "loss": 6.9876, "step": 18307 }, { "epoch": 0.7718869237093408, "grad_norm": 1.777967929840088, "learning_rate": 4.12622193072863e-05, "loss": 7.0963, "step": 18308 }, { "epoch": 0.7719290848915403, "grad_norm": 2.5105435848236084, "learning_rate": 4.126143852087823e-05, "loss": 6.7831, "step": 18309 }, { "epoch": 0.7719712460737399, "grad_norm": 1.7151321172714233, "learning_rate": 4.126065773447016e-05, "loss": 8.1752, "step": 18310 }, { "epoch": 0.7720134072559395, "grad_norm": 2.1158530712127686, "learning_rate": 4.125987694806209e-05, "loss": 6.5223, "step": 18311 }, { "epoch": 0.772055568438139, "grad_norm": 2.0589962005615234, "learning_rate": 4.125909616165402e-05, "loss": 7.1642, "step": 18312 }, { "epoch": 0.7720977296203385, "grad_norm": 2.518664598464966, "learning_rate": 4.125831537524595e-05, "loss": 8.2623, "step": 18313 }, { "epoch": 0.7721398908025381, "grad_norm": 1.2166732549667358, "learning_rate": 4.125753458883788e-05, "loss": 7.3008, "step": 18314 }, { "epoch": 0.7721820519847377, "grad_norm": 1.7817708253860474, "learning_rate": 4.125675380242981e-05, "loss": 7.05, "step": 18315 }, { "epoch": 0.7722242131669372, "grad_norm": 2.845076084136963, "learning_rate": 4.1255973016021735e-05, "loss": 7.8379, "step": 18316 }, { "epoch": 0.7722663743491367, "grad_norm": 1.2872395515441895, "learning_rate": 4.1255192229613673e-05, "loss": 6.4619, "step": 18317 }, { "epoch": 0.7723085355313363, "grad_norm": 1.1239813566207886, "learning_rate": 4.12544114432056e-05, "loss": 6.9219, "step": 18318 }, { "epoch": 0.7723506967135358, "grad_norm": 1.8076282739639282, "learning_rate": 4.125363065679753e-05, "loss": 6.9672, "step": 18319 }, { "epoch": 0.7723928578957354, "grad_norm": 1.5284233093261719, "learning_rate": 4.125284987038946e-05, "loss": 6.7538, "step": 18320 }, { "epoch": 0.772435019077935, "grad_norm": 1.7970064878463745, "learning_rate": 4.1252069083981385e-05, "loss": 6.9604, "step": 18321 }, { "epoch": 0.7724771802601345, "grad_norm": 1.0032687187194824, "learning_rate": 4.125128829757332e-05, "loss": 6.876, "step": 18322 }, { "epoch": 0.772519341442334, "grad_norm": 2.0002496242523193, "learning_rate": 4.125050751116525e-05, "loss": 6.3641, "step": 18323 }, { "epoch": 0.7725615026245336, "grad_norm": 1.2643418312072754, "learning_rate": 4.124972672475718e-05, "loss": 6.5458, "step": 18324 }, { "epoch": 0.7726036638067332, "grad_norm": 2.2083961963653564, "learning_rate": 4.124894593834911e-05, "loss": 7.3727, "step": 18325 }, { "epoch": 0.7726458249889326, "grad_norm": 2.216115951538086, "learning_rate": 4.1248165151941035e-05, "loss": 6.5286, "step": 18326 }, { "epoch": 0.7726879861711322, "grad_norm": 1.3346723318099976, "learning_rate": 4.1247384365532966e-05, "loss": 6.6884, "step": 18327 }, { "epoch": 0.7727301473533318, "grad_norm": 1.55050790309906, "learning_rate": 4.12466035791249e-05, "loss": 6.8359, "step": 18328 }, { "epoch": 0.7727723085355314, "grad_norm": 1.877524971961975, "learning_rate": 4.124582279271683e-05, "loss": 6.6016, "step": 18329 }, { "epoch": 0.7728144697177309, "grad_norm": 2.6296145915985107, "learning_rate": 4.124504200630875e-05, "loss": 6.513, "step": 18330 }, { "epoch": 0.7728566308999304, "grad_norm": 1.2946680784225464, "learning_rate": 4.1244261219900684e-05, "loss": 7.0499, "step": 18331 }, { "epoch": 0.77289879208213, "grad_norm": 1.2133092880249023, "learning_rate": 4.1243480433492615e-05, "loss": 6.7534, "step": 18332 }, { "epoch": 0.7729409532643295, "grad_norm": 4.230114459991455, "learning_rate": 4.124269964708454e-05, "loss": 9.4971, "step": 18333 }, { "epoch": 0.7729831144465291, "grad_norm": 1.8149088621139526, "learning_rate": 4.124191886067648e-05, "loss": 7.1142, "step": 18334 }, { "epoch": 0.7730252756287286, "grad_norm": 2.5272133350372314, "learning_rate": 4.12411380742684e-05, "loss": 7.2726, "step": 18335 }, { "epoch": 0.7730674368109282, "grad_norm": 4.025700092315674, "learning_rate": 4.124035728786034e-05, "loss": 9.0504, "step": 18336 }, { "epoch": 0.7731095979931277, "grad_norm": 2.5239923000335693, "learning_rate": 4.1239576501452265e-05, "loss": 6.52, "step": 18337 }, { "epoch": 0.7731517591753273, "grad_norm": 2.9809844493865967, "learning_rate": 4.1238795715044196e-05, "loss": 6.4209, "step": 18338 }, { "epoch": 0.7731939203575269, "grad_norm": 3.2485549449920654, "learning_rate": 4.123801492863613e-05, "loss": 6.9644, "step": 18339 }, { "epoch": 0.7732360815397263, "grad_norm": 2.4617762565612793, "learning_rate": 4.123723414222805e-05, "loss": 6.493, "step": 18340 }, { "epoch": 0.7732782427219259, "grad_norm": 1.1504216194152832, "learning_rate": 4.123645335581998e-05, "loss": 6.7395, "step": 18341 }, { "epoch": 0.7733204039041255, "grad_norm": 1.7412954568862915, "learning_rate": 4.1235672569411914e-05, "loss": 6.7963, "step": 18342 }, { "epoch": 0.7733625650863251, "grad_norm": 1.7541496753692627, "learning_rate": 4.1234891783003846e-05, "loss": 6.8226, "step": 18343 }, { "epoch": 0.7734047262685245, "grad_norm": 4.094711780548096, "learning_rate": 4.123411099659577e-05, "loss": 7.8135, "step": 18344 }, { "epoch": 0.7734468874507241, "grad_norm": 4.2022576332092285, "learning_rate": 4.12333302101877e-05, "loss": 6.4177, "step": 18345 }, { "epoch": 0.7734890486329237, "grad_norm": 3.247990846633911, "learning_rate": 4.123254942377963e-05, "loss": 6.5371, "step": 18346 }, { "epoch": 0.7735312098151232, "grad_norm": 1.639345645904541, "learning_rate": 4.1231768637371564e-05, "loss": 6.9373, "step": 18347 }, { "epoch": 0.7735733709973228, "grad_norm": 2.729754686355591, "learning_rate": 4.1230987850963495e-05, "loss": 6.9031, "step": 18348 }, { "epoch": 0.7736155321795223, "grad_norm": 2.0758860111236572, "learning_rate": 4.123020706455542e-05, "loss": 7.0224, "step": 18349 }, { "epoch": 0.7736576933617219, "grad_norm": 2.5441081523895264, "learning_rate": 4.122942627814735e-05, "loss": 6.732, "step": 18350 }, { "epoch": 0.7736998545439214, "grad_norm": 1.896592378616333, "learning_rate": 4.122864549173928e-05, "loss": 6.9041, "step": 18351 }, { "epoch": 0.773742015726121, "grad_norm": 1.5904978513717651, "learning_rate": 4.1227864705331207e-05, "loss": 6.7987, "step": 18352 }, { "epoch": 0.7737841769083205, "grad_norm": 1.6088793277740479, "learning_rate": 4.1227083918923145e-05, "loss": 6.4225, "step": 18353 }, { "epoch": 0.77382633809052, "grad_norm": 2.0162715911865234, "learning_rate": 4.122630313251507e-05, "loss": 6.4075, "step": 18354 }, { "epoch": 0.7738684992727196, "grad_norm": 2.044271230697632, "learning_rate": 4.1225522346107e-05, "loss": 7.1295, "step": 18355 }, { "epoch": 0.7739106604549192, "grad_norm": 4.04874324798584, "learning_rate": 4.122474155969893e-05, "loss": 6.8962, "step": 18356 }, { "epoch": 0.7739528216371188, "grad_norm": 1.1142878532409668, "learning_rate": 4.1223960773290856e-05, "loss": 6.6963, "step": 18357 }, { "epoch": 0.7739949828193182, "grad_norm": 1.2950469255447388, "learning_rate": 4.1223179986882794e-05, "loss": 6.9477, "step": 18358 }, { "epoch": 0.7740371440015178, "grad_norm": 1.902736783027649, "learning_rate": 4.122239920047472e-05, "loss": 6.7566, "step": 18359 }, { "epoch": 0.7740793051837174, "grad_norm": 1.3685283660888672, "learning_rate": 4.122161841406665e-05, "loss": 6.6621, "step": 18360 }, { "epoch": 0.7741214663659169, "grad_norm": 2.0982601642608643, "learning_rate": 4.122083762765858e-05, "loss": 6.6532, "step": 18361 }, { "epoch": 0.7741636275481164, "grad_norm": 1.7351306676864624, "learning_rate": 4.122005684125051e-05, "loss": 6.5574, "step": 18362 }, { "epoch": 0.774205788730316, "grad_norm": 1.3407185077667236, "learning_rate": 4.121927605484244e-05, "loss": 6.476, "step": 18363 }, { "epoch": 0.7742479499125156, "grad_norm": 2.958909034729004, "learning_rate": 4.121849526843437e-05, "loss": 6.9194, "step": 18364 }, { "epoch": 0.7742901110947151, "grad_norm": 3.348416805267334, "learning_rate": 4.12177144820263e-05, "loss": 7.522, "step": 18365 }, { "epoch": 0.7743322722769146, "grad_norm": 4.193795680999756, "learning_rate": 4.1216933695618224e-05, "loss": 7.7252, "step": 18366 }, { "epoch": 0.7743744334591142, "grad_norm": 1.7523058652877808, "learning_rate": 4.121615290921016e-05, "loss": 7.3267, "step": 18367 }, { "epoch": 0.7744165946413137, "grad_norm": 2.24200439453125, "learning_rate": 4.1215372122802086e-05, "loss": 7.019, "step": 18368 }, { "epoch": 0.7744587558235133, "grad_norm": 1.692135214805603, "learning_rate": 4.121459133639402e-05, "loss": 6.5507, "step": 18369 }, { "epoch": 0.7745009170057129, "grad_norm": 2.8684298992156982, "learning_rate": 4.121381054998595e-05, "loss": 7.5105, "step": 18370 }, { "epoch": 0.7745430781879123, "grad_norm": 3.4101953506469727, "learning_rate": 4.1213029763577873e-05, "loss": 7.6612, "step": 18371 }, { "epoch": 0.7745852393701119, "grad_norm": 1.6786504983901978, "learning_rate": 4.121224897716981e-05, "loss": 6.5249, "step": 18372 }, { "epoch": 0.7746274005523115, "grad_norm": 1.924653172492981, "learning_rate": 4.1211468190761736e-05, "loss": 6.4511, "step": 18373 }, { "epoch": 0.7746695617345111, "grad_norm": 2.423305034637451, "learning_rate": 4.121068740435367e-05, "loss": 6.4148, "step": 18374 }, { "epoch": 0.7747117229167105, "grad_norm": 1.0837455987930298, "learning_rate": 4.12099066179456e-05, "loss": 6.4976, "step": 18375 }, { "epoch": 0.7747538840989101, "grad_norm": 3.5521721839904785, "learning_rate": 4.120912583153752e-05, "loss": 7.9908, "step": 18376 }, { "epoch": 0.7747960452811097, "grad_norm": 1.4108994007110596, "learning_rate": 4.1208345045129454e-05, "loss": 6.9485, "step": 18377 }, { "epoch": 0.7748382064633093, "grad_norm": 2.2359530925750732, "learning_rate": 4.1207564258721385e-05, "loss": 6.9137, "step": 18378 }, { "epoch": 0.7748803676455088, "grad_norm": 1.2610971927642822, "learning_rate": 4.120678347231332e-05, "loss": 6.5257, "step": 18379 }, { "epoch": 0.7749225288277083, "grad_norm": 1.440479040145874, "learning_rate": 4.120600268590525e-05, "loss": 6.3406, "step": 18380 }, { "epoch": 0.7749646900099079, "grad_norm": 2.1910312175750732, "learning_rate": 4.120522189949718e-05, "loss": 7.07, "step": 18381 }, { "epoch": 0.7750068511921074, "grad_norm": 1.6902896165847778, "learning_rate": 4.1204441113089104e-05, "loss": 6.7762, "step": 18382 }, { "epoch": 0.775049012374307, "grad_norm": 2.6918818950653076, "learning_rate": 4.1203660326681035e-05, "loss": 7.997, "step": 18383 }, { "epoch": 0.7750911735565065, "grad_norm": 1.92423415184021, "learning_rate": 4.1202879540272966e-05, "loss": 7.2101, "step": 18384 }, { "epoch": 0.775133334738706, "grad_norm": 1.7181134223937988, "learning_rate": 4.120209875386489e-05, "loss": 6.9019, "step": 18385 }, { "epoch": 0.7751754959209056, "grad_norm": 2.080779552459717, "learning_rate": 4.120131796745683e-05, "loss": 7.0886, "step": 18386 }, { "epoch": 0.7752176571031052, "grad_norm": 2.8212318420410156, "learning_rate": 4.120053718104875e-05, "loss": 7.0693, "step": 18387 }, { "epoch": 0.7752598182853048, "grad_norm": 1.0934754610061646, "learning_rate": 4.1199756394640685e-05, "loss": 6.4094, "step": 18388 }, { "epoch": 0.7753019794675042, "grad_norm": 3.089646100997925, "learning_rate": 4.1198975608232616e-05, "loss": 7.0645, "step": 18389 }, { "epoch": 0.7753441406497038, "grad_norm": 1.5587327480316162, "learning_rate": 4.119819482182454e-05, "loss": 6.8111, "step": 18390 }, { "epoch": 0.7753863018319034, "grad_norm": 1.2944632768630981, "learning_rate": 4.119741403541648e-05, "loss": 6.7281, "step": 18391 }, { "epoch": 0.775428463014103, "grad_norm": 1.6657990217208862, "learning_rate": 4.11966332490084e-05, "loss": 7.0635, "step": 18392 }, { "epoch": 0.7754706241963024, "grad_norm": 1.6737967729568481, "learning_rate": 4.1195852462600334e-05, "loss": 6.4248, "step": 18393 }, { "epoch": 0.775512785378502, "grad_norm": 3.7967135906219482, "learning_rate": 4.1195071676192265e-05, "loss": 7.0384, "step": 18394 }, { "epoch": 0.7755549465607016, "grad_norm": 2.0030124187469482, "learning_rate": 4.119429088978419e-05, "loss": 6.6533, "step": 18395 }, { "epoch": 0.7755971077429011, "grad_norm": 1.7265112400054932, "learning_rate": 4.119351010337612e-05, "loss": 7.0767, "step": 18396 }, { "epoch": 0.7756392689251007, "grad_norm": 1.6819283962249756, "learning_rate": 4.119272931696805e-05, "loss": 6.8967, "step": 18397 }, { "epoch": 0.7756814301073002, "grad_norm": 2.7753636837005615, "learning_rate": 4.1191948530559984e-05, "loss": 6.6258, "step": 18398 }, { "epoch": 0.7757235912894997, "grad_norm": 1.8432302474975586, "learning_rate": 4.119116774415191e-05, "loss": 6.6523, "step": 18399 }, { "epoch": 0.7757657524716993, "grad_norm": 1.6480557918548584, "learning_rate": 4.1190386957743846e-05, "loss": 6.4362, "step": 18400 }, { "epoch": 0.7758079136538989, "grad_norm": 2.5213661193847656, "learning_rate": 4.118960617133577e-05, "loss": 7.1998, "step": 18401 }, { "epoch": 0.7758500748360984, "grad_norm": 1.888442039489746, "learning_rate": 4.1188825384927695e-05, "loss": 6.8872, "step": 18402 }, { "epoch": 0.7758922360182979, "grad_norm": 1.3392947912216187, "learning_rate": 4.118804459851963e-05, "loss": 6.8899, "step": 18403 }, { "epoch": 0.7759343972004975, "grad_norm": 1.6804858446121216, "learning_rate": 4.118726381211156e-05, "loss": 6.54, "step": 18404 }, { "epoch": 0.7759765583826971, "grad_norm": 1.1415542364120483, "learning_rate": 4.1186483025703496e-05, "loss": 6.606, "step": 18405 }, { "epoch": 0.7760187195648967, "grad_norm": 3.71968150138855, "learning_rate": 4.118570223929542e-05, "loss": 9.1426, "step": 18406 }, { "epoch": 0.7760608807470961, "grad_norm": 2.5338051319122314, "learning_rate": 4.118492145288735e-05, "loss": 7.1007, "step": 18407 }, { "epoch": 0.7761030419292957, "grad_norm": 3.919590711593628, "learning_rate": 4.118414066647928e-05, "loss": 8.1298, "step": 18408 }, { "epoch": 0.7761452031114953, "grad_norm": 3.285747528076172, "learning_rate": 4.118335988007121e-05, "loss": 7.5529, "step": 18409 }, { "epoch": 0.7761873642936948, "grad_norm": 1.799118161201477, "learning_rate": 4.118257909366314e-05, "loss": 6.7758, "step": 18410 }, { "epoch": 0.7762295254758943, "grad_norm": 2.720005989074707, "learning_rate": 4.118179830725507e-05, "loss": 8.6746, "step": 18411 }, { "epoch": 0.7762716866580939, "grad_norm": 2.33565354347229, "learning_rate": 4.1181017520847e-05, "loss": 6.9539, "step": 18412 }, { "epoch": 0.7763138478402934, "grad_norm": 2.976881265640259, "learning_rate": 4.1180236734438925e-05, "loss": 6.8752, "step": 18413 }, { "epoch": 0.776356009022493, "grad_norm": 2.3328962326049805, "learning_rate": 4.117945594803086e-05, "loss": 6.3892, "step": 18414 }, { "epoch": 0.7763981702046926, "grad_norm": 1.7262394428253174, "learning_rate": 4.117867516162279e-05, "loss": 6.8587, "step": 18415 }, { "epoch": 0.7764403313868921, "grad_norm": 2.692568778991699, "learning_rate": 4.117789437521472e-05, "loss": 6.6092, "step": 18416 }, { "epoch": 0.7764824925690916, "grad_norm": 3.7360434532165527, "learning_rate": 4.117711358880665e-05, "loss": 6.416, "step": 18417 }, { "epoch": 0.7765246537512912, "grad_norm": 2.162147045135498, "learning_rate": 4.1176332802398575e-05, "loss": 6.3796, "step": 18418 }, { "epoch": 0.7765668149334908, "grad_norm": 1.3985254764556885, "learning_rate": 4.1175552015990506e-05, "loss": 6.5401, "step": 18419 }, { "epoch": 0.7766089761156902, "grad_norm": 1.5176653861999512, "learning_rate": 4.117477122958244e-05, "loss": 6.4308, "step": 18420 }, { "epoch": 0.7766511372978898, "grad_norm": 0.8754496574401855, "learning_rate": 4.117399044317436e-05, "loss": 6.3902, "step": 18421 }, { "epoch": 0.7766932984800894, "grad_norm": 1.355128526687622, "learning_rate": 4.11732096567663e-05, "loss": 6.8623, "step": 18422 }, { "epoch": 0.776735459662289, "grad_norm": 1.9411003589630127, "learning_rate": 4.1172428870358224e-05, "loss": 6.5133, "step": 18423 }, { "epoch": 0.7767776208444884, "grad_norm": 0.9288472533226013, "learning_rate": 4.1171648083950156e-05, "loss": 6.7587, "step": 18424 }, { "epoch": 0.776819782026688, "grad_norm": 3.6738643646240234, "learning_rate": 4.117086729754209e-05, "loss": 9.2601, "step": 18425 }, { "epoch": 0.7768619432088876, "grad_norm": 3.0258660316467285, "learning_rate": 4.117008651113402e-05, "loss": 7.308, "step": 18426 }, { "epoch": 0.7769041043910871, "grad_norm": 3.9066131114959717, "learning_rate": 4.116930572472595e-05, "loss": 8.0728, "step": 18427 }, { "epoch": 0.7769462655732867, "grad_norm": 1.3072006702423096, "learning_rate": 4.1168524938317874e-05, "loss": 7.3563, "step": 18428 }, { "epoch": 0.7769884267554862, "grad_norm": 2.402271270751953, "learning_rate": 4.1167744151909805e-05, "loss": 6.7739, "step": 18429 }, { "epoch": 0.7770305879376858, "grad_norm": 2.869039535522461, "learning_rate": 4.1166963365501736e-05, "loss": 7.8129, "step": 18430 }, { "epoch": 0.7770727491198853, "grad_norm": 2.4587297439575195, "learning_rate": 4.116618257909367e-05, "loss": 6.5548, "step": 18431 }, { "epoch": 0.7771149103020849, "grad_norm": 1.8293815851211548, "learning_rate": 4.116540179268559e-05, "loss": 6.5512, "step": 18432 }, { "epoch": 0.7771570714842844, "grad_norm": 4.315263748168945, "learning_rate": 4.1164621006277523e-05, "loss": 7.2309, "step": 18433 }, { "epoch": 0.7771992326664839, "grad_norm": 1.8728139400482178, "learning_rate": 4.1163840219869455e-05, "loss": 6.9193, "step": 18434 }, { "epoch": 0.7772413938486835, "grad_norm": 2.3973584175109863, "learning_rate": 4.116305943346138e-05, "loss": 6.498, "step": 18435 }, { "epoch": 0.7772835550308831, "grad_norm": 1.8394489288330078, "learning_rate": 4.116227864705332e-05, "loss": 6.4989, "step": 18436 }, { "epoch": 0.7773257162130827, "grad_norm": 1.304570198059082, "learning_rate": 4.116149786064524e-05, "loss": 6.8579, "step": 18437 }, { "epoch": 0.7773678773952821, "grad_norm": 1.4808818101882935, "learning_rate": 4.116071707423717e-05, "loss": 6.4747, "step": 18438 }, { "epoch": 0.7774100385774817, "grad_norm": 3.019840955734253, "learning_rate": 4.1159936287829104e-05, "loss": 7.3482, "step": 18439 }, { "epoch": 0.7774521997596813, "grad_norm": 1.2595313787460327, "learning_rate": 4.115915550142103e-05, "loss": 6.6425, "step": 18440 }, { "epoch": 0.7774943609418808, "grad_norm": 1.1407501697540283, "learning_rate": 4.115837471501297e-05, "loss": 6.5817, "step": 18441 }, { "epoch": 0.7775365221240803, "grad_norm": 1.509445309638977, "learning_rate": 4.115759392860489e-05, "loss": 6.8367, "step": 18442 }, { "epoch": 0.7775786833062799, "grad_norm": 2.1261520385742188, "learning_rate": 4.115681314219682e-05, "loss": 6.8031, "step": 18443 }, { "epoch": 0.7776208444884795, "grad_norm": 3.3480162620544434, "learning_rate": 4.1156032355788754e-05, "loss": 8.1719, "step": 18444 }, { "epoch": 0.777663005670679, "grad_norm": 1.5395433902740479, "learning_rate": 4.1155251569380685e-05, "loss": 6.4488, "step": 18445 }, { "epoch": 0.7777051668528786, "grad_norm": 1.0254194736480713, "learning_rate": 4.115447078297261e-05, "loss": 6.5973, "step": 18446 }, { "epoch": 0.7777473280350781, "grad_norm": 1.355503797531128, "learning_rate": 4.115368999656454e-05, "loss": 6.8929, "step": 18447 }, { "epoch": 0.7777894892172776, "grad_norm": 1.6022932529449463, "learning_rate": 4.115290921015647e-05, "loss": 7.853, "step": 18448 }, { "epoch": 0.7778316503994772, "grad_norm": 2.679163932800293, "learning_rate": 4.11521284237484e-05, "loss": 7.6786, "step": 18449 }, { "epoch": 0.7778738115816768, "grad_norm": 3.9036922454833984, "learning_rate": 4.1151347637340335e-05, "loss": 7.3039, "step": 18450 }, { "epoch": 0.7779159727638763, "grad_norm": 2.3809640407562256, "learning_rate": 4.115056685093226e-05, "loss": 7.1793, "step": 18451 }, { "epoch": 0.7779581339460758, "grad_norm": 3.3170886039733887, "learning_rate": 4.114978606452419e-05, "loss": 6.7669, "step": 18452 }, { "epoch": 0.7780002951282754, "grad_norm": 1.735687017440796, "learning_rate": 4.114900527811612e-05, "loss": 6.6949, "step": 18453 }, { "epoch": 0.778042456310475, "grad_norm": 1.5672322511672974, "learning_rate": 4.1148224491708046e-05, "loss": 6.4254, "step": 18454 }, { "epoch": 0.7780846174926745, "grad_norm": 4.5983500480651855, "learning_rate": 4.1147443705299984e-05, "loss": 7.9376, "step": 18455 }, { "epoch": 0.778126778674874, "grad_norm": 2.745344877243042, "learning_rate": 4.114666291889191e-05, "loss": 6.5754, "step": 18456 }, { "epoch": 0.7781689398570736, "grad_norm": 2.932720899581909, "learning_rate": 4.114588213248384e-05, "loss": 6.4138, "step": 18457 }, { "epoch": 0.7782111010392732, "grad_norm": 2.327991008758545, "learning_rate": 4.114510134607577e-05, "loss": 6.4991, "step": 18458 }, { "epoch": 0.7782532622214727, "grad_norm": 2.21440052986145, "learning_rate": 4.1144320559667696e-05, "loss": 7.0553, "step": 18459 }, { "epoch": 0.7782954234036722, "grad_norm": 2.1071553230285645, "learning_rate": 4.1143539773259634e-05, "loss": 6.8489, "step": 18460 }, { "epoch": 0.7783375845858718, "grad_norm": 2.3015379905700684, "learning_rate": 4.114275898685156e-05, "loss": 7.2941, "step": 18461 }, { "epoch": 0.7783797457680713, "grad_norm": 3.2097136974334717, "learning_rate": 4.114197820044349e-05, "loss": 6.6845, "step": 18462 }, { "epoch": 0.7784219069502709, "grad_norm": 2.9841294288635254, "learning_rate": 4.114119741403542e-05, "loss": 6.542, "step": 18463 }, { "epoch": 0.7784640681324705, "grad_norm": 3.0875067710876465, "learning_rate": 4.1140416627627345e-05, "loss": 8.6313, "step": 18464 }, { "epoch": 0.77850622931467, "grad_norm": 2.6953346729278564, "learning_rate": 4.1139635841219276e-05, "loss": 7.6142, "step": 18465 }, { "epoch": 0.7785483904968695, "grad_norm": 1.1514606475830078, "learning_rate": 4.113885505481121e-05, "loss": 6.3867, "step": 18466 }, { "epoch": 0.7785905516790691, "grad_norm": 2.3640835285186768, "learning_rate": 4.113807426840314e-05, "loss": 6.8708, "step": 18467 }, { "epoch": 0.7786327128612687, "grad_norm": 1.790963888168335, "learning_rate": 4.113729348199506e-05, "loss": 6.4318, "step": 18468 }, { "epoch": 0.7786748740434681, "grad_norm": 1.5973687171936035, "learning_rate": 4.1136512695587e-05, "loss": 6.7061, "step": 18469 }, { "epoch": 0.7787170352256677, "grad_norm": 3.5817413330078125, "learning_rate": 4.1135731909178926e-05, "loss": 7.5916, "step": 18470 }, { "epoch": 0.7787591964078673, "grad_norm": 1.0034403800964355, "learning_rate": 4.113495112277086e-05, "loss": 6.5119, "step": 18471 }, { "epoch": 0.7788013575900669, "grad_norm": 1.8895107507705688, "learning_rate": 4.113417033636279e-05, "loss": 6.5285, "step": 18472 }, { "epoch": 0.7788435187722663, "grad_norm": 2.008904218673706, "learning_rate": 4.113338954995471e-05, "loss": 6.7772, "step": 18473 }, { "epoch": 0.7788856799544659, "grad_norm": 1.265987753868103, "learning_rate": 4.113260876354665e-05, "loss": 6.404, "step": 18474 }, { "epoch": 0.7789278411366655, "grad_norm": 1.1843748092651367, "learning_rate": 4.1131827977138575e-05, "loss": 6.3168, "step": 18475 }, { "epoch": 0.778970002318865, "grad_norm": 1.2984024286270142, "learning_rate": 4.113104719073051e-05, "loss": 6.4989, "step": 18476 }, { "epoch": 0.7790121635010646, "grad_norm": 1.2995843887329102, "learning_rate": 4.113026640432244e-05, "loss": 6.6089, "step": 18477 }, { "epoch": 0.7790543246832641, "grad_norm": 1.39424467086792, "learning_rate": 4.112948561791436e-05, "loss": 6.533, "step": 18478 }, { "epoch": 0.7790964858654637, "grad_norm": 1.4006876945495605, "learning_rate": 4.1128704831506294e-05, "loss": 6.7432, "step": 18479 }, { "epoch": 0.7791386470476632, "grad_norm": 1.1972416639328003, "learning_rate": 4.1127924045098225e-05, "loss": 6.274, "step": 18480 }, { "epoch": 0.7791808082298628, "grad_norm": 1.4809653759002686, "learning_rate": 4.1127143258690156e-05, "loss": 6.4979, "step": 18481 }, { "epoch": 0.7792229694120623, "grad_norm": 1.704818844795227, "learning_rate": 4.112636247228208e-05, "loss": 6.8833, "step": 18482 }, { "epoch": 0.7792651305942618, "grad_norm": 1.5996264219284058, "learning_rate": 4.112558168587401e-05, "loss": 6.2465, "step": 18483 }, { "epoch": 0.7793072917764614, "grad_norm": 1.6949127912521362, "learning_rate": 4.112480089946594e-05, "loss": 6.5236, "step": 18484 }, { "epoch": 0.779349452958661, "grad_norm": 2.685121536254883, "learning_rate": 4.1124020113057874e-05, "loss": 6.9306, "step": 18485 }, { "epoch": 0.7793916141408606, "grad_norm": 2.167365550994873, "learning_rate": 4.1123239326649806e-05, "loss": 6.545, "step": 18486 }, { "epoch": 0.77943377532306, "grad_norm": 5.199106693267822, "learning_rate": 4.112245854024173e-05, "loss": 7.8685, "step": 18487 }, { "epoch": 0.7794759365052596, "grad_norm": 1.544690728187561, "learning_rate": 4.112167775383367e-05, "loss": 6.4955, "step": 18488 }, { "epoch": 0.7795180976874592, "grad_norm": 1.6109269857406616, "learning_rate": 4.112089696742559e-05, "loss": 7.1559, "step": 18489 }, { "epoch": 0.7795602588696587, "grad_norm": 1.9431471824645996, "learning_rate": 4.1120116181017524e-05, "loss": 6.7059, "step": 18490 }, { "epoch": 0.7796024200518582, "grad_norm": 2.8269128799438477, "learning_rate": 4.1119335394609455e-05, "loss": 8.0578, "step": 18491 }, { "epoch": 0.7796445812340578, "grad_norm": 2.683875560760498, "learning_rate": 4.111855460820138e-05, "loss": 8.5116, "step": 18492 }, { "epoch": 0.7796867424162574, "grad_norm": 1.9722968339920044, "learning_rate": 4.111777382179331e-05, "loss": 6.7916, "step": 18493 }, { "epoch": 0.7797289035984569, "grad_norm": 2.8397793769836426, "learning_rate": 4.111699303538524e-05, "loss": 7.6339, "step": 18494 }, { "epoch": 0.7797710647806565, "grad_norm": 1.6684558391571045, "learning_rate": 4.1116212248977173e-05, "loss": 7.0611, "step": 18495 }, { "epoch": 0.779813225962856, "grad_norm": 1.4991015195846558, "learning_rate": 4.1115431462569105e-05, "loss": 6.6286, "step": 18496 }, { "epoch": 0.7798553871450555, "grad_norm": 1.4249513149261475, "learning_rate": 4.111465067616103e-05, "loss": 6.3992, "step": 18497 }, { "epoch": 0.7798975483272551, "grad_norm": 1.5686852931976318, "learning_rate": 4.111386988975296e-05, "loss": 6.5397, "step": 18498 }, { "epoch": 0.7799397095094547, "grad_norm": 1.301539659500122, "learning_rate": 4.111308910334489e-05, "loss": 6.5685, "step": 18499 }, { "epoch": 0.7799818706916541, "grad_norm": 1.2943962812423706, "learning_rate": 4.111230831693682e-05, "loss": 6.9021, "step": 18500 }, { "epoch": 0.7800240318738537, "grad_norm": 0.9283270239830017, "learning_rate": 4.111152753052875e-05, "loss": 6.4975, "step": 18501 }, { "epoch": 0.7800661930560533, "grad_norm": 2.435776710510254, "learning_rate": 4.111074674412068e-05, "loss": 6.9125, "step": 18502 }, { "epoch": 0.7801083542382529, "grad_norm": 1.3940742015838623, "learning_rate": 4.110996595771261e-05, "loss": 7.0733, "step": 18503 }, { "epoch": 0.7801505154204524, "grad_norm": 1.1439239978790283, "learning_rate": 4.1109185171304534e-05, "loss": 6.7806, "step": 18504 }, { "epoch": 0.7801926766026519, "grad_norm": 1.3863197565078735, "learning_rate": 4.110840438489647e-05, "loss": 6.6997, "step": 18505 }, { "epoch": 0.7802348377848515, "grad_norm": 1.9211621284484863, "learning_rate": 4.11076235984884e-05, "loss": 7.1245, "step": 18506 }, { "epoch": 0.780276998967051, "grad_norm": 1.104527473449707, "learning_rate": 4.1106842812080335e-05, "loss": 6.6087, "step": 18507 }, { "epoch": 0.7803191601492506, "grad_norm": 2.9865288734436035, "learning_rate": 4.110606202567226e-05, "loss": 6.2778, "step": 18508 }, { "epoch": 0.7803613213314501, "grad_norm": 1.3745677471160889, "learning_rate": 4.1105281239264184e-05, "loss": 7.1113, "step": 18509 }, { "epoch": 0.7804034825136497, "grad_norm": 1.7731175422668457, "learning_rate": 4.110450045285612e-05, "loss": 7.0796, "step": 18510 }, { "epoch": 0.7804456436958492, "grad_norm": 1.1456021070480347, "learning_rate": 4.1103719666448047e-05, "loss": 6.7943, "step": 18511 }, { "epoch": 0.7804878048780488, "grad_norm": 2.842139482498169, "learning_rate": 4.110293888003998e-05, "loss": 6.4854, "step": 18512 }, { "epoch": 0.7805299660602484, "grad_norm": 2.2062718868255615, "learning_rate": 4.110215809363191e-05, "loss": 7.3919, "step": 18513 }, { "epoch": 0.7805721272424478, "grad_norm": 2.5323421955108643, "learning_rate": 4.110137730722384e-05, "loss": 6.473, "step": 18514 }, { "epoch": 0.7806142884246474, "grad_norm": 2.6199440956115723, "learning_rate": 4.1100596520815765e-05, "loss": 8.1296, "step": 18515 }, { "epoch": 0.780656449606847, "grad_norm": 3.0458261966705322, "learning_rate": 4.1099815734407696e-05, "loss": 8.6048, "step": 18516 }, { "epoch": 0.7806986107890466, "grad_norm": 2.306265354156494, "learning_rate": 4.109903494799963e-05, "loss": 7.2979, "step": 18517 }, { "epoch": 0.780740771971246, "grad_norm": 1.5546455383300781, "learning_rate": 4.109825416159156e-05, "loss": 6.7477, "step": 18518 }, { "epoch": 0.7807829331534456, "grad_norm": 2.0256776809692383, "learning_rate": 4.109747337518349e-05, "loss": 6.5837, "step": 18519 }, { "epoch": 0.7808250943356452, "grad_norm": 1.501705288887024, "learning_rate": 4.1096692588775414e-05, "loss": 6.7511, "step": 18520 }, { "epoch": 0.7808672555178447, "grad_norm": 1.633394479751587, "learning_rate": 4.1095911802367346e-05, "loss": 6.781, "step": 18521 }, { "epoch": 0.7809094167000443, "grad_norm": 1.6121735572814941, "learning_rate": 4.109513101595928e-05, "loss": 6.7394, "step": 18522 }, { "epoch": 0.7809515778822438, "grad_norm": 4.05587100982666, "learning_rate": 4.10943502295512e-05, "loss": 7.8208, "step": 18523 }, { "epoch": 0.7809937390644434, "grad_norm": 1.6507456302642822, "learning_rate": 4.109356944314314e-05, "loss": 6.5273, "step": 18524 }, { "epoch": 0.7810359002466429, "grad_norm": 1.9218463897705078, "learning_rate": 4.1092788656735064e-05, "loss": 6.6834, "step": 18525 }, { "epoch": 0.7810780614288425, "grad_norm": 1.7198889255523682, "learning_rate": 4.1092007870326995e-05, "loss": 6.725, "step": 18526 }, { "epoch": 0.781120222611042, "grad_norm": 1.5477181673049927, "learning_rate": 4.1091227083918926e-05, "loss": 6.9679, "step": 18527 }, { "epoch": 0.7811623837932415, "grad_norm": 2.32051682472229, "learning_rate": 4.109044629751085e-05, "loss": 6.7632, "step": 18528 }, { "epoch": 0.7812045449754411, "grad_norm": 2.986142873764038, "learning_rate": 4.108966551110279e-05, "loss": 7.208, "step": 18529 }, { "epoch": 0.7812467061576407, "grad_norm": 2.424037456512451, "learning_rate": 4.108888472469471e-05, "loss": 6.9549, "step": 18530 }, { "epoch": 0.7812888673398402, "grad_norm": 1.938025951385498, "learning_rate": 4.1088103938286645e-05, "loss": 7.035, "step": 18531 }, { "epoch": 0.7813310285220397, "grad_norm": 2.5408449172973633, "learning_rate": 4.1087323151878576e-05, "loss": 7.194, "step": 18532 }, { "epoch": 0.7813731897042393, "grad_norm": 3.001702308654785, "learning_rate": 4.108654236547051e-05, "loss": 6.4712, "step": 18533 }, { "epoch": 0.7814153508864389, "grad_norm": 2.7899162769317627, "learning_rate": 4.108576157906243e-05, "loss": 7.272, "step": 18534 }, { "epoch": 0.7814575120686384, "grad_norm": 1.7364342212677002, "learning_rate": 4.108498079265436e-05, "loss": 7.2291, "step": 18535 }, { "epoch": 0.7814996732508379, "grad_norm": 1.1106451749801636, "learning_rate": 4.1084200006246294e-05, "loss": 6.6463, "step": 18536 }, { "epoch": 0.7815418344330375, "grad_norm": 1.5244734287261963, "learning_rate": 4.108341921983822e-05, "loss": 6.5977, "step": 18537 }, { "epoch": 0.7815839956152371, "grad_norm": 1.8194350004196167, "learning_rate": 4.108263843343016e-05, "loss": 7.1481, "step": 18538 }, { "epoch": 0.7816261567974366, "grad_norm": 1.8416706323623657, "learning_rate": 4.108185764702208e-05, "loss": 6.5848, "step": 18539 }, { "epoch": 0.7816683179796361, "grad_norm": 3.0957329273223877, "learning_rate": 4.108107686061401e-05, "loss": 8.0016, "step": 18540 }, { "epoch": 0.7817104791618357, "grad_norm": 1.7288752794265747, "learning_rate": 4.1080296074205944e-05, "loss": 6.3953, "step": 18541 }, { "epoch": 0.7817526403440352, "grad_norm": 1.5310213565826416, "learning_rate": 4.107951528779787e-05, "loss": 7.0738, "step": 18542 }, { "epoch": 0.7817948015262348, "grad_norm": 1.2778632640838623, "learning_rate": 4.1078734501389806e-05, "loss": 6.3935, "step": 18543 }, { "epoch": 0.7818369627084344, "grad_norm": 2.284437894821167, "learning_rate": 4.107795371498173e-05, "loss": 7.5568, "step": 18544 }, { "epoch": 0.7818791238906339, "grad_norm": 2.052727699279785, "learning_rate": 4.107717292857366e-05, "loss": 6.5073, "step": 18545 }, { "epoch": 0.7819212850728334, "grad_norm": 1.9375463724136353, "learning_rate": 4.107639214216559e-05, "loss": 7.2612, "step": 18546 }, { "epoch": 0.781963446255033, "grad_norm": 1.0969892740249634, "learning_rate": 4.107561135575752e-05, "loss": 6.5805, "step": 18547 }, { "epoch": 0.7820056074372326, "grad_norm": 0.9347144961357117, "learning_rate": 4.107483056934945e-05, "loss": 6.4531, "step": 18548 }, { "epoch": 0.782047768619432, "grad_norm": 1.3096126317977905, "learning_rate": 4.107404978294138e-05, "loss": 6.3211, "step": 18549 }, { "epoch": 0.7820899298016316, "grad_norm": 1.3704793453216553, "learning_rate": 4.107326899653331e-05, "loss": 6.5614, "step": 18550 }, { "epoch": 0.7821320909838312, "grad_norm": 2.320159673690796, "learning_rate": 4.107248821012524e-05, "loss": 8.0512, "step": 18551 }, { "epoch": 0.7821742521660308, "grad_norm": 1.5317864418029785, "learning_rate": 4.1071707423717174e-05, "loss": 6.4079, "step": 18552 }, { "epoch": 0.7822164133482303, "grad_norm": 1.4337577819824219, "learning_rate": 4.10709266373091e-05, "loss": 6.6645, "step": 18553 }, { "epoch": 0.7822585745304298, "grad_norm": 3.653236150741577, "learning_rate": 4.107014585090103e-05, "loss": 7.7032, "step": 18554 }, { "epoch": 0.7823007357126294, "grad_norm": 1.1287729740142822, "learning_rate": 4.106936506449296e-05, "loss": 6.4849, "step": 18555 }, { "epoch": 0.7823428968948289, "grad_norm": 3.3759570121765137, "learning_rate": 4.1068584278084885e-05, "loss": 7.6951, "step": 18556 }, { "epoch": 0.7823850580770285, "grad_norm": 1.242283582687378, "learning_rate": 4.1067803491676823e-05, "loss": 6.5502, "step": 18557 }, { "epoch": 0.782427219259228, "grad_norm": 1.0306841135025024, "learning_rate": 4.106702270526875e-05, "loss": 6.5408, "step": 18558 }, { "epoch": 0.7824693804414276, "grad_norm": 1.3628276586532593, "learning_rate": 4.106624191886068e-05, "loss": 6.4308, "step": 18559 }, { "epoch": 0.7825115416236271, "grad_norm": 1.829321026802063, "learning_rate": 4.106546113245261e-05, "loss": 7.0044, "step": 18560 }, { "epoch": 0.7825537028058267, "grad_norm": 3.5754964351654053, "learning_rate": 4.1064680346044535e-05, "loss": 7.3701, "step": 18561 }, { "epoch": 0.7825958639880263, "grad_norm": 1.7415800094604492, "learning_rate": 4.1063899559636466e-05, "loss": 7.007, "step": 18562 }, { "epoch": 0.7826380251702257, "grad_norm": 3.359417200088501, "learning_rate": 4.10631187732284e-05, "loss": 7.8636, "step": 18563 }, { "epoch": 0.7826801863524253, "grad_norm": 1.3775722980499268, "learning_rate": 4.106233798682033e-05, "loss": 6.6469, "step": 18564 }, { "epoch": 0.7827223475346249, "grad_norm": 3.7729761600494385, "learning_rate": 4.106155720041226e-05, "loss": 9.2224, "step": 18565 }, { "epoch": 0.7827645087168245, "grad_norm": 2.167370319366455, "learning_rate": 4.1060776414004185e-05, "loss": 7.4133, "step": 18566 }, { "epoch": 0.7828066698990239, "grad_norm": 2.1902267932891846, "learning_rate": 4.1059995627596116e-05, "loss": 8.0883, "step": 18567 }, { "epoch": 0.7828488310812235, "grad_norm": 3.0657150745391846, "learning_rate": 4.105921484118805e-05, "loss": 6.3063, "step": 18568 }, { "epoch": 0.7828909922634231, "grad_norm": 1.8508087396621704, "learning_rate": 4.105843405477998e-05, "loss": 6.3883, "step": 18569 }, { "epoch": 0.7829331534456226, "grad_norm": 1.6685205698013306, "learning_rate": 4.10576532683719e-05, "loss": 7.1278, "step": 18570 }, { "epoch": 0.7829753146278222, "grad_norm": 1.457003116607666, "learning_rate": 4.1056872481963834e-05, "loss": 7.0602, "step": 18571 }, { "epoch": 0.7830174758100217, "grad_norm": 3.393192768096924, "learning_rate": 4.1056091695555765e-05, "loss": 7.6371, "step": 18572 }, { "epoch": 0.7830596369922213, "grad_norm": 2.4718575477600098, "learning_rate": 4.105531090914769e-05, "loss": 7.0558, "step": 18573 }, { "epoch": 0.7831017981744208, "grad_norm": 1.1893270015716553, "learning_rate": 4.105453012273963e-05, "loss": 6.4152, "step": 18574 }, { "epoch": 0.7831439593566204, "grad_norm": 3.7573702335357666, "learning_rate": 4.105374933633155e-05, "loss": 7.7541, "step": 18575 }, { "epoch": 0.7831861205388199, "grad_norm": 1.4112766981124878, "learning_rate": 4.105296854992349e-05, "loss": 6.3585, "step": 18576 }, { "epoch": 0.7832282817210194, "grad_norm": 1.2906148433685303, "learning_rate": 4.1052187763515415e-05, "loss": 6.7568, "step": 18577 }, { "epoch": 0.783270442903219, "grad_norm": 1.4258016347885132, "learning_rate": 4.1051406977107346e-05, "loss": 6.3876, "step": 18578 }, { "epoch": 0.7833126040854186, "grad_norm": 1.604356050491333, "learning_rate": 4.105062619069928e-05, "loss": 6.8568, "step": 18579 }, { "epoch": 0.7833547652676182, "grad_norm": 4.2007527351379395, "learning_rate": 4.10498454042912e-05, "loss": 7.054, "step": 18580 }, { "epoch": 0.7833969264498176, "grad_norm": 1.2736586332321167, "learning_rate": 4.104906461788313e-05, "loss": 6.4163, "step": 18581 }, { "epoch": 0.7834390876320172, "grad_norm": 1.0684638023376465, "learning_rate": 4.1048283831475064e-05, "loss": 6.8217, "step": 18582 }, { "epoch": 0.7834812488142168, "grad_norm": 1.2591041326522827, "learning_rate": 4.1047503045066996e-05, "loss": 6.4874, "step": 18583 }, { "epoch": 0.7835234099964163, "grad_norm": 3.0857913494110107, "learning_rate": 4.104672225865892e-05, "loss": 8.4324, "step": 18584 }, { "epoch": 0.7835655711786158, "grad_norm": 1.5676380395889282, "learning_rate": 4.104594147225085e-05, "loss": 6.64, "step": 18585 }, { "epoch": 0.7836077323608154, "grad_norm": 1.2372828722000122, "learning_rate": 4.104516068584278e-05, "loss": 6.6156, "step": 18586 }, { "epoch": 0.783649893543015, "grad_norm": 1.3041317462921143, "learning_rate": 4.1044379899434714e-05, "loss": 6.5084, "step": 18587 }, { "epoch": 0.7836920547252145, "grad_norm": 1.1959060430526733, "learning_rate": 4.1043599113026645e-05, "loss": 6.3672, "step": 18588 }, { "epoch": 0.783734215907414, "grad_norm": 1.3865493535995483, "learning_rate": 4.104281832661857e-05, "loss": 7.0161, "step": 18589 }, { "epoch": 0.7837763770896136, "grad_norm": 2.0508248805999756, "learning_rate": 4.10420375402105e-05, "loss": 6.8249, "step": 18590 }, { "epoch": 0.7838185382718131, "grad_norm": 1.017737627029419, "learning_rate": 4.104125675380243e-05, "loss": 6.3498, "step": 18591 }, { "epoch": 0.7838606994540127, "grad_norm": 1.2903929948806763, "learning_rate": 4.1040475967394357e-05, "loss": 6.9528, "step": 18592 }, { "epoch": 0.7839028606362123, "grad_norm": 2.944633960723877, "learning_rate": 4.1039695180986295e-05, "loss": 6.3021, "step": 18593 }, { "epoch": 0.7839450218184117, "grad_norm": 2.304232120513916, "learning_rate": 4.103891439457822e-05, "loss": 6.8543, "step": 18594 }, { "epoch": 0.7839871830006113, "grad_norm": 3.6130287647247314, "learning_rate": 4.103813360817015e-05, "loss": 7.1644, "step": 18595 }, { "epoch": 0.7840293441828109, "grad_norm": 1.4596529006958008, "learning_rate": 4.103735282176208e-05, "loss": 7.3791, "step": 18596 }, { "epoch": 0.7840715053650105, "grad_norm": 1.6086416244506836, "learning_rate": 4.1036572035354006e-05, "loss": 7.3331, "step": 18597 }, { "epoch": 0.7841136665472099, "grad_norm": 1.5800310373306274, "learning_rate": 4.1035791248945944e-05, "loss": 6.9423, "step": 18598 }, { "epoch": 0.7841558277294095, "grad_norm": 1.05711829662323, "learning_rate": 4.103501046253787e-05, "loss": 6.3021, "step": 18599 }, { "epoch": 0.7841979889116091, "grad_norm": 1.61650550365448, "learning_rate": 4.10342296761298e-05, "loss": 6.6001, "step": 18600 }, { "epoch": 0.7842401500938087, "grad_norm": 2.433166742324829, "learning_rate": 4.103344888972173e-05, "loss": 7.3603, "step": 18601 }, { "epoch": 0.7842823112760082, "grad_norm": 1.212271809577942, "learning_rate": 4.103266810331366e-05, "loss": 6.4836, "step": 18602 }, { "epoch": 0.7843244724582077, "grad_norm": 1.799233078956604, "learning_rate": 4.103188731690559e-05, "loss": 6.4083, "step": 18603 }, { "epoch": 0.7843666336404073, "grad_norm": 2.3866524696350098, "learning_rate": 4.103110653049752e-05, "loss": 7.1011, "step": 18604 }, { "epoch": 0.7844087948226068, "grad_norm": 1.212237000465393, "learning_rate": 4.103032574408945e-05, "loss": 6.7385, "step": 18605 }, { "epoch": 0.7844509560048064, "grad_norm": 4.16112756729126, "learning_rate": 4.1029544957681374e-05, "loss": 7.7476, "step": 18606 }, { "epoch": 0.7844931171870059, "grad_norm": 3.143815755844116, "learning_rate": 4.102876417127331e-05, "loss": 7.2617, "step": 18607 }, { "epoch": 0.7845352783692054, "grad_norm": 3.9911837577819824, "learning_rate": 4.1027983384865236e-05, "loss": 7.27, "step": 18608 }, { "epoch": 0.784577439551405, "grad_norm": 4.273454666137695, "learning_rate": 4.102720259845717e-05, "loss": 8.2971, "step": 18609 }, { "epoch": 0.7846196007336046, "grad_norm": 1.9067575931549072, "learning_rate": 4.10264218120491e-05, "loss": 7.1933, "step": 18610 }, { "epoch": 0.7846617619158042, "grad_norm": 3.5443220138549805, "learning_rate": 4.1025641025641023e-05, "loss": 7.7813, "step": 18611 }, { "epoch": 0.7847039230980036, "grad_norm": 1.6071157455444336, "learning_rate": 4.102486023923296e-05, "loss": 6.7716, "step": 18612 }, { "epoch": 0.7847460842802032, "grad_norm": 1.949134111404419, "learning_rate": 4.1024079452824886e-05, "loss": 6.6534, "step": 18613 }, { "epoch": 0.7847882454624028, "grad_norm": 2.3279802799224854, "learning_rate": 4.102329866641682e-05, "loss": 7.3005, "step": 18614 }, { "epoch": 0.7848304066446024, "grad_norm": 1.6691356897354126, "learning_rate": 4.102251788000875e-05, "loss": 6.8469, "step": 18615 }, { "epoch": 0.7848725678268018, "grad_norm": 1.6620291471481323, "learning_rate": 4.102173709360067e-05, "loss": 7.0143, "step": 18616 }, { "epoch": 0.7849147290090014, "grad_norm": 1.3344234228134155, "learning_rate": 4.1020956307192604e-05, "loss": 6.4159, "step": 18617 }, { "epoch": 0.784956890191201, "grad_norm": 1.0481948852539062, "learning_rate": 4.1020175520784535e-05, "loss": 6.6796, "step": 18618 }, { "epoch": 0.7849990513734005, "grad_norm": 3.7179012298583984, "learning_rate": 4.101939473437647e-05, "loss": 7.5519, "step": 18619 }, { "epoch": 0.7850412125556001, "grad_norm": 4.002673625946045, "learning_rate": 4.10186139479684e-05, "loss": 6.8496, "step": 18620 }, { "epoch": 0.7850833737377996, "grad_norm": 1.9375733137130737, "learning_rate": 4.101783316156033e-05, "loss": 6.8093, "step": 18621 }, { "epoch": 0.7851255349199991, "grad_norm": 4.370033264160156, "learning_rate": 4.1017052375152254e-05, "loss": 7.3599, "step": 18622 }, { "epoch": 0.7851676961021987, "grad_norm": 3.3785808086395264, "learning_rate": 4.1016271588744185e-05, "loss": 7.3164, "step": 18623 }, { "epoch": 0.7852098572843983, "grad_norm": 1.700922966003418, "learning_rate": 4.1015490802336116e-05, "loss": 6.5733, "step": 18624 }, { "epoch": 0.7852520184665978, "grad_norm": 2.234941005706787, "learning_rate": 4.101471001592804e-05, "loss": 7.7948, "step": 18625 }, { "epoch": 0.7852941796487973, "grad_norm": 1.9617260694503784, "learning_rate": 4.101392922951998e-05, "loss": 6.4807, "step": 18626 }, { "epoch": 0.7853363408309969, "grad_norm": 1.9016058444976807, "learning_rate": 4.10131484431119e-05, "loss": 6.6365, "step": 18627 }, { "epoch": 0.7853785020131965, "grad_norm": 1.7374211549758911, "learning_rate": 4.1012367656703835e-05, "loss": 6.4779, "step": 18628 }, { "epoch": 0.785420663195396, "grad_norm": 1.5560030937194824, "learning_rate": 4.1011586870295766e-05, "loss": 6.6492, "step": 18629 }, { "epoch": 0.7854628243775955, "grad_norm": 1.653613567352295, "learning_rate": 4.101080608388769e-05, "loss": 6.3884, "step": 18630 }, { "epoch": 0.7855049855597951, "grad_norm": 1.2269527912139893, "learning_rate": 4.101002529747963e-05, "loss": 7.1796, "step": 18631 }, { "epoch": 0.7855471467419947, "grad_norm": 4.178997039794922, "learning_rate": 4.100924451107155e-05, "loss": 6.9826, "step": 18632 }, { "epoch": 0.7855893079241942, "grad_norm": 1.5873596668243408, "learning_rate": 4.1008463724663484e-05, "loss": 7.1169, "step": 18633 }, { "epoch": 0.7856314691063937, "grad_norm": 0.8805820941925049, "learning_rate": 4.1007682938255415e-05, "loss": 6.4541, "step": 18634 }, { "epoch": 0.7856736302885933, "grad_norm": 2.1236660480499268, "learning_rate": 4.100690215184734e-05, "loss": 6.8089, "step": 18635 }, { "epoch": 0.7857157914707928, "grad_norm": 1.829932689666748, "learning_rate": 4.100612136543927e-05, "loss": 6.8787, "step": 18636 }, { "epoch": 0.7857579526529924, "grad_norm": 0.9986201524734497, "learning_rate": 4.10053405790312e-05, "loss": 6.7138, "step": 18637 }, { "epoch": 0.785800113835192, "grad_norm": 1.1881754398345947, "learning_rate": 4.1004559792623134e-05, "loss": 6.6397, "step": 18638 }, { "epoch": 0.7858422750173915, "grad_norm": 2.7576100826263428, "learning_rate": 4.100377900621506e-05, "loss": 6.3175, "step": 18639 }, { "epoch": 0.785884436199591, "grad_norm": 2.1120030879974365, "learning_rate": 4.1002998219806996e-05, "loss": 7.8137, "step": 18640 }, { "epoch": 0.7859265973817906, "grad_norm": 0.9742194414138794, "learning_rate": 4.100221743339892e-05, "loss": 6.8189, "step": 18641 }, { "epoch": 0.7859687585639902, "grad_norm": 1.2035222053527832, "learning_rate": 4.1001436646990845e-05, "loss": 6.8138, "step": 18642 }, { "epoch": 0.7860109197461896, "grad_norm": 2.459852457046509, "learning_rate": 4.100065586058278e-05, "loss": 6.9537, "step": 18643 }, { "epoch": 0.7860530809283892, "grad_norm": 0.8725908994674683, "learning_rate": 4.099987507417471e-05, "loss": 6.6841, "step": 18644 }, { "epoch": 0.7860952421105888, "grad_norm": 2.858844518661499, "learning_rate": 4.0999094287766646e-05, "loss": 7.9628, "step": 18645 }, { "epoch": 0.7861374032927884, "grad_norm": 1.8682597875595093, "learning_rate": 4.099831350135857e-05, "loss": 7.0283, "step": 18646 }, { "epoch": 0.7861795644749878, "grad_norm": 1.163442611694336, "learning_rate": 4.09975327149505e-05, "loss": 7.0809, "step": 18647 }, { "epoch": 0.7862217256571874, "grad_norm": 3.732633352279663, "learning_rate": 4.099675192854243e-05, "loss": 7.808, "step": 18648 }, { "epoch": 0.786263886839387, "grad_norm": 3.189014196395874, "learning_rate": 4.099597114213436e-05, "loss": 7.8241, "step": 18649 }, { "epoch": 0.7863060480215865, "grad_norm": 1.8262287378311157, "learning_rate": 4.099519035572629e-05, "loss": 6.5505, "step": 18650 }, { "epoch": 0.7863482092037861, "grad_norm": 1.836157202720642, "learning_rate": 4.099440956931822e-05, "loss": 7.1492, "step": 18651 }, { "epoch": 0.7863903703859856, "grad_norm": 2.4018781185150146, "learning_rate": 4.099362878291015e-05, "loss": 8.0106, "step": 18652 }, { "epoch": 0.7864325315681852, "grad_norm": 2.8574721813201904, "learning_rate": 4.0992847996502075e-05, "loss": 7.9267, "step": 18653 }, { "epoch": 0.7864746927503847, "grad_norm": 1.7894989252090454, "learning_rate": 4.099206721009401e-05, "loss": 6.5453, "step": 18654 }, { "epoch": 0.7865168539325843, "grad_norm": 2.209784507751465, "learning_rate": 4.099128642368594e-05, "loss": 7.0844, "step": 18655 }, { "epoch": 0.7865590151147838, "grad_norm": 1.135143518447876, "learning_rate": 4.099050563727787e-05, "loss": 6.5556, "step": 18656 }, { "epoch": 0.7866011762969833, "grad_norm": 5.061017990112305, "learning_rate": 4.09897248508698e-05, "loss": 9.9689, "step": 18657 }, { "epoch": 0.7866433374791829, "grad_norm": 3.495944023132324, "learning_rate": 4.0988944064461725e-05, "loss": 7.8543, "step": 18658 }, { "epoch": 0.7866854986613825, "grad_norm": 1.5228666067123413, "learning_rate": 4.0988163278053656e-05, "loss": 6.4828, "step": 18659 }, { "epoch": 0.7867276598435821, "grad_norm": 1.53878915309906, "learning_rate": 4.098738249164559e-05, "loss": 6.6064, "step": 18660 }, { "epoch": 0.7867698210257815, "grad_norm": 1.67766535282135, "learning_rate": 4.098660170523751e-05, "loss": 6.9375, "step": 18661 }, { "epoch": 0.7868119822079811, "grad_norm": 2.9650204181671143, "learning_rate": 4.098582091882945e-05, "loss": 7.2375, "step": 18662 }, { "epoch": 0.7868541433901807, "grad_norm": 1.5128774642944336, "learning_rate": 4.0985040132421374e-05, "loss": 7.0241, "step": 18663 }, { "epoch": 0.7868963045723802, "grad_norm": 1.313144326210022, "learning_rate": 4.0984259346013306e-05, "loss": 6.6152, "step": 18664 }, { "epoch": 0.7869384657545797, "grad_norm": 1.712472677230835, "learning_rate": 4.098347855960524e-05, "loss": 7.1917, "step": 18665 }, { "epoch": 0.7869806269367793, "grad_norm": 2.611370325088501, "learning_rate": 4.098269777319717e-05, "loss": 7.8914, "step": 18666 }, { "epoch": 0.7870227881189789, "grad_norm": 1.6343352794647217, "learning_rate": 4.09819169867891e-05, "loss": 6.6789, "step": 18667 }, { "epoch": 0.7870649493011784, "grad_norm": 3.2757725715637207, "learning_rate": 4.0981136200381024e-05, "loss": 7.3449, "step": 18668 }, { "epoch": 0.787107110483378, "grad_norm": 2.3275716304779053, "learning_rate": 4.0980355413972955e-05, "loss": 7.3677, "step": 18669 }, { "epoch": 0.7871492716655775, "grad_norm": 1.2370747327804565, "learning_rate": 4.0979574627564886e-05, "loss": 6.4997, "step": 18670 }, { "epoch": 0.787191432847777, "grad_norm": 1.8935768604278564, "learning_rate": 4.097879384115682e-05, "loss": 6.9363, "step": 18671 }, { "epoch": 0.7872335940299766, "grad_norm": 1.5168507099151611, "learning_rate": 4.097801305474874e-05, "loss": 6.8053, "step": 18672 }, { "epoch": 0.7872757552121762, "grad_norm": 1.4421080350875854, "learning_rate": 4.0977232268340673e-05, "loss": 6.8836, "step": 18673 }, { "epoch": 0.7873179163943756, "grad_norm": 0.9867163300514221, "learning_rate": 4.0976451481932605e-05, "loss": 6.6178, "step": 18674 }, { "epoch": 0.7873600775765752, "grad_norm": 2.0899431705474854, "learning_rate": 4.097567069552453e-05, "loss": 7.3315, "step": 18675 }, { "epoch": 0.7874022387587748, "grad_norm": 1.7680970430374146, "learning_rate": 4.097488990911647e-05, "loss": 7.0309, "step": 18676 }, { "epoch": 0.7874443999409744, "grad_norm": 2.003685235977173, "learning_rate": 4.097410912270839e-05, "loss": 7.5165, "step": 18677 }, { "epoch": 0.787486561123174, "grad_norm": 1.782781720161438, "learning_rate": 4.097332833630032e-05, "loss": 6.8707, "step": 18678 }, { "epoch": 0.7875287223053734, "grad_norm": 3.013054370880127, "learning_rate": 4.0972547549892254e-05, "loss": 7.6073, "step": 18679 }, { "epoch": 0.787570883487573, "grad_norm": 3.3246777057647705, "learning_rate": 4.097176676348418e-05, "loss": 9.1048, "step": 18680 }, { "epoch": 0.7876130446697726, "grad_norm": 3.4348363876342773, "learning_rate": 4.097098597707612e-05, "loss": 6.5936, "step": 18681 }, { "epoch": 0.7876552058519721, "grad_norm": 2.162330150604248, "learning_rate": 4.097020519066804e-05, "loss": 6.4887, "step": 18682 }, { "epoch": 0.7876973670341716, "grad_norm": 2.0184576511383057, "learning_rate": 4.096942440425997e-05, "loss": 7.339, "step": 18683 }, { "epoch": 0.7877395282163712, "grad_norm": 2.0121116638183594, "learning_rate": 4.0968643617851904e-05, "loss": 6.4258, "step": 18684 }, { "epoch": 0.7877816893985707, "grad_norm": 2.894646167755127, "learning_rate": 4.0967862831443835e-05, "loss": 7.8351, "step": 18685 }, { "epoch": 0.7878238505807703, "grad_norm": 3.13322377204895, "learning_rate": 4.096708204503576e-05, "loss": 7.0563, "step": 18686 }, { "epoch": 0.7878660117629699, "grad_norm": 3.5411171913146973, "learning_rate": 4.096630125862769e-05, "loss": 7.2531, "step": 18687 }, { "epoch": 0.7879081729451693, "grad_norm": 1.2589744329452515, "learning_rate": 4.096552047221962e-05, "loss": 6.7912, "step": 18688 }, { "epoch": 0.7879503341273689, "grad_norm": 1.4780209064483643, "learning_rate": 4.096473968581155e-05, "loss": 6.7946, "step": 18689 }, { "epoch": 0.7879924953095685, "grad_norm": 1.8846386671066284, "learning_rate": 4.0963958899403485e-05, "loss": 6.5842, "step": 18690 }, { "epoch": 0.7880346564917681, "grad_norm": 2.5950868129730225, "learning_rate": 4.096317811299541e-05, "loss": 8.876, "step": 18691 }, { "epoch": 0.7880768176739675, "grad_norm": 2.0705204010009766, "learning_rate": 4.096239732658734e-05, "loss": 6.4378, "step": 18692 }, { "epoch": 0.7881189788561671, "grad_norm": 1.1197140216827393, "learning_rate": 4.096161654017927e-05, "loss": 6.9241, "step": 18693 }, { "epoch": 0.7881611400383667, "grad_norm": 2.943031072616577, "learning_rate": 4.0960835753771196e-05, "loss": 6.3131, "step": 18694 }, { "epoch": 0.7882033012205663, "grad_norm": 3.120112895965576, "learning_rate": 4.0960054967363134e-05, "loss": 6.6467, "step": 18695 }, { "epoch": 0.7882454624027658, "grad_norm": 3.2737510204315186, "learning_rate": 4.095927418095506e-05, "loss": 7.0834, "step": 18696 }, { "epoch": 0.7882876235849653, "grad_norm": 1.2584611177444458, "learning_rate": 4.095849339454699e-05, "loss": 6.3436, "step": 18697 }, { "epoch": 0.7883297847671649, "grad_norm": 2.33721923828125, "learning_rate": 4.095771260813892e-05, "loss": 6.5477, "step": 18698 }, { "epoch": 0.7883719459493644, "grad_norm": 2.0605924129486084, "learning_rate": 4.0956931821730846e-05, "loss": 7.3455, "step": 18699 }, { "epoch": 0.788414107131564, "grad_norm": 2.9950528144836426, "learning_rate": 4.0956151035322784e-05, "loss": 7.7873, "step": 18700 }, { "epoch": 0.7884562683137635, "grad_norm": 4.1468634605407715, "learning_rate": 4.095537024891471e-05, "loss": 9.3481, "step": 18701 }, { "epoch": 0.788498429495963, "grad_norm": 2.3563356399536133, "learning_rate": 4.095458946250664e-05, "loss": 6.928, "step": 18702 }, { "epoch": 0.7885405906781626, "grad_norm": 2.8453333377838135, "learning_rate": 4.095380867609857e-05, "loss": 6.868, "step": 18703 }, { "epoch": 0.7885827518603622, "grad_norm": 2.6035499572753906, "learning_rate": 4.0953027889690495e-05, "loss": 7.3714, "step": 18704 }, { "epoch": 0.7886249130425617, "grad_norm": 1.8601118326187134, "learning_rate": 4.0952247103282426e-05, "loss": 7.1304, "step": 18705 }, { "epoch": 0.7886670742247612, "grad_norm": 1.990565299987793, "learning_rate": 4.095146631687436e-05, "loss": 6.8469, "step": 18706 }, { "epoch": 0.7887092354069608, "grad_norm": 2.822401523590088, "learning_rate": 4.095068553046629e-05, "loss": 8.0328, "step": 18707 }, { "epoch": 0.7887513965891604, "grad_norm": 2.221872091293335, "learning_rate": 4.094990474405821e-05, "loss": 7.7763, "step": 18708 }, { "epoch": 0.78879355777136, "grad_norm": 1.575109601020813, "learning_rate": 4.094912395765015e-05, "loss": 6.9298, "step": 18709 }, { "epoch": 0.7888357189535594, "grad_norm": 1.91556978225708, "learning_rate": 4.0948343171242076e-05, "loss": 6.3442, "step": 18710 }, { "epoch": 0.788877880135759, "grad_norm": 2.0274710655212402, "learning_rate": 4.094756238483401e-05, "loss": 6.4094, "step": 18711 }, { "epoch": 0.7889200413179586, "grad_norm": 1.4386943578720093, "learning_rate": 4.094678159842594e-05, "loss": 6.8418, "step": 18712 }, { "epoch": 0.7889622025001581, "grad_norm": 3.1501293182373047, "learning_rate": 4.094600081201786e-05, "loss": 6.6909, "step": 18713 }, { "epoch": 0.7890043636823576, "grad_norm": 1.1904618740081787, "learning_rate": 4.09452200256098e-05, "loss": 6.7632, "step": 18714 }, { "epoch": 0.7890465248645572, "grad_norm": 1.5782179832458496, "learning_rate": 4.0944439239201725e-05, "loss": 6.5426, "step": 18715 }, { "epoch": 0.7890886860467567, "grad_norm": 2.3921375274658203, "learning_rate": 4.094365845279366e-05, "loss": 6.5813, "step": 18716 }, { "epoch": 0.7891308472289563, "grad_norm": 1.1241921186447144, "learning_rate": 4.094287766638559e-05, "loss": 6.4786, "step": 18717 }, { "epoch": 0.7891730084111559, "grad_norm": 1.2492547035217285, "learning_rate": 4.094209687997751e-05, "loss": 6.7697, "step": 18718 }, { "epoch": 0.7892151695933554, "grad_norm": 1.1989233493804932, "learning_rate": 4.0941316093569444e-05, "loss": 6.404, "step": 18719 }, { "epoch": 0.7892573307755549, "grad_norm": 1.2093160152435303, "learning_rate": 4.0940535307161375e-05, "loss": 6.3093, "step": 18720 }, { "epoch": 0.7892994919577545, "grad_norm": 1.221718192100525, "learning_rate": 4.0939754520753306e-05, "loss": 6.5671, "step": 18721 }, { "epoch": 0.7893416531399541, "grad_norm": 1.21658456325531, "learning_rate": 4.093897373434523e-05, "loss": 6.7916, "step": 18722 }, { "epoch": 0.7893838143221535, "grad_norm": 3.560572624206543, "learning_rate": 4.093819294793716e-05, "loss": 7.3176, "step": 18723 }, { "epoch": 0.7894259755043531, "grad_norm": 1.6887540817260742, "learning_rate": 4.093741216152909e-05, "loss": 6.8694, "step": 18724 }, { "epoch": 0.7894681366865527, "grad_norm": 2.0157251358032227, "learning_rate": 4.0936631375121024e-05, "loss": 6.4956, "step": 18725 }, { "epoch": 0.7895102978687523, "grad_norm": 2.938175678253174, "learning_rate": 4.0935850588712956e-05, "loss": 7.7646, "step": 18726 }, { "epoch": 0.7895524590509518, "grad_norm": 3.8972320556640625, "learning_rate": 4.093506980230488e-05, "loss": 9.2636, "step": 18727 }, { "epoch": 0.7895946202331513, "grad_norm": 1.7864757776260376, "learning_rate": 4.093428901589682e-05, "loss": 7.3627, "step": 18728 }, { "epoch": 0.7896367814153509, "grad_norm": 2.1284971237182617, "learning_rate": 4.093350822948874e-05, "loss": 7.1837, "step": 18729 }, { "epoch": 0.7896789425975504, "grad_norm": 2.0272769927978516, "learning_rate": 4.0932727443080674e-05, "loss": 6.8742, "step": 18730 }, { "epoch": 0.78972110377975, "grad_norm": 2.3536524772644043, "learning_rate": 4.0931946656672605e-05, "loss": 6.4987, "step": 18731 }, { "epoch": 0.7897632649619495, "grad_norm": 1.7585499286651611, "learning_rate": 4.093116587026453e-05, "loss": 6.376, "step": 18732 }, { "epoch": 0.7898054261441491, "grad_norm": 2.393207550048828, "learning_rate": 4.093038508385646e-05, "loss": 7.6481, "step": 18733 }, { "epoch": 0.7898475873263486, "grad_norm": 1.2699592113494873, "learning_rate": 4.092960429744839e-05, "loss": 7.2138, "step": 18734 }, { "epoch": 0.7898897485085482, "grad_norm": 2.666764259338379, "learning_rate": 4.0928823511040323e-05, "loss": 7.1885, "step": 18735 }, { "epoch": 0.7899319096907478, "grad_norm": 1.384793996810913, "learning_rate": 4.0928042724632255e-05, "loss": 6.5317, "step": 18736 }, { "epoch": 0.7899740708729472, "grad_norm": 2.589205503463745, "learning_rate": 4.092726193822418e-05, "loss": 7.9474, "step": 18737 }, { "epoch": 0.7900162320551468, "grad_norm": 1.6355657577514648, "learning_rate": 4.092648115181611e-05, "loss": 7.082, "step": 18738 }, { "epoch": 0.7900583932373464, "grad_norm": 2.51904559135437, "learning_rate": 4.092570036540804e-05, "loss": 7.2421, "step": 18739 }, { "epoch": 0.790100554419546, "grad_norm": 1.0529215335845947, "learning_rate": 4.092491957899997e-05, "loss": 6.543, "step": 18740 }, { "epoch": 0.7901427156017454, "grad_norm": 1.616868257522583, "learning_rate": 4.09241387925919e-05, "loss": 6.3927, "step": 18741 }, { "epoch": 0.790184876783945, "grad_norm": 1.162438988685608, "learning_rate": 4.092335800618383e-05, "loss": 6.6587, "step": 18742 }, { "epoch": 0.7902270379661446, "grad_norm": 1.1034232378005981, "learning_rate": 4.092257721977576e-05, "loss": 7.1773, "step": 18743 }, { "epoch": 0.7902691991483441, "grad_norm": 3.08242130279541, "learning_rate": 4.0921796433367684e-05, "loss": 7.7687, "step": 18744 }, { "epoch": 0.7903113603305437, "grad_norm": 1.6330978870391846, "learning_rate": 4.092101564695962e-05, "loss": 6.7497, "step": 18745 }, { "epoch": 0.7903535215127432, "grad_norm": 2.1206846237182617, "learning_rate": 4.092023486055155e-05, "loss": 6.7052, "step": 18746 }, { "epoch": 0.7903956826949428, "grad_norm": 1.1880366802215576, "learning_rate": 4.0919454074143485e-05, "loss": 6.6329, "step": 18747 }, { "epoch": 0.7904378438771423, "grad_norm": 2.185896396636963, "learning_rate": 4.091867328773541e-05, "loss": 7.0509, "step": 18748 }, { "epoch": 0.7904800050593419, "grad_norm": 1.584547519683838, "learning_rate": 4.0917892501327334e-05, "loss": 6.6589, "step": 18749 }, { "epoch": 0.7905221662415414, "grad_norm": 1.9000589847564697, "learning_rate": 4.091711171491927e-05, "loss": 6.5749, "step": 18750 }, { "epoch": 0.7905643274237409, "grad_norm": 1.2006027698516846, "learning_rate": 4.0916330928511197e-05, "loss": 6.8536, "step": 18751 }, { "epoch": 0.7906064886059405, "grad_norm": 1.2727999687194824, "learning_rate": 4.091555014210313e-05, "loss": 6.8498, "step": 18752 }, { "epoch": 0.7906486497881401, "grad_norm": 1.4161406755447388, "learning_rate": 4.091476935569506e-05, "loss": 6.4768, "step": 18753 }, { "epoch": 0.7906908109703397, "grad_norm": 2.536593437194824, "learning_rate": 4.091398856928699e-05, "loss": 7.5345, "step": 18754 }, { "epoch": 0.7907329721525391, "grad_norm": 2.4610533714294434, "learning_rate": 4.0913207782878915e-05, "loss": 6.5295, "step": 18755 }, { "epoch": 0.7907751333347387, "grad_norm": 2.312084674835205, "learning_rate": 4.0912426996470846e-05, "loss": 6.9972, "step": 18756 }, { "epoch": 0.7908172945169383, "grad_norm": 1.5269477367401123, "learning_rate": 4.091164621006278e-05, "loss": 7.0043, "step": 18757 }, { "epoch": 0.7908594556991378, "grad_norm": 1.6109813451766968, "learning_rate": 4.091086542365471e-05, "loss": 6.3872, "step": 18758 }, { "epoch": 0.7909016168813373, "grad_norm": 0.93494713306427, "learning_rate": 4.091008463724664e-05, "loss": 6.4816, "step": 18759 }, { "epoch": 0.7909437780635369, "grad_norm": 1.2351824045181274, "learning_rate": 4.0909303850838564e-05, "loss": 6.6531, "step": 18760 }, { "epoch": 0.7909859392457365, "grad_norm": 1.3321880102157593, "learning_rate": 4.0908523064430496e-05, "loss": 6.4373, "step": 18761 }, { "epoch": 0.791028100427936, "grad_norm": 3.260033130645752, "learning_rate": 4.090774227802243e-05, "loss": 7.3944, "step": 18762 }, { "epoch": 0.7910702616101355, "grad_norm": 2.9654579162597656, "learning_rate": 4.090696149161435e-05, "loss": 7.2576, "step": 18763 }, { "epoch": 0.7911124227923351, "grad_norm": 2.850942611694336, "learning_rate": 4.090618070520629e-05, "loss": 7.0827, "step": 18764 }, { "epoch": 0.7911545839745346, "grad_norm": 1.3591334819793701, "learning_rate": 4.0905399918798214e-05, "loss": 6.9711, "step": 18765 }, { "epoch": 0.7911967451567342, "grad_norm": 2.159282922744751, "learning_rate": 4.0904619132390145e-05, "loss": 7.9183, "step": 18766 }, { "epoch": 0.7912389063389338, "grad_norm": 1.9735982418060303, "learning_rate": 4.0903838345982076e-05, "loss": 7.291, "step": 18767 }, { "epoch": 0.7912810675211333, "grad_norm": 2.189277172088623, "learning_rate": 4.0903057559574e-05, "loss": 6.8856, "step": 18768 }, { "epoch": 0.7913232287033328, "grad_norm": 2.6695451736450195, "learning_rate": 4.090227677316594e-05, "loss": 7.3637, "step": 18769 }, { "epoch": 0.7913653898855324, "grad_norm": 7.060012340545654, "learning_rate": 4.090149598675786e-05, "loss": 6.8522, "step": 18770 }, { "epoch": 0.791407551067732, "grad_norm": 2.5113930702209473, "learning_rate": 4.0900715200349795e-05, "loss": 7.9564, "step": 18771 }, { "epoch": 0.7914497122499314, "grad_norm": 1.3918542861938477, "learning_rate": 4.0899934413941726e-05, "loss": 7.0465, "step": 18772 }, { "epoch": 0.791491873432131, "grad_norm": 2.9119889736175537, "learning_rate": 4.089915362753366e-05, "loss": 7.8254, "step": 18773 }, { "epoch": 0.7915340346143306, "grad_norm": 1.4635107517242432, "learning_rate": 4.089837284112558e-05, "loss": 6.7331, "step": 18774 }, { "epoch": 0.7915761957965302, "grad_norm": 1.3937480449676514, "learning_rate": 4.089759205471751e-05, "loss": 6.6061, "step": 18775 }, { "epoch": 0.7916183569787297, "grad_norm": 1.829406499862671, "learning_rate": 4.0896811268309444e-05, "loss": 7.2881, "step": 18776 }, { "epoch": 0.7916605181609292, "grad_norm": 1.492506980895996, "learning_rate": 4.089603048190137e-05, "loss": 6.8682, "step": 18777 }, { "epoch": 0.7917026793431288, "grad_norm": 1.285615086555481, "learning_rate": 4.089524969549331e-05, "loss": 7.2532, "step": 18778 }, { "epoch": 0.7917448405253283, "grad_norm": 1.2109715938568115, "learning_rate": 4.089446890908523e-05, "loss": 6.5485, "step": 18779 }, { "epoch": 0.7917870017075279, "grad_norm": 2.323460340499878, "learning_rate": 4.089368812267716e-05, "loss": 6.6265, "step": 18780 }, { "epoch": 0.7918291628897274, "grad_norm": 1.8837515115737915, "learning_rate": 4.0892907336269094e-05, "loss": 6.5299, "step": 18781 }, { "epoch": 0.791871324071927, "grad_norm": 2.611729383468628, "learning_rate": 4.089212654986102e-05, "loss": 7.0808, "step": 18782 }, { "epoch": 0.7919134852541265, "grad_norm": 1.0740461349487305, "learning_rate": 4.0891345763452956e-05, "loss": 6.359, "step": 18783 }, { "epoch": 0.7919556464363261, "grad_norm": 1.5864231586456299, "learning_rate": 4.089056497704488e-05, "loss": 6.434, "step": 18784 }, { "epoch": 0.7919978076185257, "grad_norm": 1.4966930150985718, "learning_rate": 4.088978419063681e-05, "loss": 6.9108, "step": 18785 }, { "epoch": 0.7920399688007251, "grad_norm": 1.4621846675872803, "learning_rate": 4.088900340422874e-05, "loss": 6.517, "step": 18786 }, { "epoch": 0.7920821299829247, "grad_norm": 1.5848437547683716, "learning_rate": 4.088822261782067e-05, "loss": 6.5916, "step": 18787 }, { "epoch": 0.7921242911651243, "grad_norm": 1.9269447326660156, "learning_rate": 4.08874418314126e-05, "loss": 7.0806, "step": 18788 }, { "epoch": 0.7921664523473239, "grad_norm": 0.8875485062599182, "learning_rate": 4.088666104500453e-05, "loss": 6.5719, "step": 18789 }, { "epoch": 0.7922086135295233, "grad_norm": 1.8061752319335938, "learning_rate": 4.088588025859646e-05, "loss": 6.3248, "step": 18790 }, { "epoch": 0.7922507747117229, "grad_norm": 3.3427748680114746, "learning_rate": 4.088509947218839e-05, "loss": 7.1499, "step": 18791 }, { "epoch": 0.7922929358939225, "grad_norm": 1.93198561668396, "learning_rate": 4.0884318685780324e-05, "loss": 7.4532, "step": 18792 }, { "epoch": 0.792335097076122, "grad_norm": 2.9969639778137207, "learning_rate": 4.088353789937225e-05, "loss": 7.7647, "step": 18793 }, { "epoch": 0.7923772582583216, "grad_norm": 1.840527057647705, "learning_rate": 4.088275711296418e-05, "loss": 6.7442, "step": 18794 }, { "epoch": 0.7924194194405211, "grad_norm": 3.3669281005859375, "learning_rate": 4.088197632655611e-05, "loss": 7.7392, "step": 18795 }, { "epoch": 0.7924615806227207, "grad_norm": 2.172140598297119, "learning_rate": 4.0881195540148035e-05, "loss": 7.5386, "step": 18796 }, { "epoch": 0.7925037418049202, "grad_norm": 1.831254005432129, "learning_rate": 4.0880414753739973e-05, "loss": 6.5696, "step": 18797 }, { "epoch": 0.7925459029871198, "grad_norm": 1.9272974729537964, "learning_rate": 4.08796339673319e-05, "loss": 6.6272, "step": 18798 }, { "epoch": 0.7925880641693193, "grad_norm": 2.976081609725952, "learning_rate": 4.087885318092383e-05, "loss": 7.8457, "step": 18799 }, { "epoch": 0.7926302253515188, "grad_norm": 1.8908660411834717, "learning_rate": 4.087807239451576e-05, "loss": 7.2088, "step": 18800 }, { "epoch": 0.7926723865337184, "grad_norm": 4.22589635848999, "learning_rate": 4.0877291608107685e-05, "loss": 7.7707, "step": 18801 }, { "epoch": 0.792714547715918, "grad_norm": 2.3604118824005127, "learning_rate": 4.0876510821699616e-05, "loss": 7.2293, "step": 18802 }, { "epoch": 0.7927567088981176, "grad_norm": 2.0792887210845947, "learning_rate": 4.087573003529155e-05, "loss": 6.801, "step": 18803 }, { "epoch": 0.792798870080317, "grad_norm": 1.6682864427566528, "learning_rate": 4.087494924888348e-05, "loss": 7.0165, "step": 18804 }, { "epoch": 0.7928410312625166, "grad_norm": 2.2390410900115967, "learning_rate": 4.087416846247541e-05, "loss": 6.5117, "step": 18805 }, { "epoch": 0.7928831924447162, "grad_norm": 2.205606460571289, "learning_rate": 4.0873387676067335e-05, "loss": 7.4379, "step": 18806 }, { "epoch": 0.7929253536269157, "grad_norm": 3.0157127380371094, "learning_rate": 4.0872606889659266e-05, "loss": 7.5996, "step": 18807 }, { "epoch": 0.7929675148091152, "grad_norm": 2.639285087585449, "learning_rate": 4.08718261032512e-05, "loss": 7.1168, "step": 18808 }, { "epoch": 0.7930096759913148, "grad_norm": 1.7006945610046387, "learning_rate": 4.087104531684313e-05, "loss": 7.0437, "step": 18809 }, { "epoch": 0.7930518371735144, "grad_norm": 1.640447735786438, "learning_rate": 4.087026453043505e-05, "loss": 7.0302, "step": 18810 }, { "epoch": 0.7930939983557139, "grad_norm": 1.824121356010437, "learning_rate": 4.0869483744026984e-05, "loss": 7.0922, "step": 18811 }, { "epoch": 0.7931361595379134, "grad_norm": 1.3445509672164917, "learning_rate": 4.0868702957618915e-05, "loss": 6.7214, "step": 18812 }, { "epoch": 0.793178320720113, "grad_norm": 1.5017120838165283, "learning_rate": 4.086792217121084e-05, "loss": 6.6321, "step": 18813 }, { "epoch": 0.7932204819023125, "grad_norm": 3.8744049072265625, "learning_rate": 4.086714138480278e-05, "loss": 7.5498, "step": 18814 }, { "epoch": 0.7932626430845121, "grad_norm": 1.7849069833755493, "learning_rate": 4.08663605983947e-05, "loss": 6.7324, "step": 18815 }, { "epoch": 0.7933048042667117, "grad_norm": 1.5344631671905518, "learning_rate": 4.086557981198664e-05, "loss": 6.5458, "step": 18816 }, { "epoch": 0.7933469654489111, "grad_norm": 1.1743013858795166, "learning_rate": 4.0864799025578565e-05, "loss": 6.5587, "step": 18817 }, { "epoch": 0.7933891266311107, "grad_norm": 1.1717495918273926, "learning_rate": 4.0864018239170496e-05, "loss": 6.7317, "step": 18818 }, { "epoch": 0.7934312878133103, "grad_norm": 1.4672248363494873, "learning_rate": 4.086323745276243e-05, "loss": 7.1714, "step": 18819 }, { "epoch": 0.7934734489955099, "grad_norm": 3.7549588680267334, "learning_rate": 4.086245666635435e-05, "loss": 7.4973, "step": 18820 }, { "epoch": 0.7935156101777093, "grad_norm": 1.1756948232650757, "learning_rate": 4.086167587994628e-05, "loss": 6.5477, "step": 18821 }, { "epoch": 0.7935577713599089, "grad_norm": 1.3977793455123901, "learning_rate": 4.0860895093538214e-05, "loss": 6.4532, "step": 18822 }, { "epoch": 0.7935999325421085, "grad_norm": 3.5005033016204834, "learning_rate": 4.0860114307130146e-05, "loss": 7.6687, "step": 18823 }, { "epoch": 0.793642093724308, "grad_norm": 2.4357388019561768, "learning_rate": 4.085933352072207e-05, "loss": 7.9427, "step": 18824 }, { "epoch": 0.7936842549065076, "grad_norm": 1.086992859840393, "learning_rate": 4.0858552734314e-05, "loss": 6.6563, "step": 18825 }, { "epoch": 0.7937264160887071, "grad_norm": 1.4857616424560547, "learning_rate": 4.085777194790593e-05, "loss": 6.5289, "step": 18826 }, { "epoch": 0.7937685772709067, "grad_norm": 1.8761074542999268, "learning_rate": 4.0856991161497864e-05, "loss": 7.2962, "step": 18827 }, { "epoch": 0.7938107384531062, "grad_norm": 1.8774056434631348, "learning_rate": 4.0856210375089795e-05, "loss": 6.6641, "step": 18828 }, { "epoch": 0.7938528996353058, "grad_norm": 2.1890361309051514, "learning_rate": 4.085542958868172e-05, "loss": 6.5142, "step": 18829 }, { "epoch": 0.7938950608175053, "grad_norm": 1.4733721017837524, "learning_rate": 4.085464880227365e-05, "loss": 6.9544, "step": 18830 }, { "epoch": 0.7939372219997048, "grad_norm": 1.4420642852783203, "learning_rate": 4.085386801586558e-05, "loss": 6.8744, "step": 18831 }, { "epoch": 0.7939793831819044, "grad_norm": 2.6704297065734863, "learning_rate": 4.0853087229457507e-05, "loss": 7.0437, "step": 18832 }, { "epoch": 0.794021544364104, "grad_norm": 1.4897780418395996, "learning_rate": 4.0852306443049445e-05, "loss": 7.0053, "step": 18833 }, { "epoch": 0.7940637055463036, "grad_norm": 1.28834068775177, "learning_rate": 4.085152565664137e-05, "loss": 6.5668, "step": 18834 }, { "epoch": 0.794105866728503, "grad_norm": 1.2766046524047852, "learning_rate": 4.08507448702333e-05, "loss": 6.5775, "step": 18835 }, { "epoch": 0.7941480279107026, "grad_norm": 1.1201541423797607, "learning_rate": 4.084996408382523e-05, "loss": 6.8308, "step": 18836 }, { "epoch": 0.7941901890929022, "grad_norm": 1.8111648559570312, "learning_rate": 4.084918329741716e-05, "loss": 7.0884, "step": 18837 }, { "epoch": 0.7942323502751018, "grad_norm": 1.8983705043792725, "learning_rate": 4.0848402511009094e-05, "loss": 6.8413, "step": 18838 }, { "epoch": 0.7942745114573012, "grad_norm": 2.284820556640625, "learning_rate": 4.084762172460102e-05, "loss": 6.5607, "step": 18839 }, { "epoch": 0.7943166726395008, "grad_norm": 1.735526442527771, "learning_rate": 4.084684093819295e-05, "loss": 6.4422, "step": 18840 }, { "epoch": 0.7943588338217004, "grad_norm": 2.3642804622650146, "learning_rate": 4.084606015178488e-05, "loss": 7.8287, "step": 18841 }, { "epoch": 0.7944009950038999, "grad_norm": 1.059130072593689, "learning_rate": 4.084527936537681e-05, "loss": 6.6847, "step": 18842 }, { "epoch": 0.7944431561860995, "grad_norm": 0.8300490975379944, "learning_rate": 4.084449857896874e-05, "loss": 6.5381, "step": 18843 }, { "epoch": 0.794485317368299, "grad_norm": 1.7927557229995728, "learning_rate": 4.084371779256067e-05, "loss": 7.0203, "step": 18844 }, { "epoch": 0.7945274785504985, "grad_norm": 4.092026233673096, "learning_rate": 4.08429370061526e-05, "loss": 7.7576, "step": 18845 }, { "epoch": 0.7945696397326981, "grad_norm": 3.212610960006714, "learning_rate": 4.0842156219744524e-05, "loss": 7.7313, "step": 18846 }, { "epoch": 0.7946118009148977, "grad_norm": 1.4015284776687622, "learning_rate": 4.084137543333646e-05, "loss": 7.0658, "step": 18847 }, { "epoch": 0.7946539620970972, "grad_norm": 1.3798907995224, "learning_rate": 4.0840594646928386e-05, "loss": 7.0551, "step": 18848 }, { "epoch": 0.7946961232792967, "grad_norm": 1.9160807132720947, "learning_rate": 4.083981386052032e-05, "loss": 6.726, "step": 18849 }, { "epoch": 0.7947382844614963, "grad_norm": 2.6466197967529297, "learning_rate": 4.083903307411225e-05, "loss": 7.6013, "step": 18850 }, { "epoch": 0.7947804456436959, "grad_norm": 1.355101227760315, "learning_rate": 4.0838252287704173e-05, "loss": 7.141, "step": 18851 }, { "epoch": 0.7948226068258955, "grad_norm": 3.6615114212036133, "learning_rate": 4.083747150129611e-05, "loss": 7.9266, "step": 18852 }, { "epoch": 0.7948647680080949, "grad_norm": 1.2888331413269043, "learning_rate": 4.0836690714888036e-05, "loss": 6.619, "step": 18853 }, { "epoch": 0.7949069291902945, "grad_norm": 3.942843437194824, "learning_rate": 4.083590992847997e-05, "loss": 8.1522, "step": 18854 }, { "epoch": 0.7949490903724941, "grad_norm": 1.4654812812805176, "learning_rate": 4.08351291420719e-05, "loss": 6.5735, "step": 18855 }, { "epoch": 0.7949912515546936, "grad_norm": 1.3653016090393066, "learning_rate": 4.083434835566382e-05, "loss": 6.9406, "step": 18856 }, { "epoch": 0.7950334127368931, "grad_norm": 2.8049206733703613, "learning_rate": 4.0833567569255754e-05, "loss": 6.7438, "step": 18857 }, { "epoch": 0.7950755739190927, "grad_norm": 1.5971441268920898, "learning_rate": 4.0832786782847685e-05, "loss": 6.6784, "step": 18858 }, { "epoch": 0.7951177351012922, "grad_norm": 2.788212776184082, "learning_rate": 4.083200599643962e-05, "loss": 7.5372, "step": 18859 }, { "epoch": 0.7951598962834918, "grad_norm": 1.5794817209243774, "learning_rate": 4.083122521003155e-05, "loss": 6.9253, "step": 18860 }, { "epoch": 0.7952020574656914, "grad_norm": 1.20391047000885, "learning_rate": 4.083044442362348e-05, "loss": 6.9198, "step": 18861 }, { "epoch": 0.7952442186478909, "grad_norm": 2.0476603507995605, "learning_rate": 4.0829663637215404e-05, "loss": 6.8075, "step": 18862 }, { "epoch": 0.7952863798300904, "grad_norm": 0.9645248651504517, "learning_rate": 4.0828882850807335e-05, "loss": 6.5598, "step": 18863 }, { "epoch": 0.79532854101229, "grad_norm": 2.867147445678711, "learning_rate": 4.0828102064399266e-05, "loss": 6.5784, "step": 18864 }, { "epoch": 0.7953707021944896, "grad_norm": 2.1103780269622803, "learning_rate": 4.082732127799119e-05, "loss": 7.3114, "step": 18865 }, { "epoch": 0.795412863376689, "grad_norm": 1.8150458335876465, "learning_rate": 4.082654049158313e-05, "loss": 6.8638, "step": 18866 }, { "epoch": 0.7954550245588886, "grad_norm": 3.8884968757629395, "learning_rate": 4.082575970517505e-05, "loss": 7.7602, "step": 18867 }, { "epoch": 0.7954971857410882, "grad_norm": 2.686835289001465, "learning_rate": 4.0824978918766985e-05, "loss": 7.5515, "step": 18868 }, { "epoch": 0.7955393469232878, "grad_norm": 1.9197258949279785, "learning_rate": 4.0824198132358916e-05, "loss": 6.4717, "step": 18869 }, { "epoch": 0.7955815081054872, "grad_norm": 1.210129737854004, "learning_rate": 4.082341734595084e-05, "loss": 6.551, "step": 18870 }, { "epoch": 0.7956236692876868, "grad_norm": 1.1347640752792358, "learning_rate": 4.082263655954277e-05, "loss": 6.607, "step": 18871 }, { "epoch": 0.7956658304698864, "grad_norm": 2.4760079383850098, "learning_rate": 4.08218557731347e-05, "loss": 6.7411, "step": 18872 }, { "epoch": 0.7957079916520859, "grad_norm": 2.7990148067474365, "learning_rate": 4.0821074986726634e-05, "loss": 7.1635, "step": 18873 }, { "epoch": 0.7957501528342855, "grad_norm": 3.0732405185699463, "learning_rate": 4.0820294200318565e-05, "loss": 6.964, "step": 18874 }, { "epoch": 0.795792314016485, "grad_norm": 2.193046808242798, "learning_rate": 4.081951341391049e-05, "loss": 6.5207, "step": 18875 }, { "epoch": 0.7958344751986846, "grad_norm": 3.0475246906280518, "learning_rate": 4.081873262750242e-05, "loss": 6.4521, "step": 18876 }, { "epoch": 0.7958766363808841, "grad_norm": 2.506441593170166, "learning_rate": 4.081795184109435e-05, "loss": 6.3828, "step": 18877 }, { "epoch": 0.7959187975630837, "grad_norm": 1.5798202753067017, "learning_rate": 4.0817171054686284e-05, "loss": 6.8004, "step": 18878 }, { "epoch": 0.7959609587452832, "grad_norm": 3.3889191150665283, "learning_rate": 4.081639026827821e-05, "loss": 7.4548, "step": 18879 }, { "epoch": 0.7960031199274827, "grad_norm": 2.5863595008850098, "learning_rate": 4.0815609481870146e-05, "loss": 7.8425, "step": 18880 }, { "epoch": 0.7960452811096823, "grad_norm": 1.8977283239364624, "learning_rate": 4.081482869546207e-05, "loss": 7.0492, "step": 18881 }, { "epoch": 0.7960874422918819, "grad_norm": 1.7812350988388062, "learning_rate": 4.0814047909053995e-05, "loss": 6.5354, "step": 18882 }, { "epoch": 0.7961296034740815, "grad_norm": 2.277252674102783, "learning_rate": 4.081326712264593e-05, "loss": 7.6251, "step": 18883 }, { "epoch": 0.7961717646562809, "grad_norm": 1.938879370689392, "learning_rate": 4.081248633623786e-05, "loss": 6.6112, "step": 18884 }, { "epoch": 0.7962139258384805, "grad_norm": 1.6755666732788086, "learning_rate": 4.0811705549829796e-05, "loss": 6.7613, "step": 18885 }, { "epoch": 0.7962560870206801, "grad_norm": 2.9451825618743896, "learning_rate": 4.081092476342172e-05, "loss": 7.5388, "step": 18886 }, { "epoch": 0.7962982482028796, "grad_norm": 2.0543556213378906, "learning_rate": 4.081014397701365e-05, "loss": 7.3703, "step": 18887 }, { "epoch": 0.7963404093850791, "grad_norm": 2.5114057064056396, "learning_rate": 4.080936319060558e-05, "loss": 7.764, "step": 18888 }, { "epoch": 0.7963825705672787, "grad_norm": 2.158036470413208, "learning_rate": 4.080858240419751e-05, "loss": 6.7188, "step": 18889 }, { "epoch": 0.7964247317494783, "grad_norm": 2.484961986541748, "learning_rate": 4.080780161778944e-05, "loss": 6.837, "step": 18890 }, { "epoch": 0.7964668929316778, "grad_norm": 3.303166151046753, "learning_rate": 4.080702083138137e-05, "loss": 6.5031, "step": 18891 }, { "epoch": 0.7965090541138774, "grad_norm": 3.5058209896087646, "learning_rate": 4.08062400449733e-05, "loss": 7.2556, "step": 18892 }, { "epoch": 0.7965512152960769, "grad_norm": 3.778541088104248, "learning_rate": 4.0805459258565225e-05, "loss": 9.4375, "step": 18893 }, { "epoch": 0.7965933764782764, "grad_norm": 1.7324076890945435, "learning_rate": 4.080467847215716e-05, "loss": 7.0011, "step": 18894 }, { "epoch": 0.796635537660476, "grad_norm": 1.503967046737671, "learning_rate": 4.080389768574909e-05, "loss": 6.6916, "step": 18895 }, { "epoch": 0.7966776988426756, "grad_norm": 1.916428804397583, "learning_rate": 4.080311689934102e-05, "loss": 6.7919, "step": 18896 }, { "epoch": 0.796719860024875, "grad_norm": 2.2414586544036865, "learning_rate": 4.080233611293295e-05, "loss": 6.7487, "step": 18897 }, { "epoch": 0.7967620212070746, "grad_norm": 1.8967194557189941, "learning_rate": 4.0801555326524875e-05, "loss": 6.3886, "step": 18898 }, { "epoch": 0.7968041823892742, "grad_norm": 3.006185531616211, "learning_rate": 4.080077454011681e-05, "loss": 7.597, "step": 18899 }, { "epoch": 0.7968463435714738, "grad_norm": 1.5353097915649414, "learning_rate": 4.079999375370874e-05, "loss": 7.2544, "step": 18900 }, { "epoch": 0.7968885047536733, "grad_norm": 2.482970952987671, "learning_rate": 4.079921296730066e-05, "loss": 7.2402, "step": 18901 }, { "epoch": 0.7969306659358728, "grad_norm": 1.9656765460968018, "learning_rate": 4.07984321808926e-05, "loss": 6.965, "step": 18902 }, { "epoch": 0.7969728271180724, "grad_norm": 2.9306938648223877, "learning_rate": 4.0797651394484524e-05, "loss": 7.6665, "step": 18903 }, { "epoch": 0.797014988300272, "grad_norm": 3.3796005249023438, "learning_rate": 4.0796870608076456e-05, "loss": 6.9927, "step": 18904 }, { "epoch": 0.7970571494824715, "grad_norm": 1.699504017829895, "learning_rate": 4.079608982166839e-05, "loss": 6.4201, "step": 18905 }, { "epoch": 0.797099310664671, "grad_norm": 3.319688081741333, "learning_rate": 4.079530903526032e-05, "loss": 7.2558, "step": 18906 }, { "epoch": 0.7971414718468706, "grad_norm": 2.7005701065063477, "learning_rate": 4.079452824885225e-05, "loss": 7.0104, "step": 18907 }, { "epoch": 0.7971836330290701, "grad_norm": 1.7944835424423218, "learning_rate": 4.0793747462444174e-05, "loss": 6.5218, "step": 18908 }, { "epoch": 0.7972257942112697, "grad_norm": 1.9298032522201538, "learning_rate": 4.0792966676036105e-05, "loss": 6.7818, "step": 18909 }, { "epoch": 0.7972679553934693, "grad_norm": 2.2584660053253174, "learning_rate": 4.0792185889628036e-05, "loss": 6.5764, "step": 18910 }, { "epoch": 0.7973101165756687, "grad_norm": 1.741815209388733, "learning_rate": 4.079140510321997e-05, "loss": 7.531, "step": 18911 }, { "epoch": 0.7973522777578683, "grad_norm": 2.205155849456787, "learning_rate": 4.079062431681189e-05, "loss": 6.8059, "step": 18912 }, { "epoch": 0.7973944389400679, "grad_norm": 1.5751111507415771, "learning_rate": 4.0789843530403823e-05, "loss": 6.6873, "step": 18913 }, { "epoch": 0.7974366001222675, "grad_norm": 4.3172383308410645, "learning_rate": 4.0789062743995755e-05, "loss": 7.9272, "step": 18914 }, { "epoch": 0.7974787613044669, "grad_norm": 2.4278407096862793, "learning_rate": 4.078828195758768e-05, "loss": 7.7152, "step": 18915 }, { "epoch": 0.7975209224866665, "grad_norm": 1.29060959815979, "learning_rate": 4.078750117117962e-05, "loss": 6.6523, "step": 18916 }, { "epoch": 0.7975630836688661, "grad_norm": 1.1070836782455444, "learning_rate": 4.078672038477154e-05, "loss": 6.6638, "step": 18917 }, { "epoch": 0.7976052448510657, "grad_norm": 1.3208585977554321, "learning_rate": 4.078593959836347e-05, "loss": 6.612, "step": 18918 }, { "epoch": 0.7976474060332652, "grad_norm": 1.1589523553848267, "learning_rate": 4.0785158811955404e-05, "loss": 6.4482, "step": 18919 }, { "epoch": 0.7976895672154647, "grad_norm": 2.0786757469177246, "learning_rate": 4.078437802554733e-05, "loss": 6.7276, "step": 18920 }, { "epoch": 0.7977317283976643, "grad_norm": 2.7268426418304443, "learning_rate": 4.078359723913927e-05, "loss": 7.6544, "step": 18921 }, { "epoch": 0.7977738895798638, "grad_norm": 2.184556722640991, "learning_rate": 4.078281645273119e-05, "loss": 7.1734, "step": 18922 }, { "epoch": 0.7978160507620634, "grad_norm": 1.3681485652923584, "learning_rate": 4.078203566632312e-05, "loss": 6.9393, "step": 18923 }, { "epoch": 0.7978582119442629, "grad_norm": 1.4055489301681519, "learning_rate": 4.0781254879915054e-05, "loss": 6.8733, "step": 18924 }, { "epoch": 0.7979003731264624, "grad_norm": 1.7195287942886353, "learning_rate": 4.0780474093506985e-05, "loss": 6.5084, "step": 18925 }, { "epoch": 0.797942534308662, "grad_norm": 1.8323782682418823, "learning_rate": 4.077969330709891e-05, "loss": 6.5819, "step": 18926 }, { "epoch": 0.7979846954908616, "grad_norm": 0.8008480072021484, "learning_rate": 4.077891252069084e-05, "loss": 6.5833, "step": 18927 }, { "epoch": 0.7980268566730611, "grad_norm": 1.5023250579833984, "learning_rate": 4.077813173428277e-05, "loss": 6.4981, "step": 18928 }, { "epoch": 0.7980690178552606, "grad_norm": 4.252406597137451, "learning_rate": 4.07773509478747e-05, "loss": 7.7162, "step": 18929 }, { "epoch": 0.7981111790374602, "grad_norm": 1.6662172079086304, "learning_rate": 4.0776570161466635e-05, "loss": 6.4671, "step": 18930 }, { "epoch": 0.7981533402196598, "grad_norm": 2.3352348804473877, "learning_rate": 4.077578937505856e-05, "loss": 7.71, "step": 18931 }, { "epoch": 0.7981955014018594, "grad_norm": 0.8201723694801331, "learning_rate": 4.077500858865049e-05, "loss": 6.4987, "step": 18932 }, { "epoch": 0.7982376625840588, "grad_norm": 2.2134838104248047, "learning_rate": 4.077422780224242e-05, "loss": 6.9495, "step": 18933 }, { "epoch": 0.7982798237662584, "grad_norm": 1.8659948110580444, "learning_rate": 4.0773447015834346e-05, "loss": 6.6458, "step": 18934 }, { "epoch": 0.798321984948458, "grad_norm": 3.423013925552368, "learning_rate": 4.0772666229426284e-05, "loss": 7.3219, "step": 18935 }, { "epoch": 0.7983641461306575, "grad_norm": 3.210784912109375, "learning_rate": 4.077188544301821e-05, "loss": 6.51, "step": 18936 }, { "epoch": 0.798406307312857, "grad_norm": 1.7131059169769287, "learning_rate": 4.077110465661014e-05, "loss": 6.6877, "step": 18937 }, { "epoch": 0.7984484684950566, "grad_norm": 3.0204577445983887, "learning_rate": 4.077032387020207e-05, "loss": 7.5767, "step": 18938 }, { "epoch": 0.7984906296772561, "grad_norm": 0.9323200583457947, "learning_rate": 4.0769543083793996e-05, "loss": 6.5079, "step": 18939 }, { "epoch": 0.7985327908594557, "grad_norm": 2.085444927215576, "learning_rate": 4.0768762297385934e-05, "loss": 6.9609, "step": 18940 }, { "epoch": 0.7985749520416553, "grad_norm": 0.8888521790504456, "learning_rate": 4.076798151097786e-05, "loss": 6.3906, "step": 18941 }, { "epoch": 0.7986171132238548, "grad_norm": 1.576478362083435, "learning_rate": 4.076720072456979e-05, "loss": 6.5778, "step": 18942 }, { "epoch": 0.7986592744060543, "grad_norm": 3.35760760307312, "learning_rate": 4.076641993816172e-05, "loss": 7.3238, "step": 18943 }, { "epoch": 0.7987014355882539, "grad_norm": 0.8964253067970276, "learning_rate": 4.0765639151753645e-05, "loss": 6.4448, "step": 18944 }, { "epoch": 0.7987435967704535, "grad_norm": 2.7348134517669678, "learning_rate": 4.0764858365345576e-05, "loss": 6.5708, "step": 18945 }, { "epoch": 0.7987857579526529, "grad_norm": 1.469983696937561, "learning_rate": 4.076407757893751e-05, "loss": 6.9023, "step": 18946 }, { "epoch": 0.7988279191348525, "grad_norm": 0.9816092252731323, "learning_rate": 4.076329679252944e-05, "loss": 6.5434, "step": 18947 }, { "epoch": 0.7988700803170521, "grad_norm": 4.372531890869141, "learning_rate": 4.076251600612136e-05, "loss": 8.239, "step": 18948 }, { "epoch": 0.7989122414992517, "grad_norm": 1.6465601921081543, "learning_rate": 4.07617352197133e-05, "loss": 7.2283, "step": 18949 }, { "epoch": 0.7989544026814512, "grad_norm": 2.211164712905884, "learning_rate": 4.0760954433305226e-05, "loss": 6.4413, "step": 18950 }, { "epoch": 0.7989965638636507, "grad_norm": 1.5826481580734253, "learning_rate": 4.076017364689716e-05, "loss": 6.3917, "step": 18951 }, { "epoch": 0.7990387250458503, "grad_norm": 1.6111304759979248, "learning_rate": 4.075939286048909e-05, "loss": 6.6695, "step": 18952 }, { "epoch": 0.7990808862280498, "grad_norm": 1.4728232622146606, "learning_rate": 4.075861207408101e-05, "loss": 6.5362, "step": 18953 }, { "epoch": 0.7991230474102494, "grad_norm": 2.5629138946533203, "learning_rate": 4.075783128767295e-05, "loss": 7.9244, "step": 18954 }, { "epoch": 0.7991652085924489, "grad_norm": 1.1726826429367065, "learning_rate": 4.0757050501264875e-05, "loss": 7.1086, "step": 18955 }, { "epoch": 0.7992073697746485, "grad_norm": 4.244454383850098, "learning_rate": 4.075626971485681e-05, "loss": 7.3822, "step": 18956 }, { "epoch": 0.799249530956848, "grad_norm": 0.7627723813056946, "learning_rate": 4.075548892844874e-05, "loss": 6.3977, "step": 18957 }, { "epoch": 0.7992916921390476, "grad_norm": 1.9169083833694458, "learning_rate": 4.075470814204066e-05, "loss": 7.0986, "step": 18958 }, { "epoch": 0.7993338533212472, "grad_norm": 2.3266470432281494, "learning_rate": 4.0753927355632594e-05, "loss": 7.9355, "step": 18959 }, { "epoch": 0.7993760145034466, "grad_norm": 2.129425287246704, "learning_rate": 4.0753146569224525e-05, "loss": 6.4754, "step": 18960 }, { "epoch": 0.7994181756856462, "grad_norm": 1.0936814546585083, "learning_rate": 4.0752365782816456e-05, "loss": 6.8848, "step": 18961 }, { "epoch": 0.7994603368678458, "grad_norm": 1.0828540325164795, "learning_rate": 4.075158499640838e-05, "loss": 6.9619, "step": 18962 }, { "epoch": 0.7995024980500454, "grad_norm": 1.713151454925537, "learning_rate": 4.075080421000031e-05, "loss": 6.4221, "step": 18963 }, { "epoch": 0.7995446592322448, "grad_norm": 1.6988695859909058, "learning_rate": 4.075002342359224e-05, "loss": 6.741, "step": 18964 }, { "epoch": 0.7995868204144444, "grad_norm": 1.4043285846710205, "learning_rate": 4.0749242637184174e-05, "loss": 6.5095, "step": 18965 }, { "epoch": 0.799628981596644, "grad_norm": 1.7109473943710327, "learning_rate": 4.0748461850776106e-05, "loss": 6.4428, "step": 18966 }, { "epoch": 0.7996711427788435, "grad_norm": 1.2716312408447266, "learning_rate": 4.074768106436803e-05, "loss": 6.6633, "step": 18967 }, { "epoch": 0.7997133039610431, "grad_norm": 3.249776601791382, "learning_rate": 4.074690027795997e-05, "loss": 7.4023, "step": 18968 }, { "epoch": 0.7997554651432426, "grad_norm": 1.2866830825805664, "learning_rate": 4.074611949155189e-05, "loss": 6.7287, "step": 18969 }, { "epoch": 0.7997976263254422, "grad_norm": 1.7714390754699707, "learning_rate": 4.0745338705143824e-05, "loss": 6.6723, "step": 18970 }, { "epoch": 0.7998397875076417, "grad_norm": 1.1558128595352173, "learning_rate": 4.0744557918735755e-05, "loss": 6.4316, "step": 18971 }, { "epoch": 0.7998819486898413, "grad_norm": 1.306443691253662, "learning_rate": 4.074377713232768e-05, "loss": 7.0243, "step": 18972 }, { "epoch": 0.7999241098720408, "grad_norm": 3.079464912414551, "learning_rate": 4.074299634591961e-05, "loss": 6.7125, "step": 18973 }, { "epoch": 0.7999662710542403, "grad_norm": 3.0490927696228027, "learning_rate": 4.074221555951154e-05, "loss": 8.2495, "step": 18974 }, { "epoch": 0.8000084322364399, "grad_norm": 2.966771364212036, "learning_rate": 4.0741434773103473e-05, "loss": 6.5889, "step": 18975 }, { "epoch": 0.8000505934186395, "grad_norm": 2.084115505218506, "learning_rate": 4.0740653986695405e-05, "loss": 6.7493, "step": 18976 }, { "epoch": 0.8000927546008391, "grad_norm": 2.440051317214966, "learning_rate": 4.073987320028733e-05, "loss": 6.6532, "step": 18977 }, { "epoch": 0.8001349157830385, "grad_norm": 5.5604658126831055, "learning_rate": 4.073909241387926e-05, "loss": 11.3345, "step": 18978 }, { "epoch": 0.8001770769652381, "grad_norm": 4.60026741027832, "learning_rate": 4.073831162747119e-05, "loss": 9.1678, "step": 18979 }, { "epoch": 0.8002192381474377, "grad_norm": 1.544248342514038, "learning_rate": 4.073753084106312e-05, "loss": 6.5067, "step": 18980 }, { "epoch": 0.8002613993296372, "grad_norm": 3.5038554668426514, "learning_rate": 4.073675005465505e-05, "loss": 6.9929, "step": 18981 }, { "epoch": 0.8003035605118367, "grad_norm": 2.7984654903411865, "learning_rate": 4.073596926824698e-05, "loss": 7.5043, "step": 18982 }, { "epoch": 0.8003457216940363, "grad_norm": 2.0602424144744873, "learning_rate": 4.073518848183891e-05, "loss": 7.498, "step": 18983 }, { "epoch": 0.8003878828762359, "grad_norm": 1.5451316833496094, "learning_rate": 4.0734407695430834e-05, "loss": 6.7515, "step": 18984 }, { "epoch": 0.8004300440584354, "grad_norm": 3.0360467433929443, "learning_rate": 4.073362690902277e-05, "loss": 6.6666, "step": 18985 }, { "epoch": 0.8004722052406349, "grad_norm": 1.5211176872253418, "learning_rate": 4.07328461226147e-05, "loss": 6.8742, "step": 18986 }, { "epoch": 0.8005143664228345, "grad_norm": 1.4797285795211792, "learning_rate": 4.0732065336206635e-05, "loss": 6.5936, "step": 18987 }, { "epoch": 0.800556527605034, "grad_norm": 1.2069059610366821, "learning_rate": 4.073128454979856e-05, "loss": 6.7079, "step": 18988 }, { "epoch": 0.8005986887872336, "grad_norm": 1.9323155879974365, "learning_rate": 4.0730503763390484e-05, "loss": 6.816, "step": 18989 }, { "epoch": 0.8006408499694332, "grad_norm": 3.491884708404541, "learning_rate": 4.072972297698242e-05, "loss": 8.1, "step": 18990 }, { "epoch": 0.8006830111516327, "grad_norm": 1.6160616874694824, "learning_rate": 4.0728942190574347e-05, "loss": 6.5111, "step": 18991 }, { "epoch": 0.8007251723338322, "grad_norm": 1.2739059925079346, "learning_rate": 4.072816140416628e-05, "loss": 6.7262, "step": 18992 }, { "epoch": 0.8007673335160318, "grad_norm": 1.180938959121704, "learning_rate": 4.072738061775821e-05, "loss": 6.5419, "step": 18993 }, { "epoch": 0.8008094946982314, "grad_norm": 2.9560887813568115, "learning_rate": 4.072659983135014e-05, "loss": 7.9957, "step": 18994 }, { "epoch": 0.8008516558804308, "grad_norm": 1.1299312114715576, "learning_rate": 4.0725819044942065e-05, "loss": 7.2115, "step": 18995 }, { "epoch": 0.8008938170626304, "grad_norm": 1.250450849533081, "learning_rate": 4.0725038258533996e-05, "loss": 6.4766, "step": 18996 }, { "epoch": 0.80093597824483, "grad_norm": 1.1846578121185303, "learning_rate": 4.072425747212593e-05, "loss": 6.6534, "step": 18997 }, { "epoch": 0.8009781394270296, "grad_norm": 3.225843906402588, "learning_rate": 4.072347668571786e-05, "loss": 6.5313, "step": 18998 }, { "epoch": 0.8010203006092291, "grad_norm": 1.690920114517212, "learning_rate": 4.072269589930979e-05, "loss": 6.9787, "step": 18999 }, { "epoch": 0.8010624617914286, "grad_norm": 1.9215625524520874, "learning_rate": 4.0721915112901714e-05, "loss": 6.9304, "step": 19000 }, { "epoch": 0.8011046229736282, "grad_norm": 3.239227294921875, "learning_rate": 4.0721134326493646e-05, "loss": 7.7899, "step": 19001 }, { "epoch": 0.8011467841558277, "grad_norm": 0.9571061730384827, "learning_rate": 4.072035354008558e-05, "loss": 6.7854, "step": 19002 }, { "epoch": 0.8011889453380273, "grad_norm": 2.6340744495391846, "learning_rate": 4.07195727536775e-05, "loss": 7.7252, "step": 19003 }, { "epoch": 0.8012311065202268, "grad_norm": 1.1733285188674927, "learning_rate": 4.071879196726944e-05, "loss": 6.805, "step": 19004 }, { "epoch": 0.8012732677024264, "grad_norm": 3.1387076377868652, "learning_rate": 4.0718011180861364e-05, "loss": 6.4026, "step": 19005 }, { "epoch": 0.8013154288846259, "grad_norm": 1.5933655500411987, "learning_rate": 4.0717230394453295e-05, "loss": 6.4096, "step": 19006 }, { "epoch": 0.8013575900668255, "grad_norm": 2.317206621170044, "learning_rate": 4.0716449608045226e-05, "loss": 7.5004, "step": 19007 }, { "epoch": 0.8013997512490251, "grad_norm": 2.436410903930664, "learning_rate": 4.071566882163715e-05, "loss": 6.8096, "step": 19008 }, { "epoch": 0.8014419124312245, "grad_norm": 1.0515276193618774, "learning_rate": 4.071488803522909e-05, "loss": 6.4347, "step": 19009 }, { "epoch": 0.8014840736134241, "grad_norm": 0.9286938309669495, "learning_rate": 4.071410724882101e-05, "loss": 6.4944, "step": 19010 }, { "epoch": 0.8015262347956237, "grad_norm": 1.7720659971237183, "learning_rate": 4.0713326462412945e-05, "loss": 6.9559, "step": 19011 }, { "epoch": 0.8015683959778233, "grad_norm": 0.7906806468963623, "learning_rate": 4.0712545676004876e-05, "loss": 6.4153, "step": 19012 }, { "epoch": 0.8016105571600227, "grad_norm": 4.311534881591797, "learning_rate": 4.071176488959681e-05, "loss": 7.7913, "step": 19013 }, { "epoch": 0.8016527183422223, "grad_norm": 2.5261783599853516, "learning_rate": 4.071098410318873e-05, "loss": 7.1328, "step": 19014 }, { "epoch": 0.8016948795244219, "grad_norm": 1.4330116510391235, "learning_rate": 4.071020331678066e-05, "loss": 6.9632, "step": 19015 }, { "epoch": 0.8017370407066214, "grad_norm": 1.4687838554382324, "learning_rate": 4.0709422530372594e-05, "loss": 6.6774, "step": 19016 }, { "epoch": 0.801779201888821, "grad_norm": 3.958674430847168, "learning_rate": 4.070864174396452e-05, "loss": 7.9096, "step": 19017 }, { "epoch": 0.8018213630710205, "grad_norm": 1.6374523639678955, "learning_rate": 4.070786095755646e-05, "loss": 6.3098, "step": 19018 }, { "epoch": 0.80186352425322, "grad_norm": 1.1767022609710693, "learning_rate": 4.070708017114838e-05, "loss": 6.5182, "step": 19019 }, { "epoch": 0.8019056854354196, "grad_norm": 1.641602635383606, "learning_rate": 4.070629938474031e-05, "loss": 7.0361, "step": 19020 }, { "epoch": 0.8019478466176192, "grad_norm": 2.738779067993164, "learning_rate": 4.0705518598332244e-05, "loss": 7.0635, "step": 19021 }, { "epoch": 0.8019900077998187, "grad_norm": 1.7392442226409912, "learning_rate": 4.070473781192417e-05, "loss": 6.6726, "step": 19022 }, { "epoch": 0.8020321689820182, "grad_norm": 3.6822431087493896, "learning_rate": 4.0703957025516106e-05, "loss": 7.7902, "step": 19023 }, { "epoch": 0.8020743301642178, "grad_norm": 1.7596663236618042, "learning_rate": 4.070317623910803e-05, "loss": 7.1611, "step": 19024 }, { "epoch": 0.8021164913464174, "grad_norm": 3.141781806945801, "learning_rate": 4.070239545269996e-05, "loss": 7.8012, "step": 19025 }, { "epoch": 0.802158652528617, "grad_norm": 2.391166925430298, "learning_rate": 4.070161466629189e-05, "loss": 6.6304, "step": 19026 }, { "epoch": 0.8022008137108164, "grad_norm": 2.8036763668060303, "learning_rate": 4.070083387988382e-05, "loss": 6.956, "step": 19027 }, { "epoch": 0.802242974893016, "grad_norm": 2.2984273433685303, "learning_rate": 4.070005309347575e-05, "loss": 7.1086, "step": 19028 }, { "epoch": 0.8022851360752156, "grad_norm": 2.194112539291382, "learning_rate": 4.069927230706768e-05, "loss": 6.5105, "step": 19029 }, { "epoch": 0.8023272972574151, "grad_norm": 2.8382363319396973, "learning_rate": 4.069849152065961e-05, "loss": 7.5506, "step": 19030 }, { "epoch": 0.8023694584396146, "grad_norm": 2.807671546936035, "learning_rate": 4.0697710734251536e-05, "loss": 6.4354, "step": 19031 }, { "epoch": 0.8024116196218142, "grad_norm": 1.4425806999206543, "learning_rate": 4.0696929947843474e-05, "loss": 6.562, "step": 19032 }, { "epoch": 0.8024537808040137, "grad_norm": 1.6863601207733154, "learning_rate": 4.06961491614354e-05, "loss": 6.6455, "step": 19033 }, { "epoch": 0.8024959419862133, "grad_norm": 2.443675994873047, "learning_rate": 4.069536837502733e-05, "loss": 7.3719, "step": 19034 }, { "epoch": 0.8025381031684129, "grad_norm": 1.6407207250595093, "learning_rate": 4.069458758861926e-05, "loss": 7.0126, "step": 19035 }, { "epoch": 0.8025802643506124, "grad_norm": 1.5720317363739014, "learning_rate": 4.0693806802211185e-05, "loss": 6.6082, "step": 19036 }, { "epoch": 0.8026224255328119, "grad_norm": 2.8525443077087402, "learning_rate": 4.0693026015803123e-05, "loss": 7.5807, "step": 19037 }, { "epoch": 0.8026645867150115, "grad_norm": 2.0935089588165283, "learning_rate": 4.069224522939505e-05, "loss": 7.525, "step": 19038 }, { "epoch": 0.8027067478972111, "grad_norm": 1.7937756776809692, "learning_rate": 4.069146444298698e-05, "loss": 7.0179, "step": 19039 }, { "epoch": 0.8027489090794105, "grad_norm": 2.1189541816711426, "learning_rate": 4.069068365657891e-05, "loss": 6.4777, "step": 19040 }, { "epoch": 0.8027910702616101, "grad_norm": 1.321091890335083, "learning_rate": 4.0689902870170835e-05, "loss": 7.1322, "step": 19041 }, { "epoch": 0.8028332314438097, "grad_norm": 1.8505443334579468, "learning_rate": 4.0689122083762766e-05, "loss": 6.9771, "step": 19042 }, { "epoch": 0.8028753926260093, "grad_norm": 2.9608535766601562, "learning_rate": 4.06883412973547e-05, "loss": 7.7788, "step": 19043 }, { "epoch": 0.8029175538082087, "grad_norm": 2.881552219390869, "learning_rate": 4.068756051094663e-05, "loss": 7.7198, "step": 19044 }, { "epoch": 0.8029597149904083, "grad_norm": 1.892065167427063, "learning_rate": 4.068677972453856e-05, "loss": 6.3565, "step": 19045 }, { "epoch": 0.8030018761726079, "grad_norm": 1.9032847881317139, "learning_rate": 4.0685998938130485e-05, "loss": 7.4335, "step": 19046 }, { "epoch": 0.8030440373548074, "grad_norm": 3.223677396774292, "learning_rate": 4.0685218151722416e-05, "loss": 6.8568, "step": 19047 }, { "epoch": 0.803086198537007, "grad_norm": 3.1992428302764893, "learning_rate": 4.068443736531435e-05, "loss": 6.524, "step": 19048 }, { "epoch": 0.8031283597192065, "grad_norm": 2.575714349746704, "learning_rate": 4.068365657890628e-05, "loss": 7.2921, "step": 19049 }, { "epoch": 0.8031705209014061, "grad_norm": 2.2760305404663086, "learning_rate": 4.06828757924982e-05, "loss": 6.7525, "step": 19050 }, { "epoch": 0.8032126820836056, "grad_norm": 1.3204880952835083, "learning_rate": 4.0682095006090134e-05, "loss": 6.5054, "step": 19051 }, { "epoch": 0.8032548432658052, "grad_norm": 0.7927941083908081, "learning_rate": 4.0681314219682065e-05, "loss": 6.3696, "step": 19052 }, { "epoch": 0.8032970044480047, "grad_norm": 2.0804543495178223, "learning_rate": 4.068053343327399e-05, "loss": 6.6678, "step": 19053 }, { "epoch": 0.8033391656302042, "grad_norm": 2.163120746612549, "learning_rate": 4.067975264686593e-05, "loss": 6.4946, "step": 19054 }, { "epoch": 0.8033813268124038, "grad_norm": 2.1761248111724854, "learning_rate": 4.067897186045785e-05, "loss": 6.5525, "step": 19055 }, { "epoch": 0.8034234879946034, "grad_norm": 1.7419605255126953, "learning_rate": 4.067819107404979e-05, "loss": 6.8265, "step": 19056 }, { "epoch": 0.803465649176803, "grad_norm": 1.9696816205978394, "learning_rate": 4.0677410287641715e-05, "loss": 7.0647, "step": 19057 }, { "epoch": 0.8035078103590024, "grad_norm": 2.9314966201782227, "learning_rate": 4.0676629501233646e-05, "loss": 7.7298, "step": 19058 }, { "epoch": 0.803549971541202, "grad_norm": 1.9235894680023193, "learning_rate": 4.067584871482558e-05, "loss": 6.7344, "step": 19059 }, { "epoch": 0.8035921327234016, "grad_norm": 2.4374382495880127, "learning_rate": 4.06750679284175e-05, "loss": 6.5426, "step": 19060 }, { "epoch": 0.8036342939056011, "grad_norm": 4.216892242431641, "learning_rate": 4.067428714200943e-05, "loss": 7.4275, "step": 19061 }, { "epoch": 0.8036764550878006, "grad_norm": 1.2682673931121826, "learning_rate": 4.0673506355601364e-05, "loss": 6.3934, "step": 19062 }, { "epoch": 0.8037186162700002, "grad_norm": 1.1402215957641602, "learning_rate": 4.0672725569193296e-05, "loss": 6.5474, "step": 19063 }, { "epoch": 0.8037607774521998, "grad_norm": 2.1902167797088623, "learning_rate": 4.067194478278522e-05, "loss": 7.0448, "step": 19064 }, { "epoch": 0.8038029386343993, "grad_norm": 3.1837849617004395, "learning_rate": 4.067116399637715e-05, "loss": 7.6045, "step": 19065 }, { "epoch": 0.8038450998165989, "grad_norm": 3.4400901794433594, "learning_rate": 4.067038320996908e-05, "loss": 6.5289, "step": 19066 }, { "epoch": 0.8038872609987984, "grad_norm": 1.5510976314544678, "learning_rate": 4.0669602423561014e-05, "loss": 6.3998, "step": 19067 }, { "epoch": 0.8039294221809979, "grad_norm": 1.6954365968704224, "learning_rate": 4.0668821637152945e-05, "loss": 7.1849, "step": 19068 }, { "epoch": 0.8039715833631975, "grad_norm": 4.463688850402832, "learning_rate": 4.066804085074487e-05, "loss": 7.1459, "step": 19069 }, { "epoch": 0.8040137445453971, "grad_norm": 0.8992925882339478, "learning_rate": 4.06672600643368e-05, "loss": 6.7461, "step": 19070 }, { "epoch": 0.8040559057275966, "grad_norm": 1.9134175777435303, "learning_rate": 4.066647927792873e-05, "loss": 7.3085, "step": 19071 }, { "epoch": 0.8040980669097961, "grad_norm": 1.1644645929336548, "learning_rate": 4.0665698491520657e-05, "loss": 6.5012, "step": 19072 }, { "epoch": 0.8041402280919957, "grad_norm": 0.808515727519989, "learning_rate": 4.0664917705112595e-05, "loss": 6.6889, "step": 19073 }, { "epoch": 0.8041823892741953, "grad_norm": 1.1347445249557495, "learning_rate": 4.066413691870452e-05, "loss": 6.4794, "step": 19074 }, { "epoch": 0.8042245504563948, "grad_norm": 1.4117271900177002, "learning_rate": 4.066335613229645e-05, "loss": 7.0167, "step": 19075 }, { "epoch": 0.8042667116385943, "grad_norm": 1.6408199071884155, "learning_rate": 4.066257534588838e-05, "loss": 6.3999, "step": 19076 }, { "epoch": 0.8043088728207939, "grad_norm": 1.0751914978027344, "learning_rate": 4.066179455948031e-05, "loss": 6.4978, "step": 19077 }, { "epoch": 0.8043510340029935, "grad_norm": 2.8943097591400146, "learning_rate": 4.0661013773072244e-05, "loss": 7.5198, "step": 19078 }, { "epoch": 0.804393195185193, "grad_norm": 2.0322272777557373, "learning_rate": 4.066023298666417e-05, "loss": 6.8896, "step": 19079 }, { "epoch": 0.8044353563673925, "grad_norm": 2.50716233253479, "learning_rate": 4.06594522002561e-05, "loss": 6.5327, "step": 19080 }, { "epoch": 0.8044775175495921, "grad_norm": 2.0854647159576416, "learning_rate": 4.065867141384803e-05, "loss": 6.9642, "step": 19081 }, { "epoch": 0.8045196787317916, "grad_norm": 1.851384162902832, "learning_rate": 4.065789062743996e-05, "loss": 7.0397, "step": 19082 }, { "epoch": 0.8045618399139912, "grad_norm": 2.2958638668060303, "learning_rate": 4.065710984103189e-05, "loss": 6.615, "step": 19083 }, { "epoch": 0.8046040010961908, "grad_norm": 1.0473906993865967, "learning_rate": 4.065632905462382e-05, "loss": 6.6144, "step": 19084 }, { "epoch": 0.8046461622783903, "grad_norm": 0.9498076438903809, "learning_rate": 4.065554826821575e-05, "loss": 6.4381, "step": 19085 }, { "epoch": 0.8046883234605898, "grad_norm": 3.1612091064453125, "learning_rate": 4.0654767481807674e-05, "loss": 6.9583, "step": 19086 }, { "epoch": 0.8047304846427894, "grad_norm": 2.7637431621551514, "learning_rate": 4.065398669539961e-05, "loss": 7.2171, "step": 19087 }, { "epoch": 0.804772645824989, "grad_norm": 1.1695959568023682, "learning_rate": 4.0653205908991536e-05, "loss": 6.5692, "step": 19088 }, { "epoch": 0.8048148070071884, "grad_norm": 1.0813086032867432, "learning_rate": 4.065242512258347e-05, "loss": 6.5687, "step": 19089 }, { "epoch": 0.804856968189388, "grad_norm": 2.45821213722229, "learning_rate": 4.06516443361754e-05, "loss": 6.8776, "step": 19090 }, { "epoch": 0.8048991293715876, "grad_norm": 2.6067235469818115, "learning_rate": 4.0650863549767323e-05, "loss": 6.9123, "step": 19091 }, { "epoch": 0.8049412905537872, "grad_norm": 3.4667787551879883, "learning_rate": 4.065008276335926e-05, "loss": 7.8695, "step": 19092 }, { "epoch": 0.8049834517359867, "grad_norm": 1.76341712474823, "learning_rate": 4.0649301976951186e-05, "loss": 6.8122, "step": 19093 }, { "epoch": 0.8050256129181862, "grad_norm": 2.769301652908325, "learning_rate": 4.064852119054312e-05, "loss": 7.3429, "step": 19094 }, { "epoch": 0.8050677741003858, "grad_norm": 3.917915105819702, "learning_rate": 4.064774040413505e-05, "loss": 7.5267, "step": 19095 }, { "epoch": 0.8051099352825853, "grad_norm": 1.0557658672332764, "learning_rate": 4.064695961772697e-05, "loss": 6.5042, "step": 19096 }, { "epoch": 0.8051520964647849, "grad_norm": 3.0504837036132812, "learning_rate": 4.0646178831318904e-05, "loss": 7.6612, "step": 19097 }, { "epoch": 0.8051942576469844, "grad_norm": 1.9783940315246582, "learning_rate": 4.0645398044910835e-05, "loss": 6.4605, "step": 19098 }, { "epoch": 0.805236418829184, "grad_norm": 3.9608616828918457, "learning_rate": 4.064461725850277e-05, "loss": 7.4274, "step": 19099 }, { "epoch": 0.8052785800113835, "grad_norm": 1.7527586221694946, "learning_rate": 4.06438364720947e-05, "loss": 7.1301, "step": 19100 }, { "epoch": 0.8053207411935831, "grad_norm": 1.6285717487335205, "learning_rate": 4.064305568568663e-05, "loss": 6.9915, "step": 19101 }, { "epoch": 0.8053629023757826, "grad_norm": 2.0113024711608887, "learning_rate": 4.0642274899278554e-05, "loss": 7.0968, "step": 19102 }, { "epoch": 0.8054050635579821, "grad_norm": 2.2535998821258545, "learning_rate": 4.0641494112870485e-05, "loss": 7.5314, "step": 19103 }, { "epoch": 0.8054472247401817, "grad_norm": 1.8047443628311157, "learning_rate": 4.0640713326462416e-05, "loss": 6.9742, "step": 19104 }, { "epoch": 0.8054893859223813, "grad_norm": 1.693264126777649, "learning_rate": 4.063993254005434e-05, "loss": 6.9189, "step": 19105 }, { "epoch": 0.8055315471045809, "grad_norm": 1.1425005197525024, "learning_rate": 4.063915175364628e-05, "loss": 6.5163, "step": 19106 }, { "epoch": 0.8055737082867803, "grad_norm": 2.489546298980713, "learning_rate": 4.06383709672382e-05, "loss": 7.0839, "step": 19107 }, { "epoch": 0.8056158694689799, "grad_norm": 1.4063602685928345, "learning_rate": 4.0637590180830135e-05, "loss": 6.372, "step": 19108 }, { "epoch": 0.8056580306511795, "grad_norm": 0.9017148017883301, "learning_rate": 4.0636809394422066e-05, "loss": 6.4685, "step": 19109 }, { "epoch": 0.805700191833379, "grad_norm": 1.7007755041122437, "learning_rate": 4.063602860801399e-05, "loss": 6.4514, "step": 19110 }, { "epoch": 0.8057423530155785, "grad_norm": 1.3229542970657349, "learning_rate": 4.063524782160592e-05, "loss": 7.0438, "step": 19111 }, { "epoch": 0.8057845141977781, "grad_norm": 1.7126405239105225, "learning_rate": 4.063446703519785e-05, "loss": 6.7991, "step": 19112 }, { "epoch": 0.8058266753799777, "grad_norm": 1.34873366355896, "learning_rate": 4.0633686248789784e-05, "loss": 7.2117, "step": 19113 }, { "epoch": 0.8058688365621772, "grad_norm": 1.1180415153503418, "learning_rate": 4.0632905462381715e-05, "loss": 6.4793, "step": 19114 }, { "epoch": 0.8059109977443768, "grad_norm": 1.0317472219467163, "learning_rate": 4.063212467597364e-05, "loss": 6.7879, "step": 19115 }, { "epoch": 0.8059531589265763, "grad_norm": 0.8365823030471802, "learning_rate": 4.063134388956557e-05, "loss": 6.607, "step": 19116 }, { "epoch": 0.8059953201087758, "grad_norm": 0.7727459073066711, "learning_rate": 4.06305631031575e-05, "loss": 6.631, "step": 19117 }, { "epoch": 0.8060374812909754, "grad_norm": 3.3675901889801025, "learning_rate": 4.0629782316749434e-05, "loss": 7.7308, "step": 19118 }, { "epoch": 0.806079642473175, "grad_norm": 1.4298452138900757, "learning_rate": 4.062900153034136e-05, "loss": 6.8475, "step": 19119 }, { "epoch": 0.8061218036553744, "grad_norm": 1.4667898416519165, "learning_rate": 4.0628220743933296e-05, "loss": 6.7591, "step": 19120 }, { "epoch": 0.806163964837574, "grad_norm": 1.1644641160964966, "learning_rate": 4.062743995752522e-05, "loss": 6.8241, "step": 19121 }, { "epoch": 0.8062061260197736, "grad_norm": 1.1586964130401611, "learning_rate": 4.062665917111715e-05, "loss": 6.6558, "step": 19122 }, { "epoch": 0.8062482872019732, "grad_norm": 1.0752313137054443, "learning_rate": 4.062587838470908e-05, "loss": 6.6251, "step": 19123 }, { "epoch": 0.8062904483841727, "grad_norm": 1.1244617700576782, "learning_rate": 4.062509759830101e-05, "loss": 6.7838, "step": 19124 }, { "epoch": 0.8063326095663722, "grad_norm": 1.0620313882827759, "learning_rate": 4.0624316811892946e-05, "loss": 6.4479, "step": 19125 }, { "epoch": 0.8063747707485718, "grad_norm": 2.417966842651367, "learning_rate": 4.062353602548487e-05, "loss": 7.6625, "step": 19126 }, { "epoch": 0.8064169319307714, "grad_norm": 1.3789442777633667, "learning_rate": 4.06227552390768e-05, "loss": 6.9996, "step": 19127 }, { "epoch": 0.8064590931129709, "grad_norm": 2.4251716136932373, "learning_rate": 4.062197445266873e-05, "loss": 7.3105, "step": 19128 }, { "epoch": 0.8065012542951704, "grad_norm": 2.2022969722747803, "learning_rate": 4.062119366626066e-05, "loss": 7.0452, "step": 19129 }, { "epoch": 0.80654341547737, "grad_norm": 1.7197892665863037, "learning_rate": 4.062041287985259e-05, "loss": 7.9362, "step": 19130 }, { "epoch": 0.8065855766595695, "grad_norm": 3.172001838684082, "learning_rate": 4.061963209344452e-05, "loss": 6.5531, "step": 19131 }, { "epoch": 0.8066277378417691, "grad_norm": 1.7093119621276855, "learning_rate": 4.061885130703645e-05, "loss": 6.7256, "step": 19132 }, { "epoch": 0.8066698990239687, "grad_norm": 3.607388496398926, "learning_rate": 4.0618070520628375e-05, "loss": 7.9391, "step": 19133 }, { "epoch": 0.8067120602061681, "grad_norm": 1.2643511295318604, "learning_rate": 4.061728973422031e-05, "loss": 6.9388, "step": 19134 }, { "epoch": 0.8067542213883677, "grad_norm": 1.3791388273239136, "learning_rate": 4.061650894781224e-05, "loss": 6.42, "step": 19135 }, { "epoch": 0.8067963825705673, "grad_norm": 3.5250344276428223, "learning_rate": 4.061572816140417e-05, "loss": 7.6909, "step": 19136 }, { "epoch": 0.8068385437527669, "grad_norm": 1.1308449506759644, "learning_rate": 4.06149473749961e-05, "loss": 6.6069, "step": 19137 }, { "epoch": 0.8068807049349663, "grad_norm": 2.484856367111206, "learning_rate": 4.0614166588588025e-05, "loss": 7.8788, "step": 19138 }, { "epoch": 0.8069228661171659, "grad_norm": 1.1169171333312988, "learning_rate": 4.061338580217996e-05, "loss": 6.5367, "step": 19139 }, { "epoch": 0.8069650272993655, "grad_norm": 1.3836382627487183, "learning_rate": 4.061260501577189e-05, "loss": 6.8778, "step": 19140 }, { "epoch": 0.807007188481565, "grad_norm": 1.356034278869629, "learning_rate": 4.061182422936381e-05, "loss": 6.7952, "step": 19141 }, { "epoch": 0.8070493496637646, "grad_norm": 3.7619214057922363, "learning_rate": 4.061104344295575e-05, "loss": 9.458, "step": 19142 }, { "epoch": 0.8070915108459641, "grad_norm": 1.3203284740447998, "learning_rate": 4.0610262656547674e-05, "loss": 7.2233, "step": 19143 }, { "epoch": 0.8071336720281637, "grad_norm": 2.820873975753784, "learning_rate": 4.0609481870139606e-05, "loss": 7.7235, "step": 19144 }, { "epoch": 0.8071758332103632, "grad_norm": 1.4031238555908203, "learning_rate": 4.060870108373154e-05, "loss": 6.9848, "step": 19145 }, { "epoch": 0.8072179943925628, "grad_norm": 2.8889834880828857, "learning_rate": 4.060792029732347e-05, "loss": 6.615, "step": 19146 }, { "epoch": 0.8072601555747623, "grad_norm": 2.645473003387451, "learning_rate": 4.06071395109154e-05, "loss": 7.0714, "step": 19147 }, { "epoch": 0.8073023167569618, "grad_norm": 1.6993675231933594, "learning_rate": 4.0606358724507324e-05, "loss": 6.6678, "step": 19148 }, { "epoch": 0.8073444779391614, "grad_norm": 2.319582223892212, "learning_rate": 4.0605577938099255e-05, "loss": 6.7839, "step": 19149 }, { "epoch": 0.807386639121361, "grad_norm": 2.014667510986328, "learning_rate": 4.0604797151691186e-05, "loss": 7.567, "step": 19150 }, { "epoch": 0.8074288003035605, "grad_norm": 2.351013422012329, "learning_rate": 4.060401636528312e-05, "loss": 7.3035, "step": 19151 }, { "epoch": 0.80747096148576, "grad_norm": 1.5242111682891846, "learning_rate": 4.060323557887504e-05, "loss": 7.0843, "step": 19152 }, { "epoch": 0.8075131226679596, "grad_norm": 1.6187008619308472, "learning_rate": 4.0602454792466973e-05, "loss": 6.994, "step": 19153 }, { "epoch": 0.8075552838501592, "grad_norm": 2.564018487930298, "learning_rate": 4.0601674006058905e-05, "loss": 6.642, "step": 19154 }, { "epoch": 0.8075974450323588, "grad_norm": 2.3312981128692627, "learning_rate": 4.060089321965083e-05, "loss": 6.6164, "step": 19155 }, { "epoch": 0.8076396062145582, "grad_norm": 2.795355796813965, "learning_rate": 4.060011243324277e-05, "loss": 7.5373, "step": 19156 }, { "epoch": 0.8076817673967578, "grad_norm": 1.4982956647872925, "learning_rate": 4.059933164683469e-05, "loss": 7.0631, "step": 19157 }, { "epoch": 0.8077239285789574, "grad_norm": 1.3326407670974731, "learning_rate": 4.059855086042662e-05, "loss": 6.9184, "step": 19158 }, { "epoch": 0.8077660897611569, "grad_norm": 2.6517181396484375, "learning_rate": 4.0597770074018554e-05, "loss": 6.6221, "step": 19159 }, { "epoch": 0.8078082509433564, "grad_norm": 1.4554741382598877, "learning_rate": 4.059698928761048e-05, "loss": 6.6571, "step": 19160 }, { "epoch": 0.807850412125556, "grad_norm": 2.4830174446105957, "learning_rate": 4.059620850120242e-05, "loss": 7.6655, "step": 19161 }, { "epoch": 0.8078925733077555, "grad_norm": 1.5921332836151123, "learning_rate": 4.059542771479434e-05, "loss": 6.8756, "step": 19162 }, { "epoch": 0.8079347344899551, "grad_norm": 1.8922336101531982, "learning_rate": 4.059464692838627e-05, "loss": 6.6473, "step": 19163 }, { "epoch": 0.8079768956721547, "grad_norm": 1.136407732963562, "learning_rate": 4.0593866141978204e-05, "loss": 7.2434, "step": 19164 }, { "epoch": 0.8080190568543542, "grad_norm": 1.3726375102996826, "learning_rate": 4.0593085355570135e-05, "loss": 6.5679, "step": 19165 }, { "epoch": 0.8080612180365537, "grad_norm": 2.8244271278381348, "learning_rate": 4.059230456916206e-05, "loss": 7.6074, "step": 19166 }, { "epoch": 0.8081033792187533, "grad_norm": 2.265928030014038, "learning_rate": 4.059152378275399e-05, "loss": 7.1506, "step": 19167 }, { "epoch": 0.8081455404009529, "grad_norm": 2.911086320877075, "learning_rate": 4.059074299634592e-05, "loss": 6.5713, "step": 19168 }, { "epoch": 0.8081877015831523, "grad_norm": 1.043605089187622, "learning_rate": 4.058996220993785e-05, "loss": 6.4774, "step": 19169 }, { "epoch": 0.8082298627653519, "grad_norm": 1.3674665689468384, "learning_rate": 4.0589181423529785e-05, "loss": 6.5141, "step": 19170 }, { "epoch": 0.8082720239475515, "grad_norm": 1.9049769639968872, "learning_rate": 4.058840063712171e-05, "loss": 7.2578, "step": 19171 }, { "epoch": 0.8083141851297511, "grad_norm": 2.0663485527038574, "learning_rate": 4.058761985071364e-05, "loss": 7.149, "step": 19172 }, { "epoch": 0.8083563463119506, "grad_norm": 1.4673640727996826, "learning_rate": 4.058683906430557e-05, "loss": 7.0075, "step": 19173 }, { "epoch": 0.8083985074941501, "grad_norm": 1.4578531980514526, "learning_rate": 4.0586058277897496e-05, "loss": 6.4036, "step": 19174 }, { "epoch": 0.8084406686763497, "grad_norm": 0.9845443964004517, "learning_rate": 4.0585277491489434e-05, "loss": 6.9082, "step": 19175 }, { "epoch": 0.8084828298585492, "grad_norm": 2.8847157955169678, "learning_rate": 4.058449670508136e-05, "loss": 7.3556, "step": 19176 }, { "epoch": 0.8085249910407488, "grad_norm": 3.8128907680511475, "learning_rate": 4.058371591867329e-05, "loss": 7.3606, "step": 19177 }, { "epoch": 0.8085671522229483, "grad_norm": 2.403017520904541, "learning_rate": 4.058293513226522e-05, "loss": 6.8917, "step": 19178 }, { "epoch": 0.8086093134051479, "grad_norm": 3.696533203125, "learning_rate": 4.0582154345857146e-05, "loss": 8.1981, "step": 19179 }, { "epoch": 0.8086514745873474, "grad_norm": 1.910807728767395, "learning_rate": 4.0581373559449084e-05, "loss": 6.7418, "step": 19180 }, { "epoch": 0.808693635769547, "grad_norm": 2.0589094161987305, "learning_rate": 4.058059277304101e-05, "loss": 7.0497, "step": 19181 }, { "epoch": 0.8087357969517466, "grad_norm": 4.024957180023193, "learning_rate": 4.057981198663294e-05, "loss": 7.6915, "step": 19182 }, { "epoch": 0.808777958133946, "grad_norm": 3.437056541442871, "learning_rate": 4.057903120022487e-05, "loss": 7.4607, "step": 19183 }, { "epoch": 0.8088201193161456, "grad_norm": 4.223163604736328, "learning_rate": 4.05782504138168e-05, "loss": 6.5409, "step": 19184 }, { "epoch": 0.8088622804983452, "grad_norm": 2.538156032562256, "learning_rate": 4.0577469627408726e-05, "loss": 6.7759, "step": 19185 }, { "epoch": 0.8089044416805448, "grad_norm": 2.498593330383301, "learning_rate": 4.057668884100066e-05, "loss": 6.4358, "step": 19186 }, { "epoch": 0.8089466028627442, "grad_norm": 1.3101836442947388, "learning_rate": 4.057590805459259e-05, "loss": 6.473, "step": 19187 }, { "epoch": 0.8089887640449438, "grad_norm": 2.4809184074401855, "learning_rate": 4.057512726818451e-05, "loss": 7.1542, "step": 19188 }, { "epoch": 0.8090309252271434, "grad_norm": 3.1340839862823486, "learning_rate": 4.057434648177645e-05, "loss": 7.0234, "step": 19189 }, { "epoch": 0.809073086409343, "grad_norm": 2.6490631103515625, "learning_rate": 4.0573565695368376e-05, "loss": 6.3749, "step": 19190 }, { "epoch": 0.8091152475915425, "grad_norm": 3.0931684970855713, "learning_rate": 4.057278490896031e-05, "loss": 8.3373, "step": 19191 }, { "epoch": 0.809157408773742, "grad_norm": 3.6505215167999268, "learning_rate": 4.057200412255224e-05, "loss": 6.8815, "step": 19192 }, { "epoch": 0.8091995699559416, "grad_norm": 1.4258407354354858, "learning_rate": 4.057122333614416e-05, "loss": 6.6283, "step": 19193 }, { "epoch": 0.8092417311381411, "grad_norm": 3.7305734157562256, "learning_rate": 4.05704425497361e-05, "loss": 6.6674, "step": 19194 }, { "epoch": 0.8092838923203407, "grad_norm": 3.3678059577941895, "learning_rate": 4.0569661763328025e-05, "loss": 7.9275, "step": 19195 }, { "epoch": 0.8093260535025402, "grad_norm": 2.377284526824951, "learning_rate": 4.056888097691996e-05, "loss": 7.6441, "step": 19196 }, { "epoch": 0.8093682146847397, "grad_norm": 2.339097261428833, "learning_rate": 4.056810019051189e-05, "loss": 7.311, "step": 19197 }, { "epoch": 0.8094103758669393, "grad_norm": 1.723900556564331, "learning_rate": 4.056731940410381e-05, "loss": 6.7496, "step": 19198 }, { "epoch": 0.8094525370491389, "grad_norm": 1.4181767702102661, "learning_rate": 4.0566538617695744e-05, "loss": 7.1063, "step": 19199 }, { "epoch": 0.8094946982313385, "grad_norm": 2.0195696353912354, "learning_rate": 4.0565757831287675e-05, "loss": 7.0995, "step": 19200 }, { "epoch": 0.8095368594135379, "grad_norm": 2.3907008171081543, "learning_rate": 4.0564977044879606e-05, "loss": 7.8765, "step": 19201 }, { "epoch": 0.8095790205957375, "grad_norm": 1.6687363386154175, "learning_rate": 4.056419625847153e-05, "loss": 6.7666, "step": 19202 }, { "epoch": 0.8096211817779371, "grad_norm": 2.6753885746002197, "learning_rate": 4.056341547206346e-05, "loss": 7.4144, "step": 19203 }, { "epoch": 0.8096633429601366, "grad_norm": 1.4258415699005127, "learning_rate": 4.056263468565539e-05, "loss": 6.7373, "step": 19204 }, { "epoch": 0.8097055041423361, "grad_norm": 1.3720073699951172, "learning_rate": 4.0561853899247324e-05, "loss": 6.7773, "step": 19205 }, { "epoch": 0.8097476653245357, "grad_norm": 1.170119285583496, "learning_rate": 4.0561073112839256e-05, "loss": 6.4811, "step": 19206 }, { "epoch": 0.8097898265067353, "grad_norm": 2.4789369106292725, "learning_rate": 4.056029232643118e-05, "loss": 8.1504, "step": 19207 }, { "epoch": 0.8098319876889348, "grad_norm": 1.4597468376159668, "learning_rate": 4.055951154002312e-05, "loss": 6.8904, "step": 19208 }, { "epoch": 0.8098741488711343, "grad_norm": 1.3250317573547363, "learning_rate": 4.055873075361504e-05, "loss": 6.5585, "step": 19209 }, { "epoch": 0.8099163100533339, "grad_norm": 1.0606235265731812, "learning_rate": 4.0557949967206974e-05, "loss": 6.5089, "step": 19210 }, { "epoch": 0.8099584712355334, "grad_norm": 2.647061586380005, "learning_rate": 4.0557169180798905e-05, "loss": 8.2187, "step": 19211 }, { "epoch": 0.810000632417733, "grad_norm": 4.1455793380737305, "learning_rate": 4.055638839439083e-05, "loss": 7.2956, "step": 19212 }, { "epoch": 0.8100427935999326, "grad_norm": 1.6643714904785156, "learning_rate": 4.055560760798276e-05, "loss": 6.6599, "step": 19213 }, { "epoch": 0.810084954782132, "grad_norm": 1.1948026418685913, "learning_rate": 4.055482682157469e-05, "loss": 6.903, "step": 19214 }, { "epoch": 0.8101271159643316, "grad_norm": 1.219396948814392, "learning_rate": 4.0554046035166623e-05, "loss": 6.651, "step": 19215 }, { "epoch": 0.8101692771465312, "grad_norm": 1.4304417371749878, "learning_rate": 4.0553265248758555e-05, "loss": 6.4093, "step": 19216 }, { "epoch": 0.8102114383287308, "grad_norm": 3.3960180282592773, "learning_rate": 4.055248446235048e-05, "loss": 7.3794, "step": 19217 }, { "epoch": 0.8102535995109302, "grad_norm": 2.5343737602233887, "learning_rate": 4.055170367594241e-05, "loss": 7.105, "step": 19218 }, { "epoch": 0.8102957606931298, "grad_norm": 1.1562637090682983, "learning_rate": 4.055092288953434e-05, "loss": 6.8568, "step": 19219 }, { "epoch": 0.8103379218753294, "grad_norm": 1.0507962703704834, "learning_rate": 4.055014210312627e-05, "loss": 6.7363, "step": 19220 }, { "epoch": 0.810380083057529, "grad_norm": 1.2247084379196167, "learning_rate": 4.05493613167182e-05, "loss": 6.4357, "step": 19221 }, { "epoch": 0.8104222442397285, "grad_norm": 1.8324511051177979, "learning_rate": 4.054858053031013e-05, "loss": 7.2073, "step": 19222 }, { "epoch": 0.810464405421928, "grad_norm": 0.921900749206543, "learning_rate": 4.054779974390206e-05, "loss": 6.5811, "step": 19223 }, { "epoch": 0.8105065666041276, "grad_norm": 0.807409942150116, "learning_rate": 4.0547018957493984e-05, "loss": 6.5311, "step": 19224 }, { "epoch": 0.8105487277863271, "grad_norm": 0.8683346509933472, "learning_rate": 4.054623817108592e-05, "loss": 6.5451, "step": 19225 }, { "epoch": 0.8105908889685267, "grad_norm": 1.1712933778762817, "learning_rate": 4.054545738467785e-05, "loss": 6.9025, "step": 19226 }, { "epoch": 0.8106330501507262, "grad_norm": 0.8114868402481079, "learning_rate": 4.0544676598269785e-05, "loss": 6.515, "step": 19227 }, { "epoch": 0.8106752113329257, "grad_norm": 2.5762906074523926, "learning_rate": 4.054389581186171e-05, "loss": 7.603, "step": 19228 }, { "epoch": 0.8107173725151253, "grad_norm": 1.299948811531067, "learning_rate": 4.0543115025453634e-05, "loss": 6.8424, "step": 19229 }, { "epoch": 0.8107595336973249, "grad_norm": 1.4045945405960083, "learning_rate": 4.054233423904557e-05, "loss": 7.1277, "step": 19230 }, { "epoch": 0.8108016948795245, "grad_norm": 3.335271120071411, "learning_rate": 4.0541553452637497e-05, "loss": 7.7315, "step": 19231 }, { "epoch": 0.8108438560617239, "grad_norm": 1.802428126335144, "learning_rate": 4.054077266622943e-05, "loss": 7.0636, "step": 19232 }, { "epoch": 0.8108860172439235, "grad_norm": 3.2751119136810303, "learning_rate": 4.053999187982136e-05, "loss": 6.5377, "step": 19233 }, { "epoch": 0.8109281784261231, "grad_norm": 2.0953097343444824, "learning_rate": 4.053921109341329e-05, "loss": 6.8496, "step": 19234 }, { "epoch": 0.8109703396083227, "grad_norm": 1.685224175453186, "learning_rate": 4.0538430307005215e-05, "loss": 6.5191, "step": 19235 }, { "epoch": 0.8110125007905221, "grad_norm": 3.0500285625457764, "learning_rate": 4.0537649520597146e-05, "loss": 7.7346, "step": 19236 }, { "epoch": 0.8110546619727217, "grad_norm": 2.217963695526123, "learning_rate": 4.053686873418908e-05, "loss": 7.0626, "step": 19237 }, { "epoch": 0.8110968231549213, "grad_norm": 2.2705078125, "learning_rate": 4.053608794778101e-05, "loss": 7.0965, "step": 19238 }, { "epoch": 0.8111389843371208, "grad_norm": 1.4844311475753784, "learning_rate": 4.053530716137294e-05, "loss": 6.6026, "step": 19239 }, { "epoch": 0.8111811455193204, "grad_norm": 2.2142021656036377, "learning_rate": 4.0534526374964864e-05, "loss": 6.4599, "step": 19240 }, { "epoch": 0.8112233067015199, "grad_norm": 0.8941373825073242, "learning_rate": 4.0533745588556796e-05, "loss": 6.6884, "step": 19241 }, { "epoch": 0.8112654678837194, "grad_norm": 2.8467984199523926, "learning_rate": 4.053296480214873e-05, "loss": 7.8692, "step": 19242 }, { "epoch": 0.811307629065919, "grad_norm": 1.4640777111053467, "learning_rate": 4.053218401574065e-05, "loss": 6.8806, "step": 19243 }, { "epoch": 0.8113497902481186, "grad_norm": 2.3624391555786133, "learning_rate": 4.053140322933259e-05, "loss": 8.0403, "step": 19244 }, { "epoch": 0.8113919514303181, "grad_norm": 2.6434221267700195, "learning_rate": 4.0530622442924514e-05, "loss": 7.0554, "step": 19245 }, { "epoch": 0.8114341126125176, "grad_norm": 2.775022268295288, "learning_rate": 4.0529841656516445e-05, "loss": 6.9611, "step": 19246 }, { "epoch": 0.8114762737947172, "grad_norm": 1.6644991636276245, "learning_rate": 4.0529060870108376e-05, "loss": 6.4152, "step": 19247 }, { "epoch": 0.8115184349769168, "grad_norm": 1.1456016302108765, "learning_rate": 4.05282800837003e-05, "loss": 6.5602, "step": 19248 }, { "epoch": 0.8115605961591164, "grad_norm": 2.491560220718384, "learning_rate": 4.052749929729224e-05, "loss": 7.0351, "step": 19249 }, { "epoch": 0.8116027573413158, "grad_norm": 3.6696829795837402, "learning_rate": 4.052671851088416e-05, "loss": 7.4206, "step": 19250 }, { "epoch": 0.8116449185235154, "grad_norm": 1.9066170454025269, "learning_rate": 4.0525937724476095e-05, "loss": 6.6905, "step": 19251 }, { "epoch": 0.811687079705715, "grad_norm": 4.2407546043396, "learning_rate": 4.0525156938068026e-05, "loss": 7.8277, "step": 19252 }, { "epoch": 0.8117292408879145, "grad_norm": 2.241612195968628, "learning_rate": 4.052437615165996e-05, "loss": 6.5461, "step": 19253 }, { "epoch": 0.811771402070114, "grad_norm": 1.5343947410583496, "learning_rate": 4.052359536525188e-05, "loss": 6.4902, "step": 19254 }, { "epoch": 0.8118135632523136, "grad_norm": 2.202686309814453, "learning_rate": 4.052281457884381e-05, "loss": 7.0167, "step": 19255 }, { "epoch": 0.8118557244345131, "grad_norm": 1.789168357849121, "learning_rate": 4.0522033792435744e-05, "loss": 6.9218, "step": 19256 }, { "epoch": 0.8118978856167127, "grad_norm": 1.734565258026123, "learning_rate": 4.052125300602767e-05, "loss": 6.7725, "step": 19257 }, { "epoch": 0.8119400467989123, "grad_norm": 2.224353075027466, "learning_rate": 4.052047221961961e-05, "loss": 7.3054, "step": 19258 }, { "epoch": 0.8119822079811118, "grad_norm": 3.1328887939453125, "learning_rate": 4.051969143321153e-05, "loss": 6.4051, "step": 19259 }, { "epoch": 0.8120243691633113, "grad_norm": 1.964169979095459, "learning_rate": 4.051891064680346e-05, "loss": 6.9734, "step": 19260 }, { "epoch": 0.8120665303455109, "grad_norm": 2.291760206222534, "learning_rate": 4.0518129860395394e-05, "loss": 6.9314, "step": 19261 }, { "epoch": 0.8121086915277105, "grad_norm": 2.6353657245635986, "learning_rate": 4.051734907398732e-05, "loss": 7.8906, "step": 19262 }, { "epoch": 0.8121508527099099, "grad_norm": 1.404653549194336, "learning_rate": 4.0516568287579256e-05, "loss": 6.6377, "step": 19263 }, { "epoch": 0.8121930138921095, "grad_norm": 2.345813274383545, "learning_rate": 4.051578750117118e-05, "loss": 6.4072, "step": 19264 }, { "epoch": 0.8122351750743091, "grad_norm": 1.3696982860565186, "learning_rate": 4.051500671476311e-05, "loss": 6.7216, "step": 19265 }, { "epoch": 0.8122773362565087, "grad_norm": 2.8365542888641357, "learning_rate": 4.051422592835504e-05, "loss": 6.9427, "step": 19266 }, { "epoch": 0.8123194974387081, "grad_norm": 1.9349557161331177, "learning_rate": 4.051344514194697e-05, "loss": 6.3932, "step": 19267 }, { "epoch": 0.8123616586209077, "grad_norm": 1.541521430015564, "learning_rate": 4.05126643555389e-05, "loss": 6.7942, "step": 19268 }, { "epoch": 0.8124038198031073, "grad_norm": 1.8333319425582886, "learning_rate": 4.051188356913083e-05, "loss": 6.761, "step": 19269 }, { "epoch": 0.8124459809853068, "grad_norm": 2.7343804836273193, "learning_rate": 4.051110278272276e-05, "loss": 6.5981, "step": 19270 }, { "epoch": 0.8124881421675064, "grad_norm": 1.1765090227127075, "learning_rate": 4.0510321996314686e-05, "loss": 6.2278, "step": 19271 }, { "epoch": 0.8125303033497059, "grad_norm": 1.535083293914795, "learning_rate": 4.0509541209906624e-05, "loss": 6.5854, "step": 19272 }, { "epoch": 0.8125724645319055, "grad_norm": 3.806744337081909, "learning_rate": 4.050876042349855e-05, "loss": 7.943, "step": 19273 }, { "epoch": 0.812614625714105, "grad_norm": 1.4364850521087646, "learning_rate": 4.050797963709048e-05, "loss": 7.5026, "step": 19274 }, { "epoch": 0.8126567868963046, "grad_norm": 2.0486228466033936, "learning_rate": 4.050719885068241e-05, "loss": 6.905, "step": 19275 }, { "epoch": 0.8126989480785041, "grad_norm": 2.2998366355895996, "learning_rate": 4.0506418064274335e-05, "loss": 7.6161, "step": 19276 }, { "epoch": 0.8127411092607036, "grad_norm": 2.2489044666290283, "learning_rate": 4.0505637277866273e-05, "loss": 7.0342, "step": 19277 }, { "epoch": 0.8127832704429032, "grad_norm": 2.4601221084594727, "learning_rate": 4.05048564914582e-05, "loss": 6.9352, "step": 19278 }, { "epoch": 0.8128254316251028, "grad_norm": 3.745499610900879, "learning_rate": 4.050407570505013e-05, "loss": 7.7796, "step": 19279 }, { "epoch": 0.8128675928073024, "grad_norm": 2.04864239692688, "learning_rate": 4.050329491864206e-05, "loss": 6.4932, "step": 19280 }, { "epoch": 0.8129097539895018, "grad_norm": 3.726436138153076, "learning_rate": 4.0502514132233985e-05, "loss": 9.5119, "step": 19281 }, { "epoch": 0.8129519151717014, "grad_norm": 1.4724104404449463, "learning_rate": 4.0501733345825916e-05, "loss": 6.3481, "step": 19282 }, { "epoch": 0.812994076353901, "grad_norm": 1.1905745267868042, "learning_rate": 4.050095255941785e-05, "loss": 6.5411, "step": 19283 }, { "epoch": 0.8130362375361005, "grad_norm": 3.201188087463379, "learning_rate": 4.050017177300978e-05, "loss": 7.7081, "step": 19284 }, { "epoch": 0.8130783987183, "grad_norm": 1.593037486076355, "learning_rate": 4.049939098660171e-05, "loss": 6.7006, "step": 19285 }, { "epoch": 0.8131205599004996, "grad_norm": 1.589821219444275, "learning_rate": 4.0498610200193635e-05, "loss": 6.7259, "step": 19286 }, { "epoch": 0.8131627210826992, "grad_norm": 3.103130578994751, "learning_rate": 4.0497829413785566e-05, "loss": 7.7374, "step": 19287 }, { "epoch": 0.8132048822648987, "grad_norm": 2.633718729019165, "learning_rate": 4.04970486273775e-05, "loss": 7.278, "step": 19288 }, { "epoch": 0.8132470434470983, "grad_norm": 2.2302355766296387, "learning_rate": 4.049626784096943e-05, "loss": 8.1242, "step": 19289 }, { "epoch": 0.8132892046292978, "grad_norm": 1.1305822134017944, "learning_rate": 4.049548705456135e-05, "loss": 6.9879, "step": 19290 }, { "epoch": 0.8133313658114973, "grad_norm": 2.839231252670288, "learning_rate": 4.0494706268153284e-05, "loss": 6.8663, "step": 19291 }, { "epoch": 0.8133735269936969, "grad_norm": 2.7694640159606934, "learning_rate": 4.0493925481745215e-05, "loss": 7.8313, "step": 19292 }, { "epoch": 0.8134156881758965, "grad_norm": 1.7514455318450928, "learning_rate": 4.049314469533714e-05, "loss": 7.1039, "step": 19293 }, { "epoch": 0.813457849358096, "grad_norm": 2.130361557006836, "learning_rate": 4.049236390892908e-05, "loss": 6.4216, "step": 19294 }, { "epoch": 0.8135000105402955, "grad_norm": 1.8529692888259888, "learning_rate": 4.0491583122521e-05, "loss": 6.7469, "step": 19295 }, { "epoch": 0.8135421717224951, "grad_norm": 1.9727892875671387, "learning_rate": 4.049080233611294e-05, "loss": 6.9314, "step": 19296 }, { "epoch": 0.8135843329046947, "grad_norm": 1.1885040998458862, "learning_rate": 4.0490021549704865e-05, "loss": 6.6824, "step": 19297 }, { "epoch": 0.8136264940868942, "grad_norm": 2.931360960006714, "learning_rate": 4.0489240763296796e-05, "loss": 7.6426, "step": 19298 }, { "epoch": 0.8136686552690937, "grad_norm": 1.84669828414917, "learning_rate": 4.048845997688873e-05, "loss": 6.7755, "step": 19299 }, { "epoch": 0.8137108164512933, "grad_norm": 1.4145219326019287, "learning_rate": 4.048767919048065e-05, "loss": 6.9037, "step": 19300 }, { "epoch": 0.8137529776334929, "grad_norm": 1.3350375890731812, "learning_rate": 4.048689840407258e-05, "loss": 6.6662, "step": 19301 }, { "epoch": 0.8137951388156924, "grad_norm": 3.4637563228607178, "learning_rate": 4.0486117617664514e-05, "loss": 7.0866, "step": 19302 }, { "epoch": 0.8138372999978919, "grad_norm": 1.2337790727615356, "learning_rate": 4.0485336831256446e-05, "loss": 6.5953, "step": 19303 }, { "epoch": 0.8138794611800915, "grad_norm": 2.5402774810791016, "learning_rate": 4.048455604484837e-05, "loss": 7.1113, "step": 19304 }, { "epoch": 0.813921622362291, "grad_norm": 1.4888426065444946, "learning_rate": 4.04837752584403e-05, "loss": 6.3643, "step": 19305 }, { "epoch": 0.8139637835444906, "grad_norm": 2.750774383544922, "learning_rate": 4.048299447203223e-05, "loss": 6.967, "step": 19306 }, { "epoch": 0.8140059447266902, "grad_norm": 2.796346426010132, "learning_rate": 4.0482213685624164e-05, "loss": 7.1091, "step": 19307 }, { "epoch": 0.8140481059088897, "grad_norm": 2.0799388885498047, "learning_rate": 4.0481432899216095e-05, "loss": 7.2902, "step": 19308 }, { "epoch": 0.8140902670910892, "grad_norm": 2.3841776847839355, "learning_rate": 4.048065211280802e-05, "loss": 7.4319, "step": 19309 }, { "epoch": 0.8141324282732888, "grad_norm": 3.243784189224243, "learning_rate": 4.047987132639995e-05, "loss": 7.6137, "step": 19310 }, { "epoch": 0.8141745894554884, "grad_norm": 1.4800130128860474, "learning_rate": 4.047909053999188e-05, "loss": 6.5311, "step": 19311 }, { "epoch": 0.8142167506376878, "grad_norm": 1.6073261499404907, "learning_rate": 4.0478309753583807e-05, "loss": 6.4178, "step": 19312 }, { "epoch": 0.8142589118198874, "grad_norm": 1.186691164970398, "learning_rate": 4.0477528967175745e-05, "loss": 6.5946, "step": 19313 }, { "epoch": 0.814301073002087, "grad_norm": 2.101834535598755, "learning_rate": 4.047674818076767e-05, "loss": 6.9048, "step": 19314 }, { "epoch": 0.8143432341842866, "grad_norm": 3.3114254474639893, "learning_rate": 4.04759673943596e-05, "loss": 8.2788, "step": 19315 }, { "epoch": 0.8143853953664861, "grad_norm": 2.4148592948913574, "learning_rate": 4.047518660795153e-05, "loss": 8.1277, "step": 19316 }, { "epoch": 0.8144275565486856, "grad_norm": 2.1781861782073975, "learning_rate": 4.047440582154346e-05, "loss": 6.66, "step": 19317 }, { "epoch": 0.8144697177308852, "grad_norm": 1.8399773836135864, "learning_rate": 4.0473625035135394e-05, "loss": 6.5779, "step": 19318 }, { "epoch": 0.8145118789130847, "grad_norm": 2.390226125717163, "learning_rate": 4.047284424872732e-05, "loss": 7.3891, "step": 19319 }, { "epoch": 0.8145540400952843, "grad_norm": 3.0132594108581543, "learning_rate": 4.047206346231925e-05, "loss": 6.4787, "step": 19320 }, { "epoch": 0.8145962012774838, "grad_norm": 2.9836411476135254, "learning_rate": 4.047128267591118e-05, "loss": 9.4758, "step": 19321 }, { "epoch": 0.8146383624596834, "grad_norm": 2.07403564453125, "learning_rate": 4.047050188950311e-05, "loss": 7.2942, "step": 19322 }, { "epoch": 0.8146805236418829, "grad_norm": 3.3449020385742188, "learning_rate": 4.046972110309504e-05, "loss": 8.2137, "step": 19323 }, { "epoch": 0.8147226848240825, "grad_norm": 3.2546379566192627, "learning_rate": 4.046894031668697e-05, "loss": 6.6238, "step": 19324 }, { "epoch": 0.814764846006282, "grad_norm": 2.9033761024475098, "learning_rate": 4.04681595302789e-05, "loss": 6.5017, "step": 19325 }, { "epoch": 0.8148070071884815, "grad_norm": 5.012976169586182, "learning_rate": 4.0467378743870824e-05, "loss": 10.6424, "step": 19326 }, { "epoch": 0.8148491683706811, "grad_norm": 1.034067153930664, "learning_rate": 4.046659795746276e-05, "loss": 6.6768, "step": 19327 }, { "epoch": 0.8148913295528807, "grad_norm": 0.951091468334198, "learning_rate": 4.0465817171054686e-05, "loss": 6.4896, "step": 19328 }, { "epoch": 0.8149334907350803, "grad_norm": 3.1464171409606934, "learning_rate": 4.046503638464662e-05, "loss": 7.693, "step": 19329 }, { "epoch": 0.8149756519172797, "grad_norm": 2.311117649078369, "learning_rate": 4.046425559823855e-05, "loss": 7.1573, "step": 19330 }, { "epoch": 0.8150178130994793, "grad_norm": 3.815955638885498, "learning_rate": 4.0463474811830473e-05, "loss": 7.0197, "step": 19331 }, { "epoch": 0.8150599742816789, "grad_norm": 1.001164436340332, "learning_rate": 4.046269402542241e-05, "loss": 6.6454, "step": 19332 }, { "epoch": 0.8151021354638784, "grad_norm": 1.0667860507965088, "learning_rate": 4.0461913239014336e-05, "loss": 6.3859, "step": 19333 }, { "epoch": 0.8151442966460779, "grad_norm": 3.1729989051818848, "learning_rate": 4.046113245260627e-05, "loss": 7.6397, "step": 19334 }, { "epoch": 0.8151864578282775, "grad_norm": 1.3323335647583008, "learning_rate": 4.04603516661982e-05, "loss": 6.655, "step": 19335 }, { "epoch": 0.815228619010477, "grad_norm": 1.581671953201294, "learning_rate": 4.045957087979012e-05, "loss": 6.336, "step": 19336 }, { "epoch": 0.8152707801926766, "grad_norm": 1.2335211038589478, "learning_rate": 4.0458790093382054e-05, "loss": 6.8379, "step": 19337 }, { "epoch": 0.8153129413748762, "grad_norm": 1.5132167339324951, "learning_rate": 4.0458009306973985e-05, "loss": 6.458, "step": 19338 }, { "epoch": 0.8153551025570757, "grad_norm": 0.9030040502548218, "learning_rate": 4.045722852056592e-05, "loss": 6.5028, "step": 19339 }, { "epoch": 0.8153972637392752, "grad_norm": 2.8504478931427, "learning_rate": 4.045644773415784e-05, "loss": 7.5435, "step": 19340 }, { "epoch": 0.8154394249214748, "grad_norm": 1.3472431898117065, "learning_rate": 4.045566694774978e-05, "loss": 6.5006, "step": 19341 }, { "epoch": 0.8154815861036744, "grad_norm": 4.590403079986572, "learning_rate": 4.0454886161341704e-05, "loss": 7.8672, "step": 19342 }, { "epoch": 0.8155237472858738, "grad_norm": 2.3914222717285156, "learning_rate": 4.0454105374933635e-05, "loss": 6.8618, "step": 19343 }, { "epoch": 0.8155659084680734, "grad_norm": 1.903286099433899, "learning_rate": 4.0453324588525566e-05, "loss": 6.6676, "step": 19344 }, { "epoch": 0.815608069650273, "grad_norm": 1.111589789390564, "learning_rate": 4.045254380211749e-05, "loss": 6.4219, "step": 19345 }, { "epoch": 0.8156502308324726, "grad_norm": 1.2288031578063965, "learning_rate": 4.045176301570943e-05, "loss": 6.705, "step": 19346 }, { "epoch": 0.8156923920146721, "grad_norm": 1.9512426853179932, "learning_rate": 4.045098222930135e-05, "loss": 6.702, "step": 19347 }, { "epoch": 0.8157345531968716, "grad_norm": 1.6310480833053589, "learning_rate": 4.0450201442893285e-05, "loss": 7.0223, "step": 19348 }, { "epoch": 0.8157767143790712, "grad_norm": 2.337454319000244, "learning_rate": 4.0449420656485216e-05, "loss": 6.4763, "step": 19349 }, { "epoch": 0.8158188755612708, "grad_norm": 1.8205710649490356, "learning_rate": 4.044863987007714e-05, "loss": 6.9202, "step": 19350 }, { "epoch": 0.8158610367434703, "grad_norm": 0.8686370253562927, "learning_rate": 4.044785908366907e-05, "loss": 6.3888, "step": 19351 }, { "epoch": 0.8159031979256698, "grad_norm": 2.136857748031616, "learning_rate": 4.0447078297261e-05, "loss": 6.8938, "step": 19352 }, { "epoch": 0.8159453591078694, "grad_norm": 4.083782196044922, "learning_rate": 4.0446297510852934e-05, "loss": 7.1502, "step": 19353 }, { "epoch": 0.8159875202900689, "grad_norm": 2.271012544631958, "learning_rate": 4.0445516724444865e-05, "loss": 6.9435, "step": 19354 }, { "epoch": 0.8160296814722685, "grad_norm": 1.8395813703536987, "learning_rate": 4.044473593803679e-05, "loss": 7.2448, "step": 19355 }, { "epoch": 0.8160718426544681, "grad_norm": 1.6915866136550903, "learning_rate": 4.044395515162872e-05, "loss": 6.4288, "step": 19356 }, { "epoch": 0.8161140038366675, "grad_norm": 2.715975761413574, "learning_rate": 4.044317436522065e-05, "loss": 6.3037, "step": 19357 }, { "epoch": 0.8161561650188671, "grad_norm": 1.7506023645401, "learning_rate": 4.0442393578812584e-05, "loss": 6.4174, "step": 19358 }, { "epoch": 0.8161983262010667, "grad_norm": 2.199950933456421, "learning_rate": 4.044161279240451e-05, "loss": 6.7139, "step": 19359 }, { "epoch": 0.8162404873832663, "grad_norm": 1.2359702587127686, "learning_rate": 4.0440832005996446e-05, "loss": 6.3246, "step": 19360 }, { "epoch": 0.8162826485654657, "grad_norm": 3.9065873622894287, "learning_rate": 4.044005121958837e-05, "loss": 8.093, "step": 19361 }, { "epoch": 0.8163248097476653, "grad_norm": 1.768311858177185, "learning_rate": 4.04392704331803e-05, "loss": 6.5246, "step": 19362 }, { "epoch": 0.8163669709298649, "grad_norm": 2.3292348384857178, "learning_rate": 4.043848964677223e-05, "loss": 6.7969, "step": 19363 }, { "epoch": 0.8164091321120645, "grad_norm": 1.9254289865493774, "learning_rate": 4.043770886036416e-05, "loss": 6.5549, "step": 19364 }, { "epoch": 0.816451293294264, "grad_norm": 1.3946424722671509, "learning_rate": 4.0436928073956096e-05, "loss": 7.1801, "step": 19365 }, { "epoch": 0.8164934544764635, "grad_norm": 1.8387269973754883, "learning_rate": 4.043614728754802e-05, "loss": 7.1679, "step": 19366 }, { "epoch": 0.8165356156586631, "grad_norm": 2.9714252948760986, "learning_rate": 4.043536650113995e-05, "loss": 8.1014, "step": 19367 }, { "epoch": 0.8165777768408626, "grad_norm": 1.7032363414764404, "learning_rate": 4.043458571473188e-05, "loss": 7.1484, "step": 19368 }, { "epoch": 0.8166199380230622, "grad_norm": 1.8135582208633423, "learning_rate": 4.043380492832381e-05, "loss": 7.4925, "step": 19369 }, { "epoch": 0.8166620992052617, "grad_norm": 1.865316390991211, "learning_rate": 4.043302414191574e-05, "loss": 6.9776, "step": 19370 }, { "epoch": 0.8167042603874612, "grad_norm": 1.9161521196365356, "learning_rate": 4.043224335550767e-05, "loss": 8.2316, "step": 19371 }, { "epoch": 0.8167464215696608, "grad_norm": 1.32956063747406, "learning_rate": 4.04314625690996e-05, "loss": 6.4895, "step": 19372 }, { "epoch": 0.8167885827518604, "grad_norm": 4.084351539611816, "learning_rate": 4.0430681782691525e-05, "loss": 7.7491, "step": 19373 }, { "epoch": 0.81683074393406, "grad_norm": 1.558130145072937, "learning_rate": 4.042990099628346e-05, "loss": 6.5511, "step": 19374 }, { "epoch": 0.8168729051162594, "grad_norm": 2.9522006511688232, "learning_rate": 4.042912020987539e-05, "loss": 6.3765, "step": 19375 }, { "epoch": 0.816915066298459, "grad_norm": 1.4876842498779297, "learning_rate": 4.042833942346732e-05, "loss": 6.3012, "step": 19376 }, { "epoch": 0.8169572274806586, "grad_norm": 2.8056864738464355, "learning_rate": 4.042755863705925e-05, "loss": 7.7525, "step": 19377 }, { "epoch": 0.8169993886628582, "grad_norm": 1.4480018615722656, "learning_rate": 4.0426777850651175e-05, "loss": 7.1915, "step": 19378 }, { "epoch": 0.8170415498450576, "grad_norm": 4.113816261291504, "learning_rate": 4.042599706424311e-05, "loss": 7.8865, "step": 19379 }, { "epoch": 0.8170837110272572, "grad_norm": 1.2066378593444824, "learning_rate": 4.042521627783504e-05, "loss": 7.1697, "step": 19380 }, { "epoch": 0.8171258722094568, "grad_norm": 1.5812757015228271, "learning_rate": 4.042443549142696e-05, "loss": 6.6673, "step": 19381 }, { "epoch": 0.8171680333916563, "grad_norm": 1.023877739906311, "learning_rate": 4.04236547050189e-05, "loss": 6.6312, "step": 19382 }, { "epoch": 0.8172101945738558, "grad_norm": 1.3282313346862793, "learning_rate": 4.0422873918610824e-05, "loss": 7.2961, "step": 19383 }, { "epoch": 0.8172523557560554, "grad_norm": 1.232318639755249, "learning_rate": 4.0422093132202756e-05, "loss": 6.9597, "step": 19384 }, { "epoch": 0.817294516938255, "grad_norm": 2.825704336166382, "learning_rate": 4.042131234579469e-05, "loss": 7.729, "step": 19385 }, { "epoch": 0.8173366781204545, "grad_norm": 3.0094332695007324, "learning_rate": 4.042053155938662e-05, "loss": 6.6341, "step": 19386 }, { "epoch": 0.8173788393026541, "grad_norm": 1.8004765510559082, "learning_rate": 4.041975077297855e-05, "loss": 6.3357, "step": 19387 }, { "epoch": 0.8174210004848536, "grad_norm": 1.5113048553466797, "learning_rate": 4.0418969986570474e-05, "loss": 6.9841, "step": 19388 }, { "epoch": 0.8174631616670531, "grad_norm": 3.1341261863708496, "learning_rate": 4.0418189200162405e-05, "loss": 7.7136, "step": 19389 }, { "epoch": 0.8175053228492527, "grad_norm": 1.5730109214782715, "learning_rate": 4.0417408413754336e-05, "loss": 6.5388, "step": 19390 }, { "epoch": 0.8175474840314523, "grad_norm": 1.5146485567092896, "learning_rate": 4.041662762734627e-05, "loss": 6.4605, "step": 19391 }, { "epoch": 0.8175896452136517, "grad_norm": 1.2447242736816406, "learning_rate": 4.041584684093819e-05, "loss": 6.5176, "step": 19392 }, { "epoch": 0.8176318063958513, "grad_norm": 3.3792619705200195, "learning_rate": 4.0415066054530123e-05, "loss": 7.5214, "step": 19393 }, { "epoch": 0.8176739675780509, "grad_norm": 3.0433995723724365, "learning_rate": 4.0414285268122055e-05, "loss": 6.5541, "step": 19394 }, { "epoch": 0.8177161287602505, "grad_norm": 1.3811806440353394, "learning_rate": 4.041350448171398e-05, "loss": 6.7319, "step": 19395 }, { "epoch": 0.81775828994245, "grad_norm": 1.1588982343673706, "learning_rate": 4.041272369530592e-05, "loss": 6.7317, "step": 19396 }, { "epoch": 0.8178004511246495, "grad_norm": 1.6779718399047852, "learning_rate": 4.041194290889784e-05, "loss": 7.0766, "step": 19397 }, { "epoch": 0.8178426123068491, "grad_norm": 1.9308302402496338, "learning_rate": 4.041116212248977e-05, "loss": 6.5634, "step": 19398 }, { "epoch": 0.8178847734890486, "grad_norm": 1.7043709754943848, "learning_rate": 4.0410381336081704e-05, "loss": 6.6751, "step": 19399 }, { "epoch": 0.8179269346712482, "grad_norm": 1.27633535861969, "learning_rate": 4.040960054967363e-05, "loss": 6.762, "step": 19400 }, { "epoch": 0.8179690958534477, "grad_norm": 2.7114200592041016, "learning_rate": 4.040881976326557e-05, "loss": 6.9814, "step": 19401 }, { "epoch": 0.8180112570356473, "grad_norm": 4.404088020324707, "learning_rate": 4.040803897685749e-05, "loss": 8.4599, "step": 19402 }, { "epoch": 0.8180534182178468, "grad_norm": 1.934536099433899, "learning_rate": 4.040725819044942e-05, "loss": 6.897, "step": 19403 }, { "epoch": 0.8180955794000464, "grad_norm": 1.9599850177764893, "learning_rate": 4.0406477404041354e-05, "loss": 6.6824, "step": 19404 }, { "epoch": 0.818137740582246, "grad_norm": 2.3189587593078613, "learning_rate": 4.0405696617633285e-05, "loss": 6.9759, "step": 19405 }, { "epoch": 0.8181799017644454, "grad_norm": 3.187324047088623, "learning_rate": 4.040491583122521e-05, "loss": 7.7649, "step": 19406 }, { "epoch": 0.818222062946645, "grad_norm": 3.0432705879211426, "learning_rate": 4.040413504481714e-05, "loss": 6.3114, "step": 19407 }, { "epoch": 0.8182642241288446, "grad_norm": 1.7299100160598755, "learning_rate": 4.040335425840907e-05, "loss": 6.4308, "step": 19408 }, { "epoch": 0.8183063853110442, "grad_norm": 2.570729970932007, "learning_rate": 4.0402573472001e-05, "loss": 6.842, "step": 19409 }, { "epoch": 0.8183485464932436, "grad_norm": 1.7825791835784912, "learning_rate": 4.0401792685592935e-05, "loss": 6.988, "step": 19410 }, { "epoch": 0.8183907076754432, "grad_norm": 1.9863311052322388, "learning_rate": 4.040101189918486e-05, "loss": 6.7743, "step": 19411 }, { "epoch": 0.8184328688576428, "grad_norm": 2.0380184650421143, "learning_rate": 4.040023111277679e-05, "loss": 6.9088, "step": 19412 }, { "epoch": 0.8184750300398423, "grad_norm": 1.359917163848877, "learning_rate": 4.039945032636872e-05, "loss": 6.5386, "step": 19413 }, { "epoch": 0.8185171912220419, "grad_norm": 1.660950779914856, "learning_rate": 4.0398669539960646e-05, "loss": 6.9391, "step": 19414 }, { "epoch": 0.8185593524042414, "grad_norm": 1.927549958229065, "learning_rate": 4.0397888753552584e-05, "loss": 7.0191, "step": 19415 }, { "epoch": 0.818601513586441, "grad_norm": 1.2270634174346924, "learning_rate": 4.039710796714451e-05, "loss": 6.5816, "step": 19416 }, { "epoch": 0.8186436747686405, "grad_norm": 3.151610851287842, "learning_rate": 4.039632718073644e-05, "loss": 6.4857, "step": 19417 }, { "epoch": 0.8186858359508401, "grad_norm": 1.9135819673538208, "learning_rate": 4.039554639432837e-05, "loss": 6.8905, "step": 19418 }, { "epoch": 0.8187279971330396, "grad_norm": 2.372382164001465, "learning_rate": 4.0394765607920296e-05, "loss": 6.5098, "step": 19419 }, { "epoch": 0.8187701583152391, "grad_norm": 2.538451910018921, "learning_rate": 4.039398482151223e-05, "loss": 6.4296, "step": 19420 }, { "epoch": 0.8188123194974387, "grad_norm": 1.759415626525879, "learning_rate": 4.039320403510416e-05, "loss": 6.8165, "step": 19421 }, { "epoch": 0.8188544806796383, "grad_norm": 2.7434000968933105, "learning_rate": 4.039242324869609e-05, "loss": 7.3277, "step": 19422 }, { "epoch": 0.8188966418618379, "grad_norm": 1.475736141204834, "learning_rate": 4.039164246228802e-05, "loss": 6.8876, "step": 19423 }, { "epoch": 0.8189388030440373, "grad_norm": 1.3587006330490112, "learning_rate": 4.039086167587995e-05, "loss": 6.4879, "step": 19424 }, { "epoch": 0.8189809642262369, "grad_norm": 1.913482904434204, "learning_rate": 4.0390080889471876e-05, "loss": 6.5671, "step": 19425 }, { "epoch": 0.8190231254084365, "grad_norm": 1.8218189477920532, "learning_rate": 4.038930010306381e-05, "loss": 7.1076, "step": 19426 }, { "epoch": 0.819065286590636, "grad_norm": 2.423377752304077, "learning_rate": 4.038851931665574e-05, "loss": 6.7184, "step": 19427 }, { "epoch": 0.8191074477728355, "grad_norm": 2.7681572437286377, "learning_rate": 4.038773853024766e-05, "loss": 7.0222, "step": 19428 }, { "epoch": 0.8191496089550351, "grad_norm": 2.403704881668091, "learning_rate": 4.03869577438396e-05, "loss": 6.7495, "step": 19429 }, { "epoch": 0.8191917701372347, "grad_norm": 1.843820571899414, "learning_rate": 4.0386176957431526e-05, "loss": 6.5085, "step": 19430 }, { "epoch": 0.8192339313194342, "grad_norm": 2.0985195636749268, "learning_rate": 4.038539617102346e-05, "loss": 6.9638, "step": 19431 }, { "epoch": 0.8192760925016337, "grad_norm": 1.3436660766601562, "learning_rate": 4.038461538461539e-05, "loss": 6.5434, "step": 19432 }, { "epoch": 0.8193182536838333, "grad_norm": 2.8327889442443848, "learning_rate": 4.038383459820731e-05, "loss": 6.923, "step": 19433 }, { "epoch": 0.8193604148660328, "grad_norm": 1.5961321592330933, "learning_rate": 4.038305381179925e-05, "loss": 7.2533, "step": 19434 }, { "epoch": 0.8194025760482324, "grad_norm": 2.8238894939422607, "learning_rate": 4.0382273025391175e-05, "loss": 6.834, "step": 19435 }, { "epoch": 0.819444737230432, "grad_norm": 4.179247856140137, "learning_rate": 4.038149223898311e-05, "loss": 7.4287, "step": 19436 }, { "epoch": 0.8194868984126314, "grad_norm": 2.743745803833008, "learning_rate": 4.038071145257504e-05, "loss": 7.8426, "step": 19437 }, { "epoch": 0.819529059594831, "grad_norm": 1.3203045129776, "learning_rate": 4.037993066616696e-05, "loss": 6.5536, "step": 19438 }, { "epoch": 0.8195712207770306, "grad_norm": 2.660999059677124, "learning_rate": 4.0379149879758894e-05, "loss": 7.5901, "step": 19439 }, { "epoch": 0.8196133819592302, "grad_norm": 2.060669422149658, "learning_rate": 4.0378369093350825e-05, "loss": 6.6267, "step": 19440 }, { "epoch": 0.8196555431414296, "grad_norm": 1.7031906843185425, "learning_rate": 4.0377588306942756e-05, "loss": 6.4756, "step": 19441 }, { "epoch": 0.8196977043236292, "grad_norm": 4.236368656158447, "learning_rate": 4.037680752053468e-05, "loss": 7.5665, "step": 19442 }, { "epoch": 0.8197398655058288, "grad_norm": 0.8695575594902039, "learning_rate": 4.037602673412661e-05, "loss": 6.4399, "step": 19443 }, { "epoch": 0.8197820266880284, "grad_norm": 1.2272851467132568, "learning_rate": 4.037524594771854e-05, "loss": 7.1993, "step": 19444 }, { "epoch": 0.8198241878702279, "grad_norm": 1.5764145851135254, "learning_rate": 4.0374465161310474e-05, "loss": 6.6669, "step": 19445 }, { "epoch": 0.8198663490524274, "grad_norm": 1.4626779556274414, "learning_rate": 4.0373684374902406e-05, "loss": 7.1763, "step": 19446 }, { "epoch": 0.819908510234627, "grad_norm": 2.5997908115386963, "learning_rate": 4.037290358849433e-05, "loss": 8.2346, "step": 19447 }, { "epoch": 0.8199506714168265, "grad_norm": 2.1029250621795654, "learning_rate": 4.037212280208627e-05, "loss": 7.0651, "step": 19448 }, { "epoch": 0.8199928325990261, "grad_norm": 2.45450758934021, "learning_rate": 4.037134201567819e-05, "loss": 7.6722, "step": 19449 }, { "epoch": 0.8200349937812256, "grad_norm": 2.724400520324707, "learning_rate": 4.0370561229270124e-05, "loss": 6.7781, "step": 19450 }, { "epoch": 0.8200771549634251, "grad_norm": 2.0734493732452393, "learning_rate": 4.0369780442862055e-05, "loss": 6.5684, "step": 19451 }, { "epoch": 0.8201193161456247, "grad_norm": 1.89979088306427, "learning_rate": 4.036899965645398e-05, "loss": 7.2679, "step": 19452 }, { "epoch": 0.8201614773278243, "grad_norm": 1.476271390914917, "learning_rate": 4.036821887004591e-05, "loss": 7.054, "step": 19453 }, { "epoch": 0.8202036385100239, "grad_norm": 2.122342586517334, "learning_rate": 4.036743808363784e-05, "loss": 6.9949, "step": 19454 }, { "epoch": 0.8202457996922233, "grad_norm": 5.32253885269165, "learning_rate": 4.0366657297229773e-05, "loss": 7.1653, "step": 19455 }, { "epoch": 0.8202879608744229, "grad_norm": 3.8187079429626465, "learning_rate": 4.0365876510821705e-05, "loss": 6.4974, "step": 19456 }, { "epoch": 0.8203301220566225, "grad_norm": 1.9649181365966797, "learning_rate": 4.036509572441363e-05, "loss": 6.8235, "step": 19457 }, { "epoch": 0.820372283238822, "grad_norm": 1.5248364210128784, "learning_rate": 4.036431493800556e-05, "loss": 6.5925, "step": 19458 }, { "epoch": 0.8204144444210215, "grad_norm": 1.4481191635131836, "learning_rate": 4.036353415159749e-05, "loss": 6.7879, "step": 19459 }, { "epoch": 0.8204566056032211, "grad_norm": 1.2784396409988403, "learning_rate": 4.036275336518942e-05, "loss": 6.7082, "step": 19460 }, { "epoch": 0.8204987667854207, "grad_norm": 1.190780520439148, "learning_rate": 4.036197257878135e-05, "loss": 7.1029, "step": 19461 }, { "epoch": 0.8205409279676202, "grad_norm": 3.774653673171997, "learning_rate": 4.036119179237328e-05, "loss": 8.1519, "step": 19462 }, { "epoch": 0.8205830891498198, "grad_norm": 2.6235597133636475, "learning_rate": 4.036041100596521e-05, "loss": 7.3518, "step": 19463 }, { "epoch": 0.8206252503320193, "grad_norm": 2.8872478008270264, "learning_rate": 4.0359630219557134e-05, "loss": 7.487, "step": 19464 }, { "epoch": 0.8206674115142188, "grad_norm": 2.365769624710083, "learning_rate": 4.035884943314907e-05, "loss": 7.3395, "step": 19465 }, { "epoch": 0.8207095726964184, "grad_norm": 3.5126519203186035, "learning_rate": 4.0358068646741e-05, "loss": 7.6956, "step": 19466 }, { "epoch": 0.820751733878618, "grad_norm": 1.4357922077178955, "learning_rate": 4.0357287860332935e-05, "loss": 7.499, "step": 19467 }, { "epoch": 0.8207938950608175, "grad_norm": 2.598407030105591, "learning_rate": 4.035650707392486e-05, "loss": 6.8532, "step": 19468 }, { "epoch": 0.820836056243017, "grad_norm": 1.861307978630066, "learning_rate": 4.035572628751679e-05, "loss": 6.4772, "step": 19469 }, { "epoch": 0.8208782174252166, "grad_norm": 1.5218397378921509, "learning_rate": 4.035494550110872e-05, "loss": 6.7143, "step": 19470 }, { "epoch": 0.8209203786074162, "grad_norm": 1.227011799812317, "learning_rate": 4.0354164714700647e-05, "loss": 6.9354, "step": 19471 }, { "epoch": 0.8209625397896158, "grad_norm": 1.364139437675476, "learning_rate": 4.035338392829258e-05, "loss": 6.4486, "step": 19472 }, { "epoch": 0.8210047009718152, "grad_norm": 1.3160758018493652, "learning_rate": 4.035260314188451e-05, "loss": 6.5372, "step": 19473 }, { "epoch": 0.8210468621540148, "grad_norm": 3.3504889011383057, "learning_rate": 4.035182235547644e-05, "loss": 6.5045, "step": 19474 }, { "epoch": 0.8210890233362144, "grad_norm": 1.3613438606262207, "learning_rate": 4.0351041569068365e-05, "loss": 6.4819, "step": 19475 }, { "epoch": 0.8211311845184139, "grad_norm": 1.049895167350769, "learning_rate": 4.0350260782660296e-05, "loss": 6.7123, "step": 19476 }, { "epoch": 0.8211733457006134, "grad_norm": 3.101396083831787, "learning_rate": 4.034947999625223e-05, "loss": 7.4469, "step": 19477 }, { "epoch": 0.821215506882813, "grad_norm": 2.4177587032318115, "learning_rate": 4.034869920984416e-05, "loss": 6.6723, "step": 19478 }, { "epoch": 0.8212576680650125, "grad_norm": 2.0755624771118164, "learning_rate": 4.034791842343609e-05, "loss": 6.9367, "step": 19479 }, { "epoch": 0.8212998292472121, "grad_norm": 2.078967571258545, "learning_rate": 4.0347137637028014e-05, "loss": 7.0001, "step": 19480 }, { "epoch": 0.8213419904294117, "grad_norm": 1.8997877836227417, "learning_rate": 4.0346356850619946e-05, "loss": 7.2267, "step": 19481 }, { "epoch": 0.8213841516116112, "grad_norm": 1.2239912748336792, "learning_rate": 4.034557606421188e-05, "loss": 6.9729, "step": 19482 }, { "epoch": 0.8214263127938107, "grad_norm": 2.267681837081909, "learning_rate": 4.03447952778038e-05, "loss": 6.5743, "step": 19483 }, { "epoch": 0.8214684739760103, "grad_norm": 3.0159213542938232, "learning_rate": 4.034401449139574e-05, "loss": 7.652, "step": 19484 }, { "epoch": 0.8215106351582099, "grad_norm": 1.7753428220748901, "learning_rate": 4.0343233704987664e-05, "loss": 6.958, "step": 19485 }, { "epoch": 0.8215527963404093, "grad_norm": 1.264154314994812, "learning_rate": 4.0342452918579595e-05, "loss": 6.7901, "step": 19486 }, { "epoch": 0.8215949575226089, "grad_norm": 1.3989919424057007, "learning_rate": 4.0341672132171526e-05, "loss": 6.4409, "step": 19487 }, { "epoch": 0.8216371187048085, "grad_norm": 1.356552243232727, "learning_rate": 4.034089134576345e-05, "loss": 6.8626, "step": 19488 }, { "epoch": 0.8216792798870081, "grad_norm": 2.6819798946380615, "learning_rate": 4.034011055935539e-05, "loss": 7.476, "step": 19489 }, { "epoch": 0.8217214410692075, "grad_norm": 2.22355580329895, "learning_rate": 4.033932977294731e-05, "loss": 6.4534, "step": 19490 }, { "epoch": 0.8217636022514071, "grad_norm": 1.6004762649536133, "learning_rate": 4.0338548986539245e-05, "loss": 7.075, "step": 19491 }, { "epoch": 0.8218057634336067, "grad_norm": 1.6552507877349854, "learning_rate": 4.0337768200131176e-05, "loss": 6.6036, "step": 19492 }, { "epoch": 0.8218479246158062, "grad_norm": 1.3551803827285767, "learning_rate": 4.033698741372311e-05, "loss": 6.5313, "step": 19493 }, { "epoch": 0.8218900857980058, "grad_norm": 1.1807425022125244, "learning_rate": 4.033620662731503e-05, "loss": 6.539, "step": 19494 }, { "epoch": 0.8219322469802053, "grad_norm": 4.11145544052124, "learning_rate": 4.033542584090696e-05, "loss": 7.8228, "step": 19495 }, { "epoch": 0.8219744081624049, "grad_norm": 2.7464091777801514, "learning_rate": 4.0334645054498894e-05, "loss": 7.0066, "step": 19496 }, { "epoch": 0.8220165693446044, "grad_norm": 3.636841297149658, "learning_rate": 4.033386426809082e-05, "loss": 7.0719, "step": 19497 }, { "epoch": 0.822058730526804, "grad_norm": 3.4967198371887207, "learning_rate": 4.033308348168276e-05, "loss": 7.6721, "step": 19498 }, { "epoch": 0.8221008917090035, "grad_norm": 1.1713279485702515, "learning_rate": 4.033230269527468e-05, "loss": 6.7197, "step": 19499 }, { "epoch": 0.822143052891203, "grad_norm": 1.4828083515167236, "learning_rate": 4.033152190886661e-05, "loss": 6.6653, "step": 19500 }, { "epoch": 0.8221852140734026, "grad_norm": 2.291895866394043, "learning_rate": 4.0330741122458544e-05, "loss": 6.4488, "step": 19501 }, { "epoch": 0.8222273752556022, "grad_norm": 2.0575969219207764, "learning_rate": 4.032996033605047e-05, "loss": 6.8925, "step": 19502 }, { "epoch": 0.8222695364378018, "grad_norm": 1.6024279594421387, "learning_rate": 4.0329179549642406e-05, "loss": 6.6395, "step": 19503 }, { "epoch": 0.8223116976200012, "grad_norm": 1.44227135181427, "learning_rate": 4.032839876323433e-05, "loss": 6.3975, "step": 19504 }, { "epoch": 0.8223538588022008, "grad_norm": 3.281376361846924, "learning_rate": 4.032761797682626e-05, "loss": 7.7958, "step": 19505 }, { "epoch": 0.8223960199844004, "grad_norm": 3.429826259613037, "learning_rate": 4.032683719041819e-05, "loss": 7.0951, "step": 19506 }, { "epoch": 0.8224381811666, "grad_norm": 2.329847812652588, "learning_rate": 4.032605640401012e-05, "loss": 6.5802, "step": 19507 }, { "epoch": 0.8224803423487994, "grad_norm": 5.268624782562256, "learning_rate": 4.032527561760205e-05, "loss": 7.7983, "step": 19508 }, { "epoch": 0.822522503530999, "grad_norm": 4.186615467071533, "learning_rate": 4.032449483119398e-05, "loss": 9.0369, "step": 19509 }, { "epoch": 0.8225646647131986, "grad_norm": 0.9316962361335754, "learning_rate": 4.032371404478591e-05, "loss": 6.5506, "step": 19510 }, { "epoch": 0.8226068258953981, "grad_norm": 1.4759165048599243, "learning_rate": 4.0322933258377836e-05, "loss": 6.9917, "step": 19511 }, { "epoch": 0.8226489870775977, "grad_norm": 1.4517139196395874, "learning_rate": 4.0322152471969774e-05, "loss": 6.5058, "step": 19512 }, { "epoch": 0.8226911482597972, "grad_norm": 2.4907875061035156, "learning_rate": 4.03213716855617e-05, "loss": 7.4168, "step": 19513 }, { "epoch": 0.8227333094419967, "grad_norm": 2.4522340297698975, "learning_rate": 4.032059089915363e-05, "loss": 6.578, "step": 19514 }, { "epoch": 0.8227754706241963, "grad_norm": 2.336902379989624, "learning_rate": 4.031981011274556e-05, "loss": 6.4792, "step": 19515 }, { "epoch": 0.8228176318063959, "grad_norm": 1.958600401878357, "learning_rate": 4.0319029326337485e-05, "loss": 6.6313, "step": 19516 }, { "epoch": 0.8228597929885954, "grad_norm": 3.2166318893432617, "learning_rate": 4.0318248539929423e-05, "loss": 7.7068, "step": 19517 }, { "epoch": 0.8229019541707949, "grad_norm": 3.3430051803588867, "learning_rate": 4.031746775352135e-05, "loss": 7.8416, "step": 19518 }, { "epoch": 0.8229441153529945, "grad_norm": 1.5466687679290771, "learning_rate": 4.031668696711328e-05, "loss": 6.8669, "step": 19519 }, { "epoch": 0.8229862765351941, "grad_norm": 1.523526906967163, "learning_rate": 4.031590618070521e-05, "loss": 6.4535, "step": 19520 }, { "epoch": 0.8230284377173936, "grad_norm": 2.572045087814331, "learning_rate": 4.0315125394297135e-05, "loss": 7.2817, "step": 19521 }, { "epoch": 0.8230705988995931, "grad_norm": 1.6642142534255981, "learning_rate": 4.0314344607889066e-05, "loss": 6.4293, "step": 19522 }, { "epoch": 0.8231127600817927, "grad_norm": 1.9341813325881958, "learning_rate": 4.0313563821481e-05, "loss": 7.0554, "step": 19523 }, { "epoch": 0.8231549212639923, "grad_norm": 2.6727070808410645, "learning_rate": 4.031278303507293e-05, "loss": 7.8699, "step": 19524 }, { "epoch": 0.8231970824461918, "grad_norm": 2.7795345783233643, "learning_rate": 4.031200224866486e-05, "loss": 7.6259, "step": 19525 }, { "epoch": 0.8232392436283913, "grad_norm": 1.8595434427261353, "learning_rate": 4.0311221462256784e-05, "loss": 6.4752, "step": 19526 }, { "epoch": 0.8232814048105909, "grad_norm": 1.885728359222412, "learning_rate": 4.0310440675848716e-05, "loss": 7.0473, "step": 19527 }, { "epoch": 0.8233235659927904, "grad_norm": 2.8965325355529785, "learning_rate": 4.030965988944065e-05, "loss": 8.034, "step": 19528 }, { "epoch": 0.82336572717499, "grad_norm": 3.0791378021240234, "learning_rate": 4.030887910303258e-05, "loss": 7.5491, "step": 19529 }, { "epoch": 0.8234078883571896, "grad_norm": 3.3499433994293213, "learning_rate": 4.03080983166245e-05, "loss": 7.9887, "step": 19530 }, { "epoch": 0.823450049539389, "grad_norm": 2.0683302879333496, "learning_rate": 4.030731753021644e-05, "loss": 7.5442, "step": 19531 }, { "epoch": 0.8234922107215886, "grad_norm": 5.831654071807861, "learning_rate": 4.0306536743808365e-05, "loss": 7.3985, "step": 19532 }, { "epoch": 0.8235343719037882, "grad_norm": 3.1481566429138184, "learning_rate": 4.030575595740029e-05, "loss": 6.4504, "step": 19533 }, { "epoch": 0.8235765330859878, "grad_norm": 2.2869582176208496, "learning_rate": 4.030497517099223e-05, "loss": 7.0007, "step": 19534 }, { "epoch": 0.8236186942681872, "grad_norm": 2.2977914810180664, "learning_rate": 4.030419438458415e-05, "loss": 7.5705, "step": 19535 }, { "epoch": 0.8236608554503868, "grad_norm": 1.9005951881408691, "learning_rate": 4.030341359817609e-05, "loss": 6.8758, "step": 19536 }, { "epoch": 0.8237030166325864, "grad_norm": 2.187598943710327, "learning_rate": 4.0302632811768015e-05, "loss": 6.5264, "step": 19537 }, { "epoch": 0.823745177814786, "grad_norm": 2.3674774169921875, "learning_rate": 4.0301852025359946e-05, "loss": 7.9154, "step": 19538 }, { "epoch": 0.8237873389969855, "grad_norm": 2.5860347747802734, "learning_rate": 4.030107123895188e-05, "loss": 6.4724, "step": 19539 }, { "epoch": 0.823829500179185, "grad_norm": 3.0006942749023438, "learning_rate": 4.03002904525438e-05, "loss": 7.6018, "step": 19540 }, { "epoch": 0.8238716613613846, "grad_norm": 2.1113474369049072, "learning_rate": 4.029950966613573e-05, "loss": 7.3285, "step": 19541 }, { "epoch": 0.8239138225435841, "grad_norm": 1.566723346710205, "learning_rate": 4.0298728879727664e-05, "loss": 6.4719, "step": 19542 }, { "epoch": 0.8239559837257837, "grad_norm": 1.844672441482544, "learning_rate": 4.0297948093319596e-05, "loss": 7.0505, "step": 19543 }, { "epoch": 0.8239981449079832, "grad_norm": 1.329781174659729, "learning_rate": 4.029716730691152e-05, "loss": 6.6493, "step": 19544 }, { "epoch": 0.8240403060901828, "grad_norm": 1.5546274185180664, "learning_rate": 4.029638652050345e-05, "loss": 6.4362, "step": 19545 }, { "epoch": 0.8240824672723823, "grad_norm": 1.7494012117385864, "learning_rate": 4.029560573409538e-05, "loss": 7.119, "step": 19546 }, { "epoch": 0.8241246284545819, "grad_norm": 1.4770148992538452, "learning_rate": 4.0294824947687314e-05, "loss": 6.5332, "step": 19547 }, { "epoch": 0.8241667896367814, "grad_norm": 4.166578769683838, "learning_rate": 4.0294044161279245e-05, "loss": 7.6249, "step": 19548 }, { "epoch": 0.8242089508189809, "grad_norm": 4.597697734832764, "learning_rate": 4.029326337487117e-05, "loss": 6.8999, "step": 19549 }, { "epoch": 0.8242511120011805, "grad_norm": 3.0797483921051025, "learning_rate": 4.02924825884631e-05, "loss": 7.5936, "step": 19550 }, { "epoch": 0.8242932731833801, "grad_norm": 3.103644371032715, "learning_rate": 4.029170180205503e-05, "loss": 7.797, "step": 19551 }, { "epoch": 0.8243354343655797, "grad_norm": 4.1356072425842285, "learning_rate": 4.0290921015646957e-05, "loss": 7.5836, "step": 19552 }, { "epoch": 0.8243775955477791, "grad_norm": 3.1105756759643555, "learning_rate": 4.0290140229238895e-05, "loss": 7.6097, "step": 19553 }, { "epoch": 0.8244197567299787, "grad_norm": 1.7259784936904907, "learning_rate": 4.028935944283082e-05, "loss": 6.6811, "step": 19554 }, { "epoch": 0.8244619179121783, "grad_norm": 1.4190187454223633, "learning_rate": 4.028857865642275e-05, "loss": 6.9492, "step": 19555 }, { "epoch": 0.8245040790943778, "grad_norm": 1.365885615348816, "learning_rate": 4.028779787001468e-05, "loss": 6.6596, "step": 19556 }, { "epoch": 0.8245462402765773, "grad_norm": 2.210167407989502, "learning_rate": 4.028701708360661e-05, "loss": 6.4585, "step": 19557 }, { "epoch": 0.8245884014587769, "grad_norm": 2.031114101409912, "learning_rate": 4.0286236297198544e-05, "loss": 6.5787, "step": 19558 }, { "epoch": 0.8246305626409764, "grad_norm": 1.5038715600967407, "learning_rate": 4.028545551079047e-05, "loss": 6.455, "step": 19559 }, { "epoch": 0.824672723823176, "grad_norm": 2.4482204914093018, "learning_rate": 4.02846747243824e-05, "loss": 7.1996, "step": 19560 }, { "epoch": 0.8247148850053756, "grad_norm": 1.4292776584625244, "learning_rate": 4.028389393797433e-05, "loss": 6.8632, "step": 19561 }, { "epoch": 0.8247570461875751, "grad_norm": 1.7462329864501953, "learning_rate": 4.028311315156626e-05, "loss": 6.932, "step": 19562 }, { "epoch": 0.8247992073697746, "grad_norm": 2.189986228942871, "learning_rate": 4.028233236515819e-05, "loss": 7.1939, "step": 19563 }, { "epoch": 0.8248413685519742, "grad_norm": 1.5515108108520508, "learning_rate": 4.028155157875012e-05, "loss": 6.5707, "step": 19564 }, { "epoch": 0.8248835297341738, "grad_norm": 2.360879898071289, "learning_rate": 4.028077079234205e-05, "loss": 7.1415, "step": 19565 }, { "epoch": 0.8249256909163732, "grad_norm": 1.3753767013549805, "learning_rate": 4.0279990005933974e-05, "loss": 6.8558, "step": 19566 }, { "epoch": 0.8249678520985728, "grad_norm": 1.224266529083252, "learning_rate": 4.027920921952591e-05, "loss": 6.5202, "step": 19567 }, { "epoch": 0.8250100132807724, "grad_norm": 3.7952053546905518, "learning_rate": 4.0278428433117836e-05, "loss": 7.6276, "step": 19568 }, { "epoch": 0.825052174462972, "grad_norm": 4.066888332366943, "learning_rate": 4.027764764670977e-05, "loss": 7.7069, "step": 19569 }, { "epoch": 0.8250943356451715, "grad_norm": 1.3780485391616821, "learning_rate": 4.02768668603017e-05, "loss": 6.5589, "step": 19570 }, { "epoch": 0.825136496827371, "grad_norm": 1.947306752204895, "learning_rate": 4.0276086073893623e-05, "loss": 6.738, "step": 19571 }, { "epoch": 0.8251786580095706, "grad_norm": 1.4590257406234741, "learning_rate": 4.027530528748556e-05, "loss": 7.0997, "step": 19572 }, { "epoch": 0.8252208191917701, "grad_norm": 2.7505764961242676, "learning_rate": 4.0274524501077486e-05, "loss": 6.3517, "step": 19573 }, { "epoch": 0.8252629803739697, "grad_norm": 2.0757646560668945, "learning_rate": 4.027374371466942e-05, "loss": 7.1279, "step": 19574 }, { "epoch": 0.8253051415561692, "grad_norm": 1.3871979713439941, "learning_rate": 4.027296292826135e-05, "loss": 6.9469, "step": 19575 }, { "epoch": 0.8253473027383688, "grad_norm": 2.8583130836486816, "learning_rate": 4.027218214185327e-05, "loss": 7.5475, "step": 19576 }, { "epoch": 0.8253894639205683, "grad_norm": 1.1472628116607666, "learning_rate": 4.0271401355445204e-05, "loss": 6.6813, "step": 19577 }, { "epoch": 0.8254316251027679, "grad_norm": 1.0198760032653809, "learning_rate": 4.0270620569037135e-05, "loss": 6.4984, "step": 19578 }, { "epoch": 0.8254737862849675, "grad_norm": 2.810539484024048, "learning_rate": 4.026983978262907e-05, "loss": 7.6616, "step": 19579 }, { "epoch": 0.8255159474671669, "grad_norm": 1.339735746383667, "learning_rate": 4.026905899622099e-05, "loss": 6.8852, "step": 19580 }, { "epoch": 0.8255581086493665, "grad_norm": 1.751046061515808, "learning_rate": 4.026827820981293e-05, "loss": 7.0723, "step": 19581 }, { "epoch": 0.8256002698315661, "grad_norm": 1.3947733640670776, "learning_rate": 4.0267497423404854e-05, "loss": 6.751, "step": 19582 }, { "epoch": 0.8256424310137657, "grad_norm": 1.2039031982421875, "learning_rate": 4.0266716636996785e-05, "loss": 6.5251, "step": 19583 }, { "epoch": 0.8256845921959651, "grad_norm": 1.414925217628479, "learning_rate": 4.0265935850588716e-05, "loss": 6.8057, "step": 19584 }, { "epoch": 0.8257267533781647, "grad_norm": 1.2614787817001343, "learning_rate": 4.026515506418064e-05, "loss": 6.5937, "step": 19585 }, { "epoch": 0.8257689145603643, "grad_norm": 1.1365735530853271, "learning_rate": 4.026437427777258e-05, "loss": 6.5458, "step": 19586 }, { "epoch": 0.8258110757425638, "grad_norm": 1.233339786529541, "learning_rate": 4.02635934913645e-05, "loss": 6.6231, "step": 19587 }, { "epoch": 0.8258532369247634, "grad_norm": 2.0549697875976562, "learning_rate": 4.0262812704956435e-05, "loss": 7.1334, "step": 19588 }, { "epoch": 0.8258953981069629, "grad_norm": 2.782757043838501, "learning_rate": 4.0262031918548366e-05, "loss": 7.1567, "step": 19589 }, { "epoch": 0.8259375592891625, "grad_norm": 2.829052448272705, "learning_rate": 4.026125113214029e-05, "loss": 7.058, "step": 19590 }, { "epoch": 0.825979720471362, "grad_norm": 2.0102410316467285, "learning_rate": 4.026047034573222e-05, "loss": 6.7338, "step": 19591 }, { "epoch": 0.8260218816535616, "grad_norm": 2.3343710899353027, "learning_rate": 4.025968955932415e-05, "loss": 7.7237, "step": 19592 }, { "epoch": 0.8260640428357611, "grad_norm": 1.3336677551269531, "learning_rate": 4.0258908772916084e-05, "loss": 6.395, "step": 19593 }, { "epoch": 0.8261062040179606, "grad_norm": 2.302757978439331, "learning_rate": 4.0258127986508015e-05, "loss": 7.0686, "step": 19594 }, { "epoch": 0.8261483652001602, "grad_norm": 2.184785842895508, "learning_rate": 4.025734720009994e-05, "loss": 7.2081, "step": 19595 }, { "epoch": 0.8261905263823598, "grad_norm": 1.6990244388580322, "learning_rate": 4.025656641369187e-05, "loss": 7.2113, "step": 19596 }, { "epoch": 0.8262326875645594, "grad_norm": 1.53084135055542, "learning_rate": 4.02557856272838e-05, "loss": 7.158, "step": 19597 }, { "epoch": 0.8262748487467588, "grad_norm": 2.19496488571167, "learning_rate": 4.0255004840875734e-05, "loss": 7.0678, "step": 19598 }, { "epoch": 0.8263170099289584, "grad_norm": 3.144613027572632, "learning_rate": 4.025422405446766e-05, "loss": 7.2178, "step": 19599 }, { "epoch": 0.826359171111158, "grad_norm": 2.467198610305786, "learning_rate": 4.0253443268059596e-05, "loss": 7.8738, "step": 19600 }, { "epoch": 0.8264013322933575, "grad_norm": 1.2579671144485474, "learning_rate": 4.025266248165152e-05, "loss": 7.1414, "step": 19601 }, { "epoch": 0.826443493475557, "grad_norm": 2.061166524887085, "learning_rate": 4.025188169524345e-05, "loss": 6.451, "step": 19602 }, { "epoch": 0.8264856546577566, "grad_norm": 1.4093523025512695, "learning_rate": 4.025110090883538e-05, "loss": 7.3113, "step": 19603 }, { "epoch": 0.8265278158399562, "grad_norm": 1.6813548803329468, "learning_rate": 4.025032012242731e-05, "loss": 6.399, "step": 19604 }, { "epoch": 0.8265699770221557, "grad_norm": 1.7051271200180054, "learning_rate": 4.0249539336019246e-05, "loss": 6.9108, "step": 19605 }, { "epoch": 0.8266121382043552, "grad_norm": 1.2608418464660645, "learning_rate": 4.024875854961117e-05, "loss": 6.5676, "step": 19606 }, { "epoch": 0.8266542993865548, "grad_norm": 1.2018206119537354, "learning_rate": 4.02479777632031e-05, "loss": 6.928, "step": 19607 }, { "epoch": 0.8266964605687543, "grad_norm": 1.5466163158416748, "learning_rate": 4.024719697679503e-05, "loss": 6.4549, "step": 19608 }, { "epoch": 0.8267386217509539, "grad_norm": 1.778745412826538, "learning_rate": 4.024641619038696e-05, "loss": 6.5054, "step": 19609 }, { "epoch": 0.8267807829331535, "grad_norm": 1.1455011367797852, "learning_rate": 4.024563540397889e-05, "loss": 6.4321, "step": 19610 }, { "epoch": 0.826822944115353, "grad_norm": 1.300803780555725, "learning_rate": 4.024485461757082e-05, "loss": 6.9922, "step": 19611 }, { "epoch": 0.8268651052975525, "grad_norm": 1.591855525970459, "learning_rate": 4.024407383116275e-05, "loss": 6.6624, "step": 19612 }, { "epoch": 0.8269072664797521, "grad_norm": 2.2946901321411133, "learning_rate": 4.0243293044754675e-05, "loss": 7.1183, "step": 19613 }, { "epoch": 0.8269494276619517, "grad_norm": 0.9897766709327698, "learning_rate": 4.024251225834661e-05, "loss": 6.7125, "step": 19614 }, { "epoch": 0.8269915888441511, "grad_norm": 1.213238000869751, "learning_rate": 4.024173147193854e-05, "loss": 6.7899, "step": 19615 }, { "epoch": 0.8270337500263507, "grad_norm": 2.1648786067962646, "learning_rate": 4.024095068553047e-05, "loss": 7.0299, "step": 19616 }, { "epoch": 0.8270759112085503, "grad_norm": 3.678227663040161, "learning_rate": 4.02401698991224e-05, "loss": 7.2802, "step": 19617 }, { "epoch": 0.8271180723907499, "grad_norm": 1.7342413663864136, "learning_rate": 4.0239389112714325e-05, "loss": 6.7355, "step": 19618 }, { "epoch": 0.8271602335729494, "grad_norm": 4.889324188232422, "learning_rate": 4.023860832630626e-05, "loss": 9.8433, "step": 19619 }, { "epoch": 0.8272023947551489, "grad_norm": 1.3945820331573486, "learning_rate": 4.023782753989819e-05, "loss": 6.84, "step": 19620 }, { "epoch": 0.8272445559373485, "grad_norm": 1.468673825263977, "learning_rate": 4.023704675349011e-05, "loss": 7.0894, "step": 19621 }, { "epoch": 0.827286717119548, "grad_norm": 1.5593538284301758, "learning_rate": 4.023626596708205e-05, "loss": 6.4834, "step": 19622 }, { "epoch": 0.8273288783017476, "grad_norm": 1.2683253288269043, "learning_rate": 4.0235485180673974e-05, "loss": 7.0985, "step": 19623 }, { "epoch": 0.8273710394839471, "grad_norm": 1.524231195449829, "learning_rate": 4.0234704394265906e-05, "loss": 6.4352, "step": 19624 }, { "epoch": 0.8274132006661467, "grad_norm": 1.210436224937439, "learning_rate": 4.023392360785784e-05, "loss": 6.7094, "step": 19625 }, { "epoch": 0.8274553618483462, "grad_norm": 1.9745604991912842, "learning_rate": 4.023314282144977e-05, "loss": 7.0584, "step": 19626 }, { "epoch": 0.8274975230305458, "grad_norm": 1.8280808925628662, "learning_rate": 4.02323620350417e-05, "loss": 6.997, "step": 19627 }, { "epoch": 0.8275396842127454, "grad_norm": 3.8341243267059326, "learning_rate": 4.0231581248633624e-05, "loss": 7.8289, "step": 19628 }, { "epoch": 0.8275818453949448, "grad_norm": 1.5664852857589722, "learning_rate": 4.0230800462225555e-05, "loss": 6.7228, "step": 19629 }, { "epoch": 0.8276240065771444, "grad_norm": 1.2218623161315918, "learning_rate": 4.0230019675817486e-05, "loss": 6.7168, "step": 19630 }, { "epoch": 0.827666167759344, "grad_norm": 3.2554519176483154, "learning_rate": 4.022923888940942e-05, "loss": 6.9964, "step": 19631 }, { "epoch": 0.8277083289415436, "grad_norm": 1.6519991159439087, "learning_rate": 4.022845810300134e-05, "loss": 6.9191, "step": 19632 }, { "epoch": 0.827750490123743, "grad_norm": 1.3175814151763916, "learning_rate": 4.0227677316593273e-05, "loss": 6.7025, "step": 19633 }, { "epoch": 0.8277926513059426, "grad_norm": 1.4659825563430786, "learning_rate": 4.0226896530185205e-05, "loss": 6.4943, "step": 19634 }, { "epoch": 0.8278348124881422, "grad_norm": 1.4908859729766846, "learning_rate": 4.022611574377713e-05, "loss": 6.3637, "step": 19635 }, { "epoch": 0.8278769736703417, "grad_norm": 1.459091067314148, "learning_rate": 4.022533495736907e-05, "loss": 6.4874, "step": 19636 }, { "epoch": 0.8279191348525413, "grad_norm": 1.2203229665756226, "learning_rate": 4.022455417096099e-05, "loss": 6.7076, "step": 19637 }, { "epoch": 0.8279612960347408, "grad_norm": 1.6605925559997559, "learning_rate": 4.022377338455292e-05, "loss": 7.0605, "step": 19638 }, { "epoch": 0.8280034572169404, "grad_norm": 2.198540449142456, "learning_rate": 4.0222992598144854e-05, "loss": 7.1533, "step": 19639 }, { "epoch": 0.8280456183991399, "grad_norm": 0.965758740901947, "learning_rate": 4.022221181173678e-05, "loss": 6.586, "step": 19640 }, { "epoch": 0.8280877795813395, "grad_norm": 1.9682128429412842, "learning_rate": 4.022143102532872e-05, "loss": 7.1857, "step": 19641 }, { "epoch": 0.828129940763539, "grad_norm": 2.87554669380188, "learning_rate": 4.022065023892064e-05, "loss": 7.7564, "step": 19642 }, { "epoch": 0.8281721019457385, "grad_norm": 1.4910187721252441, "learning_rate": 4.021986945251257e-05, "loss": 6.5164, "step": 19643 }, { "epoch": 0.8282142631279381, "grad_norm": 1.7279659509658813, "learning_rate": 4.0219088666104504e-05, "loss": 6.749, "step": 19644 }, { "epoch": 0.8282564243101377, "grad_norm": 2.4835257530212402, "learning_rate": 4.0218307879696435e-05, "loss": 7.686, "step": 19645 }, { "epoch": 0.8282985854923373, "grad_norm": 1.074658751487732, "learning_rate": 4.021752709328836e-05, "loss": 6.8762, "step": 19646 }, { "epoch": 0.8283407466745367, "grad_norm": 1.3728482723236084, "learning_rate": 4.021674630688029e-05, "loss": 6.6847, "step": 19647 }, { "epoch": 0.8283829078567363, "grad_norm": 1.6750109195709229, "learning_rate": 4.021596552047222e-05, "loss": 6.388, "step": 19648 }, { "epoch": 0.8284250690389359, "grad_norm": 2.9342145919799805, "learning_rate": 4.021518473406415e-05, "loss": 7.4543, "step": 19649 }, { "epoch": 0.8284672302211354, "grad_norm": 1.1791597604751587, "learning_rate": 4.0214403947656085e-05, "loss": 6.7916, "step": 19650 }, { "epoch": 0.8285093914033349, "grad_norm": 1.6901084184646606, "learning_rate": 4.021362316124801e-05, "loss": 7.1849, "step": 19651 }, { "epoch": 0.8285515525855345, "grad_norm": 1.264434576034546, "learning_rate": 4.021284237483994e-05, "loss": 7.1035, "step": 19652 }, { "epoch": 0.828593713767734, "grad_norm": 1.9531984329223633, "learning_rate": 4.021206158843187e-05, "loss": 6.4753, "step": 19653 }, { "epoch": 0.8286358749499336, "grad_norm": 1.5852642059326172, "learning_rate": 4.0211280802023796e-05, "loss": 7.168, "step": 19654 }, { "epoch": 0.8286780361321332, "grad_norm": 1.9650274515151978, "learning_rate": 4.0210500015615734e-05, "loss": 7.3147, "step": 19655 }, { "epoch": 0.8287201973143327, "grad_norm": 2.2783913612365723, "learning_rate": 4.020971922920766e-05, "loss": 7.4852, "step": 19656 }, { "epoch": 0.8287623584965322, "grad_norm": 4.00213623046875, "learning_rate": 4.020893844279959e-05, "loss": 7.2904, "step": 19657 }, { "epoch": 0.8288045196787318, "grad_norm": 4.051742076873779, "learning_rate": 4.020815765639152e-05, "loss": 9.4844, "step": 19658 }, { "epoch": 0.8288466808609314, "grad_norm": 1.119821548461914, "learning_rate": 4.0207376869983446e-05, "loss": 6.6951, "step": 19659 }, { "epoch": 0.8288888420431308, "grad_norm": 3.931340217590332, "learning_rate": 4.020659608357538e-05, "loss": 7.5812, "step": 19660 }, { "epoch": 0.8289310032253304, "grad_norm": 2.0315821170806885, "learning_rate": 4.020581529716731e-05, "loss": 6.7978, "step": 19661 }, { "epoch": 0.82897316440753, "grad_norm": 2.5189197063446045, "learning_rate": 4.020503451075924e-05, "loss": 6.8476, "step": 19662 }, { "epoch": 0.8290153255897296, "grad_norm": 1.496397852897644, "learning_rate": 4.020425372435117e-05, "loss": 7.0158, "step": 19663 }, { "epoch": 0.829057486771929, "grad_norm": 4.695337772369385, "learning_rate": 4.02034729379431e-05, "loss": 8.4181, "step": 19664 }, { "epoch": 0.8290996479541286, "grad_norm": 1.263579249382019, "learning_rate": 4.0202692151535026e-05, "loss": 6.803, "step": 19665 }, { "epoch": 0.8291418091363282, "grad_norm": 1.693992257118225, "learning_rate": 4.020191136512696e-05, "loss": 6.4995, "step": 19666 }, { "epoch": 0.8291839703185278, "grad_norm": 1.0217787027359009, "learning_rate": 4.020113057871889e-05, "loss": 6.5427, "step": 19667 }, { "epoch": 0.8292261315007273, "grad_norm": 1.8943126201629639, "learning_rate": 4.020034979231081e-05, "loss": 7.0001, "step": 19668 }, { "epoch": 0.8292682926829268, "grad_norm": 3.909331798553467, "learning_rate": 4.019956900590275e-05, "loss": 8.127, "step": 19669 }, { "epoch": 0.8293104538651264, "grad_norm": 1.7009892463684082, "learning_rate": 4.0198788219494676e-05, "loss": 6.3865, "step": 19670 }, { "epoch": 0.8293526150473259, "grad_norm": 2.4933083057403564, "learning_rate": 4.019800743308661e-05, "loss": 7.7163, "step": 19671 }, { "epoch": 0.8293947762295255, "grad_norm": 2.119415760040283, "learning_rate": 4.019722664667854e-05, "loss": 7.3027, "step": 19672 }, { "epoch": 0.829436937411725, "grad_norm": 2.471632242202759, "learning_rate": 4.019644586027046e-05, "loss": 6.3579, "step": 19673 }, { "epoch": 0.8294790985939245, "grad_norm": 2.356665849685669, "learning_rate": 4.01956650738624e-05, "loss": 7.7742, "step": 19674 }, { "epoch": 0.8295212597761241, "grad_norm": 1.5716729164123535, "learning_rate": 4.0194884287454325e-05, "loss": 7.086, "step": 19675 }, { "epoch": 0.8295634209583237, "grad_norm": 1.777793526649475, "learning_rate": 4.019410350104626e-05, "loss": 6.6486, "step": 19676 }, { "epoch": 0.8296055821405233, "grad_norm": 2.501906394958496, "learning_rate": 4.019332271463819e-05, "loss": 7.5055, "step": 19677 }, { "epoch": 0.8296477433227227, "grad_norm": 0.9861357808113098, "learning_rate": 4.019254192823011e-05, "loss": 6.5006, "step": 19678 }, { "epoch": 0.8296899045049223, "grad_norm": 3.6507179737091064, "learning_rate": 4.0191761141822044e-05, "loss": 7.5912, "step": 19679 }, { "epoch": 0.8297320656871219, "grad_norm": 1.720718264579773, "learning_rate": 4.0190980355413975e-05, "loss": 6.8109, "step": 19680 }, { "epoch": 0.8297742268693215, "grad_norm": 1.1266072988510132, "learning_rate": 4.0190199569005906e-05, "loss": 6.4148, "step": 19681 }, { "epoch": 0.8298163880515209, "grad_norm": 4.0045647621154785, "learning_rate": 4.018941878259783e-05, "loss": 7.409, "step": 19682 }, { "epoch": 0.8298585492337205, "grad_norm": 1.0699427127838135, "learning_rate": 4.018863799618976e-05, "loss": 6.8227, "step": 19683 }, { "epoch": 0.8299007104159201, "grad_norm": 1.1589521169662476, "learning_rate": 4.018785720978169e-05, "loss": 6.7129, "step": 19684 }, { "epoch": 0.8299428715981196, "grad_norm": 1.243919849395752, "learning_rate": 4.0187076423373624e-05, "loss": 6.8699, "step": 19685 }, { "epoch": 0.8299850327803192, "grad_norm": 1.4529452323913574, "learning_rate": 4.0186295636965556e-05, "loss": 7.1379, "step": 19686 }, { "epoch": 0.8300271939625187, "grad_norm": 1.1840749979019165, "learning_rate": 4.018551485055748e-05, "loss": 7.1035, "step": 19687 }, { "epoch": 0.8300693551447182, "grad_norm": 3.065955877304077, "learning_rate": 4.018473406414942e-05, "loss": 6.441, "step": 19688 }, { "epoch": 0.8301115163269178, "grad_norm": 2.0365982055664062, "learning_rate": 4.018395327774134e-05, "loss": 6.4548, "step": 19689 }, { "epoch": 0.8301536775091174, "grad_norm": 1.4878292083740234, "learning_rate": 4.0183172491333274e-05, "loss": 6.8217, "step": 19690 }, { "epoch": 0.8301958386913169, "grad_norm": 4.235934257507324, "learning_rate": 4.0182391704925205e-05, "loss": 7.8531, "step": 19691 }, { "epoch": 0.8302379998735164, "grad_norm": 2.38565993309021, "learning_rate": 4.018161091851713e-05, "loss": 7.1468, "step": 19692 }, { "epoch": 0.830280161055716, "grad_norm": 1.4781148433685303, "learning_rate": 4.018083013210906e-05, "loss": 7.0623, "step": 19693 }, { "epoch": 0.8303223222379156, "grad_norm": 2.340061902999878, "learning_rate": 4.018004934570099e-05, "loss": 6.8449, "step": 19694 }, { "epoch": 0.8303644834201152, "grad_norm": 2.062387228012085, "learning_rate": 4.0179268559292923e-05, "loss": 6.485, "step": 19695 }, { "epoch": 0.8304066446023146, "grad_norm": 7.28950834274292, "learning_rate": 4.0178487772884855e-05, "loss": 7.1611, "step": 19696 }, { "epoch": 0.8304488057845142, "grad_norm": 2.5683679580688477, "learning_rate": 4.017770698647678e-05, "loss": 6.6007, "step": 19697 }, { "epoch": 0.8304909669667138, "grad_norm": 1.4456963539123535, "learning_rate": 4.017692620006871e-05, "loss": 6.5447, "step": 19698 }, { "epoch": 0.8305331281489133, "grad_norm": 1.1620069742202759, "learning_rate": 4.017614541366064e-05, "loss": 7.0614, "step": 19699 }, { "epoch": 0.8305752893311128, "grad_norm": 3.0954883098602295, "learning_rate": 4.017536462725257e-05, "loss": 7.3175, "step": 19700 }, { "epoch": 0.8306174505133124, "grad_norm": 2.097227096557617, "learning_rate": 4.01745838408445e-05, "loss": 7.097, "step": 19701 }, { "epoch": 0.830659611695512, "grad_norm": 2.210463285446167, "learning_rate": 4.017380305443643e-05, "loss": 6.9896, "step": 19702 }, { "epoch": 0.8307017728777115, "grad_norm": 1.6913384199142456, "learning_rate": 4.017302226802836e-05, "loss": 6.5092, "step": 19703 }, { "epoch": 0.8307439340599111, "grad_norm": 1.6455718278884888, "learning_rate": 4.0172241481620284e-05, "loss": 6.3286, "step": 19704 }, { "epoch": 0.8307860952421106, "grad_norm": 2.337690591812134, "learning_rate": 4.017146069521222e-05, "loss": 7.862, "step": 19705 }, { "epoch": 0.8308282564243101, "grad_norm": 1.650593638420105, "learning_rate": 4.017067990880415e-05, "loss": 7.1585, "step": 19706 }, { "epoch": 0.8308704176065097, "grad_norm": 1.5579943656921387, "learning_rate": 4.0169899122396085e-05, "loss": 7.086, "step": 19707 }, { "epoch": 0.8309125787887093, "grad_norm": 1.7209861278533936, "learning_rate": 4.016911833598801e-05, "loss": 6.6512, "step": 19708 }, { "epoch": 0.8309547399709087, "grad_norm": 1.6535725593566895, "learning_rate": 4.016833754957994e-05, "loss": 6.8612, "step": 19709 }, { "epoch": 0.8309969011531083, "grad_norm": 2.262803316116333, "learning_rate": 4.016755676317187e-05, "loss": 6.499, "step": 19710 }, { "epoch": 0.8310390623353079, "grad_norm": 1.290130376815796, "learning_rate": 4.0166775976763797e-05, "loss": 6.6348, "step": 19711 }, { "epoch": 0.8310812235175075, "grad_norm": 2.7822515964508057, "learning_rate": 4.016599519035573e-05, "loss": 7.589, "step": 19712 }, { "epoch": 0.831123384699707, "grad_norm": 4.342772960662842, "learning_rate": 4.016521440394766e-05, "loss": 7.6209, "step": 19713 }, { "epoch": 0.8311655458819065, "grad_norm": 0.9859791398048401, "learning_rate": 4.016443361753959e-05, "loss": 6.4801, "step": 19714 }, { "epoch": 0.8312077070641061, "grad_norm": 1.0315144062042236, "learning_rate": 4.0163652831131515e-05, "loss": 6.7222, "step": 19715 }, { "epoch": 0.8312498682463056, "grad_norm": 1.1658313274383545, "learning_rate": 4.0162872044723446e-05, "loss": 6.7896, "step": 19716 }, { "epoch": 0.8312920294285052, "grad_norm": 1.4970591068267822, "learning_rate": 4.016209125831538e-05, "loss": 7.0582, "step": 19717 }, { "epoch": 0.8313341906107047, "grad_norm": 2.9890220165252686, "learning_rate": 4.016131047190731e-05, "loss": 7.9235, "step": 19718 }, { "epoch": 0.8313763517929043, "grad_norm": 1.3652194738388062, "learning_rate": 4.016052968549924e-05, "loss": 6.4935, "step": 19719 }, { "epoch": 0.8314185129751038, "grad_norm": 1.0943714380264282, "learning_rate": 4.0159748899091164e-05, "loss": 6.6622, "step": 19720 }, { "epoch": 0.8314606741573034, "grad_norm": 2.865574836730957, "learning_rate": 4.0158968112683096e-05, "loss": 7.2212, "step": 19721 }, { "epoch": 0.8315028353395029, "grad_norm": 1.3822438716888428, "learning_rate": 4.015818732627503e-05, "loss": 7.4826, "step": 19722 }, { "epoch": 0.8315449965217024, "grad_norm": 3.099442720413208, "learning_rate": 4.015740653986695e-05, "loss": 8.5783, "step": 19723 }, { "epoch": 0.831587157703902, "grad_norm": 1.425683617591858, "learning_rate": 4.015662575345889e-05, "loss": 6.3875, "step": 19724 }, { "epoch": 0.8316293188861016, "grad_norm": 2.0451974868774414, "learning_rate": 4.0155844967050814e-05, "loss": 6.6623, "step": 19725 }, { "epoch": 0.8316714800683012, "grad_norm": 2.6914048194885254, "learning_rate": 4.0155064180642745e-05, "loss": 7.7464, "step": 19726 }, { "epoch": 0.8317136412505006, "grad_norm": 1.97482168674469, "learning_rate": 4.0154283394234676e-05, "loss": 6.7196, "step": 19727 }, { "epoch": 0.8317558024327002, "grad_norm": 1.7392255067825317, "learning_rate": 4.01535026078266e-05, "loss": 7.1932, "step": 19728 }, { "epoch": 0.8317979636148998, "grad_norm": 3.189180612564087, "learning_rate": 4.015272182141854e-05, "loss": 7.5891, "step": 19729 }, { "epoch": 0.8318401247970993, "grad_norm": 1.7722446918487549, "learning_rate": 4.015194103501046e-05, "loss": 6.4741, "step": 19730 }, { "epoch": 0.8318822859792988, "grad_norm": 1.287550449371338, "learning_rate": 4.0151160248602395e-05, "loss": 6.5577, "step": 19731 }, { "epoch": 0.8319244471614984, "grad_norm": 2.6451070308685303, "learning_rate": 4.0150379462194326e-05, "loss": 6.4662, "step": 19732 }, { "epoch": 0.831966608343698, "grad_norm": 0.9938473701477051, "learning_rate": 4.014959867578626e-05, "loss": 6.4941, "step": 19733 }, { "epoch": 0.8320087695258975, "grad_norm": 3.978428363800049, "learning_rate": 4.014881788937818e-05, "loss": 7.6663, "step": 19734 }, { "epoch": 0.8320509307080971, "grad_norm": 3.757361888885498, "learning_rate": 4.014803710297011e-05, "loss": 7.7286, "step": 19735 }, { "epoch": 0.8320930918902966, "grad_norm": 2.793517827987671, "learning_rate": 4.0147256316562044e-05, "loss": 7.0215, "step": 19736 }, { "epoch": 0.8321352530724961, "grad_norm": 1.8295034170150757, "learning_rate": 4.014647553015397e-05, "loss": 6.786, "step": 19737 }, { "epoch": 0.8321774142546957, "grad_norm": 3.0951833724975586, "learning_rate": 4.014569474374591e-05, "loss": 7.2233, "step": 19738 }, { "epoch": 0.8322195754368953, "grad_norm": 4.697417736053467, "learning_rate": 4.014491395733783e-05, "loss": 10.2979, "step": 19739 }, { "epoch": 0.8322617366190947, "grad_norm": 1.1853188276290894, "learning_rate": 4.014413317092976e-05, "loss": 6.8571, "step": 19740 }, { "epoch": 0.8323038978012943, "grad_norm": 2.1136205196380615, "learning_rate": 4.0143352384521694e-05, "loss": 7.1128, "step": 19741 }, { "epoch": 0.8323460589834939, "grad_norm": 1.9878249168395996, "learning_rate": 4.014257159811362e-05, "loss": 6.9225, "step": 19742 }, { "epoch": 0.8323882201656935, "grad_norm": 1.436623454093933, "learning_rate": 4.0141790811705556e-05, "loss": 6.4909, "step": 19743 }, { "epoch": 0.832430381347893, "grad_norm": 2.3091440200805664, "learning_rate": 4.014101002529748e-05, "loss": 7.2544, "step": 19744 }, { "epoch": 0.8324725425300925, "grad_norm": 2.388078212738037, "learning_rate": 4.014022923888941e-05, "loss": 6.9155, "step": 19745 }, { "epoch": 0.8325147037122921, "grad_norm": 2.2558441162109375, "learning_rate": 4.013944845248134e-05, "loss": 6.6922, "step": 19746 }, { "epoch": 0.8325568648944917, "grad_norm": 1.6653547286987305, "learning_rate": 4.013866766607327e-05, "loss": 6.8557, "step": 19747 }, { "epoch": 0.8325990260766912, "grad_norm": 1.7248879671096802, "learning_rate": 4.01378868796652e-05, "loss": 6.7551, "step": 19748 }, { "epoch": 0.8326411872588907, "grad_norm": 2.833758592605591, "learning_rate": 4.013710609325713e-05, "loss": 7.9779, "step": 19749 }, { "epoch": 0.8326833484410903, "grad_norm": 4.074546813964844, "learning_rate": 4.013632530684906e-05, "loss": 7.8479, "step": 19750 }, { "epoch": 0.8327255096232898, "grad_norm": 1.381547451019287, "learning_rate": 4.0135544520440986e-05, "loss": 6.6078, "step": 19751 }, { "epoch": 0.8327676708054894, "grad_norm": 1.6241066455841064, "learning_rate": 4.0134763734032924e-05, "loss": 6.4235, "step": 19752 }, { "epoch": 0.832809831987689, "grad_norm": 0.8176561594009399, "learning_rate": 4.013398294762485e-05, "loss": 6.5623, "step": 19753 }, { "epoch": 0.8328519931698884, "grad_norm": 1.071143388748169, "learning_rate": 4.013320216121678e-05, "loss": 6.6277, "step": 19754 }, { "epoch": 0.832894154352088, "grad_norm": 1.3010635375976562, "learning_rate": 4.013242137480871e-05, "loss": 6.653, "step": 19755 }, { "epoch": 0.8329363155342876, "grad_norm": 1.3709617853164673, "learning_rate": 4.0131640588400635e-05, "loss": 6.4956, "step": 19756 }, { "epoch": 0.8329784767164872, "grad_norm": 2.632230758666992, "learning_rate": 4.0130859801992573e-05, "loss": 7.3489, "step": 19757 }, { "epoch": 0.8330206378986866, "grad_norm": 3.773951768875122, "learning_rate": 4.01300790155845e-05, "loss": 7.6766, "step": 19758 }, { "epoch": 0.8330627990808862, "grad_norm": 3.652280807495117, "learning_rate": 4.012929822917643e-05, "loss": 7.6469, "step": 19759 }, { "epoch": 0.8331049602630858, "grad_norm": 3.004822254180908, "learning_rate": 4.012851744276836e-05, "loss": 7.4628, "step": 19760 }, { "epoch": 0.8331471214452854, "grad_norm": 2.5054147243499756, "learning_rate": 4.0127736656360285e-05, "loss": 7.3746, "step": 19761 }, { "epoch": 0.8331892826274849, "grad_norm": 1.114397406578064, "learning_rate": 4.0126955869952216e-05, "loss": 6.5628, "step": 19762 }, { "epoch": 0.8332314438096844, "grad_norm": 0.9996216297149658, "learning_rate": 4.012617508354415e-05, "loss": 6.54, "step": 19763 }, { "epoch": 0.833273604991884, "grad_norm": 0.7945232391357422, "learning_rate": 4.012539429713608e-05, "loss": 6.5427, "step": 19764 }, { "epoch": 0.8333157661740835, "grad_norm": 4.056528568267822, "learning_rate": 4.012461351072801e-05, "loss": 7.7952, "step": 19765 }, { "epoch": 0.8333579273562831, "grad_norm": 0.8655804991722107, "learning_rate": 4.0123832724319934e-05, "loss": 6.6317, "step": 19766 }, { "epoch": 0.8334000885384826, "grad_norm": 1.0210824012756348, "learning_rate": 4.0123051937911866e-05, "loss": 6.8582, "step": 19767 }, { "epoch": 0.8334422497206821, "grad_norm": 2.5810210704803467, "learning_rate": 4.01222711515038e-05, "loss": 6.7517, "step": 19768 }, { "epoch": 0.8334844109028817, "grad_norm": 1.5534404516220093, "learning_rate": 4.012149036509573e-05, "loss": 6.972, "step": 19769 }, { "epoch": 0.8335265720850813, "grad_norm": 1.2726777791976929, "learning_rate": 4.012070957868765e-05, "loss": 6.6072, "step": 19770 }, { "epoch": 0.8335687332672808, "grad_norm": 1.3987594842910767, "learning_rate": 4.011992879227959e-05, "loss": 6.9589, "step": 19771 }, { "epoch": 0.8336108944494803, "grad_norm": 0.9799843430519104, "learning_rate": 4.0119148005871515e-05, "loss": 6.4864, "step": 19772 }, { "epoch": 0.8336530556316799, "grad_norm": 1.3053691387176514, "learning_rate": 4.011836721946344e-05, "loss": 6.8987, "step": 19773 }, { "epoch": 0.8336952168138795, "grad_norm": 2.1036159992218018, "learning_rate": 4.011758643305538e-05, "loss": 6.8922, "step": 19774 }, { "epoch": 0.833737377996079, "grad_norm": 1.2491321563720703, "learning_rate": 4.01168056466473e-05, "loss": 6.961, "step": 19775 }, { "epoch": 0.8337795391782785, "grad_norm": 1.44137442111969, "learning_rate": 4.011602486023924e-05, "loss": 7.1044, "step": 19776 }, { "epoch": 0.8338217003604781, "grad_norm": 1.2785183191299438, "learning_rate": 4.0115244073831165e-05, "loss": 6.3812, "step": 19777 }, { "epoch": 0.8338638615426777, "grad_norm": 1.5357446670532227, "learning_rate": 4.0114463287423096e-05, "loss": 7.2552, "step": 19778 }, { "epoch": 0.8339060227248772, "grad_norm": 2.3274641036987305, "learning_rate": 4.011368250101503e-05, "loss": 6.93, "step": 19779 }, { "epoch": 0.8339481839070767, "grad_norm": 2.2606163024902344, "learning_rate": 4.011290171460695e-05, "loss": 7.6802, "step": 19780 }, { "epoch": 0.8339903450892763, "grad_norm": 3.646569013595581, "learning_rate": 4.011212092819888e-05, "loss": 7.6323, "step": 19781 }, { "epoch": 0.8340325062714758, "grad_norm": 3.4857282638549805, "learning_rate": 4.0111340141790814e-05, "loss": 7.9435, "step": 19782 }, { "epoch": 0.8340746674536754, "grad_norm": 2.1415224075317383, "learning_rate": 4.0110559355382746e-05, "loss": 6.962, "step": 19783 }, { "epoch": 0.834116828635875, "grad_norm": 2.1281518936157227, "learning_rate": 4.010977856897467e-05, "loss": 6.4936, "step": 19784 }, { "epoch": 0.8341589898180745, "grad_norm": 1.9081965684890747, "learning_rate": 4.01089977825666e-05, "loss": 6.507, "step": 19785 }, { "epoch": 0.834201151000274, "grad_norm": 1.3170554637908936, "learning_rate": 4.010821699615853e-05, "loss": 6.5561, "step": 19786 }, { "epoch": 0.8342433121824736, "grad_norm": 1.1646013259887695, "learning_rate": 4.0107436209750464e-05, "loss": 6.724, "step": 19787 }, { "epoch": 0.8342854733646732, "grad_norm": 2.54502272605896, "learning_rate": 4.0106655423342395e-05, "loss": 7.08, "step": 19788 }, { "epoch": 0.8343276345468726, "grad_norm": 3.3658533096313477, "learning_rate": 4.010587463693432e-05, "loss": 7.0343, "step": 19789 }, { "epoch": 0.8343697957290722, "grad_norm": 2.569106340408325, "learning_rate": 4.010509385052625e-05, "loss": 7.0801, "step": 19790 }, { "epoch": 0.8344119569112718, "grad_norm": 1.305553913116455, "learning_rate": 4.010431306411818e-05, "loss": 6.6834, "step": 19791 }, { "epoch": 0.8344541180934714, "grad_norm": 2.1901257038116455, "learning_rate": 4.0103532277710107e-05, "loss": 7.4089, "step": 19792 }, { "epoch": 0.8344962792756709, "grad_norm": 2.244382619857788, "learning_rate": 4.0102751491302045e-05, "loss": 6.5947, "step": 19793 }, { "epoch": 0.8345384404578704, "grad_norm": 2.3125579357147217, "learning_rate": 4.010197070489397e-05, "loss": 6.4321, "step": 19794 }, { "epoch": 0.83458060164007, "grad_norm": 1.8157289028167725, "learning_rate": 4.01011899184859e-05, "loss": 7.1158, "step": 19795 }, { "epoch": 0.8346227628222695, "grad_norm": 1.8643836975097656, "learning_rate": 4.010040913207783e-05, "loss": 6.5648, "step": 19796 }, { "epoch": 0.8346649240044691, "grad_norm": 1.7315653562545776, "learning_rate": 4.009962834566976e-05, "loss": 6.56, "step": 19797 }, { "epoch": 0.8347070851866686, "grad_norm": 0.9705846905708313, "learning_rate": 4.0098847559261694e-05, "loss": 6.5673, "step": 19798 }, { "epoch": 0.8347492463688682, "grad_norm": 3.562216281890869, "learning_rate": 4.009806677285362e-05, "loss": 7.254, "step": 19799 }, { "epoch": 0.8347914075510677, "grad_norm": 2.9883933067321777, "learning_rate": 4.009728598644555e-05, "loss": 7.6233, "step": 19800 }, { "epoch": 0.8348335687332673, "grad_norm": 1.8923689126968384, "learning_rate": 4.009650520003748e-05, "loss": 6.7184, "step": 19801 }, { "epoch": 0.8348757299154669, "grad_norm": 1.7240781784057617, "learning_rate": 4.009572441362941e-05, "loss": 6.8773, "step": 19802 }, { "epoch": 0.8349178910976663, "grad_norm": 1.14641273021698, "learning_rate": 4.009494362722134e-05, "loss": 6.6972, "step": 19803 }, { "epoch": 0.8349600522798659, "grad_norm": 2.307234287261963, "learning_rate": 4.009416284081327e-05, "loss": 7.2912, "step": 19804 }, { "epoch": 0.8350022134620655, "grad_norm": 1.7363516092300415, "learning_rate": 4.00933820544052e-05, "loss": 6.611, "step": 19805 }, { "epoch": 0.8350443746442651, "grad_norm": 1.8956552743911743, "learning_rate": 4.0092601267997124e-05, "loss": 7.0588, "step": 19806 }, { "epoch": 0.8350865358264645, "grad_norm": 1.5747785568237305, "learning_rate": 4.009182048158906e-05, "loss": 6.5613, "step": 19807 }, { "epoch": 0.8351286970086641, "grad_norm": 1.4228445291519165, "learning_rate": 4.0091039695180986e-05, "loss": 6.7428, "step": 19808 }, { "epoch": 0.8351708581908637, "grad_norm": 2.534359931945801, "learning_rate": 4.009025890877292e-05, "loss": 7.3787, "step": 19809 }, { "epoch": 0.8352130193730632, "grad_norm": 4.341283321380615, "learning_rate": 4.008947812236485e-05, "loss": 7.6832, "step": 19810 }, { "epoch": 0.8352551805552628, "grad_norm": 2.0210628509521484, "learning_rate": 4.0088697335956773e-05, "loss": 6.4597, "step": 19811 }, { "epoch": 0.8352973417374623, "grad_norm": 1.179560661315918, "learning_rate": 4.008791654954871e-05, "loss": 6.6389, "step": 19812 }, { "epoch": 0.8353395029196619, "grad_norm": 1.4959050416946411, "learning_rate": 4.0087135763140636e-05, "loss": 7.1426, "step": 19813 }, { "epoch": 0.8353816641018614, "grad_norm": 3.824347734451294, "learning_rate": 4.008635497673257e-05, "loss": 7.814, "step": 19814 }, { "epoch": 0.835423825284061, "grad_norm": 1.9746568202972412, "learning_rate": 4.00855741903245e-05, "loss": 7.0513, "step": 19815 }, { "epoch": 0.8354659864662605, "grad_norm": 1.6378766298294067, "learning_rate": 4.008479340391643e-05, "loss": 6.578, "step": 19816 }, { "epoch": 0.83550814764846, "grad_norm": 1.0842785835266113, "learning_rate": 4.0084012617508354e-05, "loss": 6.7308, "step": 19817 }, { "epoch": 0.8355503088306596, "grad_norm": 2.695225477218628, "learning_rate": 4.0083231831100285e-05, "loss": 7.9467, "step": 19818 }, { "epoch": 0.8355924700128592, "grad_norm": 0.8646688461303711, "learning_rate": 4.008245104469222e-05, "loss": 6.8006, "step": 19819 }, { "epoch": 0.8356346311950588, "grad_norm": 1.13680899143219, "learning_rate": 4.008167025828414e-05, "loss": 6.7299, "step": 19820 }, { "epoch": 0.8356767923772582, "grad_norm": 1.850051999092102, "learning_rate": 4.008088947187608e-05, "loss": 7.0692, "step": 19821 }, { "epoch": 0.8357189535594578, "grad_norm": 1.1287763118743896, "learning_rate": 4.0080108685468004e-05, "loss": 6.6252, "step": 19822 }, { "epoch": 0.8357611147416574, "grad_norm": 3.1788887977600098, "learning_rate": 4.0079327899059935e-05, "loss": 7.6583, "step": 19823 }, { "epoch": 0.835803275923857, "grad_norm": 1.8644095659255981, "learning_rate": 4.0078547112651866e-05, "loss": 6.5767, "step": 19824 }, { "epoch": 0.8358454371060564, "grad_norm": 1.7961311340332031, "learning_rate": 4.007776632624379e-05, "loss": 7.0995, "step": 19825 }, { "epoch": 0.835887598288256, "grad_norm": 3.415325164794922, "learning_rate": 4.007698553983573e-05, "loss": 7.7524, "step": 19826 }, { "epoch": 0.8359297594704556, "grad_norm": 1.6887760162353516, "learning_rate": 4.007620475342765e-05, "loss": 6.6997, "step": 19827 }, { "epoch": 0.8359719206526551, "grad_norm": 3.2266769409179688, "learning_rate": 4.0075423967019585e-05, "loss": 6.9899, "step": 19828 }, { "epoch": 0.8360140818348546, "grad_norm": 0.9219385385513306, "learning_rate": 4.0074643180611516e-05, "loss": 6.6741, "step": 19829 }, { "epoch": 0.8360562430170542, "grad_norm": 1.5072730779647827, "learning_rate": 4.007386239420344e-05, "loss": 6.7433, "step": 19830 }, { "epoch": 0.8360984041992537, "grad_norm": 1.9490810632705688, "learning_rate": 4.007308160779537e-05, "loss": 7.0235, "step": 19831 }, { "epoch": 0.8361405653814533, "grad_norm": 1.2741693258285522, "learning_rate": 4.00723008213873e-05, "loss": 6.6504, "step": 19832 }, { "epoch": 0.8361827265636529, "grad_norm": 2.287804126739502, "learning_rate": 4.0071520034979234e-05, "loss": 6.9324, "step": 19833 }, { "epoch": 0.8362248877458524, "grad_norm": 1.2302680015563965, "learning_rate": 4.0070739248571165e-05, "loss": 6.4614, "step": 19834 }, { "epoch": 0.8362670489280519, "grad_norm": 2.591322898864746, "learning_rate": 4.006995846216309e-05, "loss": 6.934, "step": 19835 }, { "epoch": 0.8363092101102515, "grad_norm": 4.076878070831299, "learning_rate": 4.006917767575502e-05, "loss": 8.1467, "step": 19836 }, { "epoch": 0.8363513712924511, "grad_norm": 0.9193345308303833, "learning_rate": 4.006839688934695e-05, "loss": 6.966, "step": 19837 }, { "epoch": 0.8363935324746505, "grad_norm": 2.9050967693328857, "learning_rate": 4.0067616102938884e-05, "loss": 6.418, "step": 19838 }, { "epoch": 0.8364356936568501, "grad_norm": 0.919622540473938, "learning_rate": 4.006683531653081e-05, "loss": 6.668, "step": 19839 }, { "epoch": 0.8364778548390497, "grad_norm": 3.331303834915161, "learning_rate": 4.0066054530122746e-05, "loss": 7.5019, "step": 19840 }, { "epoch": 0.8365200160212493, "grad_norm": 0.9574476480484009, "learning_rate": 4.006527374371467e-05, "loss": 6.9313, "step": 19841 }, { "epoch": 0.8365621772034488, "grad_norm": 1.1535823345184326, "learning_rate": 4.00644929573066e-05, "loss": 6.5244, "step": 19842 }, { "epoch": 0.8366043383856483, "grad_norm": 1.1609715223312378, "learning_rate": 4.006371217089853e-05, "loss": 7.0097, "step": 19843 }, { "epoch": 0.8366464995678479, "grad_norm": 1.7284153699874878, "learning_rate": 4.006293138449046e-05, "loss": 7.1734, "step": 19844 }, { "epoch": 0.8366886607500474, "grad_norm": 1.186653733253479, "learning_rate": 4.0062150598082396e-05, "loss": 7.097, "step": 19845 }, { "epoch": 0.836730821932247, "grad_norm": 1.421018123626709, "learning_rate": 4.006136981167432e-05, "loss": 7.317, "step": 19846 }, { "epoch": 0.8367729831144465, "grad_norm": 2.604339838027954, "learning_rate": 4.006058902526625e-05, "loss": 7.5256, "step": 19847 }, { "epoch": 0.836815144296646, "grad_norm": 1.7078876495361328, "learning_rate": 4.005980823885818e-05, "loss": 7.0635, "step": 19848 }, { "epoch": 0.8368573054788456, "grad_norm": 1.02138352394104, "learning_rate": 4.005902745245011e-05, "loss": 6.4692, "step": 19849 }, { "epoch": 0.8368994666610452, "grad_norm": 1.4597266912460327, "learning_rate": 4.005824666604204e-05, "loss": 6.958, "step": 19850 }, { "epoch": 0.8369416278432448, "grad_norm": 1.3017932176589966, "learning_rate": 4.005746587963397e-05, "loss": 6.5033, "step": 19851 }, { "epoch": 0.8369837890254442, "grad_norm": 1.5153610706329346, "learning_rate": 4.00566850932259e-05, "loss": 6.5217, "step": 19852 }, { "epoch": 0.8370259502076438, "grad_norm": 2.664844274520874, "learning_rate": 4.0055904306817825e-05, "loss": 7.51, "step": 19853 }, { "epoch": 0.8370681113898434, "grad_norm": 1.1342564821243286, "learning_rate": 4.005512352040976e-05, "loss": 6.7246, "step": 19854 }, { "epoch": 0.837110272572043, "grad_norm": 1.8082945346832275, "learning_rate": 4.005434273400169e-05, "loss": 7.8496, "step": 19855 }, { "epoch": 0.8371524337542424, "grad_norm": 1.6340175867080688, "learning_rate": 4.005356194759362e-05, "loss": 6.5172, "step": 19856 }, { "epoch": 0.837194594936442, "grad_norm": 5.715195655822754, "learning_rate": 4.005278116118555e-05, "loss": 7.0831, "step": 19857 }, { "epoch": 0.8372367561186416, "grad_norm": 2.5981345176696777, "learning_rate": 4.0052000374777475e-05, "loss": 7.3879, "step": 19858 }, { "epoch": 0.8372789173008411, "grad_norm": 2.079756736755371, "learning_rate": 4.005121958836941e-05, "loss": 6.3652, "step": 19859 }, { "epoch": 0.8373210784830407, "grad_norm": 3.4772510528564453, "learning_rate": 4.005043880196134e-05, "loss": 7.6901, "step": 19860 }, { "epoch": 0.8373632396652402, "grad_norm": 1.4499762058258057, "learning_rate": 4.004965801555326e-05, "loss": 6.9838, "step": 19861 }, { "epoch": 0.8374054008474398, "grad_norm": 3.011542797088623, "learning_rate": 4.00488772291452e-05, "loss": 8.1792, "step": 19862 }, { "epoch": 0.8374475620296393, "grad_norm": 1.197291612625122, "learning_rate": 4.0048096442737124e-05, "loss": 6.992, "step": 19863 }, { "epoch": 0.8374897232118389, "grad_norm": 1.413246750831604, "learning_rate": 4.0047315656329056e-05, "loss": 6.7617, "step": 19864 }, { "epoch": 0.8375318843940384, "grad_norm": 1.035004734992981, "learning_rate": 4.004653486992099e-05, "loss": 6.4246, "step": 19865 }, { "epoch": 0.8375740455762379, "grad_norm": 3.715366840362549, "learning_rate": 4.004575408351292e-05, "loss": 7.6617, "step": 19866 }, { "epoch": 0.8376162067584375, "grad_norm": 1.4626262187957764, "learning_rate": 4.004497329710485e-05, "loss": 6.9552, "step": 19867 }, { "epoch": 0.8376583679406371, "grad_norm": 1.0661547183990479, "learning_rate": 4.0044192510696774e-05, "loss": 7.0357, "step": 19868 }, { "epoch": 0.8377005291228367, "grad_norm": 1.6743704080581665, "learning_rate": 4.0043411724288705e-05, "loss": 6.4303, "step": 19869 }, { "epoch": 0.8377426903050361, "grad_norm": 2.162963628768921, "learning_rate": 4.0042630937880636e-05, "loss": 6.5708, "step": 19870 }, { "epoch": 0.8377848514872357, "grad_norm": 2.819650411605835, "learning_rate": 4.004185015147257e-05, "loss": 7.6465, "step": 19871 }, { "epoch": 0.8378270126694353, "grad_norm": 3.499074935913086, "learning_rate": 4.004106936506449e-05, "loss": 8.6313, "step": 19872 }, { "epoch": 0.8378691738516348, "grad_norm": 1.6457926034927368, "learning_rate": 4.0040288578656423e-05, "loss": 6.5215, "step": 19873 }, { "epoch": 0.8379113350338343, "grad_norm": 1.2602211236953735, "learning_rate": 4.0039507792248355e-05, "loss": 6.7096, "step": 19874 }, { "epoch": 0.8379534962160339, "grad_norm": 2.763700246810913, "learning_rate": 4.003872700584028e-05, "loss": 7.0597, "step": 19875 }, { "epoch": 0.8379956573982335, "grad_norm": 0.9445474743843079, "learning_rate": 4.003794621943222e-05, "loss": 6.7873, "step": 19876 }, { "epoch": 0.838037818580433, "grad_norm": 1.7330104112625122, "learning_rate": 4.003716543302414e-05, "loss": 6.4314, "step": 19877 }, { "epoch": 0.8380799797626326, "grad_norm": 1.1391733884811401, "learning_rate": 4.003638464661607e-05, "loss": 6.57, "step": 19878 }, { "epoch": 0.8381221409448321, "grad_norm": 1.5313184261322021, "learning_rate": 4.0035603860208004e-05, "loss": 7.1192, "step": 19879 }, { "epoch": 0.8381643021270316, "grad_norm": 1.0109918117523193, "learning_rate": 4.003482307379993e-05, "loss": 6.6349, "step": 19880 }, { "epoch": 0.8382064633092312, "grad_norm": 2.4870104789733887, "learning_rate": 4.003404228739187e-05, "loss": 7.4451, "step": 19881 }, { "epoch": 0.8382486244914308, "grad_norm": 1.1443625688552856, "learning_rate": 4.003326150098379e-05, "loss": 6.7924, "step": 19882 }, { "epoch": 0.8382907856736302, "grad_norm": 4.311047077178955, "learning_rate": 4.003248071457572e-05, "loss": 8.1719, "step": 19883 }, { "epoch": 0.8383329468558298, "grad_norm": 5.087953090667725, "learning_rate": 4.0031699928167654e-05, "loss": 8.3976, "step": 19884 }, { "epoch": 0.8383751080380294, "grad_norm": 2.539414644241333, "learning_rate": 4.0030919141759585e-05, "loss": 7.7787, "step": 19885 }, { "epoch": 0.838417269220229, "grad_norm": 1.372940182685852, "learning_rate": 4.003013835535151e-05, "loss": 6.9286, "step": 19886 }, { "epoch": 0.8384594304024284, "grad_norm": 1.6375477313995361, "learning_rate": 4.002935756894344e-05, "loss": 6.9505, "step": 19887 }, { "epoch": 0.838501591584628, "grad_norm": 3.632343053817749, "learning_rate": 4.002857678253537e-05, "loss": 7.1083, "step": 19888 }, { "epoch": 0.8385437527668276, "grad_norm": 1.799448847770691, "learning_rate": 4.0027795996127297e-05, "loss": 6.6598, "step": 19889 }, { "epoch": 0.8385859139490272, "grad_norm": 1.8085451126098633, "learning_rate": 4.0027015209719235e-05, "loss": 7.0974, "step": 19890 }, { "epoch": 0.8386280751312267, "grad_norm": 2.640404462814331, "learning_rate": 4.002623442331116e-05, "loss": 6.5585, "step": 19891 }, { "epoch": 0.8386702363134262, "grad_norm": 2.2205522060394287, "learning_rate": 4.002545363690309e-05, "loss": 7.024, "step": 19892 }, { "epoch": 0.8387123974956258, "grad_norm": 2.5675618648529053, "learning_rate": 4.002467285049502e-05, "loss": 7.1113, "step": 19893 }, { "epoch": 0.8387545586778253, "grad_norm": 1.7421151399612427, "learning_rate": 4.0023892064086946e-05, "loss": 6.8581, "step": 19894 }, { "epoch": 0.8387967198600249, "grad_norm": 1.8968479633331299, "learning_rate": 4.0023111277678884e-05, "loss": 6.8827, "step": 19895 }, { "epoch": 0.8388388810422244, "grad_norm": 1.7581325769424438, "learning_rate": 4.002233049127081e-05, "loss": 6.5322, "step": 19896 }, { "epoch": 0.838881042224424, "grad_norm": 0.9773436784744263, "learning_rate": 4.002154970486274e-05, "loss": 6.8828, "step": 19897 }, { "epoch": 0.8389232034066235, "grad_norm": 1.2114158868789673, "learning_rate": 4.002076891845467e-05, "loss": 6.4947, "step": 19898 }, { "epoch": 0.8389653645888231, "grad_norm": 2.9349184036254883, "learning_rate": 4.0019988132046596e-05, "loss": 7.5458, "step": 19899 }, { "epoch": 0.8390075257710227, "grad_norm": 0.7876352667808533, "learning_rate": 4.001920734563853e-05, "loss": 6.5408, "step": 19900 }, { "epoch": 0.8390496869532221, "grad_norm": 3.5073509216308594, "learning_rate": 4.001842655923046e-05, "loss": 7.4974, "step": 19901 }, { "epoch": 0.8390918481354217, "grad_norm": 1.334705114364624, "learning_rate": 4.001764577282239e-05, "loss": 6.9146, "step": 19902 }, { "epoch": 0.8391340093176213, "grad_norm": 1.2948591709136963, "learning_rate": 4.001686498641432e-05, "loss": 6.8968, "step": 19903 }, { "epoch": 0.8391761704998209, "grad_norm": 2.6002614498138428, "learning_rate": 4.001608420000625e-05, "loss": 6.8587, "step": 19904 }, { "epoch": 0.8392183316820203, "grad_norm": 2.7603776454925537, "learning_rate": 4.0015303413598176e-05, "loss": 7.6628, "step": 19905 }, { "epoch": 0.8392604928642199, "grad_norm": 1.2104077339172363, "learning_rate": 4.001452262719011e-05, "loss": 6.5872, "step": 19906 }, { "epoch": 0.8393026540464195, "grad_norm": 2.5766537189483643, "learning_rate": 4.001374184078204e-05, "loss": 7.8751, "step": 19907 }, { "epoch": 0.839344815228619, "grad_norm": 1.2037280797958374, "learning_rate": 4.001296105437396e-05, "loss": 6.5648, "step": 19908 }, { "epoch": 0.8393869764108186, "grad_norm": 1.2806624174118042, "learning_rate": 4.00121802679659e-05, "loss": 6.5902, "step": 19909 }, { "epoch": 0.8394291375930181, "grad_norm": 1.221333622932434, "learning_rate": 4.0011399481557826e-05, "loss": 6.535, "step": 19910 }, { "epoch": 0.8394712987752176, "grad_norm": 4.069122791290283, "learning_rate": 4.001061869514976e-05, "loss": 8.303, "step": 19911 }, { "epoch": 0.8395134599574172, "grad_norm": 1.2573498487472534, "learning_rate": 4.000983790874169e-05, "loss": 6.9581, "step": 19912 }, { "epoch": 0.8395556211396168, "grad_norm": 2.625347852706909, "learning_rate": 4.000905712233361e-05, "loss": 6.5346, "step": 19913 }, { "epoch": 0.8395977823218163, "grad_norm": 1.6028105020523071, "learning_rate": 4.000827633592555e-05, "loss": 6.3798, "step": 19914 }, { "epoch": 0.8396399435040158, "grad_norm": 1.9022469520568848, "learning_rate": 4.0007495549517475e-05, "loss": 6.3578, "step": 19915 }, { "epoch": 0.8396821046862154, "grad_norm": 1.6788893938064575, "learning_rate": 4.000671476310941e-05, "loss": 6.9664, "step": 19916 }, { "epoch": 0.839724265868415, "grad_norm": 1.4261637926101685, "learning_rate": 4.000593397670134e-05, "loss": 6.3049, "step": 19917 }, { "epoch": 0.8397664270506146, "grad_norm": 1.9574445486068726, "learning_rate": 4.000515319029326e-05, "loss": 7.5234, "step": 19918 }, { "epoch": 0.839808588232814, "grad_norm": 2.9048569202423096, "learning_rate": 4.0004372403885194e-05, "loss": 7.843, "step": 19919 }, { "epoch": 0.8398507494150136, "grad_norm": 2.2789454460144043, "learning_rate": 4.0003591617477125e-05, "loss": 7.1776, "step": 19920 }, { "epoch": 0.8398929105972132, "grad_norm": 2.0657334327697754, "learning_rate": 4.0002810831069056e-05, "loss": 7.4124, "step": 19921 }, { "epoch": 0.8399350717794127, "grad_norm": 1.4180481433868408, "learning_rate": 4.000203004466098e-05, "loss": 6.944, "step": 19922 }, { "epoch": 0.8399772329616122, "grad_norm": 2.2496163845062256, "learning_rate": 4.000124925825291e-05, "loss": 6.5734, "step": 19923 }, { "epoch": 0.8400193941438118, "grad_norm": 3.1458327770233154, "learning_rate": 4.000046847184484e-05, "loss": 7.8217, "step": 19924 }, { "epoch": 0.8400615553260113, "grad_norm": 2.567922353744507, "learning_rate": 3.9999687685436774e-05, "loss": 7.0881, "step": 19925 }, { "epoch": 0.8401037165082109, "grad_norm": 2.094999074935913, "learning_rate": 3.9998906899028706e-05, "loss": 7.0964, "step": 19926 }, { "epoch": 0.8401458776904105, "grad_norm": 1.9172091484069824, "learning_rate": 3.999812611262063e-05, "loss": 6.7846, "step": 19927 }, { "epoch": 0.84018803887261, "grad_norm": 2.153630256652832, "learning_rate": 3.999734532621257e-05, "loss": 6.4679, "step": 19928 }, { "epoch": 0.8402302000548095, "grad_norm": 2.2242417335510254, "learning_rate": 3.999656453980449e-05, "loss": 6.4661, "step": 19929 }, { "epoch": 0.8402723612370091, "grad_norm": 2.5750105381011963, "learning_rate": 3.9995783753396424e-05, "loss": 7.2535, "step": 19930 }, { "epoch": 0.8403145224192087, "grad_norm": 2.683398723602295, "learning_rate": 3.9995002966988355e-05, "loss": 6.9653, "step": 19931 }, { "epoch": 0.8403566836014081, "grad_norm": 2.173516273498535, "learning_rate": 3.999422218058028e-05, "loss": 7.142, "step": 19932 }, { "epoch": 0.8403988447836077, "grad_norm": 1.3253775835037231, "learning_rate": 3.999344139417221e-05, "loss": 7.06, "step": 19933 }, { "epoch": 0.8404410059658073, "grad_norm": 2.2299578189849854, "learning_rate": 3.999266060776414e-05, "loss": 7.3115, "step": 19934 }, { "epoch": 0.8404831671480069, "grad_norm": 2.206653356552124, "learning_rate": 3.9991879821356073e-05, "loss": 6.5418, "step": 19935 }, { "epoch": 0.8405253283302064, "grad_norm": 2.4669301509857178, "learning_rate": 3.9991099034948005e-05, "loss": 6.4806, "step": 19936 }, { "epoch": 0.8405674895124059, "grad_norm": 1.4150813817977905, "learning_rate": 3.999031824853993e-05, "loss": 7.1393, "step": 19937 }, { "epoch": 0.8406096506946055, "grad_norm": 2.659130096435547, "learning_rate": 3.998953746213186e-05, "loss": 7.6383, "step": 19938 }, { "epoch": 0.840651811876805, "grad_norm": 1.046566367149353, "learning_rate": 3.998875667572379e-05, "loss": 6.5983, "step": 19939 }, { "epoch": 0.8406939730590046, "grad_norm": 1.6053556203842163, "learning_rate": 3.998797588931572e-05, "loss": 6.4331, "step": 19940 }, { "epoch": 0.8407361342412041, "grad_norm": 3.3417670726776123, "learning_rate": 3.998719510290765e-05, "loss": 7.8153, "step": 19941 }, { "epoch": 0.8407782954234037, "grad_norm": 4.366334438323975, "learning_rate": 3.998641431649958e-05, "loss": 8.3118, "step": 19942 }, { "epoch": 0.8408204566056032, "grad_norm": 1.1927236318588257, "learning_rate": 3.998563353009151e-05, "loss": 6.5018, "step": 19943 }, { "epoch": 0.8408626177878028, "grad_norm": 1.3020620346069336, "learning_rate": 3.9984852743683434e-05, "loss": 6.4622, "step": 19944 }, { "epoch": 0.8409047789700023, "grad_norm": 0.9969832301139832, "learning_rate": 3.998407195727537e-05, "loss": 6.6291, "step": 19945 }, { "epoch": 0.8409469401522018, "grad_norm": 1.0881447792053223, "learning_rate": 3.99832911708673e-05, "loss": 6.4575, "step": 19946 }, { "epoch": 0.8409891013344014, "grad_norm": 1.1758253574371338, "learning_rate": 3.9982510384459235e-05, "loss": 6.8214, "step": 19947 }, { "epoch": 0.841031262516601, "grad_norm": 3.2032477855682373, "learning_rate": 3.998172959805116e-05, "loss": 8.9018, "step": 19948 }, { "epoch": 0.8410734236988006, "grad_norm": 0.8576836585998535, "learning_rate": 3.998094881164309e-05, "loss": 6.7441, "step": 19949 }, { "epoch": 0.841115584881, "grad_norm": 1.2233061790466309, "learning_rate": 3.998016802523502e-05, "loss": 6.4628, "step": 19950 }, { "epoch": 0.8411577460631996, "grad_norm": 0.8311031460762024, "learning_rate": 3.9979387238826947e-05, "loss": 6.6775, "step": 19951 }, { "epoch": 0.8411999072453992, "grad_norm": 1.3929698467254639, "learning_rate": 3.997860645241888e-05, "loss": 7.2015, "step": 19952 }, { "epoch": 0.8412420684275987, "grad_norm": 1.7725145816802979, "learning_rate": 3.997782566601081e-05, "loss": 6.5225, "step": 19953 }, { "epoch": 0.8412842296097982, "grad_norm": 1.823838710784912, "learning_rate": 3.997704487960274e-05, "loss": 7.6424, "step": 19954 }, { "epoch": 0.8413263907919978, "grad_norm": 1.4883558750152588, "learning_rate": 3.9976264093194665e-05, "loss": 7.3092, "step": 19955 }, { "epoch": 0.8413685519741974, "grad_norm": 1.2389196157455444, "learning_rate": 3.9975483306786596e-05, "loss": 6.7795, "step": 19956 }, { "epoch": 0.8414107131563969, "grad_norm": 2.021718978881836, "learning_rate": 3.997470252037853e-05, "loss": 6.4708, "step": 19957 }, { "epoch": 0.8414528743385965, "grad_norm": 1.7213773727416992, "learning_rate": 3.997392173397046e-05, "loss": 6.5114, "step": 19958 }, { "epoch": 0.841495035520796, "grad_norm": 2.0332887172698975, "learning_rate": 3.997314094756239e-05, "loss": 6.6301, "step": 19959 }, { "epoch": 0.8415371967029955, "grad_norm": 1.2210822105407715, "learning_rate": 3.9972360161154314e-05, "loss": 6.6427, "step": 19960 }, { "epoch": 0.8415793578851951, "grad_norm": 0.9721471667289734, "learning_rate": 3.9971579374746246e-05, "loss": 6.5912, "step": 19961 }, { "epoch": 0.8416215190673947, "grad_norm": 1.3858351707458496, "learning_rate": 3.997079858833818e-05, "loss": 6.8605, "step": 19962 }, { "epoch": 0.8416636802495941, "grad_norm": 1.874006748199463, "learning_rate": 3.99700178019301e-05, "loss": 6.7766, "step": 19963 }, { "epoch": 0.8417058414317937, "grad_norm": 3.568852663040161, "learning_rate": 3.996923701552204e-05, "loss": 7.8454, "step": 19964 }, { "epoch": 0.8417480026139933, "grad_norm": 1.4987901449203491, "learning_rate": 3.9968456229113964e-05, "loss": 6.4587, "step": 19965 }, { "epoch": 0.8417901637961929, "grad_norm": 2.6234164237976074, "learning_rate": 3.9967675442705895e-05, "loss": 6.2908, "step": 19966 }, { "epoch": 0.8418323249783924, "grad_norm": 1.8079084157943726, "learning_rate": 3.9966894656297826e-05, "loss": 7.0801, "step": 19967 }, { "epoch": 0.8418744861605919, "grad_norm": 2.0754964351654053, "learning_rate": 3.996611386988975e-05, "loss": 6.996, "step": 19968 }, { "epoch": 0.8419166473427915, "grad_norm": 3.0116233825683594, "learning_rate": 3.996533308348168e-05, "loss": 6.8447, "step": 19969 }, { "epoch": 0.841958808524991, "grad_norm": 2.0558462142944336, "learning_rate": 3.996455229707361e-05, "loss": 6.449, "step": 19970 }, { "epoch": 0.8420009697071906, "grad_norm": 1.786629557609558, "learning_rate": 3.9963771510665545e-05, "loss": 6.5587, "step": 19971 }, { "epoch": 0.8420431308893901, "grad_norm": 1.6876072883605957, "learning_rate": 3.9962990724257476e-05, "loss": 6.6411, "step": 19972 }, { "epoch": 0.8420852920715897, "grad_norm": 4.166139602661133, "learning_rate": 3.996220993784941e-05, "loss": 8.0364, "step": 19973 }, { "epoch": 0.8421274532537892, "grad_norm": 2.684723138809204, "learning_rate": 3.996142915144133e-05, "loss": 6.9453, "step": 19974 }, { "epoch": 0.8421696144359888, "grad_norm": 2.216169834136963, "learning_rate": 3.996064836503326e-05, "loss": 6.7467, "step": 19975 }, { "epoch": 0.8422117756181884, "grad_norm": 2.7715537548065186, "learning_rate": 3.9959867578625194e-05, "loss": 6.6438, "step": 19976 }, { "epoch": 0.8422539368003878, "grad_norm": 4.037035942077637, "learning_rate": 3.995908679221712e-05, "loss": 7.8073, "step": 19977 }, { "epoch": 0.8422960979825874, "grad_norm": 2.5265839099884033, "learning_rate": 3.995830600580906e-05, "loss": 6.3126, "step": 19978 }, { "epoch": 0.842338259164787, "grad_norm": 1.3549871444702148, "learning_rate": 3.995752521940098e-05, "loss": 6.8702, "step": 19979 }, { "epoch": 0.8423804203469866, "grad_norm": 1.8584518432617188, "learning_rate": 3.995674443299291e-05, "loss": 7.5275, "step": 19980 }, { "epoch": 0.842422581529186, "grad_norm": 2.463770866394043, "learning_rate": 3.9955963646584844e-05, "loss": 6.45, "step": 19981 }, { "epoch": 0.8424647427113856, "grad_norm": 2.668851613998413, "learning_rate": 3.995518286017677e-05, "loss": 6.5526, "step": 19982 }, { "epoch": 0.8425069038935852, "grad_norm": 1.6972521543502808, "learning_rate": 3.9954402073768706e-05, "loss": 6.4368, "step": 19983 }, { "epoch": 0.8425490650757848, "grad_norm": 2.536884307861328, "learning_rate": 3.995362128736063e-05, "loss": 7.5933, "step": 19984 }, { "epoch": 0.8425912262579843, "grad_norm": 1.5544674396514893, "learning_rate": 3.995284050095256e-05, "loss": 6.6759, "step": 19985 }, { "epoch": 0.8426333874401838, "grad_norm": 1.4255131483078003, "learning_rate": 3.995205971454449e-05, "loss": 6.5278, "step": 19986 }, { "epoch": 0.8426755486223834, "grad_norm": 1.5712066888809204, "learning_rate": 3.995127892813642e-05, "loss": 6.3256, "step": 19987 }, { "epoch": 0.8427177098045829, "grad_norm": 2.734368085861206, "learning_rate": 3.995049814172835e-05, "loss": 6.4523, "step": 19988 }, { "epoch": 0.8427598709867825, "grad_norm": 1.8214468955993652, "learning_rate": 3.994971735532028e-05, "loss": 6.9382, "step": 19989 }, { "epoch": 0.842802032168982, "grad_norm": 1.288070559501648, "learning_rate": 3.994893656891221e-05, "loss": 7.143, "step": 19990 }, { "epoch": 0.8428441933511815, "grad_norm": 2.144486427307129, "learning_rate": 3.9948155782504136e-05, "loss": 6.5563, "step": 19991 }, { "epoch": 0.8428863545333811, "grad_norm": 1.3800705671310425, "learning_rate": 3.9947374996096074e-05, "loss": 6.9739, "step": 19992 }, { "epoch": 0.8429285157155807, "grad_norm": 2.6829867362976074, "learning_rate": 3.9946594209688e-05, "loss": 8.1712, "step": 19993 }, { "epoch": 0.8429706768977803, "grad_norm": 1.395973801612854, "learning_rate": 3.994581342327993e-05, "loss": 6.5863, "step": 19994 }, { "epoch": 0.8430128380799797, "grad_norm": 1.5365698337554932, "learning_rate": 3.994503263687186e-05, "loss": 7.5138, "step": 19995 }, { "epoch": 0.8430549992621793, "grad_norm": 1.106117606163025, "learning_rate": 3.9944251850463785e-05, "loss": 6.7754, "step": 19996 }, { "epoch": 0.8430971604443789, "grad_norm": 1.3171112537384033, "learning_rate": 3.9943471064055723e-05, "loss": 6.6797, "step": 19997 }, { "epoch": 0.8431393216265785, "grad_norm": 1.3788777589797974, "learning_rate": 3.994269027764765e-05, "loss": 6.5346, "step": 19998 }, { "epoch": 0.8431814828087779, "grad_norm": 0.9239327311515808, "learning_rate": 3.994190949123958e-05, "loss": 6.6407, "step": 19999 }, { "epoch": 0.8432236439909775, "grad_norm": 2.645777702331543, "learning_rate": 3.994112870483151e-05, "loss": 7.5383, "step": 20000 }, { "epoch": 0.8432658051731771, "grad_norm": 3.1324281692504883, "learning_rate": 3.9940347918423435e-05, "loss": 7.5483, "step": 20001 }, { "epoch": 0.8433079663553766, "grad_norm": 1.9948004484176636, "learning_rate": 3.9939567132015366e-05, "loss": 7.1444, "step": 20002 }, { "epoch": 0.8433501275375761, "grad_norm": 0.8238708972930908, "learning_rate": 3.99387863456073e-05, "loss": 6.5897, "step": 20003 }, { "epoch": 0.8433922887197757, "grad_norm": 2.009068250656128, "learning_rate": 3.993800555919923e-05, "loss": 6.8449, "step": 20004 }, { "epoch": 0.8434344499019752, "grad_norm": 1.8096307516098022, "learning_rate": 3.993722477279116e-05, "loss": 6.7451, "step": 20005 }, { "epoch": 0.8434766110841748, "grad_norm": 2.6441731452941895, "learning_rate": 3.9936443986383084e-05, "loss": 6.5955, "step": 20006 }, { "epoch": 0.8435187722663744, "grad_norm": 1.5610466003417969, "learning_rate": 3.9935663199975016e-05, "loss": 6.8472, "step": 20007 }, { "epoch": 0.8435609334485739, "grad_norm": 2.2264585494995117, "learning_rate": 3.993488241356695e-05, "loss": 7.3077, "step": 20008 }, { "epoch": 0.8436030946307734, "grad_norm": 0.9398861527442932, "learning_rate": 3.993410162715888e-05, "loss": 6.5138, "step": 20009 }, { "epoch": 0.843645255812973, "grad_norm": 1.2444344758987427, "learning_rate": 3.99333208407508e-05, "loss": 7.1193, "step": 20010 }, { "epoch": 0.8436874169951726, "grad_norm": 1.1189624071121216, "learning_rate": 3.993254005434274e-05, "loss": 7.1053, "step": 20011 }, { "epoch": 0.843729578177372, "grad_norm": 1.2856074571609497, "learning_rate": 3.9931759267934665e-05, "loss": 7.0829, "step": 20012 }, { "epoch": 0.8437717393595716, "grad_norm": 1.4736955165863037, "learning_rate": 3.993097848152659e-05, "loss": 7.0178, "step": 20013 }, { "epoch": 0.8438139005417712, "grad_norm": 0.9610186219215393, "learning_rate": 3.993019769511853e-05, "loss": 6.5978, "step": 20014 }, { "epoch": 0.8438560617239708, "grad_norm": 1.090994954109192, "learning_rate": 3.992941690871045e-05, "loss": 6.9028, "step": 20015 }, { "epoch": 0.8438982229061703, "grad_norm": 3.103464365005493, "learning_rate": 3.992863612230239e-05, "loss": 7.4059, "step": 20016 }, { "epoch": 0.8439403840883698, "grad_norm": 1.8436394929885864, "learning_rate": 3.9927855335894315e-05, "loss": 6.7512, "step": 20017 }, { "epoch": 0.8439825452705694, "grad_norm": 1.1054025888442993, "learning_rate": 3.9927074549486246e-05, "loss": 6.5524, "step": 20018 }, { "epoch": 0.844024706452769, "grad_norm": 3.2257421016693115, "learning_rate": 3.992629376307818e-05, "loss": 6.6361, "step": 20019 }, { "epoch": 0.8440668676349685, "grad_norm": 3.5169525146484375, "learning_rate": 3.99255129766701e-05, "loss": 9.5418, "step": 20020 }, { "epoch": 0.844109028817168, "grad_norm": 2.4625351428985596, "learning_rate": 3.992473219026203e-05, "loss": 6.7042, "step": 20021 }, { "epoch": 0.8441511899993676, "grad_norm": 2.041916608810425, "learning_rate": 3.9923951403853964e-05, "loss": 6.8167, "step": 20022 }, { "epoch": 0.8441933511815671, "grad_norm": 1.8296869993209839, "learning_rate": 3.9923170617445896e-05, "loss": 6.8824, "step": 20023 }, { "epoch": 0.8442355123637667, "grad_norm": 1.7039299011230469, "learning_rate": 3.992238983103782e-05, "loss": 6.6018, "step": 20024 }, { "epoch": 0.8442776735459663, "grad_norm": 2.332430839538574, "learning_rate": 3.992160904462975e-05, "loss": 7.3809, "step": 20025 }, { "epoch": 0.8443198347281657, "grad_norm": 1.253361701965332, "learning_rate": 3.992082825822168e-05, "loss": 6.6196, "step": 20026 }, { "epoch": 0.8443619959103653, "grad_norm": 4.735454082489014, "learning_rate": 3.9920047471813614e-05, "loss": 7.7118, "step": 20027 }, { "epoch": 0.8444041570925649, "grad_norm": 4.212541103363037, "learning_rate": 3.9919266685405545e-05, "loss": 7.581, "step": 20028 }, { "epoch": 0.8444463182747645, "grad_norm": 3.5966320037841797, "learning_rate": 3.991848589899747e-05, "loss": 8.0352, "step": 20029 }, { "epoch": 0.8444884794569639, "grad_norm": 1.8337671756744385, "learning_rate": 3.99177051125894e-05, "loss": 6.7724, "step": 20030 }, { "epoch": 0.8445306406391635, "grad_norm": 2.361085891723633, "learning_rate": 3.991692432618133e-05, "loss": 7.0699, "step": 20031 }, { "epoch": 0.8445728018213631, "grad_norm": 1.3433860540390015, "learning_rate": 3.9916143539773257e-05, "loss": 7.3039, "step": 20032 }, { "epoch": 0.8446149630035626, "grad_norm": 1.7531617879867554, "learning_rate": 3.9915362753365195e-05, "loss": 7.2187, "step": 20033 }, { "epoch": 0.8446571241857622, "grad_norm": 2.6400327682495117, "learning_rate": 3.991458196695712e-05, "loss": 7.6975, "step": 20034 }, { "epoch": 0.8446992853679617, "grad_norm": 3.655721664428711, "learning_rate": 3.991380118054905e-05, "loss": 6.6137, "step": 20035 }, { "epoch": 0.8447414465501613, "grad_norm": 3.3397159576416016, "learning_rate": 3.991302039414098e-05, "loss": 6.5008, "step": 20036 }, { "epoch": 0.8447836077323608, "grad_norm": 3.0324947834014893, "learning_rate": 3.991223960773291e-05, "loss": 6.853, "step": 20037 }, { "epoch": 0.8448257689145604, "grad_norm": 1.586634874343872, "learning_rate": 3.9911458821324844e-05, "loss": 7.0907, "step": 20038 }, { "epoch": 0.8448679300967599, "grad_norm": 1.9211513996124268, "learning_rate": 3.991067803491677e-05, "loss": 6.7058, "step": 20039 }, { "epoch": 0.8449100912789594, "grad_norm": 1.5936447381973267, "learning_rate": 3.99098972485087e-05, "loss": 6.872, "step": 20040 }, { "epoch": 0.844952252461159, "grad_norm": 3.6568703651428223, "learning_rate": 3.990911646210063e-05, "loss": 8.3578, "step": 20041 }, { "epoch": 0.8449944136433586, "grad_norm": 1.475848913192749, "learning_rate": 3.990833567569256e-05, "loss": 6.4743, "step": 20042 }, { "epoch": 0.8450365748255582, "grad_norm": 1.671399712562561, "learning_rate": 3.990755488928449e-05, "loss": 7.0004, "step": 20043 }, { "epoch": 0.8450787360077576, "grad_norm": 2.0209028720855713, "learning_rate": 3.990677410287642e-05, "loss": 6.5624, "step": 20044 }, { "epoch": 0.8451208971899572, "grad_norm": 1.6904575824737549, "learning_rate": 3.990599331646835e-05, "loss": 7.0449, "step": 20045 }, { "epoch": 0.8451630583721568, "grad_norm": 1.6791422367095947, "learning_rate": 3.9905212530060274e-05, "loss": 7.8569, "step": 20046 }, { "epoch": 0.8452052195543563, "grad_norm": 2.2550554275512695, "learning_rate": 3.990443174365221e-05, "loss": 6.7917, "step": 20047 }, { "epoch": 0.8452473807365558, "grad_norm": 1.935124158859253, "learning_rate": 3.9903650957244136e-05, "loss": 7.0522, "step": 20048 }, { "epoch": 0.8452895419187554, "grad_norm": 3.3204634189605713, "learning_rate": 3.990287017083607e-05, "loss": 7.7954, "step": 20049 }, { "epoch": 0.845331703100955, "grad_norm": 1.9548749923706055, "learning_rate": 3.9902089384428e-05, "loss": 6.8531, "step": 20050 }, { "epoch": 0.8453738642831545, "grad_norm": 1.9266730546951294, "learning_rate": 3.9901308598019923e-05, "loss": 6.8251, "step": 20051 }, { "epoch": 0.8454160254653541, "grad_norm": 1.110287070274353, "learning_rate": 3.990052781161186e-05, "loss": 6.6731, "step": 20052 }, { "epoch": 0.8454581866475536, "grad_norm": 2.4966325759887695, "learning_rate": 3.9899747025203786e-05, "loss": 7.6776, "step": 20053 }, { "epoch": 0.8455003478297531, "grad_norm": 0.8381664752960205, "learning_rate": 3.989896623879572e-05, "loss": 6.5832, "step": 20054 }, { "epoch": 0.8455425090119527, "grad_norm": 3.230220317840576, "learning_rate": 3.989818545238765e-05, "loss": 7.6174, "step": 20055 }, { "epoch": 0.8455846701941523, "grad_norm": 2.4424033164978027, "learning_rate": 3.989740466597958e-05, "loss": 6.4432, "step": 20056 }, { "epoch": 0.8456268313763518, "grad_norm": 1.8749297857284546, "learning_rate": 3.9896623879571504e-05, "loss": 6.9711, "step": 20057 }, { "epoch": 0.8456689925585513, "grad_norm": 1.8912818431854248, "learning_rate": 3.9895843093163435e-05, "loss": 6.594, "step": 20058 }, { "epoch": 0.8457111537407509, "grad_norm": 2.519822359085083, "learning_rate": 3.989506230675537e-05, "loss": 6.776, "step": 20059 }, { "epoch": 0.8457533149229505, "grad_norm": 1.3257358074188232, "learning_rate": 3.989428152034729e-05, "loss": 6.9877, "step": 20060 }, { "epoch": 0.8457954761051499, "grad_norm": 1.2030918598175049, "learning_rate": 3.989350073393923e-05, "loss": 6.8148, "step": 20061 }, { "epoch": 0.8458376372873495, "grad_norm": 1.9686166048049927, "learning_rate": 3.9892719947531154e-05, "loss": 7.2907, "step": 20062 }, { "epoch": 0.8458797984695491, "grad_norm": 0.8835451602935791, "learning_rate": 3.9891939161123085e-05, "loss": 6.6912, "step": 20063 }, { "epoch": 0.8459219596517487, "grad_norm": 1.2174476385116577, "learning_rate": 3.9891158374715016e-05, "loss": 6.7762, "step": 20064 }, { "epoch": 0.8459641208339482, "grad_norm": 1.6233006715774536, "learning_rate": 3.989037758830694e-05, "loss": 6.9184, "step": 20065 }, { "epoch": 0.8460062820161477, "grad_norm": 2.0069751739501953, "learning_rate": 3.988959680189888e-05, "loss": 8.0951, "step": 20066 }, { "epoch": 0.8460484431983473, "grad_norm": 1.3124723434448242, "learning_rate": 3.98888160154908e-05, "loss": 7.0497, "step": 20067 }, { "epoch": 0.8460906043805468, "grad_norm": 1.5066176652908325, "learning_rate": 3.9888035229082735e-05, "loss": 6.5451, "step": 20068 }, { "epoch": 0.8461327655627464, "grad_norm": 3.8990015983581543, "learning_rate": 3.9887254442674666e-05, "loss": 7.3706, "step": 20069 }, { "epoch": 0.8461749267449459, "grad_norm": 1.3563530445098877, "learning_rate": 3.988647365626659e-05, "loss": 6.9103, "step": 20070 }, { "epoch": 0.8462170879271455, "grad_norm": 2.97072434425354, "learning_rate": 3.988569286985852e-05, "loss": 7.6422, "step": 20071 }, { "epoch": 0.846259249109345, "grad_norm": 3.052607774734497, "learning_rate": 3.988491208345045e-05, "loss": 6.4024, "step": 20072 }, { "epoch": 0.8463014102915446, "grad_norm": 1.502578616142273, "learning_rate": 3.9884131297042384e-05, "loss": 6.7015, "step": 20073 }, { "epoch": 0.8463435714737442, "grad_norm": 1.170418381690979, "learning_rate": 3.9883350510634315e-05, "loss": 6.445, "step": 20074 }, { "epoch": 0.8463857326559436, "grad_norm": 2.758333683013916, "learning_rate": 3.988256972422624e-05, "loss": 7.6469, "step": 20075 }, { "epoch": 0.8464278938381432, "grad_norm": 0.9632310271263123, "learning_rate": 3.988178893781817e-05, "loss": 6.5014, "step": 20076 }, { "epoch": 0.8464700550203428, "grad_norm": 3.455415725708008, "learning_rate": 3.98810081514101e-05, "loss": 7.757, "step": 20077 }, { "epoch": 0.8465122162025424, "grad_norm": 1.9979444742202759, "learning_rate": 3.9880227365002034e-05, "loss": 7.4016, "step": 20078 }, { "epoch": 0.8465543773847418, "grad_norm": 2.8717615604400635, "learning_rate": 3.987944657859396e-05, "loss": 8.0581, "step": 20079 }, { "epoch": 0.8465965385669414, "grad_norm": 1.954109787940979, "learning_rate": 3.9878665792185896e-05, "loss": 6.4983, "step": 20080 }, { "epoch": 0.846638699749141, "grad_norm": 2.225770950317383, "learning_rate": 3.987788500577782e-05, "loss": 6.4351, "step": 20081 }, { "epoch": 0.8466808609313405, "grad_norm": 1.0754843950271606, "learning_rate": 3.987710421936975e-05, "loss": 6.7074, "step": 20082 }, { "epoch": 0.8467230221135401, "grad_norm": 1.9880456924438477, "learning_rate": 3.987632343296168e-05, "loss": 6.4089, "step": 20083 }, { "epoch": 0.8467651832957396, "grad_norm": 1.8169111013412476, "learning_rate": 3.987554264655361e-05, "loss": 6.9113, "step": 20084 }, { "epoch": 0.8468073444779392, "grad_norm": 2.0013227462768555, "learning_rate": 3.9874761860145546e-05, "loss": 6.6076, "step": 20085 }, { "epoch": 0.8468495056601387, "grad_norm": 4.255382061004639, "learning_rate": 3.987398107373747e-05, "loss": 8.1059, "step": 20086 }, { "epoch": 0.8468916668423383, "grad_norm": 2.9776411056518555, "learning_rate": 3.98732002873294e-05, "loss": 6.3517, "step": 20087 }, { "epoch": 0.8469338280245378, "grad_norm": 1.6638729572296143, "learning_rate": 3.987241950092133e-05, "loss": 6.5567, "step": 20088 }, { "epoch": 0.8469759892067373, "grad_norm": 2.498975992202759, "learning_rate": 3.987163871451326e-05, "loss": 6.6695, "step": 20089 }, { "epoch": 0.8470181503889369, "grad_norm": 2.9727225303649902, "learning_rate": 3.987085792810519e-05, "loss": 9.3208, "step": 20090 }, { "epoch": 0.8470603115711365, "grad_norm": 1.4348034858703613, "learning_rate": 3.987007714169712e-05, "loss": 6.3467, "step": 20091 }, { "epoch": 0.847102472753336, "grad_norm": 3.3963570594787598, "learning_rate": 3.986929635528905e-05, "loss": 7.9077, "step": 20092 }, { "epoch": 0.8471446339355355, "grad_norm": 1.5896971225738525, "learning_rate": 3.9868515568880975e-05, "loss": 6.4845, "step": 20093 }, { "epoch": 0.8471867951177351, "grad_norm": 2.02425479888916, "learning_rate": 3.986773478247291e-05, "loss": 7.0189, "step": 20094 }, { "epoch": 0.8472289562999347, "grad_norm": 1.5058684349060059, "learning_rate": 3.986695399606484e-05, "loss": 6.9812, "step": 20095 }, { "epoch": 0.8472711174821342, "grad_norm": 3.009594678878784, "learning_rate": 3.986617320965677e-05, "loss": 7.1039, "step": 20096 }, { "epoch": 0.8473132786643337, "grad_norm": 1.9049710035324097, "learning_rate": 3.98653924232487e-05, "loss": 7.0754, "step": 20097 }, { "epoch": 0.8473554398465333, "grad_norm": 1.884639024734497, "learning_rate": 3.9864611636840625e-05, "loss": 7.0224, "step": 20098 }, { "epoch": 0.8473976010287328, "grad_norm": 2.3242835998535156, "learning_rate": 3.986383085043256e-05, "loss": 6.9469, "step": 20099 }, { "epoch": 0.8474397622109324, "grad_norm": 1.4430488348007202, "learning_rate": 3.986305006402449e-05, "loss": 6.4109, "step": 20100 }, { "epoch": 0.847481923393132, "grad_norm": 1.1207090616226196, "learning_rate": 3.986226927761642e-05, "loss": 6.7315, "step": 20101 }, { "epoch": 0.8475240845753315, "grad_norm": 1.662105917930603, "learning_rate": 3.986148849120835e-05, "loss": 7.0771, "step": 20102 }, { "epoch": 0.847566245757531, "grad_norm": 3.3143351078033447, "learning_rate": 3.9860707704800274e-05, "loss": 7.678, "step": 20103 }, { "epoch": 0.8476084069397306, "grad_norm": 2.606926441192627, "learning_rate": 3.9859926918392206e-05, "loss": 7.9665, "step": 20104 }, { "epoch": 0.8476505681219302, "grad_norm": 1.8884059190750122, "learning_rate": 3.985914613198414e-05, "loss": 6.4876, "step": 20105 }, { "epoch": 0.8476927293041296, "grad_norm": 1.7254512310028076, "learning_rate": 3.985836534557607e-05, "loss": 6.82, "step": 20106 }, { "epoch": 0.8477348904863292, "grad_norm": 2.9870681762695312, "learning_rate": 3.9857584559168e-05, "loss": 7.9737, "step": 20107 }, { "epoch": 0.8477770516685288, "grad_norm": 1.7249736785888672, "learning_rate": 3.9856803772759924e-05, "loss": 6.4313, "step": 20108 }, { "epoch": 0.8478192128507284, "grad_norm": 1.6934762001037598, "learning_rate": 3.9856022986351855e-05, "loss": 6.5485, "step": 20109 }, { "epoch": 0.8478613740329278, "grad_norm": 3.246206521987915, "learning_rate": 3.9855242199943786e-05, "loss": 7.7678, "step": 20110 }, { "epoch": 0.8479035352151274, "grad_norm": 1.6640539169311523, "learning_rate": 3.985446141353572e-05, "loss": 6.4747, "step": 20111 }, { "epoch": 0.847945696397327, "grad_norm": 2.0050389766693115, "learning_rate": 3.985368062712764e-05, "loss": 6.8916, "step": 20112 }, { "epoch": 0.8479878575795265, "grad_norm": 1.3407046794891357, "learning_rate": 3.9852899840719573e-05, "loss": 6.9014, "step": 20113 }, { "epoch": 0.8480300187617261, "grad_norm": 3.026017904281616, "learning_rate": 3.9852119054311505e-05, "loss": 7.519, "step": 20114 }, { "epoch": 0.8480721799439256, "grad_norm": 1.5977247953414917, "learning_rate": 3.985133826790343e-05, "loss": 6.4152, "step": 20115 }, { "epoch": 0.8481143411261252, "grad_norm": 1.9376167058944702, "learning_rate": 3.985055748149537e-05, "loss": 6.8112, "step": 20116 }, { "epoch": 0.8481565023083247, "grad_norm": 3.117755651473999, "learning_rate": 3.984977669508729e-05, "loss": 7.6279, "step": 20117 }, { "epoch": 0.8481986634905243, "grad_norm": 1.6267695426940918, "learning_rate": 3.984899590867923e-05, "loss": 6.6513, "step": 20118 }, { "epoch": 0.8482408246727238, "grad_norm": 1.7234830856323242, "learning_rate": 3.9848215122271154e-05, "loss": 7.0529, "step": 20119 }, { "epoch": 0.8482829858549233, "grad_norm": 0.9542880654335022, "learning_rate": 3.984743433586308e-05, "loss": 7.1196, "step": 20120 }, { "epoch": 0.8483251470371229, "grad_norm": 1.6809059381484985, "learning_rate": 3.984665354945502e-05, "loss": 6.812, "step": 20121 }, { "epoch": 0.8483673082193225, "grad_norm": 2.79649019241333, "learning_rate": 3.984587276304694e-05, "loss": 7.5669, "step": 20122 }, { "epoch": 0.8484094694015221, "grad_norm": 2.4251832962036133, "learning_rate": 3.984509197663887e-05, "loss": 7.5026, "step": 20123 }, { "epoch": 0.8484516305837215, "grad_norm": 1.0303525924682617, "learning_rate": 3.9844311190230804e-05, "loss": 6.8546, "step": 20124 }, { "epoch": 0.8484937917659211, "grad_norm": 1.4222668409347534, "learning_rate": 3.9843530403822735e-05, "loss": 6.8977, "step": 20125 }, { "epoch": 0.8485359529481207, "grad_norm": 1.120806336402893, "learning_rate": 3.984274961741466e-05, "loss": 6.6922, "step": 20126 }, { "epoch": 0.8485781141303202, "grad_norm": 1.4771453142166138, "learning_rate": 3.984196883100659e-05, "loss": 6.7314, "step": 20127 }, { "epoch": 0.8486202753125197, "grad_norm": 1.1571264266967773, "learning_rate": 3.984118804459852e-05, "loss": 6.8418, "step": 20128 }, { "epoch": 0.8486624364947193, "grad_norm": 0.7653956413269043, "learning_rate": 3.9840407258190446e-05, "loss": 7.1128, "step": 20129 }, { "epoch": 0.8487045976769189, "grad_norm": 1.4275217056274414, "learning_rate": 3.9839626471782385e-05, "loss": 6.8488, "step": 20130 }, { "epoch": 0.8487467588591184, "grad_norm": 1.7011228799819946, "learning_rate": 3.983884568537431e-05, "loss": 6.4734, "step": 20131 }, { "epoch": 0.848788920041318, "grad_norm": 1.742734432220459, "learning_rate": 3.983806489896624e-05, "loss": 6.5312, "step": 20132 }, { "epoch": 0.8488310812235175, "grad_norm": 0.970687747001648, "learning_rate": 3.983728411255817e-05, "loss": 6.7637, "step": 20133 }, { "epoch": 0.848873242405717, "grad_norm": 1.6884900331497192, "learning_rate": 3.9836503326150096e-05, "loss": 6.4859, "step": 20134 }, { "epoch": 0.8489154035879166, "grad_norm": 3.90883469581604, "learning_rate": 3.9835722539742034e-05, "loss": 7.546, "step": 20135 }, { "epoch": 0.8489575647701162, "grad_norm": 3.323272466659546, "learning_rate": 3.983494175333396e-05, "loss": 7.1305, "step": 20136 }, { "epoch": 0.8489997259523157, "grad_norm": 3.527735948562622, "learning_rate": 3.983416096692589e-05, "loss": 7.9284, "step": 20137 }, { "epoch": 0.8490418871345152, "grad_norm": 1.85263192653656, "learning_rate": 3.983338018051782e-05, "loss": 6.4221, "step": 20138 }, { "epoch": 0.8490840483167148, "grad_norm": 1.6972079277038574, "learning_rate": 3.9832599394109746e-05, "loss": 6.5578, "step": 20139 }, { "epoch": 0.8491262094989144, "grad_norm": 2.1951043605804443, "learning_rate": 3.983181860770168e-05, "loss": 7.3055, "step": 20140 }, { "epoch": 0.849168370681114, "grad_norm": 3.161168098449707, "learning_rate": 3.983103782129361e-05, "loss": 7.6205, "step": 20141 }, { "epoch": 0.8492105318633134, "grad_norm": 4.6464128494262695, "learning_rate": 3.983025703488554e-05, "loss": 7.8977, "step": 20142 }, { "epoch": 0.849252693045513, "grad_norm": 3.9075636863708496, "learning_rate": 3.982947624847747e-05, "loss": 7.0921, "step": 20143 }, { "epoch": 0.8492948542277126, "grad_norm": 5.056587219238281, "learning_rate": 3.98286954620694e-05, "loss": 7.6798, "step": 20144 }, { "epoch": 0.8493370154099121, "grad_norm": 2.8094823360443115, "learning_rate": 3.9827914675661326e-05, "loss": 7.1277, "step": 20145 }, { "epoch": 0.8493791765921116, "grad_norm": 1.825136661529541, "learning_rate": 3.982713388925326e-05, "loss": 6.5405, "step": 20146 }, { "epoch": 0.8494213377743112, "grad_norm": 2.6185131072998047, "learning_rate": 3.982635310284519e-05, "loss": 7.2596, "step": 20147 }, { "epoch": 0.8494634989565107, "grad_norm": 2.6173818111419678, "learning_rate": 3.982557231643711e-05, "loss": 7.7064, "step": 20148 }, { "epoch": 0.8495056601387103, "grad_norm": 2.574282169342041, "learning_rate": 3.982479153002905e-05, "loss": 6.562, "step": 20149 }, { "epoch": 0.8495478213209099, "grad_norm": 3.338243007659912, "learning_rate": 3.9824010743620976e-05, "loss": 7.1603, "step": 20150 }, { "epoch": 0.8495899825031094, "grad_norm": 3.6096925735473633, "learning_rate": 3.982322995721291e-05, "loss": 6.9007, "step": 20151 }, { "epoch": 0.8496321436853089, "grad_norm": 2.9409937858581543, "learning_rate": 3.982244917080484e-05, "loss": 7.0121, "step": 20152 }, { "epoch": 0.8496743048675085, "grad_norm": 1.7459691762924194, "learning_rate": 3.982166838439676e-05, "loss": 7.081, "step": 20153 }, { "epoch": 0.8497164660497081, "grad_norm": 1.9800370931625366, "learning_rate": 3.98208875979887e-05, "loss": 6.8277, "step": 20154 }, { "epoch": 0.8497586272319075, "grad_norm": 1.5871001482009888, "learning_rate": 3.9820106811580625e-05, "loss": 6.741, "step": 20155 }, { "epoch": 0.8498007884141071, "grad_norm": 3.0042994022369385, "learning_rate": 3.981932602517256e-05, "loss": 7.526, "step": 20156 }, { "epoch": 0.8498429495963067, "grad_norm": 1.9818090200424194, "learning_rate": 3.981854523876449e-05, "loss": 7.1664, "step": 20157 }, { "epoch": 0.8498851107785063, "grad_norm": 3.059110403060913, "learning_rate": 3.981776445235641e-05, "loss": 6.4852, "step": 20158 }, { "epoch": 0.8499272719607058, "grad_norm": 2.6845345497131348, "learning_rate": 3.9816983665948344e-05, "loss": 6.5759, "step": 20159 }, { "epoch": 0.8499694331429053, "grad_norm": 2.1062326431274414, "learning_rate": 3.9816202879540275e-05, "loss": 6.5725, "step": 20160 }, { "epoch": 0.8500115943251049, "grad_norm": 2.3006985187530518, "learning_rate": 3.9815422093132206e-05, "loss": 7.2833, "step": 20161 }, { "epoch": 0.8500537555073044, "grad_norm": 2.0963451862335205, "learning_rate": 3.981464130672413e-05, "loss": 7.3411, "step": 20162 }, { "epoch": 0.850095916689504, "grad_norm": 3.3427038192749023, "learning_rate": 3.981386052031606e-05, "loss": 7.0206, "step": 20163 }, { "epoch": 0.8501380778717035, "grad_norm": 3.0667736530303955, "learning_rate": 3.981307973390799e-05, "loss": 6.8623, "step": 20164 }, { "epoch": 0.850180239053903, "grad_norm": 3.5332043170928955, "learning_rate": 3.9812298947499924e-05, "loss": 7.3536, "step": 20165 }, { "epoch": 0.8502224002361026, "grad_norm": 4.666004180908203, "learning_rate": 3.9811518161091856e-05, "loss": 7.4825, "step": 20166 }, { "epoch": 0.8502645614183022, "grad_norm": 3.2688496112823486, "learning_rate": 3.981073737468378e-05, "loss": 7.2633, "step": 20167 }, { "epoch": 0.8503067226005017, "grad_norm": 1.2747204303741455, "learning_rate": 3.980995658827572e-05, "loss": 6.6834, "step": 20168 }, { "epoch": 0.8503488837827012, "grad_norm": 2.3372349739074707, "learning_rate": 3.980917580186764e-05, "loss": 6.7253, "step": 20169 }, { "epoch": 0.8503910449649008, "grad_norm": 1.6266297101974487, "learning_rate": 3.9808395015459574e-05, "loss": 7.1417, "step": 20170 }, { "epoch": 0.8504332061471004, "grad_norm": 2.4982337951660156, "learning_rate": 3.9807614229051505e-05, "loss": 7.5497, "step": 20171 }, { "epoch": 0.8504753673293, "grad_norm": 2.6855409145355225, "learning_rate": 3.980683344264343e-05, "loss": 7.8374, "step": 20172 }, { "epoch": 0.8505175285114994, "grad_norm": 1.4172213077545166, "learning_rate": 3.980605265623536e-05, "loss": 6.6081, "step": 20173 }, { "epoch": 0.850559689693699, "grad_norm": 1.2774144411087036, "learning_rate": 3.980527186982729e-05, "loss": 6.7745, "step": 20174 }, { "epoch": 0.8506018508758986, "grad_norm": 2.002908706665039, "learning_rate": 3.9804491083419223e-05, "loss": 6.7164, "step": 20175 }, { "epoch": 0.8506440120580981, "grad_norm": 1.669447898864746, "learning_rate": 3.9803710297011155e-05, "loss": 6.8705, "step": 20176 }, { "epoch": 0.8506861732402976, "grad_norm": 3.3302724361419678, "learning_rate": 3.980292951060308e-05, "loss": 7.7188, "step": 20177 }, { "epoch": 0.8507283344224972, "grad_norm": 2.8534069061279297, "learning_rate": 3.980214872419501e-05, "loss": 7.2419, "step": 20178 }, { "epoch": 0.8507704956046968, "grad_norm": 1.5895169973373413, "learning_rate": 3.980136793778694e-05, "loss": 7.0707, "step": 20179 }, { "epoch": 0.8508126567868963, "grad_norm": 2.1763274669647217, "learning_rate": 3.980058715137887e-05, "loss": 7.4211, "step": 20180 }, { "epoch": 0.8508548179690959, "grad_norm": 2.923191547393799, "learning_rate": 3.97998063649708e-05, "loss": 6.6603, "step": 20181 }, { "epoch": 0.8508969791512954, "grad_norm": 1.8140839338302612, "learning_rate": 3.979902557856273e-05, "loss": 6.8167, "step": 20182 }, { "epoch": 0.8509391403334949, "grad_norm": 3.7375850677490234, "learning_rate": 3.979824479215466e-05, "loss": 7.076, "step": 20183 }, { "epoch": 0.8509813015156945, "grad_norm": 2.0861048698425293, "learning_rate": 3.9797464005746584e-05, "loss": 7.3557, "step": 20184 }, { "epoch": 0.8510234626978941, "grad_norm": 2.9853310585021973, "learning_rate": 3.979668321933852e-05, "loss": 6.5112, "step": 20185 }, { "epoch": 0.8510656238800935, "grad_norm": 2.5298149585723877, "learning_rate": 3.979590243293045e-05, "loss": 6.5422, "step": 20186 }, { "epoch": 0.8511077850622931, "grad_norm": 1.4394549131393433, "learning_rate": 3.9795121646522385e-05, "loss": 6.6027, "step": 20187 }, { "epoch": 0.8511499462444927, "grad_norm": 1.5650604963302612, "learning_rate": 3.979434086011431e-05, "loss": 7.3221, "step": 20188 }, { "epoch": 0.8511921074266923, "grad_norm": 1.457680583000183, "learning_rate": 3.979356007370624e-05, "loss": 7.1871, "step": 20189 }, { "epoch": 0.8512342686088918, "grad_norm": 2.7857086658477783, "learning_rate": 3.979277928729817e-05, "loss": 7.4976, "step": 20190 }, { "epoch": 0.8512764297910913, "grad_norm": 2.5862491130828857, "learning_rate": 3.9791998500890097e-05, "loss": 6.5701, "step": 20191 }, { "epoch": 0.8513185909732909, "grad_norm": 3.0247209072113037, "learning_rate": 3.979121771448203e-05, "loss": 7.5209, "step": 20192 }, { "epoch": 0.8513607521554905, "grad_norm": 2.920478105545044, "learning_rate": 3.979043692807396e-05, "loss": 6.6041, "step": 20193 }, { "epoch": 0.85140291333769, "grad_norm": 1.5012696981430054, "learning_rate": 3.978965614166589e-05, "loss": 6.9731, "step": 20194 }, { "epoch": 0.8514450745198895, "grad_norm": 1.5834380388259888, "learning_rate": 3.9788875355257815e-05, "loss": 6.4759, "step": 20195 }, { "epoch": 0.8514872357020891, "grad_norm": 1.4954931735992432, "learning_rate": 3.9788094568849746e-05, "loss": 6.996, "step": 20196 }, { "epoch": 0.8515293968842886, "grad_norm": 1.3942641019821167, "learning_rate": 3.978731378244168e-05, "loss": 6.6306, "step": 20197 }, { "epoch": 0.8515715580664882, "grad_norm": 4.57988977432251, "learning_rate": 3.97865329960336e-05, "loss": 7.6215, "step": 20198 }, { "epoch": 0.8516137192486878, "grad_norm": 2.810614824295044, "learning_rate": 3.978575220962554e-05, "loss": 6.7026, "step": 20199 }, { "epoch": 0.8516558804308872, "grad_norm": 3.288567066192627, "learning_rate": 3.9784971423217464e-05, "loss": 6.9388, "step": 20200 }, { "epoch": 0.8516980416130868, "grad_norm": 1.0318273305892944, "learning_rate": 3.9784190636809396e-05, "loss": 6.8352, "step": 20201 }, { "epoch": 0.8517402027952864, "grad_norm": 0.9997272491455078, "learning_rate": 3.978340985040133e-05, "loss": 6.8103, "step": 20202 }, { "epoch": 0.851782363977486, "grad_norm": 2.6270546913146973, "learning_rate": 3.978262906399325e-05, "loss": 7.7471, "step": 20203 }, { "epoch": 0.8518245251596854, "grad_norm": 1.8095630407333374, "learning_rate": 3.978184827758519e-05, "loss": 7.3146, "step": 20204 }, { "epoch": 0.851866686341885, "grad_norm": 2.2914209365844727, "learning_rate": 3.9781067491177114e-05, "loss": 7.1505, "step": 20205 }, { "epoch": 0.8519088475240846, "grad_norm": 2.601673126220703, "learning_rate": 3.9780286704769045e-05, "loss": 7.724, "step": 20206 }, { "epoch": 0.8519510087062842, "grad_norm": 2.0859768390655518, "learning_rate": 3.9779505918360976e-05, "loss": 7.739, "step": 20207 }, { "epoch": 0.8519931698884837, "grad_norm": 1.9763777256011963, "learning_rate": 3.97787251319529e-05, "loss": 6.5865, "step": 20208 }, { "epoch": 0.8520353310706832, "grad_norm": 1.1709033250808716, "learning_rate": 3.977794434554483e-05, "loss": 6.904, "step": 20209 }, { "epoch": 0.8520774922528828, "grad_norm": 1.279391884803772, "learning_rate": 3.977716355913676e-05, "loss": 7.1391, "step": 20210 }, { "epoch": 0.8521196534350823, "grad_norm": 1.435302495956421, "learning_rate": 3.9776382772728695e-05, "loss": 6.6122, "step": 20211 }, { "epoch": 0.8521618146172819, "grad_norm": 1.8946133852005005, "learning_rate": 3.9775601986320626e-05, "loss": 6.9395, "step": 20212 }, { "epoch": 0.8522039757994814, "grad_norm": 2.147580623626709, "learning_rate": 3.977482119991256e-05, "loss": 6.4262, "step": 20213 }, { "epoch": 0.852246136981681, "grad_norm": 3.252042055130005, "learning_rate": 3.977404041350448e-05, "loss": 7.5872, "step": 20214 }, { "epoch": 0.8522882981638805, "grad_norm": 1.1771851778030396, "learning_rate": 3.977325962709641e-05, "loss": 6.5635, "step": 20215 }, { "epoch": 0.8523304593460801, "grad_norm": 1.3212206363677979, "learning_rate": 3.9772478840688344e-05, "loss": 6.542, "step": 20216 }, { "epoch": 0.8523726205282797, "grad_norm": 2.6186819076538086, "learning_rate": 3.977169805428027e-05, "loss": 6.6354, "step": 20217 }, { "epoch": 0.8524147817104791, "grad_norm": 1.0702953338623047, "learning_rate": 3.977091726787221e-05, "loss": 6.8088, "step": 20218 }, { "epoch": 0.8524569428926787, "grad_norm": 1.3025405406951904, "learning_rate": 3.977013648146413e-05, "loss": 6.4837, "step": 20219 }, { "epoch": 0.8524991040748783, "grad_norm": 1.1009036302566528, "learning_rate": 3.976935569505606e-05, "loss": 7.0823, "step": 20220 }, { "epoch": 0.8525412652570779, "grad_norm": 1.0062155723571777, "learning_rate": 3.9768574908647994e-05, "loss": 6.9204, "step": 20221 }, { "epoch": 0.8525834264392773, "grad_norm": 1.016539454460144, "learning_rate": 3.976779412223992e-05, "loss": 6.608, "step": 20222 }, { "epoch": 0.8526255876214769, "grad_norm": 2.487941026687622, "learning_rate": 3.9767013335831856e-05, "loss": 7.1937, "step": 20223 }, { "epoch": 0.8526677488036765, "grad_norm": 0.7768760919570923, "learning_rate": 3.976623254942378e-05, "loss": 6.6329, "step": 20224 }, { "epoch": 0.852709909985876, "grad_norm": 2.364091396331787, "learning_rate": 3.976545176301571e-05, "loss": 7.1087, "step": 20225 }, { "epoch": 0.8527520711680755, "grad_norm": 3.730525016784668, "learning_rate": 3.976467097660764e-05, "loss": 7.7966, "step": 20226 }, { "epoch": 0.8527942323502751, "grad_norm": 1.7555732727050781, "learning_rate": 3.976389019019957e-05, "loss": 6.846, "step": 20227 }, { "epoch": 0.8528363935324746, "grad_norm": 1.8581933975219727, "learning_rate": 3.97631094037915e-05, "loss": 6.6445, "step": 20228 }, { "epoch": 0.8528785547146742, "grad_norm": 1.8903393745422363, "learning_rate": 3.976232861738343e-05, "loss": 7.1946, "step": 20229 }, { "epoch": 0.8529207158968738, "grad_norm": 4.024117946624756, "learning_rate": 3.976154783097536e-05, "loss": 6.6176, "step": 20230 }, { "epoch": 0.8529628770790733, "grad_norm": 1.4100843667984009, "learning_rate": 3.9760767044567286e-05, "loss": 6.5364, "step": 20231 }, { "epoch": 0.8530050382612728, "grad_norm": 3.6388189792633057, "learning_rate": 3.9759986258159224e-05, "loss": 7.3565, "step": 20232 }, { "epoch": 0.8530471994434724, "grad_norm": 1.2204132080078125, "learning_rate": 3.975920547175115e-05, "loss": 6.7191, "step": 20233 }, { "epoch": 0.853089360625672, "grad_norm": 1.6992613077163696, "learning_rate": 3.975842468534308e-05, "loss": 6.8147, "step": 20234 }, { "epoch": 0.8531315218078714, "grad_norm": 1.6662918329238892, "learning_rate": 3.975764389893501e-05, "loss": 7.0073, "step": 20235 }, { "epoch": 0.853173682990071, "grad_norm": 2.4999067783355713, "learning_rate": 3.9756863112526935e-05, "loss": 6.396, "step": 20236 }, { "epoch": 0.8532158441722706, "grad_norm": 1.9017248153686523, "learning_rate": 3.9756082326118873e-05, "loss": 6.2962, "step": 20237 }, { "epoch": 0.8532580053544702, "grad_norm": 1.991441011428833, "learning_rate": 3.97553015397108e-05, "loss": 7.4298, "step": 20238 }, { "epoch": 0.8533001665366697, "grad_norm": 2.1828322410583496, "learning_rate": 3.975452075330273e-05, "loss": 7.01, "step": 20239 }, { "epoch": 0.8533423277188692, "grad_norm": 2.9687294960021973, "learning_rate": 3.975373996689466e-05, "loss": 7.6375, "step": 20240 }, { "epoch": 0.8533844889010688, "grad_norm": 3.247624397277832, "learning_rate": 3.9752959180486585e-05, "loss": 7.5139, "step": 20241 }, { "epoch": 0.8534266500832683, "grad_norm": 1.1838852167129517, "learning_rate": 3.9752178394078516e-05, "loss": 7.0237, "step": 20242 }, { "epoch": 0.8534688112654679, "grad_norm": 1.5299335718154907, "learning_rate": 3.975139760767045e-05, "loss": 6.3633, "step": 20243 }, { "epoch": 0.8535109724476674, "grad_norm": 1.1853574514389038, "learning_rate": 3.975061682126238e-05, "loss": 6.7356, "step": 20244 }, { "epoch": 0.853553133629867, "grad_norm": 1.333945631980896, "learning_rate": 3.974983603485431e-05, "loss": 6.656, "step": 20245 }, { "epoch": 0.8535952948120665, "grad_norm": 1.3817001581192017, "learning_rate": 3.9749055248446234e-05, "loss": 6.8303, "step": 20246 }, { "epoch": 0.8536374559942661, "grad_norm": 2.6988699436187744, "learning_rate": 3.9748274462038166e-05, "loss": 7.6391, "step": 20247 }, { "epoch": 0.8536796171764657, "grad_norm": 1.2628933191299438, "learning_rate": 3.97474936756301e-05, "loss": 6.7458, "step": 20248 }, { "epoch": 0.8537217783586651, "grad_norm": 6.884207248687744, "learning_rate": 3.974671288922203e-05, "loss": 12.2475, "step": 20249 }, { "epoch": 0.8537639395408647, "grad_norm": 3.066579580307007, "learning_rate": 3.974593210281395e-05, "loss": 7.7277, "step": 20250 }, { "epoch": 0.8538061007230643, "grad_norm": 1.779658555984497, "learning_rate": 3.974515131640589e-05, "loss": 7.0832, "step": 20251 }, { "epoch": 0.8538482619052639, "grad_norm": 2.3154137134552, "learning_rate": 3.9744370529997815e-05, "loss": 6.7873, "step": 20252 }, { "epoch": 0.8538904230874633, "grad_norm": 2.2076964378356934, "learning_rate": 3.974358974358974e-05, "loss": 6.5156, "step": 20253 }, { "epoch": 0.8539325842696629, "grad_norm": 1.9414094686508179, "learning_rate": 3.974280895718168e-05, "loss": 6.483, "step": 20254 }, { "epoch": 0.8539747454518625, "grad_norm": 1.0999950170516968, "learning_rate": 3.97420281707736e-05, "loss": 6.9659, "step": 20255 }, { "epoch": 0.854016906634062, "grad_norm": 2.355696201324463, "learning_rate": 3.974124738436554e-05, "loss": 7.8431, "step": 20256 }, { "epoch": 0.8540590678162616, "grad_norm": 2.2702577114105225, "learning_rate": 3.9740466597957465e-05, "loss": 7.083, "step": 20257 }, { "epoch": 0.8541012289984611, "grad_norm": 1.9604547023773193, "learning_rate": 3.9739685811549396e-05, "loss": 7.0232, "step": 20258 }, { "epoch": 0.8541433901806607, "grad_norm": 2.6596148014068604, "learning_rate": 3.973890502514133e-05, "loss": 6.9234, "step": 20259 }, { "epoch": 0.8541855513628602, "grad_norm": 2.166818141937256, "learning_rate": 3.973812423873325e-05, "loss": 6.9974, "step": 20260 }, { "epoch": 0.8542277125450598, "grad_norm": 1.3939621448516846, "learning_rate": 3.973734345232518e-05, "loss": 7.1306, "step": 20261 }, { "epoch": 0.8542698737272593, "grad_norm": 2.578624725341797, "learning_rate": 3.9736562665917114e-05, "loss": 6.7388, "step": 20262 }, { "epoch": 0.8543120349094588, "grad_norm": 5.360498428344727, "learning_rate": 3.9735781879509046e-05, "loss": 7.5509, "step": 20263 }, { "epoch": 0.8543541960916584, "grad_norm": 2.0317373275756836, "learning_rate": 3.973500109310097e-05, "loss": 6.6525, "step": 20264 }, { "epoch": 0.854396357273858, "grad_norm": 2.228951930999756, "learning_rate": 3.97342203066929e-05, "loss": 7.5186, "step": 20265 }, { "epoch": 0.8544385184560576, "grad_norm": 2.030026435852051, "learning_rate": 3.973343952028483e-05, "loss": 6.6055, "step": 20266 }, { "epoch": 0.854480679638257, "grad_norm": 3.4029111862182617, "learning_rate": 3.9732658733876764e-05, "loss": 7.5568, "step": 20267 }, { "epoch": 0.8545228408204566, "grad_norm": 1.8399039506912231, "learning_rate": 3.9731877947468695e-05, "loss": 6.4797, "step": 20268 }, { "epoch": 0.8545650020026562, "grad_norm": 1.9264309406280518, "learning_rate": 3.973109716106062e-05, "loss": 6.9765, "step": 20269 }, { "epoch": 0.8546071631848557, "grad_norm": 2.9899282455444336, "learning_rate": 3.973031637465255e-05, "loss": 7.4688, "step": 20270 }, { "epoch": 0.8546493243670552, "grad_norm": 1.3440563678741455, "learning_rate": 3.972953558824448e-05, "loss": 6.6314, "step": 20271 }, { "epoch": 0.8546914855492548, "grad_norm": 1.8201290369033813, "learning_rate": 3.9728754801836407e-05, "loss": 6.4553, "step": 20272 }, { "epoch": 0.8547336467314544, "grad_norm": 1.365861177444458, "learning_rate": 3.9727974015428345e-05, "loss": 6.6187, "step": 20273 }, { "epoch": 0.8547758079136539, "grad_norm": 3.581986904144287, "learning_rate": 3.972719322902027e-05, "loss": 7.1037, "step": 20274 }, { "epoch": 0.8548179690958535, "grad_norm": 3.5931501388549805, "learning_rate": 3.97264124426122e-05, "loss": 7.5434, "step": 20275 }, { "epoch": 0.854860130278053, "grad_norm": 3.0524845123291016, "learning_rate": 3.972563165620413e-05, "loss": 7.6436, "step": 20276 }, { "epoch": 0.8549022914602525, "grad_norm": 4.089669227600098, "learning_rate": 3.972485086979606e-05, "loss": 7.8548, "step": 20277 }, { "epoch": 0.8549444526424521, "grad_norm": 1.6721372604370117, "learning_rate": 3.972407008338799e-05, "loss": 6.5436, "step": 20278 }, { "epoch": 0.8549866138246517, "grad_norm": 1.1111963987350464, "learning_rate": 3.972328929697992e-05, "loss": 6.5794, "step": 20279 }, { "epoch": 0.8550287750068511, "grad_norm": 3.410672664642334, "learning_rate": 3.972250851057185e-05, "loss": 7.2958, "step": 20280 }, { "epoch": 0.8550709361890507, "grad_norm": 2.422018051147461, "learning_rate": 3.972172772416378e-05, "loss": 6.5171, "step": 20281 }, { "epoch": 0.8551130973712503, "grad_norm": 2.192366123199463, "learning_rate": 3.972094693775571e-05, "loss": 6.6431, "step": 20282 }, { "epoch": 0.8551552585534499, "grad_norm": 1.6508278846740723, "learning_rate": 3.972016615134764e-05, "loss": 6.4701, "step": 20283 }, { "epoch": 0.8551974197356493, "grad_norm": 2.272981882095337, "learning_rate": 3.971938536493957e-05, "loss": 7.9186, "step": 20284 }, { "epoch": 0.8552395809178489, "grad_norm": 1.6559205055236816, "learning_rate": 3.97186045785315e-05, "loss": 6.9891, "step": 20285 }, { "epoch": 0.8552817421000485, "grad_norm": 1.1119545698165894, "learning_rate": 3.9717823792123424e-05, "loss": 6.5658, "step": 20286 }, { "epoch": 0.855323903282248, "grad_norm": 3.861992597579956, "learning_rate": 3.971704300571536e-05, "loss": 7.6359, "step": 20287 }, { "epoch": 0.8553660644644476, "grad_norm": 1.6437736749649048, "learning_rate": 3.9716262219307286e-05, "loss": 6.6552, "step": 20288 }, { "epoch": 0.8554082256466471, "grad_norm": 5.059242248535156, "learning_rate": 3.971548143289922e-05, "loss": 6.5207, "step": 20289 }, { "epoch": 0.8554503868288467, "grad_norm": 3.1411263942718506, "learning_rate": 3.971470064649115e-05, "loss": 7.6863, "step": 20290 }, { "epoch": 0.8554925480110462, "grad_norm": 2.7782840728759766, "learning_rate": 3.9713919860083073e-05, "loss": 7.9849, "step": 20291 }, { "epoch": 0.8555347091932458, "grad_norm": 2.0730347633361816, "learning_rate": 3.971313907367501e-05, "loss": 6.8939, "step": 20292 }, { "epoch": 0.8555768703754453, "grad_norm": 3.244117259979248, "learning_rate": 3.9712358287266936e-05, "loss": 7.5091, "step": 20293 }, { "epoch": 0.8556190315576448, "grad_norm": 2.5274360179901123, "learning_rate": 3.971157750085887e-05, "loss": 7.8107, "step": 20294 }, { "epoch": 0.8556611927398444, "grad_norm": 3.4054219722747803, "learning_rate": 3.97107967144508e-05, "loss": 6.3991, "step": 20295 }, { "epoch": 0.855703353922044, "grad_norm": 2.4762051105499268, "learning_rate": 3.971001592804273e-05, "loss": 7.7318, "step": 20296 }, { "epoch": 0.8557455151042436, "grad_norm": 3.478652238845825, "learning_rate": 3.9709235141634654e-05, "loss": 6.5679, "step": 20297 }, { "epoch": 0.855787676286443, "grad_norm": 2.6886911392211914, "learning_rate": 3.9708454355226585e-05, "loss": 6.7789, "step": 20298 }, { "epoch": 0.8558298374686426, "grad_norm": 2.5401182174682617, "learning_rate": 3.970767356881852e-05, "loss": 7.1111, "step": 20299 }, { "epoch": 0.8558719986508422, "grad_norm": 1.9814611673355103, "learning_rate": 3.970689278241044e-05, "loss": 7.9151, "step": 20300 }, { "epoch": 0.8559141598330418, "grad_norm": 1.8707798719406128, "learning_rate": 3.970611199600238e-05, "loss": 6.9487, "step": 20301 }, { "epoch": 0.8559563210152412, "grad_norm": 2.2463958263397217, "learning_rate": 3.9705331209594304e-05, "loss": 7.1788, "step": 20302 }, { "epoch": 0.8559984821974408, "grad_norm": 3.4383251667022705, "learning_rate": 3.9704550423186235e-05, "loss": 7.7961, "step": 20303 }, { "epoch": 0.8560406433796404, "grad_norm": 2.097771167755127, "learning_rate": 3.9703769636778166e-05, "loss": 6.6713, "step": 20304 }, { "epoch": 0.8560828045618399, "grad_norm": 2.2230639457702637, "learning_rate": 3.970298885037009e-05, "loss": 6.7478, "step": 20305 }, { "epoch": 0.8561249657440395, "grad_norm": 2.7802376747131348, "learning_rate": 3.970220806396203e-05, "loss": 7.2008, "step": 20306 }, { "epoch": 0.856167126926239, "grad_norm": 2.959284782409668, "learning_rate": 3.970142727755395e-05, "loss": 7.3992, "step": 20307 }, { "epoch": 0.8562092881084385, "grad_norm": 1.207852840423584, "learning_rate": 3.9700646491145885e-05, "loss": 6.9306, "step": 20308 }, { "epoch": 0.8562514492906381, "grad_norm": 1.79757559299469, "learning_rate": 3.9699865704737816e-05, "loss": 6.5701, "step": 20309 }, { "epoch": 0.8562936104728377, "grad_norm": 1.8213236331939697, "learning_rate": 3.969908491832974e-05, "loss": 6.774, "step": 20310 }, { "epoch": 0.8563357716550372, "grad_norm": 2.0552008152008057, "learning_rate": 3.969830413192167e-05, "loss": 7.7952, "step": 20311 }, { "epoch": 0.8563779328372367, "grad_norm": 1.790479302406311, "learning_rate": 3.96975233455136e-05, "loss": 6.5038, "step": 20312 }, { "epoch": 0.8564200940194363, "grad_norm": 2.6490399837493896, "learning_rate": 3.9696742559105534e-05, "loss": 7.436, "step": 20313 }, { "epoch": 0.8564622552016359, "grad_norm": 2.572411298751831, "learning_rate": 3.9695961772697465e-05, "loss": 7.5473, "step": 20314 }, { "epoch": 0.8565044163838355, "grad_norm": 1.0865358114242554, "learning_rate": 3.969518098628939e-05, "loss": 6.7597, "step": 20315 }, { "epoch": 0.8565465775660349, "grad_norm": 2.7728869915008545, "learning_rate": 3.969440019988132e-05, "loss": 7.1306, "step": 20316 }, { "epoch": 0.8565887387482345, "grad_norm": 1.638991117477417, "learning_rate": 3.969361941347325e-05, "loss": 6.9783, "step": 20317 }, { "epoch": 0.8566308999304341, "grad_norm": 1.0962047576904297, "learning_rate": 3.9692838627065184e-05, "loss": 6.6466, "step": 20318 }, { "epoch": 0.8566730611126336, "grad_norm": 2.9356014728546143, "learning_rate": 3.969205784065711e-05, "loss": 7.2055, "step": 20319 }, { "epoch": 0.8567152222948331, "grad_norm": 2.4757659435272217, "learning_rate": 3.9691277054249046e-05, "loss": 7.4322, "step": 20320 }, { "epoch": 0.8567573834770327, "grad_norm": 2.1489827632904053, "learning_rate": 3.969049626784097e-05, "loss": 6.3589, "step": 20321 }, { "epoch": 0.8567995446592322, "grad_norm": 1.9661632776260376, "learning_rate": 3.96897154814329e-05, "loss": 6.4151, "step": 20322 }, { "epoch": 0.8568417058414318, "grad_norm": 1.374448537826538, "learning_rate": 3.968893469502483e-05, "loss": 7.2619, "step": 20323 }, { "epoch": 0.8568838670236314, "grad_norm": 0.9695103764533997, "learning_rate": 3.968815390861676e-05, "loss": 6.7876, "step": 20324 }, { "epoch": 0.8569260282058309, "grad_norm": 2.3404722213745117, "learning_rate": 3.9687373122208696e-05, "loss": 6.9935, "step": 20325 }, { "epoch": 0.8569681893880304, "grad_norm": 1.5132917165756226, "learning_rate": 3.968659233580062e-05, "loss": 7.0828, "step": 20326 }, { "epoch": 0.85701035057023, "grad_norm": 3.6996309757232666, "learning_rate": 3.968581154939255e-05, "loss": 7.9056, "step": 20327 }, { "epoch": 0.8570525117524296, "grad_norm": 1.0355788469314575, "learning_rate": 3.968503076298448e-05, "loss": 6.8083, "step": 20328 }, { "epoch": 0.857094672934629, "grad_norm": 1.4954338073730469, "learning_rate": 3.968424997657641e-05, "loss": 6.9911, "step": 20329 }, { "epoch": 0.8571368341168286, "grad_norm": 0.909400224685669, "learning_rate": 3.968346919016834e-05, "loss": 6.7778, "step": 20330 }, { "epoch": 0.8571789952990282, "grad_norm": 1.1486512422561646, "learning_rate": 3.968268840376027e-05, "loss": 6.5262, "step": 20331 }, { "epoch": 0.8572211564812278, "grad_norm": 1.5805116891860962, "learning_rate": 3.96819076173522e-05, "loss": 6.3587, "step": 20332 }, { "epoch": 0.8572633176634273, "grad_norm": 3.911982536315918, "learning_rate": 3.9681126830944125e-05, "loss": 7.3618, "step": 20333 }, { "epoch": 0.8573054788456268, "grad_norm": 2.099552869796753, "learning_rate": 3.968034604453606e-05, "loss": 7.0501, "step": 20334 }, { "epoch": 0.8573476400278264, "grad_norm": 2.149034023284912, "learning_rate": 3.967956525812799e-05, "loss": 6.6383, "step": 20335 }, { "epoch": 0.857389801210026, "grad_norm": 3.267597198486328, "learning_rate": 3.967878447171992e-05, "loss": 7.5087, "step": 20336 }, { "epoch": 0.8574319623922255, "grad_norm": 1.3816790580749512, "learning_rate": 3.967800368531185e-05, "loss": 6.6851, "step": 20337 }, { "epoch": 0.857474123574425, "grad_norm": 1.9637885093688965, "learning_rate": 3.9677222898903775e-05, "loss": 7.2291, "step": 20338 }, { "epoch": 0.8575162847566246, "grad_norm": 2.730386257171631, "learning_rate": 3.967644211249571e-05, "loss": 7.5697, "step": 20339 }, { "epoch": 0.8575584459388241, "grad_norm": 1.452066421508789, "learning_rate": 3.967566132608764e-05, "loss": 7.2973, "step": 20340 }, { "epoch": 0.8576006071210237, "grad_norm": 3.0936801433563232, "learning_rate": 3.967488053967957e-05, "loss": 8.0885, "step": 20341 }, { "epoch": 0.8576427683032232, "grad_norm": 2.4748213291168213, "learning_rate": 3.96740997532715e-05, "loss": 7.3575, "step": 20342 }, { "epoch": 0.8576849294854227, "grad_norm": 1.2997784614562988, "learning_rate": 3.9673318966863424e-05, "loss": 6.5778, "step": 20343 }, { "epoch": 0.8577270906676223, "grad_norm": 1.1746481657028198, "learning_rate": 3.9672538180455356e-05, "loss": 6.6076, "step": 20344 }, { "epoch": 0.8577692518498219, "grad_norm": 1.7248034477233887, "learning_rate": 3.967175739404729e-05, "loss": 6.4526, "step": 20345 }, { "epoch": 0.8578114130320215, "grad_norm": 2.127375841140747, "learning_rate": 3.967097660763922e-05, "loss": 6.7557, "step": 20346 }, { "epoch": 0.8578535742142209, "grad_norm": 2.761432647705078, "learning_rate": 3.967019582123115e-05, "loss": 7.1987, "step": 20347 }, { "epoch": 0.8578957353964205, "grad_norm": 1.8060250282287598, "learning_rate": 3.9669415034823074e-05, "loss": 6.4947, "step": 20348 }, { "epoch": 0.8579378965786201, "grad_norm": 3.137333393096924, "learning_rate": 3.9668634248415005e-05, "loss": 7.6236, "step": 20349 }, { "epoch": 0.8579800577608196, "grad_norm": 1.437686562538147, "learning_rate": 3.9667853462006936e-05, "loss": 6.6987, "step": 20350 }, { "epoch": 0.8580222189430191, "grad_norm": 1.2796967029571533, "learning_rate": 3.966707267559887e-05, "loss": 6.7252, "step": 20351 }, { "epoch": 0.8580643801252187, "grad_norm": 1.5806496143341064, "learning_rate": 3.966629188919079e-05, "loss": 6.4705, "step": 20352 }, { "epoch": 0.8581065413074183, "grad_norm": 0.8073245286941528, "learning_rate": 3.9665511102782723e-05, "loss": 6.8322, "step": 20353 }, { "epoch": 0.8581487024896178, "grad_norm": 2.6320056915283203, "learning_rate": 3.9664730316374655e-05, "loss": 6.6465, "step": 20354 }, { "epoch": 0.8581908636718174, "grad_norm": 3.569744110107422, "learning_rate": 3.966394952996658e-05, "loss": 7.714, "step": 20355 }, { "epoch": 0.8582330248540169, "grad_norm": 1.1889512538909912, "learning_rate": 3.966316874355852e-05, "loss": 6.5042, "step": 20356 }, { "epoch": 0.8582751860362164, "grad_norm": 2.9475109577178955, "learning_rate": 3.966238795715044e-05, "loss": 7.6467, "step": 20357 }, { "epoch": 0.858317347218416, "grad_norm": 1.6694540977478027, "learning_rate": 3.966160717074237e-05, "loss": 6.7315, "step": 20358 }, { "epoch": 0.8583595084006156, "grad_norm": 1.1972843408584595, "learning_rate": 3.9660826384334304e-05, "loss": 6.6158, "step": 20359 }, { "epoch": 0.858401669582815, "grad_norm": 1.5001798868179321, "learning_rate": 3.966004559792623e-05, "loss": 6.9583, "step": 20360 }, { "epoch": 0.8584438307650146, "grad_norm": 1.3263133764266968, "learning_rate": 3.965926481151817e-05, "loss": 6.594, "step": 20361 }, { "epoch": 0.8584859919472142, "grad_norm": 1.8295258283615112, "learning_rate": 3.965848402511009e-05, "loss": 6.485, "step": 20362 }, { "epoch": 0.8585281531294138, "grad_norm": 1.4314148426055908, "learning_rate": 3.965770323870202e-05, "loss": 6.7901, "step": 20363 }, { "epoch": 0.8585703143116133, "grad_norm": 1.2060474157333374, "learning_rate": 3.9656922452293954e-05, "loss": 6.3847, "step": 20364 }, { "epoch": 0.8586124754938128, "grad_norm": 1.2087353467941284, "learning_rate": 3.9656141665885885e-05, "loss": 6.5134, "step": 20365 }, { "epoch": 0.8586546366760124, "grad_norm": 1.5695698261260986, "learning_rate": 3.965536087947781e-05, "loss": 7.3044, "step": 20366 }, { "epoch": 0.858696797858212, "grad_norm": 3.2554149627685547, "learning_rate": 3.965458009306974e-05, "loss": 7.8731, "step": 20367 }, { "epoch": 0.8587389590404115, "grad_norm": 1.3238826990127563, "learning_rate": 3.965379930666167e-05, "loss": 6.5038, "step": 20368 }, { "epoch": 0.858781120222611, "grad_norm": 1.2996147871017456, "learning_rate": 3.9653018520253596e-05, "loss": 6.6614, "step": 20369 }, { "epoch": 0.8588232814048106, "grad_norm": 1.4920889139175415, "learning_rate": 3.9652237733845535e-05, "loss": 6.5889, "step": 20370 }, { "epoch": 0.8588654425870101, "grad_norm": 1.1563026905059814, "learning_rate": 3.965145694743746e-05, "loss": 6.8329, "step": 20371 }, { "epoch": 0.8589076037692097, "grad_norm": 1.1813268661499023, "learning_rate": 3.965067616102939e-05, "loss": 6.8871, "step": 20372 }, { "epoch": 0.8589497649514093, "grad_norm": 1.3195732831954956, "learning_rate": 3.964989537462132e-05, "loss": 7.1379, "step": 20373 }, { "epoch": 0.8589919261336088, "grad_norm": 1.210866928100586, "learning_rate": 3.9649114588213246e-05, "loss": 6.6162, "step": 20374 }, { "epoch": 0.8590340873158083, "grad_norm": 2.763380289077759, "learning_rate": 3.9648333801805184e-05, "loss": 7.5551, "step": 20375 }, { "epoch": 0.8590762484980079, "grad_norm": 1.6502349376678467, "learning_rate": 3.964755301539711e-05, "loss": 7.0494, "step": 20376 }, { "epoch": 0.8591184096802075, "grad_norm": 1.8006677627563477, "learning_rate": 3.964677222898904e-05, "loss": 6.6492, "step": 20377 }, { "epoch": 0.8591605708624069, "grad_norm": 2.14363956451416, "learning_rate": 3.964599144258097e-05, "loss": 7.1236, "step": 20378 }, { "epoch": 0.8592027320446065, "grad_norm": 1.5821435451507568, "learning_rate": 3.9645210656172896e-05, "loss": 6.5407, "step": 20379 }, { "epoch": 0.8592448932268061, "grad_norm": 1.3953511714935303, "learning_rate": 3.964442986976483e-05, "loss": 6.6064, "step": 20380 }, { "epoch": 0.8592870544090057, "grad_norm": 1.4218714237213135, "learning_rate": 3.964364908335676e-05, "loss": 6.7215, "step": 20381 }, { "epoch": 0.8593292155912052, "grad_norm": 1.144484519958496, "learning_rate": 3.964286829694869e-05, "loss": 6.5528, "step": 20382 }, { "epoch": 0.8593713767734047, "grad_norm": 1.1536743640899658, "learning_rate": 3.964208751054062e-05, "loss": 6.3691, "step": 20383 }, { "epoch": 0.8594135379556043, "grad_norm": 1.1401751041412354, "learning_rate": 3.964130672413255e-05, "loss": 6.402, "step": 20384 }, { "epoch": 0.8594556991378038, "grad_norm": 0.7257580161094666, "learning_rate": 3.9640525937724476e-05, "loss": 6.5538, "step": 20385 }, { "epoch": 0.8594978603200034, "grad_norm": 1.2107959985733032, "learning_rate": 3.963974515131641e-05, "loss": 6.6294, "step": 20386 }, { "epoch": 0.8595400215022029, "grad_norm": 3.5531468391418457, "learning_rate": 3.963896436490834e-05, "loss": 9.6917, "step": 20387 }, { "epoch": 0.8595821826844025, "grad_norm": 1.9555182456970215, "learning_rate": 3.963818357850026e-05, "loss": 6.8485, "step": 20388 }, { "epoch": 0.859624343866602, "grad_norm": 1.342427372932434, "learning_rate": 3.96374027920922e-05, "loss": 7.3516, "step": 20389 }, { "epoch": 0.8596665050488016, "grad_norm": 3.2151730060577393, "learning_rate": 3.9636622005684126e-05, "loss": 9.0507, "step": 20390 }, { "epoch": 0.8597086662310011, "grad_norm": 0.7643592953681946, "learning_rate": 3.963584121927606e-05, "loss": 6.6451, "step": 20391 }, { "epoch": 0.8597508274132006, "grad_norm": 1.9386587142944336, "learning_rate": 3.963506043286799e-05, "loss": 6.9733, "step": 20392 }, { "epoch": 0.8597929885954002, "grad_norm": 1.6170579195022583, "learning_rate": 3.963427964645991e-05, "loss": 6.4642, "step": 20393 }, { "epoch": 0.8598351497775998, "grad_norm": 1.4214764833450317, "learning_rate": 3.963349886005185e-05, "loss": 7.0928, "step": 20394 }, { "epoch": 0.8598773109597994, "grad_norm": 3.0555572509765625, "learning_rate": 3.9632718073643775e-05, "loss": 7.7395, "step": 20395 }, { "epoch": 0.8599194721419988, "grad_norm": 1.383537769317627, "learning_rate": 3.963193728723571e-05, "loss": 6.7101, "step": 20396 }, { "epoch": 0.8599616333241984, "grad_norm": 1.5192190408706665, "learning_rate": 3.963115650082764e-05, "loss": 7.3072, "step": 20397 }, { "epoch": 0.860003794506398, "grad_norm": 1.1535687446594238, "learning_rate": 3.963037571441956e-05, "loss": 6.8423, "step": 20398 }, { "epoch": 0.8600459556885975, "grad_norm": 0.8398581743240356, "learning_rate": 3.9629594928011494e-05, "loss": 6.4584, "step": 20399 }, { "epoch": 0.860088116870797, "grad_norm": 3.451698064804077, "learning_rate": 3.9628814141603425e-05, "loss": 7.6137, "step": 20400 }, { "epoch": 0.8601302780529966, "grad_norm": 4.355199337005615, "learning_rate": 3.9628033355195356e-05, "loss": 8.769, "step": 20401 }, { "epoch": 0.8601724392351962, "grad_norm": 1.4946198463439941, "learning_rate": 3.962725256878728e-05, "loss": 6.9378, "step": 20402 }, { "epoch": 0.8602146004173957, "grad_norm": 2.135232448577881, "learning_rate": 3.962647178237922e-05, "loss": 6.4933, "step": 20403 }, { "epoch": 0.8602567615995953, "grad_norm": 2.9891090393066406, "learning_rate": 3.962569099597114e-05, "loss": 6.5498, "step": 20404 }, { "epoch": 0.8602989227817948, "grad_norm": 1.2528153657913208, "learning_rate": 3.9624910209563074e-05, "loss": 7.0585, "step": 20405 }, { "epoch": 0.8603410839639943, "grad_norm": 1.6230031251907349, "learning_rate": 3.9624129423155006e-05, "loss": 6.693, "step": 20406 }, { "epoch": 0.8603832451461939, "grad_norm": 1.1827183961868286, "learning_rate": 3.962334863674693e-05, "loss": 6.7856, "step": 20407 }, { "epoch": 0.8604254063283935, "grad_norm": 1.1184751987457275, "learning_rate": 3.962256785033887e-05, "loss": 6.8995, "step": 20408 }, { "epoch": 0.860467567510593, "grad_norm": 1.4078354835510254, "learning_rate": 3.962178706393079e-05, "loss": 6.5296, "step": 20409 }, { "epoch": 0.8605097286927925, "grad_norm": 1.6487761735916138, "learning_rate": 3.9621006277522724e-05, "loss": 6.5102, "step": 20410 }, { "epoch": 0.8605518898749921, "grad_norm": 1.9002485275268555, "learning_rate": 3.9620225491114655e-05, "loss": 6.872, "step": 20411 }, { "epoch": 0.8605940510571917, "grad_norm": 1.2327258586883545, "learning_rate": 3.961944470470658e-05, "loss": 6.5516, "step": 20412 }, { "epoch": 0.8606362122393912, "grad_norm": 4.678612232208252, "learning_rate": 3.961866391829851e-05, "loss": 8.5964, "step": 20413 }, { "epoch": 0.8606783734215907, "grad_norm": 1.3831751346588135, "learning_rate": 3.961788313189044e-05, "loss": 6.5257, "step": 20414 }, { "epoch": 0.8607205346037903, "grad_norm": 2.0554635524749756, "learning_rate": 3.9617102345482373e-05, "loss": 6.517, "step": 20415 }, { "epoch": 0.8607626957859899, "grad_norm": 1.494629144668579, "learning_rate": 3.9616321559074305e-05, "loss": 6.5622, "step": 20416 }, { "epoch": 0.8608048569681894, "grad_norm": 1.2294844388961792, "learning_rate": 3.961554077266623e-05, "loss": 6.5934, "step": 20417 }, { "epoch": 0.8608470181503889, "grad_norm": 2.74664044380188, "learning_rate": 3.961475998625816e-05, "loss": 7.3028, "step": 20418 }, { "epoch": 0.8608891793325885, "grad_norm": 2.165313243865967, "learning_rate": 3.961397919985009e-05, "loss": 6.9324, "step": 20419 }, { "epoch": 0.860931340514788, "grad_norm": 2.314044713973999, "learning_rate": 3.961319841344202e-05, "loss": 6.6509, "step": 20420 }, { "epoch": 0.8609735016969876, "grad_norm": 2.731680393218994, "learning_rate": 3.961241762703395e-05, "loss": 6.4485, "step": 20421 }, { "epoch": 0.8610156628791872, "grad_norm": 1.315664291381836, "learning_rate": 3.961163684062588e-05, "loss": 6.4919, "step": 20422 }, { "epoch": 0.8610578240613866, "grad_norm": 0.9865461587905884, "learning_rate": 3.961085605421781e-05, "loss": 6.4224, "step": 20423 }, { "epoch": 0.8610999852435862, "grad_norm": 1.238721251487732, "learning_rate": 3.9610075267809734e-05, "loss": 6.4869, "step": 20424 }, { "epoch": 0.8611421464257858, "grad_norm": 1.7526212930679321, "learning_rate": 3.960929448140167e-05, "loss": 6.8401, "step": 20425 }, { "epoch": 0.8611843076079854, "grad_norm": 4.983703136444092, "learning_rate": 3.96085136949936e-05, "loss": 7.1548, "step": 20426 }, { "epoch": 0.8612264687901848, "grad_norm": 2.455068826675415, "learning_rate": 3.9607732908585535e-05, "loss": 6.4066, "step": 20427 }, { "epoch": 0.8612686299723844, "grad_norm": 1.9805951118469238, "learning_rate": 3.960695212217746e-05, "loss": 7.0513, "step": 20428 }, { "epoch": 0.861310791154584, "grad_norm": 4.263855934143066, "learning_rate": 3.960617133576939e-05, "loss": 7.3479, "step": 20429 }, { "epoch": 0.8613529523367836, "grad_norm": 2.8640732765197754, "learning_rate": 3.960539054936132e-05, "loss": 7.5897, "step": 20430 }, { "epoch": 0.8613951135189831, "grad_norm": 1.6248465776443481, "learning_rate": 3.9604609762953247e-05, "loss": 7.0926, "step": 20431 }, { "epoch": 0.8614372747011826, "grad_norm": 1.6383007764816284, "learning_rate": 3.960382897654518e-05, "loss": 7.2685, "step": 20432 }, { "epoch": 0.8614794358833822, "grad_norm": 2.430919647216797, "learning_rate": 3.960304819013711e-05, "loss": 6.5665, "step": 20433 }, { "epoch": 0.8615215970655817, "grad_norm": 1.471871018409729, "learning_rate": 3.960226740372904e-05, "loss": 6.8745, "step": 20434 }, { "epoch": 0.8615637582477813, "grad_norm": 2.0042519569396973, "learning_rate": 3.9601486617320965e-05, "loss": 6.4312, "step": 20435 }, { "epoch": 0.8616059194299808, "grad_norm": 1.6189589500427246, "learning_rate": 3.9600705830912896e-05, "loss": 6.3744, "step": 20436 }, { "epoch": 0.8616480806121803, "grad_norm": 1.7999942302703857, "learning_rate": 3.959992504450483e-05, "loss": 6.884, "step": 20437 }, { "epoch": 0.8616902417943799, "grad_norm": 1.4578776359558105, "learning_rate": 3.959914425809675e-05, "loss": 7.0056, "step": 20438 }, { "epoch": 0.8617324029765795, "grad_norm": 1.4234474897384644, "learning_rate": 3.959836347168869e-05, "loss": 6.8298, "step": 20439 }, { "epoch": 0.8617745641587791, "grad_norm": 2.3890202045440674, "learning_rate": 3.9597582685280614e-05, "loss": 6.8513, "step": 20440 }, { "epoch": 0.8618167253409785, "grad_norm": 1.5820162296295166, "learning_rate": 3.9596801898872546e-05, "loss": 6.3503, "step": 20441 }, { "epoch": 0.8618588865231781, "grad_norm": 2.579685926437378, "learning_rate": 3.959602111246448e-05, "loss": 6.3055, "step": 20442 }, { "epoch": 0.8619010477053777, "grad_norm": 1.362074851989746, "learning_rate": 3.95952403260564e-05, "loss": 6.8869, "step": 20443 }, { "epoch": 0.8619432088875773, "grad_norm": 1.6628438234329224, "learning_rate": 3.959445953964834e-05, "loss": 6.5776, "step": 20444 }, { "epoch": 0.8619853700697767, "grad_norm": 1.503592848777771, "learning_rate": 3.9593678753240264e-05, "loss": 6.4983, "step": 20445 }, { "epoch": 0.8620275312519763, "grad_norm": 2.439851760864258, "learning_rate": 3.9592897966832195e-05, "loss": 6.2525, "step": 20446 }, { "epoch": 0.8620696924341759, "grad_norm": 1.06278657913208, "learning_rate": 3.9592117180424126e-05, "loss": 6.3788, "step": 20447 }, { "epoch": 0.8621118536163754, "grad_norm": 3.6691253185272217, "learning_rate": 3.959133639401605e-05, "loss": 7.7969, "step": 20448 }, { "epoch": 0.8621540147985749, "grad_norm": 1.5631327629089355, "learning_rate": 3.959055560760798e-05, "loss": 6.6607, "step": 20449 }, { "epoch": 0.8621961759807745, "grad_norm": 1.1759060621261597, "learning_rate": 3.958977482119991e-05, "loss": 6.3905, "step": 20450 }, { "epoch": 0.862238337162974, "grad_norm": 1.2915078401565552, "learning_rate": 3.9588994034791845e-05, "loss": 6.5505, "step": 20451 }, { "epoch": 0.8622804983451736, "grad_norm": 1.554639220237732, "learning_rate": 3.9588213248383776e-05, "loss": 6.2874, "step": 20452 }, { "epoch": 0.8623226595273732, "grad_norm": 2.3874616622924805, "learning_rate": 3.958743246197571e-05, "loss": 7.5532, "step": 20453 }, { "epoch": 0.8623648207095727, "grad_norm": 1.81571626663208, "learning_rate": 3.958665167556763e-05, "loss": 6.9368, "step": 20454 }, { "epoch": 0.8624069818917722, "grad_norm": 2.704141855239868, "learning_rate": 3.958587088915956e-05, "loss": 7.3476, "step": 20455 }, { "epoch": 0.8624491430739718, "grad_norm": 1.1931183338165283, "learning_rate": 3.9585090102751494e-05, "loss": 6.4645, "step": 20456 }, { "epoch": 0.8624913042561714, "grad_norm": 1.3526934385299683, "learning_rate": 3.958430931634342e-05, "loss": 6.6169, "step": 20457 }, { "epoch": 0.8625334654383708, "grad_norm": 2.546286106109619, "learning_rate": 3.958352852993536e-05, "loss": 7.5555, "step": 20458 }, { "epoch": 0.8625756266205704, "grad_norm": 1.3921107053756714, "learning_rate": 3.958274774352728e-05, "loss": 6.4494, "step": 20459 }, { "epoch": 0.86261778780277, "grad_norm": 0.8270801901817322, "learning_rate": 3.958196695711921e-05, "loss": 6.5224, "step": 20460 }, { "epoch": 0.8626599489849696, "grad_norm": 3.6484220027923584, "learning_rate": 3.9581186170711144e-05, "loss": 7.7462, "step": 20461 }, { "epoch": 0.8627021101671691, "grad_norm": 1.5846858024597168, "learning_rate": 3.958040538430307e-05, "loss": 6.6752, "step": 20462 }, { "epoch": 0.8627442713493686, "grad_norm": 3.8366622924804688, "learning_rate": 3.9579624597895006e-05, "loss": 7.7619, "step": 20463 }, { "epoch": 0.8627864325315682, "grad_norm": 1.2424370050430298, "learning_rate": 3.957884381148693e-05, "loss": 6.8089, "step": 20464 }, { "epoch": 0.8628285937137677, "grad_norm": 2.1314024925231934, "learning_rate": 3.957806302507886e-05, "loss": 7.1586, "step": 20465 }, { "epoch": 0.8628707548959673, "grad_norm": 1.9978910684585571, "learning_rate": 3.957728223867079e-05, "loss": 7.0441, "step": 20466 }, { "epoch": 0.8629129160781668, "grad_norm": 1.5299670696258545, "learning_rate": 3.957650145226272e-05, "loss": 6.7163, "step": 20467 }, { "epoch": 0.8629550772603664, "grad_norm": 1.274244785308838, "learning_rate": 3.957572066585465e-05, "loss": 8.2001, "step": 20468 }, { "epoch": 0.8629972384425659, "grad_norm": 1.6706347465515137, "learning_rate": 3.957493987944658e-05, "loss": 6.8699, "step": 20469 }, { "epoch": 0.8630393996247655, "grad_norm": 2.1866958141326904, "learning_rate": 3.957415909303851e-05, "loss": 6.5786, "step": 20470 }, { "epoch": 0.8630815608069651, "grad_norm": 3.258626937866211, "learning_rate": 3.9573378306630436e-05, "loss": 7.8984, "step": 20471 }, { "epoch": 0.8631237219891645, "grad_norm": 2.798090934753418, "learning_rate": 3.9572597520222374e-05, "loss": 7.5166, "step": 20472 }, { "epoch": 0.8631658831713641, "grad_norm": 0.8997994661331177, "learning_rate": 3.95718167338143e-05, "loss": 6.5204, "step": 20473 }, { "epoch": 0.8632080443535637, "grad_norm": 1.4094749689102173, "learning_rate": 3.957103594740623e-05, "loss": 6.4692, "step": 20474 }, { "epoch": 0.8632502055357633, "grad_norm": 2.038050413131714, "learning_rate": 3.957025516099816e-05, "loss": 7.1083, "step": 20475 }, { "epoch": 0.8632923667179627, "grad_norm": 1.6348267793655396, "learning_rate": 3.9569474374590085e-05, "loss": 6.3897, "step": 20476 }, { "epoch": 0.8633345279001623, "grad_norm": 1.7081661224365234, "learning_rate": 3.9568693588182023e-05, "loss": 6.9718, "step": 20477 }, { "epoch": 0.8633766890823619, "grad_norm": 1.587853193283081, "learning_rate": 3.956791280177395e-05, "loss": 7.1208, "step": 20478 }, { "epoch": 0.8634188502645614, "grad_norm": 2.3060901165008545, "learning_rate": 3.956713201536588e-05, "loss": 6.5133, "step": 20479 }, { "epoch": 0.863461011446761, "grad_norm": 1.5333808660507202, "learning_rate": 3.956635122895781e-05, "loss": 6.5533, "step": 20480 }, { "epoch": 0.8635031726289605, "grad_norm": 1.198446273803711, "learning_rate": 3.9565570442549735e-05, "loss": 6.7961, "step": 20481 }, { "epoch": 0.86354533381116, "grad_norm": 3.3420121669769287, "learning_rate": 3.9564789656141666e-05, "loss": 7.5561, "step": 20482 }, { "epoch": 0.8635874949933596, "grad_norm": 1.316858172416687, "learning_rate": 3.95640088697336e-05, "loss": 6.927, "step": 20483 }, { "epoch": 0.8636296561755592, "grad_norm": 1.9279335737228394, "learning_rate": 3.956322808332553e-05, "loss": 6.7434, "step": 20484 }, { "epoch": 0.8636718173577587, "grad_norm": 3.0994741916656494, "learning_rate": 3.956244729691746e-05, "loss": 7.2626, "step": 20485 }, { "epoch": 0.8637139785399582, "grad_norm": 1.4045343399047852, "learning_rate": 3.9561666510509384e-05, "loss": 6.5648, "step": 20486 }, { "epoch": 0.8637561397221578, "grad_norm": 1.4886994361877441, "learning_rate": 3.9560885724101316e-05, "loss": 7.0048, "step": 20487 }, { "epoch": 0.8637983009043574, "grad_norm": 1.7623671293258667, "learning_rate": 3.956010493769325e-05, "loss": 6.8161, "step": 20488 }, { "epoch": 0.863840462086557, "grad_norm": 1.5666840076446533, "learning_rate": 3.955932415128518e-05, "loss": 6.9522, "step": 20489 }, { "epoch": 0.8638826232687564, "grad_norm": 1.6369773149490356, "learning_rate": 3.95585433648771e-05, "loss": 6.9776, "step": 20490 }, { "epoch": 0.863924784450956, "grad_norm": 2.44851016998291, "learning_rate": 3.955776257846904e-05, "loss": 7.0283, "step": 20491 }, { "epoch": 0.8639669456331556, "grad_norm": 1.4656952619552612, "learning_rate": 3.9556981792060965e-05, "loss": 6.6605, "step": 20492 }, { "epoch": 0.8640091068153551, "grad_norm": 2.829740047454834, "learning_rate": 3.955620100565289e-05, "loss": 6.3985, "step": 20493 }, { "epoch": 0.8640512679975546, "grad_norm": 3.0433762073516846, "learning_rate": 3.955542021924483e-05, "loss": 7.7787, "step": 20494 }, { "epoch": 0.8640934291797542, "grad_norm": 2.8700764179229736, "learning_rate": 3.955463943283675e-05, "loss": 7.6564, "step": 20495 }, { "epoch": 0.8641355903619538, "grad_norm": 2.151063919067383, "learning_rate": 3.955385864642869e-05, "loss": 6.5165, "step": 20496 }, { "epoch": 0.8641777515441533, "grad_norm": 1.199105143547058, "learning_rate": 3.9553077860020615e-05, "loss": 6.6204, "step": 20497 }, { "epoch": 0.8642199127263529, "grad_norm": 2.1315295696258545, "learning_rate": 3.9552297073612546e-05, "loss": 6.9675, "step": 20498 }, { "epoch": 0.8642620739085524, "grad_norm": 2.018038034439087, "learning_rate": 3.955151628720448e-05, "loss": 7.7503, "step": 20499 }, { "epoch": 0.8643042350907519, "grad_norm": 2.1327526569366455, "learning_rate": 3.95507355007964e-05, "loss": 7.3856, "step": 20500 }, { "epoch": 0.8643463962729515, "grad_norm": 3.8584799766540527, "learning_rate": 3.954995471438833e-05, "loss": 9.3827, "step": 20501 }, { "epoch": 0.8643885574551511, "grad_norm": 1.4448829889297485, "learning_rate": 3.9549173927980264e-05, "loss": 6.926, "step": 20502 }, { "epoch": 0.8644307186373505, "grad_norm": 1.4651566743850708, "learning_rate": 3.9548393141572196e-05, "loss": 7.1357, "step": 20503 }, { "epoch": 0.8644728798195501, "grad_norm": 1.6450411081314087, "learning_rate": 3.954761235516412e-05, "loss": 6.6097, "step": 20504 }, { "epoch": 0.8645150410017497, "grad_norm": 1.5780141353607178, "learning_rate": 3.954683156875605e-05, "loss": 6.7667, "step": 20505 }, { "epoch": 0.8645572021839493, "grad_norm": 3.3572492599487305, "learning_rate": 3.954605078234798e-05, "loss": 7.5175, "step": 20506 }, { "epoch": 0.8645993633661487, "grad_norm": 3.990086555480957, "learning_rate": 3.9545269995939914e-05, "loss": 7.9448, "step": 20507 }, { "epoch": 0.8646415245483483, "grad_norm": 3.2378885746002197, "learning_rate": 3.9544489209531845e-05, "loss": 7.013, "step": 20508 }, { "epoch": 0.8646836857305479, "grad_norm": 1.9605214595794678, "learning_rate": 3.954370842312377e-05, "loss": 6.9076, "step": 20509 }, { "epoch": 0.8647258469127475, "grad_norm": 1.6769624948501587, "learning_rate": 3.95429276367157e-05, "loss": 6.9785, "step": 20510 }, { "epoch": 0.864768008094947, "grad_norm": 2.6303226947784424, "learning_rate": 3.954214685030763e-05, "loss": 7.6176, "step": 20511 }, { "epoch": 0.8648101692771465, "grad_norm": 1.7897956371307373, "learning_rate": 3.9541366063899557e-05, "loss": 6.5585, "step": 20512 }, { "epoch": 0.8648523304593461, "grad_norm": 1.348613977432251, "learning_rate": 3.9540585277491495e-05, "loss": 7.0032, "step": 20513 }, { "epoch": 0.8648944916415456, "grad_norm": 2.442265748977661, "learning_rate": 3.953980449108342e-05, "loss": 7.41, "step": 20514 }, { "epoch": 0.8649366528237452, "grad_norm": 2.7777810096740723, "learning_rate": 3.953902370467535e-05, "loss": 7.1986, "step": 20515 }, { "epoch": 0.8649788140059447, "grad_norm": 2.684901714324951, "learning_rate": 3.953824291826728e-05, "loss": 6.7019, "step": 20516 }, { "epoch": 0.8650209751881442, "grad_norm": 1.999045491218567, "learning_rate": 3.953746213185921e-05, "loss": 6.501, "step": 20517 }, { "epoch": 0.8650631363703438, "grad_norm": 1.0684361457824707, "learning_rate": 3.953668134545114e-05, "loss": 7.0232, "step": 20518 }, { "epoch": 0.8651052975525434, "grad_norm": 1.2282590866088867, "learning_rate": 3.953590055904307e-05, "loss": 6.9059, "step": 20519 }, { "epoch": 0.865147458734743, "grad_norm": 2.0843920707702637, "learning_rate": 3.9535119772635e-05, "loss": 6.9674, "step": 20520 }, { "epoch": 0.8651896199169424, "grad_norm": 2.472468137741089, "learning_rate": 3.953433898622693e-05, "loss": 6.9728, "step": 20521 }, { "epoch": 0.865231781099142, "grad_norm": 2.1360788345336914, "learning_rate": 3.953355819981886e-05, "loss": 7.2503, "step": 20522 }, { "epoch": 0.8652739422813416, "grad_norm": 1.218886137008667, "learning_rate": 3.953277741341079e-05, "loss": 7.0073, "step": 20523 }, { "epoch": 0.8653161034635412, "grad_norm": 1.6786268949508667, "learning_rate": 3.953199662700272e-05, "loss": 8.1008, "step": 20524 }, { "epoch": 0.8653582646457406, "grad_norm": 2.1108126640319824, "learning_rate": 3.953121584059465e-05, "loss": 7.0509, "step": 20525 }, { "epoch": 0.8654004258279402, "grad_norm": 3.1269330978393555, "learning_rate": 3.9530435054186574e-05, "loss": 7.6041, "step": 20526 }, { "epoch": 0.8654425870101398, "grad_norm": 2.8829522132873535, "learning_rate": 3.952965426777851e-05, "loss": 7.4606, "step": 20527 }, { "epoch": 0.8654847481923393, "grad_norm": 2.3753716945648193, "learning_rate": 3.9528873481370436e-05, "loss": 6.5377, "step": 20528 }, { "epoch": 0.8655269093745389, "grad_norm": 1.9875510931015015, "learning_rate": 3.952809269496237e-05, "loss": 6.7623, "step": 20529 }, { "epoch": 0.8655690705567384, "grad_norm": 1.4373018741607666, "learning_rate": 3.95273119085543e-05, "loss": 6.7147, "step": 20530 }, { "epoch": 0.865611231738938, "grad_norm": 2.0515336990356445, "learning_rate": 3.9526531122146223e-05, "loss": 6.4323, "step": 20531 }, { "epoch": 0.8656533929211375, "grad_norm": 3.373427391052246, "learning_rate": 3.952575033573816e-05, "loss": 6.4618, "step": 20532 }, { "epoch": 0.8656955541033371, "grad_norm": 4.160336494445801, "learning_rate": 3.9524969549330086e-05, "loss": 7.7938, "step": 20533 }, { "epoch": 0.8657377152855366, "grad_norm": 2.0518290996551514, "learning_rate": 3.952418876292202e-05, "loss": 6.7426, "step": 20534 }, { "epoch": 0.8657798764677361, "grad_norm": 3.3431949615478516, "learning_rate": 3.952340797651395e-05, "loss": 6.4119, "step": 20535 }, { "epoch": 0.8658220376499357, "grad_norm": 1.5804208517074585, "learning_rate": 3.952262719010588e-05, "loss": 7.265, "step": 20536 }, { "epoch": 0.8658641988321353, "grad_norm": 2.1094424724578857, "learning_rate": 3.9521846403697804e-05, "loss": 6.5793, "step": 20537 }, { "epoch": 0.8659063600143349, "grad_norm": 2.8615498542785645, "learning_rate": 3.9521065617289735e-05, "loss": 7.654, "step": 20538 }, { "epoch": 0.8659485211965343, "grad_norm": 1.7803791761398315, "learning_rate": 3.952028483088167e-05, "loss": 6.7127, "step": 20539 }, { "epoch": 0.8659906823787339, "grad_norm": 2.6774983406066895, "learning_rate": 3.951950404447359e-05, "loss": 9.2165, "step": 20540 }, { "epoch": 0.8660328435609335, "grad_norm": 2.165935516357422, "learning_rate": 3.951872325806553e-05, "loss": 6.6165, "step": 20541 }, { "epoch": 0.866075004743133, "grad_norm": 2.195943832397461, "learning_rate": 3.9517942471657454e-05, "loss": 6.3493, "step": 20542 }, { "epoch": 0.8661171659253325, "grad_norm": 1.9473363161087036, "learning_rate": 3.9517161685249385e-05, "loss": 7.2254, "step": 20543 }, { "epoch": 0.8661593271075321, "grad_norm": 3.2640609741210938, "learning_rate": 3.9516380898841316e-05, "loss": 7.1857, "step": 20544 }, { "epoch": 0.8662014882897316, "grad_norm": 2.752501964569092, "learning_rate": 3.951560011243324e-05, "loss": 7.7332, "step": 20545 }, { "epoch": 0.8662436494719312, "grad_norm": 2.2652587890625, "learning_rate": 3.951481932602518e-05, "loss": 6.751, "step": 20546 }, { "epoch": 0.8662858106541308, "grad_norm": 3.329676866531372, "learning_rate": 3.95140385396171e-05, "loss": 7.7567, "step": 20547 }, { "epoch": 0.8663279718363303, "grad_norm": 4.763502597808838, "learning_rate": 3.9513257753209035e-05, "loss": 7.5106, "step": 20548 }, { "epoch": 0.8663701330185298, "grad_norm": 2.9476118087768555, "learning_rate": 3.9512476966800966e-05, "loss": 7.0243, "step": 20549 }, { "epoch": 0.8664122942007294, "grad_norm": 1.8610470294952393, "learning_rate": 3.951169618039289e-05, "loss": 6.5391, "step": 20550 }, { "epoch": 0.866454455382929, "grad_norm": 2.249173164367676, "learning_rate": 3.951091539398482e-05, "loss": 7.802, "step": 20551 }, { "epoch": 0.8664966165651284, "grad_norm": 1.5065511465072632, "learning_rate": 3.951013460757675e-05, "loss": 6.5942, "step": 20552 }, { "epoch": 0.866538777747328, "grad_norm": 2.8452799320220947, "learning_rate": 3.9509353821168684e-05, "loss": 7.2587, "step": 20553 }, { "epoch": 0.8665809389295276, "grad_norm": 2.224360227584839, "learning_rate": 3.9508573034760615e-05, "loss": 7.3865, "step": 20554 }, { "epoch": 0.8666231001117272, "grad_norm": 2.5614490509033203, "learning_rate": 3.950779224835254e-05, "loss": 6.5487, "step": 20555 }, { "epoch": 0.8666652612939267, "grad_norm": 1.7213929891586304, "learning_rate": 3.950701146194447e-05, "loss": 7.1276, "step": 20556 }, { "epoch": 0.8667074224761262, "grad_norm": 1.5572508573532104, "learning_rate": 3.95062306755364e-05, "loss": 7.0022, "step": 20557 }, { "epoch": 0.8667495836583258, "grad_norm": 1.7640693187713623, "learning_rate": 3.9505449889128334e-05, "loss": 6.7968, "step": 20558 }, { "epoch": 0.8667917448405253, "grad_norm": 1.9458893537521362, "learning_rate": 3.950466910272026e-05, "loss": 6.8858, "step": 20559 }, { "epoch": 0.8668339060227249, "grad_norm": 2.1725521087646484, "learning_rate": 3.9503888316312196e-05, "loss": 6.6715, "step": 20560 }, { "epoch": 0.8668760672049244, "grad_norm": 3.384509325027466, "learning_rate": 3.950310752990412e-05, "loss": 8.2492, "step": 20561 }, { "epoch": 0.866918228387124, "grad_norm": 1.0648624897003174, "learning_rate": 3.950232674349605e-05, "loss": 6.5366, "step": 20562 }, { "epoch": 0.8669603895693235, "grad_norm": 1.082798957824707, "learning_rate": 3.950154595708798e-05, "loss": 6.6783, "step": 20563 }, { "epoch": 0.8670025507515231, "grad_norm": 1.696754813194275, "learning_rate": 3.950076517067991e-05, "loss": 6.4884, "step": 20564 }, { "epoch": 0.8670447119337226, "grad_norm": 3.0150582790374756, "learning_rate": 3.9499984384271846e-05, "loss": 7.712, "step": 20565 }, { "epoch": 0.8670868731159221, "grad_norm": 3.1283175945281982, "learning_rate": 3.949920359786377e-05, "loss": 6.492, "step": 20566 }, { "epoch": 0.8671290342981217, "grad_norm": 3.1679248809814453, "learning_rate": 3.94984228114557e-05, "loss": 6.452, "step": 20567 }, { "epoch": 0.8671711954803213, "grad_norm": 5.580162048339844, "learning_rate": 3.949764202504763e-05, "loss": 6.8144, "step": 20568 }, { "epoch": 0.8672133566625209, "grad_norm": 3.7948484420776367, "learning_rate": 3.949686123863956e-05, "loss": 7.2237, "step": 20569 }, { "epoch": 0.8672555178447203, "grad_norm": 2.122063398361206, "learning_rate": 3.949608045223149e-05, "loss": 6.9737, "step": 20570 }, { "epoch": 0.8672976790269199, "grad_norm": 2.7662041187286377, "learning_rate": 3.949529966582342e-05, "loss": 7.5294, "step": 20571 }, { "epoch": 0.8673398402091195, "grad_norm": 1.4444522857666016, "learning_rate": 3.949451887941535e-05, "loss": 7.1419, "step": 20572 }, { "epoch": 0.867382001391319, "grad_norm": 3.5171148777008057, "learning_rate": 3.9493738093007275e-05, "loss": 6.9602, "step": 20573 }, { "epoch": 0.8674241625735185, "grad_norm": 2.9487109184265137, "learning_rate": 3.9492957306599207e-05, "loss": 7.5968, "step": 20574 }, { "epoch": 0.8674663237557181, "grad_norm": 1.1592304706573486, "learning_rate": 3.949217652019114e-05, "loss": 6.8966, "step": 20575 }, { "epoch": 0.8675084849379177, "grad_norm": 3.7700319290161133, "learning_rate": 3.949139573378307e-05, "loss": 9.2747, "step": 20576 }, { "epoch": 0.8675506461201172, "grad_norm": 1.232994556427002, "learning_rate": 3.9490614947375e-05, "loss": 7.1398, "step": 20577 }, { "epoch": 0.8675928073023168, "grad_norm": 2.8720450401306152, "learning_rate": 3.9489834160966925e-05, "loss": 6.348, "step": 20578 }, { "epoch": 0.8676349684845163, "grad_norm": 1.4251949787139893, "learning_rate": 3.948905337455886e-05, "loss": 6.4596, "step": 20579 }, { "epoch": 0.8676771296667158, "grad_norm": 1.540412425994873, "learning_rate": 3.948827258815079e-05, "loss": 6.9753, "step": 20580 }, { "epoch": 0.8677192908489154, "grad_norm": 2.7056169509887695, "learning_rate": 3.948749180174272e-05, "loss": 7.1907, "step": 20581 }, { "epoch": 0.867761452031115, "grad_norm": 3.410155773162842, "learning_rate": 3.948671101533465e-05, "loss": 7.7354, "step": 20582 }, { "epoch": 0.8678036132133145, "grad_norm": 2.405700206756592, "learning_rate": 3.9485930228926574e-05, "loss": 7.0855, "step": 20583 }, { "epoch": 0.867845774395514, "grad_norm": 1.9677313566207886, "learning_rate": 3.9485149442518506e-05, "loss": 7.0766, "step": 20584 }, { "epoch": 0.8678879355777136, "grad_norm": 1.169823408126831, "learning_rate": 3.948436865611044e-05, "loss": 6.7068, "step": 20585 }, { "epoch": 0.8679300967599132, "grad_norm": 1.378501057624817, "learning_rate": 3.948358786970237e-05, "loss": 6.9646, "step": 20586 }, { "epoch": 0.8679722579421127, "grad_norm": 1.278203010559082, "learning_rate": 3.94828070832943e-05, "loss": 6.7596, "step": 20587 }, { "epoch": 0.8680144191243122, "grad_norm": 1.3547254800796509, "learning_rate": 3.9482026296886224e-05, "loss": 6.4577, "step": 20588 }, { "epoch": 0.8680565803065118, "grad_norm": 1.2962491512298584, "learning_rate": 3.9481245510478155e-05, "loss": 7.0728, "step": 20589 }, { "epoch": 0.8680987414887114, "grad_norm": 1.285634160041809, "learning_rate": 3.9480464724070086e-05, "loss": 6.6669, "step": 20590 }, { "epoch": 0.8681409026709109, "grad_norm": 1.0745415687561035, "learning_rate": 3.947968393766202e-05, "loss": 6.5344, "step": 20591 }, { "epoch": 0.8681830638531104, "grad_norm": 2.4910013675689697, "learning_rate": 3.947890315125394e-05, "loss": 6.551, "step": 20592 }, { "epoch": 0.86822522503531, "grad_norm": 2.0086069107055664, "learning_rate": 3.9478122364845873e-05, "loss": 7.36, "step": 20593 }, { "epoch": 0.8682673862175095, "grad_norm": 1.9855619668960571, "learning_rate": 3.9477341578437805e-05, "loss": 7.0981, "step": 20594 }, { "epoch": 0.8683095473997091, "grad_norm": 2.4482953548431396, "learning_rate": 3.947656079202973e-05, "loss": 6.5593, "step": 20595 }, { "epoch": 0.8683517085819087, "grad_norm": 1.6274577379226685, "learning_rate": 3.947578000562167e-05, "loss": 6.7667, "step": 20596 }, { "epoch": 0.8683938697641082, "grad_norm": 3.2737975120544434, "learning_rate": 3.947499921921359e-05, "loss": 7.7358, "step": 20597 }, { "epoch": 0.8684360309463077, "grad_norm": 3.697629451751709, "learning_rate": 3.947421843280552e-05, "loss": 8.0392, "step": 20598 }, { "epoch": 0.8684781921285073, "grad_norm": 1.8158352375030518, "learning_rate": 3.9473437646397454e-05, "loss": 6.6248, "step": 20599 }, { "epoch": 0.8685203533107069, "grad_norm": 1.2287338972091675, "learning_rate": 3.947265685998938e-05, "loss": 7.0436, "step": 20600 }, { "epoch": 0.8685625144929063, "grad_norm": 1.1819262504577637, "learning_rate": 3.947187607358132e-05, "loss": 6.6166, "step": 20601 }, { "epoch": 0.8686046756751059, "grad_norm": 4.887109279632568, "learning_rate": 3.947109528717324e-05, "loss": 10.2776, "step": 20602 }, { "epoch": 0.8686468368573055, "grad_norm": 3.9349162578582764, "learning_rate": 3.947031450076517e-05, "loss": 9.0817, "step": 20603 }, { "epoch": 0.8686889980395051, "grad_norm": 1.3322293758392334, "learning_rate": 3.9469533714357104e-05, "loss": 6.7226, "step": 20604 }, { "epoch": 0.8687311592217046, "grad_norm": 1.0695998668670654, "learning_rate": 3.9468752927949035e-05, "loss": 6.4981, "step": 20605 }, { "epoch": 0.8687733204039041, "grad_norm": 3.617053985595703, "learning_rate": 3.946797214154096e-05, "loss": 7.7452, "step": 20606 }, { "epoch": 0.8688154815861037, "grad_norm": 1.9861584901809692, "learning_rate": 3.946719135513289e-05, "loss": 6.9297, "step": 20607 }, { "epoch": 0.8688576427683032, "grad_norm": 1.50371253490448, "learning_rate": 3.946641056872482e-05, "loss": 7.5358, "step": 20608 }, { "epoch": 0.8688998039505028, "grad_norm": 1.5263792276382446, "learning_rate": 3.9465629782316746e-05, "loss": 7.185, "step": 20609 }, { "epoch": 0.8689419651327023, "grad_norm": 2.8318278789520264, "learning_rate": 3.9464848995908685e-05, "loss": 7.6869, "step": 20610 }, { "epoch": 0.8689841263149019, "grad_norm": 2.7945358753204346, "learning_rate": 3.946406820950061e-05, "loss": 6.6546, "step": 20611 }, { "epoch": 0.8690262874971014, "grad_norm": 2.8391804695129395, "learning_rate": 3.946328742309254e-05, "loss": 7.616, "step": 20612 }, { "epoch": 0.869068448679301, "grad_norm": 3.6375105381011963, "learning_rate": 3.946250663668447e-05, "loss": 6.7928, "step": 20613 }, { "epoch": 0.8691106098615006, "grad_norm": 2.9846954345703125, "learning_rate": 3.9461725850276396e-05, "loss": 8.6282, "step": 20614 }, { "epoch": 0.8691527710437, "grad_norm": 1.6335124969482422, "learning_rate": 3.9460945063868334e-05, "loss": 6.8659, "step": 20615 }, { "epoch": 0.8691949322258996, "grad_norm": 2.2651748657226562, "learning_rate": 3.946016427746026e-05, "loss": 6.8823, "step": 20616 }, { "epoch": 0.8692370934080992, "grad_norm": 3.125979423522949, "learning_rate": 3.945938349105219e-05, "loss": 6.6441, "step": 20617 }, { "epoch": 0.8692792545902988, "grad_norm": 1.6340157985687256, "learning_rate": 3.945860270464412e-05, "loss": 6.623, "step": 20618 }, { "epoch": 0.8693214157724982, "grad_norm": 2.204955577850342, "learning_rate": 3.9457821918236046e-05, "loss": 6.6488, "step": 20619 }, { "epoch": 0.8693635769546978, "grad_norm": 4.033813953399658, "learning_rate": 3.945704113182798e-05, "loss": 9.5055, "step": 20620 }, { "epoch": 0.8694057381368974, "grad_norm": 2.5993432998657227, "learning_rate": 3.945626034541991e-05, "loss": 7.5572, "step": 20621 }, { "epoch": 0.8694478993190969, "grad_norm": 1.8830361366271973, "learning_rate": 3.945547955901184e-05, "loss": 7.0953, "step": 20622 }, { "epoch": 0.8694900605012964, "grad_norm": 1.8303602933883667, "learning_rate": 3.945469877260377e-05, "loss": 6.8475, "step": 20623 }, { "epoch": 0.869532221683496, "grad_norm": 3.0972156524658203, "learning_rate": 3.94539179861957e-05, "loss": 7.7692, "step": 20624 }, { "epoch": 0.8695743828656955, "grad_norm": 2.934832811355591, "learning_rate": 3.9453137199787626e-05, "loss": 7.3591, "step": 20625 }, { "epoch": 0.8696165440478951, "grad_norm": 2.3945040702819824, "learning_rate": 3.945235641337956e-05, "loss": 7.5121, "step": 20626 }, { "epoch": 0.8696587052300947, "grad_norm": 1.2246811389923096, "learning_rate": 3.945157562697149e-05, "loss": 6.6841, "step": 20627 }, { "epoch": 0.8697008664122942, "grad_norm": 1.148310899734497, "learning_rate": 3.945079484056341e-05, "loss": 6.8359, "step": 20628 }, { "epoch": 0.8697430275944937, "grad_norm": 1.8393915891647339, "learning_rate": 3.945001405415535e-05, "loss": 6.5678, "step": 20629 }, { "epoch": 0.8697851887766933, "grad_norm": 1.3342114686965942, "learning_rate": 3.9449233267747276e-05, "loss": 6.4973, "step": 20630 }, { "epoch": 0.8698273499588929, "grad_norm": 0.9848324060440063, "learning_rate": 3.944845248133921e-05, "loss": 6.7225, "step": 20631 }, { "epoch": 0.8698695111410923, "grad_norm": 1.2845630645751953, "learning_rate": 3.944767169493114e-05, "loss": 6.6498, "step": 20632 }, { "epoch": 0.8699116723232919, "grad_norm": 1.036268711090088, "learning_rate": 3.944689090852306e-05, "loss": 6.563, "step": 20633 }, { "epoch": 0.8699538335054915, "grad_norm": 1.0846264362335205, "learning_rate": 3.9446110122115e-05, "loss": 6.6493, "step": 20634 }, { "epoch": 0.8699959946876911, "grad_norm": 1.3363585472106934, "learning_rate": 3.9445329335706925e-05, "loss": 6.4805, "step": 20635 }, { "epoch": 0.8700381558698906, "grad_norm": 1.0583584308624268, "learning_rate": 3.944454854929886e-05, "loss": 6.6426, "step": 20636 }, { "epoch": 0.8700803170520901, "grad_norm": 1.1208710670471191, "learning_rate": 3.944376776289079e-05, "loss": 6.7219, "step": 20637 }, { "epoch": 0.8701224782342897, "grad_norm": 2.465677261352539, "learning_rate": 3.944298697648271e-05, "loss": 6.6279, "step": 20638 }, { "epoch": 0.8701646394164892, "grad_norm": 2.826728582382202, "learning_rate": 3.9442206190074644e-05, "loss": 7.6803, "step": 20639 }, { "epoch": 0.8702068005986888, "grad_norm": 3.630140781402588, "learning_rate": 3.9441425403666575e-05, "loss": 8.0938, "step": 20640 }, { "epoch": 0.8702489617808883, "grad_norm": 2.2386767864227295, "learning_rate": 3.9440644617258506e-05, "loss": 6.9427, "step": 20641 }, { "epoch": 0.8702911229630879, "grad_norm": 1.6315263509750366, "learning_rate": 3.943986383085043e-05, "loss": 7.0702, "step": 20642 }, { "epoch": 0.8703332841452874, "grad_norm": 0.9845036268234253, "learning_rate": 3.943908304444237e-05, "loss": 6.5917, "step": 20643 }, { "epoch": 0.870375445327487, "grad_norm": 1.5781209468841553, "learning_rate": 3.943830225803429e-05, "loss": 6.741, "step": 20644 }, { "epoch": 0.8704176065096866, "grad_norm": 2.095754384994507, "learning_rate": 3.9437521471626224e-05, "loss": 6.8922, "step": 20645 }, { "epoch": 0.870459767691886, "grad_norm": 5.52608060836792, "learning_rate": 3.9436740685218156e-05, "loss": 10.9627, "step": 20646 }, { "epoch": 0.8705019288740856, "grad_norm": 2.189943552017212, "learning_rate": 3.943595989881008e-05, "loss": 7.0129, "step": 20647 }, { "epoch": 0.8705440900562852, "grad_norm": 1.5274982452392578, "learning_rate": 3.943517911240202e-05, "loss": 6.7057, "step": 20648 }, { "epoch": 0.8705862512384848, "grad_norm": 1.6089032888412476, "learning_rate": 3.943439832599394e-05, "loss": 7.064, "step": 20649 }, { "epoch": 0.8706284124206842, "grad_norm": 3.5095229148864746, "learning_rate": 3.9433617539585874e-05, "loss": 6.7941, "step": 20650 }, { "epoch": 0.8706705736028838, "grad_norm": 1.4418747425079346, "learning_rate": 3.9432836753177805e-05, "loss": 6.4539, "step": 20651 }, { "epoch": 0.8707127347850834, "grad_norm": 1.5374356508255005, "learning_rate": 3.943205596676973e-05, "loss": 6.5993, "step": 20652 }, { "epoch": 0.870754895967283, "grad_norm": 1.9223732948303223, "learning_rate": 3.943127518036166e-05, "loss": 7.0238, "step": 20653 }, { "epoch": 0.8707970571494825, "grad_norm": 1.5271809101104736, "learning_rate": 3.943049439395359e-05, "loss": 6.4231, "step": 20654 }, { "epoch": 0.870839218331682, "grad_norm": 3.0299458503723145, "learning_rate": 3.9429713607545523e-05, "loss": 7.1284, "step": 20655 }, { "epoch": 0.8708813795138816, "grad_norm": 2.3180673122406006, "learning_rate": 3.9428932821137455e-05, "loss": 6.6322, "step": 20656 }, { "epoch": 0.8709235406960811, "grad_norm": 3.809298515319824, "learning_rate": 3.942815203472938e-05, "loss": 7.6646, "step": 20657 }, { "epoch": 0.8709657018782807, "grad_norm": 3.2725627422332764, "learning_rate": 3.942737124832131e-05, "loss": 7.7799, "step": 20658 }, { "epoch": 0.8710078630604802, "grad_norm": 1.0545016527175903, "learning_rate": 3.942659046191324e-05, "loss": 6.5116, "step": 20659 }, { "epoch": 0.8710500242426797, "grad_norm": 1.292768955230713, "learning_rate": 3.942580967550517e-05, "loss": 6.977, "step": 20660 }, { "epoch": 0.8710921854248793, "grad_norm": 3.221912384033203, "learning_rate": 3.94250288890971e-05, "loss": 7.8892, "step": 20661 }, { "epoch": 0.8711343466070789, "grad_norm": 1.9776520729064941, "learning_rate": 3.942424810268903e-05, "loss": 6.5353, "step": 20662 }, { "epoch": 0.8711765077892785, "grad_norm": 2.025050401687622, "learning_rate": 3.942346731628096e-05, "loss": 7.3351, "step": 20663 }, { "epoch": 0.8712186689714779, "grad_norm": 1.4903976917266846, "learning_rate": 3.9422686529872884e-05, "loss": 6.9554, "step": 20664 }, { "epoch": 0.8712608301536775, "grad_norm": 1.0832935571670532, "learning_rate": 3.942190574346482e-05, "loss": 7.1202, "step": 20665 }, { "epoch": 0.8713029913358771, "grad_norm": 1.1315596103668213, "learning_rate": 3.942112495705675e-05, "loss": 6.9793, "step": 20666 }, { "epoch": 0.8713451525180766, "grad_norm": 2.111206531524658, "learning_rate": 3.942034417064868e-05, "loss": 7.325, "step": 20667 }, { "epoch": 0.8713873137002761, "grad_norm": 1.923691987991333, "learning_rate": 3.941956338424061e-05, "loss": 6.5327, "step": 20668 }, { "epoch": 0.8714294748824757, "grad_norm": 3.1068427562713623, "learning_rate": 3.941878259783254e-05, "loss": 7.4666, "step": 20669 }, { "epoch": 0.8714716360646753, "grad_norm": 2.020796775817871, "learning_rate": 3.941800181142447e-05, "loss": 6.4771, "step": 20670 }, { "epoch": 0.8715137972468748, "grad_norm": 2.3637197017669678, "learning_rate": 3.9417221025016397e-05, "loss": 6.5084, "step": 20671 }, { "epoch": 0.8715559584290744, "grad_norm": 1.318814992904663, "learning_rate": 3.941644023860833e-05, "loss": 6.8136, "step": 20672 }, { "epoch": 0.8715981196112739, "grad_norm": 1.2750434875488281, "learning_rate": 3.941565945220026e-05, "loss": 6.629, "step": 20673 }, { "epoch": 0.8716402807934734, "grad_norm": 1.4467459917068481, "learning_rate": 3.941487866579219e-05, "loss": 6.6927, "step": 20674 }, { "epoch": 0.871682441975673, "grad_norm": 1.426513433456421, "learning_rate": 3.9414097879384115e-05, "loss": 7.0439, "step": 20675 }, { "epoch": 0.8717246031578726, "grad_norm": 3.252359390258789, "learning_rate": 3.9413317092976046e-05, "loss": 6.9822, "step": 20676 }, { "epoch": 0.871766764340072, "grad_norm": 1.4483137130737305, "learning_rate": 3.941253630656798e-05, "loss": 6.5448, "step": 20677 }, { "epoch": 0.8718089255222716, "grad_norm": 2.8108413219451904, "learning_rate": 3.94117555201599e-05, "loss": 7.7159, "step": 20678 }, { "epoch": 0.8718510867044712, "grad_norm": 2.1980814933776855, "learning_rate": 3.941097473375184e-05, "loss": 7.2028, "step": 20679 }, { "epoch": 0.8718932478866708, "grad_norm": 1.0958099365234375, "learning_rate": 3.9410193947343764e-05, "loss": 7.0489, "step": 20680 }, { "epoch": 0.8719354090688702, "grad_norm": 1.4821170568466187, "learning_rate": 3.9409413160935696e-05, "loss": 7.4069, "step": 20681 }, { "epoch": 0.8719775702510698, "grad_norm": 1.6085686683654785, "learning_rate": 3.940863237452763e-05, "loss": 6.7489, "step": 20682 }, { "epoch": 0.8720197314332694, "grad_norm": 1.4190456867218018, "learning_rate": 3.940785158811955e-05, "loss": 6.6954, "step": 20683 }, { "epoch": 0.872061892615469, "grad_norm": 2.061922073364258, "learning_rate": 3.940707080171149e-05, "loss": 6.4358, "step": 20684 }, { "epoch": 0.8721040537976685, "grad_norm": 1.6989593505859375, "learning_rate": 3.9406290015303414e-05, "loss": 6.5916, "step": 20685 }, { "epoch": 0.872146214979868, "grad_norm": 4.163369178771973, "learning_rate": 3.9405509228895345e-05, "loss": 7.5785, "step": 20686 }, { "epoch": 0.8721883761620676, "grad_norm": 1.4772698879241943, "learning_rate": 3.9404728442487276e-05, "loss": 6.5224, "step": 20687 }, { "epoch": 0.8722305373442671, "grad_norm": 1.4347606897354126, "learning_rate": 3.940394765607921e-05, "loss": 6.9236, "step": 20688 }, { "epoch": 0.8722726985264667, "grad_norm": 3.3504951000213623, "learning_rate": 3.940316686967113e-05, "loss": 7.2072, "step": 20689 }, { "epoch": 0.8723148597086662, "grad_norm": 1.6483430862426758, "learning_rate": 3.940238608326306e-05, "loss": 6.9812, "step": 20690 }, { "epoch": 0.8723570208908658, "grad_norm": 1.3119542598724365, "learning_rate": 3.9401605296854995e-05, "loss": 7.2594, "step": 20691 }, { "epoch": 0.8723991820730653, "grad_norm": 1.3694952726364136, "learning_rate": 3.9400824510446926e-05, "loss": 6.6675, "step": 20692 }, { "epoch": 0.8724413432552649, "grad_norm": 3.4593377113342285, "learning_rate": 3.940004372403886e-05, "loss": 9.6418, "step": 20693 }, { "epoch": 0.8724835044374645, "grad_norm": 1.769098162651062, "learning_rate": 3.939926293763078e-05, "loss": 6.3989, "step": 20694 }, { "epoch": 0.8725256656196639, "grad_norm": 2.4227147102355957, "learning_rate": 3.939848215122271e-05, "loss": 6.9837, "step": 20695 }, { "epoch": 0.8725678268018635, "grad_norm": 1.0650010108947754, "learning_rate": 3.9397701364814644e-05, "loss": 6.542, "step": 20696 }, { "epoch": 0.8726099879840631, "grad_norm": 1.1747066974639893, "learning_rate": 3.939692057840657e-05, "loss": 6.6893, "step": 20697 }, { "epoch": 0.8726521491662627, "grad_norm": 3.7971787452697754, "learning_rate": 3.939613979199851e-05, "loss": 7.5777, "step": 20698 }, { "epoch": 0.8726943103484621, "grad_norm": 3.4384195804595947, "learning_rate": 3.939535900559043e-05, "loss": 7.5238, "step": 20699 }, { "epoch": 0.8727364715306617, "grad_norm": 1.074979305267334, "learning_rate": 3.939457821918236e-05, "loss": 6.5824, "step": 20700 }, { "epoch": 0.8727786327128613, "grad_norm": 1.375824213027954, "learning_rate": 3.9393797432774294e-05, "loss": 6.4503, "step": 20701 }, { "epoch": 0.8728207938950608, "grad_norm": 1.6498994827270508, "learning_rate": 3.939301664636622e-05, "loss": 6.9602, "step": 20702 }, { "epoch": 0.8728629550772604, "grad_norm": 3.5643908977508545, "learning_rate": 3.9392235859958156e-05, "loss": 7.6299, "step": 20703 }, { "epoch": 0.8729051162594599, "grad_norm": 1.70929753780365, "learning_rate": 3.939145507355008e-05, "loss": 7.0316, "step": 20704 }, { "epoch": 0.8729472774416595, "grad_norm": 1.8752154111862183, "learning_rate": 3.939067428714201e-05, "loss": 6.4505, "step": 20705 }, { "epoch": 0.872989438623859, "grad_norm": 2.0645675659179688, "learning_rate": 3.938989350073394e-05, "loss": 6.5193, "step": 20706 }, { "epoch": 0.8730315998060586, "grad_norm": 1.7135778665542603, "learning_rate": 3.938911271432587e-05, "loss": 6.4372, "step": 20707 }, { "epoch": 0.8730737609882581, "grad_norm": 3.1529712677001953, "learning_rate": 3.93883319279178e-05, "loss": 7.4421, "step": 20708 }, { "epoch": 0.8731159221704576, "grad_norm": 3.177058458328247, "learning_rate": 3.938755114150973e-05, "loss": 7.9248, "step": 20709 }, { "epoch": 0.8731580833526572, "grad_norm": 2.662728786468506, "learning_rate": 3.938677035510166e-05, "loss": 6.8466, "step": 20710 }, { "epoch": 0.8732002445348568, "grad_norm": 1.4610204696655273, "learning_rate": 3.9385989568693586e-05, "loss": 6.6495, "step": 20711 }, { "epoch": 0.8732424057170564, "grad_norm": 1.0810929536819458, "learning_rate": 3.9385208782285524e-05, "loss": 6.4482, "step": 20712 }, { "epoch": 0.8732845668992558, "grad_norm": 1.5435502529144287, "learning_rate": 3.938442799587745e-05, "loss": 6.6754, "step": 20713 }, { "epoch": 0.8733267280814554, "grad_norm": 0.8108063340187073, "learning_rate": 3.938364720946938e-05, "loss": 6.6325, "step": 20714 }, { "epoch": 0.873368889263655, "grad_norm": 1.7011271715164185, "learning_rate": 3.938286642306131e-05, "loss": 6.439, "step": 20715 }, { "epoch": 0.8734110504458545, "grad_norm": 1.5851927995681763, "learning_rate": 3.9382085636653235e-05, "loss": 6.9351, "step": 20716 }, { "epoch": 0.873453211628054, "grad_norm": 1.2949978113174438, "learning_rate": 3.9381304850245173e-05, "loss": 7.302, "step": 20717 }, { "epoch": 0.8734953728102536, "grad_norm": 1.9386241436004639, "learning_rate": 3.93805240638371e-05, "loss": 7.4931, "step": 20718 }, { "epoch": 0.8735375339924532, "grad_norm": 0.8049008846282959, "learning_rate": 3.937974327742903e-05, "loss": 6.7812, "step": 20719 }, { "epoch": 0.8735796951746527, "grad_norm": 0.9134001135826111, "learning_rate": 3.937896249102096e-05, "loss": 6.6194, "step": 20720 }, { "epoch": 0.8736218563568523, "grad_norm": 4.1201887130737305, "learning_rate": 3.9378181704612885e-05, "loss": 9.1044, "step": 20721 }, { "epoch": 0.8736640175390518, "grad_norm": 3.896397113800049, "learning_rate": 3.9377400918204816e-05, "loss": 7.4713, "step": 20722 }, { "epoch": 0.8737061787212513, "grad_norm": 1.463026523590088, "learning_rate": 3.937662013179675e-05, "loss": 6.9368, "step": 20723 }, { "epoch": 0.8737483399034509, "grad_norm": 1.5937609672546387, "learning_rate": 3.937583934538868e-05, "loss": 6.8291, "step": 20724 }, { "epoch": 0.8737905010856505, "grad_norm": 1.7629655599594116, "learning_rate": 3.937505855898061e-05, "loss": 6.6203, "step": 20725 }, { "epoch": 0.87383266226785, "grad_norm": 3.1565017700195312, "learning_rate": 3.9374277772572534e-05, "loss": 7.6008, "step": 20726 }, { "epoch": 0.8738748234500495, "grad_norm": 1.3275585174560547, "learning_rate": 3.9373496986164466e-05, "loss": 6.3716, "step": 20727 }, { "epoch": 0.8739169846322491, "grad_norm": 1.971954107284546, "learning_rate": 3.93727161997564e-05, "loss": 7.1622, "step": 20728 }, { "epoch": 0.8739591458144487, "grad_norm": 1.1315913200378418, "learning_rate": 3.937193541334833e-05, "loss": 6.7541, "step": 20729 }, { "epoch": 0.8740013069966481, "grad_norm": 2.674649238586426, "learning_rate": 3.937115462694025e-05, "loss": 6.986, "step": 20730 }, { "epoch": 0.8740434681788477, "grad_norm": 3.9320967197418213, "learning_rate": 3.937037384053219e-05, "loss": 8.2055, "step": 20731 }, { "epoch": 0.8740856293610473, "grad_norm": 1.4896113872528076, "learning_rate": 3.9369593054124115e-05, "loss": 6.4929, "step": 20732 }, { "epoch": 0.8741277905432469, "grad_norm": 3.5374152660369873, "learning_rate": 3.936881226771604e-05, "loss": 7.6194, "step": 20733 }, { "epoch": 0.8741699517254464, "grad_norm": 2.871419668197632, "learning_rate": 3.936803148130798e-05, "loss": 8.1242, "step": 20734 }, { "epoch": 0.8742121129076459, "grad_norm": 2.080815553665161, "learning_rate": 3.93672506948999e-05, "loss": 6.7198, "step": 20735 }, { "epoch": 0.8742542740898455, "grad_norm": 1.3571531772613525, "learning_rate": 3.936646990849184e-05, "loss": 7.2692, "step": 20736 }, { "epoch": 0.874296435272045, "grad_norm": 3.0627546310424805, "learning_rate": 3.9365689122083765e-05, "loss": 8.8707, "step": 20737 }, { "epoch": 0.8743385964542446, "grad_norm": 1.782623529434204, "learning_rate": 3.9364908335675696e-05, "loss": 7.0878, "step": 20738 }, { "epoch": 0.8743807576364441, "grad_norm": 2.12201189994812, "learning_rate": 3.936412754926763e-05, "loss": 6.4742, "step": 20739 }, { "epoch": 0.8744229188186436, "grad_norm": 1.4986132383346558, "learning_rate": 3.936334676285955e-05, "loss": 6.7384, "step": 20740 }, { "epoch": 0.8744650800008432, "grad_norm": 1.9882029294967651, "learning_rate": 3.936256597645148e-05, "loss": 6.8203, "step": 20741 }, { "epoch": 0.8745072411830428, "grad_norm": 1.187996506690979, "learning_rate": 3.9361785190043414e-05, "loss": 6.8968, "step": 20742 }, { "epoch": 0.8745494023652424, "grad_norm": 2.519944429397583, "learning_rate": 3.9361004403635346e-05, "loss": 7.5855, "step": 20743 }, { "epoch": 0.8745915635474418, "grad_norm": 0.9539662599563599, "learning_rate": 3.936022361722727e-05, "loss": 6.6498, "step": 20744 }, { "epoch": 0.8746337247296414, "grad_norm": 1.3812381029129028, "learning_rate": 3.93594428308192e-05, "loss": 6.3941, "step": 20745 }, { "epoch": 0.874675885911841, "grad_norm": 2.5012307167053223, "learning_rate": 3.935866204441113e-05, "loss": 7.0851, "step": 20746 }, { "epoch": 0.8747180470940406, "grad_norm": 0.8603503704071045, "learning_rate": 3.935788125800306e-05, "loss": 6.5031, "step": 20747 }, { "epoch": 0.87476020827624, "grad_norm": 1.1263659000396729, "learning_rate": 3.9357100471594995e-05, "loss": 6.9549, "step": 20748 }, { "epoch": 0.8748023694584396, "grad_norm": 1.0209834575653076, "learning_rate": 3.935631968518692e-05, "loss": 6.7297, "step": 20749 }, { "epoch": 0.8748445306406392, "grad_norm": 3.01065993309021, "learning_rate": 3.935553889877886e-05, "loss": 7.2391, "step": 20750 }, { "epoch": 0.8748866918228387, "grad_norm": 1.1093538999557495, "learning_rate": 3.935475811237078e-05, "loss": 6.5472, "step": 20751 }, { "epoch": 0.8749288530050383, "grad_norm": 2.383653402328491, "learning_rate": 3.9353977325962707e-05, "loss": 6.712, "step": 20752 }, { "epoch": 0.8749710141872378, "grad_norm": 4.455785751342773, "learning_rate": 3.9353196539554645e-05, "loss": 7.3448, "step": 20753 }, { "epoch": 0.8750131753694373, "grad_norm": 2.0670952796936035, "learning_rate": 3.935241575314657e-05, "loss": 6.8906, "step": 20754 }, { "epoch": 0.8750553365516369, "grad_norm": 1.3462008237838745, "learning_rate": 3.93516349667385e-05, "loss": 6.8824, "step": 20755 }, { "epoch": 0.8750974977338365, "grad_norm": 1.051581859588623, "learning_rate": 3.935085418033043e-05, "loss": 6.4784, "step": 20756 }, { "epoch": 0.875139658916036, "grad_norm": 1.5946904420852661, "learning_rate": 3.935007339392236e-05, "loss": 6.9927, "step": 20757 }, { "epoch": 0.8751818200982355, "grad_norm": 3.8528032302856445, "learning_rate": 3.934929260751429e-05, "loss": 7.5288, "step": 20758 }, { "epoch": 0.8752239812804351, "grad_norm": 2.046435594558716, "learning_rate": 3.934851182110622e-05, "loss": 6.5533, "step": 20759 }, { "epoch": 0.8752661424626347, "grad_norm": 2.184596538543701, "learning_rate": 3.934773103469815e-05, "loss": 7.195, "step": 20760 }, { "epoch": 0.8753083036448343, "grad_norm": 1.7793045043945312, "learning_rate": 3.934695024829008e-05, "loss": 6.9984, "step": 20761 }, { "epoch": 0.8753504648270337, "grad_norm": 2.452965497970581, "learning_rate": 3.934616946188201e-05, "loss": 7.394, "step": 20762 }, { "epoch": 0.8753926260092333, "grad_norm": 3.466292381286621, "learning_rate": 3.934538867547394e-05, "loss": 7.7641, "step": 20763 }, { "epoch": 0.8754347871914329, "grad_norm": 3.255638599395752, "learning_rate": 3.934460788906587e-05, "loss": 7.3455, "step": 20764 }, { "epoch": 0.8754769483736324, "grad_norm": 2.82786226272583, "learning_rate": 3.93438271026578e-05, "loss": 6.5251, "step": 20765 }, { "epoch": 0.8755191095558319, "grad_norm": 2.9130756855010986, "learning_rate": 3.9343046316249724e-05, "loss": 6.3174, "step": 20766 }, { "epoch": 0.8755612707380315, "grad_norm": 1.654036283493042, "learning_rate": 3.934226552984166e-05, "loss": 6.8592, "step": 20767 }, { "epoch": 0.875603431920231, "grad_norm": 1.4726392030715942, "learning_rate": 3.9341484743433586e-05, "loss": 6.7061, "step": 20768 }, { "epoch": 0.8756455931024306, "grad_norm": 1.484750747680664, "learning_rate": 3.934070395702552e-05, "loss": 6.608, "step": 20769 }, { "epoch": 0.8756877542846302, "grad_norm": 1.2261011600494385, "learning_rate": 3.933992317061745e-05, "loss": 6.7251, "step": 20770 }, { "epoch": 0.8757299154668297, "grad_norm": 2.951469898223877, "learning_rate": 3.9339142384209373e-05, "loss": 6.5586, "step": 20771 }, { "epoch": 0.8757720766490292, "grad_norm": 3.2293782234191895, "learning_rate": 3.933836159780131e-05, "loss": 8.0697, "step": 20772 }, { "epoch": 0.8758142378312288, "grad_norm": 2.257840156555176, "learning_rate": 3.9337580811393236e-05, "loss": 7.7221, "step": 20773 }, { "epoch": 0.8758563990134284, "grad_norm": 2.9147610664367676, "learning_rate": 3.933680002498517e-05, "loss": 7.4376, "step": 20774 }, { "epoch": 0.8758985601956278, "grad_norm": 1.9435852766036987, "learning_rate": 3.93360192385771e-05, "loss": 6.6203, "step": 20775 }, { "epoch": 0.8759407213778274, "grad_norm": 1.881324052810669, "learning_rate": 3.933523845216903e-05, "loss": 7.2046, "step": 20776 }, { "epoch": 0.875982882560027, "grad_norm": 2.0340664386749268, "learning_rate": 3.9334457665760954e-05, "loss": 6.8385, "step": 20777 }, { "epoch": 0.8760250437422266, "grad_norm": 1.0206553936004639, "learning_rate": 3.9333676879352885e-05, "loss": 6.5569, "step": 20778 }, { "epoch": 0.8760672049244261, "grad_norm": 2.361992597579956, "learning_rate": 3.933289609294482e-05, "loss": 6.4946, "step": 20779 }, { "epoch": 0.8761093661066256, "grad_norm": 2.857811450958252, "learning_rate": 3.933211530653674e-05, "loss": 8.2284, "step": 20780 }, { "epoch": 0.8761515272888252, "grad_norm": 3.150716543197632, "learning_rate": 3.933133452012868e-05, "loss": 7.5204, "step": 20781 }, { "epoch": 0.8761936884710247, "grad_norm": 1.775140404701233, "learning_rate": 3.9330553733720604e-05, "loss": 6.7279, "step": 20782 }, { "epoch": 0.8762358496532243, "grad_norm": 1.454160213470459, "learning_rate": 3.9329772947312535e-05, "loss": 7.0359, "step": 20783 }, { "epoch": 0.8762780108354238, "grad_norm": 1.5012296438217163, "learning_rate": 3.9328992160904466e-05, "loss": 6.5157, "step": 20784 }, { "epoch": 0.8763201720176234, "grad_norm": 2.2388017177581787, "learning_rate": 3.932821137449639e-05, "loss": 6.4682, "step": 20785 }, { "epoch": 0.8763623331998229, "grad_norm": 1.500243067741394, "learning_rate": 3.932743058808833e-05, "loss": 6.4054, "step": 20786 }, { "epoch": 0.8764044943820225, "grad_norm": 1.4685837030410767, "learning_rate": 3.932664980168025e-05, "loss": 6.5156, "step": 20787 }, { "epoch": 0.876446655564222, "grad_norm": 4.1687397956848145, "learning_rate": 3.9325869015272185e-05, "loss": 7.6893, "step": 20788 }, { "epoch": 0.8764888167464215, "grad_norm": 1.331605076789856, "learning_rate": 3.9325088228864116e-05, "loss": 6.6974, "step": 20789 }, { "epoch": 0.8765309779286211, "grad_norm": 3.237863540649414, "learning_rate": 3.932430744245604e-05, "loss": 8.2254, "step": 20790 }, { "epoch": 0.8765731391108207, "grad_norm": 4.594734191894531, "learning_rate": 3.932352665604797e-05, "loss": 8.9333, "step": 20791 }, { "epoch": 0.8766153002930203, "grad_norm": 1.1494990587234497, "learning_rate": 3.93227458696399e-05, "loss": 6.3499, "step": 20792 }, { "epoch": 0.8766574614752197, "grad_norm": 1.9275665283203125, "learning_rate": 3.9321965083231834e-05, "loss": 6.7954, "step": 20793 }, { "epoch": 0.8766996226574193, "grad_norm": 1.3821499347686768, "learning_rate": 3.9321184296823765e-05, "loss": 6.4975, "step": 20794 }, { "epoch": 0.8767417838396189, "grad_norm": 5.911048412322998, "learning_rate": 3.932040351041569e-05, "loss": 10.2271, "step": 20795 }, { "epoch": 0.8767839450218184, "grad_norm": 2.9710092544555664, "learning_rate": 3.931962272400762e-05, "loss": 7.166, "step": 20796 }, { "epoch": 0.8768261062040179, "grad_norm": 1.4097110033035278, "learning_rate": 3.931884193759955e-05, "loss": 6.6967, "step": 20797 }, { "epoch": 0.8768682673862175, "grad_norm": 6.331995964050293, "learning_rate": 3.9318061151191484e-05, "loss": 11.7328, "step": 20798 }, { "epoch": 0.876910428568417, "grad_norm": 1.8235259056091309, "learning_rate": 3.931728036478341e-05, "loss": 6.5517, "step": 20799 }, { "epoch": 0.8769525897506166, "grad_norm": 1.55874764919281, "learning_rate": 3.9316499578375346e-05, "loss": 6.9781, "step": 20800 }, { "epoch": 0.8769947509328162, "grad_norm": 4.659041404724121, "learning_rate": 3.931571879196727e-05, "loss": 7.0229, "step": 20801 }, { "epoch": 0.8770369121150157, "grad_norm": 1.2939187288284302, "learning_rate": 3.93149380055592e-05, "loss": 6.6191, "step": 20802 }, { "epoch": 0.8770790732972152, "grad_norm": 1.7801748514175415, "learning_rate": 3.931415721915113e-05, "loss": 7.1064, "step": 20803 }, { "epoch": 0.8771212344794148, "grad_norm": 2.661180257797241, "learning_rate": 3.931337643274306e-05, "loss": 7.4817, "step": 20804 }, { "epoch": 0.8771633956616144, "grad_norm": 1.3029329776763916, "learning_rate": 3.9312595646334996e-05, "loss": 6.5291, "step": 20805 }, { "epoch": 0.8772055568438138, "grad_norm": 1.4020270109176636, "learning_rate": 3.931181485992692e-05, "loss": 6.5806, "step": 20806 }, { "epoch": 0.8772477180260134, "grad_norm": 2.030280351638794, "learning_rate": 3.931103407351885e-05, "loss": 7.0463, "step": 20807 }, { "epoch": 0.877289879208213, "grad_norm": 2.5906126499176025, "learning_rate": 3.931025328711078e-05, "loss": 6.5572, "step": 20808 }, { "epoch": 0.8773320403904126, "grad_norm": 2.2396883964538574, "learning_rate": 3.930947250070271e-05, "loss": 6.6805, "step": 20809 }, { "epoch": 0.8773742015726121, "grad_norm": 2.13284969329834, "learning_rate": 3.930869171429464e-05, "loss": 6.5388, "step": 20810 }, { "epoch": 0.8774163627548116, "grad_norm": 3.160876989364624, "learning_rate": 3.930791092788657e-05, "loss": 6.5303, "step": 20811 }, { "epoch": 0.8774585239370112, "grad_norm": 1.686785340309143, "learning_rate": 3.93071301414785e-05, "loss": 7.0376, "step": 20812 }, { "epoch": 0.8775006851192108, "grad_norm": 2.1558024883270264, "learning_rate": 3.9306349355070425e-05, "loss": 6.52, "step": 20813 }, { "epoch": 0.8775428463014103, "grad_norm": 1.6621966361999512, "learning_rate": 3.9305568568662357e-05, "loss": 7.4323, "step": 20814 }, { "epoch": 0.8775850074836098, "grad_norm": 1.4097493886947632, "learning_rate": 3.930478778225429e-05, "loss": 6.5418, "step": 20815 }, { "epoch": 0.8776271686658094, "grad_norm": 2.8659660816192627, "learning_rate": 3.930400699584622e-05, "loss": 6.2751, "step": 20816 }, { "epoch": 0.8776693298480089, "grad_norm": 3.1516289710998535, "learning_rate": 3.930322620943815e-05, "loss": 7.4719, "step": 20817 }, { "epoch": 0.8777114910302085, "grad_norm": 1.3232423067092896, "learning_rate": 3.9302445423030075e-05, "loss": 6.5291, "step": 20818 }, { "epoch": 0.8777536522124081, "grad_norm": 1.2862563133239746, "learning_rate": 3.930166463662201e-05, "loss": 6.6969, "step": 20819 }, { "epoch": 0.8777958133946075, "grad_norm": 1.4143186807632446, "learning_rate": 3.930088385021394e-05, "loss": 7.5032, "step": 20820 }, { "epoch": 0.8778379745768071, "grad_norm": 2.496408462524414, "learning_rate": 3.930010306380587e-05, "loss": 7.2249, "step": 20821 }, { "epoch": 0.8778801357590067, "grad_norm": 2.1250202655792236, "learning_rate": 3.92993222773978e-05, "loss": 6.7905, "step": 20822 }, { "epoch": 0.8779222969412063, "grad_norm": 1.9680346250534058, "learning_rate": 3.9298541490989724e-05, "loss": 6.6845, "step": 20823 }, { "epoch": 0.8779644581234057, "grad_norm": 1.1523264646530151, "learning_rate": 3.9297760704581656e-05, "loss": 6.7077, "step": 20824 }, { "epoch": 0.8780066193056053, "grad_norm": 1.5066462755203247, "learning_rate": 3.929697991817359e-05, "loss": 6.3813, "step": 20825 }, { "epoch": 0.8780487804878049, "grad_norm": 1.5723772048950195, "learning_rate": 3.929619913176552e-05, "loss": 7.2708, "step": 20826 }, { "epoch": 0.8780909416700045, "grad_norm": 2.0232882499694824, "learning_rate": 3.929541834535744e-05, "loss": 6.8663, "step": 20827 }, { "epoch": 0.878133102852204, "grad_norm": 3.1848762035369873, "learning_rate": 3.9294637558949374e-05, "loss": 7.2916, "step": 20828 }, { "epoch": 0.8781752640344035, "grad_norm": 0.7347822785377502, "learning_rate": 3.9293856772541305e-05, "loss": 6.4888, "step": 20829 }, { "epoch": 0.8782174252166031, "grad_norm": 3.9870100021362305, "learning_rate": 3.9293075986133236e-05, "loss": 7.4178, "step": 20830 }, { "epoch": 0.8782595863988026, "grad_norm": 1.3549984693527222, "learning_rate": 3.929229519972517e-05, "loss": 6.4619, "step": 20831 }, { "epoch": 0.8783017475810022, "grad_norm": 2.663505792617798, "learning_rate": 3.929151441331709e-05, "loss": 6.7042, "step": 20832 }, { "epoch": 0.8783439087632017, "grad_norm": 1.5244438648223877, "learning_rate": 3.9290733626909023e-05, "loss": 6.8379, "step": 20833 }, { "epoch": 0.8783860699454012, "grad_norm": 0.9173007607460022, "learning_rate": 3.9289952840500955e-05, "loss": 6.8775, "step": 20834 }, { "epoch": 0.8784282311276008, "grad_norm": 1.059011697769165, "learning_rate": 3.928917205409288e-05, "loss": 6.7796, "step": 20835 }, { "epoch": 0.8784703923098004, "grad_norm": 1.3845207691192627, "learning_rate": 3.928839126768482e-05, "loss": 6.701, "step": 20836 }, { "epoch": 0.878512553492, "grad_norm": 1.6869958639144897, "learning_rate": 3.928761048127674e-05, "loss": 6.8415, "step": 20837 }, { "epoch": 0.8785547146741994, "grad_norm": 1.5190061330795288, "learning_rate": 3.928682969486867e-05, "loss": 6.7773, "step": 20838 }, { "epoch": 0.878596875856399, "grad_norm": 0.7837839126586914, "learning_rate": 3.9286048908460604e-05, "loss": 6.7603, "step": 20839 }, { "epoch": 0.8786390370385986, "grad_norm": 1.234627366065979, "learning_rate": 3.928526812205253e-05, "loss": 6.4886, "step": 20840 }, { "epoch": 0.8786811982207982, "grad_norm": 1.353551983833313, "learning_rate": 3.928448733564447e-05, "loss": 6.6316, "step": 20841 }, { "epoch": 0.8787233594029976, "grad_norm": 1.1457687616348267, "learning_rate": 3.928370654923639e-05, "loss": 6.6344, "step": 20842 }, { "epoch": 0.8787655205851972, "grad_norm": 0.9460242986679077, "learning_rate": 3.928292576282832e-05, "loss": 6.8783, "step": 20843 }, { "epoch": 0.8788076817673968, "grad_norm": 4.5964531898498535, "learning_rate": 3.9282144976420254e-05, "loss": 8.0224, "step": 20844 }, { "epoch": 0.8788498429495963, "grad_norm": 1.2441627979278564, "learning_rate": 3.9281364190012185e-05, "loss": 6.6125, "step": 20845 }, { "epoch": 0.8788920041317958, "grad_norm": 1.9203376770019531, "learning_rate": 3.928058340360411e-05, "loss": 6.9946, "step": 20846 }, { "epoch": 0.8789341653139954, "grad_norm": 3.0248491764068604, "learning_rate": 3.927980261719604e-05, "loss": 7.503, "step": 20847 }, { "epoch": 0.878976326496195, "grad_norm": 1.7735199928283691, "learning_rate": 3.927902183078797e-05, "loss": 6.7234, "step": 20848 }, { "epoch": 0.8790184876783945, "grad_norm": 5.311484336853027, "learning_rate": 3.9278241044379896e-05, "loss": 7.5063, "step": 20849 }, { "epoch": 0.8790606488605941, "grad_norm": 2.157613515853882, "learning_rate": 3.9277460257971835e-05, "loss": 7.0354, "step": 20850 }, { "epoch": 0.8791028100427936, "grad_norm": 3.0348119735717773, "learning_rate": 3.927667947156376e-05, "loss": 6.7223, "step": 20851 }, { "epoch": 0.8791449712249931, "grad_norm": 1.556320071220398, "learning_rate": 3.927589868515569e-05, "loss": 6.667, "step": 20852 }, { "epoch": 0.8791871324071927, "grad_norm": 2.382030963897705, "learning_rate": 3.927511789874762e-05, "loss": 6.9371, "step": 20853 }, { "epoch": 0.8792292935893923, "grad_norm": 1.35573410987854, "learning_rate": 3.9274337112339546e-05, "loss": 7.327, "step": 20854 }, { "epoch": 0.8792714547715917, "grad_norm": 3.0430827140808105, "learning_rate": 3.9273556325931484e-05, "loss": 6.4113, "step": 20855 }, { "epoch": 0.8793136159537913, "grad_norm": 1.692410945892334, "learning_rate": 3.927277553952341e-05, "loss": 6.5478, "step": 20856 }, { "epoch": 0.8793557771359909, "grad_norm": 2.0122642517089844, "learning_rate": 3.927199475311534e-05, "loss": 6.3474, "step": 20857 }, { "epoch": 0.8793979383181905, "grad_norm": 2.0546422004699707, "learning_rate": 3.927121396670727e-05, "loss": 7.0271, "step": 20858 }, { "epoch": 0.87944009950039, "grad_norm": 2.4856722354888916, "learning_rate": 3.9270433180299196e-05, "loss": 6.9374, "step": 20859 }, { "epoch": 0.8794822606825895, "grad_norm": 1.073453664779663, "learning_rate": 3.926965239389113e-05, "loss": 6.5417, "step": 20860 }, { "epoch": 0.8795244218647891, "grad_norm": 1.4007408618927002, "learning_rate": 3.926887160748306e-05, "loss": 6.8302, "step": 20861 }, { "epoch": 0.8795665830469886, "grad_norm": 2.1663315296173096, "learning_rate": 3.926809082107499e-05, "loss": 7.0824, "step": 20862 }, { "epoch": 0.8796087442291882, "grad_norm": 1.1109012365341187, "learning_rate": 3.926731003466692e-05, "loss": 6.519, "step": 20863 }, { "epoch": 0.8796509054113877, "grad_norm": 1.4637418985366821, "learning_rate": 3.926652924825885e-05, "loss": 6.9791, "step": 20864 }, { "epoch": 0.8796930665935873, "grad_norm": 1.7936604022979736, "learning_rate": 3.9265748461850776e-05, "loss": 7.2523, "step": 20865 }, { "epoch": 0.8797352277757868, "grad_norm": 1.698465347290039, "learning_rate": 3.926496767544271e-05, "loss": 6.7831, "step": 20866 }, { "epoch": 0.8797773889579864, "grad_norm": 1.2374125719070435, "learning_rate": 3.926418688903464e-05, "loss": 6.4985, "step": 20867 }, { "epoch": 0.879819550140186, "grad_norm": 3.4699697494506836, "learning_rate": 3.926340610262656e-05, "loss": 7.7849, "step": 20868 }, { "epoch": 0.8798617113223854, "grad_norm": 1.8303439617156982, "learning_rate": 3.92626253162185e-05, "loss": 6.6269, "step": 20869 }, { "epoch": 0.879903872504585, "grad_norm": 3.1413698196411133, "learning_rate": 3.9261844529810426e-05, "loss": 7.2886, "step": 20870 }, { "epoch": 0.8799460336867846, "grad_norm": 1.00278639793396, "learning_rate": 3.926106374340236e-05, "loss": 6.4123, "step": 20871 }, { "epoch": 0.8799881948689842, "grad_norm": 1.7535723447799683, "learning_rate": 3.926028295699429e-05, "loss": 6.5618, "step": 20872 }, { "epoch": 0.8800303560511836, "grad_norm": 1.9540218114852905, "learning_rate": 3.925950217058621e-05, "loss": 6.5214, "step": 20873 }, { "epoch": 0.8800725172333832, "grad_norm": 2.6822922229766846, "learning_rate": 3.925872138417815e-05, "loss": 6.2422, "step": 20874 }, { "epoch": 0.8801146784155828, "grad_norm": 1.0376890897750854, "learning_rate": 3.9257940597770075e-05, "loss": 6.7234, "step": 20875 }, { "epoch": 0.8801568395977823, "grad_norm": 2.206282377243042, "learning_rate": 3.925715981136201e-05, "loss": 8.3228, "step": 20876 }, { "epoch": 0.8801990007799819, "grad_norm": 1.9509925842285156, "learning_rate": 3.925637902495394e-05, "loss": 6.2524, "step": 20877 }, { "epoch": 0.8802411619621814, "grad_norm": 2.914280414581299, "learning_rate": 3.925559823854586e-05, "loss": 7.4652, "step": 20878 }, { "epoch": 0.880283323144381, "grad_norm": 5.246973037719727, "learning_rate": 3.9254817452137794e-05, "loss": 10.1044, "step": 20879 }, { "epoch": 0.8803254843265805, "grad_norm": 2.40665602684021, "learning_rate": 3.9254036665729725e-05, "loss": 6.9651, "step": 20880 }, { "epoch": 0.8803676455087801, "grad_norm": 2.9361109733581543, "learning_rate": 3.9253255879321656e-05, "loss": 8.3777, "step": 20881 }, { "epoch": 0.8804098066909796, "grad_norm": 2.1006011962890625, "learning_rate": 3.925247509291358e-05, "loss": 7.0636, "step": 20882 }, { "epoch": 0.8804519678731791, "grad_norm": 1.8104325532913208, "learning_rate": 3.925169430650552e-05, "loss": 6.7883, "step": 20883 }, { "epoch": 0.8804941290553787, "grad_norm": 2.159531354904175, "learning_rate": 3.925091352009744e-05, "loss": 7.5015, "step": 20884 }, { "epoch": 0.8805362902375783, "grad_norm": 1.9768325090408325, "learning_rate": 3.9250132733689374e-05, "loss": 7.1684, "step": 20885 }, { "epoch": 0.8805784514197779, "grad_norm": 1.6904261112213135, "learning_rate": 3.9249351947281306e-05, "loss": 6.789, "step": 20886 }, { "epoch": 0.8806206126019773, "grad_norm": 1.327824354171753, "learning_rate": 3.924857116087323e-05, "loss": 6.8762, "step": 20887 }, { "epoch": 0.8806627737841769, "grad_norm": 3.5119946002960205, "learning_rate": 3.924779037446517e-05, "loss": 7.749, "step": 20888 }, { "epoch": 0.8807049349663765, "grad_norm": 0.9823011755943298, "learning_rate": 3.924700958805709e-05, "loss": 6.6905, "step": 20889 }, { "epoch": 0.880747096148576, "grad_norm": 1.2709167003631592, "learning_rate": 3.9246228801649024e-05, "loss": 6.9944, "step": 20890 }, { "epoch": 0.8807892573307755, "grad_norm": 2.88602614402771, "learning_rate": 3.9245448015240955e-05, "loss": 6.3843, "step": 20891 }, { "epoch": 0.8808314185129751, "grad_norm": 1.1972742080688477, "learning_rate": 3.924466722883288e-05, "loss": 6.5644, "step": 20892 }, { "epoch": 0.8808735796951747, "grad_norm": 4.160239219665527, "learning_rate": 3.924388644242481e-05, "loss": 7.8032, "step": 20893 }, { "epoch": 0.8809157408773742, "grad_norm": 1.4056894779205322, "learning_rate": 3.924310565601674e-05, "loss": 7.0053, "step": 20894 }, { "epoch": 0.8809579020595738, "grad_norm": 2.352914571762085, "learning_rate": 3.9242324869608673e-05, "loss": 6.9668, "step": 20895 }, { "epoch": 0.8810000632417733, "grad_norm": 0.9985293745994568, "learning_rate": 3.9241544083200605e-05, "loss": 6.5901, "step": 20896 }, { "epoch": 0.8810422244239728, "grad_norm": 3.047959327697754, "learning_rate": 3.924076329679253e-05, "loss": 7.5437, "step": 20897 }, { "epoch": 0.8810843856061724, "grad_norm": 2.8307909965515137, "learning_rate": 3.923998251038446e-05, "loss": 7.2931, "step": 20898 }, { "epoch": 0.881126546788372, "grad_norm": 3.148224115371704, "learning_rate": 3.923920172397639e-05, "loss": 7.7231, "step": 20899 }, { "epoch": 0.8811687079705715, "grad_norm": 2.397432804107666, "learning_rate": 3.923842093756832e-05, "loss": 7.4748, "step": 20900 }, { "epoch": 0.881210869152771, "grad_norm": 0.9260403513908386, "learning_rate": 3.923764015116025e-05, "loss": 6.7193, "step": 20901 }, { "epoch": 0.8812530303349706, "grad_norm": 3.380755662918091, "learning_rate": 3.923685936475218e-05, "loss": 7.8533, "step": 20902 }, { "epoch": 0.8812951915171702, "grad_norm": 2.554992437362671, "learning_rate": 3.923607857834411e-05, "loss": 6.6686, "step": 20903 }, { "epoch": 0.8813373526993696, "grad_norm": 1.9033899307250977, "learning_rate": 3.9235297791936034e-05, "loss": 7.0551, "step": 20904 }, { "epoch": 0.8813795138815692, "grad_norm": 1.7173409461975098, "learning_rate": 3.923451700552797e-05, "loss": 6.7174, "step": 20905 }, { "epoch": 0.8814216750637688, "grad_norm": 1.7363882064819336, "learning_rate": 3.92337362191199e-05, "loss": 7.1027, "step": 20906 }, { "epoch": 0.8814638362459684, "grad_norm": 2.8650083541870117, "learning_rate": 3.923295543271183e-05, "loss": 6.4914, "step": 20907 }, { "epoch": 0.8815059974281679, "grad_norm": 1.8640443086624146, "learning_rate": 3.923217464630376e-05, "loss": 6.7294, "step": 20908 }, { "epoch": 0.8815481586103674, "grad_norm": 1.5245155096054077, "learning_rate": 3.923139385989569e-05, "loss": 6.9649, "step": 20909 }, { "epoch": 0.881590319792567, "grad_norm": 2.759556531906128, "learning_rate": 3.923061307348762e-05, "loss": 7.3843, "step": 20910 }, { "epoch": 0.8816324809747665, "grad_norm": 0.8615779876708984, "learning_rate": 3.9229832287079547e-05, "loss": 6.6792, "step": 20911 }, { "epoch": 0.8816746421569661, "grad_norm": 3.1961259841918945, "learning_rate": 3.922905150067148e-05, "loss": 7.1715, "step": 20912 }, { "epoch": 0.8817168033391656, "grad_norm": 1.2577019929885864, "learning_rate": 3.922827071426341e-05, "loss": 6.7515, "step": 20913 }, { "epoch": 0.8817589645213652, "grad_norm": 1.1225045919418335, "learning_rate": 3.922748992785534e-05, "loss": 6.4952, "step": 20914 }, { "epoch": 0.8818011257035647, "grad_norm": 1.6205065250396729, "learning_rate": 3.9226709141447265e-05, "loss": 7.1351, "step": 20915 }, { "epoch": 0.8818432868857643, "grad_norm": 1.536615014076233, "learning_rate": 3.9225928355039196e-05, "loss": 6.8161, "step": 20916 }, { "epoch": 0.8818854480679639, "grad_norm": 1.8283369541168213, "learning_rate": 3.922514756863113e-05, "loss": 6.9932, "step": 20917 }, { "epoch": 0.8819276092501633, "grad_norm": 2.1292290687561035, "learning_rate": 3.922436678222305e-05, "loss": 7.8527, "step": 20918 }, { "epoch": 0.8819697704323629, "grad_norm": 2.7545418739318848, "learning_rate": 3.922358599581499e-05, "loss": 7.6172, "step": 20919 }, { "epoch": 0.8820119316145625, "grad_norm": 1.5001550912857056, "learning_rate": 3.9222805209406914e-05, "loss": 6.6112, "step": 20920 }, { "epoch": 0.8820540927967621, "grad_norm": 1.805050253868103, "learning_rate": 3.9222024422998846e-05, "loss": 7.0066, "step": 20921 }, { "epoch": 0.8820962539789615, "grad_norm": 3.4938836097717285, "learning_rate": 3.922124363659078e-05, "loss": 7.5132, "step": 20922 }, { "epoch": 0.8821384151611611, "grad_norm": 1.8523415327072144, "learning_rate": 3.92204628501827e-05, "loss": 6.5195, "step": 20923 }, { "epoch": 0.8821805763433607, "grad_norm": 1.6625267267227173, "learning_rate": 3.921968206377464e-05, "loss": 6.6936, "step": 20924 }, { "epoch": 0.8822227375255602, "grad_norm": 2.507972002029419, "learning_rate": 3.9218901277366564e-05, "loss": 7.2863, "step": 20925 }, { "epoch": 0.8822648987077598, "grad_norm": 1.468924880027771, "learning_rate": 3.9218120490958495e-05, "loss": 6.379, "step": 20926 }, { "epoch": 0.8823070598899593, "grad_norm": 1.582263708114624, "learning_rate": 3.9217339704550426e-05, "loss": 6.9304, "step": 20927 }, { "epoch": 0.8823492210721589, "grad_norm": 1.3964065313339233, "learning_rate": 3.921655891814236e-05, "loss": 6.4825, "step": 20928 }, { "epoch": 0.8823913822543584, "grad_norm": 2.3434512615203857, "learning_rate": 3.921577813173428e-05, "loss": 7.3352, "step": 20929 }, { "epoch": 0.882433543436558, "grad_norm": 1.875316858291626, "learning_rate": 3.921499734532621e-05, "loss": 6.4413, "step": 20930 }, { "epoch": 0.8824757046187575, "grad_norm": 3.0010781288146973, "learning_rate": 3.9214216558918145e-05, "loss": 7.3817, "step": 20931 }, { "epoch": 0.882517865800957, "grad_norm": 1.8962929248809814, "learning_rate": 3.9213435772510076e-05, "loss": 6.5832, "step": 20932 }, { "epoch": 0.8825600269831566, "grad_norm": 3.7435033321380615, "learning_rate": 3.921265498610201e-05, "loss": 7.0494, "step": 20933 }, { "epoch": 0.8826021881653562, "grad_norm": 2.7866649627685547, "learning_rate": 3.921187419969393e-05, "loss": 7.637, "step": 20934 }, { "epoch": 0.8826443493475558, "grad_norm": 1.2916934490203857, "learning_rate": 3.921109341328586e-05, "loss": 6.9634, "step": 20935 }, { "epoch": 0.8826865105297552, "grad_norm": 3.2378954887390137, "learning_rate": 3.9210312626877794e-05, "loss": 7.7682, "step": 20936 }, { "epoch": 0.8827286717119548, "grad_norm": 4.450872898101807, "learning_rate": 3.920953184046972e-05, "loss": 8.1005, "step": 20937 }, { "epoch": 0.8827708328941544, "grad_norm": 2.2703540325164795, "learning_rate": 3.920875105406166e-05, "loss": 7.2, "step": 20938 }, { "epoch": 0.8828129940763539, "grad_norm": 1.0672459602355957, "learning_rate": 3.920797026765358e-05, "loss": 6.5822, "step": 20939 }, { "epoch": 0.8828551552585534, "grad_norm": 1.4196656942367554, "learning_rate": 3.920718948124551e-05, "loss": 7.039, "step": 20940 }, { "epoch": 0.882897316440753, "grad_norm": 1.2415045499801636, "learning_rate": 3.9206408694837444e-05, "loss": 6.8186, "step": 20941 }, { "epoch": 0.8829394776229526, "grad_norm": 1.2442225217819214, "learning_rate": 3.920562790842937e-05, "loss": 6.674, "step": 20942 }, { "epoch": 0.8829816388051521, "grad_norm": 1.6491669416427612, "learning_rate": 3.9204847122021306e-05, "loss": 6.5743, "step": 20943 }, { "epoch": 0.8830237999873517, "grad_norm": 1.318860650062561, "learning_rate": 3.920406633561323e-05, "loss": 7.0184, "step": 20944 }, { "epoch": 0.8830659611695512, "grad_norm": 3.0896823406219482, "learning_rate": 3.920328554920516e-05, "loss": 7.6101, "step": 20945 }, { "epoch": 0.8831081223517507, "grad_norm": 3.223360538482666, "learning_rate": 3.920250476279709e-05, "loss": 7.6458, "step": 20946 }, { "epoch": 0.8831502835339503, "grad_norm": 1.8755890130996704, "learning_rate": 3.920172397638902e-05, "loss": 6.5141, "step": 20947 }, { "epoch": 0.8831924447161499, "grad_norm": 1.0415101051330566, "learning_rate": 3.920094318998095e-05, "loss": 6.7726, "step": 20948 }, { "epoch": 0.8832346058983493, "grad_norm": 1.5164098739624023, "learning_rate": 3.920016240357288e-05, "loss": 6.847, "step": 20949 }, { "epoch": 0.8832767670805489, "grad_norm": 1.5878173112869263, "learning_rate": 3.919938161716481e-05, "loss": 6.554, "step": 20950 }, { "epoch": 0.8833189282627485, "grad_norm": 2.1012613773345947, "learning_rate": 3.9198600830756736e-05, "loss": 7.537, "step": 20951 }, { "epoch": 0.8833610894449481, "grad_norm": 1.4502041339874268, "learning_rate": 3.9197820044348674e-05, "loss": 6.5795, "step": 20952 }, { "epoch": 0.8834032506271476, "grad_norm": 2.7463338375091553, "learning_rate": 3.91970392579406e-05, "loss": 6.616, "step": 20953 }, { "epoch": 0.8834454118093471, "grad_norm": 3.1072614192962646, "learning_rate": 3.919625847153253e-05, "loss": 7.5325, "step": 20954 }, { "epoch": 0.8834875729915467, "grad_norm": 3.312847137451172, "learning_rate": 3.919547768512446e-05, "loss": 7.1651, "step": 20955 }, { "epoch": 0.8835297341737463, "grad_norm": 1.4664572477340698, "learning_rate": 3.9194696898716385e-05, "loss": 6.6859, "step": 20956 }, { "epoch": 0.8835718953559458, "grad_norm": 1.0438491106033325, "learning_rate": 3.9193916112308323e-05, "loss": 6.519, "step": 20957 }, { "epoch": 0.8836140565381453, "grad_norm": 1.3049811124801636, "learning_rate": 3.919313532590025e-05, "loss": 6.803, "step": 20958 }, { "epoch": 0.8836562177203449, "grad_norm": 1.893885850906372, "learning_rate": 3.919235453949218e-05, "loss": 6.4966, "step": 20959 }, { "epoch": 0.8836983789025444, "grad_norm": 1.1160764694213867, "learning_rate": 3.919157375308411e-05, "loss": 6.8431, "step": 20960 }, { "epoch": 0.883740540084744, "grad_norm": 1.706294059753418, "learning_rate": 3.9190792966676035e-05, "loss": 7.0477, "step": 20961 }, { "epoch": 0.8837827012669435, "grad_norm": 1.232947826385498, "learning_rate": 3.9190012180267966e-05, "loss": 6.66, "step": 20962 }, { "epoch": 0.883824862449143, "grad_norm": 1.3968075513839722, "learning_rate": 3.91892313938599e-05, "loss": 6.5868, "step": 20963 }, { "epoch": 0.8838670236313426, "grad_norm": 1.3543862104415894, "learning_rate": 3.918845060745183e-05, "loss": 6.5974, "step": 20964 }, { "epoch": 0.8839091848135422, "grad_norm": 1.1263865232467651, "learning_rate": 3.918766982104376e-05, "loss": 6.5948, "step": 20965 }, { "epoch": 0.8839513459957418, "grad_norm": 3.4353249073028564, "learning_rate": 3.9186889034635684e-05, "loss": 7.693, "step": 20966 }, { "epoch": 0.8839935071779412, "grad_norm": 3.122873306274414, "learning_rate": 3.9186108248227616e-05, "loss": 7.1574, "step": 20967 }, { "epoch": 0.8840356683601408, "grad_norm": 2.467590570449829, "learning_rate": 3.918532746181955e-05, "loss": 6.3486, "step": 20968 }, { "epoch": 0.8840778295423404, "grad_norm": 1.190736174583435, "learning_rate": 3.918454667541148e-05, "loss": 6.7014, "step": 20969 }, { "epoch": 0.88411999072454, "grad_norm": 1.8134678602218628, "learning_rate": 3.91837658890034e-05, "loss": 7.2885, "step": 20970 }, { "epoch": 0.8841621519067394, "grad_norm": 2.1349384784698486, "learning_rate": 3.918298510259534e-05, "loss": 7.2455, "step": 20971 }, { "epoch": 0.884204313088939, "grad_norm": 1.1999367475509644, "learning_rate": 3.9182204316187265e-05, "loss": 6.462, "step": 20972 }, { "epoch": 0.8842464742711386, "grad_norm": 1.4546812772750854, "learning_rate": 3.9181423529779197e-05, "loss": 6.6642, "step": 20973 }, { "epoch": 0.8842886354533381, "grad_norm": 1.7481153011322021, "learning_rate": 3.918064274337113e-05, "loss": 7.1033, "step": 20974 }, { "epoch": 0.8843307966355377, "grad_norm": 1.306848406791687, "learning_rate": 3.917986195696305e-05, "loss": 6.6673, "step": 20975 }, { "epoch": 0.8843729578177372, "grad_norm": 3.3100757598876953, "learning_rate": 3.917908117055499e-05, "loss": 7.8698, "step": 20976 }, { "epoch": 0.8844151189999367, "grad_norm": 0.9385282397270203, "learning_rate": 3.9178300384146915e-05, "loss": 6.7521, "step": 20977 }, { "epoch": 0.8844572801821363, "grad_norm": 1.8334470987319946, "learning_rate": 3.9177519597738846e-05, "loss": 6.6367, "step": 20978 }, { "epoch": 0.8844994413643359, "grad_norm": 1.3284975290298462, "learning_rate": 3.917673881133078e-05, "loss": 6.5195, "step": 20979 }, { "epoch": 0.8845416025465354, "grad_norm": 2.828925371170044, "learning_rate": 3.91759580249227e-05, "loss": 6.2851, "step": 20980 }, { "epoch": 0.8845837637287349, "grad_norm": 1.3966665267944336, "learning_rate": 3.917517723851463e-05, "loss": 6.9771, "step": 20981 }, { "epoch": 0.8846259249109345, "grad_norm": 0.9196174740791321, "learning_rate": 3.9174396452106564e-05, "loss": 6.5024, "step": 20982 }, { "epoch": 0.8846680860931341, "grad_norm": 0.9817567467689514, "learning_rate": 3.9173615665698496e-05, "loss": 6.4758, "step": 20983 }, { "epoch": 0.8847102472753336, "grad_norm": 2.439976453781128, "learning_rate": 3.917283487929042e-05, "loss": 6.6896, "step": 20984 }, { "epoch": 0.8847524084575331, "grad_norm": 1.1401686668395996, "learning_rate": 3.917205409288235e-05, "loss": 6.8999, "step": 20985 }, { "epoch": 0.8847945696397327, "grad_norm": 2.246537923812866, "learning_rate": 3.917127330647428e-05, "loss": 8.5656, "step": 20986 }, { "epoch": 0.8848367308219323, "grad_norm": 2.0240371227264404, "learning_rate": 3.917049252006621e-05, "loss": 7.1242, "step": 20987 }, { "epoch": 0.8848788920041318, "grad_norm": 1.6383304595947266, "learning_rate": 3.9169711733658145e-05, "loss": 6.9093, "step": 20988 }, { "epoch": 0.8849210531863313, "grad_norm": 1.1118600368499756, "learning_rate": 3.916893094725007e-05, "loss": 6.5225, "step": 20989 }, { "epoch": 0.8849632143685309, "grad_norm": 1.302762508392334, "learning_rate": 3.916815016084201e-05, "loss": 6.3691, "step": 20990 }, { "epoch": 0.8850053755507304, "grad_norm": 1.8644834756851196, "learning_rate": 3.916736937443393e-05, "loss": 7.4046, "step": 20991 }, { "epoch": 0.88504753673293, "grad_norm": 1.228740930557251, "learning_rate": 3.9166588588025857e-05, "loss": 7.4349, "step": 20992 }, { "epoch": 0.8850896979151296, "grad_norm": 1.6289386749267578, "learning_rate": 3.9165807801617795e-05, "loss": 6.9658, "step": 20993 }, { "epoch": 0.885131859097329, "grad_norm": 1.8611366748809814, "learning_rate": 3.916502701520972e-05, "loss": 6.7065, "step": 20994 }, { "epoch": 0.8851740202795286, "grad_norm": 3.624164581298828, "learning_rate": 3.916424622880165e-05, "loss": 7.6897, "step": 20995 }, { "epoch": 0.8852161814617282, "grad_norm": 1.018606424331665, "learning_rate": 3.916346544239358e-05, "loss": 6.7057, "step": 20996 }, { "epoch": 0.8852583426439278, "grad_norm": 1.065232753753662, "learning_rate": 3.916268465598551e-05, "loss": 6.5874, "step": 20997 }, { "epoch": 0.8853005038261272, "grad_norm": 0.8581459522247314, "learning_rate": 3.916190386957744e-05, "loss": 6.5255, "step": 20998 }, { "epoch": 0.8853426650083268, "grad_norm": 2.982363224029541, "learning_rate": 3.916112308316937e-05, "loss": 7.1406, "step": 20999 }, { "epoch": 0.8853848261905264, "grad_norm": 1.1949410438537598, "learning_rate": 3.91603422967613e-05, "loss": 6.6122, "step": 21000 }, { "epoch": 0.885426987372726, "grad_norm": 2.123248338699341, "learning_rate": 3.915956151035323e-05, "loss": 6.7938, "step": 21001 }, { "epoch": 0.8854691485549255, "grad_norm": 1.8060475587844849, "learning_rate": 3.915878072394516e-05, "loss": 6.3915, "step": 21002 }, { "epoch": 0.885511309737125, "grad_norm": 1.415529727935791, "learning_rate": 3.915799993753709e-05, "loss": 6.4608, "step": 21003 }, { "epoch": 0.8855534709193246, "grad_norm": 1.1971828937530518, "learning_rate": 3.915721915112902e-05, "loss": 6.4148, "step": 21004 }, { "epoch": 0.8855956321015241, "grad_norm": 1.2059359550476074, "learning_rate": 3.915643836472095e-05, "loss": 6.5394, "step": 21005 }, { "epoch": 0.8856377932837237, "grad_norm": 2.290811061859131, "learning_rate": 3.9155657578312874e-05, "loss": 6.4965, "step": 21006 }, { "epoch": 0.8856799544659232, "grad_norm": 1.164893388748169, "learning_rate": 3.915487679190481e-05, "loss": 6.5715, "step": 21007 }, { "epoch": 0.8857221156481228, "grad_norm": 1.9709771871566772, "learning_rate": 3.9154096005496736e-05, "loss": 6.726, "step": 21008 }, { "epoch": 0.8857642768303223, "grad_norm": 4.527133941650391, "learning_rate": 3.915331521908867e-05, "loss": 7.8174, "step": 21009 }, { "epoch": 0.8858064380125219, "grad_norm": 2.3830764293670654, "learning_rate": 3.91525344326806e-05, "loss": 7.355, "step": 21010 }, { "epoch": 0.8858485991947215, "grad_norm": 2.5477631092071533, "learning_rate": 3.9151753646272523e-05, "loss": 7.2856, "step": 21011 }, { "epoch": 0.8858907603769209, "grad_norm": 3.8431637287139893, "learning_rate": 3.915097285986446e-05, "loss": 7.7139, "step": 21012 }, { "epoch": 0.8859329215591205, "grad_norm": 1.3850749731063843, "learning_rate": 3.9150192073456386e-05, "loss": 6.6041, "step": 21013 }, { "epoch": 0.8859750827413201, "grad_norm": 1.7781944274902344, "learning_rate": 3.914941128704832e-05, "loss": 7.1113, "step": 21014 }, { "epoch": 0.8860172439235197, "grad_norm": 1.5451465845108032, "learning_rate": 3.914863050064025e-05, "loss": 6.9699, "step": 21015 }, { "epoch": 0.8860594051057191, "grad_norm": 1.6520161628723145, "learning_rate": 3.914784971423218e-05, "loss": 6.828, "step": 21016 }, { "epoch": 0.8861015662879187, "grad_norm": 2.061290979385376, "learning_rate": 3.9147068927824104e-05, "loss": 6.8944, "step": 21017 }, { "epoch": 0.8861437274701183, "grad_norm": 2.5225350856781006, "learning_rate": 3.9146288141416035e-05, "loss": 6.4212, "step": 21018 }, { "epoch": 0.8861858886523178, "grad_norm": 3.619107246398926, "learning_rate": 3.914550735500797e-05, "loss": 7.1244, "step": 21019 }, { "epoch": 0.8862280498345173, "grad_norm": 1.65958833694458, "learning_rate": 3.914472656859989e-05, "loss": 6.4899, "step": 21020 }, { "epoch": 0.8862702110167169, "grad_norm": 1.3509724140167236, "learning_rate": 3.914394578219183e-05, "loss": 7.0571, "step": 21021 }, { "epoch": 0.8863123721989165, "grad_norm": 2.2303764820098877, "learning_rate": 3.9143164995783754e-05, "loss": 6.4276, "step": 21022 }, { "epoch": 0.886354533381116, "grad_norm": 3.133275270462036, "learning_rate": 3.9142384209375685e-05, "loss": 7.0991, "step": 21023 }, { "epoch": 0.8863966945633156, "grad_norm": 2.2743263244628906, "learning_rate": 3.9141603422967616e-05, "loss": 7.0916, "step": 21024 }, { "epoch": 0.8864388557455151, "grad_norm": 2.0045366287231445, "learning_rate": 3.914082263655954e-05, "loss": 7.2317, "step": 21025 }, { "epoch": 0.8864810169277146, "grad_norm": 1.1589181423187256, "learning_rate": 3.914004185015148e-05, "loss": 6.6935, "step": 21026 }, { "epoch": 0.8865231781099142, "grad_norm": 2.6445319652557373, "learning_rate": 3.91392610637434e-05, "loss": 7.3304, "step": 21027 }, { "epoch": 0.8865653392921138, "grad_norm": 3.077951669692993, "learning_rate": 3.9138480277335335e-05, "loss": 7.8761, "step": 21028 }, { "epoch": 0.8866075004743132, "grad_norm": 2.7808125019073486, "learning_rate": 3.9137699490927266e-05, "loss": 7.6776, "step": 21029 }, { "epoch": 0.8866496616565128, "grad_norm": 2.7069644927978516, "learning_rate": 3.913691870451919e-05, "loss": 7.6519, "step": 21030 }, { "epoch": 0.8866918228387124, "grad_norm": 2.1814680099487305, "learning_rate": 3.913613791811112e-05, "loss": 7.5057, "step": 21031 }, { "epoch": 0.886733984020912, "grad_norm": 2.897203207015991, "learning_rate": 3.913535713170305e-05, "loss": 6.8817, "step": 21032 }, { "epoch": 0.8867761452031115, "grad_norm": 1.3173775672912598, "learning_rate": 3.9134576345294984e-05, "loss": 6.8432, "step": 21033 }, { "epoch": 0.886818306385311, "grad_norm": 1.2253849506378174, "learning_rate": 3.9133795558886915e-05, "loss": 7.0754, "step": 21034 }, { "epoch": 0.8868604675675106, "grad_norm": 2.666146993637085, "learning_rate": 3.9133014772478847e-05, "loss": 7.6898, "step": 21035 }, { "epoch": 0.8869026287497102, "grad_norm": 1.6923474073410034, "learning_rate": 3.913223398607077e-05, "loss": 6.7172, "step": 21036 }, { "epoch": 0.8869447899319097, "grad_norm": 1.3857110738754272, "learning_rate": 3.91314531996627e-05, "loss": 6.4767, "step": 21037 }, { "epoch": 0.8869869511141092, "grad_norm": 1.268256425857544, "learning_rate": 3.9130672413254634e-05, "loss": 6.7613, "step": 21038 }, { "epoch": 0.8870291122963088, "grad_norm": 1.300500512123108, "learning_rate": 3.912989162684656e-05, "loss": 6.5789, "step": 21039 }, { "epoch": 0.8870712734785083, "grad_norm": 3.443290948867798, "learning_rate": 3.9129110840438496e-05, "loss": 8.4811, "step": 21040 }, { "epoch": 0.8871134346607079, "grad_norm": 2.858081579208374, "learning_rate": 3.912833005403042e-05, "loss": 8.0359, "step": 21041 }, { "epoch": 0.8871555958429075, "grad_norm": 2.438829183578491, "learning_rate": 3.912754926762235e-05, "loss": 7.2255, "step": 21042 }, { "epoch": 0.887197757025107, "grad_norm": 1.5595548152923584, "learning_rate": 3.912676848121428e-05, "loss": 6.7839, "step": 21043 }, { "epoch": 0.8872399182073065, "grad_norm": 1.631237268447876, "learning_rate": 3.912598769480621e-05, "loss": 7.0379, "step": 21044 }, { "epoch": 0.8872820793895061, "grad_norm": 1.627353549003601, "learning_rate": 3.9125206908398146e-05, "loss": 7.216, "step": 21045 }, { "epoch": 0.8873242405717057, "grad_norm": 1.501978874206543, "learning_rate": 3.912442612199007e-05, "loss": 6.9868, "step": 21046 }, { "epoch": 0.8873664017539051, "grad_norm": 3.2583563327789307, "learning_rate": 3.9123645335582e-05, "loss": 7.6706, "step": 21047 }, { "epoch": 0.8874085629361047, "grad_norm": 2.059875011444092, "learning_rate": 3.912286454917393e-05, "loss": 7.58, "step": 21048 }, { "epoch": 0.8874507241183043, "grad_norm": 1.3865667581558228, "learning_rate": 3.912208376276586e-05, "loss": 6.4354, "step": 21049 }, { "epoch": 0.8874928853005039, "grad_norm": 2.6277573108673096, "learning_rate": 3.912130297635779e-05, "loss": 6.6488, "step": 21050 }, { "epoch": 0.8875350464827034, "grad_norm": 1.6871669292449951, "learning_rate": 3.912052218994972e-05, "loss": 6.5301, "step": 21051 }, { "epoch": 0.8875772076649029, "grad_norm": 1.1172066926956177, "learning_rate": 3.911974140354165e-05, "loss": 6.6219, "step": 21052 }, { "epoch": 0.8876193688471025, "grad_norm": 2.696791887283325, "learning_rate": 3.9118960617133575e-05, "loss": 7.5658, "step": 21053 }, { "epoch": 0.887661530029302, "grad_norm": 0.8737034797668457, "learning_rate": 3.9118179830725507e-05, "loss": 6.7958, "step": 21054 }, { "epoch": 0.8877036912115016, "grad_norm": 1.0415312051773071, "learning_rate": 3.911739904431744e-05, "loss": 6.5133, "step": 21055 }, { "epoch": 0.8877458523937011, "grad_norm": 2.062715768814087, "learning_rate": 3.911661825790937e-05, "loss": 7.5182, "step": 21056 }, { "epoch": 0.8877880135759006, "grad_norm": 3.8375864028930664, "learning_rate": 3.91158374715013e-05, "loss": 9.6049, "step": 21057 }, { "epoch": 0.8878301747581002, "grad_norm": 0.7116559147834778, "learning_rate": 3.9115056685093225e-05, "loss": 6.9301, "step": 21058 }, { "epoch": 0.8878723359402998, "grad_norm": 1.573067545890808, "learning_rate": 3.911427589868516e-05, "loss": 6.4347, "step": 21059 }, { "epoch": 0.8879144971224994, "grad_norm": 3.742814064025879, "learning_rate": 3.911349511227709e-05, "loss": 7.3166, "step": 21060 }, { "epoch": 0.8879566583046988, "grad_norm": 4.277873516082764, "learning_rate": 3.911271432586902e-05, "loss": 9.2021, "step": 21061 }, { "epoch": 0.8879988194868984, "grad_norm": 2.397052526473999, "learning_rate": 3.911193353946095e-05, "loss": 6.9143, "step": 21062 }, { "epoch": 0.888040980669098, "grad_norm": 1.8316411972045898, "learning_rate": 3.9111152753052874e-05, "loss": 6.4479, "step": 21063 }, { "epoch": 0.8880831418512976, "grad_norm": 1.6179444789886475, "learning_rate": 3.9110371966644806e-05, "loss": 6.4687, "step": 21064 }, { "epoch": 0.888125303033497, "grad_norm": 4.066281795501709, "learning_rate": 3.910959118023674e-05, "loss": 7.0193, "step": 21065 }, { "epoch": 0.8881674642156966, "grad_norm": 1.3179779052734375, "learning_rate": 3.910881039382867e-05, "loss": 7.0728, "step": 21066 }, { "epoch": 0.8882096253978962, "grad_norm": 1.2378075122833252, "learning_rate": 3.910802960742059e-05, "loss": 6.4779, "step": 21067 }, { "epoch": 0.8882517865800957, "grad_norm": 1.5488510131835938, "learning_rate": 3.9107248821012524e-05, "loss": 6.999, "step": 21068 }, { "epoch": 0.8882939477622952, "grad_norm": 2.108555316925049, "learning_rate": 3.9106468034604455e-05, "loss": 7.1343, "step": 21069 }, { "epoch": 0.8883361089444948, "grad_norm": 1.256263256072998, "learning_rate": 3.9105687248196386e-05, "loss": 7.1319, "step": 21070 }, { "epoch": 0.8883782701266943, "grad_norm": 2.3316774368286133, "learning_rate": 3.910490646178832e-05, "loss": 6.4679, "step": 21071 }, { "epoch": 0.8884204313088939, "grad_norm": 1.8700494766235352, "learning_rate": 3.910412567538024e-05, "loss": 6.5075, "step": 21072 }, { "epoch": 0.8884625924910935, "grad_norm": 1.748865008354187, "learning_rate": 3.9103344888972173e-05, "loss": 6.5007, "step": 21073 }, { "epoch": 0.888504753673293, "grad_norm": 2.838998317718506, "learning_rate": 3.9102564102564105e-05, "loss": 7.6326, "step": 21074 }, { "epoch": 0.8885469148554925, "grad_norm": 1.6424407958984375, "learning_rate": 3.910178331615603e-05, "loss": 7.0582, "step": 21075 }, { "epoch": 0.8885890760376921, "grad_norm": 1.636205792427063, "learning_rate": 3.910100252974797e-05, "loss": 7.0828, "step": 21076 }, { "epoch": 0.8886312372198917, "grad_norm": 2.1350934505462646, "learning_rate": 3.910022174333989e-05, "loss": 6.5424, "step": 21077 }, { "epoch": 0.8886733984020911, "grad_norm": 2.6999363899230957, "learning_rate": 3.909944095693182e-05, "loss": 7.4326, "step": 21078 }, { "epoch": 0.8887155595842907, "grad_norm": 1.4915046691894531, "learning_rate": 3.9098660170523754e-05, "loss": 6.5895, "step": 21079 }, { "epoch": 0.8887577207664903, "grad_norm": 1.6905399560928345, "learning_rate": 3.909787938411568e-05, "loss": 6.8798, "step": 21080 }, { "epoch": 0.8887998819486899, "grad_norm": 1.6100462675094604, "learning_rate": 3.909709859770762e-05, "loss": 6.5169, "step": 21081 }, { "epoch": 0.8888420431308894, "grad_norm": 1.6821209192276, "learning_rate": 3.909631781129954e-05, "loss": 6.4624, "step": 21082 }, { "epoch": 0.8888842043130889, "grad_norm": 2.559919834136963, "learning_rate": 3.909553702489147e-05, "loss": 6.3248, "step": 21083 }, { "epoch": 0.8889263654952885, "grad_norm": 2.190004825592041, "learning_rate": 3.9094756238483404e-05, "loss": 7.5231, "step": 21084 }, { "epoch": 0.888968526677488, "grad_norm": 1.5764909982681274, "learning_rate": 3.9093975452075335e-05, "loss": 6.5378, "step": 21085 }, { "epoch": 0.8890106878596876, "grad_norm": 1.4529975652694702, "learning_rate": 3.909319466566726e-05, "loss": 6.48, "step": 21086 }, { "epoch": 0.8890528490418871, "grad_norm": 1.6262489557266235, "learning_rate": 3.909241387925919e-05, "loss": 7.223, "step": 21087 }, { "epoch": 0.8890950102240867, "grad_norm": 0.9493919610977173, "learning_rate": 3.909163309285112e-05, "loss": 6.4891, "step": 21088 }, { "epoch": 0.8891371714062862, "grad_norm": 1.8026713132858276, "learning_rate": 3.9090852306443046e-05, "loss": 7.0912, "step": 21089 }, { "epoch": 0.8891793325884858, "grad_norm": 1.4000083208084106, "learning_rate": 3.9090071520034985e-05, "loss": 6.5751, "step": 21090 }, { "epoch": 0.8892214937706854, "grad_norm": 2.625030755996704, "learning_rate": 3.908929073362691e-05, "loss": 7.5849, "step": 21091 }, { "epoch": 0.8892636549528848, "grad_norm": 1.9234262704849243, "learning_rate": 3.908850994721884e-05, "loss": 6.8771, "step": 21092 }, { "epoch": 0.8893058161350844, "grad_norm": 1.9946925640106201, "learning_rate": 3.908772916081077e-05, "loss": 6.8317, "step": 21093 }, { "epoch": 0.889347977317284, "grad_norm": 1.4575660228729248, "learning_rate": 3.9086948374402696e-05, "loss": 6.812, "step": 21094 }, { "epoch": 0.8893901384994836, "grad_norm": 1.4150018692016602, "learning_rate": 3.9086167587994634e-05, "loss": 7.0015, "step": 21095 }, { "epoch": 0.889432299681683, "grad_norm": 1.5026884078979492, "learning_rate": 3.908538680158656e-05, "loss": 7.1883, "step": 21096 }, { "epoch": 0.8894744608638826, "grad_norm": 1.333764672279358, "learning_rate": 3.908460601517849e-05, "loss": 6.3277, "step": 21097 }, { "epoch": 0.8895166220460822, "grad_norm": 1.0380009412765503, "learning_rate": 3.908382522877042e-05, "loss": 6.664, "step": 21098 }, { "epoch": 0.8895587832282817, "grad_norm": 2.2228293418884277, "learning_rate": 3.9083044442362346e-05, "loss": 7.5626, "step": 21099 }, { "epoch": 0.8896009444104813, "grad_norm": 1.2865735292434692, "learning_rate": 3.908226365595428e-05, "loss": 6.6218, "step": 21100 }, { "epoch": 0.8896431055926808, "grad_norm": 1.3695145845413208, "learning_rate": 3.908148286954621e-05, "loss": 6.98, "step": 21101 }, { "epoch": 0.8896852667748804, "grad_norm": 3.517514944076538, "learning_rate": 3.908070208313814e-05, "loss": 8.8351, "step": 21102 }, { "epoch": 0.8897274279570799, "grad_norm": 2.8811194896698, "learning_rate": 3.907992129673007e-05, "loss": 6.6443, "step": 21103 }, { "epoch": 0.8897695891392795, "grad_norm": 1.1066796779632568, "learning_rate": 3.9079140510322e-05, "loss": 6.4249, "step": 21104 }, { "epoch": 0.889811750321479, "grad_norm": 1.9113377332687378, "learning_rate": 3.9078359723913926e-05, "loss": 6.7902, "step": 21105 }, { "epoch": 0.8898539115036785, "grad_norm": 1.5664644241333008, "learning_rate": 3.907757893750586e-05, "loss": 6.4892, "step": 21106 }, { "epoch": 0.8898960726858781, "grad_norm": 3.3684258460998535, "learning_rate": 3.907679815109779e-05, "loss": 7.1852, "step": 21107 }, { "epoch": 0.8899382338680777, "grad_norm": 1.2540637254714966, "learning_rate": 3.907601736468971e-05, "loss": 6.5827, "step": 21108 }, { "epoch": 0.8899803950502773, "grad_norm": 1.7015440464019775, "learning_rate": 3.907523657828165e-05, "loss": 6.333, "step": 21109 }, { "epoch": 0.8900225562324767, "grad_norm": 3.251725435256958, "learning_rate": 3.9074455791873576e-05, "loss": 7.0919, "step": 21110 }, { "epoch": 0.8900647174146763, "grad_norm": 1.225571870803833, "learning_rate": 3.907367500546551e-05, "loss": 6.4962, "step": 21111 }, { "epoch": 0.8901068785968759, "grad_norm": 4.586019992828369, "learning_rate": 3.907289421905744e-05, "loss": 10.4099, "step": 21112 }, { "epoch": 0.8901490397790754, "grad_norm": 3.131347417831421, "learning_rate": 3.907211343264936e-05, "loss": 7.7391, "step": 21113 }, { "epoch": 0.8901912009612749, "grad_norm": 2.717168092727661, "learning_rate": 3.90713326462413e-05, "loss": 7.9674, "step": 21114 }, { "epoch": 0.8902333621434745, "grad_norm": 2.7024083137512207, "learning_rate": 3.9070551859833225e-05, "loss": 7.5018, "step": 21115 }, { "epoch": 0.8902755233256741, "grad_norm": 1.4972115755081177, "learning_rate": 3.906977107342516e-05, "loss": 7.4284, "step": 21116 }, { "epoch": 0.8903176845078736, "grad_norm": 1.1503140926361084, "learning_rate": 3.906899028701709e-05, "loss": 7.5096, "step": 21117 }, { "epoch": 0.8903598456900732, "grad_norm": 1.3641642332077026, "learning_rate": 3.906820950060901e-05, "loss": 6.3689, "step": 21118 }, { "epoch": 0.8904020068722727, "grad_norm": 1.2345808744430542, "learning_rate": 3.9067428714200944e-05, "loss": 6.4579, "step": 21119 }, { "epoch": 0.8904441680544722, "grad_norm": 1.09103262424469, "learning_rate": 3.9066647927792875e-05, "loss": 6.7733, "step": 21120 }, { "epoch": 0.8904863292366718, "grad_norm": 2.9129321575164795, "learning_rate": 3.9065867141384806e-05, "loss": 7.2739, "step": 21121 }, { "epoch": 0.8905284904188714, "grad_norm": 1.3356393575668335, "learning_rate": 3.906508635497673e-05, "loss": 6.4834, "step": 21122 }, { "epoch": 0.8905706516010709, "grad_norm": 0.9532055854797363, "learning_rate": 3.906430556856867e-05, "loss": 6.6871, "step": 21123 }, { "epoch": 0.8906128127832704, "grad_norm": 2.094331741333008, "learning_rate": 3.906352478216059e-05, "loss": 7.093, "step": 21124 }, { "epoch": 0.89065497396547, "grad_norm": 3.5901923179626465, "learning_rate": 3.9062743995752524e-05, "loss": 8.056, "step": 21125 }, { "epoch": 0.8906971351476696, "grad_norm": 1.6300225257873535, "learning_rate": 3.9061963209344456e-05, "loss": 6.4666, "step": 21126 }, { "epoch": 0.890739296329869, "grad_norm": 1.7834367752075195, "learning_rate": 3.906118242293638e-05, "loss": 7.2347, "step": 21127 }, { "epoch": 0.8907814575120686, "grad_norm": 1.9524191617965698, "learning_rate": 3.906040163652832e-05, "loss": 6.4955, "step": 21128 }, { "epoch": 0.8908236186942682, "grad_norm": 1.456274151802063, "learning_rate": 3.905962085012024e-05, "loss": 6.9741, "step": 21129 }, { "epoch": 0.8908657798764678, "grad_norm": 1.4580495357513428, "learning_rate": 3.9058840063712174e-05, "loss": 6.4226, "step": 21130 }, { "epoch": 0.8909079410586673, "grad_norm": 1.295149326324463, "learning_rate": 3.9058059277304105e-05, "loss": 7.1589, "step": 21131 }, { "epoch": 0.8909501022408668, "grad_norm": 3.14969539642334, "learning_rate": 3.905727849089603e-05, "loss": 7.7567, "step": 21132 }, { "epoch": 0.8909922634230664, "grad_norm": 1.3568048477172852, "learning_rate": 3.905649770448796e-05, "loss": 6.7128, "step": 21133 }, { "epoch": 0.8910344246052659, "grad_norm": 1.3342090845108032, "learning_rate": 3.905571691807989e-05, "loss": 6.4503, "step": 21134 }, { "epoch": 0.8910765857874655, "grad_norm": 1.851548194885254, "learning_rate": 3.9054936131671823e-05, "loss": 6.9878, "step": 21135 }, { "epoch": 0.891118746969665, "grad_norm": 2.414215087890625, "learning_rate": 3.905415534526375e-05, "loss": 7.8418, "step": 21136 }, { "epoch": 0.8911609081518646, "grad_norm": 1.9180761575698853, "learning_rate": 3.905337455885568e-05, "loss": 6.6242, "step": 21137 }, { "epoch": 0.8912030693340641, "grad_norm": 1.7696008682250977, "learning_rate": 3.905259377244761e-05, "loss": 7.1525, "step": 21138 }, { "epoch": 0.8912452305162637, "grad_norm": 1.4898377656936646, "learning_rate": 3.905181298603954e-05, "loss": 7.0962, "step": 21139 }, { "epoch": 0.8912873916984633, "grad_norm": 1.6981244087219238, "learning_rate": 3.905103219963147e-05, "loss": 7.0513, "step": 21140 }, { "epoch": 0.8913295528806627, "grad_norm": 1.8654429912567139, "learning_rate": 3.90502514132234e-05, "loss": 6.4455, "step": 21141 }, { "epoch": 0.8913717140628623, "grad_norm": 1.4690321683883667, "learning_rate": 3.904947062681533e-05, "loss": 6.7642, "step": 21142 }, { "epoch": 0.8914138752450619, "grad_norm": 2.377087116241455, "learning_rate": 3.904868984040726e-05, "loss": 7.5391, "step": 21143 }, { "epoch": 0.8914560364272615, "grad_norm": 2.499048948287964, "learning_rate": 3.9047909053999184e-05, "loss": 6.4629, "step": 21144 }, { "epoch": 0.8914981976094609, "grad_norm": 4.767200469970703, "learning_rate": 3.904712826759112e-05, "loss": 7.8025, "step": 21145 }, { "epoch": 0.8915403587916605, "grad_norm": 1.660159707069397, "learning_rate": 3.904634748118305e-05, "loss": 6.3325, "step": 21146 }, { "epoch": 0.8915825199738601, "grad_norm": 1.0958709716796875, "learning_rate": 3.904556669477498e-05, "loss": 6.4191, "step": 21147 }, { "epoch": 0.8916246811560596, "grad_norm": 2.009455680847168, "learning_rate": 3.904478590836691e-05, "loss": 7.1393, "step": 21148 }, { "epoch": 0.8916668423382592, "grad_norm": 1.3176509141921997, "learning_rate": 3.904400512195884e-05, "loss": 6.4519, "step": 21149 }, { "epoch": 0.8917090035204587, "grad_norm": 2.561704397201538, "learning_rate": 3.904322433555077e-05, "loss": 7.4663, "step": 21150 }, { "epoch": 0.8917511647026582, "grad_norm": 1.7271642684936523, "learning_rate": 3.9042443549142697e-05, "loss": 6.893, "step": 21151 }, { "epoch": 0.8917933258848578, "grad_norm": 3.120751142501831, "learning_rate": 3.904166276273463e-05, "loss": 7.4633, "step": 21152 }, { "epoch": 0.8918354870670574, "grad_norm": 1.945723533630371, "learning_rate": 3.904088197632656e-05, "loss": 7.0595, "step": 21153 }, { "epoch": 0.8918776482492569, "grad_norm": 1.7905281782150269, "learning_rate": 3.904010118991849e-05, "loss": 6.6686, "step": 21154 }, { "epoch": 0.8919198094314564, "grad_norm": 1.509974718093872, "learning_rate": 3.9039320403510415e-05, "loss": 6.5741, "step": 21155 }, { "epoch": 0.891961970613656, "grad_norm": 2.013429641723633, "learning_rate": 3.9038539617102346e-05, "loss": 7.113, "step": 21156 }, { "epoch": 0.8920041317958556, "grad_norm": 2.551356315612793, "learning_rate": 3.903775883069428e-05, "loss": 7.2589, "step": 21157 }, { "epoch": 0.8920462929780552, "grad_norm": 3.7284934520721436, "learning_rate": 3.90369780442862e-05, "loss": 7.4822, "step": 21158 }, { "epoch": 0.8920884541602546, "grad_norm": 1.747839093208313, "learning_rate": 3.903619725787814e-05, "loss": 6.5749, "step": 21159 }, { "epoch": 0.8921306153424542, "grad_norm": 2.2471764087677, "learning_rate": 3.9035416471470064e-05, "loss": 7.2787, "step": 21160 }, { "epoch": 0.8921727765246538, "grad_norm": 1.2788387537002563, "learning_rate": 3.9034635685061996e-05, "loss": 6.5584, "step": 21161 }, { "epoch": 0.8922149377068533, "grad_norm": 1.5846182107925415, "learning_rate": 3.903385489865393e-05, "loss": 6.4596, "step": 21162 }, { "epoch": 0.8922570988890528, "grad_norm": 2.747694969177246, "learning_rate": 3.903307411224585e-05, "loss": 6.3521, "step": 21163 }, { "epoch": 0.8922992600712524, "grad_norm": 2.011568546295166, "learning_rate": 3.903229332583779e-05, "loss": 6.998, "step": 21164 }, { "epoch": 0.892341421253452, "grad_norm": 1.4418010711669922, "learning_rate": 3.9031512539429714e-05, "loss": 6.4527, "step": 21165 }, { "epoch": 0.8923835824356515, "grad_norm": 3.108161687850952, "learning_rate": 3.9030731753021645e-05, "loss": 7.6757, "step": 21166 }, { "epoch": 0.8924257436178511, "grad_norm": 1.7133069038391113, "learning_rate": 3.9029950966613576e-05, "loss": 7.5074, "step": 21167 }, { "epoch": 0.8924679048000506, "grad_norm": 1.467882513999939, "learning_rate": 3.902917018020551e-05, "loss": 7.2077, "step": 21168 }, { "epoch": 0.8925100659822501, "grad_norm": 2.925732374191284, "learning_rate": 3.902838939379743e-05, "loss": 7.768, "step": 21169 }, { "epoch": 0.8925522271644497, "grad_norm": 2.7514896392822266, "learning_rate": 3.902760860738936e-05, "loss": 7.6546, "step": 21170 }, { "epoch": 0.8925943883466493, "grad_norm": 1.1449143886566162, "learning_rate": 3.9026827820981295e-05, "loss": 6.8222, "step": 21171 }, { "epoch": 0.8926365495288487, "grad_norm": 1.182586669921875, "learning_rate": 3.9026047034573226e-05, "loss": 6.4852, "step": 21172 }, { "epoch": 0.8926787107110483, "grad_norm": 2.0406620502471924, "learning_rate": 3.902526624816516e-05, "loss": 7.1965, "step": 21173 }, { "epoch": 0.8927208718932479, "grad_norm": 1.8097407817840576, "learning_rate": 3.902448546175708e-05, "loss": 7.2045, "step": 21174 }, { "epoch": 0.8927630330754475, "grad_norm": 1.6530940532684326, "learning_rate": 3.902370467534901e-05, "loss": 7.352, "step": 21175 }, { "epoch": 0.892805194257647, "grad_norm": 1.7825688123703003, "learning_rate": 3.9022923888940944e-05, "loss": 6.6339, "step": 21176 }, { "epoch": 0.8928473554398465, "grad_norm": 1.4896576404571533, "learning_rate": 3.902214310253287e-05, "loss": 6.7582, "step": 21177 }, { "epoch": 0.8928895166220461, "grad_norm": 2.906756639480591, "learning_rate": 3.902136231612481e-05, "loss": 6.4396, "step": 21178 }, { "epoch": 0.8929316778042456, "grad_norm": 1.9074465036392212, "learning_rate": 3.902058152971673e-05, "loss": 7.4155, "step": 21179 }, { "epoch": 0.8929738389864452, "grad_norm": 2.470428466796875, "learning_rate": 3.901980074330866e-05, "loss": 7.9618, "step": 21180 }, { "epoch": 0.8930160001686447, "grad_norm": 1.0238193273544312, "learning_rate": 3.9019019956900594e-05, "loss": 7.0449, "step": 21181 }, { "epoch": 0.8930581613508443, "grad_norm": 1.7051700353622437, "learning_rate": 3.901823917049252e-05, "loss": 6.3979, "step": 21182 }, { "epoch": 0.8931003225330438, "grad_norm": 1.558650016784668, "learning_rate": 3.9017458384084456e-05, "loss": 6.7403, "step": 21183 }, { "epoch": 0.8931424837152434, "grad_norm": 2.7158937454223633, "learning_rate": 3.901667759767638e-05, "loss": 7.9252, "step": 21184 }, { "epoch": 0.8931846448974429, "grad_norm": 3.76632022857666, "learning_rate": 3.901589681126831e-05, "loss": 8.0574, "step": 21185 }, { "epoch": 0.8932268060796424, "grad_norm": 1.3392853736877441, "learning_rate": 3.901511602486024e-05, "loss": 6.7583, "step": 21186 }, { "epoch": 0.893268967261842, "grad_norm": 1.5599020719528198, "learning_rate": 3.901433523845217e-05, "loss": 6.8614, "step": 21187 }, { "epoch": 0.8933111284440416, "grad_norm": 2.7164416313171387, "learning_rate": 3.90135544520441e-05, "loss": 7.6825, "step": 21188 }, { "epoch": 0.8933532896262412, "grad_norm": 1.46366548538208, "learning_rate": 3.901277366563603e-05, "loss": 6.974, "step": 21189 }, { "epoch": 0.8933954508084406, "grad_norm": 1.5808297395706177, "learning_rate": 3.901199287922796e-05, "loss": 6.9043, "step": 21190 }, { "epoch": 0.8934376119906402, "grad_norm": 1.8244792222976685, "learning_rate": 3.9011212092819886e-05, "loss": 7.0111, "step": 21191 }, { "epoch": 0.8934797731728398, "grad_norm": 1.9494459629058838, "learning_rate": 3.9010431306411824e-05, "loss": 8.1192, "step": 21192 }, { "epoch": 0.8935219343550393, "grad_norm": 2.311983108520508, "learning_rate": 3.900965052000375e-05, "loss": 7.4334, "step": 21193 }, { "epoch": 0.8935640955372388, "grad_norm": 1.3149707317352295, "learning_rate": 3.900886973359568e-05, "loss": 6.527, "step": 21194 }, { "epoch": 0.8936062567194384, "grad_norm": 2.5168421268463135, "learning_rate": 3.900808894718761e-05, "loss": 7.4718, "step": 21195 }, { "epoch": 0.893648417901638, "grad_norm": 1.6544673442840576, "learning_rate": 3.9007308160779535e-05, "loss": 6.8165, "step": 21196 }, { "epoch": 0.8936905790838375, "grad_norm": 3.0374343395233154, "learning_rate": 3.9006527374371473e-05, "loss": 7.2774, "step": 21197 }, { "epoch": 0.8937327402660371, "grad_norm": 3.1767098903656006, "learning_rate": 3.90057465879634e-05, "loss": 6.6665, "step": 21198 }, { "epoch": 0.8937749014482366, "grad_norm": 2.233812093734741, "learning_rate": 3.900496580155533e-05, "loss": 7.9002, "step": 21199 }, { "epoch": 0.8938170626304361, "grad_norm": 3.2056753635406494, "learning_rate": 3.900418501514726e-05, "loss": 7.412, "step": 21200 }, { "epoch": 0.8938592238126357, "grad_norm": 1.6885513067245483, "learning_rate": 3.9003404228739185e-05, "loss": 6.6282, "step": 21201 }, { "epoch": 0.8939013849948353, "grad_norm": 3.22873854637146, "learning_rate": 3.9002623442331116e-05, "loss": 7.521, "step": 21202 }, { "epoch": 0.8939435461770348, "grad_norm": 1.8184269666671753, "learning_rate": 3.900184265592305e-05, "loss": 6.718, "step": 21203 }, { "epoch": 0.8939857073592343, "grad_norm": 2.415558099746704, "learning_rate": 3.900106186951498e-05, "loss": 6.4226, "step": 21204 }, { "epoch": 0.8940278685414339, "grad_norm": 2.4681997299194336, "learning_rate": 3.900028108310691e-05, "loss": 7.3674, "step": 21205 }, { "epoch": 0.8940700297236335, "grad_norm": 2.9957618713378906, "learning_rate": 3.8999500296698834e-05, "loss": 6.9854, "step": 21206 }, { "epoch": 0.894112190905833, "grad_norm": 3.4016928672790527, "learning_rate": 3.8998719510290766e-05, "loss": 7.5338, "step": 21207 }, { "epoch": 0.8941543520880325, "grad_norm": 2.440854787826538, "learning_rate": 3.89979387238827e-05, "loss": 6.6132, "step": 21208 }, { "epoch": 0.8941965132702321, "grad_norm": 2.4057910442352295, "learning_rate": 3.899715793747463e-05, "loss": 6.613, "step": 21209 }, { "epoch": 0.8942386744524317, "grad_norm": 1.9078477621078491, "learning_rate": 3.899637715106655e-05, "loss": 6.9953, "step": 21210 }, { "epoch": 0.8942808356346312, "grad_norm": 1.2094647884368896, "learning_rate": 3.899559636465849e-05, "loss": 6.681, "step": 21211 }, { "epoch": 0.8943229968168307, "grad_norm": 1.263365387916565, "learning_rate": 3.8994815578250415e-05, "loss": 6.5016, "step": 21212 }, { "epoch": 0.8943651579990303, "grad_norm": 1.758463740348816, "learning_rate": 3.8994034791842347e-05, "loss": 7.1415, "step": 21213 }, { "epoch": 0.8944073191812298, "grad_norm": 1.9127432107925415, "learning_rate": 3.899325400543428e-05, "loss": 6.6478, "step": 21214 }, { "epoch": 0.8944494803634294, "grad_norm": 2.914259910583496, "learning_rate": 3.89924732190262e-05, "loss": 7.6746, "step": 21215 }, { "epoch": 0.894491641545629, "grad_norm": 1.652311086654663, "learning_rate": 3.8991692432618134e-05, "loss": 7.209, "step": 21216 }, { "epoch": 0.8945338027278285, "grad_norm": 3.316305637359619, "learning_rate": 3.8990911646210065e-05, "loss": 7.8924, "step": 21217 }, { "epoch": 0.894575963910028, "grad_norm": 2.4830002784729004, "learning_rate": 3.8990130859801996e-05, "loss": 7.2233, "step": 21218 }, { "epoch": 0.8946181250922276, "grad_norm": 1.4008804559707642, "learning_rate": 3.898935007339393e-05, "loss": 6.6916, "step": 21219 }, { "epoch": 0.8946602862744272, "grad_norm": 2.2188942432403564, "learning_rate": 3.898856928698585e-05, "loss": 6.7255, "step": 21220 }, { "epoch": 0.8947024474566266, "grad_norm": 1.4328330755233765, "learning_rate": 3.898778850057778e-05, "loss": 6.3893, "step": 21221 }, { "epoch": 0.8947446086388262, "grad_norm": 1.456278681755066, "learning_rate": 3.8987007714169714e-05, "loss": 6.9628, "step": 21222 }, { "epoch": 0.8947867698210258, "grad_norm": 3.120950222015381, "learning_rate": 3.8986226927761646e-05, "loss": 8.0644, "step": 21223 }, { "epoch": 0.8948289310032254, "grad_norm": 1.5729939937591553, "learning_rate": 3.898544614135357e-05, "loss": 6.5976, "step": 21224 }, { "epoch": 0.8948710921854249, "grad_norm": 3.029754638671875, "learning_rate": 3.89846653549455e-05, "loss": 6.2617, "step": 21225 }, { "epoch": 0.8949132533676244, "grad_norm": 1.661718487739563, "learning_rate": 3.898388456853743e-05, "loss": 7.0654, "step": 21226 }, { "epoch": 0.894955414549824, "grad_norm": 2.866244316101074, "learning_rate": 3.898310378212936e-05, "loss": 6.3171, "step": 21227 }, { "epoch": 0.8949975757320235, "grad_norm": 2.903648853302002, "learning_rate": 3.8982322995721295e-05, "loss": 7.873, "step": 21228 }, { "epoch": 0.8950397369142231, "grad_norm": 1.4520666599273682, "learning_rate": 3.898154220931322e-05, "loss": 6.5461, "step": 21229 }, { "epoch": 0.8950818980964226, "grad_norm": 1.529021978378296, "learning_rate": 3.898076142290516e-05, "loss": 6.4314, "step": 21230 }, { "epoch": 0.8951240592786222, "grad_norm": 1.9356414079666138, "learning_rate": 3.897998063649708e-05, "loss": 6.8986, "step": 21231 }, { "epoch": 0.8951662204608217, "grad_norm": 1.6869150400161743, "learning_rate": 3.8979199850089007e-05, "loss": 7.1342, "step": 21232 }, { "epoch": 0.8952083816430213, "grad_norm": 1.4685746431350708, "learning_rate": 3.8978419063680945e-05, "loss": 6.7245, "step": 21233 }, { "epoch": 0.8952505428252209, "grad_norm": 2.2855112552642822, "learning_rate": 3.897763827727287e-05, "loss": 7.089, "step": 21234 }, { "epoch": 0.8952927040074203, "grad_norm": 2.7213711738586426, "learning_rate": 3.89768574908648e-05, "loss": 7.0525, "step": 21235 }, { "epoch": 0.8953348651896199, "grad_norm": 2.8855700492858887, "learning_rate": 3.897607670445673e-05, "loss": 7.3261, "step": 21236 }, { "epoch": 0.8953770263718195, "grad_norm": 3.0254478454589844, "learning_rate": 3.897529591804866e-05, "loss": 6.8564, "step": 21237 }, { "epoch": 0.8954191875540191, "grad_norm": 1.4449232816696167, "learning_rate": 3.897451513164059e-05, "loss": 6.4459, "step": 21238 }, { "epoch": 0.8954613487362185, "grad_norm": 3.116701364517212, "learning_rate": 3.897373434523252e-05, "loss": 8.0657, "step": 21239 }, { "epoch": 0.8955035099184181, "grad_norm": 3.6799306869506836, "learning_rate": 3.897295355882445e-05, "loss": 7.1562, "step": 21240 }, { "epoch": 0.8955456711006177, "grad_norm": 1.4586904048919678, "learning_rate": 3.897217277241638e-05, "loss": 7.023, "step": 21241 }, { "epoch": 0.8955878322828172, "grad_norm": 1.1426438093185425, "learning_rate": 3.897139198600831e-05, "loss": 6.9275, "step": 21242 }, { "epoch": 0.8956299934650167, "grad_norm": 1.3317323923110962, "learning_rate": 3.897061119960024e-05, "loss": 6.9445, "step": 21243 }, { "epoch": 0.8956721546472163, "grad_norm": 1.4162054061889648, "learning_rate": 3.896983041319217e-05, "loss": 7.1859, "step": 21244 }, { "epoch": 0.8957143158294159, "grad_norm": 1.4358196258544922, "learning_rate": 3.89690496267841e-05, "loss": 6.5598, "step": 21245 }, { "epoch": 0.8957564770116154, "grad_norm": 1.0374698638916016, "learning_rate": 3.8968268840376024e-05, "loss": 6.6117, "step": 21246 }, { "epoch": 0.895798638193815, "grad_norm": 1.355696678161621, "learning_rate": 3.896748805396796e-05, "loss": 6.7716, "step": 21247 }, { "epoch": 0.8958407993760145, "grad_norm": 2.9448928833007812, "learning_rate": 3.8966707267559886e-05, "loss": 6.8358, "step": 21248 }, { "epoch": 0.895882960558214, "grad_norm": 3.3123810291290283, "learning_rate": 3.896592648115182e-05, "loss": 7.4952, "step": 21249 }, { "epoch": 0.8959251217404136, "grad_norm": 2.6714060306549072, "learning_rate": 3.896514569474375e-05, "loss": 7.4423, "step": 21250 }, { "epoch": 0.8959672829226132, "grad_norm": 1.5516353845596313, "learning_rate": 3.8964364908335673e-05, "loss": 6.4085, "step": 21251 }, { "epoch": 0.8960094441048126, "grad_norm": 1.7712678909301758, "learning_rate": 3.896358412192761e-05, "loss": 6.8913, "step": 21252 }, { "epoch": 0.8960516052870122, "grad_norm": 2.332976818084717, "learning_rate": 3.8962803335519536e-05, "loss": 7.7529, "step": 21253 }, { "epoch": 0.8960937664692118, "grad_norm": 1.7453242540359497, "learning_rate": 3.896202254911147e-05, "loss": 6.4562, "step": 21254 }, { "epoch": 0.8961359276514114, "grad_norm": 1.2449854612350464, "learning_rate": 3.89612417627034e-05, "loss": 6.6579, "step": 21255 }, { "epoch": 0.8961780888336109, "grad_norm": 2.057141065597534, "learning_rate": 3.896046097629533e-05, "loss": 7.1559, "step": 21256 }, { "epoch": 0.8962202500158104, "grad_norm": 1.7538992166519165, "learning_rate": 3.8959680189887254e-05, "loss": 6.5863, "step": 21257 }, { "epoch": 0.89626241119801, "grad_norm": 1.9651179313659668, "learning_rate": 3.8958899403479185e-05, "loss": 7.1766, "step": 21258 }, { "epoch": 0.8963045723802096, "grad_norm": 2.703232526779175, "learning_rate": 3.895811861707112e-05, "loss": 6.519, "step": 21259 }, { "epoch": 0.8963467335624091, "grad_norm": 3.414402484893799, "learning_rate": 3.895733783066304e-05, "loss": 7.2894, "step": 21260 }, { "epoch": 0.8963888947446086, "grad_norm": 3.482548713684082, "learning_rate": 3.895655704425498e-05, "loss": 7.7108, "step": 21261 }, { "epoch": 0.8964310559268082, "grad_norm": 1.4926705360412598, "learning_rate": 3.8955776257846904e-05, "loss": 6.6262, "step": 21262 }, { "epoch": 0.8964732171090077, "grad_norm": 1.8242897987365723, "learning_rate": 3.8954995471438835e-05, "loss": 6.62, "step": 21263 }, { "epoch": 0.8965153782912073, "grad_norm": 0.9104161858558655, "learning_rate": 3.8954214685030766e-05, "loss": 6.4858, "step": 21264 }, { "epoch": 0.8965575394734069, "grad_norm": 2.838787317276001, "learning_rate": 3.895343389862269e-05, "loss": 7.945, "step": 21265 }, { "epoch": 0.8965997006556063, "grad_norm": 2.1829254627227783, "learning_rate": 3.895265311221463e-05, "loss": 7.1629, "step": 21266 }, { "epoch": 0.8966418618378059, "grad_norm": 2.2323896884918213, "learning_rate": 3.895187232580655e-05, "loss": 6.9646, "step": 21267 }, { "epoch": 0.8966840230200055, "grad_norm": 2.6513569355010986, "learning_rate": 3.8951091539398485e-05, "loss": 7.1418, "step": 21268 }, { "epoch": 0.8967261842022051, "grad_norm": 2.6105384826660156, "learning_rate": 3.8950310752990416e-05, "loss": 7.6789, "step": 21269 }, { "epoch": 0.8967683453844045, "grad_norm": 1.5251082181930542, "learning_rate": 3.894952996658234e-05, "loss": 7.7158, "step": 21270 }, { "epoch": 0.8968105065666041, "grad_norm": 1.8213348388671875, "learning_rate": 3.894874918017427e-05, "loss": 6.537, "step": 21271 }, { "epoch": 0.8968526677488037, "grad_norm": 1.6014450788497925, "learning_rate": 3.89479683937662e-05, "loss": 6.9568, "step": 21272 }, { "epoch": 0.8968948289310033, "grad_norm": 1.8550337553024292, "learning_rate": 3.8947187607358134e-05, "loss": 7.0123, "step": 21273 }, { "epoch": 0.8969369901132028, "grad_norm": 3.2254137992858887, "learning_rate": 3.8946406820950065e-05, "loss": 6.4286, "step": 21274 }, { "epoch": 0.8969791512954023, "grad_norm": 1.4813063144683838, "learning_rate": 3.8945626034541997e-05, "loss": 7.0371, "step": 21275 }, { "epoch": 0.8970213124776019, "grad_norm": 3.0904858112335205, "learning_rate": 3.894484524813392e-05, "loss": 7.5954, "step": 21276 }, { "epoch": 0.8970634736598014, "grad_norm": 1.5977791547775269, "learning_rate": 3.894406446172585e-05, "loss": 6.9434, "step": 21277 }, { "epoch": 0.897105634842001, "grad_norm": 1.587316632270813, "learning_rate": 3.8943283675317784e-05, "loss": 7.0904, "step": 21278 }, { "epoch": 0.8971477960242005, "grad_norm": 1.530556082725525, "learning_rate": 3.894250288890971e-05, "loss": 7.266, "step": 21279 }, { "epoch": 0.8971899572064, "grad_norm": 2.7035484313964844, "learning_rate": 3.8941722102501646e-05, "loss": 7.3709, "step": 21280 }, { "epoch": 0.8972321183885996, "grad_norm": 1.2900091409683228, "learning_rate": 3.894094131609357e-05, "loss": 6.5812, "step": 21281 }, { "epoch": 0.8972742795707992, "grad_norm": 1.5851004123687744, "learning_rate": 3.89401605296855e-05, "loss": 6.8043, "step": 21282 }, { "epoch": 0.8973164407529988, "grad_norm": 1.4630320072174072, "learning_rate": 3.893937974327743e-05, "loss": 6.5376, "step": 21283 }, { "epoch": 0.8973586019351982, "grad_norm": 2.4919748306274414, "learning_rate": 3.893859895686936e-05, "loss": 7.4762, "step": 21284 }, { "epoch": 0.8974007631173978, "grad_norm": 3.6793620586395264, "learning_rate": 3.8937818170461296e-05, "loss": 7.586, "step": 21285 }, { "epoch": 0.8974429242995974, "grad_norm": 4.327157020568848, "learning_rate": 3.893703738405322e-05, "loss": 7.9826, "step": 21286 }, { "epoch": 0.897485085481797, "grad_norm": 1.7645151615142822, "learning_rate": 3.893625659764515e-05, "loss": 7.0597, "step": 21287 }, { "epoch": 0.8975272466639964, "grad_norm": 1.795578956604004, "learning_rate": 3.893547581123708e-05, "loss": 7.6013, "step": 21288 }, { "epoch": 0.897569407846196, "grad_norm": 1.5127177238464355, "learning_rate": 3.893469502482901e-05, "loss": 7.6407, "step": 21289 }, { "epoch": 0.8976115690283956, "grad_norm": 7.356631278991699, "learning_rate": 3.893391423842094e-05, "loss": 6.7598, "step": 21290 }, { "epoch": 0.8976537302105951, "grad_norm": 2.7069053649902344, "learning_rate": 3.893313345201287e-05, "loss": 6.653, "step": 21291 }, { "epoch": 0.8976958913927947, "grad_norm": 1.9941167831420898, "learning_rate": 3.89323526656048e-05, "loss": 7.3482, "step": 21292 }, { "epoch": 0.8977380525749942, "grad_norm": 2.666874408721924, "learning_rate": 3.8931571879196725e-05, "loss": 7.119, "step": 21293 }, { "epoch": 0.8977802137571937, "grad_norm": 1.1356154680252075, "learning_rate": 3.8930791092788657e-05, "loss": 6.4953, "step": 21294 }, { "epoch": 0.8978223749393933, "grad_norm": 1.4319658279418945, "learning_rate": 3.893001030638059e-05, "loss": 7.2712, "step": 21295 }, { "epoch": 0.8978645361215929, "grad_norm": 2.7769927978515625, "learning_rate": 3.892922951997251e-05, "loss": 7.955, "step": 21296 }, { "epoch": 0.8979066973037924, "grad_norm": 3.1072747707366943, "learning_rate": 3.892844873356445e-05, "loss": 7.3872, "step": 21297 }, { "epoch": 0.8979488584859919, "grad_norm": 1.389634609222412, "learning_rate": 3.8927667947156375e-05, "loss": 6.5997, "step": 21298 }, { "epoch": 0.8979910196681915, "grad_norm": 1.2485835552215576, "learning_rate": 3.892688716074831e-05, "loss": 6.6838, "step": 21299 }, { "epoch": 0.8980331808503911, "grad_norm": 2.1746463775634766, "learning_rate": 3.892610637434024e-05, "loss": 6.9013, "step": 21300 }, { "epoch": 0.8980753420325905, "grad_norm": 1.6838064193725586, "learning_rate": 3.892532558793217e-05, "loss": 7.1108, "step": 21301 }, { "epoch": 0.8981175032147901, "grad_norm": 1.6631892919540405, "learning_rate": 3.89245448015241e-05, "loss": 6.5693, "step": 21302 }, { "epoch": 0.8981596643969897, "grad_norm": 2.3739140033721924, "learning_rate": 3.8923764015116024e-05, "loss": 7.4956, "step": 21303 }, { "epoch": 0.8982018255791893, "grad_norm": 1.2617491483688354, "learning_rate": 3.8922983228707956e-05, "loss": 6.509, "step": 21304 }, { "epoch": 0.8982439867613888, "grad_norm": 1.3668104410171509, "learning_rate": 3.892220244229989e-05, "loss": 6.9211, "step": 21305 }, { "epoch": 0.8982861479435883, "grad_norm": 1.0475904941558838, "learning_rate": 3.892142165589182e-05, "loss": 6.5502, "step": 21306 }, { "epoch": 0.8983283091257879, "grad_norm": 0.8975088000297546, "learning_rate": 3.892064086948374e-05, "loss": 6.7095, "step": 21307 }, { "epoch": 0.8983704703079874, "grad_norm": 1.3905600309371948, "learning_rate": 3.8919860083075674e-05, "loss": 7.1221, "step": 21308 }, { "epoch": 0.898412631490187, "grad_norm": 2.2060739994049072, "learning_rate": 3.8919079296667605e-05, "loss": 8.3767, "step": 21309 }, { "epoch": 0.8984547926723865, "grad_norm": 3.60487699508667, "learning_rate": 3.8918298510259536e-05, "loss": 7.4883, "step": 21310 }, { "epoch": 0.8984969538545861, "grad_norm": 2.1162328720092773, "learning_rate": 3.891751772385147e-05, "loss": 7.2258, "step": 21311 }, { "epoch": 0.8985391150367856, "grad_norm": 3.214768171310425, "learning_rate": 3.891673693744339e-05, "loss": 7.4433, "step": 21312 }, { "epoch": 0.8985812762189852, "grad_norm": 3.1780426502227783, "learning_rate": 3.8915956151035323e-05, "loss": 6.8725, "step": 21313 }, { "epoch": 0.8986234374011848, "grad_norm": 2.4094045162200928, "learning_rate": 3.8915175364627255e-05, "loss": 7.0492, "step": 21314 }, { "epoch": 0.8986655985833842, "grad_norm": 2.2902181148529053, "learning_rate": 3.891439457821918e-05, "loss": 7.3509, "step": 21315 }, { "epoch": 0.8987077597655838, "grad_norm": 2.033932685852051, "learning_rate": 3.891361379181112e-05, "loss": 7.5844, "step": 21316 }, { "epoch": 0.8987499209477834, "grad_norm": 1.018492579460144, "learning_rate": 3.891283300540304e-05, "loss": 6.9036, "step": 21317 }, { "epoch": 0.898792082129983, "grad_norm": 2.0408921241760254, "learning_rate": 3.891205221899497e-05, "loss": 6.9801, "step": 21318 }, { "epoch": 0.8988342433121824, "grad_norm": 1.6224294900894165, "learning_rate": 3.8911271432586904e-05, "loss": 6.4429, "step": 21319 }, { "epoch": 0.898876404494382, "grad_norm": 1.0864754915237427, "learning_rate": 3.8910490646178835e-05, "loss": 6.6127, "step": 21320 }, { "epoch": 0.8989185656765816, "grad_norm": 4.533295631408691, "learning_rate": 3.890970985977077e-05, "loss": 8.1903, "step": 21321 }, { "epoch": 0.8989607268587811, "grad_norm": 2.367778778076172, "learning_rate": 3.890892907336269e-05, "loss": 7.1813, "step": 21322 }, { "epoch": 0.8990028880409807, "grad_norm": 2.295609712600708, "learning_rate": 3.890814828695462e-05, "loss": 7.4349, "step": 21323 }, { "epoch": 0.8990450492231802, "grad_norm": 3.079578161239624, "learning_rate": 3.8907367500546554e-05, "loss": 7.8394, "step": 21324 }, { "epoch": 0.8990872104053798, "grad_norm": 2.321676015853882, "learning_rate": 3.8906586714138485e-05, "loss": 6.3516, "step": 21325 }, { "epoch": 0.8991293715875793, "grad_norm": 1.924968957901001, "learning_rate": 3.890580592773041e-05, "loss": 6.9442, "step": 21326 }, { "epoch": 0.8991715327697789, "grad_norm": 2.7801101207733154, "learning_rate": 3.890502514132234e-05, "loss": 6.428, "step": 21327 }, { "epoch": 0.8992136939519784, "grad_norm": 4.561000823974609, "learning_rate": 3.890424435491427e-05, "loss": 7.2226, "step": 21328 }, { "epoch": 0.8992558551341779, "grad_norm": 1.1328386068344116, "learning_rate": 3.8903463568506196e-05, "loss": 6.7083, "step": 21329 }, { "epoch": 0.8992980163163775, "grad_norm": 2.148632049560547, "learning_rate": 3.8902682782098135e-05, "loss": 7.2532, "step": 21330 }, { "epoch": 0.8993401774985771, "grad_norm": 2.621041774749756, "learning_rate": 3.890190199569006e-05, "loss": 6.9641, "step": 21331 }, { "epoch": 0.8993823386807767, "grad_norm": 2.897554636001587, "learning_rate": 3.890112120928199e-05, "loss": 6.7991, "step": 21332 }, { "epoch": 0.8994244998629761, "grad_norm": 3.8287932872772217, "learning_rate": 3.890034042287392e-05, "loss": 8.3869, "step": 21333 }, { "epoch": 0.8994666610451757, "grad_norm": 3.9749035835266113, "learning_rate": 3.8899559636465846e-05, "loss": 7.5377, "step": 21334 }, { "epoch": 0.8995088222273753, "grad_norm": 1.6099770069122314, "learning_rate": 3.8898778850057784e-05, "loss": 7.0071, "step": 21335 }, { "epoch": 0.8995509834095748, "grad_norm": 1.7343828678131104, "learning_rate": 3.889799806364971e-05, "loss": 6.8974, "step": 21336 }, { "epoch": 0.8995931445917743, "grad_norm": 3.5037219524383545, "learning_rate": 3.889721727724164e-05, "loss": 9.5982, "step": 21337 }, { "epoch": 0.8996353057739739, "grad_norm": 3.674379587173462, "learning_rate": 3.889643649083357e-05, "loss": 6.7098, "step": 21338 }, { "epoch": 0.8996774669561735, "grad_norm": 2.358083724975586, "learning_rate": 3.8895655704425496e-05, "loss": 7.0134, "step": 21339 }, { "epoch": 0.899719628138373, "grad_norm": 4.2697601318359375, "learning_rate": 3.889487491801743e-05, "loss": 7.7492, "step": 21340 }, { "epoch": 0.8997617893205726, "grad_norm": 4.117320537567139, "learning_rate": 3.889409413160936e-05, "loss": 11.2013, "step": 21341 }, { "epoch": 0.8998039505027721, "grad_norm": 2.5674378871917725, "learning_rate": 3.889331334520129e-05, "loss": 6.7391, "step": 21342 }, { "epoch": 0.8998461116849716, "grad_norm": 2.6040124893188477, "learning_rate": 3.889253255879322e-05, "loss": 7.3411, "step": 21343 }, { "epoch": 0.8998882728671712, "grad_norm": 2.2611002922058105, "learning_rate": 3.889175177238515e-05, "loss": 7.4215, "step": 21344 }, { "epoch": 0.8999304340493708, "grad_norm": 2.652665615081787, "learning_rate": 3.8890970985977076e-05, "loss": 7.4148, "step": 21345 }, { "epoch": 0.8999725952315702, "grad_norm": 4.342426300048828, "learning_rate": 3.889019019956901e-05, "loss": 7.755, "step": 21346 }, { "epoch": 0.9000147564137698, "grad_norm": 2.7205018997192383, "learning_rate": 3.888940941316094e-05, "loss": 6.5892, "step": 21347 }, { "epoch": 0.9000569175959694, "grad_norm": 3.060903549194336, "learning_rate": 3.888862862675286e-05, "loss": 6.9352, "step": 21348 }, { "epoch": 0.900099078778169, "grad_norm": 2.338864326477051, "learning_rate": 3.88878478403448e-05, "loss": 6.6556, "step": 21349 }, { "epoch": 0.9001412399603684, "grad_norm": 1.2433066368103027, "learning_rate": 3.8887067053936726e-05, "loss": 6.8306, "step": 21350 }, { "epoch": 0.900183401142568, "grad_norm": 1.4714990854263306, "learning_rate": 3.888628626752866e-05, "loss": 7.0943, "step": 21351 }, { "epoch": 0.9002255623247676, "grad_norm": 1.7743269205093384, "learning_rate": 3.888550548112059e-05, "loss": 6.846, "step": 21352 }, { "epoch": 0.9002677235069672, "grad_norm": 3.1329193115234375, "learning_rate": 3.888472469471251e-05, "loss": 7.7583, "step": 21353 }, { "epoch": 0.9003098846891667, "grad_norm": 4.4212822914123535, "learning_rate": 3.888394390830445e-05, "loss": 7.3183, "step": 21354 }, { "epoch": 0.9003520458713662, "grad_norm": 2.7453179359436035, "learning_rate": 3.8883163121896375e-05, "loss": 6.5522, "step": 21355 }, { "epoch": 0.9003942070535658, "grad_norm": 2.3730216026306152, "learning_rate": 3.888238233548831e-05, "loss": 7.7391, "step": 21356 }, { "epoch": 0.9004363682357653, "grad_norm": 1.2307742834091187, "learning_rate": 3.888160154908024e-05, "loss": 6.568, "step": 21357 }, { "epoch": 0.9004785294179649, "grad_norm": 1.2055789232254028, "learning_rate": 3.888082076267216e-05, "loss": 6.9999, "step": 21358 }, { "epoch": 0.9005206906001644, "grad_norm": 1.467167615890503, "learning_rate": 3.8880039976264094e-05, "loss": 6.6791, "step": 21359 }, { "epoch": 0.900562851782364, "grad_norm": 4.534485340118408, "learning_rate": 3.8879259189856025e-05, "loss": 8.7743, "step": 21360 }, { "epoch": 0.9006050129645635, "grad_norm": 1.4037365913391113, "learning_rate": 3.8878478403447956e-05, "loss": 6.8309, "step": 21361 }, { "epoch": 0.9006471741467631, "grad_norm": 3.127981424331665, "learning_rate": 3.887769761703988e-05, "loss": 7.4447, "step": 21362 }, { "epoch": 0.9006893353289627, "grad_norm": 1.0603023767471313, "learning_rate": 3.887691683063182e-05, "loss": 6.7284, "step": 21363 }, { "epoch": 0.9007314965111621, "grad_norm": 2.6363210678100586, "learning_rate": 3.887613604422374e-05, "loss": 7.5253, "step": 21364 }, { "epoch": 0.9007736576933617, "grad_norm": 1.5528700351715088, "learning_rate": 3.8875355257815674e-05, "loss": 7.0473, "step": 21365 }, { "epoch": 0.9008158188755613, "grad_norm": 2.0224900245666504, "learning_rate": 3.8874574471407606e-05, "loss": 6.5387, "step": 21366 }, { "epoch": 0.9008579800577609, "grad_norm": 2.5804150104522705, "learning_rate": 3.887379368499953e-05, "loss": 6.7912, "step": 21367 }, { "epoch": 0.9009001412399603, "grad_norm": 2.4039456844329834, "learning_rate": 3.887301289859147e-05, "loss": 6.8616, "step": 21368 }, { "epoch": 0.9009423024221599, "grad_norm": 2.3975651264190674, "learning_rate": 3.887223211218339e-05, "loss": 6.4911, "step": 21369 }, { "epoch": 0.9009844636043595, "grad_norm": 2.0409951210021973, "learning_rate": 3.8871451325775324e-05, "loss": 6.6112, "step": 21370 }, { "epoch": 0.901026624786559, "grad_norm": 2.1994659900665283, "learning_rate": 3.8870670539367255e-05, "loss": 7.6347, "step": 21371 }, { "epoch": 0.9010687859687586, "grad_norm": 2.3817138671875, "learning_rate": 3.886988975295918e-05, "loss": 6.5907, "step": 21372 }, { "epoch": 0.9011109471509581, "grad_norm": 2.8851144313812256, "learning_rate": 3.886910896655111e-05, "loss": 7.3545, "step": 21373 }, { "epoch": 0.9011531083331576, "grad_norm": 2.882112979888916, "learning_rate": 3.886832818014304e-05, "loss": 6.9523, "step": 21374 }, { "epoch": 0.9011952695153572, "grad_norm": 1.6438068151474, "learning_rate": 3.8867547393734973e-05, "loss": 6.9523, "step": 21375 }, { "epoch": 0.9012374306975568, "grad_norm": 1.2983896732330322, "learning_rate": 3.88667666073269e-05, "loss": 7.0599, "step": 21376 }, { "epoch": 0.9012795918797563, "grad_norm": 1.6762820482254028, "learning_rate": 3.886598582091883e-05, "loss": 6.4, "step": 21377 }, { "epoch": 0.9013217530619558, "grad_norm": 1.308832049369812, "learning_rate": 3.886520503451076e-05, "loss": 6.3903, "step": 21378 }, { "epoch": 0.9013639142441554, "grad_norm": 2.2135002613067627, "learning_rate": 3.886442424810269e-05, "loss": 7.2389, "step": 21379 }, { "epoch": 0.901406075426355, "grad_norm": 1.9013783931732178, "learning_rate": 3.886364346169462e-05, "loss": 6.8821, "step": 21380 }, { "epoch": 0.9014482366085546, "grad_norm": 1.942330241203308, "learning_rate": 3.886286267528655e-05, "loss": 7.0116, "step": 21381 }, { "epoch": 0.901490397790754, "grad_norm": 2.5864124298095703, "learning_rate": 3.8862081888878485e-05, "loss": 7.3409, "step": 21382 }, { "epoch": 0.9015325589729536, "grad_norm": 1.2288720607757568, "learning_rate": 3.886130110247041e-05, "loss": 6.6131, "step": 21383 }, { "epoch": 0.9015747201551532, "grad_norm": 3.1287806034088135, "learning_rate": 3.8860520316062334e-05, "loss": 7.7104, "step": 21384 }, { "epoch": 0.9016168813373527, "grad_norm": 2.5754003524780273, "learning_rate": 3.885973952965427e-05, "loss": 6.3417, "step": 21385 }, { "epoch": 0.9016590425195522, "grad_norm": 2.3320775032043457, "learning_rate": 3.88589587432462e-05, "loss": 7.6133, "step": 21386 }, { "epoch": 0.9017012037017518, "grad_norm": 1.2822418212890625, "learning_rate": 3.885817795683813e-05, "loss": 6.7359, "step": 21387 }, { "epoch": 0.9017433648839513, "grad_norm": 1.52607262134552, "learning_rate": 3.885739717043006e-05, "loss": 6.5106, "step": 21388 }, { "epoch": 0.9017855260661509, "grad_norm": 1.4278876781463623, "learning_rate": 3.885661638402199e-05, "loss": 7.5626, "step": 21389 }, { "epoch": 0.9018276872483505, "grad_norm": 2.1512153148651123, "learning_rate": 3.885583559761392e-05, "loss": 7.1569, "step": 21390 }, { "epoch": 0.90186984843055, "grad_norm": 3.414592742919922, "learning_rate": 3.8855054811205847e-05, "loss": 7.4664, "step": 21391 }, { "epoch": 0.9019120096127495, "grad_norm": 1.6458096504211426, "learning_rate": 3.885427402479778e-05, "loss": 6.5001, "step": 21392 }, { "epoch": 0.9019541707949491, "grad_norm": 1.3628950119018555, "learning_rate": 3.885349323838971e-05, "loss": 6.4601, "step": 21393 }, { "epoch": 0.9019963319771487, "grad_norm": 2.031233310699463, "learning_rate": 3.885271245198164e-05, "loss": 7.3039, "step": 21394 }, { "epoch": 0.9020384931593481, "grad_norm": 1.2723052501678467, "learning_rate": 3.8851931665573565e-05, "loss": 6.4434, "step": 21395 }, { "epoch": 0.9020806543415477, "grad_norm": 2.8185949325561523, "learning_rate": 3.8851150879165496e-05, "loss": 6.6488, "step": 21396 }, { "epoch": 0.9021228155237473, "grad_norm": 1.3416659832000732, "learning_rate": 3.885037009275743e-05, "loss": 6.6232, "step": 21397 }, { "epoch": 0.9021649767059469, "grad_norm": 1.8783231973648071, "learning_rate": 3.884958930634935e-05, "loss": 7.1242, "step": 21398 }, { "epoch": 0.9022071378881464, "grad_norm": 1.7497212886810303, "learning_rate": 3.884880851994129e-05, "loss": 7.207, "step": 21399 }, { "epoch": 0.9022492990703459, "grad_norm": 3.080033302307129, "learning_rate": 3.8848027733533214e-05, "loss": 8.005, "step": 21400 }, { "epoch": 0.9022914602525455, "grad_norm": 1.9787002801895142, "learning_rate": 3.8847246947125146e-05, "loss": 6.6566, "step": 21401 }, { "epoch": 0.902333621434745, "grad_norm": 1.8363946676254272, "learning_rate": 3.884646616071708e-05, "loss": 7.0603, "step": 21402 }, { "epoch": 0.9023757826169446, "grad_norm": 2.340740919113159, "learning_rate": 3.8845685374309e-05, "loss": 6.8269, "step": 21403 }, { "epoch": 0.9024179437991441, "grad_norm": 2.755977153778076, "learning_rate": 3.884490458790094e-05, "loss": 6.5206, "step": 21404 }, { "epoch": 0.9024601049813437, "grad_norm": 1.0965806245803833, "learning_rate": 3.8844123801492864e-05, "loss": 6.5688, "step": 21405 }, { "epoch": 0.9025022661635432, "grad_norm": 1.167020559310913, "learning_rate": 3.8843343015084795e-05, "loss": 6.6491, "step": 21406 }, { "epoch": 0.9025444273457428, "grad_norm": 3.365504503250122, "learning_rate": 3.8842562228676726e-05, "loss": 8.1611, "step": 21407 }, { "epoch": 0.9025865885279423, "grad_norm": 2.931962251663208, "learning_rate": 3.884178144226866e-05, "loss": 7.5019, "step": 21408 }, { "epoch": 0.9026287497101418, "grad_norm": 1.9591715335845947, "learning_rate": 3.884100065586058e-05, "loss": 6.4339, "step": 21409 }, { "epoch": 0.9026709108923414, "grad_norm": 4.598937511444092, "learning_rate": 3.884021986945251e-05, "loss": 7.9156, "step": 21410 }, { "epoch": 0.902713072074541, "grad_norm": 2.149266242980957, "learning_rate": 3.8839439083044445e-05, "loss": 7.0655, "step": 21411 }, { "epoch": 0.9027552332567406, "grad_norm": 3.7240395545959473, "learning_rate": 3.8838658296636376e-05, "loss": 7.6107, "step": 21412 }, { "epoch": 0.90279739443894, "grad_norm": 1.6381895542144775, "learning_rate": 3.883787751022831e-05, "loss": 7.2612, "step": 21413 }, { "epoch": 0.9028395556211396, "grad_norm": 1.6446092128753662, "learning_rate": 3.883709672382023e-05, "loss": 7.131, "step": 21414 }, { "epoch": 0.9028817168033392, "grad_norm": 2.526750087738037, "learning_rate": 3.883631593741216e-05, "loss": 6.4702, "step": 21415 }, { "epoch": 0.9029238779855387, "grad_norm": 1.3605984449386597, "learning_rate": 3.8835535151004094e-05, "loss": 6.9033, "step": 21416 }, { "epoch": 0.9029660391677382, "grad_norm": 2.7199766635894775, "learning_rate": 3.883475436459602e-05, "loss": 7.6495, "step": 21417 }, { "epoch": 0.9030082003499378, "grad_norm": 1.4169132709503174, "learning_rate": 3.883397357818796e-05, "loss": 6.5084, "step": 21418 }, { "epoch": 0.9030503615321374, "grad_norm": 3.0662875175476074, "learning_rate": 3.883319279177988e-05, "loss": 7.3159, "step": 21419 }, { "epoch": 0.9030925227143369, "grad_norm": 1.1261790990829468, "learning_rate": 3.883241200537181e-05, "loss": 6.6154, "step": 21420 }, { "epoch": 0.9031346838965365, "grad_norm": 2.005443572998047, "learning_rate": 3.8831631218963744e-05, "loss": 7.3665, "step": 21421 }, { "epoch": 0.903176845078736, "grad_norm": 1.737684965133667, "learning_rate": 3.883085043255567e-05, "loss": 6.5923, "step": 21422 }, { "epoch": 0.9032190062609355, "grad_norm": 1.6830291748046875, "learning_rate": 3.8830069646147606e-05, "loss": 6.6479, "step": 21423 }, { "epoch": 0.9032611674431351, "grad_norm": 1.212565302848816, "learning_rate": 3.882928885973953e-05, "loss": 6.6248, "step": 21424 }, { "epoch": 0.9033033286253347, "grad_norm": 2.6938130855560303, "learning_rate": 3.882850807333146e-05, "loss": 7.6391, "step": 21425 }, { "epoch": 0.9033454898075342, "grad_norm": 2.976944923400879, "learning_rate": 3.882772728692339e-05, "loss": 6.3625, "step": 21426 }, { "epoch": 0.9033876509897337, "grad_norm": 1.5538524389266968, "learning_rate": 3.882694650051532e-05, "loss": 6.5331, "step": 21427 }, { "epoch": 0.9034298121719333, "grad_norm": 2.6720943450927734, "learning_rate": 3.882616571410725e-05, "loss": 6.5261, "step": 21428 }, { "epoch": 0.9034719733541329, "grad_norm": 1.4500709772109985, "learning_rate": 3.882538492769918e-05, "loss": 6.538, "step": 21429 }, { "epoch": 0.9035141345363324, "grad_norm": 3.141550064086914, "learning_rate": 3.882460414129111e-05, "loss": 7.0164, "step": 21430 }, { "epoch": 0.9035562957185319, "grad_norm": 1.762546420097351, "learning_rate": 3.8823823354883036e-05, "loss": 6.9254, "step": 21431 }, { "epoch": 0.9035984569007315, "grad_norm": 2.0239245891571045, "learning_rate": 3.8823042568474974e-05, "loss": 6.7424, "step": 21432 }, { "epoch": 0.9036406180829311, "grad_norm": 2.902322769165039, "learning_rate": 3.88222617820669e-05, "loss": 6.5316, "step": 21433 }, { "epoch": 0.9036827792651306, "grad_norm": 2.0364062786102295, "learning_rate": 3.882148099565883e-05, "loss": 7.1738, "step": 21434 }, { "epoch": 0.9037249404473301, "grad_norm": 1.2945085763931274, "learning_rate": 3.882070020925076e-05, "loss": 6.7262, "step": 21435 }, { "epoch": 0.9037671016295297, "grad_norm": 3.4588141441345215, "learning_rate": 3.8819919422842685e-05, "loss": 7.6214, "step": 21436 }, { "epoch": 0.9038092628117292, "grad_norm": 1.1070990562438965, "learning_rate": 3.8819138636434623e-05, "loss": 6.714, "step": 21437 }, { "epoch": 0.9038514239939288, "grad_norm": 2.821159601211548, "learning_rate": 3.881835785002655e-05, "loss": 6.9582, "step": 21438 }, { "epoch": 0.9038935851761284, "grad_norm": 1.8542871475219727, "learning_rate": 3.881757706361848e-05, "loss": 6.8647, "step": 21439 }, { "epoch": 0.9039357463583279, "grad_norm": 1.096420168876648, "learning_rate": 3.881679627721041e-05, "loss": 6.7821, "step": 21440 }, { "epoch": 0.9039779075405274, "grad_norm": 1.6378445625305176, "learning_rate": 3.8816015490802335e-05, "loss": 7.2317, "step": 21441 }, { "epoch": 0.904020068722727, "grad_norm": 1.2464720010757446, "learning_rate": 3.8815234704394266e-05, "loss": 6.5816, "step": 21442 }, { "epoch": 0.9040622299049266, "grad_norm": 2.977750062942505, "learning_rate": 3.88144539179862e-05, "loss": 7.4128, "step": 21443 }, { "epoch": 0.904104391087126, "grad_norm": 1.4412153959274292, "learning_rate": 3.881367313157813e-05, "loss": 6.9477, "step": 21444 }, { "epoch": 0.9041465522693256, "grad_norm": 2.334498882293701, "learning_rate": 3.881289234517006e-05, "loss": 7.8501, "step": 21445 }, { "epoch": 0.9041887134515252, "grad_norm": 1.5434041023254395, "learning_rate": 3.8812111558761984e-05, "loss": 6.8993, "step": 21446 }, { "epoch": 0.9042308746337248, "grad_norm": 1.721420407295227, "learning_rate": 3.8811330772353916e-05, "loss": 7.135, "step": 21447 }, { "epoch": 0.9042730358159243, "grad_norm": 2.549867630004883, "learning_rate": 3.881054998594585e-05, "loss": 7.7288, "step": 21448 }, { "epoch": 0.9043151969981238, "grad_norm": 2.1842665672302246, "learning_rate": 3.880976919953778e-05, "loss": 6.8464, "step": 21449 }, { "epoch": 0.9043573581803234, "grad_norm": 1.4039667844772339, "learning_rate": 3.88089884131297e-05, "loss": 6.6858, "step": 21450 }, { "epoch": 0.9043995193625229, "grad_norm": 2.147397756576538, "learning_rate": 3.880820762672164e-05, "loss": 7.6646, "step": 21451 }, { "epoch": 0.9044416805447225, "grad_norm": 2.7269060611724854, "learning_rate": 3.8807426840313565e-05, "loss": 7.6272, "step": 21452 }, { "epoch": 0.904483841726922, "grad_norm": 3.0016496181488037, "learning_rate": 3.8806646053905497e-05, "loss": 6.6388, "step": 21453 }, { "epoch": 0.9045260029091216, "grad_norm": 0.93194979429245, "learning_rate": 3.880586526749743e-05, "loss": 6.6135, "step": 21454 }, { "epoch": 0.9045681640913211, "grad_norm": 3.025470018386841, "learning_rate": 3.880508448108935e-05, "loss": 8.6828, "step": 21455 }, { "epoch": 0.9046103252735207, "grad_norm": 1.024487853050232, "learning_rate": 3.8804303694681284e-05, "loss": 6.9285, "step": 21456 }, { "epoch": 0.9046524864557203, "grad_norm": 2.7287890911102295, "learning_rate": 3.8803522908273215e-05, "loss": 7.0287, "step": 21457 }, { "epoch": 0.9046946476379197, "grad_norm": 1.612859845161438, "learning_rate": 3.8802742121865146e-05, "loss": 6.4649, "step": 21458 }, { "epoch": 0.9047368088201193, "grad_norm": 3.5679221153259277, "learning_rate": 3.880196133545708e-05, "loss": 6.9598, "step": 21459 }, { "epoch": 0.9047789700023189, "grad_norm": 2.4166924953460693, "learning_rate": 3.8801180549049e-05, "loss": 7.56, "step": 21460 }, { "epoch": 0.9048211311845185, "grad_norm": 2.2390055656433105, "learning_rate": 3.880039976264093e-05, "loss": 7.3834, "step": 21461 }, { "epoch": 0.9048632923667179, "grad_norm": 1.1577935218811035, "learning_rate": 3.8799618976232864e-05, "loss": 6.9269, "step": 21462 }, { "epoch": 0.9049054535489175, "grad_norm": 4.100604057312012, "learning_rate": 3.8798838189824796e-05, "loss": 7.9912, "step": 21463 }, { "epoch": 0.9049476147311171, "grad_norm": 2.331911087036133, "learning_rate": 3.879805740341672e-05, "loss": 6.8893, "step": 21464 }, { "epoch": 0.9049897759133166, "grad_norm": 2.616408348083496, "learning_rate": 3.879727661700865e-05, "loss": 7.4499, "step": 21465 }, { "epoch": 0.9050319370955161, "grad_norm": 1.8859845399856567, "learning_rate": 3.879649583060058e-05, "loss": 6.7366, "step": 21466 }, { "epoch": 0.9050740982777157, "grad_norm": 1.4039826393127441, "learning_rate": 3.879571504419251e-05, "loss": 6.4993, "step": 21467 }, { "epoch": 0.9051162594599153, "grad_norm": 2.056154251098633, "learning_rate": 3.8794934257784445e-05, "loss": 6.6669, "step": 21468 }, { "epoch": 0.9051584206421148, "grad_norm": 1.8171894550323486, "learning_rate": 3.879415347137637e-05, "loss": 6.5365, "step": 21469 }, { "epoch": 0.9052005818243144, "grad_norm": 1.7362933158874512, "learning_rate": 3.879337268496831e-05, "loss": 6.8675, "step": 21470 }, { "epoch": 0.9052427430065139, "grad_norm": 2.1600685119628906, "learning_rate": 3.879259189856023e-05, "loss": 6.7015, "step": 21471 }, { "epoch": 0.9052849041887134, "grad_norm": 1.0793606042861938, "learning_rate": 3.8791811112152157e-05, "loss": 6.6594, "step": 21472 }, { "epoch": 0.905327065370913, "grad_norm": 2.5731356143951416, "learning_rate": 3.8791030325744095e-05, "loss": 7.6251, "step": 21473 }, { "epoch": 0.9053692265531126, "grad_norm": 1.7437232732772827, "learning_rate": 3.879024953933602e-05, "loss": 7.2213, "step": 21474 }, { "epoch": 0.905411387735312, "grad_norm": 1.521295428276062, "learning_rate": 3.878946875292795e-05, "loss": 6.84, "step": 21475 }, { "epoch": 0.9054535489175116, "grad_norm": 1.9519035816192627, "learning_rate": 3.878868796651988e-05, "loss": 7.0683, "step": 21476 }, { "epoch": 0.9054957100997112, "grad_norm": 1.0854276418685913, "learning_rate": 3.878790718011181e-05, "loss": 6.6031, "step": 21477 }, { "epoch": 0.9055378712819108, "grad_norm": 0.970849871635437, "learning_rate": 3.878712639370374e-05, "loss": 6.5671, "step": 21478 }, { "epoch": 0.9055800324641103, "grad_norm": 2.800020694732666, "learning_rate": 3.878634560729567e-05, "loss": 7.6797, "step": 21479 }, { "epoch": 0.9056221936463098, "grad_norm": 4.054070472717285, "learning_rate": 3.87855648208876e-05, "loss": 7.2966, "step": 21480 }, { "epoch": 0.9056643548285094, "grad_norm": 1.3293896913528442, "learning_rate": 3.878478403447953e-05, "loss": 6.6599, "step": 21481 }, { "epoch": 0.905706516010709, "grad_norm": 2.0107429027557373, "learning_rate": 3.878400324807146e-05, "loss": 7.401, "step": 21482 }, { "epoch": 0.9057486771929085, "grad_norm": 1.8400408029556274, "learning_rate": 3.878322246166339e-05, "loss": 7.0965, "step": 21483 }, { "epoch": 0.905790838375108, "grad_norm": 1.2680250406265259, "learning_rate": 3.878244167525532e-05, "loss": 7.032, "step": 21484 }, { "epoch": 0.9058329995573076, "grad_norm": 1.7965177297592163, "learning_rate": 3.878166088884725e-05, "loss": 6.819, "step": 21485 }, { "epoch": 0.9058751607395071, "grad_norm": 2.280296802520752, "learning_rate": 3.8780880102439174e-05, "loss": 6.4968, "step": 21486 }, { "epoch": 0.9059173219217067, "grad_norm": 2.972080945968628, "learning_rate": 3.878009931603111e-05, "loss": 9.2557, "step": 21487 }, { "epoch": 0.9059594831039063, "grad_norm": 2.8366613388061523, "learning_rate": 3.8779318529623036e-05, "loss": 6.5738, "step": 21488 }, { "epoch": 0.9060016442861057, "grad_norm": 1.4681974649429321, "learning_rate": 3.877853774321497e-05, "loss": 6.7442, "step": 21489 }, { "epoch": 0.9060438054683053, "grad_norm": 2.93340802192688, "learning_rate": 3.87777569568069e-05, "loss": 6.3983, "step": 21490 }, { "epoch": 0.9060859666505049, "grad_norm": 2.918344497680664, "learning_rate": 3.8776976170398823e-05, "loss": 7.5643, "step": 21491 }, { "epoch": 0.9061281278327045, "grad_norm": 1.1279804706573486, "learning_rate": 3.877619538399076e-05, "loss": 6.5693, "step": 21492 }, { "epoch": 0.9061702890149039, "grad_norm": 3.133471727371216, "learning_rate": 3.8775414597582686e-05, "loss": 7.4949, "step": 21493 }, { "epoch": 0.9062124501971035, "grad_norm": 3.140951633453369, "learning_rate": 3.877463381117462e-05, "loss": 7.7237, "step": 21494 }, { "epoch": 0.9062546113793031, "grad_norm": 1.8861558437347412, "learning_rate": 3.877385302476655e-05, "loss": 7.1013, "step": 21495 }, { "epoch": 0.9062967725615027, "grad_norm": 1.466921091079712, "learning_rate": 3.877307223835848e-05, "loss": 6.8096, "step": 21496 }, { "epoch": 0.9063389337437022, "grad_norm": 1.1978660821914673, "learning_rate": 3.8772291451950404e-05, "loss": 7.0643, "step": 21497 }, { "epoch": 0.9063810949259017, "grad_norm": 1.227279782295227, "learning_rate": 3.8771510665542335e-05, "loss": 6.5496, "step": 21498 }, { "epoch": 0.9064232561081013, "grad_norm": 1.5116633176803589, "learning_rate": 3.877072987913427e-05, "loss": 6.8256, "step": 21499 }, { "epoch": 0.9064654172903008, "grad_norm": 1.4789694547653198, "learning_rate": 3.876994909272619e-05, "loss": 6.5194, "step": 21500 }, { "epoch": 0.9065075784725004, "grad_norm": 1.936719536781311, "learning_rate": 3.876916830631813e-05, "loss": 6.8144, "step": 21501 }, { "epoch": 0.9065497396546999, "grad_norm": 1.2606024742126465, "learning_rate": 3.8768387519910054e-05, "loss": 6.7277, "step": 21502 }, { "epoch": 0.9065919008368994, "grad_norm": 2.762362480163574, "learning_rate": 3.8767606733501985e-05, "loss": 7.6754, "step": 21503 }, { "epoch": 0.906634062019099, "grad_norm": 2.224212169647217, "learning_rate": 3.8766825947093916e-05, "loss": 6.5367, "step": 21504 }, { "epoch": 0.9066762232012986, "grad_norm": 4.5630621910095215, "learning_rate": 3.876604516068584e-05, "loss": 7.4578, "step": 21505 }, { "epoch": 0.9067183843834982, "grad_norm": 1.9813984632492065, "learning_rate": 3.876526437427778e-05, "loss": 6.5369, "step": 21506 }, { "epoch": 0.9067605455656976, "grad_norm": 1.7794407606124878, "learning_rate": 3.87644835878697e-05, "loss": 7.1273, "step": 21507 }, { "epoch": 0.9068027067478972, "grad_norm": 1.8725584745407104, "learning_rate": 3.8763702801461635e-05, "loss": 7.5018, "step": 21508 }, { "epoch": 0.9068448679300968, "grad_norm": 2.8861076831817627, "learning_rate": 3.8762922015053566e-05, "loss": 7.7099, "step": 21509 }, { "epoch": 0.9068870291122964, "grad_norm": 2.728386402130127, "learning_rate": 3.876214122864549e-05, "loss": 8.006, "step": 21510 }, { "epoch": 0.9069291902944958, "grad_norm": 1.3116466999053955, "learning_rate": 3.876136044223742e-05, "loss": 6.3906, "step": 21511 }, { "epoch": 0.9069713514766954, "grad_norm": 1.5088884830474854, "learning_rate": 3.876057965582935e-05, "loss": 6.6598, "step": 21512 }, { "epoch": 0.907013512658895, "grad_norm": 2.407335042953491, "learning_rate": 3.8759798869421284e-05, "loss": 7.3169, "step": 21513 }, { "epoch": 0.9070556738410945, "grad_norm": 1.7429215908050537, "learning_rate": 3.8759018083013215e-05, "loss": 7.0575, "step": 21514 }, { "epoch": 0.9070978350232941, "grad_norm": 2.2302730083465576, "learning_rate": 3.8758237296605147e-05, "loss": 6.4478, "step": 21515 }, { "epoch": 0.9071399962054936, "grad_norm": 3.6396801471710205, "learning_rate": 3.875745651019707e-05, "loss": 7.7854, "step": 21516 }, { "epoch": 0.9071821573876931, "grad_norm": 4.314598083496094, "learning_rate": 3.8756675723789e-05, "loss": 7.4061, "step": 21517 }, { "epoch": 0.9072243185698927, "grad_norm": 1.9382882118225098, "learning_rate": 3.8755894937380934e-05, "loss": 6.6665, "step": 21518 }, { "epoch": 0.9072664797520923, "grad_norm": 1.790086269378662, "learning_rate": 3.875511415097286e-05, "loss": 7.2448, "step": 21519 }, { "epoch": 0.9073086409342918, "grad_norm": 2.8180675506591797, "learning_rate": 3.8754333364564796e-05, "loss": 6.6796, "step": 21520 }, { "epoch": 0.9073508021164913, "grad_norm": 2.058274507522583, "learning_rate": 3.875355257815672e-05, "loss": 6.9651, "step": 21521 }, { "epoch": 0.9073929632986909, "grad_norm": 1.8260048627853394, "learning_rate": 3.875277179174865e-05, "loss": 6.8017, "step": 21522 }, { "epoch": 0.9074351244808905, "grad_norm": 1.5706015825271606, "learning_rate": 3.875199100534058e-05, "loss": 6.5444, "step": 21523 }, { "epoch": 0.9074772856630899, "grad_norm": 1.5732791423797607, "learning_rate": 3.875121021893251e-05, "loss": 7.012, "step": 21524 }, { "epoch": 0.9075194468452895, "grad_norm": 1.1626371145248413, "learning_rate": 3.8750429432524446e-05, "loss": 6.5407, "step": 21525 }, { "epoch": 0.9075616080274891, "grad_norm": 0.81559818983078, "learning_rate": 3.874964864611637e-05, "loss": 6.8995, "step": 21526 }, { "epoch": 0.9076037692096887, "grad_norm": 1.2550560235977173, "learning_rate": 3.87488678597083e-05, "loss": 7.2396, "step": 21527 }, { "epoch": 0.9076459303918882, "grad_norm": 1.7391574382781982, "learning_rate": 3.874808707330023e-05, "loss": 6.4508, "step": 21528 }, { "epoch": 0.9076880915740877, "grad_norm": 2.256391763687134, "learning_rate": 3.874730628689216e-05, "loss": 6.3829, "step": 21529 }, { "epoch": 0.9077302527562873, "grad_norm": 0.9249299168586731, "learning_rate": 3.874652550048409e-05, "loss": 6.3443, "step": 21530 }, { "epoch": 0.9077724139384868, "grad_norm": 3.306314468383789, "learning_rate": 3.874574471407602e-05, "loss": 6.4633, "step": 21531 }, { "epoch": 0.9078145751206864, "grad_norm": 1.6525918245315552, "learning_rate": 3.874496392766795e-05, "loss": 6.6716, "step": 21532 }, { "epoch": 0.9078567363028859, "grad_norm": 2.0776469707489014, "learning_rate": 3.8744183141259875e-05, "loss": 6.9925, "step": 21533 }, { "epoch": 0.9078988974850855, "grad_norm": 2.4662859439849854, "learning_rate": 3.8743402354851807e-05, "loss": 7.0977, "step": 21534 }, { "epoch": 0.907941058667285, "grad_norm": 1.3991676568984985, "learning_rate": 3.874262156844374e-05, "loss": 6.5055, "step": 21535 }, { "epoch": 0.9079832198494846, "grad_norm": 2.428431749343872, "learning_rate": 3.874184078203566e-05, "loss": 7.8782, "step": 21536 }, { "epoch": 0.9080253810316842, "grad_norm": 3.8766932487487793, "learning_rate": 3.87410599956276e-05, "loss": 7.8676, "step": 21537 }, { "epoch": 0.9080675422138836, "grad_norm": 2.2746846675872803, "learning_rate": 3.8740279209219525e-05, "loss": 6.6697, "step": 21538 }, { "epoch": 0.9081097033960832, "grad_norm": 1.547301173210144, "learning_rate": 3.873949842281146e-05, "loss": 7.1045, "step": 21539 }, { "epoch": 0.9081518645782828, "grad_norm": 2.2713749408721924, "learning_rate": 3.873871763640339e-05, "loss": 6.6181, "step": 21540 }, { "epoch": 0.9081940257604824, "grad_norm": 1.7651848793029785, "learning_rate": 3.873793684999532e-05, "loss": 7.2647, "step": 21541 }, { "epoch": 0.9082361869426818, "grad_norm": 1.1751075983047485, "learning_rate": 3.873715606358725e-05, "loss": 6.8787, "step": 21542 }, { "epoch": 0.9082783481248814, "grad_norm": 0.7584761381149292, "learning_rate": 3.8736375277179174e-05, "loss": 6.5459, "step": 21543 }, { "epoch": 0.908320509307081, "grad_norm": 1.5856176614761353, "learning_rate": 3.8735594490771106e-05, "loss": 7.1478, "step": 21544 }, { "epoch": 0.9083626704892805, "grad_norm": 1.2830381393432617, "learning_rate": 3.873481370436304e-05, "loss": 7.09, "step": 21545 }, { "epoch": 0.9084048316714801, "grad_norm": 1.1741868257522583, "learning_rate": 3.873403291795497e-05, "loss": 6.937, "step": 21546 }, { "epoch": 0.9084469928536796, "grad_norm": 3.381620168685913, "learning_rate": 3.873325213154689e-05, "loss": 8.1766, "step": 21547 }, { "epoch": 0.9084891540358792, "grad_norm": 1.6574203968048096, "learning_rate": 3.8732471345138824e-05, "loss": 6.4945, "step": 21548 }, { "epoch": 0.9085313152180787, "grad_norm": 1.4573014974594116, "learning_rate": 3.8731690558730755e-05, "loss": 6.5354, "step": 21549 }, { "epoch": 0.9085734764002783, "grad_norm": 1.872498869895935, "learning_rate": 3.8730909772322686e-05, "loss": 6.4615, "step": 21550 }, { "epoch": 0.9086156375824778, "grad_norm": 2.9421818256378174, "learning_rate": 3.873012898591462e-05, "loss": 6.9941, "step": 21551 }, { "epoch": 0.9086577987646773, "grad_norm": 1.1626405715942383, "learning_rate": 3.872934819950654e-05, "loss": 6.5156, "step": 21552 }, { "epoch": 0.9086999599468769, "grad_norm": 1.5758217573165894, "learning_rate": 3.8728567413098473e-05, "loss": 6.5684, "step": 21553 }, { "epoch": 0.9087421211290765, "grad_norm": 2.485668659210205, "learning_rate": 3.8727786626690405e-05, "loss": 7.003, "step": 21554 }, { "epoch": 0.9087842823112761, "grad_norm": 1.9267988204956055, "learning_rate": 3.872700584028233e-05, "loss": 7.0035, "step": 21555 }, { "epoch": 0.9088264434934755, "grad_norm": 1.3379825353622437, "learning_rate": 3.872622505387427e-05, "loss": 6.4001, "step": 21556 }, { "epoch": 0.9088686046756751, "grad_norm": 1.5601468086242676, "learning_rate": 3.872544426746619e-05, "loss": 6.3492, "step": 21557 }, { "epoch": 0.9089107658578747, "grad_norm": 1.468981146812439, "learning_rate": 3.872466348105812e-05, "loss": 6.6645, "step": 21558 }, { "epoch": 0.9089529270400742, "grad_norm": 1.4415754079818726, "learning_rate": 3.8723882694650054e-05, "loss": 6.9399, "step": 21559 }, { "epoch": 0.9089950882222737, "grad_norm": 3.724226236343384, "learning_rate": 3.8723101908241985e-05, "loss": 8.1895, "step": 21560 }, { "epoch": 0.9090372494044733, "grad_norm": 1.2254379987716675, "learning_rate": 3.872232112183392e-05, "loss": 6.9142, "step": 21561 }, { "epoch": 0.9090794105866729, "grad_norm": 2.0613322257995605, "learning_rate": 3.872154033542584e-05, "loss": 7.3912, "step": 21562 }, { "epoch": 0.9091215717688724, "grad_norm": 0.9541540741920471, "learning_rate": 3.872075954901777e-05, "loss": 6.7957, "step": 21563 }, { "epoch": 0.909163732951072, "grad_norm": 2.8397560119628906, "learning_rate": 3.8719978762609704e-05, "loss": 7.8582, "step": 21564 }, { "epoch": 0.9092058941332715, "grad_norm": 1.434647560119629, "learning_rate": 3.8719197976201635e-05, "loss": 6.8996, "step": 21565 }, { "epoch": 0.909248055315471, "grad_norm": 1.1981892585754395, "learning_rate": 3.871841718979356e-05, "loss": 6.7072, "step": 21566 }, { "epoch": 0.9092902164976706, "grad_norm": 2.1295392513275146, "learning_rate": 3.871763640338549e-05, "loss": 7.1319, "step": 21567 }, { "epoch": 0.9093323776798702, "grad_norm": 3.3326644897460938, "learning_rate": 3.871685561697742e-05, "loss": 8.0893, "step": 21568 }, { "epoch": 0.9093745388620696, "grad_norm": 2.7203667163848877, "learning_rate": 3.8716074830569346e-05, "loss": 8.5944, "step": 21569 }, { "epoch": 0.9094167000442692, "grad_norm": 1.039757490158081, "learning_rate": 3.8715294044161285e-05, "loss": 6.5647, "step": 21570 }, { "epoch": 0.9094588612264688, "grad_norm": 1.2388173341751099, "learning_rate": 3.871451325775321e-05, "loss": 6.5782, "step": 21571 }, { "epoch": 0.9095010224086684, "grad_norm": 2.2744204998016357, "learning_rate": 3.871373247134514e-05, "loss": 6.4571, "step": 21572 }, { "epoch": 0.9095431835908679, "grad_norm": 2.2212514877319336, "learning_rate": 3.871295168493707e-05, "loss": 6.9796, "step": 21573 }, { "epoch": 0.9095853447730674, "grad_norm": 1.598610758781433, "learning_rate": 3.8712170898528996e-05, "loss": 6.8259, "step": 21574 }, { "epoch": 0.909627505955267, "grad_norm": 1.7611113786697388, "learning_rate": 3.8711390112120934e-05, "loss": 6.8566, "step": 21575 }, { "epoch": 0.9096696671374666, "grad_norm": 2.120136022567749, "learning_rate": 3.871060932571286e-05, "loss": 7.0865, "step": 21576 }, { "epoch": 0.9097118283196661, "grad_norm": 1.8355166912078857, "learning_rate": 3.870982853930479e-05, "loss": 7.2584, "step": 21577 }, { "epoch": 0.9097539895018656, "grad_norm": 3.307790517807007, "learning_rate": 3.870904775289672e-05, "loss": 7.3824, "step": 21578 }, { "epoch": 0.9097961506840652, "grad_norm": 1.85451340675354, "learning_rate": 3.8708266966488646e-05, "loss": 7.1811, "step": 21579 }, { "epoch": 0.9098383118662647, "grad_norm": 2.737252950668335, "learning_rate": 3.870748618008058e-05, "loss": 6.4578, "step": 21580 }, { "epoch": 0.9098804730484643, "grad_norm": 2.0175387859344482, "learning_rate": 3.870670539367251e-05, "loss": 7.3249, "step": 21581 }, { "epoch": 0.9099226342306638, "grad_norm": 1.6518704891204834, "learning_rate": 3.870592460726444e-05, "loss": 6.6026, "step": 21582 }, { "epoch": 0.9099647954128633, "grad_norm": 3.3967270851135254, "learning_rate": 3.870514382085637e-05, "loss": 7.7192, "step": 21583 }, { "epoch": 0.9100069565950629, "grad_norm": 1.429630994796753, "learning_rate": 3.87043630344483e-05, "loss": 6.5564, "step": 21584 }, { "epoch": 0.9100491177772625, "grad_norm": 2.6356394290924072, "learning_rate": 3.8703582248040226e-05, "loss": 6.7348, "step": 21585 }, { "epoch": 0.9100912789594621, "grad_norm": 2.7587056159973145, "learning_rate": 3.870280146163216e-05, "loss": 7.5882, "step": 21586 }, { "epoch": 0.9101334401416615, "grad_norm": 2.6319520473480225, "learning_rate": 3.870202067522409e-05, "loss": 7.7486, "step": 21587 }, { "epoch": 0.9101756013238611, "grad_norm": 1.4387180805206299, "learning_rate": 3.870123988881601e-05, "loss": 6.5015, "step": 21588 }, { "epoch": 0.9102177625060607, "grad_norm": 1.1006778478622437, "learning_rate": 3.870045910240795e-05, "loss": 7.2484, "step": 21589 }, { "epoch": 0.9102599236882603, "grad_norm": 2.5951762199401855, "learning_rate": 3.8699678315999876e-05, "loss": 6.5675, "step": 21590 }, { "epoch": 0.9103020848704597, "grad_norm": 2.1999382972717285, "learning_rate": 3.869889752959181e-05, "loss": 7.3074, "step": 21591 }, { "epoch": 0.9103442460526593, "grad_norm": 1.4190655946731567, "learning_rate": 3.869811674318374e-05, "loss": 7.2259, "step": 21592 }, { "epoch": 0.9103864072348589, "grad_norm": 1.4444164037704468, "learning_rate": 3.869733595677566e-05, "loss": 6.5384, "step": 21593 }, { "epoch": 0.9104285684170584, "grad_norm": 2.0206775665283203, "learning_rate": 3.86965551703676e-05, "loss": 7.252, "step": 21594 }, { "epoch": 0.910470729599258, "grad_norm": 1.2149944305419922, "learning_rate": 3.8695774383959525e-05, "loss": 6.4042, "step": 21595 }, { "epoch": 0.9105128907814575, "grad_norm": 1.6889010667800903, "learning_rate": 3.869499359755146e-05, "loss": 6.7082, "step": 21596 }, { "epoch": 0.910555051963657, "grad_norm": 4.132036209106445, "learning_rate": 3.869421281114339e-05, "loss": 7.6037, "step": 21597 }, { "epoch": 0.9105972131458566, "grad_norm": 1.2682124376296997, "learning_rate": 3.869343202473531e-05, "loss": 6.4582, "step": 21598 }, { "epoch": 0.9106393743280562, "grad_norm": 2.2823548316955566, "learning_rate": 3.8692651238327244e-05, "loss": 7.0975, "step": 21599 }, { "epoch": 0.9106815355102557, "grad_norm": 1.9218050241470337, "learning_rate": 3.8691870451919175e-05, "loss": 6.93, "step": 21600 }, { "epoch": 0.9107236966924552, "grad_norm": 1.6228781938552856, "learning_rate": 3.8691089665511106e-05, "loss": 6.5415, "step": 21601 }, { "epoch": 0.9107658578746548, "grad_norm": 1.5234624147415161, "learning_rate": 3.869030887910303e-05, "loss": 6.8625, "step": 21602 }, { "epoch": 0.9108080190568544, "grad_norm": 2.5068812370300293, "learning_rate": 3.868952809269497e-05, "loss": 6.5406, "step": 21603 }, { "epoch": 0.910850180239054, "grad_norm": 2.1153790950775146, "learning_rate": 3.868874730628689e-05, "loss": 6.2648, "step": 21604 }, { "epoch": 0.9108923414212534, "grad_norm": 2.4193129539489746, "learning_rate": 3.8687966519878824e-05, "loss": 6.5449, "step": 21605 }, { "epoch": 0.910934502603453, "grad_norm": 2.475950002670288, "learning_rate": 3.8687185733470756e-05, "loss": 7.2543, "step": 21606 }, { "epoch": 0.9109766637856526, "grad_norm": 1.0287976264953613, "learning_rate": 3.868640494706268e-05, "loss": 6.8548, "step": 21607 }, { "epoch": 0.9110188249678521, "grad_norm": 2.1498525142669678, "learning_rate": 3.868562416065462e-05, "loss": 6.9099, "step": 21608 }, { "epoch": 0.9110609861500516, "grad_norm": 0.7597833275794983, "learning_rate": 3.868484337424654e-05, "loss": 6.4405, "step": 21609 }, { "epoch": 0.9111031473322512, "grad_norm": 3.3281307220458984, "learning_rate": 3.8684062587838474e-05, "loss": 7.9073, "step": 21610 }, { "epoch": 0.9111453085144507, "grad_norm": 1.5787339210510254, "learning_rate": 3.8683281801430405e-05, "loss": 6.5504, "step": 21611 }, { "epoch": 0.9111874696966503, "grad_norm": 1.5569647550582886, "learning_rate": 3.868250101502233e-05, "loss": 6.5816, "step": 21612 }, { "epoch": 0.9112296308788499, "grad_norm": 4.524538993835449, "learning_rate": 3.868172022861426e-05, "loss": 7.8683, "step": 21613 }, { "epoch": 0.9112717920610494, "grad_norm": 1.2686095237731934, "learning_rate": 3.868093944220619e-05, "loss": 6.4582, "step": 21614 }, { "epoch": 0.9113139532432489, "grad_norm": 1.2459198236465454, "learning_rate": 3.8680158655798123e-05, "loss": 6.4062, "step": 21615 }, { "epoch": 0.9113561144254485, "grad_norm": 1.334520936012268, "learning_rate": 3.867937786939005e-05, "loss": 6.7631, "step": 21616 }, { "epoch": 0.9113982756076481, "grad_norm": 1.2989380359649658, "learning_rate": 3.867859708298198e-05, "loss": 6.6961, "step": 21617 }, { "epoch": 0.9114404367898475, "grad_norm": 1.2762681245803833, "learning_rate": 3.867781629657391e-05, "loss": 6.8465, "step": 21618 }, { "epoch": 0.9114825979720471, "grad_norm": 3.20997953414917, "learning_rate": 3.867703551016584e-05, "loss": 7.7052, "step": 21619 }, { "epoch": 0.9115247591542467, "grad_norm": 2.0550100803375244, "learning_rate": 3.867625472375777e-05, "loss": 7.5858, "step": 21620 }, { "epoch": 0.9115669203364463, "grad_norm": 2.7213046550750732, "learning_rate": 3.86754739373497e-05, "loss": 7.7573, "step": 21621 }, { "epoch": 0.9116090815186458, "grad_norm": 2.444284200668335, "learning_rate": 3.8674693150941635e-05, "loss": 8.901, "step": 21622 }, { "epoch": 0.9116512427008453, "grad_norm": 2.6574394702911377, "learning_rate": 3.867391236453356e-05, "loss": 7.5302, "step": 21623 }, { "epoch": 0.9116934038830449, "grad_norm": 3.018742799758911, "learning_rate": 3.8673131578125484e-05, "loss": 7.6682, "step": 21624 }, { "epoch": 0.9117355650652444, "grad_norm": 1.4630545377731323, "learning_rate": 3.867235079171742e-05, "loss": 6.6038, "step": 21625 }, { "epoch": 0.911777726247444, "grad_norm": 1.1549484729766846, "learning_rate": 3.867157000530935e-05, "loss": 6.5746, "step": 21626 }, { "epoch": 0.9118198874296435, "grad_norm": 3.474107027053833, "learning_rate": 3.867078921890128e-05, "loss": 7.525, "step": 21627 }, { "epoch": 0.9118620486118431, "grad_norm": 3.7139999866485596, "learning_rate": 3.867000843249321e-05, "loss": 7.7063, "step": 21628 }, { "epoch": 0.9119042097940426, "grad_norm": 2.5348715782165527, "learning_rate": 3.866922764608514e-05, "loss": 7.7664, "step": 21629 }, { "epoch": 0.9119463709762422, "grad_norm": 1.3675616979599, "learning_rate": 3.866844685967707e-05, "loss": 7.4029, "step": 21630 }, { "epoch": 0.9119885321584418, "grad_norm": 2.032743453979492, "learning_rate": 3.8667666073268997e-05, "loss": 7.6888, "step": 21631 }, { "epoch": 0.9120306933406412, "grad_norm": 2.4046733379364014, "learning_rate": 3.866688528686093e-05, "loss": 6.8833, "step": 21632 }, { "epoch": 0.9120728545228408, "grad_norm": 2.7292003631591797, "learning_rate": 3.866610450045286e-05, "loss": 7.7734, "step": 21633 }, { "epoch": 0.9121150157050404, "grad_norm": 1.3582075834274292, "learning_rate": 3.866532371404479e-05, "loss": 6.9496, "step": 21634 }, { "epoch": 0.91215717688724, "grad_norm": 1.6948413848876953, "learning_rate": 3.8664542927636715e-05, "loss": 6.9951, "step": 21635 }, { "epoch": 0.9121993380694394, "grad_norm": 2.978180170059204, "learning_rate": 3.8663762141228646e-05, "loss": 7.5465, "step": 21636 }, { "epoch": 0.912241499251639, "grad_norm": 1.3510652780532837, "learning_rate": 3.866298135482058e-05, "loss": 6.5582, "step": 21637 }, { "epoch": 0.9122836604338386, "grad_norm": 1.5443806648254395, "learning_rate": 3.86622005684125e-05, "loss": 7.0628, "step": 21638 }, { "epoch": 0.9123258216160381, "grad_norm": 3.421211004257202, "learning_rate": 3.866141978200444e-05, "loss": 7.6878, "step": 21639 }, { "epoch": 0.9123679827982376, "grad_norm": 1.3190302848815918, "learning_rate": 3.8660638995596364e-05, "loss": 6.7725, "step": 21640 }, { "epoch": 0.9124101439804372, "grad_norm": 1.6748192310333252, "learning_rate": 3.8659858209188296e-05, "loss": 6.5366, "step": 21641 }, { "epoch": 0.9124523051626368, "grad_norm": 1.6277433633804321, "learning_rate": 3.865907742278023e-05, "loss": 6.447, "step": 21642 }, { "epoch": 0.9124944663448363, "grad_norm": 2.7613141536712646, "learning_rate": 3.865829663637215e-05, "loss": 8.2339, "step": 21643 }, { "epoch": 0.9125366275270359, "grad_norm": 1.3331412076950073, "learning_rate": 3.865751584996409e-05, "loss": 7.248, "step": 21644 }, { "epoch": 0.9125787887092354, "grad_norm": 1.2621084451675415, "learning_rate": 3.8656735063556014e-05, "loss": 6.9459, "step": 21645 }, { "epoch": 0.9126209498914349, "grad_norm": 1.1354905366897583, "learning_rate": 3.8655954277147945e-05, "loss": 6.522, "step": 21646 }, { "epoch": 0.9126631110736345, "grad_norm": 1.4949380159378052, "learning_rate": 3.8655173490739876e-05, "loss": 6.7652, "step": 21647 }, { "epoch": 0.9127052722558341, "grad_norm": 2.43498158454895, "learning_rate": 3.865439270433181e-05, "loss": 6.5611, "step": 21648 }, { "epoch": 0.9127474334380336, "grad_norm": 2.983016014099121, "learning_rate": 3.865361191792373e-05, "loss": 7.6741, "step": 21649 }, { "epoch": 0.9127895946202331, "grad_norm": 4.867161750793457, "learning_rate": 3.865283113151566e-05, "loss": 10.6666, "step": 21650 }, { "epoch": 0.9128317558024327, "grad_norm": 1.6021126508712769, "learning_rate": 3.8652050345107595e-05, "loss": 6.5311, "step": 21651 }, { "epoch": 0.9128739169846323, "grad_norm": 1.7660919427871704, "learning_rate": 3.8651269558699526e-05, "loss": 6.4422, "step": 21652 }, { "epoch": 0.9129160781668318, "grad_norm": 3.2250804901123047, "learning_rate": 3.865048877229146e-05, "loss": 7.0984, "step": 21653 }, { "epoch": 0.9129582393490313, "grad_norm": 2.4759702682495117, "learning_rate": 3.864970798588338e-05, "loss": 7.5312, "step": 21654 }, { "epoch": 0.9130004005312309, "grad_norm": 1.290359616279602, "learning_rate": 3.864892719947531e-05, "loss": 6.5302, "step": 21655 }, { "epoch": 0.9130425617134305, "grad_norm": 2.4327850341796875, "learning_rate": 3.8648146413067244e-05, "loss": 6.5658, "step": 21656 }, { "epoch": 0.91308472289563, "grad_norm": 1.6029022932052612, "learning_rate": 3.864736562665917e-05, "loss": 7.0674, "step": 21657 }, { "epoch": 0.9131268840778295, "grad_norm": 3.75410532951355, "learning_rate": 3.864658484025111e-05, "loss": 7.6985, "step": 21658 }, { "epoch": 0.9131690452600291, "grad_norm": 1.6679457426071167, "learning_rate": 3.864580405384303e-05, "loss": 6.7671, "step": 21659 }, { "epoch": 0.9132112064422286, "grad_norm": 1.0757322311401367, "learning_rate": 3.864502326743496e-05, "loss": 6.5047, "step": 21660 }, { "epoch": 0.9132533676244282, "grad_norm": 1.4195024967193604, "learning_rate": 3.8644242481026894e-05, "loss": 6.6497, "step": 21661 }, { "epoch": 0.9132955288066278, "grad_norm": 1.7600008249282837, "learning_rate": 3.864346169461882e-05, "loss": 7.0242, "step": 21662 }, { "epoch": 0.9133376899888273, "grad_norm": 6.014064788818359, "learning_rate": 3.8642680908210756e-05, "loss": 11.7913, "step": 21663 }, { "epoch": 0.9133798511710268, "grad_norm": 2.5226900577545166, "learning_rate": 3.864190012180268e-05, "loss": 7.3586, "step": 21664 }, { "epoch": 0.9134220123532264, "grad_norm": 1.3190231323242188, "learning_rate": 3.864111933539461e-05, "loss": 6.5904, "step": 21665 }, { "epoch": 0.913464173535426, "grad_norm": 2.1992294788360596, "learning_rate": 3.864033854898654e-05, "loss": 6.5453, "step": 21666 }, { "epoch": 0.9135063347176254, "grad_norm": 2.322791814804077, "learning_rate": 3.8639557762578474e-05, "loss": 7.0482, "step": 21667 }, { "epoch": 0.913548495899825, "grad_norm": 0.9672174453735352, "learning_rate": 3.86387769761704e-05, "loss": 6.3994, "step": 21668 }, { "epoch": 0.9135906570820246, "grad_norm": 1.8882129192352295, "learning_rate": 3.863799618976233e-05, "loss": 6.8025, "step": 21669 }, { "epoch": 0.9136328182642242, "grad_norm": 1.3966436386108398, "learning_rate": 3.863721540335426e-05, "loss": 6.6717, "step": 21670 }, { "epoch": 0.9136749794464237, "grad_norm": 2.143596649169922, "learning_rate": 3.8636434616946186e-05, "loss": 6.8418, "step": 21671 }, { "epoch": 0.9137171406286232, "grad_norm": 2.6265687942504883, "learning_rate": 3.8635653830538124e-05, "loss": 7.6392, "step": 21672 }, { "epoch": 0.9137593018108228, "grad_norm": 1.4206186532974243, "learning_rate": 3.863487304413005e-05, "loss": 7.2099, "step": 21673 }, { "epoch": 0.9138014629930223, "grad_norm": 1.0585850477218628, "learning_rate": 3.863409225772198e-05, "loss": 6.7267, "step": 21674 }, { "epoch": 0.9138436241752219, "grad_norm": 2.2552287578582764, "learning_rate": 3.863331147131391e-05, "loss": 7.7664, "step": 21675 }, { "epoch": 0.9138857853574214, "grad_norm": 1.6438060998916626, "learning_rate": 3.8632530684905835e-05, "loss": 6.6239, "step": 21676 }, { "epoch": 0.913927946539621, "grad_norm": 1.211810827255249, "learning_rate": 3.8631749898497773e-05, "loss": 6.6084, "step": 21677 }, { "epoch": 0.9139701077218205, "grad_norm": 1.2339375019073486, "learning_rate": 3.86309691120897e-05, "loss": 7.5352, "step": 21678 }, { "epoch": 0.9140122689040201, "grad_norm": 1.348930835723877, "learning_rate": 3.863018832568163e-05, "loss": 6.5233, "step": 21679 }, { "epoch": 0.9140544300862197, "grad_norm": 3.351712226867676, "learning_rate": 3.862940753927356e-05, "loss": 8.6151, "step": 21680 }, { "epoch": 0.9140965912684191, "grad_norm": 1.2405133247375488, "learning_rate": 3.8628626752865485e-05, "loss": 6.8932, "step": 21681 }, { "epoch": 0.9141387524506187, "grad_norm": 1.5476374626159668, "learning_rate": 3.8627845966457416e-05, "loss": 6.3893, "step": 21682 }, { "epoch": 0.9141809136328183, "grad_norm": 2.2111833095550537, "learning_rate": 3.862706518004935e-05, "loss": 6.8831, "step": 21683 }, { "epoch": 0.9142230748150179, "grad_norm": 1.101781964302063, "learning_rate": 3.862628439364128e-05, "loss": 6.4712, "step": 21684 }, { "epoch": 0.9142652359972173, "grad_norm": 1.0321155786514282, "learning_rate": 3.86255036072332e-05, "loss": 6.7119, "step": 21685 }, { "epoch": 0.9143073971794169, "grad_norm": 1.5601245164871216, "learning_rate": 3.8624722820825134e-05, "loss": 6.912, "step": 21686 }, { "epoch": 0.9143495583616165, "grad_norm": 0.7196832299232483, "learning_rate": 3.8623942034417066e-05, "loss": 6.5221, "step": 21687 }, { "epoch": 0.914391719543816, "grad_norm": 3.743543863296509, "learning_rate": 3.8623161248009e-05, "loss": 7.9106, "step": 21688 }, { "epoch": 0.9144338807260155, "grad_norm": 0.7974070310592651, "learning_rate": 3.862238046160093e-05, "loss": 6.8503, "step": 21689 }, { "epoch": 0.9144760419082151, "grad_norm": 2.0202629566192627, "learning_rate": 3.862159967519285e-05, "loss": 7.3801, "step": 21690 }, { "epoch": 0.9145182030904146, "grad_norm": 1.5370268821716309, "learning_rate": 3.862081888878479e-05, "loss": 6.4933, "step": 21691 }, { "epoch": 0.9145603642726142, "grad_norm": 1.2530653476715088, "learning_rate": 3.8620038102376715e-05, "loss": 6.7506, "step": 21692 }, { "epoch": 0.9146025254548138, "grad_norm": 1.4558707475662231, "learning_rate": 3.8619257315968647e-05, "loss": 7.0024, "step": 21693 }, { "epoch": 0.9146446866370133, "grad_norm": 1.2812613248825073, "learning_rate": 3.861847652956058e-05, "loss": 7.1222, "step": 21694 }, { "epoch": 0.9146868478192128, "grad_norm": 1.740563988685608, "learning_rate": 3.86176957431525e-05, "loss": 6.8724, "step": 21695 }, { "epoch": 0.9147290090014124, "grad_norm": 3.445765495300293, "learning_rate": 3.8616914956744434e-05, "loss": 7.7491, "step": 21696 }, { "epoch": 0.914771170183612, "grad_norm": 0.9759005308151245, "learning_rate": 3.8616134170336365e-05, "loss": 6.5439, "step": 21697 }, { "epoch": 0.9148133313658114, "grad_norm": 4.241878986358643, "learning_rate": 3.8615353383928296e-05, "loss": 6.9833, "step": 21698 }, { "epoch": 0.914855492548011, "grad_norm": 1.3444726467132568, "learning_rate": 3.861457259752023e-05, "loss": 6.545, "step": 21699 }, { "epoch": 0.9148976537302106, "grad_norm": 0.9428860545158386, "learning_rate": 3.861379181111215e-05, "loss": 6.6627, "step": 21700 }, { "epoch": 0.9149398149124102, "grad_norm": 2.455280065536499, "learning_rate": 3.861301102470408e-05, "loss": 7.4858, "step": 21701 }, { "epoch": 0.9149819760946097, "grad_norm": 1.26904296875, "learning_rate": 3.8612230238296014e-05, "loss": 6.9925, "step": 21702 }, { "epoch": 0.9150241372768092, "grad_norm": 1.3316668272018433, "learning_rate": 3.8611449451887946e-05, "loss": 6.8434, "step": 21703 }, { "epoch": 0.9150662984590088, "grad_norm": 1.4284236431121826, "learning_rate": 3.861066866547987e-05, "loss": 7.2033, "step": 21704 }, { "epoch": 0.9151084596412083, "grad_norm": 1.5020544528961182, "learning_rate": 3.86098878790718e-05, "loss": 6.6498, "step": 21705 }, { "epoch": 0.9151506208234079, "grad_norm": 2.133784294128418, "learning_rate": 3.860910709266373e-05, "loss": 7.606, "step": 21706 }, { "epoch": 0.9151927820056074, "grad_norm": 1.7453597784042358, "learning_rate": 3.860832630625566e-05, "loss": 6.7048, "step": 21707 }, { "epoch": 0.915234943187807, "grad_norm": 1.9519225358963013, "learning_rate": 3.8607545519847595e-05, "loss": 7.0953, "step": 21708 }, { "epoch": 0.9152771043700065, "grad_norm": 3.0889785289764404, "learning_rate": 3.860676473343952e-05, "loss": 6.6647, "step": 21709 }, { "epoch": 0.9153192655522061, "grad_norm": 1.6508595943450928, "learning_rate": 3.860598394703146e-05, "loss": 7.0158, "step": 21710 }, { "epoch": 0.9153614267344057, "grad_norm": 1.436000943183899, "learning_rate": 3.860520316062338e-05, "loss": 6.5666, "step": 21711 }, { "epoch": 0.9154035879166051, "grad_norm": 2.6663811206817627, "learning_rate": 3.8604422374215307e-05, "loss": 6.3514, "step": 21712 }, { "epoch": 0.9154457490988047, "grad_norm": 1.4480175971984863, "learning_rate": 3.8603641587807245e-05, "loss": 6.5109, "step": 21713 }, { "epoch": 0.9154879102810043, "grad_norm": 2.8281872272491455, "learning_rate": 3.860286080139917e-05, "loss": 8.0079, "step": 21714 }, { "epoch": 0.9155300714632039, "grad_norm": 1.341221809387207, "learning_rate": 3.86020800149911e-05, "loss": 6.5046, "step": 21715 }, { "epoch": 0.9155722326454033, "grad_norm": 3.1721889972686768, "learning_rate": 3.860129922858303e-05, "loss": 7.4262, "step": 21716 }, { "epoch": 0.9156143938276029, "grad_norm": 1.7269197702407837, "learning_rate": 3.860051844217496e-05, "loss": 7.3415, "step": 21717 }, { "epoch": 0.9156565550098025, "grad_norm": 1.6851942539215088, "learning_rate": 3.859973765576689e-05, "loss": 6.9029, "step": 21718 }, { "epoch": 0.915698716192002, "grad_norm": 1.928292155265808, "learning_rate": 3.859895686935882e-05, "loss": 6.4298, "step": 21719 }, { "epoch": 0.9157408773742016, "grad_norm": 2.9647388458251953, "learning_rate": 3.859817608295075e-05, "loss": 7.2672, "step": 21720 }, { "epoch": 0.9157830385564011, "grad_norm": 1.2496075630187988, "learning_rate": 3.859739529654268e-05, "loss": 6.5705, "step": 21721 }, { "epoch": 0.9158251997386007, "grad_norm": 1.826047658920288, "learning_rate": 3.859661451013461e-05, "loss": 7.0511, "step": 21722 }, { "epoch": 0.9158673609208002, "grad_norm": 4.271119594573975, "learning_rate": 3.859583372372654e-05, "loss": 7.1919, "step": 21723 }, { "epoch": 0.9159095221029998, "grad_norm": 1.4718222618103027, "learning_rate": 3.859505293731847e-05, "loss": 6.8967, "step": 21724 }, { "epoch": 0.9159516832851993, "grad_norm": 3.068251609802246, "learning_rate": 3.85942721509104e-05, "loss": 8.1946, "step": 21725 }, { "epoch": 0.9159938444673988, "grad_norm": 4.005949974060059, "learning_rate": 3.8593491364502324e-05, "loss": 7.1966, "step": 21726 }, { "epoch": 0.9160360056495984, "grad_norm": 1.7090319395065308, "learning_rate": 3.859271057809426e-05, "loss": 7.2824, "step": 21727 }, { "epoch": 0.916078166831798, "grad_norm": 2.7289743423461914, "learning_rate": 3.8591929791686186e-05, "loss": 7.6633, "step": 21728 }, { "epoch": 0.9161203280139976, "grad_norm": 1.4871870279312134, "learning_rate": 3.859114900527812e-05, "loss": 6.55, "step": 21729 }, { "epoch": 0.916162489196197, "grad_norm": 1.1449923515319824, "learning_rate": 3.859036821887005e-05, "loss": 6.9948, "step": 21730 }, { "epoch": 0.9162046503783966, "grad_norm": 1.5739775896072388, "learning_rate": 3.8589587432461973e-05, "loss": 6.74, "step": 21731 }, { "epoch": 0.9162468115605962, "grad_norm": 1.6016335487365723, "learning_rate": 3.858880664605391e-05, "loss": 6.672, "step": 21732 }, { "epoch": 0.9162889727427957, "grad_norm": 1.5797587633132935, "learning_rate": 3.8588025859645836e-05, "loss": 6.9872, "step": 21733 }, { "epoch": 0.9163311339249952, "grad_norm": 1.92033052444458, "learning_rate": 3.858724507323777e-05, "loss": 6.8974, "step": 21734 }, { "epoch": 0.9163732951071948, "grad_norm": 2.880048990249634, "learning_rate": 3.85864642868297e-05, "loss": 6.4654, "step": 21735 }, { "epoch": 0.9164154562893944, "grad_norm": 1.1943395137786865, "learning_rate": 3.858568350042163e-05, "loss": 6.5119, "step": 21736 }, { "epoch": 0.9164576174715939, "grad_norm": 0.9912394285202026, "learning_rate": 3.8584902714013554e-05, "loss": 6.5644, "step": 21737 }, { "epoch": 0.9164997786537935, "grad_norm": 3.912787437438965, "learning_rate": 3.8584121927605485e-05, "loss": 7.3609, "step": 21738 }, { "epoch": 0.916541939835993, "grad_norm": 1.5831401348114014, "learning_rate": 3.858334114119742e-05, "loss": 6.3099, "step": 21739 }, { "epoch": 0.9165841010181925, "grad_norm": 1.225265622138977, "learning_rate": 3.858256035478934e-05, "loss": 6.8078, "step": 21740 }, { "epoch": 0.9166262622003921, "grad_norm": 2.7288081645965576, "learning_rate": 3.858177956838128e-05, "loss": 6.3847, "step": 21741 }, { "epoch": 0.9166684233825917, "grad_norm": 1.364486575126648, "learning_rate": 3.8580998781973204e-05, "loss": 7.0658, "step": 21742 }, { "epoch": 0.9167105845647912, "grad_norm": 1.4989954233169556, "learning_rate": 3.8580217995565135e-05, "loss": 6.5154, "step": 21743 }, { "epoch": 0.9167527457469907, "grad_norm": 3.692167282104492, "learning_rate": 3.8579437209157066e-05, "loss": 7.7749, "step": 21744 }, { "epoch": 0.9167949069291903, "grad_norm": 1.699447751045227, "learning_rate": 3.857865642274899e-05, "loss": 7.9895, "step": 21745 }, { "epoch": 0.9168370681113899, "grad_norm": 1.7342016696929932, "learning_rate": 3.857787563634093e-05, "loss": 7.0144, "step": 21746 }, { "epoch": 0.9168792292935893, "grad_norm": 2.7038862705230713, "learning_rate": 3.857709484993285e-05, "loss": 7.8144, "step": 21747 }, { "epoch": 0.9169213904757889, "grad_norm": 1.9260399341583252, "learning_rate": 3.8576314063524785e-05, "loss": 6.4664, "step": 21748 }, { "epoch": 0.9169635516579885, "grad_norm": 1.8633376359939575, "learning_rate": 3.8575533277116716e-05, "loss": 7.0301, "step": 21749 }, { "epoch": 0.9170057128401881, "grad_norm": 1.821460247039795, "learning_rate": 3.857475249070864e-05, "loss": 6.6281, "step": 21750 }, { "epoch": 0.9170478740223876, "grad_norm": 2.4506306648254395, "learning_rate": 3.857397170430057e-05, "loss": 6.4583, "step": 21751 }, { "epoch": 0.9170900352045871, "grad_norm": 1.7753100395202637, "learning_rate": 3.85731909178925e-05, "loss": 6.7282, "step": 21752 }, { "epoch": 0.9171321963867867, "grad_norm": 1.517520785331726, "learning_rate": 3.8572410131484434e-05, "loss": 6.5445, "step": 21753 }, { "epoch": 0.9171743575689862, "grad_norm": 2.291642904281616, "learning_rate": 3.8571629345076365e-05, "loss": 7.0752, "step": 21754 }, { "epoch": 0.9172165187511858, "grad_norm": 2.952176809310913, "learning_rate": 3.8570848558668297e-05, "loss": 7.8357, "step": 21755 }, { "epoch": 0.9172586799333853, "grad_norm": 1.9987319707870483, "learning_rate": 3.857006777226022e-05, "loss": 6.9384, "step": 21756 }, { "epoch": 0.9173008411155849, "grad_norm": 1.2060072422027588, "learning_rate": 3.856928698585215e-05, "loss": 6.7049, "step": 21757 }, { "epoch": 0.9173430022977844, "grad_norm": 1.5398321151733398, "learning_rate": 3.8568506199444084e-05, "loss": 6.7754, "step": 21758 }, { "epoch": 0.917385163479984, "grad_norm": 1.5909349918365479, "learning_rate": 3.856772541303601e-05, "loss": 7.1696, "step": 21759 }, { "epoch": 0.9174273246621836, "grad_norm": 1.4445191621780396, "learning_rate": 3.8566944626627946e-05, "loss": 6.6744, "step": 21760 }, { "epoch": 0.917469485844383, "grad_norm": 3.1309497356414795, "learning_rate": 3.856616384021987e-05, "loss": 7.7068, "step": 21761 }, { "epoch": 0.9175116470265826, "grad_norm": 2.032339096069336, "learning_rate": 3.85653830538118e-05, "loss": 7.0005, "step": 21762 }, { "epoch": 0.9175538082087822, "grad_norm": 2.158083915710449, "learning_rate": 3.856460226740373e-05, "loss": 6.8941, "step": 21763 }, { "epoch": 0.9175959693909818, "grad_norm": 2.5311737060546875, "learning_rate": 3.856382148099566e-05, "loss": 7.5738, "step": 21764 }, { "epoch": 0.9176381305731812, "grad_norm": 1.0298864841461182, "learning_rate": 3.856304069458759e-05, "loss": 6.8277, "step": 21765 }, { "epoch": 0.9176802917553808, "grad_norm": 1.6839519739151, "learning_rate": 3.856225990817952e-05, "loss": 6.4993, "step": 21766 }, { "epoch": 0.9177224529375804, "grad_norm": 1.2506160736083984, "learning_rate": 3.856147912177145e-05, "loss": 6.4396, "step": 21767 }, { "epoch": 0.9177646141197799, "grad_norm": 2.902526378631592, "learning_rate": 3.856069833536338e-05, "loss": 6.5898, "step": 21768 }, { "epoch": 0.9178067753019795, "grad_norm": 5.0756072998046875, "learning_rate": 3.855991754895531e-05, "loss": 10.1098, "step": 21769 }, { "epoch": 0.917848936484179, "grad_norm": 2.3319242000579834, "learning_rate": 3.855913676254724e-05, "loss": 7.6961, "step": 21770 }, { "epoch": 0.9178910976663786, "grad_norm": 1.3100425004959106, "learning_rate": 3.855835597613917e-05, "loss": 7.2368, "step": 21771 }, { "epoch": 0.9179332588485781, "grad_norm": 1.5314996242523193, "learning_rate": 3.85575751897311e-05, "loss": 6.3998, "step": 21772 }, { "epoch": 0.9179754200307777, "grad_norm": 1.501596212387085, "learning_rate": 3.8556794403323025e-05, "loss": 6.5176, "step": 21773 }, { "epoch": 0.9180175812129772, "grad_norm": 2.546726703643799, "learning_rate": 3.8556013616914957e-05, "loss": 7.3438, "step": 21774 }, { "epoch": 0.9180597423951767, "grad_norm": 1.4926520586013794, "learning_rate": 3.855523283050689e-05, "loss": 6.579, "step": 21775 }, { "epoch": 0.9181019035773763, "grad_norm": 3.013244390487671, "learning_rate": 3.855445204409881e-05, "loss": 7.638, "step": 21776 }, { "epoch": 0.9181440647595759, "grad_norm": 1.2162058353424072, "learning_rate": 3.855367125769075e-05, "loss": 6.8199, "step": 21777 }, { "epoch": 0.9181862259417755, "grad_norm": 1.2591640949249268, "learning_rate": 3.8552890471282675e-05, "loss": 6.6175, "step": 21778 }, { "epoch": 0.9182283871239749, "grad_norm": 2.1027634143829346, "learning_rate": 3.855210968487461e-05, "loss": 7.1602, "step": 21779 }, { "epoch": 0.9182705483061745, "grad_norm": 3.0300676822662354, "learning_rate": 3.855132889846654e-05, "loss": 8.4639, "step": 21780 }, { "epoch": 0.9183127094883741, "grad_norm": 2.3719260692596436, "learning_rate": 3.855054811205847e-05, "loss": 7.1879, "step": 21781 }, { "epoch": 0.9183548706705736, "grad_norm": 2.7855637073516846, "learning_rate": 3.85497673256504e-05, "loss": 7.1171, "step": 21782 }, { "epoch": 0.9183970318527731, "grad_norm": 1.6467671394348145, "learning_rate": 3.8548986539242324e-05, "loss": 6.6604, "step": 21783 }, { "epoch": 0.9184391930349727, "grad_norm": 1.527217984199524, "learning_rate": 3.8548205752834256e-05, "loss": 6.5873, "step": 21784 }, { "epoch": 0.9184813542171723, "grad_norm": 1.1201114654541016, "learning_rate": 3.854742496642619e-05, "loss": 6.7172, "step": 21785 }, { "epoch": 0.9185235153993718, "grad_norm": 2.78306245803833, "learning_rate": 3.854664418001812e-05, "loss": 7.71, "step": 21786 }, { "epoch": 0.9185656765815714, "grad_norm": 1.759809136390686, "learning_rate": 3.854586339361004e-05, "loss": 6.8476, "step": 21787 }, { "epoch": 0.9186078377637709, "grad_norm": 1.8946231603622437, "learning_rate": 3.8545082607201974e-05, "loss": 6.7147, "step": 21788 }, { "epoch": 0.9186499989459704, "grad_norm": 2.8074471950531006, "learning_rate": 3.8544301820793905e-05, "loss": 6.8274, "step": 21789 }, { "epoch": 0.91869216012817, "grad_norm": 2.70491886138916, "learning_rate": 3.8543521034385836e-05, "loss": 7.8127, "step": 21790 }, { "epoch": 0.9187343213103696, "grad_norm": 1.9234516620635986, "learning_rate": 3.854274024797777e-05, "loss": 6.8812, "step": 21791 }, { "epoch": 0.918776482492569, "grad_norm": 1.4367148876190186, "learning_rate": 3.854195946156969e-05, "loss": 6.5575, "step": 21792 }, { "epoch": 0.9188186436747686, "grad_norm": 5.049142360687256, "learning_rate": 3.8541178675161623e-05, "loss": 10.499, "step": 21793 }, { "epoch": 0.9188608048569682, "grad_norm": 1.0702669620513916, "learning_rate": 3.8540397888753555e-05, "loss": 6.872, "step": 21794 }, { "epoch": 0.9189029660391678, "grad_norm": 0.9802191853523254, "learning_rate": 3.853961710234548e-05, "loss": 6.9515, "step": 21795 }, { "epoch": 0.9189451272213673, "grad_norm": 1.0029183626174927, "learning_rate": 3.853883631593742e-05, "loss": 6.4489, "step": 21796 }, { "epoch": 0.9189872884035668, "grad_norm": 2.301706314086914, "learning_rate": 3.853805552952934e-05, "loss": 7.1855, "step": 21797 }, { "epoch": 0.9190294495857664, "grad_norm": 2.7567412853240967, "learning_rate": 3.853727474312127e-05, "loss": 7.6028, "step": 21798 }, { "epoch": 0.919071610767966, "grad_norm": 1.6717209815979004, "learning_rate": 3.8536493956713204e-05, "loss": 6.5797, "step": 21799 }, { "epoch": 0.9191137719501655, "grad_norm": 1.2412277460098267, "learning_rate": 3.8535713170305135e-05, "loss": 6.7971, "step": 21800 }, { "epoch": 0.919155933132365, "grad_norm": 2.4854347705841064, "learning_rate": 3.853493238389707e-05, "loss": 8.2404, "step": 21801 }, { "epoch": 0.9191980943145646, "grad_norm": 2.2861554622650146, "learning_rate": 3.853415159748899e-05, "loss": 7.6236, "step": 21802 }, { "epoch": 0.9192402554967641, "grad_norm": 2.634842872619629, "learning_rate": 3.853337081108092e-05, "loss": 6.5481, "step": 21803 }, { "epoch": 0.9192824166789637, "grad_norm": 2.7792153358459473, "learning_rate": 3.8532590024672854e-05, "loss": 6.4936, "step": 21804 }, { "epoch": 0.9193245778611632, "grad_norm": 2.4889233112335205, "learning_rate": 3.8531809238264785e-05, "loss": 7.7281, "step": 21805 }, { "epoch": 0.9193667390433627, "grad_norm": 2.0508010387420654, "learning_rate": 3.853102845185671e-05, "loss": 6.8488, "step": 21806 }, { "epoch": 0.9194089002255623, "grad_norm": 1.4886993169784546, "learning_rate": 3.853024766544864e-05, "loss": 6.7978, "step": 21807 }, { "epoch": 0.9194510614077619, "grad_norm": 1.0662165880203247, "learning_rate": 3.852946687904057e-05, "loss": 6.664, "step": 21808 }, { "epoch": 0.9194932225899615, "grad_norm": 1.1722692251205444, "learning_rate": 3.8528686092632496e-05, "loss": 6.8171, "step": 21809 }, { "epoch": 0.9195353837721609, "grad_norm": 2.6155576705932617, "learning_rate": 3.8527905306224435e-05, "loss": 7.7753, "step": 21810 }, { "epoch": 0.9195775449543605, "grad_norm": 1.2993892431259155, "learning_rate": 3.852712451981636e-05, "loss": 6.3675, "step": 21811 }, { "epoch": 0.9196197061365601, "grad_norm": 3.001281261444092, "learning_rate": 3.852634373340829e-05, "loss": 7.6919, "step": 21812 }, { "epoch": 0.9196618673187597, "grad_norm": 3.221130132675171, "learning_rate": 3.852556294700022e-05, "loss": 6.4307, "step": 21813 }, { "epoch": 0.9197040285009591, "grad_norm": 1.3789088726043701, "learning_rate": 3.8524782160592146e-05, "loss": 6.4899, "step": 21814 }, { "epoch": 0.9197461896831587, "grad_norm": 2.8954734802246094, "learning_rate": 3.8524001374184084e-05, "loss": 7.472, "step": 21815 }, { "epoch": 0.9197883508653583, "grad_norm": 1.2354435920715332, "learning_rate": 3.852322058777601e-05, "loss": 6.7473, "step": 21816 }, { "epoch": 0.9198305120475578, "grad_norm": 3.440342426300049, "learning_rate": 3.852243980136794e-05, "loss": 8.3928, "step": 21817 }, { "epoch": 0.9198726732297574, "grad_norm": 1.3333972692489624, "learning_rate": 3.852165901495987e-05, "loss": 6.9935, "step": 21818 }, { "epoch": 0.9199148344119569, "grad_norm": 2.5536370277404785, "learning_rate": 3.8520878228551796e-05, "loss": 8.1536, "step": 21819 }, { "epoch": 0.9199569955941564, "grad_norm": 2.1692726612091064, "learning_rate": 3.852009744214373e-05, "loss": 6.4844, "step": 21820 }, { "epoch": 0.919999156776356, "grad_norm": 3.0376474857330322, "learning_rate": 3.851931665573566e-05, "loss": 8.7128, "step": 21821 }, { "epoch": 0.9200413179585556, "grad_norm": 5.34818172454834, "learning_rate": 3.851853586932759e-05, "loss": 8.2011, "step": 21822 }, { "epoch": 0.9200834791407551, "grad_norm": 2.558727741241455, "learning_rate": 3.851775508291952e-05, "loss": 7.3071, "step": 21823 }, { "epoch": 0.9201256403229546, "grad_norm": 1.338441014289856, "learning_rate": 3.851697429651145e-05, "loss": 7.1403, "step": 21824 }, { "epoch": 0.9201678015051542, "grad_norm": 1.56855046749115, "learning_rate": 3.8516193510103376e-05, "loss": 6.6467, "step": 21825 }, { "epoch": 0.9202099626873538, "grad_norm": 2.0489015579223633, "learning_rate": 3.851541272369531e-05, "loss": 6.4599, "step": 21826 }, { "epoch": 0.9202521238695534, "grad_norm": 1.7585667371749878, "learning_rate": 3.851463193728724e-05, "loss": 6.8586, "step": 21827 }, { "epoch": 0.9202942850517528, "grad_norm": 2.723907709121704, "learning_rate": 3.851385115087916e-05, "loss": 7.2405, "step": 21828 }, { "epoch": 0.9203364462339524, "grad_norm": 1.7166705131530762, "learning_rate": 3.85130703644711e-05, "loss": 6.9853, "step": 21829 }, { "epoch": 0.920378607416152, "grad_norm": 1.4278136491775513, "learning_rate": 3.8512289578063026e-05, "loss": 6.9997, "step": 21830 }, { "epoch": 0.9204207685983515, "grad_norm": 2.3082222938537598, "learning_rate": 3.851150879165496e-05, "loss": 7.8642, "step": 21831 }, { "epoch": 0.920462929780551, "grad_norm": 2.6704261302948, "learning_rate": 3.851072800524689e-05, "loss": 7.6567, "step": 21832 }, { "epoch": 0.9205050909627506, "grad_norm": 1.8073945045471191, "learning_rate": 3.850994721883881e-05, "loss": 6.6352, "step": 21833 }, { "epoch": 0.9205472521449501, "grad_norm": 2.880220651626587, "learning_rate": 3.850916643243075e-05, "loss": 7.7754, "step": 21834 }, { "epoch": 0.9205894133271497, "grad_norm": 2.2985455989837646, "learning_rate": 3.8508385646022675e-05, "loss": 8.2498, "step": 21835 }, { "epoch": 0.9206315745093493, "grad_norm": 1.3128094673156738, "learning_rate": 3.850760485961461e-05, "loss": 6.8186, "step": 21836 }, { "epoch": 0.9206737356915488, "grad_norm": 1.749513864517212, "learning_rate": 3.850682407320654e-05, "loss": 6.4026, "step": 21837 }, { "epoch": 0.9207158968737483, "grad_norm": 1.928289771080017, "learning_rate": 3.850604328679846e-05, "loss": 7.3952, "step": 21838 }, { "epoch": 0.9207580580559479, "grad_norm": 5.421963691711426, "learning_rate": 3.8505262500390394e-05, "loss": 10.3427, "step": 21839 }, { "epoch": 0.9208002192381475, "grad_norm": 3.4399499893188477, "learning_rate": 3.8504481713982325e-05, "loss": 7.6252, "step": 21840 }, { "epoch": 0.9208423804203469, "grad_norm": 1.3822661638259888, "learning_rate": 3.8503700927574256e-05, "loss": 6.8207, "step": 21841 }, { "epoch": 0.9208845416025465, "grad_norm": 1.919330358505249, "learning_rate": 3.850292014116618e-05, "loss": 6.8943, "step": 21842 }, { "epoch": 0.9209267027847461, "grad_norm": 1.776617169380188, "learning_rate": 3.850213935475812e-05, "loss": 6.8076, "step": 21843 }, { "epoch": 0.9209688639669457, "grad_norm": 1.014795184135437, "learning_rate": 3.850135856835004e-05, "loss": 7.1487, "step": 21844 }, { "epoch": 0.9210110251491452, "grad_norm": 3.192716121673584, "learning_rate": 3.8500577781941974e-05, "loss": 7.4721, "step": 21845 }, { "epoch": 0.9210531863313447, "grad_norm": 1.6682511568069458, "learning_rate": 3.8499796995533906e-05, "loss": 6.6375, "step": 21846 }, { "epoch": 0.9210953475135443, "grad_norm": 1.3219935894012451, "learning_rate": 3.849901620912583e-05, "loss": 6.9985, "step": 21847 }, { "epoch": 0.9211375086957438, "grad_norm": 1.9195833206176758, "learning_rate": 3.849823542271777e-05, "loss": 6.6322, "step": 21848 }, { "epoch": 0.9211796698779434, "grad_norm": 1.7644617557525635, "learning_rate": 3.849745463630969e-05, "loss": 7.0658, "step": 21849 }, { "epoch": 0.9212218310601429, "grad_norm": 1.600294589996338, "learning_rate": 3.8496673849901624e-05, "loss": 7.5024, "step": 21850 }, { "epoch": 0.9212639922423425, "grad_norm": 1.4801702499389648, "learning_rate": 3.8495893063493555e-05, "loss": 6.7022, "step": 21851 }, { "epoch": 0.921306153424542, "grad_norm": 1.3551127910614014, "learning_rate": 3.849511227708548e-05, "loss": 6.9071, "step": 21852 }, { "epoch": 0.9213483146067416, "grad_norm": 1.4359158277511597, "learning_rate": 3.849433149067741e-05, "loss": 6.9891, "step": 21853 }, { "epoch": 0.9213904757889412, "grad_norm": 0.9338351488113403, "learning_rate": 3.849355070426934e-05, "loss": 6.6344, "step": 21854 }, { "epoch": 0.9214326369711406, "grad_norm": 2.451836585998535, "learning_rate": 3.8492769917861273e-05, "loss": 6.6849, "step": 21855 }, { "epoch": 0.9214747981533402, "grad_norm": 4.174998760223389, "learning_rate": 3.84919891314532e-05, "loss": 9.4378, "step": 21856 }, { "epoch": 0.9215169593355398, "grad_norm": 3.0077624320983887, "learning_rate": 3.849120834504513e-05, "loss": 7.7065, "step": 21857 }, { "epoch": 0.9215591205177394, "grad_norm": 1.4634946584701538, "learning_rate": 3.849042755863706e-05, "loss": 6.3729, "step": 21858 }, { "epoch": 0.9216012816999388, "grad_norm": 1.6980212926864624, "learning_rate": 3.848964677222899e-05, "loss": 6.6725, "step": 21859 }, { "epoch": 0.9216434428821384, "grad_norm": 2.351919412612915, "learning_rate": 3.848886598582092e-05, "loss": 6.698, "step": 21860 }, { "epoch": 0.921685604064338, "grad_norm": 1.0898582935333252, "learning_rate": 3.848808519941285e-05, "loss": 6.5321, "step": 21861 }, { "epoch": 0.9217277652465375, "grad_norm": 2.8288071155548096, "learning_rate": 3.8487304413004785e-05, "loss": 7.9589, "step": 21862 }, { "epoch": 0.921769926428737, "grad_norm": 0.8559487462043762, "learning_rate": 3.848652362659671e-05, "loss": 6.4663, "step": 21863 }, { "epoch": 0.9218120876109366, "grad_norm": 1.5643930435180664, "learning_rate": 3.8485742840188634e-05, "loss": 6.3969, "step": 21864 }, { "epoch": 0.9218542487931362, "grad_norm": 2.4593312740325928, "learning_rate": 3.848496205378057e-05, "loss": 7.936, "step": 21865 }, { "epoch": 0.9218964099753357, "grad_norm": 2.7662017345428467, "learning_rate": 3.84841812673725e-05, "loss": 7.5751, "step": 21866 }, { "epoch": 0.9219385711575353, "grad_norm": 1.044370174407959, "learning_rate": 3.848340048096443e-05, "loss": 6.5498, "step": 21867 }, { "epoch": 0.9219807323397348, "grad_norm": 3.5789754390716553, "learning_rate": 3.848261969455636e-05, "loss": 7.7508, "step": 21868 }, { "epoch": 0.9220228935219343, "grad_norm": 3.5965287685394287, "learning_rate": 3.848183890814829e-05, "loss": 7.4451, "step": 21869 }, { "epoch": 0.9220650547041339, "grad_norm": 2.2258145809173584, "learning_rate": 3.848105812174022e-05, "loss": 6.9395, "step": 21870 }, { "epoch": 0.9221072158863335, "grad_norm": 2.6197171211242676, "learning_rate": 3.8480277335332147e-05, "loss": 7.6262, "step": 21871 }, { "epoch": 0.922149377068533, "grad_norm": 2.405282735824585, "learning_rate": 3.847949654892408e-05, "loss": 8.0626, "step": 21872 }, { "epoch": 0.9221915382507325, "grad_norm": 2.8517589569091797, "learning_rate": 3.847871576251601e-05, "loss": 7.4261, "step": 21873 }, { "epoch": 0.9222336994329321, "grad_norm": 4.115591049194336, "learning_rate": 3.847793497610794e-05, "loss": 8.0256, "step": 21874 }, { "epoch": 0.9222758606151317, "grad_norm": 1.8056361675262451, "learning_rate": 3.8477154189699865e-05, "loss": 7.1594, "step": 21875 }, { "epoch": 0.9223180217973312, "grad_norm": 1.631034255027771, "learning_rate": 3.8476373403291796e-05, "loss": 6.5741, "step": 21876 }, { "epoch": 0.9223601829795307, "grad_norm": 1.3754379749298096, "learning_rate": 3.847559261688373e-05, "loss": 6.8683, "step": 21877 }, { "epoch": 0.9224023441617303, "grad_norm": 1.4185981750488281, "learning_rate": 3.847481183047565e-05, "loss": 7.0614, "step": 21878 }, { "epoch": 0.9224445053439299, "grad_norm": 3.6031668186187744, "learning_rate": 3.847403104406759e-05, "loss": 7.294, "step": 21879 }, { "epoch": 0.9224866665261294, "grad_norm": 1.588379979133606, "learning_rate": 3.8473250257659514e-05, "loss": 6.5924, "step": 21880 }, { "epoch": 0.9225288277083289, "grad_norm": 2.872285842895508, "learning_rate": 3.8472469471251446e-05, "loss": 7.7344, "step": 21881 }, { "epoch": 0.9225709888905285, "grad_norm": 1.5162886381149292, "learning_rate": 3.847168868484338e-05, "loss": 6.5707, "step": 21882 }, { "epoch": 0.922613150072728, "grad_norm": 1.4015231132507324, "learning_rate": 3.84709078984353e-05, "loss": 6.5052, "step": 21883 }, { "epoch": 0.9226553112549276, "grad_norm": 3.022043466567993, "learning_rate": 3.847012711202724e-05, "loss": 7.5001, "step": 21884 }, { "epoch": 0.9226974724371272, "grad_norm": 2.0404787063598633, "learning_rate": 3.8469346325619164e-05, "loss": 6.5308, "step": 21885 }, { "epoch": 0.9227396336193266, "grad_norm": 1.404708981513977, "learning_rate": 3.8468565539211095e-05, "loss": 6.8935, "step": 21886 }, { "epoch": 0.9227817948015262, "grad_norm": 1.275659203529358, "learning_rate": 3.8467784752803026e-05, "loss": 6.5019, "step": 21887 }, { "epoch": 0.9228239559837258, "grad_norm": 3.834730625152588, "learning_rate": 3.846700396639496e-05, "loss": 7.5906, "step": 21888 }, { "epoch": 0.9228661171659254, "grad_norm": 3.037384510040283, "learning_rate": 3.846622317998688e-05, "loss": 7.7281, "step": 21889 }, { "epoch": 0.9229082783481248, "grad_norm": 1.2943533658981323, "learning_rate": 3.846544239357881e-05, "loss": 6.5762, "step": 21890 }, { "epoch": 0.9229504395303244, "grad_norm": 2.4578697681427, "learning_rate": 3.8464661607170745e-05, "loss": 7.189, "step": 21891 }, { "epoch": 0.922992600712524, "grad_norm": 1.9933140277862549, "learning_rate": 3.8463880820762676e-05, "loss": 7.2733, "step": 21892 }, { "epoch": 0.9230347618947236, "grad_norm": 1.7918510437011719, "learning_rate": 3.846310003435461e-05, "loss": 6.9487, "step": 21893 }, { "epoch": 0.9230769230769231, "grad_norm": 2.0061521530151367, "learning_rate": 3.846231924794653e-05, "loss": 6.6076, "step": 21894 }, { "epoch": 0.9231190842591226, "grad_norm": 2.0043628215789795, "learning_rate": 3.846153846153846e-05, "loss": 6.6321, "step": 21895 }, { "epoch": 0.9231612454413222, "grad_norm": 1.4848109483718872, "learning_rate": 3.8460757675130394e-05, "loss": 6.4729, "step": 21896 }, { "epoch": 0.9232034066235217, "grad_norm": 2.0227181911468506, "learning_rate": 3.845997688872232e-05, "loss": 7.0426, "step": 21897 }, { "epoch": 0.9232455678057213, "grad_norm": 1.9204890727996826, "learning_rate": 3.845919610231426e-05, "loss": 6.4746, "step": 21898 }, { "epoch": 0.9232877289879208, "grad_norm": 1.6349303722381592, "learning_rate": 3.845841531590618e-05, "loss": 7.0956, "step": 21899 }, { "epoch": 0.9233298901701203, "grad_norm": 1.6041651964187622, "learning_rate": 3.845763452949811e-05, "loss": 6.7442, "step": 21900 }, { "epoch": 0.9233720513523199, "grad_norm": 2.0632007122039795, "learning_rate": 3.8456853743090044e-05, "loss": 7.0676, "step": 21901 }, { "epoch": 0.9234142125345195, "grad_norm": 3.0889339447021484, "learning_rate": 3.845607295668197e-05, "loss": 7.9027, "step": 21902 }, { "epoch": 0.9234563737167191, "grad_norm": 1.8217084407806396, "learning_rate": 3.8455292170273906e-05, "loss": 6.5548, "step": 21903 }, { "epoch": 0.9234985348989185, "grad_norm": 2.3963847160339355, "learning_rate": 3.845451138386583e-05, "loss": 7.1177, "step": 21904 }, { "epoch": 0.9235406960811181, "grad_norm": 1.5751099586486816, "learning_rate": 3.845373059745776e-05, "loss": 6.5947, "step": 21905 }, { "epoch": 0.9235828572633177, "grad_norm": 3.733151912689209, "learning_rate": 3.845294981104969e-05, "loss": 7.334, "step": 21906 }, { "epoch": 0.9236250184455173, "grad_norm": 1.4110032320022583, "learning_rate": 3.8452169024641624e-05, "loss": 6.6697, "step": 21907 }, { "epoch": 0.9236671796277167, "grad_norm": 2.632458209991455, "learning_rate": 3.845138823823355e-05, "loss": 6.3915, "step": 21908 }, { "epoch": 0.9237093408099163, "grad_norm": 2.710129499435425, "learning_rate": 3.845060745182548e-05, "loss": 7.7125, "step": 21909 }, { "epoch": 0.9237515019921159, "grad_norm": 2.1196577548980713, "learning_rate": 3.844982666541741e-05, "loss": 6.7212, "step": 21910 }, { "epoch": 0.9237936631743154, "grad_norm": 2.3687033653259277, "learning_rate": 3.8449045879009336e-05, "loss": 6.5761, "step": 21911 }, { "epoch": 0.923835824356515, "grad_norm": 1.4633070230484009, "learning_rate": 3.8448265092601274e-05, "loss": 6.6849, "step": 21912 }, { "epoch": 0.9238779855387145, "grad_norm": 1.370388388633728, "learning_rate": 3.84474843061932e-05, "loss": 7.1628, "step": 21913 }, { "epoch": 0.923920146720914, "grad_norm": 1.0498020648956299, "learning_rate": 3.844670351978513e-05, "loss": 6.5181, "step": 21914 }, { "epoch": 0.9239623079031136, "grad_norm": 3.826317071914673, "learning_rate": 3.844592273337706e-05, "loss": 6.3933, "step": 21915 }, { "epoch": 0.9240044690853132, "grad_norm": 1.3783332109451294, "learning_rate": 3.8445141946968985e-05, "loss": 6.6236, "step": 21916 }, { "epoch": 0.9240466302675127, "grad_norm": 3.160989284515381, "learning_rate": 3.8444361160560923e-05, "loss": 7.6888, "step": 21917 }, { "epoch": 0.9240887914497122, "grad_norm": 1.2916234731674194, "learning_rate": 3.844358037415285e-05, "loss": 6.9537, "step": 21918 }, { "epoch": 0.9241309526319118, "grad_norm": 2.8807735443115234, "learning_rate": 3.844279958774478e-05, "loss": 7.4122, "step": 21919 }, { "epoch": 0.9241731138141114, "grad_norm": 2.632493019104004, "learning_rate": 3.844201880133671e-05, "loss": 8.5329, "step": 21920 }, { "epoch": 0.9242152749963108, "grad_norm": 1.5939396619796753, "learning_rate": 3.8441238014928635e-05, "loss": 7.2034, "step": 21921 }, { "epoch": 0.9242574361785104, "grad_norm": 2.147993326187134, "learning_rate": 3.8440457228520566e-05, "loss": 7.0542, "step": 21922 }, { "epoch": 0.92429959736071, "grad_norm": 2.7425434589385986, "learning_rate": 3.84396764421125e-05, "loss": 7.7203, "step": 21923 }, { "epoch": 0.9243417585429096, "grad_norm": 2.8640944957733154, "learning_rate": 3.843889565570443e-05, "loss": 6.6511, "step": 21924 }, { "epoch": 0.9243839197251091, "grad_norm": 2.1418910026550293, "learning_rate": 3.843811486929635e-05, "loss": 6.5731, "step": 21925 }, { "epoch": 0.9244260809073086, "grad_norm": 3.567117691040039, "learning_rate": 3.8437334082888284e-05, "loss": 8.2383, "step": 21926 }, { "epoch": 0.9244682420895082, "grad_norm": 1.102354645729065, "learning_rate": 3.8436553296480216e-05, "loss": 7.095, "step": 21927 }, { "epoch": 0.9245104032717077, "grad_norm": 1.955613374710083, "learning_rate": 3.843577251007215e-05, "loss": 6.5575, "step": 21928 }, { "epoch": 0.9245525644539073, "grad_norm": 1.9112950563430786, "learning_rate": 3.843499172366408e-05, "loss": 6.6499, "step": 21929 }, { "epoch": 0.9245947256361068, "grad_norm": 3.338630437850952, "learning_rate": 3.8434210937256e-05, "loss": 6.717, "step": 21930 }, { "epoch": 0.9246368868183064, "grad_norm": 2.8250679969787598, "learning_rate": 3.843343015084794e-05, "loss": 6.645, "step": 21931 }, { "epoch": 0.9246790480005059, "grad_norm": 1.6162663698196411, "learning_rate": 3.8432649364439865e-05, "loss": 6.5095, "step": 21932 }, { "epoch": 0.9247212091827055, "grad_norm": 1.7570289373397827, "learning_rate": 3.8431868578031797e-05, "loss": 6.81, "step": 21933 }, { "epoch": 0.9247633703649051, "grad_norm": 1.5233747959136963, "learning_rate": 3.843108779162373e-05, "loss": 6.6681, "step": 21934 }, { "epoch": 0.9248055315471045, "grad_norm": 3.5377748012542725, "learning_rate": 3.843030700521565e-05, "loss": 7.4837, "step": 21935 }, { "epoch": 0.9248476927293041, "grad_norm": 1.2487342357635498, "learning_rate": 3.8429526218807584e-05, "loss": 6.6069, "step": 21936 }, { "epoch": 0.9248898539115037, "grad_norm": 1.8677150011062622, "learning_rate": 3.8428745432399515e-05, "loss": 6.9608, "step": 21937 }, { "epoch": 0.9249320150937033, "grad_norm": 1.717647910118103, "learning_rate": 3.8427964645991446e-05, "loss": 7.3067, "step": 21938 }, { "epoch": 0.9249741762759027, "grad_norm": 1.286840558052063, "learning_rate": 3.842718385958338e-05, "loss": 6.7615, "step": 21939 }, { "epoch": 0.9250163374581023, "grad_norm": 3.308300495147705, "learning_rate": 3.84264030731753e-05, "loss": 7.6191, "step": 21940 }, { "epoch": 0.9250584986403019, "grad_norm": 2.108535051345825, "learning_rate": 3.842562228676723e-05, "loss": 7.1092, "step": 21941 }, { "epoch": 0.9251006598225014, "grad_norm": 1.0914530754089355, "learning_rate": 3.8424841500359164e-05, "loss": 6.9649, "step": 21942 }, { "epoch": 0.925142821004701, "grad_norm": 3.107243061065674, "learning_rate": 3.8424060713951096e-05, "loss": 7.1709, "step": 21943 }, { "epoch": 0.9251849821869005, "grad_norm": 4.718821048736572, "learning_rate": 3.842327992754302e-05, "loss": 8.2996, "step": 21944 }, { "epoch": 0.9252271433691001, "grad_norm": 1.7862101793289185, "learning_rate": 3.842249914113495e-05, "loss": 7.3961, "step": 21945 }, { "epoch": 0.9252693045512996, "grad_norm": 1.9223856925964355, "learning_rate": 3.842171835472688e-05, "loss": 7.0882, "step": 21946 }, { "epoch": 0.9253114657334992, "grad_norm": 1.5738195180892944, "learning_rate": 3.842093756831881e-05, "loss": 6.7546, "step": 21947 }, { "epoch": 0.9253536269156987, "grad_norm": 3.577737808227539, "learning_rate": 3.8420156781910745e-05, "loss": 8.2038, "step": 21948 }, { "epoch": 0.9253957880978982, "grad_norm": 3.1038050651550293, "learning_rate": 3.841937599550267e-05, "loss": 7.3712, "step": 21949 }, { "epoch": 0.9254379492800978, "grad_norm": 2.7794156074523926, "learning_rate": 3.841859520909461e-05, "loss": 7.634, "step": 21950 }, { "epoch": 0.9254801104622974, "grad_norm": 2.3237297534942627, "learning_rate": 3.841781442268653e-05, "loss": 7.4721, "step": 21951 }, { "epoch": 0.925522271644497, "grad_norm": 1.3813961744308472, "learning_rate": 3.841703363627846e-05, "loss": 6.5069, "step": 21952 }, { "epoch": 0.9255644328266964, "grad_norm": 2.724682092666626, "learning_rate": 3.8416252849870395e-05, "loss": 6.6598, "step": 21953 }, { "epoch": 0.925606594008896, "grad_norm": 1.7576011419296265, "learning_rate": 3.841547206346232e-05, "loss": 6.594, "step": 21954 }, { "epoch": 0.9256487551910956, "grad_norm": 1.414008617401123, "learning_rate": 3.841469127705425e-05, "loss": 6.9697, "step": 21955 }, { "epoch": 0.9256909163732951, "grad_norm": 1.3391000032424927, "learning_rate": 3.841391049064618e-05, "loss": 6.7912, "step": 21956 }, { "epoch": 0.9257330775554946, "grad_norm": 2.2273335456848145, "learning_rate": 3.841312970423811e-05, "loss": 6.4785, "step": 21957 }, { "epoch": 0.9257752387376942, "grad_norm": 1.8229727745056152, "learning_rate": 3.841234891783004e-05, "loss": 6.8438, "step": 21958 }, { "epoch": 0.9258173999198938, "grad_norm": 1.0546351671218872, "learning_rate": 3.841156813142197e-05, "loss": 7.2636, "step": 21959 }, { "epoch": 0.9258595611020933, "grad_norm": 2.64393949508667, "learning_rate": 3.84107873450139e-05, "loss": 7.234, "step": 21960 }, { "epoch": 0.9259017222842929, "grad_norm": 1.4661846160888672, "learning_rate": 3.841000655860583e-05, "loss": 6.4962, "step": 21961 }, { "epoch": 0.9259438834664924, "grad_norm": 1.3470137119293213, "learning_rate": 3.840922577219776e-05, "loss": 6.5575, "step": 21962 }, { "epoch": 0.9259860446486919, "grad_norm": 2.431405782699585, "learning_rate": 3.840844498578969e-05, "loss": 7.2355, "step": 21963 }, { "epoch": 0.9260282058308915, "grad_norm": 3.239041328430176, "learning_rate": 3.840766419938162e-05, "loss": 8.6516, "step": 21964 }, { "epoch": 0.9260703670130911, "grad_norm": 2.4653451442718506, "learning_rate": 3.840688341297355e-05, "loss": 7.0125, "step": 21965 }, { "epoch": 0.9261125281952906, "grad_norm": 3.326345682144165, "learning_rate": 3.8406102626565474e-05, "loss": 9.439, "step": 21966 }, { "epoch": 0.9261546893774901, "grad_norm": 3.850355625152588, "learning_rate": 3.840532184015741e-05, "loss": 7.7249, "step": 21967 }, { "epoch": 0.9261968505596897, "grad_norm": 1.7073603868484497, "learning_rate": 3.8404541053749336e-05, "loss": 7.2149, "step": 21968 }, { "epoch": 0.9262390117418893, "grad_norm": 1.6521648168563843, "learning_rate": 3.840376026734127e-05, "loss": 6.5875, "step": 21969 }, { "epoch": 0.9262811729240887, "grad_norm": 2.2788026332855225, "learning_rate": 3.84029794809332e-05, "loss": 6.93, "step": 21970 }, { "epoch": 0.9263233341062883, "grad_norm": 1.6956769227981567, "learning_rate": 3.8402198694525123e-05, "loss": 6.9723, "step": 21971 }, { "epoch": 0.9263654952884879, "grad_norm": 1.9049535989761353, "learning_rate": 3.840141790811706e-05, "loss": 6.7718, "step": 21972 }, { "epoch": 0.9264076564706875, "grad_norm": 3.0117762088775635, "learning_rate": 3.8400637121708986e-05, "loss": 7.6169, "step": 21973 }, { "epoch": 0.926449817652887, "grad_norm": 3.281780481338501, "learning_rate": 3.839985633530092e-05, "loss": 7.5506, "step": 21974 }, { "epoch": 0.9264919788350865, "grad_norm": 2.836576223373413, "learning_rate": 3.839907554889285e-05, "loss": 7.5384, "step": 21975 }, { "epoch": 0.9265341400172861, "grad_norm": 2.167388439178467, "learning_rate": 3.839829476248478e-05, "loss": 6.6095, "step": 21976 }, { "epoch": 0.9265763011994856, "grad_norm": 2.4319674968719482, "learning_rate": 3.8397513976076704e-05, "loss": 6.5994, "step": 21977 }, { "epoch": 0.9266184623816852, "grad_norm": 2.0900957584381104, "learning_rate": 3.8396733189668635e-05, "loss": 6.5392, "step": 21978 }, { "epoch": 0.9266606235638847, "grad_norm": 1.238588809967041, "learning_rate": 3.839595240326057e-05, "loss": 7.1239, "step": 21979 }, { "epoch": 0.9267027847460843, "grad_norm": 1.1963934898376465, "learning_rate": 3.839517161685249e-05, "loss": 6.5446, "step": 21980 }, { "epoch": 0.9267449459282838, "grad_norm": 1.8872474431991577, "learning_rate": 3.839439083044443e-05, "loss": 7.0649, "step": 21981 }, { "epoch": 0.9267871071104834, "grad_norm": 1.4462101459503174, "learning_rate": 3.8393610044036354e-05, "loss": 6.8729, "step": 21982 }, { "epoch": 0.926829268292683, "grad_norm": 2.3401944637298584, "learning_rate": 3.8392829257628285e-05, "loss": 7.4382, "step": 21983 }, { "epoch": 0.9268714294748824, "grad_norm": 1.5220551490783691, "learning_rate": 3.8392048471220216e-05, "loss": 6.6886, "step": 21984 }, { "epoch": 0.926913590657082, "grad_norm": 1.402760624885559, "learning_rate": 3.839126768481214e-05, "loss": 7.3632, "step": 21985 }, { "epoch": 0.9269557518392816, "grad_norm": 1.6190190315246582, "learning_rate": 3.839048689840408e-05, "loss": 7.2415, "step": 21986 }, { "epoch": 0.9269979130214812, "grad_norm": 1.5562480688095093, "learning_rate": 3.8389706111996e-05, "loss": 7.1474, "step": 21987 }, { "epoch": 0.9270400742036806, "grad_norm": 2.093874454498291, "learning_rate": 3.8388925325587935e-05, "loss": 7.3675, "step": 21988 }, { "epoch": 0.9270822353858802, "grad_norm": 3.6379520893096924, "learning_rate": 3.8388144539179866e-05, "loss": 7.6901, "step": 21989 }, { "epoch": 0.9271243965680798, "grad_norm": 1.4237138032913208, "learning_rate": 3.838736375277179e-05, "loss": 6.5535, "step": 21990 }, { "epoch": 0.9271665577502793, "grad_norm": 1.7017958164215088, "learning_rate": 3.838658296636372e-05, "loss": 7.0088, "step": 21991 }, { "epoch": 0.9272087189324789, "grad_norm": 2.4412550926208496, "learning_rate": 3.838580217995565e-05, "loss": 8.6221, "step": 21992 }, { "epoch": 0.9272508801146784, "grad_norm": 3.3719987869262695, "learning_rate": 3.8385021393547584e-05, "loss": 7.2398, "step": 21993 }, { "epoch": 0.927293041296878, "grad_norm": 1.9148192405700684, "learning_rate": 3.8384240607139515e-05, "loss": 6.6147, "step": 21994 }, { "epoch": 0.9273352024790775, "grad_norm": 2.720064640045166, "learning_rate": 3.8383459820731447e-05, "loss": 7.2151, "step": 21995 }, { "epoch": 0.9273773636612771, "grad_norm": 3.427354097366333, "learning_rate": 3.838267903432337e-05, "loss": 7.2666, "step": 21996 }, { "epoch": 0.9274195248434766, "grad_norm": 1.5991934537887573, "learning_rate": 3.83818982479153e-05, "loss": 6.4885, "step": 21997 }, { "epoch": 0.9274616860256761, "grad_norm": 1.5109775066375732, "learning_rate": 3.8381117461507234e-05, "loss": 6.7245, "step": 21998 }, { "epoch": 0.9275038472078757, "grad_norm": 2.2932586669921875, "learning_rate": 3.838033667509916e-05, "loss": 6.9372, "step": 21999 }, { "epoch": 0.9275460083900753, "grad_norm": 1.5058116912841797, "learning_rate": 3.8379555888691096e-05, "loss": 6.9979, "step": 22000 }, { "epoch": 0.9275881695722749, "grad_norm": 3.0864205360412598, "learning_rate": 3.837877510228302e-05, "loss": 7.4976, "step": 22001 }, { "epoch": 0.9276303307544743, "grad_norm": 3.0322134494781494, "learning_rate": 3.837799431587495e-05, "loss": 7.7306, "step": 22002 }, { "epoch": 0.9276724919366739, "grad_norm": 2.3246912956237793, "learning_rate": 3.837721352946688e-05, "loss": 7.7792, "step": 22003 }, { "epoch": 0.9277146531188735, "grad_norm": 3.5634517669677734, "learning_rate": 3.837643274305881e-05, "loss": 7.8913, "step": 22004 }, { "epoch": 0.927756814301073, "grad_norm": 2.5210459232330322, "learning_rate": 3.837565195665074e-05, "loss": 6.4713, "step": 22005 }, { "epoch": 0.9277989754832725, "grad_norm": 2.9960272312164307, "learning_rate": 3.837487117024267e-05, "loss": 7.0848, "step": 22006 }, { "epoch": 0.9278411366654721, "grad_norm": 2.2667455673217773, "learning_rate": 3.83740903838346e-05, "loss": 6.6291, "step": 22007 }, { "epoch": 0.9278832978476717, "grad_norm": 2.1530654430389404, "learning_rate": 3.837330959742653e-05, "loss": 6.5461, "step": 22008 }, { "epoch": 0.9279254590298712, "grad_norm": 1.5165165662765503, "learning_rate": 3.837252881101846e-05, "loss": 6.5963, "step": 22009 }, { "epoch": 0.9279676202120708, "grad_norm": 1.4404852390289307, "learning_rate": 3.837174802461039e-05, "loss": 6.9809, "step": 22010 }, { "epoch": 0.9280097813942703, "grad_norm": 4.058290481567383, "learning_rate": 3.837096723820232e-05, "loss": 8.2349, "step": 22011 }, { "epoch": 0.9280519425764698, "grad_norm": 2.8019514083862305, "learning_rate": 3.837018645179425e-05, "loss": 6.4444, "step": 22012 }, { "epoch": 0.9280941037586694, "grad_norm": 2.641179323196411, "learning_rate": 3.8369405665386175e-05, "loss": 6.6571, "step": 22013 }, { "epoch": 0.928136264940869, "grad_norm": 1.823119878768921, "learning_rate": 3.8368624878978107e-05, "loss": 6.681, "step": 22014 }, { "epoch": 0.9281784261230684, "grad_norm": 1.894079327583313, "learning_rate": 3.836784409257004e-05, "loss": 6.9989, "step": 22015 }, { "epoch": 0.928220587305268, "grad_norm": 1.6247198581695557, "learning_rate": 3.836706330616196e-05, "loss": 6.6491, "step": 22016 }, { "epoch": 0.9282627484874676, "grad_norm": 3.1617214679718018, "learning_rate": 3.83662825197539e-05, "loss": 6.764, "step": 22017 }, { "epoch": 0.9283049096696672, "grad_norm": 2.2682201862335205, "learning_rate": 3.8365501733345825e-05, "loss": 6.6106, "step": 22018 }, { "epoch": 0.9283470708518667, "grad_norm": 2.3354992866516113, "learning_rate": 3.836472094693776e-05, "loss": 6.5008, "step": 22019 }, { "epoch": 0.9283892320340662, "grad_norm": 2.056326389312744, "learning_rate": 3.836394016052969e-05, "loss": 6.7551, "step": 22020 }, { "epoch": 0.9284313932162658, "grad_norm": 1.5111910104751587, "learning_rate": 3.836315937412162e-05, "loss": 6.554, "step": 22021 }, { "epoch": 0.9284735543984654, "grad_norm": 0.9305917620658875, "learning_rate": 3.836237858771355e-05, "loss": 6.5201, "step": 22022 }, { "epoch": 0.9285157155806649, "grad_norm": 1.722785234451294, "learning_rate": 3.8361597801305474e-05, "loss": 6.7523, "step": 22023 }, { "epoch": 0.9285578767628644, "grad_norm": 3.542191505432129, "learning_rate": 3.8360817014897406e-05, "loss": 7.6824, "step": 22024 }, { "epoch": 0.928600037945064, "grad_norm": 2.341832399368286, "learning_rate": 3.836003622848934e-05, "loss": 6.5328, "step": 22025 }, { "epoch": 0.9286421991272635, "grad_norm": 4.435296058654785, "learning_rate": 3.835925544208127e-05, "loss": 7.7817, "step": 22026 }, { "epoch": 0.9286843603094631, "grad_norm": 2.803562641143799, "learning_rate": 3.835847465567319e-05, "loss": 8.0193, "step": 22027 }, { "epoch": 0.9287265214916626, "grad_norm": 2.195164680480957, "learning_rate": 3.8357693869265124e-05, "loss": 7.2007, "step": 22028 }, { "epoch": 0.9287686826738621, "grad_norm": 1.666283130645752, "learning_rate": 3.8356913082857055e-05, "loss": 6.9273, "step": 22029 }, { "epoch": 0.9288108438560617, "grad_norm": 2.1003952026367188, "learning_rate": 3.8356132296448986e-05, "loss": 6.5289, "step": 22030 }, { "epoch": 0.9288530050382613, "grad_norm": 2.84938907623291, "learning_rate": 3.835535151004092e-05, "loss": 7.6304, "step": 22031 }, { "epoch": 0.9288951662204609, "grad_norm": 2.3579018115997314, "learning_rate": 3.835457072363284e-05, "loss": 6.6485, "step": 22032 }, { "epoch": 0.9289373274026603, "grad_norm": 2.4511196613311768, "learning_rate": 3.8353789937224773e-05, "loss": 7.1661, "step": 22033 }, { "epoch": 0.9289794885848599, "grad_norm": 1.451316237449646, "learning_rate": 3.8353009150816705e-05, "loss": 7.3447, "step": 22034 }, { "epoch": 0.9290216497670595, "grad_norm": 2.005898952484131, "learning_rate": 3.835222836440863e-05, "loss": 6.8427, "step": 22035 }, { "epoch": 0.929063810949259, "grad_norm": 1.1278268098831177, "learning_rate": 3.835144757800057e-05, "loss": 6.9269, "step": 22036 }, { "epoch": 0.9291059721314585, "grad_norm": 1.545606017112732, "learning_rate": 3.835066679159249e-05, "loss": 6.9238, "step": 22037 }, { "epoch": 0.9291481333136581, "grad_norm": 1.6031261682510376, "learning_rate": 3.834988600518442e-05, "loss": 6.7989, "step": 22038 }, { "epoch": 0.9291902944958577, "grad_norm": 0.9447096586227417, "learning_rate": 3.8349105218776354e-05, "loss": 6.4348, "step": 22039 }, { "epoch": 0.9292324556780572, "grad_norm": 3.51053786277771, "learning_rate": 3.8348324432368285e-05, "loss": 7.6649, "step": 22040 }, { "epoch": 0.9292746168602568, "grad_norm": 3.3296213150024414, "learning_rate": 3.834754364596022e-05, "loss": 7.8374, "step": 22041 }, { "epoch": 0.9293167780424563, "grad_norm": 0.9958630204200745, "learning_rate": 3.834676285955214e-05, "loss": 6.904, "step": 22042 }, { "epoch": 0.9293589392246558, "grad_norm": 1.1637059450149536, "learning_rate": 3.834598207314407e-05, "loss": 6.774, "step": 22043 }, { "epoch": 0.9294011004068554, "grad_norm": 1.8424668312072754, "learning_rate": 3.8345201286736004e-05, "loss": 7.1123, "step": 22044 }, { "epoch": 0.929443261589055, "grad_norm": 1.9786492586135864, "learning_rate": 3.8344420500327935e-05, "loss": 6.672, "step": 22045 }, { "epoch": 0.9294854227712545, "grad_norm": 2.4758810997009277, "learning_rate": 3.834363971391986e-05, "loss": 6.4837, "step": 22046 }, { "epoch": 0.929527583953454, "grad_norm": 2.7040557861328125, "learning_rate": 3.834285892751179e-05, "loss": 7.3965, "step": 22047 }, { "epoch": 0.9295697451356536, "grad_norm": 1.4902716875076294, "learning_rate": 3.834207814110372e-05, "loss": 6.627, "step": 22048 }, { "epoch": 0.9296119063178532, "grad_norm": 1.493716835975647, "learning_rate": 3.8341297354695646e-05, "loss": 7.0248, "step": 22049 }, { "epoch": 0.9296540675000527, "grad_norm": 3.859656572341919, "learning_rate": 3.8340516568287585e-05, "loss": 7.7085, "step": 22050 }, { "epoch": 0.9296962286822522, "grad_norm": 3.5879271030426025, "learning_rate": 3.833973578187951e-05, "loss": 6.479, "step": 22051 }, { "epoch": 0.9297383898644518, "grad_norm": 0.8370879888534546, "learning_rate": 3.833895499547144e-05, "loss": 6.4512, "step": 22052 }, { "epoch": 0.9297805510466514, "grad_norm": 1.1774500608444214, "learning_rate": 3.833817420906337e-05, "loss": 6.9892, "step": 22053 }, { "epoch": 0.9298227122288509, "grad_norm": 2.8546392917633057, "learning_rate": 3.8337393422655296e-05, "loss": 6.6063, "step": 22054 }, { "epoch": 0.9298648734110504, "grad_norm": 2.5039823055267334, "learning_rate": 3.8336612636247234e-05, "loss": 7.9248, "step": 22055 }, { "epoch": 0.92990703459325, "grad_norm": 2.1549248695373535, "learning_rate": 3.833583184983916e-05, "loss": 7.1902, "step": 22056 }, { "epoch": 0.9299491957754495, "grad_norm": 2.19728946685791, "learning_rate": 3.833505106343109e-05, "loss": 6.5935, "step": 22057 }, { "epoch": 0.9299913569576491, "grad_norm": 1.3733848333358765, "learning_rate": 3.833427027702302e-05, "loss": 7.2362, "step": 22058 }, { "epoch": 0.9300335181398487, "grad_norm": 1.153516411781311, "learning_rate": 3.8333489490614946e-05, "loss": 6.8195, "step": 22059 }, { "epoch": 0.9300756793220482, "grad_norm": 2.552903413772583, "learning_rate": 3.833270870420688e-05, "loss": 6.4262, "step": 22060 }, { "epoch": 0.9301178405042477, "grad_norm": 2.4279403686523438, "learning_rate": 3.833192791779881e-05, "loss": 7.1167, "step": 22061 }, { "epoch": 0.9301600016864473, "grad_norm": 2.5608417987823486, "learning_rate": 3.833114713139074e-05, "loss": 7.4032, "step": 22062 }, { "epoch": 0.9302021628686469, "grad_norm": 1.4849902391433716, "learning_rate": 3.833036634498267e-05, "loss": 6.678, "step": 22063 }, { "epoch": 0.9302443240508463, "grad_norm": 4.351219177246094, "learning_rate": 3.83295855585746e-05, "loss": 7.6613, "step": 22064 }, { "epoch": 0.9302864852330459, "grad_norm": 2.1123721599578857, "learning_rate": 3.8328804772166526e-05, "loss": 6.6667, "step": 22065 }, { "epoch": 0.9303286464152455, "grad_norm": 1.5219248533248901, "learning_rate": 3.832802398575846e-05, "loss": 6.6392, "step": 22066 }, { "epoch": 0.9303708075974451, "grad_norm": 2.9583778381347656, "learning_rate": 3.832724319935039e-05, "loss": 7.5204, "step": 22067 }, { "epoch": 0.9304129687796446, "grad_norm": 3.3853132724761963, "learning_rate": 3.832646241294231e-05, "loss": 8.0041, "step": 22068 }, { "epoch": 0.9304551299618441, "grad_norm": 1.912239670753479, "learning_rate": 3.832568162653425e-05, "loss": 6.877, "step": 22069 }, { "epoch": 0.9304972911440437, "grad_norm": 1.4268029928207397, "learning_rate": 3.8324900840126176e-05, "loss": 6.6311, "step": 22070 }, { "epoch": 0.9305394523262432, "grad_norm": 1.7261667251586914, "learning_rate": 3.832412005371811e-05, "loss": 6.5875, "step": 22071 }, { "epoch": 0.9305816135084428, "grad_norm": 2.7352871894836426, "learning_rate": 3.832333926731004e-05, "loss": 7.5935, "step": 22072 }, { "epoch": 0.9306237746906423, "grad_norm": 1.8412081003189087, "learning_rate": 3.832255848090196e-05, "loss": 7.6284, "step": 22073 }, { "epoch": 0.9306659358728419, "grad_norm": 2.659212112426758, "learning_rate": 3.8321777694493894e-05, "loss": 8.4819, "step": 22074 }, { "epoch": 0.9307080970550414, "grad_norm": 0.8001270890235901, "learning_rate": 3.8320996908085825e-05, "loss": 6.4622, "step": 22075 }, { "epoch": 0.930750258237241, "grad_norm": 1.7548587322235107, "learning_rate": 3.832021612167776e-05, "loss": 6.999, "step": 22076 }, { "epoch": 0.9307924194194406, "grad_norm": 1.1060271263122559, "learning_rate": 3.831943533526969e-05, "loss": 6.6344, "step": 22077 }, { "epoch": 0.93083458060164, "grad_norm": 3.2372817993164062, "learning_rate": 3.831865454886161e-05, "loss": 7.8208, "step": 22078 }, { "epoch": 0.9308767417838396, "grad_norm": 1.2794982194900513, "learning_rate": 3.8317873762453544e-05, "loss": 6.7533, "step": 22079 }, { "epoch": 0.9309189029660392, "grad_norm": 1.5608083009719849, "learning_rate": 3.8317092976045475e-05, "loss": 6.5611, "step": 22080 }, { "epoch": 0.9309610641482388, "grad_norm": 2.292180299758911, "learning_rate": 3.8316312189637406e-05, "loss": 7.229, "step": 22081 }, { "epoch": 0.9310032253304382, "grad_norm": 1.5160311460494995, "learning_rate": 3.831553140322933e-05, "loss": 7.9394, "step": 22082 }, { "epoch": 0.9310453865126378, "grad_norm": 1.3932310342788696, "learning_rate": 3.831475061682127e-05, "loss": 6.6124, "step": 22083 }, { "epoch": 0.9310875476948374, "grad_norm": 3.569898843765259, "learning_rate": 3.831396983041319e-05, "loss": 6.943, "step": 22084 }, { "epoch": 0.9311297088770369, "grad_norm": 3.469982147216797, "learning_rate": 3.8313189044005124e-05, "loss": 7.3367, "step": 22085 }, { "epoch": 0.9311718700592364, "grad_norm": 4.479605674743652, "learning_rate": 3.8312408257597056e-05, "loss": 7.6362, "step": 22086 }, { "epoch": 0.931214031241436, "grad_norm": 1.2042406797409058, "learning_rate": 3.831162747118898e-05, "loss": 6.7692, "step": 22087 }, { "epoch": 0.9312561924236356, "grad_norm": 1.3633902072906494, "learning_rate": 3.831084668478092e-05, "loss": 6.4285, "step": 22088 }, { "epoch": 0.9312983536058351, "grad_norm": 2.0627787113189697, "learning_rate": 3.831006589837284e-05, "loss": 7.0195, "step": 22089 }, { "epoch": 0.9313405147880347, "grad_norm": 3.7832329273223877, "learning_rate": 3.8309285111964774e-05, "loss": 7.6935, "step": 22090 }, { "epoch": 0.9313826759702342, "grad_norm": 2.345494031906128, "learning_rate": 3.8308504325556705e-05, "loss": 7.2555, "step": 22091 }, { "epoch": 0.9314248371524337, "grad_norm": 4.176864147186279, "learning_rate": 3.830772353914863e-05, "loss": 6.987, "step": 22092 }, { "epoch": 0.9314669983346333, "grad_norm": 1.5088573694229126, "learning_rate": 3.830694275274056e-05, "loss": 6.8009, "step": 22093 }, { "epoch": 0.9315091595168329, "grad_norm": 1.2801085710525513, "learning_rate": 3.830616196633249e-05, "loss": 6.6143, "step": 22094 }, { "epoch": 0.9315513206990323, "grad_norm": 2.445277452468872, "learning_rate": 3.8305381179924423e-05, "loss": 7.0815, "step": 22095 }, { "epoch": 0.9315934818812319, "grad_norm": 3.735215902328491, "learning_rate": 3.830460039351635e-05, "loss": 7.7981, "step": 22096 }, { "epoch": 0.9316356430634315, "grad_norm": 1.895569086074829, "learning_rate": 3.830381960710828e-05, "loss": 6.9238, "step": 22097 }, { "epoch": 0.9316778042456311, "grad_norm": 1.4225544929504395, "learning_rate": 3.830303882070021e-05, "loss": 6.4533, "step": 22098 }, { "epoch": 0.9317199654278306, "grad_norm": 1.232500433921814, "learning_rate": 3.830225803429214e-05, "loss": 7.1591, "step": 22099 }, { "epoch": 0.9317621266100301, "grad_norm": 2.7898290157318115, "learning_rate": 3.830147724788407e-05, "loss": 7.7294, "step": 22100 }, { "epoch": 0.9318042877922297, "grad_norm": 2.831501007080078, "learning_rate": 3.8300696461476e-05, "loss": 7.67, "step": 22101 }, { "epoch": 0.9318464489744293, "grad_norm": 3.4534850120544434, "learning_rate": 3.8299915675067935e-05, "loss": 7.6374, "step": 22102 }, { "epoch": 0.9318886101566288, "grad_norm": 1.7545154094696045, "learning_rate": 3.829913488865986e-05, "loss": 6.54, "step": 22103 }, { "epoch": 0.9319307713388283, "grad_norm": 2.2913265228271484, "learning_rate": 3.8298354102251784e-05, "loss": 6.5089, "step": 22104 }, { "epoch": 0.9319729325210279, "grad_norm": 3.81341814994812, "learning_rate": 3.829757331584372e-05, "loss": 7.4818, "step": 22105 }, { "epoch": 0.9320150937032274, "grad_norm": 1.2404109239578247, "learning_rate": 3.829679252943565e-05, "loss": 7.132, "step": 22106 }, { "epoch": 0.932057254885427, "grad_norm": 1.381909728050232, "learning_rate": 3.829601174302758e-05, "loss": 6.7228, "step": 22107 }, { "epoch": 0.9320994160676266, "grad_norm": 4.625416278839111, "learning_rate": 3.829523095661951e-05, "loss": 8.2568, "step": 22108 }, { "epoch": 0.932141577249826, "grad_norm": 2.09576153755188, "learning_rate": 3.829445017021144e-05, "loss": 7.1971, "step": 22109 }, { "epoch": 0.9321837384320256, "grad_norm": 2.216209888458252, "learning_rate": 3.829366938380337e-05, "loss": 6.5141, "step": 22110 }, { "epoch": 0.9322258996142252, "grad_norm": 1.633594274520874, "learning_rate": 3.8292888597395297e-05, "loss": 6.3962, "step": 22111 }, { "epoch": 0.9322680607964248, "grad_norm": 2.1890673637390137, "learning_rate": 3.829210781098723e-05, "loss": 6.7974, "step": 22112 }, { "epoch": 0.9323102219786242, "grad_norm": 2.5165650844573975, "learning_rate": 3.829132702457916e-05, "loss": 7.3096, "step": 22113 }, { "epoch": 0.9323523831608238, "grad_norm": 1.363759160041809, "learning_rate": 3.829054623817109e-05, "loss": 6.4664, "step": 22114 }, { "epoch": 0.9323945443430234, "grad_norm": 1.5690548419952393, "learning_rate": 3.8289765451763015e-05, "loss": 7.0821, "step": 22115 }, { "epoch": 0.932436705525223, "grad_norm": 1.2304863929748535, "learning_rate": 3.8288984665354946e-05, "loss": 6.953, "step": 22116 }, { "epoch": 0.9324788667074225, "grad_norm": 2.934720039367676, "learning_rate": 3.828820387894688e-05, "loss": 8.1224, "step": 22117 }, { "epoch": 0.932521027889622, "grad_norm": 2.8150737285614014, "learning_rate": 3.82874230925388e-05, "loss": 7.6689, "step": 22118 }, { "epoch": 0.9325631890718216, "grad_norm": 1.0949190855026245, "learning_rate": 3.828664230613074e-05, "loss": 6.8561, "step": 22119 }, { "epoch": 0.9326053502540211, "grad_norm": 1.171372890472412, "learning_rate": 3.8285861519722664e-05, "loss": 6.5787, "step": 22120 }, { "epoch": 0.9326475114362207, "grad_norm": 3.775700092315674, "learning_rate": 3.8285080733314596e-05, "loss": 7.6415, "step": 22121 }, { "epoch": 0.9326896726184202, "grad_norm": 1.6578341722488403, "learning_rate": 3.828429994690653e-05, "loss": 6.525, "step": 22122 }, { "epoch": 0.9327318338006197, "grad_norm": 0.8797221183776855, "learning_rate": 3.828351916049845e-05, "loss": 6.88, "step": 22123 }, { "epoch": 0.9327739949828193, "grad_norm": 3.128831386566162, "learning_rate": 3.828273837409039e-05, "loss": 7.965, "step": 22124 }, { "epoch": 0.9328161561650189, "grad_norm": 1.8311820030212402, "learning_rate": 3.8281957587682314e-05, "loss": 6.8841, "step": 22125 }, { "epoch": 0.9328583173472185, "grad_norm": 1.8286354541778564, "learning_rate": 3.8281176801274245e-05, "loss": 6.7835, "step": 22126 }, { "epoch": 0.9329004785294179, "grad_norm": 1.750769019126892, "learning_rate": 3.8280396014866176e-05, "loss": 6.4161, "step": 22127 }, { "epoch": 0.9329426397116175, "grad_norm": 3.5436668395996094, "learning_rate": 3.827961522845811e-05, "loss": 7.7046, "step": 22128 }, { "epoch": 0.9329848008938171, "grad_norm": 1.4890213012695312, "learning_rate": 3.827883444205003e-05, "loss": 7.0423, "step": 22129 }, { "epoch": 0.9330269620760167, "grad_norm": 0.7989181876182556, "learning_rate": 3.827805365564196e-05, "loss": 6.6397, "step": 22130 }, { "epoch": 0.9330691232582161, "grad_norm": 1.1659091711044312, "learning_rate": 3.8277272869233895e-05, "loss": 6.5945, "step": 22131 }, { "epoch": 0.9331112844404157, "grad_norm": 1.071334719657898, "learning_rate": 3.8276492082825826e-05, "loss": 6.6314, "step": 22132 }, { "epoch": 0.9331534456226153, "grad_norm": 1.389841914176941, "learning_rate": 3.827571129641776e-05, "loss": 6.5876, "step": 22133 }, { "epoch": 0.9331956068048148, "grad_norm": 3.537123441696167, "learning_rate": 3.827493051000968e-05, "loss": 8.9404, "step": 22134 }, { "epoch": 0.9332377679870144, "grad_norm": 0.8014655113220215, "learning_rate": 3.827414972360161e-05, "loss": 6.5252, "step": 22135 }, { "epoch": 0.9332799291692139, "grad_norm": 0.9501339793205261, "learning_rate": 3.8273368937193544e-05, "loss": 6.9694, "step": 22136 }, { "epoch": 0.9333220903514134, "grad_norm": 1.3797836303710938, "learning_rate": 3.827258815078547e-05, "loss": 7.1096, "step": 22137 }, { "epoch": 0.933364251533613, "grad_norm": 2.33318829536438, "learning_rate": 3.827180736437741e-05, "loss": 6.7101, "step": 22138 }, { "epoch": 0.9334064127158126, "grad_norm": 1.386484146118164, "learning_rate": 3.827102657796933e-05, "loss": 6.7048, "step": 22139 }, { "epoch": 0.9334485738980121, "grad_norm": 2.5888359546661377, "learning_rate": 3.827024579156126e-05, "loss": 8.8644, "step": 22140 }, { "epoch": 0.9334907350802116, "grad_norm": 1.1636947393417358, "learning_rate": 3.8269465005153194e-05, "loss": 6.4874, "step": 22141 }, { "epoch": 0.9335328962624112, "grad_norm": 1.5184985399246216, "learning_rate": 3.826868421874512e-05, "loss": 7.1377, "step": 22142 }, { "epoch": 0.9335750574446108, "grad_norm": 2.036000967025757, "learning_rate": 3.8267903432337056e-05, "loss": 7.6372, "step": 22143 }, { "epoch": 0.9336172186268102, "grad_norm": 0.8814712166786194, "learning_rate": 3.826712264592898e-05, "loss": 6.936, "step": 22144 }, { "epoch": 0.9336593798090098, "grad_norm": 1.4037305116653442, "learning_rate": 3.826634185952091e-05, "loss": 6.5701, "step": 22145 }, { "epoch": 0.9337015409912094, "grad_norm": 1.1981528997421265, "learning_rate": 3.826556107311284e-05, "loss": 6.6943, "step": 22146 }, { "epoch": 0.933743702173409, "grad_norm": 1.793269395828247, "learning_rate": 3.8264780286704774e-05, "loss": 6.7731, "step": 22147 }, { "epoch": 0.9337858633556085, "grad_norm": 3.298924446105957, "learning_rate": 3.82639995002967e-05, "loss": 7.8045, "step": 22148 }, { "epoch": 0.933828024537808, "grad_norm": 2.6622934341430664, "learning_rate": 3.826321871388863e-05, "loss": 7.6929, "step": 22149 }, { "epoch": 0.9338701857200076, "grad_norm": 4.652163505554199, "learning_rate": 3.826243792748056e-05, "loss": 7.7284, "step": 22150 }, { "epoch": 0.9339123469022071, "grad_norm": 1.4170042276382446, "learning_rate": 3.8261657141072486e-05, "loss": 6.5915, "step": 22151 }, { "epoch": 0.9339545080844067, "grad_norm": 0.9923080205917358, "learning_rate": 3.8260876354664424e-05, "loss": 6.4369, "step": 22152 }, { "epoch": 0.9339966692666062, "grad_norm": 1.3351651430130005, "learning_rate": 3.826009556825635e-05, "loss": 6.5286, "step": 22153 }, { "epoch": 0.9340388304488058, "grad_norm": 1.140662670135498, "learning_rate": 3.825931478184828e-05, "loss": 6.7119, "step": 22154 }, { "epoch": 0.9340809916310053, "grad_norm": 2.9847464561462402, "learning_rate": 3.825853399544021e-05, "loss": 7.2456, "step": 22155 }, { "epoch": 0.9341231528132049, "grad_norm": 0.7979452013969421, "learning_rate": 3.8257753209032135e-05, "loss": 6.6154, "step": 22156 }, { "epoch": 0.9341653139954045, "grad_norm": 3.720210552215576, "learning_rate": 3.8256972422624073e-05, "loss": 7.4307, "step": 22157 }, { "epoch": 0.9342074751776039, "grad_norm": 1.2000998258590698, "learning_rate": 3.8256191636216e-05, "loss": 6.4405, "step": 22158 }, { "epoch": 0.9342496363598035, "grad_norm": 3.176811456680298, "learning_rate": 3.825541084980793e-05, "loss": 7.4431, "step": 22159 }, { "epoch": 0.9342917975420031, "grad_norm": 1.5318868160247803, "learning_rate": 3.825463006339986e-05, "loss": 6.4382, "step": 22160 }, { "epoch": 0.9343339587242027, "grad_norm": 1.5269345045089722, "learning_rate": 3.8253849276991785e-05, "loss": 6.4259, "step": 22161 }, { "epoch": 0.9343761199064021, "grad_norm": 1.3061953783035278, "learning_rate": 3.8253068490583716e-05, "loss": 6.5487, "step": 22162 }, { "epoch": 0.9344182810886017, "grad_norm": 3.0846736431121826, "learning_rate": 3.825228770417565e-05, "loss": 7.6721, "step": 22163 }, { "epoch": 0.9344604422708013, "grad_norm": 1.5444267988204956, "learning_rate": 3.825150691776758e-05, "loss": 7.3, "step": 22164 }, { "epoch": 0.9345026034530008, "grad_norm": 1.225520372390747, "learning_rate": 3.82507261313595e-05, "loss": 7.6236, "step": 22165 }, { "epoch": 0.9345447646352004, "grad_norm": 1.521456003189087, "learning_rate": 3.8249945344951434e-05, "loss": 7.2635, "step": 22166 }, { "epoch": 0.9345869258173999, "grad_norm": 1.2022546529769897, "learning_rate": 3.8249164558543366e-05, "loss": 6.6803, "step": 22167 }, { "epoch": 0.9346290869995995, "grad_norm": 1.6612550020217896, "learning_rate": 3.82483837721353e-05, "loss": 7.0634, "step": 22168 }, { "epoch": 0.934671248181799, "grad_norm": 3.282315492630005, "learning_rate": 3.824760298572723e-05, "loss": 7.4541, "step": 22169 }, { "epoch": 0.9347134093639986, "grad_norm": 3.267202615737915, "learning_rate": 3.824682219931915e-05, "loss": 7.8068, "step": 22170 }, { "epoch": 0.9347555705461981, "grad_norm": 3.632634162902832, "learning_rate": 3.824604141291109e-05, "loss": 7.6878, "step": 22171 }, { "epoch": 0.9347977317283976, "grad_norm": 3.493624687194824, "learning_rate": 3.8245260626503015e-05, "loss": 7.7759, "step": 22172 }, { "epoch": 0.9348398929105972, "grad_norm": 1.4246530532836914, "learning_rate": 3.8244479840094947e-05, "loss": 7.1036, "step": 22173 }, { "epoch": 0.9348820540927968, "grad_norm": 2.2430572509765625, "learning_rate": 3.824369905368688e-05, "loss": 7.2357, "step": 22174 }, { "epoch": 0.9349242152749964, "grad_norm": 1.731938362121582, "learning_rate": 3.82429182672788e-05, "loss": 6.6433, "step": 22175 }, { "epoch": 0.9349663764571958, "grad_norm": 2.195913553237915, "learning_rate": 3.8242137480870734e-05, "loss": 6.5474, "step": 22176 }, { "epoch": 0.9350085376393954, "grad_norm": 2.2924206256866455, "learning_rate": 3.8241356694462665e-05, "loss": 7.5959, "step": 22177 }, { "epoch": 0.935050698821595, "grad_norm": 1.2986958026885986, "learning_rate": 3.8240575908054596e-05, "loss": 7.0547, "step": 22178 }, { "epoch": 0.9350928600037945, "grad_norm": 1.1120506525039673, "learning_rate": 3.823979512164653e-05, "loss": 6.8827, "step": 22179 }, { "epoch": 0.935135021185994, "grad_norm": 1.6587848663330078, "learning_rate": 3.823901433523845e-05, "loss": 6.842, "step": 22180 }, { "epoch": 0.9351771823681936, "grad_norm": 2.0818464756011963, "learning_rate": 3.823823354883038e-05, "loss": 6.6783, "step": 22181 }, { "epoch": 0.9352193435503932, "grad_norm": 1.792763113975525, "learning_rate": 3.8237452762422314e-05, "loss": 6.8415, "step": 22182 }, { "epoch": 0.9352615047325927, "grad_norm": 3.1737236976623535, "learning_rate": 3.8236671976014246e-05, "loss": 7.2921, "step": 22183 }, { "epoch": 0.9353036659147923, "grad_norm": 1.1442313194274902, "learning_rate": 3.823589118960617e-05, "loss": 6.7125, "step": 22184 }, { "epoch": 0.9353458270969918, "grad_norm": 2.833725929260254, "learning_rate": 3.82351104031981e-05, "loss": 7.1672, "step": 22185 }, { "epoch": 0.9353879882791913, "grad_norm": 1.1651803255081177, "learning_rate": 3.823432961679003e-05, "loss": 6.7333, "step": 22186 }, { "epoch": 0.9354301494613909, "grad_norm": 1.3900266885757446, "learning_rate": 3.823354883038196e-05, "loss": 6.6289, "step": 22187 }, { "epoch": 0.9354723106435905, "grad_norm": 1.5094531774520874, "learning_rate": 3.8232768043973895e-05, "loss": 6.4221, "step": 22188 }, { "epoch": 0.93551447182579, "grad_norm": 1.6286638975143433, "learning_rate": 3.823198725756582e-05, "loss": 7.3828, "step": 22189 }, { "epoch": 0.9355566330079895, "grad_norm": 3.0258004665374756, "learning_rate": 3.823120647115776e-05, "loss": 7.6133, "step": 22190 }, { "epoch": 0.9355987941901891, "grad_norm": 0.8710245490074158, "learning_rate": 3.823042568474968e-05, "loss": 6.8557, "step": 22191 }, { "epoch": 0.9356409553723887, "grad_norm": 3.1848435401916504, "learning_rate": 3.822964489834161e-05, "loss": 8.0857, "step": 22192 }, { "epoch": 0.9356831165545882, "grad_norm": 2.490648031234741, "learning_rate": 3.8228864111933545e-05, "loss": 6.4459, "step": 22193 }, { "epoch": 0.9357252777367877, "grad_norm": 3.353593587875366, "learning_rate": 3.822808332552547e-05, "loss": 7.7113, "step": 22194 }, { "epoch": 0.9357674389189873, "grad_norm": 2.3086297512054443, "learning_rate": 3.82273025391174e-05, "loss": 6.7321, "step": 22195 }, { "epoch": 0.9358096001011869, "grad_norm": 1.486194133758545, "learning_rate": 3.822652175270933e-05, "loss": 6.4373, "step": 22196 }, { "epoch": 0.9358517612833864, "grad_norm": 1.5838708877563477, "learning_rate": 3.822574096630126e-05, "loss": 7.1014, "step": 22197 }, { "epoch": 0.9358939224655859, "grad_norm": 1.328542709350586, "learning_rate": 3.822496017989319e-05, "loss": 6.6752, "step": 22198 }, { "epoch": 0.9359360836477855, "grad_norm": 1.2846204042434692, "learning_rate": 3.822417939348512e-05, "loss": 6.5139, "step": 22199 }, { "epoch": 0.935978244829985, "grad_norm": 1.165569543838501, "learning_rate": 3.822339860707705e-05, "loss": 6.7989, "step": 22200 }, { "epoch": 0.9360204060121846, "grad_norm": 1.988750696182251, "learning_rate": 3.822261782066898e-05, "loss": 7.0831, "step": 22201 }, { "epoch": 0.9360625671943841, "grad_norm": 1.4357376098632812, "learning_rate": 3.822183703426091e-05, "loss": 7.1185, "step": 22202 }, { "epoch": 0.9361047283765837, "grad_norm": 0.8849081993103027, "learning_rate": 3.822105624785284e-05, "loss": 6.6812, "step": 22203 }, { "epoch": 0.9361468895587832, "grad_norm": 3.288243293762207, "learning_rate": 3.822027546144477e-05, "loss": 7.9837, "step": 22204 }, { "epoch": 0.9361890507409828, "grad_norm": 0.9018632769584656, "learning_rate": 3.82194946750367e-05, "loss": 6.6965, "step": 22205 }, { "epoch": 0.9362312119231824, "grad_norm": 1.6720693111419678, "learning_rate": 3.8218713888628624e-05, "loss": 6.3855, "step": 22206 }, { "epoch": 0.9362733731053818, "grad_norm": 0.985428512096405, "learning_rate": 3.821793310222056e-05, "loss": 6.6956, "step": 22207 }, { "epoch": 0.9363155342875814, "grad_norm": 2.650055170059204, "learning_rate": 3.8217152315812486e-05, "loss": 7.0054, "step": 22208 }, { "epoch": 0.936357695469781, "grad_norm": 2.40446138381958, "learning_rate": 3.821637152940442e-05, "loss": 7.533, "step": 22209 }, { "epoch": 0.9363998566519806, "grad_norm": 1.1491732597351074, "learning_rate": 3.821559074299635e-05, "loss": 6.5058, "step": 22210 }, { "epoch": 0.93644201783418, "grad_norm": 1.8859455585479736, "learning_rate": 3.8214809956588273e-05, "loss": 6.9818, "step": 22211 }, { "epoch": 0.9364841790163796, "grad_norm": 1.0655232667922974, "learning_rate": 3.821402917018021e-05, "loss": 6.8475, "step": 22212 }, { "epoch": 0.9365263401985792, "grad_norm": 1.6706266403198242, "learning_rate": 3.8213248383772136e-05, "loss": 6.962, "step": 22213 }, { "epoch": 0.9365685013807787, "grad_norm": 1.2574042081832886, "learning_rate": 3.821246759736407e-05, "loss": 7.0413, "step": 22214 }, { "epoch": 0.9366106625629783, "grad_norm": 0.8699590563774109, "learning_rate": 3.8211686810956e-05, "loss": 6.7378, "step": 22215 }, { "epoch": 0.9366528237451778, "grad_norm": 1.0264341831207275, "learning_rate": 3.821090602454793e-05, "loss": 6.7523, "step": 22216 }, { "epoch": 0.9366949849273773, "grad_norm": 2.534550905227661, "learning_rate": 3.8210125238139854e-05, "loss": 7.5228, "step": 22217 }, { "epoch": 0.9367371461095769, "grad_norm": 1.2219884395599365, "learning_rate": 3.8209344451731785e-05, "loss": 6.5407, "step": 22218 }, { "epoch": 0.9367793072917765, "grad_norm": 1.2769166231155396, "learning_rate": 3.820856366532372e-05, "loss": 6.6532, "step": 22219 }, { "epoch": 0.936821468473976, "grad_norm": 1.432570457458496, "learning_rate": 3.820778287891564e-05, "loss": 6.6571, "step": 22220 }, { "epoch": 0.9368636296561755, "grad_norm": 0.7990376353263855, "learning_rate": 3.820700209250758e-05, "loss": 6.4785, "step": 22221 }, { "epoch": 0.9369057908383751, "grad_norm": 0.787766695022583, "learning_rate": 3.8206221306099504e-05, "loss": 6.5817, "step": 22222 }, { "epoch": 0.9369479520205747, "grad_norm": 2.4390347003936768, "learning_rate": 3.8205440519691435e-05, "loss": 7.0569, "step": 22223 }, { "epoch": 0.9369901132027743, "grad_norm": 1.5503805875778198, "learning_rate": 3.8204659733283366e-05, "loss": 6.6832, "step": 22224 }, { "epoch": 0.9370322743849737, "grad_norm": 1.678337812423706, "learning_rate": 3.820387894687529e-05, "loss": 6.502, "step": 22225 }, { "epoch": 0.9370744355671733, "grad_norm": 0.998681366443634, "learning_rate": 3.820309816046723e-05, "loss": 6.7233, "step": 22226 }, { "epoch": 0.9371165967493729, "grad_norm": 0.8559120297431946, "learning_rate": 3.820231737405915e-05, "loss": 6.613, "step": 22227 }, { "epoch": 0.9371587579315724, "grad_norm": 1.5455573797225952, "learning_rate": 3.8201536587651084e-05, "loss": 7.2001, "step": 22228 }, { "epoch": 0.9372009191137719, "grad_norm": 1.527658462524414, "learning_rate": 3.8200755801243016e-05, "loss": 6.9171, "step": 22229 }, { "epoch": 0.9372430802959715, "grad_norm": 1.0288410186767578, "learning_rate": 3.819997501483494e-05, "loss": 7.0146, "step": 22230 }, { "epoch": 0.937285241478171, "grad_norm": 0.9719711542129517, "learning_rate": 3.819919422842687e-05, "loss": 6.9872, "step": 22231 }, { "epoch": 0.9373274026603706, "grad_norm": 2.4262781143188477, "learning_rate": 3.81984134420188e-05, "loss": 7.6603, "step": 22232 }, { "epoch": 0.9373695638425702, "grad_norm": 0.9997983574867249, "learning_rate": 3.8197632655610734e-05, "loss": 6.8691, "step": 22233 }, { "epoch": 0.9374117250247697, "grad_norm": 2.985045909881592, "learning_rate": 3.819685186920266e-05, "loss": 7.2977, "step": 22234 }, { "epoch": 0.9374538862069692, "grad_norm": 0.7721250653266907, "learning_rate": 3.8196071082794597e-05, "loss": 6.5072, "step": 22235 }, { "epoch": 0.9374960473891688, "grad_norm": 1.8710596561431885, "learning_rate": 3.819529029638652e-05, "loss": 6.496, "step": 22236 }, { "epoch": 0.9375382085713684, "grad_norm": 2.214130163192749, "learning_rate": 3.819450950997845e-05, "loss": 7.9711, "step": 22237 }, { "epoch": 0.9375803697535678, "grad_norm": 0.9125571250915527, "learning_rate": 3.8193728723570384e-05, "loss": 6.4592, "step": 22238 }, { "epoch": 0.9376225309357674, "grad_norm": 1.0481235980987549, "learning_rate": 3.819294793716231e-05, "loss": 6.9813, "step": 22239 }, { "epoch": 0.937664692117967, "grad_norm": 2.364213705062866, "learning_rate": 3.8192167150754246e-05, "loss": 7.1167, "step": 22240 }, { "epoch": 0.9377068533001666, "grad_norm": 3.6428816318511963, "learning_rate": 3.819138636434617e-05, "loss": 8.0357, "step": 22241 }, { "epoch": 0.9377490144823661, "grad_norm": 1.9059584140777588, "learning_rate": 3.81906055779381e-05, "loss": 7.2877, "step": 22242 }, { "epoch": 0.9377911756645656, "grad_norm": 1.0533092021942139, "learning_rate": 3.818982479153003e-05, "loss": 6.6038, "step": 22243 }, { "epoch": 0.9378333368467652, "grad_norm": 3.2715489864349365, "learning_rate": 3.818904400512196e-05, "loss": 7.4123, "step": 22244 }, { "epoch": 0.9378754980289647, "grad_norm": 1.419510841369629, "learning_rate": 3.818826321871389e-05, "loss": 6.9085, "step": 22245 }, { "epoch": 0.9379176592111643, "grad_norm": 2.9353346824645996, "learning_rate": 3.818748243230582e-05, "loss": 7.8196, "step": 22246 }, { "epoch": 0.9379598203933638, "grad_norm": 1.0027799606323242, "learning_rate": 3.818670164589775e-05, "loss": 6.4557, "step": 22247 }, { "epoch": 0.9380019815755634, "grad_norm": 1.0564218759536743, "learning_rate": 3.818592085948968e-05, "loss": 6.4951, "step": 22248 }, { "epoch": 0.9380441427577629, "grad_norm": 3.6032183170318604, "learning_rate": 3.818514007308161e-05, "loss": 8.084, "step": 22249 }, { "epoch": 0.9380863039399625, "grad_norm": 3.329744815826416, "learning_rate": 3.818435928667354e-05, "loss": 7.295, "step": 22250 }, { "epoch": 0.9381284651221621, "grad_norm": 2.0925984382629395, "learning_rate": 3.818357850026547e-05, "loss": 6.8873, "step": 22251 }, { "epoch": 0.9381706263043615, "grad_norm": 1.5607585906982422, "learning_rate": 3.81827977138574e-05, "loss": 6.4979, "step": 22252 }, { "epoch": 0.9382127874865611, "grad_norm": 2.757925510406494, "learning_rate": 3.8182016927449325e-05, "loss": 6.4251, "step": 22253 }, { "epoch": 0.9382549486687607, "grad_norm": 1.364263653755188, "learning_rate": 3.818123614104126e-05, "loss": 6.5895, "step": 22254 }, { "epoch": 0.9382971098509603, "grad_norm": 1.4690155982971191, "learning_rate": 3.818045535463319e-05, "loss": 6.9727, "step": 22255 }, { "epoch": 0.9383392710331597, "grad_norm": 1.805475115776062, "learning_rate": 3.817967456822511e-05, "loss": 7.1417, "step": 22256 }, { "epoch": 0.9383814322153593, "grad_norm": 1.840659737586975, "learning_rate": 3.817889378181705e-05, "loss": 7.0581, "step": 22257 }, { "epoch": 0.9384235933975589, "grad_norm": 1.47542405128479, "learning_rate": 3.8178112995408975e-05, "loss": 6.8931, "step": 22258 }, { "epoch": 0.9384657545797584, "grad_norm": 1.900275707244873, "learning_rate": 3.817733220900091e-05, "loss": 7.0134, "step": 22259 }, { "epoch": 0.9385079157619579, "grad_norm": 1.680717945098877, "learning_rate": 3.817655142259284e-05, "loss": 7.1919, "step": 22260 }, { "epoch": 0.9385500769441575, "grad_norm": 1.2603808641433716, "learning_rate": 3.817577063618477e-05, "loss": 6.735, "step": 22261 }, { "epoch": 0.9385922381263571, "grad_norm": 4.308348655700684, "learning_rate": 3.81749898497767e-05, "loss": 7.6556, "step": 22262 }, { "epoch": 0.9386343993085566, "grad_norm": 1.3755509853363037, "learning_rate": 3.8174209063368624e-05, "loss": 6.9187, "step": 22263 }, { "epoch": 0.9386765604907562, "grad_norm": 1.5104255676269531, "learning_rate": 3.8173428276960556e-05, "loss": 6.9707, "step": 22264 }, { "epoch": 0.9387187216729557, "grad_norm": 1.4384161233901978, "learning_rate": 3.817264749055249e-05, "loss": 6.5293, "step": 22265 }, { "epoch": 0.9387608828551552, "grad_norm": 4.004322528839111, "learning_rate": 3.817186670414442e-05, "loss": 9.8455, "step": 22266 }, { "epoch": 0.9388030440373548, "grad_norm": 1.4449045658111572, "learning_rate": 3.817108591773634e-05, "loss": 6.5275, "step": 22267 }, { "epoch": 0.9388452052195544, "grad_norm": 2.3883817195892334, "learning_rate": 3.8170305131328274e-05, "loss": 7.8086, "step": 22268 }, { "epoch": 0.9388873664017539, "grad_norm": 3.0648443698883057, "learning_rate": 3.8169524344920205e-05, "loss": 8.0009, "step": 22269 }, { "epoch": 0.9389295275839534, "grad_norm": 1.6712579727172852, "learning_rate": 3.8168743558512136e-05, "loss": 6.6395, "step": 22270 }, { "epoch": 0.938971688766153, "grad_norm": 2.873840808868408, "learning_rate": 3.816796277210407e-05, "loss": 7.6666, "step": 22271 }, { "epoch": 0.9390138499483526, "grad_norm": 2.3020482063293457, "learning_rate": 3.816718198569599e-05, "loss": 7.2383, "step": 22272 }, { "epoch": 0.9390560111305521, "grad_norm": 1.608938217163086, "learning_rate": 3.8166401199287923e-05, "loss": 6.6185, "step": 22273 }, { "epoch": 0.9390981723127516, "grad_norm": 1.3555608987808228, "learning_rate": 3.8165620412879855e-05, "loss": 7.0753, "step": 22274 }, { "epoch": 0.9391403334949512, "grad_norm": 1.2157559394836426, "learning_rate": 3.816483962647178e-05, "loss": 6.8388, "step": 22275 }, { "epoch": 0.9391824946771508, "grad_norm": 1.5234668254852295, "learning_rate": 3.816405884006372e-05, "loss": 6.5882, "step": 22276 }, { "epoch": 0.9392246558593503, "grad_norm": 2.7873823642730713, "learning_rate": 3.816327805365564e-05, "loss": 7.2701, "step": 22277 }, { "epoch": 0.9392668170415498, "grad_norm": 1.4411697387695312, "learning_rate": 3.816249726724757e-05, "loss": 6.8434, "step": 22278 }, { "epoch": 0.9393089782237494, "grad_norm": 1.4385771751403809, "learning_rate": 3.8161716480839504e-05, "loss": 6.4533, "step": 22279 }, { "epoch": 0.9393511394059489, "grad_norm": 1.498476505279541, "learning_rate": 3.8160935694431435e-05, "loss": 6.627, "step": 22280 }, { "epoch": 0.9393933005881485, "grad_norm": 1.002902865409851, "learning_rate": 3.816015490802337e-05, "loss": 6.8239, "step": 22281 }, { "epoch": 0.9394354617703481, "grad_norm": 1.6539379358291626, "learning_rate": 3.815937412161529e-05, "loss": 6.9616, "step": 22282 }, { "epoch": 0.9394776229525476, "grad_norm": 1.4723962545394897, "learning_rate": 3.815859333520722e-05, "loss": 6.5163, "step": 22283 }, { "epoch": 0.9395197841347471, "grad_norm": 2.851461410522461, "learning_rate": 3.8157812548799154e-05, "loss": 6.3725, "step": 22284 }, { "epoch": 0.9395619453169467, "grad_norm": 2.582505702972412, "learning_rate": 3.8157031762391085e-05, "loss": 7.9537, "step": 22285 }, { "epoch": 0.9396041064991463, "grad_norm": 2.2670438289642334, "learning_rate": 3.815625097598301e-05, "loss": 6.5024, "step": 22286 }, { "epoch": 0.9396462676813457, "grad_norm": 2.5517876148223877, "learning_rate": 3.815547018957494e-05, "loss": 7.2316, "step": 22287 }, { "epoch": 0.9396884288635453, "grad_norm": 1.1372652053833008, "learning_rate": 3.815468940316687e-05, "loss": 6.6614, "step": 22288 }, { "epoch": 0.9397305900457449, "grad_norm": 1.7969486713409424, "learning_rate": 3.8153908616758796e-05, "loss": 7.0458, "step": 22289 }, { "epoch": 0.9397727512279445, "grad_norm": 2.3237743377685547, "learning_rate": 3.8153127830350735e-05, "loss": 7.0824, "step": 22290 }, { "epoch": 0.939814912410144, "grad_norm": 1.2686421871185303, "learning_rate": 3.815234704394266e-05, "loss": 6.5119, "step": 22291 }, { "epoch": 0.9398570735923435, "grad_norm": 2.140695810317993, "learning_rate": 3.815156625753459e-05, "loss": 6.509, "step": 22292 }, { "epoch": 0.9398992347745431, "grad_norm": 1.0731912851333618, "learning_rate": 3.815078547112652e-05, "loss": 6.6226, "step": 22293 }, { "epoch": 0.9399413959567426, "grad_norm": 0.9306350946426392, "learning_rate": 3.8150004684718446e-05, "loss": 6.4874, "step": 22294 }, { "epoch": 0.9399835571389422, "grad_norm": 2.8005173206329346, "learning_rate": 3.8149223898310384e-05, "loss": 7.5543, "step": 22295 }, { "epoch": 0.9400257183211417, "grad_norm": 3.1179842948913574, "learning_rate": 3.814844311190231e-05, "loss": 7.688, "step": 22296 }, { "epoch": 0.9400678795033413, "grad_norm": 1.1679128408432007, "learning_rate": 3.814766232549424e-05, "loss": 6.5247, "step": 22297 }, { "epoch": 0.9401100406855408, "grad_norm": 1.3901182413101196, "learning_rate": 3.814688153908617e-05, "loss": 6.4758, "step": 22298 }, { "epoch": 0.9401522018677404, "grad_norm": 4.056795120239258, "learning_rate": 3.8146100752678096e-05, "loss": 7.7636, "step": 22299 }, { "epoch": 0.94019436304994, "grad_norm": 2.274756908416748, "learning_rate": 3.814531996627003e-05, "loss": 6.5166, "step": 22300 }, { "epoch": 0.9402365242321394, "grad_norm": 2.0525121688842773, "learning_rate": 3.814453917986196e-05, "loss": 6.6003, "step": 22301 }, { "epoch": 0.940278685414339, "grad_norm": 1.6285232305526733, "learning_rate": 3.814375839345389e-05, "loss": 6.5359, "step": 22302 }, { "epoch": 0.9403208465965386, "grad_norm": 1.970358967781067, "learning_rate": 3.814297760704582e-05, "loss": 6.9688, "step": 22303 }, { "epoch": 0.9403630077787382, "grad_norm": 1.6742464303970337, "learning_rate": 3.814219682063775e-05, "loss": 7.3502, "step": 22304 }, { "epoch": 0.9404051689609376, "grad_norm": 1.142007827758789, "learning_rate": 3.8141416034229676e-05, "loss": 6.4961, "step": 22305 }, { "epoch": 0.9404473301431372, "grad_norm": 3.1723108291625977, "learning_rate": 3.814063524782161e-05, "loss": 8.0775, "step": 22306 }, { "epoch": 0.9404894913253368, "grad_norm": 2.998425006866455, "learning_rate": 3.813985446141354e-05, "loss": 7.6825, "step": 22307 }, { "epoch": 0.9405316525075363, "grad_norm": 0.895326554775238, "learning_rate": 3.813907367500546e-05, "loss": 6.6112, "step": 22308 }, { "epoch": 0.9405738136897358, "grad_norm": 3.268871307373047, "learning_rate": 3.81382928885974e-05, "loss": 7.5973, "step": 22309 }, { "epoch": 0.9406159748719354, "grad_norm": 1.6435247659683228, "learning_rate": 3.8137512102189326e-05, "loss": 7.1558, "step": 22310 }, { "epoch": 0.940658136054135, "grad_norm": 2.3885653018951416, "learning_rate": 3.813673131578126e-05, "loss": 7.3187, "step": 22311 }, { "epoch": 0.9407002972363345, "grad_norm": 2.1544764041900635, "learning_rate": 3.813595052937319e-05, "loss": 7.4114, "step": 22312 }, { "epoch": 0.9407424584185341, "grad_norm": 2.1780309677124023, "learning_rate": 3.813516974296511e-05, "loss": 6.9281, "step": 22313 }, { "epoch": 0.9407846196007336, "grad_norm": 1.8715863227844238, "learning_rate": 3.8134388956557044e-05, "loss": 6.6563, "step": 22314 }, { "epoch": 0.9408267807829331, "grad_norm": 1.4793702363967896, "learning_rate": 3.8133608170148975e-05, "loss": 6.9817, "step": 22315 }, { "epoch": 0.9408689419651327, "grad_norm": 1.2994098663330078, "learning_rate": 3.813282738374091e-05, "loss": 6.5801, "step": 22316 }, { "epoch": 0.9409111031473323, "grad_norm": 1.4323091506958008, "learning_rate": 3.813204659733284e-05, "loss": 6.9032, "step": 22317 }, { "epoch": 0.9409532643295317, "grad_norm": 1.6231783628463745, "learning_rate": 3.813126581092476e-05, "loss": 7.0655, "step": 22318 }, { "epoch": 0.9409954255117313, "grad_norm": 1.4326913356781006, "learning_rate": 3.8130485024516694e-05, "loss": 6.5488, "step": 22319 }, { "epoch": 0.9410375866939309, "grad_norm": 2.6826839447021484, "learning_rate": 3.8129704238108625e-05, "loss": 7.7282, "step": 22320 }, { "epoch": 0.9410797478761305, "grad_norm": 2.0493626594543457, "learning_rate": 3.8128923451700556e-05, "loss": 7.1974, "step": 22321 }, { "epoch": 0.94112190905833, "grad_norm": 1.2939910888671875, "learning_rate": 3.812814266529248e-05, "loss": 6.5256, "step": 22322 }, { "epoch": 0.9411640702405295, "grad_norm": 1.263620138168335, "learning_rate": 3.812736187888442e-05, "loss": 6.7561, "step": 22323 }, { "epoch": 0.9412062314227291, "grad_norm": 1.4234750270843506, "learning_rate": 3.812658109247634e-05, "loss": 6.5675, "step": 22324 }, { "epoch": 0.9412483926049287, "grad_norm": 2.2975656986236572, "learning_rate": 3.8125800306068274e-05, "loss": 6.6918, "step": 22325 }, { "epoch": 0.9412905537871282, "grad_norm": 1.2676868438720703, "learning_rate": 3.8125019519660206e-05, "loss": 6.8735, "step": 22326 }, { "epoch": 0.9413327149693277, "grad_norm": 3.7213480472564697, "learning_rate": 3.812423873325213e-05, "loss": 7.6132, "step": 22327 }, { "epoch": 0.9413748761515273, "grad_norm": 1.5766592025756836, "learning_rate": 3.812345794684407e-05, "loss": 6.9524, "step": 22328 }, { "epoch": 0.9414170373337268, "grad_norm": 0.951612651348114, "learning_rate": 3.812267716043599e-05, "loss": 6.7844, "step": 22329 }, { "epoch": 0.9414591985159264, "grad_norm": 2.129852533340454, "learning_rate": 3.8121896374027924e-05, "loss": 6.5586, "step": 22330 }, { "epoch": 0.941501359698126, "grad_norm": 2.2157957553863525, "learning_rate": 3.8121115587619855e-05, "loss": 7.8757, "step": 22331 }, { "epoch": 0.9415435208803254, "grad_norm": 1.0122578144073486, "learning_rate": 3.812033480121178e-05, "loss": 6.6691, "step": 22332 }, { "epoch": 0.941585682062525, "grad_norm": 2.0417962074279785, "learning_rate": 3.811955401480371e-05, "loss": 6.9102, "step": 22333 }, { "epoch": 0.9416278432447246, "grad_norm": 1.062808871269226, "learning_rate": 3.811877322839564e-05, "loss": 6.9455, "step": 22334 }, { "epoch": 0.9416700044269242, "grad_norm": 3.092594623565674, "learning_rate": 3.8117992441987573e-05, "loss": 7.6362, "step": 22335 }, { "epoch": 0.9417121656091236, "grad_norm": 1.1934367418289185, "learning_rate": 3.81172116555795e-05, "loss": 6.5186, "step": 22336 }, { "epoch": 0.9417543267913232, "grad_norm": 1.1483362913131714, "learning_rate": 3.811643086917143e-05, "loss": 6.4661, "step": 22337 }, { "epoch": 0.9417964879735228, "grad_norm": 1.579981803894043, "learning_rate": 3.811565008276336e-05, "loss": 6.7903, "step": 22338 }, { "epoch": 0.9418386491557224, "grad_norm": 1.1038986444473267, "learning_rate": 3.811486929635529e-05, "loss": 6.8336, "step": 22339 }, { "epoch": 0.9418808103379219, "grad_norm": 2.443345546722412, "learning_rate": 3.811408850994722e-05, "loss": 7.4285, "step": 22340 }, { "epoch": 0.9419229715201214, "grad_norm": 1.1629817485809326, "learning_rate": 3.811330772353915e-05, "loss": 6.4533, "step": 22341 }, { "epoch": 0.941965132702321, "grad_norm": 2.0955312252044678, "learning_rate": 3.8112526937131085e-05, "loss": 6.6448, "step": 22342 }, { "epoch": 0.9420072938845205, "grad_norm": 3.318713665008545, "learning_rate": 3.811174615072301e-05, "loss": 7.8749, "step": 22343 }, { "epoch": 0.9420494550667201, "grad_norm": 1.1623761653900146, "learning_rate": 3.8110965364314934e-05, "loss": 6.7737, "step": 22344 }, { "epoch": 0.9420916162489196, "grad_norm": 1.5558844804763794, "learning_rate": 3.811018457790687e-05, "loss": 6.6984, "step": 22345 }, { "epoch": 0.9421337774311191, "grad_norm": 1.638006567955017, "learning_rate": 3.81094037914988e-05, "loss": 7.1275, "step": 22346 }, { "epoch": 0.9421759386133187, "grad_norm": 2.374549388885498, "learning_rate": 3.810862300509073e-05, "loss": 6.5511, "step": 22347 }, { "epoch": 0.9422180997955183, "grad_norm": 1.2982778549194336, "learning_rate": 3.810784221868266e-05, "loss": 6.493, "step": 22348 }, { "epoch": 0.9422602609777179, "grad_norm": 2.1606948375701904, "learning_rate": 3.810706143227459e-05, "loss": 7.0866, "step": 22349 }, { "epoch": 0.9423024221599173, "grad_norm": 2.623194694519043, "learning_rate": 3.810628064586652e-05, "loss": 7.5319, "step": 22350 }, { "epoch": 0.9423445833421169, "grad_norm": 1.0853149890899658, "learning_rate": 3.8105499859458447e-05, "loss": 6.8145, "step": 22351 }, { "epoch": 0.9423867445243165, "grad_norm": 1.9411309957504272, "learning_rate": 3.810471907305038e-05, "loss": 7.1513, "step": 22352 }, { "epoch": 0.942428905706516, "grad_norm": 2.094541549682617, "learning_rate": 3.810393828664231e-05, "loss": 7.1379, "step": 22353 }, { "epoch": 0.9424710668887155, "grad_norm": 1.8256298303604126, "learning_rate": 3.810315750023424e-05, "loss": 6.9571, "step": 22354 }, { "epoch": 0.9425132280709151, "grad_norm": 3.2251389026641846, "learning_rate": 3.8102376713826165e-05, "loss": 7.6947, "step": 22355 }, { "epoch": 0.9425553892531147, "grad_norm": 1.464700698852539, "learning_rate": 3.8101595927418096e-05, "loss": 6.5068, "step": 22356 }, { "epoch": 0.9425975504353142, "grad_norm": 1.6891433000564575, "learning_rate": 3.810081514101003e-05, "loss": 6.3226, "step": 22357 }, { "epoch": 0.9426397116175138, "grad_norm": 1.1997171640396118, "learning_rate": 3.810003435460195e-05, "loss": 6.974, "step": 22358 }, { "epoch": 0.9426818727997133, "grad_norm": 4.362879276275635, "learning_rate": 3.809925356819389e-05, "loss": 10.1393, "step": 22359 }, { "epoch": 0.9427240339819128, "grad_norm": 1.2587944269180298, "learning_rate": 3.8098472781785814e-05, "loss": 7.0439, "step": 22360 }, { "epoch": 0.9427661951641124, "grad_norm": 2.8803155422210693, "learning_rate": 3.8097691995377746e-05, "loss": 7.3451, "step": 22361 }, { "epoch": 0.942808356346312, "grad_norm": 3.4386651515960693, "learning_rate": 3.809691120896968e-05, "loss": 8.0376, "step": 22362 }, { "epoch": 0.9428505175285115, "grad_norm": 1.2677401304244995, "learning_rate": 3.80961304225616e-05, "loss": 6.5778, "step": 22363 }, { "epoch": 0.942892678710711, "grad_norm": 2.3025054931640625, "learning_rate": 3.809534963615354e-05, "loss": 7.6526, "step": 22364 }, { "epoch": 0.9429348398929106, "grad_norm": 1.4402021169662476, "learning_rate": 3.8094568849745464e-05, "loss": 6.9192, "step": 22365 }, { "epoch": 0.9429770010751102, "grad_norm": 1.9396014213562012, "learning_rate": 3.8093788063337395e-05, "loss": 6.4277, "step": 22366 }, { "epoch": 0.9430191622573096, "grad_norm": 1.475052833557129, "learning_rate": 3.8093007276929326e-05, "loss": 7.0644, "step": 22367 }, { "epoch": 0.9430613234395092, "grad_norm": 1.0635274648666382, "learning_rate": 3.809222649052126e-05, "loss": 6.5361, "step": 22368 }, { "epoch": 0.9431034846217088, "grad_norm": 2.1126151084899902, "learning_rate": 3.809144570411318e-05, "loss": 7.167, "step": 22369 }, { "epoch": 0.9431456458039084, "grad_norm": 1.81272292137146, "learning_rate": 3.809066491770511e-05, "loss": 7.2336, "step": 22370 }, { "epoch": 0.9431878069861079, "grad_norm": 1.0349605083465576, "learning_rate": 3.8089884131297045e-05, "loss": 6.4453, "step": 22371 }, { "epoch": 0.9432299681683074, "grad_norm": 1.3960461616516113, "learning_rate": 3.8089103344888976e-05, "loss": 6.5132, "step": 22372 }, { "epoch": 0.943272129350507, "grad_norm": 1.1963454484939575, "learning_rate": 3.808832255848091e-05, "loss": 6.6444, "step": 22373 }, { "epoch": 0.9433142905327065, "grad_norm": 1.9146103858947754, "learning_rate": 3.808754177207283e-05, "loss": 7.0053, "step": 22374 }, { "epoch": 0.9433564517149061, "grad_norm": 3.125300645828247, "learning_rate": 3.808676098566476e-05, "loss": 7.7031, "step": 22375 }, { "epoch": 0.9433986128971056, "grad_norm": 3.7004480361938477, "learning_rate": 3.8085980199256694e-05, "loss": 7.811, "step": 22376 }, { "epoch": 0.9434407740793052, "grad_norm": 1.71905517578125, "learning_rate": 3.808519941284862e-05, "loss": 6.588, "step": 22377 }, { "epoch": 0.9434829352615047, "grad_norm": 1.4235550165176392, "learning_rate": 3.808441862644056e-05, "loss": 6.5583, "step": 22378 }, { "epoch": 0.9435250964437043, "grad_norm": 1.5399290323257446, "learning_rate": 3.808363784003248e-05, "loss": 6.4473, "step": 22379 }, { "epoch": 0.9435672576259039, "grad_norm": 2.0164222717285156, "learning_rate": 3.808285705362441e-05, "loss": 7.1367, "step": 22380 }, { "epoch": 0.9436094188081033, "grad_norm": 1.4521517753601074, "learning_rate": 3.8082076267216344e-05, "loss": 6.4198, "step": 22381 }, { "epoch": 0.9436515799903029, "grad_norm": 2.8829073905944824, "learning_rate": 3.808129548080827e-05, "loss": 7.4232, "step": 22382 }, { "epoch": 0.9436937411725025, "grad_norm": 0.9209998846054077, "learning_rate": 3.8080514694400206e-05, "loss": 6.3934, "step": 22383 }, { "epoch": 0.9437359023547021, "grad_norm": 1.6948884725570679, "learning_rate": 3.807973390799213e-05, "loss": 6.9764, "step": 22384 }, { "epoch": 0.9437780635369015, "grad_norm": 2.183770179748535, "learning_rate": 3.807895312158406e-05, "loss": 7.0696, "step": 22385 }, { "epoch": 0.9438202247191011, "grad_norm": 3.8923897743225098, "learning_rate": 3.807817233517599e-05, "loss": 7.7174, "step": 22386 }, { "epoch": 0.9438623859013007, "grad_norm": 2.2178525924682617, "learning_rate": 3.8077391548767924e-05, "loss": 6.9942, "step": 22387 }, { "epoch": 0.9439045470835002, "grad_norm": 2.5652737617492676, "learning_rate": 3.807661076235985e-05, "loss": 7.3059, "step": 22388 }, { "epoch": 0.9439467082656998, "grad_norm": 3.1961300373077393, "learning_rate": 3.807582997595178e-05, "loss": 6.8955, "step": 22389 }, { "epoch": 0.9439888694478993, "grad_norm": 1.2232449054718018, "learning_rate": 3.807504918954371e-05, "loss": 6.9926, "step": 22390 }, { "epoch": 0.9440310306300989, "grad_norm": 1.9716482162475586, "learning_rate": 3.8074268403135636e-05, "loss": 6.9681, "step": 22391 }, { "epoch": 0.9440731918122984, "grad_norm": 1.7694945335388184, "learning_rate": 3.8073487616727574e-05, "loss": 6.7529, "step": 22392 }, { "epoch": 0.944115352994498, "grad_norm": 2.094855785369873, "learning_rate": 3.80727068303195e-05, "loss": 7.1955, "step": 22393 }, { "epoch": 0.9441575141766975, "grad_norm": 1.7887078523635864, "learning_rate": 3.807192604391143e-05, "loss": 6.7656, "step": 22394 }, { "epoch": 0.944199675358897, "grad_norm": 1.889835000038147, "learning_rate": 3.807114525750336e-05, "loss": 6.8303, "step": 22395 }, { "epoch": 0.9442418365410966, "grad_norm": 1.9883345365524292, "learning_rate": 3.8070364471095285e-05, "loss": 6.5743, "step": 22396 }, { "epoch": 0.9442839977232962, "grad_norm": 2.2870192527770996, "learning_rate": 3.8069583684687223e-05, "loss": 7.2068, "step": 22397 }, { "epoch": 0.9443261589054958, "grad_norm": 1.5865391492843628, "learning_rate": 3.806880289827915e-05, "loss": 6.5151, "step": 22398 }, { "epoch": 0.9443683200876952, "grad_norm": 2.2196052074432373, "learning_rate": 3.806802211187108e-05, "loss": 6.9666, "step": 22399 }, { "epoch": 0.9444104812698948, "grad_norm": 1.2062162160873413, "learning_rate": 3.806724132546301e-05, "loss": 7.038, "step": 22400 }, { "epoch": 0.9444526424520944, "grad_norm": 1.56979501247406, "learning_rate": 3.8066460539054935e-05, "loss": 7.1621, "step": 22401 }, { "epoch": 0.944494803634294, "grad_norm": 1.6237189769744873, "learning_rate": 3.8065679752646866e-05, "loss": 6.9806, "step": 22402 }, { "epoch": 0.9445369648164934, "grad_norm": 1.1809695959091187, "learning_rate": 3.80648989662388e-05, "loss": 6.3967, "step": 22403 }, { "epoch": 0.944579125998693, "grad_norm": 1.3292893171310425, "learning_rate": 3.806411817983073e-05, "loss": 6.6053, "step": 22404 }, { "epoch": 0.9446212871808926, "grad_norm": 1.6433515548706055, "learning_rate": 3.806333739342265e-05, "loss": 7.1267, "step": 22405 }, { "epoch": 0.9446634483630921, "grad_norm": 1.3368759155273438, "learning_rate": 3.8062556607014584e-05, "loss": 6.8783, "step": 22406 }, { "epoch": 0.9447056095452917, "grad_norm": 1.3560086488723755, "learning_rate": 3.8061775820606516e-05, "loss": 6.4962, "step": 22407 }, { "epoch": 0.9447477707274912, "grad_norm": 1.13113272190094, "learning_rate": 3.806099503419845e-05, "loss": 6.8923, "step": 22408 }, { "epoch": 0.9447899319096907, "grad_norm": 1.1711093187332153, "learning_rate": 3.806021424779038e-05, "loss": 6.5747, "step": 22409 }, { "epoch": 0.9448320930918903, "grad_norm": 1.2830561399459839, "learning_rate": 3.80594334613823e-05, "loss": 6.9324, "step": 22410 }, { "epoch": 0.9448742542740899, "grad_norm": 1.1609715223312378, "learning_rate": 3.805865267497424e-05, "loss": 6.9969, "step": 22411 }, { "epoch": 0.9449164154562893, "grad_norm": 1.1638975143432617, "learning_rate": 3.8057871888566165e-05, "loss": 6.8362, "step": 22412 }, { "epoch": 0.9449585766384889, "grad_norm": 1.0124869346618652, "learning_rate": 3.8057091102158097e-05, "loss": 6.908, "step": 22413 }, { "epoch": 0.9450007378206885, "grad_norm": 0.9290131330490112, "learning_rate": 3.805631031575003e-05, "loss": 6.5687, "step": 22414 }, { "epoch": 0.9450428990028881, "grad_norm": 3.0125439167022705, "learning_rate": 3.805552952934195e-05, "loss": 7.6836, "step": 22415 }, { "epoch": 0.9450850601850876, "grad_norm": 1.2716902494430542, "learning_rate": 3.8054748742933884e-05, "loss": 7.0392, "step": 22416 }, { "epoch": 0.9451272213672871, "grad_norm": 2.4595730304718018, "learning_rate": 3.8053967956525815e-05, "loss": 7.7635, "step": 22417 }, { "epoch": 0.9451693825494867, "grad_norm": 3.004037857055664, "learning_rate": 3.8053187170117746e-05, "loss": 7.8946, "step": 22418 }, { "epoch": 0.9452115437316863, "grad_norm": 1.6842893362045288, "learning_rate": 3.805240638370968e-05, "loss": 7.3545, "step": 22419 }, { "epoch": 0.9452537049138858, "grad_norm": 2.92747163772583, "learning_rate": 3.80516255973016e-05, "loss": 6.4423, "step": 22420 }, { "epoch": 0.9452958660960853, "grad_norm": 2.719536304473877, "learning_rate": 3.805084481089353e-05, "loss": 8.1732, "step": 22421 }, { "epoch": 0.9453380272782849, "grad_norm": 2.172492265701294, "learning_rate": 3.8050064024485464e-05, "loss": 6.5077, "step": 22422 }, { "epoch": 0.9453801884604844, "grad_norm": 4.4710893630981445, "learning_rate": 3.8049283238077396e-05, "loss": 9.0143, "step": 22423 }, { "epoch": 0.945422349642684, "grad_norm": 0.9498534798622131, "learning_rate": 3.804850245166932e-05, "loss": 7.0503, "step": 22424 }, { "epoch": 0.9454645108248835, "grad_norm": 1.1010226011276245, "learning_rate": 3.804772166526125e-05, "loss": 6.4846, "step": 22425 }, { "epoch": 0.945506672007083, "grad_norm": 1.7995738983154297, "learning_rate": 3.804694087885318e-05, "loss": 7.0217, "step": 22426 }, { "epoch": 0.9455488331892826, "grad_norm": 2.690570831298828, "learning_rate": 3.804616009244511e-05, "loss": 7.0093, "step": 22427 }, { "epoch": 0.9455909943714822, "grad_norm": 1.6258935928344727, "learning_rate": 3.8045379306037045e-05, "loss": 6.7264, "step": 22428 }, { "epoch": 0.9456331555536818, "grad_norm": 1.8290529251098633, "learning_rate": 3.804459851962897e-05, "loss": 6.9072, "step": 22429 }, { "epoch": 0.9456753167358812, "grad_norm": 1.3971080780029297, "learning_rate": 3.804381773322091e-05, "loss": 7.168, "step": 22430 }, { "epoch": 0.9457174779180808, "grad_norm": 1.6560320854187012, "learning_rate": 3.804303694681283e-05, "loss": 6.7358, "step": 22431 }, { "epoch": 0.9457596391002804, "grad_norm": 1.822265863418579, "learning_rate": 3.804225616040476e-05, "loss": 6.547, "step": 22432 }, { "epoch": 0.94580180028248, "grad_norm": 3.1863341331481934, "learning_rate": 3.8041475373996695e-05, "loss": 7.964, "step": 22433 }, { "epoch": 0.9458439614646794, "grad_norm": 1.6811057329177856, "learning_rate": 3.804069458758862e-05, "loss": 7.0656, "step": 22434 }, { "epoch": 0.945886122646879, "grad_norm": 2.9723424911499023, "learning_rate": 3.803991380118055e-05, "loss": 7.8255, "step": 22435 }, { "epoch": 0.9459282838290786, "grad_norm": 1.8666630983352661, "learning_rate": 3.803913301477248e-05, "loss": 6.958, "step": 22436 }, { "epoch": 0.9459704450112781, "grad_norm": 1.5879360437393188, "learning_rate": 3.803835222836441e-05, "loss": 6.6692, "step": 22437 }, { "epoch": 0.9460126061934777, "grad_norm": 1.3694818019866943, "learning_rate": 3.803757144195634e-05, "loss": 6.6307, "step": 22438 }, { "epoch": 0.9460547673756772, "grad_norm": 1.2232098579406738, "learning_rate": 3.803679065554827e-05, "loss": 7.0147, "step": 22439 }, { "epoch": 0.9460969285578767, "grad_norm": 1.0641263723373413, "learning_rate": 3.80360098691402e-05, "loss": 6.5557, "step": 22440 }, { "epoch": 0.9461390897400763, "grad_norm": 1.3986045122146606, "learning_rate": 3.803522908273213e-05, "loss": 6.7451, "step": 22441 }, { "epoch": 0.9461812509222759, "grad_norm": 1.583362102508545, "learning_rate": 3.803444829632406e-05, "loss": 6.3958, "step": 22442 }, { "epoch": 0.9462234121044754, "grad_norm": 2.670156955718994, "learning_rate": 3.803366750991599e-05, "loss": 6.9597, "step": 22443 }, { "epoch": 0.9462655732866749, "grad_norm": 3.371896982192993, "learning_rate": 3.803288672350792e-05, "loss": 7.8682, "step": 22444 }, { "epoch": 0.9463077344688745, "grad_norm": 1.7989041805267334, "learning_rate": 3.803210593709985e-05, "loss": 7.3066, "step": 22445 }, { "epoch": 0.9463498956510741, "grad_norm": 1.935401201248169, "learning_rate": 3.8031325150691774e-05, "loss": 6.5861, "step": 22446 }, { "epoch": 0.9463920568332737, "grad_norm": 1.8620471954345703, "learning_rate": 3.803054436428371e-05, "loss": 6.5324, "step": 22447 }, { "epoch": 0.9464342180154731, "grad_norm": 2.893280506134033, "learning_rate": 3.8029763577875636e-05, "loss": 6.3558, "step": 22448 }, { "epoch": 0.9464763791976727, "grad_norm": 2.5605273246765137, "learning_rate": 3.802898279146757e-05, "loss": 7.5908, "step": 22449 }, { "epoch": 0.9465185403798723, "grad_norm": 2.4457716941833496, "learning_rate": 3.80282020050595e-05, "loss": 7.2908, "step": 22450 }, { "epoch": 0.9465607015620718, "grad_norm": 4.2917962074279785, "learning_rate": 3.8027421218651423e-05, "loss": 8.4581, "step": 22451 }, { "epoch": 0.9466028627442713, "grad_norm": 1.8947094678878784, "learning_rate": 3.802664043224336e-05, "loss": 6.4922, "step": 22452 }, { "epoch": 0.9466450239264709, "grad_norm": 2.00040340423584, "learning_rate": 3.8025859645835286e-05, "loss": 6.9364, "step": 22453 }, { "epoch": 0.9466871851086704, "grad_norm": 2.3238677978515625, "learning_rate": 3.802507885942722e-05, "loss": 7.3675, "step": 22454 }, { "epoch": 0.94672934629087, "grad_norm": 1.2059407234191895, "learning_rate": 3.802429807301915e-05, "loss": 6.7322, "step": 22455 }, { "epoch": 0.9467715074730696, "grad_norm": 0.8647521734237671, "learning_rate": 3.802351728661108e-05, "loss": 6.392, "step": 22456 }, { "epoch": 0.9468136686552691, "grad_norm": 1.6074062585830688, "learning_rate": 3.8022736500203004e-05, "loss": 6.6606, "step": 22457 }, { "epoch": 0.9468558298374686, "grad_norm": 2.634897232055664, "learning_rate": 3.8021955713794935e-05, "loss": 7.2537, "step": 22458 }, { "epoch": 0.9468979910196682, "grad_norm": 1.0688581466674805, "learning_rate": 3.802117492738687e-05, "loss": 6.4127, "step": 22459 }, { "epoch": 0.9469401522018678, "grad_norm": 0.8408176302909851, "learning_rate": 3.802039414097879e-05, "loss": 6.5702, "step": 22460 }, { "epoch": 0.9469823133840672, "grad_norm": 1.4215186834335327, "learning_rate": 3.801961335457073e-05, "loss": 6.5623, "step": 22461 }, { "epoch": 0.9470244745662668, "grad_norm": 0.8617687821388245, "learning_rate": 3.8018832568162654e-05, "loss": 6.4658, "step": 22462 }, { "epoch": 0.9470666357484664, "grad_norm": 1.6533215045928955, "learning_rate": 3.8018051781754585e-05, "loss": 6.7769, "step": 22463 }, { "epoch": 0.947108796930666, "grad_norm": 3.15236759185791, "learning_rate": 3.8017270995346516e-05, "loss": 8.2812, "step": 22464 }, { "epoch": 0.9471509581128655, "grad_norm": 1.736204981803894, "learning_rate": 3.801649020893844e-05, "loss": 6.7919, "step": 22465 }, { "epoch": 0.947193119295065, "grad_norm": 0.8708930611610413, "learning_rate": 3.801570942253038e-05, "loss": 6.4912, "step": 22466 }, { "epoch": 0.9472352804772646, "grad_norm": 1.5816580057144165, "learning_rate": 3.80149286361223e-05, "loss": 6.6999, "step": 22467 }, { "epoch": 0.9472774416594641, "grad_norm": 1.4004591703414917, "learning_rate": 3.8014147849714234e-05, "loss": 6.6964, "step": 22468 }, { "epoch": 0.9473196028416637, "grad_norm": 1.1248799562454224, "learning_rate": 3.8013367063306166e-05, "loss": 6.618, "step": 22469 }, { "epoch": 0.9473617640238632, "grad_norm": 1.9896599054336548, "learning_rate": 3.801258627689809e-05, "loss": 7.3435, "step": 22470 }, { "epoch": 0.9474039252060628, "grad_norm": 2.281765937805176, "learning_rate": 3.801180549049002e-05, "loss": 6.3636, "step": 22471 }, { "epoch": 0.9474460863882623, "grad_norm": 3.5887815952301025, "learning_rate": 3.801102470408195e-05, "loss": 7.8988, "step": 22472 }, { "epoch": 0.9474882475704619, "grad_norm": 2.5252790451049805, "learning_rate": 3.8010243917673884e-05, "loss": 7.1458, "step": 22473 }, { "epoch": 0.9475304087526615, "grad_norm": 1.687546730041504, "learning_rate": 3.800946313126581e-05, "loss": 6.4281, "step": 22474 }, { "epoch": 0.9475725699348609, "grad_norm": 2.1955924034118652, "learning_rate": 3.8008682344857747e-05, "loss": 6.6812, "step": 22475 }, { "epoch": 0.9476147311170605, "grad_norm": 1.8679455518722534, "learning_rate": 3.800790155844967e-05, "loss": 7.2959, "step": 22476 }, { "epoch": 0.9476568922992601, "grad_norm": 1.9542829990386963, "learning_rate": 3.80071207720416e-05, "loss": 6.3496, "step": 22477 }, { "epoch": 0.9476990534814597, "grad_norm": 1.7508258819580078, "learning_rate": 3.8006339985633534e-05, "loss": 6.8513, "step": 22478 }, { "epoch": 0.9477412146636591, "grad_norm": 2.9711432456970215, "learning_rate": 3.800555919922546e-05, "loss": 7.7373, "step": 22479 }, { "epoch": 0.9477833758458587, "grad_norm": 2.1006717681884766, "learning_rate": 3.8004778412817396e-05, "loss": 7.193, "step": 22480 }, { "epoch": 0.9478255370280583, "grad_norm": 1.3474076986312866, "learning_rate": 3.800399762640932e-05, "loss": 6.8881, "step": 22481 }, { "epoch": 0.9478676982102578, "grad_norm": 3.368514060974121, "learning_rate": 3.800321684000125e-05, "loss": 8.2334, "step": 22482 }, { "epoch": 0.9479098593924573, "grad_norm": 1.6865140199661255, "learning_rate": 3.800243605359318e-05, "loss": 7.1903, "step": 22483 }, { "epoch": 0.9479520205746569, "grad_norm": 1.9317916631698608, "learning_rate": 3.800165526718511e-05, "loss": 6.6275, "step": 22484 }, { "epoch": 0.9479941817568565, "grad_norm": 2.1669890880584717, "learning_rate": 3.800087448077704e-05, "loss": 6.505, "step": 22485 }, { "epoch": 0.948036342939056, "grad_norm": 2.300443172454834, "learning_rate": 3.800009369436897e-05, "loss": 7.8772, "step": 22486 }, { "epoch": 0.9480785041212556, "grad_norm": 1.516557216644287, "learning_rate": 3.79993129079609e-05, "loss": 6.9737, "step": 22487 }, { "epoch": 0.9481206653034551, "grad_norm": 2.7990918159484863, "learning_rate": 3.799853212155283e-05, "loss": 7.5234, "step": 22488 }, { "epoch": 0.9481628264856546, "grad_norm": 2.0807907581329346, "learning_rate": 3.799775133514476e-05, "loss": 6.5336, "step": 22489 }, { "epoch": 0.9482049876678542, "grad_norm": 1.0930432081222534, "learning_rate": 3.799697054873669e-05, "loss": 6.4832, "step": 22490 }, { "epoch": 0.9482471488500538, "grad_norm": 2.208153247833252, "learning_rate": 3.799618976232862e-05, "loss": 7.7532, "step": 22491 }, { "epoch": 0.9482893100322533, "grad_norm": 1.6109784841537476, "learning_rate": 3.799540897592055e-05, "loss": 7.2278, "step": 22492 }, { "epoch": 0.9483314712144528, "grad_norm": 2.3244705200195312, "learning_rate": 3.7994628189512475e-05, "loss": 7.4715, "step": 22493 }, { "epoch": 0.9483736323966524, "grad_norm": 1.7112095355987549, "learning_rate": 3.799384740310441e-05, "loss": 7.1246, "step": 22494 }, { "epoch": 0.948415793578852, "grad_norm": 2.4440057277679443, "learning_rate": 3.799306661669634e-05, "loss": 6.6722, "step": 22495 }, { "epoch": 0.9484579547610515, "grad_norm": 1.3985631465911865, "learning_rate": 3.799228583028826e-05, "loss": 6.7481, "step": 22496 }, { "epoch": 0.948500115943251, "grad_norm": 1.7065716981887817, "learning_rate": 3.79915050438802e-05, "loss": 6.4731, "step": 22497 }, { "epoch": 0.9485422771254506, "grad_norm": 1.268823266029358, "learning_rate": 3.7990724257472125e-05, "loss": 6.9434, "step": 22498 }, { "epoch": 0.9485844383076502, "grad_norm": 1.0229085683822632, "learning_rate": 3.798994347106406e-05, "loss": 7.0672, "step": 22499 }, { "epoch": 0.9486265994898497, "grad_norm": 1.3916544914245605, "learning_rate": 3.798916268465599e-05, "loss": 6.7483, "step": 22500 }, { "epoch": 0.9486687606720492, "grad_norm": 1.3916627168655396, "learning_rate": 3.798838189824792e-05, "loss": 7.1284, "step": 22501 }, { "epoch": 0.9487109218542488, "grad_norm": 1.2779046297073364, "learning_rate": 3.798760111183985e-05, "loss": 6.6112, "step": 22502 }, { "epoch": 0.9487530830364483, "grad_norm": 2.4084043502807617, "learning_rate": 3.7986820325431774e-05, "loss": 6.512, "step": 22503 }, { "epoch": 0.9487952442186479, "grad_norm": 1.2338271141052246, "learning_rate": 3.7986039539023706e-05, "loss": 6.5161, "step": 22504 }, { "epoch": 0.9488374054008475, "grad_norm": 2.0903475284576416, "learning_rate": 3.798525875261564e-05, "loss": 6.5915, "step": 22505 }, { "epoch": 0.948879566583047, "grad_norm": 1.2655141353607178, "learning_rate": 3.798447796620757e-05, "loss": 6.6178, "step": 22506 }, { "epoch": 0.9489217277652465, "grad_norm": 2.2717514038085938, "learning_rate": 3.798369717979949e-05, "loss": 6.9277, "step": 22507 }, { "epoch": 0.9489638889474461, "grad_norm": 2.5833709239959717, "learning_rate": 3.7982916393391424e-05, "loss": 7.0769, "step": 22508 }, { "epoch": 0.9490060501296457, "grad_norm": 2.6911568641662598, "learning_rate": 3.7982135606983355e-05, "loss": 6.8901, "step": 22509 }, { "epoch": 0.9490482113118451, "grad_norm": 1.1697273254394531, "learning_rate": 3.7981354820575286e-05, "loss": 6.5787, "step": 22510 }, { "epoch": 0.9490903724940447, "grad_norm": 1.0793551206588745, "learning_rate": 3.798057403416722e-05, "loss": 6.9014, "step": 22511 }, { "epoch": 0.9491325336762443, "grad_norm": 2.2438197135925293, "learning_rate": 3.797979324775914e-05, "loss": 7.3146, "step": 22512 }, { "epoch": 0.9491746948584439, "grad_norm": 1.329463005065918, "learning_rate": 3.7979012461351073e-05, "loss": 6.8624, "step": 22513 }, { "epoch": 0.9492168560406434, "grad_norm": 1.3541514873504639, "learning_rate": 3.7978231674943005e-05, "loss": 6.6088, "step": 22514 }, { "epoch": 0.9492590172228429, "grad_norm": 2.9653046131134033, "learning_rate": 3.797745088853493e-05, "loss": 7.385, "step": 22515 }, { "epoch": 0.9493011784050425, "grad_norm": 1.8731707334518433, "learning_rate": 3.797667010212687e-05, "loss": 6.9762, "step": 22516 }, { "epoch": 0.949343339587242, "grad_norm": 1.8289178609848022, "learning_rate": 3.797588931571879e-05, "loss": 6.9354, "step": 22517 }, { "epoch": 0.9493855007694416, "grad_norm": 0.9343076348304749, "learning_rate": 3.797510852931072e-05, "loss": 6.8313, "step": 22518 }, { "epoch": 0.9494276619516411, "grad_norm": 2.6695492267608643, "learning_rate": 3.7974327742902654e-05, "loss": 6.6636, "step": 22519 }, { "epoch": 0.9494698231338407, "grad_norm": 2.471944570541382, "learning_rate": 3.7973546956494585e-05, "loss": 7.6162, "step": 22520 }, { "epoch": 0.9495119843160402, "grad_norm": 1.258259892463684, "learning_rate": 3.797276617008652e-05, "loss": 6.6743, "step": 22521 }, { "epoch": 0.9495541454982398, "grad_norm": 2.073519706726074, "learning_rate": 3.797198538367844e-05, "loss": 7.3562, "step": 22522 }, { "epoch": 0.9495963066804394, "grad_norm": 2.9082651138305664, "learning_rate": 3.797120459727037e-05, "loss": 7.0144, "step": 22523 }, { "epoch": 0.9496384678626388, "grad_norm": 1.4165948629379272, "learning_rate": 3.7970423810862304e-05, "loss": 7.3384, "step": 22524 }, { "epoch": 0.9496806290448384, "grad_norm": 2.4895403385162354, "learning_rate": 3.7969643024454235e-05, "loss": 7.5319, "step": 22525 }, { "epoch": 0.949722790227038, "grad_norm": 0.9539237022399902, "learning_rate": 3.796886223804616e-05, "loss": 6.6357, "step": 22526 }, { "epoch": 0.9497649514092376, "grad_norm": 2.319438934326172, "learning_rate": 3.796808145163809e-05, "loss": 7.5257, "step": 22527 }, { "epoch": 0.949807112591437, "grad_norm": 3.0606329441070557, "learning_rate": 3.796730066523002e-05, "loss": 7.3914, "step": 22528 }, { "epoch": 0.9498492737736366, "grad_norm": 1.8723357915878296, "learning_rate": 3.7966519878821946e-05, "loss": 6.9912, "step": 22529 }, { "epoch": 0.9498914349558362, "grad_norm": 1.3818820714950562, "learning_rate": 3.7965739092413885e-05, "loss": 6.5359, "step": 22530 }, { "epoch": 0.9499335961380357, "grad_norm": 2.403164863586426, "learning_rate": 3.796495830600581e-05, "loss": 6.5731, "step": 22531 }, { "epoch": 0.9499757573202353, "grad_norm": 2.193488836288452, "learning_rate": 3.796417751959774e-05, "loss": 6.7725, "step": 22532 }, { "epoch": 0.9500179185024348, "grad_norm": 1.4193657636642456, "learning_rate": 3.796339673318967e-05, "loss": 7.794, "step": 22533 }, { "epoch": 0.9500600796846344, "grad_norm": 2.402783155441284, "learning_rate": 3.7962615946781596e-05, "loss": 7.5812, "step": 22534 }, { "epoch": 0.9501022408668339, "grad_norm": 1.3215681314468384, "learning_rate": 3.7961835160373534e-05, "loss": 6.5389, "step": 22535 }, { "epoch": 0.9501444020490335, "grad_norm": 1.1903685331344604, "learning_rate": 3.796105437396546e-05, "loss": 6.7601, "step": 22536 }, { "epoch": 0.950186563231233, "grad_norm": 1.597265601158142, "learning_rate": 3.796027358755739e-05, "loss": 6.9934, "step": 22537 }, { "epoch": 0.9502287244134325, "grad_norm": 1.4129135608673096, "learning_rate": 3.795949280114932e-05, "loss": 7.1477, "step": 22538 }, { "epoch": 0.9502708855956321, "grad_norm": 3.1950459480285645, "learning_rate": 3.795871201474125e-05, "loss": 8.011, "step": 22539 }, { "epoch": 0.9503130467778317, "grad_norm": 3.2089309692382812, "learning_rate": 3.795793122833318e-05, "loss": 6.6609, "step": 22540 }, { "epoch": 0.9503552079600311, "grad_norm": 2.1117444038391113, "learning_rate": 3.795715044192511e-05, "loss": 7.2139, "step": 22541 }, { "epoch": 0.9503973691422307, "grad_norm": 2.5800282955169678, "learning_rate": 3.795636965551704e-05, "loss": 7.5005, "step": 22542 }, { "epoch": 0.9504395303244303, "grad_norm": 2.7815468311309814, "learning_rate": 3.7955588869108964e-05, "loss": 7.3622, "step": 22543 }, { "epoch": 0.9504816915066299, "grad_norm": 1.3258780241012573, "learning_rate": 3.79548080827009e-05, "loss": 7.2339, "step": 22544 }, { "epoch": 0.9505238526888294, "grad_norm": 2.261791229248047, "learning_rate": 3.7954027296292826e-05, "loss": 6.7144, "step": 22545 }, { "epoch": 0.9505660138710289, "grad_norm": 2.3241260051727295, "learning_rate": 3.795324650988476e-05, "loss": 6.4823, "step": 22546 }, { "epoch": 0.9506081750532285, "grad_norm": 1.424773931503296, "learning_rate": 3.795246572347669e-05, "loss": 6.6058, "step": 22547 }, { "epoch": 0.950650336235428, "grad_norm": 1.6378045082092285, "learning_rate": 3.795168493706861e-05, "loss": 6.8096, "step": 22548 }, { "epoch": 0.9506924974176276, "grad_norm": 1.9237275123596191, "learning_rate": 3.795090415066055e-05, "loss": 6.5846, "step": 22549 }, { "epoch": 0.9507346585998271, "grad_norm": 2.590223789215088, "learning_rate": 3.7950123364252476e-05, "loss": 7.3786, "step": 22550 }, { "epoch": 0.9507768197820267, "grad_norm": 1.4780858755111694, "learning_rate": 3.794934257784441e-05, "loss": 6.5808, "step": 22551 }, { "epoch": 0.9508189809642262, "grad_norm": 2.226309061050415, "learning_rate": 3.794856179143634e-05, "loss": 6.8868, "step": 22552 }, { "epoch": 0.9508611421464258, "grad_norm": 1.5285841226577759, "learning_rate": 3.794778100502826e-05, "loss": 6.5822, "step": 22553 }, { "epoch": 0.9509033033286254, "grad_norm": 1.8794840574264526, "learning_rate": 3.7947000218620194e-05, "loss": 6.6692, "step": 22554 }, { "epoch": 0.9509454645108248, "grad_norm": 1.1980230808258057, "learning_rate": 3.7946219432212125e-05, "loss": 6.5924, "step": 22555 }, { "epoch": 0.9509876256930244, "grad_norm": 2.2833194732666016, "learning_rate": 3.794543864580406e-05, "loss": 6.7517, "step": 22556 }, { "epoch": 0.951029786875224, "grad_norm": 3.390963554382324, "learning_rate": 3.794465785939599e-05, "loss": 7.2298, "step": 22557 }, { "epoch": 0.9510719480574236, "grad_norm": 2.3528857231140137, "learning_rate": 3.794387707298791e-05, "loss": 7.2554, "step": 22558 }, { "epoch": 0.951114109239623, "grad_norm": 1.9633204936981201, "learning_rate": 3.7943096286579844e-05, "loss": 7.0408, "step": 22559 }, { "epoch": 0.9511562704218226, "grad_norm": 1.4737273454666138, "learning_rate": 3.7942315500171775e-05, "loss": 7.327, "step": 22560 }, { "epoch": 0.9511984316040222, "grad_norm": 1.5601445436477661, "learning_rate": 3.7941534713763706e-05, "loss": 6.4488, "step": 22561 }, { "epoch": 0.9512405927862218, "grad_norm": 1.1930046081542969, "learning_rate": 3.794075392735563e-05, "loss": 6.7074, "step": 22562 }, { "epoch": 0.9512827539684213, "grad_norm": 1.1899012327194214, "learning_rate": 3.793997314094757e-05, "loss": 7.0648, "step": 22563 }, { "epoch": 0.9513249151506208, "grad_norm": 1.4971452951431274, "learning_rate": 3.793919235453949e-05, "loss": 6.6023, "step": 22564 }, { "epoch": 0.9513670763328204, "grad_norm": 2.004092216491699, "learning_rate": 3.7938411568131424e-05, "loss": 6.7888, "step": 22565 }, { "epoch": 0.9514092375150199, "grad_norm": 2.9091272354125977, "learning_rate": 3.7937630781723356e-05, "loss": 8.1565, "step": 22566 }, { "epoch": 0.9514513986972195, "grad_norm": 3.8843002319335938, "learning_rate": 3.793684999531528e-05, "loss": 9.7415, "step": 22567 }, { "epoch": 0.951493559879419, "grad_norm": 3.50801944732666, "learning_rate": 3.793606920890722e-05, "loss": 7.6135, "step": 22568 }, { "epoch": 0.9515357210616185, "grad_norm": 1.7940855026245117, "learning_rate": 3.793528842249914e-05, "loss": 7.6228, "step": 22569 }, { "epoch": 0.9515778822438181, "grad_norm": 1.98732328414917, "learning_rate": 3.7934507636091074e-05, "loss": 6.6147, "step": 22570 }, { "epoch": 0.9516200434260177, "grad_norm": 1.8101155757904053, "learning_rate": 3.7933726849683005e-05, "loss": 7.1727, "step": 22571 }, { "epoch": 0.9516622046082173, "grad_norm": 2.496295928955078, "learning_rate": 3.793294606327493e-05, "loss": 7.912, "step": 22572 }, { "epoch": 0.9517043657904167, "grad_norm": 3.9590070247650146, "learning_rate": 3.793216527686686e-05, "loss": 7.4694, "step": 22573 }, { "epoch": 0.9517465269726163, "grad_norm": 2.738098382949829, "learning_rate": 3.793138449045879e-05, "loss": 8.0573, "step": 22574 }, { "epoch": 0.9517886881548159, "grad_norm": 2.4418885707855225, "learning_rate": 3.7930603704050723e-05, "loss": 7.6411, "step": 22575 }, { "epoch": 0.9518308493370154, "grad_norm": 1.044925570487976, "learning_rate": 3.792982291764265e-05, "loss": 7.0461, "step": 22576 }, { "epoch": 0.9518730105192149, "grad_norm": 2.5526068210601807, "learning_rate": 3.792904213123458e-05, "loss": 6.6934, "step": 22577 }, { "epoch": 0.9519151717014145, "grad_norm": 1.1888186931610107, "learning_rate": 3.792826134482651e-05, "loss": 7.0678, "step": 22578 }, { "epoch": 0.9519573328836141, "grad_norm": 1.318532109260559, "learning_rate": 3.792748055841844e-05, "loss": 6.5852, "step": 22579 }, { "epoch": 0.9519994940658136, "grad_norm": 1.1030982732772827, "learning_rate": 3.792669977201037e-05, "loss": 6.635, "step": 22580 }, { "epoch": 0.9520416552480132, "grad_norm": 1.727339267730713, "learning_rate": 3.79259189856023e-05, "loss": 6.5278, "step": 22581 }, { "epoch": 0.9520838164302127, "grad_norm": 0.9927193522453308, "learning_rate": 3.7925138199194235e-05, "loss": 6.9241, "step": 22582 }, { "epoch": 0.9521259776124122, "grad_norm": 1.2003980875015259, "learning_rate": 3.792435741278616e-05, "loss": 6.5186, "step": 22583 }, { "epoch": 0.9521681387946118, "grad_norm": 3.1633951663970947, "learning_rate": 3.7923576626378084e-05, "loss": 6.4932, "step": 22584 }, { "epoch": 0.9522102999768114, "grad_norm": 1.1298655271530151, "learning_rate": 3.792279583997002e-05, "loss": 6.6217, "step": 22585 }, { "epoch": 0.9522524611590109, "grad_norm": 1.2653446197509766, "learning_rate": 3.792201505356195e-05, "loss": 7.0007, "step": 22586 }, { "epoch": 0.9522946223412104, "grad_norm": 1.1980725526809692, "learning_rate": 3.792123426715388e-05, "loss": 6.6768, "step": 22587 }, { "epoch": 0.95233678352341, "grad_norm": 1.5201618671417236, "learning_rate": 3.792045348074581e-05, "loss": 6.9615, "step": 22588 }, { "epoch": 0.9523789447056096, "grad_norm": 1.0315228700637817, "learning_rate": 3.791967269433774e-05, "loss": 6.6634, "step": 22589 }, { "epoch": 0.9524211058878091, "grad_norm": 1.397991418838501, "learning_rate": 3.791889190792967e-05, "loss": 7.0415, "step": 22590 }, { "epoch": 0.9524632670700086, "grad_norm": 2.855605125427246, "learning_rate": 3.7918111121521597e-05, "loss": 6.6112, "step": 22591 }, { "epoch": 0.9525054282522082, "grad_norm": 1.2845202684402466, "learning_rate": 3.791733033511353e-05, "loss": 6.796, "step": 22592 }, { "epoch": 0.9525475894344078, "grad_norm": 0.8968666195869446, "learning_rate": 3.791654954870546e-05, "loss": 6.6853, "step": 22593 }, { "epoch": 0.9525897506166073, "grad_norm": 2.312650442123413, "learning_rate": 3.791576876229739e-05, "loss": 7.099, "step": 22594 }, { "epoch": 0.9526319117988068, "grad_norm": 1.8761143684387207, "learning_rate": 3.7914987975889315e-05, "loss": 7.0182, "step": 22595 }, { "epoch": 0.9526740729810064, "grad_norm": 3.1968495845794678, "learning_rate": 3.7914207189481246e-05, "loss": 7.3723, "step": 22596 }, { "epoch": 0.9527162341632059, "grad_norm": 0.9721919894218445, "learning_rate": 3.791342640307318e-05, "loss": 6.76, "step": 22597 }, { "epoch": 0.9527583953454055, "grad_norm": 1.5637476444244385, "learning_rate": 3.79126456166651e-05, "loss": 6.4322, "step": 22598 }, { "epoch": 0.952800556527605, "grad_norm": 3.3148770332336426, "learning_rate": 3.791186483025704e-05, "loss": 8.6059, "step": 22599 }, { "epoch": 0.9528427177098046, "grad_norm": 1.6033095121383667, "learning_rate": 3.7911084043848964e-05, "loss": 6.5235, "step": 22600 }, { "epoch": 0.9528848788920041, "grad_norm": 1.347951054573059, "learning_rate": 3.79103032574409e-05, "loss": 6.7557, "step": 22601 }, { "epoch": 0.9529270400742037, "grad_norm": 1.2860628366470337, "learning_rate": 3.790952247103283e-05, "loss": 6.7351, "step": 22602 }, { "epoch": 0.9529692012564033, "grad_norm": 0.9670464992523193, "learning_rate": 3.790874168462475e-05, "loss": 6.8786, "step": 22603 }, { "epoch": 0.9530113624386027, "grad_norm": 1.0311927795410156, "learning_rate": 3.790796089821669e-05, "loss": 6.4907, "step": 22604 }, { "epoch": 0.9530535236208023, "grad_norm": 4.1253156661987305, "learning_rate": 3.7907180111808614e-05, "loss": 6.5292, "step": 22605 }, { "epoch": 0.9530956848030019, "grad_norm": 2.388110399246216, "learning_rate": 3.7906399325400545e-05, "loss": 6.453, "step": 22606 }, { "epoch": 0.9531378459852015, "grad_norm": 2.917093276977539, "learning_rate": 3.7905618538992476e-05, "loss": 7.5296, "step": 22607 }, { "epoch": 0.9531800071674009, "grad_norm": 1.187200903892517, "learning_rate": 3.790483775258441e-05, "loss": 6.4604, "step": 22608 }, { "epoch": 0.9532221683496005, "grad_norm": 1.4158029556274414, "learning_rate": 3.790405696617633e-05, "loss": 7.5677, "step": 22609 }, { "epoch": 0.9532643295318001, "grad_norm": 0.986183226108551, "learning_rate": 3.790327617976826e-05, "loss": 6.5318, "step": 22610 }, { "epoch": 0.9533064907139996, "grad_norm": 2.9597206115722656, "learning_rate": 3.7902495393360195e-05, "loss": 6.3606, "step": 22611 }, { "epoch": 0.9533486518961992, "grad_norm": 1.2501763105392456, "learning_rate": 3.7901714606952126e-05, "loss": 7.6933, "step": 22612 }, { "epoch": 0.9533908130783987, "grad_norm": 2.080676317214966, "learning_rate": 3.790093382054406e-05, "loss": 7.1457, "step": 22613 }, { "epoch": 0.9534329742605983, "grad_norm": 2.0090270042419434, "learning_rate": 3.790015303413598e-05, "loss": 7.0027, "step": 22614 }, { "epoch": 0.9534751354427978, "grad_norm": 2.4806854724884033, "learning_rate": 3.789937224772791e-05, "loss": 7.1386, "step": 22615 }, { "epoch": 0.9535172966249974, "grad_norm": 3.5331435203552246, "learning_rate": 3.7898591461319844e-05, "loss": 7.9784, "step": 22616 }, { "epoch": 0.9535594578071969, "grad_norm": 2.040743589401245, "learning_rate": 3.789781067491177e-05, "loss": 6.5806, "step": 22617 }, { "epoch": 0.9536016189893964, "grad_norm": 2.88986873626709, "learning_rate": 3.789702988850371e-05, "loss": 7.6513, "step": 22618 }, { "epoch": 0.953643780171596, "grad_norm": 3.52836012840271, "learning_rate": 3.789624910209563e-05, "loss": 8.1672, "step": 22619 }, { "epoch": 0.9536859413537956, "grad_norm": 1.0160173177719116, "learning_rate": 3.789546831568756e-05, "loss": 6.531, "step": 22620 }, { "epoch": 0.9537281025359952, "grad_norm": 1.1493968963623047, "learning_rate": 3.7894687529279494e-05, "loss": 7.2419, "step": 22621 }, { "epoch": 0.9537702637181946, "grad_norm": 1.7043066024780273, "learning_rate": 3.789390674287142e-05, "loss": 7.2227, "step": 22622 }, { "epoch": 0.9538124249003942, "grad_norm": 1.332276701927185, "learning_rate": 3.789312595646335e-05, "loss": 6.9939, "step": 22623 }, { "epoch": 0.9538545860825938, "grad_norm": 1.797560691833496, "learning_rate": 3.789234517005528e-05, "loss": 6.7527, "step": 22624 }, { "epoch": 0.9538967472647933, "grad_norm": 3.7494935989379883, "learning_rate": 3.789156438364721e-05, "loss": 7.7327, "step": 22625 }, { "epoch": 0.9539389084469928, "grad_norm": 1.6330960988998413, "learning_rate": 3.789078359723914e-05, "loss": 6.6065, "step": 22626 }, { "epoch": 0.9539810696291924, "grad_norm": 1.1765152215957642, "learning_rate": 3.7890002810831074e-05, "loss": 7.0208, "step": 22627 }, { "epoch": 0.954023230811392, "grad_norm": 2.6566483974456787, "learning_rate": 3.7889222024423e-05, "loss": 8.221, "step": 22628 }, { "epoch": 0.9540653919935915, "grad_norm": 1.6731350421905518, "learning_rate": 3.788844123801493e-05, "loss": 6.5614, "step": 22629 }, { "epoch": 0.9541075531757911, "grad_norm": 1.0571564435958862, "learning_rate": 3.788766045160686e-05, "loss": 6.6443, "step": 22630 }, { "epoch": 0.9541497143579906, "grad_norm": 1.3965357542037964, "learning_rate": 3.7886879665198786e-05, "loss": 6.8064, "step": 22631 }, { "epoch": 0.9541918755401901, "grad_norm": 2.8584916591644287, "learning_rate": 3.7886098878790724e-05, "loss": 7.4545, "step": 22632 }, { "epoch": 0.9542340367223897, "grad_norm": 1.8376632928848267, "learning_rate": 3.788531809238265e-05, "loss": 6.4322, "step": 22633 }, { "epoch": 0.9542761979045893, "grad_norm": 2.9133803844451904, "learning_rate": 3.788453730597458e-05, "loss": 8.1145, "step": 22634 }, { "epoch": 0.9543183590867887, "grad_norm": 1.0898243188858032, "learning_rate": 3.788375651956651e-05, "loss": 7.186, "step": 22635 }, { "epoch": 0.9543605202689883, "grad_norm": 1.5768593549728394, "learning_rate": 3.7882975733158435e-05, "loss": 6.4111, "step": 22636 }, { "epoch": 0.9544026814511879, "grad_norm": 1.7753074169158936, "learning_rate": 3.7882194946750373e-05, "loss": 6.713, "step": 22637 }, { "epoch": 0.9544448426333875, "grad_norm": 1.419613003730774, "learning_rate": 3.78814141603423e-05, "loss": 6.5046, "step": 22638 }, { "epoch": 0.954487003815587, "grad_norm": 2.7523913383483887, "learning_rate": 3.788063337393423e-05, "loss": 7.5821, "step": 22639 }, { "epoch": 0.9545291649977865, "grad_norm": 2.235276222229004, "learning_rate": 3.787985258752616e-05, "loss": 7.0638, "step": 22640 }, { "epoch": 0.9545713261799861, "grad_norm": 2.736391305923462, "learning_rate": 3.7879071801118085e-05, "loss": 7.5024, "step": 22641 }, { "epoch": 0.9546134873621857, "grad_norm": 3.367236614227295, "learning_rate": 3.7878291014710016e-05, "loss": 8.2519, "step": 22642 }, { "epoch": 0.9546556485443852, "grad_norm": 1.2816418409347534, "learning_rate": 3.787751022830195e-05, "loss": 6.6236, "step": 22643 }, { "epoch": 0.9546978097265847, "grad_norm": 1.7328099012374878, "learning_rate": 3.787672944189388e-05, "loss": 7.2091, "step": 22644 }, { "epoch": 0.9547399709087843, "grad_norm": 1.8762586116790771, "learning_rate": 3.78759486554858e-05, "loss": 6.6896, "step": 22645 }, { "epoch": 0.9547821320909838, "grad_norm": 1.1932412385940552, "learning_rate": 3.7875167869077734e-05, "loss": 6.7128, "step": 22646 }, { "epoch": 0.9548242932731834, "grad_norm": 1.3892195224761963, "learning_rate": 3.7874387082669666e-05, "loss": 6.8675, "step": 22647 }, { "epoch": 0.9548664544553829, "grad_norm": 3.679504156112671, "learning_rate": 3.78736062962616e-05, "loss": 7.8387, "step": 22648 }, { "epoch": 0.9549086156375824, "grad_norm": 1.472273826599121, "learning_rate": 3.787282550985353e-05, "loss": 6.6695, "step": 22649 }, { "epoch": 0.954950776819782, "grad_norm": 1.3806177377700806, "learning_rate": 3.787204472344545e-05, "loss": 6.7507, "step": 22650 }, { "epoch": 0.9549929380019816, "grad_norm": 1.5855374336242676, "learning_rate": 3.787126393703739e-05, "loss": 6.5486, "step": 22651 }, { "epoch": 0.9550350991841812, "grad_norm": 1.2898505926132202, "learning_rate": 3.7870483150629315e-05, "loss": 6.64, "step": 22652 }, { "epoch": 0.9550772603663806, "grad_norm": 3.1523170471191406, "learning_rate": 3.7869702364221247e-05, "loss": 7.4942, "step": 22653 }, { "epoch": 0.9551194215485802, "grad_norm": 1.7460803985595703, "learning_rate": 3.786892157781318e-05, "loss": 6.4849, "step": 22654 }, { "epoch": 0.9551615827307798, "grad_norm": 1.9040604829788208, "learning_rate": 3.78681407914051e-05, "loss": 6.5161, "step": 22655 }, { "epoch": 0.9552037439129794, "grad_norm": 1.6194641590118408, "learning_rate": 3.7867360004997034e-05, "loss": 6.4651, "step": 22656 }, { "epoch": 0.9552459050951788, "grad_norm": 2.405789613723755, "learning_rate": 3.7866579218588965e-05, "loss": 7.8424, "step": 22657 }, { "epoch": 0.9552880662773784, "grad_norm": 2.4710817337036133, "learning_rate": 3.7865798432180896e-05, "loss": 7.2635, "step": 22658 }, { "epoch": 0.955330227459578, "grad_norm": 1.296931505203247, "learning_rate": 3.786501764577283e-05, "loss": 6.5354, "step": 22659 }, { "epoch": 0.9553723886417775, "grad_norm": 1.9520378112792969, "learning_rate": 3.786423685936475e-05, "loss": 6.6075, "step": 22660 }, { "epoch": 0.9554145498239771, "grad_norm": 2.165837287902832, "learning_rate": 3.786345607295668e-05, "loss": 6.9917, "step": 22661 }, { "epoch": 0.9554567110061766, "grad_norm": 1.302312970161438, "learning_rate": 3.7862675286548614e-05, "loss": 6.9367, "step": 22662 }, { "epoch": 0.9554988721883761, "grad_norm": 2.5313711166381836, "learning_rate": 3.7861894500140546e-05, "loss": 7.7053, "step": 22663 }, { "epoch": 0.9555410333705757, "grad_norm": 1.4748841524124146, "learning_rate": 3.786111371373247e-05, "loss": 7.2028, "step": 22664 }, { "epoch": 0.9555831945527753, "grad_norm": 5.453263759613037, "learning_rate": 3.78603329273244e-05, "loss": 7.1168, "step": 22665 }, { "epoch": 0.9556253557349748, "grad_norm": 1.1835813522338867, "learning_rate": 3.785955214091633e-05, "loss": 6.621, "step": 22666 }, { "epoch": 0.9556675169171743, "grad_norm": 1.6082850694656372, "learning_rate": 3.785877135450826e-05, "loss": 6.793, "step": 22667 }, { "epoch": 0.9557096780993739, "grad_norm": 1.6634217500686646, "learning_rate": 3.7857990568100195e-05, "loss": 6.4696, "step": 22668 }, { "epoch": 0.9557518392815735, "grad_norm": 2.808187484741211, "learning_rate": 3.785720978169212e-05, "loss": 7.6625, "step": 22669 }, { "epoch": 0.955794000463773, "grad_norm": 1.9425528049468994, "learning_rate": 3.785642899528406e-05, "loss": 7.2417, "step": 22670 }, { "epoch": 0.9558361616459725, "grad_norm": 2.0635199546813965, "learning_rate": 3.785564820887598e-05, "loss": 6.5764, "step": 22671 }, { "epoch": 0.9558783228281721, "grad_norm": 1.7937999963760376, "learning_rate": 3.785486742246791e-05, "loss": 7.3868, "step": 22672 }, { "epoch": 0.9559204840103717, "grad_norm": 1.2827730178833008, "learning_rate": 3.7854086636059845e-05, "loss": 6.8058, "step": 22673 }, { "epoch": 0.9559626451925712, "grad_norm": 2.6005873680114746, "learning_rate": 3.785330584965177e-05, "loss": 7.96, "step": 22674 }, { "epoch": 0.9560048063747707, "grad_norm": 2.116919755935669, "learning_rate": 3.78525250632437e-05, "loss": 6.6996, "step": 22675 }, { "epoch": 0.9560469675569703, "grad_norm": 2.6529572010040283, "learning_rate": 3.785174427683563e-05, "loss": 6.3968, "step": 22676 }, { "epoch": 0.9560891287391698, "grad_norm": 1.8587636947631836, "learning_rate": 3.785096349042756e-05, "loss": 7.0977, "step": 22677 }, { "epoch": 0.9561312899213694, "grad_norm": 4.080406665802002, "learning_rate": 3.785018270401949e-05, "loss": 7.4661, "step": 22678 }, { "epoch": 0.956173451103569, "grad_norm": 0.9892644882202148, "learning_rate": 3.784940191761142e-05, "loss": 6.9372, "step": 22679 }, { "epoch": 0.9562156122857685, "grad_norm": 2.681485414505005, "learning_rate": 3.784862113120335e-05, "loss": 7.6987, "step": 22680 }, { "epoch": 0.956257773467968, "grad_norm": 2.0477371215820312, "learning_rate": 3.784784034479528e-05, "loss": 6.9855, "step": 22681 }, { "epoch": 0.9562999346501676, "grad_norm": 3.3077588081359863, "learning_rate": 3.784705955838721e-05, "loss": 7.4452, "step": 22682 }, { "epoch": 0.9563420958323672, "grad_norm": 1.7453217506408691, "learning_rate": 3.784627877197914e-05, "loss": 6.5301, "step": 22683 }, { "epoch": 0.9563842570145666, "grad_norm": 1.0003758668899536, "learning_rate": 3.784549798557107e-05, "loss": 6.6137, "step": 22684 }, { "epoch": 0.9564264181967662, "grad_norm": 1.3880730867385864, "learning_rate": 3.7844717199163e-05, "loss": 6.5024, "step": 22685 }, { "epoch": 0.9564685793789658, "grad_norm": 3.2313122749328613, "learning_rate": 3.7843936412754924e-05, "loss": 7.621, "step": 22686 }, { "epoch": 0.9565107405611654, "grad_norm": 2.119551658630371, "learning_rate": 3.784315562634686e-05, "loss": 6.6183, "step": 22687 }, { "epoch": 0.9565529017433649, "grad_norm": 1.526527762413025, "learning_rate": 3.7842374839938786e-05, "loss": 6.9311, "step": 22688 }, { "epoch": 0.9565950629255644, "grad_norm": 2.3847122192382812, "learning_rate": 3.784159405353072e-05, "loss": 6.822, "step": 22689 }, { "epoch": 0.956637224107764, "grad_norm": 1.2974975109100342, "learning_rate": 3.784081326712265e-05, "loss": 7.2555, "step": 22690 }, { "epoch": 0.9566793852899635, "grad_norm": 2.1646292209625244, "learning_rate": 3.7840032480714573e-05, "loss": 7.2283, "step": 22691 }, { "epoch": 0.9567215464721631, "grad_norm": 1.0111103057861328, "learning_rate": 3.783925169430651e-05, "loss": 6.9772, "step": 22692 }, { "epoch": 0.9567637076543626, "grad_norm": 3.0866825580596924, "learning_rate": 3.7838470907898436e-05, "loss": 7.6352, "step": 22693 }, { "epoch": 0.9568058688365622, "grad_norm": 1.9487929344177246, "learning_rate": 3.783769012149037e-05, "loss": 6.577, "step": 22694 }, { "epoch": 0.9568480300187617, "grad_norm": 6.734535217285156, "learning_rate": 3.78369093350823e-05, "loss": 11.8922, "step": 22695 }, { "epoch": 0.9568901912009613, "grad_norm": 2.1346662044525146, "learning_rate": 3.783612854867423e-05, "loss": 6.4812, "step": 22696 }, { "epoch": 0.9569323523831609, "grad_norm": 2.751964569091797, "learning_rate": 3.7835347762266154e-05, "loss": 7.4125, "step": 22697 }, { "epoch": 0.9569745135653603, "grad_norm": 7.726218223571777, "learning_rate": 3.7834566975858085e-05, "loss": 13.625, "step": 22698 }, { "epoch": 0.9570166747475599, "grad_norm": 2.259854793548584, "learning_rate": 3.783378618945002e-05, "loss": 8.0151, "step": 22699 }, { "epoch": 0.9570588359297595, "grad_norm": 1.7472398281097412, "learning_rate": 3.783300540304194e-05, "loss": 6.7238, "step": 22700 }, { "epoch": 0.9571009971119591, "grad_norm": 1.5695987939834595, "learning_rate": 3.783222461663388e-05, "loss": 6.4661, "step": 22701 }, { "epoch": 0.9571431582941585, "grad_norm": 2.62296724319458, "learning_rate": 3.7831443830225804e-05, "loss": 7.6958, "step": 22702 }, { "epoch": 0.9571853194763581, "grad_norm": 1.983044981956482, "learning_rate": 3.7830663043817735e-05, "loss": 6.8933, "step": 22703 }, { "epoch": 0.9572274806585577, "grad_norm": 3.096994638442993, "learning_rate": 3.7829882257409666e-05, "loss": 7.3382, "step": 22704 }, { "epoch": 0.9572696418407572, "grad_norm": 2.1476080417633057, "learning_rate": 3.782910147100159e-05, "loss": 7.6568, "step": 22705 }, { "epoch": 0.9573118030229567, "grad_norm": 0.9824667572975159, "learning_rate": 3.782832068459353e-05, "loss": 6.9436, "step": 22706 }, { "epoch": 0.9573539642051563, "grad_norm": 1.2766520977020264, "learning_rate": 3.782753989818545e-05, "loss": 6.742, "step": 22707 }, { "epoch": 0.9573961253873559, "grad_norm": 2.9282007217407227, "learning_rate": 3.7826759111777384e-05, "loss": 8.2536, "step": 22708 }, { "epoch": 0.9574382865695554, "grad_norm": 1.2915503978729248, "learning_rate": 3.7825978325369316e-05, "loss": 6.666, "step": 22709 }, { "epoch": 0.957480447751755, "grad_norm": 1.4075238704681396, "learning_rate": 3.782519753896124e-05, "loss": 6.5277, "step": 22710 }, { "epoch": 0.9575226089339545, "grad_norm": 1.8564915657043457, "learning_rate": 3.782441675255317e-05, "loss": 7.0382, "step": 22711 }, { "epoch": 0.957564770116154, "grad_norm": 1.3813585042953491, "learning_rate": 3.78236359661451e-05, "loss": 6.5078, "step": 22712 }, { "epoch": 0.9576069312983536, "grad_norm": 2.063279151916504, "learning_rate": 3.7822855179737034e-05, "loss": 6.5367, "step": 22713 }, { "epoch": 0.9576490924805532, "grad_norm": 1.8564374446868896, "learning_rate": 3.782207439332896e-05, "loss": 7.5436, "step": 22714 }, { "epoch": 0.9576912536627527, "grad_norm": 1.095546007156372, "learning_rate": 3.7821293606920897e-05, "loss": 6.6886, "step": 22715 }, { "epoch": 0.9577334148449522, "grad_norm": 2.221007823944092, "learning_rate": 3.782051282051282e-05, "loss": 7.1788, "step": 22716 }, { "epoch": 0.9577755760271518, "grad_norm": 3.5246169567108154, "learning_rate": 3.781973203410475e-05, "loss": 7.779, "step": 22717 }, { "epoch": 0.9578177372093514, "grad_norm": 1.3215540647506714, "learning_rate": 3.7818951247696684e-05, "loss": 7.0604, "step": 22718 }, { "epoch": 0.957859898391551, "grad_norm": 1.172269344329834, "learning_rate": 3.781817046128861e-05, "loss": 7.1322, "step": 22719 }, { "epoch": 0.9579020595737504, "grad_norm": 1.6280848979949951, "learning_rate": 3.7817389674880546e-05, "loss": 6.7861, "step": 22720 }, { "epoch": 0.95794422075595, "grad_norm": 1.0902572870254517, "learning_rate": 3.781660888847247e-05, "loss": 6.824, "step": 22721 }, { "epoch": 0.9579863819381496, "grad_norm": 1.2780613899230957, "learning_rate": 3.78158281020644e-05, "loss": 6.8068, "step": 22722 }, { "epoch": 0.9580285431203491, "grad_norm": 1.1335511207580566, "learning_rate": 3.781504731565633e-05, "loss": 7.1898, "step": 22723 }, { "epoch": 0.9580707043025486, "grad_norm": 2.68082594871521, "learning_rate": 3.781426652924826e-05, "loss": 6.6181, "step": 22724 }, { "epoch": 0.9581128654847482, "grad_norm": 1.9532586336135864, "learning_rate": 3.781348574284019e-05, "loss": 6.7574, "step": 22725 }, { "epoch": 0.9581550266669477, "grad_norm": 3.2245635986328125, "learning_rate": 3.781270495643212e-05, "loss": 7.6961, "step": 22726 }, { "epoch": 0.9581971878491473, "grad_norm": 5.721651077270508, "learning_rate": 3.781192417002405e-05, "loss": 11.4669, "step": 22727 }, { "epoch": 0.9582393490313469, "grad_norm": 2.299468994140625, "learning_rate": 3.781114338361598e-05, "loss": 6.4961, "step": 22728 }, { "epoch": 0.9582815102135464, "grad_norm": 1.3870553970336914, "learning_rate": 3.781036259720791e-05, "loss": 6.6406, "step": 22729 }, { "epoch": 0.9583236713957459, "grad_norm": 2.3327014446258545, "learning_rate": 3.780958181079984e-05, "loss": 7.4673, "step": 22730 }, { "epoch": 0.9583658325779455, "grad_norm": 1.417500615119934, "learning_rate": 3.780880102439177e-05, "loss": 6.897, "step": 22731 }, { "epoch": 0.9584079937601451, "grad_norm": 1.5882271528244019, "learning_rate": 3.78080202379837e-05, "loss": 6.5783, "step": 22732 }, { "epoch": 0.9584501549423445, "grad_norm": 1.2925899028778076, "learning_rate": 3.7807239451575625e-05, "loss": 6.7306, "step": 22733 }, { "epoch": 0.9584923161245441, "grad_norm": 1.6248141527175903, "learning_rate": 3.780645866516756e-05, "loss": 7.2167, "step": 22734 }, { "epoch": 0.9585344773067437, "grad_norm": 1.5118151903152466, "learning_rate": 3.780567787875949e-05, "loss": 6.4609, "step": 22735 }, { "epoch": 0.9585766384889433, "grad_norm": 2.1544594764709473, "learning_rate": 3.780489709235141e-05, "loss": 6.5055, "step": 22736 }, { "epoch": 0.9586187996711428, "grad_norm": 2.142935276031494, "learning_rate": 3.780411630594335e-05, "loss": 7.2313, "step": 22737 }, { "epoch": 0.9586609608533423, "grad_norm": 2.3542773723602295, "learning_rate": 3.7803335519535275e-05, "loss": 7.8897, "step": 22738 }, { "epoch": 0.9587031220355419, "grad_norm": 4.113974571228027, "learning_rate": 3.780255473312721e-05, "loss": 7.7243, "step": 22739 }, { "epoch": 0.9587452832177414, "grad_norm": 3.384472131729126, "learning_rate": 3.780177394671914e-05, "loss": 7.8069, "step": 22740 }, { "epoch": 0.958787444399941, "grad_norm": 3.6250052452087402, "learning_rate": 3.780099316031107e-05, "loss": 8.0164, "step": 22741 }, { "epoch": 0.9588296055821405, "grad_norm": 1.3484481573104858, "learning_rate": 3.7800212373903e-05, "loss": 7.1031, "step": 22742 }, { "epoch": 0.95887176676434, "grad_norm": 1.7747604846954346, "learning_rate": 3.7799431587494924e-05, "loss": 6.6285, "step": 22743 }, { "epoch": 0.9589139279465396, "grad_norm": 2.3798229694366455, "learning_rate": 3.7798650801086856e-05, "loss": 7.6554, "step": 22744 }, { "epoch": 0.9589560891287392, "grad_norm": 1.4486619234085083, "learning_rate": 3.779787001467879e-05, "loss": 7.0147, "step": 22745 }, { "epoch": 0.9589982503109388, "grad_norm": 4.449791431427002, "learning_rate": 3.779708922827072e-05, "loss": 7.7746, "step": 22746 }, { "epoch": 0.9590404114931382, "grad_norm": 2.3798301219940186, "learning_rate": 3.779630844186264e-05, "loss": 7.2679, "step": 22747 }, { "epoch": 0.9590825726753378, "grad_norm": 3.2140095233917236, "learning_rate": 3.7795527655454574e-05, "loss": 8.6104, "step": 22748 }, { "epoch": 0.9591247338575374, "grad_norm": 3.5325050354003906, "learning_rate": 3.7794746869046505e-05, "loss": 6.4935, "step": 22749 }, { "epoch": 0.959166895039737, "grad_norm": 3.1425294876098633, "learning_rate": 3.7793966082638436e-05, "loss": 8.5093, "step": 22750 }, { "epoch": 0.9592090562219364, "grad_norm": 1.2282453775405884, "learning_rate": 3.779318529623037e-05, "loss": 6.5987, "step": 22751 }, { "epoch": 0.959251217404136, "grad_norm": 2.2648701667785645, "learning_rate": 3.779240450982229e-05, "loss": 6.9895, "step": 22752 }, { "epoch": 0.9592933785863356, "grad_norm": 3.4227027893066406, "learning_rate": 3.7791623723414223e-05, "loss": 7.9448, "step": 22753 }, { "epoch": 0.9593355397685351, "grad_norm": 1.7575976848602295, "learning_rate": 3.7790842937006155e-05, "loss": 6.5457, "step": 22754 }, { "epoch": 0.9593777009507347, "grad_norm": 2.6132652759552, "learning_rate": 3.779006215059808e-05, "loss": 7.509, "step": 22755 }, { "epoch": 0.9594198621329342, "grad_norm": 1.7451311349868774, "learning_rate": 3.778928136419002e-05, "loss": 6.6052, "step": 22756 }, { "epoch": 0.9594620233151337, "grad_norm": 1.881410837173462, "learning_rate": 3.778850057778194e-05, "loss": 6.8718, "step": 22757 }, { "epoch": 0.9595041844973333, "grad_norm": 2.1251533031463623, "learning_rate": 3.778771979137387e-05, "loss": 6.9782, "step": 22758 }, { "epoch": 0.9595463456795329, "grad_norm": 2.528045415878296, "learning_rate": 3.7786939004965804e-05, "loss": 7.5019, "step": 22759 }, { "epoch": 0.9595885068617324, "grad_norm": 1.1484123468399048, "learning_rate": 3.7786158218557735e-05, "loss": 6.5802, "step": 22760 }, { "epoch": 0.9596306680439319, "grad_norm": 3.1785974502563477, "learning_rate": 3.778537743214967e-05, "loss": 7.4778, "step": 22761 }, { "epoch": 0.9596728292261315, "grad_norm": 1.9125388860702515, "learning_rate": 3.778459664574159e-05, "loss": 7.1272, "step": 22762 }, { "epoch": 0.9597149904083311, "grad_norm": 2.1418073177337646, "learning_rate": 3.778381585933352e-05, "loss": 6.624, "step": 22763 }, { "epoch": 0.9597571515905305, "grad_norm": 1.4285447597503662, "learning_rate": 3.7783035072925454e-05, "loss": 6.7135, "step": 22764 }, { "epoch": 0.9597993127727301, "grad_norm": 0.9399662017822266, "learning_rate": 3.7782254286517385e-05, "loss": 7.2157, "step": 22765 }, { "epoch": 0.9598414739549297, "grad_norm": 2.7272331714630127, "learning_rate": 3.778147350010931e-05, "loss": 7.4819, "step": 22766 }, { "epoch": 0.9598836351371293, "grad_norm": 2.693143367767334, "learning_rate": 3.778069271370124e-05, "loss": 7.8715, "step": 22767 }, { "epoch": 0.9599257963193288, "grad_norm": 3.743364095687866, "learning_rate": 3.777991192729317e-05, "loss": 6.7715, "step": 22768 }, { "epoch": 0.9599679575015283, "grad_norm": 2.4867806434631348, "learning_rate": 3.7779131140885096e-05, "loss": 6.7968, "step": 22769 }, { "epoch": 0.9600101186837279, "grad_norm": 1.9877935647964478, "learning_rate": 3.7778350354477035e-05, "loss": 6.5931, "step": 22770 }, { "epoch": 0.9600522798659274, "grad_norm": 1.8230832815170288, "learning_rate": 3.777756956806896e-05, "loss": 7.704, "step": 22771 }, { "epoch": 0.960094441048127, "grad_norm": 1.8497873544692993, "learning_rate": 3.777678878166089e-05, "loss": 6.8763, "step": 22772 }, { "epoch": 0.9601366022303265, "grad_norm": 1.846433401107788, "learning_rate": 3.777600799525282e-05, "loss": 6.7787, "step": 22773 }, { "epoch": 0.9601787634125261, "grad_norm": 1.8704718351364136, "learning_rate": 3.7775227208844746e-05, "loss": 6.6603, "step": 22774 }, { "epoch": 0.9602209245947256, "grad_norm": 1.299026370048523, "learning_rate": 3.7774446422436684e-05, "loss": 6.5996, "step": 22775 }, { "epoch": 0.9602630857769252, "grad_norm": 1.1083866357803345, "learning_rate": 3.777366563602861e-05, "loss": 6.6796, "step": 22776 }, { "epoch": 0.9603052469591248, "grad_norm": 1.6042265892028809, "learning_rate": 3.777288484962054e-05, "loss": 6.5223, "step": 22777 }, { "epoch": 0.9603474081413242, "grad_norm": 2.0007376670837402, "learning_rate": 3.777210406321247e-05, "loss": 6.6816, "step": 22778 }, { "epoch": 0.9603895693235238, "grad_norm": 1.70357084274292, "learning_rate": 3.77713232768044e-05, "loss": 6.6156, "step": 22779 }, { "epoch": 0.9604317305057234, "grad_norm": 1.4903852939605713, "learning_rate": 3.777054249039633e-05, "loss": 7.1035, "step": 22780 }, { "epoch": 0.960473891687923, "grad_norm": 2.7869019508361816, "learning_rate": 3.776976170398826e-05, "loss": 7.9322, "step": 22781 }, { "epoch": 0.9605160528701224, "grad_norm": 1.2772283554077148, "learning_rate": 3.776898091758019e-05, "loss": 6.5748, "step": 22782 }, { "epoch": 0.960558214052322, "grad_norm": 1.1559234857559204, "learning_rate": 3.7768200131172114e-05, "loss": 6.5499, "step": 22783 }, { "epoch": 0.9606003752345216, "grad_norm": 1.6834450960159302, "learning_rate": 3.776741934476405e-05, "loss": 7.1402, "step": 22784 }, { "epoch": 0.9606425364167211, "grad_norm": 3.488393783569336, "learning_rate": 3.7766638558355976e-05, "loss": 7.9057, "step": 22785 }, { "epoch": 0.9606846975989207, "grad_norm": 2.098616361618042, "learning_rate": 3.776585777194791e-05, "loss": 6.3427, "step": 22786 }, { "epoch": 0.9607268587811202, "grad_norm": 1.4075349569320679, "learning_rate": 3.776507698553984e-05, "loss": 6.7014, "step": 22787 }, { "epoch": 0.9607690199633198, "grad_norm": 1.9200985431671143, "learning_rate": 3.776429619913176e-05, "loss": 6.7956, "step": 22788 }, { "epoch": 0.9608111811455193, "grad_norm": 1.8667813539505005, "learning_rate": 3.77635154127237e-05, "loss": 7.1621, "step": 22789 }, { "epoch": 0.9608533423277189, "grad_norm": 1.5081738233566284, "learning_rate": 3.7762734626315626e-05, "loss": 6.7656, "step": 22790 }, { "epoch": 0.9608955035099184, "grad_norm": 2.0806117057800293, "learning_rate": 3.776195383990756e-05, "loss": 8.1543, "step": 22791 }, { "epoch": 0.9609376646921179, "grad_norm": 4.017653465270996, "learning_rate": 3.776117305349949e-05, "loss": 10.0959, "step": 22792 }, { "epoch": 0.9609798258743175, "grad_norm": 2.3388824462890625, "learning_rate": 3.776039226709141e-05, "loss": 6.7335, "step": 22793 }, { "epoch": 0.9610219870565171, "grad_norm": 1.7566744089126587, "learning_rate": 3.7759611480683344e-05, "loss": 7.0145, "step": 22794 }, { "epoch": 0.9610641482387167, "grad_norm": 1.6359525918960571, "learning_rate": 3.7758830694275275e-05, "loss": 7.2825, "step": 22795 }, { "epoch": 0.9611063094209161, "grad_norm": 2.3914694786071777, "learning_rate": 3.775804990786721e-05, "loss": 7.5176, "step": 22796 }, { "epoch": 0.9611484706031157, "grad_norm": 1.5522452592849731, "learning_rate": 3.775726912145914e-05, "loss": 6.9454, "step": 22797 }, { "epoch": 0.9611906317853153, "grad_norm": 1.7682685852050781, "learning_rate": 3.775648833505106e-05, "loss": 7.1063, "step": 22798 }, { "epoch": 0.9612327929675148, "grad_norm": 3.1376373767852783, "learning_rate": 3.7755707548642994e-05, "loss": 6.7457, "step": 22799 }, { "epoch": 0.9612749541497143, "grad_norm": 2.964752197265625, "learning_rate": 3.7754926762234925e-05, "loss": 6.8109, "step": 22800 }, { "epoch": 0.9613171153319139, "grad_norm": 1.1079181432724, "learning_rate": 3.7754145975826856e-05, "loss": 6.7794, "step": 22801 }, { "epoch": 0.9613592765141135, "grad_norm": 3.248309373855591, "learning_rate": 3.775336518941878e-05, "loss": 6.6019, "step": 22802 }, { "epoch": 0.961401437696313, "grad_norm": 1.1825919151306152, "learning_rate": 3.775258440301072e-05, "loss": 6.5406, "step": 22803 }, { "epoch": 0.9614435988785126, "grad_norm": 2.8211171627044678, "learning_rate": 3.775180361660264e-05, "loss": 6.5242, "step": 22804 }, { "epoch": 0.9614857600607121, "grad_norm": 1.4632617235183716, "learning_rate": 3.7751022830194574e-05, "loss": 8.2367, "step": 22805 }, { "epoch": 0.9615279212429116, "grad_norm": 1.4547197818756104, "learning_rate": 3.7750242043786506e-05, "loss": 6.6273, "step": 22806 }, { "epoch": 0.9615700824251112, "grad_norm": 1.55574369430542, "learning_rate": 3.774946125737843e-05, "loss": 7.2042, "step": 22807 }, { "epoch": 0.9616122436073108, "grad_norm": 1.1139658689498901, "learning_rate": 3.774868047097037e-05, "loss": 7.2414, "step": 22808 }, { "epoch": 0.9616544047895103, "grad_norm": 0.9576252102851868, "learning_rate": 3.774789968456229e-05, "loss": 6.6034, "step": 22809 }, { "epoch": 0.9616965659717098, "grad_norm": 2.6690165996551514, "learning_rate": 3.7747118898154224e-05, "loss": 6.6463, "step": 22810 }, { "epoch": 0.9617387271539094, "grad_norm": 1.2339611053466797, "learning_rate": 3.7746338111746155e-05, "loss": 6.9796, "step": 22811 }, { "epoch": 0.961780888336109, "grad_norm": 1.1681801080703735, "learning_rate": 3.774555732533808e-05, "loss": 6.6041, "step": 22812 }, { "epoch": 0.9618230495183085, "grad_norm": 1.1260348558425903, "learning_rate": 3.774477653893001e-05, "loss": 6.484, "step": 22813 }, { "epoch": 0.961865210700508, "grad_norm": 1.6901949644088745, "learning_rate": 3.774399575252194e-05, "loss": 6.9487, "step": 22814 }, { "epoch": 0.9619073718827076, "grad_norm": 3.005095958709717, "learning_rate": 3.7743214966113873e-05, "loss": 7.4148, "step": 22815 }, { "epoch": 0.9619495330649072, "grad_norm": 2.764486074447632, "learning_rate": 3.77424341797058e-05, "loss": 7.7272, "step": 22816 }, { "epoch": 0.9619916942471067, "grad_norm": 1.8909116983413696, "learning_rate": 3.774165339329773e-05, "loss": 6.7401, "step": 22817 }, { "epoch": 0.9620338554293062, "grad_norm": 1.7707394361495972, "learning_rate": 3.774087260688966e-05, "loss": 6.5361, "step": 22818 }, { "epoch": 0.9620760166115058, "grad_norm": 3.1973109245300293, "learning_rate": 3.774009182048159e-05, "loss": 7.57, "step": 22819 }, { "epoch": 0.9621181777937053, "grad_norm": 1.8138320446014404, "learning_rate": 3.773931103407352e-05, "loss": 7.4279, "step": 22820 }, { "epoch": 0.9621603389759049, "grad_norm": 2.6756186485290527, "learning_rate": 3.773853024766545e-05, "loss": 7.6531, "step": 22821 }, { "epoch": 0.9622025001581044, "grad_norm": 1.866391658782959, "learning_rate": 3.7737749461257385e-05, "loss": 7.158, "step": 22822 }, { "epoch": 0.962244661340304, "grad_norm": 1.5870287418365479, "learning_rate": 3.773696867484931e-05, "loss": 6.4712, "step": 22823 }, { "epoch": 0.9622868225225035, "grad_norm": 1.5429060459136963, "learning_rate": 3.773618788844124e-05, "loss": 6.7908, "step": 22824 }, { "epoch": 0.9623289837047031, "grad_norm": 1.5105639696121216, "learning_rate": 3.773540710203317e-05, "loss": 7.1267, "step": 22825 }, { "epoch": 0.9623711448869027, "grad_norm": 1.9592376947402954, "learning_rate": 3.77346263156251e-05, "loss": 7.9304, "step": 22826 }, { "epoch": 0.9624133060691021, "grad_norm": 1.3620179891586304, "learning_rate": 3.773384552921703e-05, "loss": 6.6038, "step": 22827 }, { "epoch": 0.9624554672513017, "grad_norm": 1.0321584939956665, "learning_rate": 3.773306474280896e-05, "loss": 6.4813, "step": 22828 }, { "epoch": 0.9624976284335013, "grad_norm": 2.3489818572998047, "learning_rate": 3.773228395640089e-05, "loss": 7.1096, "step": 22829 }, { "epoch": 0.9625397896157009, "grad_norm": 3.343353033065796, "learning_rate": 3.773150316999282e-05, "loss": 7.8055, "step": 22830 }, { "epoch": 0.9625819507979003, "grad_norm": 1.4371294975280762, "learning_rate": 3.7730722383584746e-05, "loss": 6.5757, "step": 22831 }, { "epoch": 0.9626241119800999, "grad_norm": 2.8778727054595947, "learning_rate": 3.772994159717668e-05, "loss": 6.4044, "step": 22832 }, { "epoch": 0.9626662731622995, "grad_norm": 1.4358819723129272, "learning_rate": 3.772916081076861e-05, "loss": 6.8137, "step": 22833 }, { "epoch": 0.962708434344499, "grad_norm": 1.7161725759506226, "learning_rate": 3.772838002436054e-05, "loss": 6.5088, "step": 22834 }, { "epoch": 0.9627505955266986, "grad_norm": 1.7125566005706787, "learning_rate": 3.7727599237952465e-05, "loss": 6.5536, "step": 22835 }, { "epoch": 0.9627927567088981, "grad_norm": 4.862085819244385, "learning_rate": 3.7726818451544396e-05, "loss": 7.1417, "step": 22836 }, { "epoch": 0.9628349178910977, "grad_norm": 2.192042112350464, "learning_rate": 3.772603766513633e-05, "loss": 6.5053, "step": 22837 }, { "epoch": 0.9628770790732972, "grad_norm": 3.2661185264587402, "learning_rate": 3.772525687872825e-05, "loss": 7.7294, "step": 22838 }, { "epoch": 0.9629192402554968, "grad_norm": 4.472472667694092, "learning_rate": 3.772447609232019e-05, "loss": 6.9433, "step": 22839 }, { "epoch": 0.9629614014376963, "grad_norm": 2.6657967567443848, "learning_rate": 3.7723695305912114e-05, "loss": 6.6521, "step": 22840 }, { "epoch": 0.9630035626198958, "grad_norm": 3.4614555835723877, "learning_rate": 3.772291451950405e-05, "loss": 6.5978, "step": 22841 }, { "epoch": 0.9630457238020954, "grad_norm": 2.927241086959839, "learning_rate": 3.772213373309598e-05, "loss": 7.8596, "step": 22842 }, { "epoch": 0.963087884984295, "grad_norm": 2.9333791732788086, "learning_rate": 3.77213529466879e-05, "loss": 7.449, "step": 22843 }, { "epoch": 0.9631300461664946, "grad_norm": 1.7450804710388184, "learning_rate": 3.772057216027984e-05, "loss": 6.6556, "step": 22844 }, { "epoch": 0.963172207348694, "grad_norm": 3.4785094261169434, "learning_rate": 3.7719791373871764e-05, "loss": 10.3545, "step": 22845 }, { "epoch": 0.9632143685308936, "grad_norm": 2.6475839614868164, "learning_rate": 3.7719010587463695e-05, "loss": 7.6909, "step": 22846 }, { "epoch": 0.9632565297130932, "grad_norm": 2.4580347537994385, "learning_rate": 3.7718229801055626e-05, "loss": 7.101, "step": 22847 }, { "epoch": 0.9632986908952927, "grad_norm": 3.0347914695739746, "learning_rate": 3.771744901464756e-05, "loss": 7.5069, "step": 22848 }, { "epoch": 0.9633408520774922, "grad_norm": 2.646233558654785, "learning_rate": 3.771666822823948e-05, "loss": 6.7733, "step": 22849 }, { "epoch": 0.9633830132596918, "grad_norm": 4.0605926513671875, "learning_rate": 3.771588744183141e-05, "loss": 7.3653, "step": 22850 }, { "epoch": 0.9634251744418914, "grad_norm": 1.2201337814331055, "learning_rate": 3.7715106655423345e-05, "loss": 6.7233, "step": 22851 }, { "epoch": 0.9634673356240909, "grad_norm": 1.3451721668243408, "learning_rate": 3.7714325869015276e-05, "loss": 6.4656, "step": 22852 }, { "epoch": 0.9635094968062905, "grad_norm": 4.090140342712402, "learning_rate": 3.771354508260721e-05, "loss": 7.5218, "step": 22853 }, { "epoch": 0.96355165798849, "grad_norm": 2.4527883529663086, "learning_rate": 3.771276429619913e-05, "loss": 7.2619, "step": 22854 }, { "epoch": 0.9635938191706895, "grad_norm": 4.001238822937012, "learning_rate": 3.771198350979106e-05, "loss": 7.9642, "step": 22855 }, { "epoch": 0.9636359803528891, "grad_norm": 3.7159101963043213, "learning_rate": 3.7711202723382994e-05, "loss": 7.2175, "step": 22856 }, { "epoch": 0.9636781415350887, "grad_norm": 2.9275221824645996, "learning_rate": 3.771042193697492e-05, "loss": 7.6302, "step": 22857 }, { "epoch": 0.9637203027172881, "grad_norm": 1.4324461221694946, "learning_rate": 3.770964115056686e-05, "loss": 6.8041, "step": 22858 }, { "epoch": 0.9637624638994877, "grad_norm": 2.5778427124023438, "learning_rate": 3.770886036415878e-05, "loss": 7.2883, "step": 22859 }, { "epoch": 0.9638046250816873, "grad_norm": 2.612137794494629, "learning_rate": 3.770807957775071e-05, "loss": 6.5872, "step": 22860 }, { "epoch": 0.9638467862638869, "grad_norm": 2.0682997703552246, "learning_rate": 3.7707298791342644e-05, "loss": 7.1921, "step": 22861 }, { "epoch": 0.9638889474460864, "grad_norm": 2.1558997631073, "learning_rate": 3.770651800493457e-05, "loss": 6.7378, "step": 22862 }, { "epoch": 0.9639311086282859, "grad_norm": 2.5841872692108154, "learning_rate": 3.77057372185265e-05, "loss": 6.4899, "step": 22863 }, { "epoch": 0.9639732698104855, "grad_norm": 1.4480767250061035, "learning_rate": 3.770495643211843e-05, "loss": 7.0307, "step": 22864 }, { "epoch": 0.964015430992685, "grad_norm": 1.3716493844985962, "learning_rate": 3.770417564571036e-05, "loss": 6.8161, "step": 22865 }, { "epoch": 0.9640575921748846, "grad_norm": 3.4373903274536133, "learning_rate": 3.770339485930229e-05, "loss": 7.6312, "step": 22866 }, { "epoch": 0.9640997533570841, "grad_norm": 2.875598669052124, "learning_rate": 3.7702614072894224e-05, "loss": 6.9065, "step": 22867 }, { "epoch": 0.9641419145392837, "grad_norm": 2.5240397453308105, "learning_rate": 3.770183328648615e-05, "loss": 7.1124, "step": 22868 }, { "epoch": 0.9641840757214832, "grad_norm": 3.623793601989746, "learning_rate": 3.770105250007808e-05, "loss": 7.7565, "step": 22869 }, { "epoch": 0.9642262369036828, "grad_norm": 2.517387866973877, "learning_rate": 3.770027171367001e-05, "loss": 6.8753, "step": 22870 }, { "epoch": 0.9642683980858824, "grad_norm": 1.1824896335601807, "learning_rate": 3.7699490927261936e-05, "loss": 6.7561, "step": 22871 }, { "epoch": 0.9643105592680818, "grad_norm": 1.810579776763916, "learning_rate": 3.7698710140853874e-05, "loss": 6.5025, "step": 22872 }, { "epoch": 0.9643527204502814, "grad_norm": 4.034529685974121, "learning_rate": 3.76979293544458e-05, "loss": 9.6699, "step": 22873 }, { "epoch": 0.964394881632481, "grad_norm": 1.5980327129364014, "learning_rate": 3.769714856803773e-05, "loss": 6.7699, "step": 22874 }, { "epoch": 0.9644370428146806, "grad_norm": 2.6957387924194336, "learning_rate": 3.769636778162966e-05, "loss": 7.1593, "step": 22875 }, { "epoch": 0.96447920399688, "grad_norm": 1.4868366718292236, "learning_rate": 3.7695586995221585e-05, "loss": 7.4749, "step": 22876 }, { "epoch": 0.9645213651790796, "grad_norm": 2.307474374771118, "learning_rate": 3.7694806208813523e-05, "loss": 7.7214, "step": 22877 }, { "epoch": 0.9645635263612792, "grad_norm": 1.2030420303344727, "learning_rate": 3.769402542240545e-05, "loss": 7.0387, "step": 22878 }, { "epoch": 0.9646056875434788, "grad_norm": 0.7402406334877014, "learning_rate": 3.769324463599738e-05, "loss": 6.4668, "step": 22879 }, { "epoch": 0.9646478487256782, "grad_norm": 2.4676201343536377, "learning_rate": 3.769246384958931e-05, "loss": 7.1728, "step": 22880 }, { "epoch": 0.9646900099078778, "grad_norm": 3.0833420753479004, "learning_rate": 3.7691683063181235e-05, "loss": 7.3798, "step": 22881 }, { "epoch": 0.9647321710900774, "grad_norm": 1.4289000034332275, "learning_rate": 3.7690902276773166e-05, "loss": 6.8301, "step": 22882 }, { "epoch": 0.9647743322722769, "grad_norm": 1.0248901844024658, "learning_rate": 3.76901214903651e-05, "loss": 6.8222, "step": 22883 }, { "epoch": 0.9648164934544765, "grad_norm": 3.2718865871429443, "learning_rate": 3.768934070395703e-05, "loss": 7.6008, "step": 22884 }, { "epoch": 0.964858654636676, "grad_norm": 2.4300739765167236, "learning_rate": 3.768855991754895e-05, "loss": 6.663, "step": 22885 }, { "epoch": 0.9649008158188755, "grad_norm": 1.73336923122406, "learning_rate": 3.768777913114089e-05, "loss": 6.5483, "step": 22886 }, { "epoch": 0.9649429770010751, "grad_norm": 1.515459656715393, "learning_rate": 3.7686998344732816e-05, "loss": 6.538, "step": 22887 }, { "epoch": 0.9649851381832747, "grad_norm": 0.9191434979438782, "learning_rate": 3.768621755832475e-05, "loss": 6.7635, "step": 22888 }, { "epoch": 0.9650272993654742, "grad_norm": 3.601458787918091, "learning_rate": 3.768543677191668e-05, "loss": 7.7522, "step": 22889 }, { "epoch": 0.9650694605476737, "grad_norm": 3.723207950592041, "learning_rate": 3.76846559855086e-05, "loss": 6.9873, "step": 22890 }, { "epoch": 0.9651116217298733, "grad_norm": 3.5770323276519775, "learning_rate": 3.768387519910054e-05, "loss": 7.7856, "step": 22891 }, { "epoch": 0.9651537829120729, "grad_norm": 1.141927719116211, "learning_rate": 3.7683094412692465e-05, "loss": 6.5933, "step": 22892 }, { "epoch": 0.9651959440942725, "grad_norm": 1.3288849592208862, "learning_rate": 3.7682313626284397e-05, "loss": 6.4671, "step": 22893 }, { "epoch": 0.9652381052764719, "grad_norm": 3.1814589500427246, "learning_rate": 3.768153283987633e-05, "loss": 7.9705, "step": 22894 }, { "epoch": 0.9652802664586715, "grad_norm": 3.379958391189575, "learning_rate": 3.768075205346825e-05, "loss": 7.7251, "step": 22895 }, { "epoch": 0.9653224276408711, "grad_norm": 1.0013489723205566, "learning_rate": 3.7679971267060184e-05, "loss": 6.7451, "step": 22896 }, { "epoch": 0.9653645888230706, "grad_norm": 1.4793773889541626, "learning_rate": 3.7679190480652115e-05, "loss": 6.9736, "step": 22897 }, { "epoch": 0.9654067500052701, "grad_norm": 1.021897554397583, "learning_rate": 3.7678409694244046e-05, "loss": 6.7995, "step": 22898 }, { "epoch": 0.9654489111874697, "grad_norm": 1.680151104927063, "learning_rate": 3.767762890783598e-05, "loss": 6.4845, "step": 22899 }, { "epoch": 0.9654910723696692, "grad_norm": 1.541441559791565, "learning_rate": 3.76768481214279e-05, "loss": 6.9159, "step": 22900 }, { "epoch": 0.9655332335518688, "grad_norm": 2.1701419353485107, "learning_rate": 3.767606733501983e-05, "loss": 6.5204, "step": 22901 }, { "epoch": 0.9655753947340684, "grad_norm": 2.815314531326294, "learning_rate": 3.7675286548611764e-05, "loss": 7.0849, "step": 22902 }, { "epoch": 0.9656175559162679, "grad_norm": 1.46829092502594, "learning_rate": 3.7674505762203696e-05, "loss": 7.0116, "step": 22903 }, { "epoch": 0.9656597170984674, "grad_norm": 1.427188754081726, "learning_rate": 3.767372497579562e-05, "loss": 7.0127, "step": 22904 }, { "epoch": 0.965701878280667, "grad_norm": 1.8514535427093506, "learning_rate": 3.767294418938755e-05, "loss": 6.5123, "step": 22905 }, { "epoch": 0.9657440394628666, "grad_norm": 1.3360711336135864, "learning_rate": 3.767216340297948e-05, "loss": 6.9127, "step": 22906 }, { "epoch": 0.965786200645066, "grad_norm": 2.597306728363037, "learning_rate": 3.767138261657141e-05, "loss": 7.9682, "step": 22907 }, { "epoch": 0.9658283618272656, "grad_norm": 3.671644687652588, "learning_rate": 3.7670601830163345e-05, "loss": 7.9665, "step": 22908 }, { "epoch": 0.9658705230094652, "grad_norm": 1.5871886014938354, "learning_rate": 3.766982104375527e-05, "loss": 6.9207, "step": 22909 }, { "epoch": 0.9659126841916648, "grad_norm": 2.0648882389068604, "learning_rate": 3.766904025734721e-05, "loss": 6.9196, "step": 22910 }, { "epoch": 0.9659548453738643, "grad_norm": 1.6440006494522095, "learning_rate": 3.766825947093913e-05, "loss": 6.8019, "step": 22911 }, { "epoch": 0.9659970065560638, "grad_norm": 1.5793325901031494, "learning_rate": 3.766747868453106e-05, "loss": 6.4092, "step": 22912 }, { "epoch": 0.9660391677382634, "grad_norm": 2.427722454071045, "learning_rate": 3.7666697898122995e-05, "loss": 7.6698, "step": 22913 }, { "epoch": 0.966081328920463, "grad_norm": 1.5098850727081299, "learning_rate": 3.766591711171492e-05, "loss": 6.4124, "step": 22914 }, { "epoch": 0.9661234901026625, "grad_norm": 1.7853741645812988, "learning_rate": 3.766513632530685e-05, "loss": 6.4531, "step": 22915 }, { "epoch": 0.966165651284862, "grad_norm": 4.2858099937438965, "learning_rate": 3.766435553889878e-05, "loss": 7.2659, "step": 22916 }, { "epoch": 0.9662078124670616, "grad_norm": 2.792637348175049, "learning_rate": 3.766357475249071e-05, "loss": 7.4161, "step": 22917 }, { "epoch": 0.9662499736492611, "grad_norm": 2.150564670562744, "learning_rate": 3.766279396608264e-05, "loss": 6.9622, "step": 22918 }, { "epoch": 0.9662921348314607, "grad_norm": 1.7417902946472168, "learning_rate": 3.766201317967457e-05, "loss": 6.7901, "step": 22919 }, { "epoch": 0.9663342960136603, "grad_norm": 1.4336035251617432, "learning_rate": 3.76612323932665e-05, "loss": 6.5642, "step": 22920 }, { "epoch": 0.9663764571958597, "grad_norm": 1.3415287733078003, "learning_rate": 3.766045160685843e-05, "loss": 6.5878, "step": 22921 }, { "epoch": 0.9664186183780593, "grad_norm": 1.7762209177017212, "learning_rate": 3.765967082045036e-05, "loss": 7.3665, "step": 22922 }, { "epoch": 0.9664607795602589, "grad_norm": 3.629530429840088, "learning_rate": 3.765889003404229e-05, "loss": 7.4451, "step": 22923 }, { "epoch": 0.9665029407424585, "grad_norm": 1.4923264980316162, "learning_rate": 3.765810924763422e-05, "loss": 6.5285, "step": 22924 }, { "epoch": 0.9665451019246579, "grad_norm": 3.1232597827911377, "learning_rate": 3.765732846122615e-05, "loss": 7.0693, "step": 22925 }, { "epoch": 0.9665872631068575, "grad_norm": 1.6658551692962646, "learning_rate": 3.7656547674818074e-05, "loss": 7.0954, "step": 22926 }, { "epoch": 0.9666294242890571, "grad_norm": 1.549568772315979, "learning_rate": 3.765576688841001e-05, "loss": 6.8016, "step": 22927 }, { "epoch": 0.9666715854712566, "grad_norm": 2.6225459575653076, "learning_rate": 3.7654986102001936e-05, "loss": 8.0326, "step": 22928 }, { "epoch": 0.9667137466534561, "grad_norm": 1.271270751953125, "learning_rate": 3.765420531559387e-05, "loss": 6.519, "step": 22929 }, { "epoch": 0.9667559078356557, "grad_norm": 1.523208498954773, "learning_rate": 3.76534245291858e-05, "loss": 6.6351, "step": 22930 }, { "epoch": 0.9667980690178553, "grad_norm": 1.4485009908676147, "learning_rate": 3.7652643742777723e-05, "loss": 6.765, "step": 22931 }, { "epoch": 0.9668402302000548, "grad_norm": 3.1123478412628174, "learning_rate": 3.7651862956369655e-05, "loss": 6.7146, "step": 22932 }, { "epoch": 0.9668823913822544, "grad_norm": 2.6431000232696533, "learning_rate": 3.7651082169961586e-05, "loss": 7.4292, "step": 22933 }, { "epoch": 0.9669245525644539, "grad_norm": 1.6271892786026, "learning_rate": 3.765030138355352e-05, "loss": 7.0474, "step": 22934 }, { "epoch": 0.9669667137466534, "grad_norm": 1.7396200895309448, "learning_rate": 3.764952059714545e-05, "loss": 6.9978, "step": 22935 }, { "epoch": 0.967008874928853, "grad_norm": 2.8469526767730713, "learning_rate": 3.764873981073738e-05, "loss": 6.4327, "step": 22936 }, { "epoch": 0.9670510361110526, "grad_norm": 2.426330327987671, "learning_rate": 3.7647959024329304e-05, "loss": 7.2482, "step": 22937 }, { "epoch": 0.967093197293252, "grad_norm": 1.2954671382904053, "learning_rate": 3.7647178237921235e-05, "loss": 7.3096, "step": 22938 }, { "epoch": 0.9671353584754516, "grad_norm": 1.2035398483276367, "learning_rate": 3.764639745151317e-05, "loss": 6.6311, "step": 22939 }, { "epoch": 0.9671775196576512, "grad_norm": 1.2299572229385376, "learning_rate": 3.764561666510509e-05, "loss": 6.6651, "step": 22940 }, { "epoch": 0.9672196808398508, "grad_norm": 1.0768311023712158, "learning_rate": 3.764483587869703e-05, "loss": 6.7641, "step": 22941 }, { "epoch": 0.9672618420220503, "grad_norm": 0.9841904640197754, "learning_rate": 3.7644055092288954e-05, "loss": 6.7245, "step": 22942 }, { "epoch": 0.9673040032042498, "grad_norm": 1.0818088054656982, "learning_rate": 3.7643274305880885e-05, "loss": 6.6255, "step": 22943 }, { "epoch": 0.9673461643864494, "grad_norm": 0.9415982365608215, "learning_rate": 3.7642493519472816e-05, "loss": 6.5445, "step": 22944 }, { "epoch": 0.967388325568649, "grad_norm": 1.0413401126861572, "learning_rate": 3.764171273306474e-05, "loss": 6.4391, "step": 22945 }, { "epoch": 0.9674304867508485, "grad_norm": 1.7413625717163086, "learning_rate": 3.764093194665668e-05, "loss": 7.0667, "step": 22946 }, { "epoch": 0.967472647933048, "grad_norm": 2.7593424320220947, "learning_rate": 3.76401511602486e-05, "loss": 7.257, "step": 22947 }, { "epoch": 0.9675148091152476, "grad_norm": 1.0146589279174805, "learning_rate": 3.7639370373840534e-05, "loss": 6.8469, "step": 22948 }, { "epoch": 0.9675569702974471, "grad_norm": 1.3671690225601196, "learning_rate": 3.7638589587432466e-05, "loss": 6.6178, "step": 22949 }, { "epoch": 0.9675991314796467, "grad_norm": 1.4334239959716797, "learning_rate": 3.763780880102439e-05, "loss": 6.5921, "step": 22950 }, { "epoch": 0.9676412926618463, "grad_norm": 3.0038952827453613, "learning_rate": 3.763702801461632e-05, "loss": 7.9863, "step": 22951 }, { "epoch": 0.9676834538440457, "grad_norm": 1.5673331022262573, "learning_rate": 3.763624722820825e-05, "loss": 6.6053, "step": 22952 }, { "epoch": 0.9677256150262453, "grad_norm": 2.8228659629821777, "learning_rate": 3.7635466441800184e-05, "loss": 7.3436, "step": 22953 }, { "epoch": 0.9677677762084449, "grad_norm": 0.9331772923469543, "learning_rate": 3.763468565539211e-05, "loss": 6.8068, "step": 22954 }, { "epoch": 0.9678099373906445, "grad_norm": 2.084080934524536, "learning_rate": 3.7633904868984047e-05, "loss": 7.1742, "step": 22955 }, { "epoch": 0.9678520985728439, "grad_norm": 0.7717289924621582, "learning_rate": 3.763312408257597e-05, "loss": 6.6363, "step": 22956 }, { "epoch": 0.9678942597550435, "grad_norm": 1.6606146097183228, "learning_rate": 3.76323432961679e-05, "loss": 6.5566, "step": 22957 }, { "epoch": 0.9679364209372431, "grad_norm": 2.504664421081543, "learning_rate": 3.7631562509759834e-05, "loss": 7.0397, "step": 22958 }, { "epoch": 0.9679785821194427, "grad_norm": 2.024589776992798, "learning_rate": 3.763078172335176e-05, "loss": 7.4182, "step": 22959 }, { "epoch": 0.9680207433016422, "grad_norm": 2.93174409866333, "learning_rate": 3.7630000936943696e-05, "loss": 8.6956, "step": 22960 }, { "epoch": 0.9680629044838417, "grad_norm": 2.4007976055145264, "learning_rate": 3.762922015053562e-05, "loss": 6.5468, "step": 22961 }, { "epoch": 0.9681050656660413, "grad_norm": 1.5307375192642212, "learning_rate": 3.762843936412755e-05, "loss": 6.8938, "step": 22962 }, { "epoch": 0.9681472268482408, "grad_norm": 2.0481977462768555, "learning_rate": 3.762765857771948e-05, "loss": 6.4964, "step": 22963 }, { "epoch": 0.9681893880304404, "grad_norm": 1.749233365058899, "learning_rate": 3.762687779131141e-05, "loss": 6.9566, "step": 22964 }, { "epoch": 0.9682315492126399, "grad_norm": 1.7441691160202026, "learning_rate": 3.762609700490334e-05, "loss": 7.2326, "step": 22965 }, { "epoch": 0.9682737103948394, "grad_norm": 1.1165056228637695, "learning_rate": 3.762531621849527e-05, "loss": 6.7473, "step": 22966 }, { "epoch": 0.968315871577039, "grad_norm": 1.2240487337112427, "learning_rate": 3.76245354320872e-05, "loss": 6.4073, "step": 22967 }, { "epoch": 0.9683580327592386, "grad_norm": 2.1918137073516846, "learning_rate": 3.762375464567913e-05, "loss": 6.8354, "step": 22968 }, { "epoch": 0.9684001939414382, "grad_norm": 1.1601605415344238, "learning_rate": 3.762297385927106e-05, "loss": 6.6356, "step": 22969 }, { "epoch": 0.9684423551236376, "grad_norm": 3.6482656002044678, "learning_rate": 3.762219307286299e-05, "loss": 7.9609, "step": 22970 }, { "epoch": 0.9684845163058372, "grad_norm": 1.2207467555999756, "learning_rate": 3.762141228645492e-05, "loss": 6.3925, "step": 22971 }, { "epoch": 0.9685266774880368, "grad_norm": 1.298954725265503, "learning_rate": 3.762063150004685e-05, "loss": 6.8063, "step": 22972 }, { "epoch": 0.9685688386702364, "grad_norm": 1.6289453506469727, "learning_rate": 3.7619850713638775e-05, "loss": 6.8333, "step": 22973 }, { "epoch": 0.9686109998524358, "grad_norm": 1.1339813470840454, "learning_rate": 3.761906992723071e-05, "loss": 6.6957, "step": 22974 }, { "epoch": 0.9686531610346354, "grad_norm": 1.040144681930542, "learning_rate": 3.761828914082264e-05, "loss": 6.7865, "step": 22975 }, { "epoch": 0.968695322216835, "grad_norm": 1.9474002122879028, "learning_rate": 3.761750835441456e-05, "loss": 7.0651, "step": 22976 }, { "epoch": 0.9687374833990345, "grad_norm": 2.1171629428863525, "learning_rate": 3.76167275680065e-05, "loss": 7.2178, "step": 22977 }, { "epoch": 0.9687796445812341, "grad_norm": 1.6888251304626465, "learning_rate": 3.7615946781598425e-05, "loss": 6.7472, "step": 22978 }, { "epoch": 0.9688218057634336, "grad_norm": 2.4318056106567383, "learning_rate": 3.761516599519036e-05, "loss": 6.457, "step": 22979 }, { "epoch": 0.9688639669456331, "grad_norm": 1.422258973121643, "learning_rate": 3.761438520878229e-05, "loss": 7.0397, "step": 22980 }, { "epoch": 0.9689061281278327, "grad_norm": 1.3905205726623535, "learning_rate": 3.761360442237422e-05, "loss": 6.9826, "step": 22981 }, { "epoch": 0.9689482893100323, "grad_norm": 2.698352336883545, "learning_rate": 3.761282363596615e-05, "loss": 8.6927, "step": 22982 }, { "epoch": 0.9689904504922318, "grad_norm": 2.001504421234131, "learning_rate": 3.7612042849558074e-05, "loss": 6.4353, "step": 22983 }, { "epoch": 0.9690326116744313, "grad_norm": 3.127095937728882, "learning_rate": 3.7611262063150006e-05, "loss": 7.8408, "step": 22984 }, { "epoch": 0.9690747728566309, "grad_norm": 1.9044878482818604, "learning_rate": 3.761048127674194e-05, "loss": 6.5409, "step": 22985 }, { "epoch": 0.9691169340388305, "grad_norm": 2.3459415435791016, "learning_rate": 3.760970049033387e-05, "loss": 6.8043, "step": 22986 }, { "epoch": 0.9691590952210299, "grad_norm": 2.7253448963165283, "learning_rate": 3.760891970392579e-05, "loss": 8.4949, "step": 22987 }, { "epoch": 0.9692012564032295, "grad_norm": 1.1748994588851929, "learning_rate": 3.7608138917517724e-05, "loss": 6.6979, "step": 22988 }, { "epoch": 0.9692434175854291, "grad_norm": 2.66103458404541, "learning_rate": 3.7607358131109655e-05, "loss": 6.7639, "step": 22989 }, { "epoch": 0.9692855787676287, "grad_norm": 1.3416011333465576, "learning_rate": 3.7606577344701586e-05, "loss": 6.6935, "step": 22990 }, { "epoch": 0.9693277399498282, "grad_norm": 3.2363486289978027, "learning_rate": 3.760579655829352e-05, "loss": 6.3093, "step": 22991 }, { "epoch": 0.9693699011320277, "grad_norm": 2.310612916946411, "learning_rate": 3.760501577188544e-05, "loss": 7.0096, "step": 22992 }, { "epoch": 0.9694120623142273, "grad_norm": 1.5360078811645508, "learning_rate": 3.7604234985477373e-05, "loss": 6.9226, "step": 22993 }, { "epoch": 0.9694542234964268, "grad_norm": 1.3936424255371094, "learning_rate": 3.7603454199069305e-05, "loss": 7.0867, "step": 22994 }, { "epoch": 0.9694963846786264, "grad_norm": 1.083217978477478, "learning_rate": 3.760267341266123e-05, "loss": 6.9761, "step": 22995 }, { "epoch": 0.9695385458608259, "grad_norm": 2.9396796226501465, "learning_rate": 3.760189262625317e-05, "loss": 7.6397, "step": 22996 }, { "epoch": 0.9695807070430255, "grad_norm": 2.137558937072754, "learning_rate": 3.760111183984509e-05, "loss": 7.0117, "step": 22997 }, { "epoch": 0.969622868225225, "grad_norm": 1.5737031698226929, "learning_rate": 3.760033105343702e-05, "loss": 6.6234, "step": 22998 }, { "epoch": 0.9696650294074246, "grad_norm": 1.0920466184616089, "learning_rate": 3.7599550267028954e-05, "loss": 6.8003, "step": 22999 }, { "epoch": 0.9697071905896242, "grad_norm": 3.532829761505127, "learning_rate": 3.7598769480620885e-05, "loss": 7.7083, "step": 23000 }, { "epoch": 0.9697493517718236, "grad_norm": 3.5470755100250244, "learning_rate": 3.759798869421282e-05, "loss": 8.0412, "step": 23001 }, { "epoch": 0.9697915129540232, "grad_norm": 3.8556907176971436, "learning_rate": 3.759720790780474e-05, "loss": 8.1569, "step": 23002 }, { "epoch": 0.9698336741362228, "grad_norm": 1.5577061176300049, "learning_rate": 3.759642712139667e-05, "loss": 6.841, "step": 23003 }, { "epoch": 0.9698758353184224, "grad_norm": 1.385818362236023, "learning_rate": 3.7595646334988604e-05, "loss": 6.816, "step": 23004 }, { "epoch": 0.9699179965006218, "grad_norm": 1.5176070928573608, "learning_rate": 3.7594865548580535e-05, "loss": 7.189, "step": 23005 }, { "epoch": 0.9699601576828214, "grad_norm": 2.2979848384857178, "learning_rate": 3.759408476217246e-05, "loss": 8.0671, "step": 23006 }, { "epoch": 0.970002318865021, "grad_norm": 1.6820292472839355, "learning_rate": 3.759330397576439e-05, "loss": 7.0264, "step": 23007 }, { "epoch": 0.9700444800472205, "grad_norm": 2.1458492279052734, "learning_rate": 3.759252318935632e-05, "loss": 7.2003, "step": 23008 }, { "epoch": 0.9700866412294201, "grad_norm": 3.4328343868255615, "learning_rate": 3.7591742402948246e-05, "loss": 7.6027, "step": 23009 }, { "epoch": 0.9701288024116196, "grad_norm": 1.7779558897018433, "learning_rate": 3.7590961616540185e-05, "loss": 7.2157, "step": 23010 }, { "epoch": 0.9701709635938192, "grad_norm": 1.2227543592453003, "learning_rate": 3.759018083013211e-05, "loss": 7.0647, "step": 23011 }, { "epoch": 0.9702131247760187, "grad_norm": 1.5696176290512085, "learning_rate": 3.758940004372404e-05, "loss": 7.0595, "step": 23012 }, { "epoch": 0.9702552859582183, "grad_norm": 0.8927922248840332, "learning_rate": 3.758861925731597e-05, "loss": 6.9836, "step": 23013 }, { "epoch": 0.9702974471404178, "grad_norm": 1.5052143335342407, "learning_rate": 3.7587838470907896e-05, "loss": 6.5156, "step": 23014 }, { "epoch": 0.9703396083226173, "grad_norm": 1.626152515411377, "learning_rate": 3.7587057684499834e-05, "loss": 6.5352, "step": 23015 }, { "epoch": 0.9703817695048169, "grad_norm": 3.4870665073394775, "learning_rate": 3.758627689809176e-05, "loss": 7.3185, "step": 23016 }, { "epoch": 0.9704239306870165, "grad_norm": 1.159681797027588, "learning_rate": 3.758549611168369e-05, "loss": 7.041, "step": 23017 }, { "epoch": 0.9704660918692161, "grad_norm": 1.4738264083862305, "learning_rate": 3.758471532527562e-05, "loss": 6.5848, "step": 23018 }, { "epoch": 0.9705082530514155, "grad_norm": 1.8047153949737549, "learning_rate": 3.758393453886755e-05, "loss": 6.8488, "step": 23019 }, { "epoch": 0.9705504142336151, "grad_norm": 1.147979497909546, "learning_rate": 3.758315375245948e-05, "loss": 6.589, "step": 23020 }, { "epoch": 0.9705925754158147, "grad_norm": 3.346705198287964, "learning_rate": 3.758237296605141e-05, "loss": 6.4734, "step": 23021 }, { "epoch": 0.9706347365980142, "grad_norm": 2.532040596008301, "learning_rate": 3.758159217964334e-05, "loss": 7.1716, "step": 23022 }, { "epoch": 0.9706768977802137, "grad_norm": 3.78041934967041, "learning_rate": 3.7580811393235264e-05, "loss": 6.6389, "step": 23023 }, { "epoch": 0.9707190589624133, "grad_norm": 2.2764978408813477, "learning_rate": 3.75800306068272e-05, "loss": 6.9836, "step": 23024 }, { "epoch": 0.9707612201446129, "grad_norm": 2.185234546661377, "learning_rate": 3.7579249820419126e-05, "loss": 7.7662, "step": 23025 }, { "epoch": 0.9708033813268124, "grad_norm": 2.9359283447265625, "learning_rate": 3.757846903401106e-05, "loss": 7.752, "step": 23026 }, { "epoch": 0.970845542509012, "grad_norm": 1.955293893814087, "learning_rate": 3.757768824760299e-05, "loss": 7.1207, "step": 23027 }, { "epoch": 0.9708877036912115, "grad_norm": 1.520362377166748, "learning_rate": 3.757690746119491e-05, "loss": 6.6308, "step": 23028 }, { "epoch": 0.970929864873411, "grad_norm": 2.447005271911621, "learning_rate": 3.757612667478685e-05, "loss": 6.6949, "step": 23029 }, { "epoch": 0.9709720260556106, "grad_norm": 3.5087645053863525, "learning_rate": 3.7575345888378776e-05, "loss": 7.7396, "step": 23030 }, { "epoch": 0.9710141872378102, "grad_norm": 2.621946334838867, "learning_rate": 3.757456510197071e-05, "loss": 6.6389, "step": 23031 }, { "epoch": 0.9710563484200097, "grad_norm": 1.8668220043182373, "learning_rate": 3.757378431556264e-05, "loss": 6.5625, "step": 23032 }, { "epoch": 0.9710985096022092, "grad_norm": 2.4448602199554443, "learning_rate": 3.757300352915456e-05, "loss": 7.4034, "step": 23033 }, { "epoch": 0.9711406707844088, "grad_norm": 2.9157512187957764, "learning_rate": 3.7572222742746494e-05, "loss": 7.6056, "step": 23034 }, { "epoch": 0.9711828319666084, "grad_norm": 3.2576169967651367, "learning_rate": 3.7571441956338425e-05, "loss": 7.3337, "step": 23035 }, { "epoch": 0.971224993148808, "grad_norm": 3.5575172901153564, "learning_rate": 3.7570661169930357e-05, "loss": 7.4756, "step": 23036 }, { "epoch": 0.9712671543310074, "grad_norm": 1.9762729406356812, "learning_rate": 3.756988038352229e-05, "loss": 6.582, "step": 23037 }, { "epoch": 0.971309315513207, "grad_norm": 2.2804501056671143, "learning_rate": 3.756909959711421e-05, "loss": 6.7336, "step": 23038 }, { "epoch": 0.9713514766954066, "grad_norm": 1.8473645448684692, "learning_rate": 3.7568318810706144e-05, "loss": 6.5263, "step": 23039 }, { "epoch": 0.9713936378776061, "grad_norm": 1.4205392599105835, "learning_rate": 3.7567538024298075e-05, "loss": 6.5287, "step": 23040 }, { "epoch": 0.9714357990598056, "grad_norm": 1.3152505159378052, "learning_rate": 3.7566757237890006e-05, "loss": 6.8766, "step": 23041 }, { "epoch": 0.9714779602420052, "grad_norm": 3.2810332775115967, "learning_rate": 3.756597645148193e-05, "loss": 7.4968, "step": 23042 }, { "epoch": 0.9715201214242047, "grad_norm": 2.4136879444122314, "learning_rate": 3.756519566507387e-05, "loss": 6.5193, "step": 23043 }, { "epoch": 0.9715622826064043, "grad_norm": 2.623493194580078, "learning_rate": 3.756441487866579e-05, "loss": 7.1575, "step": 23044 }, { "epoch": 0.9716044437886038, "grad_norm": 2.0989267826080322, "learning_rate": 3.7563634092257724e-05, "loss": 7.3326, "step": 23045 }, { "epoch": 0.9716466049708034, "grad_norm": 3.138190746307373, "learning_rate": 3.7562853305849656e-05, "loss": 8.882, "step": 23046 }, { "epoch": 0.9716887661530029, "grad_norm": 1.7429141998291016, "learning_rate": 3.756207251944158e-05, "loss": 6.6685, "step": 23047 }, { "epoch": 0.9717309273352025, "grad_norm": 1.2741373777389526, "learning_rate": 3.756129173303352e-05, "loss": 6.5752, "step": 23048 }, { "epoch": 0.9717730885174021, "grad_norm": 1.5363516807556152, "learning_rate": 3.756051094662544e-05, "loss": 7.2066, "step": 23049 }, { "epoch": 0.9718152496996015, "grad_norm": 2.9339070320129395, "learning_rate": 3.7559730160217374e-05, "loss": 6.8714, "step": 23050 }, { "epoch": 0.9718574108818011, "grad_norm": 1.7943768501281738, "learning_rate": 3.7558949373809305e-05, "loss": 6.6426, "step": 23051 }, { "epoch": 0.9718995720640007, "grad_norm": 4.211813926696777, "learning_rate": 3.755816858740123e-05, "loss": 8.0203, "step": 23052 }, { "epoch": 0.9719417332462003, "grad_norm": 2.274561882019043, "learning_rate": 3.755738780099316e-05, "loss": 6.9569, "step": 23053 }, { "epoch": 0.9719838944283997, "grad_norm": 1.2322940826416016, "learning_rate": 3.755660701458509e-05, "loss": 6.6089, "step": 23054 }, { "epoch": 0.9720260556105993, "grad_norm": 2.395925998687744, "learning_rate": 3.7555826228177023e-05, "loss": 7.0389, "step": 23055 }, { "epoch": 0.9720682167927989, "grad_norm": 3.3083086013793945, "learning_rate": 3.755504544176895e-05, "loss": 8.5528, "step": 23056 }, { "epoch": 0.9721103779749984, "grad_norm": 2.0198726654052734, "learning_rate": 3.755426465536088e-05, "loss": 6.7553, "step": 23057 }, { "epoch": 0.972152539157198, "grad_norm": 3.0547935962677, "learning_rate": 3.755348386895281e-05, "loss": 7.748, "step": 23058 }, { "epoch": 0.9721947003393975, "grad_norm": 2.9893577098846436, "learning_rate": 3.755270308254474e-05, "loss": 7.064, "step": 23059 }, { "epoch": 0.972236861521597, "grad_norm": 1.0396751165390015, "learning_rate": 3.755192229613667e-05, "loss": 6.8725, "step": 23060 }, { "epoch": 0.9722790227037966, "grad_norm": 1.0123653411865234, "learning_rate": 3.75511415097286e-05, "loss": 6.3777, "step": 23061 }, { "epoch": 0.9723211838859962, "grad_norm": 2.931110382080078, "learning_rate": 3.7550360723320535e-05, "loss": 7.326, "step": 23062 }, { "epoch": 0.9723633450681957, "grad_norm": 3.743975877761841, "learning_rate": 3.754957993691246e-05, "loss": 7.7915, "step": 23063 }, { "epoch": 0.9724055062503952, "grad_norm": 2.114159345626831, "learning_rate": 3.754879915050439e-05, "loss": 6.7502, "step": 23064 }, { "epoch": 0.9724476674325948, "grad_norm": 1.7322657108306885, "learning_rate": 3.754801836409632e-05, "loss": 6.6352, "step": 23065 }, { "epoch": 0.9724898286147944, "grad_norm": 3.3036460876464844, "learning_rate": 3.754723757768825e-05, "loss": 8.1761, "step": 23066 }, { "epoch": 0.972531989796994, "grad_norm": 1.743237018585205, "learning_rate": 3.754645679128018e-05, "loss": 6.3728, "step": 23067 }, { "epoch": 0.9725741509791934, "grad_norm": 1.7827064990997314, "learning_rate": 3.754567600487211e-05, "loss": 6.6466, "step": 23068 }, { "epoch": 0.972616312161393, "grad_norm": 1.891318917274475, "learning_rate": 3.754489521846404e-05, "loss": 6.7675, "step": 23069 }, { "epoch": 0.9726584733435926, "grad_norm": 1.4141063690185547, "learning_rate": 3.754411443205597e-05, "loss": 6.6904, "step": 23070 }, { "epoch": 0.9727006345257921, "grad_norm": 1.4427597522735596, "learning_rate": 3.7543333645647896e-05, "loss": 6.9185, "step": 23071 }, { "epoch": 0.9727427957079916, "grad_norm": 2.4470810890197754, "learning_rate": 3.754255285923983e-05, "loss": 7.04, "step": 23072 }, { "epoch": 0.9727849568901912, "grad_norm": 1.4875226020812988, "learning_rate": 3.754177207283176e-05, "loss": 6.4721, "step": 23073 }, { "epoch": 0.9728271180723908, "grad_norm": 1.1221461296081543, "learning_rate": 3.754099128642369e-05, "loss": 6.6434, "step": 23074 }, { "epoch": 0.9728692792545903, "grad_norm": 2.047606945037842, "learning_rate": 3.7540210500015615e-05, "loss": 6.4494, "step": 23075 }, { "epoch": 0.9729114404367899, "grad_norm": 1.8952916860580444, "learning_rate": 3.7539429713607546e-05, "loss": 6.5518, "step": 23076 }, { "epoch": 0.9729536016189894, "grad_norm": 1.440730333328247, "learning_rate": 3.753864892719948e-05, "loss": 6.991, "step": 23077 }, { "epoch": 0.9729957628011889, "grad_norm": 1.4938485622406006, "learning_rate": 3.75378681407914e-05, "loss": 6.8202, "step": 23078 }, { "epoch": 0.9730379239833885, "grad_norm": 2.231172800064087, "learning_rate": 3.753708735438334e-05, "loss": 7.2917, "step": 23079 }, { "epoch": 0.9730800851655881, "grad_norm": 3.8152568340301514, "learning_rate": 3.7536306567975264e-05, "loss": 7.3421, "step": 23080 }, { "epoch": 0.9731222463477875, "grad_norm": 3.5914511680603027, "learning_rate": 3.75355257815672e-05, "loss": 7.847, "step": 23081 }, { "epoch": 0.9731644075299871, "grad_norm": 1.6291080713272095, "learning_rate": 3.753474499515913e-05, "loss": 6.8155, "step": 23082 }, { "epoch": 0.9732065687121867, "grad_norm": 1.2310686111450195, "learning_rate": 3.753396420875105e-05, "loss": 6.8208, "step": 23083 }, { "epoch": 0.9732487298943863, "grad_norm": 2.2922592163085938, "learning_rate": 3.753318342234299e-05, "loss": 7.5902, "step": 23084 }, { "epoch": 0.9732908910765858, "grad_norm": 1.2472209930419922, "learning_rate": 3.7532402635934914e-05, "loss": 7.0768, "step": 23085 }, { "epoch": 0.9733330522587853, "grad_norm": 1.029679298400879, "learning_rate": 3.7531621849526845e-05, "loss": 6.7257, "step": 23086 }, { "epoch": 0.9733752134409849, "grad_norm": 2.4238767623901367, "learning_rate": 3.7530841063118776e-05, "loss": 6.6051, "step": 23087 }, { "epoch": 0.9734173746231845, "grad_norm": 1.0115796327590942, "learning_rate": 3.753006027671071e-05, "loss": 6.5142, "step": 23088 }, { "epoch": 0.973459535805384, "grad_norm": 3.9786548614501953, "learning_rate": 3.752927949030263e-05, "loss": 7.9671, "step": 23089 }, { "epoch": 0.9735016969875835, "grad_norm": 1.632228136062622, "learning_rate": 3.752849870389456e-05, "loss": 6.4994, "step": 23090 }, { "epoch": 0.9735438581697831, "grad_norm": 1.367977499961853, "learning_rate": 3.7527717917486495e-05, "loss": 6.868, "step": 23091 }, { "epoch": 0.9735860193519826, "grad_norm": 1.0893250703811646, "learning_rate": 3.752693713107842e-05, "loss": 6.5952, "step": 23092 }, { "epoch": 0.9736281805341822, "grad_norm": 1.7460676431655884, "learning_rate": 3.752615634467036e-05, "loss": 6.3903, "step": 23093 }, { "epoch": 0.9736703417163818, "grad_norm": 2.6730759143829346, "learning_rate": 3.752537555826228e-05, "loss": 6.8086, "step": 23094 }, { "epoch": 0.9737125028985812, "grad_norm": 0.9467185735702515, "learning_rate": 3.752459477185421e-05, "loss": 6.8903, "step": 23095 }, { "epoch": 0.9737546640807808, "grad_norm": 1.198876976966858, "learning_rate": 3.7523813985446144e-05, "loss": 6.5829, "step": 23096 }, { "epoch": 0.9737968252629804, "grad_norm": 0.9906158447265625, "learning_rate": 3.752303319903807e-05, "loss": 6.6107, "step": 23097 }, { "epoch": 0.97383898644518, "grad_norm": 0.9917349219322205, "learning_rate": 3.752225241263001e-05, "loss": 6.526, "step": 23098 }, { "epoch": 0.9738811476273794, "grad_norm": 2.8604588508605957, "learning_rate": 3.752147162622193e-05, "loss": 6.422, "step": 23099 }, { "epoch": 0.973923308809579, "grad_norm": 1.6867164373397827, "learning_rate": 3.752069083981386e-05, "loss": 6.9448, "step": 23100 }, { "epoch": 0.9739654699917786, "grad_norm": 1.5892188549041748, "learning_rate": 3.7519910053405794e-05, "loss": 6.4892, "step": 23101 }, { "epoch": 0.9740076311739782, "grad_norm": 2.3970093727111816, "learning_rate": 3.751912926699772e-05, "loss": 7.3864, "step": 23102 }, { "epoch": 0.9740497923561776, "grad_norm": 0.9113509654998779, "learning_rate": 3.751834848058965e-05, "loss": 6.6891, "step": 23103 }, { "epoch": 0.9740919535383772, "grad_norm": 2.999950408935547, "learning_rate": 3.751756769418158e-05, "loss": 7.6901, "step": 23104 }, { "epoch": 0.9741341147205768, "grad_norm": 1.6416641473770142, "learning_rate": 3.751678690777351e-05, "loss": 6.5053, "step": 23105 }, { "epoch": 0.9741762759027763, "grad_norm": 5.138742923736572, "learning_rate": 3.751600612136544e-05, "loss": 7.3557, "step": 23106 }, { "epoch": 0.9742184370849759, "grad_norm": 1.5834197998046875, "learning_rate": 3.7515225334957374e-05, "loss": 7.1494, "step": 23107 }, { "epoch": 0.9742605982671754, "grad_norm": 3.447129011154175, "learning_rate": 3.75144445485493e-05, "loss": 7.769, "step": 23108 }, { "epoch": 0.9743027594493749, "grad_norm": 1.3419378995895386, "learning_rate": 3.751366376214123e-05, "loss": 6.32, "step": 23109 }, { "epoch": 0.9743449206315745, "grad_norm": 1.1743228435516357, "learning_rate": 3.751288297573316e-05, "loss": 6.5585, "step": 23110 }, { "epoch": 0.9743870818137741, "grad_norm": 1.2316118478775024, "learning_rate": 3.7512102189325086e-05, "loss": 6.3609, "step": 23111 }, { "epoch": 0.9744292429959736, "grad_norm": 1.6509655714035034, "learning_rate": 3.7511321402917024e-05, "loss": 7.0043, "step": 23112 }, { "epoch": 0.9744714041781731, "grad_norm": 1.2533224821090698, "learning_rate": 3.751054061650895e-05, "loss": 6.597, "step": 23113 }, { "epoch": 0.9745135653603727, "grad_norm": 1.4394627809524536, "learning_rate": 3.750975983010088e-05, "loss": 6.3079, "step": 23114 }, { "epoch": 0.9745557265425723, "grad_norm": 1.1213874816894531, "learning_rate": 3.750897904369281e-05, "loss": 6.5949, "step": 23115 }, { "epoch": 0.9745978877247718, "grad_norm": 1.4707705974578857, "learning_rate": 3.7508198257284735e-05, "loss": 6.6306, "step": 23116 }, { "epoch": 0.9746400489069713, "grad_norm": 0.9268597960472107, "learning_rate": 3.7507417470876673e-05, "loss": 6.4092, "step": 23117 }, { "epoch": 0.9746822100891709, "grad_norm": 1.7177313566207886, "learning_rate": 3.75066366844686e-05, "loss": 7.2498, "step": 23118 }, { "epoch": 0.9747243712713705, "grad_norm": 0.9361318349838257, "learning_rate": 3.750585589806053e-05, "loss": 6.5842, "step": 23119 }, { "epoch": 0.97476653245357, "grad_norm": 1.2075433731079102, "learning_rate": 3.750507511165246e-05, "loss": 6.879, "step": 23120 }, { "epoch": 0.9748086936357695, "grad_norm": 3.0858631134033203, "learning_rate": 3.7504294325244385e-05, "loss": 7.8695, "step": 23121 }, { "epoch": 0.9748508548179691, "grad_norm": 2.8754355907440186, "learning_rate": 3.7503513538836316e-05, "loss": 7.6664, "step": 23122 }, { "epoch": 0.9748930160001686, "grad_norm": 3.8718693256378174, "learning_rate": 3.750273275242825e-05, "loss": 8.1667, "step": 23123 }, { "epoch": 0.9749351771823682, "grad_norm": 4.41261100769043, "learning_rate": 3.750195196602018e-05, "loss": 8.1299, "step": 23124 }, { "epoch": 0.9749773383645678, "grad_norm": 3.5275776386260986, "learning_rate": 3.75011711796121e-05, "loss": 7.183, "step": 23125 }, { "epoch": 0.9750194995467673, "grad_norm": 2.8757948875427246, "learning_rate": 3.750039039320404e-05, "loss": 7.7135, "step": 23126 }, { "epoch": 0.9750616607289668, "grad_norm": 3.0260446071624756, "learning_rate": 3.7499609606795966e-05, "loss": 7.409, "step": 23127 }, { "epoch": 0.9751038219111664, "grad_norm": 0.9975391030311584, "learning_rate": 3.74988288203879e-05, "loss": 6.6467, "step": 23128 }, { "epoch": 0.975145983093366, "grad_norm": 1.6525521278381348, "learning_rate": 3.749804803397983e-05, "loss": 8.0889, "step": 23129 }, { "epoch": 0.9751881442755654, "grad_norm": 1.2840087413787842, "learning_rate": 3.749726724757175e-05, "loss": 7.0782, "step": 23130 }, { "epoch": 0.975230305457765, "grad_norm": 1.1367735862731934, "learning_rate": 3.749648646116369e-05, "loss": 6.9527, "step": 23131 }, { "epoch": 0.9752724666399646, "grad_norm": 1.06613028049469, "learning_rate": 3.7495705674755615e-05, "loss": 6.5921, "step": 23132 }, { "epoch": 0.9753146278221642, "grad_norm": 1.0043461322784424, "learning_rate": 3.7494924888347547e-05, "loss": 6.5683, "step": 23133 }, { "epoch": 0.9753567890043637, "grad_norm": 3.507849931716919, "learning_rate": 3.749414410193948e-05, "loss": 8.0503, "step": 23134 }, { "epoch": 0.9753989501865632, "grad_norm": 2.8753035068511963, "learning_rate": 3.74933633155314e-05, "loss": 6.8067, "step": 23135 }, { "epoch": 0.9754411113687628, "grad_norm": 2.3042919635772705, "learning_rate": 3.7492582529123334e-05, "loss": 7.1359, "step": 23136 }, { "epoch": 0.9754832725509623, "grad_norm": 3.0442497730255127, "learning_rate": 3.7491801742715265e-05, "loss": 8.4558, "step": 23137 }, { "epoch": 0.9755254337331619, "grad_norm": 3.088562488555908, "learning_rate": 3.7491020956307196e-05, "loss": 7.8898, "step": 23138 }, { "epoch": 0.9755675949153614, "grad_norm": 2.1835858821868896, "learning_rate": 3.749024016989913e-05, "loss": 7.0725, "step": 23139 }, { "epoch": 0.975609756097561, "grad_norm": 2.101773977279663, "learning_rate": 3.748945938349105e-05, "loss": 6.6611, "step": 23140 }, { "epoch": 0.9756519172797605, "grad_norm": 1.2763205766677856, "learning_rate": 3.748867859708298e-05, "loss": 7.4117, "step": 23141 }, { "epoch": 0.9756940784619601, "grad_norm": 2.8460826873779297, "learning_rate": 3.7487897810674914e-05, "loss": 6.4263, "step": 23142 }, { "epoch": 0.9757362396441597, "grad_norm": 3.6125736236572266, "learning_rate": 3.7487117024266846e-05, "loss": 7.8138, "step": 23143 }, { "epoch": 0.9757784008263591, "grad_norm": 2.3723413944244385, "learning_rate": 3.748633623785877e-05, "loss": 7.555, "step": 23144 }, { "epoch": 0.9758205620085587, "grad_norm": 1.1926121711730957, "learning_rate": 3.74855554514507e-05, "loss": 7.0181, "step": 23145 }, { "epoch": 0.9758627231907583, "grad_norm": 1.3213214874267578, "learning_rate": 3.748477466504263e-05, "loss": 6.6298, "step": 23146 }, { "epoch": 0.9759048843729579, "grad_norm": 1.6422983407974243, "learning_rate": 3.748399387863456e-05, "loss": 7.0178, "step": 23147 }, { "epoch": 0.9759470455551573, "grad_norm": 1.383230447769165, "learning_rate": 3.7483213092226495e-05, "loss": 6.9957, "step": 23148 }, { "epoch": 0.9759892067373569, "grad_norm": 4.641666412353516, "learning_rate": 3.748243230581842e-05, "loss": 7.6681, "step": 23149 }, { "epoch": 0.9760313679195565, "grad_norm": 0.7933807373046875, "learning_rate": 3.748165151941036e-05, "loss": 6.5498, "step": 23150 }, { "epoch": 0.976073529101756, "grad_norm": 1.003298282623291, "learning_rate": 3.748087073300228e-05, "loss": 6.8696, "step": 23151 }, { "epoch": 0.9761156902839556, "grad_norm": 1.114989161491394, "learning_rate": 3.748008994659421e-05, "loss": 6.9995, "step": 23152 }, { "epoch": 0.9761578514661551, "grad_norm": 2.4238853454589844, "learning_rate": 3.7479309160186145e-05, "loss": 6.5866, "step": 23153 }, { "epoch": 0.9762000126483547, "grad_norm": 1.0430574417114258, "learning_rate": 3.747852837377807e-05, "loss": 6.3999, "step": 23154 }, { "epoch": 0.9762421738305542, "grad_norm": 1.2025707960128784, "learning_rate": 3.747774758737e-05, "loss": 6.622, "step": 23155 }, { "epoch": 0.9762843350127538, "grad_norm": 1.2144321203231812, "learning_rate": 3.747696680096193e-05, "loss": 6.9611, "step": 23156 }, { "epoch": 0.9763264961949533, "grad_norm": 1.134216547012329, "learning_rate": 3.747618601455386e-05, "loss": 7.0737, "step": 23157 }, { "epoch": 0.9763686573771528, "grad_norm": 1.3791043758392334, "learning_rate": 3.747540522814579e-05, "loss": 6.869, "step": 23158 }, { "epoch": 0.9764108185593524, "grad_norm": 2.3615238666534424, "learning_rate": 3.747462444173772e-05, "loss": 6.732, "step": 23159 }, { "epoch": 0.976452979741552, "grad_norm": 1.6648242473602295, "learning_rate": 3.747384365532965e-05, "loss": 6.8489, "step": 23160 }, { "epoch": 0.9764951409237514, "grad_norm": 1.909266471862793, "learning_rate": 3.747306286892158e-05, "loss": 7.2642, "step": 23161 }, { "epoch": 0.976537302105951, "grad_norm": 2.265110969543457, "learning_rate": 3.747228208251351e-05, "loss": 7.3047, "step": 23162 }, { "epoch": 0.9765794632881506, "grad_norm": 1.4169795513153076, "learning_rate": 3.747150129610544e-05, "loss": 6.4375, "step": 23163 }, { "epoch": 0.9766216244703502, "grad_norm": 0.9378324747085571, "learning_rate": 3.747072050969737e-05, "loss": 6.8913, "step": 23164 }, { "epoch": 0.9766637856525497, "grad_norm": 3.3888370990753174, "learning_rate": 3.74699397232893e-05, "loss": 7.6861, "step": 23165 }, { "epoch": 0.9767059468347492, "grad_norm": 2.6696841716766357, "learning_rate": 3.7469158936881224e-05, "loss": 6.6711, "step": 23166 }, { "epoch": 0.9767481080169488, "grad_norm": 2.470571994781494, "learning_rate": 3.746837815047316e-05, "loss": 7.6344, "step": 23167 }, { "epoch": 0.9767902691991484, "grad_norm": 1.040015697479248, "learning_rate": 3.7467597364065086e-05, "loss": 6.8484, "step": 23168 }, { "epoch": 0.9768324303813479, "grad_norm": 0.7666773200035095, "learning_rate": 3.746681657765702e-05, "loss": 6.5287, "step": 23169 }, { "epoch": 0.9768745915635474, "grad_norm": 1.1457226276397705, "learning_rate": 3.746603579124895e-05, "loss": 6.6411, "step": 23170 }, { "epoch": 0.976916752745747, "grad_norm": 3.430572986602783, "learning_rate": 3.746525500484088e-05, "loss": 7.8206, "step": 23171 }, { "epoch": 0.9769589139279465, "grad_norm": 1.0936236381530762, "learning_rate": 3.7464474218432805e-05, "loss": 6.7897, "step": 23172 }, { "epoch": 0.9770010751101461, "grad_norm": 1.2514727115631104, "learning_rate": 3.7463693432024736e-05, "loss": 7.3583, "step": 23173 }, { "epoch": 0.9770432362923457, "grad_norm": 1.0899873971939087, "learning_rate": 3.746291264561667e-05, "loss": 6.5554, "step": 23174 }, { "epoch": 0.9770853974745451, "grad_norm": 1.1251662969589233, "learning_rate": 3.74621318592086e-05, "loss": 7.1683, "step": 23175 }, { "epoch": 0.9771275586567447, "grad_norm": 1.1465466022491455, "learning_rate": 3.746135107280053e-05, "loss": 6.5429, "step": 23176 }, { "epoch": 0.9771697198389443, "grad_norm": 0.8572984933853149, "learning_rate": 3.7460570286392454e-05, "loss": 6.5283, "step": 23177 }, { "epoch": 0.9772118810211439, "grad_norm": 1.0037963390350342, "learning_rate": 3.7459789499984385e-05, "loss": 6.6488, "step": 23178 }, { "epoch": 0.9772540422033433, "grad_norm": 3.0616772174835205, "learning_rate": 3.745900871357632e-05, "loss": 7.7277, "step": 23179 }, { "epoch": 0.9772962033855429, "grad_norm": 1.7850291728973389, "learning_rate": 3.745822792716824e-05, "loss": 7.4826, "step": 23180 }, { "epoch": 0.9773383645677425, "grad_norm": 2.022472858428955, "learning_rate": 3.745744714076018e-05, "loss": 7.5454, "step": 23181 }, { "epoch": 0.977380525749942, "grad_norm": 1.7667768001556396, "learning_rate": 3.7456666354352104e-05, "loss": 7.072, "step": 23182 }, { "epoch": 0.9774226869321416, "grad_norm": 2.021094799041748, "learning_rate": 3.7455885567944035e-05, "loss": 7.4237, "step": 23183 }, { "epoch": 0.9774648481143411, "grad_norm": 3.006678342819214, "learning_rate": 3.7455104781535966e-05, "loss": 6.8271, "step": 23184 }, { "epoch": 0.9775070092965407, "grad_norm": 2.3332860469818115, "learning_rate": 3.745432399512789e-05, "loss": 6.6345, "step": 23185 }, { "epoch": 0.9775491704787402, "grad_norm": 1.4359666109085083, "learning_rate": 3.745354320871983e-05, "loss": 6.5072, "step": 23186 }, { "epoch": 0.9775913316609398, "grad_norm": 1.0743168592453003, "learning_rate": 3.745276242231175e-05, "loss": 6.6704, "step": 23187 }, { "epoch": 0.9776334928431393, "grad_norm": 1.343580722808838, "learning_rate": 3.7451981635903684e-05, "loss": 6.9015, "step": 23188 }, { "epoch": 0.9776756540253388, "grad_norm": 2.0478718280792236, "learning_rate": 3.7451200849495616e-05, "loss": 6.7519, "step": 23189 }, { "epoch": 0.9777178152075384, "grad_norm": 3.607654094696045, "learning_rate": 3.745042006308754e-05, "loss": 6.5981, "step": 23190 }, { "epoch": 0.977759976389738, "grad_norm": 1.489771842956543, "learning_rate": 3.744963927667947e-05, "loss": 6.6302, "step": 23191 }, { "epoch": 0.9778021375719376, "grad_norm": 1.6142082214355469, "learning_rate": 3.74488584902714e-05, "loss": 6.4459, "step": 23192 }, { "epoch": 0.977844298754137, "grad_norm": 3.346853017807007, "learning_rate": 3.7448077703863334e-05, "loss": 7.7635, "step": 23193 }, { "epoch": 0.9778864599363366, "grad_norm": 1.3300522565841675, "learning_rate": 3.744729691745526e-05, "loss": 6.9886, "step": 23194 }, { "epoch": 0.9779286211185362, "grad_norm": 1.8162174224853516, "learning_rate": 3.7446516131047197e-05, "loss": 6.9096, "step": 23195 }, { "epoch": 0.9779707823007358, "grad_norm": 4.377413749694824, "learning_rate": 3.744573534463912e-05, "loss": 8.0097, "step": 23196 }, { "epoch": 0.9780129434829352, "grad_norm": 2.619381904602051, "learning_rate": 3.744495455823105e-05, "loss": 6.5315, "step": 23197 }, { "epoch": 0.9780551046651348, "grad_norm": 3.4914462566375732, "learning_rate": 3.7444173771822984e-05, "loss": 7.8693, "step": 23198 }, { "epoch": 0.9780972658473344, "grad_norm": 1.8622889518737793, "learning_rate": 3.744339298541491e-05, "loss": 6.5342, "step": 23199 }, { "epoch": 0.9781394270295339, "grad_norm": 3.3168833255767822, "learning_rate": 3.7442612199006846e-05, "loss": 7.5598, "step": 23200 }, { "epoch": 0.9781815882117335, "grad_norm": 3.189598798751831, "learning_rate": 3.744183141259877e-05, "loss": 7.5819, "step": 23201 }, { "epoch": 0.978223749393933, "grad_norm": 1.8287551403045654, "learning_rate": 3.74410506261907e-05, "loss": 6.6209, "step": 23202 }, { "epoch": 0.9782659105761325, "grad_norm": 1.6048519611358643, "learning_rate": 3.744026983978263e-05, "loss": 6.3382, "step": 23203 }, { "epoch": 0.9783080717583321, "grad_norm": 3.8787083625793457, "learning_rate": 3.743948905337456e-05, "loss": 7.5224, "step": 23204 }, { "epoch": 0.9783502329405317, "grad_norm": 3.5569164752960205, "learning_rate": 3.743870826696649e-05, "loss": 7.7185, "step": 23205 }, { "epoch": 0.9783923941227312, "grad_norm": 1.9189270734786987, "learning_rate": 3.743792748055842e-05, "loss": 6.709, "step": 23206 }, { "epoch": 0.9784345553049307, "grad_norm": 0.8724825382232666, "learning_rate": 3.743714669415035e-05, "loss": 6.5438, "step": 23207 }, { "epoch": 0.9784767164871303, "grad_norm": 1.5929332971572876, "learning_rate": 3.743636590774228e-05, "loss": 6.433, "step": 23208 }, { "epoch": 0.9785188776693299, "grad_norm": 2.8231747150421143, "learning_rate": 3.743558512133421e-05, "loss": 7.4439, "step": 23209 }, { "epoch": 0.9785610388515295, "grad_norm": 4.2170329093933105, "learning_rate": 3.743480433492614e-05, "loss": 7.0621, "step": 23210 }, { "epoch": 0.9786032000337289, "grad_norm": 2.1447651386260986, "learning_rate": 3.743402354851807e-05, "loss": 7.7896, "step": 23211 }, { "epoch": 0.9786453612159285, "grad_norm": 2.8293440341949463, "learning_rate": 3.743324276211e-05, "loss": 7.9161, "step": 23212 }, { "epoch": 0.9786875223981281, "grad_norm": 3.508504867553711, "learning_rate": 3.7432461975701925e-05, "loss": 8.0757, "step": 23213 }, { "epoch": 0.9787296835803276, "grad_norm": 1.3084840774536133, "learning_rate": 3.743168118929386e-05, "loss": 6.7979, "step": 23214 }, { "epoch": 0.9787718447625271, "grad_norm": 1.2549680471420288, "learning_rate": 3.743090040288579e-05, "loss": 6.8864, "step": 23215 }, { "epoch": 0.9788140059447267, "grad_norm": 2.0454022884368896, "learning_rate": 3.743011961647771e-05, "loss": 7.5088, "step": 23216 }, { "epoch": 0.9788561671269262, "grad_norm": 1.1402984857559204, "learning_rate": 3.742933883006965e-05, "loss": 6.5348, "step": 23217 }, { "epoch": 0.9788983283091258, "grad_norm": 2.548781156539917, "learning_rate": 3.7428558043661575e-05, "loss": 7.3218, "step": 23218 }, { "epoch": 0.9789404894913253, "grad_norm": 2.62015700340271, "learning_rate": 3.742777725725351e-05, "loss": 7.0643, "step": 23219 }, { "epoch": 0.9789826506735249, "grad_norm": 2.254009246826172, "learning_rate": 3.742699647084544e-05, "loss": 6.8469, "step": 23220 }, { "epoch": 0.9790248118557244, "grad_norm": 2.537044048309326, "learning_rate": 3.742621568443737e-05, "loss": 6.5558, "step": 23221 }, { "epoch": 0.979066973037924, "grad_norm": 1.7275989055633545, "learning_rate": 3.74254348980293e-05, "loss": 7.0096, "step": 23222 }, { "epoch": 0.9791091342201236, "grad_norm": 1.6259292364120483, "learning_rate": 3.7424654111621224e-05, "loss": 7.1495, "step": 23223 }, { "epoch": 0.979151295402323, "grad_norm": 2.1208434104919434, "learning_rate": 3.7423873325213156e-05, "loss": 7.4228, "step": 23224 }, { "epoch": 0.9791934565845226, "grad_norm": 2.342869520187378, "learning_rate": 3.742309253880509e-05, "loss": 6.6811, "step": 23225 }, { "epoch": 0.9792356177667222, "grad_norm": 2.4655044078826904, "learning_rate": 3.742231175239702e-05, "loss": 8.0997, "step": 23226 }, { "epoch": 0.9792777789489218, "grad_norm": 1.201320767402649, "learning_rate": 3.742153096598894e-05, "loss": 6.5735, "step": 23227 }, { "epoch": 0.9793199401311212, "grad_norm": 1.9360777139663696, "learning_rate": 3.7420750179580874e-05, "loss": 7.0834, "step": 23228 }, { "epoch": 0.9793621013133208, "grad_norm": 2.448863983154297, "learning_rate": 3.7419969393172805e-05, "loss": 7.5116, "step": 23229 }, { "epoch": 0.9794042624955204, "grad_norm": 1.8463170528411865, "learning_rate": 3.7419188606764736e-05, "loss": 6.6132, "step": 23230 }, { "epoch": 0.97944642367772, "grad_norm": 1.331239938735962, "learning_rate": 3.741840782035667e-05, "loss": 6.7102, "step": 23231 }, { "epoch": 0.9794885848599195, "grad_norm": 1.129465937614441, "learning_rate": 3.741762703394859e-05, "loss": 6.4971, "step": 23232 }, { "epoch": 0.979530746042119, "grad_norm": 1.4381300210952759, "learning_rate": 3.7416846247540523e-05, "loss": 6.9784, "step": 23233 }, { "epoch": 0.9795729072243186, "grad_norm": 1.5918408632278442, "learning_rate": 3.7416065461132455e-05, "loss": 6.8385, "step": 23234 }, { "epoch": 0.9796150684065181, "grad_norm": 1.7489699125289917, "learning_rate": 3.741528467472438e-05, "loss": 7.1424, "step": 23235 }, { "epoch": 0.9796572295887177, "grad_norm": 3.3608927726745605, "learning_rate": 3.741450388831632e-05, "loss": 9.2435, "step": 23236 }, { "epoch": 0.9796993907709172, "grad_norm": 1.709855556488037, "learning_rate": 3.741372310190824e-05, "loss": 6.3636, "step": 23237 }, { "epoch": 0.9797415519531167, "grad_norm": 2.6936440467834473, "learning_rate": 3.741294231550017e-05, "loss": 7.5173, "step": 23238 }, { "epoch": 0.9797837131353163, "grad_norm": 2.011854648590088, "learning_rate": 3.7412161529092104e-05, "loss": 6.5309, "step": 23239 }, { "epoch": 0.9798258743175159, "grad_norm": 1.4379684925079346, "learning_rate": 3.7411380742684035e-05, "loss": 6.9261, "step": 23240 }, { "epoch": 0.9798680354997155, "grad_norm": 1.4793015718460083, "learning_rate": 3.741059995627597e-05, "loss": 6.8254, "step": 23241 }, { "epoch": 0.9799101966819149, "grad_norm": 0.9491938352584839, "learning_rate": 3.740981916986789e-05, "loss": 6.509, "step": 23242 }, { "epoch": 0.9799523578641145, "grad_norm": 2.342341184616089, "learning_rate": 3.740903838345982e-05, "loss": 6.4968, "step": 23243 }, { "epoch": 0.9799945190463141, "grad_norm": 4.4011712074279785, "learning_rate": 3.7408257597051754e-05, "loss": 7.8779, "step": 23244 }, { "epoch": 0.9800366802285136, "grad_norm": 2.5283045768737793, "learning_rate": 3.7407476810643685e-05, "loss": 7.3116, "step": 23245 }, { "epoch": 0.9800788414107131, "grad_norm": 1.5490052700042725, "learning_rate": 3.740669602423561e-05, "loss": 7.1026, "step": 23246 }, { "epoch": 0.9801210025929127, "grad_norm": 1.6006813049316406, "learning_rate": 3.740591523782754e-05, "loss": 7.1837, "step": 23247 }, { "epoch": 0.9801631637751123, "grad_norm": 1.991033673286438, "learning_rate": 3.740513445141947e-05, "loss": 6.4769, "step": 23248 }, { "epoch": 0.9802053249573118, "grad_norm": 1.042791485786438, "learning_rate": 3.7404353665011396e-05, "loss": 7.0269, "step": 23249 }, { "epoch": 0.9802474861395114, "grad_norm": 1.3263680934906006, "learning_rate": 3.7403572878603335e-05, "loss": 6.6017, "step": 23250 }, { "epoch": 0.9802896473217109, "grad_norm": 1.4426132440567017, "learning_rate": 3.740279209219526e-05, "loss": 7.218, "step": 23251 }, { "epoch": 0.9803318085039104, "grad_norm": 1.47370183467865, "learning_rate": 3.740201130578719e-05, "loss": 6.5553, "step": 23252 }, { "epoch": 0.98037396968611, "grad_norm": 2.342583656311035, "learning_rate": 3.740123051937912e-05, "loss": 8.1458, "step": 23253 }, { "epoch": 0.9804161308683096, "grad_norm": 3.213892936706543, "learning_rate": 3.7400449732971046e-05, "loss": 7.7273, "step": 23254 }, { "epoch": 0.980458292050509, "grad_norm": 1.7201625108718872, "learning_rate": 3.7399668946562984e-05, "loss": 7.2437, "step": 23255 }, { "epoch": 0.9805004532327086, "grad_norm": 1.4280807971954346, "learning_rate": 3.739888816015491e-05, "loss": 6.8023, "step": 23256 }, { "epoch": 0.9805426144149082, "grad_norm": 4.622534275054932, "learning_rate": 3.739810737374684e-05, "loss": 6.9323, "step": 23257 }, { "epoch": 0.9805847755971078, "grad_norm": 1.1075124740600586, "learning_rate": 3.739732658733877e-05, "loss": 6.6032, "step": 23258 }, { "epoch": 0.9806269367793073, "grad_norm": 2.188067674636841, "learning_rate": 3.73965458009307e-05, "loss": 6.6093, "step": 23259 }, { "epoch": 0.9806690979615068, "grad_norm": 1.4451022148132324, "learning_rate": 3.739576501452263e-05, "loss": 6.5715, "step": 23260 }, { "epoch": 0.9807112591437064, "grad_norm": 1.9955415725708008, "learning_rate": 3.739498422811456e-05, "loss": 6.526, "step": 23261 }, { "epoch": 0.980753420325906, "grad_norm": 1.2083756923675537, "learning_rate": 3.739420344170649e-05, "loss": 6.8787, "step": 23262 }, { "epoch": 0.9807955815081055, "grad_norm": 1.9870985746383667, "learning_rate": 3.7393422655298414e-05, "loss": 6.6284, "step": 23263 }, { "epoch": 0.980837742690305, "grad_norm": 1.1751247644424438, "learning_rate": 3.739264186889035e-05, "loss": 6.7233, "step": 23264 }, { "epoch": 0.9808799038725046, "grad_norm": 1.2937464714050293, "learning_rate": 3.7391861082482276e-05, "loss": 6.8299, "step": 23265 }, { "epoch": 0.9809220650547041, "grad_norm": 1.671110987663269, "learning_rate": 3.739108029607421e-05, "loss": 6.526, "step": 23266 }, { "epoch": 0.9809642262369037, "grad_norm": 1.8345149755477905, "learning_rate": 3.739029950966614e-05, "loss": 6.4289, "step": 23267 }, { "epoch": 0.9810063874191032, "grad_norm": 1.1245927810668945, "learning_rate": 3.738951872325806e-05, "loss": 7.0812, "step": 23268 }, { "epoch": 0.9810485486013028, "grad_norm": 1.6504344940185547, "learning_rate": 3.738873793685e-05, "loss": 7.0508, "step": 23269 }, { "epoch": 0.9810907097835023, "grad_norm": 1.8439685106277466, "learning_rate": 3.7387957150441926e-05, "loss": 7.11, "step": 23270 }, { "epoch": 0.9811328709657019, "grad_norm": 0.9753923416137695, "learning_rate": 3.738717636403386e-05, "loss": 6.6455, "step": 23271 }, { "epoch": 0.9811750321479015, "grad_norm": 1.4074054956436157, "learning_rate": 3.738639557762579e-05, "loss": 6.919, "step": 23272 }, { "epoch": 0.9812171933301009, "grad_norm": 0.8205609321594238, "learning_rate": 3.738561479121771e-05, "loss": 6.4072, "step": 23273 }, { "epoch": 0.9812593545123005, "grad_norm": 2.1185569763183594, "learning_rate": 3.7384834004809644e-05, "loss": 7.538, "step": 23274 }, { "epoch": 0.9813015156945001, "grad_norm": 4.383954048156738, "learning_rate": 3.7384053218401575e-05, "loss": 8.2717, "step": 23275 }, { "epoch": 0.9813436768766997, "grad_norm": 2.4288785457611084, "learning_rate": 3.7383272431993507e-05, "loss": 6.711, "step": 23276 }, { "epoch": 0.9813858380588991, "grad_norm": 2.2428536415100098, "learning_rate": 3.738249164558544e-05, "loss": 7.1295, "step": 23277 }, { "epoch": 0.9814279992410987, "grad_norm": 2.5202476978302, "learning_rate": 3.738171085917736e-05, "loss": 6.6338, "step": 23278 }, { "epoch": 0.9814701604232983, "grad_norm": 4.319442272186279, "learning_rate": 3.7380930072769294e-05, "loss": 8.0485, "step": 23279 }, { "epoch": 0.9815123216054978, "grad_norm": 1.639559268951416, "learning_rate": 3.7380149286361225e-05, "loss": 7.0887, "step": 23280 }, { "epoch": 0.9815544827876974, "grad_norm": 2.1255671977996826, "learning_rate": 3.7379368499953156e-05, "loss": 6.5914, "step": 23281 }, { "epoch": 0.9815966439698969, "grad_norm": 1.8063229322433472, "learning_rate": 3.737858771354508e-05, "loss": 6.3722, "step": 23282 }, { "epoch": 0.9816388051520964, "grad_norm": 2.556330442428589, "learning_rate": 3.737780692713702e-05, "loss": 6.9142, "step": 23283 }, { "epoch": 0.981680966334296, "grad_norm": 3.2006852626800537, "learning_rate": 3.737702614072894e-05, "loss": 8.0111, "step": 23284 }, { "epoch": 0.9817231275164956, "grad_norm": 3.851628541946411, "learning_rate": 3.7376245354320874e-05, "loss": 7.7623, "step": 23285 }, { "epoch": 0.9817652886986951, "grad_norm": 2.1662473678588867, "learning_rate": 3.7375464567912806e-05, "loss": 7.3934, "step": 23286 }, { "epoch": 0.9818074498808946, "grad_norm": 2.452519416809082, "learning_rate": 3.737468378150473e-05, "loss": 7.7676, "step": 23287 }, { "epoch": 0.9818496110630942, "grad_norm": 2.184950828552246, "learning_rate": 3.737390299509667e-05, "loss": 7.3584, "step": 23288 }, { "epoch": 0.9818917722452938, "grad_norm": 3.052211046218872, "learning_rate": 3.737312220868859e-05, "loss": 7.6409, "step": 23289 }, { "epoch": 0.9819339334274934, "grad_norm": 2.6974854469299316, "learning_rate": 3.7372341422280524e-05, "loss": 7.6521, "step": 23290 }, { "epoch": 0.9819760946096928, "grad_norm": 1.902244210243225, "learning_rate": 3.7371560635872455e-05, "loss": 6.6398, "step": 23291 }, { "epoch": 0.9820182557918924, "grad_norm": 2.9231605529785156, "learning_rate": 3.737077984946438e-05, "loss": 8.3554, "step": 23292 }, { "epoch": 0.982060416974092, "grad_norm": 2.6842257976531982, "learning_rate": 3.736999906305631e-05, "loss": 7.7392, "step": 23293 }, { "epoch": 0.9821025781562915, "grad_norm": 3.0987062454223633, "learning_rate": 3.736921827664824e-05, "loss": 7.8385, "step": 23294 }, { "epoch": 0.982144739338491, "grad_norm": 2.0268354415893555, "learning_rate": 3.7368437490240173e-05, "loss": 6.4033, "step": 23295 }, { "epoch": 0.9821869005206906, "grad_norm": 2.3616178035736084, "learning_rate": 3.73676567038321e-05, "loss": 6.5223, "step": 23296 }, { "epoch": 0.9822290617028901, "grad_norm": 1.4309911727905273, "learning_rate": 3.736687591742403e-05, "loss": 7.6432, "step": 23297 }, { "epoch": 0.9822712228850897, "grad_norm": 1.2794971466064453, "learning_rate": 3.736609513101596e-05, "loss": 6.9444, "step": 23298 }, { "epoch": 0.9823133840672893, "grad_norm": 1.5351111888885498, "learning_rate": 3.736531434460789e-05, "loss": 7.2219, "step": 23299 }, { "epoch": 0.9823555452494888, "grad_norm": 1.4956772327423096, "learning_rate": 3.736453355819982e-05, "loss": 7.0605, "step": 23300 }, { "epoch": 0.9823977064316883, "grad_norm": 1.2787092924118042, "learning_rate": 3.736375277179175e-05, "loss": 6.683, "step": 23301 }, { "epoch": 0.9824398676138879, "grad_norm": 1.6914652585983276, "learning_rate": 3.7362971985383685e-05, "loss": 6.4974, "step": 23302 }, { "epoch": 0.9824820287960875, "grad_norm": 1.3642585277557373, "learning_rate": 3.736219119897561e-05, "loss": 6.7232, "step": 23303 }, { "epoch": 0.9825241899782869, "grad_norm": 1.8751152753829956, "learning_rate": 3.736141041256754e-05, "loss": 7.1707, "step": 23304 }, { "epoch": 0.9825663511604865, "grad_norm": 1.3179007768630981, "learning_rate": 3.736062962615947e-05, "loss": 6.571, "step": 23305 }, { "epoch": 0.9826085123426861, "grad_norm": 1.3977948427200317, "learning_rate": 3.73598488397514e-05, "loss": 6.5815, "step": 23306 }, { "epoch": 0.9826506735248857, "grad_norm": 1.2801892757415771, "learning_rate": 3.735906805334333e-05, "loss": 6.935, "step": 23307 }, { "epoch": 0.9826928347070852, "grad_norm": 1.2351115942001343, "learning_rate": 3.735828726693526e-05, "loss": 6.7057, "step": 23308 }, { "epoch": 0.9827349958892847, "grad_norm": 0.8373124599456787, "learning_rate": 3.735750648052719e-05, "loss": 6.5378, "step": 23309 }, { "epoch": 0.9827771570714843, "grad_norm": 2.3519015312194824, "learning_rate": 3.735672569411912e-05, "loss": 7.4517, "step": 23310 }, { "epoch": 0.9828193182536838, "grad_norm": 1.1941983699798584, "learning_rate": 3.7355944907711046e-05, "loss": 6.6528, "step": 23311 }, { "epoch": 0.9828614794358834, "grad_norm": 2.8721487522125244, "learning_rate": 3.735516412130298e-05, "loss": 7.3057, "step": 23312 }, { "epoch": 0.9829036406180829, "grad_norm": 1.7305152416229248, "learning_rate": 3.735438333489491e-05, "loss": 7.1505, "step": 23313 }, { "epoch": 0.9829458018002825, "grad_norm": 0.9508771896362305, "learning_rate": 3.735360254848684e-05, "loss": 6.6009, "step": 23314 }, { "epoch": 0.982987962982482, "grad_norm": 1.1250720024108887, "learning_rate": 3.7352821762078765e-05, "loss": 6.7317, "step": 23315 }, { "epoch": 0.9830301241646816, "grad_norm": 1.2940592765808105, "learning_rate": 3.7352040975670696e-05, "loss": 7.1625, "step": 23316 }, { "epoch": 0.9830722853468812, "grad_norm": 0.9578319191932678, "learning_rate": 3.735126018926263e-05, "loss": 6.6839, "step": 23317 }, { "epoch": 0.9831144465290806, "grad_norm": 1.1173890829086304, "learning_rate": 3.735047940285455e-05, "loss": 6.6271, "step": 23318 }, { "epoch": 0.9831566077112802, "grad_norm": 1.3110181093215942, "learning_rate": 3.734969861644649e-05, "loss": 7.0358, "step": 23319 }, { "epoch": 0.9831987688934798, "grad_norm": 0.9354126453399658, "learning_rate": 3.7348917830038414e-05, "loss": 6.5247, "step": 23320 }, { "epoch": 0.9832409300756794, "grad_norm": 3.749922513961792, "learning_rate": 3.734813704363035e-05, "loss": 7.7562, "step": 23321 }, { "epoch": 0.9832830912578788, "grad_norm": 1.38972806930542, "learning_rate": 3.734735625722228e-05, "loss": 6.8011, "step": 23322 }, { "epoch": 0.9833252524400784, "grad_norm": 1.9011807441711426, "learning_rate": 3.73465754708142e-05, "loss": 7.2364, "step": 23323 }, { "epoch": 0.983367413622278, "grad_norm": 1.080142617225647, "learning_rate": 3.734579468440614e-05, "loss": 6.92, "step": 23324 }, { "epoch": 0.9834095748044775, "grad_norm": 1.6283625364303589, "learning_rate": 3.7345013897998064e-05, "loss": 6.4928, "step": 23325 }, { "epoch": 0.983451735986677, "grad_norm": 1.5107407569885254, "learning_rate": 3.7344233111589995e-05, "loss": 6.4279, "step": 23326 }, { "epoch": 0.9834938971688766, "grad_norm": 2.502450704574585, "learning_rate": 3.7343452325181926e-05, "loss": 6.5366, "step": 23327 }, { "epoch": 0.9835360583510762, "grad_norm": 3.1306238174438477, "learning_rate": 3.734267153877386e-05, "loss": 7.7313, "step": 23328 }, { "epoch": 0.9835782195332757, "grad_norm": 1.7036234140396118, "learning_rate": 3.734189075236578e-05, "loss": 7.5051, "step": 23329 }, { "epoch": 0.9836203807154753, "grad_norm": 1.7770673036575317, "learning_rate": 3.734110996595771e-05, "loss": 7.1356, "step": 23330 }, { "epoch": 0.9836625418976748, "grad_norm": 2.0637054443359375, "learning_rate": 3.7340329179549645e-05, "loss": 6.5168, "step": 23331 }, { "epoch": 0.9837047030798743, "grad_norm": 2.443295955657959, "learning_rate": 3.733954839314157e-05, "loss": 6.4687, "step": 23332 }, { "epoch": 0.9837468642620739, "grad_norm": 1.535211205482483, "learning_rate": 3.733876760673351e-05, "loss": 6.4887, "step": 23333 }, { "epoch": 0.9837890254442735, "grad_norm": 1.1160866022109985, "learning_rate": 3.733798682032543e-05, "loss": 6.5742, "step": 23334 }, { "epoch": 0.983831186626473, "grad_norm": 2.564257860183716, "learning_rate": 3.733720603391736e-05, "loss": 7.411, "step": 23335 }, { "epoch": 0.9838733478086725, "grad_norm": 1.6034114360809326, "learning_rate": 3.7336425247509294e-05, "loss": 6.4088, "step": 23336 }, { "epoch": 0.9839155089908721, "grad_norm": 2.1911392211914062, "learning_rate": 3.733564446110122e-05, "loss": 7.1376, "step": 23337 }, { "epoch": 0.9839576701730717, "grad_norm": 1.9891786575317383, "learning_rate": 3.733486367469316e-05, "loss": 7.1426, "step": 23338 }, { "epoch": 0.9839998313552712, "grad_norm": 1.9946796894073486, "learning_rate": 3.733408288828508e-05, "loss": 6.2624, "step": 23339 }, { "epoch": 0.9840419925374707, "grad_norm": 1.6646941900253296, "learning_rate": 3.733330210187701e-05, "loss": 6.7761, "step": 23340 }, { "epoch": 0.9840841537196703, "grad_norm": 1.4570389986038208, "learning_rate": 3.7332521315468944e-05, "loss": 7.7808, "step": 23341 }, { "epoch": 0.9841263149018699, "grad_norm": 1.3462151288986206, "learning_rate": 3.733174052906087e-05, "loss": 6.7452, "step": 23342 }, { "epoch": 0.9841684760840694, "grad_norm": 2.06147837638855, "learning_rate": 3.73309597426528e-05, "loss": 7.2742, "step": 23343 }, { "epoch": 0.9842106372662689, "grad_norm": 3.095020294189453, "learning_rate": 3.733017895624473e-05, "loss": 7.5452, "step": 23344 }, { "epoch": 0.9842527984484685, "grad_norm": 1.5545716285705566, "learning_rate": 3.732939816983666e-05, "loss": 6.3972, "step": 23345 }, { "epoch": 0.984294959630668, "grad_norm": 2.8833866119384766, "learning_rate": 3.732861738342859e-05, "loss": 7.9454, "step": 23346 }, { "epoch": 0.9843371208128676, "grad_norm": 1.7235133647918701, "learning_rate": 3.7327836597020524e-05, "loss": 6.4875, "step": 23347 }, { "epoch": 0.9843792819950672, "grad_norm": 1.8233574628829956, "learning_rate": 3.732705581061245e-05, "loss": 7.0619, "step": 23348 }, { "epoch": 0.9844214431772667, "grad_norm": 3.0621535778045654, "learning_rate": 3.732627502420438e-05, "loss": 7.785, "step": 23349 }, { "epoch": 0.9844636043594662, "grad_norm": 1.08195960521698, "learning_rate": 3.732549423779631e-05, "loss": 6.3271, "step": 23350 }, { "epoch": 0.9845057655416658, "grad_norm": 1.065194845199585, "learning_rate": 3.7324713451388236e-05, "loss": 6.7796, "step": 23351 }, { "epoch": 0.9845479267238654, "grad_norm": 1.543623924255371, "learning_rate": 3.7323932664980174e-05, "loss": 6.7397, "step": 23352 }, { "epoch": 0.9845900879060648, "grad_norm": 0.9756635427474976, "learning_rate": 3.73231518785721e-05, "loss": 6.3915, "step": 23353 }, { "epoch": 0.9846322490882644, "grad_norm": 1.712327480316162, "learning_rate": 3.732237109216403e-05, "loss": 6.7247, "step": 23354 }, { "epoch": 0.984674410270464, "grad_norm": 1.7382770776748657, "learning_rate": 3.732159030575596e-05, "loss": 6.7777, "step": 23355 }, { "epoch": 0.9847165714526636, "grad_norm": 2.8215060234069824, "learning_rate": 3.7320809519347885e-05, "loss": 7.6988, "step": 23356 }, { "epoch": 0.9847587326348631, "grad_norm": 1.7694939374923706, "learning_rate": 3.7320028732939823e-05, "loss": 6.4237, "step": 23357 }, { "epoch": 0.9848008938170626, "grad_norm": 1.4465327262878418, "learning_rate": 3.731924794653175e-05, "loss": 6.8123, "step": 23358 }, { "epoch": 0.9848430549992622, "grad_norm": 0.6831939816474915, "learning_rate": 3.731846716012368e-05, "loss": 6.5681, "step": 23359 }, { "epoch": 0.9848852161814617, "grad_norm": 1.303507685661316, "learning_rate": 3.731768637371561e-05, "loss": 6.873, "step": 23360 }, { "epoch": 0.9849273773636613, "grad_norm": 1.897664189338684, "learning_rate": 3.7316905587307535e-05, "loss": 7.0491, "step": 23361 }, { "epoch": 0.9849695385458608, "grad_norm": 1.6923484802246094, "learning_rate": 3.7316124800899466e-05, "loss": 6.8984, "step": 23362 }, { "epoch": 0.9850116997280604, "grad_norm": 1.652124047279358, "learning_rate": 3.73153440144914e-05, "loss": 7.0514, "step": 23363 }, { "epoch": 0.9850538609102599, "grad_norm": 1.8880231380462646, "learning_rate": 3.731456322808333e-05, "loss": 7.0953, "step": 23364 }, { "epoch": 0.9850960220924595, "grad_norm": 1.0624768733978271, "learning_rate": 3.731378244167525e-05, "loss": 6.383, "step": 23365 }, { "epoch": 0.9851381832746591, "grad_norm": 3.089980363845825, "learning_rate": 3.731300165526719e-05, "loss": 7.1961, "step": 23366 }, { "epoch": 0.9851803444568585, "grad_norm": 3.301734447479248, "learning_rate": 3.7312220868859116e-05, "loss": 7.1322, "step": 23367 }, { "epoch": 0.9852225056390581, "grad_norm": 1.3823833465576172, "learning_rate": 3.731144008245105e-05, "loss": 6.9664, "step": 23368 }, { "epoch": 0.9852646668212577, "grad_norm": 2.0382533073425293, "learning_rate": 3.731065929604298e-05, "loss": 6.6582, "step": 23369 }, { "epoch": 0.9853068280034573, "grad_norm": 1.8058857917785645, "learning_rate": 3.73098785096349e-05, "loss": 7.0204, "step": 23370 }, { "epoch": 0.9853489891856567, "grad_norm": 1.55250084400177, "learning_rate": 3.730909772322684e-05, "loss": 7.176, "step": 23371 }, { "epoch": 0.9853911503678563, "grad_norm": 3.566216468811035, "learning_rate": 3.7308316936818765e-05, "loss": 8.3258, "step": 23372 }, { "epoch": 0.9854333115500559, "grad_norm": 0.9789084196090698, "learning_rate": 3.7307536150410697e-05, "loss": 6.5223, "step": 23373 }, { "epoch": 0.9854754727322554, "grad_norm": 0.9470322728157043, "learning_rate": 3.730675536400263e-05, "loss": 6.4699, "step": 23374 }, { "epoch": 0.985517633914455, "grad_norm": 3.3746373653411865, "learning_rate": 3.730597457759455e-05, "loss": 7.2939, "step": 23375 }, { "epoch": 0.9855597950966545, "grad_norm": 1.3498774766921997, "learning_rate": 3.7305193791186484e-05, "loss": 6.774, "step": 23376 }, { "epoch": 0.985601956278854, "grad_norm": 3.072901964187622, "learning_rate": 3.7304413004778415e-05, "loss": 6.3785, "step": 23377 }, { "epoch": 0.9856441174610536, "grad_norm": 0.7113720774650574, "learning_rate": 3.7303632218370346e-05, "loss": 6.622, "step": 23378 }, { "epoch": 0.9856862786432532, "grad_norm": 3.683084487915039, "learning_rate": 3.730285143196228e-05, "loss": 7.7839, "step": 23379 }, { "epoch": 0.9857284398254527, "grad_norm": 1.568166732788086, "learning_rate": 3.73020706455542e-05, "loss": 6.5209, "step": 23380 }, { "epoch": 0.9857706010076522, "grad_norm": 1.6384470462799072, "learning_rate": 3.730128985914613e-05, "loss": 7.0927, "step": 23381 }, { "epoch": 0.9858127621898518, "grad_norm": 2.634708881378174, "learning_rate": 3.7300509072738064e-05, "loss": 7.3336, "step": 23382 }, { "epoch": 0.9858549233720514, "grad_norm": 2.431145429611206, "learning_rate": 3.7299728286329996e-05, "loss": 7.2056, "step": 23383 }, { "epoch": 0.9858970845542508, "grad_norm": 4.249538898468018, "learning_rate": 3.729894749992192e-05, "loss": 7.0712, "step": 23384 }, { "epoch": 0.9859392457364504, "grad_norm": 1.9316943883895874, "learning_rate": 3.729816671351385e-05, "loss": 6.9013, "step": 23385 }, { "epoch": 0.98598140691865, "grad_norm": 5.236485481262207, "learning_rate": 3.729738592710578e-05, "loss": 7.2145, "step": 23386 }, { "epoch": 0.9860235681008496, "grad_norm": 3.1184277534484863, "learning_rate": 3.729660514069771e-05, "loss": 7.4781, "step": 23387 }, { "epoch": 0.9860657292830491, "grad_norm": 2.5329835414886475, "learning_rate": 3.7295824354289645e-05, "loss": 6.5333, "step": 23388 }, { "epoch": 0.9861078904652486, "grad_norm": 2.718987464904785, "learning_rate": 3.729504356788157e-05, "loss": 7.0992, "step": 23389 }, { "epoch": 0.9861500516474482, "grad_norm": 2.324681043624878, "learning_rate": 3.729426278147351e-05, "loss": 6.9779, "step": 23390 }, { "epoch": 0.9861922128296478, "grad_norm": 2.7693967819213867, "learning_rate": 3.729348199506543e-05, "loss": 8.4824, "step": 23391 }, { "epoch": 0.9862343740118473, "grad_norm": 1.524718999862671, "learning_rate": 3.729270120865736e-05, "loss": 6.5089, "step": 23392 }, { "epoch": 0.9862765351940468, "grad_norm": 1.3979554176330566, "learning_rate": 3.7291920422249295e-05, "loss": 7.3187, "step": 23393 }, { "epoch": 0.9863186963762464, "grad_norm": 1.3718849420547485, "learning_rate": 3.729113963584122e-05, "loss": 6.696, "step": 23394 }, { "epoch": 0.9863608575584459, "grad_norm": 1.9438542127609253, "learning_rate": 3.729035884943315e-05, "loss": 6.5798, "step": 23395 }, { "epoch": 0.9864030187406455, "grad_norm": 4.601085186004639, "learning_rate": 3.728957806302508e-05, "loss": 7.2364, "step": 23396 }, { "epoch": 0.9864451799228451, "grad_norm": 1.3571934700012207, "learning_rate": 3.728879727661701e-05, "loss": 6.5543, "step": 23397 }, { "epoch": 0.9864873411050445, "grad_norm": 1.6921403408050537, "learning_rate": 3.728801649020894e-05, "loss": 6.9765, "step": 23398 }, { "epoch": 0.9865295022872441, "grad_norm": 3.239208698272705, "learning_rate": 3.728723570380087e-05, "loss": 7.454, "step": 23399 }, { "epoch": 0.9865716634694437, "grad_norm": 1.419570803642273, "learning_rate": 3.72864549173928e-05, "loss": 6.3819, "step": 23400 }, { "epoch": 0.9866138246516433, "grad_norm": 3.185947895050049, "learning_rate": 3.7285674130984724e-05, "loss": 7.5335, "step": 23401 }, { "epoch": 0.9866559858338427, "grad_norm": 3.3011412620544434, "learning_rate": 3.728489334457666e-05, "loss": 7.7191, "step": 23402 }, { "epoch": 0.9866981470160423, "grad_norm": 0.9629467129707336, "learning_rate": 3.728411255816859e-05, "loss": 6.4138, "step": 23403 }, { "epoch": 0.9867403081982419, "grad_norm": 1.9283376932144165, "learning_rate": 3.728333177176052e-05, "loss": 6.9492, "step": 23404 }, { "epoch": 0.9867824693804415, "grad_norm": 1.3740003108978271, "learning_rate": 3.728255098535245e-05, "loss": 6.513, "step": 23405 }, { "epoch": 0.986824630562641, "grad_norm": 1.4056110382080078, "learning_rate": 3.7281770198944374e-05, "loss": 6.3965, "step": 23406 }, { "epoch": 0.9868667917448405, "grad_norm": 3.871253490447998, "learning_rate": 3.728098941253631e-05, "loss": 7.8809, "step": 23407 }, { "epoch": 0.9869089529270401, "grad_norm": 0.9010334610939026, "learning_rate": 3.7280208626128236e-05, "loss": 6.5532, "step": 23408 }, { "epoch": 0.9869511141092396, "grad_norm": 3.9042508602142334, "learning_rate": 3.727942783972017e-05, "loss": 7.8279, "step": 23409 }, { "epoch": 0.9869932752914392, "grad_norm": 1.0700055360794067, "learning_rate": 3.72786470533121e-05, "loss": 6.5476, "step": 23410 }, { "epoch": 0.9870354364736387, "grad_norm": 0.9947236180305481, "learning_rate": 3.727786626690403e-05, "loss": 6.6217, "step": 23411 }, { "epoch": 0.9870775976558382, "grad_norm": 2.2621965408325195, "learning_rate": 3.7277085480495955e-05, "loss": 8.3792, "step": 23412 }, { "epoch": 0.9871197588380378, "grad_norm": 1.632616400718689, "learning_rate": 3.7276304694087886e-05, "loss": 7.1262, "step": 23413 }, { "epoch": 0.9871619200202374, "grad_norm": 1.6615673303604126, "learning_rate": 3.727552390767982e-05, "loss": 7.4238, "step": 23414 }, { "epoch": 0.987204081202437, "grad_norm": 1.8613227605819702, "learning_rate": 3.727474312127175e-05, "loss": 6.9047, "step": 23415 }, { "epoch": 0.9872462423846364, "grad_norm": 2.3371472358703613, "learning_rate": 3.727396233486368e-05, "loss": 7.3853, "step": 23416 }, { "epoch": 0.987288403566836, "grad_norm": 1.5557358264923096, "learning_rate": 3.7273181548455604e-05, "loss": 6.8706, "step": 23417 }, { "epoch": 0.9873305647490356, "grad_norm": 1.4953017234802246, "learning_rate": 3.7272400762047535e-05, "loss": 6.556, "step": 23418 }, { "epoch": 0.9873727259312352, "grad_norm": 1.4578062295913696, "learning_rate": 3.727161997563947e-05, "loss": 6.9888, "step": 23419 }, { "epoch": 0.9874148871134346, "grad_norm": 1.156786561012268, "learning_rate": 3.727083918923139e-05, "loss": 6.466, "step": 23420 }, { "epoch": 0.9874570482956342, "grad_norm": 1.6097346544265747, "learning_rate": 3.727005840282333e-05, "loss": 6.5517, "step": 23421 }, { "epoch": 0.9874992094778338, "grad_norm": 0.9500367641448975, "learning_rate": 3.7269277616415254e-05, "loss": 6.7835, "step": 23422 }, { "epoch": 0.9875413706600333, "grad_norm": 2.2529592514038086, "learning_rate": 3.7268496830007185e-05, "loss": 7.0196, "step": 23423 }, { "epoch": 0.9875835318422329, "grad_norm": 1.984585165977478, "learning_rate": 3.7267716043599116e-05, "loss": 7.3985, "step": 23424 }, { "epoch": 0.9876256930244324, "grad_norm": 1.5571061372756958, "learning_rate": 3.726693525719104e-05, "loss": 7.0289, "step": 23425 }, { "epoch": 0.987667854206632, "grad_norm": 4.357640743255615, "learning_rate": 3.726615447078298e-05, "loss": 7.6875, "step": 23426 }, { "epoch": 0.9877100153888315, "grad_norm": 1.1257922649383545, "learning_rate": 3.72653736843749e-05, "loss": 7.0064, "step": 23427 }, { "epoch": 0.9877521765710311, "grad_norm": 1.1143676042556763, "learning_rate": 3.7264592897966834e-05, "loss": 6.6697, "step": 23428 }, { "epoch": 0.9877943377532306, "grad_norm": 1.0898189544677734, "learning_rate": 3.7263812111558766e-05, "loss": 6.6781, "step": 23429 }, { "epoch": 0.9878364989354301, "grad_norm": 1.1167066097259521, "learning_rate": 3.726303132515069e-05, "loss": 6.6498, "step": 23430 }, { "epoch": 0.9878786601176297, "grad_norm": 1.1235320568084717, "learning_rate": 3.726225053874262e-05, "loss": 6.7848, "step": 23431 }, { "epoch": 0.9879208212998293, "grad_norm": 2.5151526927948, "learning_rate": 3.726146975233455e-05, "loss": 7.8255, "step": 23432 }, { "epoch": 0.9879629824820289, "grad_norm": 1.8793152570724487, "learning_rate": 3.7260688965926484e-05, "loss": 6.7924, "step": 23433 }, { "epoch": 0.9880051436642283, "grad_norm": 0.8564863801002502, "learning_rate": 3.725990817951841e-05, "loss": 6.7132, "step": 23434 }, { "epoch": 0.9880473048464279, "grad_norm": 1.1633960008621216, "learning_rate": 3.7259127393110347e-05, "loss": 6.5158, "step": 23435 }, { "epoch": 0.9880894660286275, "grad_norm": 2.476870059967041, "learning_rate": 3.725834660670227e-05, "loss": 7.7124, "step": 23436 }, { "epoch": 0.988131627210827, "grad_norm": 1.7647696733474731, "learning_rate": 3.72575658202942e-05, "loss": 6.8675, "step": 23437 }, { "epoch": 0.9881737883930265, "grad_norm": 3.115764856338501, "learning_rate": 3.7256785033886134e-05, "loss": 7.7586, "step": 23438 }, { "epoch": 0.9882159495752261, "grad_norm": 2.8459384441375732, "learning_rate": 3.725600424747806e-05, "loss": 8.8362, "step": 23439 }, { "epoch": 0.9882581107574256, "grad_norm": 2.8488430976867676, "learning_rate": 3.7255223461069996e-05, "loss": 8.0575, "step": 23440 }, { "epoch": 0.9883002719396252, "grad_norm": 1.8561252355575562, "learning_rate": 3.725444267466192e-05, "loss": 7.0718, "step": 23441 }, { "epoch": 0.9883424331218247, "grad_norm": 1.385772466659546, "learning_rate": 3.725366188825385e-05, "loss": 6.5198, "step": 23442 }, { "epoch": 0.9883845943040243, "grad_norm": 2.730271577835083, "learning_rate": 3.725288110184578e-05, "loss": 7.7812, "step": 23443 }, { "epoch": 0.9884267554862238, "grad_norm": 2.913416862487793, "learning_rate": 3.725210031543771e-05, "loss": 6.516, "step": 23444 }, { "epoch": 0.9884689166684234, "grad_norm": 1.644434928894043, "learning_rate": 3.725131952902964e-05, "loss": 6.415, "step": 23445 }, { "epoch": 0.988511077850623, "grad_norm": 1.4418865442276, "learning_rate": 3.725053874262157e-05, "loss": 6.4524, "step": 23446 }, { "epoch": 0.9885532390328224, "grad_norm": 3.3094711303710938, "learning_rate": 3.72497579562135e-05, "loss": 7.8126, "step": 23447 }, { "epoch": 0.988595400215022, "grad_norm": 1.1009886264801025, "learning_rate": 3.724897716980543e-05, "loss": 6.4466, "step": 23448 }, { "epoch": 0.9886375613972216, "grad_norm": 3.280489921569824, "learning_rate": 3.724819638339736e-05, "loss": 7.6826, "step": 23449 }, { "epoch": 0.9886797225794212, "grad_norm": 1.5223815441131592, "learning_rate": 3.724741559698929e-05, "loss": 6.9682, "step": 23450 }, { "epoch": 0.9887218837616206, "grad_norm": 1.392661213874817, "learning_rate": 3.724663481058122e-05, "loss": 6.6786, "step": 23451 }, { "epoch": 0.9887640449438202, "grad_norm": 1.7605268955230713, "learning_rate": 3.724585402417315e-05, "loss": 7.0785, "step": 23452 }, { "epoch": 0.9888062061260198, "grad_norm": 1.8310154676437378, "learning_rate": 3.7245073237765075e-05, "loss": 7.041, "step": 23453 }, { "epoch": 0.9888483673082193, "grad_norm": 1.1748642921447754, "learning_rate": 3.724429245135701e-05, "loss": 6.6543, "step": 23454 }, { "epoch": 0.9888905284904189, "grad_norm": 1.3358994722366333, "learning_rate": 3.724351166494894e-05, "loss": 7.3901, "step": 23455 }, { "epoch": 0.9889326896726184, "grad_norm": 1.309476375579834, "learning_rate": 3.724273087854087e-05, "loss": 7.1601, "step": 23456 }, { "epoch": 0.988974850854818, "grad_norm": 4.419305324554443, "learning_rate": 3.72419500921328e-05, "loss": 7.559, "step": 23457 }, { "epoch": 0.9890170120370175, "grad_norm": 1.2153563499450684, "learning_rate": 3.7241169305724725e-05, "loss": 6.7226, "step": 23458 }, { "epoch": 0.9890591732192171, "grad_norm": 2.521791934967041, "learning_rate": 3.724038851931666e-05, "loss": 6.5457, "step": 23459 }, { "epoch": 0.9891013344014166, "grad_norm": 3.337202787399292, "learning_rate": 3.723960773290859e-05, "loss": 6.4085, "step": 23460 }, { "epoch": 0.9891434955836161, "grad_norm": 1.8059369325637817, "learning_rate": 3.723882694650052e-05, "loss": 6.5355, "step": 23461 }, { "epoch": 0.9891856567658157, "grad_norm": 1.394283413887024, "learning_rate": 3.723804616009245e-05, "loss": 6.6061, "step": 23462 }, { "epoch": 0.9892278179480153, "grad_norm": 1.7601172924041748, "learning_rate": 3.7237265373684374e-05, "loss": 7.2443, "step": 23463 }, { "epoch": 0.9892699791302149, "grad_norm": 1.699780821800232, "learning_rate": 3.7236484587276306e-05, "loss": 6.3978, "step": 23464 }, { "epoch": 0.9893121403124143, "grad_norm": 1.101041555404663, "learning_rate": 3.723570380086824e-05, "loss": 6.5005, "step": 23465 }, { "epoch": 0.9893543014946139, "grad_norm": 3.6341333389282227, "learning_rate": 3.723492301446017e-05, "loss": 7.8244, "step": 23466 }, { "epoch": 0.9893964626768135, "grad_norm": 3.103571653366089, "learning_rate": 3.723414222805209e-05, "loss": 8.8799, "step": 23467 }, { "epoch": 0.989438623859013, "grad_norm": 1.0630615949630737, "learning_rate": 3.7233361441644024e-05, "loss": 6.6912, "step": 23468 }, { "epoch": 0.9894807850412125, "grad_norm": 1.7211772203445435, "learning_rate": 3.7232580655235955e-05, "loss": 6.9295, "step": 23469 }, { "epoch": 0.9895229462234121, "grad_norm": 1.3405159711837769, "learning_rate": 3.7231799868827886e-05, "loss": 6.6187, "step": 23470 }, { "epoch": 0.9895651074056117, "grad_norm": 2.0322763919830322, "learning_rate": 3.723101908241982e-05, "loss": 6.8531, "step": 23471 }, { "epoch": 0.9896072685878112, "grad_norm": 1.231965184211731, "learning_rate": 3.723023829601174e-05, "loss": 6.8375, "step": 23472 }, { "epoch": 0.9896494297700108, "grad_norm": 3.8332064151763916, "learning_rate": 3.722945750960368e-05, "loss": 7.6918, "step": 23473 }, { "epoch": 0.9896915909522103, "grad_norm": 1.086049199104309, "learning_rate": 3.7228676723195605e-05, "loss": 6.4826, "step": 23474 }, { "epoch": 0.9897337521344098, "grad_norm": 1.9081830978393555, "learning_rate": 3.722789593678753e-05, "loss": 7.203, "step": 23475 }, { "epoch": 0.9897759133166094, "grad_norm": 1.1897388696670532, "learning_rate": 3.722711515037947e-05, "loss": 6.5479, "step": 23476 }, { "epoch": 0.989818074498809, "grad_norm": 1.2896456718444824, "learning_rate": 3.722633436397139e-05, "loss": 7.6082, "step": 23477 }, { "epoch": 0.9898602356810084, "grad_norm": 1.3184911012649536, "learning_rate": 3.722555357756332e-05, "loss": 6.8611, "step": 23478 }, { "epoch": 0.989902396863208, "grad_norm": 1.275219440460205, "learning_rate": 3.7224772791155254e-05, "loss": 6.449, "step": 23479 }, { "epoch": 0.9899445580454076, "grad_norm": 3.081946611404419, "learning_rate": 3.7223992004747185e-05, "loss": 7.7704, "step": 23480 }, { "epoch": 0.9899867192276072, "grad_norm": 1.7194139957427979, "learning_rate": 3.722321121833911e-05, "loss": 7.124, "step": 23481 }, { "epoch": 0.9900288804098067, "grad_norm": 3.092576742172241, "learning_rate": 3.722243043193104e-05, "loss": 7.8149, "step": 23482 }, { "epoch": 0.9900710415920062, "grad_norm": 1.6192649602890015, "learning_rate": 3.722164964552297e-05, "loss": 6.4184, "step": 23483 }, { "epoch": 0.9901132027742058, "grad_norm": 1.6821948289871216, "learning_rate": 3.7220868859114904e-05, "loss": 7.0, "step": 23484 }, { "epoch": 0.9901553639564054, "grad_norm": 1.5288549661636353, "learning_rate": 3.7220088072706835e-05, "loss": 6.4394, "step": 23485 }, { "epoch": 0.9901975251386049, "grad_norm": 1.1311147212982178, "learning_rate": 3.721930728629876e-05, "loss": 6.663, "step": 23486 }, { "epoch": 0.9902396863208044, "grad_norm": 4.036046028137207, "learning_rate": 3.721852649989069e-05, "loss": 7.4413, "step": 23487 }, { "epoch": 0.990281847503004, "grad_norm": 1.0012136697769165, "learning_rate": 3.721774571348262e-05, "loss": 6.6536, "step": 23488 }, { "epoch": 0.9903240086852035, "grad_norm": 3.4942216873168945, "learning_rate": 3.7216964927074546e-05, "loss": 8.2539, "step": 23489 }, { "epoch": 0.9903661698674031, "grad_norm": 1.3758755922317505, "learning_rate": 3.7216184140666485e-05, "loss": 6.908, "step": 23490 }, { "epoch": 0.9904083310496027, "grad_norm": 2.5726120471954346, "learning_rate": 3.721540335425841e-05, "loss": 7.6287, "step": 23491 }, { "epoch": 0.9904504922318021, "grad_norm": 5.267165184020996, "learning_rate": 3.721462256785034e-05, "loss": 7.3327, "step": 23492 }, { "epoch": 0.9904926534140017, "grad_norm": 1.2013928890228271, "learning_rate": 3.721384178144227e-05, "loss": 6.5698, "step": 23493 }, { "epoch": 0.9905348145962013, "grad_norm": 1.04582679271698, "learning_rate": 3.7213060995034196e-05, "loss": 6.4591, "step": 23494 }, { "epoch": 0.9905769757784009, "grad_norm": 1.4082393646240234, "learning_rate": 3.7212280208626134e-05, "loss": 6.8192, "step": 23495 }, { "epoch": 0.9906191369606003, "grad_norm": 3.532261848449707, "learning_rate": 3.721149942221806e-05, "loss": 7.8358, "step": 23496 }, { "epoch": 0.9906612981427999, "grad_norm": 1.4347355365753174, "learning_rate": 3.721071863580999e-05, "loss": 6.5701, "step": 23497 }, { "epoch": 0.9907034593249995, "grad_norm": 2.5629007816314697, "learning_rate": 3.720993784940192e-05, "loss": 7.6603, "step": 23498 }, { "epoch": 0.990745620507199, "grad_norm": 1.9830785989761353, "learning_rate": 3.720915706299385e-05, "loss": 6.8603, "step": 23499 }, { "epoch": 0.9907877816893985, "grad_norm": 1.431662678718567, "learning_rate": 3.720837627658578e-05, "loss": 6.5957, "step": 23500 } ], "logging_steps": 1, "max_steps": 71154, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.87113849612051e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }