{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.999680034129693, "eval_steps": 500, "global_step": 9375, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0003199658703071672, "grad_norm": 7.29244327545166, "learning_rate": 1.0660980810234543e-08, "loss": 1.0554, "step": 1 }, { "epoch": 0.0006399317406143345, "grad_norm": 6.493714332580566, "learning_rate": 2.1321961620469085e-08, "loss": 1.0764, "step": 2 }, { "epoch": 0.0009598976109215017, "grad_norm": 6.872972011566162, "learning_rate": 3.1982942430703625e-08, "loss": 1.053, "step": 3 }, { "epoch": 0.001279863481228669, "grad_norm": 7.674505233764648, "learning_rate": 4.264392324093817e-08, "loss": 1.1434, "step": 4 }, { "epoch": 0.0015998293515358361, "grad_norm": 7.034660816192627, "learning_rate": 5.330490405117271e-08, "loss": 1.0999, "step": 5 }, { "epoch": 0.0019197952218430034, "grad_norm": 6.8419647216796875, "learning_rate": 6.396588486140725e-08, "loss": 1.0741, "step": 6 }, { "epoch": 0.0022397610921501706, "grad_norm": 7.366310119628906, "learning_rate": 7.462686567164179e-08, "loss": 1.1859, "step": 7 }, { "epoch": 0.002559726962457338, "grad_norm": 7.265549659729004, "learning_rate": 8.528784648187634e-08, "loss": 1.0836, "step": 8 }, { "epoch": 0.002879692832764505, "grad_norm": 7.0671586990356445, "learning_rate": 9.59488272921109e-08, "loss": 1.0631, "step": 9 }, { "epoch": 0.0031996587030716723, "grad_norm": 6.615237712860107, "learning_rate": 1.0660980810234542e-07, "loss": 0.9951, "step": 10 }, { "epoch": 0.0035196245733788395, "grad_norm": 7.323988914489746, "learning_rate": 1.1727078891257997e-07, "loss": 1.1467, "step": 11 }, { "epoch": 0.0038395904436860067, "grad_norm": 7.808572292327881, "learning_rate": 1.279317697228145e-07, "loss": 1.1524, "step": 12 }, { "epoch": 0.004159556313993174, "grad_norm": 7.407500267028809, "learning_rate": 1.3859275053304905e-07, "loss": 1.0871, "step": 13 }, { "epoch": 0.004479522184300341, "grad_norm": 7.308926105499268, "learning_rate": 1.4925373134328358e-07, "loss": 1.1039, "step": 14 }, { "epoch": 0.004799488054607509, "grad_norm": 7.3174004554748535, "learning_rate": 1.5991471215351813e-07, "loss": 1.145, "step": 15 }, { "epoch": 0.005119453924914676, "grad_norm": 7.225617408752441, "learning_rate": 1.7057569296375268e-07, "loss": 1.1349, "step": 16 }, { "epoch": 0.005439419795221843, "grad_norm": 6.9253668785095215, "learning_rate": 1.812366737739872e-07, "loss": 1.0338, "step": 17 }, { "epoch": 0.00575938566552901, "grad_norm": 7.221156597137451, "learning_rate": 1.918976545842218e-07, "loss": 1.1135, "step": 18 }, { "epoch": 0.006079351535836178, "grad_norm": 6.97838020324707, "learning_rate": 2.0255863539445632e-07, "loss": 1.054, "step": 19 }, { "epoch": 0.0063993174061433445, "grad_norm": 6.8710455894470215, "learning_rate": 2.1321961620469084e-07, "loss": 1.0712, "step": 20 }, { "epoch": 0.006719283276450512, "grad_norm": 7.243351936340332, "learning_rate": 2.2388059701492537e-07, "loss": 1.0826, "step": 21 }, { "epoch": 0.007039249146757679, "grad_norm": 6.599764347076416, "learning_rate": 2.3454157782515995e-07, "loss": 1.1181, "step": 22 }, { "epoch": 0.007359215017064847, "grad_norm": 6.794438362121582, "learning_rate": 2.4520255863539447e-07, "loss": 1.0853, "step": 23 }, { "epoch": 0.007679180887372013, "grad_norm": 6.50425386428833, "learning_rate": 2.55863539445629e-07, "loss": 1.0862, "step": 24 }, { "epoch": 0.00799914675767918, "grad_norm": 6.425209045410156, "learning_rate": 2.665245202558635e-07, "loss": 1.0684, "step": 25 }, { "epoch": 0.008319112627986349, "grad_norm": 6.9282708168029785, "learning_rate": 2.771855010660981e-07, "loss": 1.1113, "step": 26 }, { "epoch": 0.008639078498293516, "grad_norm": 6.989166259765625, "learning_rate": 2.8784648187633263e-07, "loss": 1.0686, "step": 27 }, { "epoch": 0.008959044368600682, "grad_norm": 7.104398727416992, "learning_rate": 2.9850746268656716e-07, "loss": 1.101, "step": 28 }, { "epoch": 0.009279010238907849, "grad_norm": 6.411932468414307, "learning_rate": 3.0916844349680174e-07, "loss": 1.076, "step": 29 }, { "epoch": 0.009598976109215018, "grad_norm": 5.924315452575684, "learning_rate": 3.1982942430703626e-07, "loss": 1.0632, "step": 30 }, { "epoch": 0.009918941979522184, "grad_norm": 6.04094123840332, "learning_rate": 3.3049040511727084e-07, "loss": 1.0843, "step": 31 }, { "epoch": 0.010238907849829351, "grad_norm": 5.795870304107666, "learning_rate": 3.4115138592750537e-07, "loss": 1.0012, "step": 32 }, { "epoch": 0.010558873720136518, "grad_norm": 5.819528579711914, "learning_rate": 3.518123667377399e-07, "loss": 1.0604, "step": 33 }, { "epoch": 0.010878839590443687, "grad_norm": 5.332779884338379, "learning_rate": 3.624733475479744e-07, "loss": 1.0586, "step": 34 }, { "epoch": 0.011198805460750853, "grad_norm": 5.464365005493164, "learning_rate": 3.7313432835820895e-07, "loss": 1.1297, "step": 35 }, { "epoch": 0.01151877133105802, "grad_norm": 5.652714729309082, "learning_rate": 3.837953091684436e-07, "loss": 1.0346, "step": 36 }, { "epoch": 0.011838737201365187, "grad_norm": 5.79964542388916, "learning_rate": 3.944562899786781e-07, "loss": 1.0312, "step": 37 }, { "epoch": 0.012158703071672355, "grad_norm": 5.450277805328369, "learning_rate": 4.0511727078891263e-07, "loss": 1.0336, "step": 38 }, { "epoch": 0.012478668941979522, "grad_norm": 4.844827175140381, "learning_rate": 4.1577825159914716e-07, "loss": 0.9825, "step": 39 }, { "epoch": 0.012798634812286689, "grad_norm": 3.600152015686035, "learning_rate": 4.264392324093817e-07, "loss": 0.9513, "step": 40 }, { "epoch": 0.013118600682593856, "grad_norm": 3.4587929248809814, "learning_rate": 4.371002132196162e-07, "loss": 0.9135, "step": 41 }, { "epoch": 0.013438566552901024, "grad_norm": 2.950002431869507, "learning_rate": 4.4776119402985074e-07, "loss": 0.9524, "step": 42 }, { "epoch": 0.013758532423208191, "grad_norm": 3.295619249343872, "learning_rate": 4.5842217484008537e-07, "loss": 0.9572, "step": 43 }, { "epoch": 0.014078498293515358, "grad_norm": 2.7801365852355957, "learning_rate": 4.690831556503199e-07, "loss": 0.9389, "step": 44 }, { "epoch": 0.014398464163822526, "grad_norm": 2.845637321472168, "learning_rate": 4.797441364605544e-07, "loss": 0.9185, "step": 45 }, { "epoch": 0.014718430034129693, "grad_norm": 2.6487975120544434, "learning_rate": 4.904051172707889e-07, "loss": 0.9893, "step": 46 }, { "epoch": 0.01503839590443686, "grad_norm": 2.6472787857055664, "learning_rate": 5.010660980810235e-07, "loss": 0.9619, "step": 47 }, { "epoch": 0.015358361774744027, "grad_norm": 2.818723440170288, "learning_rate": 5.11727078891258e-07, "loss": 0.9178, "step": 48 }, { "epoch": 0.015678327645051195, "grad_norm": 2.410865306854248, "learning_rate": 5.223880597014925e-07, "loss": 0.9459, "step": 49 }, { "epoch": 0.01599829351535836, "grad_norm": 2.5202646255493164, "learning_rate": 5.33049040511727e-07, "loss": 0.9107, "step": 50 }, { "epoch": 0.01631825938566553, "grad_norm": 2.329850196838379, "learning_rate": 5.437100213219617e-07, "loss": 0.9128, "step": 51 }, { "epoch": 0.016638225255972697, "grad_norm": 2.3990280628204346, "learning_rate": 5.543710021321962e-07, "loss": 0.9408, "step": 52 }, { "epoch": 0.016958191126279862, "grad_norm": 2.338702440261841, "learning_rate": 5.650319829424307e-07, "loss": 0.9071, "step": 53 }, { "epoch": 0.01727815699658703, "grad_norm": 2.126091957092285, "learning_rate": 5.756929637526653e-07, "loss": 0.9262, "step": 54 }, { "epoch": 0.0175981228668942, "grad_norm": 2.0692224502563477, "learning_rate": 5.863539445628998e-07, "loss": 0.9058, "step": 55 }, { "epoch": 0.017918088737201365, "grad_norm": 2.1477673053741455, "learning_rate": 5.970149253731343e-07, "loss": 0.8858, "step": 56 }, { "epoch": 0.018238054607508533, "grad_norm": 2.15913462638855, "learning_rate": 6.076759061833689e-07, "loss": 0.8942, "step": 57 }, { "epoch": 0.018558020477815698, "grad_norm": 2.243981122970581, "learning_rate": 6.183368869936035e-07, "loss": 0.8919, "step": 58 }, { "epoch": 0.018877986348122867, "grad_norm": 2.0859246253967285, "learning_rate": 6.28997867803838e-07, "loss": 0.9206, "step": 59 }, { "epoch": 0.019197952218430035, "grad_norm": 2.022251844406128, "learning_rate": 6.396588486140725e-07, "loss": 0.8063, "step": 60 }, { "epoch": 0.0195179180887372, "grad_norm": 2.079070806503296, "learning_rate": 6.50319829424307e-07, "loss": 0.8856, "step": 61 }, { "epoch": 0.01983788395904437, "grad_norm": 1.8525385856628418, "learning_rate": 6.609808102345417e-07, "loss": 0.8495, "step": 62 }, { "epoch": 0.020157849829351537, "grad_norm": 1.8756672143936157, "learning_rate": 6.716417910447762e-07, "loss": 0.8234, "step": 63 }, { "epoch": 0.020477815699658702, "grad_norm": 1.7776811122894287, "learning_rate": 6.823027718550107e-07, "loss": 0.899, "step": 64 }, { "epoch": 0.02079778156996587, "grad_norm": 1.680071234703064, "learning_rate": 6.929637526652453e-07, "loss": 0.9274, "step": 65 }, { "epoch": 0.021117747440273036, "grad_norm": 1.493758201599121, "learning_rate": 7.036247334754798e-07, "loss": 0.8671, "step": 66 }, { "epoch": 0.021437713310580205, "grad_norm": 1.4272754192352295, "learning_rate": 7.142857142857143e-07, "loss": 0.8085, "step": 67 }, { "epoch": 0.021757679180887373, "grad_norm": 1.404809594154358, "learning_rate": 7.249466950959488e-07, "loss": 0.8463, "step": 68 }, { "epoch": 0.022077645051194538, "grad_norm": 1.2687548398971558, "learning_rate": 7.356076759061834e-07, "loss": 0.8656, "step": 69 }, { "epoch": 0.022397610921501707, "grad_norm": 1.2395495176315308, "learning_rate": 7.462686567164179e-07, "loss": 0.8554, "step": 70 }, { "epoch": 0.022717576791808875, "grad_norm": 1.2378782033920288, "learning_rate": 7.569296375266526e-07, "loss": 0.848, "step": 71 }, { "epoch": 0.02303754266211604, "grad_norm": 1.4011038541793823, "learning_rate": 7.675906183368872e-07, "loss": 0.826, "step": 72 }, { "epoch": 0.02335750853242321, "grad_norm": 1.261948585510254, "learning_rate": 7.782515991471217e-07, "loss": 0.842, "step": 73 }, { "epoch": 0.023677474402730374, "grad_norm": 1.0836261510849, "learning_rate": 7.889125799573562e-07, "loss": 0.8043, "step": 74 }, { "epoch": 0.023997440273037542, "grad_norm": 1.2844727039337158, "learning_rate": 7.995735607675907e-07, "loss": 0.8443, "step": 75 }, { "epoch": 0.02431740614334471, "grad_norm": 1.0323970317840576, "learning_rate": 8.102345415778253e-07, "loss": 0.7497, "step": 76 }, { "epoch": 0.024637372013651876, "grad_norm": 1.0269299745559692, "learning_rate": 8.208955223880598e-07, "loss": 0.7616, "step": 77 }, { "epoch": 0.024957337883959044, "grad_norm": 1.0544461011886597, "learning_rate": 8.315565031982943e-07, "loss": 0.7697, "step": 78 }, { "epoch": 0.025277303754266213, "grad_norm": 1.0795966386795044, "learning_rate": 8.422174840085288e-07, "loss": 0.768, "step": 79 }, { "epoch": 0.025597269624573378, "grad_norm": 0.9690151214599609, "learning_rate": 8.528784648187634e-07, "loss": 0.7908, "step": 80 }, { "epoch": 0.025917235494880547, "grad_norm": 1.0053969621658325, "learning_rate": 8.635394456289979e-07, "loss": 0.8469, "step": 81 }, { "epoch": 0.02623720136518771, "grad_norm": 0.928873598575592, "learning_rate": 8.742004264392324e-07, "loss": 0.76, "step": 82 }, { "epoch": 0.02655716723549488, "grad_norm": 0.9888681769371033, "learning_rate": 8.848614072494669e-07, "loss": 0.7964, "step": 83 }, { "epoch": 0.02687713310580205, "grad_norm": 0.8391404747962952, "learning_rate": 8.955223880597015e-07, "loss": 0.6921, "step": 84 }, { "epoch": 0.027197098976109214, "grad_norm": 0.9005293846130371, "learning_rate": 9.06183368869936e-07, "loss": 0.7879, "step": 85 }, { "epoch": 0.027517064846416382, "grad_norm": 0.8991631269454956, "learning_rate": 9.168443496801707e-07, "loss": 0.7402, "step": 86 }, { "epoch": 0.02783703071672355, "grad_norm": 0.8436546921730042, "learning_rate": 9.275053304904053e-07, "loss": 0.7411, "step": 87 }, { "epoch": 0.028156996587030716, "grad_norm": 0.8756288290023804, "learning_rate": 9.381663113006398e-07, "loss": 0.7711, "step": 88 }, { "epoch": 0.028476962457337884, "grad_norm": 0.7887997031211853, "learning_rate": 9.488272921108743e-07, "loss": 0.7588, "step": 89 }, { "epoch": 0.028796928327645053, "grad_norm": 0.8141859769821167, "learning_rate": 9.594882729211088e-07, "loss": 0.7743, "step": 90 }, { "epoch": 0.029116894197952218, "grad_norm": 0.8312386274337769, "learning_rate": 9.701492537313434e-07, "loss": 0.7418, "step": 91 }, { "epoch": 0.029436860068259386, "grad_norm": 0.8288974165916443, "learning_rate": 9.808102345415779e-07, "loss": 0.7381, "step": 92 }, { "epoch": 0.02975682593856655, "grad_norm": 0.8249732255935669, "learning_rate": 9.914712153518124e-07, "loss": 0.7541, "step": 93 }, { "epoch": 0.03007679180887372, "grad_norm": 0.7712000608444214, "learning_rate": 1.002132196162047e-06, "loss": 0.7684, "step": 94 }, { "epoch": 0.03039675767918089, "grad_norm": 0.7715640664100647, "learning_rate": 1.0127931769722815e-06, "loss": 0.8098, "step": 95 }, { "epoch": 0.030716723549488054, "grad_norm": 0.7638954520225525, "learning_rate": 1.023454157782516e-06, "loss": 0.7735, "step": 96 }, { "epoch": 0.031036689419795222, "grad_norm": 0.7430364489555359, "learning_rate": 1.0341151385927505e-06, "loss": 0.745, "step": 97 }, { "epoch": 0.03135665529010239, "grad_norm": 0.7673352956771851, "learning_rate": 1.044776119402985e-06, "loss": 0.756, "step": 98 }, { "epoch": 0.031676621160409556, "grad_norm": 0.8949801921844482, "learning_rate": 1.0554371002132196e-06, "loss": 0.7666, "step": 99 }, { "epoch": 0.03199658703071672, "grad_norm": 0.8077754974365234, "learning_rate": 1.066098081023454e-06, "loss": 0.7326, "step": 100 }, { "epoch": 0.03231655290102389, "grad_norm": 0.7525066137313843, "learning_rate": 1.0767590618336886e-06, "loss": 0.7114, "step": 101 }, { "epoch": 0.03263651877133106, "grad_norm": 0.7346336841583252, "learning_rate": 1.0874200426439234e-06, "loss": 0.7588, "step": 102 }, { "epoch": 0.03295648464163822, "grad_norm": 0.8265507221221924, "learning_rate": 1.0980810234541579e-06, "loss": 0.7152, "step": 103 }, { "epoch": 0.033276450511945395, "grad_norm": 0.7094560861587524, "learning_rate": 1.1087420042643924e-06, "loss": 0.715, "step": 104 }, { "epoch": 0.03359641638225256, "grad_norm": 0.7683343887329102, "learning_rate": 1.119402985074627e-06, "loss": 0.7334, "step": 105 }, { "epoch": 0.033916382252559725, "grad_norm": 0.7544600963592529, "learning_rate": 1.1300639658848615e-06, "loss": 0.7015, "step": 106 }, { "epoch": 0.0342363481228669, "grad_norm": 0.7648451924324036, "learning_rate": 1.140724946695096e-06, "loss": 0.7146, "step": 107 }, { "epoch": 0.03455631399317406, "grad_norm": 0.7159047722816467, "learning_rate": 1.1513859275053305e-06, "loss": 0.6871, "step": 108 }, { "epoch": 0.03487627986348123, "grad_norm": 0.6753017902374268, "learning_rate": 1.162046908315565e-06, "loss": 0.6809, "step": 109 }, { "epoch": 0.0351962457337884, "grad_norm": 0.6882673501968384, "learning_rate": 1.1727078891257996e-06, "loss": 0.7309, "step": 110 }, { "epoch": 0.035516211604095564, "grad_norm": 0.6664094924926758, "learning_rate": 1.183368869936034e-06, "loss": 0.7106, "step": 111 }, { "epoch": 0.03583617747440273, "grad_norm": 0.7424576878547668, "learning_rate": 1.1940298507462686e-06, "loss": 0.7242, "step": 112 }, { "epoch": 0.036156143344709894, "grad_norm": 0.7396127581596375, "learning_rate": 1.2046908315565034e-06, "loss": 0.719, "step": 113 }, { "epoch": 0.036476109215017066, "grad_norm": 0.6853981018066406, "learning_rate": 1.2153518123667379e-06, "loss": 0.6913, "step": 114 }, { "epoch": 0.03679607508532423, "grad_norm": 0.7118975520133972, "learning_rate": 1.2260127931769724e-06, "loss": 0.7491, "step": 115 }, { "epoch": 0.037116040955631396, "grad_norm": 0.6694145798683167, "learning_rate": 1.236673773987207e-06, "loss": 0.6652, "step": 116 }, { "epoch": 0.03743600682593857, "grad_norm": 0.6602712273597717, "learning_rate": 1.2473347547974415e-06, "loss": 0.7452, "step": 117 }, { "epoch": 0.03775597269624573, "grad_norm": 0.6255103349685669, "learning_rate": 1.257995735607676e-06, "loss": 0.7292, "step": 118 }, { "epoch": 0.0380759385665529, "grad_norm": 0.738187313079834, "learning_rate": 1.2686567164179105e-06, "loss": 0.7191, "step": 119 }, { "epoch": 0.03839590443686007, "grad_norm": 0.6636806130409241, "learning_rate": 1.279317697228145e-06, "loss": 0.6932, "step": 120 }, { "epoch": 0.038715870307167236, "grad_norm": 0.6825982332229614, "learning_rate": 1.2899786780383796e-06, "loss": 0.7343, "step": 121 }, { "epoch": 0.0390358361774744, "grad_norm": 0.740442156791687, "learning_rate": 1.300639658848614e-06, "loss": 0.7303, "step": 122 }, { "epoch": 0.03935580204778157, "grad_norm": 0.5918875336647034, "learning_rate": 1.3113006396588488e-06, "loss": 0.7187, "step": 123 }, { "epoch": 0.03967576791808874, "grad_norm": 0.7803345918655396, "learning_rate": 1.3219616204690834e-06, "loss": 0.8031, "step": 124 }, { "epoch": 0.0399957337883959, "grad_norm": 0.6779199838638306, "learning_rate": 1.3326226012793179e-06, "loss": 0.7096, "step": 125 }, { "epoch": 0.040315699658703075, "grad_norm": 0.6925889253616333, "learning_rate": 1.3432835820895524e-06, "loss": 0.6824, "step": 126 }, { "epoch": 0.04063566552901024, "grad_norm": 0.7514821290969849, "learning_rate": 1.353944562899787e-06, "loss": 0.7371, "step": 127 }, { "epoch": 0.040955631399317405, "grad_norm": 0.6688573360443115, "learning_rate": 1.3646055437100215e-06, "loss": 0.7117, "step": 128 }, { "epoch": 0.04127559726962458, "grad_norm": 0.6557217240333557, "learning_rate": 1.375266524520256e-06, "loss": 0.6998, "step": 129 }, { "epoch": 0.04159556313993174, "grad_norm": 0.7434080243110657, "learning_rate": 1.3859275053304905e-06, "loss": 0.7321, "step": 130 }, { "epoch": 0.04191552901023891, "grad_norm": 0.669056236743927, "learning_rate": 1.396588486140725e-06, "loss": 0.7098, "step": 131 }, { "epoch": 0.04223549488054607, "grad_norm": 0.5930835604667664, "learning_rate": 1.4072494669509596e-06, "loss": 0.6943, "step": 132 }, { "epoch": 0.042555460750853244, "grad_norm": 0.6236996054649353, "learning_rate": 1.417910447761194e-06, "loss": 0.6928, "step": 133 }, { "epoch": 0.04287542662116041, "grad_norm": 0.6656635403633118, "learning_rate": 1.4285714285714286e-06, "loss": 0.6943, "step": 134 }, { "epoch": 0.043195392491467574, "grad_norm": 0.7729142308235168, "learning_rate": 1.4392324093816632e-06, "loss": 0.7283, "step": 135 }, { "epoch": 0.043515358361774746, "grad_norm": 0.6872455477714539, "learning_rate": 1.4498933901918977e-06, "loss": 0.706, "step": 136 }, { "epoch": 0.04383532423208191, "grad_norm": 0.6097860336303711, "learning_rate": 1.4605543710021322e-06, "loss": 0.6711, "step": 137 }, { "epoch": 0.044155290102389076, "grad_norm": 0.7215981483459473, "learning_rate": 1.4712153518123667e-06, "loss": 0.75, "step": 138 }, { "epoch": 0.04447525597269625, "grad_norm": 0.6652256846427917, "learning_rate": 1.4818763326226013e-06, "loss": 0.7266, "step": 139 }, { "epoch": 0.04479522184300341, "grad_norm": 0.7119171619415283, "learning_rate": 1.4925373134328358e-06, "loss": 0.6318, "step": 140 }, { "epoch": 0.04511518771331058, "grad_norm": 0.7838914394378662, "learning_rate": 1.5031982942430705e-06, "loss": 0.6842, "step": 141 }, { "epoch": 0.04543515358361775, "grad_norm": 0.7664074897766113, "learning_rate": 1.5138592750533053e-06, "loss": 0.7572, "step": 142 }, { "epoch": 0.045755119453924915, "grad_norm": 0.6630170941352844, "learning_rate": 1.5245202558635398e-06, "loss": 0.6776, "step": 143 }, { "epoch": 0.04607508532423208, "grad_norm": 0.6675658822059631, "learning_rate": 1.5351812366737743e-06, "loss": 0.7035, "step": 144 }, { "epoch": 0.04639505119453925, "grad_norm": 0.6449645161628723, "learning_rate": 1.5458422174840088e-06, "loss": 0.7138, "step": 145 }, { "epoch": 0.04671501706484642, "grad_norm": 0.6699844002723694, "learning_rate": 1.5565031982942434e-06, "loss": 0.6748, "step": 146 }, { "epoch": 0.04703498293515358, "grad_norm": 0.6595117449760437, "learning_rate": 1.5671641791044779e-06, "loss": 0.6752, "step": 147 }, { "epoch": 0.04735494880546075, "grad_norm": 0.6826941967010498, "learning_rate": 1.5778251599147124e-06, "loss": 0.7394, "step": 148 }, { "epoch": 0.04767491467576792, "grad_norm": 0.7005380392074585, "learning_rate": 1.588486140724947e-06, "loss": 0.6968, "step": 149 }, { "epoch": 0.047994880546075085, "grad_norm": 0.608120322227478, "learning_rate": 1.5991471215351815e-06, "loss": 0.6177, "step": 150 }, { "epoch": 0.04831484641638225, "grad_norm": 0.7013968825340271, "learning_rate": 1.609808102345416e-06, "loss": 0.6854, "step": 151 }, { "epoch": 0.04863481228668942, "grad_norm": 0.6775135397911072, "learning_rate": 1.6204690831556505e-06, "loss": 0.7177, "step": 152 }, { "epoch": 0.04895477815699659, "grad_norm": 0.7195072174072266, "learning_rate": 1.631130063965885e-06, "loss": 0.6894, "step": 153 }, { "epoch": 0.04927474402730375, "grad_norm": 0.7061877846717834, "learning_rate": 1.6417910447761196e-06, "loss": 0.7465, "step": 154 }, { "epoch": 0.049594709897610924, "grad_norm": 0.6255117058753967, "learning_rate": 1.652452025586354e-06, "loss": 0.6758, "step": 155 }, { "epoch": 0.04991467576791809, "grad_norm": 0.7507020235061646, "learning_rate": 1.6631130063965886e-06, "loss": 0.7578, "step": 156 }, { "epoch": 0.050234641638225254, "grad_norm": 0.6676733493804932, "learning_rate": 1.6737739872068232e-06, "loss": 0.6807, "step": 157 }, { "epoch": 0.050554607508532426, "grad_norm": 0.6160469651222229, "learning_rate": 1.6844349680170577e-06, "loss": 0.662, "step": 158 }, { "epoch": 0.05087457337883959, "grad_norm": 0.6552683711051941, "learning_rate": 1.6950959488272922e-06, "loss": 0.6405, "step": 159 }, { "epoch": 0.051194539249146756, "grad_norm": 0.6595558524131775, "learning_rate": 1.7057569296375267e-06, "loss": 0.6813, "step": 160 }, { "epoch": 0.05151450511945393, "grad_norm": 0.6022697687149048, "learning_rate": 1.7164179104477613e-06, "loss": 0.6718, "step": 161 }, { "epoch": 0.05183447098976109, "grad_norm": 0.631741464138031, "learning_rate": 1.7270788912579958e-06, "loss": 0.6571, "step": 162 }, { "epoch": 0.05215443686006826, "grad_norm": 0.712632417678833, "learning_rate": 1.7377398720682303e-06, "loss": 0.6367, "step": 163 }, { "epoch": 0.05247440273037542, "grad_norm": 0.691933274269104, "learning_rate": 1.7484008528784648e-06, "loss": 0.6726, "step": 164 }, { "epoch": 0.052794368600682595, "grad_norm": 0.648144543170929, "learning_rate": 1.7590618336886994e-06, "loss": 0.7117, "step": 165 }, { "epoch": 0.05311433447098976, "grad_norm": 0.7375291585922241, "learning_rate": 1.7697228144989339e-06, "loss": 0.7251, "step": 166 }, { "epoch": 0.053434300341296925, "grad_norm": 0.6844941973686218, "learning_rate": 1.7803837953091684e-06, "loss": 0.6893, "step": 167 }, { "epoch": 0.0537542662116041, "grad_norm": 0.7131636142730713, "learning_rate": 1.791044776119403e-06, "loss": 0.7042, "step": 168 }, { "epoch": 0.05407423208191126, "grad_norm": 0.7594255208969116, "learning_rate": 1.8017057569296375e-06, "loss": 0.7, "step": 169 }, { "epoch": 0.05439419795221843, "grad_norm": 0.7435464859008789, "learning_rate": 1.812366737739872e-06, "loss": 0.697, "step": 170 }, { "epoch": 0.0547141638225256, "grad_norm": 0.7214502692222595, "learning_rate": 1.8230277185501067e-06, "loss": 0.7183, "step": 171 }, { "epoch": 0.055034129692832764, "grad_norm": 0.6937075257301331, "learning_rate": 1.8336886993603415e-06, "loss": 0.6498, "step": 172 }, { "epoch": 0.05535409556313993, "grad_norm": 0.6126171350479126, "learning_rate": 1.844349680170576e-06, "loss": 0.6458, "step": 173 }, { "epoch": 0.0556740614334471, "grad_norm": 0.697283148765564, "learning_rate": 1.8550106609808105e-06, "loss": 0.6651, "step": 174 }, { "epoch": 0.05599402730375427, "grad_norm": 0.6571756601333618, "learning_rate": 1.865671641791045e-06, "loss": 0.6113, "step": 175 }, { "epoch": 0.05631399317406143, "grad_norm": 0.658406674861908, "learning_rate": 1.8763326226012796e-06, "loss": 0.6706, "step": 176 }, { "epoch": 0.056633959044368604, "grad_norm": 0.7434971332550049, "learning_rate": 1.886993603411514e-06, "loss": 0.6733, "step": 177 }, { "epoch": 0.05695392491467577, "grad_norm": 0.6445094347000122, "learning_rate": 1.8976545842217486e-06, "loss": 0.6544, "step": 178 }, { "epoch": 0.057273890784982934, "grad_norm": 0.6629904508590698, "learning_rate": 1.908315565031983e-06, "loss": 0.7276, "step": 179 }, { "epoch": 0.057593856655290106, "grad_norm": 0.699326753616333, "learning_rate": 1.9189765458422177e-06, "loss": 0.6755, "step": 180 }, { "epoch": 0.05791382252559727, "grad_norm": 0.8005274534225464, "learning_rate": 1.929637526652452e-06, "loss": 0.6551, "step": 181 }, { "epoch": 0.058233788395904436, "grad_norm": 0.7194087505340576, "learning_rate": 1.9402985074626867e-06, "loss": 0.6459, "step": 182 }, { "epoch": 0.0585537542662116, "grad_norm": 0.7291178703308105, "learning_rate": 1.9509594882729213e-06, "loss": 0.7131, "step": 183 }, { "epoch": 0.05887372013651877, "grad_norm": 0.6386391520500183, "learning_rate": 1.9616204690831558e-06, "loss": 0.6826, "step": 184 }, { "epoch": 0.05919368600682594, "grad_norm": 0.6472675800323486, "learning_rate": 1.9722814498933903e-06, "loss": 0.6299, "step": 185 }, { "epoch": 0.0595136518771331, "grad_norm": 0.6684521436691284, "learning_rate": 1.982942430703625e-06, "loss": 0.6535, "step": 186 }, { "epoch": 0.059833617747440275, "grad_norm": 0.6821104884147644, "learning_rate": 1.9936034115138594e-06, "loss": 0.6948, "step": 187 }, { "epoch": 0.06015358361774744, "grad_norm": 0.6314494013786316, "learning_rate": 2.004264392324094e-06, "loss": 0.664, "step": 188 }, { "epoch": 0.060473549488054605, "grad_norm": 0.7831369638442993, "learning_rate": 2.0149253731343284e-06, "loss": 0.7075, "step": 189 }, { "epoch": 0.06079351535836178, "grad_norm": 0.6174335479736328, "learning_rate": 2.025586353944563e-06, "loss": 0.6666, "step": 190 }, { "epoch": 0.06111348122866894, "grad_norm": 0.6839204430580139, "learning_rate": 2.0362473347547975e-06, "loss": 0.6791, "step": 191 }, { "epoch": 0.06143344709897611, "grad_norm": 0.6346204280853271, "learning_rate": 2.046908315565032e-06, "loss": 0.6579, "step": 192 }, { "epoch": 0.06175341296928328, "grad_norm": 0.7199333906173706, "learning_rate": 2.0575692963752665e-06, "loss": 0.6369, "step": 193 }, { "epoch": 0.062073378839590444, "grad_norm": 0.6710346937179565, "learning_rate": 2.068230277185501e-06, "loss": 0.6483, "step": 194 }, { "epoch": 0.06239334470989761, "grad_norm": 0.7106640338897705, "learning_rate": 2.0788912579957356e-06, "loss": 0.662, "step": 195 }, { "epoch": 0.06271331058020478, "grad_norm": 0.8064489364624023, "learning_rate": 2.08955223880597e-06, "loss": 0.6897, "step": 196 }, { "epoch": 0.06303327645051195, "grad_norm": 0.7295928597450256, "learning_rate": 2.1002132196162046e-06, "loss": 0.6644, "step": 197 }, { "epoch": 0.06335324232081911, "grad_norm": 0.7068210244178772, "learning_rate": 2.110874200426439e-06, "loss": 0.6854, "step": 198 }, { "epoch": 0.06367320819112628, "grad_norm": 0.7519869208335876, "learning_rate": 2.1215351812366737e-06, "loss": 0.6667, "step": 199 }, { "epoch": 0.06399317406143344, "grad_norm": 0.7251495122909546, "learning_rate": 2.132196162046908e-06, "loss": 0.6764, "step": 200 }, { "epoch": 0.06431313993174062, "grad_norm": 0.5924597382545471, "learning_rate": 2.1428571428571427e-06, "loss": 0.6447, "step": 201 }, { "epoch": 0.06463310580204779, "grad_norm": 0.7425607442855835, "learning_rate": 2.1535181236673773e-06, "loss": 0.6867, "step": 202 }, { "epoch": 0.06495307167235495, "grad_norm": 0.662427544593811, "learning_rate": 2.1641791044776118e-06, "loss": 0.6814, "step": 203 }, { "epoch": 0.06527303754266212, "grad_norm": 0.7188402414321899, "learning_rate": 2.1748400852878467e-06, "loss": 0.7082, "step": 204 }, { "epoch": 0.06559300341296928, "grad_norm": 0.6829525828361511, "learning_rate": 2.1855010660980813e-06, "loss": 0.6382, "step": 205 }, { "epoch": 0.06591296928327645, "grad_norm": 0.713671088218689, "learning_rate": 2.1961620469083158e-06, "loss": 0.6556, "step": 206 }, { "epoch": 0.06623293515358361, "grad_norm": 0.7159529328346252, "learning_rate": 2.2068230277185503e-06, "loss": 0.6594, "step": 207 }, { "epoch": 0.06655290102389079, "grad_norm": 0.856481671333313, "learning_rate": 2.217484008528785e-06, "loss": 0.6609, "step": 208 }, { "epoch": 0.06687286689419795, "grad_norm": 0.6801454424858093, "learning_rate": 2.2281449893390194e-06, "loss": 0.619, "step": 209 }, { "epoch": 0.06719283276450512, "grad_norm": 0.7858098149299622, "learning_rate": 2.238805970149254e-06, "loss": 0.6505, "step": 210 }, { "epoch": 0.06751279863481228, "grad_norm": 0.6824889183044434, "learning_rate": 2.2494669509594884e-06, "loss": 0.6682, "step": 211 }, { "epoch": 0.06783276450511945, "grad_norm": 0.691604733467102, "learning_rate": 2.260127931769723e-06, "loss": 0.7058, "step": 212 }, { "epoch": 0.06815273037542662, "grad_norm": 0.732110321521759, "learning_rate": 2.2707889125799575e-06, "loss": 0.6431, "step": 213 }, { "epoch": 0.0684726962457338, "grad_norm": 0.6957806348800659, "learning_rate": 2.281449893390192e-06, "loss": 0.6827, "step": 214 }, { "epoch": 0.06879266211604096, "grad_norm": 0.7264412641525269, "learning_rate": 2.2921108742004265e-06, "loss": 0.6411, "step": 215 }, { "epoch": 0.06911262798634812, "grad_norm": 0.6580216884613037, "learning_rate": 2.302771855010661e-06, "loss": 0.6748, "step": 216 }, { "epoch": 0.06943259385665529, "grad_norm": 0.7710157036781311, "learning_rate": 2.3134328358208956e-06, "loss": 0.6895, "step": 217 }, { "epoch": 0.06975255972696245, "grad_norm": 0.769323468208313, "learning_rate": 2.32409381663113e-06, "loss": 0.6593, "step": 218 }, { "epoch": 0.07007252559726962, "grad_norm": 0.6831722259521484, "learning_rate": 2.3347547974413646e-06, "loss": 0.6512, "step": 219 }, { "epoch": 0.0703924914675768, "grad_norm": 0.7421032786369324, "learning_rate": 2.345415778251599e-06, "loss": 0.6667, "step": 220 }, { "epoch": 0.07071245733788396, "grad_norm": 0.8579223155975342, "learning_rate": 2.3560767590618337e-06, "loss": 0.6856, "step": 221 }, { "epoch": 0.07103242320819113, "grad_norm": 0.7443462610244751, "learning_rate": 2.366737739872068e-06, "loss": 0.7142, "step": 222 }, { "epoch": 0.0713523890784983, "grad_norm": 0.6016785502433777, "learning_rate": 2.3773987206823027e-06, "loss": 0.6414, "step": 223 }, { "epoch": 0.07167235494880546, "grad_norm": 0.6656649708747864, "learning_rate": 2.3880597014925373e-06, "loss": 0.6648, "step": 224 }, { "epoch": 0.07199232081911262, "grad_norm": 0.6443098187446594, "learning_rate": 2.398720682302772e-06, "loss": 0.637, "step": 225 }, { "epoch": 0.07231228668941979, "grad_norm": 0.6778454780578613, "learning_rate": 2.4093816631130067e-06, "loss": 0.6501, "step": 226 }, { "epoch": 0.07263225255972697, "grad_norm": 0.6492361426353455, "learning_rate": 2.4200426439232413e-06, "loss": 0.7013, "step": 227 }, { "epoch": 0.07295221843003413, "grad_norm": 0.8034964203834534, "learning_rate": 2.4307036247334758e-06, "loss": 0.7234, "step": 228 }, { "epoch": 0.0732721843003413, "grad_norm": 0.6665212512016296, "learning_rate": 2.4413646055437103e-06, "loss": 0.6744, "step": 229 }, { "epoch": 0.07359215017064846, "grad_norm": 0.6973935961723328, "learning_rate": 2.452025586353945e-06, "loss": 0.6579, "step": 230 }, { "epoch": 0.07391211604095563, "grad_norm": 0.698112964630127, "learning_rate": 2.4626865671641794e-06, "loss": 0.6221, "step": 231 }, { "epoch": 0.07423208191126279, "grad_norm": 0.8335263133049011, "learning_rate": 2.473347547974414e-06, "loss": 0.6552, "step": 232 }, { "epoch": 0.07455204778156997, "grad_norm": 0.6335951685905457, "learning_rate": 2.4840085287846484e-06, "loss": 0.6366, "step": 233 }, { "epoch": 0.07487201365187714, "grad_norm": 0.7226898074150085, "learning_rate": 2.494669509594883e-06, "loss": 0.6684, "step": 234 }, { "epoch": 0.0751919795221843, "grad_norm": 0.7174168825149536, "learning_rate": 2.5053304904051175e-06, "loss": 0.6656, "step": 235 }, { "epoch": 0.07551194539249147, "grad_norm": 0.7624191045761108, "learning_rate": 2.515991471215352e-06, "loss": 0.6451, "step": 236 }, { "epoch": 0.07583191126279863, "grad_norm": 0.7072004675865173, "learning_rate": 2.5266524520255865e-06, "loss": 0.6608, "step": 237 }, { "epoch": 0.0761518771331058, "grad_norm": 0.6365002989768982, "learning_rate": 2.537313432835821e-06, "loss": 0.6563, "step": 238 }, { "epoch": 0.07647184300341298, "grad_norm": 0.7351357340812683, "learning_rate": 2.5479744136460556e-06, "loss": 0.6749, "step": 239 }, { "epoch": 0.07679180887372014, "grad_norm": 0.6643660068511963, "learning_rate": 2.55863539445629e-06, "loss": 0.621, "step": 240 }, { "epoch": 0.0771117747440273, "grad_norm": 0.767565131187439, "learning_rate": 2.5692963752665246e-06, "loss": 0.6579, "step": 241 }, { "epoch": 0.07743174061433447, "grad_norm": 0.6990678906440735, "learning_rate": 2.579957356076759e-06, "loss": 0.6516, "step": 242 }, { "epoch": 0.07775170648464164, "grad_norm": 0.6894639134407043, "learning_rate": 2.5906183368869937e-06, "loss": 0.6645, "step": 243 }, { "epoch": 0.0780716723549488, "grad_norm": 0.7961663007736206, "learning_rate": 2.601279317697228e-06, "loss": 0.6722, "step": 244 }, { "epoch": 0.07839163822525597, "grad_norm": 0.6465956568717957, "learning_rate": 2.6119402985074627e-06, "loss": 0.6401, "step": 245 }, { "epoch": 0.07871160409556315, "grad_norm": 0.6665163040161133, "learning_rate": 2.6226012793176977e-06, "loss": 0.6665, "step": 246 }, { "epoch": 0.07903156996587031, "grad_norm": 0.6720198392868042, "learning_rate": 2.6332622601279318e-06, "loss": 0.6444, "step": 247 }, { "epoch": 0.07935153583617748, "grad_norm": 0.6681743860244751, "learning_rate": 2.6439232409381667e-06, "loss": 0.6919, "step": 248 }, { "epoch": 0.07967150170648464, "grad_norm": 0.7217524647712708, "learning_rate": 2.654584221748401e-06, "loss": 0.6661, "step": 249 }, { "epoch": 0.0799914675767918, "grad_norm": 0.6475330591201782, "learning_rate": 2.6652452025586358e-06, "loss": 0.6383, "step": 250 }, { "epoch": 0.08031143344709897, "grad_norm": 0.6881884336471558, "learning_rate": 2.6759061833688703e-06, "loss": 0.6602, "step": 251 }, { "epoch": 0.08063139931740615, "grad_norm": 0.6619604825973511, "learning_rate": 2.686567164179105e-06, "loss": 0.6399, "step": 252 }, { "epoch": 0.08095136518771331, "grad_norm": 0.6265876889228821, "learning_rate": 2.6972281449893394e-06, "loss": 0.5992, "step": 253 }, { "epoch": 0.08127133105802048, "grad_norm": 0.675489604473114, "learning_rate": 2.707889125799574e-06, "loss": 0.6298, "step": 254 }, { "epoch": 0.08159129692832764, "grad_norm": 0.6868592500686646, "learning_rate": 2.7185501066098084e-06, "loss": 0.6043, "step": 255 }, { "epoch": 0.08191126279863481, "grad_norm": 0.6755011677742004, "learning_rate": 2.729211087420043e-06, "loss": 0.6325, "step": 256 }, { "epoch": 0.08223122866894197, "grad_norm": 0.680072009563446, "learning_rate": 2.7398720682302775e-06, "loss": 0.6401, "step": 257 }, { "epoch": 0.08255119453924915, "grad_norm": 0.6612681746482849, "learning_rate": 2.750533049040512e-06, "loss": 0.6355, "step": 258 }, { "epoch": 0.08287116040955632, "grad_norm": 0.7801066637039185, "learning_rate": 2.7611940298507465e-06, "loss": 0.6424, "step": 259 }, { "epoch": 0.08319112627986348, "grad_norm": 0.663550615310669, "learning_rate": 2.771855010660981e-06, "loss": 0.6217, "step": 260 }, { "epoch": 0.08351109215017065, "grad_norm": 0.641058623790741, "learning_rate": 2.7825159914712156e-06, "loss": 0.6314, "step": 261 }, { "epoch": 0.08383105802047781, "grad_norm": 0.8522260189056396, "learning_rate": 2.79317697228145e-06, "loss": 0.6258, "step": 262 }, { "epoch": 0.08415102389078498, "grad_norm": 0.6692214012145996, "learning_rate": 2.8038379530916846e-06, "loss": 0.6255, "step": 263 }, { "epoch": 0.08447098976109214, "grad_norm": 0.6521351933479309, "learning_rate": 2.814498933901919e-06, "loss": 0.6366, "step": 264 }, { "epoch": 0.08479095563139932, "grad_norm": 0.6812143921852112, "learning_rate": 2.825159914712154e-06, "loss": 0.6331, "step": 265 }, { "epoch": 0.08511092150170649, "grad_norm": 0.6537390351295471, "learning_rate": 2.835820895522388e-06, "loss": 0.667, "step": 266 }, { "epoch": 0.08543088737201365, "grad_norm": 0.6082245111465454, "learning_rate": 2.846481876332623e-06, "loss": 0.6321, "step": 267 }, { "epoch": 0.08575085324232082, "grad_norm": 0.6908940672874451, "learning_rate": 2.8571428571428573e-06, "loss": 0.6164, "step": 268 }, { "epoch": 0.08607081911262798, "grad_norm": 0.7633948922157288, "learning_rate": 2.867803837953092e-06, "loss": 0.6714, "step": 269 }, { "epoch": 0.08639078498293515, "grad_norm": 0.6619597673416138, "learning_rate": 2.8784648187633263e-06, "loss": 0.6453, "step": 270 }, { "epoch": 0.08671075085324233, "grad_norm": 0.7117516398429871, "learning_rate": 2.8891257995735613e-06, "loss": 0.6731, "step": 271 }, { "epoch": 0.08703071672354949, "grad_norm": 0.7704693078994751, "learning_rate": 2.8997867803837954e-06, "loss": 0.646, "step": 272 }, { "epoch": 0.08735068259385666, "grad_norm": 0.6669108867645264, "learning_rate": 2.9104477611940303e-06, "loss": 0.6283, "step": 273 }, { "epoch": 0.08767064846416382, "grad_norm": 0.7196122407913208, "learning_rate": 2.9211087420042644e-06, "loss": 0.6208, "step": 274 }, { "epoch": 0.08799061433447099, "grad_norm": 0.7552131414413452, "learning_rate": 2.9317697228144994e-06, "loss": 0.6619, "step": 275 }, { "epoch": 0.08831058020477815, "grad_norm": 0.6823918223381042, "learning_rate": 2.9424307036247335e-06, "loss": 0.6141, "step": 276 }, { "epoch": 0.08863054607508532, "grad_norm": 0.7164825797080994, "learning_rate": 2.9530916844349684e-06, "loss": 0.6555, "step": 277 }, { "epoch": 0.0889505119453925, "grad_norm": 0.6870175004005432, "learning_rate": 2.9637526652452025e-06, "loss": 0.6142, "step": 278 }, { "epoch": 0.08927047781569966, "grad_norm": 0.7144311666488647, "learning_rate": 2.9744136460554375e-06, "loss": 0.6431, "step": 279 }, { "epoch": 0.08959044368600683, "grad_norm": 0.6681182384490967, "learning_rate": 2.9850746268656716e-06, "loss": 0.682, "step": 280 }, { "epoch": 0.08991040955631399, "grad_norm": 0.6359437704086304, "learning_rate": 2.9957356076759065e-06, "loss": 0.6527, "step": 281 }, { "epoch": 0.09023037542662116, "grad_norm": 0.718871533870697, "learning_rate": 3.006396588486141e-06, "loss": 0.6613, "step": 282 }, { "epoch": 0.09055034129692832, "grad_norm": 0.7304879426956177, "learning_rate": 3.0170575692963756e-06, "loss": 0.6771, "step": 283 }, { "epoch": 0.0908703071672355, "grad_norm": 0.6561257839202881, "learning_rate": 3.0277185501066105e-06, "loss": 0.5928, "step": 284 }, { "epoch": 0.09119027303754267, "grad_norm": 0.7376087307929993, "learning_rate": 3.0383795309168446e-06, "loss": 0.6923, "step": 285 }, { "epoch": 0.09151023890784983, "grad_norm": 0.6895893216133118, "learning_rate": 3.0490405117270796e-06, "loss": 0.6311, "step": 286 }, { "epoch": 0.091830204778157, "grad_norm": 0.7078286409378052, "learning_rate": 3.0597014925373137e-06, "loss": 0.645, "step": 287 }, { "epoch": 0.09215017064846416, "grad_norm": 0.8378876447677612, "learning_rate": 3.0703624733475486e-06, "loss": 0.6613, "step": 288 }, { "epoch": 0.09247013651877133, "grad_norm": 0.7053190469741821, "learning_rate": 3.0810234541577827e-06, "loss": 0.5924, "step": 289 }, { "epoch": 0.0927901023890785, "grad_norm": 0.6344333291053772, "learning_rate": 3.0916844349680177e-06, "loss": 0.6003, "step": 290 }, { "epoch": 0.09311006825938567, "grad_norm": 0.6472283601760864, "learning_rate": 3.1023454157782518e-06, "loss": 0.6304, "step": 291 }, { "epoch": 0.09343003412969283, "grad_norm": 0.6516737341880798, "learning_rate": 3.1130063965884867e-06, "loss": 0.6816, "step": 292 }, { "epoch": 0.09375, "grad_norm": 0.730661153793335, "learning_rate": 3.123667377398721e-06, "loss": 0.624, "step": 293 }, { "epoch": 0.09406996587030717, "grad_norm": 0.7991647720336914, "learning_rate": 3.1343283582089558e-06, "loss": 0.637, "step": 294 }, { "epoch": 0.09438993174061433, "grad_norm": 0.6912937760353088, "learning_rate": 3.14498933901919e-06, "loss": 0.6645, "step": 295 }, { "epoch": 0.0947098976109215, "grad_norm": 0.7326075434684753, "learning_rate": 3.155650319829425e-06, "loss": 0.6634, "step": 296 }, { "epoch": 0.09502986348122867, "grad_norm": 0.736692488193512, "learning_rate": 3.166311300639659e-06, "loss": 0.6611, "step": 297 }, { "epoch": 0.09534982935153584, "grad_norm": 0.6914559602737427, "learning_rate": 3.176972281449894e-06, "loss": 0.6449, "step": 298 }, { "epoch": 0.095669795221843, "grad_norm": 0.7045454382896423, "learning_rate": 3.187633262260128e-06, "loss": 0.6794, "step": 299 }, { "epoch": 0.09598976109215017, "grad_norm": 0.7817652821540833, "learning_rate": 3.198294243070363e-06, "loss": 0.6839, "step": 300 }, { "epoch": 0.09630972696245733, "grad_norm": 0.7341583967208862, "learning_rate": 3.208955223880597e-06, "loss": 0.6486, "step": 301 }, { "epoch": 0.0966296928327645, "grad_norm": 0.6466298699378967, "learning_rate": 3.219616204690832e-06, "loss": 0.6218, "step": 302 }, { "epoch": 0.09694965870307168, "grad_norm": 0.7359744906425476, "learning_rate": 3.230277185501066e-06, "loss": 0.6357, "step": 303 }, { "epoch": 0.09726962457337884, "grad_norm": 0.7613527774810791, "learning_rate": 3.240938166311301e-06, "loss": 0.662, "step": 304 }, { "epoch": 0.09758959044368601, "grad_norm": 0.7327111959457397, "learning_rate": 3.251599147121535e-06, "loss": 0.6443, "step": 305 }, { "epoch": 0.09790955631399317, "grad_norm": 0.7447808384895325, "learning_rate": 3.26226012793177e-06, "loss": 0.6432, "step": 306 }, { "epoch": 0.09822952218430034, "grad_norm": 0.705234706401825, "learning_rate": 3.272921108742004e-06, "loss": 0.6817, "step": 307 }, { "epoch": 0.0985494880546075, "grad_norm": 0.743067741394043, "learning_rate": 3.283582089552239e-06, "loss": 0.6824, "step": 308 }, { "epoch": 0.09886945392491468, "grad_norm": 0.6528238654136658, "learning_rate": 3.2942430703624733e-06, "loss": 0.6466, "step": 309 }, { "epoch": 0.09918941979522185, "grad_norm": 0.6550853252410889, "learning_rate": 3.304904051172708e-06, "loss": 0.5895, "step": 310 }, { "epoch": 0.09950938566552901, "grad_norm": 0.6330170631408691, "learning_rate": 3.3155650319829423e-06, "loss": 0.6457, "step": 311 }, { "epoch": 0.09982935153583618, "grad_norm": 0.7128297686576843, "learning_rate": 3.3262260127931773e-06, "loss": 0.6464, "step": 312 }, { "epoch": 0.10014931740614334, "grad_norm": 0.636508047580719, "learning_rate": 3.336886993603412e-06, "loss": 0.6088, "step": 313 }, { "epoch": 0.10046928327645051, "grad_norm": 0.8207855224609375, "learning_rate": 3.3475479744136463e-06, "loss": 0.7245, "step": 314 }, { "epoch": 0.10078924914675767, "grad_norm": 0.6568029522895813, "learning_rate": 3.3582089552238813e-06, "loss": 0.6385, "step": 315 }, { "epoch": 0.10110921501706485, "grad_norm": 0.7073377966880798, "learning_rate": 3.3688699360341154e-06, "loss": 0.6677, "step": 316 }, { "epoch": 0.10142918088737202, "grad_norm": 0.6592832207679749, "learning_rate": 3.3795309168443503e-06, "loss": 0.6273, "step": 317 }, { "epoch": 0.10174914675767918, "grad_norm": 0.6786140203475952, "learning_rate": 3.3901918976545844e-06, "loss": 0.6669, "step": 318 }, { "epoch": 0.10206911262798635, "grad_norm": 0.6719084978103638, "learning_rate": 3.4008528784648194e-06, "loss": 0.6335, "step": 319 }, { "epoch": 0.10238907849829351, "grad_norm": 0.6851851940155029, "learning_rate": 3.4115138592750535e-06, "loss": 0.6011, "step": 320 }, { "epoch": 0.10270904436860068, "grad_norm": 0.6099117994308472, "learning_rate": 3.4221748400852884e-06, "loss": 0.6332, "step": 321 }, { "epoch": 0.10302901023890786, "grad_norm": 0.6866176724433899, "learning_rate": 3.4328358208955225e-06, "loss": 0.6017, "step": 322 }, { "epoch": 0.10334897610921502, "grad_norm": 0.7074094414710999, "learning_rate": 3.4434968017057575e-06, "loss": 0.6177, "step": 323 }, { "epoch": 0.10366894197952219, "grad_norm": 0.6538453102111816, "learning_rate": 3.4541577825159916e-06, "loss": 0.644, "step": 324 }, { "epoch": 0.10398890784982935, "grad_norm": 0.6807860732078552, "learning_rate": 3.4648187633262265e-06, "loss": 0.6325, "step": 325 }, { "epoch": 0.10430887372013652, "grad_norm": 0.8274717330932617, "learning_rate": 3.4754797441364606e-06, "loss": 0.6364, "step": 326 }, { "epoch": 0.10462883959044368, "grad_norm": 0.7771089673042297, "learning_rate": 3.4861407249466956e-06, "loss": 0.6476, "step": 327 }, { "epoch": 0.10494880546075085, "grad_norm": 0.7471724152565002, "learning_rate": 3.4968017057569297e-06, "loss": 0.6656, "step": 328 }, { "epoch": 0.10526877133105803, "grad_norm": 0.7084169387817383, "learning_rate": 3.5074626865671646e-06, "loss": 0.6722, "step": 329 }, { "epoch": 0.10558873720136519, "grad_norm": 0.700748085975647, "learning_rate": 3.5181236673773987e-06, "loss": 0.622, "step": 330 }, { "epoch": 0.10590870307167236, "grad_norm": 0.8253344893455505, "learning_rate": 3.5287846481876337e-06, "loss": 0.6405, "step": 331 }, { "epoch": 0.10622866894197952, "grad_norm": 0.7408985495567322, "learning_rate": 3.5394456289978678e-06, "loss": 0.6814, "step": 332 }, { "epoch": 0.10654863481228669, "grad_norm": 0.9821094274520874, "learning_rate": 3.5501066098081027e-06, "loss": 0.6398, "step": 333 }, { "epoch": 0.10686860068259385, "grad_norm": 0.6583428978919983, "learning_rate": 3.560767590618337e-06, "loss": 0.6206, "step": 334 }, { "epoch": 0.10718856655290103, "grad_norm": 0.5934454798698425, "learning_rate": 3.5714285714285718e-06, "loss": 0.6149, "step": 335 }, { "epoch": 0.1075085324232082, "grad_norm": 0.6851022839546204, "learning_rate": 3.582089552238806e-06, "loss": 0.6245, "step": 336 }, { "epoch": 0.10782849829351536, "grad_norm": 0.6943828463554382, "learning_rate": 3.592750533049041e-06, "loss": 0.6542, "step": 337 }, { "epoch": 0.10814846416382252, "grad_norm": 0.6568377017974854, "learning_rate": 3.603411513859275e-06, "loss": 0.6493, "step": 338 }, { "epoch": 0.10846843003412969, "grad_norm": 0.6676375865936279, "learning_rate": 3.61407249466951e-06, "loss": 0.6573, "step": 339 }, { "epoch": 0.10878839590443685, "grad_norm": 0.6609268188476562, "learning_rate": 3.624733475479744e-06, "loss": 0.6402, "step": 340 }, { "epoch": 0.10910836177474403, "grad_norm": 0.6636278033256531, "learning_rate": 3.635394456289979e-06, "loss": 0.6256, "step": 341 }, { "epoch": 0.1094283276450512, "grad_norm": 0.6692773699760437, "learning_rate": 3.6460554371002135e-06, "loss": 0.6314, "step": 342 }, { "epoch": 0.10974829351535836, "grad_norm": 0.7020422220230103, "learning_rate": 3.656716417910448e-06, "loss": 0.6009, "step": 343 }, { "epoch": 0.11006825938566553, "grad_norm": 0.5868972539901733, "learning_rate": 3.667377398720683e-06, "loss": 0.627, "step": 344 }, { "epoch": 0.1103882252559727, "grad_norm": 0.6617975831031799, "learning_rate": 3.678038379530917e-06, "loss": 0.6157, "step": 345 }, { "epoch": 0.11070819112627986, "grad_norm": 0.6820520758628845, "learning_rate": 3.688699360341152e-06, "loss": 0.6337, "step": 346 }, { "epoch": 0.11102815699658702, "grad_norm": 0.7438987493515015, "learning_rate": 3.699360341151386e-06, "loss": 0.6758, "step": 347 }, { "epoch": 0.1113481228668942, "grad_norm": 0.7671152353286743, "learning_rate": 3.710021321961621e-06, "loss": 0.6857, "step": 348 }, { "epoch": 0.11166808873720137, "grad_norm": 0.7241283655166626, "learning_rate": 3.720682302771855e-06, "loss": 0.6179, "step": 349 }, { "epoch": 0.11198805460750853, "grad_norm": 0.6965861916542053, "learning_rate": 3.73134328358209e-06, "loss": 0.6618, "step": 350 }, { "epoch": 0.1123080204778157, "grad_norm": 0.7182785868644714, "learning_rate": 3.742004264392324e-06, "loss": 0.6689, "step": 351 }, { "epoch": 0.11262798634812286, "grad_norm": 0.7257066369056702, "learning_rate": 3.752665245202559e-06, "loss": 0.6041, "step": 352 }, { "epoch": 0.11294795221843003, "grad_norm": 0.6591728925704956, "learning_rate": 3.7633262260127933e-06, "loss": 0.6401, "step": 353 }, { "epoch": 0.11326791808873721, "grad_norm": 0.6794602870941162, "learning_rate": 3.773987206823028e-06, "loss": 0.6376, "step": 354 }, { "epoch": 0.11358788395904437, "grad_norm": 0.7658209204673767, "learning_rate": 3.7846481876332623e-06, "loss": 0.634, "step": 355 }, { "epoch": 0.11390784982935154, "grad_norm": 0.6292558312416077, "learning_rate": 3.7953091684434973e-06, "loss": 0.6086, "step": 356 }, { "epoch": 0.1142278156996587, "grad_norm": 0.6457000374794006, "learning_rate": 3.8059701492537314e-06, "loss": 0.605, "step": 357 }, { "epoch": 0.11454778156996587, "grad_norm": 0.7265068292617798, "learning_rate": 3.816631130063966e-06, "loss": 0.6344, "step": 358 }, { "epoch": 0.11486774744027303, "grad_norm": 0.6964762806892395, "learning_rate": 3.827292110874201e-06, "loss": 0.6327, "step": 359 }, { "epoch": 0.11518771331058021, "grad_norm": 0.717154860496521, "learning_rate": 3.837953091684435e-06, "loss": 0.6271, "step": 360 }, { "epoch": 0.11550767918088738, "grad_norm": 0.687904953956604, "learning_rate": 3.84861407249467e-06, "loss": 0.6407, "step": 361 }, { "epoch": 0.11582764505119454, "grad_norm": 0.6947673559188843, "learning_rate": 3.859275053304904e-06, "loss": 0.614, "step": 362 }, { "epoch": 0.1161476109215017, "grad_norm": 0.7783513069152832, "learning_rate": 3.869936034115139e-06, "loss": 0.622, "step": 363 }, { "epoch": 0.11646757679180887, "grad_norm": 0.7462539076805115, "learning_rate": 3.8805970149253735e-06, "loss": 0.6867, "step": 364 }, { "epoch": 0.11678754266211604, "grad_norm": 0.6062483787536621, "learning_rate": 3.891257995735608e-06, "loss": 0.6055, "step": 365 }, { "epoch": 0.1171075085324232, "grad_norm": 0.7094764113426208, "learning_rate": 3.9019189765458425e-06, "loss": 0.63, "step": 366 }, { "epoch": 0.11742747440273038, "grad_norm": 0.7285301089286804, "learning_rate": 3.912579957356077e-06, "loss": 0.6302, "step": 367 }, { "epoch": 0.11774744027303755, "grad_norm": 0.6298773884773254, "learning_rate": 3.9232409381663116e-06, "loss": 0.6001, "step": 368 }, { "epoch": 0.11806740614334471, "grad_norm": 0.7302676439285278, "learning_rate": 3.933901918976546e-06, "loss": 0.6395, "step": 369 }, { "epoch": 0.11838737201365188, "grad_norm": 0.6924223303794861, "learning_rate": 3.944562899786781e-06, "loss": 0.5949, "step": 370 }, { "epoch": 0.11870733788395904, "grad_norm": 0.6996256709098816, "learning_rate": 3.955223880597015e-06, "loss": 0.6594, "step": 371 }, { "epoch": 0.1190273037542662, "grad_norm": 0.7116991281509399, "learning_rate": 3.96588486140725e-06, "loss": 0.6228, "step": 372 }, { "epoch": 0.11934726962457338, "grad_norm": 0.6261963844299316, "learning_rate": 3.976545842217484e-06, "loss": 0.606, "step": 373 }, { "epoch": 0.11966723549488055, "grad_norm": 0.7136417627334595, "learning_rate": 3.987206823027719e-06, "loss": 0.5899, "step": 374 }, { "epoch": 0.11998720136518772, "grad_norm": 0.6391706466674805, "learning_rate": 3.997867803837953e-06, "loss": 0.6355, "step": 375 }, { "epoch": 0.12030716723549488, "grad_norm": 0.6646369099617004, "learning_rate": 4.008528784648188e-06, "loss": 0.6122, "step": 376 }, { "epoch": 0.12062713310580205, "grad_norm": 0.6632401347160339, "learning_rate": 4.019189765458423e-06, "loss": 0.6471, "step": 377 }, { "epoch": 0.12094709897610921, "grad_norm": 0.6242071390151978, "learning_rate": 4.029850746268657e-06, "loss": 0.5711, "step": 378 }, { "epoch": 0.12126706484641639, "grad_norm": 0.641707181930542, "learning_rate": 4.040511727078892e-06, "loss": 0.6318, "step": 379 }, { "epoch": 0.12158703071672355, "grad_norm": 0.7148804068565369, "learning_rate": 4.051172707889126e-06, "loss": 0.6299, "step": 380 }, { "epoch": 0.12190699658703072, "grad_norm": 0.7760037183761597, "learning_rate": 4.061833688699361e-06, "loss": 0.6155, "step": 381 }, { "epoch": 0.12222696245733788, "grad_norm": 0.679989755153656, "learning_rate": 4.072494669509595e-06, "loss": 0.6409, "step": 382 }, { "epoch": 0.12254692832764505, "grad_norm": 0.6069170236587524, "learning_rate": 4.08315565031983e-06, "loss": 0.607, "step": 383 }, { "epoch": 0.12286689419795221, "grad_norm": 0.6047013401985168, "learning_rate": 4.093816631130064e-06, "loss": 0.6175, "step": 384 }, { "epoch": 0.12318686006825938, "grad_norm": 0.6415553689002991, "learning_rate": 4.104477611940299e-06, "loss": 0.6168, "step": 385 }, { "epoch": 0.12350682593856656, "grad_norm": 0.7326337695121765, "learning_rate": 4.115138592750533e-06, "loss": 0.6247, "step": 386 }, { "epoch": 0.12382679180887372, "grad_norm": 0.7146400809288025, "learning_rate": 4.125799573560768e-06, "loss": 0.62, "step": 387 }, { "epoch": 0.12414675767918089, "grad_norm": 0.6480631828308105, "learning_rate": 4.136460554371002e-06, "loss": 0.6511, "step": 388 }, { "epoch": 0.12446672354948805, "grad_norm": 0.6476372480392456, "learning_rate": 4.1471215351812375e-06, "loss": 0.6027, "step": 389 }, { "epoch": 0.12478668941979522, "grad_norm": 0.644311785697937, "learning_rate": 4.157782515991471e-06, "loss": 0.6371, "step": 390 }, { "epoch": 0.1251066552901024, "grad_norm": 0.6093184351921082, "learning_rate": 4.1684434968017065e-06, "loss": 0.638, "step": 391 }, { "epoch": 0.12542662116040956, "grad_norm": 0.6396116018295288, "learning_rate": 4.17910447761194e-06, "loss": 0.6051, "step": 392 }, { "epoch": 0.12574658703071673, "grad_norm": 0.6135868430137634, "learning_rate": 4.1897654584221756e-06, "loss": 0.6662, "step": 393 }, { "epoch": 0.1260665529010239, "grad_norm": 0.661790132522583, "learning_rate": 4.200426439232409e-06, "loss": 0.623, "step": 394 }, { "epoch": 0.12638651877133106, "grad_norm": 0.687401533126831, "learning_rate": 4.211087420042645e-06, "loss": 0.6448, "step": 395 }, { "epoch": 0.12670648464163822, "grad_norm": 0.6437802910804749, "learning_rate": 4.221748400852878e-06, "loss": 0.6103, "step": 396 }, { "epoch": 0.1270264505119454, "grad_norm": 0.7061891555786133, "learning_rate": 4.232409381663114e-06, "loss": 0.634, "step": 397 }, { "epoch": 0.12734641638225255, "grad_norm": 0.6025312542915344, "learning_rate": 4.243070362473347e-06, "loss": 0.6258, "step": 398 }, { "epoch": 0.12766638225255972, "grad_norm": 0.9356108903884888, "learning_rate": 4.253731343283583e-06, "loss": 0.5903, "step": 399 }, { "epoch": 0.12798634812286688, "grad_norm": 0.6520940661430359, "learning_rate": 4.264392324093816e-06, "loss": 0.5926, "step": 400 }, { "epoch": 0.12830631399317405, "grad_norm": 0.6317100524902344, "learning_rate": 4.275053304904052e-06, "loss": 0.5891, "step": 401 }, { "epoch": 0.12862627986348124, "grad_norm": 0.6645535826683044, "learning_rate": 4.2857142857142855e-06, "loss": 0.572, "step": 402 }, { "epoch": 0.1289462457337884, "grad_norm": 0.6676063537597656, "learning_rate": 4.296375266524521e-06, "loss": 0.6334, "step": 403 }, { "epoch": 0.12926621160409557, "grad_norm": 0.6608290672302246, "learning_rate": 4.3070362473347545e-06, "loss": 0.6003, "step": 404 }, { "epoch": 0.12958617747440274, "grad_norm": 0.7465037107467651, "learning_rate": 4.31769722814499e-06, "loss": 0.6297, "step": 405 }, { "epoch": 0.1299061433447099, "grad_norm": 0.7195454239845276, "learning_rate": 4.3283582089552236e-06, "loss": 0.6276, "step": 406 }, { "epoch": 0.13022610921501707, "grad_norm": 0.7333062887191772, "learning_rate": 4.339019189765459e-06, "loss": 0.5858, "step": 407 }, { "epoch": 0.13054607508532423, "grad_norm": 0.7343773245811462, "learning_rate": 4.3496801705756935e-06, "loss": 0.6304, "step": 408 }, { "epoch": 0.1308660409556314, "grad_norm": 0.7145223021507263, "learning_rate": 4.360341151385928e-06, "loss": 0.676, "step": 409 }, { "epoch": 0.13118600682593856, "grad_norm": 0.6226415634155273, "learning_rate": 4.3710021321961625e-06, "loss": 0.6018, "step": 410 }, { "epoch": 0.13150597269624573, "grad_norm": 0.7167772650718689, "learning_rate": 4.381663113006397e-06, "loss": 0.6159, "step": 411 }, { "epoch": 0.1318259385665529, "grad_norm": 0.6907845139503479, "learning_rate": 4.3923240938166316e-06, "loss": 0.6303, "step": 412 }, { "epoch": 0.13214590443686006, "grad_norm": 0.729068398475647, "learning_rate": 4.402985074626866e-06, "loss": 0.603, "step": 413 }, { "epoch": 0.13246587030716722, "grad_norm": 0.6759557127952576, "learning_rate": 4.413646055437101e-06, "loss": 0.6493, "step": 414 }, { "epoch": 0.13278583617747441, "grad_norm": 0.6795688271522522, "learning_rate": 4.424307036247335e-06, "loss": 0.631, "step": 415 }, { "epoch": 0.13310580204778158, "grad_norm": 0.6279520988464355, "learning_rate": 4.43496801705757e-06, "loss": 0.596, "step": 416 }, { "epoch": 0.13342576791808874, "grad_norm": 0.6398487091064453, "learning_rate": 4.445628997867804e-06, "loss": 0.6141, "step": 417 }, { "epoch": 0.1337457337883959, "grad_norm": 0.6193283200263977, "learning_rate": 4.456289978678039e-06, "loss": 0.6011, "step": 418 }, { "epoch": 0.13406569965870307, "grad_norm": 0.6516242027282715, "learning_rate": 4.466950959488273e-06, "loss": 0.6447, "step": 419 }, { "epoch": 0.13438566552901024, "grad_norm": 0.6823792457580566, "learning_rate": 4.477611940298508e-06, "loss": 0.6377, "step": 420 }, { "epoch": 0.1347056313993174, "grad_norm": 0.6426964402198792, "learning_rate": 4.488272921108742e-06, "loss": 0.5983, "step": 421 }, { "epoch": 0.13502559726962457, "grad_norm": 0.6207419633865356, "learning_rate": 4.498933901918977e-06, "loss": 0.632, "step": 422 }, { "epoch": 0.13534556313993173, "grad_norm": 0.6856409907341003, "learning_rate": 4.509594882729211e-06, "loss": 0.6397, "step": 423 }, { "epoch": 0.1356655290102389, "grad_norm": 0.666517972946167, "learning_rate": 4.520255863539446e-06, "loss": 0.6226, "step": 424 }, { "epoch": 0.13598549488054607, "grad_norm": 0.7025275230407715, "learning_rate": 4.53091684434968e-06, "loss": 0.6276, "step": 425 }, { "epoch": 0.13630546075085323, "grad_norm": 0.7011261582374573, "learning_rate": 4.541577825159915e-06, "loss": 0.6618, "step": 426 }, { "epoch": 0.13662542662116042, "grad_norm": 0.6024221777915955, "learning_rate": 4.5522388059701495e-06, "loss": 0.574, "step": 427 }, { "epoch": 0.1369453924914676, "grad_norm": 0.6258330345153809, "learning_rate": 4.562899786780384e-06, "loss": 0.6353, "step": 428 }, { "epoch": 0.13726535836177475, "grad_norm": 0.633340060710907, "learning_rate": 4.5735607675906185e-06, "loss": 0.5696, "step": 429 }, { "epoch": 0.13758532423208192, "grad_norm": 0.6408478617668152, "learning_rate": 4.584221748400853e-06, "loss": 0.6102, "step": 430 }, { "epoch": 0.13790529010238908, "grad_norm": 0.7019578814506531, "learning_rate": 4.5948827292110876e-06, "loss": 0.651, "step": 431 }, { "epoch": 0.13822525597269625, "grad_norm": 0.6981320381164551, "learning_rate": 4.605543710021322e-06, "loss": 0.6474, "step": 432 }, { "epoch": 0.1385452218430034, "grad_norm": 0.6756162643432617, "learning_rate": 4.616204690831557e-06, "loss": 0.5541, "step": 433 }, { "epoch": 0.13886518771331058, "grad_norm": 0.7130482792854309, "learning_rate": 4.626865671641791e-06, "loss": 0.5995, "step": 434 }, { "epoch": 0.13918515358361774, "grad_norm": 0.6483554244041443, "learning_rate": 4.637526652452026e-06, "loss": 0.5764, "step": 435 }, { "epoch": 0.1395051194539249, "grad_norm": 0.7095510959625244, "learning_rate": 4.64818763326226e-06, "loss": 0.6112, "step": 436 }, { "epoch": 0.13982508532423207, "grad_norm": 0.8624890446662903, "learning_rate": 4.658848614072495e-06, "loss": 0.6447, "step": 437 }, { "epoch": 0.14014505119453924, "grad_norm": 0.7431426644325256, "learning_rate": 4.669509594882729e-06, "loss": 0.6282, "step": 438 }, { "epoch": 0.1404650170648464, "grad_norm": 0.6454997062683105, "learning_rate": 4.680170575692965e-06, "loss": 0.6507, "step": 439 }, { "epoch": 0.1407849829351536, "grad_norm": 0.7619973421096802, "learning_rate": 4.690831556503198e-06, "loss": 0.6045, "step": 440 }, { "epoch": 0.14110494880546076, "grad_norm": 0.678459644317627, "learning_rate": 4.701492537313434e-06, "loss": 0.6377, "step": 441 }, { "epoch": 0.14142491467576793, "grad_norm": 0.6140632033348083, "learning_rate": 4.712153518123667e-06, "loss": 0.6237, "step": 442 }, { "epoch": 0.1417448805460751, "grad_norm": 0.6932960748672485, "learning_rate": 4.722814498933903e-06, "loss": 0.6598, "step": 443 }, { "epoch": 0.14206484641638226, "grad_norm": 0.8086076974868774, "learning_rate": 4.733475479744136e-06, "loss": 0.6339, "step": 444 }, { "epoch": 0.14238481228668942, "grad_norm": 0.7320109009742737, "learning_rate": 4.744136460554372e-06, "loss": 0.5965, "step": 445 }, { "epoch": 0.1427047781569966, "grad_norm": 0.6901565790176392, "learning_rate": 4.7547974413646055e-06, "loss": 0.6033, "step": 446 }, { "epoch": 0.14302474402730375, "grad_norm": 0.7285332679748535, "learning_rate": 4.765458422174841e-06, "loss": 0.6248, "step": 447 }, { "epoch": 0.14334470989761092, "grad_norm": 0.6956148147583008, "learning_rate": 4.7761194029850745e-06, "loss": 0.6356, "step": 448 }, { "epoch": 0.14366467576791808, "grad_norm": 0.673285186290741, "learning_rate": 4.78678038379531e-06, "loss": 0.6328, "step": 449 }, { "epoch": 0.14398464163822525, "grad_norm": 0.7397879958152771, "learning_rate": 4.797441364605544e-06, "loss": 0.6247, "step": 450 }, { "epoch": 0.1443046075085324, "grad_norm": 0.825232982635498, "learning_rate": 4.808102345415779e-06, "loss": 0.6125, "step": 451 }, { "epoch": 0.14462457337883958, "grad_norm": 0.727387011051178, "learning_rate": 4.8187633262260135e-06, "loss": 0.6363, "step": 452 }, { "epoch": 0.14494453924914677, "grad_norm": 0.6681481599807739, "learning_rate": 4.829424307036248e-06, "loss": 0.6234, "step": 453 }, { "epoch": 0.14526450511945393, "grad_norm": 0.7631570100784302, "learning_rate": 4.8400852878464825e-06, "loss": 0.6238, "step": 454 }, { "epoch": 0.1455844709897611, "grad_norm": 0.6824873089790344, "learning_rate": 4.850746268656717e-06, "loss": 0.6299, "step": 455 }, { "epoch": 0.14590443686006827, "grad_norm": 0.6673473119735718, "learning_rate": 4.8614072494669516e-06, "loss": 0.6139, "step": 456 }, { "epoch": 0.14622440273037543, "grad_norm": 0.6729328036308289, "learning_rate": 4.872068230277186e-06, "loss": 0.5745, "step": 457 }, { "epoch": 0.1465443686006826, "grad_norm": 0.7668410539627075, "learning_rate": 4.882729211087421e-06, "loss": 0.6114, "step": 458 }, { "epoch": 0.14686433447098976, "grad_norm": 0.657248318195343, "learning_rate": 4.893390191897655e-06, "loss": 0.639, "step": 459 }, { "epoch": 0.14718430034129693, "grad_norm": 0.6458208560943604, "learning_rate": 4.90405117270789e-06, "loss": 0.6091, "step": 460 }, { "epoch": 0.1475042662116041, "grad_norm": 0.6650400161743164, "learning_rate": 4.914712153518124e-06, "loss": 0.5947, "step": 461 }, { "epoch": 0.14782423208191126, "grad_norm": 0.6707201600074768, "learning_rate": 4.925373134328359e-06, "loss": 0.592, "step": 462 }, { "epoch": 0.14814419795221842, "grad_norm": 0.7198519706726074, "learning_rate": 4.936034115138593e-06, "loss": 0.5947, "step": 463 }, { "epoch": 0.14846416382252559, "grad_norm": 0.7003646492958069, "learning_rate": 4.946695095948828e-06, "loss": 0.6417, "step": 464 }, { "epoch": 0.14878412969283278, "grad_norm": 0.6954401731491089, "learning_rate": 4.957356076759062e-06, "loss": 0.6477, "step": 465 }, { "epoch": 0.14910409556313994, "grad_norm": 0.7278227806091309, "learning_rate": 4.968017057569297e-06, "loss": 0.651, "step": 466 }, { "epoch": 0.1494240614334471, "grad_norm": 0.8354657888412476, "learning_rate": 4.978678038379531e-06, "loss": 0.5986, "step": 467 }, { "epoch": 0.14974402730375427, "grad_norm": 0.7701458930969238, "learning_rate": 4.989339019189766e-06, "loss": 0.6188, "step": 468 }, { "epoch": 0.15006399317406144, "grad_norm": 0.6838735938072205, "learning_rate": 5e-06, "loss": 0.6067, "step": 469 }, { "epoch": 0.1503839590443686, "grad_norm": 0.6737359762191772, "learning_rate": 5.010660980810235e-06, "loss": 0.632, "step": 470 }, { "epoch": 0.15070392491467577, "grad_norm": 0.7790951132774353, "learning_rate": 5.02132196162047e-06, "loss": 0.5928, "step": 471 }, { "epoch": 0.15102389078498293, "grad_norm": 0.7367972135543823, "learning_rate": 5.031982942430704e-06, "loss": 0.6217, "step": 472 }, { "epoch": 0.1513438566552901, "grad_norm": 0.7044655084609985, "learning_rate": 5.0426439232409385e-06, "loss": 0.6131, "step": 473 }, { "epoch": 0.15166382252559726, "grad_norm": 0.6171327233314514, "learning_rate": 5.053304904051173e-06, "loss": 0.6232, "step": 474 }, { "epoch": 0.15198378839590443, "grad_norm": 0.7287469506263733, "learning_rate": 5.063965884861408e-06, "loss": 0.5999, "step": 475 }, { "epoch": 0.1523037542662116, "grad_norm": 0.7412554621696472, "learning_rate": 5.074626865671642e-06, "loss": 0.6647, "step": 476 }, { "epoch": 0.15262372013651876, "grad_norm": 0.7282645106315613, "learning_rate": 5.085287846481877e-06, "loss": 0.6364, "step": 477 }, { "epoch": 0.15294368600682595, "grad_norm": 0.7131372094154358, "learning_rate": 5.095948827292111e-06, "loss": 0.6549, "step": 478 }, { "epoch": 0.15326365187713312, "grad_norm": 0.7518019676208496, "learning_rate": 5.1066098081023465e-06, "loss": 0.6298, "step": 479 }, { "epoch": 0.15358361774744028, "grad_norm": 0.6324292421340942, "learning_rate": 5.11727078891258e-06, "loss": 0.6214, "step": 480 }, { "epoch": 0.15390358361774745, "grad_norm": 0.692251980304718, "learning_rate": 5.127931769722815e-06, "loss": 0.5769, "step": 481 }, { "epoch": 0.1542235494880546, "grad_norm": 0.6692104339599609, "learning_rate": 5.138592750533049e-06, "loss": 0.5986, "step": 482 }, { "epoch": 0.15454351535836178, "grad_norm": 0.6897379159927368, "learning_rate": 5.149253731343285e-06, "loss": 0.6281, "step": 483 }, { "epoch": 0.15486348122866894, "grad_norm": 0.6690213680267334, "learning_rate": 5.159914712153518e-06, "loss": 0.574, "step": 484 }, { "epoch": 0.1551834470989761, "grad_norm": 0.6694300770759583, "learning_rate": 5.170575692963753e-06, "loss": 0.6249, "step": 485 }, { "epoch": 0.15550341296928327, "grad_norm": 0.6471107006072998, "learning_rate": 5.181236673773987e-06, "loss": 0.6231, "step": 486 }, { "epoch": 0.15582337883959044, "grad_norm": 0.618867039680481, "learning_rate": 5.191897654584223e-06, "loss": 0.6061, "step": 487 }, { "epoch": 0.1561433447098976, "grad_norm": 0.6762124300003052, "learning_rate": 5.202558635394456e-06, "loss": 0.6264, "step": 488 }, { "epoch": 0.15646331058020477, "grad_norm": 0.6503360867500305, "learning_rate": 5.213219616204691e-06, "loss": 0.6345, "step": 489 }, { "epoch": 0.15678327645051193, "grad_norm": 0.7969846725463867, "learning_rate": 5.2238805970149255e-06, "loss": 0.6233, "step": 490 }, { "epoch": 0.15710324232081913, "grad_norm": 0.7003441452980042, "learning_rate": 5.234541577825161e-06, "loss": 0.634, "step": 491 }, { "epoch": 0.1574232081911263, "grad_norm": 0.6595831513404846, "learning_rate": 5.245202558635395e-06, "loss": 0.6253, "step": 492 }, { "epoch": 0.15774317406143346, "grad_norm": 0.7349994778633118, "learning_rate": 5.255863539445629e-06, "loss": 0.6214, "step": 493 }, { "epoch": 0.15806313993174062, "grad_norm": 0.7195835113525391, "learning_rate": 5.2665245202558636e-06, "loss": 0.5856, "step": 494 }, { "epoch": 0.15838310580204779, "grad_norm": 0.6510567665100098, "learning_rate": 5.277185501066099e-06, "loss": 0.6374, "step": 495 }, { "epoch": 0.15870307167235495, "grad_norm": 0.7259697318077087, "learning_rate": 5.2878464818763335e-06, "loss": 0.6446, "step": 496 }, { "epoch": 0.15902303754266212, "grad_norm": 0.6870812177658081, "learning_rate": 5.298507462686567e-06, "loss": 0.5873, "step": 497 }, { "epoch": 0.15934300341296928, "grad_norm": 0.8159638047218323, "learning_rate": 5.309168443496802e-06, "loss": 0.6401, "step": 498 }, { "epoch": 0.15966296928327645, "grad_norm": 0.6727353930473328, "learning_rate": 5.319829424307037e-06, "loss": 0.5834, "step": 499 }, { "epoch": 0.1599829351535836, "grad_norm": 0.6303347945213318, "learning_rate": 5.3304904051172716e-06, "loss": 0.5554, "step": 500 }, { "epoch": 0.16030290102389078, "grad_norm": 0.6501518487930298, "learning_rate": 5.341151385927505e-06, "loss": 0.6153, "step": 501 }, { "epoch": 0.16062286689419794, "grad_norm": 0.7419226169586182, "learning_rate": 5.351812366737741e-06, "loss": 0.6119, "step": 502 }, { "epoch": 0.1609428327645051, "grad_norm": 0.6859214901924133, "learning_rate": 5.362473347547975e-06, "loss": 0.6041, "step": 503 }, { "epoch": 0.1612627986348123, "grad_norm": 0.68569016456604, "learning_rate": 5.37313432835821e-06, "loss": 0.6317, "step": 504 }, { "epoch": 0.16158276450511946, "grad_norm": 0.6246910095214844, "learning_rate": 5.383795309168443e-06, "loss": 0.6128, "step": 505 }, { "epoch": 0.16190273037542663, "grad_norm": 0.7596530914306641, "learning_rate": 5.394456289978679e-06, "loss": 0.6383, "step": 506 }, { "epoch": 0.1622226962457338, "grad_norm": 0.6581939458847046, "learning_rate": 5.405117270788913e-06, "loss": 0.5816, "step": 507 }, { "epoch": 0.16254266211604096, "grad_norm": 0.6166406273841858, "learning_rate": 5.415778251599148e-06, "loss": 0.6065, "step": 508 }, { "epoch": 0.16286262798634812, "grad_norm": 0.7130981683731079, "learning_rate": 5.4264392324093815e-06, "loss": 0.6438, "step": 509 }, { "epoch": 0.1631825938566553, "grad_norm": 0.7229422926902771, "learning_rate": 5.437100213219617e-06, "loss": 0.6152, "step": 510 }, { "epoch": 0.16350255972696245, "grad_norm": 0.6925520896911621, "learning_rate": 5.447761194029851e-06, "loss": 0.6183, "step": 511 }, { "epoch": 0.16382252559726962, "grad_norm": 0.7318209409713745, "learning_rate": 5.458422174840086e-06, "loss": 0.6151, "step": 512 }, { "epoch": 0.16414249146757678, "grad_norm": 0.7651650905609131, "learning_rate": 5.4690831556503196e-06, "loss": 0.6298, "step": 513 }, { "epoch": 0.16446245733788395, "grad_norm": 0.7928692698478699, "learning_rate": 5.479744136460555e-06, "loss": 0.6212, "step": 514 }, { "epoch": 0.16478242320819111, "grad_norm": 0.7141736745834351, "learning_rate": 5.4904051172707895e-06, "loss": 0.6023, "step": 515 }, { "epoch": 0.1651023890784983, "grad_norm": 0.6715232133865356, "learning_rate": 5.501066098081024e-06, "loss": 0.6173, "step": 516 }, { "epoch": 0.16542235494880547, "grad_norm": 0.6879045963287354, "learning_rate": 5.511727078891258e-06, "loss": 0.6054, "step": 517 }, { "epoch": 0.16574232081911264, "grad_norm": 0.7811073660850525, "learning_rate": 5.522388059701493e-06, "loss": 0.5491, "step": 518 }, { "epoch": 0.1660622866894198, "grad_norm": 0.6594018340110779, "learning_rate": 5.5330490405117276e-06, "loss": 0.5957, "step": 519 }, { "epoch": 0.16638225255972697, "grad_norm": 0.6076728701591492, "learning_rate": 5.543710021321962e-06, "loss": 0.5769, "step": 520 }, { "epoch": 0.16670221843003413, "grad_norm": 0.6521137952804565, "learning_rate": 5.554371002132196e-06, "loss": 0.6136, "step": 521 }, { "epoch": 0.1670221843003413, "grad_norm": 0.6845177412033081, "learning_rate": 5.565031982942431e-06, "loss": 0.5994, "step": 522 }, { "epoch": 0.16734215017064846, "grad_norm": 0.690648078918457, "learning_rate": 5.575692963752666e-06, "loss": 0.6106, "step": 523 }, { "epoch": 0.16766211604095563, "grad_norm": 0.6770439743995667, "learning_rate": 5.5863539445629e-06, "loss": 0.5759, "step": 524 }, { "epoch": 0.1679820819112628, "grad_norm": 0.7323155403137207, "learning_rate": 5.597014925373134e-06, "loss": 0.6161, "step": 525 }, { "epoch": 0.16830204778156996, "grad_norm": 0.6350544691085815, "learning_rate": 5.607675906183369e-06, "loss": 0.5853, "step": 526 }, { "epoch": 0.16862201365187712, "grad_norm": 0.6570029854774475, "learning_rate": 5.618336886993604e-06, "loss": 0.6346, "step": 527 }, { "epoch": 0.1689419795221843, "grad_norm": 0.7343910932540894, "learning_rate": 5.628997867803838e-06, "loss": 0.6016, "step": 528 }, { "epoch": 0.16926194539249148, "grad_norm": 0.6532744765281677, "learning_rate": 5.639658848614073e-06, "loss": 0.6323, "step": 529 }, { "epoch": 0.16958191126279865, "grad_norm": 0.617245078086853, "learning_rate": 5.650319829424308e-06, "loss": 0.6216, "step": 530 }, { "epoch": 0.1699018771331058, "grad_norm": 0.6686919927597046, "learning_rate": 5.660980810234542e-06, "loss": 0.5816, "step": 531 }, { "epoch": 0.17022184300341298, "grad_norm": 0.6508616805076599, "learning_rate": 5.671641791044776e-06, "loss": 0.6232, "step": 532 }, { "epoch": 0.17054180887372014, "grad_norm": 0.6479763984680176, "learning_rate": 5.682302771855012e-06, "loss": 0.6242, "step": 533 }, { "epoch": 0.1708617747440273, "grad_norm": 0.6898424029350281, "learning_rate": 5.692963752665246e-06, "loss": 0.6153, "step": 534 }, { "epoch": 0.17118174061433447, "grad_norm": 0.631033718585968, "learning_rate": 5.70362473347548e-06, "loss": 0.5956, "step": 535 }, { "epoch": 0.17150170648464164, "grad_norm": 0.7228789329528809, "learning_rate": 5.7142857142857145e-06, "loss": 0.6204, "step": 536 }, { "epoch": 0.1718216723549488, "grad_norm": 0.7352203726768494, "learning_rate": 5.72494669509595e-06, "loss": 0.6203, "step": 537 }, { "epoch": 0.17214163822525597, "grad_norm": 0.6712698936462402, "learning_rate": 5.735607675906184e-06, "loss": 0.5926, "step": 538 }, { "epoch": 0.17246160409556313, "grad_norm": 0.7073250412940979, "learning_rate": 5.746268656716418e-06, "loss": 0.616, "step": 539 }, { "epoch": 0.1727815699658703, "grad_norm": 0.6370400786399841, "learning_rate": 5.756929637526653e-06, "loss": 0.576, "step": 540 }, { "epoch": 0.17310153583617746, "grad_norm": 0.8138288259506226, "learning_rate": 5.767590618336888e-06, "loss": 0.5674, "step": 541 }, { "epoch": 0.17342150170648465, "grad_norm": 0.6876887083053589, "learning_rate": 5.7782515991471225e-06, "loss": 0.6106, "step": 542 }, { "epoch": 0.17374146757679182, "grad_norm": 0.6476050615310669, "learning_rate": 5.788912579957356e-06, "loss": 0.6193, "step": 543 }, { "epoch": 0.17406143344709898, "grad_norm": 0.6442128419876099, "learning_rate": 5.799573560767591e-06, "loss": 0.5954, "step": 544 }, { "epoch": 0.17438139931740615, "grad_norm": 0.7466898560523987, "learning_rate": 5.810234541577826e-06, "loss": 0.6207, "step": 545 }, { "epoch": 0.17470136518771331, "grad_norm": 0.8017808794975281, "learning_rate": 5.820895522388061e-06, "loss": 0.5983, "step": 546 }, { "epoch": 0.17502133105802048, "grad_norm": 0.6698926091194153, "learning_rate": 5.831556503198294e-06, "loss": 0.6028, "step": 547 }, { "epoch": 0.17534129692832764, "grad_norm": 0.6686583757400513, "learning_rate": 5.842217484008529e-06, "loss": 0.6043, "step": 548 }, { "epoch": 0.1756612627986348, "grad_norm": 0.7330739498138428, "learning_rate": 5.852878464818764e-06, "loss": 0.5881, "step": 549 }, { "epoch": 0.17598122866894197, "grad_norm": 0.7499303221702576, "learning_rate": 5.863539445628999e-06, "loss": 0.5953, "step": 550 }, { "epoch": 0.17630119453924914, "grad_norm": 0.6294394135475159, "learning_rate": 5.874200426439232e-06, "loss": 0.591, "step": 551 }, { "epoch": 0.1766211604095563, "grad_norm": 0.6672900319099426, "learning_rate": 5.884861407249467e-06, "loss": 0.5864, "step": 552 }, { "epoch": 0.17694112627986347, "grad_norm": 0.7012150883674622, "learning_rate": 5.895522388059702e-06, "loss": 0.5784, "step": 553 }, { "epoch": 0.17726109215017063, "grad_norm": 0.6924875378608704, "learning_rate": 5.906183368869937e-06, "loss": 0.6435, "step": 554 }, { "epoch": 0.17758105802047783, "grad_norm": 0.8030469417572021, "learning_rate": 5.9168443496801705e-06, "loss": 0.6521, "step": 555 }, { "epoch": 0.177901023890785, "grad_norm": 0.6786168813705444, "learning_rate": 5.927505330490405e-06, "loss": 0.5929, "step": 556 }, { "epoch": 0.17822098976109216, "grad_norm": 0.6793124079704285, "learning_rate": 5.93816631130064e-06, "loss": 0.6022, "step": 557 }, { "epoch": 0.17854095563139932, "grad_norm": 0.7449572086334229, "learning_rate": 5.948827292110875e-06, "loss": 0.6223, "step": 558 }, { "epoch": 0.1788609215017065, "grad_norm": 0.6794044971466064, "learning_rate": 5.959488272921109e-06, "loss": 0.555, "step": 559 }, { "epoch": 0.17918088737201365, "grad_norm": 0.6000784635543823, "learning_rate": 5.970149253731343e-06, "loss": 0.5957, "step": 560 }, { "epoch": 0.17950085324232082, "grad_norm": 0.7043874859809875, "learning_rate": 5.9808102345415785e-06, "loss": 0.5595, "step": 561 }, { "epoch": 0.17982081911262798, "grad_norm": 0.7883280515670776, "learning_rate": 5.991471215351813e-06, "loss": 0.5963, "step": 562 }, { "epoch": 0.18014078498293515, "grad_norm": 0.6450311541557312, "learning_rate": 6.002132196162047e-06, "loss": 0.6307, "step": 563 }, { "epoch": 0.1804607508532423, "grad_norm": 0.6373298764228821, "learning_rate": 6.012793176972282e-06, "loss": 0.5746, "step": 564 }, { "epoch": 0.18078071672354948, "grad_norm": 0.8034327030181885, "learning_rate": 6.023454157782517e-06, "loss": 0.6252, "step": 565 }, { "epoch": 0.18110068259385664, "grad_norm": 0.6212670207023621, "learning_rate": 6.034115138592751e-06, "loss": 0.627, "step": 566 }, { "epoch": 0.18142064846416384, "grad_norm": 0.7424978017807007, "learning_rate": 6.044776119402986e-06, "loss": 0.6091, "step": 567 }, { "epoch": 0.181740614334471, "grad_norm": 0.6831108331680298, "learning_rate": 6.055437100213221e-06, "loss": 0.5944, "step": 568 }, { "epoch": 0.18206058020477817, "grad_norm": 0.6674385666847229, "learning_rate": 6.066098081023455e-06, "loss": 0.6182, "step": 569 }, { "epoch": 0.18238054607508533, "grad_norm": 0.6555317044258118, "learning_rate": 6.076759061833689e-06, "loss": 0.614, "step": 570 }, { "epoch": 0.1827005119453925, "grad_norm": 0.6911100745201111, "learning_rate": 6.087420042643924e-06, "loss": 0.6122, "step": 571 }, { "epoch": 0.18302047781569966, "grad_norm": 0.6371263861656189, "learning_rate": 6.098081023454159e-06, "loss": 0.6368, "step": 572 }, { "epoch": 0.18334044368600683, "grad_norm": 0.6716217398643494, "learning_rate": 6.108742004264393e-06, "loss": 0.5839, "step": 573 }, { "epoch": 0.183660409556314, "grad_norm": 0.7504799962043762, "learning_rate": 6.119402985074627e-06, "loss": 0.6295, "step": 574 }, { "epoch": 0.18398037542662116, "grad_norm": 0.678804337978363, "learning_rate": 6.130063965884862e-06, "loss": 0.6007, "step": 575 }, { "epoch": 0.18430034129692832, "grad_norm": 0.7213124632835388, "learning_rate": 6.140724946695097e-06, "loss": 0.6269, "step": 576 }, { "epoch": 0.1846203071672355, "grad_norm": 0.6804861426353455, "learning_rate": 6.151385927505331e-06, "loss": 0.5593, "step": 577 }, { "epoch": 0.18494027303754265, "grad_norm": 0.744810163974762, "learning_rate": 6.1620469083155655e-06, "loss": 0.6072, "step": 578 }, { "epoch": 0.18526023890784982, "grad_norm": 0.6842663288116455, "learning_rate": 6.1727078891258e-06, "loss": 0.6033, "step": 579 }, { "epoch": 0.185580204778157, "grad_norm": 0.7466291189193726, "learning_rate": 6.183368869936035e-06, "loss": 0.6218, "step": 580 }, { "epoch": 0.18590017064846417, "grad_norm": 0.7174020409584045, "learning_rate": 6.194029850746269e-06, "loss": 0.6095, "step": 581 }, { "epoch": 0.18622013651877134, "grad_norm": 0.7337679862976074, "learning_rate": 6.2046908315565036e-06, "loss": 0.6252, "step": 582 }, { "epoch": 0.1865401023890785, "grad_norm": 0.6675230264663696, "learning_rate": 6.215351812366738e-06, "loss": 0.6061, "step": 583 }, { "epoch": 0.18686006825938567, "grad_norm": 0.7611439228057861, "learning_rate": 6.2260127931769735e-06, "loss": 0.6308, "step": 584 }, { "epoch": 0.18718003412969283, "grad_norm": 0.7200325131416321, "learning_rate": 6.236673773987207e-06, "loss": 0.6027, "step": 585 }, { "epoch": 0.1875, "grad_norm": 0.6008200645446777, "learning_rate": 6.247334754797442e-06, "loss": 0.5991, "step": 586 }, { "epoch": 0.18781996587030717, "grad_norm": 0.7074070572853088, "learning_rate": 6.257995735607676e-06, "loss": 0.5954, "step": 587 }, { "epoch": 0.18813993174061433, "grad_norm": 0.6179446578025818, "learning_rate": 6.2686567164179116e-06, "loss": 0.5912, "step": 588 }, { "epoch": 0.1884598976109215, "grad_norm": 0.6689445376396179, "learning_rate": 6.279317697228145e-06, "loss": 0.6092, "step": 589 }, { "epoch": 0.18877986348122866, "grad_norm": 0.688685417175293, "learning_rate": 6.28997867803838e-06, "loss": 0.6181, "step": 590 }, { "epoch": 0.18909982935153583, "grad_norm": 0.8675934672355652, "learning_rate": 6.300639658848614e-06, "loss": 0.6463, "step": 591 }, { "epoch": 0.189419795221843, "grad_norm": 0.714806854724884, "learning_rate": 6.31130063965885e-06, "loss": 0.6366, "step": 592 }, { "epoch": 0.18973976109215018, "grad_norm": 0.7190007567405701, "learning_rate": 6.321961620469083e-06, "loss": 0.6041, "step": 593 }, { "epoch": 0.19005972696245735, "grad_norm": 0.7643619179725647, "learning_rate": 6.332622601279318e-06, "loss": 0.6026, "step": 594 }, { "epoch": 0.1903796928327645, "grad_norm": 0.6657800078392029, "learning_rate": 6.343283582089553e-06, "loss": 0.6191, "step": 595 }, { "epoch": 0.19069965870307168, "grad_norm": 0.6771308779716492, "learning_rate": 6.353944562899788e-06, "loss": 0.6488, "step": 596 }, { "epoch": 0.19101962457337884, "grad_norm": 0.7485832571983337, "learning_rate": 6.3646055437100215e-06, "loss": 0.6351, "step": 597 }, { "epoch": 0.191339590443686, "grad_norm": 0.6575362682342529, "learning_rate": 6.375266524520256e-06, "loss": 0.5739, "step": 598 }, { "epoch": 0.19165955631399317, "grad_norm": 0.6349213123321533, "learning_rate": 6.385927505330491e-06, "loss": 0.5851, "step": 599 }, { "epoch": 0.19197952218430034, "grad_norm": 0.6870304942131042, "learning_rate": 6.396588486140726e-06, "loss": 0.6312, "step": 600 }, { "epoch": 0.1922994880546075, "grad_norm": 0.5886594653129578, "learning_rate": 6.4072494669509596e-06, "loss": 0.6055, "step": 601 }, { "epoch": 0.19261945392491467, "grad_norm": 0.7129337191581726, "learning_rate": 6.417910447761194e-06, "loss": 0.558, "step": 602 }, { "epoch": 0.19293941979522183, "grad_norm": 0.6680113077163696, "learning_rate": 6.4285714285714295e-06, "loss": 0.5845, "step": 603 }, { "epoch": 0.193259385665529, "grad_norm": 0.7426353096961975, "learning_rate": 6.439232409381664e-06, "loss": 0.6856, "step": 604 }, { "epoch": 0.19357935153583616, "grad_norm": 0.7182149291038513, "learning_rate": 6.449893390191898e-06, "loss": 0.5754, "step": 605 }, { "epoch": 0.19389931740614336, "grad_norm": 0.6413835883140564, "learning_rate": 6.460554371002132e-06, "loss": 0.62, "step": 606 }, { "epoch": 0.19421928327645052, "grad_norm": 0.7080693244934082, "learning_rate": 6.4712153518123676e-06, "loss": 0.6363, "step": 607 }, { "epoch": 0.1945392491467577, "grad_norm": 0.6945396661758423, "learning_rate": 6.481876332622602e-06, "loss": 0.6092, "step": 608 }, { "epoch": 0.19485921501706485, "grad_norm": 0.7453746199607849, "learning_rate": 6.492537313432837e-06, "loss": 0.6184, "step": 609 }, { "epoch": 0.19517918088737202, "grad_norm": 0.7010712623596191, "learning_rate": 6.50319829424307e-06, "loss": 0.6062, "step": 610 }, { "epoch": 0.19549914675767918, "grad_norm": 0.626008927822113, "learning_rate": 6.513859275053306e-06, "loss": 0.6378, "step": 611 }, { "epoch": 0.19581911262798635, "grad_norm": 0.6873940229415894, "learning_rate": 6.52452025586354e-06, "loss": 0.6199, "step": 612 }, { "epoch": 0.1961390784982935, "grad_norm": 0.7339004278182983, "learning_rate": 6.535181236673775e-06, "loss": 0.609, "step": 613 }, { "epoch": 0.19645904436860068, "grad_norm": 0.7159809470176697, "learning_rate": 6.545842217484008e-06, "loss": 0.6354, "step": 614 }, { "epoch": 0.19677901023890784, "grad_norm": 0.7771121263504028, "learning_rate": 6.556503198294244e-06, "loss": 0.6504, "step": 615 }, { "epoch": 0.197098976109215, "grad_norm": 0.7306795120239258, "learning_rate": 6.567164179104478e-06, "loss": 0.6061, "step": 616 }, { "epoch": 0.19741894197952217, "grad_norm": 0.651380181312561, "learning_rate": 6.577825159914713e-06, "loss": 0.5689, "step": 617 }, { "epoch": 0.19773890784982937, "grad_norm": 0.6777538061141968, "learning_rate": 6.5884861407249465e-06, "loss": 0.5912, "step": 618 }, { "epoch": 0.19805887372013653, "grad_norm": 0.6833059191703796, "learning_rate": 6.599147121535182e-06, "loss": 0.5937, "step": 619 }, { "epoch": 0.1983788395904437, "grad_norm": 0.6252380609512329, "learning_rate": 6.609808102345416e-06, "loss": 0.5579, "step": 620 }, { "epoch": 0.19869880546075086, "grad_norm": 0.7200431823730469, "learning_rate": 6.620469083155651e-06, "loss": 0.579, "step": 621 }, { "epoch": 0.19901877133105803, "grad_norm": 0.6220011115074158, "learning_rate": 6.631130063965885e-06, "loss": 0.5765, "step": 622 }, { "epoch": 0.1993387372013652, "grad_norm": 0.6783512234687805, "learning_rate": 6.64179104477612e-06, "loss": 0.5552, "step": 623 }, { "epoch": 0.19965870307167236, "grad_norm": 0.6065102815628052, "learning_rate": 6.6524520255863545e-06, "loss": 0.6104, "step": 624 }, { "epoch": 0.19997866894197952, "grad_norm": 0.789917528629303, "learning_rate": 6.663113006396589e-06, "loss": 0.5819, "step": 625 }, { "epoch": 0.20029863481228669, "grad_norm": 0.7503900527954102, "learning_rate": 6.673773987206824e-06, "loss": 0.6099, "step": 626 }, { "epoch": 0.20061860068259385, "grad_norm": 0.7439480423927307, "learning_rate": 6.684434968017058e-06, "loss": 0.6757, "step": 627 }, { "epoch": 0.20093856655290102, "grad_norm": 0.7218211889266968, "learning_rate": 6.695095948827293e-06, "loss": 0.5991, "step": 628 }, { "epoch": 0.20125853242320818, "grad_norm": 0.7553074359893799, "learning_rate": 6.705756929637527e-06, "loss": 0.6263, "step": 629 }, { "epoch": 0.20157849829351535, "grad_norm": 0.7821144461631775, "learning_rate": 6.7164179104477625e-06, "loss": 0.6385, "step": 630 }, { "epoch": 0.20189846416382254, "grad_norm": 0.7380728125572205, "learning_rate": 6.727078891257996e-06, "loss": 0.6083, "step": 631 }, { "epoch": 0.2022184300341297, "grad_norm": 0.6631566286087036, "learning_rate": 6.737739872068231e-06, "loss": 0.5956, "step": 632 }, { "epoch": 0.20253839590443687, "grad_norm": 0.7742174863815308, "learning_rate": 6.748400852878465e-06, "loss": 0.5878, "step": 633 }, { "epoch": 0.20285836177474403, "grad_norm": 0.7614405751228333, "learning_rate": 6.759061833688701e-06, "loss": 0.5883, "step": 634 }, { "epoch": 0.2031783276450512, "grad_norm": 0.6650667786598206, "learning_rate": 6.769722814498934e-06, "loss": 0.5952, "step": 635 }, { "epoch": 0.20349829351535836, "grad_norm": 0.710259735584259, "learning_rate": 6.780383795309169e-06, "loss": 0.5938, "step": 636 }, { "epoch": 0.20381825938566553, "grad_norm": 0.6996580362319946, "learning_rate": 6.791044776119403e-06, "loss": 0.6285, "step": 637 }, { "epoch": 0.2041382252559727, "grad_norm": 0.7071582078933716, "learning_rate": 6.801705756929639e-06, "loss": 0.5862, "step": 638 }, { "epoch": 0.20445819112627986, "grad_norm": 0.7259476780891418, "learning_rate": 6.812366737739872e-06, "loss": 0.608, "step": 639 }, { "epoch": 0.20477815699658702, "grad_norm": 0.7430290579795837, "learning_rate": 6.823027718550107e-06, "loss": 0.6497, "step": 640 }, { "epoch": 0.2050981228668942, "grad_norm": 0.7619306445121765, "learning_rate": 6.8336886993603415e-06, "loss": 0.6543, "step": 641 }, { "epoch": 0.20541808873720135, "grad_norm": 0.8050456643104553, "learning_rate": 6.844349680170577e-06, "loss": 0.6278, "step": 642 }, { "epoch": 0.20573805460750852, "grad_norm": 0.7514400482177734, "learning_rate": 6.8550106609808105e-06, "loss": 0.6198, "step": 643 }, { "epoch": 0.2060580204778157, "grad_norm": 0.7751818299293518, "learning_rate": 6.865671641791045e-06, "loss": 0.6206, "step": 644 }, { "epoch": 0.20637798634812288, "grad_norm": 0.7462335824966431, "learning_rate": 6.8763326226012796e-06, "loss": 0.6169, "step": 645 }, { "epoch": 0.20669795221843004, "grad_norm": 0.6527459621429443, "learning_rate": 6.886993603411515e-06, "loss": 0.6172, "step": 646 }, { "epoch": 0.2070179180887372, "grad_norm": 0.7508776187896729, "learning_rate": 6.8976545842217495e-06, "loss": 0.5796, "step": 647 }, { "epoch": 0.20733788395904437, "grad_norm": 0.6763598918914795, "learning_rate": 6.908315565031983e-06, "loss": 0.6227, "step": 648 }, { "epoch": 0.20765784982935154, "grad_norm": 0.7581509351730347, "learning_rate": 6.918976545842218e-06, "loss": 0.6063, "step": 649 }, { "epoch": 0.2079778156996587, "grad_norm": 0.7387362122535706, "learning_rate": 6.929637526652453e-06, "loss": 0.5696, "step": 650 }, { "epoch": 0.20829778156996587, "grad_norm": 0.7688075304031372, "learning_rate": 6.9402985074626876e-06, "loss": 0.5871, "step": 651 }, { "epoch": 0.20861774744027303, "grad_norm": 0.6743102073669434, "learning_rate": 6.950959488272921e-06, "loss": 0.61, "step": 652 }, { "epoch": 0.2089377133105802, "grad_norm": 0.7545280456542969, "learning_rate": 6.961620469083156e-06, "loss": 0.614, "step": 653 }, { "epoch": 0.20925767918088736, "grad_norm": 0.694983720779419, "learning_rate": 6.972281449893391e-06, "loss": 0.6409, "step": 654 }, { "epoch": 0.20957764505119453, "grad_norm": 0.7406476736068726, "learning_rate": 6.982942430703626e-06, "loss": 0.6163, "step": 655 }, { "epoch": 0.2098976109215017, "grad_norm": 0.6860409379005432, "learning_rate": 6.993603411513859e-06, "loss": 0.5756, "step": 656 }, { "epoch": 0.21021757679180889, "grad_norm": 0.662331759929657, "learning_rate": 7.004264392324095e-06, "loss": 0.644, "step": 657 }, { "epoch": 0.21053754266211605, "grad_norm": 0.7123412489891052, "learning_rate": 7.014925373134329e-06, "loss": 0.6, "step": 658 }, { "epoch": 0.21085750853242322, "grad_norm": 0.7488113641738892, "learning_rate": 7.025586353944564e-06, "loss": 0.5972, "step": 659 }, { "epoch": 0.21117747440273038, "grad_norm": 0.7227650880813599, "learning_rate": 7.0362473347547975e-06, "loss": 0.622, "step": 660 }, { "epoch": 0.21149744027303755, "grad_norm": 0.7823781967163086, "learning_rate": 7.046908315565033e-06, "loss": 0.6404, "step": 661 }, { "epoch": 0.2118174061433447, "grad_norm": 0.6650978922843933, "learning_rate": 7.057569296375267e-06, "loss": 0.6049, "step": 662 }, { "epoch": 0.21213737201365188, "grad_norm": 0.6199343204498291, "learning_rate": 7.068230277185502e-06, "loss": 0.5808, "step": 663 }, { "epoch": 0.21245733788395904, "grad_norm": 0.7609500885009766, "learning_rate": 7.0788912579957356e-06, "loss": 0.6099, "step": 664 }, { "epoch": 0.2127773037542662, "grad_norm": 0.6560154557228088, "learning_rate": 7.089552238805971e-06, "loss": 0.59, "step": 665 }, { "epoch": 0.21309726962457337, "grad_norm": 0.6989269852638245, "learning_rate": 7.1002132196162055e-06, "loss": 0.6321, "step": 666 }, { "epoch": 0.21341723549488054, "grad_norm": 0.7098025679588318, "learning_rate": 7.11087420042644e-06, "loss": 0.5795, "step": 667 }, { "epoch": 0.2137372013651877, "grad_norm": 0.6742767691612244, "learning_rate": 7.121535181236674e-06, "loss": 0.6138, "step": 668 }, { "epoch": 0.2140571672354949, "grad_norm": 0.7612442970275879, "learning_rate": 7.132196162046909e-06, "loss": 0.6684, "step": 669 }, { "epoch": 0.21437713310580206, "grad_norm": 0.7280893325805664, "learning_rate": 7.1428571428571436e-06, "loss": 0.6422, "step": 670 }, { "epoch": 0.21469709897610922, "grad_norm": 0.6309712529182434, "learning_rate": 7.153518123667378e-06, "loss": 0.5463, "step": 671 }, { "epoch": 0.2150170648464164, "grad_norm": 0.6505151987075806, "learning_rate": 7.164179104477612e-06, "loss": 0.6028, "step": 672 }, { "epoch": 0.21533703071672355, "grad_norm": 0.6818749308586121, "learning_rate": 7.174840085287847e-06, "loss": 0.6038, "step": 673 }, { "epoch": 0.21565699658703072, "grad_norm": 0.6691836714744568, "learning_rate": 7.185501066098082e-06, "loss": 0.6081, "step": 674 }, { "epoch": 0.21597696245733788, "grad_norm": 0.6752950549125671, "learning_rate": 7.196162046908316e-06, "loss": 0.6063, "step": 675 }, { "epoch": 0.21629692832764505, "grad_norm": 0.7235035300254822, "learning_rate": 7.20682302771855e-06, "loss": 0.6094, "step": 676 }, { "epoch": 0.21661689419795221, "grad_norm": 0.6782038807868958, "learning_rate": 7.217484008528785e-06, "loss": 0.6016, "step": 677 }, { "epoch": 0.21693686006825938, "grad_norm": 0.7983608841896057, "learning_rate": 7.22814498933902e-06, "loss": 0.6242, "step": 678 }, { "epoch": 0.21725682593856654, "grad_norm": 0.8160451650619507, "learning_rate": 7.238805970149254e-06, "loss": 0.6348, "step": 679 }, { "epoch": 0.2175767918088737, "grad_norm": 0.808131754398346, "learning_rate": 7.249466950959488e-06, "loss": 0.6194, "step": 680 }, { "epoch": 0.21789675767918087, "grad_norm": 0.8144751787185669, "learning_rate": 7.260127931769723e-06, "loss": 0.6214, "step": 681 }, { "epoch": 0.21821672354948807, "grad_norm": 0.8087030053138733, "learning_rate": 7.270788912579958e-06, "loss": 0.6095, "step": 682 }, { "epoch": 0.21853668941979523, "grad_norm": 0.8073561191558838, "learning_rate": 7.281449893390192e-06, "loss": 0.6004, "step": 683 }, { "epoch": 0.2188566552901024, "grad_norm": 0.7714781761169434, "learning_rate": 7.292110874200427e-06, "loss": 0.5931, "step": 684 }, { "epoch": 0.21917662116040956, "grad_norm": 0.7731906771659851, "learning_rate": 7.302771855010662e-06, "loss": 0.5817, "step": 685 }, { "epoch": 0.21949658703071673, "grad_norm": 0.8933697938919067, "learning_rate": 7.313432835820896e-06, "loss": 0.6729, "step": 686 }, { "epoch": 0.2198165529010239, "grad_norm": 0.8103617429733276, "learning_rate": 7.3240938166311305e-06, "loss": 0.6505, "step": 687 }, { "epoch": 0.22013651877133106, "grad_norm": 0.6836149096488953, "learning_rate": 7.334754797441366e-06, "loss": 0.5906, "step": 688 }, { "epoch": 0.22045648464163822, "grad_norm": 0.8601536750793457, "learning_rate": 7.3454157782516e-06, "loss": 0.6226, "step": 689 }, { "epoch": 0.2207764505119454, "grad_norm": 0.8050239086151123, "learning_rate": 7.356076759061834e-06, "loss": 0.582, "step": 690 }, { "epoch": 0.22109641638225255, "grad_norm": 0.7658442258834839, "learning_rate": 7.366737739872069e-06, "loss": 0.5756, "step": 691 }, { "epoch": 0.22141638225255972, "grad_norm": 0.9899559020996094, "learning_rate": 7.377398720682304e-06, "loss": 0.661, "step": 692 }, { "epoch": 0.22173634812286688, "grad_norm": 0.663729190826416, "learning_rate": 7.3880597014925385e-06, "loss": 0.5754, "step": 693 }, { "epoch": 0.22205631399317405, "grad_norm": 0.6831468939781189, "learning_rate": 7.398720682302772e-06, "loss": 0.6276, "step": 694 }, { "epoch": 0.22237627986348124, "grad_norm": 0.7177518606185913, "learning_rate": 7.409381663113007e-06, "loss": 0.5654, "step": 695 }, { "epoch": 0.2226962457337884, "grad_norm": 0.8334975838661194, "learning_rate": 7.420042643923242e-06, "loss": 0.5664, "step": 696 }, { "epoch": 0.22301621160409557, "grad_norm": 0.7577218413352966, "learning_rate": 7.430703624733477e-06, "loss": 0.6004, "step": 697 }, { "epoch": 0.22333617747440274, "grad_norm": 0.8322723507881165, "learning_rate": 7.44136460554371e-06, "loss": 0.6105, "step": 698 }, { "epoch": 0.2236561433447099, "grad_norm": 0.8639805316925049, "learning_rate": 7.452025586353945e-06, "loss": 0.6352, "step": 699 }, { "epoch": 0.22397610921501707, "grad_norm": 0.7729711532592773, "learning_rate": 7.46268656716418e-06, "loss": 0.6066, "step": 700 }, { "epoch": 0.22429607508532423, "grad_norm": 0.906555712223053, "learning_rate": 7.473347547974415e-06, "loss": 0.6466, "step": 701 }, { "epoch": 0.2246160409556314, "grad_norm": 0.8130903244018555, "learning_rate": 7.484008528784648e-06, "loss": 0.5961, "step": 702 }, { "epoch": 0.22493600682593856, "grad_norm": 0.8586740493774414, "learning_rate": 7.494669509594883e-06, "loss": 0.5668, "step": 703 }, { "epoch": 0.22525597269624573, "grad_norm": 0.952433705329895, "learning_rate": 7.505330490405118e-06, "loss": 0.5882, "step": 704 }, { "epoch": 0.2255759385665529, "grad_norm": 0.7945285439491272, "learning_rate": 7.515991471215353e-06, "loss": 0.6012, "step": 705 }, { "epoch": 0.22589590443686006, "grad_norm": 0.7891091108322144, "learning_rate": 7.5266524520255865e-06, "loss": 0.6197, "step": 706 }, { "epoch": 0.22621587030716722, "grad_norm": 0.7546164393424988, "learning_rate": 7.537313432835821e-06, "loss": 0.5517, "step": 707 }, { "epoch": 0.22653583617747441, "grad_norm": 0.7216648459434509, "learning_rate": 7.547974413646056e-06, "loss": 0.6037, "step": 708 }, { "epoch": 0.22685580204778158, "grad_norm": 0.7756451368331909, "learning_rate": 7.558635394456291e-06, "loss": 0.6059, "step": 709 }, { "epoch": 0.22717576791808874, "grad_norm": 0.6714305281639099, "learning_rate": 7.569296375266525e-06, "loss": 0.6041, "step": 710 }, { "epoch": 0.2274957337883959, "grad_norm": 0.698634147644043, "learning_rate": 7.579957356076759e-06, "loss": 0.6272, "step": 711 }, { "epoch": 0.22781569965870307, "grad_norm": 0.7570346593856812, "learning_rate": 7.5906183368869945e-06, "loss": 0.5946, "step": 712 }, { "epoch": 0.22813566552901024, "grad_norm": 0.7321618795394897, "learning_rate": 7.601279317697229e-06, "loss": 0.5984, "step": 713 }, { "epoch": 0.2284556313993174, "grad_norm": 0.7243728637695312, "learning_rate": 7.611940298507463e-06, "loss": 0.6318, "step": 714 }, { "epoch": 0.22877559726962457, "grad_norm": 0.7606075406074524, "learning_rate": 7.622601279317697e-06, "loss": 0.577, "step": 715 }, { "epoch": 0.22909556313993173, "grad_norm": 0.5844424366950989, "learning_rate": 7.633262260127933e-06, "loss": 0.5897, "step": 716 }, { "epoch": 0.2294155290102389, "grad_norm": 0.6975888609886169, "learning_rate": 7.643923240938167e-06, "loss": 0.5822, "step": 717 }, { "epoch": 0.22973549488054607, "grad_norm": 0.7610205411911011, "learning_rate": 7.654584221748402e-06, "loss": 0.6291, "step": 718 }, { "epoch": 0.23005546075085323, "grad_norm": 0.6593003869056702, "learning_rate": 7.665245202558636e-06, "loss": 0.6267, "step": 719 }, { "epoch": 0.23037542662116042, "grad_norm": 0.7518547773361206, "learning_rate": 7.67590618336887e-06, "loss": 0.6667, "step": 720 }, { "epoch": 0.2306953924914676, "grad_norm": 0.6882115006446838, "learning_rate": 7.686567164179105e-06, "loss": 0.5997, "step": 721 }, { "epoch": 0.23101535836177475, "grad_norm": 0.647657036781311, "learning_rate": 7.69722814498934e-06, "loss": 0.622, "step": 722 }, { "epoch": 0.23133532423208192, "grad_norm": 0.6777190566062927, "learning_rate": 7.707889125799574e-06, "loss": 0.5917, "step": 723 }, { "epoch": 0.23165529010238908, "grad_norm": 0.7947143316268921, "learning_rate": 7.718550106609809e-06, "loss": 0.5911, "step": 724 }, { "epoch": 0.23197525597269625, "grad_norm": 0.6618469953536987, "learning_rate": 7.729211087420043e-06, "loss": 0.593, "step": 725 }, { "epoch": 0.2322952218430034, "grad_norm": 0.6656261682510376, "learning_rate": 7.739872068230278e-06, "loss": 0.6148, "step": 726 }, { "epoch": 0.23261518771331058, "grad_norm": 0.7834561467170715, "learning_rate": 7.750533049040512e-06, "loss": 0.6014, "step": 727 }, { "epoch": 0.23293515358361774, "grad_norm": 0.7411160469055176, "learning_rate": 7.761194029850747e-06, "loss": 0.5903, "step": 728 }, { "epoch": 0.2332551194539249, "grad_norm": 0.684743344783783, "learning_rate": 7.771855010660981e-06, "loss": 0.597, "step": 729 }, { "epoch": 0.23357508532423207, "grad_norm": 0.6873225569725037, "learning_rate": 7.782515991471216e-06, "loss": 0.5827, "step": 730 }, { "epoch": 0.23389505119453924, "grad_norm": 0.6318769454956055, "learning_rate": 7.79317697228145e-06, "loss": 0.5915, "step": 731 }, { "epoch": 0.2342150170648464, "grad_norm": 0.6759801506996155, "learning_rate": 7.803837953091685e-06, "loss": 0.6325, "step": 732 }, { "epoch": 0.2345349829351536, "grad_norm": 0.6570876836776733, "learning_rate": 7.81449893390192e-06, "loss": 0.616, "step": 733 }, { "epoch": 0.23485494880546076, "grad_norm": 0.7084457278251648, "learning_rate": 7.825159914712154e-06, "loss": 0.583, "step": 734 }, { "epoch": 0.23517491467576793, "grad_norm": 0.7025005221366882, "learning_rate": 7.835820895522389e-06, "loss": 0.5998, "step": 735 }, { "epoch": 0.2354948805460751, "grad_norm": 0.7843663692474365, "learning_rate": 7.846481876332623e-06, "loss": 0.6449, "step": 736 }, { "epoch": 0.23581484641638226, "grad_norm": 0.6961464285850525, "learning_rate": 7.857142857142858e-06, "loss": 0.6293, "step": 737 }, { "epoch": 0.23613481228668942, "grad_norm": 0.6443120837211609, "learning_rate": 7.867803837953092e-06, "loss": 0.5757, "step": 738 }, { "epoch": 0.2364547781569966, "grad_norm": 0.7188411951065063, "learning_rate": 7.878464818763327e-06, "loss": 0.5754, "step": 739 }, { "epoch": 0.23677474402730375, "grad_norm": 0.7644008994102478, "learning_rate": 7.889125799573561e-06, "loss": 0.6046, "step": 740 }, { "epoch": 0.23709470989761092, "grad_norm": 0.7195684313774109, "learning_rate": 7.899786780383796e-06, "loss": 0.5944, "step": 741 }, { "epoch": 0.23741467576791808, "grad_norm": 0.7057017087936401, "learning_rate": 7.91044776119403e-06, "loss": 0.5895, "step": 742 }, { "epoch": 0.23773464163822525, "grad_norm": 0.7110957503318787, "learning_rate": 7.921108742004265e-06, "loss": 0.6243, "step": 743 }, { "epoch": 0.2380546075085324, "grad_norm": 0.6075982451438904, "learning_rate": 7.9317697228145e-06, "loss": 0.5831, "step": 744 }, { "epoch": 0.23837457337883958, "grad_norm": 0.6880046129226685, "learning_rate": 7.942430703624734e-06, "loss": 0.56, "step": 745 }, { "epoch": 0.23869453924914677, "grad_norm": 0.7850193381309509, "learning_rate": 7.953091684434968e-06, "loss": 0.6023, "step": 746 }, { "epoch": 0.23901450511945393, "grad_norm": 0.8188655376434326, "learning_rate": 7.963752665245203e-06, "loss": 0.6267, "step": 747 }, { "epoch": 0.2393344709897611, "grad_norm": 0.7466493248939514, "learning_rate": 7.974413646055437e-06, "loss": 0.6166, "step": 748 }, { "epoch": 0.23965443686006827, "grad_norm": 0.6109229922294617, "learning_rate": 7.985074626865672e-06, "loss": 0.5878, "step": 749 }, { "epoch": 0.23997440273037543, "grad_norm": 0.7116419672966003, "learning_rate": 7.995735607675907e-06, "loss": 0.5625, "step": 750 }, { "epoch": 0.2402943686006826, "grad_norm": 0.7238773107528687, "learning_rate": 8.006396588486141e-06, "loss": 0.6143, "step": 751 }, { "epoch": 0.24061433447098976, "grad_norm": 0.611035943031311, "learning_rate": 8.017057569296376e-06, "loss": 0.6131, "step": 752 }, { "epoch": 0.24093430034129693, "grad_norm": 0.6717967391014099, "learning_rate": 8.02771855010661e-06, "loss": 0.5653, "step": 753 }, { "epoch": 0.2412542662116041, "grad_norm": 0.6449785828590393, "learning_rate": 8.038379530916846e-06, "loss": 0.5731, "step": 754 }, { "epoch": 0.24157423208191126, "grad_norm": 0.5837206840515137, "learning_rate": 8.049040511727079e-06, "loss": 0.5718, "step": 755 }, { "epoch": 0.24189419795221842, "grad_norm": 0.6526334285736084, "learning_rate": 8.059701492537314e-06, "loss": 0.6007, "step": 756 }, { "epoch": 0.24221416382252559, "grad_norm": 0.6099703907966614, "learning_rate": 8.070362473347548e-06, "loss": 0.5914, "step": 757 }, { "epoch": 0.24253412969283278, "grad_norm": 0.6639705896377563, "learning_rate": 8.081023454157784e-06, "loss": 0.601, "step": 758 }, { "epoch": 0.24285409556313994, "grad_norm": 0.66612309217453, "learning_rate": 8.091684434968017e-06, "loss": 0.5493, "step": 759 }, { "epoch": 0.2431740614334471, "grad_norm": 0.6732496023178101, "learning_rate": 8.102345415778252e-06, "loss": 0.5995, "step": 760 }, { "epoch": 0.24349402730375427, "grad_norm": 0.6970329284667969, "learning_rate": 8.113006396588486e-06, "loss": 0.5815, "step": 761 }, { "epoch": 0.24381399317406144, "grad_norm": 0.7264329791069031, "learning_rate": 8.123667377398723e-06, "loss": 0.6032, "step": 762 }, { "epoch": 0.2441339590443686, "grad_norm": 0.6926742792129517, "learning_rate": 8.134328358208955e-06, "loss": 0.571, "step": 763 }, { "epoch": 0.24445392491467577, "grad_norm": 0.7230720520019531, "learning_rate": 8.14498933901919e-06, "loss": 0.6081, "step": 764 }, { "epoch": 0.24477389078498293, "grad_norm": 0.7081652879714966, "learning_rate": 8.155650319829424e-06, "loss": 0.5771, "step": 765 }, { "epoch": 0.2450938566552901, "grad_norm": 0.6964278817176819, "learning_rate": 8.16631130063966e-06, "loss": 0.6184, "step": 766 }, { "epoch": 0.24541382252559726, "grad_norm": 0.6587214469909668, "learning_rate": 8.176972281449893e-06, "loss": 0.5783, "step": 767 }, { "epoch": 0.24573378839590443, "grad_norm": 0.7075352668762207, "learning_rate": 8.187633262260128e-06, "loss": 0.5982, "step": 768 }, { "epoch": 0.2460537542662116, "grad_norm": 0.6418300271034241, "learning_rate": 8.198294243070363e-06, "loss": 0.5858, "step": 769 }, { "epoch": 0.24637372013651876, "grad_norm": 0.5982336401939392, "learning_rate": 8.208955223880599e-06, "loss": 0.5749, "step": 770 }, { "epoch": 0.24669368600682595, "grad_norm": 0.696170449256897, "learning_rate": 8.219616204690832e-06, "loss": 0.5657, "step": 771 }, { "epoch": 0.24701365187713312, "grad_norm": 0.6634019017219543, "learning_rate": 8.230277185501066e-06, "loss": 0.5397, "step": 772 }, { "epoch": 0.24733361774744028, "grad_norm": 0.6658785939216614, "learning_rate": 8.2409381663113e-06, "loss": 0.5805, "step": 773 }, { "epoch": 0.24765358361774745, "grad_norm": 0.6904900670051575, "learning_rate": 8.251599147121537e-06, "loss": 0.5917, "step": 774 }, { "epoch": 0.2479735494880546, "grad_norm": 0.6993851065635681, "learning_rate": 8.26226012793177e-06, "loss": 0.5809, "step": 775 }, { "epoch": 0.24829351535836178, "grad_norm": 0.7204084396362305, "learning_rate": 8.272921108742004e-06, "loss": 0.5708, "step": 776 }, { "epoch": 0.24861348122866894, "grad_norm": 0.6606250405311584, "learning_rate": 8.283582089552239e-06, "loss": 0.5826, "step": 777 }, { "epoch": 0.2489334470989761, "grad_norm": 0.5864621996879578, "learning_rate": 8.294243070362475e-06, "loss": 0.5777, "step": 778 }, { "epoch": 0.24925341296928327, "grad_norm": 0.6498123407363892, "learning_rate": 8.304904051172708e-06, "loss": 0.6237, "step": 779 }, { "epoch": 0.24957337883959044, "grad_norm": 0.6379949450492859, "learning_rate": 8.315565031982942e-06, "loss": 0.5634, "step": 780 }, { "epoch": 0.2498933447098976, "grad_norm": 0.6023986339569092, "learning_rate": 8.326226012793177e-06, "loss": 0.5861, "step": 781 }, { "epoch": 0.2502133105802048, "grad_norm": 0.7081618309020996, "learning_rate": 8.336886993603413e-06, "loss": 0.5642, "step": 782 }, { "epoch": 0.25053327645051193, "grad_norm": 0.5907355546951294, "learning_rate": 8.347547974413648e-06, "loss": 0.5535, "step": 783 }, { "epoch": 0.2508532423208191, "grad_norm": 0.7312473058700562, "learning_rate": 8.35820895522388e-06, "loss": 0.5675, "step": 784 }, { "epoch": 0.25117320819112626, "grad_norm": 0.6056268215179443, "learning_rate": 8.368869936034117e-06, "loss": 0.5623, "step": 785 }, { "epoch": 0.25149317406143346, "grad_norm": 0.5965462923049927, "learning_rate": 8.379530916844351e-06, "loss": 0.6102, "step": 786 }, { "epoch": 0.2518131399317406, "grad_norm": 0.6213732957839966, "learning_rate": 8.390191897654586e-06, "loss": 0.575, "step": 787 }, { "epoch": 0.2521331058020478, "grad_norm": 0.6415621042251587, "learning_rate": 8.400852878464819e-06, "loss": 0.595, "step": 788 }, { "epoch": 0.2524530716723549, "grad_norm": 0.6275827884674072, "learning_rate": 8.411513859275055e-06, "loss": 0.5753, "step": 789 }, { "epoch": 0.2527730375426621, "grad_norm": 0.7514165639877319, "learning_rate": 8.42217484008529e-06, "loss": 0.6222, "step": 790 }, { "epoch": 0.2530930034129693, "grad_norm": 0.6514053344726562, "learning_rate": 8.432835820895524e-06, "loss": 0.5706, "step": 791 }, { "epoch": 0.25341296928327645, "grad_norm": 0.7108455300331116, "learning_rate": 8.443496801705757e-06, "loss": 0.6316, "step": 792 }, { "epoch": 0.25373293515358364, "grad_norm": 0.6999474763870239, "learning_rate": 8.454157782515993e-06, "loss": 0.5585, "step": 793 }, { "epoch": 0.2540529010238908, "grad_norm": 0.6626673936843872, "learning_rate": 8.464818763326227e-06, "loss": 0.5968, "step": 794 }, { "epoch": 0.25437286689419797, "grad_norm": 0.6229158043861389, "learning_rate": 8.475479744136462e-06, "loss": 0.6297, "step": 795 }, { "epoch": 0.2546928327645051, "grad_norm": 0.6660380959510803, "learning_rate": 8.486140724946695e-06, "loss": 0.6002, "step": 796 }, { "epoch": 0.2550127986348123, "grad_norm": 0.6904699206352234, "learning_rate": 8.496801705756931e-06, "loss": 0.605, "step": 797 }, { "epoch": 0.25533276450511944, "grad_norm": 0.7006338834762573, "learning_rate": 8.507462686567165e-06, "loss": 0.5965, "step": 798 }, { "epoch": 0.25565273037542663, "grad_norm": 0.6655489206314087, "learning_rate": 8.5181236673774e-06, "loss": 0.6186, "step": 799 }, { "epoch": 0.25597269624573377, "grad_norm": 0.7310318946838379, "learning_rate": 8.528784648187633e-06, "loss": 0.6423, "step": 800 }, { "epoch": 0.25629266211604096, "grad_norm": 0.737553060054779, "learning_rate": 8.539445628997869e-06, "loss": 0.6044, "step": 801 }, { "epoch": 0.2566126279863481, "grad_norm": 0.7621052265167236, "learning_rate": 8.550106609808104e-06, "loss": 0.6067, "step": 802 }, { "epoch": 0.2569325938566553, "grad_norm": 0.7435135245323181, "learning_rate": 8.560767590618338e-06, "loss": 0.5814, "step": 803 }, { "epoch": 0.2572525597269625, "grad_norm": 0.7449132204055786, "learning_rate": 8.571428571428571e-06, "loss": 0.6206, "step": 804 }, { "epoch": 0.2575725255972696, "grad_norm": 0.6484292149543762, "learning_rate": 8.582089552238807e-06, "loss": 0.6, "step": 805 }, { "epoch": 0.2578924914675768, "grad_norm": 0.6599040031433105, "learning_rate": 8.592750533049042e-06, "loss": 0.5526, "step": 806 }, { "epoch": 0.25821245733788395, "grad_norm": 0.7006849646568298, "learning_rate": 8.603411513859276e-06, "loss": 0.6108, "step": 807 }, { "epoch": 0.25853242320819114, "grad_norm": 0.6509926319122314, "learning_rate": 8.614072494669509e-06, "loss": 0.5835, "step": 808 }, { "epoch": 0.2588523890784983, "grad_norm": 0.6741862297058105, "learning_rate": 8.624733475479745e-06, "loss": 0.5516, "step": 809 }, { "epoch": 0.2591723549488055, "grad_norm": 0.6708396673202515, "learning_rate": 8.63539445628998e-06, "loss": 0.5815, "step": 810 }, { "epoch": 0.2594923208191126, "grad_norm": 0.7102819085121155, "learning_rate": 8.646055437100214e-06, "loss": 0.6043, "step": 811 }, { "epoch": 0.2598122866894198, "grad_norm": 0.6952850222587585, "learning_rate": 8.656716417910447e-06, "loss": 0.5943, "step": 812 }, { "epoch": 0.26013225255972694, "grad_norm": 0.7286235094070435, "learning_rate": 8.667377398720683e-06, "loss": 0.5735, "step": 813 }, { "epoch": 0.26045221843003413, "grad_norm": 0.6984187960624695, "learning_rate": 8.678038379530918e-06, "loss": 0.5924, "step": 814 }, { "epoch": 0.26077218430034127, "grad_norm": 0.6510578393936157, "learning_rate": 8.688699360341152e-06, "loss": 0.5999, "step": 815 }, { "epoch": 0.26109215017064846, "grad_norm": 0.6672604084014893, "learning_rate": 8.699360341151387e-06, "loss": 0.6006, "step": 816 }, { "epoch": 0.26141211604095566, "grad_norm": 0.6719098687171936, "learning_rate": 8.710021321961621e-06, "loss": 0.5706, "step": 817 }, { "epoch": 0.2617320819112628, "grad_norm": 0.6405746340751648, "learning_rate": 8.720682302771856e-06, "loss": 0.5847, "step": 818 }, { "epoch": 0.26205204778157, "grad_norm": 0.6537840962409973, "learning_rate": 8.73134328358209e-06, "loss": 0.5908, "step": 819 }, { "epoch": 0.2623720136518771, "grad_norm": 0.685456395149231, "learning_rate": 8.742004264392325e-06, "loss": 0.5854, "step": 820 }, { "epoch": 0.2626919795221843, "grad_norm": 0.7683976888656616, "learning_rate": 8.75266524520256e-06, "loss": 0.6454, "step": 821 }, { "epoch": 0.26301194539249145, "grad_norm": 0.6867132782936096, "learning_rate": 8.763326226012794e-06, "loss": 0.562, "step": 822 }, { "epoch": 0.26333191126279865, "grad_norm": 0.6225679516792297, "learning_rate": 8.773987206823029e-06, "loss": 0.6013, "step": 823 }, { "epoch": 0.2636518771331058, "grad_norm": 0.7322930097579956, "learning_rate": 8.784648187633263e-06, "loss": 0.5666, "step": 824 }, { "epoch": 0.263971843003413, "grad_norm": 0.6259191632270813, "learning_rate": 8.795309168443498e-06, "loss": 0.5527, "step": 825 }, { "epoch": 0.2642918088737201, "grad_norm": 0.7102184295654297, "learning_rate": 8.805970149253732e-06, "loss": 0.6152, "step": 826 }, { "epoch": 0.2646117747440273, "grad_norm": 0.7099477648735046, "learning_rate": 8.816631130063967e-06, "loss": 0.6105, "step": 827 }, { "epoch": 0.26493174061433444, "grad_norm": 0.7012109160423279, "learning_rate": 8.827292110874201e-06, "loss": 0.6027, "step": 828 }, { "epoch": 0.26525170648464164, "grad_norm": 0.765407145023346, "learning_rate": 8.837953091684436e-06, "loss": 0.5808, "step": 829 }, { "epoch": 0.26557167235494883, "grad_norm": 0.6566329598426819, "learning_rate": 8.84861407249467e-06, "loss": 0.5603, "step": 830 }, { "epoch": 0.26589163822525597, "grad_norm": 0.7441750764846802, "learning_rate": 8.859275053304905e-06, "loss": 0.5913, "step": 831 }, { "epoch": 0.26621160409556316, "grad_norm": 0.6909970045089722, "learning_rate": 8.86993603411514e-06, "loss": 0.602, "step": 832 }, { "epoch": 0.2665315699658703, "grad_norm": 0.6908165812492371, "learning_rate": 8.880597014925374e-06, "loss": 0.6129, "step": 833 }, { "epoch": 0.2668515358361775, "grad_norm": 0.73782879114151, "learning_rate": 8.891257995735608e-06, "loss": 0.568, "step": 834 }, { "epoch": 0.2671715017064846, "grad_norm": 0.6002166867256165, "learning_rate": 8.901918976545843e-06, "loss": 0.5844, "step": 835 }, { "epoch": 0.2674914675767918, "grad_norm": 0.5845295190811157, "learning_rate": 8.912579957356077e-06, "loss": 0.594, "step": 836 }, { "epoch": 0.26781143344709896, "grad_norm": 0.783158004283905, "learning_rate": 8.923240938166312e-06, "loss": 0.633, "step": 837 }, { "epoch": 0.26813139931740615, "grad_norm": 0.6837108135223389, "learning_rate": 8.933901918976547e-06, "loss": 0.5938, "step": 838 }, { "epoch": 0.2684513651877133, "grad_norm": 0.6747543811798096, "learning_rate": 8.944562899786781e-06, "loss": 0.5842, "step": 839 }, { "epoch": 0.2687713310580205, "grad_norm": 0.7062118053436279, "learning_rate": 8.955223880597016e-06, "loss": 0.5453, "step": 840 }, { "epoch": 0.2690912969283277, "grad_norm": 0.7308565974235535, "learning_rate": 8.96588486140725e-06, "loss": 0.6113, "step": 841 }, { "epoch": 0.2694112627986348, "grad_norm": 0.6708758473396301, "learning_rate": 8.976545842217485e-06, "loss": 0.6004, "step": 842 }, { "epoch": 0.269731228668942, "grad_norm": 0.7115057706832886, "learning_rate": 8.987206823027719e-06, "loss": 0.5802, "step": 843 }, { "epoch": 0.27005119453924914, "grad_norm": 0.6746699810028076, "learning_rate": 8.997867803837954e-06, "loss": 0.6277, "step": 844 }, { "epoch": 0.27037116040955633, "grad_norm": 0.7693876028060913, "learning_rate": 9.008528784648188e-06, "loss": 0.6117, "step": 845 }, { "epoch": 0.27069112627986347, "grad_norm": 0.6683522462844849, "learning_rate": 9.019189765458423e-06, "loss": 0.5984, "step": 846 }, { "epoch": 0.27101109215017066, "grad_norm": 0.6747117638587952, "learning_rate": 9.029850746268657e-06, "loss": 0.6316, "step": 847 }, { "epoch": 0.2713310580204778, "grad_norm": 0.6540120244026184, "learning_rate": 9.040511727078892e-06, "loss": 0.5743, "step": 848 }, { "epoch": 0.271651023890785, "grad_norm": 0.6651832461357117, "learning_rate": 9.051172707889126e-06, "loss": 0.615, "step": 849 }, { "epoch": 0.27197098976109213, "grad_norm": 0.5970577001571655, "learning_rate": 9.06183368869936e-06, "loss": 0.574, "step": 850 }, { "epoch": 0.2722909556313993, "grad_norm": 0.6609485745429993, "learning_rate": 9.072494669509595e-06, "loss": 0.5525, "step": 851 }, { "epoch": 0.27261092150170646, "grad_norm": 0.6449378132820129, "learning_rate": 9.08315565031983e-06, "loss": 0.6593, "step": 852 }, { "epoch": 0.27293088737201365, "grad_norm": 0.7078551054000854, "learning_rate": 9.093816631130064e-06, "loss": 0.6115, "step": 853 }, { "epoch": 0.27325085324232085, "grad_norm": 0.6257254481315613, "learning_rate": 9.104477611940299e-06, "loss": 0.568, "step": 854 }, { "epoch": 0.273570819112628, "grad_norm": 0.70318603515625, "learning_rate": 9.115138592750533e-06, "loss": 0.588, "step": 855 }, { "epoch": 0.2738907849829352, "grad_norm": 0.6586565971374512, "learning_rate": 9.125799573560768e-06, "loss": 0.5872, "step": 856 }, { "epoch": 0.2742107508532423, "grad_norm": 0.7630588412284851, "learning_rate": 9.136460554371003e-06, "loss": 0.6025, "step": 857 }, { "epoch": 0.2745307167235495, "grad_norm": 0.6974326968193054, "learning_rate": 9.147121535181237e-06, "loss": 0.6098, "step": 858 }, { "epoch": 0.27485068259385664, "grad_norm": 0.7954372763633728, "learning_rate": 9.157782515991472e-06, "loss": 0.6076, "step": 859 }, { "epoch": 0.27517064846416384, "grad_norm": 0.717241644859314, "learning_rate": 9.168443496801706e-06, "loss": 0.6285, "step": 860 }, { "epoch": 0.275490614334471, "grad_norm": 0.7189328074455261, "learning_rate": 9.17910447761194e-06, "loss": 0.605, "step": 861 }, { "epoch": 0.27581058020477817, "grad_norm": 0.8277390003204346, "learning_rate": 9.189765458422175e-06, "loss": 0.589, "step": 862 }, { "epoch": 0.2761305460750853, "grad_norm": 0.6832088232040405, "learning_rate": 9.200426439232411e-06, "loss": 0.6062, "step": 863 }, { "epoch": 0.2764505119453925, "grad_norm": 0.653217077255249, "learning_rate": 9.211087420042644e-06, "loss": 0.5561, "step": 864 }, { "epoch": 0.27677047781569963, "grad_norm": 0.759894073009491, "learning_rate": 9.221748400852879e-06, "loss": 0.6274, "step": 865 }, { "epoch": 0.2770904436860068, "grad_norm": 0.7022303938865662, "learning_rate": 9.232409381663113e-06, "loss": 0.5755, "step": 866 }, { "epoch": 0.277410409556314, "grad_norm": 0.7041699886322021, "learning_rate": 9.24307036247335e-06, "loss": 0.5774, "step": 867 }, { "epoch": 0.27773037542662116, "grad_norm": 0.6180068850517273, "learning_rate": 9.253731343283582e-06, "loss": 0.5866, "step": 868 }, { "epoch": 0.27805034129692835, "grad_norm": 0.7739641666412354, "learning_rate": 9.264392324093817e-06, "loss": 0.6054, "step": 869 }, { "epoch": 0.2783703071672355, "grad_norm": 0.7242615818977356, "learning_rate": 9.275053304904051e-06, "loss": 0.5805, "step": 870 }, { "epoch": 0.2786902730375427, "grad_norm": 0.8257843255996704, "learning_rate": 9.285714285714288e-06, "loss": 0.5771, "step": 871 }, { "epoch": 0.2790102389078498, "grad_norm": 0.8289785385131836, "learning_rate": 9.29637526652452e-06, "loss": 0.6556, "step": 872 }, { "epoch": 0.279330204778157, "grad_norm": 0.6636307835578918, "learning_rate": 9.307036247334755e-06, "loss": 0.6051, "step": 873 }, { "epoch": 0.27965017064846415, "grad_norm": 0.7275445461273193, "learning_rate": 9.31769722814499e-06, "loss": 0.5676, "step": 874 }, { "epoch": 0.27997013651877134, "grad_norm": 0.6746795773506165, "learning_rate": 9.328358208955226e-06, "loss": 0.5767, "step": 875 }, { "epoch": 0.2802901023890785, "grad_norm": 0.6168851256370544, "learning_rate": 9.339019189765458e-06, "loss": 0.6061, "step": 876 }, { "epoch": 0.28061006825938567, "grad_norm": 0.7217643857002258, "learning_rate": 9.349680170575693e-06, "loss": 0.6003, "step": 877 }, { "epoch": 0.2809300341296928, "grad_norm": 0.6689902544021606, "learning_rate": 9.36034115138593e-06, "loss": 0.6247, "step": 878 }, { "epoch": 0.28125, "grad_norm": 0.6647174954414368, "learning_rate": 9.371002132196164e-06, "loss": 0.603, "step": 879 }, { "epoch": 0.2815699658703072, "grad_norm": 0.662013828754425, "learning_rate": 9.381663113006397e-06, "loss": 0.6026, "step": 880 }, { "epoch": 0.28188993174061433, "grad_norm": 0.6372009515762329, "learning_rate": 9.392324093816631e-06, "loss": 0.5461, "step": 881 }, { "epoch": 0.2822098976109215, "grad_norm": 0.7451372146606445, "learning_rate": 9.402985074626867e-06, "loss": 0.6046, "step": 882 }, { "epoch": 0.28252986348122866, "grad_norm": 0.6218172907829285, "learning_rate": 9.413646055437102e-06, "loss": 0.5474, "step": 883 }, { "epoch": 0.28284982935153585, "grad_norm": 0.5694857835769653, "learning_rate": 9.424307036247335e-06, "loss": 0.5917, "step": 884 }, { "epoch": 0.283169795221843, "grad_norm": 0.5968525409698486, "learning_rate": 9.43496801705757e-06, "loss": 0.5733, "step": 885 }, { "epoch": 0.2834897610921502, "grad_norm": 0.7009020447731018, "learning_rate": 9.445628997867805e-06, "loss": 0.6102, "step": 886 }, { "epoch": 0.2838097269624573, "grad_norm": 0.654617965221405, "learning_rate": 9.45628997867804e-06, "loss": 0.6114, "step": 887 }, { "epoch": 0.2841296928327645, "grad_norm": 0.6574299931526184, "learning_rate": 9.466950959488273e-06, "loss": 0.6115, "step": 888 }, { "epoch": 0.28444965870307165, "grad_norm": 0.7500622868537903, "learning_rate": 9.477611940298507e-06, "loss": 0.6424, "step": 889 }, { "epoch": 0.28476962457337884, "grad_norm": 0.6697449684143066, "learning_rate": 9.488272921108744e-06, "loss": 0.5415, "step": 890 }, { "epoch": 0.285089590443686, "grad_norm": 0.6367616653442383, "learning_rate": 9.498933901918978e-06, "loss": 0.5739, "step": 891 }, { "epoch": 0.2854095563139932, "grad_norm": 0.6737374663352966, "learning_rate": 9.509594882729211e-06, "loss": 0.5941, "step": 892 }, { "epoch": 0.28572952218430037, "grad_norm": 0.6352681517601013, "learning_rate": 9.520255863539445e-06, "loss": 0.588, "step": 893 }, { "epoch": 0.2860494880546075, "grad_norm": 0.6935127377510071, "learning_rate": 9.530916844349682e-06, "loss": 0.5845, "step": 894 }, { "epoch": 0.2863694539249147, "grad_norm": 0.6597230434417725, "learning_rate": 9.541577825159916e-06, "loss": 0.6159, "step": 895 }, { "epoch": 0.28668941979522183, "grad_norm": 0.7148585319519043, "learning_rate": 9.552238805970149e-06, "loss": 0.5976, "step": 896 }, { "epoch": 0.287009385665529, "grad_norm": 0.7132214307785034, "learning_rate": 9.562899786780384e-06, "loss": 0.6219, "step": 897 }, { "epoch": 0.28732935153583616, "grad_norm": 0.73183673620224, "learning_rate": 9.57356076759062e-06, "loss": 0.5652, "step": 898 }, { "epoch": 0.28764931740614336, "grad_norm": 0.7999463677406311, "learning_rate": 9.584221748400854e-06, "loss": 0.558, "step": 899 }, { "epoch": 0.2879692832764505, "grad_norm": 0.8034552335739136, "learning_rate": 9.594882729211089e-06, "loss": 0.6331, "step": 900 }, { "epoch": 0.2882892491467577, "grad_norm": 0.660112738609314, "learning_rate": 9.605543710021322e-06, "loss": 0.5843, "step": 901 }, { "epoch": 0.2886092150170648, "grad_norm": 0.8363476395606995, "learning_rate": 9.616204690831558e-06, "loss": 0.5984, "step": 902 }, { "epoch": 0.288929180887372, "grad_norm": 0.7044869065284729, "learning_rate": 9.626865671641792e-06, "loss": 0.584, "step": 903 }, { "epoch": 0.28924914675767915, "grad_norm": 0.6434544920921326, "learning_rate": 9.637526652452027e-06, "loss": 0.5776, "step": 904 }, { "epoch": 0.28956911262798635, "grad_norm": 0.7166516780853271, "learning_rate": 9.64818763326226e-06, "loss": 0.5585, "step": 905 }, { "epoch": 0.28988907849829354, "grad_norm": 0.6601057052612305, "learning_rate": 9.658848614072496e-06, "loss": 0.5893, "step": 906 }, { "epoch": 0.2902090443686007, "grad_norm": 0.6507968902587891, "learning_rate": 9.66950959488273e-06, "loss": 0.633, "step": 907 }, { "epoch": 0.29052901023890787, "grad_norm": 0.6383233070373535, "learning_rate": 9.680170575692965e-06, "loss": 0.5559, "step": 908 }, { "epoch": 0.290848976109215, "grad_norm": 0.6499894857406616, "learning_rate": 9.6908315565032e-06, "loss": 0.5959, "step": 909 }, { "epoch": 0.2911689419795222, "grad_norm": 0.62410569190979, "learning_rate": 9.701492537313434e-06, "loss": 0.5943, "step": 910 }, { "epoch": 0.29148890784982934, "grad_norm": 0.7339252829551697, "learning_rate": 9.712153518123669e-06, "loss": 0.5782, "step": 911 }, { "epoch": 0.29180887372013653, "grad_norm": 0.6893714666366577, "learning_rate": 9.722814498933903e-06, "loss": 0.6149, "step": 912 }, { "epoch": 0.29212883959044367, "grad_norm": 0.6200568079948425, "learning_rate": 9.733475479744138e-06, "loss": 0.5349, "step": 913 }, { "epoch": 0.29244880546075086, "grad_norm": 0.6621658205986023, "learning_rate": 9.744136460554372e-06, "loss": 0.5936, "step": 914 }, { "epoch": 0.292768771331058, "grad_norm": 0.6932737827301025, "learning_rate": 9.754797441364607e-06, "loss": 0.6119, "step": 915 }, { "epoch": 0.2930887372013652, "grad_norm": 0.6547825932502747, "learning_rate": 9.765458422174841e-06, "loss": 0.557, "step": 916 }, { "epoch": 0.2934087030716723, "grad_norm": 0.7168145179748535, "learning_rate": 9.776119402985076e-06, "loss": 0.591, "step": 917 }, { "epoch": 0.2937286689419795, "grad_norm": 0.7716021537780762, "learning_rate": 9.78678038379531e-06, "loss": 0.5465, "step": 918 }, { "epoch": 0.2940486348122867, "grad_norm": 0.6409813165664673, "learning_rate": 9.797441364605545e-06, "loss": 0.6201, "step": 919 }, { "epoch": 0.29436860068259385, "grad_norm": 0.724183201789856, "learning_rate": 9.80810234541578e-06, "loss": 0.5796, "step": 920 }, { "epoch": 0.29468856655290104, "grad_norm": 0.7610300183296204, "learning_rate": 9.818763326226014e-06, "loss": 0.5688, "step": 921 }, { "epoch": 0.2950085324232082, "grad_norm": 0.6418836116790771, "learning_rate": 9.829424307036248e-06, "loss": 0.6027, "step": 922 }, { "epoch": 0.2953284982935154, "grad_norm": 0.678600549697876, "learning_rate": 9.840085287846483e-06, "loss": 0.5505, "step": 923 }, { "epoch": 0.2956484641638225, "grad_norm": 0.7092781066894531, "learning_rate": 9.850746268656717e-06, "loss": 0.5945, "step": 924 }, { "epoch": 0.2959684300341297, "grad_norm": 0.6756407022476196, "learning_rate": 9.861407249466952e-06, "loss": 0.5891, "step": 925 }, { "epoch": 0.29628839590443684, "grad_norm": 0.6679978966712952, "learning_rate": 9.872068230277187e-06, "loss": 0.5535, "step": 926 }, { "epoch": 0.29660836177474403, "grad_norm": 0.6598780751228333, "learning_rate": 9.882729211087421e-06, "loss": 0.612, "step": 927 }, { "epoch": 0.29692832764505117, "grad_norm": 0.6880948543548584, "learning_rate": 9.893390191897656e-06, "loss": 0.6034, "step": 928 }, { "epoch": 0.29724829351535836, "grad_norm": 0.7000877857208252, "learning_rate": 9.90405117270789e-06, "loss": 0.5231, "step": 929 }, { "epoch": 0.29756825938566556, "grad_norm": 0.7053589820861816, "learning_rate": 9.914712153518125e-06, "loss": 0.5982, "step": 930 }, { "epoch": 0.2978882252559727, "grad_norm": 0.7295544743537903, "learning_rate": 9.925373134328359e-06, "loss": 0.5584, "step": 931 }, { "epoch": 0.2982081911262799, "grad_norm": 0.657467782497406, "learning_rate": 9.936034115138594e-06, "loss": 0.568, "step": 932 }, { "epoch": 0.298528156996587, "grad_norm": 0.716752290725708, "learning_rate": 9.946695095948828e-06, "loss": 0.5728, "step": 933 }, { "epoch": 0.2988481228668942, "grad_norm": 0.6869081854820251, "learning_rate": 9.957356076759063e-06, "loss": 0.5801, "step": 934 }, { "epoch": 0.29916808873720135, "grad_norm": 0.6525644063949585, "learning_rate": 9.968017057569297e-06, "loss": 0.5332, "step": 935 }, { "epoch": 0.29948805460750855, "grad_norm": 0.7127978205680847, "learning_rate": 9.978678038379532e-06, "loss": 0.6113, "step": 936 }, { "epoch": 0.2998080204778157, "grad_norm": 0.7229118943214417, "learning_rate": 9.989339019189766e-06, "loss": 0.6064, "step": 937 }, { "epoch": 0.3001279863481229, "grad_norm": 0.8196436762809753, "learning_rate": 1e-05, "loss": 0.5811, "step": 938 }, { "epoch": 0.30044795221843, "grad_norm": 0.8263665437698364, "learning_rate": 9.999999653371993e-06, "loss": 0.6201, "step": 939 }, { "epoch": 0.3007679180887372, "grad_norm": 0.84651118516922, "learning_rate": 9.999998613488017e-06, "loss": 0.594, "step": 940 }, { "epoch": 0.30108788395904434, "grad_norm": 0.7862319350242615, "learning_rate": 9.999996880348222e-06, "loss": 0.6073, "step": 941 }, { "epoch": 0.30140784982935154, "grad_norm": 0.7739344239234924, "learning_rate": 9.999994453952842e-06, "loss": 0.6434, "step": 942 }, { "epoch": 0.30172781569965873, "grad_norm": 0.6697371006011963, "learning_rate": 9.999991334302215e-06, "loss": 0.5778, "step": 943 }, { "epoch": 0.30204778156996587, "grad_norm": 0.7090111970901489, "learning_rate": 9.999987521396774e-06, "loss": 0.5715, "step": 944 }, { "epoch": 0.30236774744027306, "grad_norm": 0.7254164218902588, "learning_rate": 9.99998301523705e-06, "loss": 0.5908, "step": 945 }, { "epoch": 0.3026877133105802, "grad_norm": 0.74472975730896, "learning_rate": 9.999977815823665e-06, "loss": 0.6066, "step": 946 }, { "epoch": 0.3030076791808874, "grad_norm": 0.7350637316703796, "learning_rate": 9.99997192315734e-06, "loss": 0.5776, "step": 947 }, { "epoch": 0.3033276450511945, "grad_norm": 0.6912806034088135, "learning_rate": 9.999965337238892e-06, "loss": 0.5896, "step": 948 }, { "epoch": 0.3036476109215017, "grad_norm": 0.7248418927192688, "learning_rate": 9.999958058069237e-06, "loss": 0.6134, "step": 949 }, { "epoch": 0.30396757679180886, "grad_norm": 0.7095966339111328, "learning_rate": 9.999950085649382e-06, "loss": 0.5876, "step": 950 }, { "epoch": 0.30428754266211605, "grad_norm": 0.7419503331184387, "learning_rate": 9.999941419980432e-06, "loss": 0.5915, "step": 951 }, { "epoch": 0.3046075085324232, "grad_norm": 0.6916894912719727, "learning_rate": 9.99993206106359e-06, "loss": 0.5961, "step": 952 }, { "epoch": 0.3049274744027304, "grad_norm": 0.7504989504814148, "learning_rate": 9.99992200890015e-06, "loss": 0.6087, "step": 953 }, { "epoch": 0.3052474402730375, "grad_norm": 0.624904453754425, "learning_rate": 9.99991126349151e-06, "loss": 0.6079, "step": 954 }, { "epoch": 0.3055674061433447, "grad_norm": 0.671826958656311, "learning_rate": 9.999899824839159e-06, "loss": 0.5845, "step": 955 }, { "epoch": 0.3058873720136519, "grad_norm": 0.7129992246627808, "learning_rate": 9.999887692944681e-06, "loss": 0.5992, "step": 956 }, { "epoch": 0.30620733788395904, "grad_norm": 0.6746633052825928, "learning_rate": 9.99987486780976e-06, "loss": 0.5832, "step": 957 }, { "epoch": 0.30652730375426623, "grad_norm": 0.6484116911888123, "learning_rate": 9.999861349436174e-06, "loss": 0.5421, "step": 958 }, { "epoch": 0.30684726962457337, "grad_norm": 0.7942398190498352, "learning_rate": 9.999847137825795e-06, "loss": 0.5889, "step": 959 }, { "epoch": 0.30716723549488056, "grad_norm": 0.6215122938156128, "learning_rate": 9.999832232980598e-06, "loss": 0.5349, "step": 960 }, { "epoch": 0.3074872013651877, "grad_norm": 0.6754631400108337, "learning_rate": 9.999816634902645e-06, "loss": 0.5662, "step": 961 }, { "epoch": 0.3078071672354949, "grad_norm": 0.6522238850593567, "learning_rate": 9.999800343594101e-06, "loss": 0.6217, "step": 962 }, { "epoch": 0.30812713310580203, "grad_norm": 0.6306779384613037, "learning_rate": 9.999783359057226e-06, "loss": 0.5367, "step": 963 }, { "epoch": 0.3084470989761092, "grad_norm": 0.6738232374191284, "learning_rate": 9.999765681294371e-06, "loss": 0.6207, "step": 964 }, { "epoch": 0.30876706484641636, "grad_norm": 0.774307906627655, "learning_rate": 9.999747310307993e-06, "loss": 0.6226, "step": 965 }, { "epoch": 0.30908703071672355, "grad_norm": 0.6412689685821533, "learning_rate": 9.999728246100632e-06, "loss": 0.5872, "step": 966 }, { "epoch": 0.3094069965870307, "grad_norm": 0.6333892941474915, "learning_rate": 9.999708488674936e-06, "loss": 0.5863, "step": 967 }, { "epoch": 0.3097269624573379, "grad_norm": 0.6409503817558289, "learning_rate": 9.999688038033643e-06, "loss": 0.5536, "step": 968 }, { "epoch": 0.3100469283276451, "grad_norm": 0.6600871086120605, "learning_rate": 9.99966689417959e-06, "loss": 0.5638, "step": 969 }, { "epoch": 0.3103668941979522, "grad_norm": 0.6778677701950073, "learning_rate": 9.999645057115706e-06, "loss": 0.5806, "step": 970 }, { "epoch": 0.3106868600682594, "grad_norm": 0.6836385726928711, "learning_rate": 9.99962252684502e-06, "loss": 0.5788, "step": 971 }, { "epoch": 0.31100682593856654, "grad_norm": 0.6408164501190186, "learning_rate": 9.999599303370658e-06, "loss": 0.5407, "step": 972 }, { "epoch": 0.31132679180887374, "grad_norm": 0.7385551929473877, "learning_rate": 9.999575386695834e-06, "loss": 0.5931, "step": 973 }, { "epoch": 0.3116467576791809, "grad_norm": 0.7113818526268005, "learning_rate": 9.999550776823872e-06, "loss": 0.5816, "step": 974 }, { "epoch": 0.31196672354948807, "grad_norm": 0.6312804818153381, "learning_rate": 9.999525473758178e-06, "loss": 0.5758, "step": 975 }, { "epoch": 0.3122866894197952, "grad_norm": 0.6690606474876404, "learning_rate": 9.999499477502263e-06, "loss": 0.594, "step": 976 }, { "epoch": 0.3126066552901024, "grad_norm": 0.7593414187431335, "learning_rate": 9.999472788059732e-06, "loss": 0.6296, "step": 977 }, { "epoch": 0.31292662116040953, "grad_norm": 0.6702037453651428, "learning_rate": 9.999445405434283e-06, "loss": 0.5851, "step": 978 }, { "epoch": 0.3132465870307167, "grad_norm": 0.6028419137001038, "learning_rate": 9.999417329629715e-06, "loss": 0.5407, "step": 979 }, { "epoch": 0.31356655290102387, "grad_norm": 0.7517277002334595, "learning_rate": 9.999388560649919e-06, "loss": 0.6294, "step": 980 }, { "epoch": 0.31388651877133106, "grad_norm": 0.6471433639526367, "learning_rate": 9.999359098498884e-06, "loss": 0.5681, "step": 981 }, { "epoch": 0.31420648464163825, "grad_norm": 0.6184797286987305, "learning_rate": 9.999328943180696e-06, "loss": 0.5607, "step": 982 }, { "epoch": 0.3145264505119454, "grad_norm": 0.652275562286377, "learning_rate": 9.999298094699538e-06, "loss": 0.57, "step": 983 }, { "epoch": 0.3148464163822526, "grad_norm": 0.6612683534622192, "learning_rate": 9.999266553059683e-06, "loss": 0.5986, "step": 984 }, { "epoch": 0.3151663822525597, "grad_norm": 0.6565192937850952, "learning_rate": 9.999234318265507e-06, "loss": 0.5933, "step": 985 }, { "epoch": 0.3154863481228669, "grad_norm": 0.6525236964225769, "learning_rate": 9.999201390321476e-06, "loss": 0.574, "step": 986 }, { "epoch": 0.31580631399317405, "grad_norm": 0.6861425638198853, "learning_rate": 9.999167769232162e-06, "loss": 0.6212, "step": 987 }, { "epoch": 0.31612627986348124, "grad_norm": 0.5889517068862915, "learning_rate": 9.999133455002219e-06, "loss": 0.5596, "step": 988 }, { "epoch": 0.3164462457337884, "grad_norm": 0.5472261905670166, "learning_rate": 9.999098447636411e-06, "loss": 0.5732, "step": 989 }, { "epoch": 0.31676621160409557, "grad_norm": 0.6761401891708374, "learning_rate": 9.999062747139587e-06, "loss": 0.5878, "step": 990 }, { "epoch": 0.3170861774744027, "grad_norm": 0.7099294662475586, "learning_rate": 9.999026353516702e-06, "loss": 0.6142, "step": 991 }, { "epoch": 0.3174061433447099, "grad_norm": 0.7079955339431763, "learning_rate": 9.998989266772797e-06, "loss": 0.5645, "step": 992 }, { "epoch": 0.31772610921501704, "grad_norm": 0.6157618761062622, "learning_rate": 9.998951486913015e-06, "loss": 0.5805, "step": 993 }, { "epoch": 0.31804607508532423, "grad_norm": 0.6414501667022705, "learning_rate": 9.998913013942597e-06, "loss": 0.5719, "step": 994 }, { "epoch": 0.3183660409556314, "grad_norm": 0.7088600993156433, "learning_rate": 9.998873847866876e-06, "loss": 0.5769, "step": 995 }, { "epoch": 0.31868600682593856, "grad_norm": 0.6993066668510437, "learning_rate": 9.998833988691281e-06, "loss": 0.5656, "step": 996 }, { "epoch": 0.31900597269624575, "grad_norm": 0.700746476650238, "learning_rate": 9.998793436421342e-06, "loss": 0.6116, "step": 997 }, { "epoch": 0.3193259385665529, "grad_norm": 0.7170496582984924, "learning_rate": 9.998752191062678e-06, "loss": 0.6043, "step": 998 }, { "epoch": 0.3196459044368601, "grad_norm": 0.7763385772705078, "learning_rate": 9.998710252621009e-06, "loss": 0.5861, "step": 999 }, { "epoch": 0.3199658703071672, "grad_norm": 0.6285117268562317, "learning_rate": 9.99866762110215e-06, "loss": 0.5729, "step": 1000 }, { "epoch": 0.3202858361774744, "grad_norm": 0.7778967618942261, "learning_rate": 9.998624296512011e-06, "loss": 0.6236, "step": 1001 }, { "epoch": 0.32060580204778155, "grad_norm": 0.6404998302459717, "learning_rate": 9.9985802788566e-06, "loss": 0.5518, "step": 1002 }, { "epoch": 0.32092576791808874, "grad_norm": 0.7212223410606384, "learning_rate": 9.99853556814202e-06, "loss": 0.5896, "step": 1003 }, { "epoch": 0.3212457337883959, "grad_norm": 0.634592592716217, "learning_rate": 9.998490164374472e-06, "loss": 0.6042, "step": 1004 }, { "epoch": 0.3215656996587031, "grad_norm": 0.676922082901001, "learning_rate": 9.998444067560247e-06, "loss": 0.6182, "step": 1005 }, { "epoch": 0.3218856655290102, "grad_norm": 0.6989039182662964, "learning_rate": 9.998397277705741e-06, "loss": 0.5595, "step": 1006 }, { "epoch": 0.3222056313993174, "grad_norm": 0.628720223903656, "learning_rate": 9.998349794817436e-06, "loss": 0.5595, "step": 1007 }, { "epoch": 0.3225255972696246, "grad_norm": 0.7136312127113342, "learning_rate": 9.998301618901923e-06, "loss": 0.594, "step": 1008 }, { "epoch": 0.32284556313993173, "grad_norm": 0.5778558254241943, "learning_rate": 9.998252749965873e-06, "loss": 0.5656, "step": 1009 }, { "epoch": 0.3231655290102389, "grad_norm": 0.713370144367218, "learning_rate": 9.99820318801607e-06, "loss": 0.6236, "step": 1010 }, { "epoch": 0.32348549488054607, "grad_norm": 0.6834118366241455, "learning_rate": 9.99815293305938e-06, "loss": 0.629, "step": 1011 }, { "epoch": 0.32380546075085326, "grad_norm": 0.6257147789001465, "learning_rate": 9.998101985102773e-06, "loss": 0.5491, "step": 1012 }, { "epoch": 0.3241254266211604, "grad_norm": 0.7445933818817139, "learning_rate": 9.998050344153313e-06, "loss": 0.6007, "step": 1013 }, { "epoch": 0.3244453924914676, "grad_norm": 0.7208285927772522, "learning_rate": 9.99799801021816e-06, "loss": 0.5681, "step": 1014 }, { "epoch": 0.3247653583617747, "grad_norm": 0.6081100106239319, "learning_rate": 9.99794498330457e-06, "loss": 0.5463, "step": 1015 }, { "epoch": 0.3250853242320819, "grad_norm": 0.6651808619499207, "learning_rate": 9.997891263419896e-06, "loss": 0.581, "step": 1016 }, { "epoch": 0.32540529010238906, "grad_norm": 0.5771706700325012, "learning_rate": 9.997836850571584e-06, "loss": 0.5767, "step": 1017 }, { "epoch": 0.32572525597269625, "grad_norm": 0.7077763080596924, "learning_rate": 9.997781744767182e-06, "loss": 0.5861, "step": 1018 }, { "epoch": 0.3260452218430034, "grad_norm": 0.6261406540870667, "learning_rate": 9.997725946014328e-06, "loss": 0.5665, "step": 1019 }, { "epoch": 0.3263651877133106, "grad_norm": 0.6279311776161194, "learning_rate": 9.997669454320759e-06, "loss": 0.5788, "step": 1020 }, { "epoch": 0.32668515358361777, "grad_norm": 0.7319290041923523, "learning_rate": 9.997612269694308e-06, "loss": 0.5647, "step": 1021 }, { "epoch": 0.3270051194539249, "grad_norm": 0.5971276760101318, "learning_rate": 9.997554392142903e-06, "loss": 0.5847, "step": 1022 }, { "epoch": 0.3273250853242321, "grad_norm": 0.6742677092552185, "learning_rate": 9.997495821674572e-06, "loss": 0.6257, "step": 1023 }, { "epoch": 0.32764505119453924, "grad_norm": 0.7395401000976562, "learning_rate": 9.99743655829743e-06, "loss": 0.6391, "step": 1024 }, { "epoch": 0.32796501706484643, "grad_norm": 0.6834313273429871, "learning_rate": 9.997376602019698e-06, "loss": 0.5521, "step": 1025 }, { "epoch": 0.32828498293515357, "grad_norm": 0.7318760752677917, "learning_rate": 9.99731595284969e-06, "loss": 0.589, "step": 1026 }, { "epoch": 0.32860494880546076, "grad_norm": 0.6454777717590332, "learning_rate": 9.99725461079581e-06, "loss": 0.6045, "step": 1027 }, { "epoch": 0.3289249146757679, "grad_norm": 0.6907580494880676, "learning_rate": 9.997192575866569e-06, "loss": 0.5911, "step": 1028 }, { "epoch": 0.3292448805460751, "grad_norm": 0.7321333885192871, "learning_rate": 9.997129848070563e-06, "loss": 0.5866, "step": 1029 }, { "epoch": 0.32956484641638223, "grad_norm": 0.6222965717315674, "learning_rate": 9.997066427416494e-06, "loss": 0.5784, "step": 1030 }, { "epoch": 0.3298848122866894, "grad_norm": 0.7147840261459351, "learning_rate": 9.997002313913153e-06, "loss": 0.5809, "step": 1031 }, { "epoch": 0.3302047781569966, "grad_norm": 0.7093286514282227, "learning_rate": 9.996937507569429e-06, "loss": 0.5888, "step": 1032 }, { "epoch": 0.33052474402730375, "grad_norm": 0.6205602884292603, "learning_rate": 9.996872008394307e-06, "loss": 0.5634, "step": 1033 }, { "epoch": 0.33084470989761094, "grad_norm": 0.6194776892662048, "learning_rate": 9.996805816396873e-06, "loss": 0.5492, "step": 1034 }, { "epoch": 0.3311646757679181, "grad_norm": 0.6360750198364258, "learning_rate": 9.996738931586298e-06, "loss": 0.6012, "step": 1035 }, { "epoch": 0.3314846416382253, "grad_norm": 0.646390438079834, "learning_rate": 9.996671353971861e-06, "loss": 0.5748, "step": 1036 }, { "epoch": 0.3318046075085324, "grad_norm": 0.6864686608314514, "learning_rate": 9.996603083562928e-06, "loss": 0.5952, "step": 1037 }, { "epoch": 0.3321245733788396, "grad_norm": 0.6476268172264099, "learning_rate": 9.996534120368968e-06, "loss": 0.5659, "step": 1038 }, { "epoch": 0.33244453924914674, "grad_norm": 0.651563823223114, "learning_rate": 9.99646446439954e-06, "loss": 0.6185, "step": 1039 }, { "epoch": 0.33276450511945393, "grad_norm": 0.6136869192123413, "learning_rate": 9.996394115664302e-06, "loss": 0.5927, "step": 1040 }, { "epoch": 0.3330844709897611, "grad_norm": 0.6799325346946716, "learning_rate": 9.996323074173012e-06, "loss": 0.6027, "step": 1041 }, { "epoch": 0.33340443686006827, "grad_norm": 0.6832480430603027, "learning_rate": 9.996251339935517e-06, "loss": 0.6224, "step": 1042 }, { "epoch": 0.3337244027303754, "grad_norm": 0.6809660196304321, "learning_rate": 9.996178912961761e-06, "loss": 0.5949, "step": 1043 }, { "epoch": 0.3340443686006826, "grad_norm": 0.654651403427124, "learning_rate": 9.996105793261787e-06, "loss": 0.55, "step": 1044 }, { "epoch": 0.3343643344709898, "grad_norm": 0.6393281817436218, "learning_rate": 9.996031980845737e-06, "loss": 0.5628, "step": 1045 }, { "epoch": 0.3346843003412969, "grad_norm": 0.691877007484436, "learning_rate": 9.995957475723844e-06, "loss": 0.6134, "step": 1046 }, { "epoch": 0.3350042662116041, "grad_norm": 0.6541376113891602, "learning_rate": 9.995882277906432e-06, "loss": 0.5532, "step": 1047 }, { "epoch": 0.33532423208191126, "grad_norm": 0.6597897410392761, "learning_rate": 9.995806387403935e-06, "loss": 0.5665, "step": 1048 }, { "epoch": 0.33564419795221845, "grad_norm": 0.6826539635658264, "learning_rate": 9.995729804226873e-06, "loss": 0.5538, "step": 1049 }, { "epoch": 0.3359641638225256, "grad_norm": 0.6731455326080322, "learning_rate": 9.995652528385861e-06, "loss": 0.5992, "step": 1050 }, { "epoch": 0.3362841296928328, "grad_norm": 0.7085751295089722, "learning_rate": 9.995574559891618e-06, "loss": 0.5933, "step": 1051 }, { "epoch": 0.3366040955631399, "grad_norm": 0.7039170861244202, "learning_rate": 9.995495898754952e-06, "loss": 0.6041, "step": 1052 }, { "epoch": 0.3369240614334471, "grad_norm": 0.6699726581573486, "learning_rate": 9.99541654498677e-06, "loss": 0.6477, "step": 1053 }, { "epoch": 0.33724402730375425, "grad_norm": 0.7025903463363647, "learning_rate": 9.995336498598075e-06, "loss": 0.599, "step": 1054 }, { "epoch": 0.33756399317406144, "grad_norm": 0.6295349597930908, "learning_rate": 9.995255759599963e-06, "loss": 0.6067, "step": 1055 }, { "epoch": 0.3378839590443686, "grad_norm": 0.7312242388725281, "learning_rate": 9.99517432800363e-06, "loss": 0.6038, "step": 1056 }, { "epoch": 0.33820392491467577, "grad_norm": 0.6863213181495667, "learning_rate": 9.99509220382037e-06, "loss": 0.5819, "step": 1057 }, { "epoch": 0.33852389078498296, "grad_norm": 0.5715763568878174, "learning_rate": 9.995009387061564e-06, "loss": 0.5727, "step": 1058 }, { "epoch": 0.3388438566552901, "grad_norm": 0.6914387345314026, "learning_rate": 9.9949258777387e-06, "loss": 0.6037, "step": 1059 }, { "epoch": 0.3391638225255973, "grad_norm": 0.6119387149810791, "learning_rate": 9.994841675863352e-06, "loss": 0.5823, "step": 1060 }, { "epoch": 0.33948378839590443, "grad_norm": 0.6641659736633301, "learning_rate": 9.994756781447197e-06, "loss": 0.5566, "step": 1061 }, { "epoch": 0.3398037542662116, "grad_norm": 0.736589252948761, "learning_rate": 9.994671194502006e-06, "loss": 0.6084, "step": 1062 }, { "epoch": 0.34012372013651876, "grad_norm": 0.6191753149032593, "learning_rate": 9.994584915039644e-06, "loss": 0.6031, "step": 1063 }, { "epoch": 0.34044368600682595, "grad_norm": 0.6466872692108154, "learning_rate": 9.994497943072076e-06, "loss": 0.544, "step": 1064 }, { "epoch": 0.3407636518771331, "grad_norm": 0.7583022117614746, "learning_rate": 9.99441027861136e-06, "loss": 0.5698, "step": 1065 }, { "epoch": 0.3410836177474403, "grad_norm": 0.6253617405891418, "learning_rate": 9.99432192166965e-06, "loss": 0.5644, "step": 1066 }, { "epoch": 0.3414035836177474, "grad_norm": 0.6997573375701904, "learning_rate": 9.994232872259197e-06, "loss": 0.6113, "step": 1067 }, { "epoch": 0.3417235494880546, "grad_norm": 0.8160573840141296, "learning_rate": 9.99414313039235e-06, "loss": 0.5807, "step": 1068 }, { "epoch": 0.34204351535836175, "grad_norm": 0.7675189971923828, "learning_rate": 9.994052696081546e-06, "loss": 0.5939, "step": 1069 }, { "epoch": 0.34236348122866894, "grad_norm": 0.6997158527374268, "learning_rate": 9.99396156933933e-06, "loss": 0.5717, "step": 1070 }, { "epoch": 0.34268344709897613, "grad_norm": 0.7480242252349854, "learning_rate": 9.993869750178334e-06, "loss": 0.5421, "step": 1071 }, { "epoch": 0.3430034129692833, "grad_norm": 0.637569010257721, "learning_rate": 9.993777238611292e-06, "loss": 0.5676, "step": 1072 }, { "epoch": 0.34332337883959047, "grad_norm": 0.6248916983604431, "learning_rate": 9.993684034651024e-06, "loss": 0.5755, "step": 1073 }, { "epoch": 0.3436433447098976, "grad_norm": 0.7089166045188904, "learning_rate": 9.993590138310458e-06, "loss": 0.5866, "step": 1074 }, { "epoch": 0.3439633105802048, "grad_norm": 0.7098347544670105, "learning_rate": 9.993495549602615e-06, "loss": 0.5986, "step": 1075 }, { "epoch": 0.34428327645051193, "grad_norm": 0.7500614523887634, "learning_rate": 9.993400268540603e-06, "loss": 0.5966, "step": 1076 }, { "epoch": 0.3446032423208191, "grad_norm": 0.841600775718689, "learning_rate": 9.993304295137638e-06, "loss": 0.5983, "step": 1077 }, { "epoch": 0.34492320819112626, "grad_norm": 0.6450603008270264, "learning_rate": 9.993207629407026e-06, "loss": 0.5679, "step": 1078 }, { "epoch": 0.34524317406143346, "grad_norm": 0.717484176158905, "learning_rate": 9.993110271362168e-06, "loss": 0.5646, "step": 1079 }, { "epoch": 0.3455631399317406, "grad_norm": 0.63707035779953, "learning_rate": 9.993012221016565e-06, "loss": 0.5142, "step": 1080 }, { "epoch": 0.3458831058020478, "grad_norm": 0.6633713841438293, "learning_rate": 9.99291347838381e-06, "loss": 0.5782, "step": 1081 }, { "epoch": 0.3462030716723549, "grad_norm": 0.6746697425842285, "learning_rate": 9.992814043477595e-06, "loss": 0.5736, "step": 1082 }, { "epoch": 0.3465230375426621, "grad_norm": 0.6532283425331116, "learning_rate": 9.992713916311707e-06, "loss": 0.5737, "step": 1083 }, { "epoch": 0.3468430034129693, "grad_norm": 0.6815126538276672, "learning_rate": 9.992613096900027e-06, "loss": 0.5999, "step": 1084 }, { "epoch": 0.34716296928327645, "grad_norm": 0.7754561901092529, "learning_rate": 9.992511585256537e-06, "loss": 0.5579, "step": 1085 }, { "epoch": 0.34748293515358364, "grad_norm": 0.7235525250434875, "learning_rate": 9.992409381395306e-06, "loss": 0.578, "step": 1086 }, { "epoch": 0.3478029010238908, "grad_norm": 0.7848513126373291, "learning_rate": 9.992306485330511e-06, "loss": 0.586, "step": 1087 }, { "epoch": 0.34812286689419797, "grad_norm": 0.6942896246910095, "learning_rate": 9.992202897076416e-06, "loss": 0.5753, "step": 1088 }, { "epoch": 0.3484428327645051, "grad_norm": 0.7098321318626404, "learning_rate": 9.992098616647384e-06, "loss": 0.5201, "step": 1089 }, { "epoch": 0.3487627986348123, "grad_norm": 0.7250136137008667, "learning_rate": 9.99199364405787e-06, "loss": 0.5745, "step": 1090 }, { "epoch": 0.34908276450511944, "grad_norm": 0.6978463530540466, "learning_rate": 9.991887979322434e-06, "loss": 0.5581, "step": 1091 }, { "epoch": 0.34940273037542663, "grad_norm": 0.6456747055053711, "learning_rate": 9.991781622455723e-06, "loss": 0.5557, "step": 1092 }, { "epoch": 0.34972269624573377, "grad_norm": 0.6560348272323608, "learning_rate": 9.991674573472488e-06, "loss": 0.5489, "step": 1093 }, { "epoch": 0.35004266211604096, "grad_norm": 0.6581658720970154, "learning_rate": 9.991566832387564e-06, "loss": 0.5708, "step": 1094 }, { "epoch": 0.3503626279863481, "grad_norm": 0.6679834127426147, "learning_rate": 9.991458399215897e-06, "loss": 0.6162, "step": 1095 }, { "epoch": 0.3506825938566553, "grad_norm": 0.644548773765564, "learning_rate": 9.991349273972517e-06, "loss": 0.5614, "step": 1096 }, { "epoch": 0.3510025597269625, "grad_norm": 0.645207941532135, "learning_rate": 9.991239456672555e-06, "loss": 0.595, "step": 1097 }, { "epoch": 0.3513225255972696, "grad_norm": 0.6927055716514587, "learning_rate": 9.991128947331236e-06, "loss": 0.6157, "step": 1098 }, { "epoch": 0.3516424914675768, "grad_norm": 0.66288822889328, "learning_rate": 9.991017745963886e-06, "loss": 0.5706, "step": 1099 }, { "epoch": 0.35196245733788395, "grad_norm": 0.5749547481536865, "learning_rate": 9.990905852585922e-06, "loss": 0.5642, "step": 1100 }, { "epoch": 0.35228242320819114, "grad_norm": 0.726407527923584, "learning_rate": 9.990793267212857e-06, "loss": 0.5749, "step": 1101 }, { "epoch": 0.3526023890784983, "grad_norm": 0.6601400971412659, "learning_rate": 9.9906799898603e-06, "loss": 0.6701, "step": 1102 }, { "epoch": 0.3529223549488055, "grad_norm": 0.614708662033081, "learning_rate": 9.99056602054396e-06, "loss": 0.5712, "step": 1103 }, { "epoch": 0.3532423208191126, "grad_norm": 0.6331604719161987, "learning_rate": 9.990451359279636e-06, "loss": 0.5656, "step": 1104 }, { "epoch": 0.3535622866894198, "grad_norm": 0.6421919465065002, "learning_rate": 9.990336006083228e-06, "loss": 0.5533, "step": 1105 }, { "epoch": 0.35388225255972694, "grad_norm": 0.6681429147720337, "learning_rate": 9.99021996097073e-06, "loss": 0.5741, "step": 1106 }, { "epoch": 0.35420221843003413, "grad_norm": 0.7112416625022888, "learning_rate": 9.99010322395823e-06, "loss": 0.5495, "step": 1107 }, { "epoch": 0.35452218430034127, "grad_norm": 0.705014169216156, "learning_rate": 9.989985795061914e-06, "loss": 0.59, "step": 1108 }, { "epoch": 0.35484215017064846, "grad_norm": 0.7376593351364136, "learning_rate": 9.989867674298066e-06, "loss": 0.5733, "step": 1109 }, { "epoch": 0.35516211604095566, "grad_norm": 0.6666154861450195, "learning_rate": 9.989748861683061e-06, "loss": 0.5523, "step": 1110 }, { "epoch": 0.3554820819112628, "grad_norm": 0.7035896182060242, "learning_rate": 9.989629357233375e-06, "loss": 0.5774, "step": 1111 }, { "epoch": 0.35580204778157, "grad_norm": 0.6515471935272217, "learning_rate": 9.989509160965575e-06, "loss": 0.5701, "step": 1112 }, { "epoch": 0.3561220136518771, "grad_norm": 0.6917015314102173, "learning_rate": 9.989388272896328e-06, "loss": 0.5746, "step": 1113 }, { "epoch": 0.3564419795221843, "grad_norm": 0.6861441731452942, "learning_rate": 9.989266693042394e-06, "loss": 0.605, "step": 1114 }, { "epoch": 0.35676194539249145, "grad_norm": 0.5988866686820984, "learning_rate": 9.98914442142063e-06, "loss": 0.5874, "step": 1115 }, { "epoch": 0.35708191126279865, "grad_norm": 0.644522488117218, "learning_rate": 9.989021458047992e-06, "loss": 0.5827, "step": 1116 }, { "epoch": 0.3574018771331058, "grad_norm": 0.7073920369148254, "learning_rate": 9.988897802941523e-06, "loss": 0.5459, "step": 1117 }, { "epoch": 0.357721843003413, "grad_norm": 0.5924735069274902, "learning_rate": 9.988773456118376e-06, "loss": 0.5737, "step": 1118 }, { "epoch": 0.3580418088737201, "grad_norm": 0.8429334163665771, "learning_rate": 9.988648417595785e-06, "loss": 0.626, "step": 1119 }, { "epoch": 0.3583617747440273, "grad_norm": 0.7303470969200134, "learning_rate": 9.988522687391092e-06, "loss": 0.5688, "step": 1120 }, { "epoch": 0.35868174061433444, "grad_norm": 0.6042421460151672, "learning_rate": 9.988396265521725e-06, "loss": 0.5737, "step": 1121 }, { "epoch": 0.35900170648464164, "grad_norm": 0.7154580950737, "learning_rate": 9.988269152005215e-06, "loss": 0.6298, "step": 1122 }, { "epoch": 0.35932167235494883, "grad_norm": 0.6817564368247986, "learning_rate": 9.988141346859186e-06, "loss": 0.6032, "step": 1123 }, { "epoch": 0.35964163822525597, "grad_norm": 0.7040557861328125, "learning_rate": 9.988012850101359e-06, "loss": 0.592, "step": 1124 }, { "epoch": 0.35996160409556316, "grad_norm": 0.6956594586372375, "learning_rate": 9.987883661749548e-06, "loss": 0.5889, "step": 1125 }, { "epoch": 0.3602815699658703, "grad_norm": 0.6449658870697021, "learning_rate": 9.987753781821667e-06, "loss": 0.6034, "step": 1126 }, { "epoch": 0.3606015358361775, "grad_norm": 0.7003522515296936, "learning_rate": 9.987623210335725e-06, "loss": 0.5691, "step": 1127 }, { "epoch": 0.3609215017064846, "grad_norm": 0.6019575595855713, "learning_rate": 9.987491947309825e-06, "loss": 0.526, "step": 1128 }, { "epoch": 0.3612414675767918, "grad_norm": 0.7500554323196411, "learning_rate": 9.987359992762165e-06, "loss": 0.622, "step": 1129 }, { "epoch": 0.36156143344709896, "grad_norm": 0.6906517148017883, "learning_rate": 9.987227346711043e-06, "loss": 0.5249, "step": 1130 }, { "epoch": 0.36188139931740615, "grad_norm": 0.6303286552429199, "learning_rate": 9.987094009174849e-06, "loss": 0.5833, "step": 1131 }, { "epoch": 0.3622013651877133, "grad_norm": 0.6838951706886292, "learning_rate": 9.98695998017207e-06, "loss": 0.556, "step": 1132 }, { "epoch": 0.3625213310580205, "grad_norm": 0.7039484977722168, "learning_rate": 9.986825259721292e-06, "loss": 0.5779, "step": 1133 }, { "epoch": 0.3628412969283277, "grad_norm": 0.6559299826622009, "learning_rate": 9.986689847841192e-06, "loss": 0.5974, "step": 1134 }, { "epoch": 0.3631612627986348, "grad_norm": 0.7664186954498291, "learning_rate": 9.986553744550546e-06, "loss": 0.5808, "step": 1135 }, { "epoch": 0.363481228668942, "grad_norm": 0.6697313189506531, "learning_rate": 9.986416949868223e-06, "loss": 0.5865, "step": 1136 }, { "epoch": 0.36380119453924914, "grad_norm": 0.709850013256073, "learning_rate": 9.986279463813192e-06, "loss": 0.5888, "step": 1137 }, { "epoch": 0.36412116040955633, "grad_norm": 0.7761910557746887, "learning_rate": 9.986141286404514e-06, "loss": 0.6414, "step": 1138 }, { "epoch": 0.36444112627986347, "grad_norm": 0.6782816052436829, "learning_rate": 9.98600241766135e-06, "loss": 0.5639, "step": 1139 }, { "epoch": 0.36476109215017066, "grad_norm": 0.6862764954566956, "learning_rate": 9.985862857602951e-06, "loss": 0.5562, "step": 1140 }, { "epoch": 0.3650810580204778, "grad_norm": 0.7213887572288513, "learning_rate": 9.985722606248669e-06, "loss": 0.6093, "step": 1141 }, { "epoch": 0.365401023890785, "grad_norm": 0.747785210609436, "learning_rate": 9.985581663617948e-06, "loss": 0.5812, "step": 1142 }, { "epoch": 0.36572098976109213, "grad_norm": 0.5790359973907471, "learning_rate": 9.985440029730335e-06, "loss": 0.5696, "step": 1143 }, { "epoch": 0.3660409556313993, "grad_norm": 0.653598964214325, "learning_rate": 9.985297704605461e-06, "loss": 0.5617, "step": 1144 }, { "epoch": 0.36636092150170646, "grad_norm": 0.6908425092697144, "learning_rate": 9.985154688263065e-06, "loss": 0.5403, "step": 1145 }, { "epoch": 0.36668088737201365, "grad_norm": 0.5721466541290283, "learning_rate": 9.985010980722974e-06, "loss": 0.5788, "step": 1146 }, { "epoch": 0.36700085324232085, "grad_norm": 0.6806579232215881, "learning_rate": 9.984866582005112e-06, "loss": 0.5775, "step": 1147 }, { "epoch": 0.367320819112628, "grad_norm": 0.6131672859191895, "learning_rate": 9.9847214921295e-06, "loss": 0.5577, "step": 1148 }, { "epoch": 0.3676407849829352, "grad_norm": 0.6523510217666626, "learning_rate": 9.98457571111626e-06, "loss": 0.5776, "step": 1149 }, { "epoch": 0.3679607508532423, "grad_norm": 0.6214600801467896, "learning_rate": 9.9844292389856e-06, "loss": 0.5581, "step": 1150 }, { "epoch": 0.3682807167235495, "grad_norm": 0.6363633871078491, "learning_rate": 9.98428207575783e-06, "loss": 0.6062, "step": 1151 }, { "epoch": 0.36860068259385664, "grad_norm": 0.7081869840621948, "learning_rate": 9.984134221453352e-06, "loss": 0.5975, "step": 1152 }, { "epoch": 0.36892064846416384, "grad_norm": 0.6446593403816223, "learning_rate": 9.98398567609267e-06, "loss": 0.6211, "step": 1153 }, { "epoch": 0.369240614334471, "grad_norm": 0.7253109216690063, "learning_rate": 9.983836439696378e-06, "loss": 0.6514, "step": 1154 }, { "epoch": 0.36956058020477817, "grad_norm": 0.6849066019058228, "learning_rate": 9.983686512285166e-06, "loss": 0.5941, "step": 1155 }, { "epoch": 0.3698805460750853, "grad_norm": 0.7073627710342407, "learning_rate": 9.983535893879825e-06, "loss": 0.627, "step": 1156 }, { "epoch": 0.3702005119453925, "grad_norm": 0.7770823240280151, "learning_rate": 9.983384584501236e-06, "loss": 0.639, "step": 1157 }, { "epoch": 0.37052047781569963, "grad_norm": 0.6505449414253235, "learning_rate": 9.98323258417038e-06, "loss": 0.5884, "step": 1158 }, { "epoch": 0.3708404436860068, "grad_norm": 0.6462435722351074, "learning_rate": 9.983079892908332e-06, "loss": 0.5755, "step": 1159 }, { "epoch": 0.371160409556314, "grad_norm": 0.6581071019172668, "learning_rate": 9.982926510736261e-06, "loss": 0.5478, "step": 1160 }, { "epoch": 0.37148037542662116, "grad_norm": 0.6703523397445679, "learning_rate": 9.982772437675436e-06, "loss": 0.6015, "step": 1161 }, { "epoch": 0.37180034129692835, "grad_norm": 0.6885740160942078, "learning_rate": 9.982617673747218e-06, "loss": 0.5669, "step": 1162 }, { "epoch": 0.3721203071672355, "grad_norm": 0.7145320177078247, "learning_rate": 9.982462218973064e-06, "loss": 0.6137, "step": 1163 }, { "epoch": 0.3724402730375427, "grad_norm": 0.6247287392616272, "learning_rate": 9.98230607337453e-06, "loss": 0.5928, "step": 1164 }, { "epoch": 0.3727602389078498, "grad_norm": 0.6764070987701416, "learning_rate": 9.982149236973264e-06, "loss": 0.5466, "step": 1165 }, { "epoch": 0.373080204778157, "grad_norm": 0.6221293807029724, "learning_rate": 9.981991709791012e-06, "loss": 0.5462, "step": 1166 }, { "epoch": 0.37340017064846415, "grad_norm": 0.6133841276168823, "learning_rate": 9.981833491849618e-06, "loss": 0.5722, "step": 1167 }, { "epoch": 0.37372013651877134, "grad_norm": 0.7736768126487732, "learning_rate": 9.981674583171017e-06, "loss": 0.6171, "step": 1168 }, { "epoch": 0.3740401023890785, "grad_norm": 0.575372040271759, "learning_rate": 9.98151498377724e-06, "loss": 0.572, "step": 1169 }, { "epoch": 0.37436006825938567, "grad_norm": 0.6014509201049805, "learning_rate": 9.98135469369042e-06, "loss": 0.5583, "step": 1170 }, { "epoch": 0.3746800341296928, "grad_norm": 0.6997368335723877, "learning_rate": 9.981193712932778e-06, "loss": 0.5789, "step": 1171 }, { "epoch": 0.375, "grad_norm": 0.6987643837928772, "learning_rate": 9.981032041526635e-06, "loss": 0.5355, "step": 1172 }, { "epoch": 0.3753199658703072, "grad_norm": 0.5802475214004517, "learning_rate": 9.980869679494407e-06, "loss": 0.5657, "step": 1173 }, { "epoch": 0.37563993174061433, "grad_norm": 0.6369011402130127, "learning_rate": 9.980706626858607e-06, "loss": 0.5567, "step": 1174 }, { "epoch": 0.3759598976109215, "grad_norm": 0.5437308549880981, "learning_rate": 9.980542883641839e-06, "loss": 0.5418, "step": 1175 }, { "epoch": 0.37627986348122866, "grad_norm": 0.5967769622802734, "learning_rate": 9.98037844986681e-06, "loss": 0.5721, "step": 1176 }, { "epoch": 0.37659982935153585, "grad_norm": 0.7096673250198364, "learning_rate": 9.980213325556319e-06, "loss": 0.6103, "step": 1177 }, { "epoch": 0.376919795221843, "grad_norm": 0.5754175186157227, "learning_rate": 9.980047510733257e-06, "loss": 0.5682, "step": 1178 }, { "epoch": 0.3772397610921502, "grad_norm": 0.6432886123657227, "learning_rate": 9.979881005420615e-06, "loss": 0.5549, "step": 1179 }, { "epoch": 0.3775597269624573, "grad_norm": 0.6181249618530273, "learning_rate": 9.979713809641482e-06, "loss": 0.5719, "step": 1180 }, { "epoch": 0.3778796928327645, "grad_norm": 0.7071648240089417, "learning_rate": 9.97954592341904e-06, "loss": 0.575, "step": 1181 }, { "epoch": 0.37819965870307165, "grad_norm": 0.6493441462516785, "learning_rate": 9.979377346776564e-06, "loss": 0.5015, "step": 1182 }, { "epoch": 0.37851962457337884, "grad_norm": 0.6092111468315125, "learning_rate": 9.979208079737428e-06, "loss": 0.5714, "step": 1183 }, { "epoch": 0.378839590443686, "grad_norm": 0.6894551515579224, "learning_rate": 9.979038122325103e-06, "loss": 0.6282, "step": 1184 }, { "epoch": 0.3791595563139932, "grad_norm": 0.6173356175422668, "learning_rate": 9.978867474563151e-06, "loss": 0.5896, "step": 1185 }, { "epoch": 0.37947952218430037, "grad_norm": 0.6287574172019958, "learning_rate": 9.978696136475235e-06, "loss": 0.5881, "step": 1186 }, { "epoch": 0.3797994880546075, "grad_norm": 0.6383042335510254, "learning_rate": 9.97852410808511e-06, "loss": 0.598, "step": 1187 }, { "epoch": 0.3801194539249147, "grad_norm": 0.6773865818977356, "learning_rate": 9.978351389416626e-06, "loss": 0.6118, "step": 1188 }, { "epoch": 0.38043941979522183, "grad_norm": 0.6683553457260132, "learning_rate": 9.978177980493736e-06, "loss": 0.6278, "step": 1189 }, { "epoch": 0.380759385665529, "grad_norm": 0.6252800822257996, "learning_rate": 9.978003881340478e-06, "loss": 0.5507, "step": 1190 }, { "epoch": 0.38107935153583616, "grad_norm": 0.6656942367553711, "learning_rate": 9.977829091980996e-06, "loss": 0.6245, "step": 1191 }, { "epoch": 0.38139931740614336, "grad_norm": 0.6880273222923279, "learning_rate": 9.977653612439518e-06, "loss": 0.5634, "step": 1192 }, { "epoch": 0.3817192832764505, "grad_norm": 0.6544172167778015, "learning_rate": 9.97747744274038e-06, "loss": 0.5966, "step": 1193 }, { "epoch": 0.3820392491467577, "grad_norm": 0.6122605204582214, "learning_rate": 9.977300582908007e-06, "loss": 0.6154, "step": 1194 }, { "epoch": 0.3823592150170648, "grad_norm": 0.739066481590271, "learning_rate": 9.97712303296692e-06, "loss": 0.5971, "step": 1195 }, { "epoch": 0.382679180887372, "grad_norm": 0.6078752875328064, "learning_rate": 9.976944792941736e-06, "loss": 0.561, "step": 1196 }, { "epoch": 0.38299914675767915, "grad_norm": 0.6688116788864136, "learning_rate": 9.976765862857171e-06, "loss": 0.5797, "step": 1197 }, { "epoch": 0.38331911262798635, "grad_norm": 0.6825562715530396, "learning_rate": 9.976586242738032e-06, "loss": 0.5775, "step": 1198 }, { "epoch": 0.38363907849829354, "grad_norm": 0.6867298483848572, "learning_rate": 9.976405932609221e-06, "loss": 0.581, "step": 1199 }, { "epoch": 0.3839590443686007, "grad_norm": 0.7204470634460449, "learning_rate": 9.976224932495741e-06, "loss": 0.5947, "step": 1200 }, { "epoch": 0.38427901023890787, "grad_norm": 0.8085878491401672, "learning_rate": 9.97604324242269e-06, "loss": 0.6117, "step": 1201 }, { "epoch": 0.384598976109215, "grad_norm": 0.6830131411552429, "learning_rate": 9.975860862415255e-06, "loss": 0.58, "step": 1202 }, { "epoch": 0.3849189419795222, "grad_norm": 0.6403172016143799, "learning_rate": 9.975677792498725e-06, "loss": 0.5923, "step": 1203 }, { "epoch": 0.38523890784982934, "grad_norm": 0.6732902526855469, "learning_rate": 9.975494032698485e-06, "loss": 0.5615, "step": 1204 }, { "epoch": 0.38555887372013653, "grad_norm": 0.726446270942688, "learning_rate": 9.975309583040012e-06, "loss": 0.5781, "step": 1205 }, { "epoch": 0.38587883959044367, "grad_norm": 0.7127211093902588, "learning_rate": 9.975124443548878e-06, "loss": 0.573, "step": 1206 }, { "epoch": 0.38619880546075086, "grad_norm": 0.6206179261207581, "learning_rate": 9.974938614250755e-06, "loss": 0.5707, "step": 1207 }, { "epoch": 0.386518771331058, "grad_norm": 0.8432472944259644, "learning_rate": 9.974752095171409e-06, "loss": 0.5884, "step": 1208 }, { "epoch": 0.3868387372013652, "grad_norm": 0.8524286150932312, "learning_rate": 9.9745648863367e-06, "loss": 0.6203, "step": 1209 }, { "epoch": 0.3871587030716723, "grad_norm": 0.6269184947013855, "learning_rate": 9.974376987772584e-06, "loss": 0.5799, "step": 1210 }, { "epoch": 0.3874786689419795, "grad_norm": 0.7409312129020691, "learning_rate": 9.974188399505116e-06, "loss": 0.5737, "step": 1211 }, { "epoch": 0.3877986348122867, "grad_norm": 0.6829506754875183, "learning_rate": 9.97399912156044e-06, "loss": 0.5729, "step": 1212 }, { "epoch": 0.38811860068259385, "grad_norm": 0.7614309191703796, "learning_rate": 9.973809153964803e-06, "loss": 0.559, "step": 1213 }, { "epoch": 0.38843856655290104, "grad_norm": 0.6158562898635864, "learning_rate": 9.973618496744544e-06, "loss": 0.5986, "step": 1214 }, { "epoch": 0.3887585324232082, "grad_norm": 0.5494482517242432, "learning_rate": 9.973427149926098e-06, "loss": 0.5347, "step": 1215 }, { "epoch": 0.3890784982935154, "grad_norm": 0.7281996607780457, "learning_rate": 9.973235113535992e-06, "loss": 0.566, "step": 1216 }, { "epoch": 0.3893984641638225, "grad_norm": 0.7141414284706116, "learning_rate": 9.973042387600856e-06, "loss": 0.5825, "step": 1217 }, { "epoch": 0.3897184300341297, "grad_norm": 0.5919783115386963, "learning_rate": 9.972848972147408e-06, "loss": 0.5699, "step": 1218 }, { "epoch": 0.39003839590443684, "grad_norm": 0.6098409295082092, "learning_rate": 9.97265486720247e-06, "loss": 0.5485, "step": 1219 }, { "epoch": 0.39035836177474403, "grad_norm": 0.5585481524467468, "learning_rate": 9.972460072792951e-06, "loss": 0.5359, "step": 1220 }, { "epoch": 0.39067832764505117, "grad_norm": 0.620502233505249, "learning_rate": 9.972264588945863e-06, "loss": 0.6053, "step": 1221 }, { "epoch": 0.39099829351535836, "grad_norm": 0.6805775165557861, "learning_rate": 9.972068415688307e-06, "loss": 0.6028, "step": 1222 }, { "epoch": 0.39131825938566556, "grad_norm": 0.6690160632133484, "learning_rate": 9.971871553047484e-06, "loss": 0.5774, "step": 1223 }, { "epoch": 0.3916382252559727, "grad_norm": 0.7072525024414062, "learning_rate": 9.971674001050687e-06, "loss": 0.6017, "step": 1224 }, { "epoch": 0.3919581911262799, "grad_norm": 0.5692431926727295, "learning_rate": 9.971475759725308e-06, "loss": 0.574, "step": 1225 }, { "epoch": 0.392278156996587, "grad_norm": 0.6000913977622986, "learning_rate": 9.971276829098837e-06, "loss": 0.5986, "step": 1226 }, { "epoch": 0.3925981228668942, "grad_norm": 0.7071651816368103, "learning_rate": 9.971077209198851e-06, "loss": 0.5838, "step": 1227 }, { "epoch": 0.39291808873720135, "grad_norm": 0.6404964327812195, "learning_rate": 9.970876900053031e-06, "loss": 0.5755, "step": 1228 }, { "epoch": 0.39323805460750855, "grad_norm": 0.5926943421363831, "learning_rate": 9.970675901689148e-06, "loss": 0.5661, "step": 1229 }, { "epoch": 0.3935580204778157, "grad_norm": 0.6548840999603271, "learning_rate": 9.970474214135072e-06, "loss": 0.6253, "step": 1230 }, { "epoch": 0.3938779863481229, "grad_norm": 0.5912249684333801, "learning_rate": 9.970271837418765e-06, "loss": 0.5632, "step": 1231 }, { "epoch": 0.39419795221843, "grad_norm": 0.5992922782897949, "learning_rate": 9.97006877156829e-06, "loss": 0.535, "step": 1232 }, { "epoch": 0.3945179180887372, "grad_norm": 0.6178451776504517, "learning_rate": 9.9698650166118e-06, "loss": 0.5984, "step": 1233 }, { "epoch": 0.39483788395904434, "grad_norm": 0.6016383767127991, "learning_rate": 9.969660572577548e-06, "loss": 0.5452, "step": 1234 }, { "epoch": 0.39515784982935154, "grad_norm": 0.5990430116653442, "learning_rate": 9.969455439493877e-06, "loss": 0.5966, "step": 1235 }, { "epoch": 0.39547781569965873, "grad_norm": 0.6854697465896606, "learning_rate": 9.969249617389232e-06, "loss": 0.6192, "step": 1236 }, { "epoch": 0.39579778156996587, "grad_norm": 0.558914065361023, "learning_rate": 9.969043106292149e-06, "loss": 0.589, "step": 1237 }, { "epoch": 0.39611774744027306, "grad_norm": 0.6286885738372803, "learning_rate": 9.96883590623126e-06, "loss": 0.5764, "step": 1238 }, { "epoch": 0.3964377133105802, "grad_norm": 0.6284793615341187, "learning_rate": 9.968628017235297e-06, "loss": 0.5925, "step": 1239 }, { "epoch": 0.3967576791808874, "grad_norm": 0.6545090079307556, "learning_rate": 9.96841943933308e-06, "loss": 0.6065, "step": 1240 }, { "epoch": 0.3970776450511945, "grad_norm": 0.5711103677749634, "learning_rate": 9.968210172553533e-06, "loss": 0.5231, "step": 1241 }, { "epoch": 0.3973976109215017, "grad_norm": 0.6072283983230591, "learning_rate": 9.968000216925668e-06, "loss": 0.5538, "step": 1242 }, { "epoch": 0.39771757679180886, "grad_norm": 0.6144881248474121, "learning_rate": 9.967789572478596e-06, "loss": 0.5628, "step": 1243 }, { "epoch": 0.39803754266211605, "grad_norm": 0.6187753677368164, "learning_rate": 9.967578239241523e-06, "loss": 0.555, "step": 1244 }, { "epoch": 0.3983575085324232, "grad_norm": 0.6248455047607422, "learning_rate": 9.967366217243752e-06, "loss": 0.5834, "step": 1245 }, { "epoch": 0.3986774744027304, "grad_norm": 0.6310631632804871, "learning_rate": 9.967153506514677e-06, "loss": 0.5562, "step": 1246 }, { "epoch": 0.3989974402730375, "grad_norm": 0.6546112298965454, "learning_rate": 9.966940107083795e-06, "loss": 0.5814, "step": 1247 }, { "epoch": 0.3993174061433447, "grad_norm": 0.5986621975898743, "learning_rate": 9.96672601898069e-06, "loss": 0.5235, "step": 1248 }, { "epoch": 0.3996373720136519, "grad_norm": 0.7620200514793396, "learning_rate": 9.966511242235048e-06, "loss": 0.5646, "step": 1249 }, { "epoch": 0.39995733788395904, "grad_norm": 0.6559775471687317, "learning_rate": 9.966295776876648e-06, "loss": 0.5526, "step": 1250 }, { "epoch": 0.40027730375426623, "grad_norm": 0.544110119342804, "learning_rate": 9.966079622935363e-06, "loss": 0.5504, "step": 1251 }, { "epoch": 0.40059726962457337, "grad_norm": 0.6554135680198669, "learning_rate": 9.965862780441165e-06, "loss": 0.5796, "step": 1252 }, { "epoch": 0.40091723549488056, "grad_norm": 0.6634418368339539, "learning_rate": 9.965645249424118e-06, "loss": 0.6091, "step": 1253 }, { "epoch": 0.4012372013651877, "grad_norm": 0.57181316614151, "learning_rate": 9.965427029914383e-06, "loss": 0.5736, "step": 1254 }, { "epoch": 0.4015571672354949, "grad_norm": 0.6040608286857605, "learning_rate": 9.965208121942217e-06, "loss": 0.5524, "step": 1255 }, { "epoch": 0.40187713310580203, "grad_norm": 0.7191135287284851, "learning_rate": 9.964988525537971e-06, "loss": 0.5519, "step": 1256 }, { "epoch": 0.4021970989761092, "grad_norm": 0.599594235420227, "learning_rate": 9.964768240732094e-06, "loss": 0.5674, "step": 1257 }, { "epoch": 0.40251706484641636, "grad_norm": 0.6317551136016846, "learning_rate": 9.964547267555127e-06, "loss": 0.5634, "step": 1258 }, { "epoch": 0.40283703071672355, "grad_norm": 0.6445105075836182, "learning_rate": 9.964325606037709e-06, "loss": 0.5879, "step": 1259 }, { "epoch": 0.4031569965870307, "grad_norm": 0.6589673161506653, "learning_rate": 9.964103256210573e-06, "loss": 0.5316, "step": 1260 }, { "epoch": 0.4034769624573379, "grad_norm": 0.6195188164710999, "learning_rate": 9.96388021810455e-06, "loss": 0.6039, "step": 1261 }, { "epoch": 0.4037969283276451, "grad_norm": 0.6745509505271912, "learning_rate": 9.963656491750562e-06, "loss": 0.559, "step": 1262 }, { "epoch": 0.4041168941979522, "grad_norm": 0.5846664905548096, "learning_rate": 9.963432077179629e-06, "loss": 0.5752, "step": 1263 }, { "epoch": 0.4044368600682594, "grad_norm": 0.6875172853469849, "learning_rate": 9.963206974422868e-06, "loss": 0.5997, "step": 1264 }, { "epoch": 0.40475682593856654, "grad_norm": 0.6228593587875366, "learning_rate": 9.962981183511491e-06, "loss": 0.5657, "step": 1265 }, { "epoch": 0.40507679180887374, "grad_norm": 0.5766012072563171, "learning_rate": 9.9627547044768e-06, "loss": 0.5441, "step": 1266 }, { "epoch": 0.4053967576791809, "grad_norm": 0.6164464354515076, "learning_rate": 9.9625275373502e-06, "loss": 0.5954, "step": 1267 }, { "epoch": 0.40571672354948807, "grad_norm": 0.6023998856544495, "learning_rate": 9.962299682163185e-06, "loss": 0.5657, "step": 1268 }, { "epoch": 0.4060366894197952, "grad_norm": 0.6143546104431152, "learning_rate": 9.962071138947352e-06, "loss": 0.5329, "step": 1269 }, { "epoch": 0.4063566552901024, "grad_norm": 0.6571676731109619, "learning_rate": 9.961841907734385e-06, "loss": 0.5595, "step": 1270 }, { "epoch": 0.40667662116040953, "grad_norm": 0.591964602470398, "learning_rate": 9.961611988556067e-06, "loss": 0.5509, "step": 1271 }, { "epoch": 0.4069965870307167, "grad_norm": 0.6351963877677917, "learning_rate": 9.96138138144428e-06, "loss": 0.5731, "step": 1272 }, { "epoch": 0.40731655290102387, "grad_norm": 0.598535418510437, "learning_rate": 9.961150086430995e-06, "loss": 0.5716, "step": 1273 }, { "epoch": 0.40763651877133106, "grad_norm": 0.8259955644607544, "learning_rate": 9.960918103548281e-06, "loss": 0.6108, "step": 1274 }, { "epoch": 0.40795648464163825, "grad_norm": 0.6393885016441345, "learning_rate": 9.960685432828305e-06, "loss": 0.5721, "step": 1275 }, { "epoch": 0.4082764505119454, "grad_norm": 0.7137302756309509, "learning_rate": 9.960452074303327e-06, "loss": 0.5791, "step": 1276 }, { "epoch": 0.4085964163822526, "grad_norm": 0.7044225335121155, "learning_rate": 9.960218028005699e-06, "loss": 0.5997, "step": 1277 }, { "epoch": 0.4089163822525597, "grad_norm": 0.6670733094215393, "learning_rate": 9.959983293967875e-06, "loss": 0.5852, "step": 1278 }, { "epoch": 0.4092363481228669, "grad_norm": 0.6231539845466614, "learning_rate": 9.9597478722224e-06, "loss": 0.5777, "step": 1279 }, { "epoch": 0.40955631399317405, "grad_norm": 0.6410960555076599, "learning_rate": 9.959511762801916e-06, "loss": 0.5573, "step": 1280 }, { "epoch": 0.40987627986348124, "grad_norm": 0.67142254114151, "learning_rate": 9.95927496573916e-06, "loss": 0.5713, "step": 1281 }, { "epoch": 0.4101962457337884, "grad_norm": 0.6224493980407715, "learning_rate": 9.959037481066964e-06, "loss": 0.5823, "step": 1282 }, { "epoch": 0.41051621160409557, "grad_norm": 0.7081955075263977, "learning_rate": 9.958799308818254e-06, "loss": 0.5603, "step": 1283 }, { "epoch": 0.4108361774744027, "grad_norm": 0.6348280310630798, "learning_rate": 9.958560449026055e-06, "loss": 0.594, "step": 1284 }, { "epoch": 0.4111561433447099, "grad_norm": 0.689378023147583, "learning_rate": 9.958320901723487e-06, "loss": 0.5512, "step": 1285 }, { "epoch": 0.41147610921501704, "grad_norm": 0.7600042223930359, "learning_rate": 9.958080666943757e-06, "loss": 0.5554, "step": 1286 }, { "epoch": 0.41179607508532423, "grad_norm": 0.5788906216621399, "learning_rate": 9.95783974472018e-06, "loss": 0.5857, "step": 1287 }, { "epoch": 0.4121160409556314, "grad_norm": 0.6882662177085876, "learning_rate": 9.957598135086157e-06, "loss": 0.5523, "step": 1288 }, { "epoch": 0.41243600682593856, "grad_norm": 0.7677709460258484, "learning_rate": 9.957355838075188e-06, "loss": 0.556, "step": 1289 }, { "epoch": 0.41275597269624575, "grad_norm": 0.6081621646881104, "learning_rate": 9.95711285372087e-06, "loss": 0.5768, "step": 1290 }, { "epoch": 0.4130759385665529, "grad_norm": 0.6662126779556274, "learning_rate": 9.95686918205689e-06, "loss": 0.6074, "step": 1291 }, { "epoch": 0.4133959044368601, "grad_norm": 0.7383588552474976, "learning_rate": 9.956624823117036e-06, "loss": 0.6067, "step": 1292 }, { "epoch": 0.4137158703071672, "grad_norm": 0.6374850869178772, "learning_rate": 9.956379776935185e-06, "loss": 0.5683, "step": 1293 }, { "epoch": 0.4140358361774744, "grad_norm": 0.6523234844207764, "learning_rate": 9.956134043545317e-06, "loss": 0.6137, "step": 1294 }, { "epoch": 0.41435580204778155, "grad_norm": 0.7293864488601685, "learning_rate": 9.955887622981501e-06, "loss": 0.599, "step": 1295 }, { "epoch": 0.41467576791808874, "grad_norm": 0.7201833724975586, "learning_rate": 9.955640515277905e-06, "loss": 0.5998, "step": 1296 }, { "epoch": 0.4149957337883959, "grad_norm": 0.6556185483932495, "learning_rate": 9.955392720468789e-06, "loss": 0.5898, "step": 1297 }, { "epoch": 0.4153156996587031, "grad_norm": 0.6244340538978577, "learning_rate": 9.95514423858851e-06, "loss": 0.6161, "step": 1298 }, { "epoch": 0.4156356655290102, "grad_norm": 0.7177112698554993, "learning_rate": 9.954895069671523e-06, "loss": 0.5785, "step": 1299 }, { "epoch": 0.4159556313993174, "grad_norm": 0.6149005889892578, "learning_rate": 9.954645213752372e-06, "loss": 0.5764, "step": 1300 }, { "epoch": 0.4162755972696246, "grad_norm": 0.6709874868392944, "learning_rate": 9.954394670865701e-06, "loss": 0.6015, "step": 1301 }, { "epoch": 0.41659556313993173, "grad_norm": 0.5673186779022217, "learning_rate": 9.95414344104625e-06, "loss": 0.5538, "step": 1302 }, { "epoch": 0.4169155290102389, "grad_norm": 0.5811302065849304, "learning_rate": 9.953891524328851e-06, "loss": 0.5742, "step": 1303 }, { "epoch": 0.41723549488054607, "grad_norm": 0.7059139609336853, "learning_rate": 9.953638920748432e-06, "loss": 0.563, "step": 1304 }, { "epoch": 0.41755546075085326, "grad_norm": 0.7833511233329773, "learning_rate": 9.953385630340016e-06, "loss": 0.615, "step": 1305 }, { "epoch": 0.4178754266211604, "grad_norm": 0.7137935161590576, "learning_rate": 9.953131653138726e-06, "loss": 0.5711, "step": 1306 }, { "epoch": 0.4181953924914676, "grad_norm": 0.5633505582809448, "learning_rate": 9.952876989179772e-06, "loss": 0.5677, "step": 1307 }, { "epoch": 0.4185153583617747, "grad_norm": 0.5868337750434875, "learning_rate": 9.952621638498465e-06, "loss": 0.5379, "step": 1308 }, { "epoch": 0.4188353242320819, "grad_norm": 0.6548082828521729, "learning_rate": 9.95236560113021e-06, "loss": 0.6171, "step": 1309 }, { "epoch": 0.41915529010238906, "grad_norm": 0.6385195851325989, "learning_rate": 9.952108877110505e-06, "loss": 0.5671, "step": 1310 }, { "epoch": 0.41947525597269625, "grad_norm": 0.6664000153541565, "learning_rate": 9.951851466474949e-06, "loss": 0.5631, "step": 1311 }, { "epoch": 0.4197952218430034, "grad_norm": 0.6223545074462891, "learning_rate": 9.95159336925923e-06, "loss": 0.5661, "step": 1312 }, { "epoch": 0.4201151877133106, "grad_norm": 0.6739969253540039, "learning_rate": 9.95133458549913e-06, "loss": 0.5956, "step": 1313 }, { "epoch": 0.42043515358361777, "grad_norm": 0.6768685579299927, "learning_rate": 9.951075115230537e-06, "loss": 0.5835, "step": 1314 }, { "epoch": 0.4207551194539249, "grad_norm": 0.615283727645874, "learning_rate": 9.950814958489421e-06, "loss": 0.5621, "step": 1315 }, { "epoch": 0.4210750853242321, "grad_norm": 0.5856695175170898, "learning_rate": 9.950554115311855e-06, "loss": 0.5479, "step": 1316 }, { "epoch": 0.42139505119453924, "grad_norm": 0.6586756110191345, "learning_rate": 9.950292585734007e-06, "loss": 0.5547, "step": 1317 }, { "epoch": 0.42171501706484643, "grad_norm": 0.6436223387718201, "learning_rate": 9.950030369792133e-06, "loss": 0.5959, "step": 1318 }, { "epoch": 0.42203498293515357, "grad_norm": 0.5667495727539062, "learning_rate": 9.949767467522597e-06, "loss": 0.5868, "step": 1319 }, { "epoch": 0.42235494880546076, "grad_norm": 0.6080712080001831, "learning_rate": 9.949503878961846e-06, "loss": 0.5511, "step": 1320 }, { "epoch": 0.4226749146757679, "grad_norm": 0.6536353826522827, "learning_rate": 9.949239604146427e-06, "loss": 0.5547, "step": 1321 }, { "epoch": 0.4229948805460751, "grad_norm": 0.6362671256065369, "learning_rate": 9.948974643112984e-06, "loss": 0.5693, "step": 1322 }, { "epoch": 0.42331484641638223, "grad_norm": 0.6655951738357544, "learning_rate": 9.948708995898251e-06, "loss": 0.6167, "step": 1323 }, { "epoch": 0.4236348122866894, "grad_norm": 0.678941011428833, "learning_rate": 9.948442662539064e-06, "loss": 0.5996, "step": 1324 }, { "epoch": 0.4239547781569966, "grad_norm": 0.6312676668167114, "learning_rate": 9.948175643072347e-06, "loss": 0.575, "step": 1325 }, { "epoch": 0.42427474402730375, "grad_norm": 0.663619875907898, "learning_rate": 9.947907937535127e-06, "loss": 0.5598, "step": 1326 }, { "epoch": 0.42459470989761094, "grad_norm": 0.6370262503623962, "learning_rate": 9.947639545964516e-06, "loss": 0.5459, "step": 1327 }, { "epoch": 0.4249146757679181, "grad_norm": 0.6334279775619507, "learning_rate": 9.947370468397731e-06, "loss": 0.5924, "step": 1328 }, { "epoch": 0.4252346416382253, "grad_norm": 0.6385119557380676, "learning_rate": 9.947100704872079e-06, "loss": 0.521, "step": 1329 }, { "epoch": 0.4255546075085324, "grad_norm": 0.5783963799476624, "learning_rate": 9.946830255424964e-06, "loss": 0.5408, "step": 1330 }, { "epoch": 0.4258745733788396, "grad_norm": 0.6530171036720276, "learning_rate": 9.94655912009388e-06, "loss": 0.567, "step": 1331 }, { "epoch": 0.42619453924914674, "grad_norm": 0.6820838451385498, "learning_rate": 9.946287298916426e-06, "loss": 0.6293, "step": 1332 }, { "epoch": 0.42651450511945393, "grad_norm": 0.6430705785751343, "learning_rate": 9.946014791930285e-06, "loss": 0.5808, "step": 1333 }, { "epoch": 0.4268344709897611, "grad_norm": 0.69588702917099, "learning_rate": 9.945741599173244e-06, "loss": 0.5724, "step": 1334 }, { "epoch": 0.42715443686006827, "grad_norm": 0.6550391912460327, "learning_rate": 9.94546772068318e-06, "loss": 0.5921, "step": 1335 }, { "epoch": 0.4274744027303754, "grad_norm": 0.6094030737876892, "learning_rate": 9.945193156498068e-06, "loss": 0.5448, "step": 1336 }, { "epoch": 0.4277943686006826, "grad_norm": 0.7154574394226074, "learning_rate": 9.944917906655976e-06, "loss": 0.615, "step": 1337 }, { "epoch": 0.4281143344709898, "grad_norm": 0.5963408946990967, "learning_rate": 9.944641971195065e-06, "loss": 0.5823, "step": 1338 }, { "epoch": 0.4284343003412969, "grad_norm": 0.6101605892181396, "learning_rate": 9.944365350153598e-06, "loss": 0.5914, "step": 1339 }, { "epoch": 0.4287542662116041, "grad_norm": 0.655142068862915, "learning_rate": 9.944088043569925e-06, "loss": 0.5732, "step": 1340 }, { "epoch": 0.42907423208191126, "grad_norm": 0.6475589275360107, "learning_rate": 9.9438100514825e-06, "loss": 0.5754, "step": 1341 }, { "epoch": 0.42939419795221845, "grad_norm": 0.6456246376037598, "learning_rate": 9.943531373929863e-06, "loss": 0.5733, "step": 1342 }, { "epoch": 0.4297141638225256, "grad_norm": 0.7405171990394592, "learning_rate": 9.943252010950653e-06, "loss": 0.5914, "step": 1343 }, { "epoch": 0.4300341296928328, "grad_norm": 0.5655635595321655, "learning_rate": 9.942971962583605e-06, "loss": 0.5319, "step": 1344 }, { "epoch": 0.4303540955631399, "grad_norm": 0.6160438656806946, "learning_rate": 9.942691228867548e-06, "loss": 0.5907, "step": 1345 }, { "epoch": 0.4306740614334471, "grad_norm": 0.6776102781295776, "learning_rate": 9.942409809841404e-06, "loss": 0.5641, "step": 1346 }, { "epoch": 0.43099402730375425, "grad_norm": 0.6737133860588074, "learning_rate": 9.942127705544196e-06, "loss": 0.5942, "step": 1347 }, { "epoch": 0.43131399317406144, "grad_norm": 0.6324644684791565, "learning_rate": 9.941844916015035e-06, "loss": 0.601, "step": 1348 }, { "epoch": 0.4316339590443686, "grad_norm": 0.620689868927002, "learning_rate": 9.941561441293134e-06, "loss": 0.5364, "step": 1349 }, { "epoch": 0.43195392491467577, "grad_norm": 0.6571909189224243, "learning_rate": 9.941277281417792e-06, "loss": 0.6054, "step": 1350 }, { "epoch": 0.43227389078498296, "grad_norm": 0.5578046441078186, "learning_rate": 9.94099243642841e-06, "loss": 0.5415, "step": 1351 }, { "epoch": 0.4325938566552901, "grad_norm": 0.5772393345832825, "learning_rate": 9.940706906364484e-06, "loss": 0.5384, "step": 1352 }, { "epoch": 0.4329138225255973, "grad_norm": 0.6861337423324585, "learning_rate": 9.9404206912656e-06, "loss": 0.5939, "step": 1353 }, { "epoch": 0.43323378839590443, "grad_norm": 0.6963403820991516, "learning_rate": 9.940133791171445e-06, "loss": 0.5802, "step": 1354 }, { "epoch": 0.4335537542662116, "grad_norm": 0.564021646976471, "learning_rate": 9.939846206121795e-06, "loss": 0.5282, "step": 1355 }, { "epoch": 0.43387372013651876, "grad_norm": 0.6498206257820129, "learning_rate": 9.939557936156527e-06, "loss": 0.5259, "step": 1356 }, { "epoch": 0.43419368600682595, "grad_norm": 0.6931615471839905, "learning_rate": 9.939268981315609e-06, "loss": 0.5777, "step": 1357 }, { "epoch": 0.4345136518771331, "grad_norm": 0.6055909395217896, "learning_rate": 9.938979341639105e-06, "loss": 0.5769, "step": 1358 }, { "epoch": 0.4348336177474403, "grad_norm": 0.7766366004943848, "learning_rate": 9.938689017167173e-06, "loss": 0.573, "step": 1359 }, { "epoch": 0.4351535836177474, "grad_norm": 0.6288790106773376, "learning_rate": 9.938398007940067e-06, "loss": 0.5826, "step": 1360 }, { "epoch": 0.4354735494880546, "grad_norm": 0.6430128216743469, "learning_rate": 9.938106313998136e-06, "loss": 0.5781, "step": 1361 }, { "epoch": 0.43579351535836175, "grad_norm": 0.6517099142074585, "learning_rate": 9.937813935381823e-06, "loss": 0.5561, "step": 1362 }, { "epoch": 0.43611348122866894, "grad_norm": 0.6719871163368225, "learning_rate": 9.93752087213167e-06, "loss": 0.57, "step": 1363 }, { "epoch": 0.43643344709897613, "grad_norm": 0.6535536050796509, "learning_rate": 9.937227124288306e-06, "loss": 0.5627, "step": 1364 }, { "epoch": 0.4367534129692833, "grad_norm": 0.689322829246521, "learning_rate": 9.936932691892463e-06, "loss": 0.5457, "step": 1365 }, { "epoch": 0.43707337883959047, "grad_norm": 0.6880068182945251, "learning_rate": 9.936637574984964e-06, "loss": 0.5631, "step": 1366 }, { "epoch": 0.4373933447098976, "grad_norm": 0.7001758217811584, "learning_rate": 9.936341773606723e-06, "loss": 0.6106, "step": 1367 }, { "epoch": 0.4377133105802048, "grad_norm": 0.6887964010238647, "learning_rate": 9.936045287798759e-06, "loss": 0.5849, "step": 1368 }, { "epoch": 0.43803327645051193, "grad_norm": 0.7089389562606812, "learning_rate": 9.935748117602178e-06, "loss": 0.5952, "step": 1369 }, { "epoch": 0.4383532423208191, "grad_norm": 0.6172699928283691, "learning_rate": 9.93545026305818e-06, "loss": 0.5147, "step": 1370 }, { "epoch": 0.43867320819112626, "grad_norm": 0.6297218799591064, "learning_rate": 9.935151724208068e-06, "loss": 0.572, "step": 1371 }, { "epoch": 0.43899317406143346, "grad_norm": 0.7111726999282837, "learning_rate": 9.93485250109323e-06, "loss": 0.579, "step": 1372 }, { "epoch": 0.4393131399317406, "grad_norm": 0.6565432548522949, "learning_rate": 9.934552593755158e-06, "loss": 0.615, "step": 1373 }, { "epoch": 0.4396331058020478, "grad_norm": 0.6117081642150879, "learning_rate": 9.934252002235433e-06, "loss": 0.5565, "step": 1374 }, { "epoch": 0.4399530716723549, "grad_norm": 0.6642066836357117, "learning_rate": 9.93395072657573e-06, "loss": 0.5598, "step": 1375 }, { "epoch": 0.4402730375426621, "grad_norm": 0.6202808022499084, "learning_rate": 9.933648766817823e-06, "loss": 0.5716, "step": 1376 }, { "epoch": 0.4405930034129693, "grad_norm": 0.5759875178337097, "learning_rate": 9.933346123003581e-06, "loss": 0.5462, "step": 1377 }, { "epoch": 0.44091296928327645, "grad_norm": 0.7093290686607361, "learning_rate": 9.933042795174964e-06, "loss": 0.5705, "step": 1378 }, { "epoch": 0.44123293515358364, "grad_norm": 0.6566049456596375, "learning_rate": 9.932738783374028e-06, "loss": 0.5653, "step": 1379 }, { "epoch": 0.4415529010238908, "grad_norm": 0.631485641002655, "learning_rate": 9.932434087642924e-06, "loss": 0.5672, "step": 1380 }, { "epoch": 0.44187286689419797, "grad_norm": 0.6641365885734558, "learning_rate": 9.932128708023901e-06, "loss": 0.538, "step": 1381 }, { "epoch": 0.4421928327645051, "grad_norm": 0.574982762336731, "learning_rate": 9.931822644559298e-06, "loss": 0.5212, "step": 1382 }, { "epoch": 0.4425127986348123, "grad_norm": 0.6498568058013916, "learning_rate": 9.931515897291555e-06, "loss": 0.5818, "step": 1383 }, { "epoch": 0.44283276450511944, "grad_norm": 0.6279152035713196, "learning_rate": 9.931208466263198e-06, "loss": 0.5548, "step": 1384 }, { "epoch": 0.44315273037542663, "grad_norm": 0.6007976531982422, "learning_rate": 9.930900351516857e-06, "loss": 0.5456, "step": 1385 }, { "epoch": 0.44347269624573377, "grad_norm": 0.6235578060150146, "learning_rate": 9.930591553095248e-06, "loss": 0.6002, "step": 1386 }, { "epoch": 0.44379266211604096, "grad_norm": 0.6414409279823303, "learning_rate": 9.930282071041188e-06, "loss": 0.5304, "step": 1387 }, { "epoch": 0.4441126279863481, "grad_norm": 0.6062139272689819, "learning_rate": 9.92997190539759e-06, "loss": 0.5881, "step": 1388 }, { "epoch": 0.4444325938566553, "grad_norm": 0.6121253967285156, "learning_rate": 9.929661056207455e-06, "loss": 0.6046, "step": 1389 }, { "epoch": 0.4447525597269625, "grad_norm": 0.6918778419494629, "learning_rate": 9.929349523513884e-06, "loss": 0.5734, "step": 1390 }, { "epoch": 0.4450725255972696, "grad_norm": 0.6338698267936707, "learning_rate": 9.929037307360073e-06, "loss": 0.5541, "step": 1391 }, { "epoch": 0.4453924914675768, "grad_norm": 0.5875884890556335, "learning_rate": 9.928724407789307e-06, "loss": 0.5652, "step": 1392 }, { "epoch": 0.44571245733788395, "grad_norm": 0.6555237770080566, "learning_rate": 9.928410824844974e-06, "loss": 0.5742, "step": 1393 }, { "epoch": 0.44603242320819114, "grad_norm": 0.5992658734321594, "learning_rate": 9.92809655857055e-06, "loss": 0.5878, "step": 1394 }, { "epoch": 0.4463523890784983, "grad_norm": 0.6018292307853699, "learning_rate": 9.927781609009611e-06, "loss": 0.5827, "step": 1395 }, { "epoch": 0.4466723549488055, "grad_norm": 0.6248888969421387, "learning_rate": 9.927465976205823e-06, "loss": 0.5768, "step": 1396 }, { "epoch": 0.4469923208191126, "grad_norm": 0.6832228899002075, "learning_rate": 9.927149660202949e-06, "loss": 0.5906, "step": 1397 }, { "epoch": 0.4473122866894198, "grad_norm": 0.5579515695571899, "learning_rate": 9.926832661044848e-06, "loss": 0.5318, "step": 1398 }, { "epoch": 0.44763225255972694, "grad_norm": 0.7638499736785889, "learning_rate": 9.92651497877547e-06, "loss": 0.557, "step": 1399 }, { "epoch": 0.44795221843003413, "grad_norm": 0.5726019740104675, "learning_rate": 9.926196613438866e-06, "loss": 0.5835, "step": 1400 }, { "epoch": 0.44827218430034127, "grad_norm": 0.669023871421814, "learning_rate": 9.925877565079172e-06, "loss": 0.5944, "step": 1401 }, { "epoch": 0.44859215017064846, "grad_norm": 0.6804897785186768, "learning_rate": 9.92555783374063e-06, "loss": 0.5345, "step": 1402 }, { "epoch": 0.44891211604095566, "grad_norm": 0.7297967076301575, "learning_rate": 9.925237419467567e-06, "loss": 0.6176, "step": 1403 }, { "epoch": 0.4492320819112628, "grad_norm": 0.6806580424308777, "learning_rate": 9.924916322304413e-06, "loss": 0.5738, "step": 1404 }, { "epoch": 0.44955204778157, "grad_norm": 0.597823977470398, "learning_rate": 9.924594542295684e-06, "loss": 0.5565, "step": 1405 }, { "epoch": 0.4498720136518771, "grad_norm": 0.6245985627174377, "learning_rate": 9.924272079485996e-06, "loss": 0.5889, "step": 1406 }, { "epoch": 0.4501919795221843, "grad_norm": 0.6150023937225342, "learning_rate": 9.923948933920064e-06, "loss": 0.5652, "step": 1407 }, { "epoch": 0.45051194539249145, "grad_norm": 0.6009223461151123, "learning_rate": 9.923625105642687e-06, "loss": 0.5729, "step": 1408 }, { "epoch": 0.45083191126279865, "grad_norm": 0.5567916631698608, "learning_rate": 9.923300594698765e-06, "loss": 0.5315, "step": 1409 }, { "epoch": 0.4511518771331058, "grad_norm": 0.612503707408905, "learning_rate": 9.922975401133292e-06, "loss": 0.5928, "step": 1410 }, { "epoch": 0.451471843003413, "grad_norm": 0.6755174398422241, "learning_rate": 9.922649524991359e-06, "loss": 0.5171, "step": 1411 }, { "epoch": 0.4517918088737201, "grad_norm": 0.6449228525161743, "learning_rate": 9.922322966318148e-06, "loss": 0.6091, "step": 1412 }, { "epoch": 0.4521117747440273, "grad_norm": 0.6059353947639465, "learning_rate": 9.921995725158934e-06, "loss": 0.5472, "step": 1413 }, { "epoch": 0.45243174061433444, "grad_norm": 0.6328505277633667, "learning_rate": 9.921667801559092e-06, "loss": 0.5961, "step": 1414 }, { "epoch": 0.45275170648464164, "grad_norm": 0.6243288516998291, "learning_rate": 9.921339195564088e-06, "loss": 0.5667, "step": 1415 }, { "epoch": 0.45307167235494883, "grad_norm": 0.5473213791847229, "learning_rate": 9.921009907219485e-06, "loss": 0.5603, "step": 1416 }, { "epoch": 0.45339163822525597, "grad_norm": 0.5441519021987915, "learning_rate": 9.920679936570937e-06, "loss": 0.6045, "step": 1417 }, { "epoch": 0.45371160409556316, "grad_norm": 0.5660945773124695, "learning_rate": 9.920349283664197e-06, "loss": 0.5489, "step": 1418 }, { "epoch": 0.4540315699658703, "grad_norm": 0.5919684767723083, "learning_rate": 9.920017948545109e-06, "loss": 0.5622, "step": 1419 }, { "epoch": 0.4543515358361775, "grad_norm": 0.6301105618476868, "learning_rate": 9.919685931259615e-06, "loss": 0.5791, "step": 1420 }, { "epoch": 0.4546715017064846, "grad_norm": 0.6944140195846558, "learning_rate": 9.919353231853746e-06, "loss": 0.5882, "step": 1421 }, { "epoch": 0.4549914675767918, "grad_norm": 0.6283155083656311, "learning_rate": 9.919019850373635e-06, "loss": 0.5745, "step": 1422 }, { "epoch": 0.45531143344709896, "grad_norm": 0.5469874143600464, "learning_rate": 9.918685786865503e-06, "loss": 0.53, "step": 1423 }, { "epoch": 0.45563139931740615, "grad_norm": 0.5966683030128479, "learning_rate": 9.91835104137567e-06, "loss": 0.5867, "step": 1424 }, { "epoch": 0.4559513651877133, "grad_norm": 0.6650198101997375, "learning_rate": 9.918015613950548e-06, "loss": 0.5862, "step": 1425 }, { "epoch": 0.4562713310580205, "grad_norm": 0.6126238107681274, "learning_rate": 9.917679504636644e-06, "loss": 0.5622, "step": 1426 }, { "epoch": 0.4565912969283277, "grad_norm": 0.6327012777328491, "learning_rate": 9.91734271348056e-06, "loss": 0.5256, "step": 1427 }, { "epoch": 0.4569112627986348, "grad_norm": 0.608069896697998, "learning_rate": 9.917005240528994e-06, "loss": 0.5402, "step": 1428 }, { "epoch": 0.457231228668942, "grad_norm": 0.6601377725601196, "learning_rate": 9.916667085828735e-06, "loss": 0.562, "step": 1429 }, { "epoch": 0.45755119453924914, "grad_norm": 0.7051287889480591, "learning_rate": 9.916328249426671e-06, "loss": 0.5794, "step": 1430 }, { "epoch": 0.45787116040955633, "grad_norm": 0.5601382255554199, "learning_rate": 9.915988731369782e-06, "loss": 0.5622, "step": 1431 }, { "epoch": 0.45819112627986347, "grad_norm": 0.5865865349769592, "learning_rate": 9.91564853170514e-06, "loss": 0.5529, "step": 1432 }, { "epoch": 0.45851109215017066, "grad_norm": 0.6290679574012756, "learning_rate": 9.915307650479915e-06, "loss": 0.5504, "step": 1433 }, { "epoch": 0.4588310580204778, "grad_norm": 0.6417455077171326, "learning_rate": 9.91496608774137e-06, "loss": 0.5786, "step": 1434 }, { "epoch": 0.459151023890785, "grad_norm": 0.638545572757721, "learning_rate": 9.914623843536866e-06, "loss": 0.5837, "step": 1435 }, { "epoch": 0.45947098976109213, "grad_norm": 0.5751261711120605, "learning_rate": 9.914280917913851e-06, "loss": 0.5992, "step": 1436 }, { "epoch": 0.4597909556313993, "grad_norm": 0.6094741225242615, "learning_rate": 9.913937310919877e-06, "loss": 0.5261, "step": 1437 }, { "epoch": 0.46011092150170646, "grad_norm": 0.5998948812484741, "learning_rate": 9.913593022602583e-06, "loss": 0.5936, "step": 1438 }, { "epoch": 0.46043088737201365, "grad_norm": 0.7354366779327393, "learning_rate": 9.913248053009705e-06, "loss": 0.6342, "step": 1439 }, { "epoch": 0.46075085324232085, "grad_norm": 0.6196147799491882, "learning_rate": 9.912902402189075e-06, "loss": 0.576, "step": 1440 }, { "epoch": 0.461070819112628, "grad_norm": 0.6172345280647278, "learning_rate": 9.912556070188615e-06, "loss": 0.5871, "step": 1441 }, { "epoch": 0.4613907849829352, "grad_norm": 0.7067185044288635, "learning_rate": 9.912209057056345e-06, "loss": 0.5775, "step": 1442 }, { "epoch": 0.4617107508532423, "grad_norm": 0.607168436050415, "learning_rate": 9.911861362840382e-06, "loss": 0.579, "step": 1443 }, { "epoch": 0.4620307167235495, "grad_norm": 0.6715491414070129, "learning_rate": 9.911512987588932e-06, "loss": 0.6184, "step": 1444 }, { "epoch": 0.46235068259385664, "grad_norm": 0.6675828695297241, "learning_rate": 9.911163931350296e-06, "loss": 0.5802, "step": 1445 }, { "epoch": 0.46267064846416384, "grad_norm": 0.5708353519439697, "learning_rate": 9.910814194172872e-06, "loss": 0.5371, "step": 1446 }, { "epoch": 0.462990614334471, "grad_norm": 0.599196195602417, "learning_rate": 9.910463776105154e-06, "loss": 0.5975, "step": 1447 }, { "epoch": 0.46331058020477817, "grad_norm": 0.635504961013794, "learning_rate": 9.910112677195724e-06, "loss": 0.5694, "step": 1448 }, { "epoch": 0.4636305460750853, "grad_norm": 0.6071615815162659, "learning_rate": 9.909760897493267e-06, "loss": 0.5395, "step": 1449 }, { "epoch": 0.4639505119453925, "grad_norm": 0.6165795922279358, "learning_rate": 9.909408437046553e-06, "loss": 0.5666, "step": 1450 }, { "epoch": 0.46427047781569963, "grad_norm": 0.6419926285743713, "learning_rate": 9.909055295904453e-06, "loss": 0.5628, "step": 1451 }, { "epoch": 0.4645904436860068, "grad_norm": 0.6607449650764465, "learning_rate": 9.908701474115931e-06, "loss": 0.5526, "step": 1452 }, { "epoch": 0.464910409556314, "grad_norm": 0.6067793369293213, "learning_rate": 9.908346971730045e-06, "loss": 0.5443, "step": 1453 }, { "epoch": 0.46523037542662116, "grad_norm": 0.6271002292633057, "learning_rate": 9.907991788795945e-06, "loss": 0.5797, "step": 1454 }, { "epoch": 0.46555034129692835, "grad_norm": 0.6588297486305237, "learning_rate": 9.90763592536288e-06, "loss": 0.5894, "step": 1455 }, { "epoch": 0.4658703071672355, "grad_norm": 0.6160594820976257, "learning_rate": 9.90727938148019e-06, "loss": 0.5708, "step": 1456 }, { "epoch": 0.4661902730375427, "grad_norm": 0.742313027381897, "learning_rate": 9.90692215719731e-06, "loss": 0.5786, "step": 1457 }, { "epoch": 0.4665102389078498, "grad_norm": 0.6778494119644165, "learning_rate": 9.906564252563769e-06, "loss": 0.5975, "step": 1458 }, { "epoch": 0.466830204778157, "grad_norm": 0.6413978934288025, "learning_rate": 9.906205667629193e-06, "loss": 0.5772, "step": 1459 }, { "epoch": 0.46715017064846415, "grad_norm": 0.6484549641609192, "learning_rate": 9.905846402443297e-06, "loss": 0.5799, "step": 1460 }, { "epoch": 0.46747013651877134, "grad_norm": 0.5865228176116943, "learning_rate": 9.905486457055897e-06, "loss": 0.5508, "step": 1461 }, { "epoch": 0.4677901023890785, "grad_norm": 0.6155765652656555, "learning_rate": 9.905125831516899e-06, "loss": 0.5502, "step": 1462 }, { "epoch": 0.46811006825938567, "grad_norm": 0.5499103665351868, "learning_rate": 9.904764525876301e-06, "loss": 0.5334, "step": 1463 }, { "epoch": 0.4684300341296928, "grad_norm": 0.5945428609848022, "learning_rate": 9.904402540184203e-06, "loss": 0.6077, "step": 1464 }, { "epoch": 0.46875, "grad_norm": 0.6026771068572998, "learning_rate": 9.904039874490791e-06, "loss": 0.5386, "step": 1465 }, { "epoch": 0.4690699658703072, "grad_norm": 0.564043402671814, "learning_rate": 9.903676528846353e-06, "loss": 0.5591, "step": 1466 }, { "epoch": 0.46938993174061433, "grad_norm": 0.6140351295471191, "learning_rate": 9.903312503301263e-06, "loss": 0.5584, "step": 1467 }, { "epoch": 0.4697098976109215, "grad_norm": 0.564509391784668, "learning_rate": 9.902947797905997e-06, "loss": 0.5536, "step": 1468 }, { "epoch": 0.47002986348122866, "grad_norm": 0.5968620777130127, "learning_rate": 9.90258241271112e-06, "loss": 0.6038, "step": 1469 }, { "epoch": 0.47034982935153585, "grad_norm": 0.7169108986854553, "learning_rate": 9.902216347767292e-06, "loss": 0.5891, "step": 1470 }, { "epoch": 0.470669795221843, "grad_norm": 0.6527320146560669, "learning_rate": 9.901849603125271e-06, "loss": 0.6206, "step": 1471 }, { "epoch": 0.4709897610921502, "grad_norm": 0.695573091506958, "learning_rate": 9.901482178835907e-06, "loss": 0.5959, "step": 1472 }, { "epoch": 0.4713097269624573, "grad_norm": 0.6598088145256042, "learning_rate": 9.90111407495014e-06, "loss": 0.5587, "step": 1473 }, { "epoch": 0.4716296928327645, "grad_norm": 0.6323714852333069, "learning_rate": 9.90074529151901e-06, "loss": 0.5583, "step": 1474 }, { "epoch": 0.47194965870307165, "grad_norm": 0.6153364181518555, "learning_rate": 9.90037582859365e-06, "loss": 0.5852, "step": 1475 }, { "epoch": 0.47226962457337884, "grad_norm": 0.6652483344078064, "learning_rate": 9.900005686225288e-06, "loss": 0.587, "step": 1476 }, { "epoch": 0.472589590443686, "grad_norm": 0.6001269817352295, "learning_rate": 9.899634864465242e-06, "loss": 0.5545, "step": 1477 }, { "epoch": 0.4729095563139932, "grad_norm": 0.6979308128356934, "learning_rate": 9.899263363364926e-06, "loss": 0.5771, "step": 1478 }, { "epoch": 0.47322952218430037, "grad_norm": 0.6638851165771484, "learning_rate": 9.898891182975852e-06, "loss": 0.602, "step": 1479 }, { "epoch": 0.4735494880546075, "grad_norm": 0.593870997428894, "learning_rate": 9.898518323349621e-06, "loss": 0.5897, "step": 1480 }, { "epoch": 0.4738694539249147, "grad_norm": 0.5697798132896423, "learning_rate": 9.898144784537933e-06, "loss": 0.5594, "step": 1481 }, { "epoch": 0.47418941979522183, "grad_norm": 0.642444908618927, "learning_rate": 9.897770566592578e-06, "loss": 0.5916, "step": 1482 }, { "epoch": 0.474509385665529, "grad_norm": 0.6214845776557922, "learning_rate": 9.897395669565439e-06, "loss": 0.5523, "step": 1483 }, { "epoch": 0.47482935153583616, "grad_norm": 0.6055586934089661, "learning_rate": 9.897020093508502e-06, "loss": 0.5692, "step": 1484 }, { "epoch": 0.47514931740614336, "grad_norm": 0.7527763843536377, "learning_rate": 9.896643838473836e-06, "loss": 0.5906, "step": 1485 }, { "epoch": 0.4754692832764505, "grad_norm": 0.7165666818618774, "learning_rate": 9.896266904513612e-06, "loss": 0.596, "step": 1486 }, { "epoch": 0.4757892491467577, "grad_norm": 0.5976629257202148, "learning_rate": 9.89588929168009e-06, "loss": 0.5478, "step": 1487 }, { "epoch": 0.4761092150170648, "grad_norm": 0.6347126960754395, "learning_rate": 9.89551100002563e-06, "loss": 0.5746, "step": 1488 }, { "epoch": 0.476429180887372, "grad_norm": 0.6473086476325989, "learning_rate": 9.895132029602679e-06, "loss": 0.62, "step": 1489 }, { "epoch": 0.47674914675767915, "grad_norm": 0.664999783039093, "learning_rate": 9.894752380463782e-06, "loss": 0.561, "step": 1490 }, { "epoch": 0.47706911262798635, "grad_norm": 0.5583751201629639, "learning_rate": 9.894372052661581e-06, "loss": 0.5326, "step": 1491 }, { "epoch": 0.47738907849829354, "grad_norm": 0.6377359628677368, "learning_rate": 9.893991046248806e-06, "loss": 0.5871, "step": 1492 }, { "epoch": 0.4777090443686007, "grad_norm": 0.6565425992012024, "learning_rate": 9.893609361278284e-06, "loss": 0.5779, "step": 1493 }, { "epoch": 0.47802901023890787, "grad_norm": 0.6273458003997803, "learning_rate": 9.893226997802936e-06, "loss": 0.5339, "step": 1494 }, { "epoch": 0.478348976109215, "grad_norm": 0.5752052068710327, "learning_rate": 9.892843955875782e-06, "loss": 0.5491, "step": 1495 }, { "epoch": 0.4786689419795222, "grad_norm": 0.5871261954307556, "learning_rate": 9.892460235549923e-06, "loss": 0.6143, "step": 1496 }, { "epoch": 0.47898890784982934, "grad_norm": 0.6665970683097839, "learning_rate": 9.89207583687857e-06, "loss": 0.591, "step": 1497 }, { "epoch": 0.47930887372013653, "grad_norm": 0.6098381876945496, "learning_rate": 9.891690759915016e-06, "loss": 0.5687, "step": 1498 }, { "epoch": 0.47962883959044367, "grad_norm": 0.6258179545402527, "learning_rate": 9.891305004712652e-06, "loss": 0.5505, "step": 1499 }, { "epoch": 0.47994880546075086, "grad_norm": 0.6166390180587769, "learning_rate": 9.890918571324966e-06, "loss": 0.5425, "step": 1500 }, { "epoch": 0.480268771331058, "grad_norm": 0.7183076739311218, "learning_rate": 9.890531459805535e-06, "loss": 0.5511, "step": 1501 }, { "epoch": 0.4805887372013652, "grad_norm": 0.6725587248802185, "learning_rate": 9.890143670208034e-06, "loss": 0.56, "step": 1502 }, { "epoch": 0.4809087030716723, "grad_norm": 0.6727641820907593, "learning_rate": 9.88975520258623e-06, "loss": 0.5736, "step": 1503 }, { "epoch": 0.4812286689419795, "grad_norm": 0.6162008047103882, "learning_rate": 9.889366056993985e-06, "loss": 0.5851, "step": 1504 }, { "epoch": 0.4815486348122867, "grad_norm": 0.6258243322372437, "learning_rate": 9.888976233485254e-06, "loss": 0.5272, "step": 1505 }, { "epoch": 0.48186860068259385, "grad_norm": 0.6168959736824036, "learning_rate": 9.888585732114086e-06, "loss": 0.5632, "step": 1506 }, { "epoch": 0.48218856655290104, "grad_norm": 0.6138393878936768, "learning_rate": 9.888194552934627e-06, "loss": 0.5812, "step": 1507 }, { "epoch": 0.4825085324232082, "grad_norm": 0.5661769509315491, "learning_rate": 9.887802696001112e-06, "loss": 0.562, "step": 1508 }, { "epoch": 0.4828284982935154, "grad_norm": 0.6120560765266418, "learning_rate": 9.887410161367873e-06, "loss": 0.5828, "step": 1509 }, { "epoch": 0.4831484641638225, "grad_norm": 0.6099132895469666, "learning_rate": 9.887016949089334e-06, "loss": 0.5826, "step": 1510 }, { "epoch": 0.4834684300341297, "grad_norm": 0.6493592858314514, "learning_rate": 9.886623059220018e-06, "loss": 0.5573, "step": 1511 }, { "epoch": 0.48378839590443684, "grad_norm": 0.6510331034660339, "learning_rate": 9.886228491814533e-06, "loss": 0.6168, "step": 1512 }, { "epoch": 0.48410836177474403, "grad_norm": 0.6180525422096252, "learning_rate": 9.885833246927592e-06, "loss": 0.5541, "step": 1513 }, { "epoch": 0.48442832764505117, "grad_norm": 0.6133224964141846, "learning_rate": 9.885437324613993e-06, "loss": 0.5828, "step": 1514 }, { "epoch": 0.48474829351535836, "grad_norm": 0.6018075942993164, "learning_rate": 9.88504072492863e-06, "loss": 0.5662, "step": 1515 }, { "epoch": 0.48506825938566556, "grad_norm": 0.5534994006156921, "learning_rate": 9.884643447926496e-06, "loss": 0.5982, "step": 1516 }, { "epoch": 0.4853882252559727, "grad_norm": 0.5935896039009094, "learning_rate": 9.88424549366267e-06, "loss": 0.5697, "step": 1517 }, { "epoch": 0.4857081911262799, "grad_norm": 0.5847989320755005, "learning_rate": 9.88384686219233e-06, "loss": 0.5552, "step": 1518 }, { "epoch": 0.486028156996587, "grad_norm": 0.647017240524292, "learning_rate": 9.883447553570749e-06, "loss": 0.5672, "step": 1519 }, { "epoch": 0.4863481228668942, "grad_norm": 0.634661078453064, "learning_rate": 9.88304756785329e-06, "loss": 0.5312, "step": 1520 }, { "epoch": 0.48666808873720135, "grad_norm": 0.5421854257583618, "learning_rate": 9.88264690509541e-06, "loss": 0.5452, "step": 1521 }, { "epoch": 0.48698805460750855, "grad_norm": 0.6323463320732117, "learning_rate": 9.882245565352662e-06, "loss": 0.5914, "step": 1522 }, { "epoch": 0.4873080204778157, "grad_norm": 0.6671409010887146, "learning_rate": 9.881843548680694e-06, "loss": 0.5507, "step": 1523 }, { "epoch": 0.4876279863481229, "grad_norm": 0.6813701391220093, "learning_rate": 9.881440855135245e-06, "loss": 0.6005, "step": 1524 }, { "epoch": 0.48794795221843, "grad_norm": 0.5512477159500122, "learning_rate": 9.881037484772148e-06, "loss": 0.5743, "step": 1525 }, { "epoch": 0.4882679180887372, "grad_norm": 0.6768247485160828, "learning_rate": 9.880633437647333e-06, "loss": 0.5925, "step": 1526 }, { "epoch": 0.48858788395904434, "grad_norm": 0.6516999006271362, "learning_rate": 9.880228713816819e-06, "loss": 0.5655, "step": 1527 }, { "epoch": 0.48890784982935154, "grad_norm": 0.5851047039031982, "learning_rate": 9.879823313336723e-06, "loss": 0.5573, "step": 1528 }, { "epoch": 0.48922781569965873, "grad_norm": 0.6633613705635071, "learning_rate": 9.879417236263255e-06, "loss": 0.5639, "step": 1529 }, { "epoch": 0.48954778156996587, "grad_norm": 0.5621408224105835, "learning_rate": 9.879010482652716e-06, "loss": 0.5566, "step": 1530 }, { "epoch": 0.48986774744027306, "grad_norm": 0.7079163789749146, "learning_rate": 9.878603052561503e-06, "loss": 0.5656, "step": 1531 }, { "epoch": 0.4901877133105802, "grad_norm": 0.6103261709213257, "learning_rate": 9.87819494604611e-06, "loss": 0.5482, "step": 1532 }, { "epoch": 0.4905076791808874, "grad_norm": 0.6498193144798279, "learning_rate": 9.877786163163119e-06, "loss": 0.5401, "step": 1533 }, { "epoch": 0.4908276450511945, "grad_norm": 0.6479784250259399, "learning_rate": 9.877376703969207e-06, "loss": 0.5726, "step": 1534 }, { "epoch": 0.4911476109215017, "grad_norm": 0.6524590253829956, "learning_rate": 9.876966568521148e-06, "loss": 0.5684, "step": 1535 }, { "epoch": 0.49146757679180886, "grad_norm": 0.5773913264274597, "learning_rate": 9.876555756875807e-06, "loss": 0.5931, "step": 1536 }, { "epoch": 0.49178754266211605, "grad_norm": 0.7117459177970886, "learning_rate": 9.876144269090144e-06, "loss": 0.5374, "step": 1537 }, { "epoch": 0.4921075085324232, "grad_norm": 0.5589005351066589, "learning_rate": 9.875732105221211e-06, "loss": 0.5523, "step": 1538 }, { "epoch": 0.4924274744027304, "grad_norm": 0.6223816275596619, "learning_rate": 9.875319265326158e-06, "loss": 0.5407, "step": 1539 }, { "epoch": 0.4927474402730375, "grad_norm": 0.6707935333251953, "learning_rate": 9.874905749462222e-06, "loss": 0.5952, "step": 1540 }, { "epoch": 0.4930674061433447, "grad_norm": 0.5833324790000916, "learning_rate": 9.87449155768674e-06, "loss": 0.5847, "step": 1541 }, { "epoch": 0.4933873720136519, "grad_norm": 0.6765121817588806, "learning_rate": 9.874076690057139e-06, "loss": 0.5881, "step": 1542 }, { "epoch": 0.49370733788395904, "grad_norm": 0.6047101020812988, "learning_rate": 9.87366114663094e-06, "loss": 0.58, "step": 1543 }, { "epoch": 0.49402730375426623, "grad_norm": 0.6373699307441711, "learning_rate": 9.873244927465762e-06, "loss": 0.5745, "step": 1544 }, { "epoch": 0.49434726962457337, "grad_norm": 0.6877090334892273, "learning_rate": 9.87282803261931e-06, "loss": 0.5312, "step": 1545 }, { "epoch": 0.49466723549488056, "grad_norm": 0.6187264919281006, "learning_rate": 9.872410462149392e-06, "loss": 0.5746, "step": 1546 }, { "epoch": 0.4949872013651877, "grad_norm": 0.5987083911895752, "learning_rate": 9.8719922161139e-06, "loss": 0.5413, "step": 1547 }, { "epoch": 0.4953071672354949, "grad_norm": 0.662311315536499, "learning_rate": 9.871573294570825e-06, "loss": 0.6106, "step": 1548 }, { "epoch": 0.49562713310580203, "grad_norm": 0.6173791289329529, "learning_rate": 9.871153697578254e-06, "loss": 0.586, "step": 1549 }, { "epoch": 0.4959470989761092, "grad_norm": 0.587100625038147, "learning_rate": 9.870733425194362e-06, "loss": 0.5345, "step": 1550 }, { "epoch": 0.49626706484641636, "grad_norm": 0.6461461782455444, "learning_rate": 9.870312477477421e-06, "loss": 0.5861, "step": 1551 }, { "epoch": 0.49658703071672355, "grad_norm": 0.6375117897987366, "learning_rate": 9.869890854485795e-06, "loss": 0.6194, "step": 1552 }, { "epoch": 0.4969069965870307, "grad_norm": 0.6479952931404114, "learning_rate": 9.869468556277945e-06, "loss": 0.5479, "step": 1553 }, { "epoch": 0.4972269624573379, "grad_norm": 0.6575143337249756, "learning_rate": 9.869045582912419e-06, "loss": 0.5822, "step": 1554 }, { "epoch": 0.4975469283276451, "grad_norm": 0.5700595378875732, "learning_rate": 9.868621934447869e-06, "loss": 0.5529, "step": 1555 }, { "epoch": 0.4978668941979522, "grad_norm": 0.62436842918396, "learning_rate": 9.868197610943028e-06, "loss": 0.5779, "step": 1556 }, { "epoch": 0.4981868600682594, "grad_norm": 0.6780308485031128, "learning_rate": 9.867772612456731e-06, "loss": 0.5869, "step": 1557 }, { "epoch": 0.49850682593856654, "grad_norm": 0.6406908631324768, "learning_rate": 9.867346939047909e-06, "loss": 0.5717, "step": 1558 }, { "epoch": 0.49882679180887374, "grad_norm": 0.6032212376594543, "learning_rate": 9.866920590775575e-06, "loss": 0.5775, "step": 1559 }, { "epoch": 0.4991467576791809, "grad_norm": 0.5806117653846741, "learning_rate": 9.866493567698847e-06, "loss": 0.5663, "step": 1560 }, { "epoch": 0.49946672354948807, "grad_norm": 0.5774424076080322, "learning_rate": 9.866065869876932e-06, "loss": 0.5421, "step": 1561 }, { "epoch": 0.4997866894197952, "grad_norm": 0.5617887377738953, "learning_rate": 9.86563749736913e-06, "loss": 0.5653, "step": 1562 }, { "epoch": 0.5001066552901023, "grad_norm": 0.6689897179603577, "learning_rate": 9.865208450234834e-06, "loss": 0.5195, "step": 1563 }, { "epoch": 0.5004266211604096, "grad_norm": 0.6852822303771973, "learning_rate": 9.864778728533535e-06, "loss": 0.5569, "step": 1564 }, { "epoch": 0.5007465870307167, "grad_norm": 0.5707159638404846, "learning_rate": 9.864348332324811e-06, "loss": 0.5549, "step": 1565 }, { "epoch": 0.5010665529010239, "grad_norm": 0.7322942018508911, "learning_rate": 9.86391726166834e-06, "loss": 0.6051, "step": 1566 }, { "epoch": 0.5013865187713311, "grad_norm": 0.602660596370697, "learning_rate": 9.863485516623889e-06, "loss": 0.5289, "step": 1567 }, { "epoch": 0.5017064846416383, "grad_norm": 0.6694923043251038, "learning_rate": 9.86305309725132e-06, "loss": 0.5844, "step": 1568 }, { "epoch": 0.5020264505119454, "grad_norm": 0.6712828874588013, "learning_rate": 9.86262000361059e-06, "loss": 0.6245, "step": 1569 }, { "epoch": 0.5023464163822525, "grad_norm": 0.5861441493034363, "learning_rate": 9.862186235761746e-06, "loss": 0.582, "step": 1570 }, { "epoch": 0.5026663822525598, "grad_norm": 0.5756668448448181, "learning_rate": 9.86175179376493e-06, "loss": 0.5677, "step": 1571 }, { "epoch": 0.5029863481228669, "grad_norm": 0.6047782301902771, "learning_rate": 9.861316677680379e-06, "loss": 0.5764, "step": 1572 }, { "epoch": 0.503306313993174, "grad_norm": 0.6280213594436646, "learning_rate": 9.860880887568423e-06, "loss": 0.613, "step": 1573 }, { "epoch": 0.5036262798634812, "grad_norm": 0.645076334476471, "learning_rate": 9.860444423489485e-06, "loss": 0.5216, "step": 1574 }, { "epoch": 0.5039462457337884, "grad_norm": 0.6336789727210999, "learning_rate": 9.860007285504079e-06, "loss": 0.5584, "step": 1575 }, { "epoch": 0.5042662116040956, "grad_norm": 0.5808457732200623, "learning_rate": 9.859569473672816e-06, "loss": 0.5566, "step": 1576 }, { "epoch": 0.5045861774744027, "grad_norm": 0.5337081551551819, "learning_rate": 9.859130988056402e-06, "loss": 0.5731, "step": 1577 }, { "epoch": 0.5049061433447098, "grad_norm": 0.6150102019309998, "learning_rate": 9.858691828715627e-06, "loss": 0.56, "step": 1578 }, { "epoch": 0.5052261092150171, "grad_norm": 0.5708062052726746, "learning_rate": 9.858251995711388e-06, "loss": 0.5633, "step": 1579 }, { "epoch": 0.5055460750853242, "grad_norm": 0.6003345251083374, "learning_rate": 9.857811489104663e-06, "loss": 0.5849, "step": 1580 }, { "epoch": 0.5058660409556314, "grad_norm": 0.5903301239013672, "learning_rate": 9.857370308956532e-06, "loss": 0.5836, "step": 1581 }, { "epoch": 0.5061860068259386, "grad_norm": 0.5505856275558472, "learning_rate": 9.856928455328163e-06, "loss": 0.5519, "step": 1582 }, { "epoch": 0.5065059726962458, "grad_norm": 0.6287134289741516, "learning_rate": 9.856485928280822e-06, "loss": 0.5945, "step": 1583 }, { "epoch": 0.5068259385665529, "grad_norm": 0.6269447207450867, "learning_rate": 9.856042727875864e-06, "loss": 0.6088, "step": 1584 }, { "epoch": 0.50714590443686, "grad_norm": 0.6402657628059387, "learning_rate": 9.85559885417474e-06, "loss": 0.5714, "step": 1585 }, { "epoch": 0.5074658703071673, "grad_norm": 0.6432064771652222, "learning_rate": 9.855154307238993e-06, "loss": 0.5953, "step": 1586 }, { "epoch": 0.5077858361774744, "grad_norm": 0.5881626605987549, "learning_rate": 9.854709087130261e-06, "loss": 0.5384, "step": 1587 }, { "epoch": 0.5081058020477816, "grad_norm": 0.6792723536491394, "learning_rate": 9.854263193910274e-06, "loss": 0.6163, "step": 1588 }, { "epoch": 0.5084257679180887, "grad_norm": 0.6320458054542542, "learning_rate": 9.853816627640854e-06, "loss": 0.6078, "step": 1589 }, { "epoch": 0.5087457337883959, "grad_norm": 0.6325360536575317, "learning_rate": 9.85336938838392e-06, "loss": 0.6139, "step": 1590 }, { "epoch": 0.5090656996587031, "grad_norm": 0.5258231163024902, "learning_rate": 9.852921476201482e-06, "loss": 0.5716, "step": 1591 }, { "epoch": 0.5093856655290102, "grad_norm": 0.575778067111969, "learning_rate": 9.852472891155642e-06, "loss": 0.5962, "step": 1592 }, { "epoch": 0.5097056313993175, "grad_norm": 0.6357591152191162, "learning_rate": 9.852023633308599e-06, "loss": 0.5391, "step": 1593 }, { "epoch": 0.5100255972696246, "grad_norm": 0.6247155070304871, "learning_rate": 9.851573702722641e-06, "loss": 0.5548, "step": 1594 }, { "epoch": 0.5103455631399317, "grad_norm": 0.6489351987838745, "learning_rate": 9.851123099460153e-06, "loss": 0.5942, "step": 1595 }, { "epoch": 0.5106655290102389, "grad_norm": 0.5720632672309875, "learning_rate": 9.850671823583614e-06, "loss": 0.5263, "step": 1596 }, { "epoch": 0.5109854948805461, "grad_norm": 0.6225847601890564, "learning_rate": 9.850219875155586e-06, "loss": 0.5797, "step": 1597 }, { "epoch": 0.5113054607508533, "grad_norm": 0.669680655002594, "learning_rate": 9.849767254238741e-06, "loss": 0.5773, "step": 1598 }, { "epoch": 0.5116254266211604, "grad_norm": 0.6217535138130188, "learning_rate": 9.84931396089583e-06, "loss": 0.53, "step": 1599 }, { "epoch": 0.5119453924914675, "grad_norm": 0.5527267456054688, "learning_rate": 9.848859995189705e-06, "loss": 0.5335, "step": 1600 }, { "epoch": 0.5122653583617748, "grad_norm": 0.640916645526886, "learning_rate": 9.84840535718331e-06, "loss": 0.5725, "step": 1601 }, { "epoch": 0.5125853242320819, "grad_norm": 0.6258885264396667, "learning_rate": 9.847950046939679e-06, "loss": 0.5809, "step": 1602 }, { "epoch": 0.512905290102389, "grad_norm": 0.6089369058609009, "learning_rate": 9.84749406452194e-06, "loss": 0.5531, "step": 1603 }, { "epoch": 0.5132252559726962, "grad_norm": 0.6907296180725098, "learning_rate": 9.847037409993318e-06, "loss": 0.5515, "step": 1604 }, { "epoch": 0.5135452218430034, "grad_norm": 0.6565805077552795, "learning_rate": 9.846580083417128e-06, "loss": 0.5793, "step": 1605 }, { "epoch": 0.5138651877133106, "grad_norm": 0.5883509516716003, "learning_rate": 9.84612208485678e-06, "loss": 0.5424, "step": 1606 }, { "epoch": 0.5141851535836177, "grad_norm": 0.6544880867004395, "learning_rate": 9.845663414375772e-06, "loss": 0.569, "step": 1607 }, { "epoch": 0.514505119453925, "grad_norm": 0.6848540306091309, "learning_rate": 9.845204072037705e-06, "loss": 0.573, "step": 1608 }, { "epoch": 0.5148250853242321, "grad_norm": 0.6440815329551697, "learning_rate": 9.844744057906263e-06, "loss": 0.5743, "step": 1609 }, { "epoch": 0.5151450511945392, "grad_norm": 0.5966008901596069, "learning_rate": 9.844283372045231e-06, "loss": 0.5526, "step": 1610 }, { "epoch": 0.5154650170648464, "grad_norm": 0.6927117705345154, "learning_rate": 9.843822014518479e-06, "loss": 0.581, "step": 1611 }, { "epoch": 0.5157849829351536, "grad_norm": 0.6129295229911804, "learning_rate": 9.843359985389978e-06, "loss": 0.6324, "step": 1612 }, { "epoch": 0.5161049488054608, "grad_norm": 0.5667604207992554, "learning_rate": 9.84289728472379e-06, "loss": 0.5525, "step": 1613 }, { "epoch": 0.5164249146757679, "grad_norm": 0.6120064854621887, "learning_rate": 9.842433912584066e-06, "loss": 0.6035, "step": 1614 }, { "epoch": 0.516744880546075, "grad_norm": 0.5590425133705139, "learning_rate": 9.841969869035055e-06, "loss": 0.5454, "step": 1615 }, { "epoch": 0.5170648464163823, "grad_norm": 0.5539948344230652, "learning_rate": 9.841505154141095e-06, "loss": 0.5396, "step": 1616 }, { "epoch": 0.5173848122866894, "grad_norm": 0.5885478258132935, "learning_rate": 9.841039767966622e-06, "loss": 0.5496, "step": 1617 }, { "epoch": 0.5177047781569966, "grad_norm": 0.5543431043624878, "learning_rate": 9.840573710576163e-06, "loss": 0.5502, "step": 1618 }, { "epoch": 0.5180247440273038, "grad_norm": 0.5092489123344421, "learning_rate": 9.840106982034335e-06, "loss": 0.5557, "step": 1619 }, { "epoch": 0.518344709897611, "grad_norm": 0.5805311799049377, "learning_rate": 9.83963958240585e-06, "loss": 0.5748, "step": 1620 }, { "epoch": 0.5186646757679181, "grad_norm": 0.5589832663536072, "learning_rate": 9.839171511755514e-06, "loss": 0.5251, "step": 1621 }, { "epoch": 0.5189846416382252, "grad_norm": 0.6321656703948975, "learning_rate": 9.838702770148227e-06, "loss": 0.5652, "step": 1622 }, { "epoch": 0.5193046075085325, "grad_norm": 0.5689617991447449, "learning_rate": 9.83823335764898e-06, "loss": 0.5823, "step": 1623 }, { "epoch": 0.5196245733788396, "grad_norm": 0.6367678642272949, "learning_rate": 9.837763274322858e-06, "loss": 0.5428, "step": 1624 }, { "epoch": 0.5199445392491467, "grad_norm": 0.6572225093841553, "learning_rate": 9.837292520235037e-06, "loss": 0.5924, "step": 1625 }, { "epoch": 0.5202645051194539, "grad_norm": 0.5359474420547485, "learning_rate": 9.836821095450789e-06, "loss": 0.5428, "step": 1626 }, { "epoch": 0.5205844709897611, "grad_norm": 0.6847219467163086, "learning_rate": 9.836349000035477e-06, "loss": 0.5699, "step": 1627 }, { "epoch": 0.5209044368600683, "grad_norm": 0.5682468414306641, "learning_rate": 9.835876234054558e-06, "loss": 0.5705, "step": 1628 }, { "epoch": 0.5212244027303754, "grad_norm": 0.6233001947402954, "learning_rate": 9.835402797573582e-06, "loss": 0.602, "step": 1629 }, { "epoch": 0.5215443686006825, "grad_norm": 0.6758689880371094, "learning_rate": 9.834928690658191e-06, "loss": 0.5808, "step": 1630 }, { "epoch": 0.5218643344709898, "grad_norm": 0.6670978665351868, "learning_rate": 9.83445391337412e-06, "loss": 0.5779, "step": 1631 }, { "epoch": 0.5221843003412969, "grad_norm": 0.5305650234222412, "learning_rate": 9.833978465787198e-06, "loss": 0.5356, "step": 1632 }, { "epoch": 0.5225042662116041, "grad_norm": 0.7348790764808655, "learning_rate": 9.833502347963346e-06, "loss": 0.539, "step": 1633 }, { "epoch": 0.5228242320819113, "grad_norm": 0.7539340257644653, "learning_rate": 9.83302555996858e-06, "loss": 0.566, "step": 1634 }, { "epoch": 0.5231441979522184, "grad_norm": 0.6037124395370483, "learning_rate": 9.832548101869004e-06, "loss": 0.5512, "step": 1635 }, { "epoch": 0.5234641638225256, "grad_norm": 0.6238608956336975, "learning_rate": 9.832069973730821e-06, "loss": 0.5866, "step": 1636 }, { "epoch": 0.5237841296928327, "grad_norm": 0.6644582152366638, "learning_rate": 9.831591175620323e-06, "loss": 0.5994, "step": 1637 }, { "epoch": 0.52410409556314, "grad_norm": 0.6156083345413208, "learning_rate": 9.831111707603896e-06, "loss": 0.5531, "step": 1638 }, { "epoch": 0.5244240614334471, "grad_norm": 0.6902807950973511, "learning_rate": 9.830631569748018e-06, "loss": 0.5558, "step": 1639 }, { "epoch": 0.5247440273037542, "grad_norm": 0.6093671321868896, "learning_rate": 9.83015076211926e-06, "loss": 0.5589, "step": 1640 }, { "epoch": 0.5250639931740614, "grad_norm": 0.6647526025772095, "learning_rate": 9.829669284784292e-06, "loss": 0.5418, "step": 1641 }, { "epoch": 0.5253839590443686, "grad_norm": 0.6453307271003723, "learning_rate": 9.829187137809865e-06, "loss": 0.5448, "step": 1642 }, { "epoch": 0.5257039249146758, "grad_norm": 0.5826766490936279, "learning_rate": 9.828704321262833e-06, "loss": 0.5196, "step": 1643 }, { "epoch": 0.5260238907849829, "grad_norm": 0.6424474716186523, "learning_rate": 9.828220835210136e-06, "loss": 0.5954, "step": 1644 }, { "epoch": 0.5263438566552902, "grad_norm": 0.5636090040206909, "learning_rate": 9.827736679718813e-06, "loss": 0.5072, "step": 1645 }, { "epoch": 0.5266638225255973, "grad_norm": 0.5572488307952881, "learning_rate": 9.827251854855992e-06, "loss": 0.5685, "step": 1646 }, { "epoch": 0.5269837883959044, "grad_norm": 0.6091232895851135, "learning_rate": 9.826766360688893e-06, "loss": 0.5586, "step": 1647 }, { "epoch": 0.5273037542662116, "grad_norm": 0.5716792941093445, "learning_rate": 9.82628019728483e-06, "loss": 0.547, "step": 1648 }, { "epoch": 0.5276237201365188, "grad_norm": 0.5463572144508362, "learning_rate": 9.825793364711214e-06, "loss": 0.5681, "step": 1649 }, { "epoch": 0.527943686006826, "grad_norm": 0.6867672204971313, "learning_rate": 9.825305863035542e-06, "loss": 0.6118, "step": 1650 }, { "epoch": 0.5282636518771331, "grad_norm": 0.608964204788208, "learning_rate": 9.824817692325406e-06, "loss": 0.5524, "step": 1651 }, { "epoch": 0.5285836177474402, "grad_norm": 0.6046398282051086, "learning_rate": 9.824328852648493e-06, "loss": 0.5688, "step": 1652 }, { "epoch": 0.5289035836177475, "grad_norm": 0.5403018593788147, "learning_rate": 9.823839344072582e-06, "loss": 0.5252, "step": 1653 }, { "epoch": 0.5292235494880546, "grad_norm": 0.5631974935531616, "learning_rate": 9.82334916666554e-06, "loss": 0.5623, "step": 1654 }, { "epoch": 0.5295435153583617, "grad_norm": 0.6233628392219543, "learning_rate": 9.822858320495335e-06, "loss": 0.5735, "step": 1655 }, { "epoch": 0.5298634812286689, "grad_norm": 0.5567300915718079, "learning_rate": 9.822366805630022e-06, "loss": 0.534, "step": 1656 }, { "epoch": 0.5301834470989761, "grad_norm": 0.6214247345924377, "learning_rate": 9.82187462213775e-06, "loss": 0.5767, "step": 1657 }, { "epoch": 0.5305034129692833, "grad_norm": 0.6399384140968323, "learning_rate": 9.82138177008676e-06, "loss": 0.5787, "step": 1658 }, { "epoch": 0.5308233788395904, "grad_norm": 0.584986686706543, "learning_rate": 9.820888249545388e-06, "loss": 0.5501, "step": 1659 }, { "epoch": 0.5311433447098977, "grad_norm": 0.5978248715400696, "learning_rate": 9.82039406058206e-06, "loss": 0.5829, "step": 1660 }, { "epoch": 0.5314633105802048, "grad_norm": 0.6715003252029419, "learning_rate": 9.819899203265298e-06, "loss": 0.5932, "step": 1661 }, { "epoch": 0.5317832764505119, "grad_norm": 0.622571587562561, "learning_rate": 9.81940367766371e-06, "loss": 0.6027, "step": 1662 }, { "epoch": 0.5321032423208191, "grad_norm": 0.646557092666626, "learning_rate": 9.818907483846007e-06, "loss": 0.5819, "step": 1663 }, { "epoch": 0.5324232081911263, "grad_norm": 0.6905471086502075, "learning_rate": 9.818410621880983e-06, "loss": 0.609, "step": 1664 }, { "epoch": 0.5327431740614335, "grad_norm": 0.6914578676223755, "learning_rate": 9.81791309183753e-06, "loss": 0.5885, "step": 1665 }, { "epoch": 0.5330631399317406, "grad_norm": 0.5626682043075562, "learning_rate": 9.81741489378463e-06, "loss": 0.5536, "step": 1666 }, { "epoch": 0.5333831058020477, "grad_norm": 0.663912296295166, "learning_rate": 9.816916027791359e-06, "loss": 0.5744, "step": 1667 }, { "epoch": 0.533703071672355, "grad_norm": 0.6111547946929932, "learning_rate": 9.816416493926887e-06, "loss": 0.5534, "step": 1668 }, { "epoch": 0.5340230375426621, "grad_norm": 0.5859062671661377, "learning_rate": 9.815916292260474e-06, "loss": 0.572, "step": 1669 }, { "epoch": 0.5343430034129693, "grad_norm": 0.5649756789207458, "learning_rate": 9.815415422861472e-06, "loss": 0.5531, "step": 1670 }, { "epoch": 0.5346629692832765, "grad_norm": 0.5535812973976135, "learning_rate": 9.81491388579933e-06, "loss": 0.5256, "step": 1671 }, { "epoch": 0.5349829351535836, "grad_norm": 0.6786409020423889, "learning_rate": 9.814411681143585e-06, "loss": 0.5855, "step": 1672 }, { "epoch": 0.5353029010238908, "grad_norm": 0.5840369462966919, "learning_rate": 9.813908808963868e-06, "loss": 0.5619, "step": 1673 }, { "epoch": 0.5356228668941979, "grad_norm": 0.5695275664329529, "learning_rate": 9.813405269329903e-06, "loss": 0.5961, "step": 1674 }, { "epoch": 0.5359428327645052, "grad_norm": 0.607681155204773, "learning_rate": 9.812901062311508e-06, "loss": 0.5416, "step": 1675 }, { "epoch": 0.5362627986348123, "grad_norm": 0.5815897583961487, "learning_rate": 9.81239618797859e-06, "loss": 0.5771, "step": 1676 }, { "epoch": 0.5365827645051194, "grad_norm": 0.544284999370575, "learning_rate": 9.81189064640115e-06, "loss": 0.5203, "step": 1677 }, { "epoch": 0.5369027303754266, "grad_norm": 0.5429071187973022, "learning_rate": 9.811384437649285e-06, "loss": 0.5808, "step": 1678 }, { "epoch": 0.5372226962457338, "grad_norm": 0.5531368851661682, "learning_rate": 9.810877561793178e-06, "loss": 0.5617, "step": 1679 }, { "epoch": 0.537542662116041, "grad_norm": 0.7014932036399841, "learning_rate": 9.81037001890311e-06, "loss": 0.5843, "step": 1680 }, { "epoch": 0.5378626279863481, "grad_norm": 0.6051618456840515, "learning_rate": 9.809861809049453e-06, "loss": 0.56, "step": 1681 }, { "epoch": 0.5381825938566553, "grad_norm": 0.5951033234596252, "learning_rate": 9.809352932302666e-06, "loss": 0.5714, "step": 1682 }, { "epoch": 0.5385025597269625, "grad_norm": 0.5959492325782776, "learning_rate": 9.808843388733314e-06, "loss": 0.553, "step": 1683 }, { "epoch": 0.5388225255972696, "grad_norm": 0.5903566479682922, "learning_rate": 9.808333178412038e-06, "loss": 0.5638, "step": 1684 }, { "epoch": 0.5391424914675768, "grad_norm": 0.5624989867210388, "learning_rate": 9.807822301409583e-06, "loss": 0.5699, "step": 1685 }, { "epoch": 0.539462457337884, "grad_norm": 0.6225293278694153, "learning_rate": 9.807310757796782e-06, "loss": 0.5659, "step": 1686 }, { "epoch": 0.5397824232081911, "grad_norm": 0.6644200682640076, "learning_rate": 9.80679854764456e-06, "loss": 0.5929, "step": 1687 }, { "epoch": 0.5401023890784983, "grad_norm": 0.6093822121620178, "learning_rate": 9.806285671023938e-06, "loss": 0.5704, "step": 1688 }, { "epoch": 0.5404223549488054, "grad_norm": 0.6620171666145325, "learning_rate": 9.805772128006027e-06, "loss": 0.5991, "step": 1689 }, { "epoch": 0.5407423208191127, "grad_norm": 0.6101434230804443, "learning_rate": 9.805257918662026e-06, "loss": 0.5798, "step": 1690 }, { "epoch": 0.5410622866894198, "grad_norm": 0.5370322465896606, "learning_rate": 9.804743043063235e-06, "loss": 0.5582, "step": 1691 }, { "epoch": 0.5413822525597269, "grad_norm": 0.7034540176391602, "learning_rate": 9.804227501281041e-06, "loss": 0.6284, "step": 1692 }, { "epoch": 0.5417022184300341, "grad_norm": 0.6034532785415649, "learning_rate": 9.803711293386925e-06, "loss": 0.5597, "step": 1693 }, { "epoch": 0.5420221843003413, "grad_norm": 0.6423254013061523, "learning_rate": 9.803194419452457e-06, "loss": 0.5648, "step": 1694 }, { "epoch": 0.5423421501706485, "grad_norm": 0.6289602518081665, "learning_rate": 9.802676879549307e-06, "loss": 0.5663, "step": 1695 }, { "epoch": 0.5426621160409556, "grad_norm": 0.5871692895889282, "learning_rate": 9.802158673749226e-06, "loss": 0.5469, "step": 1696 }, { "epoch": 0.5429820819112628, "grad_norm": 0.6468052864074707, "learning_rate": 9.801639802124073e-06, "loss": 0.5633, "step": 1697 }, { "epoch": 0.54330204778157, "grad_norm": 0.6076865196228027, "learning_rate": 9.80112026474578e-06, "loss": 0.5436, "step": 1698 }, { "epoch": 0.5436220136518771, "grad_norm": 0.6306731700897217, "learning_rate": 9.80060006168639e-06, "loss": 0.5826, "step": 1699 }, { "epoch": 0.5439419795221843, "grad_norm": 0.5897573828697205, "learning_rate": 9.800079193018025e-06, "loss": 0.5587, "step": 1700 }, { "epoch": 0.5442619453924915, "grad_norm": 0.6413190364837646, "learning_rate": 9.799557658812903e-06, "loss": 0.5394, "step": 1701 }, { "epoch": 0.5445819112627986, "grad_norm": 0.5940037965774536, "learning_rate": 9.79903545914334e-06, "loss": 0.5601, "step": 1702 }, { "epoch": 0.5449018771331058, "grad_norm": 0.6278110146522522, "learning_rate": 9.798512594081737e-06, "loss": 0.5293, "step": 1703 }, { "epoch": 0.5452218430034129, "grad_norm": 0.7648538947105408, "learning_rate": 9.797989063700591e-06, "loss": 0.5833, "step": 1704 }, { "epoch": 0.5455418088737202, "grad_norm": 0.5689882040023804, "learning_rate": 9.797464868072489e-06, "loss": 0.5842, "step": 1705 }, { "epoch": 0.5458617747440273, "grad_norm": 0.5978536605834961, "learning_rate": 9.79694000727011e-06, "loss": 0.609, "step": 1706 }, { "epoch": 0.5461817406143344, "grad_norm": 0.7378367185592651, "learning_rate": 9.796414481366227e-06, "loss": 0.5615, "step": 1707 }, { "epoch": 0.5465017064846417, "grad_norm": 0.7020463347434998, "learning_rate": 9.795888290433709e-06, "loss": 0.6027, "step": 1708 }, { "epoch": 0.5468216723549488, "grad_norm": 0.5986933708190918, "learning_rate": 9.795361434545508e-06, "loss": 0.5398, "step": 1709 }, { "epoch": 0.547141638225256, "grad_norm": 0.7171380519866943, "learning_rate": 9.794833913774678e-06, "loss": 0.5354, "step": 1710 }, { "epoch": 0.5474616040955631, "grad_norm": 0.647392749786377, "learning_rate": 9.794305728194353e-06, "loss": 0.5626, "step": 1711 }, { "epoch": 0.5477815699658704, "grad_norm": 0.670254647731781, "learning_rate": 9.793776877877774e-06, "loss": 0.578, "step": 1712 }, { "epoch": 0.5481015358361775, "grad_norm": 0.6468953490257263, "learning_rate": 9.793247362898264e-06, "loss": 0.5564, "step": 1713 }, { "epoch": 0.5484215017064846, "grad_norm": 0.6417179107666016, "learning_rate": 9.79271718332924e-06, "loss": 0.5667, "step": 1714 }, { "epoch": 0.5487414675767918, "grad_norm": 0.603004515171051, "learning_rate": 9.792186339244211e-06, "loss": 0.5482, "step": 1715 }, { "epoch": 0.549061433447099, "grad_norm": 0.6157732009887695, "learning_rate": 9.791654830716784e-06, "loss": 0.5758, "step": 1716 }, { "epoch": 0.5493813993174061, "grad_norm": 0.6211690902709961, "learning_rate": 9.79112265782065e-06, "loss": 0.5542, "step": 1717 }, { "epoch": 0.5497013651877133, "grad_norm": 0.6685905456542969, "learning_rate": 9.790589820629594e-06, "loss": 0.5799, "step": 1718 }, { "epoch": 0.5500213310580204, "grad_norm": 0.6300346255302429, "learning_rate": 9.790056319217496e-06, "loss": 0.6207, "step": 1719 }, { "epoch": 0.5503412969283277, "grad_norm": 0.6636313796043396, "learning_rate": 9.789522153658328e-06, "loss": 0.6008, "step": 1720 }, { "epoch": 0.5506612627986348, "grad_norm": 0.6509947180747986, "learning_rate": 9.78898732402615e-06, "loss": 0.5185, "step": 1721 }, { "epoch": 0.550981228668942, "grad_norm": 0.553341269493103, "learning_rate": 9.78845183039512e-06, "loss": 0.5621, "step": 1722 }, { "epoch": 0.5513011945392492, "grad_norm": 0.5741779804229736, "learning_rate": 9.78791567283948e-06, "loss": 0.5271, "step": 1723 }, { "epoch": 0.5516211604095563, "grad_norm": 0.5845010876655579, "learning_rate": 9.787378851433575e-06, "loss": 0.5661, "step": 1724 }, { "epoch": 0.5519411262798635, "grad_norm": 0.5319338440895081, "learning_rate": 9.786841366251832e-06, "loss": 0.5589, "step": 1725 }, { "epoch": 0.5522610921501706, "grad_norm": 0.6268966197967529, "learning_rate": 9.786303217368776e-06, "loss": 0.5764, "step": 1726 }, { "epoch": 0.5525810580204779, "grad_norm": 0.6318864226341248, "learning_rate": 9.78576440485902e-06, "loss": 0.6013, "step": 1727 }, { "epoch": 0.552901023890785, "grad_norm": 0.5269508957862854, "learning_rate": 9.78522492879727e-06, "loss": 0.5649, "step": 1728 }, { "epoch": 0.5532209897610921, "grad_norm": 0.5737921595573425, "learning_rate": 9.78468478925833e-06, "loss": 0.5322, "step": 1729 }, { "epoch": 0.5535409556313993, "grad_norm": 0.6844146847724915, "learning_rate": 9.784143986317084e-06, "loss": 0.5964, "step": 1730 }, { "epoch": 0.5538609215017065, "grad_norm": 0.5287778377532959, "learning_rate": 9.783602520048524e-06, "loss": 0.5648, "step": 1731 }, { "epoch": 0.5541808873720137, "grad_norm": 0.599004328250885, "learning_rate": 9.783060390527715e-06, "loss": 0.5581, "step": 1732 }, { "epoch": 0.5545008532423208, "grad_norm": 0.6111088395118713, "learning_rate": 9.78251759782983e-06, "loss": 0.5774, "step": 1733 }, { "epoch": 0.554820819112628, "grad_norm": 0.6105920672416687, "learning_rate": 9.781974142030128e-06, "loss": 0.5592, "step": 1734 }, { "epoch": 0.5551407849829352, "grad_norm": 0.583831250667572, "learning_rate": 9.781430023203957e-06, "loss": 0.5793, "step": 1735 }, { "epoch": 0.5554607508532423, "grad_norm": 0.5499464273452759, "learning_rate": 9.780885241426762e-06, "loss": 0.5533, "step": 1736 }, { "epoch": 0.5557807167235495, "grad_norm": 0.6046625971794128, "learning_rate": 9.780339796774076e-06, "loss": 0.5736, "step": 1737 }, { "epoch": 0.5561006825938567, "grad_norm": 0.5838026404380798, "learning_rate": 9.779793689321528e-06, "loss": 0.561, "step": 1738 }, { "epoch": 0.5564206484641638, "grad_norm": 0.5161740183830261, "learning_rate": 9.779246919144833e-06, "loss": 0.5488, "step": 1739 }, { "epoch": 0.556740614334471, "grad_norm": 0.5640830993652344, "learning_rate": 9.778699486319803e-06, "loss": 0.5456, "step": 1740 }, { "epoch": 0.5570605802047781, "grad_norm": 0.6633821725845337, "learning_rate": 9.77815139092234e-06, "loss": 0.5319, "step": 1741 }, { "epoch": 0.5573805460750854, "grad_norm": 0.5820072889328003, "learning_rate": 9.77760263302844e-06, "loss": 0.5577, "step": 1742 }, { "epoch": 0.5577005119453925, "grad_norm": 0.6018913388252258, "learning_rate": 9.777053212714188e-06, "loss": 0.5492, "step": 1743 }, { "epoch": 0.5580204778156996, "grad_norm": 0.6947749853134155, "learning_rate": 9.776503130055758e-06, "loss": 0.5873, "step": 1744 }, { "epoch": 0.5583404436860068, "grad_norm": 0.6588101387023926, "learning_rate": 9.775952385129423e-06, "loss": 0.5815, "step": 1745 }, { "epoch": 0.558660409556314, "grad_norm": 0.5936509370803833, "learning_rate": 9.775400978011547e-06, "loss": 0.6015, "step": 1746 }, { "epoch": 0.5589803754266212, "grad_norm": 0.6646673679351807, "learning_rate": 9.774848908778577e-06, "loss": 0.5585, "step": 1747 }, { "epoch": 0.5593003412969283, "grad_norm": 0.5602323412895203, "learning_rate": 9.774296177507065e-06, "loss": 0.5442, "step": 1748 }, { "epoch": 0.5596203071672355, "grad_norm": 0.5378860235214233, "learning_rate": 9.773742784273643e-06, "loss": 0.5255, "step": 1749 }, { "epoch": 0.5599402730375427, "grad_norm": 0.6054778099060059, "learning_rate": 9.773188729155042e-06, "loss": 0.5897, "step": 1750 }, { "epoch": 0.5602602389078498, "grad_norm": 0.5988175272941589, "learning_rate": 9.77263401222808e-06, "loss": 0.6037, "step": 1751 }, { "epoch": 0.560580204778157, "grad_norm": 0.6248955726623535, "learning_rate": 9.772078633569671e-06, "loss": 0.5725, "step": 1752 }, { "epoch": 0.5609001706484642, "grad_norm": 0.6076432466506958, "learning_rate": 9.771522593256817e-06, "loss": 0.563, "step": 1753 }, { "epoch": 0.5612201365187713, "grad_norm": 0.612916111946106, "learning_rate": 9.770965891366617e-06, "loss": 0.596, "step": 1754 }, { "epoch": 0.5615401023890785, "grad_norm": 0.598166286945343, "learning_rate": 9.770408527976257e-06, "loss": 0.5782, "step": 1755 }, { "epoch": 0.5618600682593856, "grad_norm": 0.5937659740447998, "learning_rate": 9.769850503163015e-06, "loss": 0.5321, "step": 1756 }, { "epoch": 0.5621800341296929, "grad_norm": 0.5992322564125061, "learning_rate": 9.76929181700426e-06, "loss": 0.5703, "step": 1757 }, { "epoch": 0.5625, "grad_norm": 0.6490613222122192, "learning_rate": 9.76873246957746e-06, "loss": 0.5482, "step": 1758 }, { "epoch": 0.5628199658703071, "grad_norm": 0.642821192741394, "learning_rate": 9.768172460960166e-06, "loss": 0.5629, "step": 1759 }, { "epoch": 0.5631399317406144, "grad_norm": 0.6676160097122192, "learning_rate": 9.767611791230023e-06, "loss": 0.5194, "step": 1760 }, { "epoch": 0.5634598976109215, "grad_norm": 0.6234229207038879, "learning_rate": 9.767050460464769e-06, "loss": 0.5573, "step": 1761 }, { "epoch": 0.5637798634812287, "grad_norm": 0.6466017365455627, "learning_rate": 9.766488468742235e-06, "loss": 0.5885, "step": 1762 }, { "epoch": 0.5640998293515358, "grad_norm": 0.559739887714386, "learning_rate": 9.76592581614034e-06, "loss": 0.5735, "step": 1763 }, { "epoch": 0.564419795221843, "grad_norm": 0.6247774958610535, "learning_rate": 9.765362502737098e-06, "loss": 0.5771, "step": 1764 }, { "epoch": 0.5647397610921502, "grad_norm": 0.5745558738708496, "learning_rate": 9.764798528610612e-06, "loss": 0.5715, "step": 1765 }, { "epoch": 0.5650597269624573, "grad_norm": 0.5948444604873657, "learning_rate": 9.764233893839077e-06, "loss": 0.5802, "step": 1766 }, { "epoch": 0.5653796928327645, "grad_norm": 0.5603572726249695, "learning_rate": 9.763668598500782e-06, "loss": 0.5397, "step": 1767 }, { "epoch": 0.5656996587030717, "grad_norm": 0.5986836552619934, "learning_rate": 9.763102642674104e-06, "loss": 0.5411, "step": 1768 }, { "epoch": 0.5660196245733788, "grad_norm": 0.6480838060379028, "learning_rate": 9.762536026437515e-06, "loss": 0.5741, "step": 1769 }, { "epoch": 0.566339590443686, "grad_norm": 0.6260193586349487, "learning_rate": 9.761968749869576e-06, "loss": 0.5964, "step": 1770 }, { "epoch": 0.5666595563139932, "grad_norm": 0.6947192549705505, "learning_rate": 9.76140081304894e-06, "loss": 0.5766, "step": 1771 }, { "epoch": 0.5669795221843004, "grad_norm": 0.5749015808105469, "learning_rate": 9.760832216054356e-06, "loss": 0.5477, "step": 1772 }, { "epoch": 0.5672994880546075, "grad_norm": 0.6033229231834412, "learning_rate": 9.760262958964656e-06, "loss": 0.5639, "step": 1773 }, { "epoch": 0.5676194539249146, "grad_norm": 0.6450707912445068, "learning_rate": 9.75969304185877e-06, "loss": 0.5523, "step": 1774 }, { "epoch": 0.5679394197952219, "grad_norm": 0.6025317907333374, "learning_rate": 9.759122464815718e-06, "loss": 0.5907, "step": 1775 }, { "epoch": 0.568259385665529, "grad_norm": 0.7295170426368713, "learning_rate": 9.75855122791461e-06, "loss": 0.5803, "step": 1776 }, { "epoch": 0.5685793515358362, "grad_norm": 0.613014280796051, "learning_rate": 9.757979331234651e-06, "loss": 0.5558, "step": 1777 }, { "epoch": 0.5688993174061433, "grad_norm": 0.5929551720619202, "learning_rate": 9.757406774855134e-06, "loss": 0.5863, "step": 1778 }, { "epoch": 0.5692192832764505, "grad_norm": 0.6426234841346741, "learning_rate": 9.756833558855443e-06, "loss": 0.5739, "step": 1779 }, { "epoch": 0.5695392491467577, "grad_norm": 0.719906747341156, "learning_rate": 9.756259683315056e-06, "loss": 0.5883, "step": 1780 }, { "epoch": 0.5698592150170648, "grad_norm": 0.6483733057975769, "learning_rate": 9.755685148313544e-06, "loss": 0.6078, "step": 1781 }, { "epoch": 0.570179180887372, "grad_norm": 0.6298717260360718, "learning_rate": 9.755109953930564e-06, "loss": 0.5331, "step": 1782 }, { "epoch": 0.5704991467576792, "grad_norm": 0.7005184888839722, "learning_rate": 9.754534100245867e-06, "loss": 0.5949, "step": 1783 }, { "epoch": 0.5708191126279863, "grad_norm": 0.5551364421844482, "learning_rate": 9.7539575873393e-06, "loss": 0.5433, "step": 1784 }, { "epoch": 0.5711390784982935, "grad_norm": 0.6557787656784058, "learning_rate": 9.753380415290793e-06, "loss": 0.5892, "step": 1785 }, { "epoch": 0.5714590443686007, "grad_norm": 0.718512237071991, "learning_rate": 9.752802584180372e-06, "loss": 0.6151, "step": 1786 }, { "epoch": 0.5717790102389079, "grad_norm": 0.5826222896575928, "learning_rate": 9.752224094088156e-06, "loss": 0.5744, "step": 1787 }, { "epoch": 0.572098976109215, "grad_norm": 0.695049524307251, "learning_rate": 9.751644945094352e-06, "loss": 0.5994, "step": 1788 }, { "epoch": 0.5724189419795221, "grad_norm": 0.6085717082023621, "learning_rate": 9.751065137279261e-06, "loss": 0.5471, "step": 1789 }, { "epoch": 0.5727389078498294, "grad_norm": 0.5604434013366699, "learning_rate": 9.750484670723272e-06, "loss": 0.5563, "step": 1790 }, { "epoch": 0.5730588737201365, "grad_norm": 0.6504515409469604, "learning_rate": 9.749903545506868e-06, "loss": 0.5488, "step": 1791 }, { "epoch": 0.5733788395904437, "grad_norm": 0.7121344208717346, "learning_rate": 9.749321761710625e-06, "loss": 0.5697, "step": 1792 }, { "epoch": 0.5736988054607508, "grad_norm": 0.6262866258621216, "learning_rate": 9.748739319415203e-06, "loss": 0.5913, "step": 1793 }, { "epoch": 0.574018771331058, "grad_norm": 0.6223553419113159, "learning_rate": 9.748156218701363e-06, "loss": 0.5731, "step": 1794 }, { "epoch": 0.5743387372013652, "grad_norm": 0.693714439868927, "learning_rate": 9.747572459649952e-06, "loss": 0.5876, "step": 1795 }, { "epoch": 0.5746587030716723, "grad_norm": 0.5756598114967346, "learning_rate": 9.746988042341907e-06, "loss": 0.5376, "step": 1796 }, { "epoch": 0.5749786689419796, "grad_norm": 0.6898940801620483, "learning_rate": 9.746402966858259e-06, "loss": 0.5538, "step": 1797 }, { "epoch": 0.5752986348122867, "grad_norm": 0.62108314037323, "learning_rate": 9.74581723328013e-06, "loss": 0.5493, "step": 1798 }, { "epoch": 0.5756186006825939, "grad_norm": 0.6687491536140442, "learning_rate": 9.745230841688733e-06, "loss": 0.5688, "step": 1799 }, { "epoch": 0.575938566552901, "grad_norm": 0.6858484148979187, "learning_rate": 9.74464379216537e-06, "loss": 0.5609, "step": 1800 }, { "epoch": 0.5762585324232082, "grad_norm": 0.6215513944625854, "learning_rate": 9.744056084791439e-06, "loss": 0.5785, "step": 1801 }, { "epoch": 0.5765784982935154, "grad_norm": 0.6347957849502563, "learning_rate": 9.743467719648423e-06, "loss": 0.5399, "step": 1802 }, { "epoch": 0.5768984641638225, "grad_norm": 0.6881147623062134, "learning_rate": 9.742878696817901e-06, "loss": 0.5333, "step": 1803 }, { "epoch": 0.5772184300341296, "grad_norm": 0.7048043608665466, "learning_rate": 9.742289016381542e-06, "loss": 0.5787, "step": 1804 }, { "epoch": 0.5775383959044369, "grad_norm": 0.6560038924217224, "learning_rate": 9.741698678421108e-06, "loss": 0.5611, "step": 1805 }, { "epoch": 0.577858361774744, "grad_norm": 0.6992453932762146, "learning_rate": 9.741107683018444e-06, "loss": 0.5972, "step": 1806 }, { "epoch": 0.5781783276450512, "grad_norm": 0.6837847232818604, "learning_rate": 9.7405160302555e-06, "loss": 0.563, "step": 1807 }, { "epoch": 0.5784982935153583, "grad_norm": 0.8289478421211243, "learning_rate": 9.739923720214305e-06, "loss": 0.5804, "step": 1808 }, { "epoch": 0.5788182593856656, "grad_norm": 0.666740357875824, "learning_rate": 9.739330752976981e-06, "loss": 0.5798, "step": 1809 }, { "epoch": 0.5791382252559727, "grad_norm": 0.6751397252082825, "learning_rate": 9.738737128625748e-06, "loss": 0.5375, "step": 1810 }, { "epoch": 0.5794581911262798, "grad_norm": 0.6747352480888367, "learning_rate": 9.73814284724291e-06, "loss": 0.5687, "step": 1811 }, { "epoch": 0.5797781569965871, "grad_norm": 0.6946969628334045, "learning_rate": 9.737547908910867e-06, "loss": 0.5897, "step": 1812 }, { "epoch": 0.5800981228668942, "grad_norm": 0.6373754739761353, "learning_rate": 9.736952313712108e-06, "loss": 0.5496, "step": 1813 }, { "epoch": 0.5804180887372014, "grad_norm": 0.6169047355651855, "learning_rate": 9.736356061729212e-06, "loss": 0.575, "step": 1814 }, { "epoch": 0.5807380546075085, "grad_norm": 0.6824852228164673, "learning_rate": 9.735759153044849e-06, "loss": 0.5506, "step": 1815 }, { "epoch": 0.5810580204778157, "grad_norm": 0.5575690865516663, "learning_rate": 9.735161587741782e-06, "loss": 0.5717, "step": 1816 }, { "epoch": 0.5813779863481229, "grad_norm": 0.5755314230918884, "learning_rate": 9.734563365902865e-06, "loss": 0.5444, "step": 1817 }, { "epoch": 0.58169795221843, "grad_norm": 0.6323144435882568, "learning_rate": 9.733964487611044e-06, "loss": 0.5629, "step": 1818 }, { "epoch": 0.5820179180887372, "grad_norm": 0.5643534064292908, "learning_rate": 9.733364952949348e-06, "loss": 0.5611, "step": 1819 }, { "epoch": 0.5823378839590444, "grad_norm": 0.6021159887313843, "learning_rate": 9.732764762000908e-06, "loss": 0.5562, "step": 1820 }, { "epoch": 0.5826578498293515, "grad_norm": 0.5586191415786743, "learning_rate": 9.73216391484894e-06, "loss": 0.5187, "step": 1821 }, { "epoch": 0.5829778156996587, "grad_norm": 0.6042790412902832, "learning_rate": 9.731562411576751e-06, "loss": 0.5652, "step": 1822 }, { "epoch": 0.5832977815699659, "grad_norm": 0.5898236036300659, "learning_rate": 9.730960252267744e-06, "loss": 0.5713, "step": 1823 }, { "epoch": 0.5836177474402731, "grad_norm": 0.5529911518096924, "learning_rate": 9.730357437005404e-06, "loss": 0.5775, "step": 1824 }, { "epoch": 0.5839377133105802, "grad_norm": 0.5990542769432068, "learning_rate": 9.729753965873316e-06, "loss": 0.6122, "step": 1825 }, { "epoch": 0.5842576791808873, "grad_norm": 0.6193751692771912, "learning_rate": 9.729149838955152e-06, "loss": 0.5542, "step": 1826 }, { "epoch": 0.5845776450511946, "grad_norm": 0.5816980004310608, "learning_rate": 9.728545056334672e-06, "loss": 0.5646, "step": 1827 }, { "epoch": 0.5848976109215017, "grad_norm": 0.6045491099357605, "learning_rate": 9.72793961809573e-06, "loss": 0.5797, "step": 1828 }, { "epoch": 0.5852175767918089, "grad_norm": 0.6036240458488464, "learning_rate": 9.727333524322274e-06, "loss": 0.5432, "step": 1829 }, { "epoch": 0.585537542662116, "grad_norm": 0.516349732875824, "learning_rate": 9.726726775098338e-06, "loss": 0.5456, "step": 1830 }, { "epoch": 0.5858575085324232, "grad_norm": 0.655970573425293, "learning_rate": 9.726119370508048e-06, "loss": 0.5891, "step": 1831 }, { "epoch": 0.5861774744027304, "grad_norm": 0.5936410427093506, "learning_rate": 9.72551131063562e-06, "loss": 0.5418, "step": 1832 }, { "epoch": 0.5864974402730375, "grad_norm": 0.573919951915741, "learning_rate": 9.724902595565367e-06, "loss": 0.5551, "step": 1833 }, { "epoch": 0.5868174061433447, "grad_norm": 0.593414843082428, "learning_rate": 9.724293225381681e-06, "loss": 0.5491, "step": 1834 }, { "epoch": 0.5871373720136519, "grad_norm": 0.6796587109565735, "learning_rate": 9.723683200169059e-06, "loss": 0.564, "step": 1835 }, { "epoch": 0.587457337883959, "grad_norm": 0.7001802921295166, "learning_rate": 9.723072520012077e-06, "loss": 0.5585, "step": 1836 }, { "epoch": 0.5877773037542662, "grad_norm": 0.5853090882301331, "learning_rate": 9.722461184995408e-06, "loss": 0.5905, "step": 1837 }, { "epoch": 0.5880972696245734, "grad_norm": 0.6028174161911011, "learning_rate": 9.721849195203814e-06, "loss": 0.5431, "step": 1838 }, { "epoch": 0.5884172354948806, "grad_norm": 0.6703921556472778, "learning_rate": 9.721236550722152e-06, "loss": 0.5828, "step": 1839 }, { "epoch": 0.5887372013651877, "grad_norm": 0.5544477105140686, "learning_rate": 9.720623251635358e-06, "loss": 0.5307, "step": 1840 }, { "epoch": 0.5890571672354948, "grad_norm": 0.5607783794403076, "learning_rate": 9.720009298028473e-06, "loss": 0.5603, "step": 1841 }, { "epoch": 0.5893771331058021, "grad_norm": 0.5793670415878296, "learning_rate": 9.71939468998662e-06, "loss": 0.5483, "step": 1842 }, { "epoch": 0.5896970989761092, "grad_norm": 0.6787139773368835, "learning_rate": 9.718779427595017e-06, "loss": 0.5509, "step": 1843 }, { "epoch": 0.5900170648464164, "grad_norm": 0.613737165927887, "learning_rate": 9.71816351093897e-06, "loss": 0.5673, "step": 1844 }, { "epoch": 0.5903370307167235, "grad_norm": 0.5790629982948303, "learning_rate": 9.717546940103874e-06, "loss": 0.5594, "step": 1845 }, { "epoch": 0.5906569965870307, "grad_norm": 0.5192556381225586, "learning_rate": 9.71692971517522e-06, "loss": 0.5669, "step": 1846 }, { "epoch": 0.5909769624573379, "grad_norm": 0.5302321314811707, "learning_rate": 9.716311836238585e-06, "loss": 0.5579, "step": 1847 }, { "epoch": 0.591296928327645, "grad_norm": 0.5794715285301208, "learning_rate": 9.715693303379643e-06, "loss": 0.5409, "step": 1848 }, { "epoch": 0.5916168941979523, "grad_norm": 0.6204904913902283, "learning_rate": 9.71507411668415e-06, "loss": 0.6097, "step": 1849 }, { "epoch": 0.5919368600682594, "grad_norm": 0.6074978709220886, "learning_rate": 9.714454276237958e-06, "loss": 0.5656, "step": 1850 }, { "epoch": 0.5922568259385665, "grad_norm": 0.6031474471092224, "learning_rate": 9.713833782127008e-06, "loss": 0.5845, "step": 1851 }, { "epoch": 0.5925767918088737, "grad_norm": 0.5594843626022339, "learning_rate": 9.713212634437337e-06, "loss": 0.5373, "step": 1852 }, { "epoch": 0.5928967576791809, "grad_norm": 0.5694454908370972, "learning_rate": 9.71259083325506e-06, "loss": 0.5959, "step": 1853 }, { "epoch": 0.5932167235494881, "grad_norm": 0.6202699542045593, "learning_rate": 9.711968378666396e-06, "loss": 0.5807, "step": 1854 }, { "epoch": 0.5935366894197952, "grad_norm": 0.5497620701789856, "learning_rate": 9.711345270757649e-06, "loss": 0.597, "step": 1855 }, { "epoch": 0.5938566552901023, "grad_norm": 0.597712516784668, "learning_rate": 9.71072150961521e-06, "loss": 0.5641, "step": 1856 }, { "epoch": 0.5941766211604096, "grad_norm": 0.5572855472564697, "learning_rate": 9.71009709532557e-06, "loss": 0.5706, "step": 1857 }, { "epoch": 0.5944965870307167, "grad_norm": 0.614568829536438, "learning_rate": 9.709472027975299e-06, "loss": 0.5732, "step": 1858 }, { "epoch": 0.5948165529010239, "grad_norm": 0.5438926815986633, "learning_rate": 9.708846307651066e-06, "loss": 0.5323, "step": 1859 }, { "epoch": 0.5951365187713311, "grad_norm": 0.5313812494277954, "learning_rate": 9.708219934439627e-06, "loss": 0.5536, "step": 1860 }, { "epoch": 0.5954564846416383, "grad_norm": 0.61684250831604, "learning_rate": 9.70759290842783e-06, "loss": 0.5556, "step": 1861 }, { "epoch": 0.5957764505119454, "grad_norm": 0.5726536512374878, "learning_rate": 9.706965229702614e-06, "loss": 0.5442, "step": 1862 }, { "epoch": 0.5960964163822525, "grad_norm": 0.6879222393035889, "learning_rate": 9.706336898351005e-06, "loss": 0.5752, "step": 1863 }, { "epoch": 0.5964163822525598, "grad_norm": 0.5403534770011902, "learning_rate": 9.705707914460122e-06, "loss": 0.5306, "step": 1864 }, { "epoch": 0.5967363481228669, "grad_norm": 0.5859735608100891, "learning_rate": 9.70507827811718e-06, "loss": 0.5654, "step": 1865 }, { "epoch": 0.597056313993174, "grad_norm": 0.6655783653259277, "learning_rate": 9.704447989409472e-06, "loss": 0.5517, "step": 1866 }, { "epoch": 0.5973762798634812, "grad_norm": 0.7582557201385498, "learning_rate": 9.703817048424389e-06, "loss": 0.6227, "step": 1867 }, { "epoch": 0.5976962457337884, "grad_norm": 0.5447322726249695, "learning_rate": 9.703185455249416e-06, "loss": 0.5584, "step": 1868 }, { "epoch": 0.5980162116040956, "grad_norm": 0.6286244988441467, "learning_rate": 9.70255320997212e-06, "loss": 0.5837, "step": 1869 }, { "epoch": 0.5983361774744027, "grad_norm": 0.6168583631515503, "learning_rate": 9.701920312680163e-06, "loss": 0.5357, "step": 1870 }, { "epoch": 0.5986561433447098, "grad_norm": 0.5949482321739197, "learning_rate": 9.7012867634613e-06, "loss": 0.5357, "step": 1871 }, { "epoch": 0.5989761092150171, "grad_norm": 0.5717710256576538, "learning_rate": 9.70065256240337e-06, "loss": 0.5891, "step": 1872 }, { "epoch": 0.5992960750853242, "grad_norm": 0.6137530207633972, "learning_rate": 9.70001770959431e-06, "loss": 0.5506, "step": 1873 }, { "epoch": 0.5996160409556314, "grad_norm": 0.623816192150116, "learning_rate": 9.699382205122138e-06, "loss": 0.569, "step": 1874 }, { "epoch": 0.5999360068259386, "grad_norm": 0.5570641160011292, "learning_rate": 9.69874604907497e-06, "loss": 0.5678, "step": 1875 }, { "epoch": 0.6002559726962458, "grad_norm": 0.5365951657295227, "learning_rate": 9.698109241541009e-06, "loss": 0.5651, "step": 1876 }, { "epoch": 0.6005759385665529, "grad_norm": 0.5629282593727112, "learning_rate": 9.697471782608552e-06, "loss": 0.5368, "step": 1877 }, { "epoch": 0.60089590443686, "grad_norm": 0.5601006150245667, "learning_rate": 9.69683367236598e-06, "loss": 0.5681, "step": 1878 }, { "epoch": 0.6012158703071673, "grad_norm": 0.5880118012428284, "learning_rate": 9.696194910901766e-06, "loss": 0.5722, "step": 1879 }, { "epoch": 0.6015358361774744, "grad_norm": 0.5651369690895081, "learning_rate": 9.695555498304481e-06, "loss": 0.5514, "step": 1880 }, { "epoch": 0.6018558020477816, "grad_norm": 0.614946722984314, "learning_rate": 9.694915434662777e-06, "loss": 0.506, "step": 1881 }, { "epoch": 0.6021757679180887, "grad_norm": 0.5546241402626038, "learning_rate": 9.6942747200654e-06, "loss": 0.5185, "step": 1882 }, { "epoch": 0.6024957337883959, "grad_norm": 0.6413306593894958, "learning_rate": 9.693633354601187e-06, "loss": 0.6012, "step": 1883 }, { "epoch": 0.6028156996587031, "grad_norm": 0.5133163332939148, "learning_rate": 9.69299133835906e-06, "loss": 0.5371, "step": 1884 }, { "epoch": 0.6031356655290102, "grad_norm": 0.5547313094139099, "learning_rate": 9.692348671428038e-06, "loss": 0.5699, "step": 1885 }, { "epoch": 0.6034556313993175, "grad_norm": 0.5818104147911072, "learning_rate": 9.691705353897229e-06, "loss": 0.5507, "step": 1886 }, { "epoch": 0.6037755972696246, "grad_norm": 0.5582051873207092, "learning_rate": 9.69106138585583e-06, "loss": 0.5771, "step": 1887 }, { "epoch": 0.6040955631399317, "grad_norm": 0.66030353307724, "learning_rate": 9.690416767393124e-06, "loss": 0.5983, "step": 1888 }, { "epoch": 0.6044155290102389, "grad_norm": 0.5691258907318115, "learning_rate": 9.689771498598491e-06, "loss": 0.5335, "step": 1889 }, { "epoch": 0.6047354948805461, "grad_norm": 0.663033664226532, "learning_rate": 9.689125579561398e-06, "loss": 0.5597, "step": 1890 }, { "epoch": 0.6050554607508533, "grad_norm": 0.6218370795249939, "learning_rate": 9.688479010371403e-06, "loss": 0.5757, "step": 1891 }, { "epoch": 0.6053754266211604, "grad_norm": 0.6317211389541626, "learning_rate": 9.687831791118151e-06, "loss": 0.6012, "step": 1892 }, { "epoch": 0.6056953924914675, "grad_norm": 0.5636645555496216, "learning_rate": 9.687183921891383e-06, "loss": 0.5671, "step": 1893 }, { "epoch": 0.6060153583617748, "grad_norm": 0.5526756644248962, "learning_rate": 9.686535402780926e-06, "loss": 0.5263, "step": 1894 }, { "epoch": 0.6063353242320819, "grad_norm": 0.5666709542274475, "learning_rate": 9.685886233876696e-06, "loss": 0.5286, "step": 1895 }, { "epoch": 0.606655290102389, "grad_norm": 0.579018235206604, "learning_rate": 9.685236415268702e-06, "loss": 0.5574, "step": 1896 }, { "epoch": 0.6069752559726962, "grad_norm": 0.5683245658874512, "learning_rate": 9.684585947047045e-06, "loss": 0.582, "step": 1897 }, { "epoch": 0.6072952218430034, "grad_norm": 0.6050066351890564, "learning_rate": 9.683934829301909e-06, "loss": 0.5789, "step": 1898 }, { "epoch": 0.6076151877133106, "grad_norm": 0.5764254927635193, "learning_rate": 9.683283062123574e-06, "loss": 0.572, "step": 1899 }, { "epoch": 0.6079351535836177, "grad_norm": 0.4990369975566864, "learning_rate": 9.682630645602409e-06, "loss": 0.5779, "step": 1900 }, { "epoch": 0.608255119453925, "grad_norm": 0.5741258859634399, "learning_rate": 9.681977579828869e-06, "loss": 0.5449, "step": 1901 }, { "epoch": 0.6085750853242321, "grad_norm": 0.570561945438385, "learning_rate": 9.681323864893509e-06, "loss": 0.5505, "step": 1902 }, { "epoch": 0.6088950511945392, "grad_norm": 0.6162533164024353, "learning_rate": 9.68066950088696e-06, "loss": 0.5623, "step": 1903 }, { "epoch": 0.6092150170648464, "grad_norm": 0.6171571016311646, "learning_rate": 9.680014487899957e-06, "loss": 0.5874, "step": 1904 }, { "epoch": 0.6095349829351536, "grad_norm": 0.5293908715248108, "learning_rate": 9.679358826023312e-06, "loss": 0.5515, "step": 1905 }, { "epoch": 0.6098549488054608, "grad_norm": 0.5890594720840454, "learning_rate": 9.678702515347937e-06, "loss": 0.57, "step": 1906 }, { "epoch": 0.6101749146757679, "grad_norm": 0.5866312384605408, "learning_rate": 9.678045555964831e-06, "loss": 0.575, "step": 1907 }, { "epoch": 0.610494880546075, "grad_norm": 0.6178193092346191, "learning_rate": 9.67738794796508e-06, "loss": 0.5976, "step": 1908 }, { "epoch": 0.6108148464163823, "grad_norm": 0.594685971736908, "learning_rate": 9.676729691439864e-06, "loss": 0.5369, "step": 1909 }, { "epoch": 0.6111348122866894, "grad_norm": 0.6227867603302002, "learning_rate": 9.676070786480447e-06, "loss": 0.5444, "step": 1910 }, { "epoch": 0.6114547781569966, "grad_norm": 0.6000764966011047, "learning_rate": 9.675411233178193e-06, "loss": 0.53, "step": 1911 }, { "epoch": 0.6117747440273038, "grad_norm": 0.6180450916290283, "learning_rate": 9.674751031624546e-06, "loss": 0.5318, "step": 1912 }, { "epoch": 0.612094709897611, "grad_norm": 0.572875440120697, "learning_rate": 9.674090181911044e-06, "loss": 0.5393, "step": 1913 }, { "epoch": 0.6124146757679181, "grad_norm": 0.6435962915420532, "learning_rate": 9.673428684129317e-06, "loss": 0.5432, "step": 1914 }, { "epoch": 0.6127346416382252, "grad_norm": 0.6471882462501526, "learning_rate": 9.672766538371079e-06, "loss": 0.5597, "step": 1915 }, { "epoch": 0.6130546075085325, "grad_norm": 0.621203601360321, "learning_rate": 9.67210374472814e-06, "loss": 0.5464, "step": 1916 }, { "epoch": 0.6133745733788396, "grad_norm": 0.5741633176803589, "learning_rate": 9.671440303292395e-06, "loss": 0.5824, "step": 1917 }, { "epoch": 0.6136945392491467, "grad_norm": 0.559482991695404, "learning_rate": 9.670776214155833e-06, "loss": 0.5802, "step": 1918 }, { "epoch": 0.6140145051194539, "grad_norm": 0.5610204935073853, "learning_rate": 9.670111477410528e-06, "loss": 0.5608, "step": 1919 }, { "epoch": 0.6143344709897611, "grad_norm": 0.6152501106262207, "learning_rate": 9.66944609314865e-06, "loss": 0.5532, "step": 1920 }, { "epoch": 0.6146544368600683, "grad_norm": 0.599856436252594, "learning_rate": 9.668780061462455e-06, "loss": 0.5923, "step": 1921 }, { "epoch": 0.6149744027303754, "grad_norm": 0.661961555480957, "learning_rate": 9.668113382444287e-06, "loss": 0.5971, "step": 1922 }, { "epoch": 0.6152943686006825, "grad_norm": 0.605215847492218, "learning_rate": 9.66744605618658e-06, "loss": 0.5498, "step": 1923 }, { "epoch": 0.6156143344709898, "grad_norm": 0.5865672826766968, "learning_rate": 9.666778082781865e-06, "loss": 0.5347, "step": 1924 }, { "epoch": 0.6159343003412969, "grad_norm": 0.6111371517181396, "learning_rate": 9.666109462322753e-06, "loss": 0.59, "step": 1925 }, { "epoch": 0.6162542662116041, "grad_norm": 0.5758267641067505, "learning_rate": 9.665440194901951e-06, "loss": 0.5529, "step": 1926 }, { "epoch": 0.6165742320819113, "grad_norm": 0.6464254260063171, "learning_rate": 9.664770280612255e-06, "loss": 0.5979, "step": 1927 }, { "epoch": 0.6168941979522184, "grad_norm": 0.6565837264060974, "learning_rate": 9.664099719546547e-06, "loss": 0.5606, "step": 1928 }, { "epoch": 0.6172141638225256, "grad_norm": 0.5908191800117493, "learning_rate": 9.663428511797801e-06, "loss": 0.5834, "step": 1929 }, { "epoch": 0.6175341296928327, "grad_norm": 0.5503283143043518, "learning_rate": 9.66275665745908e-06, "loss": 0.5443, "step": 1930 }, { "epoch": 0.61785409556314, "grad_norm": 0.6072626113891602, "learning_rate": 9.66208415662354e-06, "loss": 0.5551, "step": 1931 }, { "epoch": 0.6181740614334471, "grad_norm": 0.5384520292282104, "learning_rate": 9.661411009384426e-06, "loss": 0.5487, "step": 1932 }, { "epoch": 0.6184940273037542, "grad_norm": 0.5442841649055481, "learning_rate": 9.660737215835066e-06, "loss": 0.5345, "step": 1933 }, { "epoch": 0.6188139931740614, "grad_norm": 0.5890656113624573, "learning_rate": 9.660062776068883e-06, "loss": 0.5655, "step": 1934 }, { "epoch": 0.6191339590443686, "grad_norm": 0.5688875317573547, "learning_rate": 9.659387690179392e-06, "loss": 0.5339, "step": 1935 }, { "epoch": 0.6194539249146758, "grad_norm": 0.5539190769195557, "learning_rate": 9.658711958260189e-06, "loss": 0.5863, "step": 1936 }, { "epoch": 0.6197738907849829, "grad_norm": 0.6129565834999084, "learning_rate": 9.658035580404972e-06, "loss": 0.5742, "step": 1937 }, { "epoch": 0.6200938566552902, "grad_norm": 0.5914084911346436, "learning_rate": 9.657358556707516e-06, "loss": 0.5388, "step": 1938 }, { "epoch": 0.6204138225255973, "grad_norm": 0.5551041960716248, "learning_rate": 9.656680887261693e-06, "loss": 0.5513, "step": 1939 }, { "epoch": 0.6207337883959044, "grad_norm": 0.6423928141593933, "learning_rate": 9.656002572161463e-06, "loss": 0.5914, "step": 1940 }, { "epoch": 0.6210537542662116, "grad_norm": 0.585381031036377, "learning_rate": 9.655323611500876e-06, "loss": 0.5434, "step": 1941 }, { "epoch": 0.6213737201365188, "grad_norm": 0.6105824708938599, "learning_rate": 9.65464400537407e-06, "loss": 0.5868, "step": 1942 }, { "epoch": 0.621693686006826, "grad_norm": 0.5563498139381409, "learning_rate": 9.65396375387527e-06, "loss": 0.5621, "step": 1943 }, { "epoch": 0.6220136518771331, "grad_norm": 0.606756329536438, "learning_rate": 9.6532828570988e-06, "loss": 0.545, "step": 1944 }, { "epoch": 0.6223336177474402, "grad_norm": 0.4880354106426239, "learning_rate": 9.652601315139063e-06, "loss": 0.5487, "step": 1945 }, { "epoch": 0.6226535836177475, "grad_norm": 0.527872622013092, "learning_rate": 9.651919128090557e-06, "loss": 0.5838, "step": 1946 }, { "epoch": 0.6229735494880546, "grad_norm": 0.5351196527481079, "learning_rate": 9.651236296047866e-06, "loss": 0.549, "step": 1947 }, { "epoch": 0.6232935153583617, "grad_norm": 0.6507715582847595, "learning_rate": 9.650552819105667e-06, "loss": 0.591, "step": 1948 }, { "epoch": 0.6236134812286689, "grad_norm": 0.6204143762588501, "learning_rate": 9.649868697358727e-06, "loss": 0.5482, "step": 1949 }, { "epoch": 0.6239334470989761, "grad_norm": 0.568878173828125, "learning_rate": 9.649183930901895e-06, "loss": 0.5811, "step": 1950 }, { "epoch": 0.6242534129692833, "grad_norm": 0.6354591846466064, "learning_rate": 9.64849851983012e-06, "loss": 0.5746, "step": 1951 }, { "epoch": 0.6245733788395904, "grad_norm": 0.6112775802612305, "learning_rate": 9.647812464238434e-06, "loss": 0.5806, "step": 1952 }, { "epoch": 0.6248933447098977, "grad_norm": 0.5902199745178223, "learning_rate": 9.647125764221956e-06, "loss": 0.5939, "step": 1953 }, { "epoch": 0.6252133105802048, "grad_norm": 0.544347882270813, "learning_rate": 9.646438419875901e-06, "loss": 0.585, "step": 1954 }, { "epoch": 0.6255332764505119, "grad_norm": 0.5880687832832336, "learning_rate": 9.645750431295568e-06, "loss": 0.5594, "step": 1955 }, { "epoch": 0.6258532423208191, "grad_norm": 0.711134135723114, "learning_rate": 9.645061798576352e-06, "loss": 0.5548, "step": 1956 }, { "epoch": 0.6261732081911263, "grad_norm": 0.5626425743103027, "learning_rate": 9.644372521813728e-06, "loss": 0.5494, "step": 1957 }, { "epoch": 0.6264931740614335, "grad_norm": 0.692373514175415, "learning_rate": 9.643682601103265e-06, "loss": 0.6095, "step": 1958 }, { "epoch": 0.6268131399317406, "grad_norm": 0.5923463702201843, "learning_rate": 9.642992036540624e-06, "loss": 0.5349, "step": 1959 }, { "epoch": 0.6271331058020477, "grad_norm": 0.6308784484863281, "learning_rate": 9.64230082822155e-06, "loss": 0.5608, "step": 1960 }, { "epoch": 0.627453071672355, "grad_norm": 0.6047257781028748, "learning_rate": 9.641608976241883e-06, "loss": 0.5927, "step": 1961 }, { "epoch": 0.6277730375426621, "grad_norm": 0.7537582516670227, "learning_rate": 9.640916480697549e-06, "loss": 0.5708, "step": 1962 }, { "epoch": 0.6280930034129693, "grad_norm": 0.6602718234062195, "learning_rate": 9.640223341684559e-06, "loss": 0.5749, "step": 1963 }, { "epoch": 0.6284129692832765, "grad_norm": 0.5211908221244812, "learning_rate": 9.639529559299022e-06, "loss": 0.5495, "step": 1964 }, { "epoch": 0.6287329351535836, "grad_norm": 0.6491508483886719, "learning_rate": 9.638835133637129e-06, "loss": 0.5618, "step": 1965 }, { "epoch": 0.6290529010238908, "grad_norm": 0.6506140828132629, "learning_rate": 9.638140064795165e-06, "loss": 0.5724, "step": 1966 }, { "epoch": 0.6293728668941979, "grad_norm": 0.6765151619911194, "learning_rate": 9.6374443528695e-06, "loss": 0.6029, "step": 1967 }, { "epoch": 0.6296928327645052, "grad_norm": 0.5773376822471619, "learning_rate": 9.636747997956597e-06, "loss": 0.5702, "step": 1968 }, { "epoch": 0.6300127986348123, "grad_norm": 0.6763083338737488, "learning_rate": 9.636051000153006e-06, "loss": 0.5499, "step": 1969 }, { "epoch": 0.6303327645051194, "grad_norm": 0.6451863050460815, "learning_rate": 9.635353359555366e-06, "loss": 0.5408, "step": 1970 }, { "epoch": 0.6306527303754266, "grad_norm": 0.6031057834625244, "learning_rate": 9.634655076260409e-06, "loss": 0.5545, "step": 1971 }, { "epoch": 0.6309726962457338, "grad_norm": 0.6576459407806396, "learning_rate": 9.633956150364948e-06, "loss": 0.5346, "step": 1972 }, { "epoch": 0.631292662116041, "grad_norm": 0.6280539035797119, "learning_rate": 9.63325658196589e-06, "loss": 0.5397, "step": 1973 }, { "epoch": 0.6316126279863481, "grad_norm": 0.5733145475387573, "learning_rate": 9.632556371160235e-06, "loss": 0.58, "step": 1974 }, { "epoch": 0.6319325938566553, "grad_norm": 0.642741858959198, "learning_rate": 9.631855518045066e-06, "loss": 0.5692, "step": 1975 }, { "epoch": 0.6322525597269625, "grad_norm": 0.6114027500152588, "learning_rate": 9.631154022717559e-06, "loss": 0.5559, "step": 1976 }, { "epoch": 0.6325725255972696, "grad_norm": 0.525458812713623, "learning_rate": 9.630451885274972e-06, "loss": 0.5216, "step": 1977 }, { "epoch": 0.6328924914675768, "grad_norm": 0.6430639624595642, "learning_rate": 9.629749105814664e-06, "loss": 0.5731, "step": 1978 }, { "epoch": 0.633212457337884, "grad_norm": 0.6785262823104858, "learning_rate": 9.629045684434069e-06, "loss": 0.5897, "step": 1979 }, { "epoch": 0.6335324232081911, "grad_norm": 0.5230071544647217, "learning_rate": 9.628341621230724e-06, "loss": 0.5355, "step": 1980 }, { "epoch": 0.6338523890784983, "grad_norm": 0.6526293754577637, "learning_rate": 9.627636916302244e-06, "loss": 0.5579, "step": 1981 }, { "epoch": 0.6341723549488054, "grad_norm": 0.5032544136047363, "learning_rate": 9.626931569746337e-06, "loss": 0.5699, "step": 1982 }, { "epoch": 0.6344923208191127, "grad_norm": 0.5933758616447449, "learning_rate": 9.626225581660802e-06, "loss": 0.531, "step": 1983 }, { "epoch": 0.6348122866894198, "grad_norm": 0.692962110042572, "learning_rate": 9.625518952143526e-06, "loss": 0.5833, "step": 1984 }, { "epoch": 0.6351322525597269, "grad_norm": 0.6170587539672852, "learning_rate": 9.624811681292482e-06, "loss": 0.5408, "step": 1985 }, { "epoch": 0.6354522184300341, "grad_norm": 0.589077889919281, "learning_rate": 9.624103769205735e-06, "loss": 0.5583, "step": 1986 }, { "epoch": 0.6357721843003413, "grad_norm": 0.5993129014968872, "learning_rate": 9.623395215981438e-06, "loss": 0.558, "step": 1987 }, { "epoch": 0.6360921501706485, "grad_norm": 0.5902277231216431, "learning_rate": 9.62268602171783e-06, "loss": 0.5258, "step": 1988 }, { "epoch": 0.6364121160409556, "grad_norm": 0.6470785140991211, "learning_rate": 9.621976186513246e-06, "loss": 0.5756, "step": 1989 }, { "epoch": 0.6367320819112628, "grad_norm": 0.6402467489242554, "learning_rate": 9.621265710466103e-06, "loss": 0.549, "step": 1990 }, { "epoch": 0.63705204778157, "grad_norm": 0.6703201532363892, "learning_rate": 9.62055459367491e-06, "loss": 0.5687, "step": 1991 }, { "epoch": 0.6373720136518771, "grad_norm": 0.5826023817062378, "learning_rate": 9.619842836238262e-06, "loss": 0.576, "step": 1992 }, { "epoch": 0.6376919795221843, "grad_norm": 0.5884040594100952, "learning_rate": 9.619130438254848e-06, "loss": 0.5591, "step": 1993 }, { "epoch": 0.6380119453924915, "grad_norm": 0.612905740737915, "learning_rate": 9.618417399823442e-06, "loss": 0.5536, "step": 1994 }, { "epoch": 0.6383319112627986, "grad_norm": 0.6277275681495667, "learning_rate": 9.617703721042908e-06, "loss": 0.5701, "step": 1995 }, { "epoch": 0.6386518771331058, "grad_norm": 0.5961500406265259, "learning_rate": 9.616989402012197e-06, "loss": 0.5349, "step": 1996 }, { "epoch": 0.6389718430034129, "grad_norm": 0.7005389928817749, "learning_rate": 9.616274442830351e-06, "loss": 0.5559, "step": 1997 }, { "epoch": 0.6392918088737202, "grad_norm": 0.5778456926345825, "learning_rate": 9.615558843596499e-06, "loss": 0.5772, "step": 1998 }, { "epoch": 0.6396117747440273, "grad_norm": 0.6484526991844177, "learning_rate": 9.614842604409863e-06, "loss": 0.5679, "step": 1999 }, { "epoch": 0.6399317406143344, "grad_norm": 0.6132345795631409, "learning_rate": 9.614125725369748e-06, "loss": 0.6166, "step": 2000 }, { "epoch": 0.6402517064846417, "grad_norm": 0.6483284831047058, "learning_rate": 9.61340820657555e-06, "loss": 0.5865, "step": 2001 }, { "epoch": 0.6405716723549488, "grad_norm": 0.6628165245056152, "learning_rate": 9.612690048126751e-06, "loss": 0.5806, "step": 2002 }, { "epoch": 0.640891638225256, "grad_norm": 0.651597261428833, "learning_rate": 9.611971250122932e-06, "loss": 0.5479, "step": 2003 }, { "epoch": 0.6412116040955631, "grad_norm": 0.6118643283843994, "learning_rate": 9.611251812663748e-06, "loss": 0.5947, "step": 2004 }, { "epoch": 0.6415315699658704, "grad_norm": 0.6847550272941589, "learning_rate": 9.610531735848953e-06, "loss": 0.5963, "step": 2005 }, { "epoch": 0.6418515358361775, "grad_norm": 0.6172077059745789, "learning_rate": 9.609811019778388e-06, "loss": 0.5258, "step": 2006 }, { "epoch": 0.6421715017064846, "grad_norm": 0.7246551513671875, "learning_rate": 9.609089664551976e-06, "loss": 0.5895, "step": 2007 }, { "epoch": 0.6424914675767918, "grad_norm": 0.5504890084266663, "learning_rate": 9.60836767026974e-06, "loss": 0.5393, "step": 2008 }, { "epoch": 0.642811433447099, "grad_norm": 0.6666256189346313, "learning_rate": 9.607645037031783e-06, "loss": 0.5494, "step": 2009 }, { "epoch": 0.6431313993174061, "grad_norm": 0.6688473224639893, "learning_rate": 9.606921764938297e-06, "loss": 0.5617, "step": 2010 }, { "epoch": 0.6434513651877133, "grad_norm": 0.6392967104911804, "learning_rate": 9.606197854089565e-06, "loss": 0.5448, "step": 2011 }, { "epoch": 0.6437713310580204, "grad_norm": 0.5535218715667725, "learning_rate": 9.60547330458596e-06, "loss": 0.5338, "step": 2012 }, { "epoch": 0.6440912969283277, "grad_norm": 0.6675567626953125, "learning_rate": 9.604748116527941e-06, "loss": 0.5717, "step": 2013 }, { "epoch": 0.6444112627986348, "grad_norm": 0.7083868384361267, "learning_rate": 9.604022290016058e-06, "loss": 0.5345, "step": 2014 }, { "epoch": 0.644731228668942, "grad_norm": 0.6258317232131958, "learning_rate": 9.603295825150945e-06, "loss": 0.5257, "step": 2015 }, { "epoch": 0.6450511945392492, "grad_norm": 0.6635270118713379, "learning_rate": 9.602568722033325e-06, "loss": 0.5718, "step": 2016 }, { "epoch": 0.6453711604095563, "grad_norm": 0.5791119933128357, "learning_rate": 9.601840980764016e-06, "loss": 0.5769, "step": 2017 }, { "epoch": 0.6456911262798635, "grad_norm": 0.5586481094360352, "learning_rate": 9.601112601443921e-06, "loss": 0.5455, "step": 2018 }, { "epoch": 0.6460110921501706, "grad_norm": 0.6536003947257996, "learning_rate": 9.600383584174026e-06, "loss": 0.5536, "step": 2019 }, { "epoch": 0.6463310580204779, "grad_norm": 0.5298084616661072, "learning_rate": 9.599653929055415e-06, "loss": 0.541, "step": 2020 }, { "epoch": 0.646651023890785, "grad_norm": 0.5338242650032043, "learning_rate": 9.598923636189252e-06, "loss": 0.5341, "step": 2021 }, { "epoch": 0.6469709897610921, "grad_norm": 0.5623722076416016, "learning_rate": 9.598192705676793e-06, "loss": 0.5432, "step": 2022 }, { "epoch": 0.6472909556313993, "grad_norm": 0.6378763914108276, "learning_rate": 9.597461137619385e-06, "loss": 0.5895, "step": 2023 }, { "epoch": 0.6476109215017065, "grad_norm": 0.5924339294433594, "learning_rate": 9.59672893211846e-06, "loss": 0.5071, "step": 2024 }, { "epoch": 0.6479308873720137, "grad_norm": 0.5660730600357056, "learning_rate": 9.595996089275536e-06, "loss": 0.5536, "step": 2025 }, { "epoch": 0.6482508532423208, "grad_norm": 0.6002275347709656, "learning_rate": 9.595262609192227e-06, "loss": 0.5031, "step": 2026 }, { "epoch": 0.648570819112628, "grad_norm": 0.6197472214698792, "learning_rate": 9.594528491970228e-06, "loss": 0.5654, "step": 2027 }, { "epoch": 0.6488907849829352, "grad_norm": 0.5491067171096802, "learning_rate": 9.593793737711328e-06, "loss": 0.5654, "step": 2028 }, { "epoch": 0.6492107508532423, "grad_norm": 0.6247221827507019, "learning_rate": 9.593058346517399e-06, "loss": 0.559, "step": 2029 }, { "epoch": 0.6495307167235495, "grad_norm": 0.636074960231781, "learning_rate": 9.592322318490404e-06, "loss": 0.5868, "step": 2030 }, { "epoch": 0.6498506825938567, "grad_norm": 0.5542947053909302, "learning_rate": 9.591585653732396e-06, "loss": 0.5166, "step": 2031 }, { "epoch": 0.6501706484641638, "grad_norm": 0.5761454105377197, "learning_rate": 9.590848352345512e-06, "loss": 0.5599, "step": 2032 }, { "epoch": 0.650490614334471, "grad_norm": 0.6053789258003235, "learning_rate": 9.590110414431981e-06, "loss": 0.5886, "step": 2033 }, { "epoch": 0.6508105802047781, "grad_norm": 0.5515697598457336, "learning_rate": 9.58937184009412e-06, "loss": 0.5624, "step": 2034 }, { "epoch": 0.6511305460750854, "grad_norm": 0.6260466575622559, "learning_rate": 9.588632629434331e-06, "loss": 0.5709, "step": 2035 }, { "epoch": 0.6514505119453925, "grad_norm": 0.5678718090057373, "learning_rate": 9.587892782555109e-06, "loss": 0.5663, "step": 2036 }, { "epoch": 0.6517704778156996, "grad_norm": 0.6218145489692688, "learning_rate": 9.587152299559033e-06, "loss": 0.5648, "step": 2037 }, { "epoch": 0.6520904436860068, "grad_norm": 0.5651326179504395, "learning_rate": 9.586411180548771e-06, "loss": 0.5434, "step": 2038 }, { "epoch": 0.652410409556314, "grad_norm": 0.5573369264602661, "learning_rate": 9.585669425627084e-06, "loss": 0.5597, "step": 2039 }, { "epoch": 0.6527303754266212, "grad_norm": 0.6214823126792908, "learning_rate": 9.584927034896811e-06, "loss": 0.5441, "step": 2040 }, { "epoch": 0.6530503412969283, "grad_norm": 0.6007993221282959, "learning_rate": 9.58418400846089e-06, "loss": 0.5621, "step": 2041 }, { "epoch": 0.6533703071672355, "grad_norm": 0.5891932249069214, "learning_rate": 9.583440346422343e-06, "loss": 0.5827, "step": 2042 }, { "epoch": 0.6536902730375427, "grad_norm": 0.6093569993972778, "learning_rate": 9.582696048884277e-06, "loss": 0.5996, "step": 2043 }, { "epoch": 0.6540102389078498, "grad_norm": 0.616129457950592, "learning_rate": 9.58195111594989e-06, "loss": 0.5865, "step": 2044 }, { "epoch": 0.654330204778157, "grad_norm": 0.6123115420341492, "learning_rate": 9.581205547722469e-06, "loss": 0.5576, "step": 2045 }, { "epoch": 0.6546501706484642, "grad_norm": 0.6001349687576294, "learning_rate": 9.580459344305388e-06, "loss": 0.5784, "step": 2046 }, { "epoch": 0.6549701365187713, "grad_norm": 0.5365661382675171, "learning_rate": 9.579712505802108e-06, "loss": 0.5567, "step": 2047 }, { "epoch": 0.6552901023890785, "grad_norm": 0.5674573183059692, "learning_rate": 9.578965032316179e-06, "loss": 0.4916, "step": 2048 }, { "epoch": 0.6556100682593856, "grad_norm": 0.6029785871505737, "learning_rate": 9.57821692395124e-06, "loss": 0.5384, "step": 2049 }, { "epoch": 0.6559300341296929, "grad_norm": 0.6047776341438293, "learning_rate": 9.577468180811018e-06, "loss": 0.6, "step": 2050 }, { "epoch": 0.65625, "grad_norm": 0.6041454076766968, "learning_rate": 9.576718802999323e-06, "loss": 0.5408, "step": 2051 }, { "epoch": 0.6565699658703071, "grad_norm": 0.6480724215507507, "learning_rate": 9.575968790620062e-06, "loss": 0.6052, "step": 2052 }, { "epoch": 0.6568899317406144, "grad_norm": 0.6151446104049683, "learning_rate": 9.575218143777223e-06, "loss": 0.5528, "step": 2053 }, { "epoch": 0.6572098976109215, "grad_norm": 0.608056366443634, "learning_rate": 9.574466862574885e-06, "loss": 0.5174, "step": 2054 }, { "epoch": 0.6575298634812287, "grad_norm": 0.6194968819618225, "learning_rate": 9.573714947117211e-06, "loss": 0.5555, "step": 2055 }, { "epoch": 0.6578498293515358, "grad_norm": 0.5183257460594177, "learning_rate": 9.57296239750846e-06, "loss": 0.5358, "step": 2056 }, { "epoch": 0.658169795221843, "grad_norm": 0.6629782915115356, "learning_rate": 9.57220921385297e-06, "loss": 0.6058, "step": 2057 }, { "epoch": 0.6584897610921502, "grad_norm": 0.6532469987869263, "learning_rate": 9.571455396255172e-06, "loss": 0.6021, "step": 2058 }, { "epoch": 0.6588097269624573, "grad_norm": 0.6884270906448364, "learning_rate": 9.570700944819584e-06, "loss": 0.5639, "step": 2059 }, { "epoch": 0.6591296928327645, "grad_norm": 0.5459734797477722, "learning_rate": 9.56994585965081e-06, "loss": 0.5185, "step": 2060 }, { "epoch": 0.6594496587030717, "grad_norm": 0.5665500164031982, "learning_rate": 9.569190140853547e-06, "loss": 0.5767, "step": 2061 }, { "epoch": 0.6597696245733788, "grad_norm": 0.5893674492835999, "learning_rate": 9.568433788532571e-06, "loss": 0.5901, "step": 2062 }, { "epoch": 0.660089590443686, "grad_norm": 0.6321619749069214, "learning_rate": 9.567676802792758e-06, "loss": 0.556, "step": 2063 }, { "epoch": 0.6604095563139932, "grad_norm": 0.6531881093978882, "learning_rate": 9.566919183739058e-06, "loss": 0.5742, "step": 2064 }, { "epoch": 0.6607295221843004, "grad_norm": 0.7623351216316223, "learning_rate": 9.566160931476522e-06, "loss": 0.5853, "step": 2065 }, { "epoch": 0.6610494880546075, "grad_norm": 0.5670693516731262, "learning_rate": 9.565402046110277e-06, "loss": 0.5645, "step": 2066 }, { "epoch": 0.6613694539249146, "grad_norm": 0.5728259086608887, "learning_rate": 9.564642527745546e-06, "loss": 0.5466, "step": 2067 }, { "epoch": 0.6616894197952219, "grad_norm": 0.625257134437561, "learning_rate": 9.563882376487638e-06, "loss": 0.5377, "step": 2068 }, { "epoch": 0.662009385665529, "grad_norm": 0.6368067264556885, "learning_rate": 9.563121592441949e-06, "loss": 0.6021, "step": 2069 }, { "epoch": 0.6623293515358362, "grad_norm": 0.6265316605567932, "learning_rate": 9.562360175713962e-06, "loss": 0.5594, "step": 2070 }, { "epoch": 0.6626493174061433, "grad_norm": 0.550437867641449, "learning_rate": 9.561598126409247e-06, "loss": 0.5333, "step": 2071 }, { "epoch": 0.6629692832764505, "grad_norm": 0.5962300300598145, "learning_rate": 9.560835444633463e-06, "loss": 0.5759, "step": 2072 }, { "epoch": 0.6632892491467577, "grad_norm": 0.5362560153007507, "learning_rate": 9.56007213049236e-06, "loss": 0.5436, "step": 2073 }, { "epoch": 0.6636092150170648, "grad_norm": 0.5743743181228638, "learning_rate": 9.559308184091768e-06, "loss": 0.5559, "step": 2074 }, { "epoch": 0.663929180887372, "grad_norm": 0.5720234513282776, "learning_rate": 9.55854360553761e-06, "loss": 0.5366, "step": 2075 }, { "epoch": 0.6642491467576792, "grad_norm": 0.6186221837997437, "learning_rate": 9.557778394935901e-06, "loss": 0.6093, "step": 2076 }, { "epoch": 0.6645691126279863, "grad_norm": 0.5654708743095398, "learning_rate": 9.557012552392733e-06, "loss": 0.5436, "step": 2077 }, { "epoch": 0.6648890784982935, "grad_norm": 0.6198699474334717, "learning_rate": 9.55624607801429e-06, "loss": 0.5633, "step": 2078 }, { "epoch": 0.6652090443686007, "grad_norm": 0.5168883800506592, "learning_rate": 9.55547897190685e-06, "loss": 0.5425, "step": 2079 }, { "epoch": 0.6655290102389079, "grad_norm": 0.5414618849754333, "learning_rate": 9.55471123417677e-06, "loss": 0.54, "step": 2080 }, { "epoch": 0.665848976109215, "grad_norm": 0.5594053268432617, "learning_rate": 9.553942864930496e-06, "loss": 0.5526, "step": 2081 }, { "epoch": 0.6661689419795221, "grad_norm": 0.5626882910728455, "learning_rate": 9.553173864274567e-06, "loss": 0.5354, "step": 2082 }, { "epoch": 0.6664889078498294, "grad_norm": 0.5668383240699768, "learning_rate": 9.552404232315602e-06, "loss": 0.542, "step": 2083 }, { "epoch": 0.6668088737201365, "grad_norm": 0.548861563205719, "learning_rate": 9.551633969160316e-06, "loss": 0.5673, "step": 2084 }, { "epoch": 0.6671288395904437, "grad_norm": 0.5764604806900024, "learning_rate": 9.550863074915505e-06, "loss": 0.5761, "step": 2085 }, { "epoch": 0.6674488054607508, "grad_norm": 0.5341320633888245, "learning_rate": 9.550091549688049e-06, "loss": 0.533, "step": 2086 }, { "epoch": 0.667768771331058, "grad_norm": 0.5438635349273682, "learning_rate": 9.54931939358493e-06, "loss": 0.5435, "step": 2087 }, { "epoch": 0.6680887372013652, "grad_norm": 0.5967687368392944, "learning_rate": 9.548546606713203e-06, "loss": 0.5621, "step": 2088 }, { "epoch": 0.6684087030716723, "grad_norm": 0.5669074654579163, "learning_rate": 9.547773189180018e-06, "loss": 0.5309, "step": 2089 }, { "epoch": 0.6687286689419796, "grad_norm": 0.5866972208023071, "learning_rate": 9.546999141092608e-06, "loss": 0.5896, "step": 2090 }, { "epoch": 0.6690486348122867, "grad_norm": 0.5700693130493164, "learning_rate": 9.546224462558297e-06, "loss": 0.5485, "step": 2091 }, { "epoch": 0.6693686006825939, "grad_norm": 0.566959798336029, "learning_rate": 9.545449153684495e-06, "loss": 0.5671, "step": 2092 }, { "epoch": 0.669688566552901, "grad_norm": 0.5802577137947083, "learning_rate": 9.544673214578699e-06, "loss": 0.5132, "step": 2093 }, { "epoch": 0.6700085324232082, "grad_norm": 0.6248685121536255, "learning_rate": 9.543896645348496e-06, "loss": 0.5444, "step": 2094 }, { "epoch": 0.6703284982935154, "grad_norm": 0.5355051159858704, "learning_rate": 9.543119446101556e-06, "loss": 0.5325, "step": 2095 }, { "epoch": 0.6706484641638225, "grad_norm": 0.5653332471847534, "learning_rate": 9.542341616945638e-06, "loss": 0.583, "step": 2096 }, { "epoch": 0.6709684300341296, "grad_norm": 0.6538986563682556, "learning_rate": 9.541563157988591e-06, "loss": 0.5806, "step": 2097 }, { "epoch": 0.6712883959044369, "grad_norm": 0.563230037689209, "learning_rate": 9.540784069338349e-06, "loss": 0.571, "step": 2098 }, { "epoch": 0.671608361774744, "grad_norm": 0.6563031077384949, "learning_rate": 9.540004351102933e-06, "loss": 0.568, "step": 2099 }, { "epoch": 0.6719283276450512, "grad_norm": 0.6009393334388733, "learning_rate": 9.539224003390453e-06, "loss": 0.5772, "step": 2100 }, { "epoch": 0.6722482935153583, "grad_norm": 0.5868406891822815, "learning_rate": 9.538443026309102e-06, "loss": 0.5695, "step": 2101 }, { "epoch": 0.6725682593856656, "grad_norm": 0.5810580253601074, "learning_rate": 9.537661419967167e-06, "loss": 0.5536, "step": 2102 }, { "epoch": 0.6728882252559727, "grad_norm": 0.576774537563324, "learning_rate": 9.536879184473016e-06, "loss": 0.5695, "step": 2103 }, { "epoch": 0.6732081911262798, "grad_norm": 0.5885899066925049, "learning_rate": 9.536096319935109e-06, "loss": 0.5558, "step": 2104 }, { "epoch": 0.6735281569965871, "grad_norm": 0.548465371131897, "learning_rate": 9.53531282646199e-06, "loss": 0.558, "step": 2105 }, { "epoch": 0.6738481228668942, "grad_norm": 0.6229990124702454, "learning_rate": 9.53452870416229e-06, "loss": 0.5522, "step": 2106 }, { "epoch": 0.6741680887372014, "grad_norm": 0.6606165170669556, "learning_rate": 9.533743953144733e-06, "loss": 0.5813, "step": 2107 }, { "epoch": 0.6744880546075085, "grad_norm": 0.5467153787612915, "learning_rate": 9.532958573518121e-06, "loss": 0.5552, "step": 2108 }, { "epoch": 0.6748080204778157, "grad_norm": 0.6055881977081299, "learning_rate": 9.532172565391348e-06, "loss": 0.5708, "step": 2109 }, { "epoch": 0.6751279863481229, "grad_norm": 0.656714916229248, "learning_rate": 9.5313859288734e-06, "loss": 0.6132, "step": 2110 }, { "epoch": 0.67544795221843, "grad_norm": 0.6124818325042725, "learning_rate": 9.530598664073339e-06, "loss": 0.5514, "step": 2111 }, { "epoch": 0.6757679180887372, "grad_norm": 0.6045283675193787, "learning_rate": 9.529810771100323e-06, "loss": 0.5938, "step": 2112 }, { "epoch": 0.6760878839590444, "grad_norm": 0.5918228626251221, "learning_rate": 9.529022250063596e-06, "loss": 0.5707, "step": 2113 }, { "epoch": 0.6764078498293515, "grad_norm": 0.6131752729415894, "learning_rate": 9.528233101072484e-06, "loss": 0.5446, "step": 2114 }, { "epoch": 0.6767278156996587, "grad_norm": 0.5795077681541443, "learning_rate": 9.527443324236403e-06, "loss": 0.5858, "step": 2115 }, { "epoch": 0.6770477815699659, "grad_norm": 0.5680365562438965, "learning_rate": 9.526652919664862e-06, "loss": 0.5822, "step": 2116 }, { "epoch": 0.6773677474402731, "grad_norm": 0.5775081515312195, "learning_rate": 9.525861887467445e-06, "loss": 0.5118, "step": 2117 }, { "epoch": 0.6776877133105802, "grad_norm": 0.5150967240333557, "learning_rate": 9.525070227753835e-06, "loss": 0.5428, "step": 2118 }, { "epoch": 0.6780076791808873, "grad_norm": 0.5972749590873718, "learning_rate": 9.524277940633793e-06, "loss": 0.5869, "step": 2119 }, { "epoch": 0.6783276450511946, "grad_norm": 0.5536697506904602, "learning_rate": 9.523485026217173e-06, "loss": 0.5436, "step": 2120 }, { "epoch": 0.6786476109215017, "grad_norm": 0.5641371011734009, "learning_rate": 9.52269148461391e-06, "loss": 0.5613, "step": 2121 }, { "epoch": 0.6789675767918089, "grad_norm": 0.5990756750106812, "learning_rate": 9.521897315934032e-06, "loss": 0.5683, "step": 2122 }, { "epoch": 0.679287542662116, "grad_norm": 0.6080614924430847, "learning_rate": 9.521102520287653e-06, "loss": 0.5252, "step": 2123 }, { "epoch": 0.6796075085324232, "grad_norm": 0.6255257725715637, "learning_rate": 9.520307097784968e-06, "loss": 0.5752, "step": 2124 }, { "epoch": 0.6799274744027304, "grad_norm": 0.5562875866889954, "learning_rate": 9.519511048536268e-06, "loss": 0.5629, "step": 2125 }, { "epoch": 0.6802474402730375, "grad_norm": 0.6114696860313416, "learning_rate": 9.518714372651922e-06, "loss": 0.5415, "step": 2126 }, { "epoch": 0.6805674061433447, "grad_norm": 0.5472313761711121, "learning_rate": 9.517917070242395e-06, "loss": 0.5656, "step": 2127 }, { "epoch": 0.6808873720136519, "grad_norm": 0.558124840259552, "learning_rate": 9.51711914141823e-06, "loss": 0.5919, "step": 2128 }, { "epoch": 0.681207337883959, "grad_norm": 0.518718421459198, "learning_rate": 9.51632058629006e-06, "loss": 0.5501, "step": 2129 }, { "epoch": 0.6815273037542662, "grad_norm": 0.6277896165847778, "learning_rate": 9.515521404968608e-06, "loss": 0.5744, "step": 2130 }, { "epoch": 0.6818472696245734, "grad_norm": 0.5866926312446594, "learning_rate": 9.514721597564683e-06, "loss": 0.5584, "step": 2131 }, { "epoch": 0.6821672354948806, "grad_norm": 0.6311371326446533, "learning_rate": 9.513921164189174e-06, "loss": 0.5514, "step": 2132 }, { "epoch": 0.6824872013651877, "grad_norm": 0.5464537739753723, "learning_rate": 9.513120104953068e-06, "loss": 0.5536, "step": 2133 }, { "epoch": 0.6828071672354948, "grad_norm": 0.6255695223808289, "learning_rate": 9.512318419967427e-06, "loss": 0.5714, "step": 2134 }, { "epoch": 0.6831271331058021, "grad_norm": 0.6228389143943787, "learning_rate": 9.511516109343413e-06, "loss": 0.5756, "step": 2135 }, { "epoch": 0.6834470989761092, "grad_norm": 0.5070335865020752, "learning_rate": 9.51071317319226e-06, "loss": 0.5195, "step": 2136 }, { "epoch": 0.6837670648464164, "grad_norm": 0.6195150017738342, "learning_rate": 9.509909611625298e-06, "loss": 0.5495, "step": 2137 }, { "epoch": 0.6840870307167235, "grad_norm": 0.5539630651473999, "learning_rate": 9.509105424753945e-06, "loss": 0.5532, "step": 2138 }, { "epoch": 0.6844069965870307, "grad_norm": 0.5826463103294373, "learning_rate": 9.5083006126897e-06, "loss": 0.5402, "step": 2139 }, { "epoch": 0.6847269624573379, "grad_norm": 0.5683773159980774, "learning_rate": 9.507495175544151e-06, "loss": 0.6041, "step": 2140 }, { "epoch": 0.685046928327645, "grad_norm": 0.5582982301712036, "learning_rate": 9.506689113428974e-06, "loss": 0.5553, "step": 2141 }, { "epoch": 0.6853668941979523, "grad_norm": 0.5302207469940186, "learning_rate": 9.505882426455928e-06, "loss": 0.5365, "step": 2142 }, { "epoch": 0.6856868600682594, "grad_norm": 0.6130880117416382, "learning_rate": 9.505075114736864e-06, "loss": 0.5793, "step": 2143 }, { "epoch": 0.6860068259385665, "grad_norm": 0.6494556665420532, "learning_rate": 9.504267178383715e-06, "loss": 0.5663, "step": 2144 }, { "epoch": 0.6863267918088737, "grad_norm": 0.5702915787696838, "learning_rate": 9.503458617508504e-06, "loss": 0.5071, "step": 2145 }, { "epoch": 0.6866467576791809, "grad_norm": 0.6269530057907104, "learning_rate": 9.502649432223337e-06, "loss": 0.5841, "step": 2146 }, { "epoch": 0.6869667235494881, "grad_norm": 0.6300735473632812, "learning_rate": 9.50183962264041e-06, "loss": 0.5813, "step": 2147 }, { "epoch": 0.6872866894197952, "grad_norm": 0.6130741238594055, "learning_rate": 9.501029188872005e-06, "loss": 0.5486, "step": 2148 }, { "epoch": 0.6876066552901023, "grad_norm": 0.5809904932975769, "learning_rate": 9.500218131030485e-06, "loss": 0.5095, "step": 2149 }, { "epoch": 0.6879266211604096, "grad_norm": 0.6523924469947815, "learning_rate": 9.499406449228308e-06, "loss": 0.5684, "step": 2150 }, { "epoch": 0.6882465870307167, "grad_norm": 0.6566171050071716, "learning_rate": 9.498594143578015e-06, "loss": 0.6229, "step": 2151 }, { "epoch": 0.6885665529010239, "grad_norm": 0.6797991394996643, "learning_rate": 9.49778121419223e-06, "loss": 0.5686, "step": 2152 }, { "epoch": 0.6888865187713311, "grad_norm": 0.6882073283195496, "learning_rate": 9.496967661183672e-06, "loss": 0.5598, "step": 2153 }, { "epoch": 0.6892064846416383, "grad_norm": 0.5378338098526001, "learning_rate": 9.496153484665136e-06, "loss": 0.5425, "step": 2154 }, { "epoch": 0.6895264505119454, "grad_norm": 0.5711609125137329, "learning_rate": 9.49533868474951e-06, "loss": 0.5526, "step": 2155 }, { "epoch": 0.6898464163822525, "grad_norm": 0.6440864205360413, "learning_rate": 9.494523261549767e-06, "loss": 0.5619, "step": 2156 }, { "epoch": 0.6901663822525598, "grad_norm": 0.5898805856704712, "learning_rate": 9.493707215178969e-06, "loss": 0.5894, "step": 2157 }, { "epoch": 0.6904863481228669, "grad_norm": 0.6344678997993469, "learning_rate": 9.492890545750256e-06, "loss": 0.6375, "step": 2158 }, { "epoch": 0.690806313993174, "grad_norm": 0.6756842732429504, "learning_rate": 9.492073253376865e-06, "loss": 0.6162, "step": 2159 }, { "epoch": 0.6911262798634812, "grad_norm": 0.5748379826545715, "learning_rate": 9.491255338172116e-06, "loss": 0.5633, "step": 2160 }, { "epoch": 0.6914462457337884, "grad_norm": 0.48376211524009705, "learning_rate": 9.490436800249408e-06, "loss": 0.5135, "step": 2161 }, { "epoch": 0.6917662116040956, "grad_norm": 0.5189322829246521, "learning_rate": 9.489617639722238e-06, "loss": 0.5291, "step": 2162 }, { "epoch": 0.6920861774744027, "grad_norm": 0.6013830900192261, "learning_rate": 9.48879785670418e-06, "loss": 0.5762, "step": 2163 }, { "epoch": 0.6924061433447098, "grad_norm": 0.5476438999176025, "learning_rate": 9.487977451308899e-06, "loss": 0.5396, "step": 2164 }, { "epoch": 0.6927261092150171, "grad_norm": 0.6140022873878479, "learning_rate": 9.487156423650145e-06, "loss": 0.5812, "step": 2165 }, { "epoch": 0.6930460750853242, "grad_norm": 0.6048492193222046, "learning_rate": 9.486334773841754e-06, "loss": 0.5541, "step": 2166 }, { "epoch": 0.6933660409556314, "grad_norm": 0.5617876648902893, "learning_rate": 9.48551250199765e-06, "loss": 0.5525, "step": 2167 }, { "epoch": 0.6936860068259386, "grad_norm": 0.5541356205940247, "learning_rate": 9.484689608231841e-06, "loss": 0.5662, "step": 2168 }, { "epoch": 0.6940059726962458, "grad_norm": 0.5808799862861633, "learning_rate": 9.483866092658424e-06, "loss": 0.5356, "step": 2169 }, { "epoch": 0.6943259385665529, "grad_norm": 0.5347186326980591, "learning_rate": 9.48304195539158e-06, "loss": 0.5437, "step": 2170 }, { "epoch": 0.69464590443686, "grad_norm": 0.5472405552864075, "learning_rate": 9.482217196545573e-06, "loss": 0.5391, "step": 2171 }, { "epoch": 0.6949658703071673, "grad_norm": 0.6138960123062134, "learning_rate": 9.481391816234759e-06, "loss": 0.5718, "step": 2172 }, { "epoch": 0.6952858361774744, "grad_norm": 0.5508807301521301, "learning_rate": 9.48056581457358e-06, "loss": 0.5382, "step": 2173 }, { "epoch": 0.6956058020477816, "grad_norm": 0.49969929456710815, "learning_rate": 9.47973919167656e-06, "loss": 0.525, "step": 2174 }, { "epoch": 0.6959257679180887, "grad_norm": 0.5662696361541748, "learning_rate": 9.478911947658312e-06, "loss": 0.5612, "step": 2175 }, { "epoch": 0.6962457337883959, "grad_norm": 0.6077730655670166, "learning_rate": 9.478084082633535e-06, "loss": 0.5865, "step": 2176 }, { "epoch": 0.6965656996587031, "grad_norm": 0.5491179823875427, "learning_rate": 9.477255596717012e-06, "loss": 0.5485, "step": 2177 }, { "epoch": 0.6968856655290102, "grad_norm": 0.571593165397644, "learning_rate": 9.476426490023614e-06, "loss": 0.5466, "step": 2178 }, { "epoch": 0.6972056313993175, "grad_norm": 0.5554285049438477, "learning_rate": 9.475596762668298e-06, "loss": 0.527, "step": 2179 }, { "epoch": 0.6975255972696246, "grad_norm": 0.5728554129600525, "learning_rate": 9.474766414766108e-06, "loss": 0.5425, "step": 2180 }, { "epoch": 0.6978455631399317, "grad_norm": 0.5726357698440552, "learning_rate": 9.473935446432169e-06, "loss": 0.5428, "step": 2181 }, { "epoch": 0.6981655290102389, "grad_norm": 0.5684338808059692, "learning_rate": 9.4731038577817e-06, "loss": 0.5561, "step": 2182 }, { "epoch": 0.6984854948805461, "grad_norm": 0.632646381855011, "learning_rate": 9.472271648929998e-06, "loss": 0.5511, "step": 2183 }, { "epoch": 0.6988054607508533, "grad_norm": 0.6783565878868103, "learning_rate": 9.471438819992453e-06, "loss": 0.5756, "step": 2184 }, { "epoch": 0.6991254266211604, "grad_norm": 0.6838875412940979, "learning_rate": 9.470605371084535e-06, "loss": 0.5466, "step": 2185 }, { "epoch": 0.6994453924914675, "grad_norm": 0.6280619502067566, "learning_rate": 9.469771302321806e-06, "loss": 0.5683, "step": 2186 }, { "epoch": 0.6997653583617748, "grad_norm": 0.5844177007675171, "learning_rate": 9.468936613819908e-06, "loss": 0.5491, "step": 2187 }, { "epoch": 0.7000853242320819, "grad_norm": 0.6486324667930603, "learning_rate": 9.468101305694571e-06, "loss": 0.5942, "step": 2188 }, { "epoch": 0.700405290102389, "grad_norm": 0.723613440990448, "learning_rate": 9.467265378061614e-06, "loss": 0.5893, "step": 2189 }, { "epoch": 0.7007252559726962, "grad_norm": 0.569793701171875, "learning_rate": 9.466428831036938e-06, "loss": 0.5499, "step": 2190 }, { "epoch": 0.7010452218430034, "grad_norm": 0.6333669424057007, "learning_rate": 9.46559166473653e-06, "loss": 0.5642, "step": 2191 }, { "epoch": 0.7013651877133106, "grad_norm": 0.5689230561256409, "learning_rate": 9.464753879276467e-06, "loss": 0.5523, "step": 2192 }, { "epoch": 0.7016851535836177, "grad_norm": 0.612147867679596, "learning_rate": 9.463915474772908e-06, "loss": 0.5986, "step": 2193 }, { "epoch": 0.702005119453925, "grad_norm": 0.6131147742271423, "learning_rate": 9.463076451342096e-06, "loss": 0.5693, "step": 2194 }, { "epoch": 0.7023250853242321, "grad_norm": 0.583795964717865, "learning_rate": 9.462236809100366e-06, "loss": 0.5704, "step": 2195 }, { "epoch": 0.7026450511945392, "grad_norm": 0.4957069754600525, "learning_rate": 9.461396548164134e-06, "loss": 0.5479, "step": 2196 }, { "epoch": 0.7029650170648464, "grad_norm": 0.581490159034729, "learning_rate": 9.460555668649902e-06, "loss": 0.5683, "step": 2197 }, { "epoch": 0.7032849829351536, "grad_norm": 0.5183387398719788, "learning_rate": 9.459714170674264e-06, "loss": 0.5238, "step": 2198 }, { "epoch": 0.7036049488054608, "grad_norm": 0.559424102306366, "learning_rate": 9.458872054353888e-06, "loss": 0.561, "step": 2199 }, { "epoch": 0.7039249146757679, "grad_norm": 0.5429365634918213, "learning_rate": 9.458029319805539e-06, "loss": 0.5593, "step": 2200 }, { "epoch": 0.704244880546075, "grad_norm": 0.5553236603736877, "learning_rate": 9.45718596714606e-06, "loss": 0.5499, "step": 2201 }, { "epoch": 0.7045648464163823, "grad_norm": 0.5307144522666931, "learning_rate": 9.456341996492387e-06, "loss": 0.5841, "step": 2202 }, { "epoch": 0.7048848122866894, "grad_norm": 0.5576136708259583, "learning_rate": 9.455497407961533e-06, "loss": 0.5509, "step": 2203 }, { "epoch": 0.7052047781569966, "grad_norm": 0.5416161417961121, "learning_rate": 9.454652201670603e-06, "loss": 0.5637, "step": 2204 }, { "epoch": 0.7055247440273038, "grad_norm": 0.5229343175888062, "learning_rate": 9.453806377736788e-06, "loss": 0.5459, "step": 2205 }, { "epoch": 0.705844709897611, "grad_norm": 0.5486166477203369, "learning_rate": 9.452959936277359e-06, "loss": 0.4968, "step": 2206 }, { "epoch": 0.7061646757679181, "grad_norm": 0.578492283821106, "learning_rate": 9.452112877409678e-06, "loss": 0.5824, "step": 2207 }, { "epoch": 0.7064846416382252, "grad_norm": 0.5441255569458008, "learning_rate": 9.451265201251193e-06, "loss": 0.5638, "step": 2208 }, { "epoch": 0.7068046075085325, "grad_norm": 0.5505200028419495, "learning_rate": 9.450416907919431e-06, "loss": 0.5713, "step": 2209 }, { "epoch": 0.7071245733788396, "grad_norm": 0.5709050893783569, "learning_rate": 9.44956799753201e-06, "loss": 0.5245, "step": 2210 }, { "epoch": 0.7074445392491467, "grad_norm": 0.5804630517959595, "learning_rate": 9.448718470206633e-06, "loss": 0.5786, "step": 2211 }, { "epoch": 0.7077645051194539, "grad_norm": 0.5691149830818176, "learning_rate": 9.44786832606109e-06, "loss": 0.5291, "step": 2212 }, { "epoch": 0.7080844709897611, "grad_norm": 0.6125223636627197, "learning_rate": 9.447017565213252e-06, "loss": 0.5594, "step": 2213 }, { "epoch": 0.7084044368600683, "grad_norm": 0.6221138834953308, "learning_rate": 9.446166187781077e-06, "loss": 0.5753, "step": 2214 }, { "epoch": 0.7087244027303754, "grad_norm": 0.5770483016967773, "learning_rate": 9.445314193882614e-06, "loss": 0.5283, "step": 2215 }, { "epoch": 0.7090443686006825, "grad_norm": 0.5975071787834167, "learning_rate": 9.444461583635988e-06, "loss": 0.5723, "step": 2216 }, { "epoch": 0.7093643344709898, "grad_norm": 0.5773232579231262, "learning_rate": 9.443608357159417e-06, "loss": 0.5518, "step": 2217 }, { "epoch": 0.7096843003412969, "grad_norm": 0.6178917288780212, "learning_rate": 9.442754514571202e-06, "loss": 0.5543, "step": 2218 }, { "epoch": 0.7100042662116041, "grad_norm": 0.6687595248222351, "learning_rate": 9.441900055989728e-06, "loss": 0.5567, "step": 2219 }, { "epoch": 0.7103242320819113, "grad_norm": 0.5844324231147766, "learning_rate": 9.441044981533468e-06, "loss": 0.5789, "step": 2220 }, { "epoch": 0.7106441979522184, "grad_norm": 0.5794469714164734, "learning_rate": 9.44018929132098e-06, "loss": 0.5428, "step": 2221 }, { "epoch": 0.7109641638225256, "grad_norm": 0.6569241881370544, "learning_rate": 9.439332985470903e-06, "loss": 0.53, "step": 2222 }, { "epoch": 0.7112841296928327, "grad_norm": 0.5866938233375549, "learning_rate": 9.438476064101968e-06, "loss": 0.5437, "step": 2223 }, { "epoch": 0.71160409556314, "grad_norm": 0.5498455762863159, "learning_rate": 9.437618527332987e-06, "loss": 0.5631, "step": 2224 }, { "epoch": 0.7119240614334471, "grad_norm": 0.6395000219345093, "learning_rate": 9.436760375282858e-06, "loss": 0.5654, "step": 2225 }, { "epoch": 0.7122440273037542, "grad_norm": 0.6275232434272766, "learning_rate": 9.435901608070567e-06, "loss": 0.5715, "step": 2226 }, { "epoch": 0.7125639931740614, "grad_norm": 0.49817797541618347, "learning_rate": 9.43504222581518e-06, "loss": 0.5151, "step": 2227 }, { "epoch": 0.7128839590443686, "grad_norm": 0.5261150598526001, "learning_rate": 9.434182228635854e-06, "loss": 0.5704, "step": 2228 }, { "epoch": 0.7132039249146758, "grad_norm": 0.5595937967300415, "learning_rate": 9.433321616651826e-06, "loss": 0.5495, "step": 2229 }, { "epoch": 0.7135238907849829, "grad_norm": 0.5996530055999756, "learning_rate": 9.432460389982423e-06, "loss": 0.5793, "step": 2230 }, { "epoch": 0.7138438566552902, "grad_norm": 0.5601897835731506, "learning_rate": 9.431598548747055e-06, "loss": 0.5147, "step": 2231 }, { "epoch": 0.7141638225255973, "grad_norm": 0.5980433225631714, "learning_rate": 9.430736093065217e-06, "loss": 0.5913, "step": 2232 }, { "epoch": 0.7144837883959044, "grad_norm": 0.5890451669692993, "learning_rate": 9.429873023056488e-06, "loss": 0.5444, "step": 2233 }, { "epoch": 0.7148037542662116, "grad_norm": 0.4962788224220276, "learning_rate": 9.429009338840538e-06, "loss": 0.5403, "step": 2234 }, { "epoch": 0.7151237201365188, "grad_norm": 0.5949501395225525, "learning_rate": 9.428145040537112e-06, "loss": 0.5499, "step": 2235 }, { "epoch": 0.715443686006826, "grad_norm": 0.7749754786491394, "learning_rate": 9.427280128266049e-06, "loss": 0.5708, "step": 2236 }, { "epoch": 0.7157636518771331, "grad_norm": 0.5890344977378845, "learning_rate": 9.426414602147272e-06, "loss": 0.5684, "step": 2237 }, { "epoch": 0.7160836177474402, "grad_norm": 0.5542466640472412, "learning_rate": 9.425548462300784e-06, "loss": 0.5601, "step": 2238 }, { "epoch": 0.7164035836177475, "grad_norm": 0.5782498121261597, "learning_rate": 9.424681708846679e-06, "loss": 0.545, "step": 2239 }, { "epoch": 0.7167235494880546, "grad_norm": 0.6824494004249573, "learning_rate": 9.423814341905131e-06, "loss": 0.5361, "step": 2240 }, { "epoch": 0.7170435153583617, "grad_norm": 0.6495462656021118, "learning_rate": 9.422946361596403e-06, "loss": 0.5401, "step": 2241 }, { "epoch": 0.7173634812286689, "grad_norm": 0.5456177592277527, "learning_rate": 9.422077768040841e-06, "loss": 0.5466, "step": 2242 }, { "epoch": 0.7176834470989761, "grad_norm": 0.6046304106712341, "learning_rate": 9.421208561358876e-06, "loss": 0.5417, "step": 2243 }, { "epoch": 0.7180034129692833, "grad_norm": 0.589859127998352, "learning_rate": 9.420338741671025e-06, "loss": 0.5742, "step": 2244 }, { "epoch": 0.7183233788395904, "grad_norm": 0.6276467442512512, "learning_rate": 9.41946830909789e-06, "loss": 0.5462, "step": 2245 }, { "epoch": 0.7186433447098977, "grad_norm": 0.6103612184524536, "learning_rate": 9.418597263760156e-06, "loss": 0.6013, "step": 2246 }, { "epoch": 0.7189633105802048, "grad_norm": 0.5994572043418884, "learning_rate": 9.417725605778599e-06, "loss": 0.5452, "step": 2247 }, { "epoch": 0.7192832764505119, "grad_norm": 0.6904470324516296, "learning_rate": 9.416853335274068e-06, "loss": 0.5627, "step": 2248 }, { "epoch": 0.7196032423208191, "grad_norm": 0.5794340372085571, "learning_rate": 9.415980452367511e-06, "loss": 0.5332, "step": 2249 }, { "epoch": 0.7199232081911263, "grad_norm": 0.6280128359794617, "learning_rate": 9.415106957179952e-06, "loss": 0.5445, "step": 2250 }, { "epoch": 0.7202431740614335, "grad_norm": 0.6575672030448914, "learning_rate": 9.414232849832501e-06, "loss": 0.5578, "step": 2251 }, { "epoch": 0.7205631399317406, "grad_norm": 0.6580616235733032, "learning_rate": 9.413358130446356e-06, "loss": 0.5861, "step": 2252 }, { "epoch": 0.7208831058020477, "grad_norm": 0.5627638101577759, "learning_rate": 9.412482799142795e-06, "loss": 0.5496, "step": 2253 }, { "epoch": 0.721203071672355, "grad_norm": 0.49028298258781433, "learning_rate": 9.411606856043187e-06, "loss": 0.5239, "step": 2254 }, { "epoch": 0.7215230375426621, "grad_norm": 0.6451398730278015, "learning_rate": 9.41073030126898e-06, "loss": 0.5666, "step": 2255 }, { "epoch": 0.7218430034129693, "grad_norm": 0.6588875651359558, "learning_rate": 9.40985313494171e-06, "loss": 0.5748, "step": 2256 }, { "epoch": 0.7221629692832765, "grad_norm": 0.5587199926376343, "learning_rate": 9.408975357182998e-06, "loss": 0.5774, "step": 2257 }, { "epoch": 0.7224829351535836, "grad_norm": 0.6241255402565002, "learning_rate": 9.40809696811455e-06, "loss": 0.5683, "step": 2258 }, { "epoch": 0.7228029010238908, "grad_norm": 0.6500443816184998, "learning_rate": 9.407217967858154e-06, "loss": 0.5878, "step": 2259 }, { "epoch": 0.7231228668941979, "grad_norm": 0.6004780530929565, "learning_rate": 9.406338356535686e-06, "loss": 0.5402, "step": 2260 }, { "epoch": 0.7234428327645052, "grad_norm": 0.5574365258216858, "learning_rate": 9.405458134269101e-06, "loss": 0.564, "step": 2261 }, { "epoch": 0.7237627986348123, "grad_norm": 0.5756843686103821, "learning_rate": 9.404577301180447e-06, "loss": 0.5, "step": 2262 }, { "epoch": 0.7240827645051194, "grad_norm": 0.6266509294509888, "learning_rate": 9.403695857391851e-06, "loss": 0.5331, "step": 2263 }, { "epoch": 0.7244027303754266, "grad_norm": 0.623327910900116, "learning_rate": 9.402813803025526e-06, "loss": 0.5532, "step": 2264 }, { "epoch": 0.7247226962457338, "grad_norm": 0.5519808530807495, "learning_rate": 9.401931138203771e-06, "loss": 0.5492, "step": 2265 }, { "epoch": 0.725042662116041, "grad_norm": 0.6137987375259399, "learning_rate": 9.401047863048969e-06, "loss": 0.5459, "step": 2266 }, { "epoch": 0.7253626279863481, "grad_norm": 0.6169253587722778, "learning_rate": 9.400163977683586e-06, "loss": 0.5627, "step": 2267 }, { "epoch": 0.7256825938566553, "grad_norm": 0.54496169090271, "learning_rate": 9.399279482230174e-06, "loss": 0.5469, "step": 2268 }, { "epoch": 0.7260025597269625, "grad_norm": 0.5488664507865906, "learning_rate": 9.398394376811368e-06, "loss": 0.5397, "step": 2269 }, { "epoch": 0.7263225255972696, "grad_norm": 0.6857829689979553, "learning_rate": 9.397508661549892e-06, "loss": 0.5724, "step": 2270 }, { "epoch": 0.7266424914675768, "grad_norm": 0.5773136615753174, "learning_rate": 9.396622336568548e-06, "loss": 0.5493, "step": 2271 }, { "epoch": 0.726962457337884, "grad_norm": 0.6079044938087463, "learning_rate": 9.395735401990229e-06, "loss": 0.5908, "step": 2272 }, { "epoch": 0.7272824232081911, "grad_norm": 0.6028911471366882, "learning_rate": 9.394847857937907e-06, "loss": 0.5518, "step": 2273 }, { "epoch": 0.7276023890784983, "grad_norm": 0.5717269778251648, "learning_rate": 9.393959704534644e-06, "loss": 0.5436, "step": 2274 }, { "epoch": 0.7279223549488054, "grad_norm": 0.5948625206947327, "learning_rate": 9.393070941903581e-06, "loss": 0.5364, "step": 2275 }, { "epoch": 0.7282423208191127, "grad_norm": 0.5648112893104553, "learning_rate": 9.39218157016795e-06, "loss": 0.5549, "step": 2276 }, { "epoch": 0.7285622866894198, "grad_norm": 0.5189826488494873, "learning_rate": 9.391291589451056e-06, "loss": 0.5415, "step": 2277 }, { "epoch": 0.7288822525597269, "grad_norm": 0.5410546660423279, "learning_rate": 9.390400999876301e-06, "loss": 0.5813, "step": 2278 }, { "epoch": 0.7292022184300341, "grad_norm": 0.5291692614555359, "learning_rate": 9.389509801567166e-06, "loss": 0.5725, "step": 2279 }, { "epoch": 0.7295221843003413, "grad_norm": 0.5540335178375244, "learning_rate": 9.388617994647217e-06, "loss": 0.535, "step": 2280 }, { "epoch": 0.7298421501706485, "grad_norm": 0.5406812429428101, "learning_rate": 9.387725579240103e-06, "loss": 0.5461, "step": 2281 }, { "epoch": 0.7301621160409556, "grad_norm": 0.694333553314209, "learning_rate": 9.386832555469559e-06, "loss": 0.5458, "step": 2282 }, { "epoch": 0.7304820819112628, "grad_norm": 0.6259264349937439, "learning_rate": 9.385938923459405e-06, "loss": 0.5644, "step": 2283 }, { "epoch": 0.73080204778157, "grad_norm": 0.5157386064529419, "learning_rate": 9.385044683333542e-06, "loss": 0.5576, "step": 2284 }, { "epoch": 0.7311220136518771, "grad_norm": 0.6391353011131287, "learning_rate": 9.384149835215956e-06, "loss": 0.5555, "step": 2285 }, { "epoch": 0.7314419795221843, "grad_norm": 0.5598930716514587, "learning_rate": 9.383254379230724e-06, "loss": 0.5507, "step": 2286 }, { "epoch": 0.7317619453924915, "grad_norm": 0.5639938712120056, "learning_rate": 9.382358315501999e-06, "loss": 0.5908, "step": 2287 }, { "epoch": 0.7320819112627986, "grad_norm": 0.5010425448417664, "learning_rate": 9.38146164415402e-06, "loss": 0.5409, "step": 2288 }, { "epoch": 0.7324018771331058, "grad_norm": 0.5784039497375488, "learning_rate": 9.380564365311113e-06, "loss": 0.5346, "step": 2289 }, { "epoch": 0.7327218430034129, "grad_norm": 0.5947515964508057, "learning_rate": 9.379666479097688e-06, "loss": 0.586, "step": 2290 }, { "epoch": 0.7330418088737202, "grad_norm": 0.5534799098968506, "learning_rate": 9.378767985638236e-06, "loss": 0.5561, "step": 2291 }, { "epoch": 0.7333617747440273, "grad_norm": 0.5672453045845032, "learning_rate": 9.377868885057335e-06, "loss": 0.5461, "step": 2292 }, { "epoch": 0.7336817406143344, "grad_norm": 0.5631408095359802, "learning_rate": 9.376969177479647e-06, "loss": 0.5254, "step": 2293 }, { "epoch": 0.7340017064846417, "grad_norm": 0.5236619710922241, "learning_rate": 9.376068863029916e-06, "loss": 0.5147, "step": 2294 }, { "epoch": 0.7343216723549488, "grad_norm": 0.5404022336006165, "learning_rate": 9.375167941832974e-06, "loss": 0.4949, "step": 2295 }, { "epoch": 0.734641638225256, "grad_norm": 0.5881881713867188, "learning_rate": 9.374266414013731e-06, "loss": 0.5972, "step": 2296 }, { "epoch": 0.7349616040955631, "grad_norm": 0.4837624430656433, "learning_rate": 9.37336427969719e-06, "loss": 0.5481, "step": 2297 }, { "epoch": 0.7352815699658704, "grad_norm": 0.5485632419586182, "learning_rate": 9.372461539008429e-06, "loss": 0.5446, "step": 2298 }, { "epoch": 0.7356015358361775, "grad_norm": 0.587704598903656, "learning_rate": 9.371558192072615e-06, "loss": 0.5284, "step": 2299 }, { "epoch": 0.7359215017064846, "grad_norm": 0.5816254615783691, "learning_rate": 9.370654239014999e-06, "loss": 0.5594, "step": 2300 }, { "epoch": 0.7362414675767918, "grad_norm": 0.6278566122055054, "learning_rate": 9.369749679960916e-06, "loss": 0.5864, "step": 2301 }, { "epoch": 0.736561433447099, "grad_norm": 0.5978617668151855, "learning_rate": 9.36884451503578e-06, "loss": 0.594, "step": 2302 }, { "epoch": 0.7368813993174061, "grad_norm": 0.5780118703842163, "learning_rate": 9.3679387443651e-06, "loss": 0.5678, "step": 2303 }, { "epoch": 0.7372013651877133, "grad_norm": 0.5417023301124573, "learning_rate": 9.367032368074455e-06, "loss": 0.5372, "step": 2304 }, { "epoch": 0.7375213310580204, "grad_norm": 0.6295386552810669, "learning_rate": 9.36612538628952e-06, "loss": 0.5751, "step": 2305 }, { "epoch": 0.7378412969283277, "grad_norm": 0.5656825304031372, "learning_rate": 9.365217799136047e-06, "loss": 0.556, "step": 2306 }, { "epoch": 0.7381612627986348, "grad_norm": 0.5957443714141846, "learning_rate": 9.364309606739875e-06, "loss": 0.555, "step": 2307 }, { "epoch": 0.738481228668942, "grad_norm": 0.5462073683738708, "learning_rate": 9.363400809226925e-06, "loss": 0.521, "step": 2308 }, { "epoch": 0.7388011945392492, "grad_norm": 0.5976278185844421, "learning_rate": 9.362491406723204e-06, "loss": 0.5796, "step": 2309 }, { "epoch": 0.7391211604095563, "grad_norm": 0.5917383432388306, "learning_rate": 9.361581399354803e-06, "loss": 0.5599, "step": 2310 }, { "epoch": 0.7394411262798635, "grad_norm": 0.5707705020904541, "learning_rate": 9.360670787247893e-06, "loss": 0.5708, "step": 2311 }, { "epoch": 0.7397610921501706, "grad_norm": 0.5376531481742859, "learning_rate": 9.359759570528731e-06, "loss": 0.5616, "step": 2312 }, { "epoch": 0.7400810580204779, "grad_norm": 0.5556191205978394, "learning_rate": 9.35884774932366e-06, "loss": 0.5563, "step": 2313 }, { "epoch": 0.740401023890785, "grad_norm": 0.6098505854606628, "learning_rate": 9.357935323759104e-06, "loss": 0.5677, "step": 2314 }, { "epoch": 0.7407209897610921, "grad_norm": 0.5319908857345581, "learning_rate": 9.357022293961574e-06, "loss": 0.5497, "step": 2315 }, { "epoch": 0.7410409556313993, "grad_norm": 0.5752114057540894, "learning_rate": 9.356108660057662e-06, "loss": 0.5249, "step": 2316 }, { "epoch": 0.7413609215017065, "grad_norm": 0.5302920937538147, "learning_rate": 9.355194422174042e-06, "loss": 0.5319, "step": 2317 }, { "epoch": 0.7416808873720137, "grad_norm": 0.5611221790313721, "learning_rate": 9.354279580437476e-06, "loss": 0.5687, "step": 2318 }, { "epoch": 0.7420008532423208, "grad_norm": 0.5666431188583374, "learning_rate": 9.35336413497481e-06, "loss": 0.5119, "step": 2319 }, { "epoch": 0.742320819112628, "grad_norm": 0.5902916789054871, "learning_rate": 9.352448085912966e-06, "loss": 0.5731, "step": 2320 }, { "epoch": 0.7426407849829352, "grad_norm": 0.4963330626487732, "learning_rate": 9.351531433378963e-06, "loss": 0.5328, "step": 2321 }, { "epoch": 0.7429607508532423, "grad_norm": 0.5711694955825806, "learning_rate": 9.350614177499887e-06, "loss": 0.5581, "step": 2322 }, { "epoch": 0.7432807167235495, "grad_norm": 0.5014269948005676, "learning_rate": 9.349696318402926e-06, "loss": 0.5519, "step": 2323 }, { "epoch": 0.7436006825938567, "grad_norm": 0.4947138726711273, "learning_rate": 9.348777856215335e-06, "loss": 0.5613, "step": 2324 }, { "epoch": 0.7439206484641638, "grad_norm": 0.6116201281547546, "learning_rate": 9.347858791064462e-06, "loss": 0.5401, "step": 2325 }, { "epoch": 0.744240614334471, "grad_norm": 0.5416233539581299, "learning_rate": 9.346939123077738e-06, "loss": 0.5463, "step": 2326 }, { "epoch": 0.7445605802047781, "grad_norm": 0.5299511551856995, "learning_rate": 9.346018852382676e-06, "loss": 0.5404, "step": 2327 }, { "epoch": 0.7448805460750854, "grad_norm": 0.5581720471382141, "learning_rate": 9.34509797910687e-06, "loss": 0.5422, "step": 2328 }, { "epoch": 0.7452005119453925, "grad_norm": 0.6147540211677551, "learning_rate": 9.344176503378003e-06, "loss": 0.5996, "step": 2329 }, { "epoch": 0.7455204778156996, "grad_norm": 0.5862179398536682, "learning_rate": 9.343254425323837e-06, "loss": 0.5432, "step": 2330 }, { "epoch": 0.7458404436860068, "grad_norm": 0.6720561981201172, "learning_rate": 9.34233174507222e-06, "loss": 0.5424, "step": 2331 }, { "epoch": 0.746160409556314, "grad_norm": 0.582324743270874, "learning_rate": 9.341408462751084e-06, "loss": 0.5591, "step": 2332 }, { "epoch": 0.7464803754266212, "grad_norm": 0.5627602338790894, "learning_rate": 9.34048457848844e-06, "loss": 0.545, "step": 2333 }, { "epoch": 0.7468003412969283, "grad_norm": 0.66405189037323, "learning_rate": 9.339560092412387e-06, "loss": 0.5512, "step": 2334 }, { "epoch": 0.7471203071672355, "grad_norm": 0.6157020330429077, "learning_rate": 9.338635004651108e-06, "loss": 0.556, "step": 2335 }, { "epoch": 0.7474402730375427, "grad_norm": 0.5703638792037964, "learning_rate": 9.337709315332867e-06, "loss": 0.5773, "step": 2336 }, { "epoch": 0.7477602389078498, "grad_norm": 0.6717652082443237, "learning_rate": 9.336783024586009e-06, "loss": 0.5652, "step": 2337 }, { "epoch": 0.748080204778157, "grad_norm": 0.6250460147857666, "learning_rate": 9.335856132538968e-06, "loss": 0.5611, "step": 2338 }, { "epoch": 0.7484001706484642, "grad_norm": 0.5538761019706726, "learning_rate": 9.33492863932026e-06, "loss": 0.5491, "step": 2339 }, { "epoch": 0.7487201365187713, "grad_norm": 0.5979110598564148, "learning_rate": 9.334000545058479e-06, "loss": 0.553, "step": 2340 }, { "epoch": 0.7490401023890785, "grad_norm": 0.6450609564781189, "learning_rate": 9.333071849882308e-06, "loss": 0.5387, "step": 2341 }, { "epoch": 0.7493600682593856, "grad_norm": 0.5491853952407837, "learning_rate": 9.332142553920513e-06, "loss": 0.5501, "step": 2342 }, { "epoch": 0.7496800341296929, "grad_norm": 0.5262613296508789, "learning_rate": 9.331212657301942e-06, "loss": 0.5415, "step": 2343 }, { "epoch": 0.75, "grad_norm": 0.6032441258430481, "learning_rate": 9.330282160155524e-06, "loss": 0.5409, "step": 2344 }, { "epoch": 0.7503199658703071, "grad_norm": 0.5730162858963013, "learning_rate": 9.329351062610277e-06, "loss": 0.5634, "step": 2345 }, { "epoch": 0.7506399317406144, "grad_norm": 0.673250138759613, "learning_rate": 9.328419364795295e-06, "loss": 0.5421, "step": 2346 }, { "epoch": 0.7509598976109215, "grad_norm": 0.6810570955276489, "learning_rate": 9.32748706683976e-06, "loss": 0.6039, "step": 2347 }, { "epoch": 0.7512798634812287, "grad_norm": 0.5421538352966309, "learning_rate": 9.326554168872939e-06, "loss": 0.535, "step": 2348 }, { "epoch": 0.7515998293515358, "grad_norm": 0.5963729023933411, "learning_rate": 9.325620671024179e-06, "loss": 0.5228, "step": 2349 }, { "epoch": 0.751919795221843, "grad_norm": 0.6112815141677856, "learning_rate": 9.324686573422905e-06, "loss": 0.5389, "step": 2350 }, { "epoch": 0.7522397610921502, "grad_norm": 0.558083713054657, "learning_rate": 9.323751876198638e-06, "loss": 0.5221, "step": 2351 }, { "epoch": 0.7525597269624573, "grad_norm": 0.6266518831253052, "learning_rate": 9.32281657948097e-06, "loss": 0.5705, "step": 2352 }, { "epoch": 0.7528796928327645, "grad_norm": 0.5977521538734436, "learning_rate": 9.321880683399585e-06, "loss": 0.5114, "step": 2353 }, { "epoch": 0.7531996587030717, "grad_norm": 0.5047455430030823, "learning_rate": 9.320944188084241e-06, "loss": 0.5552, "step": 2354 }, { "epoch": 0.7535196245733788, "grad_norm": 0.6396503448486328, "learning_rate": 9.32000709366479e-06, "loss": 0.5598, "step": 2355 }, { "epoch": 0.753839590443686, "grad_norm": 0.6784676313400269, "learning_rate": 9.319069400271158e-06, "loss": 0.6056, "step": 2356 }, { "epoch": 0.7541595563139932, "grad_norm": 0.5095100402832031, "learning_rate": 9.318131108033355e-06, "loss": 0.5418, "step": 2357 }, { "epoch": 0.7544795221843004, "grad_norm": 0.5125514268875122, "learning_rate": 9.31719221708148e-06, "loss": 0.532, "step": 2358 }, { "epoch": 0.7547994880546075, "grad_norm": 0.6723603010177612, "learning_rate": 9.316252727545712e-06, "loss": 0.5809, "step": 2359 }, { "epoch": 0.7551194539249146, "grad_norm": 0.6266522407531738, "learning_rate": 9.31531263955631e-06, "loss": 0.5725, "step": 2360 }, { "epoch": 0.7554394197952219, "grad_norm": 0.7055239677429199, "learning_rate": 9.314371953243619e-06, "loss": 0.5895, "step": 2361 }, { "epoch": 0.755759385665529, "grad_norm": 0.5571537613868713, "learning_rate": 9.313430668738067e-06, "loss": 0.5023, "step": 2362 }, { "epoch": 0.7560793515358362, "grad_norm": 0.5370808839797974, "learning_rate": 9.312488786170162e-06, "loss": 0.5384, "step": 2363 }, { "epoch": 0.7563993174061433, "grad_norm": 0.7225911021232605, "learning_rate": 9.311546305670499e-06, "loss": 0.5611, "step": 2364 }, { "epoch": 0.7567192832764505, "grad_norm": 0.6097559928894043, "learning_rate": 9.310603227369754e-06, "loss": 0.5769, "step": 2365 }, { "epoch": 0.7570392491467577, "grad_norm": 0.5765969753265381, "learning_rate": 9.309659551398687e-06, "loss": 0.5401, "step": 2366 }, { "epoch": 0.7573592150170648, "grad_norm": 0.5765284299850464, "learning_rate": 9.308715277888138e-06, "loss": 0.5554, "step": 2367 }, { "epoch": 0.757679180887372, "grad_norm": 0.5664956569671631, "learning_rate": 9.307770406969032e-06, "loss": 0.5521, "step": 2368 }, { "epoch": 0.7579991467576792, "grad_norm": 0.6559858322143555, "learning_rate": 9.306824938772376e-06, "loss": 0.5975, "step": 2369 }, { "epoch": 0.7583191126279863, "grad_norm": 0.5683647990226746, "learning_rate": 9.305878873429262e-06, "loss": 0.5589, "step": 2370 }, { "epoch": 0.7586390784982935, "grad_norm": 0.5724915266036987, "learning_rate": 9.30493221107086e-06, "loss": 0.5845, "step": 2371 }, { "epoch": 0.7589590443686007, "grad_norm": 0.5758845210075378, "learning_rate": 9.303984951828432e-06, "loss": 0.5192, "step": 2372 }, { "epoch": 0.7592790102389079, "grad_norm": 0.6145362257957458, "learning_rate": 9.30303709583331e-06, "loss": 0.5919, "step": 2373 }, { "epoch": 0.759598976109215, "grad_norm": 0.5690919756889343, "learning_rate": 9.302088643216917e-06, "loss": 0.5514, "step": 2374 }, { "epoch": 0.7599189419795221, "grad_norm": 0.5775154232978821, "learning_rate": 9.301139594110759e-06, "loss": 0.5562, "step": 2375 }, { "epoch": 0.7602389078498294, "grad_norm": 0.5660016536712646, "learning_rate": 9.300189948646422e-06, "loss": 0.5245, "step": 2376 }, { "epoch": 0.7605588737201365, "grad_norm": 0.5814532041549683, "learning_rate": 9.299239706955576e-06, "loss": 0.567, "step": 2377 }, { "epoch": 0.7608788395904437, "grad_norm": 0.567104160785675, "learning_rate": 9.298288869169973e-06, "loss": 0.5791, "step": 2378 }, { "epoch": 0.7611988054607508, "grad_norm": 0.5601080060005188, "learning_rate": 9.297337435421447e-06, "loss": 0.5262, "step": 2379 }, { "epoch": 0.761518771331058, "grad_norm": 0.6034749150276184, "learning_rate": 9.296385405841915e-06, "loss": 0.5797, "step": 2380 }, { "epoch": 0.7618387372013652, "grad_norm": 0.5768266916275024, "learning_rate": 9.295432780563378e-06, "loss": 0.5407, "step": 2381 }, { "epoch": 0.7621587030716723, "grad_norm": 0.641884446144104, "learning_rate": 9.294479559717918e-06, "loss": 0.5483, "step": 2382 }, { "epoch": 0.7624786689419796, "grad_norm": 0.5506654381752014, "learning_rate": 9.2935257434377e-06, "loss": 0.5332, "step": 2383 }, { "epoch": 0.7627986348122867, "grad_norm": 0.6289987564086914, "learning_rate": 9.292571331854975e-06, "loss": 0.5984, "step": 2384 }, { "epoch": 0.7631186006825939, "grad_norm": 0.5896539092063904, "learning_rate": 9.291616325102068e-06, "loss": 0.5669, "step": 2385 }, { "epoch": 0.763438566552901, "grad_norm": 0.5433265566825867, "learning_rate": 9.290660723311396e-06, "loss": 0.5449, "step": 2386 }, { "epoch": 0.7637585324232082, "grad_norm": 0.6118497252464294, "learning_rate": 9.289704526615453e-06, "loss": 0.5488, "step": 2387 }, { "epoch": 0.7640784982935154, "grad_norm": 0.5877235531806946, "learning_rate": 9.288747735146817e-06, "loss": 0.5455, "step": 2388 }, { "epoch": 0.7643984641638225, "grad_norm": 0.5190421938896179, "learning_rate": 9.287790349038148e-06, "loss": 0.5325, "step": 2389 }, { "epoch": 0.7647184300341296, "grad_norm": 0.6026919484138489, "learning_rate": 9.286832368422188e-06, "loss": 0.5318, "step": 2390 }, { "epoch": 0.7650383959044369, "grad_norm": 0.5628033876419067, "learning_rate": 9.285873793431764e-06, "loss": 0.5363, "step": 2391 }, { "epoch": 0.765358361774744, "grad_norm": 0.6197019219398499, "learning_rate": 9.284914624199782e-06, "loss": 0.5844, "step": 2392 }, { "epoch": 0.7656783276450512, "grad_norm": 0.5734281539916992, "learning_rate": 9.28395486085923e-06, "loss": 0.5388, "step": 2393 }, { "epoch": 0.7659982935153583, "grad_norm": 0.5641579627990723, "learning_rate": 9.282994503543185e-06, "loss": 0.5716, "step": 2394 }, { "epoch": 0.7663182593856656, "grad_norm": 0.5311891436576843, "learning_rate": 9.282033552384802e-06, "loss": 0.5338, "step": 2395 }, { "epoch": 0.7666382252559727, "grad_norm": 0.5733810067176819, "learning_rate": 9.281072007517312e-06, "loss": 0.589, "step": 2396 }, { "epoch": 0.7669581911262798, "grad_norm": 0.5777380466461182, "learning_rate": 9.28010986907404e-06, "loss": 0.5742, "step": 2397 }, { "epoch": 0.7672781569965871, "grad_norm": 0.547640860080719, "learning_rate": 9.279147137188383e-06, "loss": 0.5362, "step": 2398 }, { "epoch": 0.7675981228668942, "grad_norm": 0.5929672718048096, "learning_rate": 9.278183811993829e-06, "loss": 0.5552, "step": 2399 }, { "epoch": 0.7679180887372014, "grad_norm": 0.5406671762466431, "learning_rate": 9.277219893623944e-06, "loss": 0.5282, "step": 2400 }, { "epoch": 0.7682380546075085, "grad_norm": 0.6314170360565186, "learning_rate": 9.276255382212374e-06, "loss": 0.5592, "step": 2401 }, { "epoch": 0.7685580204778157, "grad_norm": 0.5819299817085266, "learning_rate": 9.27529027789285e-06, "loss": 0.55, "step": 2402 }, { "epoch": 0.7688779863481229, "grad_norm": 0.5875648856163025, "learning_rate": 9.274324580799187e-06, "loss": 0.5232, "step": 2403 }, { "epoch": 0.76919795221843, "grad_norm": 0.6291683912277222, "learning_rate": 9.273358291065278e-06, "loss": 0.581, "step": 2404 }, { "epoch": 0.7695179180887372, "grad_norm": 0.5859588384628296, "learning_rate": 9.2723914088251e-06, "loss": 0.5647, "step": 2405 }, { "epoch": 0.7698378839590444, "grad_norm": 0.6533288955688477, "learning_rate": 9.271423934212716e-06, "loss": 0.5455, "step": 2406 }, { "epoch": 0.7701578498293515, "grad_norm": 0.6398971676826477, "learning_rate": 9.270455867362262e-06, "loss": 0.5864, "step": 2407 }, { "epoch": 0.7704778156996587, "grad_norm": 0.651111900806427, "learning_rate": 9.269487208407965e-06, "loss": 0.5817, "step": 2408 }, { "epoch": 0.7707977815699659, "grad_norm": 0.6151703000068665, "learning_rate": 9.268517957484132e-06, "loss": 0.5491, "step": 2409 }, { "epoch": 0.7711177474402731, "grad_norm": 0.5498530864715576, "learning_rate": 9.26754811472515e-06, "loss": 0.5109, "step": 2410 }, { "epoch": 0.7714377133105802, "grad_norm": 0.5751739144325256, "learning_rate": 9.266577680265485e-06, "loss": 0.5357, "step": 2411 }, { "epoch": 0.7717576791808873, "grad_norm": 0.5931308269500732, "learning_rate": 9.265606654239691e-06, "loss": 0.5096, "step": 2412 }, { "epoch": 0.7720776450511946, "grad_norm": 0.5704831480979919, "learning_rate": 9.264635036782406e-06, "loss": 0.5619, "step": 2413 }, { "epoch": 0.7723976109215017, "grad_norm": 0.5751301646232605, "learning_rate": 9.263662828028342e-06, "loss": 0.564, "step": 2414 }, { "epoch": 0.7727175767918089, "grad_norm": 0.6460949778556824, "learning_rate": 9.262690028112296e-06, "loss": 0.5846, "step": 2415 }, { "epoch": 0.773037542662116, "grad_norm": 0.6896305680274963, "learning_rate": 9.261716637169149e-06, "loss": 0.5768, "step": 2416 }, { "epoch": 0.7733575085324232, "grad_norm": 0.622558057308197, "learning_rate": 9.260742655333866e-06, "loss": 0.5444, "step": 2417 }, { "epoch": 0.7736774744027304, "grad_norm": 0.5928201079368591, "learning_rate": 9.259768082741487e-06, "loss": 0.5549, "step": 2418 }, { "epoch": 0.7739974402730375, "grad_norm": 0.601613461971283, "learning_rate": 9.258792919527138e-06, "loss": 0.5898, "step": 2419 }, { "epoch": 0.7743174061433447, "grad_norm": 0.5642574429512024, "learning_rate": 9.257817165826027e-06, "loss": 0.5639, "step": 2420 }, { "epoch": 0.7746373720136519, "grad_norm": 0.559229850769043, "learning_rate": 9.256840821773444e-06, "loss": 0.5087, "step": 2421 }, { "epoch": 0.774957337883959, "grad_norm": 0.5684958100318909, "learning_rate": 9.25586388750476e-06, "loss": 0.5474, "step": 2422 }, { "epoch": 0.7752773037542662, "grad_norm": 0.5171270370483398, "learning_rate": 9.254886363155429e-06, "loss": 0.4506, "step": 2423 }, { "epoch": 0.7755972696245734, "grad_norm": 0.5402911305427551, "learning_rate": 9.253908248860983e-06, "loss": 0.5472, "step": 2424 }, { "epoch": 0.7759172354948806, "grad_norm": 0.5100975632667542, "learning_rate": 9.252929544757043e-06, "loss": 0.5254, "step": 2425 }, { "epoch": 0.7762372013651877, "grad_norm": 0.500893771648407, "learning_rate": 9.251950250979303e-06, "loss": 0.5522, "step": 2426 }, { "epoch": 0.7765571672354948, "grad_norm": 0.5404722094535828, "learning_rate": 9.250970367663548e-06, "loss": 0.5315, "step": 2427 }, { "epoch": 0.7768771331058021, "grad_norm": 0.5347235202789307, "learning_rate": 9.249989894945635e-06, "loss": 0.5537, "step": 2428 }, { "epoch": 0.7771970989761092, "grad_norm": 0.5374565720558167, "learning_rate": 9.249008832961512e-06, "loss": 0.5449, "step": 2429 }, { "epoch": 0.7775170648464164, "grad_norm": 0.5726111531257629, "learning_rate": 9.248027181847202e-06, "loss": 0.5544, "step": 2430 }, { "epoch": 0.7778370307167235, "grad_norm": 0.6014326214790344, "learning_rate": 9.247044941738812e-06, "loss": 0.5675, "step": 2431 }, { "epoch": 0.7781569965870307, "grad_norm": 0.6092072129249573, "learning_rate": 9.246062112772532e-06, "loss": 0.5617, "step": 2432 }, { "epoch": 0.7784769624573379, "grad_norm": 0.5175899863243103, "learning_rate": 9.245078695084632e-06, "loss": 0.5597, "step": 2433 }, { "epoch": 0.778796928327645, "grad_norm": 0.6158414483070374, "learning_rate": 9.244094688811464e-06, "loss": 0.5519, "step": 2434 }, { "epoch": 0.7791168941979523, "grad_norm": 0.5460450053215027, "learning_rate": 9.243110094089462e-06, "loss": 0.5298, "step": 2435 }, { "epoch": 0.7794368600682594, "grad_norm": 0.622840404510498, "learning_rate": 9.242124911055141e-06, "loss": 0.5494, "step": 2436 }, { "epoch": 0.7797568259385665, "grad_norm": 0.6064757108688354, "learning_rate": 9.241139139845098e-06, "loss": 0.532, "step": 2437 }, { "epoch": 0.7800767918088737, "grad_norm": 0.5535280704498291, "learning_rate": 9.24015278059601e-06, "loss": 0.5737, "step": 2438 }, { "epoch": 0.7803967576791809, "grad_norm": 0.5520623922348022, "learning_rate": 9.239165833444637e-06, "loss": 0.5546, "step": 2439 }, { "epoch": 0.7807167235494881, "grad_norm": 0.5182672739028931, "learning_rate": 9.238178298527825e-06, "loss": 0.5435, "step": 2440 }, { "epoch": 0.7810366894197952, "grad_norm": 0.5491767525672913, "learning_rate": 9.23719017598249e-06, "loss": 0.5614, "step": 2441 }, { "epoch": 0.7813566552901023, "grad_norm": 0.6395105123519897, "learning_rate": 9.23620146594564e-06, "loss": 0.5591, "step": 2442 }, { "epoch": 0.7816766211604096, "grad_norm": 0.542748749256134, "learning_rate": 9.235212168554361e-06, "loss": 0.5607, "step": 2443 }, { "epoch": 0.7819965870307167, "grad_norm": 0.5202548503875732, "learning_rate": 9.234222283945819e-06, "loss": 0.5218, "step": 2444 }, { "epoch": 0.7823165529010239, "grad_norm": 0.5446274280548096, "learning_rate": 9.233231812257266e-06, "loss": 0.5106, "step": 2445 }, { "epoch": 0.7826365187713311, "grad_norm": 0.575524091720581, "learning_rate": 9.232240753626027e-06, "loss": 0.5425, "step": 2446 }, { "epoch": 0.7829564846416383, "grad_norm": 0.5616278052330017, "learning_rate": 9.231249108189517e-06, "loss": 0.5324, "step": 2447 }, { "epoch": 0.7832764505119454, "grad_norm": 0.6591216325759888, "learning_rate": 9.230256876085227e-06, "loss": 0.5237, "step": 2448 }, { "epoch": 0.7835964163822525, "grad_norm": 0.5643269419670105, "learning_rate": 9.229264057450733e-06, "loss": 0.5073, "step": 2449 }, { "epoch": 0.7839163822525598, "grad_norm": 0.6130651831626892, "learning_rate": 9.22827065242369e-06, "loss": 0.5591, "step": 2450 }, { "epoch": 0.7842363481228669, "grad_norm": 0.6077596545219421, "learning_rate": 9.227276661141831e-06, "loss": 0.5416, "step": 2451 }, { "epoch": 0.784556313993174, "grad_norm": 0.5985930562019348, "learning_rate": 9.22628208374298e-06, "loss": 0.5855, "step": 2452 }, { "epoch": 0.7848762798634812, "grad_norm": 0.5189453363418579, "learning_rate": 9.225286920365034e-06, "loss": 0.5481, "step": 2453 }, { "epoch": 0.7851962457337884, "grad_norm": 0.5798327326774597, "learning_rate": 9.224291171145973e-06, "loss": 0.5645, "step": 2454 }, { "epoch": 0.7855162116040956, "grad_norm": 0.6887683868408203, "learning_rate": 9.223294836223859e-06, "loss": 0.5703, "step": 2455 }, { "epoch": 0.7858361774744027, "grad_norm": 0.4903722107410431, "learning_rate": 9.222297915736835e-06, "loss": 0.5485, "step": 2456 }, { "epoch": 0.7861561433447098, "grad_norm": 0.48105481266975403, "learning_rate": 9.221300409823127e-06, "loss": 0.5185, "step": 2457 }, { "epoch": 0.7864761092150171, "grad_norm": 0.6211410760879517, "learning_rate": 9.220302318621036e-06, "loss": 0.5486, "step": 2458 }, { "epoch": 0.7867960750853242, "grad_norm": 0.5718651413917542, "learning_rate": 9.219303642268953e-06, "loss": 0.5382, "step": 2459 }, { "epoch": 0.7871160409556314, "grad_norm": 0.5989018678665161, "learning_rate": 9.218304380905343e-06, "loss": 0.538, "step": 2460 }, { "epoch": 0.7874360068259386, "grad_norm": 0.5832222104072571, "learning_rate": 9.217304534668758e-06, "loss": 0.5708, "step": 2461 }, { "epoch": 0.7877559726962458, "grad_norm": 0.5775060653686523, "learning_rate": 9.216304103697823e-06, "loss": 0.5366, "step": 2462 }, { "epoch": 0.7880759385665529, "grad_norm": 0.5242322087287903, "learning_rate": 9.215303088131255e-06, "loss": 0.5273, "step": 2463 }, { "epoch": 0.78839590443686, "grad_norm": 0.5479870438575745, "learning_rate": 9.21430148810784e-06, "loss": 0.5312, "step": 2464 }, { "epoch": 0.7887158703071673, "grad_norm": 0.6071968078613281, "learning_rate": 9.213299303766455e-06, "loss": 0.5651, "step": 2465 }, { "epoch": 0.7890358361774744, "grad_norm": 0.5586081743240356, "learning_rate": 9.212296535246053e-06, "loss": 0.5379, "step": 2466 }, { "epoch": 0.7893558020477816, "grad_norm": 0.6063216924667358, "learning_rate": 9.211293182685668e-06, "loss": 0.5698, "step": 2467 }, { "epoch": 0.7896757679180887, "grad_norm": 0.5611611008644104, "learning_rate": 9.210289246224416e-06, "loss": 0.5793, "step": 2468 }, { "epoch": 0.7899957337883959, "grad_norm": 0.557030200958252, "learning_rate": 9.209284726001497e-06, "loss": 0.544, "step": 2469 }, { "epoch": 0.7903156996587031, "grad_norm": 0.5873385667800903, "learning_rate": 9.208279622156187e-06, "loss": 0.6026, "step": 2470 }, { "epoch": 0.7906356655290102, "grad_norm": 0.5318636894226074, "learning_rate": 9.207273934827843e-06, "loss": 0.5339, "step": 2471 }, { "epoch": 0.7909556313993175, "grad_norm": 0.5742864608764648, "learning_rate": 9.206267664155906e-06, "loss": 0.57, "step": 2472 }, { "epoch": 0.7912755972696246, "grad_norm": 0.577001690864563, "learning_rate": 9.205260810279898e-06, "loss": 0.5645, "step": 2473 }, { "epoch": 0.7915955631399317, "grad_norm": 0.5809705257415771, "learning_rate": 9.20425337333942e-06, "loss": 0.545, "step": 2474 }, { "epoch": 0.7919155290102389, "grad_norm": 0.5854853391647339, "learning_rate": 9.203245353474155e-06, "loss": 0.613, "step": 2475 }, { "epoch": 0.7922354948805461, "grad_norm": 0.5700336694717407, "learning_rate": 9.202236750823863e-06, "loss": 0.574, "step": 2476 }, { "epoch": 0.7925554607508533, "grad_norm": 0.6192684769630432, "learning_rate": 9.201227565528391e-06, "loss": 0.5922, "step": 2477 }, { "epoch": 0.7928754266211604, "grad_norm": 0.571351945400238, "learning_rate": 9.200217797727663e-06, "loss": 0.5931, "step": 2478 }, { "epoch": 0.7931953924914675, "grad_norm": 0.6180151700973511, "learning_rate": 9.199207447561685e-06, "loss": 0.5079, "step": 2479 }, { "epoch": 0.7935153583617748, "grad_norm": 0.5922881364822388, "learning_rate": 9.198196515170541e-06, "loss": 0.5711, "step": 2480 }, { "epoch": 0.7938353242320819, "grad_norm": 0.5363336801528931, "learning_rate": 9.1971850006944e-06, "loss": 0.5138, "step": 2481 }, { "epoch": 0.794155290102389, "grad_norm": 0.6847931742668152, "learning_rate": 9.19617290427351e-06, "loss": 0.5606, "step": 2482 }, { "epoch": 0.7944752559726962, "grad_norm": 0.5617700815200806, "learning_rate": 9.195160226048198e-06, "loss": 0.5244, "step": 2483 }, { "epoch": 0.7947952218430034, "grad_norm": 0.5189504027366638, "learning_rate": 9.194146966158875e-06, "loss": 0.5361, "step": 2484 }, { "epoch": 0.7951151877133106, "grad_norm": 0.5778887271881104, "learning_rate": 9.193133124746029e-06, "loss": 0.5416, "step": 2485 }, { "epoch": 0.7954351535836177, "grad_norm": 0.5868055820465088, "learning_rate": 9.19211870195023e-06, "loss": 0.543, "step": 2486 }, { "epoch": 0.795755119453925, "grad_norm": 0.5262547731399536, "learning_rate": 9.19110369791213e-06, "loss": 0.5511, "step": 2487 }, { "epoch": 0.7960750853242321, "grad_norm": 0.4826159179210663, "learning_rate": 9.190088112772461e-06, "loss": 0.5329, "step": 2488 }, { "epoch": 0.7963950511945392, "grad_norm": 0.5700838565826416, "learning_rate": 9.189071946672034e-06, "loss": 0.5597, "step": 2489 }, { "epoch": 0.7967150170648464, "grad_norm": 0.6034910082817078, "learning_rate": 9.188055199751743e-06, "loss": 0.5425, "step": 2490 }, { "epoch": 0.7970349829351536, "grad_norm": 0.4894741475582123, "learning_rate": 9.187037872152558e-06, "loss": 0.5411, "step": 2491 }, { "epoch": 0.7973549488054608, "grad_norm": 0.5268144607543945, "learning_rate": 9.186019964015536e-06, "loss": 0.5332, "step": 2492 }, { "epoch": 0.7976749146757679, "grad_norm": 0.5782467126846313, "learning_rate": 9.18500147548181e-06, "loss": 0.5761, "step": 2493 }, { "epoch": 0.797994880546075, "grad_norm": 0.6057249903678894, "learning_rate": 9.183982406692593e-06, "loss": 0.5691, "step": 2494 }, { "epoch": 0.7983148464163823, "grad_norm": 0.5561552047729492, "learning_rate": 9.182962757789184e-06, "loss": 0.5671, "step": 2495 }, { "epoch": 0.7986348122866894, "grad_norm": 0.6109516024589539, "learning_rate": 9.181942528912957e-06, "loss": 0.5615, "step": 2496 }, { "epoch": 0.7989547781569966, "grad_norm": 0.6077998280525208, "learning_rate": 9.180921720205365e-06, "loss": 0.5502, "step": 2497 }, { "epoch": 0.7992747440273038, "grad_norm": 0.6387701630592346, "learning_rate": 9.179900331807949e-06, "loss": 0.5795, "step": 2498 }, { "epoch": 0.799594709897611, "grad_norm": 0.5575502514839172, "learning_rate": 9.178878363862323e-06, "loss": 0.5485, "step": 2499 }, { "epoch": 0.7999146757679181, "grad_norm": 0.5736941695213318, "learning_rate": 9.177855816510184e-06, "loss": 0.539, "step": 2500 }, { "epoch": 0.8002346416382252, "grad_norm": 0.5884807109832764, "learning_rate": 9.176832689893308e-06, "loss": 0.5695, "step": 2501 }, { "epoch": 0.8005546075085325, "grad_norm": 0.6237932443618774, "learning_rate": 9.175808984153558e-06, "loss": 0.5824, "step": 2502 }, { "epoch": 0.8008745733788396, "grad_norm": 0.6127126216888428, "learning_rate": 9.174784699432869e-06, "loss": 0.5423, "step": 2503 }, { "epoch": 0.8011945392491467, "grad_norm": 0.5360780954360962, "learning_rate": 9.173759835873257e-06, "loss": 0.5472, "step": 2504 }, { "epoch": 0.8015145051194539, "grad_norm": 0.5300256013870239, "learning_rate": 9.172734393616823e-06, "loss": 0.5536, "step": 2505 }, { "epoch": 0.8018344709897611, "grad_norm": 0.548160970211029, "learning_rate": 9.171708372805745e-06, "loss": 0.5125, "step": 2506 }, { "epoch": 0.8021544368600683, "grad_norm": 0.5152131915092468, "learning_rate": 9.170681773582283e-06, "loss": 0.518, "step": 2507 }, { "epoch": 0.8024744027303754, "grad_norm": 0.5524023771286011, "learning_rate": 9.169654596088775e-06, "loss": 0.501, "step": 2508 }, { "epoch": 0.8027943686006825, "grad_norm": 0.5697095394134521, "learning_rate": 9.168626840467642e-06, "loss": 0.5725, "step": 2509 }, { "epoch": 0.8031143344709898, "grad_norm": 0.48423081636428833, "learning_rate": 9.167598506861381e-06, "loss": 0.5171, "step": 2510 }, { "epoch": 0.8034343003412969, "grad_norm": 0.5414108037948608, "learning_rate": 9.166569595412576e-06, "loss": 0.5041, "step": 2511 }, { "epoch": 0.8037542662116041, "grad_norm": 0.6320025324821472, "learning_rate": 9.165540106263881e-06, "loss": 0.547, "step": 2512 }, { "epoch": 0.8040742320819113, "grad_norm": 0.6039816737174988, "learning_rate": 9.16451003955804e-06, "loss": 0.5548, "step": 2513 }, { "epoch": 0.8043941979522184, "grad_norm": 0.6364385485649109, "learning_rate": 9.163479395437874e-06, "loss": 0.5447, "step": 2514 }, { "epoch": 0.8047141638225256, "grad_norm": 0.5636100769042969, "learning_rate": 9.16244817404628e-06, "loss": 0.5541, "step": 2515 }, { "epoch": 0.8050341296928327, "grad_norm": 0.5757590532302856, "learning_rate": 9.161416375526239e-06, "loss": 0.5308, "step": 2516 }, { "epoch": 0.80535409556314, "grad_norm": 0.603026807308197, "learning_rate": 9.16038400002081e-06, "loss": 0.5636, "step": 2517 }, { "epoch": 0.8056740614334471, "grad_norm": 0.5327634215354919, "learning_rate": 9.159351047673135e-06, "loss": 0.5385, "step": 2518 }, { "epoch": 0.8059940273037542, "grad_norm": 0.5723426342010498, "learning_rate": 9.158317518626434e-06, "loss": 0.5588, "step": 2519 }, { "epoch": 0.8063139931740614, "grad_norm": 0.5705613493919373, "learning_rate": 9.157283413024007e-06, "loss": 0.5524, "step": 2520 }, { "epoch": 0.8066339590443686, "grad_norm": 0.5529534816741943, "learning_rate": 9.156248731009232e-06, "loss": 0.5022, "step": 2521 }, { "epoch": 0.8069539249146758, "grad_norm": 0.49362483620643616, "learning_rate": 9.155213472725572e-06, "loss": 0.4861, "step": 2522 }, { "epoch": 0.8072738907849829, "grad_norm": 0.5816895961761475, "learning_rate": 9.154177638316563e-06, "loss": 0.5934, "step": 2523 }, { "epoch": 0.8075938566552902, "grad_norm": 0.6569143533706665, "learning_rate": 9.153141227925828e-06, "loss": 0.5722, "step": 2524 }, { "epoch": 0.8079138225255973, "grad_norm": 0.5592034459114075, "learning_rate": 9.152104241697066e-06, "loss": 0.602, "step": 2525 }, { "epoch": 0.8082337883959044, "grad_norm": 0.5290538668632507, "learning_rate": 9.151066679774054e-06, "loss": 0.5583, "step": 2526 }, { "epoch": 0.8085537542662116, "grad_norm": 0.6192449927330017, "learning_rate": 9.150028542300654e-06, "loss": 0.5751, "step": 2527 }, { "epoch": 0.8088737201365188, "grad_norm": 0.6423482894897461, "learning_rate": 9.148989829420805e-06, "loss": 0.5344, "step": 2528 }, { "epoch": 0.809193686006826, "grad_norm": 0.5743381977081299, "learning_rate": 9.147950541278523e-06, "loss": 0.5625, "step": 2529 }, { "epoch": 0.8095136518771331, "grad_norm": 0.6718060970306396, "learning_rate": 9.14691067801791e-06, "loss": 0.603, "step": 2530 }, { "epoch": 0.8098336177474402, "grad_norm": 0.5860727429389954, "learning_rate": 9.145870239783143e-06, "loss": 0.5766, "step": 2531 }, { "epoch": 0.8101535836177475, "grad_norm": 0.5706322193145752, "learning_rate": 9.144829226718479e-06, "loss": 0.5721, "step": 2532 }, { "epoch": 0.8104735494880546, "grad_norm": 0.5903095006942749, "learning_rate": 9.143787638968255e-06, "loss": 0.5558, "step": 2533 }, { "epoch": 0.8107935153583617, "grad_norm": 0.5164182186126709, "learning_rate": 9.142745476676891e-06, "loss": 0.5865, "step": 2534 }, { "epoch": 0.8111134812286689, "grad_norm": 0.5815507173538208, "learning_rate": 9.141702739988882e-06, "loss": 0.5819, "step": 2535 }, { "epoch": 0.8114334470989761, "grad_norm": 0.5640064477920532, "learning_rate": 9.140659429048807e-06, "loss": 0.5121, "step": 2536 }, { "epoch": 0.8117534129692833, "grad_norm": 0.5752261281013489, "learning_rate": 9.139615544001319e-06, "loss": 0.5497, "step": 2537 }, { "epoch": 0.8120733788395904, "grad_norm": 0.555300235748291, "learning_rate": 9.138571084991157e-06, "loss": 0.5443, "step": 2538 }, { "epoch": 0.8123933447098977, "grad_norm": 0.6510416865348816, "learning_rate": 9.137526052163135e-06, "loss": 0.5879, "step": 2539 }, { "epoch": 0.8127133105802048, "grad_norm": 0.5922650694847107, "learning_rate": 9.136480445662147e-06, "loss": 0.4972, "step": 2540 }, { "epoch": 0.8130332764505119, "grad_norm": 0.5528703927993774, "learning_rate": 9.13543426563317e-06, "loss": 0.5555, "step": 2541 }, { "epoch": 0.8133532423208191, "grad_norm": 0.5601704716682434, "learning_rate": 9.134387512221258e-06, "loss": 0.5401, "step": 2542 }, { "epoch": 0.8136732081911263, "grad_norm": 0.5348582863807678, "learning_rate": 9.133340185571541e-06, "loss": 0.5659, "step": 2543 }, { "epoch": 0.8139931740614335, "grad_norm": 0.578411877155304, "learning_rate": 9.132292285829237e-06, "loss": 0.583, "step": 2544 }, { "epoch": 0.8143131399317406, "grad_norm": 0.5427941083908081, "learning_rate": 9.131243813139636e-06, "loss": 0.5231, "step": 2545 }, { "epoch": 0.8146331058020477, "grad_norm": 0.5223431587219238, "learning_rate": 9.130194767648111e-06, "loss": 0.5417, "step": 2546 }, { "epoch": 0.814953071672355, "grad_norm": 0.49521204829216003, "learning_rate": 9.129145149500111e-06, "loss": 0.5165, "step": 2547 }, { "epoch": 0.8152730375426621, "grad_norm": 0.6073542237281799, "learning_rate": 9.12809495884117e-06, "loss": 0.5444, "step": 2548 }, { "epoch": 0.8155930034129693, "grad_norm": 0.5415356159210205, "learning_rate": 9.127044195816895e-06, "loss": 0.5444, "step": 2549 }, { "epoch": 0.8159129692832765, "grad_norm": 0.5234550833702087, "learning_rate": 9.125992860572979e-06, "loss": 0.5263, "step": 2550 }, { "epoch": 0.8162329351535836, "grad_norm": 0.5407820343971252, "learning_rate": 9.124940953255188e-06, "loss": 0.5136, "step": 2551 }, { "epoch": 0.8165529010238908, "grad_norm": 0.5394188761711121, "learning_rate": 9.123888474009373e-06, "loss": 0.5811, "step": 2552 }, { "epoch": 0.8168728668941979, "grad_norm": 0.5949886441230774, "learning_rate": 9.122835422981458e-06, "loss": 0.6073, "step": 2553 }, { "epoch": 0.8171928327645052, "grad_norm": 0.5322878956794739, "learning_rate": 9.121781800317453e-06, "loss": 0.5135, "step": 2554 }, { "epoch": 0.8175127986348123, "grad_norm": 0.5521470308303833, "learning_rate": 9.120727606163443e-06, "loss": 0.5549, "step": 2555 }, { "epoch": 0.8178327645051194, "grad_norm": 0.5179667472839355, "learning_rate": 9.119672840665592e-06, "loss": 0.5182, "step": 2556 }, { "epoch": 0.8181527303754266, "grad_norm": 0.5399349331855774, "learning_rate": 9.118617503970146e-06, "loss": 0.5662, "step": 2557 }, { "epoch": 0.8184726962457338, "grad_norm": 0.5927430391311646, "learning_rate": 9.117561596223428e-06, "loss": 0.5615, "step": 2558 }, { "epoch": 0.818792662116041, "grad_norm": 0.5932854413986206, "learning_rate": 9.116505117571841e-06, "loss": 0.5695, "step": 2559 }, { "epoch": 0.8191126279863481, "grad_norm": 0.5340254306793213, "learning_rate": 9.115448068161867e-06, "loss": 0.5032, "step": 2560 }, { "epoch": 0.8194325938566553, "grad_norm": 0.5510215759277344, "learning_rate": 9.114390448140068e-06, "loss": 0.5229, "step": 2561 }, { "epoch": 0.8197525597269625, "grad_norm": 0.5657440423965454, "learning_rate": 9.113332257653083e-06, "loss": 0.5447, "step": 2562 }, { "epoch": 0.8200725255972696, "grad_norm": 0.5934768319129944, "learning_rate": 9.112273496847633e-06, "loss": 0.542, "step": 2563 }, { "epoch": 0.8203924914675768, "grad_norm": 0.5292602777481079, "learning_rate": 9.111214165870515e-06, "loss": 0.5284, "step": 2564 }, { "epoch": 0.820712457337884, "grad_norm": 0.6175116896629333, "learning_rate": 9.110154264868607e-06, "loss": 0.5362, "step": 2565 }, { "epoch": 0.8210324232081911, "grad_norm": 0.645007312297821, "learning_rate": 9.109093793988866e-06, "loss": 0.5563, "step": 2566 }, { "epoch": 0.8213523890784983, "grad_norm": 0.5482667088508606, "learning_rate": 9.108032753378326e-06, "loss": 0.5697, "step": 2567 }, { "epoch": 0.8216723549488054, "grad_norm": 0.5520942807197571, "learning_rate": 9.106971143184103e-06, "loss": 0.538, "step": 2568 }, { "epoch": 0.8219923208191127, "grad_norm": 0.6108903884887695, "learning_rate": 9.10590896355339e-06, "loss": 0.5506, "step": 2569 }, { "epoch": 0.8223122866894198, "grad_norm": 0.6232043504714966, "learning_rate": 9.10484621463346e-06, "loss": 0.5474, "step": 2570 }, { "epoch": 0.8226322525597269, "grad_norm": 0.6807348728179932, "learning_rate": 9.103782896571664e-06, "loss": 0.5816, "step": 2571 }, { "epoch": 0.8229522184300341, "grad_norm": 0.5627373456954956, "learning_rate": 9.102719009515433e-06, "loss": 0.5305, "step": 2572 }, { "epoch": 0.8232721843003413, "grad_norm": 0.5594233870506287, "learning_rate": 9.101654553612276e-06, "loss": 0.5761, "step": 2573 }, { "epoch": 0.8235921501706485, "grad_norm": 0.5685380101203918, "learning_rate": 9.100589529009781e-06, "loss": 0.5471, "step": 2574 }, { "epoch": 0.8239121160409556, "grad_norm": 0.5186659693717957, "learning_rate": 9.099523935855612e-06, "loss": 0.5139, "step": 2575 }, { "epoch": 0.8242320819112628, "grad_norm": 0.5025923848152161, "learning_rate": 9.09845777429752e-06, "loss": 0.549, "step": 2576 }, { "epoch": 0.82455204778157, "grad_norm": 0.560330331325531, "learning_rate": 9.097391044483325e-06, "loss": 0.5302, "step": 2577 }, { "epoch": 0.8248720136518771, "grad_norm": 0.5975551009178162, "learning_rate": 9.096323746560934e-06, "loss": 0.5809, "step": 2578 }, { "epoch": 0.8251919795221843, "grad_norm": 0.6584590077400208, "learning_rate": 9.095255880678327e-06, "loss": 0.566, "step": 2579 }, { "epoch": 0.8255119453924915, "grad_norm": 0.5146849751472473, "learning_rate": 9.094187446983567e-06, "loss": 0.5378, "step": 2580 }, { "epoch": 0.8258319112627986, "grad_norm": 0.5368896126747131, "learning_rate": 9.09311844562479e-06, "loss": 0.521, "step": 2581 }, { "epoch": 0.8261518771331058, "grad_norm": 0.5094373822212219, "learning_rate": 9.092048876750217e-06, "loss": 0.5257, "step": 2582 }, { "epoch": 0.8264718430034129, "grad_norm": 0.5936737060546875, "learning_rate": 9.090978740508144e-06, "loss": 0.5851, "step": 2583 }, { "epoch": 0.8267918088737202, "grad_norm": 0.5526201128959656, "learning_rate": 9.089908037046947e-06, "loss": 0.5503, "step": 2584 }, { "epoch": 0.8271117747440273, "grad_norm": 0.49123311042785645, "learning_rate": 9.088836766515081e-06, "loss": 0.5558, "step": 2585 }, { "epoch": 0.8274317406143344, "grad_norm": 0.5265716910362244, "learning_rate": 9.087764929061078e-06, "loss": 0.5121, "step": 2586 }, { "epoch": 0.8277517064846417, "grad_norm": 0.557817280292511, "learning_rate": 9.086692524833551e-06, "loss": 0.5553, "step": 2587 }, { "epoch": 0.8280716723549488, "grad_norm": 0.5713567733764648, "learning_rate": 9.085619553981186e-06, "loss": 0.5635, "step": 2588 }, { "epoch": 0.828391638225256, "grad_norm": 0.522912323474884, "learning_rate": 9.084546016652758e-06, "loss": 0.5343, "step": 2589 }, { "epoch": 0.8287116040955631, "grad_norm": 0.5856422781944275, "learning_rate": 9.08347191299711e-06, "loss": 0.5829, "step": 2590 }, { "epoch": 0.8290315699658704, "grad_norm": 0.49362123012542725, "learning_rate": 9.082397243163167e-06, "loss": 0.5586, "step": 2591 }, { "epoch": 0.8293515358361775, "grad_norm": 0.5722553133964539, "learning_rate": 9.081322007299937e-06, "loss": 0.5474, "step": 2592 }, { "epoch": 0.8296715017064846, "grad_norm": 0.5332806706428528, "learning_rate": 9.0802462055565e-06, "loss": 0.5405, "step": 2593 }, { "epoch": 0.8299914675767918, "grad_norm": 0.6191956400871277, "learning_rate": 9.079169838082018e-06, "loss": 0.5523, "step": 2594 }, { "epoch": 0.830311433447099, "grad_norm": 0.5539155602455139, "learning_rate": 9.07809290502573e-06, "loss": 0.5548, "step": 2595 }, { "epoch": 0.8306313993174061, "grad_norm": 0.5115573406219482, "learning_rate": 9.077015406536957e-06, "loss": 0.5252, "step": 2596 }, { "epoch": 0.8309513651877133, "grad_norm": 0.58519446849823, "learning_rate": 9.075937342765091e-06, "loss": 0.5527, "step": 2597 }, { "epoch": 0.8312713310580204, "grad_norm": 0.5422502756118774, "learning_rate": 9.074858713859609e-06, "loss": 0.5694, "step": 2598 }, { "epoch": 0.8315912969283277, "grad_norm": 0.6382519006729126, "learning_rate": 9.073779519970066e-06, "loss": 0.5818, "step": 2599 }, { "epoch": 0.8319112627986348, "grad_norm": 0.5833666324615479, "learning_rate": 9.072699761246088e-06, "loss": 0.5681, "step": 2600 }, { "epoch": 0.832231228668942, "grad_norm": 0.4834451675415039, "learning_rate": 9.071619437837392e-06, "loss": 0.5375, "step": 2601 }, { "epoch": 0.8325511945392492, "grad_norm": 0.621788740158081, "learning_rate": 9.070538549893762e-06, "loss": 0.5985, "step": 2602 }, { "epoch": 0.8328711604095563, "grad_norm": 0.5683957934379578, "learning_rate": 9.069457097565064e-06, "loss": 0.5245, "step": 2603 }, { "epoch": 0.8331911262798635, "grad_norm": 0.5451465249061584, "learning_rate": 9.068375081001243e-06, "loss": 0.5411, "step": 2604 }, { "epoch": 0.8335110921501706, "grad_norm": 0.5952214002609253, "learning_rate": 9.067292500352325e-06, "loss": 0.5121, "step": 2605 }, { "epoch": 0.8338310580204779, "grad_norm": 0.5726862549781799, "learning_rate": 9.066209355768408e-06, "loss": 0.5588, "step": 2606 }, { "epoch": 0.834151023890785, "grad_norm": 0.5016576051712036, "learning_rate": 9.065125647399671e-06, "loss": 0.5518, "step": 2607 }, { "epoch": 0.8344709897610921, "grad_norm": 0.5270254611968994, "learning_rate": 9.064041375396373e-06, "loss": 0.5436, "step": 2608 }, { "epoch": 0.8347909556313993, "grad_norm": 0.627449095249176, "learning_rate": 9.062956539908849e-06, "loss": 0.5619, "step": 2609 }, { "epoch": 0.8351109215017065, "grad_norm": 0.6327446103096008, "learning_rate": 9.061871141087514e-06, "loss": 0.5936, "step": 2610 }, { "epoch": 0.8354308873720137, "grad_norm": 0.5677685141563416, "learning_rate": 9.060785179082859e-06, "loss": 0.5432, "step": 2611 }, { "epoch": 0.8357508532423208, "grad_norm": 0.6317307949066162, "learning_rate": 9.059698654045452e-06, "loss": 0.5711, "step": 2612 }, { "epoch": 0.836070819112628, "grad_norm": 0.6746708750724792, "learning_rate": 9.058611566125943e-06, "loss": 0.5124, "step": 2613 }, { "epoch": 0.8363907849829352, "grad_norm": 0.5662692189216614, "learning_rate": 9.057523915475058e-06, "loss": 0.522, "step": 2614 }, { "epoch": 0.8367107508532423, "grad_norm": 0.6253296732902527, "learning_rate": 9.056435702243601e-06, "loss": 0.5421, "step": 2615 }, { "epoch": 0.8370307167235495, "grad_norm": 0.6256198287010193, "learning_rate": 9.055346926582453e-06, "loss": 0.5609, "step": 2616 }, { "epoch": 0.8373506825938567, "grad_norm": 0.616001307964325, "learning_rate": 9.054257588642577e-06, "loss": 0.5958, "step": 2617 }, { "epoch": 0.8376706484641638, "grad_norm": 0.576518714427948, "learning_rate": 9.053167688575007e-06, "loss": 0.5546, "step": 2618 }, { "epoch": 0.837990614334471, "grad_norm": 0.5338309407234192, "learning_rate": 9.052077226530862e-06, "loss": 0.5613, "step": 2619 }, { "epoch": 0.8383105802047781, "grad_norm": 0.5424176454544067, "learning_rate": 9.050986202661334e-06, "loss": 0.5502, "step": 2620 }, { "epoch": 0.8386305460750854, "grad_norm": 0.5512865781784058, "learning_rate": 9.049894617117697e-06, "loss": 0.4723, "step": 2621 }, { "epoch": 0.8389505119453925, "grad_norm": 0.5608572363853455, "learning_rate": 9.048802470051296e-06, "loss": 0.5575, "step": 2622 }, { "epoch": 0.8392704778156996, "grad_norm": 0.5869863033294678, "learning_rate": 9.047709761613565e-06, "loss": 0.5445, "step": 2623 }, { "epoch": 0.8395904436860068, "grad_norm": 0.5191389918327332, "learning_rate": 9.046616491956003e-06, "loss": 0.5522, "step": 2624 }, { "epoch": 0.839910409556314, "grad_norm": 0.5765413045883179, "learning_rate": 9.045522661230199e-06, "loss": 0.5665, "step": 2625 }, { "epoch": 0.8402303754266212, "grad_norm": 0.6026390194892883, "learning_rate": 9.044428269587811e-06, "loss": 0.6141, "step": 2626 }, { "epoch": 0.8405503412969283, "grad_norm": 0.6081013083457947, "learning_rate": 9.043333317180576e-06, "loss": 0.5959, "step": 2627 }, { "epoch": 0.8408703071672355, "grad_norm": 0.5654779076576233, "learning_rate": 9.042237804160313e-06, "loss": 0.5416, "step": 2628 }, { "epoch": 0.8411902730375427, "grad_norm": 0.5690180659294128, "learning_rate": 9.041141730678916e-06, "loss": 0.5304, "step": 2629 }, { "epoch": 0.8415102389078498, "grad_norm": 0.5598645210266113, "learning_rate": 9.040045096888356e-06, "loss": 0.5438, "step": 2630 }, { "epoch": 0.841830204778157, "grad_norm": 0.5394770503044128, "learning_rate": 9.038947902940684e-06, "loss": 0.5373, "step": 2631 }, { "epoch": 0.8421501706484642, "grad_norm": 0.6138805150985718, "learning_rate": 9.037850148988025e-06, "loss": 0.5637, "step": 2632 }, { "epoch": 0.8424701365187713, "grad_norm": 0.5667594075202942, "learning_rate": 9.036751835182586e-06, "loss": 0.513, "step": 2633 }, { "epoch": 0.8427901023890785, "grad_norm": 0.5401478409767151, "learning_rate": 9.035652961676649e-06, "loss": 0.5349, "step": 2634 }, { "epoch": 0.8431100682593856, "grad_norm": 0.5945031642913818, "learning_rate": 9.034553528622572e-06, "loss": 0.5904, "step": 2635 }, { "epoch": 0.8434300341296929, "grad_norm": 0.5829552412033081, "learning_rate": 9.033453536172796e-06, "loss": 0.5686, "step": 2636 }, { "epoch": 0.84375, "grad_norm": 0.5223750472068787, "learning_rate": 9.032352984479836e-06, "loss": 0.5471, "step": 2637 }, { "epoch": 0.8440699658703071, "grad_norm": 0.5635274052619934, "learning_rate": 9.031251873696281e-06, "loss": 0.6, "step": 2638 }, { "epoch": 0.8443899317406144, "grad_norm": 0.4674994647502899, "learning_rate": 9.030150203974806e-06, "loss": 0.5353, "step": 2639 }, { "epoch": 0.8447098976109215, "grad_norm": 0.5379506945610046, "learning_rate": 9.029047975468156e-06, "loss": 0.5438, "step": 2640 }, { "epoch": 0.8450298634812287, "grad_norm": 0.5450134873390198, "learning_rate": 9.027945188329157e-06, "loss": 0.5392, "step": 2641 }, { "epoch": 0.8453498293515358, "grad_norm": 0.507026731967926, "learning_rate": 9.026841842710711e-06, "loss": 0.5031, "step": 2642 }, { "epoch": 0.845669795221843, "grad_norm": 0.5834566950798035, "learning_rate": 9.025737938765803e-06, "loss": 0.5483, "step": 2643 }, { "epoch": 0.8459897610921502, "grad_norm": 0.5272031426429749, "learning_rate": 9.024633476647483e-06, "loss": 0.5276, "step": 2644 }, { "epoch": 0.8463097269624573, "grad_norm": 0.4700632691383362, "learning_rate": 9.02352845650889e-06, "loss": 0.4892, "step": 2645 }, { "epoch": 0.8466296928327645, "grad_norm": 0.5325533747673035, "learning_rate": 9.022422878503237e-06, "loss": 0.5452, "step": 2646 }, { "epoch": 0.8469496587030717, "grad_norm": 0.6028921008110046, "learning_rate": 9.021316742783813e-06, "loss": 0.562, "step": 2647 }, { "epoch": 0.8472696245733788, "grad_norm": 0.5505808591842651, "learning_rate": 9.020210049503985e-06, "loss": 0.5695, "step": 2648 }, { "epoch": 0.847589590443686, "grad_norm": 0.5052388906478882, "learning_rate": 9.019102798817196e-06, "loss": 0.5528, "step": 2649 }, { "epoch": 0.8479095563139932, "grad_norm": 0.5818324685096741, "learning_rate": 9.017994990876971e-06, "loss": 0.5459, "step": 2650 }, { "epoch": 0.8482295221843004, "grad_norm": 0.6578103303909302, "learning_rate": 9.016886625836905e-06, "loss": 0.5379, "step": 2651 }, { "epoch": 0.8485494880546075, "grad_norm": 0.5725463628768921, "learning_rate": 9.015777703850675e-06, "loss": 0.5805, "step": 2652 }, { "epoch": 0.8488694539249146, "grad_norm": 0.5812577605247498, "learning_rate": 9.014668225072036e-06, "loss": 0.5674, "step": 2653 }, { "epoch": 0.8491894197952219, "grad_norm": 0.5736384987831116, "learning_rate": 9.013558189654819e-06, "loss": 0.5737, "step": 2654 }, { "epoch": 0.849509385665529, "grad_norm": 0.5640150904655457, "learning_rate": 9.01244759775293e-06, "loss": 0.5291, "step": 2655 }, { "epoch": 0.8498293515358362, "grad_norm": 0.5893073081970215, "learning_rate": 9.011336449520354e-06, "loss": 0.5637, "step": 2656 }, { "epoch": 0.8501493174061433, "grad_norm": 0.566746175289154, "learning_rate": 9.010224745111153e-06, "loss": 0.5164, "step": 2657 }, { "epoch": 0.8504692832764505, "grad_norm": 0.6648749113082886, "learning_rate": 9.009112484679468e-06, "loss": 0.5534, "step": 2658 }, { "epoch": 0.8507892491467577, "grad_norm": 0.5884232521057129, "learning_rate": 9.007999668379514e-06, "loss": 0.5423, "step": 2659 }, { "epoch": 0.8511092150170648, "grad_norm": 0.6019272208213806, "learning_rate": 9.006886296365585e-06, "loss": 0.5717, "step": 2660 }, { "epoch": 0.851429180887372, "grad_norm": 0.5348987579345703, "learning_rate": 9.005772368792048e-06, "loss": 0.5093, "step": 2661 }, { "epoch": 0.8517491467576792, "grad_norm": 0.6644459962844849, "learning_rate": 9.004657885813354e-06, "loss": 0.54, "step": 2662 }, { "epoch": 0.8520691126279863, "grad_norm": 0.5964223146438599, "learning_rate": 9.003542847584026e-06, "loss": 0.5232, "step": 2663 }, { "epoch": 0.8523890784982935, "grad_norm": 0.5980588793754578, "learning_rate": 9.002427254258668e-06, "loss": 0.5612, "step": 2664 }, { "epoch": 0.8527090443686007, "grad_norm": 0.5639047026634216, "learning_rate": 9.001311105991955e-06, "loss": 0.5391, "step": 2665 }, { "epoch": 0.8530290102389079, "grad_norm": 0.5928334593772888, "learning_rate": 9.000194402938643e-06, "loss": 0.5579, "step": 2666 }, { "epoch": 0.853348976109215, "grad_norm": 0.6095545291900635, "learning_rate": 8.999077145253564e-06, "loss": 0.569, "step": 2667 }, { "epoch": 0.8536689419795221, "grad_norm": 0.6068754196166992, "learning_rate": 8.997959333091628e-06, "loss": 0.5654, "step": 2668 }, { "epoch": 0.8539889078498294, "grad_norm": 0.5735090970993042, "learning_rate": 8.996840966607822e-06, "loss": 0.5874, "step": 2669 }, { "epoch": 0.8543088737201365, "grad_norm": 0.5521860718727112, "learning_rate": 8.995722045957207e-06, "loss": 0.541, "step": 2670 }, { "epoch": 0.8546288395904437, "grad_norm": 0.5921656489372253, "learning_rate": 8.994602571294923e-06, "loss": 0.6061, "step": 2671 }, { "epoch": 0.8549488054607508, "grad_norm": 0.4652414619922638, "learning_rate": 8.993482542776189e-06, "loss": 0.5335, "step": 2672 }, { "epoch": 0.855268771331058, "grad_norm": 0.6044648885726929, "learning_rate": 8.992361960556293e-06, "loss": 0.581, "step": 2673 }, { "epoch": 0.8555887372013652, "grad_norm": 0.5695050954818726, "learning_rate": 8.99124082479061e-06, "loss": 0.5629, "step": 2674 }, { "epoch": 0.8559087030716723, "grad_norm": 0.5469666719436646, "learning_rate": 8.990119135634585e-06, "loss": 0.5034, "step": 2675 }, { "epoch": 0.8562286689419796, "grad_norm": 0.5787575244903564, "learning_rate": 8.988996893243742e-06, "loss": 0.5616, "step": 2676 }, { "epoch": 0.8565486348122867, "grad_norm": 0.56728196144104, "learning_rate": 8.98787409777368e-06, "loss": 0.5243, "step": 2677 }, { "epoch": 0.8568686006825939, "grad_norm": 0.5428320169448853, "learning_rate": 8.986750749380077e-06, "loss": 0.4988, "step": 2678 }, { "epoch": 0.857188566552901, "grad_norm": 0.5656844973564148, "learning_rate": 8.985626848218685e-06, "loss": 0.5216, "step": 2679 }, { "epoch": 0.8575085324232082, "grad_norm": 0.5635480880737305, "learning_rate": 8.984502394445338e-06, "loss": 0.5733, "step": 2680 }, { "epoch": 0.8578284982935154, "grad_norm": 0.6129047870635986, "learning_rate": 8.98337738821594e-06, "loss": 0.5393, "step": 2681 }, { "epoch": 0.8581484641638225, "grad_norm": 0.5660448670387268, "learning_rate": 8.982251829686474e-06, "loss": 0.5313, "step": 2682 }, { "epoch": 0.8584684300341296, "grad_norm": 0.5053395628929138, "learning_rate": 8.981125719013002e-06, "loss": 0.5444, "step": 2683 }, { "epoch": 0.8587883959044369, "grad_norm": 0.5126214027404785, "learning_rate": 8.97999905635166e-06, "loss": 0.5096, "step": 2684 }, { "epoch": 0.859108361774744, "grad_norm": 0.5766167044639587, "learning_rate": 8.97887184185866e-06, "loss": 0.5686, "step": 2685 }, { "epoch": 0.8594283276450512, "grad_norm": 0.5058866739273071, "learning_rate": 8.977744075690292e-06, "loss": 0.5559, "step": 2686 }, { "epoch": 0.8597482935153583, "grad_norm": 0.5399185419082642, "learning_rate": 8.976615758002924e-06, "loss": 0.5541, "step": 2687 }, { "epoch": 0.8600682593856656, "grad_norm": 0.5576114058494568, "learning_rate": 8.975486888952995e-06, "loss": 0.5613, "step": 2688 }, { "epoch": 0.8603882252559727, "grad_norm": 0.5862126350402832, "learning_rate": 8.974357468697029e-06, "loss": 0.5576, "step": 2689 }, { "epoch": 0.8607081911262798, "grad_norm": 0.5256986021995544, "learning_rate": 8.973227497391617e-06, "loss": 0.5533, "step": 2690 }, { "epoch": 0.8610281569965871, "grad_norm": 0.5526366829872131, "learning_rate": 8.972096975193432e-06, "loss": 0.5787, "step": 2691 }, { "epoch": 0.8613481228668942, "grad_norm": 0.521307110786438, "learning_rate": 8.970965902259225e-06, "loss": 0.5153, "step": 2692 }, { "epoch": 0.8616680887372014, "grad_norm": 0.5446743369102478, "learning_rate": 8.969834278745817e-06, "loss": 0.5052, "step": 2693 }, { "epoch": 0.8619880546075085, "grad_norm": 0.5722483396530151, "learning_rate": 8.96870210481011e-06, "loss": 0.5212, "step": 2694 }, { "epoch": 0.8623080204778157, "grad_norm": 0.6217164397239685, "learning_rate": 8.967569380609082e-06, "loss": 0.5632, "step": 2695 }, { "epoch": 0.8626279863481229, "grad_norm": 0.5332825183868408, "learning_rate": 8.966436106299787e-06, "loss": 0.5366, "step": 2696 }, { "epoch": 0.86294795221843, "grad_norm": 0.637496829032898, "learning_rate": 8.965302282039355e-06, "loss": 0.5572, "step": 2697 }, { "epoch": 0.8632679180887372, "grad_norm": 0.5120847821235657, "learning_rate": 8.964167907984989e-06, "loss": 0.5332, "step": 2698 }, { "epoch": 0.8635878839590444, "grad_norm": 0.5647338628768921, "learning_rate": 8.963032984293974e-06, "loss": 0.5273, "step": 2699 }, { "epoch": 0.8639078498293515, "grad_norm": 0.5611434578895569, "learning_rate": 8.961897511123668e-06, "loss": 0.5539, "step": 2700 }, { "epoch": 0.8642278156996587, "grad_norm": 0.5692500472068787, "learning_rate": 8.960761488631507e-06, "loss": 0.5587, "step": 2701 }, { "epoch": 0.8645477815699659, "grad_norm": 0.5833654999732971, "learning_rate": 8.959624916975001e-06, "loss": 0.5765, "step": 2702 }, { "epoch": 0.8648677474402731, "grad_norm": 0.5490713715553284, "learning_rate": 8.958487796311735e-06, "loss": 0.5488, "step": 2703 }, { "epoch": 0.8651877133105802, "grad_norm": 0.5373678803443909, "learning_rate": 8.957350126799373e-06, "loss": 0.5698, "step": 2704 }, { "epoch": 0.8655076791808873, "grad_norm": 0.514848530292511, "learning_rate": 8.956211908595658e-06, "loss": 0.5174, "step": 2705 }, { "epoch": 0.8658276450511946, "grad_norm": 0.6037289500236511, "learning_rate": 8.955073141858401e-06, "loss": 0.5398, "step": 2706 }, { "epoch": 0.8661476109215017, "grad_norm": 0.5680069327354431, "learning_rate": 8.953933826745494e-06, "loss": 0.5443, "step": 2707 }, { "epoch": 0.8664675767918089, "grad_norm": 0.5605342984199524, "learning_rate": 8.952793963414908e-06, "loss": 0.5377, "step": 2708 }, { "epoch": 0.866787542662116, "grad_norm": 0.5577074885368347, "learning_rate": 8.951653552024681e-06, "loss": 0.5356, "step": 2709 }, { "epoch": 0.8671075085324232, "grad_norm": 0.623981237411499, "learning_rate": 8.950512592732934e-06, "loss": 0.6175, "step": 2710 }, { "epoch": 0.8674274744027304, "grad_norm": 0.5811691880226135, "learning_rate": 8.949371085697867e-06, "loss": 0.5212, "step": 2711 }, { "epoch": 0.8677474402730375, "grad_norm": 0.5302852392196655, "learning_rate": 8.948229031077747e-06, "loss": 0.5181, "step": 2712 }, { "epoch": 0.8680674061433447, "grad_norm": 0.49263468384742737, "learning_rate": 8.94708642903092e-06, "loss": 0.5354, "step": 2713 }, { "epoch": 0.8683873720136519, "grad_norm": 0.5899059176445007, "learning_rate": 8.945943279715812e-06, "loss": 0.5281, "step": 2714 }, { "epoch": 0.868707337883959, "grad_norm": 0.552376389503479, "learning_rate": 8.944799583290921e-06, "loss": 0.5609, "step": 2715 }, { "epoch": 0.8690273037542662, "grad_norm": 0.6187187433242798, "learning_rate": 8.943655339914822e-06, "loss": 0.5563, "step": 2716 }, { "epoch": 0.8693472696245734, "grad_norm": 0.5890134572982788, "learning_rate": 8.942510549746167e-06, "loss": 0.553, "step": 2717 }, { "epoch": 0.8696672354948806, "grad_norm": 0.513079047203064, "learning_rate": 8.94136521294368e-06, "loss": 0.5575, "step": 2718 }, { "epoch": 0.8699872013651877, "grad_norm": 0.6054332852363586, "learning_rate": 8.940219329666167e-06, "loss": 0.581, "step": 2719 }, { "epoch": 0.8703071672354948, "grad_norm": 0.6034533381462097, "learning_rate": 8.939072900072501e-06, "loss": 0.5318, "step": 2720 }, { "epoch": 0.8706271331058021, "grad_norm": 0.5352435111999512, "learning_rate": 8.93792592432164e-06, "loss": 0.536, "step": 2721 }, { "epoch": 0.8709470989761092, "grad_norm": 0.5317062735557556, "learning_rate": 8.936778402572612e-06, "loss": 0.5402, "step": 2722 }, { "epoch": 0.8712670648464164, "grad_norm": 0.568576455116272, "learning_rate": 8.935630334984522e-06, "loss": 0.554, "step": 2723 }, { "epoch": 0.8715870307167235, "grad_norm": 0.6381053924560547, "learning_rate": 8.93448172171655e-06, "loss": 0.5739, "step": 2724 }, { "epoch": 0.8719069965870307, "grad_norm": 0.5327437520027161, "learning_rate": 8.933332562927957e-06, "loss": 0.5542, "step": 2725 }, { "epoch": 0.8722269624573379, "grad_norm": 0.55152827501297, "learning_rate": 8.932182858778068e-06, "loss": 0.5835, "step": 2726 }, { "epoch": 0.872546928327645, "grad_norm": 0.47426605224609375, "learning_rate": 8.9310326094263e-06, "loss": 0.5204, "step": 2727 }, { "epoch": 0.8728668941979523, "grad_norm": 0.6053329706192017, "learning_rate": 8.92988181503213e-06, "loss": 0.5297, "step": 2728 }, { "epoch": 0.8731868600682594, "grad_norm": 0.5693221092224121, "learning_rate": 8.928730475755117e-06, "loss": 0.5053, "step": 2729 }, { "epoch": 0.8735068259385665, "grad_norm": 0.5186700820922852, "learning_rate": 8.927578591754897e-06, "loss": 0.5076, "step": 2730 }, { "epoch": 0.8738267918088737, "grad_norm": 0.6362303495407104, "learning_rate": 8.926426163191181e-06, "loss": 0.5471, "step": 2731 }, { "epoch": 0.8741467576791809, "grad_norm": 0.5959476232528687, "learning_rate": 8.925273190223756e-06, "loss": 0.5924, "step": 2732 }, { "epoch": 0.8744667235494881, "grad_norm": 0.6008637547492981, "learning_rate": 8.92411967301248e-06, "loss": 0.5632, "step": 2733 }, { "epoch": 0.8747866894197952, "grad_norm": 0.5581860542297363, "learning_rate": 8.92296561171729e-06, "loss": 0.5579, "step": 2734 }, { "epoch": 0.8751066552901023, "grad_norm": 0.5555343627929688, "learning_rate": 8.921811006498196e-06, "loss": 0.5167, "step": 2735 }, { "epoch": 0.8754266211604096, "grad_norm": 0.5411742329597473, "learning_rate": 8.92065585751529e-06, "loss": 0.5516, "step": 2736 }, { "epoch": 0.8757465870307167, "grad_norm": 0.577962338924408, "learning_rate": 8.919500164928734e-06, "loss": 0.5955, "step": 2737 }, { "epoch": 0.8760665529010239, "grad_norm": 0.5593159794807434, "learning_rate": 8.918343928898763e-06, "loss": 0.5599, "step": 2738 }, { "epoch": 0.8763865187713311, "grad_norm": 0.5594756007194519, "learning_rate": 8.917187149585693e-06, "loss": 0.5772, "step": 2739 }, { "epoch": 0.8767064846416383, "grad_norm": 0.5813189744949341, "learning_rate": 8.916029827149913e-06, "loss": 0.5239, "step": 2740 }, { "epoch": 0.8770264505119454, "grad_norm": 0.47838127613067627, "learning_rate": 8.914871961751885e-06, "loss": 0.5125, "step": 2741 }, { "epoch": 0.8773464163822525, "grad_norm": 0.6236173510551453, "learning_rate": 8.91371355355215e-06, "loss": 0.5797, "step": 2742 }, { "epoch": 0.8776663822525598, "grad_norm": 0.5655630826950073, "learning_rate": 8.912554602711322e-06, "loss": 0.5769, "step": 2743 }, { "epoch": 0.8779863481228669, "grad_norm": 0.5529646873474121, "learning_rate": 8.911395109390092e-06, "loss": 0.5223, "step": 2744 }, { "epoch": 0.878306313993174, "grad_norm": 0.6270757913589478, "learning_rate": 8.910235073749226e-06, "loss": 0.5334, "step": 2745 }, { "epoch": 0.8786262798634812, "grad_norm": 0.6319167613983154, "learning_rate": 8.909074495949562e-06, "loss": 0.5783, "step": 2746 }, { "epoch": 0.8789462457337884, "grad_norm": 0.6389390826225281, "learning_rate": 8.907913376152016e-06, "loss": 0.5596, "step": 2747 }, { "epoch": 0.8792662116040956, "grad_norm": 0.5657618045806885, "learning_rate": 8.90675171451758e-06, "loss": 0.5306, "step": 2748 }, { "epoch": 0.8795861774744027, "grad_norm": 0.5682520270347595, "learning_rate": 8.905589511207318e-06, "loss": 0.5704, "step": 2749 }, { "epoch": 0.8799061433447098, "grad_norm": 0.499196320772171, "learning_rate": 8.904426766382372e-06, "loss": 0.4701, "step": 2750 }, { "epoch": 0.8802261092150171, "grad_norm": 0.5992483496665955, "learning_rate": 8.903263480203956e-06, "loss": 0.5276, "step": 2751 }, { "epoch": 0.8805460750853242, "grad_norm": 0.5560103058815002, "learning_rate": 8.902099652833364e-06, "loss": 0.5456, "step": 2752 }, { "epoch": 0.8808660409556314, "grad_norm": 0.5642792582511902, "learning_rate": 8.900935284431962e-06, "loss": 0.5321, "step": 2753 }, { "epoch": 0.8811860068259386, "grad_norm": 0.516201376914978, "learning_rate": 8.89977037516119e-06, "loss": 0.55, "step": 2754 }, { "epoch": 0.8815059726962458, "grad_norm": 0.5581544041633606, "learning_rate": 8.89860492518256e-06, "loss": 0.5317, "step": 2755 }, { "epoch": 0.8818259385665529, "grad_norm": 0.5785426497459412, "learning_rate": 8.897438934657672e-06, "loss": 0.5578, "step": 2756 }, { "epoch": 0.88214590443686, "grad_norm": 0.5539271831512451, "learning_rate": 8.896272403748185e-06, "loss": 0.5888, "step": 2757 }, { "epoch": 0.8824658703071673, "grad_norm": 0.5707237124443054, "learning_rate": 8.895105332615841e-06, "loss": 0.5474, "step": 2758 }, { "epoch": 0.8827858361774744, "grad_norm": 0.5254892706871033, "learning_rate": 8.893937721422458e-06, "loss": 0.5439, "step": 2759 }, { "epoch": 0.8831058020477816, "grad_norm": 0.5878953337669373, "learning_rate": 8.892769570329924e-06, "loss": 0.5552, "step": 2760 }, { "epoch": 0.8834257679180887, "grad_norm": 0.5976927280426025, "learning_rate": 8.891600879500206e-06, "loss": 0.5769, "step": 2761 }, { "epoch": 0.8837457337883959, "grad_norm": 0.5668741464614868, "learning_rate": 8.890431649095346e-06, "loss": 0.5564, "step": 2762 }, { "epoch": 0.8840656996587031, "grad_norm": 0.5591926574707031, "learning_rate": 8.889261879277458e-06, "loss": 0.5558, "step": 2763 }, { "epoch": 0.8843856655290102, "grad_norm": 0.5671588182449341, "learning_rate": 8.88809157020873e-06, "loss": 0.5211, "step": 2764 }, { "epoch": 0.8847056313993175, "grad_norm": 0.5909327864646912, "learning_rate": 8.886920722051428e-06, "loss": 0.5776, "step": 2765 }, { "epoch": 0.8850255972696246, "grad_norm": 0.6187429428100586, "learning_rate": 8.885749334967893e-06, "loss": 0.5325, "step": 2766 }, { "epoch": 0.8853455631399317, "grad_norm": 0.5392321348190308, "learning_rate": 8.884577409120535e-06, "loss": 0.5435, "step": 2767 }, { "epoch": 0.8856655290102389, "grad_norm": 0.5093667507171631, "learning_rate": 8.88340494467185e-06, "loss": 0.5234, "step": 2768 }, { "epoch": 0.8859854948805461, "grad_norm": 0.6120261549949646, "learning_rate": 8.882231941784397e-06, "loss": 0.5452, "step": 2769 }, { "epoch": 0.8863054607508533, "grad_norm": 0.5771980285644531, "learning_rate": 8.881058400620813e-06, "loss": 0.5603, "step": 2770 }, { "epoch": 0.8866254266211604, "grad_norm": 0.5893442034721375, "learning_rate": 8.879884321343813e-06, "loss": 0.545, "step": 2771 }, { "epoch": 0.8869453924914675, "grad_norm": 0.52390456199646, "learning_rate": 8.878709704116185e-06, "loss": 0.5536, "step": 2772 }, { "epoch": 0.8872653583617748, "grad_norm": 0.5993599891662598, "learning_rate": 8.877534549100791e-06, "loss": 0.5893, "step": 2773 }, { "epoch": 0.8875853242320819, "grad_norm": 0.49362796545028687, "learning_rate": 8.876358856460565e-06, "loss": 0.5173, "step": 2774 }, { "epoch": 0.887905290102389, "grad_norm": 0.5340424180030823, "learning_rate": 8.87518262635852e-06, "loss": 0.5477, "step": 2775 }, { "epoch": 0.8882252559726962, "grad_norm": 0.5215820670127869, "learning_rate": 8.874005858957743e-06, "loss": 0.5443, "step": 2776 }, { "epoch": 0.8885452218430034, "grad_norm": 0.5687538385391235, "learning_rate": 8.872828554421393e-06, "loss": 0.579, "step": 2777 }, { "epoch": 0.8888651877133106, "grad_norm": 0.5075172185897827, "learning_rate": 8.871650712912705e-06, "loss": 0.5353, "step": 2778 }, { "epoch": 0.8891851535836177, "grad_norm": 0.5134161114692688, "learning_rate": 8.870472334594988e-06, "loss": 0.5732, "step": 2779 }, { "epoch": 0.889505119453925, "grad_norm": 0.53742516040802, "learning_rate": 8.869293419631626e-06, "loss": 0.5513, "step": 2780 }, { "epoch": 0.8898250853242321, "grad_norm": 0.5549346804618835, "learning_rate": 8.868113968186077e-06, "loss": 0.5563, "step": 2781 }, { "epoch": 0.8901450511945392, "grad_norm": 0.5468842387199402, "learning_rate": 8.86693398042187e-06, "loss": 0.5268, "step": 2782 }, { "epoch": 0.8904650170648464, "grad_norm": 0.5684927105903625, "learning_rate": 8.865753456502617e-06, "loss": 0.5227, "step": 2783 }, { "epoch": 0.8907849829351536, "grad_norm": 0.6799585819244385, "learning_rate": 8.864572396591996e-06, "loss": 0.5657, "step": 2784 }, { "epoch": 0.8911049488054608, "grad_norm": 0.5659300684928894, "learning_rate": 8.863390800853765e-06, "loss": 0.5481, "step": 2785 }, { "epoch": 0.8914249146757679, "grad_norm": 0.5571771860122681, "learning_rate": 8.862208669451748e-06, "loss": 0.5719, "step": 2786 }, { "epoch": 0.891744880546075, "grad_norm": 0.6032619476318359, "learning_rate": 8.861026002549855e-06, "loss": 0.5787, "step": 2787 }, { "epoch": 0.8920648464163823, "grad_norm": 0.5702298283576965, "learning_rate": 8.85984280031206e-06, "loss": 0.5556, "step": 2788 }, { "epoch": 0.8923848122866894, "grad_norm": 0.5752246975898743, "learning_rate": 8.85865906290242e-06, "loss": 0.5418, "step": 2789 }, { "epoch": 0.8927047781569966, "grad_norm": 0.6689168214797974, "learning_rate": 8.857474790485057e-06, "loss": 0.5781, "step": 2790 }, { "epoch": 0.8930247440273038, "grad_norm": 0.5612762570381165, "learning_rate": 8.856289983224173e-06, "loss": 0.569, "step": 2791 }, { "epoch": 0.893344709897611, "grad_norm": 0.5860694050788879, "learning_rate": 8.855104641284045e-06, "loss": 0.5601, "step": 2792 }, { "epoch": 0.8936646757679181, "grad_norm": 0.5637890100479126, "learning_rate": 8.853918764829019e-06, "loss": 0.5481, "step": 2793 }, { "epoch": 0.8939846416382252, "grad_norm": 0.6216111779212952, "learning_rate": 8.85273235402352e-06, "loss": 0.5817, "step": 2794 }, { "epoch": 0.8943046075085325, "grad_norm": 0.5583057999610901, "learning_rate": 8.851545409032047e-06, "loss": 0.5643, "step": 2795 }, { "epoch": 0.8946245733788396, "grad_norm": 0.5592483878135681, "learning_rate": 8.850357930019169e-06, "loss": 0.551, "step": 2796 }, { "epoch": 0.8949445392491467, "grad_norm": 0.5292885303497314, "learning_rate": 8.849169917149532e-06, "loss": 0.5708, "step": 2797 }, { "epoch": 0.8952645051194539, "grad_norm": 0.5390934944152832, "learning_rate": 8.847981370587855e-06, "loss": 0.5364, "step": 2798 }, { "epoch": 0.8955844709897611, "grad_norm": 0.577752411365509, "learning_rate": 8.84679229049893e-06, "loss": 0.5575, "step": 2799 }, { "epoch": 0.8959044368600683, "grad_norm": 0.5969122052192688, "learning_rate": 8.845602677047629e-06, "loss": 0.5527, "step": 2800 }, { "epoch": 0.8962244027303754, "grad_norm": 0.5382528305053711, "learning_rate": 8.84441253039889e-06, "loss": 0.5714, "step": 2801 }, { "epoch": 0.8965443686006825, "grad_norm": 0.5916634798049927, "learning_rate": 8.843221850717726e-06, "loss": 0.5521, "step": 2802 }, { "epoch": 0.8968643344709898, "grad_norm": 0.554932177066803, "learning_rate": 8.842030638169232e-06, "loss": 0.5056, "step": 2803 }, { "epoch": 0.8971843003412969, "grad_norm": 0.5452033281326294, "learning_rate": 8.840838892918568e-06, "loss": 0.5701, "step": 2804 }, { "epoch": 0.8975042662116041, "grad_norm": 0.56942218542099, "learning_rate": 8.83964661513097e-06, "loss": 0.5709, "step": 2805 }, { "epoch": 0.8978242320819113, "grad_norm": 0.5929978489875793, "learning_rate": 8.838453804971747e-06, "loss": 0.5407, "step": 2806 }, { "epoch": 0.8981441979522184, "grad_norm": 0.6034539341926575, "learning_rate": 8.83726046260629e-06, "loss": 0.5651, "step": 2807 }, { "epoch": 0.8984641638225256, "grad_norm": 0.6311516165733337, "learning_rate": 8.836066588200052e-06, "loss": 0.5282, "step": 2808 }, { "epoch": 0.8987841296928327, "grad_norm": 0.5441495180130005, "learning_rate": 8.834872181918566e-06, "loss": 0.549, "step": 2809 }, { "epoch": 0.89910409556314, "grad_norm": 0.5885206460952759, "learning_rate": 8.833677243927439e-06, "loss": 0.5435, "step": 2810 }, { "epoch": 0.8994240614334471, "grad_norm": 0.5274768471717834, "learning_rate": 8.832481774392349e-06, "loss": 0.5029, "step": 2811 }, { "epoch": 0.8997440273037542, "grad_norm": 0.5387116074562073, "learning_rate": 8.831285773479053e-06, "loss": 0.5111, "step": 2812 }, { "epoch": 0.9000639931740614, "grad_norm": 0.5455957651138306, "learning_rate": 8.830089241353375e-06, "loss": 0.5528, "step": 2813 }, { "epoch": 0.9003839590443686, "grad_norm": 0.6203275918960571, "learning_rate": 8.828892178181215e-06, "loss": 0.5325, "step": 2814 }, { "epoch": 0.9007039249146758, "grad_norm": 0.5818467736244202, "learning_rate": 8.827694584128547e-06, "loss": 0.5351, "step": 2815 }, { "epoch": 0.9010238907849829, "grad_norm": 0.5505111217498779, "learning_rate": 8.82649645936142e-06, "loss": 0.5425, "step": 2816 }, { "epoch": 0.9013438566552902, "grad_norm": 0.6561931371688843, "learning_rate": 8.825297804045958e-06, "loss": 0.6024, "step": 2817 }, { "epoch": 0.9016638225255973, "grad_norm": 0.5352081060409546, "learning_rate": 8.824098618348353e-06, "loss": 0.5039, "step": 2818 }, { "epoch": 0.9019837883959044, "grad_norm": 0.5815563797950745, "learning_rate": 8.822898902434873e-06, "loss": 0.5709, "step": 2819 }, { "epoch": 0.9023037542662116, "grad_norm": 0.5135438442230225, "learning_rate": 8.821698656471863e-06, "loss": 0.5081, "step": 2820 }, { "epoch": 0.9026237201365188, "grad_norm": 0.6108431816101074, "learning_rate": 8.820497880625733e-06, "loss": 0.5468, "step": 2821 }, { "epoch": 0.902943686006826, "grad_norm": 0.5745890140533447, "learning_rate": 8.819296575062978e-06, "loss": 0.5556, "step": 2822 }, { "epoch": 0.9032636518771331, "grad_norm": 0.5635342001914978, "learning_rate": 8.818094739950157e-06, "loss": 0.524, "step": 2823 }, { "epoch": 0.9035836177474402, "grad_norm": 0.5868473649024963, "learning_rate": 8.81689237545391e-06, "loss": 0.543, "step": 2824 }, { "epoch": 0.9039035836177475, "grad_norm": 0.6234641671180725, "learning_rate": 8.815689481740942e-06, "loss": 0.5314, "step": 2825 }, { "epoch": 0.9042235494880546, "grad_norm": 0.6335117816925049, "learning_rate": 8.814486058978035e-06, "loss": 0.545, "step": 2826 }, { "epoch": 0.9045435153583617, "grad_norm": 0.5347116589546204, "learning_rate": 8.81328210733205e-06, "loss": 0.5594, "step": 2827 }, { "epoch": 0.9048634812286689, "grad_norm": 0.560544490814209, "learning_rate": 8.81207762696991e-06, "loss": 0.5276, "step": 2828 }, { "epoch": 0.9051834470989761, "grad_norm": 0.5208972692489624, "learning_rate": 8.810872618058622e-06, "loss": 0.5337, "step": 2829 }, { "epoch": 0.9055034129692833, "grad_norm": 0.5120471715927124, "learning_rate": 8.809667080765262e-06, "loss": 0.5361, "step": 2830 }, { "epoch": 0.9058233788395904, "grad_norm": 0.551671028137207, "learning_rate": 8.808461015256976e-06, "loss": 0.5768, "step": 2831 }, { "epoch": 0.9061433447098977, "grad_norm": 0.6156903505325317, "learning_rate": 8.807254421700991e-06, "loss": 0.521, "step": 2832 }, { "epoch": 0.9064633105802048, "grad_norm": 0.5635914206504822, "learning_rate": 8.8060473002646e-06, "loss": 0.5314, "step": 2833 }, { "epoch": 0.9067832764505119, "grad_norm": 0.5528858304023743, "learning_rate": 8.804839651115171e-06, "loss": 0.547, "step": 2834 }, { "epoch": 0.9071032423208191, "grad_norm": 0.5680487751960754, "learning_rate": 8.803631474420146e-06, "loss": 0.5885, "step": 2835 }, { "epoch": 0.9074232081911263, "grad_norm": 0.6367501616477966, "learning_rate": 8.802422770347044e-06, "loss": 0.6022, "step": 2836 }, { "epoch": 0.9077431740614335, "grad_norm": 0.5560848712921143, "learning_rate": 8.801213539063448e-06, "loss": 0.5123, "step": 2837 }, { "epoch": 0.9080631399317406, "grad_norm": 0.5595853328704834, "learning_rate": 8.800003780737024e-06, "loss": 0.5648, "step": 2838 }, { "epoch": 0.9083831058020477, "grad_norm": 0.5442648530006409, "learning_rate": 8.798793495535503e-06, "loss": 0.5222, "step": 2839 }, { "epoch": 0.908703071672355, "grad_norm": 0.5956857204437256, "learning_rate": 8.797582683626693e-06, "loss": 0.5404, "step": 2840 }, { "epoch": 0.9090230375426621, "grad_norm": 0.6557626128196716, "learning_rate": 8.796371345178477e-06, "loss": 0.5639, "step": 2841 }, { "epoch": 0.9093430034129693, "grad_norm": 0.5646318793296814, "learning_rate": 8.795159480358806e-06, "loss": 0.5396, "step": 2842 }, { "epoch": 0.9096629692832765, "grad_norm": 0.5170820951461792, "learning_rate": 8.793947089335709e-06, "loss": 0.5738, "step": 2843 }, { "epoch": 0.9099829351535836, "grad_norm": 0.520595133304596, "learning_rate": 8.792734172277282e-06, "loss": 0.5519, "step": 2844 }, { "epoch": 0.9103029010238908, "grad_norm": 0.562983512878418, "learning_rate": 8.791520729351699e-06, "loss": 0.5641, "step": 2845 }, { "epoch": 0.9106228668941979, "grad_norm": 0.6088793873786926, "learning_rate": 8.790306760727206e-06, "loss": 0.5708, "step": 2846 }, { "epoch": 0.9109428327645052, "grad_norm": 0.5645493268966675, "learning_rate": 8.78909226657212e-06, "loss": 0.5257, "step": 2847 }, { "epoch": 0.9112627986348123, "grad_norm": 0.6144366264343262, "learning_rate": 8.787877247054835e-06, "loss": 0.5757, "step": 2848 }, { "epoch": 0.9115827645051194, "grad_norm": 0.6192265152931213, "learning_rate": 8.786661702343811e-06, "loss": 0.5786, "step": 2849 }, { "epoch": 0.9119027303754266, "grad_norm": 0.5839664340019226, "learning_rate": 8.785445632607587e-06, "loss": 0.5418, "step": 2850 }, { "epoch": 0.9122226962457338, "grad_norm": 0.5896580219268799, "learning_rate": 8.784229038014772e-06, "loss": 0.5292, "step": 2851 }, { "epoch": 0.912542662116041, "grad_norm": 0.594308614730835, "learning_rate": 8.783011918734048e-06, "loss": 0.5923, "step": 2852 }, { "epoch": 0.9128626279863481, "grad_norm": 0.5671731233596802, "learning_rate": 8.78179427493417e-06, "loss": 0.5543, "step": 2853 }, { "epoch": 0.9131825938566553, "grad_norm": 0.5983908772468567, "learning_rate": 8.780576106783968e-06, "loss": 0.5705, "step": 2854 }, { "epoch": 0.9135025597269625, "grad_norm": 0.6325986385345459, "learning_rate": 8.77935741445234e-06, "loss": 0.5402, "step": 2855 }, { "epoch": 0.9138225255972696, "grad_norm": 0.542965292930603, "learning_rate": 8.778138198108259e-06, "loss": 0.5163, "step": 2856 }, { "epoch": 0.9141424914675768, "grad_norm": 0.5294291973114014, "learning_rate": 8.776918457920772e-06, "loss": 0.5433, "step": 2857 }, { "epoch": 0.914462457337884, "grad_norm": 0.6218854784965515, "learning_rate": 8.775698194058996e-06, "loss": 0.5748, "step": 2858 }, { "epoch": 0.9147824232081911, "grad_norm": 0.5364216566085815, "learning_rate": 8.774477406692125e-06, "loss": 0.5315, "step": 2859 }, { "epoch": 0.9151023890784983, "grad_norm": 0.5388673543930054, "learning_rate": 8.77325609598942e-06, "loss": 0.569, "step": 2860 }, { "epoch": 0.9154223549488054, "grad_norm": 0.5232530236244202, "learning_rate": 8.772034262120219e-06, "loss": 0.5583, "step": 2861 }, { "epoch": 0.9157423208191127, "grad_norm": 0.5909742116928101, "learning_rate": 8.770811905253929e-06, "loss": 0.5538, "step": 2862 }, { "epoch": 0.9160622866894198, "grad_norm": 0.547285795211792, "learning_rate": 8.76958902556003e-06, "loss": 0.5687, "step": 2863 }, { "epoch": 0.9163822525597269, "grad_norm": 0.5807123780250549, "learning_rate": 8.768365623208079e-06, "loss": 0.5542, "step": 2864 }, { "epoch": 0.9167022184300341, "grad_norm": 0.5130710601806641, "learning_rate": 8.767141698367701e-06, "loss": 0.5538, "step": 2865 }, { "epoch": 0.9170221843003413, "grad_norm": 0.5519154071807861, "learning_rate": 8.765917251208595e-06, "loss": 0.5351, "step": 2866 }, { "epoch": 0.9173421501706485, "grad_norm": 0.5024713277816772, "learning_rate": 8.764692281900531e-06, "loss": 0.5335, "step": 2867 }, { "epoch": 0.9176621160409556, "grad_norm": 0.5541590452194214, "learning_rate": 8.763466790613354e-06, "loss": 0.5737, "step": 2868 }, { "epoch": 0.9179820819112628, "grad_norm": 0.5962385535240173, "learning_rate": 8.762240777516979e-06, "loss": 0.5615, "step": 2869 }, { "epoch": 0.91830204778157, "grad_norm": 0.6544550657272339, "learning_rate": 8.761014242781392e-06, "loss": 0.5484, "step": 2870 }, { "epoch": 0.9186220136518771, "grad_norm": 0.5623205304145813, "learning_rate": 8.759787186576659e-06, "loss": 0.539, "step": 2871 }, { "epoch": 0.9189419795221843, "grad_norm": 0.5766005516052246, "learning_rate": 8.758559609072906e-06, "loss": 0.5491, "step": 2872 }, { "epoch": 0.9192619453924915, "grad_norm": 0.5268698334693909, "learning_rate": 8.757331510440343e-06, "loss": 0.5364, "step": 2873 }, { "epoch": 0.9195819112627986, "grad_norm": 0.4956420660018921, "learning_rate": 8.756102890849246e-06, "loss": 0.5355, "step": 2874 }, { "epoch": 0.9199018771331058, "grad_norm": 0.5091249942779541, "learning_rate": 8.754873750469964e-06, "loss": 0.5329, "step": 2875 }, { "epoch": 0.9202218430034129, "grad_norm": 0.6004356145858765, "learning_rate": 8.753644089472921e-06, "loss": 0.5531, "step": 2876 }, { "epoch": 0.9205418088737202, "grad_norm": 0.5854215621948242, "learning_rate": 8.752413908028608e-06, "loss": 0.5494, "step": 2877 }, { "epoch": 0.9208617747440273, "grad_norm": 0.5781176090240479, "learning_rate": 8.751183206307592e-06, "loss": 0.5567, "step": 2878 }, { "epoch": 0.9211817406143344, "grad_norm": 0.5860122442245483, "learning_rate": 8.749951984480511e-06, "loss": 0.5464, "step": 2879 }, { "epoch": 0.9215017064846417, "grad_norm": 0.5667566061019897, "learning_rate": 8.748720242718077e-06, "loss": 0.5346, "step": 2880 }, { "epoch": 0.9218216723549488, "grad_norm": 0.5337679386138916, "learning_rate": 8.747487981191072e-06, "loss": 0.5867, "step": 2881 }, { "epoch": 0.922141638225256, "grad_norm": 0.5574522614479065, "learning_rate": 8.74625520007035e-06, "loss": 0.5468, "step": 2882 }, { "epoch": 0.9224616040955631, "grad_norm": 0.5934003591537476, "learning_rate": 8.745021899526836e-06, "loss": 0.5639, "step": 2883 }, { "epoch": 0.9227815699658704, "grad_norm": 0.5360896587371826, "learning_rate": 8.743788079731533e-06, "loss": 0.5547, "step": 2884 }, { "epoch": 0.9231015358361775, "grad_norm": 0.544714629650116, "learning_rate": 8.742553740855507e-06, "loss": 0.5821, "step": 2885 }, { "epoch": 0.9234215017064846, "grad_norm": 0.576442301273346, "learning_rate": 8.741318883069903e-06, "loss": 0.5365, "step": 2886 }, { "epoch": 0.9237414675767918, "grad_norm": 0.508274257183075, "learning_rate": 8.740083506545933e-06, "loss": 0.5391, "step": 2887 }, { "epoch": 0.924061433447099, "grad_norm": 0.6314295530319214, "learning_rate": 8.738847611454887e-06, "loss": 0.5923, "step": 2888 }, { "epoch": 0.9243813993174061, "grad_norm": 0.5180608630180359, "learning_rate": 8.737611197968123e-06, "loss": 0.5535, "step": 2889 }, { "epoch": 0.9247013651877133, "grad_norm": 0.5145260691642761, "learning_rate": 8.736374266257069e-06, "loss": 0.5231, "step": 2890 }, { "epoch": 0.9250213310580204, "grad_norm": 0.5807346701622009, "learning_rate": 8.735136816493227e-06, "loss": 0.5227, "step": 2891 }, { "epoch": 0.9253412969283277, "grad_norm": 0.5591275691986084, "learning_rate": 8.733898848848172e-06, "loss": 0.5787, "step": 2892 }, { "epoch": 0.9256612627986348, "grad_norm": 0.5673110485076904, "learning_rate": 8.732660363493551e-06, "loss": 0.5172, "step": 2893 }, { "epoch": 0.925981228668942, "grad_norm": 0.49727028608322144, "learning_rate": 8.73142136060108e-06, "loss": 0.4982, "step": 2894 }, { "epoch": 0.9263011945392492, "grad_norm": 0.5635839700698853, "learning_rate": 8.730181840342547e-06, "loss": 0.5463, "step": 2895 }, { "epoch": 0.9266211604095563, "grad_norm": 0.5192051529884338, "learning_rate": 8.728941802889816e-06, "loss": 0.5239, "step": 2896 }, { "epoch": 0.9269411262798635, "grad_norm": 0.5240466594696045, "learning_rate": 8.727701248414816e-06, "loss": 0.5677, "step": 2897 }, { "epoch": 0.9272610921501706, "grad_norm": 0.5680010318756104, "learning_rate": 8.726460177089555e-06, "loss": 0.5549, "step": 2898 }, { "epoch": 0.9275810580204779, "grad_norm": 0.5307434797286987, "learning_rate": 8.725218589086107e-06, "loss": 0.5343, "step": 2899 }, { "epoch": 0.927901023890785, "grad_norm": 0.5652339458465576, "learning_rate": 8.72397648457662e-06, "loss": 0.581, "step": 2900 }, { "epoch": 0.9282209897610921, "grad_norm": 0.5568917393684387, "learning_rate": 8.722733863733314e-06, "loss": 0.5728, "step": 2901 }, { "epoch": 0.9285409556313993, "grad_norm": 0.5474474430084229, "learning_rate": 8.721490726728477e-06, "loss": 0.525, "step": 2902 }, { "epoch": 0.9288609215017065, "grad_norm": 0.5851140022277832, "learning_rate": 8.720247073734477e-06, "loss": 0.5398, "step": 2903 }, { "epoch": 0.9291808873720137, "grad_norm": 0.5718165040016174, "learning_rate": 8.719002904923742e-06, "loss": 0.5565, "step": 2904 }, { "epoch": 0.9295008532423208, "grad_norm": 0.5905211567878723, "learning_rate": 8.717758220468781e-06, "loss": 0.5636, "step": 2905 }, { "epoch": 0.929820819112628, "grad_norm": 0.5033652186393738, "learning_rate": 8.71651302054217e-06, "loss": 0.5419, "step": 2906 }, { "epoch": 0.9301407849829352, "grad_norm": 0.5867974162101746, "learning_rate": 8.715267305316559e-06, "loss": 0.5789, "step": 2907 }, { "epoch": 0.9304607508532423, "grad_norm": 0.5967637300491333, "learning_rate": 8.714021074964665e-06, "loss": 0.5306, "step": 2908 }, { "epoch": 0.9307807167235495, "grad_norm": 0.632807195186615, "learning_rate": 8.712774329659282e-06, "loss": 0.5807, "step": 2909 }, { "epoch": 0.9311006825938567, "grad_norm": 0.6260687708854675, "learning_rate": 8.71152706957327e-06, "loss": 0.5246, "step": 2910 }, { "epoch": 0.9314206484641638, "grad_norm": 0.5414279103279114, "learning_rate": 8.710279294879565e-06, "loss": 0.5595, "step": 2911 }, { "epoch": 0.931740614334471, "grad_norm": 0.5461296439170837, "learning_rate": 8.709031005751173e-06, "loss": 0.5444, "step": 2912 }, { "epoch": 0.9320605802047781, "grad_norm": 0.5828396677970886, "learning_rate": 8.707782202361171e-06, "loss": 0.5554, "step": 2913 }, { "epoch": 0.9323805460750854, "grad_norm": 0.6433313488960266, "learning_rate": 8.706532884882704e-06, "loss": 0.5773, "step": 2914 }, { "epoch": 0.9327005119453925, "grad_norm": 0.5411011576652527, "learning_rate": 8.705283053488994e-06, "loss": 0.5083, "step": 2915 }, { "epoch": 0.9330204778156996, "grad_norm": 0.5453394055366516, "learning_rate": 8.704032708353331e-06, "loss": 0.5247, "step": 2916 }, { "epoch": 0.9333404436860068, "grad_norm": 0.5410546064376831, "learning_rate": 8.702781849649078e-06, "loss": 0.5624, "step": 2917 }, { "epoch": 0.933660409556314, "grad_norm": 0.644980788230896, "learning_rate": 8.701530477549666e-06, "loss": 0.5544, "step": 2918 }, { "epoch": 0.9339803754266212, "grad_norm": 0.5591293573379517, "learning_rate": 8.7002785922286e-06, "loss": 0.5541, "step": 2919 }, { "epoch": 0.9343003412969283, "grad_norm": 0.5224248766899109, "learning_rate": 8.699026193859457e-06, "loss": 0.5256, "step": 2920 }, { "epoch": 0.9346203071672355, "grad_norm": 0.5552311539649963, "learning_rate": 8.697773282615881e-06, "loss": 0.5575, "step": 2921 }, { "epoch": 0.9349402730375427, "grad_norm": 0.5218842029571533, "learning_rate": 8.69651985867159e-06, "loss": 0.4947, "step": 2922 }, { "epoch": 0.9352602389078498, "grad_norm": 0.4850265681743622, "learning_rate": 8.695265922200376e-06, "loss": 0.5369, "step": 2923 }, { "epoch": 0.935580204778157, "grad_norm": 0.48894792795181274, "learning_rate": 8.694011473376094e-06, "loss": 0.5039, "step": 2924 }, { "epoch": 0.9359001706484642, "grad_norm": 0.4918539226055145, "learning_rate": 8.69275651237268e-06, "loss": 0.5237, "step": 2925 }, { "epoch": 0.9362201365187713, "grad_norm": 0.5880904197692871, "learning_rate": 8.69150103936413e-06, "loss": 0.5804, "step": 2926 }, { "epoch": 0.9365401023890785, "grad_norm": 0.4817546010017395, "learning_rate": 8.690245054524522e-06, "loss": 0.5256, "step": 2927 }, { "epoch": 0.9368600682593856, "grad_norm": 0.5519566535949707, "learning_rate": 8.688988558027997e-06, "loss": 0.5757, "step": 2928 }, { "epoch": 0.9371800341296929, "grad_norm": 0.5464499592781067, "learning_rate": 8.68773155004877e-06, "loss": 0.5341, "step": 2929 }, { "epoch": 0.9375, "grad_norm": 0.517829954624176, "learning_rate": 8.68647403076113e-06, "loss": 0.5241, "step": 2930 }, { "epoch": 0.9378199658703071, "grad_norm": 0.5223374366760254, "learning_rate": 8.685216000339426e-06, "loss": 0.5226, "step": 2931 }, { "epoch": 0.9381399317406144, "grad_norm": 0.6255651116371155, "learning_rate": 8.683957458958093e-06, "loss": 0.5509, "step": 2932 }, { "epoch": 0.9384598976109215, "grad_norm": 0.5552142858505249, "learning_rate": 8.682698406791627e-06, "loss": 0.5545, "step": 2933 }, { "epoch": 0.9387798634812287, "grad_norm": 0.52936851978302, "learning_rate": 8.681438844014595e-06, "loss": 0.5503, "step": 2934 }, { "epoch": 0.9390998293515358, "grad_norm": 0.5493532419204712, "learning_rate": 8.680178770801639e-06, "loss": 0.5101, "step": 2935 }, { "epoch": 0.939419795221843, "grad_norm": 0.5778977870941162, "learning_rate": 8.678918187327467e-06, "loss": 0.5461, "step": 2936 }, { "epoch": 0.9397397610921502, "grad_norm": 0.49943238496780396, "learning_rate": 8.677657093766865e-06, "loss": 0.5331, "step": 2937 }, { "epoch": 0.9400597269624573, "grad_norm": 0.5420438051223755, "learning_rate": 8.676395490294683e-06, "loss": 0.5324, "step": 2938 }, { "epoch": 0.9403796928327645, "grad_norm": 0.501508355140686, "learning_rate": 8.675133377085842e-06, "loss": 0.5687, "step": 2939 }, { "epoch": 0.9406996587030717, "grad_norm": 0.5415540337562561, "learning_rate": 8.673870754315336e-06, "loss": 0.5755, "step": 2940 }, { "epoch": 0.9410196245733788, "grad_norm": 0.5656829476356506, "learning_rate": 8.672607622158232e-06, "loss": 0.5624, "step": 2941 }, { "epoch": 0.941339590443686, "grad_norm": 0.5423877239227295, "learning_rate": 8.671343980789664e-06, "loss": 0.5723, "step": 2942 }, { "epoch": 0.9416595563139932, "grad_norm": 0.5485814809799194, "learning_rate": 8.670079830384834e-06, "loss": 0.5432, "step": 2943 }, { "epoch": 0.9419795221843004, "grad_norm": 0.5483697652816772, "learning_rate": 8.66881517111902e-06, "loss": 0.5533, "step": 2944 }, { "epoch": 0.9422994880546075, "grad_norm": 0.5534570217132568, "learning_rate": 8.667550003167571e-06, "loss": 0.5706, "step": 2945 }, { "epoch": 0.9426194539249146, "grad_norm": 0.5758525133132935, "learning_rate": 8.6662843267059e-06, "loss": 0.5817, "step": 2946 }, { "epoch": 0.9429394197952219, "grad_norm": 0.5041936039924622, "learning_rate": 8.665018141909498e-06, "loss": 0.5227, "step": 2947 }, { "epoch": 0.943259385665529, "grad_norm": 0.5714983344078064, "learning_rate": 8.663751448953921e-06, "loss": 0.5125, "step": 2948 }, { "epoch": 0.9435793515358362, "grad_norm": 0.5733338594436646, "learning_rate": 8.662484248014798e-06, "loss": 0.5499, "step": 2949 }, { "epoch": 0.9438993174061433, "grad_norm": 0.5094056725502014, "learning_rate": 8.661216539267827e-06, "loss": 0.5274, "step": 2950 }, { "epoch": 0.9442192832764505, "grad_norm": 0.49745941162109375, "learning_rate": 8.659948322888778e-06, "loss": 0.5243, "step": 2951 }, { "epoch": 0.9445392491467577, "grad_norm": 0.5140569806098938, "learning_rate": 8.658679599053493e-06, "loss": 0.511, "step": 2952 }, { "epoch": 0.9448592150170648, "grad_norm": 0.49003928899765015, "learning_rate": 8.65741036793788e-06, "loss": 0.5207, "step": 2953 }, { "epoch": 0.945179180887372, "grad_norm": 0.5427951216697693, "learning_rate": 8.656140629717918e-06, "loss": 0.5698, "step": 2954 }, { "epoch": 0.9454991467576792, "grad_norm": 0.5415076017379761, "learning_rate": 8.65487038456966e-06, "loss": 0.5584, "step": 2955 }, { "epoch": 0.9458191126279863, "grad_norm": 0.5324149131774902, "learning_rate": 8.653599632669225e-06, "loss": 0.5395, "step": 2956 }, { "epoch": 0.9461390784982935, "grad_norm": 0.538913369178772, "learning_rate": 8.652328374192807e-06, "loss": 0.5625, "step": 2957 }, { "epoch": 0.9464590443686007, "grad_norm": 0.5423220992088318, "learning_rate": 8.651056609316666e-06, "loss": 0.5491, "step": 2958 }, { "epoch": 0.9467790102389079, "grad_norm": 0.5532695651054382, "learning_rate": 8.649784338217133e-06, "loss": 0.545, "step": 2959 }, { "epoch": 0.947098976109215, "grad_norm": 0.5493027567863464, "learning_rate": 8.648511561070611e-06, "loss": 0.5579, "step": 2960 }, { "epoch": 0.9474189419795221, "grad_norm": 0.5765998363494873, "learning_rate": 8.647238278053572e-06, "loss": 0.5608, "step": 2961 }, { "epoch": 0.9477389078498294, "grad_norm": 0.5316721796989441, "learning_rate": 8.645964489342558e-06, "loss": 0.5451, "step": 2962 }, { "epoch": 0.9480588737201365, "grad_norm": 0.5517355799674988, "learning_rate": 8.644690195114183e-06, "loss": 0.5615, "step": 2963 }, { "epoch": 0.9483788395904437, "grad_norm": 0.5976579189300537, "learning_rate": 8.643415395545125e-06, "loss": 0.5925, "step": 2964 }, { "epoch": 0.9486988054607508, "grad_norm": 0.5386163592338562, "learning_rate": 8.642140090812141e-06, "loss": 0.5727, "step": 2965 }, { "epoch": 0.949018771331058, "grad_norm": 0.6165236234664917, "learning_rate": 8.640864281092051e-06, "loss": 0.5379, "step": 2966 }, { "epoch": 0.9493387372013652, "grad_norm": 0.5166876316070557, "learning_rate": 8.639587966561748e-06, "loss": 0.538, "step": 2967 }, { "epoch": 0.9496587030716723, "grad_norm": 0.5584707856178284, "learning_rate": 8.638311147398195e-06, "loss": 0.5364, "step": 2968 }, { "epoch": 0.9499786689419796, "grad_norm": 0.5661454796791077, "learning_rate": 8.637033823778426e-06, "loss": 0.5683, "step": 2969 }, { "epoch": 0.9502986348122867, "grad_norm": 0.43857452273368835, "learning_rate": 8.63575599587954e-06, "loss": 0.5328, "step": 2970 }, { "epoch": 0.9506186006825939, "grad_norm": 0.5182543396949768, "learning_rate": 8.634477663878714e-06, "loss": 0.5386, "step": 2971 }, { "epoch": 0.950938566552901, "grad_norm": 0.6035263538360596, "learning_rate": 8.633198827953185e-06, "loss": 0.5329, "step": 2972 }, { "epoch": 0.9512585324232082, "grad_norm": 0.6187506318092346, "learning_rate": 8.631919488280267e-06, "loss": 0.6131, "step": 2973 }, { "epoch": 0.9515784982935154, "grad_norm": 0.6119416356086731, "learning_rate": 8.630639645037345e-06, "loss": 0.5745, "step": 2974 }, { "epoch": 0.9518984641638225, "grad_norm": 0.6179429292678833, "learning_rate": 8.629359298401866e-06, "loss": 0.5084, "step": 2975 }, { "epoch": 0.9522184300341296, "grad_norm": 0.5868006348609924, "learning_rate": 8.628078448551355e-06, "loss": 0.5159, "step": 2976 }, { "epoch": 0.9525383959044369, "grad_norm": 0.5683136582374573, "learning_rate": 8.626797095663403e-06, "loss": 0.5664, "step": 2977 }, { "epoch": 0.952858361774744, "grad_norm": 0.5581501722335815, "learning_rate": 8.62551523991567e-06, "loss": 0.5361, "step": 2978 }, { "epoch": 0.9531783276450512, "grad_norm": 0.5560288429260254, "learning_rate": 8.624232881485887e-06, "loss": 0.5436, "step": 2979 }, { "epoch": 0.9534982935153583, "grad_norm": 0.536558985710144, "learning_rate": 8.622950020551857e-06, "loss": 0.5327, "step": 2980 }, { "epoch": 0.9538182593856656, "grad_norm": 0.5403125286102295, "learning_rate": 8.621666657291446e-06, "loss": 0.5641, "step": 2981 }, { "epoch": 0.9541382252559727, "grad_norm": 0.5872126221656799, "learning_rate": 8.620382791882597e-06, "loss": 0.5457, "step": 2982 }, { "epoch": 0.9544581911262798, "grad_norm": 0.6266401410102844, "learning_rate": 8.619098424503318e-06, "loss": 0.5543, "step": 2983 }, { "epoch": 0.9547781569965871, "grad_norm": 0.5969432592391968, "learning_rate": 8.61781355533169e-06, "loss": 0.5553, "step": 2984 }, { "epoch": 0.9550981228668942, "grad_norm": 0.5646101832389832, "learning_rate": 8.616528184545858e-06, "loss": 0.5469, "step": 2985 }, { "epoch": 0.9554180887372014, "grad_norm": 0.5284569263458252, "learning_rate": 8.615242312324043e-06, "loss": 0.5331, "step": 2986 }, { "epoch": 0.9557380546075085, "grad_norm": 0.5586323738098145, "learning_rate": 8.613955938844533e-06, "loss": 0.5167, "step": 2987 }, { "epoch": 0.9560580204778157, "grad_norm": 0.612792432308197, "learning_rate": 8.612669064285684e-06, "loss": 0.5447, "step": 2988 }, { "epoch": 0.9563779863481229, "grad_norm": 0.5168460011482239, "learning_rate": 8.611381688825924e-06, "loss": 0.524, "step": 2989 }, { "epoch": 0.95669795221843, "grad_norm": 0.5096966028213501, "learning_rate": 8.610093812643747e-06, "loss": 0.527, "step": 2990 }, { "epoch": 0.9570179180887372, "grad_norm": 0.5304121375083923, "learning_rate": 8.60880543591772e-06, "loss": 0.5298, "step": 2991 }, { "epoch": 0.9573378839590444, "grad_norm": 0.5534718036651611, "learning_rate": 8.607516558826477e-06, "loss": 0.5516, "step": 2992 }, { "epoch": 0.9576578498293515, "grad_norm": 0.5174388289451599, "learning_rate": 8.606227181548725e-06, "loss": 0.5036, "step": 2993 }, { "epoch": 0.9579778156996587, "grad_norm": 0.6583805084228516, "learning_rate": 8.604937304263234e-06, "loss": 0.5903, "step": 2994 }, { "epoch": 0.9582977815699659, "grad_norm": 0.5533691644668579, "learning_rate": 8.60364692714885e-06, "loss": 0.5468, "step": 2995 }, { "epoch": 0.9586177474402731, "grad_norm": 0.6131950616836548, "learning_rate": 8.602356050384483e-06, "loss": 0.5772, "step": 2996 }, { "epoch": 0.9589377133105802, "grad_norm": 0.5823208689689636, "learning_rate": 8.601064674149118e-06, "loss": 0.5806, "step": 2997 }, { "epoch": 0.9592576791808873, "grad_norm": 0.4916262626647949, "learning_rate": 8.599772798621804e-06, "loss": 0.516, "step": 2998 }, { "epoch": 0.9595776450511946, "grad_norm": 0.6060745120048523, "learning_rate": 8.598480423981658e-06, "loss": 0.5884, "step": 2999 }, { "epoch": 0.9598976109215017, "grad_norm": 0.618219792842865, "learning_rate": 8.597187550407875e-06, "loss": 0.5709, "step": 3000 }, { "epoch": 0.9602175767918089, "grad_norm": 0.5556377172470093, "learning_rate": 8.59589417807971e-06, "loss": 0.5586, "step": 3001 }, { "epoch": 0.960537542662116, "grad_norm": 0.5143526196479797, "learning_rate": 8.594600307176489e-06, "loss": 0.5287, "step": 3002 }, { "epoch": 0.9608575085324232, "grad_norm": 0.6290923357009888, "learning_rate": 8.593305937877614e-06, "loss": 0.5302, "step": 3003 }, { "epoch": 0.9611774744027304, "grad_norm": 0.5228386521339417, "learning_rate": 8.592011070362546e-06, "loss": 0.5407, "step": 3004 }, { "epoch": 0.9614974402730375, "grad_norm": 0.579010009765625, "learning_rate": 8.590715704810823e-06, "loss": 0.5843, "step": 3005 }, { "epoch": 0.9618174061433447, "grad_norm": 0.5646539926528931, "learning_rate": 8.589419841402046e-06, "loss": 0.551, "step": 3006 }, { "epoch": 0.9621373720136519, "grad_norm": 0.5978478193283081, "learning_rate": 8.58812348031589e-06, "loss": 0.5851, "step": 3007 }, { "epoch": 0.962457337883959, "grad_norm": 0.5651780962944031, "learning_rate": 8.586826621732099e-06, "loss": 0.5137, "step": 3008 }, { "epoch": 0.9627773037542662, "grad_norm": 0.4923568665981293, "learning_rate": 8.58552926583048e-06, "loss": 0.5288, "step": 3009 }, { "epoch": 0.9630972696245734, "grad_norm": 0.5681024789810181, "learning_rate": 8.584231412790915e-06, "loss": 0.5577, "step": 3010 }, { "epoch": 0.9634172354948806, "grad_norm": 0.6538505554199219, "learning_rate": 8.582933062793353e-06, "loss": 0.5595, "step": 3011 }, { "epoch": 0.9637372013651877, "grad_norm": 0.6075664758682251, "learning_rate": 8.581634216017812e-06, "loss": 0.585, "step": 3012 }, { "epoch": 0.9640571672354948, "grad_norm": 0.5595016479492188, "learning_rate": 8.580334872644379e-06, "loss": 0.5538, "step": 3013 }, { "epoch": 0.9643771331058021, "grad_norm": 0.532981812953949, "learning_rate": 8.579035032853207e-06, "loss": 0.5603, "step": 3014 }, { "epoch": 0.9646970989761092, "grad_norm": 0.6255946159362793, "learning_rate": 8.577734696824523e-06, "loss": 0.5478, "step": 3015 }, { "epoch": 0.9650170648464164, "grad_norm": 0.5752984881401062, "learning_rate": 8.576433864738618e-06, "loss": 0.5203, "step": 3016 }, { "epoch": 0.9653370307167235, "grad_norm": 0.5297114849090576, "learning_rate": 8.575132536775854e-06, "loss": 0.5227, "step": 3017 }, { "epoch": 0.9656569965870307, "grad_norm": 0.581031322479248, "learning_rate": 8.573830713116663e-06, "loss": 0.5667, "step": 3018 }, { "epoch": 0.9659769624573379, "grad_norm": 0.5882019996643066, "learning_rate": 8.572528393941547e-06, "loss": 0.5482, "step": 3019 }, { "epoch": 0.966296928327645, "grad_norm": 0.5825169086456299, "learning_rate": 8.57122557943107e-06, "loss": 0.5323, "step": 3020 }, { "epoch": 0.9666168941979523, "grad_norm": 0.5378094911575317, "learning_rate": 8.56992226976587e-06, "loss": 0.536, "step": 3021 }, { "epoch": 0.9669368600682594, "grad_norm": 0.5341657400131226, "learning_rate": 8.568618465126653e-06, "loss": 0.5173, "step": 3022 }, { "epoch": 0.9672568259385665, "grad_norm": 0.5237258076667786, "learning_rate": 8.567314165694192e-06, "loss": 0.5211, "step": 3023 }, { "epoch": 0.9675767918088737, "grad_norm": 0.5663238763809204, "learning_rate": 8.566009371649331e-06, "loss": 0.5642, "step": 3024 }, { "epoch": 0.9678967576791809, "grad_norm": 0.5667411684989929, "learning_rate": 8.56470408317298e-06, "loss": 0.5888, "step": 3025 }, { "epoch": 0.9682167235494881, "grad_norm": 0.5694569945335388, "learning_rate": 8.56339830044612e-06, "loss": 0.5472, "step": 3026 }, { "epoch": 0.9685366894197952, "grad_norm": 0.5204916596412659, "learning_rate": 8.562092023649797e-06, "loss": 0.5311, "step": 3027 }, { "epoch": 0.9688566552901023, "grad_norm": 0.5056095123291016, "learning_rate": 8.560785252965131e-06, "loss": 0.4955, "step": 3028 }, { "epoch": 0.9691766211604096, "grad_norm": 0.5372504591941833, "learning_rate": 8.559477988573305e-06, "loss": 0.5407, "step": 3029 }, { "epoch": 0.9694965870307167, "grad_norm": 0.5474112033843994, "learning_rate": 8.558170230655576e-06, "loss": 0.5968, "step": 3030 }, { "epoch": 0.9698165529010239, "grad_norm": 0.5350422859191895, "learning_rate": 8.556861979393263e-06, "loss": 0.5182, "step": 3031 }, { "epoch": 0.9701365187713311, "grad_norm": 0.546489417552948, "learning_rate": 8.555553234967757e-06, "loss": 0.5249, "step": 3032 }, { "epoch": 0.9704564846416383, "grad_norm": 0.5644913911819458, "learning_rate": 8.554243997560517e-06, "loss": 0.5295, "step": 3033 }, { "epoch": 0.9707764505119454, "grad_norm": 0.673241913318634, "learning_rate": 8.552934267353072e-06, "loss": 0.541, "step": 3034 }, { "epoch": 0.9710964163822525, "grad_norm": 0.6079564094543457, "learning_rate": 8.551624044527016e-06, "loss": 0.5735, "step": 3035 }, { "epoch": 0.9714163822525598, "grad_norm": 0.5430492162704468, "learning_rate": 8.550313329264015e-06, "loss": 0.5315, "step": 3036 }, { "epoch": 0.9717363481228669, "grad_norm": 0.5448910593986511, "learning_rate": 8.549002121745798e-06, "loss": 0.5193, "step": 3037 }, { "epoch": 0.972056313993174, "grad_norm": 0.568170964717865, "learning_rate": 8.547690422154167e-06, "loss": 0.5585, "step": 3038 }, { "epoch": 0.9723762798634812, "grad_norm": 0.536199152469635, "learning_rate": 8.546378230670992e-06, "loss": 0.5278, "step": 3039 }, { "epoch": 0.9726962457337884, "grad_norm": 0.6083328723907471, "learning_rate": 8.545065547478209e-06, "loss": 0.5808, "step": 3040 }, { "epoch": 0.9730162116040956, "grad_norm": 0.5674001574516296, "learning_rate": 8.543752372757822e-06, "loss": 0.5648, "step": 3041 }, { "epoch": 0.9733361774744027, "grad_norm": 0.5693421363830566, "learning_rate": 8.542438706691906e-06, "loss": 0.5582, "step": 3042 }, { "epoch": 0.9736561433447098, "grad_norm": 0.44867393374443054, "learning_rate": 8.541124549462601e-06, "loss": 0.5147, "step": 3043 }, { "epoch": 0.9739761092150171, "grad_norm": 0.6155474185943604, "learning_rate": 8.539809901252118e-06, "loss": 0.5725, "step": 3044 }, { "epoch": 0.9742960750853242, "grad_norm": 0.4862614870071411, "learning_rate": 8.538494762242733e-06, "loss": 0.558, "step": 3045 }, { "epoch": 0.9746160409556314, "grad_norm": 0.45874154567718506, "learning_rate": 8.537179132616794e-06, "loss": 0.5087, "step": 3046 }, { "epoch": 0.9749360068259386, "grad_norm": 0.5131855607032776, "learning_rate": 8.53586301255671e-06, "loss": 0.49, "step": 3047 }, { "epoch": 0.9752559726962458, "grad_norm": 0.5835806727409363, "learning_rate": 8.534546402244968e-06, "loss": 0.5597, "step": 3048 }, { "epoch": 0.9755759385665529, "grad_norm": 0.559389054775238, "learning_rate": 8.533229301864114e-06, "loss": 0.5037, "step": 3049 }, { "epoch": 0.97589590443686, "grad_norm": 0.5053697824478149, "learning_rate": 8.531911711596767e-06, "loss": 0.5171, "step": 3050 }, { "epoch": 0.9762158703071673, "grad_norm": 0.586311936378479, "learning_rate": 8.530593631625611e-06, "loss": 0.5929, "step": 3051 }, { "epoch": 0.9765358361774744, "grad_norm": 0.5995824933052063, "learning_rate": 8.529275062133404e-06, "loss": 0.5303, "step": 3052 }, { "epoch": 0.9768558020477816, "grad_norm": 0.5481569170951843, "learning_rate": 8.527956003302961e-06, "loss": 0.5559, "step": 3053 }, { "epoch": 0.9771757679180887, "grad_norm": 0.5199539065361023, "learning_rate": 8.526636455317174e-06, "loss": 0.5406, "step": 3054 }, { "epoch": 0.9774957337883959, "grad_norm": 0.5552107095718384, "learning_rate": 8.525316418359e-06, "loss": 0.5756, "step": 3055 }, { "epoch": 0.9778156996587031, "grad_norm": 0.47679877281188965, "learning_rate": 8.523995892611465e-06, "loss": 0.5354, "step": 3056 }, { "epoch": 0.9781356655290102, "grad_norm": 0.5298394560813904, "learning_rate": 8.522674878257658e-06, "loss": 0.5243, "step": 3057 }, { "epoch": 0.9784556313993175, "grad_norm": 0.5738407373428345, "learning_rate": 8.521353375480743e-06, "loss": 0.5903, "step": 3058 }, { "epoch": 0.9787755972696246, "grad_norm": 0.6375570893287659, "learning_rate": 8.520031384463945e-06, "loss": 0.6056, "step": 3059 }, { "epoch": 0.9790955631399317, "grad_norm": 0.512746274471283, "learning_rate": 8.518708905390562e-06, "loss": 0.5555, "step": 3060 }, { "epoch": 0.9794155290102389, "grad_norm": 0.544510006904602, "learning_rate": 8.517385938443955e-06, "loss": 0.5579, "step": 3061 }, { "epoch": 0.9797354948805461, "grad_norm": 0.5213636755943298, "learning_rate": 8.516062483807556e-06, "loss": 0.5501, "step": 3062 }, { "epoch": 0.9800554607508533, "grad_norm": 0.5775883793830872, "learning_rate": 8.514738541664865e-06, "loss": 0.5589, "step": 3063 }, { "epoch": 0.9803754266211604, "grad_norm": 0.5627555251121521, "learning_rate": 8.513414112199445e-06, "loss": 0.5033, "step": 3064 }, { "epoch": 0.9806953924914675, "grad_norm": 0.5240724682807922, "learning_rate": 8.512089195594933e-06, "loss": 0.5415, "step": 3065 }, { "epoch": 0.9810153583617748, "grad_norm": 0.4943949282169342, "learning_rate": 8.510763792035029e-06, "loss": 0.5319, "step": 3066 }, { "epoch": 0.9813353242320819, "grad_norm": 0.5783194303512573, "learning_rate": 8.509437901703501e-06, "loss": 0.5346, "step": 3067 }, { "epoch": 0.981655290102389, "grad_norm": 0.5179439187049866, "learning_rate": 8.508111524784186e-06, "loss": 0.52, "step": 3068 }, { "epoch": 0.9819752559726962, "grad_norm": 0.557386577129364, "learning_rate": 8.506784661460987e-06, "loss": 0.5643, "step": 3069 }, { "epoch": 0.9822952218430034, "grad_norm": 0.5388240814208984, "learning_rate": 8.505457311917878e-06, "loss": 0.5375, "step": 3070 }, { "epoch": 0.9826151877133106, "grad_norm": 0.5537760257720947, "learning_rate": 8.504129476338893e-06, "loss": 0.588, "step": 3071 }, { "epoch": 0.9829351535836177, "grad_norm": 0.5869219303131104, "learning_rate": 8.502801154908142e-06, "loss": 0.6043, "step": 3072 }, { "epoch": 0.983255119453925, "grad_norm": 0.5651918649673462, "learning_rate": 8.501472347809799e-06, "loss": 0.5446, "step": 3073 }, { "epoch": 0.9835750853242321, "grad_norm": 0.5868792533874512, "learning_rate": 8.500143055228098e-06, "loss": 0.5518, "step": 3074 }, { "epoch": 0.9838950511945392, "grad_norm": 0.5043837428092957, "learning_rate": 8.498813277347355e-06, "loss": 0.5071, "step": 3075 }, { "epoch": 0.9842150170648464, "grad_norm": 0.5354411602020264, "learning_rate": 8.497483014351941e-06, "loss": 0.5535, "step": 3076 }, { "epoch": 0.9845349829351536, "grad_norm": 0.5286612510681152, "learning_rate": 8.4961522664263e-06, "loss": 0.5606, "step": 3077 }, { "epoch": 0.9848549488054608, "grad_norm": 0.5994420647621155, "learning_rate": 8.49482103375494e-06, "loss": 0.5705, "step": 3078 }, { "epoch": 0.9851749146757679, "grad_norm": 0.602558970451355, "learning_rate": 8.49348931652244e-06, "loss": 0.5416, "step": 3079 }, { "epoch": 0.985494880546075, "grad_norm": 0.556376039981842, "learning_rate": 8.492157114913444e-06, "loss": 0.5471, "step": 3080 }, { "epoch": 0.9858148464163823, "grad_norm": 0.5554435849189758, "learning_rate": 8.490824429112664e-06, "loss": 0.508, "step": 3081 }, { "epoch": 0.9861348122866894, "grad_norm": 0.540307343006134, "learning_rate": 8.489491259304874e-06, "loss": 0.5097, "step": 3082 }, { "epoch": 0.9864547781569966, "grad_norm": 0.5240446925163269, "learning_rate": 8.488157605674924e-06, "loss": 0.5212, "step": 3083 }, { "epoch": 0.9867747440273038, "grad_norm": 0.5979939699172974, "learning_rate": 8.486823468407727e-06, "loss": 0.5393, "step": 3084 }, { "epoch": 0.987094709897611, "grad_norm": 0.5482702851295471, "learning_rate": 8.48548884768826e-06, "loss": 0.5371, "step": 3085 }, { "epoch": 0.9874146757679181, "grad_norm": 0.5796502828598022, "learning_rate": 8.484153743701572e-06, "loss": 0.5188, "step": 3086 }, { "epoch": 0.9877346416382252, "grad_norm": 0.5995625853538513, "learning_rate": 8.482818156632776e-06, "loss": 0.5373, "step": 3087 }, { "epoch": 0.9880546075085325, "grad_norm": 0.6242316365242004, "learning_rate": 8.481482086667052e-06, "loss": 0.5614, "step": 3088 }, { "epoch": 0.9883745733788396, "grad_norm": 0.5484504103660583, "learning_rate": 8.480145533989648e-06, "loss": 0.55, "step": 3089 }, { "epoch": 0.9886945392491467, "grad_norm": 0.5537928342819214, "learning_rate": 8.478808498785879e-06, "loss": 0.4946, "step": 3090 }, { "epoch": 0.9890145051194539, "grad_norm": 0.5466088056564331, "learning_rate": 8.47747098124113e-06, "loss": 0.5628, "step": 3091 }, { "epoch": 0.9893344709897611, "grad_norm": 0.6240554451942444, "learning_rate": 8.476132981540842e-06, "loss": 0.5386, "step": 3092 }, { "epoch": 0.9896544368600683, "grad_norm": 0.6110613346099854, "learning_rate": 8.474794499870535e-06, "loss": 0.5548, "step": 3093 }, { "epoch": 0.9899744027303754, "grad_norm": 0.5479440689086914, "learning_rate": 8.47345553641579e-06, "loss": 0.5283, "step": 3094 }, { "epoch": 0.9902943686006825, "grad_norm": 0.4982355237007141, "learning_rate": 8.472116091362256e-06, "loss": 0.5143, "step": 3095 }, { "epoch": 0.9906143344709898, "grad_norm": 0.577256441116333, "learning_rate": 8.47077616489565e-06, "loss": 0.5033, "step": 3096 }, { "epoch": 0.9909343003412969, "grad_norm": 0.6515082120895386, "learning_rate": 8.469435757201754e-06, "loss": 0.6009, "step": 3097 }, { "epoch": 0.9912542662116041, "grad_norm": 0.5402945280075073, "learning_rate": 8.468094868466416e-06, "loss": 0.542, "step": 3098 }, { "epoch": 0.9915742320819113, "grad_norm": 0.5361272096633911, "learning_rate": 8.466753498875551e-06, "loss": 0.5378, "step": 3099 }, { "epoch": 0.9918941979522184, "grad_norm": 0.6215559244155884, "learning_rate": 8.465411648615144e-06, "loss": 0.5756, "step": 3100 }, { "epoch": 0.9922141638225256, "grad_norm": 0.540403425693512, "learning_rate": 8.464069317871242e-06, "loss": 0.5224, "step": 3101 }, { "epoch": 0.9925341296928327, "grad_norm": 0.6040542125701904, "learning_rate": 8.46272650682996e-06, "loss": 0.5539, "step": 3102 }, { "epoch": 0.99285409556314, "grad_norm": 0.5860587954521179, "learning_rate": 8.461383215677486e-06, "loss": 0.5314, "step": 3103 }, { "epoch": 0.9931740614334471, "grad_norm": 0.5857371687889099, "learning_rate": 8.460039444600063e-06, "loss": 0.564, "step": 3104 }, { "epoch": 0.9934940273037542, "grad_norm": 0.5509449243545532, "learning_rate": 8.45869519378401e-06, "loss": 0.5033, "step": 3105 }, { "epoch": 0.9938139931740614, "grad_norm": 0.5352454781532288, "learning_rate": 8.457350463415706e-06, "loss": 0.5561, "step": 3106 }, { "epoch": 0.9941339590443686, "grad_norm": 0.5762051343917847, "learning_rate": 8.456005253681601e-06, "loss": 0.5428, "step": 3107 }, { "epoch": 0.9944539249146758, "grad_norm": 0.5870938897132874, "learning_rate": 8.45465956476821e-06, "loss": 0.507, "step": 3108 }, { "epoch": 0.9947738907849829, "grad_norm": 0.5523510575294495, "learning_rate": 8.453313396862113e-06, "loss": 0.5233, "step": 3109 }, { "epoch": 0.9950938566552902, "grad_norm": 0.5425226092338562, "learning_rate": 8.45196675014996e-06, "loss": 0.5139, "step": 3110 }, { "epoch": 0.9954138225255973, "grad_norm": 0.6074098348617554, "learning_rate": 8.450619624818465e-06, "loss": 0.5366, "step": 3111 }, { "epoch": 0.9957337883959044, "grad_norm": 0.6013935208320618, "learning_rate": 8.449272021054407e-06, "loss": 0.5685, "step": 3112 }, { "epoch": 0.9960537542662116, "grad_norm": 0.5533202290534973, "learning_rate": 8.447923939044636e-06, "loss": 0.506, "step": 3113 }, { "epoch": 0.9963737201365188, "grad_norm": 0.5740543603897095, "learning_rate": 8.44657537897606e-06, "loss": 0.5556, "step": 3114 }, { "epoch": 0.996693686006826, "grad_norm": 0.539481520652771, "learning_rate": 8.445226341035662e-06, "loss": 0.5381, "step": 3115 }, { "epoch": 0.9970136518771331, "grad_norm": 0.5156455636024475, "learning_rate": 8.443876825410488e-06, "loss": 0.5027, "step": 3116 }, { "epoch": 0.9973336177474402, "grad_norm": 0.5651331543922424, "learning_rate": 8.442526832287649e-06, "loss": 0.5773, "step": 3117 }, { "epoch": 0.9976535836177475, "grad_norm": 0.5460453629493713, "learning_rate": 8.441176361854322e-06, "loss": 0.5315, "step": 3118 }, { "epoch": 0.9979735494880546, "grad_norm": 0.5436340570449829, "learning_rate": 8.439825414297755e-06, "loss": 0.5222, "step": 3119 }, { "epoch": 0.9982935153583617, "grad_norm": 0.5273920893669128, "learning_rate": 8.438473989805253e-06, "loss": 0.5286, "step": 3120 }, { "epoch": 0.9986134812286689, "grad_norm": 0.5575355291366577, "learning_rate": 8.437122088564197e-06, "loss": 0.5435, "step": 3121 }, { "epoch": 0.9989334470989761, "grad_norm": 0.5628740191459656, "learning_rate": 8.43576971076203e-06, "loss": 0.5183, "step": 3122 }, { "epoch": 0.9992534129692833, "grad_norm": 0.48700088262557983, "learning_rate": 8.434416856586258e-06, "loss": 0.5431, "step": 3123 }, { "epoch": 0.9995733788395904, "grad_norm": 0.5751681923866272, "learning_rate": 8.433063526224456e-06, "loss": 0.5581, "step": 3124 }, { "epoch": 0.9998933447098977, "grad_norm": 0.5555768609046936, "learning_rate": 8.431709719864268e-06, "loss": 0.5391, "step": 3125 }, { "epoch": 1.0002133105802047, "grad_norm": 1.2219892740249634, "learning_rate": 8.430355437693398e-06, "loss": 0.9491, "step": 3126 }, { "epoch": 1.000533276450512, "grad_norm": 0.4970657229423523, "learning_rate": 8.42900067989962e-06, "loss": 0.4525, "step": 3127 }, { "epoch": 1.0008532423208192, "grad_norm": 0.5354446172714233, "learning_rate": 8.427645446670772e-06, "loss": 0.5274, "step": 3128 }, { "epoch": 1.0011732081911262, "grad_norm": 0.5338842272758484, "learning_rate": 8.426289738194759e-06, "loss": 0.5127, "step": 3129 }, { "epoch": 1.0014931740614335, "grad_norm": 0.5222037434577942, "learning_rate": 8.424933554659554e-06, "loss": 0.4923, "step": 3130 }, { "epoch": 1.0018131399317407, "grad_norm": 0.5694612264633179, "learning_rate": 8.42357689625319e-06, "loss": 0.505, "step": 3131 }, { "epoch": 1.0021331058020477, "grad_norm": 0.5298541784286499, "learning_rate": 8.422219763163769e-06, "loss": 0.4549, "step": 3132 }, { "epoch": 1.002453071672355, "grad_norm": 0.5489261150360107, "learning_rate": 8.420862155579463e-06, "loss": 0.5095, "step": 3133 }, { "epoch": 1.0027730375426622, "grad_norm": 0.5825126767158508, "learning_rate": 8.419504073688503e-06, "loss": 0.455, "step": 3134 }, { "epoch": 1.0030930034129693, "grad_norm": 0.5706313848495483, "learning_rate": 8.418145517679188e-06, "loss": 0.532, "step": 3135 }, { "epoch": 1.0034129692832765, "grad_norm": 0.5857523083686829, "learning_rate": 8.416786487739888e-06, "loss": 0.4992, "step": 3136 }, { "epoch": 1.0037329351535835, "grad_norm": 0.5802491903305054, "learning_rate": 8.41542698405903e-06, "loss": 0.4995, "step": 3137 }, { "epoch": 1.0040529010238908, "grad_norm": 0.5400286316871643, "learning_rate": 8.414067006825108e-06, "loss": 0.5103, "step": 3138 }, { "epoch": 1.004372866894198, "grad_norm": 0.5207788944244385, "learning_rate": 8.412706556226694e-06, "loss": 0.4338, "step": 3139 }, { "epoch": 1.004692832764505, "grad_norm": 0.5851981043815613, "learning_rate": 8.411345632452405e-06, "loss": 0.5164, "step": 3140 }, { "epoch": 1.0050127986348123, "grad_norm": 0.6366373300552368, "learning_rate": 8.409984235690945e-06, "loss": 0.5192, "step": 3141 }, { "epoch": 1.0053327645051195, "grad_norm": 0.5763211846351624, "learning_rate": 8.408622366131067e-06, "loss": 0.5273, "step": 3142 }, { "epoch": 1.0056527303754266, "grad_norm": 0.6516720652580261, "learning_rate": 8.407260023961594e-06, "loss": 0.5475, "step": 3143 }, { "epoch": 1.0059726962457338, "grad_norm": 0.5470252633094788, "learning_rate": 8.405897209371424e-06, "loss": 0.477, "step": 3144 }, { "epoch": 1.0062926621160408, "grad_norm": 0.5717004537582397, "learning_rate": 8.404533922549506e-06, "loss": 0.4527, "step": 3145 }, { "epoch": 1.006612627986348, "grad_norm": 0.614112913608551, "learning_rate": 8.403170163684864e-06, "loss": 0.4917, "step": 3146 }, { "epoch": 1.0069325938566553, "grad_norm": 0.5276660919189453, "learning_rate": 8.401805932966585e-06, "loss": 0.5293, "step": 3147 }, { "epoch": 1.0072525597269624, "grad_norm": 0.5499048829078674, "learning_rate": 8.400441230583822e-06, "loss": 0.4858, "step": 3148 }, { "epoch": 1.0075725255972696, "grad_norm": 0.6219374537467957, "learning_rate": 8.39907605672579e-06, "loss": 0.5287, "step": 3149 }, { "epoch": 1.0078924914675769, "grad_norm": 0.5626710057258606, "learning_rate": 8.397710411581774e-06, "loss": 0.5166, "step": 3150 }, { "epoch": 1.008212457337884, "grad_norm": 0.5865727663040161, "learning_rate": 8.396344295341124e-06, "loss": 0.5228, "step": 3151 }, { "epoch": 1.0085324232081911, "grad_norm": 0.6077318787574768, "learning_rate": 8.394977708193248e-06, "loss": 0.5257, "step": 3152 }, { "epoch": 1.0088523890784984, "grad_norm": 0.4857766628265381, "learning_rate": 8.39361065032763e-06, "loss": 0.4495, "step": 3153 }, { "epoch": 1.0091723549488054, "grad_norm": 0.5425410866737366, "learning_rate": 8.392243121933815e-06, "loss": 0.5265, "step": 3154 }, { "epoch": 1.0094923208191127, "grad_norm": 0.5726216435432434, "learning_rate": 8.390875123201408e-06, "loss": 0.4532, "step": 3155 }, { "epoch": 1.0098122866894197, "grad_norm": 0.5400950908660889, "learning_rate": 8.389506654320085e-06, "loss": 0.5149, "step": 3156 }, { "epoch": 1.010132252559727, "grad_norm": 0.5174579620361328, "learning_rate": 8.388137715479587e-06, "loss": 0.5041, "step": 3157 }, { "epoch": 1.0104522184300342, "grad_norm": 0.5398826599121094, "learning_rate": 8.38676830686972e-06, "loss": 0.524, "step": 3158 }, { "epoch": 1.0107721843003412, "grad_norm": 0.5821525454521179, "learning_rate": 8.385398428680353e-06, "loss": 0.4714, "step": 3159 }, { "epoch": 1.0110921501706485, "grad_norm": 0.5623863935470581, "learning_rate": 8.384028081101419e-06, "loss": 0.5151, "step": 3160 }, { "epoch": 1.0114121160409557, "grad_norm": 0.5194283127784729, "learning_rate": 8.382657264322924e-06, "loss": 0.4981, "step": 3161 }, { "epoch": 1.0117320819112627, "grad_norm": 0.5321395397186279, "learning_rate": 8.381285978534925e-06, "loss": 0.4672, "step": 3162 }, { "epoch": 1.01205204778157, "grad_norm": 0.48655349016189575, "learning_rate": 8.379914223927562e-06, "loss": 0.4755, "step": 3163 }, { "epoch": 1.0123720136518772, "grad_norm": 0.5503526329994202, "learning_rate": 8.378542000691024e-06, "loss": 0.5549, "step": 3164 }, { "epoch": 1.0126919795221843, "grad_norm": 0.5182765126228333, "learning_rate": 8.377169309015571e-06, "loss": 0.4444, "step": 3165 }, { "epoch": 1.0130119453924915, "grad_norm": 0.5637997984886169, "learning_rate": 8.375796149091534e-06, "loss": 0.5394, "step": 3166 }, { "epoch": 1.0133319112627985, "grad_norm": 0.5851975083351135, "learning_rate": 8.374422521109298e-06, "loss": 0.5391, "step": 3167 }, { "epoch": 1.0136518771331058, "grad_norm": 0.5644164681434631, "learning_rate": 8.373048425259319e-06, "loss": 0.4577, "step": 3168 }, { "epoch": 1.013971843003413, "grad_norm": 0.5628954768180847, "learning_rate": 8.371673861732119e-06, "loss": 0.4758, "step": 3169 }, { "epoch": 1.01429180887372, "grad_norm": 0.5550352931022644, "learning_rate": 8.370298830718283e-06, "loss": 0.5186, "step": 3170 }, { "epoch": 1.0146117747440273, "grad_norm": 0.5640912652015686, "learning_rate": 8.368923332408458e-06, "loss": 0.487, "step": 3171 }, { "epoch": 1.0149317406143346, "grad_norm": 0.6037182211875916, "learning_rate": 8.36754736699336e-06, "loss": 0.5129, "step": 3172 }, { "epoch": 1.0152517064846416, "grad_norm": 0.6077131628990173, "learning_rate": 8.366170934663768e-06, "loss": 0.5372, "step": 3173 }, { "epoch": 1.0155716723549488, "grad_norm": 0.5096414089202881, "learning_rate": 8.364794035610527e-06, "loss": 0.5031, "step": 3174 }, { "epoch": 1.015891638225256, "grad_norm": 0.5736592411994934, "learning_rate": 8.363416670024546e-06, "loss": 0.4667, "step": 3175 }, { "epoch": 1.016211604095563, "grad_norm": 0.6692733764648438, "learning_rate": 8.362038838096797e-06, "loss": 0.5971, "step": 3176 }, { "epoch": 1.0165315699658704, "grad_norm": 0.555644154548645, "learning_rate": 8.360660540018317e-06, "loss": 0.4759, "step": 3177 }, { "epoch": 1.0168515358361774, "grad_norm": 0.5540698766708374, "learning_rate": 8.35928177598021e-06, "loss": 0.4867, "step": 3178 }, { "epoch": 1.0171715017064846, "grad_norm": 0.5888321399688721, "learning_rate": 8.357902546173645e-06, "loss": 0.5535, "step": 3179 }, { "epoch": 1.0174914675767919, "grad_norm": 0.516242265701294, "learning_rate": 8.356522850789852e-06, "loss": 0.4709, "step": 3180 }, { "epoch": 1.017811433447099, "grad_norm": 0.5233179926872253, "learning_rate": 8.355142690020128e-06, "loss": 0.5062, "step": 3181 }, { "epoch": 1.0181313993174061, "grad_norm": 0.5436132550239563, "learning_rate": 8.353762064055832e-06, "loss": 0.535, "step": 3182 }, { "epoch": 1.0184513651877134, "grad_norm": 0.5038325190544128, "learning_rate": 8.352380973088394e-06, "loss": 0.4532, "step": 3183 }, { "epoch": 1.0187713310580204, "grad_norm": 0.5422745943069458, "learning_rate": 8.350999417309298e-06, "loss": 0.4596, "step": 3184 }, { "epoch": 1.0190912969283277, "grad_norm": 0.5560320019721985, "learning_rate": 8.349617396910104e-06, "loss": 0.4887, "step": 3185 }, { "epoch": 1.019411262798635, "grad_norm": 0.5979790687561035, "learning_rate": 8.348234912082427e-06, "loss": 0.4871, "step": 3186 }, { "epoch": 1.019731228668942, "grad_norm": 0.5235424637794495, "learning_rate": 8.346851963017952e-06, "loss": 0.4893, "step": 3187 }, { "epoch": 1.0200511945392492, "grad_norm": 0.5454728007316589, "learning_rate": 8.345468549908425e-06, "loss": 0.5642, "step": 3188 }, { "epoch": 1.0203711604095562, "grad_norm": 0.46851009130477905, "learning_rate": 8.344084672945659e-06, "loss": 0.5052, "step": 3189 }, { "epoch": 1.0206911262798635, "grad_norm": 0.531676709651947, "learning_rate": 8.342700332321531e-06, "loss": 0.4974, "step": 3190 }, { "epoch": 1.0210110921501707, "grad_norm": 0.5864760279655457, "learning_rate": 8.34131552822798e-06, "loss": 0.5055, "step": 3191 }, { "epoch": 1.0213310580204777, "grad_norm": 0.5224139094352722, "learning_rate": 8.339930260857011e-06, "loss": 0.5447, "step": 3192 }, { "epoch": 1.021651023890785, "grad_norm": 0.5960366725921631, "learning_rate": 8.338544530400693e-06, "loss": 0.5089, "step": 3193 }, { "epoch": 1.0219709897610922, "grad_norm": 0.5359147787094116, "learning_rate": 8.337158337051161e-06, "loss": 0.5099, "step": 3194 }, { "epoch": 1.0222909556313993, "grad_norm": 0.48335564136505127, "learning_rate": 8.33577168100061e-06, "loss": 0.4905, "step": 3195 }, { "epoch": 1.0226109215017065, "grad_norm": 0.5697606801986694, "learning_rate": 8.334384562441302e-06, "loss": 0.47, "step": 3196 }, { "epoch": 1.0229308873720138, "grad_norm": 0.5768377780914307, "learning_rate": 8.332996981565564e-06, "loss": 0.4789, "step": 3197 }, { "epoch": 1.0232508532423208, "grad_norm": 0.5663846731185913, "learning_rate": 8.331608938565782e-06, "loss": 0.5589, "step": 3198 }, { "epoch": 1.023570819112628, "grad_norm": 0.5420123934745789, "learning_rate": 8.330220433634416e-06, "loss": 0.5203, "step": 3199 }, { "epoch": 1.023890784982935, "grad_norm": 0.49670645594596863, "learning_rate": 8.32883146696398e-06, "loss": 0.4672, "step": 3200 }, { "epoch": 1.0242107508532423, "grad_norm": 0.5433145761489868, "learning_rate": 8.327442038747055e-06, "loss": 0.4974, "step": 3201 }, { "epoch": 1.0245307167235496, "grad_norm": 0.567774772644043, "learning_rate": 8.32605214917629e-06, "loss": 0.5517, "step": 3202 }, { "epoch": 1.0248506825938566, "grad_norm": 0.5511053204536438, "learning_rate": 8.324661798444391e-06, "loss": 0.4643, "step": 3203 }, { "epoch": 1.0251706484641638, "grad_norm": 0.5236759781837463, "learning_rate": 8.323270986744137e-06, "loss": 0.5015, "step": 3204 }, { "epoch": 1.025490614334471, "grad_norm": 0.5212015509605408, "learning_rate": 8.321879714268361e-06, "loss": 0.5431, "step": 3205 }, { "epoch": 1.025810580204778, "grad_norm": 0.5333899855613708, "learning_rate": 8.320487981209966e-06, "loss": 0.4933, "step": 3206 }, { "epoch": 1.0261305460750854, "grad_norm": 0.5734988451004028, "learning_rate": 8.319095787761918e-06, "loss": 0.5144, "step": 3207 }, { "epoch": 1.0264505119453924, "grad_norm": 0.5054695010185242, "learning_rate": 8.317703134117245e-06, "loss": 0.4651, "step": 3208 }, { "epoch": 1.0267704778156996, "grad_norm": 0.5281773209571838, "learning_rate": 8.316310020469043e-06, "loss": 0.5086, "step": 3209 }, { "epoch": 1.0270904436860069, "grad_norm": 0.556444525718689, "learning_rate": 8.314916447010467e-06, "loss": 0.472, "step": 3210 }, { "epoch": 1.027410409556314, "grad_norm": 0.4962175488471985, "learning_rate": 8.313522413934736e-06, "loss": 0.4704, "step": 3211 }, { "epoch": 1.0277303754266212, "grad_norm": 0.589185357093811, "learning_rate": 8.312127921435137e-06, "loss": 0.5101, "step": 3212 }, { "epoch": 1.0280503412969284, "grad_norm": 0.5513185858726501, "learning_rate": 8.310732969705018e-06, "loss": 0.4972, "step": 3213 }, { "epoch": 1.0283703071672354, "grad_norm": 0.5592458844184875, "learning_rate": 8.309337558937789e-06, "loss": 0.4905, "step": 3214 }, { "epoch": 1.0286902730375427, "grad_norm": 0.5382454991340637, "learning_rate": 8.307941689326926e-06, "loss": 0.4726, "step": 3215 }, { "epoch": 1.02901023890785, "grad_norm": 0.6426695585250854, "learning_rate": 8.306545361065968e-06, "loss": 0.5357, "step": 3216 }, { "epoch": 1.029330204778157, "grad_norm": 0.5265448093414307, "learning_rate": 8.305148574348519e-06, "loss": 0.4265, "step": 3217 }, { "epoch": 1.0296501706484642, "grad_norm": 0.5393176078796387, "learning_rate": 8.303751329368242e-06, "loss": 0.5238, "step": 3218 }, { "epoch": 1.0299701365187712, "grad_norm": 0.5781740546226501, "learning_rate": 8.30235362631887e-06, "loss": 0.5115, "step": 3219 }, { "epoch": 1.0302901023890785, "grad_norm": 0.5424593687057495, "learning_rate": 8.300955465394197e-06, "loss": 0.4784, "step": 3220 }, { "epoch": 1.0306100682593857, "grad_norm": 0.5782645344734192, "learning_rate": 8.299556846788074e-06, "loss": 0.5157, "step": 3221 }, { "epoch": 1.0309300341296928, "grad_norm": 0.5042377710342407, "learning_rate": 8.298157770694427e-06, "loss": 0.4234, "step": 3222 }, { "epoch": 1.03125, "grad_norm": 0.5630826354026794, "learning_rate": 8.296758237307234e-06, "loss": 0.5474, "step": 3223 }, { "epoch": 1.0315699658703072, "grad_norm": 0.5753233432769775, "learning_rate": 8.295358246820548e-06, "loss": 0.4971, "step": 3224 }, { "epoch": 1.0318899317406143, "grad_norm": 0.5464937090873718, "learning_rate": 8.293957799428477e-06, "loss": 0.4847, "step": 3225 }, { "epoch": 1.0322098976109215, "grad_norm": 0.5086600184440613, "learning_rate": 8.292556895325195e-06, "loss": 0.5046, "step": 3226 }, { "epoch": 1.0325298634812288, "grad_norm": 0.5309995412826538, "learning_rate": 8.291155534704937e-06, "loss": 0.4632, "step": 3227 }, { "epoch": 1.0328498293515358, "grad_norm": 0.5247173309326172, "learning_rate": 8.289753717762004e-06, "loss": 0.4877, "step": 3228 }, { "epoch": 1.033169795221843, "grad_norm": 0.5032480955123901, "learning_rate": 8.28835144469076e-06, "loss": 0.5124, "step": 3229 }, { "epoch": 1.03348976109215, "grad_norm": 0.5265656113624573, "learning_rate": 8.286948715685635e-06, "loss": 0.5028, "step": 3230 }, { "epoch": 1.0338097269624573, "grad_norm": 0.5498801469802856, "learning_rate": 8.285545530941114e-06, "loss": 0.4715, "step": 3231 }, { "epoch": 1.0341296928327646, "grad_norm": 0.5443584322929382, "learning_rate": 8.284141890651754e-06, "loss": 0.4765, "step": 3232 }, { "epoch": 1.0344496587030716, "grad_norm": 0.5465605854988098, "learning_rate": 8.282737795012169e-06, "loss": 0.4875, "step": 3233 }, { "epoch": 1.0347696245733788, "grad_norm": 0.5705323815345764, "learning_rate": 8.281333244217041e-06, "loss": 0.5012, "step": 3234 }, { "epoch": 1.035089590443686, "grad_norm": 0.5770228505134583, "learning_rate": 8.279928238461108e-06, "loss": 0.5451, "step": 3235 }, { "epoch": 1.0354095563139931, "grad_norm": 0.5116875767707825, "learning_rate": 8.278522777939181e-06, "loss": 0.423, "step": 3236 }, { "epoch": 1.0357295221843004, "grad_norm": 0.5620558857917786, "learning_rate": 8.277116862846127e-06, "loss": 0.5358, "step": 3237 }, { "epoch": 1.0360494880546076, "grad_norm": 0.5042481422424316, "learning_rate": 8.275710493376876e-06, "loss": 0.4488, "step": 3238 }, { "epoch": 1.0363694539249146, "grad_norm": 0.5645015239715576, "learning_rate": 8.274303669726427e-06, "loss": 0.51, "step": 3239 }, { "epoch": 1.036689419795222, "grad_norm": 0.5250040888786316, "learning_rate": 8.272896392089833e-06, "loss": 0.5062, "step": 3240 }, { "epoch": 1.037009385665529, "grad_norm": 0.528319239616394, "learning_rate": 8.271488660662217e-06, "loss": 0.4766, "step": 3241 }, { "epoch": 1.0373293515358362, "grad_norm": 0.5562976598739624, "learning_rate": 8.270080475638762e-06, "loss": 0.4982, "step": 3242 }, { "epoch": 1.0376493174061434, "grad_norm": 0.5111921429634094, "learning_rate": 8.268671837214717e-06, "loss": 0.472, "step": 3243 }, { "epoch": 1.0379692832764504, "grad_norm": 0.5360894203186035, "learning_rate": 8.267262745585387e-06, "loss": 0.5228, "step": 3244 }, { "epoch": 1.0382892491467577, "grad_norm": 0.543817937374115, "learning_rate": 8.265853200946147e-06, "loss": 0.4887, "step": 3245 }, { "epoch": 1.038609215017065, "grad_norm": 0.5790839791297913, "learning_rate": 8.264443203492435e-06, "loss": 0.4692, "step": 3246 }, { "epoch": 1.038929180887372, "grad_norm": 0.5253699421882629, "learning_rate": 8.263032753419741e-06, "loss": 0.5429, "step": 3247 }, { "epoch": 1.0392491467576792, "grad_norm": 0.48694974184036255, "learning_rate": 8.261621850923634e-06, "loss": 0.5038, "step": 3248 }, { "epoch": 1.0395691126279862, "grad_norm": 0.5342028737068176, "learning_rate": 8.260210496199732e-06, "loss": 0.4863, "step": 3249 }, { "epoch": 1.0398890784982935, "grad_norm": 0.5062534213066101, "learning_rate": 8.258798689443724e-06, "loss": 0.4391, "step": 3250 }, { "epoch": 1.0402090443686007, "grad_norm": 0.613639771938324, "learning_rate": 8.257386430851355e-06, "loss": 0.5303, "step": 3251 }, { "epoch": 1.0405290102389078, "grad_norm": 0.5553945302963257, "learning_rate": 8.255973720618438e-06, "loss": 0.5305, "step": 3252 }, { "epoch": 1.040848976109215, "grad_norm": 0.5541419982910156, "learning_rate": 8.25456055894085e-06, "loss": 0.5362, "step": 3253 }, { "epoch": 1.0411689419795223, "grad_norm": 0.552528977394104, "learning_rate": 8.253146946014525e-06, "loss": 0.499, "step": 3254 }, { "epoch": 1.0414889078498293, "grad_norm": 0.5524160861968994, "learning_rate": 8.25173288203546e-06, "loss": 0.5571, "step": 3255 }, { "epoch": 1.0418088737201365, "grad_norm": 0.5329165458679199, "learning_rate": 8.250318367199722e-06, "loss": 0.4503, "step": 3256 }, { "epoch": 1.0421288395904438, "grad_norm": 0.5469471216201782, "learning_rate": 8.24890340170343e-06, "loss": 0.473, "step": 3257 }, { "epoch": 1.0424488054607508, "grad_norm": 0.5729274749755859, "learning_rate": 8.247487985742774e-06, "loss": 0.4979, "step": 3258 }, { "epoch": 1.042768771331058, "grad_norm": 0.5663588047027588, "learning_rate": 8.246072119514002e-06, "loss": 0.5106, "step": 3259 }, { "epoch": 1.0430887372013653, "grad_norm": 0.5898635983467102, "learning_rate": 8.244655803213425e-06, "loss": 0.4891, "step": 3260 }, { "epoch": 1.0434087030716723, "grad_norm": 0.5540909171104431, "learning_rate": 8.243239037037418e-06, "loss": 0.5196, "step": 3261 }, { "epoch": 1.0437286689419796, "grad_norm": 0.5461093783378601, "learning_rate": 8.241821821182417e-06, "loss": 0.4959, "step": 3262 }, { "epoch": 1.0440486348122866, "grad_norm": 0.5747105479240417, "learning_rate": 8.240404155844919e-06, "loss": 0.4654, "step": 3263 }, { "epoch": 1.0443686006825939, "grad_norm": 0.5178041458129883, "learning_rate": 8.238986041221485e-06, "loss": 0.4613, "step": 3264 }, { "epoch": 1.044688566552901, "grad_norm": 0.5070105791091919, "learning_rate": 8.237567477508744e-06, "loss": 0.4415, "step": 3265 }, { "epoch": 1.0450085324232081, "grad_norm": 0.5689228177070618, "learning_rate": 8.236148464903374e-06, "loss": 0.4882, "step": 3266 }, { "epoch": 1.0453284982935154, "grad_norm": 0.6333630084991455, "learning_rate": 8.234729003602128e-06, "loss": 0.5739, "step": 3267 }, { "epoch": 1.0456484641638226, "grad_norm": 0.5439204573631287, "learning_rate": 8.233309093801815e-06, "loss": 0.5014, "step": 3268 }, { "epoch": 1.0459684300341296, "grad_norm": 0.673212468624115, "learning_rate": 8.231888735699305e-06, "loss": 0.501, "step": 3269 }, { "epoch": 1.046288395904437, "grad_norm": 0.5605301260948181, "learning_rate": 8.230467929491533e-06, "loss": 0.4604, "step": 3270 }, { "epoch": 1.046608361774744, "grad_norm": 0.49820882081985474, "learning_rate": 8.229046675375498e-06, "loss": 0.4852, "step": 3271 }, { "epoch": 1.0469283276450512, "grad_norm": 0.5804089307785034, "learning_rate": 8.227624973548256e-06, "loss": 0.486, "step": 3272 }, { "epoch": 1.0472482935153584, "grad_norm": 0.6037084460258484, "learning_rate": 8.226202824206929e-06, "loss": 0.4991, "step": 3273 }, { "epoch": 1.0475682593856654, "grad_norm": 0.567973792552948, "learning_rate": 8.224780227548698e-06, "loss": 0.5378, "step": 3274 }, { "epoch": 1.0478882252559727, "grad_norm": 0.5975142121315002, "learning_rate": 8.22335718377081e-06, "loss": 0.4967, "step": 3275 }, { "epoch": 1.04820819112628, "grad_norm": 0.5783261060714722, "learning_rate": 8.221933693070569e-06, "loss": 0.4535, "step": 3276 }, { "epoch": 1.048528156996587, "grad_norm": 0.5568932890892029, "learning_rate": 8.220509755645348e-06, "loss": 0.5058, "step": 3277 }, { "epoch": 1.0488481228668942, "grad_norm": 0.524492084980011, "learning_rate": 8.219085371692573e-06, "loss": 0.4632, "step": 3278 }, { "epoch": 1.0491680887372015, "grad_norm": 0.534509003162384, "learning_rate": 8.21766054140974e-06, "loss": 0.4929, "step": 3279 }, { "epoch": 1.0494880546075085, "grad_norm": 0.5752255916595459, "learning_rate": 8.216235264994402e-06, "loss": 0.544, "step": 3280 }, { "epoch": 1.0498080204778157, "grad_norm": 0.5233433842658997, "learning_rate": 8.214809542644173e-06, "loss": 0.4828, "step": 3281 }, { "epoch": 1.0501279863481228, "grad_norm": 0.6086320281028748, "learning_rate": 8.213383374556736e-06, "loss": 0.4538, "step": 3282 }, { "epoch": 1.05044795221843, "grad_norm": 0.609158456325531, "learning_rate": 8.211956760929827e-06, "loss": 0.4731, "step": 3283 }, { "epoch": 1.0507679180887373, "grad_norm": 0.5908544659614563, "learning_rate": 8.210529701961248e-06, "loss": 0.5263, "step": 3284 }, { "epoch": 1.0510878839590443, "grad_norm": 0.6520845293998718, "learning_rate": 8.209102197848866e-06, "loss": 0.5137, "step": 3285 }, { "epoch": 1.0514078498293515, "grad_norm": 0.6237879395484924, "learning_rate": 8.207674248790602e-06, "loss": 0.4624, "step": 3286 }, { "epoch": 1.0517278156996588, "grad_norm": 0.5166109800338745, "learning_rate": 8.206245854984445e-06, "loss": 0.5023, "step": 3287 }, { "epoch": 1.0520477815699658, "grad_norm": 0.5419546961784363, "learning_rate": 8.20481701662844e-06, "loss": 0.4643, "step": 3288 }, { "epoch": 1.052367747440273, "grad_norm": 0.6240190863609314, "learning_rate": 8.203387733920704e-06, "loss": 0.5036, "step": 3289 }, { "epoch": 1.0526877133105803, "grad_norm": 0.6003880500793457, "learning_rate": 8.201958007059405e-06, "loss": 0.5088, "step": 3290 }, { "epoch": 1.0530076791808873, "grad_norm": 0.5665394067764282, "learning_rate": 8.200527836242775e-06, "loss": 0.5574, "step": 3291 }, { "epoch": 1.0533276450511946, "grad_norm": 0.4948168098926544, "learning_rate": 8.19909722166911e-06, "loss": 0.4787, "step": 3292 }, { "epoch": 1.0536476109215016, "grad_norm": 0.5346024632453918, "learning_rate": 8.197666163536768e-06, "loss": 0.468, "step": 3293 }, { "epoch": 1.0539675767918089, "grad_norm": 0.535876989364624, "learning_rate": 8.196234662044164e-06, "loss": 0.4379, "step": 3294 }, { "epoch": 1.054287542662116, "grad_norm": 0.5247263312339783, "learning_rate": 8.19480271738978e-06, "loss": 0.5081, "step": 3295 }, { "epoch": 1.0546075085324231, "grad_norm": 0.5306622385978699, "learning_rate": 8.193370329772154e-06, "loss": 0.4969, "step": 3296 }, { "epoch": 1.0549274744027304, "grad_norm": 0.49779438972473145, "learning_rate": 8.191937499389892e-06, "loss": 0.4608, "step": 3297 }, { "epoch": 1.0552474402730376, "grad_norm": 0.555260181427002, "learning_rate": 8.190504226441654e-06, "loss": 0.5152, "step": 3298 }, { "epoch": 1.0555674061433447, "grad_norm": 0.5562401413917542, "learning_rate": 8.18907051112617e-06, "loss": 0.5202, "step": 3299 }, { "epoch": 1.055887372013652, "grad_norm": 0.49704766273498535, "learning_rate": 8.187636353642218e-06, "loss": 0.4844, "step": 3300 }, { "epoch": 1.0562073378839592, "grad_norm": 0.5293060541152954, "learning_rate": 8.186201754188655e-06, "loss": 0.4755, "step": 3301 }, { "epoch": 1.0565273037542662, "grad_norm": 0.5252447128295898, "learning_rate": 8.184766712964385e-06, "loss": 0.5046, "step": 3302 }, { "epoch": 1.0568472696245734, "grad_norm": 0.4727516174316406, "learning_rate": 8.183331230168376e-06, "loss": 0.4682, "step": 3303 }, { "epoch": 1.0571672354948805, "grad_norm": 0.5746709108352661, "learning_rate": 8.181895305999665e-06, "loss": 0.5164, "step": 3304 }, { "epoch": 1.0574872013651877, "grad_norm": 0.5763015747070312, "learning_rate": 8.180458940657343e-06, "loss": 0.5427, "step": 3305 }, { "epoch": 1.057807167235495, "grad_norm": 0.5482608079910278, "learning_rate": 8.17902213434056e-06, "loss": 0.463, "step": 3306 }, { "epoch": 1.058127133105802, "grad_norm": 0.5401831269264221, "learning_rate": 8.177584887248536e-06, "loss": 0.5213, "step": 3307 }, { "epoch": 1.0584470989761092, "grad_norm": 0.5248244404792786, "learning_rate": 8.176147199580542e-06, "loss": 0.5038, "step": 3308 }, { "epoch": 1.0587670648464165, "grad_norm": 0.5097712278366089, "learning_rate": 8.17470907153592e-06, "loss": 0.4701, "step": 3309 }, { "epoch": 1.0590870307167235, "grad_norm": 0.47569549083709717, "learning_rate": 8.173270503314063e-06, "loss": 0.4556, "step": 3310 }, { "epoch": 1.0594069965870307, "grad_norm": 0.5200427770614624, "learning_rate": 8.171831495114436e-06, "loss": 0.4838, "step": 3311 }, { "epoch": 1.0597269624573378, "grad_norm": 0.5401687026023865, "learning_rate": 8.170392047136555e-06, "loss": 0.5279, "step": 3312 }, { "epoch": 1.060046928327645, "grad_norm": 0.5490127205848694, "learning_rate": 8.168952159580004e-06, "loss": 0.5255, "step": 3313 }, { "epoch": 1.0603668941979523, "grad_norm": 0.5770564675331116, "learning_rate": 8.167511832644423e-06, "loss": 0.4994, "step": 3314 }, { "epoch": 1.0606868600682593, "grad_norm": 0.5831413269042969, "learning_rate": 8.166071066529515e-06, "loss": 0.5379, "step": 3315 }, { "epoch": 1.0610068259385665, "grad_norm": 0.5414169430732727, "learning_rate": 8.164629861435047e-06, "loss": 0.4899, "step": 3316 }, { "epoch": 1.0613267918088738, "grad_norm": 0.5539081692695618, "learning_rate": 8.16318821756084e-06, "loss": 0.4502, "step": 3317 }, { "epoch": 1.0616467576791808, "grad_norm": 0.5647209882736206, "learning_rate": 8.16174613510678e-06, "loss": 0.5359, "step": 3318 }, { "epoch": 1.061966723549488, "grad_norm": 0.49853506684303284, "learning_rate": 8.160303614272817e-06, "loss": 0.4405, "step": 3319 }, { "epoch": 1.0622866894197953, "grad_norm": 0.483504056930542, "learning_rate": 8.158860655258955e-06, "loss": 0.4475, "step": 3320 }, { "epoch": 1.0626066552901023, "grad_norm": 0.5061872005462646, "learning_rate": 8.157417258265263e-06, "loss": 0.4942, "step": 3321 }, { "epoch": 1.0629266211604096, "grad_norm": 0.5668355226516724, "learning_rate": 8.155973423491868e-06, "loss": 0.5019, "step": 3322 }, { "epoch": 1.0632465870307168, "grad_norm": 0.6143890023231506, "learning_rate": 8.154529151138964e-06, "loss": 0.5261, "step": 3323 }, { "epoch": 1.0635665529010239, "grad_norm": 0.4885425269603729, "learning_rate": 8.153084441406797e-06, "loss": 0.4456, "step": 3324 }, { "epoch": 1.0638865187713311, "grad_norm": 0.5878362059593201, "learning_rate": 8.151639294495678e-06, "loss": 0.5305, "step": 3325 }, { "epoch": 1.0642064846416381, "grad_norm": 0.5964034199714661, "learning_rate": 8.15019371060598e-06, "loss": 0.5334, "step": 3326 }, { "epoch": 1.0645264505119454, "grad_norm": 0.5398755669593811, "learning_rate": 8.148747689938135e-06, "loss": 0.4528, "step": 3327 }, { "epoch": 1.0648464163822526, "grad_norm": 0.5904126167297363, "learning_rate": 8.147301232692634e-06, "loss": 0.5541, "step": 3328 }, { "epoch": 1.0651663822525597, "grad_norm": 0.54847651720047, "learning_rate": 8.14585433907003e-06, "loss": 0.536, "step": 3329 }, { "epoch": 1.065486348122867, "grad_norm": 0.5282127857208252, "learning_rate": 8.144407009270939e-06, "loss": 0.5024, "step": 3330 }, { "epoch": 1.0658063139931742, "grad_norm": 0.5941870808601379, "learning_rate": 8.142959243496032e-06, "loss": 0.4766, "step": 3331 }, { "epoch": 1.0661262798634812, "grad_norm": 0.519917368888855, "learning_rate": 8.141511041946046e-06, "loss": 0.5628, "step": 3332 }, { "epoch": 1.0664462457337884, "grad_norm": 0.5521838665008545, "learning_rate": 8.140062404821773e-06, "loss": 0.5019, "step": 3333 }, { "epoch": 1.0667662116040955, "grad_norm": 0.643412709236145, "learning_rate": 8.138613332324072e-06, "loss": 0.5178, "step": 3334 }, { "epoch": 1.0670861774744027, "grad_norm": 0.6367138028144836, "learning_rate": 8.137163824653855e-06, "loss": 0.4894, "step": 3335 }, { "epoch": 1.06740614334471, "grad_norm": 0.5044212341308594, "learning_rate": 8.135713882012102e-06, "loss": 0.4632, "step": 3336 }, { "epoch": 1.067726109215017, "grad_norm": 0.602066159248352, "learning_rate": 8.134263504599844e-06, "loss": 0.4942, "step": 3337 }, { "epoch": 1.0680460750853242, "grad_norm": 0.6120120286941528, "learning_rate": 8.132812692618181e-06, "loss": 0.5182, "step": 3338 }, { "epoch": 1.0683660409556315, "grad_norm": 0.5152328014373779, "learning_rate": 8.131361446268271e-06, "loss": 0.4354, "step": 3339 }, { "epoch": 1.0686860068259385, "grad_norm": 0.5362467169761658, "learning_rate": 8.129909765751326e-06, "loss": 0.5225, "step": 3340 }, { "epoch": 1.0690059726962458, "grad_norm": 0.5970868468284607, "learning_rate": 8.128457651268631e-06, "loss": 0.5388, "step": 3341 }, { "epoch": 1.069325938566553, "grad_norm": 0.53805011510849, "learning_rate": 8.127005103021516e-06, "loss": 0.4494, "step": 3342 }, { "epoch": 1.06964590443686, "grad_norm": 0.5005940794944763, "learning_rate": 8.125552121211385e-06, "loss": 0.5064, "step": 3343 }, { "epoch": 1.0699658703071673, "grad_norm": 0.537519097328186, "learning_rate": 8.124098706039687e-06, "loss": 0.5224, "step": 3344 }, { "epoch": 1.0702858361774743, "grad_norm": 0.5686160922050476, "learning_rate": 8.122644857707949e-06, "loss": 0.5255, "step": 3345 }, { "epoch": 1.0706058020477816, "grad_norm": 0.5152590870857239, "learning_rate": 8.121190576417744e-06, "loss": 0.5161, "step": 3346 }, { "epoch": 1.0709257679180888, "grad_norm": 0.5477904081344604, "learning_rate": 8.11973586237071e-06, "loss": 0.5362, "step": 3347 }, { "epoch": 1.0712457337883958, "grad_norm": 0.5535791516304016, "learning_rate": 8.118280715768546e-06, "loss": 0.4732, "step": 3348 }, { "epoch": 1.071565699658703, "grad_norm": 0.5651800632476807, "learning_rate": 8.116825136813009e-06, "loss": 0.5736, "step": 3349 }, { "epoch": 1.0718856655290103, "grad_norm": 0.5503290295600891, "learning_rate": 8.115369125705919e-06, "loss": 0.4364, "step": 3350 }, { "epoch": 1.0722056313993173, "grad_norm": 0.5717608332633972, "learning_rate": 8.11391268264915e-06, "loss": 0.4905, "step": 3351 }, { "epoch": 1.0725255972696246, "grad_norm": 0.6119275689125061, "learning_rate": 8.112455807844642e-06, "loss": 0.4943, "step": 3352 }, { "epoch": 1.0728455631399316, "grad_norm": 0.5548709034919739, "learning_rate": 8.110998501494391e-06, "loss": 0.4723, "step": 3353 }, { "epoch": 1.0731655290102389, "grad_norm": 0.5616167783737183, "learning_rate": 8.109540763800454e-06, "loss": 0.5413, "step": 3354 }, { "epoch": 1.0734854948805461, "grad_norm": 0.5117178559303284, "learning_rate": 8.108082594964951e-06, "loss": 0.4356, "step": 3355 }, { "epoch": 1.0738054607508531, "grad_norm": 0.6019683480262756, "learning_rate": 8.106623995190058e-06, "loss": 0.5237, "step": 3356 }, { "epoch": 1.0741254266211604, "grad_norm": 0.5925198197364807, "learning_rate": 8.105164964678009e-06, "loss": 0.6085, "step": 3357 }, { "epoch": 1.0744453924914676, "grad_norm": 0.5321702361106873, "learning_rate": 8.103705503631103e-06, "loss": 0.4576, "step": 3358 }, { "epoch": 1.0747653583617747, "grad_norm": 0.578241229057312, "learning_rate": 8.102245612251696e-06, "loss": 0.5489, "step": 3359 }, { "epoch": 1.075085324232082, "grad_norm": 0.609102189540863, "learning_rate": 8.100785290742202e-06, "loss": 0.5153, "step": 3360 }, { "epoch": 1.0754052901023892, "grad_norm": 0.5429154634475708, "learning_rate": 8.099324539305096e-06, "loss": 0.5183, "step": 3361 }, { "epoch": 1.0757252559726962, "grad_norm": 0.5318170785903931, "learning_rate": 8.097863358142914e-06, "loss": 0.4384, "step": 3362 }, { "epoch": 1.0760452218430034, "grad_norm": 0.5602610111236572, "learning_rate": 8.096401747458253e-06, "loss": 0.4774, "step": 3363 }, { "epoch": 1.0763651877133107, "grad_norm": 0.6025781631469727, "learning_rate": 8.094939707453762e-06, "loss": 0.5688, "step": 3364 }, { "epoch": 1.0766851535836177, "grad_norm": 0.5115510821342468, "learning_rate": 8.09347723833216e-06, "loss": 0.4436, "step": 3365 }, { "epoch": 1.077005119453925, "grad_norm": 0.5866360664367676, "learning_rate": 8.092014340296215e-06, "loss": 0.5131, "step": 3366 }, { "epoch": 1.077325085324232, "grad_norm": 0.5072371363639832, "learning_rate": 8.090551013548762e-06, "loss": 0.4412, "step": 3367 }, { "epoch": 1.0776450511945392, "grad_norm": 0.525693953037262, "learning_rate": 8.089087258292693e-06, "loss": 0.4848, "step": 3368 }, { "epoch": 1.0779650170648465, "grad_norm": 0.5036314129829407, "learning_rate": 8.08762307473096e-06, "loss": 0.5591, "step": 3369 }, { "epoch": 1.0782849829351535, "grad_norm": 0.5551751852035522, "learning_rate": 8.086158463066574e-06, "loss": 0.4882, "step": 3370 }, { "epoch": 1.0786049488054608, "grad_norm": 0.6021180748939514, "learning_rate": 8.084693423502602e-06, "loss": 0.5101, "step": 3371 }, { "epoch": 1.078924914675768, "grad_norm": 0.5439380407333374, "learning_rate": 8.083227956242179e-06, "loss": 0.4716, "step": 3372 }, { "epoch": 1.079244880546075, "grad_norm": 0.5912663340568542, "learning_rate": 8.081762061488489e-06, "loss": 0.515, "step": 3373 }, { "epoch": 1.0795648464163823, "grad_norm": 0.5152432322502136, "learning_rate": 8.080295739444782e-06, "loss": 0.4317, "step": 3374 }, { "epoch": 1.0798848122866893, "grad_norm": 0.5591632127761841, "learning_rate": 8.078828990314364e-06, "loss": 0.5434, "step": 3375 }, { "epoch": 1.0802047781569966, "grad_norm": 0.5563610792160034, "learning_rate": 8.077361814300602e-06, "loss": 0.5485, "step": 3376 }, { "epoch": 1.0805247440273038, "grad_norm": 0.5657890439033508, "learning_rate": 8.075894211606925e-06, "loss": 0.4926, "step": 3377 }, { "epoch": 1.0808447098976108, "grad_norm": 0.5712653994560242, "learning_rate": 8.074426182436812e-06, "loss": 0.4851, "step": 3378 }, { "epoch": 1.081164675767918, "grad_norm": 0.6388885378837585, "learning_rate": 8.072957726993813e-06, "loss": 0.55, "step": 3379 }, { "epoch": 1.0814846416382253, "grad_norm": 0.5559564232826233, "learning_rate": 8.071488845481528e-06, "loss": 0.5251, "step": 3380 }, { "epoch": 1.0818046075085324, "grad_norm": 0.48583507537841797, "learning_rate": 8.070019538103617e-06, "loss": 0.4192, "step": 3381 }, { "epoch": 1.0821245733788396, "grad_norm": 0.6477534174919128, "learning_rate": 8.068549805063806e-06, "loss": 0.519, "step": 3382 }, { "epoch": 1.0824445392491469, "grad_norm": 0.6126707196235657, "learning_rate": 8.067079646565872e-06, "loss": 0.52, "step": 3383 }, { "epoch": 1.0827645051194539, "grad_norm": 0.5714107155799866, "learning_rate": 8.065609062813657e-06, "loss": 0.4332, "step": 3384 }, { "epoch": 1.0830844709897611, "grad_norm": 0.5787184238433838, "learning_rate": 8.064138054011056e-06, "loss": 0.4626, "step": 3385 }, { "epoch": 1.0834044368600684, "grad_norm": 0.5252422094345093, "learning_rate": 8.062666620362028e-06, "loss": 0.545, "step": 3386 }, { "epoch": 1.0837244027303754, "grad_norm": 0.5058318972587585, "learning_rate": 8.061194762070588e-06, "loss": 0.494, "step": 3387 }, { "epoch": 1.0840443686006827, "grad_norm": 0.6317284107208252, "learning_rate": 8.05972247934081e-06, "loss": 0.5148, "step": 3388 }, { "epoch": 1.0843643344709897, "grad_norm": 0.5603262782096863, "learning_rate": 8.058249772376831e-06, "loss": 0.4689, "step": 3389 }, { "epoch": 1.084684300341297, "grad_norm": 0.5218910574913025, "learning_rate": 8.056776641382844e-06, "loss": 0.495, "step": 3390 }, { "epoch": 1.0850042662116042, "grad_norm": 0.5703060626983643, "learning_rate": 8.055303086563095e-06, "loss": 0.5145, "step": 3391 }, { "epoch": 1.0853242320819112, "grad_norm": 0.5872288942337036, "learning_rate": 8.0538291081219e-06, "loss": 0.5014, "step": 3392 }, { "epoch": 1.0856441979522184, "grad_norm": 0.5883344411849976, "learning_rate": 8.052354706263624e-06, "loss": 0.5359, "step": 3393 }, { "epoch": 1.0859641638225257, "grad_norm": 0.5529842376708984, "learning_rate": 8.050879881192695e-06, "loss": 0.4936, "step": 3394 }, { "epoch": 1.0862841296928327, "grad_norm": 0.5603022575378418, "learning_rate": 8.0494046331136e-06, "loss": 0.5348, "step": 3395 }, { "epoch": 1.08660409556314, "grad_norm": 0.5294363498687744, "learning_rate": 8.047928962230886e-06, "loss": 0.4998, "step": 3396 }, { "epoch": 1.086924061433447, "grad_norm": 0.5693865418434143, "learning_rate": 8.046452868749155e-06, "loss": 0.4972, "step": 3397 }, { "epoch": 1.0872440273037542, "grad_norm": 0.5111364722251892, "learning_rate": 8.044976352873066e-06, "loss": 0.4359, "step": 3398 }, { "epoch": 1.0875639931740615, "grad_norm": 0.5233534574508667, "learning_rate": 8.043499414807344e-06, "loss": 0.5259, "step": 3399 }, { "epoch": 1.0878839590443685, "grad_norm": 0.4962422549724579, "learning_rate": 8.042022054756765e-06, "loss": 0.4379, "step": 3400 }, { "epoch": 1.0882039249146758, "grad_norm": 0.5874216556549072, "learning_rate": 8.040544272926171e-06, "loss": 0.5349, "step": 3401 }, { "epoch": 1.088523890784983, "grad_norm": 0.6073318123817444, "learning_rate": 8.039066069520455e-06, "loss": 0.5709, "step": 3402 }, { "epoch": 1.08884385665529, "grad_norm": 0.5323451161384583, "learning_rate": 8.037587444744573e-06, "loss": 0.4819, "step": 3403 }, { "epoch": 1.0891638225255973, "grad_norm": 0.5540988445281982, "learning_rate": 8.036108398803536e-06, "loss": 0.4548, "step": 3404 }, { "epoch": 1.0894837883959045, "grad_norm": 0.545438289642334, "learning_rate": 8.034628931902419e-06, "loss": 0.4836, "step": 3405 }, { "epoch": 1.0898037542662116, "grad_norm": 0.5775328278541565, "learning_rate": 8.033149044246348e-06, "loss": 0.5394, "step": 3406 }, { "epoch": 1.0901237201365188, "grad_norm": 0.49468064308166504, "learning_rate": 8.031668736040515e-06, "loss": 0.4658, "step": 3407 }, { "epoch": 1.0904436860068258, "grad_norm": 0.5362755656242371, "learning_rate": 8.030188007490164e-06, "loss": 0.5005, "step": 3408 }, { "epoch": 1.090763651877133, "grad_norm": 0.5736452341079712, "learning_rate": 8.028706858800598e-06, "loss": 0.5454, "step": 3409 }, { "epoch": 1.0910836177474403, "grad_norm": 0.5717540383338928, "learning_rate": 8.027225290177182e-06, "loss": 0.4439, "step": 3410 }, { "epoch": 1.0914035836177474, "grad_norm": 0.5134483575820923, "learning_rate": 8.025743301825342e-06, "loss": 0.4817, "step": 3411 }, { "epoch": 1.0917235494880546, "grad_norm": 0.488443523645401, "learning_rate": 8.02426089395055e-06, "loss": 0.4982, "step": 3412 }, { "epoch": 1.0920435153583619, "grad_norm": 0.5470604300498962, "learning_rate": 8.022778066758348e-06, "loss": 0.503, "step": 3413 }, { "epoch": 1.0923634812286689, "grad_norm": 0.5137178301811218, "learning_rate": 8.02129482045433e-06, "loss": 0.5184, "step": 3414 }, { "epoch": 1.0926834470989761, "grad_norm": 0.5206054449081421, "learning_rate": 8.01981115524415e-06, "loss": 0.5315, "step": 3415 }, { "epoch": 1.0930034129692832, "grad_norm": 0.49802935123443604, "learning_rate": 8.018327071333521e-06, "loss": 0.4993, "step": 3416 }, { "epoch": 1.0933233788395904, "grad_norm": 0.4849085509777069, "learning_rate": 8.016842568928213e-06, "loss": 0.4391, "step": 3417 }, { "epoch": 1.0936433447098977, "grad_norm": 0.5910800695419312, "learning_rate": 8.015357648234052e-06, "loss": 0.5363, "step": 3418 }, { "epoch": 1.0939633105802047, "grad_norm": 0.6028000712394714, "learning_rate": 8.013872309456928e-06, "loss": 0.507, "step": 3419 }, { "epoch": 1.094283276450512, "grad_norm": 0.5333077907562256, "learning_rate": 8.01238655280278e-06, "loss": 0.5316, "step": 3420 }, { "epoch": 1.0946032423208192, "grad_norm": 0.5566905736923218, "learning_rate": 8.010900378477612e-06, "loss": 0.5112, "step": 3421 }, { "epoch": 1.0949232081911262, "grad_norm": 0.5991218686103821, "learning_rate": 8.009413786687486e-06, "loss": 0.481, "step": 3422 }, { "epoch": 1.0952431740614335, "grad_norm": 0.5846459865570068, "learning_rate": 8.007926777638519e-06, "loss": 0.5164, "step": 3423 }, { "epoch": 1.0955631399317407, "grad_norm": 0.603992760181427, "learning_rate": 8.006439351536884e-06, "loss": 0.51, "step": 3424 }, { "epoch": 1.0958831058020477, "grad_norm": 0.5503547787666321, "learning_rate": 8.004951508588817e-06, "loss": 0.4707, "step": 3425 }, { "epoch": 1.096203071672355, "grad_norm": 0.5133087635040283, "learning_rate": 8.003463249000608e-06, "loss": 0.5167, "step": 3426 }, { "epoch": 1.0965230375426622, "grad_norm": 0.5910055637359619, "learning_rate": 8.001974572978605e-06, "loss": 0.485, "step": 3427 }, { "epoch": 1.0968430034129693, "grad_norm": 0.5813104510307312, "learning_rate": 8.000485480729217e-06, "loss": 0.4728, "step": 3428 }, { "epoch": 1.0971629692832765, "grad_norm": 0.5279648303985596, "learning_rate": 7.998995972458907e-06, "loss": 0.4882, "step": 3429 }, { "epoch": 1.0974829351535835, "grad_norm": 0.5567468404769897, "learning_rate": 7.997506048374198e-06, "loss": 0.4974, "step": 3430 }, { "epoch": 1.0978029010238908, "grad_norm": 0.5733569264411926, "learning_rate": 7.996015708681671e-06, "loss": 0.5255, "step": 3431 }, { "epoch": 1.098122866894198, "grad_norm": 0.46804070472717285, "learning_rate": 7.994524953587959e-06, "loss": 0.4432, "step": 3432 }, { "epoch": 1.098442832764505, "grad_norm": 0.5812718272209167, "learning_rate": 7.99303378329976e-06, "loss": 0.5642, "step": 3433 }, { "epoch": 1.0987627986348123, "grad_norm": 0.5907002687454224, "learning_rate": 7.991542198023827e-06, "loss": 0.5194, "step": 3434 }, { "epoch": 1.0990827645051195, "grad_norm": 0.5492275357246399, "learning_rate": 7.99005019796697e-06, "loss": 0.4919, "step": 3435 }, { "epoch": 1.0994027303754266, "grad_norm": 0.586290717124939, "learning_rate": 7.988557783336055e-06, "loss": 0.4975, "step": 3436 }, { "epoch": 1.0997226962457338, "grad_norm": 0.6038715839385986, "learning_rate": 7.987064954338007e-06, "loss": 0.5042, "step": 3437 }, { "epoch": 1.1000426621160408, "grad_norm": 0.5693135857582092, "learning_rate": 7.985571711179812e-06, "loss": 0.5071, "step": 3438 }, { "epoch": 1.100362627986348, "grad_norm": 0.5785577297210693, "learning_rate": 7.984078054068505e-06, "loss": 0.4957, "step": 3439 }, { "epoch": 1.1006825938566553, "grad_norm": 0.5374515056610107, "learning_rate": 7.982583983211187e-06, "loss": 0.5037, "step": 3440 }, { "epoch": 1.1010025597269624, "grad_norm": 0.5297296643257141, "learning_rate": 7.981089498815013e-06, "loss": 0.4822, "step": 3441 }, { "epoch": 1.1013225255972696, "grad_norm": 0.5651179552078247, "learning_rate": 7.979594601087191e-06, "loss": 0.496, "step": 3442 }, { "epoch": 1.1016424914675769, "grad_norm": 0.5637895464897156, "learning_rate": 7.978099290234995e-06, "loss": 0.4916, "step": 3443 }, { "epoch": 1.101962457337884, "grad_norm": 0.5521149039268494, "learning_rate": 7.976603566465747e-06, "loss": 0.4709, "step": 3444 }, { "epoch": 1.1022824232081911, "grad_norm": 0.4833112359046936, "learning_rate": 7.975107429986836e-06, "loss": 0.4566, "step": 3445 }, { "epoch": 1.1026023890784984, "grad_norm": 0.5734673142433167, "learning_rate": 7.973610881005702e-06, "loss": 0.518, "step": 3446 }, { "epoch": 1.1029223549488054, "grad_norm": 0.5697218775749207, "learning_rate": 7.97211391972984e-06, "loss": 0.515, "step": 3447 }, { "epoch": 1.1032423208191127, "grad_norm": 0.5977779030799866, "learning_rate": 7.970616546366809e-06, "loss": 0.5046, "step": 3448 }, { "epoch": 1.1035622866894197, "grad_norm": 0.5142936110496521, "learning_rate": 7.969118761124217e-06, "loss": 0.4758, "step": 3449 }, { "epoch": 1.103882252559727, "grad_norm": 0.5793342590332031, "learning_rate": 7.96762056420974e-06, "loss": 0.5176, "step": 3450 }, { "epoch": 1.1042022184300342, "grad_norm": 0.5544641613960266, "learning_rate": 7.9661219558311e-06, "loss": 0.4681, "step": 3451 }, { "epoch": 1.1045221843003412, "grad_norm": 0.5324788689613342, "learning_rate": 7.964622936196082e-06, "loss": 0.4927, "step": 3452 }, { "epoch": 1.1048421501706485, "grad_norm": 0.5800503492355347, "learning_rate": 7.963123505512529e-06, "loss": 0.501, "step": 3453 }, { "epoch": 1.1051621160409557, "grad_norm": 0.6092519164085388, "learning_rate": 7.961623663988336e-06, "loss": 0.4742, "step": 3454 }, { "epoch": 1.1054820819112627, "grad_norm": 0.5879743695259094, "learning_rate": 7.960123411831459e-06, "loss": 0.6033, "step": 3455 }, { "epoch": 1.10580204778157, "grad_norm": 0.5198939442634583, "learning_rate": 7.958622749249911e-06, "loss": 0.4904, "step": 3456 }, { "epoch": 1.1061220136518772, "grad_norm": 0.48006105422973633, "learning_rate": 7.95712167645176e-06, "loss": 0.4164, "step": 3457 }, { "epoch": 1.1064419795221843, "grad_norm": 0.5257875323295593, "learning_rate": 7.95562019364513e-06, "loss": 0.4833, "step": 3458 }, { "epoch": 1.1067619453924915, "grad_norm": 0.5582365989685059, "learning_rate": 7.954118301038203e-06, "loss": 0.521, "step": 3459 }, { "epoch": 1.1070819112627985, "grad_norm": 0.5165651440620422, "learning_rate": 7.952615998839222e-06, "loss": 0.5118, "step": 3460 }, { "epoch": 1.1074018771331058, "grad_norm": 0.5487995147705078, "learning_rate": 7.95111328725648e-06, "loss": 0.4862, "step": 3461 }, { "epoch": 1.107721843003413, "grad_norm": 0.5739307999610901, "learning_rate": 7.949610166498329e-06, "loss": 0.5425, "step": 3462 }, { "epoch": 1.10804180887372, "grad_norm": 0.5141011476516724, "learning_rate": 7.948106636773184e-06, "loss": 0.4642, "step": 3463 }, { "epoch": 1.1083617747440273, "grad_norm": 0.4784911274909973, "learning_rate": 7.946602698289503e-06, "loss": 0.4384, "step": 3464 }, { "epoch": 1.1086817406143346, "grad_norm": 0.5456899404525757, "learning_rate": 7.945098351255814e-06, "loss": 0.5388, "step": 3465 }, { "epoch": 1.1090017064846416, "grad_norm": 0.520692765712738, "learning_rate": 7.943593595880696e-06, "loss": 0.5118, "step": 3466 }, { "epoch": 1.1093216723549488, "grad_norm": 0.5141918063163757, "learning_rate": 7.942088432372784e-06, "loss": 0.4171, "step": 3467 }, { "epoch": 1.109641638225256, "grad_norm": 0.5195292234420776, "learning_rate": 7.940582860940771e-06, "loss": 0.5243, "step": 3468 }, { "epoch": 1.109961604095563, "grad_norm": 0.5189877152442932, "learning_rate": 7.939076881793408e-06, "loss": 0.4453, "step": 3469 }, { "epoch": 1.1102815699658704, "grad_norm": 0.5618454813957214, "learning_rate": 7.937570495139499e-06, "loss": 0.5297, "step": 3470 }, { "epoch": 1.1106015358361774, "grad_norm": 0.509519636631012, "learning_rate": 7.936063701187908e-06, "loss": 0.4521, "step": 3471 }, { "epoch": 1.1109215017064846, "grad_norm": 0.5387877225875854, "learning_rate": 7.934556500147553e-06, "loss": 0.5053, "step": 3472 }, { "epoch": 1.1112414675767919, "grad_norm": 0.5411096811294556, "learning_rate": 7.933048892227406e-06, "loss": 0.516, "step": 3473 }, { "epoch": 1.111561433447099, "grad_norm": 0.6244606375694275, "learning_rate": 7.931540877636503e-06, "loss": 0.5462, "step": 3474 }, { "epoch": 1.1118813993174061, "grad_norm": 0.5785732865333557, "learning_rate": 7.930032456583931e-06, "loss": 0.4832, "step": 3475 }, { "epoch": 1.1122013651877134, "grad_norm": 0.5496439337730408, "learning_rate": 7.928523629278835e-06, "loss": 0.5213, "step": 3476 }, { "epoch": 1.1125213310580204, "grad_norm": 0.5574624538421631, "learning_rate": 7.927014395930412e-06, "loss": 0.4723, "step": 3477 }, { "epoch": 1.1128412969283277, "grad_norm": 0.5714210271835327, "learning_rate": 7.925504756747924e-06, "loss": 0.5104, "step": 3478 }, { "epoch": 1.1131612627986347, "grad_norm": 0.5857492685317993, "learning_rate": 7.923994711940682e-06, "loss": 0.6138, "step": 3479 }, { "epoch": 1.113481228668942, "grad_norm": 0.5344730615615845, "learning_rate": 7.922484261718055e-06, "loss": 0.4384, "step": 3480 }, { "epoch": 1.1138011945392492, "grad_norm": 0.528183102607727, "learning_rate": 7.920973406289469e-06, "loss": 0.5313, "step": 3481 }, { "epoch": 1.1141211604095562, "grad_norm": 0.6133708953857422, "learning_rate": 7.919462145864405e-06, "loss": 0.509, "step": 3482 }, { "epoch": 1.1144411262798635, "grad_norm": 0.5474709868431091, "learning_rate": 7.917950480652403e-06, "loss": 0.4831, "step": 3483 }, { "epoch": 1.1147610921501707, "grad_norm": 0.6181241869926453, "learning_rate": 7.916438410863058e-06, "loss": 0.5362, "step": 3484 }, { "epoch": 1.1150810580204777, "grad_norm": 0.5898253321647644, "learning_rate": 7.914925936706018e-06, "loss": 0.49, "step": 3485 }, { "epoch": 1.115401023890785, "grad_norm": 0.619894802570343, "learning_rate": 7.913413058390989e-06, "loss": 0.504, "step": 3486 }, { "epoch": 1.1157209897610922, "grad_norm": 0.5624905228614807, "learning_rate": 7.911899776127736e-06, "loss": 0.523, "step": 3487 }, { "epoch": 1.1160409556313993, "grad_norm": 0.5253486037254333, "learning_rate": 7.910386090126076e-06, "loss": 0.4276, "step": 3488 }, { "epoch": 1.1163609215017065, "grad_norm": 0.6467494964599609, "learning_rate": 7.908872000595882e-06, "loss": 0.5575, "step": 3489 }, { "epoch": 1.1166808873720138, "grad_norm": 0.5251191854476929, "learning_rate": 7.907357507747087e-06, "loss": 0.4818, "step": 3490 }, { "epoch": 1.1170008532423208, "grad_norm": 0.636448860168457, "learning_rate": 7.905842611789675e-06, "loss": 0.5822, "step": 3491 }, { "epoch": 1.117320819112628, "grad_norm": 0.5673524141311646, "learning_rate": 7.904327312933688e-06, "loss": 0.522, "step": 3492 }, { "epoch": 1.117640784982935, "grad_norm": 0.569740355014801, "learning_rate": 7.902811611389227e-06, "loss": 0.4544, "step": 3493 }, { "epoch": 1.1179607508532423, "grad_norm": 0.613383412361145, "learning_rate": 7.901295507366445e-06, "loss": 0.5084, "step": 3494 }, { "epoch": 1.1182807167235496, "grad_norm": 0.5107038617134094, "learning_rate": 7.899779001075547e-06, "loss": 0.4043, "step": 3495 }, { "epoch": 1.1186006825938566, "grad_norm": 0.5148925185203552, "learning_rate": 7.898262092726805e-06, "loss": 0.5596, "step": 3496 }, { "epoch": 1.1189206484641638, "grad_norm": 0.5359509587287903, "learning_rate": 7.896744782530535e-06, "loss": 0.5055, "step": 3497 }, { "epoch": 1.119240614334471, "grad_norm": 0.5600014328956604, "learning_rate": 7.895227070697118e-06, "loss": 0.4898, "step": 3498 }, { "epoch": 1.119560580204778, "grad_norm": 0.5266281962394714, "learning_rate": 7.893708957436982e-06, "loss": 0.4846, "step": 3499 }, { "epoch": 1.1198805460750854, "grad_norm": 0.5474272966384888, "learning_rate": 7.892190442960621e-06, "loss": 0.4776, "step": 3500 }, { "epoch": 1.1202005119453924, "grad_norm": 0.5400792956352234, "learning_rate": 7.890671527478575e-06, "loss": 0.5099, "step": 3501 }, { "epoch": 1.1205204778156996, "grad_norm": 0.528954267501831, "learning_rate": 7.889152211201442e-06, "loss": 0.5321, "step": 3502 }, { "epoch": 1.1208404436860069, "grad_norm": 0.5811314582824707, "learning_rate": 7.887632494339881e-06, "loss": 0.4551, "step": 3503 }, { "epoch": 1.121160409556314, "grad_norm": 0.5416238903999329, "learning_rate": 7.8861123771046e-06, "loss": 0.4848, "step": 3504 }, { "epoch": 1.1214803754266212, "grad_norm": 0.5491629838943481, "learning_rate": 7.884591859706365e-06, "loss": 0.5012, "step": 3505 }, { "epoch": 1.1218003412969284, "grad_norm": 0.5045695900917053, "learning_rate": 7.883070942356001e-06, "loss": 0.4729, "step": 3506 }, { "epoch": 1.1221203071672354, "grad_norm": 0.5594595074653625, "learning_rate": 7.88154962526438e-06, "loss": 0.524, "step": 3507 }, { "epoch": 1.1224402730375427, "grad_norm": 0.5319141745567322, "learning_rate": 7.880027908642439e-06, "loss": 0.5206, "step": 3508 }, { "epoch": 1.12276023890785, "grad_norm": 0.5062196850776672, "learning_rate": 7.878505792701162e-06, "loss": 0.5039, "step": 3509 }, { "epoch": 1.123080204778157, "grad_norm": 0.5017806887626648, "learning_rate": 7.876983277651592e-06, "loss": 0.4911, "step": 3510 }, { "epoch": 1.1234001706484642, "grad_norm": 0.45875391364097595, "learning_rate": 7.875460363704834e-06, "loss": 0.4715, "step": 3511 }, { "epoch": 1.1237201365187712, "grad_norm": 0.4955386817455292, "learning_rate": 7.873937051072037e-06, "loss": 0.5038, "step": 3512 }, { "epoch": 1.1240401023890785, "grad_norm": 0.555153489112854, "learning_rate": 7.872413339964407e-06, "loss": 0.4824, "step": 3513 }, { "epoch": 1.1243600682593857, "grad_norm": 0.5983140468597412, "learning_rate": 7.870889230593213e-06, "loss": 0.5636, "step": 3514 }, { "epoch": 1.1246800341296928, "grad_norm": 0.5900697112083435, "learning_rate": 7.869364723169774e-06, "loss": 0.5056, "step": 3515 }, { "epoch": 1.125, "grad_norm": 0.5284932851791382, "learning_rate": 7.867839817905466e-06, "loss": 0.4837, "step": 3516 }, { "epoch": 1.1253199658703072, "grad_norm": 0.6077491641044617, "learning_rate": 7.866314515011713e-06, "loss": 0.5123, "step": 3517 }, { "epoch": 1.1256399317406143, "grad_norm": 0.5993654131889343, "learning_rate": 7.864788814700006e-06, "loss": 0.5094, "step": 3518 }, { "epoch": 1.1259598976109215, "grad_norm": 0.6185998320579529, "learning_rate": 7.863262717181884e-06, "loss": 0.548, "step": 3519 }, { "epoch": 1.1262798634812285, "grad_norm": 0.5267618894577026, "learning_rate": 7.861736222668943e-06, "loss": 0.542, "step": 3520 }, { "epoch": 1.1265998293515358, "grad_norm": 0.5616210103034973, "learning_rate": 7.86020933137283e-06, "loss": 0.5353, "step": 3521 }, { "epoch": 1.126919795221843, "grad_norm": 0.5184577703475952, "learning_rate": 7.858682043505253e-06, "loss": 0.4946, "step": 3522 }, { "epoch": 1.12723976109215, "grad_norm": 0.5412434339523315, "learning_rate": 7.857154359277972e-06, "loss": 0.512, "step": 3523 }, { "epoch": 1.1275597269624573, "grad_norm": 0.5075342655181885, "learning_rate": 7.855626278902799e-06, "loss": 0.4858, "step": 3524 }, { "epoch": 1.1278796928327646, "grad_norm": 0.6242489814758301, "learning_rate": 7.85409780259161e-06, "loss": 0.5285, "step": 3525 }, { "epoch": 1.1281996587030716, "grad_norm": 0.5738124847412109, "learning_rate": 7.852568930556328e-06, "loss": 0.4731, "step": 3526 }, { "epoch": 1.1285196245733788, "grad_norm": 0.551703691482544, "learning_rate": 7.851039663008929e-06, "loss": 0.5013, "step": 3527 }, { "epoch": 1.128839590443686, "grad_norm": 0.5525521039962769, "learning_rate": 7.849510000161454e-06, "loss": 0.5382, "step": 3528 }, { "epoch": 1.1291595563139931, "grad_norm": 0.5638213157653809, "learning_rate": 7.847979942225985e-06, "loss": 0.4689, "step": 3529 }, { "epoch": 1.1294795221843004, "grad_norm": 0.5558946132659912, "learning_rate": 7.846449489414675e-06, "loss": 0.4681, "step": 3530 }, { "epoch": 1.1297994880546076, "grad_norm": 0.583815336227417, "learning_rate": 7.844918641939717e-06, "loss": 0.5684, "step": 3531 }, { "epoch": 1.1301194539249146, "grad_norm": 0.6019387245178223, "learning_rate": 7.843387400013367e-06, "loss": 0.4637, "step": 3532 }, { "epoch": 1.130439419795222, "grad_norm": 0.5469242334365845, "learning_rate": 7.841855763847934e-06, "loss": 0.5238, "step": 3533 }, { "epoch": 1.130759385665529, "grad_norm": 0.5761325359344482, "learning_rate": 7.84032373365578e-06, "loss": 0.508, "step": 3534 }, { "epoch": 1.1310793515358362, "grad_norm": 0.6610817909240723, "learning_rate": 7.838791309649324e-06, "loss": 0.5173, "step": 3535 }, { "epoch": 1.1313993174061434, "grad_norm": 0.588883101940155, "learning_rate": 7.837258492041036e-06, "loss": 0.4501, "step": 3536 }, { "epoch": 1.1317192832764504, "grad_norm": 0.5269465446472168, "learning_rate": 7.835725281043447e-06, "loss": 0.5181, "step": 3537 }, { "epoch": 1.1320392491467577, "grad_norm": 0.6146968007087708, "learning_rate": 7.834191676869135e-06, "loss": 0.5121, "step": 3538 }, { "epoch": 1.132359215017065, "grad_norm": 0.568915843963623, "learning_rate": 7.832657679730738e-06, "loss": 0.4973, "step": 3539 }, { "epoch": 1.132679180887372, "grad_norm": 0.5202858448028564, "learning_rate": 7.831123289840946e-06, "loss": 0.4665, "step": 3540 }, { "epoch": 1.1329991467576792, "grad_norm": 0.5461797118186951, "learning_rate": 7.829588507412503e-06, "loss": 0.4784, "step": 3541 }, { "epoch": 1.1333191126279862, "grad_norm": 0.5784962773323059, "learning_rate": 7.82805333265821e-06, "loss": 0.4863, "step": 3542 }, { "epoch": 1.1336390784982935, "grad_norm": 0.5158717036247253, "learning_rate": 7.826517765790919e-06, "loss": 0.4341, "step": 3543 }, { "epoch": 1.1339590443686007, "grad_norm": 0.6058448553085327, "learning_rate": 7.82498180702354e-06, "loss": 0.5447, "step": 3544 }, { "epoch": 1.1342790102389078, "grad_norm": 0.545375645160675, "learning_rate": 7.823445456569036e-06, "loss": 0.4691, "step": 3545 }, { "epoch": 1.134598976109215, "grad_norm": 0.6026395559310913, "learning_rate": 7.821908714640421e-06, "loss": 0.5708, "step": 3546 }, { "epoch": 1.1349189419795223, "grad_norm": 0.5096762776374817, "learning_rate": 7.82037158145077e-06, "loss": 0.446, "step": 3547 }, { "epoch": 1.1352389078498293, "grad_norm": 0.5593512058258057, "learning_rate": 7.818834057213205e-06, "loss": 0.4804, "step": 3548 }, { "epoch": 1.1355588737201365, "grad_norm": 0.5331721901893616, "learning_rate": 7.817296142140907e-06, "loss": 0.526, "step": 3549 }, { "epoch": 1.1358788395904438, "grad_norm": 0.5482791066169739, "learning_rate": 7.81575783644711e-06, "loss": 0.499, "step": 3550 }, { "epoch": 1.1361988054607508, "grad_norm": 0.48665091395378113, "learning_rate": 7.8142191403451e-06, "loss": 0.4608, "step": 3551 }, { "epoch": 1.136518771331058, "grad_norm": 0.5491358041763306, "learning_rate": 7.812680054048223e-06, "loss": 0.5322, "step": 3552 }, { "epoch": 1.1368387372013653, "grad_norm": 0.5035708546638489, "learning_rate": 7.81114057776987e-06, "loss": 0.4729, "step": 3553 }, { "epoch": 1.1371587030716723, "grad_norm": 0.5579684972763062, "learning_rate": 7.809600711723497e-06, "loss": 0.533, "step": 3554 }, { "epoch": 1.1374786689419796, "grad_norm": 0.5616936683654785, "learning_rate": 7.808060456122603e-06, "loss": 0.499, "step": 3555 }, { "epoch": 1.1377986348122866, "grad_norm": 0.5179516077041626, "learning_rate": 7.80651981118075e-06, "loss": 0.4782, "step": 3556 }, { "epoch": 1.1381186006825939, "grad_norm": 0.5304527878761292, "learning_rate": 7.804978777111549e-06, "loss": 0.4879, "step": 3557 }, { "epoch": 1.138438566552901, "grad_norm": 0.5091323256492615, "learning_rate": 7.803437354128665e-06, "loss": 0.4899, "step": 3558 }, { "epoch": 1.1387585324232081, "grad_norm": 0.5082446336746216, "learning_rate": 7.801895542445819e-06, "loss": 0.5065, "step": 3559 }, { "epoch": 1.1390784982935154, "grad_norm": 0.4766141176223755, "learning_rate": 7.800353342276787e-06, "loss": 0.469, "step": 3560 }, { "epoch": 1.1393984641638226, "grad_norm": 0.5594205856323242, "learning_rate": 7.798810753835392e-06, "loss": 0.5247, "step": 3561 }, { "epoch": 1.1397184300341296, "grad_norm": 0.5137618780136108, "learning_rate": 7.797267777335523e-06, "loss": 0.4822, "step": 3562 }, { "epoch": 1.140038395904437, "grad_norm": 0.5178728699684143, "learning_rate": 7.795724412991108e-06, "loss": 0.5045, "step": 3563 }, { "epoch": 1.140358361774744, "grad_norm": 0.5358498692512512, "learning_rate": 7.794180661016143e-06, "loss": 0.4849, "step": 3564 }, { "epoch": 1.1406783276450512, "grad_norm": 0.4935292601585388, "learning_rate": 7.792636521624665e-06, "loss": 0.4566, "step": 3565 }, { "epoch": 1.1409982935153584, "grad_norm": 0.5339186191558838, "learning_rate": 7.791091995030775e-06, "loss": 0.5159, "step": 3566 }, { "epoch": 1.1413182593856654, "grad_norm": 0.5401326417922974, "learning_rate": 7.789547081448622e-06, "loss": 0.4636, "step": 3567 }, { "epoch": 1.1416382252559727, "grad_norm": 0.6003397703170776, "learning_rate": 7.78800178109241e-06, "loss": 0.5024, "step": 3568 }, { "epoch": 1.14195819112628, "grad_norm": 0.5978726744651794, "learning_rate": 7.786456094176398e-06, "loss": 0.4962, "step": 3569 }, { "epoch": 1.142278156996587, "grad_norm": 0.6008307933807373, "learning_rate": 7.784910020914895e-06, "loss": 0.5261, "step": 3570 }, { "epoch": 1.1425981228668942, "grad_norm": 0.5893910527229309, "learning_rate": 7.783363561522269e-06, "loss": 0.5164, "step": 3571 }, { "epoch": 1.1429180887372015, "grad_norm": 0.5838620662689209, "learning_rate": 7.781816716212935e-06, "loss": 0.5484, "step": 3572 }, { "epoch": 1.1432380546075085, "grad_norm": 0.5458717942237854, "learning_rate": 7.780269485201366e-06, "loss": 0.4492, "step": 3573 }, { "epoch": 1.1435580204778157, "grad_norm": 0.5298654437065125, "learning_rate": 7.778721868702091e-06, "loss": 0.4869, "step": 3574 }, { "epoch": 1.143877986348123, "grad_norm": 0.6417195796966553, "learning_rate": 7.777173866929682e-06, "loss": 0.5718, "step": 3575 }, { "epoch": 1.14419795221843, "grad_norm": 0.5706183910369873, "learning_rate": 7.775625480098778e-06, "loss": 0.4595, "step": 3576 }, { "epoch": 1.1445179180887373, "grad_norm": 0.5745950937271118, "learning_rate": 7.774076708424062e-06, "loss": 0.5629, "step": 3577 }, { "epoch": 1.1448378839590443, "grad_norm": 0.5032234787940979, "learning_rate": 7.772527552120274e-06, "loss": 0.4597, "step": 3578 }, { "epoch": 1.1451578498293515, "grad_norm": 0.5396010279655457, "learning_rate": 7.770978011402204e-06, "loss": 0.5109, "step": 3579 }, { "epoch": 1.1454778156996588, "grad_norm": 0.5810889601707458, "learning_rate": 7.769428086484699e-06, "loss": 0.4488, "step": 3580 }, { "epoch": 1.1457977815699658, "grad_norm": 0.6284360289573669, "learning_rate": 7.767877777582659e-06, "loss": 0.524, "step": 3581 }, { "epoch": 1.146117747440273, "grad_norm": 0.545884370803833, "learning_rate": 7.766327084911034e-06, "loss": 0.5034, "step": 3582 }, { "epoch": 1.14643771331058, "grad_norm": 0.5152574777603149, "learning_rate": 7.764776008684832e-06, "loss": 0.4454, "step": 3583 }, { "epoch": 1.1467576791808873, "grad_norm": 0.5437384843826294, "learning_rate": 7.763224549119111e-06, "loss": 0.4885, "step": 3584 }, { "epoch": 1.1470776450511946, "grad_norm": 0.5401154160499573, "learning_rate": 7.76167270642898e-06, "loss": 0.4683, "step": 3585 }, { "epoch": 1.1473976109215016, "grad_norm": 0.6209553480148315, "learning_rate": 7.760120480829606e-06, "loss": 0.5572, "step": 3586 }, { "epoch": 1.1477175767918089, "grad_norm": 0.5519247055053711, "learning_rate": 7.758567872536208e-06, "loss": 0.4868, "step": 3587 }, { "epoch": 1.148037542662116, "grad_norm": 0.5403838157653809, "learning_rate": 7.757014881764056e-06, "loss": 0.4766, "step": 3588 }, { "epoch": 1.1483575085324231, "grad_norm": 0.5646221041679382, "learning_rate": 7.755461508728473e-06, "loss": 0.5081, "step": 3589 }, { "epoch": 1.1486774744027304, "grad_norm": 0.5375046730041504, "learning_rate": 7.753907753644835e-06, "loss": 0.4697, "step": 3590 }, { "epoch": 1.1489974402730376, "grad_norm": 0.5387952327728271, "learning_rate": 7.752353616728579e-06, "loss": 0.4507, "step": 3591 }, { "epoch": 1.1493174061433447, "grad_norm": 0.5580173134803772, "learning_rate": 7.750799098195179e-06, "loss": 0.51, "step": 3592 }, { "epoch": 1.149637372013652, "grad_norm": 0.5314825773239136, "learning_rate": 7.749244198260175e-06, "loss": 0.5039, "step": 3593 }, { "epoch": 1.1499573378839592, "grad_norm": 0.6043481826782227, "learning_rate": 7.747688917139158e-06, "loss": 0.5187, "step": 3594 }, { "epoch": 1.1502773037542662, "grad_norm": 0.5296799540519714, "learning_rate": 7.746133255047764e-06, "loss": 0.4991, "step": 3595 }, { "epoch": 1.1505972696245734, "grad_norm": 0.5214899182319641, "learning_rate": 7.744577212201692e-06, "loss": 0.4695, "step": 3596 }, { "epoch": 1.1509172354948805, "grad_norm": 0.5552440881729126, "learning_rate": 7.743020788816688e-06, "loss": 0.5271, "step": 3597 }, { "epoch": 1.1512372013651877, "grad_norm": 0.6425376534461975, "learning_rate": 7.741463985108551e-06, "loss": 0.5198, "step": 3598 }, { "epoch": 1.151557167235495, "grad_norm": 0.5188428163528442, "learning_rate": 7.739906801293136e-06, "loss": 0.4517, "step": 3599 }, { "epoch": 1.151877133105802, "grad_norm": 0.5039048790931702, "learning_rate": 7.738349237586344e-06, "loss": 0.4634, "step": 3600 }, { "epoch": 1.1521970989761092, "grad_norm": 0.6108112931251526, "learning_rate": 7.736791294204138e-06, "loss": 0.5629, "step": 3601 }, { "epoch": 1.1525170648464165, "grad_norm": 0.6057872176170349, "learning_rate": 7.735232971362525e-06, "loss": 0.4689, "step": 3602 }, { "epoch": 1.1528370307167235, "grad_norm": 0.6644954681396484, "learning_rate": 7.733674269277572e-06, "loss": 0.5443, "step": 3603 }, { "epoch": 1.1531569965870307, "grad_norm": 0.5029540061950684, "learning_rate": 7.73211518816539e-06, "loss": 0.4546, "step": 3604 }, { "epoch": 1.1534769624573378, "grad_norm": 0.6114902496337891, "learning_rate": 7.730555728242154e-06, "loss": 0.522, "step": 3605 }, { "epoch": 1.153796928327645, "grad_norm": 0.5663495063781738, "learning_rate": 7.728995889724078e-06, "loss": 0.5292, "step": 3606 }, { "epoch": 1.1541168941979523, "grad_norm": 0.5311685800552368, "learning_rate": 7.72743567282744e-06, "loss": 0.4955, "step": 3607 }, { "epoch": 1.1544368600682593, "grad_norm": 0.5237186551094055, "learning_rate": 7.725875077768565e-06, "loss": 0.483, "step": 3608 }, { "epoch": 1.1547568259385665, "grad_norm": 0.5001371502876282, "learning_rate": 7.724314104763832e-06, "loss": 0.4528, "step": 3609 }, { "epoch": 1.1550767918088738, "grad_norm": 0.556122362613678, "learning_rate": 7.72275275402967e-06, "loss": 0.4903, "step": 3610 }, { "epoch": 1.1553967576791808, "grad_norm": 0.5611037611961365, "learning_rate": 7.721191025782563e-06, "loss": 0.4774, "step": 3611 }, { "epoch": 1.155716723549488, "grad_norm": 0.5195589661598206, "learning_rate": 7.719628920239047e-06, "loss": 0.4798, "step": 3612 }, { "epoch": 1.1560366894197953, "grad_norm": 0.5992692112922668, "learning_rate": 7.71806643761571e-06, "loss": 0.5276, "step": 3613 }, { "epoch": 1.1563566552901023, "grad_norm": 0.604767918586731, "learning_rate": 7.71650357812919e-06, "loss": 0.532, "step": 3614 }, { "epoch": 1.1566766211604096, "grad_norm": 0.5662189722061157, "learning_rate": 7.714940341996182e-06, "loss": 0.4664, "step": 3615 }, { "epoch": 1.1569965870307168, "grad_norm": 0.5372429490089417, "learning_rate": 7.71337672943343e-06, "loss": 0.4649, "step": 3616 }, { "epoch": 1.1573165529010239, "grad_norm": 0.5533605217933655, "learning_rate": 7.71181274065773e-06, "loss": 0.4805, "step": 3617 }, { "epoch": 1.1576365187713311, "grad_norm": 0.7256330251693726, "learning_rate": 7.71024837588593e-06, "loss": 0.5416, "step": 3618 }, { "epoch": 1.1579564846416381, "grad_norm": 0.5054036378860474, "learning_rate": 7.70868363533493e-06, "loss": 0.4353, "step": 3619 }, { "epoch": 1.1582764505119454, "grad_norm": 0.6028924584388733, "learning_rate": 7.70711851922169e-06, "loss": 0.5202, "step": 3620 }, { "epoch": 1.1585964163822526, "grad_norm": 0.5548515319824219, "learning_rate": 7.705553027763208e-06, "loss": 0.5081, "step": 3621 }, { "epoch": 1.1589163822525597, "grad_norm": 0.576492965221405, "learning_rate": 7.703987161176545e-06, "loss": 0.4899, "step": 3622 }, { "epoch": 1.159236348122867, "grad_norm": 0.5903343558311462, "learning_rate": 7.702420919678808e-06, "loss": 0.5091, "step": 3623 }, { "epoch": 1.159556313993174, "grad_norm": 0.5502042770385742, "learning_rate": 7.70085430348716e-06, "loss": 0.5349, "step": 3624 }, { "epoch": 1.1598762798634812, "grad_norm": 0.5461248159408569, "learning_rate": 7.699287312818812e-06, "loss": 0.4802, "step": 3625 }, { "epoch": 1.1601962457337884, "grad_norm": 0.5498080253601074, "learning_rate": 7.697719947891032e-06, "loss": 0.4375, "step": 3626 }, { "epoch": 1.1605162116040955, "grad_norm": 0.5582046508789062, "learning_rate": 7.696152208921137e-06, "loss": 0.5167, "step": 3627 }, { "epoch": 1.1608361774744027, "grad_norm": 0.5449654459953308, "learning_rate": 7.694584096126493e-06, "loss": 0.4915, "step": 3628 }, { "epoch": 1.16115614334471, "grad_norm": 0.5606305003166199, "learning_rate": 7.693015609724524e-06, "loss": 0.5151, "step": 3629 }, { "epoch": 1.161476109215017, "grad_norm": 0.5650143623352051, "learning_rate": 7.6914467499327e-06, "loss": 0.481, "step": 3630 }, { "epoch": 1.1617960750853242, "grad_norm": 0.5574379563331604, "learning_rate": 7.689877516968546e-06, "loss": 0.5309, "step": 3631 }, { "epoch": 1.1621160409556315, "grad_norm": 0.5204976797103882, "learning_rate": 7.688307911049639e-06, "loss": 0.4789, "step": 3632 }, { "epoch": 1.1624360068259385, "grad_norm": 0.5226881504058838, "learning_rate": 7.686737932393606e-06, "loss": 0.4807, "step": 3633 }, { "epoch": 1.1627559726962458, "grad_norm": 0.5622045993804932, "learning_rate": 7.685167581218125e-06, "loss": 0.4955, "step": 3634 }, { "epoch": 1.163075938566553, "grad_norm": 0.4869825839996338, "learning_rate": 7.683596857740929e-06, "loss": 0.49, "step": 3635 }, { "epoch": 1.16339590443686, "grad_norm": 0.5154707431793213, "learning_rate": 7.682025762179801e-06, "loss": 0.4621, "step": 3636 }, { "epoch": 1.1637158703071673, "grad_norm": 0.5349481105804443, "learning_rate": 7.680454294752574e-06, "loss": 0.5108, "step": 3637 }, { "epoch": 1.1640358361774745, "grad_norm": 0.5748913288116455, "learning_rate": 7.678882455677135e-06, "loss": 0.5004, "step": 3638 }, { "epoch": 1.1643558020477816, "grad_norm": 0.5858158469200134, "learning_rate": 7.67731024517142e-06, "loss": 0.4905, "step": 3639 }, { "epoch": 1.1646757679180888, "grad_norm": 0.5105371475219727, "learning_rate": 7.67573766345342e-06, "loss": 0.4657, "step": 3640 }, { "epoch": 1.1649957337883958, "grad_norm": 0.5406975150108337, "learning_rate": 7.674164710741172e-06, "loss": 0.4788, "step": 3641 }, { "epoch": 1.165315699658703, "grad_norm": 0.6753333806991577, "learning_rate": 7.672591387252773e-06, "loss": 0.5787, "step": 3642 }, { "epoch": 1.1656356655290103, "grad_norm": 0.529922366142273, "learning_rate": 7.671017693206358e-06, "loss": 0.5105, "step": 3643 }, { "epoch": 1.1659556313993173, "grad_norm": 0.6094779968261719, "learning_rate": 7.66944362882013e-06, "loss": 0.474, "step": 3644 }, { "epoch": 1.1662755972696246, "grad_norm": 0.5302444100379944, "learning_rate": 7.667869194312331e-06, "loss": 0.4518, "step": 3645 }, { "epoch": 1.1665955631399316, "grad_norm": 0.5726618766784668, "learning_rate": 7.666294389901258e-06, "loss": 0.4765, "step": 3646 }, { "epoch": 1.1669155290102389, "grad_norm": 0.5927723050117493, "learning_rate": 7.66471921580526e-06, "loss": 0.4839, "step": 3647 }, { "epoch": 1.1672354948805461, "grad_norm": 0.5504419207572937, "learning_rate": 7.663143672242739e-06, "loss": 0.5065, "step": 3648 }, { "epoch": 1.1675554607508531, "grad_norm": 0.5538895726203918, "learning_rate": 7.661567759432142e-06, "loss": 0.5142, "step": 3649 }, { "epoch": 1.1678754266211604, "grad_norm": 0.550895094871521, "learning_rate": 7.659991477591973e-06, "loss": 0.4485, "step": 3650 }, { "epoch": 1.1681953924914676, "grad_norm": 0.496063768863678, "learning_rate": 7.658414826940788e-06, "loss": 0.4729, "step": 3651 }, { "epoch": 1.1685153583617747, "grad_norm": 0.5706762075424194, "learning_rate": 7.656837807697187e-06, "loss": 0.5289, "step": 3652 }, { "epoch": 1.168835324232082, "grad_norm": 0.5495587587356567, "learning_rate": 7.655260420079827e-06, "loss": 0.4959, "step": 3653 }, { "epoch": 1.1691552901023892, "grad_norm": 0.5080969929695129, "learning_rate": 7.653682664307419e-06, "loss": 0.452, "step": 3654 }, { "epoch": 1.1694752559726962, "grad_norm": 0.531419038772583, "learning_rate": 7.652104540598712e-06, "loss": 0.4954, "step": 3655 }, { "epoch": 1.1697952218430034, "grad_norm": 0.5374448895454407, "learning_rate": 7.650526049172524e-06, "loss": 0.4881, "step": 3656 }, { "epoch": 1.1701151877133107, "grad_norm": 0.5262492895126343, "learning_rate": 7.648947190247709e-06, "loss": 0.4666, "step": 3657 }, { "epoch": 1.1704351535836177, "grad_norm": 0.5371428728103638, "learning_rate": 7.64736796404318e-06, "loss": 0.5373, "step": 3658 }, { "epoch": 1.170755119453925, "grad_norm": 0.5738182663917542, "learning_rate": 7.645788370777898e-06, "loss": 0.5619, "step": 3659 }, { "epoch": 1.171075085324232, "grad_norm": 0.6016698479652405, "learning_rate": 7.644208410670874e-06, "loss": 0.4897, "step": 3660 }, { "epoch": 1.1713950511945392, "grad_norm": 0.5825162529945374, "learning_rate": 7.642628083941172e-06, "loss": 0.5264, "step": 3661 }, { "epoch": 1.1717150170648465, "grad_norm": 0.5533202290534973, "learning_rate": 7.64104739080791e-06, "loss": 0.4943, "step": 3662 }, { "epoch": 1.1720349829351535, "grad_norm": 0.5478944778442383, "learning_rate": 7.639466331490248e-06, "loss": 0.5156, "step": 3663 }, { "epoch": 1.1723549488054608, "grad_norm": 0.5522754788398743, "learning_rate": 7.637884906207406e-06, "loss": 0.4601, "step": 3664 }, { "epoch": 1.172674914675768, "grad_norm": 0.5325472950935364, "learning_rate": 7.636303115178646e-06, "loss": 0.4643, "step": 3665 }, { "epoch": 1.172994880546075, "grad_norm": 0.5285465717315674, "learning_rate": 7.634720958623287e-06, "loss": 0.4952, "step": 3666 }, { "epoch": 1.1733148464163823, "grad_norm": 0.5523001551628113, "learning_rate": 7.633138436760702e-06, "loss": 0.5429, "step": 3667 }, { "epoch": 1.1736348122866893, "grad_norm": 0.5258068442344666, "learning_rate": 7.6315555498103e-06, "loss": 0.474, "step": 3668 }, { "epoch": 1.1739547781569966, "grad_norm": 0.5632274746894836, "learning_rate": 7.629972297991558e-06, "loss": 0.5199, "step": 3669 }, { "epoch": 1.1742747440273038, "grad_norm": 0.5483708381652832, "learning_rate": 7.62838868152399e-06, "loss": 0.5031, "step": 3670 }, { "epoch": 1.1745947098976108, "grad_norm": 0.5653208494186401, "learning_rate": 7.626804700627172e-06, "loss": 0.4494, "step": 3671 }, { "epoch": 1.174914675767918, "grad_norm": 0.530471920967102, "learning_rate": 7.625220355520722e-06, "loss": 0.521, "step": 3672 }, { "epoch": 1.1752346416382253, "grad_norm": 0.47522956132888794, "learning_rate": 7.62363564642431e-06, "loss": 0.4676, "step": 3673 }, { "epoch": 1.1755546075085324, "grad_norm": 0.583890438079834, "learning_rate": 7.62205057355766e-06, "loss": 0.5447, "step": 3674 }, { "epoch": 1.1758745733788396, "grad_norm": 0.5340621471405029, "learning_rate": 7.620465137140543e-06, "loss": 0.4847, "step": 3675 }, { "epoch": 1.1761945392491469, "grad_norm": 0.5879509449005127, "learning_rate": 7.6188793373927835e-06, "loss": 0.4956, "step": 3676 }, { "epoch": 1.1765145051194539, "grad_norm": 0.5591239333152771, "learning_rate": 7.617293174534252e-06, "loss": 0.5683, "step": 3677 }, { "epoch": 1.1768344709897611, "grad_norm": 0.5671486258506775, "learning_rate": 7.6157066487848745e-06, "loss": 0.4741, "step": 3678 }, { "epoch": 1.1771544368600684, "grad_norm": 0.5293091535568237, "learning_rate": 7.6141197603646225e-06, "loss": 0.438, "step": 3679 }, { "epoch": 1.1774744027303754, "grad_norm": 0.5160754323005676, "learning_rate": 7.612532509493521e-06, "loss": 0.4938, "step": 3680 }, { "epoch": 1.1777943686006827, "grad_norm": 0.5765207409858704, "learning_rate": 7.610944896391644e-06, "loss": 0.5045, "step": 3681 }, { "epoch": 1.1781143344709897, "grad_norm": 0.48545050621032715, "learning_rate": 7.609356921279116e-06, "loss": 0.4778, "step": 3682 }, { "epoch": 1.178434300341297, "grad_norm": 0.5499753355979919, "learning_rate": 7.607768584376112e-06, "loss": 0.5101, "step": 3683 }, { "epoch": 1.1787542662116042, "grad_norm": 0.6321439146995544, "learning_rate": 7.606179885902858e-06, "loss": 0.4998, "step": 3684 }, { "epoch": 1.1790742320819112, "grad_norm": 0.5480753779411316, "learning_rate": 7.604590826079625e-06, "loss": 0.4849, "step": 3685 }, { "epoch": 1.1793941979522184, "grad_norm": 0.5435191988945007, "learning_rate": 7.603001405126742e-06, "loss": 0.5338, "step": 3686 }, { "epoch": 1.1797141638225255, "grad_norm": 0.5904504060745239, "learning_rate": 7.601411623264582e-06, "loss": 0.4498, "step": 3687 }, { "epoch": 1.1800341296928327, "grad_norm": 0.5386573076248169, "learning_rate": 7.599821480713571e-06, "loss": 0.4663, "step": 3688 }, { "epoch": 1.18035409556314, "grad_norm": 0.5477933287620544, "learning_rate": 7.598230977694185e-06, "loss": 0.5051, "step": 3689 }, { "epoch": 1.180674061433447, "grad_norm": 0.5635031461715698, "learning_rate": 7.596640114426948e-06, "loss": 0.4759, "step": 3690 }, { "epoch": 1.1809940273037542, "grad_norm": 0.5476839542388916, "learning_rate": 7.5950488911324325e-06, "loss": 0.4915, "step": 3691 }, { "epoch": 1.1813139931740615, "grad_norm": 0.5442044734954834, "learning_rate": 7.593457308031269e-06, "loss": 0.515, "step": 3692 }, { "epoch": 1.1816339590443685, "grad_norm": 0.5077284574508667, "learning_rate": 7.591865365344129e-06, "loss": 0.4912, "step": 3693 }, { "epoch": 1.1819539249146758, "grad_norm": 0.46649423241615295, "learning_rate": 7.5902730632917395e-06, "loss": 0.4025, "step": 3694 }, { "epoch": 1.182273890784983, "grad_norm": 0.5748319029808044, "learning_rate": 7.5886804020948725e-06, "loss": 0.5741, "step": 3695 }, { "epoch": 1.18259385665529, "grad_norm": 0.5408945679664612, "learning_rate": 7.587087381974353e-06, "loss": 0.5067, "step": 3696 }, { "epoch": 1.1829138225255973, "grad_norm": 0.5399074554443359, "learning_rate": 7.585494003151055e-06, "loss": 0.4849, "step": 3697 }, { "epoch": 1.1832337883959045, "grad_norm": 0.5231366753578186, "learning_rate": 7.583900265845905e-06, "loss": 0.5007, "step": 3698 }, { "epoch": 1.1835537542662116, "grad_norm": 0.590194582939148, "learning_rate": 7.582306170279873e-06, "loss": 0.4894, "step": 3699 }, { "epoch": 1.1838737201365188, "grad_norm": 0.5386995673179626, "learning_rate": 7.580711716673985e-06, "loss": 0.5246, "step": 3700 }, { "epoch": 1.184193686006826, "grad_norm": 0.514801025390625, "learning_rate": 7.579116905249313e-06, "loss": 0.4177, "step": 3701 }, { "epoch": 1.184513651877133, "grad_norm": 0.5471378564834595, "learning_rate": 7.5775217362269785e-06, "loss": 0.5378, "step": 3702 }, { "epoch": 1.1848336177474403, "grad_norm": 0.5445446372032166, "learning_rate": 7.575926209828154e-06, "loss": 0.5192, "step": 3703 }, { "epoch": 1.1851535836177474, "grad_norm": 0.4958280324935913, "learning_rate": 7.574330326274062e-06, "loss": 0.5271, "step": 3704 }, { "epoch": 1.1854735494880546, "grad_norm": 0.5592751502990723, "learning_rate": 7.572734085785973e-06, "loss": 0.5534, "step": 3705 }, { "epoch": 1.1857935153583619, "grad_norm": 0.48439303040504456, "learning_rate": 7.57113748858521e-06, "loss": 0.4323, "step": 3706 }, { "epoch": 1.1861134812286689, "grad_norm": 0.5728739500045776, "learning_rate": 7.569540534893139e-06, "loss": 0.5274, "step": 3707 }, { "epoch": 1.1864334470989761, "grad_norm": 0.5586733222007751, "learning_rate": 7.567943224931182e-06, "loss": 0.484, "step": 3708 }, { "epoch": 1.1867534129692832, "grad_norm": 0.6303009986877441, "learning_rate": 7.566345558920807e-06, "loss": 0.4873, "step": 3709 }, { "epoch": 1.1870733788395904, "grad_norm": 0.5872066020965576, "learning_rate": 7.564747537083533e-06, "loss": 0.5474, "step": 3710 }, { "epoch": 1.1873933447098977, "grad_norm": 0.5944347977638245, "learning_rate": 7.563149159640929e-06, "loss": 0.5085, "step": 3711 }, { "epoch": 1.1877133105802047, "grad_norm": 0.5409294366836548, "learning_rate": 7.56155042681461e-06, "loss": 0.4931, "step": 3712 }, { "epoch": 1.188033276450512, "grad_norm": 0.49820324778556824, "learning_rate": 7.559951338826242e-06, "loss": 0.4691, "step": 3713 }, { "epoch": 1.1883532423208192, "grad_norm": 0.560949981212616, "learning_rate": 7.558351895897541e-06, "loss": 0.4733, "step": 3714 }, { "epoch": 1.1886732081911262, "grad_norm": 0.5346348285675049, "learning_rate": 7.556752098250272e-06, "loss": 0.5654, "step": 3715 }, { "epoch": 1.1889931740614335, "grad_norm": 0.5067862868309021, "learning_rate": 7.555151946106249e-06, "loss": 0.4771, "step": 3716 }, { "epoch": 1.1893131399317407, "grad_norm": 0.47606727480888367, "learning_rate": 7.553551439687335e-06, "loss": 0.4608, "step": 3717 }, { "epoch": 1.1896331058020477, "grad_norm": 0.6058302521705627, "learning_rate": 7.551950579215441e-06, "loss": 0.5519, "step": 3718 }, { "epoch": 1.189953071672355, "grad_norm": 0.5280526280403137, "learning_rate": 7.55034936491253e-06, "loss": 0.4647, "step": 3719 }, { "epoch": 1.1902730375426622, "grad_norm": 0.5691516995429993, "learning_rate": 7.548747797000611e-06, "loss": 0.529, "step": 3720 }, { "epoch": 1.1905930034129693, "grad_norm": 0.5301852226257324, "learning_rate": 7.547145875701744e-06, "loss": 0.489, "step": 3721 }, { "epoch": 1.1909129692832765, "grad_norm": 0.5888407230377197, "learning_rate": 7.545543601238037e-06, "loss": 0.49, "step": 3722 }, { "epoch": 1.1912329351535835, "grad_norm": 0.5133174061775208, "learning_rate": 7.543940973831647e-06, "loss": 0.5068, "step": 3723 }, { "epoch": 1.1915529010238908, "grad_norm": 0.4624699354171753, "learning_rate": 7.54233799370478e-06, "loss": 0.4747, "step": 3724 }, { "epoch": 1.191872866894198, "grad_norm": 0.6100415587425232, "learning_rate": 7.540734661079694e-06, "loss": 0.5976, "step": 3725 }, { "epoch": 1.192192832764505, "grad_norm": 0.5251874923706055, "learning_rate": 7.539130976178688e-06, "loss": 0.4889, "step": 3726 }, { "epoch": 1.1925127986348123, "grad_norm": 0.5343571901321411, "learning_rate": 7.5375269392241205e-06, "loss": 0.5279, "step": 3727 }, { "epoch": 1.1928327645051195, "grad_norm": 0.5854262113571167, "learning_rate": 7.535922550438387e-06, "loss": 0.4543, "step": 3728 }, { "epoch": 1.1931527303754266, "grad_norm": 0.5820112824440002, "learning_rate": 7.534317810043943e-06, "loss": 0.5291, "step": 3729 }, { "epoch": 1.1934726962457338, "grad_norm": 0.4850411117076874, "learning_rate": 7.532712718263285e-06, "loss": 0.478, "step": 3730 }, { "epoch": 1.1937926621160408, "grad_norm": 0.515326738357544, "learning_rate": 7.5311072753189604e-06, "loss": 0.4698, "step": 3731 }, { "epoch": 1.194112627986348, "grad_norm": 0.5624679923057556, "learning_rate": 7.52950148143357e-06, "loss": 0.5063, "step": 3732 }, { "epoch": 1.1944325938566553, "grad_norm": 0.556932270526886, "learning_rate": 7.527895336829754e-06, "loss": 0.5218, "step": 3733 }, { "epoch": 1.1947525597269624, "grad_norm": 0.565220296382904, "learning_rate": 7.526288841730209e-06, "loss": 0.4744, "step": 3734 }, { "epoch": 1.1950725255972696, "grad_norm": 0.5757796168327332, "learning_rate": 7.524681996357677e-06, "loss": 0.4932, "step": 3735 }, { "epoch": 1.1953924914675769, "grad_norm": 0.5540634989738464, "learning_rate": 7.523074800934948e-06, "loss": 0.53, "step": 3736 }, { "epoch": 1.195712457337884, "grad_norm": 0.518076479434967, "learning_rate": 7.5214672556848645e-06, "loss": 0.495, "step": 3737 }, { "epoch": 1.1960324232081911, "grad_norm": 0.5275093913078308, "learning_rate": 7.5198593608303106e-06, "loss": 0.5026, "step": 3738 }, { "epoch": 1.1963523890784984, "grad_norm": 0.5814018249511719, "learning_rate": 7.5182511165942264e-06, "loss": 0.5303, "step": 3739 }, { "epoch": 1.1966723549488054, "grad_norm": 0.5568544268608093, "learning_rate": 7.516642523199595e-06, "loss": 0.4707, "step": 3740 }, { "epoch": 1.1969923208191127, "grad_norm": 0.4932362139225006, "learning_rate": 7.51503358086945e-06, "loss": 0.4604, "step": 3741 }, { "epoch": 1.19731228668942, "grad_norm": 0.5840187072753906, "learning_rate": 7.513424289826874e-06, "loss": 0.5553, "step": 3742 }, { "epoch": 1.197632252559727, "grad_norm": 0.5801661610603333, "learning_rate": 7.511814650294994e-06, "loss": 0.5175, "step": 3743 }, { "epoch": 1.1979522184300342, "grad_norm": 0.5177812576293945, "learning_rate": 7.510204662496994e-06, "loss": 0.4402, "step": 3744 }, { "epoch": 1.1982721843003412, "grad_norm": 0.5612710118293762, "learning_rate": 7.508594326656096e-06, "loss": 0.5158, "step": 3745 }, { "epoch": 1.1985921501706485, "grad_norm": 0.5415658950805664, "learning_rate": 7.506983642995576e-06, "loss": 0.4528, "step": 3746 }, { "epoch": 1.1989121160409557, "grad_norm": 0.5858944058418274, "learning_rate": 7.50537261173876e-06, "loss": 0.5192, "step": 3747 }, { "epoch": 1.1992320819112627, "grad_norm": 0.5906977653503418, "learning_rate": 7.503761233109015e-06, "loss": 0.5243, "step": 3748 }, { "epoch": 1.19955204778157, "grad_norm": 0.5039230585098267, "learning_rate": 7.502149507329766e-06, "loss": 0.4875, "step": 3749 }, { "epoch": 1.199872013651877, "grad_norm": 0.5558406710624695, "learning_rate": 7.5005374346244755e-06, "loss": 0.4915, "step": 3750 }, { "epoch": 1.2001919795221843, "grad_norm": 0.5710980296134949, "learning_rate": 7.498925015216662e-06, "loss": 0.5284, "step": 3751 }, { "epoch": 1.2005119453924915, "grad_norm": 0.5781745910644531, "learning_rate": 7.497312249329889e-06, "loss": 0.4762, "step": 3752 }, { "epoch": 1.2008319112627985, "grad_norm": 0.5416224002838135, "learning_rate": 7.495699137187767e-06, "loss": 0.4787, "step": 3753 }, { "epoch": 1.2011518771331058, "grad_norm": 0.531855046749115, "learning_rate": 7.494085679013959e-06, "loss": 0.5433, "step": 3754 }, { "epoch": 1.201471843003413, "grad_norm": 0.5295029878616333, "learning_rate": 7.49247187503217e-06, "loss": 0.5452, "step": 3755 }, { "epoch": 1.20179180887372, "grad_norm": 0.5794286131858826, "learning_rate": 7.490857725466157e-06, "loss": 0.4521, "step": 3756 }, { "epoch": 1.2021117747440273, "grad_norm": 0.587912917137146, "learning_rate": 7.489243230539726e-06, "loss": 0.5056, "step": 3757 }, { "epoch": 1.2024317406143346, "grad_norm": 0.5290972590446472, "learning_rate": 7.487628390476724e-06, "loss": 0.5188, "step": 3758 }, { "epoch": 1.2027517064846416, "grad_norm": 0.5166901350021362, "learning_rate": 7.486013205501053e-06, "loss": 0.4563, "step": 3759 }, { "epoch": 1.2030716723549488, "grad_norm": 0.5781247615814209, "learning_rate": 7.484397675836661e-06, "loss": 0.5591, "step": 3760 }, { "epoch": 1.203391638225256, "grad_norm": 0.4947353005409241, "learning_rate": 7.482781801707543e-06, "loss": 0.4994, "step": 3761 }, { "epoch": 1.203711604095563, "grad_norm": 0.6305955648422241, "learning_rate": 7.481165583337741e-06, "loss": 0.5184, "step": 3762 }, { "epoch": 1.2040315699658704, "grad_norm": 0.5540664792060852, "learning_rate": 7.479549020951345e-06, "loss": 0.5269, "step": 3763 }, { "epoch": 1.2043515358361774, "grad_norm": 0.5958874225616455, "learning_rate": 7.477932114772495e-06, "loss": 0.4759, "step": 3764 }, { "epoch": 1.2046715017064846, "grad_norm": 0.6359137296676636, "learning_rate": 7.476314865025376e-06, "loss": 0.509, "step": 3765 }, { "epoch": 1.2049914675767919, "grad_norm": 0.5720431208610535, "learning_rate": 7.474697271934221e-06, "loss": 0.5066, "step": 3766 }, { "epoch": 1.205311433447099, "grad_norm": 0.5656705498695374, "learning_rate": 7.473079335723314e-06, "loss": 0.5232, "step": 3767 }, { "epoch": 1.2056313993174061, "grad_norm": 0.5155625939369202, "learning_rate": 7.471461056616981e-06, "loss": 0.4721, "step": 3768 }, { "epoch": 1.2059513651877134, "grad_norm": 0.5236969590187073, "learning_rate": 7.469842434839601e-06, "loss": 0.5029, "step": 3769 }, { "epoch": 1.2062713310580204, "grad_norm": 0.518295168876648, "learning_rate": 7.468223470615593e-06, "loss": 0.4919, "step": 3770 }, { "epoch": 1.2065912969283277, "grad_norm": 0.5845280289649963, "learning_rate": 7.466604164169433e-06, "loss": 0.514, "step": 3771 }, { "epoch": 1.2069112627986347, "grad_norm": 0.559059202671051, "learning_rate": 7.464984515725638e-06, "loss": 0.5455, "step": 3772 }, { "epoch": 1.207231228668942, "grad_norm": 0.588559091091156, "learning_rate": 7.463364525508775e-06, "loss": 0.5324, "step": 3773 }, { "epoch": 1.2075511945392492, "grad_norm": 0.5640997886657715, "learning_rate": 7.461744193743457e-06, "loss": 0.486, "step": 3774 }, { "epoch": 1.2078711604095562, "grad_norm": 0.5955579876899719, "learning_rate": 7.460123520654343e-06, "loss": 0.4913, "step": 3775 }, { "epoch": 1.2081911262798635, "grad_norm": 0.576140820980072, "learning_rate": 7.458502506466146e-06, "loss": 0.5175, "step": 3776 }, { "epoch": 1.2085110921501707, "grad_norm": 0.5625115633010864, "learning_rate": 7.456881151403618e-06, "loss": 0.4463, "step": 3777 }, { "epoch": 1.2088310580204777, "grad_norm": 0.7174860835075378, "learning_rate": 7.455259455691562e-06, "loss": 0.5819, "step": 3778 }, { "epoch": 1.209151023890785, "grad_norm": 0.5854070782661438, "learning_rate": 7.45363741955483e-06, "loss": 0.5497, "step": 3779 }, { "epoch": 1.2094709897610922, "grad_norm": 0.5234859585762024, "learning_rate": 7.4520150432183156e-06, "loss": 0.4448, "step": 3780 }, { "epoch": 1.2097909556313993, "grad_norm": 0.5454064011573792, "learning_rate": 7.4503923269069675e-06, "loss": 0.4655, "step": 3781 }, { "epoch": 1.2101109215017065, "grad_norm": 0.6177207827568054, "learning_rate": 7.448769270845776e-06, "loss": 0.5141, "step": 3782 }, { "epoch": 1.2104308873720138, "grad_norm": 0.5759219527244568, "learning_rate": 7.447145875259778e-06, "loss": 0.3989, "step": 3783 }, { "epoch": 1.2107508532423208, "grad_norm": 0.589008092880249, "learning_rate": 7.4455221403740595e-06, "loss": 0.5068, "step": 3784 }, { "epoch": 1.211070819112628, "grad_norm": 0.6944248080253601, "learning_rate": 7.443898066413755e-06, "loss": 0.5725, "step": 3785 }, { "epoch": 1.211390784982935, "grad_norm": 0.6175182461738586, "learning_rate": 7.442273653604045e-06, "loss": 0.5015, "step": 3786 }, { "epoch": 1.2117107508532423, "grad_norm": 0.5488205552101135, "learning_rate": 7.440648902170153e-06, "loss": 0.498, "step": 3787 }, { "epoch": 1.2120307167235496, "grad_norm": 0.5426703095436096, "learning_rate": 7.439023812337356e-06, "loss": 0.5147, "step": 3788 }, { "epoch": 1.2123506825938566, "grad_norm": 0.6002992987632751, "learning_rate": 7.4373983843309716e-06, "loss": 0.51, "step": 3789 }, { "epoch": 1.2126706484641638, "grad_norm": 0.5468200445175171, "learning_rate": 7.435772618376369e-06, "loss": 0.4511, "step": 3790 }, { "epoch": 1.212990614334471, "grad_norm": 0.5139322280883789, "learning_rate": 7.4341465146989645e-06, "loss": 0.5221, "step": 3791 }, { "epoch": 1.213310580204778, "grad_norm": 0.5981809496879578, "learning_rate": 7.432520073524217e-06, "loss": 0.504, "step": 3792 }, { "epoch": 1.2136305460750854, "grad_norm": 0.6306090950965881, "learning_rate": 7.4308932950776325e-06, "loss": 0.5581, "step": 3793 }, { "epoch": 1.2139505119453924, "grad_norm": 0.526813805103302, "learning_rate": 7.42926617958477e-06, "loss": 0.4759, "step": 3794 }, { "epoch": 1.2142704778156996, "grad_norm": 0.46746328473091125, "learning_rate": 7.4276387272712295e-06, "loss": 0.4742, "step": 3795 }, { "epoch": 1.2145904436860069, "grad_norm": 0.5398551821708679, "learning_rate": 7.426010938362657e-06, "loss": 0.4754, "step": 3796 }, { "epoch": 1.214910409556314, "grad_norm": 0.6205205917358398, "learning_rate": 7.424382813084751e-06, "loss": 0.5797, "step": 3797 }, { "epoch": 1.2152303754266212, "grad_norm": 0.4543765187263489, "learning_rate": 7.422754351663252e-06, "loss": 0.4051, "step": 3798 }, { "epoch": 1.2155503412969284, "grad_norm": 0.5468537211418152, "learning_rate": 7.421125554323945e-06, "loss": 0.5062, "step": 3799 }, { "epoch": 1.2158703071672354, "grad_norm": 0.5804492831230164, "learning_rate": 7.419496421292669e-06, "loss": 0.4949, "step": 3800 }, { "epoch": 1.2161902730375427, "grad_norm": 0.5590612292289734, "learning_rate": 7.417866952795304e-06, "loss": 0.5176, "step": 3801 }, { "epoch": 1.21651023890785, "grad_norm": 0.6057924032211304, "learning_rate": 7.416237149057776e-06, "loss": 0.5109, "step": 3802 }, { "epoch": 1.216830204778157, "grad_norm": 0.5891520380973816, "learning_rate": 7.414607010306062e-06, "loss": 0.4933, "step": 3803 }, { "epoch": 1.2171501706484642, "grad_norm": 0.5428804159164429, "learning_rate": 7.41297653676618e-06, "loss": 0.4571, "step": 3804 }, { "epoch": 1.2174701365187715, "grad_norm": 0.5450373888015747, "learning_rate": 7.411345728664198e-06, "loss": 0.4901, "step": 3805 }, { "epoch": 1.2177901023890785, "grad_norm": 0.5642934441566467, "learning_rate": 7.4097145862262295e-06, "loss": 0.5008, "step": 3806 }, { "epoch": 1.2181100682593857, "grad_norm": 0.5875202417373657, "learning_rate": 7.408083109678437e-06, "loss": 0.4655, "step": 3807 }, { "epoch": 1.2184300341296928, "grad_norm": 0.6120683550834656, "learning_rate": 7.406451299247022e-06, "loss": 0.5183, "step": 3808 }, { "epoch": 1.21875, "grad_norm": 0.5873705744743347, "learning_rate": 7.40481915515824e-06, "loss": 0.5272, "step": 3809 }, { "epoch": 1.2190699658703072, "grad_norm": 0.633432924747467, "learning_rate": 7.40318667763839e-06, "loss": 0.5113, "step": 3810 }, { "epoch": 1.2193899317406143, "grad_norm": 0.5159578323364258, "learning_rate": 7.4015538669138144e-06, "loss": 0.4249, "step": 3811 }, { "epoch": 1.2197098976109215, "grad_norm": 0.6399247646331787, "learning_rate": 7.399920723210907e-06, "loss": 0.5455, "step": 3812 }, { "epoch": 1.2200298634812285, "grad_norm": 0.5966647267341614, "learning_rate": 7.398287246756104e-06, "loss": 0.5282, "step": 3813 }, { "epoch": 1.2203498293515358, "grad_norm": 0.6498501896858215, "learning_rate": 7.396653437775889e-06, "loss": 0.5312, "step": 3814 }, { "epoch": 1.220669795221843, "grad_norm": 0.5518560409545898, "learning_rate": 7.395019296496792e-06, "loss": 0.4957, "step": 3815 }, { "epoch": 1.22098976109215, "grad_norm": 0.5691536068916321, "learning_rate": 7.393384823145386e-06, "loss": 0.4695, "step": 3816 }, { "epoch": 1.2213097269624573, "grad_norm": 0.506877064704895, "learning_rate": 7.391750017948298e-06, "loss": 0.4619, "step": 3817 }, { "epoch": 1.2216296928327646, "grad_norm": 0.5100228786468506, "learning_rate": 7.390114881132192e-06, "loss": 0.4562, "step": 3818 }, { "epoch": 1.2219496587030716, "grad_norm": 0.5075163841247559, "learning_rate": 7.388479412923782e-06, "loss": 0.5437, "step": 3819 }, { "epoch": 1.2222696245733788, "grad_norm": 0.5212882161140442, "learning_rate": 7.386843613549828e-06, "loss": 0.5016, "step": 3820 }, { "epoch": 1.222589590443686, "grad_norm": 0.5214887857437134, "learning_rate": 7.385207483237135e-06, "loss": 0.4902, "step": 3821 }, { "epoch": 1.2229095563139931, "grad_norm": 0.5080110430717468, "learning_rate": 7.383571022212555e-06, "loss": 0.4779, "step": 3822 }, { "epoch": 1.2232295221843004, "grad_norm": 0.5242145657539368, "learning_rate": 7.381934230702985e-06, "loss": 0.5419, "step": 3823 }, { "epoch": 1.2235494880546076, "grad_norm": 0.5515463948249817, "learning_rate": 7.3802971089353696e-06, "loss": 0.4978, "step": 3824 }, { "epoch": 1.2238694539249146, "grad_norm": 0.4667942523956299, "learning_rate": 7.378659657136696e-06, "loss": 0.4749, "step": 3825 }, { "epoch": 1.224189419795222, "grad_norm": 0.49958017468452454, "learning_rate": 7.377021875533998e-06, "loss": 0.4846, "step": 3826 }, { "epoch": 1.224509385665529, "grad_norm": 0.49523359537124634, "learning_rate": 7.37538376435436e-06, "loss": 0.4785, "step": 3827 }, { "epoch": 1.2248293515358362, "grad_norm": 0.5057116150856018, "learning_rate": 7.373745323824902e-06, "loss": 0.4996, "step": 3828 }, { "epoch": 1.2251493174061434, "grad_norm": 0.5707506537437439, "learning_rate": 7.372106554172802e-06, "loss": 0.4677, "step": 3829 }, { "epoch": 1.2254692832764504, "grad_norm": 0.5567911863327026, "learning_rate": 7.370467455625274e-06, "loss": 0.5527, "step": 3830 }, { "epoch": 1.2257892491467577, "grad_norm": 0.539019763469696, "learning_rate": 7.368828028409581e-06, "loss": 0.5028, "step": 3831 }, { "epoch": 1.226109215017065, "grad_norm": 0.5459343194961548, "learning_rate": 7.367188272753033e-06, "loss": 0.5102, "step": 3832 }, { "epoch": 1.226429180887372, "grad_norm": 0.550582766532898, "learning_rate": 7.365548188882982e-06, "loss": 0.5098, "step": 3833 }, { "epoch": 1.2267491467576792, "grad_norm": 0.5117180347442627, "learning_rate": 7.363907777026829e-06, "loss": 0.4085, "step": 3834 }, { "epoch": 1.2270691126279862, "grad_norm": 0.6057801842689514, "learning_rate": 7.36226703741202e-06, "loss": 0.6115, "step": 3835 }, { "epoch": 1.2273890784982935, "grad_norm": 0.5283210277557373, "learning_rate": 7.360625970266042e-06, "loss": 0.5066, "step": 3836 }, { "epoch": 1.2277090443686007, "grad_norm": 0.5724843740463257, "learning_rate": 7.358984575816437e-06, "loss": 0.5366, "step": 3837 }, { "epoch": 1.2280290102389078, "grad_norm": 0.49431341886520386, "learning_rate": 7.3573428542907784e-06, "loss": 0.4815, "step": 3838 }, { "epoch": 1.228348976109215, "grad_norm": 0.5193344950675964, "learning_rate": 7.3557008059167e-06, "loss": 0.4899, "step": 3839 }, { "epoch": 1.2286689419795223, "grad_norm": 0.5036170482635498, "learning_rate": 7.35405843092187e-06, "loss": 0.4829, "step": 3840 }, { "epoch": 1.2289889078498293, "grad_norm": 0.5176819562911987, "learning_rate": 7.352415729534006e-06, "loss": 0.5814, "step": 3841 }, { "epoch": 1.2293088737201365, "grad_norm": 0.46437740325927734, "learning_rate": 7.350772701980873e-06, "loss": 0.467, "step": 3842 }, { "epoch": 1.2296288395904438, "grad_norm": 0.520893394947052, "learning_rate": 7.3491293484902735e-06, "loss": 0.4686, "step": 3843 }, { "epoch": 1.2299488054607508, "grad_norm": 0.5439760088920593, "learning_rate": 7.347485669290067e-06, "loss": 0.5304, "step": 3844 }, { "epoch": 1.230268771331058, "grad_norm": 0.5253164768218994, "learning_rate": 7.345841664608146e-06, "loss": 0.4339, "step": 3845 }, { "epoch": 1.2305887372013653, "grad_norm": 0.5876121520996094, "learning_rate": 7.344197334672458e-06, "loss": 0.4885, "step": 3846 }, { "epoch": 1.2309087030716723, "grad_norm": 0.5066995024681091, "learning_rate": 7.34255267971099e-06, "loss": 0.5047, "step": 3847 }, { "epoch": 1.2312286689419796, "grad_norm": 0.530464231967926, "learning_rate": 7.3409076999517715e-06, "loss": 0.5129, "step": 3848 }, { "epoch": 1.2315486348122866, "grad_norm": 0.5482179522514343, "learning_rate": 7.339262395622887e-06, "loss": 0.4811, "step": 3849 }, { "epoch": 1.2318686006825939, "grad_norm": 0.5003216862678528, "learning_rate": 7.337616766952455e-06, "loss": 0.4204, "step": 3850 }, { "epoch": 1.232188566552901, "grad_norm": 0.5469177961349487, "learning_rate": 7.3359708141686484e-06, "loss": 0.5048, "step": 3851 }, { "epoch": 1.2325085324232081, "grad_norm": 0.6300505995750427, "learning_rate": 7.334324537499677e-06, "loss": 0.467, "step": 3852 }, { "epoch": 1.2328284982935154, "grad_norm": 0.5806970000267029, "learning_rate": 7.3326779371738e-06, "loss": 0.5221, "step": 3853 }, { "epoch": 1.2331484641638226, "grad_norm": 0.4897526502609253, "learning_rate": 7.3310310134193215e-06, "loss": 0.472, "step": 3854 }, { "epoch": 1.2334684300341296, "grad_norm": 0.5181820392608643, "learning_rate": 7.329383766464588e-06, "loss": 0.4576, "step": 3855 }, { "epoch": 1.233788395904437, "grad_norm": 0.60699862241745, "learning_rate": 7.3277361965379936e-06, "loss": 0.472, "step": 3856 }, { "epoch": 1.234108361774744, "grad_norm": 0.5165917873382568, "learning_rate": 7.326088303867974e-06, "loss": 0.4934, "step": 3857 }, { "epoch": 1.2344283276450512, "grad_norm": 0.4755984842777252, "learning_rate": 7.324440088683014e-06, "loss": 0.4642, "step": 3858 }, { "epoch": 1.2347482935153584, "grad_norm": 0.5469703078269958, "learning_rate": 7.322791551211639e-06, "loss": 0.4932, "step": 3859 }, { "epoch": 1.2350682593856654, "grad_norm": 0.5659158825874329, "learning_rate": 7.321142691682421e-06, "loss": 0.4837, "step": 3860 }, { "epoch": 1.2353882252559727, "grad_norm": 0.5701580047607422, "learning_rate": 7.319493510323976e-06, "loss": 0.493, "step": 3861 }, { "epoch": 1.23570819112628, "grad_norm": 0.5682664513587952, "learning_rate": 7.317844007364966e-06, "loss": 0.5328, "step": 3862 }, { "epoch": 1.236028156996587, "grad_norm": 0.5998944044113159, "learning_rate": 7.316194183034096e-06, "loss": 0.4956, "step": 3863 }, { "epoch": 1.2363481228668942, "grad_norm": 0.5686836242675781, "learning_rate": 7.314544037560115e-06, "loss": 0.5208, "step": 3864 }, { "epoch": 1.2366680887372015, "grad_norm": 0.5470593571662903, "learning_rate": 7.3128935711718185e-06, "loss": 0.5055, "step": 3865 }, { "epoch": 1.2369880546075085, "grad_norm": 0.5307252407073975, "learning_rate": 7.311242784098047e-06, "loss": 0.5105, "step": 3866 }, { "epoch": 1.2373080204778157, "grad_norm": 0.544973611831665, "learning_rate": 7.309591676567683e-06, "loss": 0.4639, "step": 3867 }, { "epoch": 1.237627986348123, "grad_norm": 0.6316813230514526, "learning_rate": 7.307940248809655e-06, "loss": 0.567, "step": 3868 }, { "epoch": 1.23794795221843, "grad_norm": 0.5119458436965942, "learning_rate": 7.306288501052934e-06, "loss": 0.4561, "step": 3869 }, { "epoch": 1.2382679180887373, "grad_norm": 0.5519663095474243, "learning_rate": 7.3046364335265375e-06, "loss": 0.532, "step": 3870 }, { "epoch": 1.2385878839590443, "grad_norm": 0.568614661693573, "learning_rate": 7.302984046459528e-06, "loss": 0.4663, "step": 3871 }, { "epoch": 1.2389078498293515, "grad_norm": 0.6015342473983765, "learning_rate": 7.301331340081009e-06, "loss": 0.5595, "step": 3872 }, { "epoch": 1.2392278156996588, "grad_norm": 0.5088731050491333, "learning_rate": 7.299678314620132e-06, "loss": 0.4932, "step": 3873 }, { "epoch": 1.2395477815699658, "grad_norm": 0.5520064234733582, "learning_rate": 7.298024970306089e-06, "loss": 0.5142, "step": 3874 }, { "epoch": 1.239867747440273, "grad_norm": 0.5585522055625916, "learning_rate": 7.29637130736812e-06, "loss": 0.4941, "step": 3875 }, { "epoch": 1.24018771331058, "grad_norm": 0.5988473892211914, "learning_rate": 7.294717326035508e-06, "loss": 0.4555, "step": 3876 }, { "epoch": 1.2405076791808873, "grad_norm": 0.5895554423332214, "learning_rate": 7.293063026537575e-06, "loss": 0.536, "step": 3877 }, { "epoch": 1.2408276450511946, "grad_norm": 0.5260162353515625, "learning_rate": 7.291408409103698e-06, "loss": 0.5072, "step": 3878 }, { "epoch": 1.2411476109215016, "grad_norm": 0.5081711411476135, "learning_rate": 7.289753473963286e-06, "loss": 0.4796, "step": 3879 }, { "epoch": 1.2414675767918089, "grad_norm": 0.5044803023338318, "learning_rate": 7.2880982213458e-06, "loss": 0.5022, "step": 3880 }, { "epoch": 1.241787542662116, "grad_norm": 0.5551183223724365, "learning_rate": 7.2864426514807455e-06, "loss": 0.5554, "step": 3881 }, { "epoch": 1.2421075085324231, "grad_norm": 0.5451104044914246, "learning_rate": 7.2847867645976656e-06, "loss": 0.4554, "step": 3882 }, { "epoch": 1.2424274744027304, "grad_norm": 0.5335094928741455, "learning_rate": 7.2831305609261514e-06, "loss": 0.5072, "step": 3883 }, { "epoch": 1.2427474402730376, "grad_norm": 0.5673237442970276, "learning_rate": 7.281474040695839e-06, "loss": 0.4995, "step": 3884 }, { "epoch": 1.2430674061433447, "grad_norm": 0.5887155532836914, "learning_rate": 7.279817204136405e-06, "loss": 0.5045, "step": 3885 }, { "epoch": 1.243387372013652, "grad_norm": 0.5778360366821289, "learning_rate": 7.278160051477574e-06, "loss": 0.4836, "step": 3886 }, { "epoch": 1.2437073378839592, "grad_norm": 0.5329510569572449, "learning_rate": 7.2765025829491105e-06, "loss": 0.5502, "step": 3887 }, { "epoch": 1.2440273037542662, "grad_norm": 0.5407595038414001, "learning_rate": 7.274844798780826e-06, "loss": 0.4718, "step": 3888 }, { "epoch": 1.2443472696245734, "grad_norm": 0.5253113508224487, "learning_rate": 7.273186699202572e-06, "loss": 0.4936, "step": 3889 }, { "epoch": 1.2446672354948805, "grad_norm": 0.5383972525596619, "learning_rate": 7.27152828444425e-06, "loss": 0.5175, "step": 3890 }, { "epoch": 1.2449872013651877, "grad_norm": 0.5071119666099548, "learning_rate": 7.269869554735796e-06, "loss": 0.5104, "step": 3891 }, { "epoch": 1.245307167235495, "grad_norm": 0.5174180865287781, "learning_rate": 7.2682105103072e-06, "loss": 0.5488, "step": 3892 }, { "epoch": 1.245627133105802, "grad_norm": 0.45115968585014343, "learning_rate": 7.266551151388485e-06, "loss": 0.4492, "step": 3893 }, { "epoch": 1.2459470989761092, "grad_norm": 0.5595500469207764, "learning_rate": 7.264891478209729e-06, "loss": 0.533, "step": 3894 }, { "epoch": 1.2462670648464165, "grad_norm": 0.5710468292236328, "learning_rate": 7.263231491001044e-06, "loss": 0.4815, "step": 3895 }, { "epoch": 1.2465870307167235, "grad_norm": 0.5594239830970764, "learning_rate": 7.261571189992589e-06, "loss": 0.4924, "step": 3896 }, { "epoch": 1.2469069965870307, "grad_norm": 0.5303643345832825, "learning_rate": 7.259910575414569e-06, "loss": 0.468, "step": 3897 }, { "epoch": 1.2472269624573378, "grad_norm": 0.5882477760314941, "learning_rate": 7.258249647497228e-06, "loss": 0.498, "step": 3898 }, { "epoch": 1.247546928327645, "grad_norm": 0.578035831451416, "learning_rate": 7.256588406470857e-06, "loss": 0.4829, "step": 3899 }, { "epoch": 1.2478668941979523, "grad_norm": 0.4835994243621826, "learning_rate": 7.254926852565789e-06, "loss": 0.486, "step": 3900 }, { "epoch": 1.2481868600682593, "grad_norm": 0.5448126196861267, "learning_rate": 7.253264986012399e-06, "loss": 0.468, "step": 3901 }, { "epoch": 1.2485068259385665, "grad_norm": 0.6015421152114868, "learning_rate": 7.251602807041111e-06, "loss": 0.54, "step": 3902 }, { "epoch": 1.2488267918088738, "grad_norm": 0.5436009168624878, "learning_rate": 7.249940315882382e-06, "loss": 0.5052, "step": 3903 }, { "epoch": 1.2491467576791808, "grad_norm": 0.595392644405365, "learning_rate": 7.248277512766722e-06, "loss": 0.5412, "step": 3904 }, { "epoch": 1.249466723549488, "grad_norm": 0.5022367238998413, "learning_rate": 7.24661439792468e-06, "loss": 0.4956, "step": 3905 }, { "epoch": 1.2497866894197953, "grad_norm": 0.5141124725341797, "learning_rate": 7.244950971586849e-06, "loss": 0.4603, "step": 3906 }, { "epoch": 1.2501066552901023, "grad_norm": 0.5934104323387146, "learning_rate": 7.243287233983866e-06, "loss": 0.5731, "step": 3907 }, { "epoch": 1.2504266211604096, "grad_norm": 0.5501481294631958, "learning_rate": 7.241623185346409e-06, "loss": 0.4424, "step": 3908 }, { "epoch": 1.2507465870307168, "grad_norm": 0.5349768400192261, "learning_rate": 7.239958825905201e-06, "loss": 0.5224, "step": 3909 }, { "epoch": 1.2510665529010239, "grad_norm": 0.5530389547348022, "learning_rate": 7.238294155891006e-06, "loss": 0.5149, "step": 3910 }, { "epoch": 1.2513865187713311, "grad_norm": 0.5477389693260193, "learning_rate": 7.2366291755346344e-06, "loss": 0.4818, "step": 3911 }, { "epoch": 1.2517064846416384, "grad_norm": 0.5573341250419617, "learning_rate": 7.234963885066937e-06, "loss": 0.4673, "step": 3912 }, { "epoch": 1.2520264505119454, "grad_norm": 0.5481014251708984, "learning_rate": 7.233298284718806e-06, "loss": 0.5064, "step": 3913 }, { "epoch": 1.2523464163822526, "grad_norm": 0.5816885828971863, "learning_rate": 7.231632374721184e-06, "loss": 0.4764, "step": 3914 }, { "epoch": 1.2526663822525597, "grad_norm": 0.6380035281181335, "learning_rate": 7.2299661553050474e-06, "loss": 0.5305, "step": 3915 }, { "epoch": 1.252986348122867, "grad_norm": 0.571526825428009, "learning_rate": 7.22829962670142e-06, "loss": 0.5471, "step": 3916 }, { "epoch": 1.253306313993174, "grad_norm": 0.5838145613670349, "learning_rate": 7.2266327891413705e-06, "loss": 0.503, "step": 3917 }, { "epoch": 1.2536262798634812, "grad_norm": 0.5403713583946228, "learning_rate": 7.224965642856003e-06, "loss": 0.4958, "step": 3918 }, { "epoch": 1.2539462457337884, "grad_norm": 0.5514960289001465, "learning_rate": 7.223298188076475e-06, "loss": 0.4782, "step": 3919 }, { "epoch": 1.2542662116040955, "grad_norm": 0.6443050503730774, "learning_rate": 7.221630425033977e-06, "loss": 0.501, "step": 3920 }, { "epoch": 1.2545861774744027, "grad_norm": 0.5583950281143188, "learning_rate": 7.219962353959749e-06, "loss": 0.4591, "step": 3921 }, { "epoch": 1.25490614334471, "grad_norm": 0.5959684252738953, "learning_rate": 7.218293975085071e-06, "loss": 0.5533, "step": 3922 }, { "epoch": 1.255226109215017, "grad_norm": 0.5673621296882629, "learning_rate": 7.2166252886412614e-06, "loss": 0.4859, "step": 3923 }, { "epoch": 1.2555460750853242, "grad_norm": 0.5439527034759521, "learning_rate": 7.2149562948596914e-06, "loss": 0.445, "step": 3924 }, { "epoch": 1.2558660409556315, "grad_norm": 0.6172842383384705, "learning_rate": 7.213286993971764e-06, "loss": 0.5214, "step": 3925 }, { "epoch": 1.2561860068259385, "grad_norm": 0.5660737156867981, "learning_rate": 7.211617386208936e-06, "loss": 0.5096, "step": 3926 }, { "epoch": 1.2565059726962458, "grad_norm": 0.554425060749054, "learning_rate": 7.209947471802694e-06, "loss": 0.5024, "step": 3927 }, { "epoch": 1.256825938566553, "grad_norm": 0.6696584224700928, "learning_rate": 7.208277250984577e-06, "loss": 0.5449, "step": 3928 }, { "epoch": 1.25714590443686, "grad_norm": 0.5374483466148376, "learning_rate": 7.206606723986164e-06, "loss": 0.483, "step": 3929 }, { "epoch": 1.2574658703071673, "grad_norm": 0.5405775308609009, "learning_rate": 7.204935891039071e-06, "loss": 0.4617, "step": 3930 }, { "epoch": 1.2577858361774745, "grad_norm": 0.636476993560791, "learning_rate": 7.203264752374968e-06, "loss": 0.5509, "step": 3931 }, { "epoch": 1.2581058020477816, "grad_norm": 0.5011911392211914, "learning_rate": 7.201593308225554e-06, "loss": 0.4196, "step": 3932 }, { "epoch": 1.2584257679180888, "grad_norm": 0.5440027117729187, "learning_rate": 7.199921558822579e-06, "loss": 0.5234, "step": 3933 }, { "epoch": 1.2587457337883958, "grad_norm": 0.6314454674720764, "learning_rate": 7.198249504397834e-06, "loss": 0.4913, "step": 3934 }, { "epoch": 1.259065699658703, "grad_norm": 0.5794150233268738, "learning_rate": 7.19657714518315e-06, "loss": 0.5514, "step": 3935 }, { "epoch": 1.25938566552901, "grad_norm": 0.5919234156608582, "learning_rate": 7.194904481410403e-06, "loss": 0.5185, "step": 3936 }, { "epoch": 1.2597056313993173, "grad_norm": 0.5429485440254211, "learning_rate": 7.193231513311509e-06, "loss": 0.5403, "step": 3937 }, { "epoch": 1.2600255972696246, "grad_norm": 0.5316311717033386, "learning_rate": 7.191558241118427e-06, "loss": 0.4117, "step": 3938 }, { "epoch": 1.2603455631399316, "grad_norm": 0.5937775373458862, "learning_rate": 7.18988466506316e-06, "loss": 0.5259, "step": 3939 }, { "epoch": 1.2606655290102389, "grad_norm": 0.5185807943344116, "learning_rate": 7.188210785377748e-06, "loss": 0.5076, "step": 3940 }, { "epoch": 1.2609854948805461, "grad_norm": 0.5359899401664734, "learning_rate": 7.186536602294278e-06, "loss": 0.5209, "step": 3941 }, { "epoch": 1.2613054607508531, "grad_norm": 0.5379454493522644, "learning_rate": 7.184862116044877e-06, "loss": 0.5135, "step": 3942 }, { "epoch": 1.2616254266211604, "grad_norm": 0.5110536217689514, "learning_rate": 7.183187326861717e-06, "loss": 0.4563, "step": 3943 }, { "epoch": 1.2619453924914676, "grad_norm": 0.5227304697036743, "learning_rate": 7.1815122349770075e-06, "loss": 0.4812, "step": 3944 }, { "epoch": 1.2622653583617747, "grad_norm": 0.5106444954872131, "learning_rate": 7.1798368406230004e-06, "loss": 0.4683, "step": 3945 }, { "epoch": 1.262585324232082, "grad_norm": 0.5713802576065063, "learning_rate": 7.178161144031994e-06, "loss": 0.5206, "step": 3946 }, { "epoch": 1.2629052901023892, "grad_norm": 0.6216138601303101, "learning_rate": 7.176485145436325e-06, "loss": 0.4972, "step": 3947 }, { "epoch": 1.2632252559726962, "grad_norm": 0.5910817980766296, "learning_rate": 7.174808845068372e-06, "loss": 0.4889, "step": 3948 }, { "epoch": 1.2635452218430034, "grad_norm": 0.5474295020103455, "learning_rate": 7.173132243160557e-06, "loss": 0.5213, "step": 3949 }, { "epoch": 1.2638651877133107, "grad_norm": 0.48133257031440735, "learning_rate": 7.171455339945339e-06, "loss": 0.4447, "step": 3950 }, { "epoch": 1.2641851535836177, "grad_norm": 0.5086634755134583, "learning_rate": 7.1697781356552295e-06, "loss": 0.4936, "step": 3951 }, { "epoch": 1.264505119453925, "grad_norm": 0.5170502662658691, "learning_rate": 7.168100630522769e-06, "loss": 0.4732, "step": 3952 }, { "epoch": 1.2648250853242322, "grad_norm": 0.5468535423278809, "learning_rate": 7.166422824780549e-06, "loss": 0.499, "step": 3953 }, { "epoch": 1.2651450511945392, "grad_norm": 0.5571072101593018, "learning_rate": 7.164744718661198e-06, "loss": 0.4899, "step": 3954 }, { "epoch": 1.2654650170648465, "grad_norm": 0.5490705966949463, "learning_rate": 7.163066312397386e-06, "loss": 0.492, "step": 3955 }, { "epoch": 1.2657849829351535, "grad_norm": 0.567050039768219, "learning_rate": 7.16138760622183e-06, "loss": 0.5362, "step": 3956 }, { "epoch": 1.2661049488054608, "grad_norm": 0.5776382684707642, "learning_rate": 7.15970860036728e-06, "loss": 0.466, "step": 3957 }, { "epoch": 1.2664249146757678, "grad_norm": 0.4942425787448883, "learning_rate": 7.158029295066535e-06, "loss": 0.4487, "step": 3958 }, { "epoch": 1.266744880546075, "grad_norm": 0.4956960082054138, "learning_rate": 7.156349690552433e-06, "loss": 0.5155, "step": 3959 }, { "epoch": 1.2670648464163823, "grad_norm": 0.5051069259643555, "learning_rate": 7.154669787057851e-06, "loss": 0.4963, "step": 3960 }, { "epoch": 1.2673848122866893, "grad_norm": 0.5262790322303772, "learning_rate": 7.152989584815712e-06, "loss": 0.4625, "step": 3961 }, { "epoch": 1.2677047781569966, "grad_norm": 0.592454731464386, "learning_rate": 7.151309084058976e-06, "loss": 0.5454, "step": 3962 }, { "epoch": 1.2680247440273038, "grad_norm": 0.5562735199928284, "learning_rate": 7.149628285020647e-06, "loss": 0.5272, "step": 3963 }, { "epoch": 1.2683447098976108, "grad_norm": 0.5365687608718872, "learning_rate": 7.147947187933772e-06, "loss": 0.5179, "step": 3964 }, { "epoch": 1.268664675767918, "grad_norm": 0.49221140146255493, "learning_rate": 7.146265793031434e-06, "loss": 0.4755, "step": 3965 }, { "epoch": 1.2689846416382253, "grad_norm": 0.5778467655181885, "learning_rate": 7.144584100546762e-06, "loss": 0.4927, "step": 3966 }, { "epoch": 1.2693046075085324, "grad_norm": 0.5484925508499146, "learning_rate": 7.142902110712925e-06, "loss": 0.4975, "step": 3967 }, { "epoch": 1.2696245733788396, "grad_norm": 0.5711588263511658, "learning_rate": 7.141219823763132e-06, "loss": 0.513, "step": 3968 }, { "epoch": 1.2699445392491469, "grad_norm": 0.4926115572452545, "learning_rate": 7.139537239930634e-06, "loss": 0.462, "step": 3969 }, { "epoch": 1.2702645051194539, "grad_norm": 0.5405696034431458, "learning_rate": 7.137854359448723e-06, "loss": 0.5335, "step": 3970 }, { "epoch": 1.2705844709897611, "grad_norm": 0.493924617767334, "learning_rate": 7.136171182550736e-06, "loss": 0.4639, "step": 3971 }, { "epoch": 1.2709044368600684, "grad_norm": 0.5897502303123474, "learning_rate": 7.1344877094700425e-06, "loss": 0.518, "step": 3972 }, { "epoch": 1.2712244027303754, "grad_norm": 0.5166211128234863, "learning_rate": 7.132803940440061e-06, "loss": 0.5369, "step": 3973 }, { "epoch": 1.2715443686006827, "grad_norm": 0.5344966053962708, "learning_rate": 7.131119875694246e-06, "loss": 0.4817, "step": 3974 }, { "epoch": 1.27186433447099, "grad_norm": 0.5253636240959167, "learning_rate": 7.129435515466098e-06, "loss": 0.4572, "step": 3975 }, { "epoch": 1.272184300341297, "grad_norm": 0.5483711361885071, "learning_rate": 7.127750859989154e-06, "loss": 0.5079, "step": 3976 }, { "epoch": 1.2725042662116042, "grad_norm": 0.5688130855560303, "learning_rate": 7.126065909496993e-06, "loss": 0.5126, "step": 3977 }, { "epoch": 1.2728242320819112, "grad_norm": 0.5963906049728394, "learning_rate": 7.124380664223236e-06, "loss": 0.5238, "step": 3978 }, { "epoch": 1.2731441979522184, "grad_norm": 0.6051427721977234, "learning_rate": 7.122695124401544e-06, "loss": 0.4873, "step": 3979 }, { "epoch": 1.2734641638225255, "grad_norm": 0.5014509558677673, "learning_rate": 7.121009290265619e-06, "loss": 0.4829, "step": 3980 }, { "epoch": 1.2737841296928327, "grad_norm": 0.551353931427002, "learning_rate": 7.1193231620492055e-06, "loss": 0.5161, "step": 3981 }, { "epoch": 1.27410409556314, "grad_norm": 0.5647012591362, "learning_rate": 7.117636739986086e-06, "loss": 0.4926, "step": 3982 }, { "epoch": 1.274424061433447, "grad_norm": 0.5722877383232117, "learning_rate": 7.115950024310083e-06, "loss": 0.5451, "step": 3983 }, { "epoch": 1.2747440273037542, "grad_norm": 0.5415180325508118, "learning_rate": 7.114263015255065e-06, "loss": 0.5067, "step": 3984 }, { "epoch": 1.2750639931740615, "grad_norm": 0.5215157866477966, "learning_rate": 7.112575713054937e-06, "loss": 0.4927, "step": 3985 }, { "epoch": 1.2753839590443685, "grad_norm": 0.490604043006897, "learning_rate": 7.110888117943643e-06, "loss": 0.481, "step": 3986 }, { "epoch": 1.2757039249146758, "grad_norm": 0.5944197177886963, "learning_rate": 7.109200230155173e-06, "loss": 0.4671, "step": 3987 }, { "epoch": 1.276023890784983, "grad_norm": 0.5736401677131653, "learning_rate": 7.107512049923555e-06, "loss": 0.5181, "step": 3988 }, { "epoch": 1.27634385665529, "grad_norm": 0.5174470543861389, "learning_rate": 7.105823577482853e-06, "loss": 0.5242, "step": 3989 }, { "epoch": 1.2766638225255973, "grad_norm": 0.48873624205589294, "learning_rate": 7.10413481306718e-06, "loss": 0.4692, "step": 3990 }, { "epoch": 1.2769837883959045, "grad_norm": 0.5451329350471497, "learning_rate": 7.102445756910685e-06, "loss": 0.5023, "step": 3991 }, { "epoch": 1.2773037542662116, "grad_norm": 0.6022747159004211, "learning_rate": 7.100756409247556e-06, "loss": 0.521, "step": 3992 }, { "epoch": 1.2776237201365188, "grad_norm": 0.5710732936859131, "learning_rate": 7.099066770312023e-06, "loss": 0.4571, "step": 3993 }, { "epoch": 1.277943686006826, "grad_norm": 0.5511924624443054, "learning_rate": 7.097376840338357e-06, "loss": 0.5359, "step": 3994 }, { "epoch": 1.278263651877133, "grad_norm": 0.5261993408203125, "learning_rate": 7.095686619560868e-06, "loss": 0.4652, "step": 3995 }, { "epoch": 1.2785836177474403, "grad_norm": 0.5718681216239929, "learning_rate": 7.09399610821391e-06, "loss": 0.5136, "step": 3996 }, { "epoch": 1.2789035836177474, "grad_norm": 0.5797334909439087, "learning_rate": 7.092305306531872e-06, "loss": 0.4918, "step": 3997 }, { "epoch": 1.2792235494880546, "grad_norm": 0.5252419710159302, "learning_rate": 7.090614214749185e-06, "loss": 0.4283, "step": 3998 }, { "epoch": 1.2795435153583616, "grad_norm": 0.576372504234314, "learning_rate": 7.0889228331003236e-06, "loss": 0.5202, "step": 3999 }, { "epoch": 1.2798634812286689, "grad_norm": 0.5281257629394531, "learning_rate": 7.087231161819796e-06, "loss": 0.5165, "step": 4000 }, { "epoch": 1.2801834470989761, "grad_norm": 0.5718040466308594, "learning_rate": 7.085539201142159e-06, "loss": 0.4998, "step": 4001 }, { "epoch": 1.2805034129692832, "grad_norm": 0.5610496401786804, "learning_rate": 7.083846951302002e-06, "loss": 0.4913, "step": 4002 }, { "epoch": 1.2808233788395904, "grad_norm": 0.48426106572151184, "learning_rate": 7.0821544125339585e-06, "loss": 0.4525, "step": 4003 }, { "epoch": 1.2811433447098977, "grad_norm": 0.563601553440094, "learning_rate": 7.080461585072701e-06, "loss": 0.5032, "step": 4004 }, { "epoch": 1.2814633105802047, "grad_norm": 0.5950374603271484, "learning_rate": 7.078768469152941e-06, "loss": 0.5532, "step": 4005 }, { "epoch": 1.281783276450512, "grad_norm": 0.5124794244766235, "learning_rate": 7.0770750650094335e-06, "loss": 0.4315, "step": 4006 }, { "epoch": 1.2821032423208192, "grad_norm": 0.5625861883163452, "learning_rate": 7.075381372876969e-06, "loss": 0.4925, "step": 4007 }, { "epoch": 1.2824232081911262, "grad_norm": 0.5750341415405273, "learning_rate": 7.073687392990379e-06, "loss": 0.5136, "step": 4008 }, { "epoch": 1.2827431740614335, "grad_norm": 0.519352376461029, "learning_rate": 7.071993125584537e-06, "loss": 0.5323, "step": 4009 }, { "epoch": 1.2830631399317407, "grad_norm": 0.5940561294555664, "learning_rate": 7.070298570894357e-06, "loss": 0.5413, "step": 4010 }, { "epoch": 1.2833831058020477, "grad_norm": 0.5244735479354858, "learning_rate": 7.068603729154789e-06, "loss": 0.458, "step": 4011 }, { "epoch": 1.283703071672355, "grad_norm": 0.48761364817619324, "learning_rate": 7.066908600600826e-06, "loss": 0.4501, "step": 4012 }, { "epoch": 1.2840230375426622, "grad_norm": 0.5354412794113159, "learning_rate": 7.065213185467497e-06, "loss": 0.4823, "step": 4013 }, { "epoch": 1.2843430034129693, "grad_norm": 0.5936446785926819, "learning_rate": 7.063517483989879e-06, "loss": 0.5013, "step": 4014 }, { "epoch": 1.2846629692832765, "grad_norm": 0.5855966210365295, "learning_rate": 7.061821496403075e-06, "loss": 0.5154, "step": 4015 }, { "epoch": 1.2849829351535837, "grad_norm": 0.5463988780975342, "learning_rate": 7.0601252229422435e-06, "loss": 0.5012, "step": 4016 }, { "epoch": 1.2853029010238908, "grad_norm": 0.6146540641784668, "learning_rate": 7.05842866384257e-06, "loss": 0.5459, "step": 4017 }, { "epoch": 1.285622866894198, "grad_norm": 0.5338971018791199, "learning_rate": 7.056731819339287e-06, "loss": 0.4941, "step": 4018 }, { "epoch": 1.285942832764505, "grad_norm": 0.5357498526573181, "learning_rate": 7.055034689667661e-06, "loss": 0.4963, "step": 4019 }, { "epoch": 1.2862627986348123, "grad_norm": 0.5381874442100525, "learning_rate": 7.053337275063004e-06, "loss": 0.4578, "step": 4020 }, { "epoch": 1.2865827645051193, "grad_norm": 0.642769455909729, "learning_rate": 7.051639575760664e-06, "loss": 0.5292, "step": 4021 }, { "epoch": 1.2869027303754266, "grad_norm": 0.5198947191238403, "learning_rate": 7.0499415919960276e-06, "loss": 0.4965, "step": 4022 }, { "epoch": 1.2872226962457338, "grad_norm": 0.5457560420036316, "learning_rate": 7.048243324004525e-06, "loss": 0.4504, "step": 4023 }, { "epoch": 1.2875426621160408, "grad_norm": 0.5783088207244873, "learning_rate": 7.046544772021621e-06, "loss": 0.4804, "step": 4024 }, { "epoch": 1.287862627986348, "grad_norm": 0.5254973769187927, "learning_rate": 7.0448459362828206e-06, "loss": 0.4466, "step": 4025 }, { "epoch": 1.2881825938566553, "grad_norm": 0.5481970906257629, "learning_rate": 7.043146817023673e-06, "loss": 0.5046, "step": 4026 }, { "epoch": 1.2885025597269624, "grad_norm": 0.579662024974823, "learning_rate": 7.04144741447976e-06, "loss": 0.4715, "step": 4027 }, { "epoch": 1.2888225255972696, "grad_norm": 0.5224183797836304, "learning_rate": 7.039747728886708e-06, "loss": 0.5318, "step": 4028 }, { "epoch": 1.2891424914675769, "grad_norm": 0.4587222635746002, "learning_rate": 7.0380477604801786e-06, "loss": 0.3968, "step": 4029 }, { "epoch": 1.289462457337884, "grad_norm": 0.6167943477630615, "learning_rate": 7.036347509495875e-06, "loss": 0.5936, "step": 4030 }, { "epoch": 1.2897824232081911, "grad_norm": 0.5406517386436462, "learning_rate": 7.034646976169541e-06, "loss": 0.4896, "step": 4031 }, { "epoch": 1.2901023890784984, "grad_norm": 0.4828827381134033, "learning_rate": 7.032946160736956e-06, "loss": 0.4759, "step": 4032 }, { "epoch": 1.2904223549488054, "grad_norm": 0.5524978041648865, "learning_rate": 7.031245063433938e-06, "loss": 0.5368, "step": 4033 }, { "epoch": 1.2907423208191127, "grad_norm": 0.5164719223976135, "learning_rate": 7.029543684496352e-06, "loss": 0.4773, "step": 4034 }, { "epoch": 1.29106228668942, "grad_norm": 0.5399884581565857, "learning_rate": 7.027842024160087e-06, "loss": 0.4917, "step": 4035 }, { "epoch": 1.291382252559727, "grad_norm": 0.5169853568077087, "learning_rate": 7.0261400826610904e-06, "loss": 0.4577, "step": 4036 }, { "epoch": 1.2917022184300342, "grad_norm": 0.576972246170044, "learning_rate": 7.0244378602353334e-06, "loss": 0.5175, "step": 4037 }, { "epoch": 1.2920221843003412, "grad_norm": 0.6214761137962341, "learning_rate": 7.02273535711883e-06, "loss": 0.5466, "step": 4038 }, { "epoch": 1.2923421501706485, "grad_norm": 0.5590415000915527, "learning_rate": 7.021032573547637e-06, "loss": 0.4417, "step": 4039 }, { "epoch": 1.2926621160409557, "grad_norm": 0.5316540598869324, "learning_rate": 7.019329509757845e-06, "loss": 0.4974, "step": 4040 }, { "epoch": 1.2929820819112627, "grad_norm": 0.5390881896018982, "learning_rate": 7.017626165985589e-06, "loss": 0.4569, "step": 4041 }, { "epoch": 1.29330204778157, "grad_norm": 0.5673624873161316, "learning_rate": 7.015922542467038e-06, "loss": 0.5331, "step": 4042 }, { "epoch": 1.293622013651877, "grad_norm": 0.5370771884918213, "learning_rate": 7.014218639438399e-06, "loss": 0.506, "step": 4043 }, { "epoch": 1.2939419795221843, "grad_norm": 0.5172004699707031, "learning_rate": 7.012514457135924e-06, "loss": 0.4921, "step": 4044 }, { "epoch": 1.2942619453924915, "grad_norm": 0.5819820761680603, "learning_rate": 7.010809995795897e-06, "loss": 0.487, "step": 4045 }, { "epoch": 1.2945819112627985, "grad_norm": 0.5423453450202942, "learning_rate": 7.009105255654647e-06, "loss": 0.4607, "step": 4046 }, { "epoch": 1.2949018771331058, "grad_norm": 0.5723846554756165, "learning_rate": 7.007400236948536e-06, "loss": 0.5675, "step": 4047 }, { "epoch": 1.295221843003413, "grad_norm": 0.5019378066062927, "learning_rate": 7.005694939913966e-06, "loss": 0.4632, "step": 4048 }, { "epoch": 1.29554180887372, "grad_norm": 0.6267055869102478, "learning_rate": 7.00398936478738e-06, "loss": 0.5521, "step": 4049 }, { "epoch": 1.2958617747440273, "grad_norm": 0.573279082775116, "learning_rate": 7.002283511805258e-06, "loss": 0.5095, "step": 4050 }, { "epoch": 1.2961817406143346, "grad_norm": 0.5108147859573364, "learning_rate": 7.000577381204119e-06, "loss": 0.5046, "step": 4051 }, { "epoch": 1.2965017064846416, "grad_norm": 0.4877328872680664, "learning_rate": 6.998870973220519e-06, "loss": 0.4508, "step": 4052 }, { "epoch": 1.2968216723549488, "grad_norm": 0.5888170599937439, "learning_rate": 6.997164288091053e-06, "loss": 0.5318, "step": 4053 }, { "epoch": 1.297141638225256, "grad_norm": 0.5399531722068787, "learning_rate": 6.995457326052355e-06, "loss": 0.5082, "step": 4054 }, { "epoch": 1.297461604095563, "grad_norm": 0.5535011291503906, "learning_rate": 6.9937500873411e-06, "loss": 0.53, "step": 4055 }, { "epoch": 1.2977815699658704, "grad_norm": 0.522777795791626, "learning_rate": 6.992042572193997e-06, "loss": 0.4785, "step": 4056 }, { "epoch": 1.2981015358361776, "grad_norm": 0.5063053369522095, "learning_rate": 6.990334780847794e-06, "loss": 0.4694, "step": 4057 }, { "epoch": 1.2984215017064846, "grad_norm": 0.5899012088775635, "learning_rate": 6.98862671353928e-06, "loss": 0.5359, "step": 4058 }, { "epoch": 1.2987414675767919, "grad_norm": 0.5376660823822021, "learning_rate": 6.9869183705052775e-06, "loss": 0.4811, "step": 4059 }, { "epoch": 1.299061433447099, "grad_norm": 0.5922868251800537, "learning_rate": 6.985209751982655e-06, "loss": 0.5346, "step": 4060 }, { "epoch": 1.2993813993174061, "grad_norm": 0.5054578185081482, "learning_rate": 6.983500858208312e-06, "loss": 0.4288, "step": 4061 }, { "epoch": 1.2997013651877132, "grad_norm": 0.5970536470413208, "learning_rate": 6.981791689419187e-06, "loss": 0.5166, "step": 4062 }, { "epoch": 1.3000213310580204, "grad_norm": 0.5270934700965881, "learning_rate": 6.980082245852262e-06, "loss": 0.4704, "step": 4063 }, { "epoch": 1.3003412969283277, "grad_norm": 0.5640825033187866, "learning_rate": 6.97837252774455e-06, "loss": 0.4985, "step": 4064 }, { "epoch": 1.3006612627986347, "grad_norm": 0.5541390776634216, "learning_rate": 6.976662535333107e-06, "loss": 0.5034, "step": 4065 }, { "epoch": 1.300981228668942, "grad_norm": 0.5418012738227844, "learning_rate": 6.974952268855028e-06, "loss": 0.4515, "step": 4066 }, { "epoch": 1.3013011945392492, "grad_norm": 0.5639212727546692, "learning_rate": 6.9732417285474376e-06, "loss": 0.482, "step": 4067 }, { "epoch": 1.3016211604095562, "grad_norm": 0.5673478245735168, "learning_rate": 6.971530914647511e-06, "loss": 0.5271, "step": 4068 }, { "epoch": 1.3019411262798635, "grad_norm": 0.5235918760299683, "learning_rate": 6.969819827392449e-06, "loss": 0.4988, "step": 4069 }, { "epoch": 1.3022610921501707, "grad_norm": 0.4916918873786926, "learning_rate": 6.9681084670194984e-06, "loss": 0.4936, "step": 4070 }, { "epoch": 1.3025810580204777, "grad_norm": 0.5602163672447205, "learning_rate": 6.966396833765941e-06, "loss": 0.5123, "step": 4071 }, { "epoch": 1.302901023890785, "grad_norm": 0.5419959425926208, "learning_rate": 6.964684927869098e-06, "loss": 0.4499, "step": 4072 }, { "epoch": 1.3032209897610922, "grad_norm": 0.5200892686843872, "learning_rate": 6.9629727495663265e-06, "loss": 0.4945, "step": 4073 }, { "epoch": 1.3035409556313993, "grad_norm": 0.5808398127555847, "learning_rate": 6.961260299095021e-06, "loss": 0.4693, "step": 4074 }, { "epoch": 1.3038609215017065, "grad_norm": 0.6151460409164429, "learning_rate": 6.959547576692616e-06, "loss": 0.5331, "step": 4075 }, { "epoch": 1.3041808873720138, "grad_norm": 0.608762264251709, "learning_rate": 6.9578345825965805e-06, "loss": 0.52, "step": 4076 }, { "epoch": 1.3045008532423208, "grad_norm": 0.5830143094062805, "learning_rate": 6.956121317044426e-06, "loss": 0.5071, "step": 4077 }, { "epoch": 1.304820819112628, "grad_norm": 0.5641476511955261, "learning_rate": 6.954407780273697e-06, "loss": 0.5199, "step": 4078 }, { "epoch": 1.3051407849829353, "grad_norm": 0.5512235164642334, "learning_rate": 6.95269397252198e-06, "loss": 0.4253, "step": 4079 }, { "epoch": 1.3054607508532423, "grad_norm": 0.5953153371810913, "learning_rate": 6.950979894026893e-06, "loss": 0.5585, "step": 4080 }, { "epoch": 1.3057807167235496, "grad_norm": 0.5572149157524109, "learning_rate": 6.949265545026096e-06, "loss": 0.4667, "step": 4081 }, { "epoch": 1.3061006825938566, "grad_norm": 0.5659805536270142, "learning_rate": 6.947550925757287e-06, "loss": 0.5031, "step": 4082 }, { "epoch": 1.3064206484641638, "grad_norm": 0.4794667363166809, "learning_rate": 6.9458360364582e-06, "loss": 0.4365, "step": 4083 }, { "epoch": 1.3067406143344709, "grad_norm": 0.5645291209220886, "learning_rate": 6.944120877366605e-06, "loss": 0.6142, "step": 4084 }, { "epoch": 1.307060580204778, "grad_norm": 0.5863152742385864, "learning_rate": 6.942405448720311e-06, "loss": 0.501, "step": 4085 }, { "epoch": 1.3073805460750854, "grad_norm": 0.5941131711006165, "learning_rate": 6.940689750757163e-06, "loss": 0.5084, "step": 4086 }, { "epoch": 1.3077005119453924, "grad_norm": 0.5305261611938477, "learning_rate": 6.938973783715049e-06, "loss": 0.5205, "step": 4087 }, { "epoch": 1.3080204778156996, "grad_norm": 0.5063480138778687, "learning_rate": 6.937257547831887e-06, "loss": 0.4882, "step": 4088 }, { "epoch": 1.3083404436860069, "grad_norm": 0.5177851915359497, "learning_rate": 6.935541043345636e-06, "loss": 0.5213, "step": 4089 }, { "epoch": 1.308660409556314, "grad_norm": 0.5293291807174683, "learning_rate": 6.933824270494289e-06, "loss": 0.4641, "step": 4090 }, { "epoch": 1.3089803754266212, "grad_norm": 0.5412163734436035, "learning_rate": 6.932107229515881e-06, "loss": 0.5234, "step": 4091 }, { "epoch": 1.3093003412969284, "grad_norm": 0.5220082998275757, "learning_rate": 6.930389920648481e-06, "loss": 0.4499, "step": 4092 }, { "epoch": 1.3096203071672354, "grad_norm": 0.631682276725769, "learning_rate": 6.928672344130197e-06, "loss": 0.5726, "step": 4093 }, { "epoch": 1.3099402730375427, "grad_norm": 0.5830767154693604, "learning_rate": 6.9269545001991735e-06, "loss": 0.5176, "step": 4094 }, { "epoch": 1.31026023890785, "grad_norm": 0.5310470461845398, "learning_rate": 6.925236389093588e-06, "loss": 0.4576, "step": 4095 }, { "epoch": 1.310580204778157, "grad_norm": 0.5761337876319885, "learning_rate": 6.923518011051662e-06, "loss": 0.5753, "step": 4096 }, { "epoch": 1.3109001706484642, "grad_norm": 0.5443764328956604, "learning_rate": 6.92179936631165e-06, "loss": 0.4763, "step": 4097 }, { "epoch": 1.3112201365187715, "grad_norm": 0.5663328766822815, "learning_rate": 6.920080455111843e-06, "loss": 0.5491, "step": 4098 }, { "epoch": 1.3115401023890785, "grad_norm": 0.540813148021698, "learning_rate": 6.918361277690572e-06, "loss": 0.4574, "step": 4099 }, { "epoch": 1.3118600682593857, "grad_norm": 0.4971042573451996, "learning_rate": 6.916641834286202e-06, "loss": 0.4582, "step": 4100 }, { "epoch": 1.3121800341296928, "grad_norm": 0.5601802468299866, "learning_rate": 6.914922125137135e-06, "loss": 0.5056, "step": 4101 }, { "epoch": 1.3125, "grad_norm": 0.582165539264679, "learning_rate": 6.9132021504818115e-06, "loss": 0.5102, "step": 4102 }, { "epoch": 1.3128199658703072, "grad_norm": 0.5567464232444763, "learning_rate": 6.9114819105587085e-06, "loss": 0.5152, "step": 4103 }, { "epoch": 1.3131399317406143, "grad_norm": 0.5217399597167969, "learning_rate": 6.90976140560634e-06, "loss": 0.4812, "step": 4104 }, { "epoch": 1.3134598976109215, "grad_norm": 0.5689828395843506, "learning_rate": 6.9080406358632535e-06, "loss": 0.4781, "step": 4105 }, { "epoch": 1.3137798634812285, "grad_norm": 0.5900806784629822, "learning_rate": 6.906319601568039e-06, "loss": 0.5043, "step": 4106 }, { "epoch": 1.3140998293515358, "grad_norm": 0.49958544969558716, "learning_rate": 6.904598302959318e-06, "loss": 0.4601, "step": 4107 }, { "epoch": 1.314419795221843, "grad_norm": 0.5210117101669312, "learning_rate": 6.90287674027575e-06, "loss": 0.5583, "step": 4108 }, { "epoch": 1.31473976109215, "grad_norm": 0.5850756764411926, "learning_rate": 6.901154913756034e-06, "loss": 0.5359, "step": 4109 }, { "epoch": 1.3150597269624573, "grad_norm": 0.5848261713981628, "learning_rate": 6.8994328236389006e-06, "loss": 0.4849, "step": 4110 }, { "epoch": 1.3153796928327646, "grad_norm": 0.5249413251876831, "learning_rate": 6.8977104701631226e-06, "loss": 0.5132, "step": 4111 }, { "epoch": 1.3156996587030716, "grad_norm": 0.5254254937171936, "learning_rate": 6.895987853567504e-06, "loss": 0.4501, "step": 4112 }, { "epoch": 1.3160196245733788, "grad_norm": 0.6095935702323914, "learning_rate": 6.894264974090889e-06, "loss": 0.5633, "step": 4113 }, { "epoch": 1.316339590443686, "grad_norm": 0.5580556392669678, "learning_rate": 6.892541831972157e-06, "loss": 0.5085, "step": 4114 }, { "epoch": 1.3166595563139931, "grad_norm": 0.5535349249839783, "learning_rate": 6.890818427450221e-06, "loss": 0.5184, "step": 4115 }, { "epoch": 1.3169795221843004, "grad_norm": 0.5419041514396667, "learning_rate": 6.8890947607640376e-06, "loss": 0.4825, "step": 4116 }, { "epoch": 1.3172994880546076, "grad_norm": 0.5677555799484253, "learning_rate": 6.8873708321525925e-06, "loss": 0.5134, "step": 4117 }, { "epoch": 1.3176194539249146, "grad_norm": 0.5081667900085449, "learning_rate": 6.88564664185491e-06, "loss": 0.5181, "step": 4118 }, { "epoch": 1.317939419795222, "grad_norm": 0.449253648519516, "learning_rate": 6.883922190110054e-06, "loss": 0.4367, "step": 4119 }, { "epoch": 1.3182593856655291, "grad_norm": 0.5432609915733337, "learning_rate": 6.882197477157118e-06, "loss": 0.5034, "step": 4120 }, { "epoch": 1.3185793515358362, "grad_norm": 0.5242462754249573, "learning_rate": 6.880472503235238e-06, "loss": 0.4599, "step": 4121 }, { "epoch": 1.3188993174061434, "grad_norm": 0.5869811177253723, "learning_rate": 6.878747268583584e-06, "loss": 0.5204, "step": 4122 }, { "epoch": 1.3192192832764504, "grad_norm": 0.6002717018127441, "learning_rate": 6.8770217734413606e-06, "loss": 0.5206, "step": 4123 }, { "epoch": 1.3195392491467577, "grad_norm": 0.5224626064300537, "learning_rate": 6.87529601804781e-06, "loss": 0.4443, "step": 4124 }, { "epoch": 1.3198592150170647, "grad_norm": 0.5698513984680176, "learning_rate": 6.873570002642209e-06, "loss": 0.5324, "step": 4125 }, { "epoch": 1.320179180887372, "grad_norm": 0.555054247379303, "learning_rate": 6.871843727463876e-06, "loss": 0.4941, "step": 4126 }, { "epoch": 1.3204991467576792, "grad_norm": 0.6183741092681885, "learning_rate": 6.870117192752157e-06, "loss": 0.5522, "step": 4127 }, { "epoch": 1.3208191126279862, "grad_norm": 0.5069617629051208, "learning_rate": 6.86839039874644e-06, "loss": 0.4479, "step": 4128 }, { "epoch": 1.3211390784982935, "grad_norm": 0.5739780068397522, "learning_rate": 6.866663345686147e-06, "loss": 0.4713, "step": 4129 }, { "epoch": 1.3214590443686007, "grad_norm": 0.6382153034210205, "learning_rate": 6.8649360338107315e-06, "loss": 0.5124, "step": 4130 }, { "epoch": 1.3217790102389078, "grad_norm": 0.614399254322052, "learning_rate": 6.8632084633596964e-06, "loss": 0.5227, "step": 4131 }, { "epoch": 1.322098976109215, "grad_norm": 0.5141477584838867, "learning_rate": 6.861480634572564e-06, "loss": 0.4562, "step": 4132 }, { "epoch": 1.3224189419795223, "grad_norm": 0.5629687309265137, "learning_rate": 6.859752547688902e-06, "loss": 0.4857, "step": 4133 }, { "epoch": 1.3227389078498293, "grad_norm": 0.607196569442749, "learning_rate": 6.858024202948312e-06, "loss": 0.5035, "step": 4134 }, { "epoch": 1.3230588737201365, "grad_norm": 0.5607357621192932, "learning_rate": 6.85629560059043e-06, "loss": 0.4983, "step": 4135 }, { "epoch": 1.3233788395904438, "grad_norm": 0.5918465852737427, "learning_rate": 6.854566740854932e-06, "loss": 0.5253, "step": 4136 }, { "epoch": 1.3236988054607508, "grad_norm": 0.5611516237258911, "learning_rate": 6.852837623981523e-06, "loss": 0.465, "step": 4137 }, { "epoch": 1.324018771331058, "grad_norm": 0.5750157833099365, "learning_rate": 6.851108250209949e-06, "loss": 0.531, "step": 4138 }, { "epoch": 1.3243387372013653, "grad_norm": 0.5478917360305786, "learning_rate": 6.849378619779989e-06, "loss": 0.4976, "step": 4139 }, { "epoch": 1.3246587030716723, "grad_norm": 0.4693235456943512, "learning_rate": 6.847648732931458e-06, "loss": 0.4736, "step": 4140 }, { "epoch": 1.3249786689419796, "grad_norm": 0.5775055289268494, "learning_rate": 6.845918589904209e-06, "loss": 0.4965, "step": 4141 }, { "epoch": 1.3252986348122868, "grad_norm": 0.6024038195610046, "learning_rate": 6.844188190938125e-06, "loss": 0.496, "step": 4142 }, { "epoch": 1.3256186006825939, "grad_norm": 0.6416028738021851, "learning_rate": 6.84245753627313e-06, "loss": 0.494, "step": 4143 }, { "epoch": 1.325938566552901, "grad_norm": 0.5104652643203735, "learning_rate": 6.840726626149182e-06, "loss": 0.4953, "step": 4144 }, { "epoch": 1.3262585324232081, "grad_norm": 0.5593465566635132, "learning_rate": 6.838995460806272e-06, "loss": 0.4754, "step": 4145 }, { "epoch": 1.3265784982935154, "grad_norm": 0.5791028141975403, "learning_rate": 6.837264040484431e-06, "loss": 0.4951, "step": 4146 }, { "epoch": 1.3268984641638224, "grad_norm": 0.5533764958381653, "learning_rate": 6.835532365423716e-06, "loss": 0.5162, "step": 4147 }, { "epoch": 1.3272184300341296, "grad_norm": 0.5973724722862244, "learning_rate": 6.833800435864235e-06, "loss": 0.506, "step": 4148 }, { "epoch": 1.327538395904437, "grad_norm": 0.5919985771179199, "learning_rate": 6.832068252046116e-06, "loss": 0.4638, "step": 4149 }, { "epoch": 1.327858361774744, "grad_norm": 0.5745266079902649, "learning_rate": 6.830335814209527e-06, "loss": 0.481, "step": 4150 }, { "epoch": 1.3281783276450512, "grad_norm": 0.5681392550468445, "learning_rate": 6.828603122594679e-06, "loss": 0.4785, "step": 4151 }, { "epoch": 1.3284982935153584, "grad_norm": 0.5746228098869324, "learning_rate": 6.826870177441807e-06, "loss": 0.5011, "step": 4152 }, { "epoch": 1.3288182593856654, "grad_norm": 0.5923866629600525, "learning_rate": 6.825136978991187e-06, "loss": 0.5005, "step": 4153 }, { "epoch": 1.3291382252559727, "grad_norm": 0.5296605825424194, "learning_rate": 6.82340352748313e-06, "loss": 0.5353, "step": 4154 }, { "epoch": 1.32945819112628, "grad_norm": 0.570958137512207, "learning_rate": 6.821669823157979e-06, "loss": 0.4754, "step": 4155 }, { "epoch": 1.329778156996587, "grad_norm": 0.5462219715118408, "learning_rate": 6.819935866256116e-06, "loss": 0.479, "step": 4156 }, { "epoch": 1.3300981228668942, "grad_norm": 0.5859329700469971, "learning_rate": 6.818201657017955e-06, "loss": 0.5163, "step": 4157 }, { "epoch": 1.3304180887372015, "grad_norm": 0.5124255418777466, "learning_rate": 6.816467195683949e-06, "loss": 0.4648, "step": 4158 }, { "epoch": 1.3307380546075085, "grad_norm": 0.520322859287262, "learning_rate": 6.814732482494579e-06, "loss": 0.5181, "step": 4159 }, { "epoch": 1.3310580204778157, "grad_norm": 0.5678321719169617, "learning_rate": 6.812997517690368e-06, "loss": 0.5017, "step": 4160 }, { "epoch": 1.331377986348123, "grad_norm": 0.5411324501037598, "learning_rate": 6.8112623015118695e-06, "loss": 0.4702, "step": 4161 }, { "epoch": 1.33169795221843, "grad_norm": 0.5200678706169128, "learning_rate": 6.809526834199675e-06, "loss": 0.491, "step": 4162 }, { "epoch": 1.3320179180887373, "grad_norm": 0.4977427124977112, "learning_rate": 6.807791115994408e-06, "loss": 0.5364, "step": 4163 }, { "epoch": 1.3323378839590443, "grad_norm": 0.5493211150169373, "learning_rate": 6.806055147136727e-06, "loss": 0.5077, "step": 4164 }, { "epoch": 1.3326578498293515, "grad_norm": 0.5668837428092957, "learning_rate": 6.804318927867327e-06, "loss": 0.4819, "step": 4165 }, { "epoch": 1.3329778156996588, "grad_norm": 0.6170305609703064, "learning_rate": 6.802582458426937e-06, "loss": 0.5639, "step": 4166 }, { "epoch": 1.3332977815699658, "grad_norm": 0.46503856778144836, "learning_rate": 6.800845739056321e-06, "loss": 0.4498, "step": 4167 }, { "epoch": 1.333617747440273, "grad_norm": 0.4986816942691803, "learning_rate": 6.799108769996279e-06, "loss": 0.5055, "step": 4168 }, { "epoch": 1.33393771331058, "grad_norm": 0.5867021083831787, "learning_rate": 6.797371551487638e-06, "loss": 0.5206, "step": 4169 }, { "epoch": 1.3342576791808873, "grad_norm": 0.5245837569236755, "learning_rate": 6.795634083771271e-06, "loss": 0.4289, "step": 4170 }, { "epoch": 1.3345776450511946, "grad_norm": 0.5654990077018738, "learning_rate": 6.793896367088077e-06, "loss": 0.53, "step": 4171 }, { "epoch": 1.3348976109215016, "grad_norm": 0.5152663588523865, "learning_rate": 6.7921584016789945e-06, "loss": 0.5477, "step": 4172 }, { "epoch": 1.3352175767918089, "grad_norm": 0.5061206221580505, "learning_rate": 6.790420187784993e-06, "loss": 0.4932, "step": 4173 }, { "epoch": 1.335537542662116, "grad_norm": 0.5259354710578918, "learning_rate": 6.788681725647079e-06, "loss": 0.5231, "step": 4174 }, { "epoch": 1.3358575085324231, "grad_norm": 0.45490285754203796, "learning_rate": 6.786943015506292e-06, "loss": 0.4411, "step": 4175 }, { "epoch": 1.3361774744027304, "grad_norm": 0.5072759389877319, "learning_rate": 6.785204057603706e-06, "loss": 0.4343, "step": 4176 }, { "epoch": 1.3364974402730376, "grad_norm": 0.5490756630897522, "learning_rate": 6.78346485218043e-06, "loss": 0.4856, "step": 4177 }, { "epoch": 1.3368174061433447, "grad_norm": 0.5347392559051514, "learning_rate": 6.781725399477606e-06, "loss": 0.5403, "step": 4178 }, { "epoch": 1.337137372013652, "grad_norm": 0.48878544569015503, "learning_rate": 6.779985699736413e-06, "loss": 0.4763, "step": 4179 }, { "epoch": 1.3374573378839592, "grad_norm": 0.5018070340156555, "learning_rate": 6.778245753198061e-06, "loss": 0.5074, "step": 4180 }, { "epoch": 1.3377773037542662, "grad_norm": 0.5557435750961304, "learning_rate": 6.776505560103795e-06, "loss": 0.453, "step": 4181 }, { "epoch": 1.3380972696245734, "grad_norm": 0.6862342357635498, "learning_rate": 6.774765120694896e-06, "loss": 0.5549, "step": 4182 }, { "epoch": 1.3384172354948807, "grad_norm": 0.5224938988685608, "learning_rate": 6.773024435212678e-06, "loss": 0.4618, "step": 4183 }, { "epoch": 1.3387372013651877, "grad_norm": 0.5360044240951538, "learning_rate": 6.771283503898492e-06, "loss": 0.5176, "step": 4184 }, { "epoch": 1.339057167235495, "grad_norm": 0.5369170904159546, "learning_rate": 6.769542326993713e-06, "loss": 0.534, "step": 4185 }, { "epoch": 1.339377133105802, "grad_norm": 0.48003724217414856, "learning_rate": 6.767800904739764e-06, "loss": 0.4751, "step": 4186 }, { "epoch": 1.3396970989761092, "grad_norm": 0.5560697913169861, "learning_rate": 6.766059237378092e-06, "loss": 0.4935, "step": 4187 }, { "epoch": 1.3400170648464163, "grad_norm": 0.5447647571563721, "learning_rate": 6.764317325150183e-06, "loss": 0.5242, "step": 4188 }, { "epoch": 1.3403370307167235, "grad_norm": 0.4936341345310211, "learning_rate": 6.762575168297554e-06, "loss": 0.4576, "step": 4189 }, { "epoch": 1.3406569965870307, "grad_norm": 0.6130675077438354, "learning_rate": 6.760832767061756e-06, "loss": 0.5732, "step": 4190 }, { "epoch": 1.3409769624573378, "grad_norm": 0.5237568616867065, "learning_rate": 6.7590901216843775e-06, "loss": 0.4388, "step": 4191 }, { "epoch": 1.341296928327645, "grad_norm": 0.5610059499740601, "learning_rate": 6.757347232407037e-06, "loss": 0.5656, "step": 4192 }, { "epoch": 1.3416168941979523, "grad_norm": 0.5444913506507874, "learning_rate": 6.75560409947139e-06, "loss": 0.466, "step": 4193 }, { "epoch": 1.3419368600682593, "grad_norm": 0.5468993186950684, "learning_rate": 6.753860723119122e-06, "loss": 0.5077, "step": 4194 }, { "epoch": 1.3422568259385665, "grad_norm": 0.5734399557113647, "learning_rate": 6.752117103591953e-06, "loss": 0.5201, "step": 4195 }, { "epoch": 1.3425767918088738, "grad_norm": 0.491445392370224, "learning_rate": 6.750373241131643e-06, "loss": 0.4913, "step": 4196 }, { "epoch": 1.3428967576791808, "grad_norm": 0.5285682082176208, "learning_rate": 6.748629135979975e-06, "loss": 0.5144, "step": 4197 }, { "epoch": 1.343216723549488, "grad_norm": 0.47706079483032227, "learning_rate": 6.746884788378774e-06, "loss": 0.475, "step": 4198 }, { "epoch": 1.3435366894197953, "grad_norm": 0.5360599160194397, "learning_rate": 6.745140198569897e-06, "loss": 0.5166, "step": 4199 }, { "epoch": 1.3438566552901023, "grad_norm": 0.4963415265083313, "learning_rate": 6.743395366795229e-06, "loss": 0.4649, "step": 4200 }, { "epoch": 1.3441766211604096, "grad_norm": 0.4736952483654022, "learning_rate": 6.7416502932967e-06, "loss": 0.4616, "step": 4201 }, { "epoch": 1.3444965870307168, "grad_norm": 0.5033711194992065, "learning_rate": 6.73990497831626e-06, "loss": 0.4607, "step": 4202 }, { "epoch": 1.3448165529010239, "grad_norm": 0.515525221824646, "learning_rate": 6.738159422095902e-06, "loss": 0.5365, "step": 4203 }, { "epoch": 1.3451365187713311, "grad_norm": 0.5089691281318665, "learning_rate": 6.736413624877649e-06, "loss": 0.4572, "step": 4204 }, { "epoch": 1.3454564846416384, "grad_norm": 0.5544673204421997, "learning_rate": 6.734667586903557e-06, "loss": 0.495, "step": 4205 }, { "epoch": 1.3457764505119454, "grad_norm": 0.533311128616333, "learning_rate": 6.732921308415719e-06, "loss": 0.5134, "step": 4206 }, { "epoch": 1.3460964163822526, "grad_norm": 0.549479067325592, "learning_rate": 6.7311747896562565e-06, "loss": 0.4981, "step": 4207 }, { "epoch": 1.3464163822525597, "grad_norm": 0.5087676048278809, "learning_rate": 6.729428030867325e-06, "loss": 0.4681, "step": 4208 }, { "epoch": 1.346736348122867, "grad_norm": 0.5456358790397644, "learning_rate": 6.72768103229112e-06, "loss": 0.4963, "step": 4209 }, { "epoch": 1.347056313993174, "grad_norm": 0.539222776889801, "learning_rate": 6.725933794169858e-06, "loss": 0.4599, "step": 4210 }, { "epoch": 1.3473762798634812, "grad_norm": 0.5186460018157959, "learning_rate": 6.724186316745801e-06, "loss": 0.5047, "step": 4211 }, { "epoch": 1.3476962457337884, "grad_norm": 0.5008431673049927, "learning_rate": 6.722438600261234e-06, "loss": 0.4849, "step": 4212 }, { "epoch": 1.3480162116040955, "grad_norm": 0.525508463382721, "learning_rate": 6.7206906449584875e-06, "loss": 0.4877, "step": 4213 }, { "epoch": 1.3483361774744027, "grad_norm": 0.522044837474823, "learning_rate": 6.718942451079911e-06, "loss": 0.5136, "step": 4214 }, { "epoch": 1.34865614334471, "grad_norm": 0.4838278293609619, "learning_rate": 6.7171940188678945e-06, "loss": 0.4731, "step": 4215 }, { "epoch": 1.348976109215017, "grad_norm": 0.5309022068977356, "learning_rate": 6.715445348564863e-06, "loss": 0.5491, "step": 4216 }, { "epoch": 1.3492960750853242, "grad_norm": 0.5318649411201477, "learning_rate": 6.713696440413269e-06, "loss": 0.5208, "step": 4217 }, { "epoch": 1.3496160409556315, "grad_norm": 0.5618005394935608, "learning_rate": 6.711947294655603e-06, "loss": 0.4938, "step": 4218 }, { "epoch": 1.3499360068259385, "grad_norm": 0.5245568752288818, "learning_rate": 6.710197911534384e-06, "loss": 0.4842, "step": 4219 }, { "epoch": 1.3502559726962458, "grad_norm": 0.5300405025482178, "learning_rate": 6.708448291292168e-06, "loss": 0.4699, "step": 4220 }, { "epoch": 1.350575938566553, "grad_norm": 0.5263601541519165, "learning_rate": 6.706698434171542e-06, "loss": 0.4887, "step": 4221 }, { "epoch": 1.35089590443686, "grad_norm": 0.5631449818611145, "learning_rate": 6.704948340415122e-06, "loss": 0.4885, "step": 4222 }, { "epoch": 1.3512158703071673, "grad_norm": 0.5387476682662964, "learning_rate": 6.703198010265566e-06, "loss": 0.4892, "step": 4223 }, { "epoch": 1.3515358361774745, "grad_norm": 0.5017451643943787, "learning_rate": 6.701447443965557e-06, "loss": 0.5033, "step": 4224 }, { "epoch": 1.3518558020477816, "grad_norm": 0.48120734095573425, "learning_rate": 6.699696641757811e-06, "loss": 0.5109, "step": 4225 }, { "epoch": 1.3521757679180888, "grad_norm": 0.48019859194755554, "learning_rate": 6.697945603885082e-06, "loss": 0.4445, "step": 4226 }, { "epoch": 1.3524957337883958, "grad_norm": 0.529334545135498, "learning_rate": 6.6961943305901515e-06, "loss": 0.507, "step": 4227 }, { "epoch": 1.352815699658703, "grad_norm": 0.5923269391059875, "learning_rate": 6.694442822115839e-06, "loss": 0.5348, "step": 4228 }, { "epoch": 1.35313566552901, "grad_norm": 0.5207008123397827, "learning_rate": 6.692691078704988e-06, "loss": 0.5012, "step": 4229 }, { "epoch": 1.3534556313993173, "grad_norm": 0.5076130032539368, "learning_rate": 6.690939100600483e-06, "loss": 0.477, "step": 4230 }, { "epoch": 1.3537755972696246, "grad_norm": 0.5776696801185608, "learning_rate": 6.689186888045239e-06, "loss": 0.4944, "step": 4231 }, { "epoch": 1.3540955631399316, "grad_norm": 0.549407422542572, "learning_rate": 6.687434441282199e-06, "loss": 0.4745, "step": 4232 }, { "epoch": 1.3544155290102389, "grad_norm": 0.5382245182991028, "learning_rate": 6.685681760554346e-06, "loss": 0.5273, "step": 4233 }, { "epoch": 1.3547354948805461, "grad_norm": 0.5502383708953857, "learning_rate": 6.683928846104687e-06, "loss": 0.4917, "step": 4234 }, { "epoch": 1.3550554607508531, "grad_norm": 0.5238730311393738, "learning_rate": 6.6821756981762674e-06, "loss": 0.5307, "step": 4235 }, { "epoch": 1.3553754266211604, "grad_norm": 0.5505203008651733, "learning_rate": 6.680422317012165e-06, "loss": 0.4522, "step": 4236 }, { "epoch": 1.3556953924914676, "grad_norm": 0.5909242630004883, "learning_rate": 6.678668702855485e-06, "loss": 0.5243, "step": 4237 }, { "epoch": 1.3560153583617747, "grad_norm": 0.5503251552581787, "learning_rate": 6.676914855949372e-06, "loss": 0.5046, "step": 4238 }, { "epoch": 1.356335324232082, "grad_norm": 0.516450822353363, "learning_rate": 6.675160776536996e-06, "loss": 0.4225, "step": 4239 }, { "epoch": 1.3566552901023892, "grad_norm": 0.5583712458610535, "learning_rate": 6.673406464861563e-06, "loss": 0.5473, "step": 4240 }, { "epoch": 1.3569752559726962, "grad_norm": 0.4808334708213806, "learning_rate": 6.671651921166311e-06, "loss": 0.4481, "step": 4241 }, { "epoch": 1.3572952218430034, "grad_norm": 0.5560965538024902, "learning_rate": 6.669897145694507e-06, "loss": 0.5281, "step": 4242 }, { "epoch": 1.3576151877133107, "grad_norm": 0.5860428810119629, "learning_rate": 6.668142138689458e-06, "loss": 0.5322, "step": 4243 }, { "epoch": 1.3579351535836177, "grad_norm": 0.5544723272323608, "learning_rate": 6.666386900394493e-06, "loss": 0.4332, "step": 4244 }, { "epoch": 1.358255119453925, "grad_norm": 0.5531538128852844, "learning_rate": 6.664631431052978e-06, "loss": 0.5327, "step": 4245 }, { "epoch": 1.3585750853242322, "grad_norm": 0.5179780125617981, "learning_rate": 6.662875730908313e-06, "loss": 0.4247, "step": 4246 }, { "epoch": 1.3588950511945392, "grad_norm": 0.6032951474189758, "learning_rate": 6.661119800203929e-06, "loss": 0.5362, "step": 4247 }, { "epoch": 1.3592150170648465, "grad_norm": 0.6043846011161804, "learning_rate": 6.659363639183287e-06, "loss": 0.5117, "step": 4248 }, { "epoch": 1.3595349829351535, "grad_norm": 0.5354533195495605, "learning_rate": 6.657607248089879e-06, "loss": 0.4898, "step": 4249 }, { "epoch": 1.3598549488054608, "grad_norm": 0.541419506072998, "learning_rate": 6.65585062716723e-06, "loss": 0.4887, "step": 4250 }, { "epoch": 1.3601749146757678, "grad_norm": 0.5310930609703064, "learning_rate": 6.654093776658902e-06, "loss": 0.452, "step": 4251 }, { "epoch": 1.360494880546075, "grad_norm": 0.6092569828033447, "learning_rate": 6.652336696808481e-06, "loss": 0.5665, "step": 4252 }, { "epoch": 1.3608148464163823, "grad_norm": 0.5411043167114258, "learning_rate": 6.65057938785959e-06, "loss": 0.4336, "step": 4253 }, { "epoch": 1.3611348122866893, "grad_norm": 0.627030074596405, "learning_rate": 6.648821850055881e-06, "loss": 0.5466, "step": 4254 }, { "epoch": 1.3614547781569966, "grad_norm": 0.5255821943283081, "learning_rate": 6.6470640836410385e-06, "loss": 0.4861, "step": 4255 }, { "epoch": 1.3617747440273038, "grad_norm": 0.5354859232902527, "learning_rate": 6.6453060888587785e-06, "loss": 0.5151, "step": 4256 }, { "epoch": 1.3620947098976108, "grad_norm": 0.47319358587265015, "learning_rate": 6.6435478659528505e-06, "loss": 0.4744, "step": 4257 }, { "epoch": 1.362414675767918, "grad_norm": 0.4874507486820221, "learning_rate": 6.641789415167035e-06, "loss": 0.4122, "step": 4258 }, { "epoch": 1.3627346416382253, "grad_norm": 0.679909348487854, "learning_rate": 6.64003073674514e-06, "loss": 0.6077, "step": 4259 }, { "epoch": 1.3630546075085324, "grad_norm": 0.57915860414505, "learning_rate": 6.638271830931011e-06, "loss": 0.4758, "step": 4260 }, { "epoch": 1.3633745733788396, "grad_norm": 0.5016781091690063, "learning_rate": 6.6365126979685226e-06, "loss": 0.4749, "step": 4261 }, { "epoch": 1.3636945392491469, "grad_norm": 0.5453657507896423, "learning_rate": 6.634753338101578e-06, "loss": 0.5459, "step": 4262 }, { "epoch": 1.3640145051194539, "grad_norm": 0.5285664796829224, "learning_rate": 6.632993751574119e-06, "loss": 0.4907, "step": 4263 }, { "epoch": 1.3643344709897611, "grad_norm": 0.5597354173660278, "learning_rate": 6.6312339386301106e-06, "loss": 0.5013, "step": 4264 }, { "epoch": 1.3646544368600684, "grad_norm": 0.47672930359840393, "learning_rate": 6.6294738995135545e-06, "loss": 0.4441, "step": 4265 }, { "epoch": 1.3649744027303754, "grad_norm": 0.5141209959983826, "learning_rate": 6.62771363446848e-06, "loss": 0.4799, "step": 4266 }, { "epoch": 1.3652943686006827, "grad_norm": 0.5273243188858032, "learning_rate": 6.625953143738954e-06, "loss": 0.4501, "step": 4267 }, { "epoch": 1.36561433447099, "grad_norm": 0.5881574749946594, "learning_rate": 6.624192427569067e-06, "loss": 0.5606, "step": 4268 }, { "epoch": 1.365934300341297, "grad_norm": 0.5605153441429138, "learning_rate": 6.622431486202948e-06, "loss": 0.505, "step": 4269 }, { "epoch": 1.3662542662116042, "grad_norm": 0.5217699408531189, "learning_rate": 6.62067031988475e-06, "loss": 0.5574, "step": 4270 }, { "epoch": 1.3665742320819112, "grad_norm": 0.5430951714515686, "learning_rate": 6.618908928858663e-06, "loss": 0.4656, "step": 4271 }, { "epoch": 1.3668941979522184, "grad_norm": 0.5407613515853882, "learning_rate": 6.6171473133689054e-06, "loss": 0.4657, "step": 4272 }, { "epoch": 1.3672141638225255, "grad_norm": 0.5902913212776184, "learning_rate": 6.615385473659726e-06, "loss": 0.5697, "step": 4273 }, { "epoch": 1.3675341296928327, "grad_norm": 0.4799916744232178, "learning_rate": 6.61362340997541e-06, "loss": 0.4782, "step": 4274 }, { "epoch": 1.36785409556314, "grad_norm": 0.48387208580970764, "learning_rate": 6.611861122560266e-06, "loss": 0.4736, "step": 4275 }, { "epoch": 1.368174061433447, "grad_norm": 0.5645774006843567, "learning_rate": 6.610098611658637e-06, "loss": 0.4813, "step": 4276 }, { "epoch": 1.3684940273037542, "grad_norm": 0.6046881079673767, "learning_rate": 6.6083358775149e-06, "loss": 0.5181, "step": 4277 }, { "epoch": 1.3688139931740615, "grad_norm": 0.48757874965667725, "learning_rate": 6.606572920373457e-06, "loss": 0.4181, "step": 4278 }, { "epoch": 1.3691339590443685, "grad_norm": 0.5753948092460632, "learning_rate": 6.604809740478748e-06, "loss": 0.5632, "step": 4279 }, { "epoch": 1.3694539249146758, "grad_norm": 0.5969665050506592, "learning_rate": 6.6030463380752355e-06, "loss": 0.5201, "step": 4280 }, { "epoch": 1.369773890784983, "grad_norm": 0.5898136496543884, "learning_rate": 6.601282713407421e-06, "loss": 0.5166, "step": 4281 }, { "epoch": 1.37009385665529, "grad_norm": 0.5735839605331421, "learning_rate": 6.599518866719832e-06, "loss": 0.5284, "step": 4282 }, { "epoch": 1.3704138225255973, "grad_norm": 0.49882665276527405, "learning_rate": 6.597754798257028e-06, "loss": 0.4944, "step": 4283 }, { "epoch": 1.3707337883959045, "grad_norm": 0.6059930324554443, "learning_rate": 6.595990508263599e-06, "loss": 0.5447, "step": 4284 }, { "epoch": 1.3710537542662116, "grad_norm": 0.48686307668685913, "learning_rate": 6.594225996984165e-06, "loss": 0.4422, "step": 4285 }, { "epoch": 1.3713737201365188, "grad_norm": 0.492951899766922, "learning_rate": 6.592461264663379e-06, "loss": 0.4841, "step": 4286 }, { "epoch": 1.371693686006826, "grad_norm": 0.5390515327453613, "learning_rate": 6.590696311545923e-06, "loss": 0.5077, "step": 4287 }, { "epoch": 1.372013651877133, "grad_norm": 0.5391637086868286, "learning_rate": 6.588931137876511e-06, "loss": 0.4964, "step": 4288 }, { "epoch": 1.3723336177474403, "grad_norm": 0.5298265218734741, "learning_rate": 6.587165743899884e-06, "loss": 0.4985, "step": 4289 }, { "epoch": 1.3726535836177474, "grad_norm": 0.5959964990615845, "learning_rate": 6.585400129860816e-06, "loss": 0.535, "step": 4290 }, { "epoch": 1.3729735494880546, "grad_norm": 0.4907185733318329, "learning_rate": 6.583634296004116e-06, "loss": 0.4545, "step": 4291 }, { "epoch": 1.3732935153583616, "grad_norm": 0.4811200499534607, "learning_rate": 6.581868242574613e-06, "loss": 0.4626, "step": 4292 }, { "epoch": 1.3736134812286689, "grad_norm": 0.5177947282791138, "learning_rate": 6.580101969817176e-06, "loss": 0.4521, "step": 4293 }, { "epoch": 1.3739334470989761, "grad_norm": 0.5208389163017273, "learning_rate": 6.578335477976699e-06, "loss": 0.4762, "step": 4294 }, { "epoch": 1.3742534129692832, "grad_norm": 0.6004898548126221, "learning_rate": 6.5765687672981104e-06, "loss": 0.4973, "step": 4295 }, { "epoch": 1.3745733788395904, "grad_norm": 0.5464395880699158, "learning_rate": 6.574801838026366e-06, "loss": 0.5322, "step": 4296 }, { "epoch": 1.3748933447098977, "grad_norm": 0.5355855822563171, "learning_rate": 6.57303469040645e-06, "loss": 0.4759, "step": 4297 }, { "epoch": 1.3752133105802047, "grad_norm": 0.6042685508728027, "learning_rate": 6.571267324683383e-06, "loss": 0.5253, "step": 4298 }, { "epoch": 1.375533276450512, "grad_norm": 0.5859010219573975, "learning_rate": 6.5694997411022125e-06, "loss": 0.4887, "step": 4299 }, { "epoch": 1.3758532423208192, "grad_norm": 0.5045570731163025, "learning_rate": 6.567731939908013e-06, "loss": 0.4559, "step": 4300 }, { "epoch": 1.3761732081911262, "grad_norm": 0.5356442332267761, "learning_rate": 6.565963921345896e-06, "loss": 0.5476, "step": 4301 }, { "epoch": 1.3764931740614335, "grad_norm": 0.4713646173477173, "learning_rate": 6.564195685660994e-06, "loss": 0.4164, "step": 4302 }, { "epoch": 1.3768131399317407, "grad_norm": 0.5834218263626099, "learning_rate": 6.562427233098482e-06, "loss": 0.536, "step": 4303 }, { "epoch": 1.3771331058020477, "grad_norm": 0.4851571023464203, "learning_rate": 6.560658563903552e-06, "loss": 0.3982, "step": 4304 }, { "epoch": 1.377453071672355, "grad_norm": 0.6995872259140015, "learning_rate": 6.558889678321436e-06, "loss": 0.534, "step": 4305 }, { "epoch": 1.3777730375426622, "grad_norm": 0.5719860792160034, "learning_rate": 6.557120576597391e-06, "loss": 0.516, "step": 4306 }, { "epoch": 1.3780930034129693, "grad_norm": 0.5261143445968628, "learning_rate": 6.5553512589767035e-06, "loss": 0.495, "step": 4307 }, { "epoch": 1.3784129692832765, "grad_norm": 0.5705453157424927, "learning_rate": 6.553581725704694e-06, "loss": 0.4526, "step": 4308 }, { "epoch": 1.3787329351535837, "grad_norm": 0.5274971127510071, "learning_rate": 6.551811977026707e-06, "loss": 0.4554, "step": 4309 }, { "epoch": 1.3790529010238908, "grad_norm": 0.47913479804992676, "learning_rate": 6.550042013188125e-06, "loss": 0.4548, "step": 4310 }, { "epoch": 1.379372866894198, "grad_norm": 0.6112393140792847, "learning_rate": 6.548271834434352e-06, "loss": 0.5369, "step": 4311 }, { "epoch": 1.379692832764505, "grad_norm": 0.5392977595329285, "learning_rate": 6.546501441010826e-06, "loss": 0.4944, "step": 4312 }, { "epoch": 1.3800127986348123, "grad_norm": 0.6074196696281433, "learning_rate": 6.544730833163016e-06, "loss": 0.4915, "step": 4313 }, { "epoch": 1.3803327645051193, "grad_norm": 0.5954985618591309, "learning_rate": 6.542960011136418e-06, "loss": 0.4676, "step": 4314 }, { "epoch": 1.3806527303754266, "grad_norm": 0.5220882296562195, "learning_rate": 6.541188975176557e-06, "loss": 0.488, "step": 4315 }, { "epoch": 1.3809726962457338, "grad_norm": 0.49383068084716797, "learning_rate": 6.539417725528992e-06, "loss": 0.447, "step": 4316 }, { "epoch": 1.3812926621160408, "grad_norm": 0.5188145041465759, "learning_rate": 6.5376462624393045e-06, "loss": 0.4711, "step": 4317 }, { "epoch": 1.381612627986348, "grad_norm": 0.5797480344772339, "learning_rate": 6.535874586153115e-06, "loss": 0.5739, "step": 4318 }, { "epoch": 1.3819325938566553, "grad_norm": 0.5276670455932617, "learning_rate": 6.534102696916066e-06, "loss": 0.4939, "step": 4319 }, { "epoch": 1.3822525597269624, "grad_norm": 0.5280518531799316, "learning_rate": 6.532330594973833e-06, "loss": 0.5221, "step": 4320 }, { "epoch": 1.3825725255972696, "grad_norm": 0.4720649719238281, "learning_rate": 6.530558280572119e-06, "loss": 0.4313, "step": 4321 }, { "epoch": 1.3828924914675769, "grad_norm": 0.524074375629425, "learning_rate": 6.528785753956657e-06, "loss": 0.5351, "step": 4322 }, { "epoch": 1.383212457337884, "grad_norm": 0.543278157711029, "learning_rate": 6.5270130153732135e-06, "loss": 0.486, "step": 4323 }, { "epoch": 1.3835324232081911, "grad_norm": 0.5051714181900024, "learning_rate": 6.525240065067576e-06, "loss": 0.4621, "step": 4324 }, { "epoch": 1.3838523890784984, "grad_norm": 0.5485959649085999, "learning_rate": 6.523466903285571e-06, "loss": 0.5552, "step": 4325 }, { "epoch": 1.3841723549488054, "grad_norm": 0.47958627343177795, "learning_rate": 6.521693530273046e-06, "loss": 0.4453, "step": 4326 }, { "epoch": 1.3844923208191127, "grad_norm": 0.5448989272117615, "learning_rate": 6.519919946275881e-06, "loss": 0.504, "step": 4327 }, { "epoch": 1.38481228668942, "grad_norm": 0.48901090025901794, "learning_rate": 6.518146151539988e-06, "loss": 0.474, "step": 4328 }, { "epoch": 1.385132252559727, "grad_norm": 0.48510274291038513, "learning_rate": 6.516372146311305e-06, "loss": 0.4333, "step": 4329 }, { "epoch": 1.3854522184300342, "grad_norm": 0.5269479155540466, "learning_rate": 6.5145979308358e-06, "loss": 0.5249, "step": 4330 }, { "epoch": 1.3857721843003412, "grad_norm": 0.5366206765174866, "learning_rate": 6.512823505359469e-06, "loss": 0.4811, "step": 4331 }, { "epoch": 1.3860921501706485, "grad_norm": 0.5333788394927979, "learning_rate": 6.511048870128338e-06, "loss": 0.527, "step": 4332 }, { "epoch": 1.3864121160409557, "grad_norm": 0.5139570832252502, "learning_rate": 6.509274025388466e-06, "loss": 0.4685, "step": 4333 }, { "epoch": 1.3867320819112627, "grad_norm": 0.5171332955360413, "learning_rate": 6.5074989713859325e-06, "loss": 0.4854, "step": 4334 }, { "epoch": 1.38705204778157, "grad_norm": 0.5386876463890076, "learning_rate": 6.505723708366854e-06, "loss": 0.5118, "step": 4335 }, { "epoch": 1.387372013651877, "grad_norm": 0.5003767013549805, "learning_rate": 6.503948236577371e-06, "loss": 0.4569, "step": 4336 }, { "epoch": 1.3876919795221843, "grad_norm": 0.5198277235031128, "learning_rate": 6.502172556263656e-06, "loss": 0.5561, "step": 4337 }, { "epoch": 1.3880119453924915, "grad_norm": 0.48306384682655334, "learning_rate": 6.500396667671911e-06, "loss": 0.4224, "step": 4338 }, { "epoch": 1.3883319112627985, "grad_norm": 0.5466201305389404, "learning_rate": 6.498620571048361e-06, "loss": 0.5258, "step": 4339 }, { "epoch": 1.3886518771331058, "grad_norm": 0.5272469520568848, "learning_rate": 6.496844266639266e-06, "loss": 0.5222, "step": 4340 }, { "epoch": 1.388971843003413, "grad_norm": 0.519819974899292, "learning_rate": 6.4950677546909125e-06, "loss": 0.5233, "step": 4341 }, { "epoch": 1.38929180887372, "grad_norm": 0.5432008504867554, "learning_rate": 6.493291035449616e-06, "loss": 0.5145, "step": 4342 }, { "epoch": 1.3896117747440273, "grad_norm": 0.44557735323905945, "learning_rate": 6.491514109161723e-06, "loss": 0.4508, "step": 4343 }, { "epoch": 1.3899317406143346, "grad_norm": 0.5467312335968018, "learning_rate": 6.489736976073603e-06, "loss": 0.5767, "step": 4344 }, { "epoch": 1.3902517064846416, "grad_norm": 0.5047017335891724, "learning_rate": 6.487959636431659e-06, "loss": 0.4873, "step": 4345 }, { "epoch": 1.3905716723549488, "grad_norm": 0.46947747468948364, "learning_rate": 6.486182090482322e-06, "loss": 0.4477, "step": 4346 }, { "epoch": 1.390891638225256, "grad_norm": 0.5482416152954102, "learning_rate": 6.484404338472049e-06, "loss": 0.5166, "step": 4347 }, { "epoch": 1.391211604095563, "grad_norm": 0.4624686539173126, "learning_rate": 6.48262638064733e-06, "loss": 0.4461, "step": 4348 }, { "epoch": 1.3915315699658704, "grad_norm": 0.5751667618751526, "learning_rate": 6.480848217254678e-06, "loss": 0.5293, "step": 4349 }, { "epoch": 1.3918515358361776, "grad_norm": 0.5793061852455139, "learning_rate": 6.479069848540641e-06, "loss": 0.5203, "step": 4350 }, { "epoch": 1.3921715017064846, "grad_norm": 0.5687735676765442, "learning_rate": 6.477291274751789e-06, "loss": 0.5057, "step": 4351 }, { "epoch": 1.3924914675767919, "grad_norm": 0.5311597585678101, "learning_rate": 6.4755124961347245e-06, "loss": 0.5907, "step": 4352 }, { "epoch": 1.392811433447099, "grad_norm": 0.5056808590888977, "learning_rate": 6.473733512936078e-06, "loss": 0.4645, "step": 4353 }, { "epoch": 1.3931313993174061, "grad_norm": 0.5219675302505493, "learning_rate": 6.471954325402508e-06, "loss": 0.5018, "step": 4354 }, { "epoch": 1.3934513651877132, "grad_norm": 0.5257090330123901, "learning_rate": 6.4701749337806975e-06, "loss": 0.4795, "step": 4355 }, { "epoch": 1.3937713310580204, "grad_norm": 0.52811199426651, "learning_rate": 6.468395338317365e-06, "loss": 0.5357, "step": 4356 }, { "epoch": 1.3940912969283277, "grad_norm": 0.5537838339805603, "learning_rate": 6.466615539259252e-06, "loss": 0.4896, "step": 4357 }, { "epoch": 1.3944112627986347, "grad_norm": 0.5262822508811951, "learning_rate": 6.46483553685313e-06, "loss": 0.526, "step": 4358 }, { "epoch": 1.394731228668942, "grad_norm": 0.5014103055000305, "learning_rate": 6.4630553313457985e-06, "loss": 0.4792, "step": 4359 }, { "epoch": 1.3950511945392492, "grad_norm": 0.4720224142074585, "learning_rate": 6.461274922984087e-06, "loss": 0.4431, "step": 4360 }, { "epoch": 1.3953711604095562, "grad_norm": 0.536048173904419, "learning_rate": 6.459494312014848e-06, "loss": 0.535, "step": 4361 }, { "epoch": 1.3956911262798635, "grad_norm": 0.5579946637153625, "learning_rate": 6.457713498684966e-06, "loss": 0.5141, "step": 4362 }, { "epoch": 1.3960110921501707, "grad_norm": 0.5562567710876465, "learning_rate": 6.455932483241355e-06, "loss": 0.4696, "step": 4363 }, { "epoch": 1.3963310580204777, "grad_norm": 0.5763883590698242, "learning_rate": 6.454151265930953e-06, "loss": 0.5313, "step": 4364 }, { "epoch": 1.396651023890785, "grad_norm": 0.5138348937034607, "learning_rate": 6.45236984700073e-06, "loss": 0.4556, "step": 4365 }, { "epoch": 1.3969709897610922, "grad_norm": 0.5289807319641113, "learning_rate": 6.450588226697679e-06, "loss": 0.4584, "step": 4366 }, { "epoch": 1.3972909556313993, "grad_norm": 0.6209538578987122, "learning_rate": 6.448806405268826e-06, "loss": 0.6022, "step": 4367 }, { "epoch": 1.3976109215017065, "grad_norm": 0.5078195333480835, "learning_rate": 6.4470243829612215e-06, "loss": 0.4854, "step": 4368 }, { "epoch": 1.3979308873720138, "grad_norm": 0.504351019859314, "learning_rate": 6.445242160021945e-06, "loss": 0.512, "step": 4369 }, { "epoch": 1.3982508532423208, "grad_norm": 0.5405347943305969, "learning_rate": 6.443459736698106e-06, "loss": 0.4986, "step": 4370 }, { "epoch": 1.398570819112628, "grad_norm": 0.5360117554664612, "learning_rate": 6.441677113236837e-06, "loss": 0.4989, "step": 4371 }, { "epoch": 1.3988907849829353, "grad_norm": 0.4768439829349518, "learning_rate": 6.439894289885303e-06, "loss": 0.5027, "step": 4372 }, { "epoch": 1.3992107508532423, "grad_norm": 0.533896267414093, "learning_rate": 6.438111266890691e-06, "loss": 0.472, "step": 4373 }, { "epoch": 1.3995307167235496, "grad_norm": 0.6232820749282837, "learning_rate": 6.436328044500223e-06, "loss": 0.5015, "step": 4374 }, { "epoch": 1.3998506825938566, "grad_norm": 0.6136919856071472, "learning_rate": 6.434544622961143e-06, "loss": 0.4755, "step": 4375 }, { "epoch": 1.4001706484641638, "grad_norm": 0.5707723498344421, "learning_rate": 6.4327610025207265e-06, "loss": 0.4811, "step": 4376 }, { "epoch": 1.4004906143344709, "grad_norm": 0.5362030267715454, "learning_rate": 6.430977183426273e-06, "loss": 0.5322, "step": 4377 }, { "epoch": 1.400810580204778, "grad_norm": 0.5937615036964417, "learning_rate": 6.4291931659251115e-06, "loss": 0.4952, "step": 4378 }, { "epoch": 1.4011305460750854, "grad_norm": 0.5517301559448242, "learning_rate": 6.427408950264597e-06, "loss": 0.5075, "step": 4379 }, { "epoch": 1.4014505119453924, "grad_norm": 0.5565688610076904, "learning_rate": 6.425624536692116e-06, "loss": 0.4979, "step": 4380 }, { "epoch": 1.4017704778156996, "grad_norm": 0.5186704993247986, "learning_rate": 6.423839925455078e-06, "loss": 0.4758, "step": 4381 }, { "epoch": 1.4020904436860069, "grad_norm": 0.4568119943141937, "learning_rate": 6.422055116800921e-06, "loss": 0.4475, "step": 4382 }, { "epoch": 1.402410409556314, "grad_norm": 0.5236671566963196, "learning_rate": 6.4202701109771105e-06, "loss": 0.4868, "step": 4383 }, { "epoch": 1.4027303754266212, "grad_norm": 0.4925619661808014, "learning_rate": 6.41848490823114e-06, "loss": 0.4861, "step": 4384 }, { "epoch": 1.4030503412969284, "grad_norm": 0.5599465370178223, "learning_rate": 6.416699508810531e-06, "loss": 0.4979, "step": 4385 }, { "epoch": 1.4033703071672354, "grad_norm": 0.5127217173576355, "learning_rate": 6.414913912962832e-06, "loss": 0.4933, "step": 4386 }, { "epoch": 1.4036902730375427, "grad_norm": 0.49806344509124756, "learning_rate": 6.413128120935616e-06, "loss": 0.4744, "step": 4387 }, { "epoch": 1.40401023890785, "grad_norm": 0.5179780125617981, "learning_rate": 6.411342132976486e-06, "loss": 0.5146, "step": 4388 }, { "epoch": 1.404330204778157, "grad_norm": 0.49689823389053345, "learning_rate": 6.409555949333072e-06, "loss": 0.517, "step": 4389 }, { "epoch": 1.4046501706484642, "grad_norm": 0.52426677942276, "learning_rate": 6.4077695702530285e-06, "loss": 0.4991, "step": 4390 }, { "epoch": 1.4049701365187715, "grad_norm": 0.5141168236732483, "learning_rate": 6.405982995984043e-06, "loss": 0.4354, "step": 4391 }, { "epoch": 1.4052901023890785, "grad_norm": 0.565462052822113, "learning_rate": 6.40419622677382e-06, "loss": 0.5294, "step": 4392 }, { "epoch": 1.4056100682593857, "grad_norm": 0.5194235444068909, "learning_rate": 6.402409262870104e-06, "loss": 0.5367, "step": 4393 }, { "epoch": 1.4059300341296928, "grad_norm": 0.5440438389778137, "learning_rate": 6.400622104520655e-06, "loss": 0.5077, "step": 4394 }, { "epoch": 1.40625, "grad_norm": 0.44700324535369873, "learning_rate": 6.398834751973266e-06, "loss": 0.4344, "step": 4395 }, { "epoch": 1.4065699658703072, "grad_norm": 0.4888385236263275, "learning_rate": 6.397047205475757e-06, "loss": 0.4749, "step": 4396 }, { "epoch": 1.4068899317406143, "grad_norm": 0.5607296824455261, "learning_rate": 6.395259465275971e-06, "loss": 0.5707, "step": 4397 }, { "epoch": 1.4072098976109215, "grad_norm": 0.5562505722045898, "learning_rate": 6.393471531621783e-06, "loss": 0.5076, "step": 4398 }, { "epoch": 1.4075298634812285, "grad_norm": 0.5209715366363525, "learning_rate": 6.39168340476109e-06, "loss": 0.4721, "step": 4399 }, { "epoch": 1.4078498293515358, "grad_norm": 0.5225011706352234, "learning_rate": 6.3898950849418184e-06, "loss": 0.4992, "step": 4400 }, { "epoch": 1.408169795221843, "grad_norm": 0.4573118984699249, "learning_rate": 6.388106572411922e-06, "loss": 0.4788, "step": 4401 }, { "epoch": 1.40848976109215, "grad_norm": 0.5134564638137817, "learning_rate": 6.386317867419379e-06, "loss": 0.479, "step": 4402 }, { "epoch": 1.4088097269624573, "grad_norm": 0.5742200016975403, "learning_rate": 6.384528970212196e-06, "loss": 0.5452, "step": 4403 }, { "epoch": 1.4091296928327646, "grad_norm": 0.4976794123649597, "learning_rate": 6.3827398810384045e-06, "loss": 0.4799, "step": 4404 }, { "epoch": 1.4094496587030716, "grad_norm": 0.5459603667259216, "learning_rate": 6.380950600146066e-06, "loss": 0.4634, "step": 4405 }, { "epoch": 1.4097696245733788, "grad_norm": 0.5783873200416565, "learning_rate": 6.379161127783267e-06, "loss": 0.536, "step": 4406 }, { "epoch": 1.410089590443686, "grad_norm": 0.5288960337638855, "learning_rate": 6.377371464198114e-06, "loss": 0.4809, "step": 4407 }, { "epoch": 1.4104095563139931, "grad_norm": 0.5651437044143677, "learning_rate": 6.375581609638754e-06, "loss": 0.5454, "step": 4408 }, { "epoch": 1.4107295221843004, "grad_norm": 0.4710017144680023, "learning_rate": 6.3737915643533484e-06, "loss": 0.45, "step": 4409 }, { "epoch": 1.4110494880546076, "grad_norm": 0.4859638512134552, "learning_rate": 6.3720013285900905e-06, "loss": 0.4643, "step": 4410 }, { "epoch": 1.4113694539249146, "grad_norm": 0.5344699621200562, "learning_rate": 6.370210902597198e-06, "loss": 0.5429, "step": 4411 }, { "epoch": 1.411689419795222, "grad_norm": 0.5420651435852051, "learning_rate": 6.368420286622913e-06, "loss": 0.4771, "step": 4412 }, { "epoch": 1.4120093856655291, "grad_norm": 0.5548089146614075, "learning_rate": 6.366629480915512e-06, "loss": 0.527, "step": 4413 }, { "epoch": 1.4123293515358362, "grad_norm": 0.48039743304252625, "learning_rate": 6.364838485723286e-06, "loss": 0.492, "step": 4414 }, { "epoch": 1.4126493174061434, "grad_norm": 0.513484537601471, "learning_rate": 6.363047301294564e-06, "loss": 0.5029, "step": 4415 }, { "epoch": 1.4129692832764504, "grad_norm": 0.5703129768371582, "learning_rate": 6.361255927877694e-06, "loss": 0.5472, "step": 4416 }, { "epoch": 1.4132892491467577, "grad_norm": 0.5068913102149963, "learning_rate": 6.359464365721051e-06, "loss": 0.4668, "step": 4417 }, { "epoch": 1.4136092150170647, "grad_norm": 0.517008364200592, "learning_rate": 6.357672615073039e-06, "loss": 0.5086, "step": 4418 }, { "epoch": 1.413929180887372, "grad_norm": 0.49030840396881104, "learning_rate": 6.355880676182086e-06, "loss": 0.4621, "step": 4419 }, { "epoch": 1.4142491467576792, "grad_norm": 0.5618411302566528, "learning_rate": 6.354088549296645e-06, "loss": 0.5257, "step": 4420 }, { "epoch": 1.4145691126279862, "grad_norm": 0.5886921286582947, "learning_rate": 6.3522962346651985e-06, "loss": 0.5134, "step": 4421 }, { "epoch": 1.4148890784982935, "grad_norm": 0.5557438731193542, "learning_rate": 6.3505037325362515e-06, "loss": 0.4778, "step": 4422 }, { "epoch": 1.4152090443686007, "grad_norm": 0.5813466310501099, "learning_rate": 6.348711043158339e-06, "loss": 0.558, "step": 4423 }, { "epoch": 1.4155290102389078, "grad_norm": 0.5316801071166992, "learning_rate": 6.346918166780016e-06, "loss": 0.5242, "step": 4424 }, { "epoch": 1.415848976109215, "grad_norm": 0.4986865520477295, "learning_rate": 6.34512510364987e-06, "loss": 0.4553, "step": 4425 }, { "epoch": 1.4161689419795223, "grad_norm": 0.6381187438964844, "learning_rate": 6.3433318540165074e-06, "loss": 0.5365, "step": 4426 }, { "epoch": 1.4164889078498293, "grad_norm": 0.5615751147270203, "learning_rate": 6.34153841812857e-06, "loss": 0.5152, "step": 4427 }, { "epoch": 1.4168088737201365, "grad_norm": 0.5620642900466919, "learning_rate": 6.339744796234717e-06, "loss": 0.5363, "step": 4428 }, { "epoch": 1.4171288395904438, "grad_norm": 0.5495388507843018, "learning_rate": 6.337950988583634e-06, "loss": 0.4252, "step": 4429 }, { "epoch": 1.4174488054607508, "grad_norm": 0.5447618961334229, "learning_rate": 6.336156995424039e-06, "loss": 0.4448, "step": 4430 }, { "epoch": 1.417768771331058, "grad_norm": 0.5106366872787476, "learning_rate": 6.334362817004669e-06, "loss": 0.4755, "step": 4431 }, { "epoch": 1.4180887372013653, "grad_norm": 0.553810179233551, "learning_rate": 6.332568453574288e-06, "loss": 0.5131, "step": 4432 }, { "epoch": 1.4184087030716723, "grad_norm": 0.48646387457847595, "learning_rate": 6.330773905381689e-06, "loss": 0.5283, "step": 4433 }, { "epoch": 1.4187286689419796, "grad_norm": 0.529070258140564, "learning_rate": 6.328979172675685e-06, "loss": 0.4845, "step": 4434 }, { "epoch": 1.4190486348122868, "grad_norm": 0.5212682485580444, "learning_rate": 6.327184255705123e-06, "loss": 0.4631, "step": 4435 }, { "epoch": 1.4193686006825939, "grad_norm": 0.5363638401031494, "learning_rate": 6.325389154718865e-06, "loss": 0.5055, "step": 4436 }, { "epoch": 1.419688566552901, "grad_norm": 0.5298949480056763, "learning_rate": 6.323593869965807e-06, "loss": 0.4998, "step": 4437 }, { "epoch": 1.4200085324232081, "grad_norm": 0.5294914245605469, "learning_rate": 6.3217984016948675e-06, "loss": 0.4782, "step": 4438 }, { "epoch": 1.4203284982935154, "grad_norm": 0.5212501287460327, "learning_rate": 6.320002750154988e-06, "loss": 0.4807, "step": 4439 }, { "epoch": 1.4206484641638224, "grad_norm": 0.5917192697525024, "learning_rate": 6.3182069155951394e-06, "loss": 0.5038, "step": 4440 }, { "epoch": 1.4209684300341296, "grad_norm": 0.5383630394935608, "learning_rate": 6.3164108982643155e-06, "loss": 0.5086, "step": 4441 }, { "epoch": 1.421288395904437, "grad_norm": 0.5691002011299133, "learning_rate": 6.314614698411539e-06, "loss": 0.4683, "step": 4442 }, { "epoch": 1.421608361774744, "grad_norm": 0.5580256581306458, "learning_rate": 6.3128183162858515e-06, "loss": 0.5137, "step": 4443 }, { "epoch": 1.4219283276450512, "grad_norm": 0.5336185693740845, "learning_rate": 6.311021752136325e-06, "loss": 0.5186, "step": 4444 }, { "epoch": 1.4222482935153584, "grad_norm": 0.567459225654602, "learning_rate": 6.309225006212056e-06, "loss": 0.535, "step": 4445 }, { "epoch": 1.4225682593856654, "grad_norm": 0.5339890122413635, "learning_rate": 6.307428078762165e-06, "loss": 0.4494, "step": 4446 }, { "epoch": 1.4228882252559727, "grad_norm": 0.49709126353263855, "learning_rate": 6.305630970035796e-06, "loss": 0.4704, "step": 4447 }, { "epoch": 1.42320819112628, "grad_norm": 0.6274470090866089, "learning_rate": 6.303833680282125e-06, "loss": 0.5234, "step": 4448 }, { "epoch": 1.423528156996587, "grad_norm": 0.5845829844474792, "learning_rate": 6.302036209750344e-06, "loss": 0.4932, "step": 4449 }, { "epoch": 1.4238481228668942, "grad_norm": 0.5564723014831543, "learning_rate": 6.300238558689678e-06, "loss": 0.4898, "step": 4450 }, { "epoch": 1.4241680887372015, "grad_norm": 0.5889071226119995, "learning_rate": 6.2984407273493695e-06, "loss": 0.4758, "step": 4451 }, { "epoch": 1.4244880546075085, "grad_norm": 0.6283670663833618, "learning_rate": 6.296642715978693e-06, "loss": 0.5183, "step": 4452 }, { "epoch": 1.4248080204778157, "grad_norm": 0.5578067302703857, "learning_rate": 6.294844524826944e-06, "loss": 0.476, "step": 4453 }, { "epoch": 1.425127986348123, "grad_norm": 0.54828280210495, "learning_rate": 6.2930461541434426e-06, "loss": 0.526, "step": 4454 }, { "epoch": 1.42544795221843, "grad_norm": 0.5079402327537537, "learning_rate": 6.291247604177539e-06, "loss": 0.4554, "step": 4455 }, { "epoch": 1.4257679180887373, "grad_norm": 0.5570620894432068, "learning_rate": 6.2894488751786e-06, "loss": 0.515, "step": 4456 }, { "epoch": 1.4260878839590443, "grad_norm": 0.5727689266204834, "learning_rate": 6.287649967396024e-06, "loss": 0.4815, "step": 4457 }, { "epoch": 1.4264078498293515, "grad_norm": 0.5417753458023071, "learning_rate": 6.285850881079229e-06, "loss": 0.4616, "step": 4458 }, { "epoch": 1.4267278156996588, "grad_norm": 0.5604990720748901, "learning_rate": 6.284051616477664e-06, "loss": 0.5118, "step": 4459 }, { "epoch": 1.4270477815699658, "grad_norm": 0.48993706703186035, "learning_rate": 6.282252173840797e-06, "loss": 0.5136, "step": 4460 }, { "epoch": 1.427367747440273, "grad_norm": 0.5203538537025452, "learning_rate": 6.280452553418126e-06, "loss": 0.5236, "step": 4461 }, { "epoch": 1.42768771331058, "grad_norm": 0.5383204817771912, "learning_rate": 6.278652755459165e-06, "loss": 0.4484, "step": 4462 }, { "epoch": 1.4280076791808873, "grad_norm": 0.541811466217041, "learning_rate": 6.276852780213463e-06, "loss": 0.5496, "step": 4463 }, { "epoch": 1.4283276450511946, "grad_norm": 0.5643213987350464, "learning_rate": 6.275052627930585e-06, "loss": 0.4532, "step": 4464 }, { "epoch": 1.4286476109215016, "grad_norm": 0.5392845273017883, "learning_rate": 6.273252298860127e-06, "loss": 0.4213, "step": 4465 }, { "epoch": 1.4289675767918089, "grad_norm": 0.5217435956001282, "learning_rate": 6.271451793251707e-06, "loss": 0.5221, "step": 4466 }, { "epoch": 1.429287542662116, "grad_norm": 0.5222741365432739, "learning_rate": 6.269651111354967e-06, "loss": 0.4767, "step": 4467 }, { "epoch": 1.4296075085324231, "grad_norm": 0.5568684339523315, "learning_rate": 6.267850253419571e-06, "loss": 0.4401, "step": 4468 }, { "epoch": 1.4299274744027304, "grad_norm": 0.6233351230621338, "learning_rate": 6.266049219695211e-06, "loss": 0.5179, "step": 4469 }, { "epoch": 1.4302474402730376, "grad_norm": 0.5469030141830444, "learning_rate": 6.264248010431606e-06, "loss": 0.5201, "step": 4470 }, { "epoch": 1.4305674061433447, "grad_norm": 0.5641871094703674, "learning_rate": 6.262446625878493e-06, "loss": 0.4582, "step": 4471 }, { "epoch": 1.430887372013652, "grad_norm": 0.5134510397911072, "learning_rate": 6.260645066285637e-06, "loss": 0.459, "step": 4472 }, { "epoch": 1.4312073378839592, "grad_norm": 0.5164428353309631, "learning_rate": 6.258843331902824e-06, "loss": 0.4997, "step": 4473 }, { "epoch": 1.4315273037542662, "grad_norm": 0.5034913420677185, "learning_rate": 6.257041422979871e-06, "loss": 0.445, "step": 4474 }, { "epoch": 1.4318472696245734, "grad_norm": 0.549512505531311, "learning_rate": 6.25523933976661e-06, "loss": 0.5495, "step": 4475 }, { "epoch": 1.4321672354948807, "grad_norm": 0.5255251526832581, "learning_rate": 6.253437082512908e-06, "loss": 0.4362, "step": 4476 }, { "epoch": 1.4324872013651877, "grad_norm": 0.5761882066726685, "learning_rate": 6.251634651468642e-06, "loss": 0.5355, "step": 4477 }, { "epoch": 1.432807167235495, "grad_norm": 0.4902086555957794, "learning_rate": 6.249832046883729e-06, "loss": 0.4468, "step": 4478 }, { "epoch": 1.433127133105802, "grad_norm": 0.5528424382209778, "learning_rate": 6.248029269008098e-06, "loss": 0.523, "step": 4479 }, { "epoch": 1.4334470989761092, "grad_norm": 0.563441276550293, "learning_rate": 6.246226318091708e-06, "loss": 0.504, "step": 4480 }, { "epoch": 1.4337670648464163, "grad_norm": 0.5314658880233765, "learning_rate": 6.24442319438454e-06, "loss": 0.5354, "step": 4481 }, { "epoch": 1.4340870307167235, "grad_norm": 0.5264476537704468, "learning_rate": 6.242619898136597e-06, "loss": 0.4423, "step": 4482 }, { "epoch": 1.4344069965870307, "grad_norm": 0.5332199335098267, "learning_rate": 6.240816429597913e-06, "loss": 0.5381, "step": 4483 }, { "epoch": 1.4347269624573378, "grad_norm": 0.563339114189148, "learning_rate": 6.239012789018537e-06, "loss": 0.5371, "step": 4484 }, { "epoch": 1.435046928327645, "grad_norm": 0.5181301832199097, "learning_rate": 6.237208976648547e-06, "loss": 0.4461, "step": 4485 }, { "epoch": 1.4353668941979523, "grad_norm": 0.5529221892356873, "learning_rate": 6.235404992738044e-06, "loss": 0.5006, "step": 4486 }, { "epoch": 1.4356868600682593, "grad_norm": 0.5713469386100769, "learning_rate": 6.233600837537153e-06, "loss": 0.547, "step": 4487 }, { "epoch": 1.4360068259385665, "grad_norm": 0.46673092246055603, "learning_rate": 6.231796511296022e-06, "loss": 0.4204, "step": 4488 }, { "epoch": 1.4363267918088738, "grad_norm": 0.5404127240180969, "learning_rate": 6.229992014264822e-06, "loss": 0.4984, "step": 4489 }, { "epoch": 1.4366467576791808, "grad_norm": 0.5437692403793335, "learning_rate": 6.228187346693749e-06, "loss": 0.5257, "step": 4490 }, { "epoch": 1.436966723549488, "grad_norm": 0.57470703125, "learning_rate": 6.2263825088330254e-06, "loss": 0.5077, "step": 4491 }, { "epoch": 1.4372866894197953, "grad_norm": 0.528522789478302, "learning_rate": 6.224577500932888e-06, "loss": 0.4906, "step": 4492 }, { "epoch": 1.4376066552901023, "grad_norm": 0.4943845272064209, "learning_rate": 6.22277232324361e-06, "loss": 0.4605, "step": 4493 }, { "epoch": 1.4379266211604096, "grad_norm": 0.5202118754386902, "learning_rate": 6.220966976015477e-06, "loss": 0.572, "step": 4494 }, { "epoch": 1.4382465870307168, "grad_norm": 0.4969080090522766, "learning_rate": 6.219161459498804e-06, "loss": 0.4913, "step": 4495 }, { "epoch": 1.4385665529010239, "grad_norm": 0.5071524381637573, "learning_rate": 6.217355773943928e-06, "loss": 0.4422, "step": 4496 }, { "epoch": 1.4388865187713311, "grad_norm": 0.47950607538223267, "learning_rate": 6.215549919601209e-06, "loss": 0.4634, "step": 4497 }, { "epoch": 1.4392064846416384, "grad_norm": 0.5288113355636597, "learning_rate": 6.213743896721033e-06, "loss": 0.4738, "step": 4498 }, { "epoch": 1.4395264505119454, "grad_norm": 0.5222241282463074, "learning_rate": 6.2119377055538025e-06, "loss": 0.4789, "step": 4499 }, { "epoch": 1.4398464163822526, "grad_norm": 0.5527712106704712, "learning_rate": 6.210131346349953e-06, "loss": 0.5543, "step": 4500 }, { "epoch": 1.4401663822525597, "grad_norm": 0.48620420694351196, "learning_rate": 6.208324819359939e-06, "loss": 0.4742, "step": 4501 }, { "epoch": 1.440486348122867, "grad_norm": 0.542361855506897, "learning_rate": 6.206518124834231e-06, "loss": 0.5524, "step": 4502 }, { "epoch": 1.440806313993174, "grad_norm": 0.5217463374137878, "learning_rate": 6.204711263023335e-06, "loss": 0.4876, "step": 4503 }, { "epoch": 1.4411262798634812, "grad_norm": 0.5138661861419678, "learning_rate": 6.202904234177772e-06, "loss": 0.4902, "step": 4504 }, { "epoch": 1.4414462457337884, "grad_norm": 0.48745661973953247, "learning_rate": 6.201097038548092e-06, "loss": 0.4861, "step": 4505 }, { "epoch": 1.4417662116040955, "grad_norm": 0.5008990168571472, "learning_rate": 6.199289676384863e-06, "loss": 0.4723, "step": 4506 }, { "epoch": 1.4420861774744027, "grad_norm": 0.565268874168396, "learning_rate": 6.197482147938675e-06, "loss": 0.5103, "step": 4507 }, { "epoch": 1.44240614334471, "grad_norm": 0.49539247155189514, "learning_rate": 6.195674453460148e-06, "loss": 0.4721, "step": 4508 }, { "epoch": 1.442726109215017, "grad_norm": 0.5051780343055725, "learning_rate": 6.193866593199919e-06, "loss": 0.5116, "step": 4509 }, { "epoch": 1.4430460750853242, "grad_norm": 0.4699634909629822, "learning_rate": 6.192058567408652e-06, "loss": 0.4729, "step": 4510 }, { "epoch": 1.4433660409556315, "grad_norm": 0.5447596311569214, "learning_rate": 6.190250376337028e-06, "loss": 0.5723, "step": 4511 }, { "epoch": 1.4436860068259385, "grad_norm": 0.5196283459663391, "learning_rate": 6.18844202023576e-06, "loss": 0.4707, "step": 4512 }, { "epoch": 1.4440059726962458, "grad_norm": 0.5027531981468201, "learning_rate": 6.186633499355576e-06, "loss": 0.4685, "step": 4513 }, { "epoch": 1.444325938566553, "grad_norm": 0.5171422362327576, "learning_rate": 6.1848248139472265e-06, "loss": 0.5018, "step": 4514 }, { "epoch": 1.44464590443686, "grad_norm": 0.5533860921859741, "learning_rate": 6.183015964261495e-06, "loss": 0.5076, "step": 4515 }, { "epoch": 1.4449658703071673, "grad_norm": 0.4939151108264923, "learning_rate": 6.181206950549175e-06, "loss": 0.4772, "step": 4516 }, { "epoch": 1.4452858361774745, "grad_norm": 0.5183248519897461, "learning_rate": 6.179397773061091e-06, "loss": 0.5044, "step": 4517 }, { "epoch": 1.4456058020477816, "grad_norm": 0.5419015884399414, "learning_rate": 6.177588432048087e-06, "loss": 0.4741, "step": 4518 }, { "epoch": 1.4459257679180888, "grad_norm": 0.570713460445404, "learning_rate": 6.175778927761029e-06, "loss": 0.4863, "step": 4519 }, { "epoch": 1.4462457337883958, "grad_norm": 0.49967116117477417, "learning_rate": 6.173969260450809e-06, "loss": 0.4739, "step": 4520 }, { "epoch": 1.446565699658703, "grad_norm": 0.5816405415534973, "learning_rate": 6.172159430368339e-06, "loss": 0.5197, "step": 4521 }, { "epoch": 1.44688566552901, "grad_norm": 0.5697169899940491, "learning_rate": 6.170349437764554e-06, "loss": 0.5279, "step": 4522 }, { "epoch": 1.4472056313993173, "grad_norm": 0.5516278743743896, "learning_rate": 6.1685392828904115e-06, "loss": 0.5231, "step": 4523 }, { "epoch": 1.4475255972696246, "grad_norm": 0.4786688983440399, "learning_rate": 6.16672896599689e-06, "loss": 0.43, "step": 4524 }, { "epoch": 1.4478455631399316, "grad_norm": 0.5326284170150757, "learning_rate": 6.1649184873349955e-06, "loss": 0.5344, "step": 4525 }, { "epoch": 1.4481655290102389, "grad_norm": 0.6142365336418152, "learning_rate": 6.16310784715575e-06, "loss": 0.5301, "step": 4526 }, { "epoch": 1.4484854948805461, "grad_norm": 0.47505971789360046, "learning_rate": 6.161297045710203e-06, "loss": 0.4368, "step": 4527 }, { "epoch": 1.4488054607508531, "grad_norm": 0.5444201827049255, "learning_rate": 6.159486083249424e-06, "loss": 0.5113, "step": 4528 }, { "epoch": 1.4491254266211604, "grad_norm": 0.5231147408485413, "learning_rate": 6.157674960024503e-06, "loss": 0.5127, "step": 4529 }, { "epoch": 1.4494453924914676, "grad_norm": 0.5051755309104919, "learning_rate": 6.155863676286559e-06, "loss": 0.4394, "step": 4530 }, { "epoch": 1.4497653583617747, "grad_norm": 0.5654588937759399, "learning_rate": 6.154052232286724e-06, "loss": 0.6037, "step": 4531 }, { "epoch": 1.450085324232082, "grad_norm": 0.44511279463768005, "learning_rate": 6.152240628276158e-06, "loss": 0.4571, "step": 4532 }, { "epoch": 1.4504052901023892, "grad_norm": 0.49236172437667847, "learning_rate": 6.150428864506043e-06, "loss": 0.4829, "step": 4533 }, { "epoch": 1.4507252559726962, "grad_norm": 0.5119214057922363, "learning_rate": 6.148616941227582e-06, "loss": 0.5184, "step": 4534 }, { "epoch": 1.4510452218430034, "grad_norm": 0.5311828851699829, "learning_rate": 6.146804858692001e-06, "loss": 0.4858, "step": 4535 }, { "epoch": 1.4513651877133107, "grad_norm": 0.5040050745010376, "learning_rate": 6.144992617150547e-06, "loss": 0.5039, "step": 4536 }, { "epoch": 1.4516851535836177, "grad_norm": 0.4733341932296753, "learning_rate": 6.143180216854488e-06, "loss": 0.4451, "step": 4537 }, { "epoch": 1.452005119453925, "grad_norm": 0.5829097628593445, "learning_rate": 6.141367658055116e-06, "loss": 0.5119, "step": 4538 }, { "epoch": 1.4523250853242322, "grad_norm": 0.533114492893219, "learning_rate": 6.139554941003747e-06, "loss": 0.498, "step": 4539 }, { "epoch": 1.4526450511945392, "grad_norm": 0.5461969971656799, "learning_rate": 6.137742065951714e-06, "loss": 0.5466, "step": 4540 }, { "epoch": 1.4529650170648465, "grad_norm": 0.4805932641029358, "learning_rate": 6.135929033150374e-06, "loss": 0.4623, "step": 4541 }, { "epoch": 1.4532849829351535, "grad_norm": 0.5171344876289368, "learning_rate": 6.134115842851107e-06, "loss": 0.4539, "step": 4542 }, { "epoch": 1.4536049488054608, "grad_norm": 0.5133865475654602, "learning_rate": 6.1323024953053155e-06, "loss": 0.4817, "step": 4543 }, { "epoch": 1.4539249146757678, "grad_norm": 0.5248680710792542, "learning_rate": 6.130488990764419e-06, "loss": 0.492, "step": 4544 }, { "epoch": 1.454244880546075, "grad_norm": 0.5106906294822693, "learning_rate": 6.128675329479865e-06, "loss": 0.4955, "step": 4545 }, { "epoch": 1.4545648464163823, "grad_norm": 0.5252261757850647, "learning_rate": 6.126861511703119e-06, "loss": 0.4828, "step": 4546 }, { "epoch": 1.4548848122866893, "grad_norm": 0.5696649551391602, "learning_rate": 6.125047537685668e-06, "loss": 0.4814, "step": 4547 }, { "epoch": 1.4552047781569966, "grad_norm": 0.6102914214134216, "learning_rate": 6.123233407679022e-06, "loss": 0.5212, "step": 4548 }, { "epoch": 1.4555247440273038, "grad_norm": 0.5728787183761597, "learning_rate": 6.121419121934714e-06, "loss": 0.485, "step": 4549 }, { "epoch": 1.4558447098976108, "grad_norm": 0.5707157850265503, "learning_rate": 6.119604680704295e-06, "loss": 0.4785, "step": 4550 }, { "epoch": 1.456164675767918, "grad_norm": 0.5385659337043762, "learning_rate": 6.117790084239341e-06, "loss": 0.5202, "step": 4551 }, { "epoch": 1.4564846416382253, "grad_norm": 0.5896328687667847, "learning_rate": 6.115975332791446e-06, "loss": 0.515, "step": 4552 }, { "epoch": 1.4568046075085324, "grad_norm": 0.510565996170044, "learning_rate": 6.114160426612229e-06, "loss": 0.4589, "step": 4553 }, { "epoch": 1.4571245733788396, "grad_norm": 0.5548239946365356, "learning_rate": 6.112345365953327e-06, "loss": 0.5438, "step": 4554 }, { "epoch": 1.4574445392491469, "grad_norm": 0.4849468171596527, "learning_rate": 6.110530151066404e-06, "loss": 0.4236, "step": 4555 }, { "epoch": 1.4577645051194539, "grad_norm": 0.5229912996292114, "learning_rate": 6.108714782203139e-06, "loss": 0.4821, "step": 4556 }, { "epoch": 1.4580844709897611, "grad_norm": 0.4885832965373993, "learning_rate": 6.1068992596152355e-06, "loss": 0.4836, "step": 4557 }, { "epoch": 1.4584044368600684, "grad_norm": 0.47455790638923645, "learning_rate": 6.1050835835544174e-06, "loss": 0.4684, "step": 4558 }, { "epoch": 1.4587244027303754, "grad_norm": 0.583892822265625, "learning_rate": 6.103267754272431e-06, "loss": 0.5626, "step": 4559 }, { "epoch": 1.4590443686006827, "grad_norm": 0.5615590214729309, "learning_rate": 6.1014517720210444e-06, "loss": 0.4889, "step": 4560 }, { "epoch": 1.45936433447099, "grad_norm": 0.5332754254341125, "learning_rate": 6.099635637052044e-06, "loss": 0.5188, "step": 4561 }, { "epoch": 1.459684300341297, "grad_norm": 0.5477079153060913, "learning_rate": 6.097819349617239e-06, "loss": 0.5068, "step": 4562 }, { "epoch": 1.4600042662116042, "grad_norm": 0.5035220980644226, "learning_rate": 6.096002909968461e-06, "loss": 0.4824, "step": 4563 }, { "epoch": 1.4603242320819112, "grad_norm": 0.5831959843635559, "learning_rate": 6.094186318357561e-06, "loss": 0.5456, "step": 4564 }, { "epoch": 1.4606441979522184, "grad_norm": 0.46603646874427795, "learning_rate": 6.092369575036411e-06, "loss": 0.4154, "step": 4565 }, { "epoch": 1.4609641638225255, "grad_norm": 0.5741982460021973, "learning_rate": 6.090552680256907e-06, "loss": 0.5277, "step": 4566 }, { "epoch": 1.4612841296928327, "grad_norm": 0.4673265814781189, "learning_rate": 6.088735634270959e-06, "loss": 0.4391, "step": 4567 }, { "epoch": 1.46160409556314, "grad_norm": 0.4824434816837311, "learning_rate": 6.086918437330508e-06, "loss": 0.48, "step": 4568 }, { "epoch": 1.461924061433447, "grad_norm": 0.5617666244506836, "learning_rate": 6.0851010896875074e-06, "loss": 0.4755, "step": 4569 }, { "epoch": 1.4622440273037542, "grad_norm": 0.5845708250999451, "learning_rate": 6.0832835915939346e-06, "loss": 0.5971, "step": 4570 }, { "epoch": 1.4625639931740615, "grad_norm": 0.5152300596237183, "learning_rate": 6.08146594330179e-06, "loss": 0.5032, "step": 4571 }, { "epoch": 1.4628839590443685, "grad_norm": 0.4916346073150635, "learning_rate": 6.079648145063089e-06, "loss": 0.4346, "step": 4572 }, { "epoch": 1.4632039249146758, "grad_norm": 0.5176913738250732, "learning_rate": 6.077830197129877e-06, "loss": 0.5062, "step": 4573 }, { "epoch": 1.463523890784983, "grad_norm": 0.5129452347755432, "learning_rate": 6.076012099754209e-06, "loss": 0.4645, "step": 4574 }, { "epoch": 1.46384385665529, "grad_norm": 0.5412490963935852, "learning_rate": 6.074193853188171e-06, "loss": 0.5391, "step": 4575 }, { "epoch": 1.4641638225255973, "grad_norm": 0.534808337688446, "learning_rate": 6.072375457683861e-06, "loss": 0.4695, "step": 4576 }, { "epoch": 1.4644837883959045, "grad_norm": 0.5926088094711304, "learning_rate": 6.070556913493404e-06, "loss": 0.566, "step": 4577 }, { "epoch": 1.4648037542662116, "grad_norm": 0.5333240032196045, "learning_rate": 6.068738220868944e-06, "loss": 0.4734, "step": 4578 }, { "epoch": 1.4651237201365188, "grad_norm": 0.5552252531051636, "learning_rate": 6.066919380062644e-06, "loss": 0.5371, "step": 4579 }, { "epoch": 1.465443686006826, "grad_norm": 0.5706639289855957, "learning_rate": 6.065100391326687e-06, "loss": 0.4958, "step": 4580 }, { "epoch": 1.465763651877133, "grad_norm": 0.557990550994873, "learning_rate": 6.06328125491328e-06, "loss": 0.4884, "step": 4581 }, { "epoch": 1.4660836177474403, "grad_norm": 0.4787043035030365, "learning_rate": 6.061461971074648e-06, "loss": 0.4394, "step": 4582 }, { "epoch": 1.4664035836177474, "grad_norm": 0.5325111746788025, "learning_rate": 6.059642540063038e-06, "loss": 0.5368, "step": 4583 }, { "epoch": 1.4667235494880546, "grad_norm": 0.5462748408317566, "learning_rate": 6.0578229621307115e-06, "loss": 0.4885, "step": 4584 }, { "epoch": 1.4670435153583616, "grad_norm": 0.5821309089660645, "learning_rate": 6.056003237529963e-06, "loss": 0.514, "step": 4585 }, { "epoch": 1.4673634812286689, "grad_norm": 0.5710198879241943, "learning_rate": 6.054183366513092e-06, "loss": 0.4664, "step": 4586 }, { "epoch": 1.4676834470989761, "grad_norm": 0.5251044631004333, "learning_rate": 6.0523633493324306e-06, "loss": 0.4987, "step": 4587 }, { "epoch": 1.4680034129692832, "grad_norm": 0.4900965392589569, "learning_rate": 6.050543186240325e-06, "loss": 0.4147, "step": 4588 }, { "epoch": 1.4683233788395904, "grad_norm": 0.556492030620575, "learning_rate": 6.048722877489141e-06, "loss": 0.5776, "step": 4589 }, { "epoch": 1.4686433447098977, "grad_norm": 0.5008504390716553, "learning_rate": 6.046902423331271e-06, "loss": 0.4207, "step": 4590 }, { "epoch": 1.4689633105802047, "grad_norm": 0.5561898350715637, "learning_rate": 6.045081824019119e-06, "loss": 0.5348, "step": 4591 }, { "epoch": 1.469283276450512, "grad_norm": 0.5290460586547852, "learning_rate": 6.043261079805115e-06, "loss": 0.4541, "step": 4592 }, { "epoch": 1.4696032423208192, "grad_norm": 0.5425384640693665, "learning_rate": 6.041440190941707e-06, "loss": 0.4747, "step": 4593 }, { "epoch": 1.4699232081911262, "grad_norm": 0.5726215243339539, "learning_rate": 6.039619157681363e-06, "loss": 0.5583, "step": 4594 }, { "epoch": 1.4702431740614335, "grad_norm": 0.5443554520606995, "learning_rate": 6.037797980276574e-06, "loss": 0.4511, "step": 4595 }, { "epoch": 1.4705631399317407, "grad_norm": 0.550445020198822, "learning_rate": 6.035976658979846e-06, "loss": 0.5567, "step": 4596 }, { "epoch": 1.4708831058020477, "grad_norm": 0.5292520523071289, "learning_rate": 6.034155194043707e-06, "loss": 0.4503, "step": 4597 }, { "epoch": 1.471203071672355, "grad_norm": 0.5710574984550476, "learning_rate": 6.032333585720707e-06, "loss": 0.4981, "step": 4598 }, { "epoch": 1.4715230375426622, "grad_norm": 0.5461126565933228, "learning_rate": 6.0305118342634125e-06, "loss": 0.5255, "step": 4599 }, { "epoch": 1.4718430034129693, "grad_norm": 0.5360023975372314, "learning_rate": 6.028689939924413e-06, "loss": 0.4961, "step": 4600 }, { "epoch": 1.4721629692832765, "grad_norm": 0.6214020252227783, "learning_rate": 6.026867902956317e-06, "loss": 0.4996, "step": 4601 }, { "epoch": 1.4724829351535837, "grad_norm": 0.5595095157623291, "learning_rate": 6.025045723611749e-06, "loss": 0.5351, "step": 4602 }, { "epoch": 1.4728029010238908, "grad_norm": 0.5649869441986084, "learning_rate": 6.02322340214336e-06, "loss": 0.4554, "step": 4603 }, { "epoch": 1.473122866894198, "grad_norm": 0.6052053570747375, "learning_rate": 6.021400938803813e-06, "loss": 0.5152, "step": 4604 }, { "epoch": 1.473442832764505, "grad_norm": 0.5253227949142456, "learning_rate": 6.019578333845799e-06, "loss": 0.5051, "step": 4605 }, { "epoch": 1.4737627986348123, "grad_norm": 0.5365368723869324, "learning_rate": 6.0177555875220215e-06, "loss": 0.4785, "step": 4606 }, { "epoch": 1.4740827645051193, "grad_norm": 0.6048970818519592, "learning_rate": 6.015932700085207e-06, "loss": 0.485, "step": 4607 }, { "epoch": 1.4744027303754266, "grad_norm": 0.5239445567131042, "learning_rate": 6.014109671788105e-06, "loss": 0.4881, "step": 4608 }, { "epoch": 1.4747226962457338, "grad_norm": 0.5779982209205627, "learning_rate": 6.012286502883472e-06, "loss": 0.5032, "step": 4609 }, { "epoch": 1.4750426621160408, "grad_norm": 0.6136361360549927, "learning_rate": 6.010463193624101e-06, "loss": 0.5191, "step": 4610 }, { "epoch": 1.475362627986348, "grad_norm": 0.5763174295425415, "learning_rate": 6.008639744262791e-06, "loss": 0.4517, "step": 4611 }, { "epoch": 1.4756825938566553, "grad_norm": 0.6773867607116699, "learning_rate": 6.006816155052366e-06, "loss": 0.5122, "step": 4612 }, { "epoch": 1.4760025597269624, "grad_norm": 0.5779669284820557, "learning_rate": 6.004992426245673e-06, "loss": 0.4881, "step": 4613 }, { "epoch": 1.4763225255972696, "grad_norm": 0.5112670063972473, "learning_rate": 6.003168558095568e-06, "loss": 0.4937, "step": 4614 }, { "epoch": 1.4766424914675769, "grad_norm": 0.5309809446334839, "learning_rate": 6.001344550854937e-06, "loss": 0.4664, "step": 4615 }, { "epoch": 1.476962457337884, "grad_norm": 0.6619424223899841, "learning_rate": 5.99952040477668e-06, "loss": 0.5895, "step": 4616 }, { "epoch": 1.4772824232081911, "grad_norm": 0.5281962752342224, "learning_rate": 5.9976961201137155e-06, "loss": 0.4549, "step": 4617 }, { "epoch": 1.4776023890784984, "grad_norm": 0.504805326461792, "learning_rate": 5.995871697118983e-06, "loss": 0.5001, "step": 4618 }, { "epoch": 1.4779223549488054, "grad_norm": 0.5093720555305481, "learning_rate": 5.994047136045442e-06, "loss": 0.4665, "step": 4619 }, { "epoch": 1.4782423208191127, "grad_norm": 0.4796576499938965, "learning_rate": 5.992222437146072e-06, "loss": 0.5012, "step": 4620 }, { "epoch": 1.47856228668942, "grad_norm": 0.505854606628418, "learning_rate": 5.990397600673865e-06, "loss": 0.4341, "step": 4621 }, { "epoch": 1.478882252559727, "grad_norm": 0.5118453502655029, "learning_rate": 5.9885726268818385e-06, "loss": 0.5302, "step": 4622 }, { "epoch": 1.4792022184300342, "grad_norm": 0.5388093590736389, "learning_rate": 5.986747516023031e-06, "loss": 0.4738, "step": 4623 }, { "epoch": 1.4795221843003412, "grad_norm": 0.529425859451294, "learning_rate": 5.98492226835049e-06, "loss": 0.5502, "step": 4624 }, { "epoch": 1.4798421501706485, "grad_norm": 0.5454614758491516, "learning_rate": 5.983096884117294e-06, "loss": 0.4309, "step": 4625 }, { "epoch": 1.4801621160409557, "grad_norm": 0.5218547582626343, "learning_rate": 5.9812713635765315e-06, "loss": 0.5107, "step": 4626 }, { "epoch": 1.4804820819112627, "grad_norm": 0.5063565969467163, "learning_rate": 5.979445706981314e-06, "loss": 0.4839, "step": 4627 }, { "epoch": 1.48080204778157, "grad_norm": 0.5817762613296509, "learning_rate": 5.977619914584771e-06, "loss": 0.4742, "step": 4628 }, { "epoch": 1.481122013651877, "grad_norm": 0.5883211493492126, "learning_rate": 5.975793986640051e-06, "loss": 0.5371, "step": 4629 }, { "epoch": 1.4814419795221843, "grad_norm": 0.5384506583213806, "learning_rate": 5.973967923400321e-06, "loss": 0.5137, "step": 4630 }, { "epoch": 1.4817619453924915, "grad_norm": 0.47395968437194824, "learning_rate": 5.9721417251187655e-06, "loss": 0.4571, "step": 4631 }, { "epoch": 1.4820819112627985, "grad_norm": 0.5172684192657471, "learning_rate": 5.970315392048591e-06, "loss": 0.5435, "step": 4632 }, { "epoch": 1.4824018771331058, "grad_norm": 0.5081861615180969, "learning_rate": 5.96848892444302e-06, "loss": 0.4908, "step": 4633 }, { "epoch": 1.482721843003413, "grad_norm": 0.5542958378791809, "learning_rate": 5.966662322555295e-06, "loss": 0.4779, "step": 4634 }, { "epoch": 1.48304180887372, "grad_norm": 0.4761788547039032, "learning_rate": 5.964835586638674e-06, "loss": 0.4499, "step": 4635 }, { "epoch": 1.4833617747440273, "grad_norm": 0.4623819887638092, "learning_rate": 5.963008716946441e-06, "loss": 0.4911, "step": 4636 }, { "epoch": 1.4836817406143346, "grad_norm": 0.4585903584957123, "learning_rate": 5.96118171373189e-06, "loss": 0.4528, "step": 4637 }, { "epoch": 1.4840017064846416, "grad_norm": 0.5143521428108215, "learning_rate": 5.959354577248337e-06, "loss": 0.5496, "step": 4638 }, { "epoch": 1.4843216723549488, "grad_norm": 0.4974868595600128, "learning_rate": 5.957527307749118e-06, "loss": 0.4704, "step": 4639 }, { "epoch": 1.484641638225256, "grad_norm": 0.5384208559989929, "learning_rate": 5.9556999054875864e-06, "loss": 0.4347, "step": 4640 }, { "epoch": 1.484961604095563, "grad_norm": 0.528374433517456, "learning_rate": 5.953872370717114e-06, "loss": 0.5509, "step": 4641 }, { "epoch": 1.4852815699658704, "grad_norm": 0.45676034688949585, "learning_rate": 5.952044703691089e-06, "loss": 0.4405, "step": 4642 }, { "epoch": 1.4856015358361776, "grad_norm": 0.5470004677772522, "learning_rate": 5.95021690466292e-06, "loss": 0.5111, "step": 4643 }, { "epoch": 1.4859215017064846, "grad_norm": 0.5926286578178406, "learning_rate": 5.948388973886035e-06, "loss": 0.4974, "step": 4644 }, { "epoch": 1.4862414675767919, "grad_norm": 0.5255135893821716, "learning_rate": 5.946560911613877e-06, "loss": 0.5402, "step": 4645 }, { "epoch": 1.486561433447099, "grad_norm": 0.49791911244392395, "learning_rate": 5.94473271809991e-06, "loss": 0.4569, "step": 4646 }, { "epoch": 1.4868813993174061, "grad_norm": 0.45523884892463684, "learning_rate": 5.942904393597616e-06, "loss": 0.4204, "step": 4647 }, { "epoch": 1.4872013651877132, "grad_norm": 0.5056770443916321, "learning_rate": 5.9410759383604925e-06, "loss": 0.55, "step": 4648 }, { "epoch": 1.4875213310580204, "grad_norm": 0.49918273091316223, "learning_rate": 5.939247352642058e-06, "loss": 0.5166, "step": 4649 }, { "epoch": 1.4878412969283277, "grad_norm": 0.512938916683197, "learning_rate": 5.937418636695849e-06, "loss": 0.5038, "step": 4650 }, { "epoch": 1.4881612627986347, "grad_norm": 0.5014441013336182, "learning_rate": 5.935589790775417e-06, "loss": 0.4865, "step": 4651 }, { "epoch": 1.488481228668942, "grad_norm": 0.4981122314929962, "learning_rate": 5.933760815134335e-06, "loss": 0.4903, "step": 4652 }, { "epoch": 1.4888011945392492, "grad_norm": 0.552680253982544, "learning_rate": 5.9319317100261944e-06, "loss": 0.5292, "step": 4653 }, { "epoch": 1.4891211604095562, "grad_norm": 0.4756479561328888, "learning_rate": 5.9301024757046e-06, "loss": 0.4578, "step": 4654 }, { "epoch": 1.4894411262798635, "grad_norm": 0.511592447757721, "learning_rate": 5.928273112423177e-06, "loss": 0.481, "step": 4655 }, { "epoch": 1.4897610921501707, "grad_norm": 0.5142168402671814, "learning_rate": 5.926443620435572e-06, "loss": 0.5207, "step": 4656 }, { "epoch": 1.4900810580204777, "grad_norm": 0.5315825343132019, "learning_rate": 5.924613999995443e-06, "loss": 0.4931, "step": 4657 }, { "epoch": 1.490401023890785, "grad_norm": 0.5098580121994019, "learning_rate": 5.922784251356474e-06, "loss": 0.4662, "step": 4658 }, { "epoch": 1.4907209897610922, "grad_norm": 0.5362950563430786, "learning_rate": 5.920954374772356e-06, "loss": 0.5303, "step": 4659 }, { "epoch": 1.4910409556313993, "grad_norm": 0.5265049338340759, "learning_rate": 5.919124370496805e-06, "loss": 0.4547, "step": 4660 }, { "epoch": 1.4913609215017065, "grad_norm": 0.4737069010734558, "learning_rate": 5.917294238783555e-06, "loss": 0.4993, "step": 4661 }, { "epoch": 1.4916808873720138, "grad_norm": 0.4527972936630249, "learning_rate": 5.915463979886356e-06, "loss": 0.4522, "step": 4662 }, { "epoch": 1.4920008532423208, "grad_norm": 0.5019689798355103, "learning_rate": 5.913633594058975e-06, "loss": 0.4868, "step": 4663 }, { "epoch": 1.492320819112628, "grad_norm": 0.530006468296051, "learning_rate": 5.911803081555197e-06, "loss": 0.4841, "step": 4664 }, { "epoch": 1.4926407849829353, "grad_norm": 0.5467323064804077, "learning_rate": 5.909972442628823e-06, "loss": 0.449, "step": 4665 }, { "epoch": 1.4929607508532423, "grad_norm": 0.5686629414558411, "learning_rate": 5.908141677533676e-06, "loss": 0.508, "step": 4666 }, { "epoch": 1.4932807167235496, "grad_norm": 0.5066972970962524, "learning_rate": 5.906310786523594e-06, "loss": 0.4984, "step": 4667 }, { "epoch": 1.4936006825938566, "grad_norm": 0.5746550559997559, "learning_rate": 5.90447976985243e-06, "loss": 0.4575, "step": 4668 }, { "epoch": 1.4939206484641638, "grad_norm": 0.6012100577354431, "learning_rate": 5.902648627774059e-06, "loss": 0.5184, "step": 4669 }, { "epoch": 1.4942406143344709, "grad_norm": 0.5028852820396423, "learning_rate": 5.900817360542369e-06, "loss": 0.4661, "step": 4670 }, { "epoch": 1.494560580204778, "grad_norm": 0.5793988108634949, "learning_rate": 5.898985968411268e-06, "loss": 0.5398, "step": 4671 }, { "epoch": 1.4948805460750854, "grad_norm": 0.5978303551673889, "learning_rate": 5.897154451634681e-06, "loss": 0.505, "step": 4672 }, { "epoch": 1.4952005119453924, "grad_norm": 0.49366962909698486, "learning_rate": 5.8953228104665505e-06, "loss": 0.4814, "step": 4673 }, { "epoch": 1.4955204778156996, "grad_norm": 0.4858211278915405, "learning_rate": 5.893491045160834e-06, "loss": 0.4624, "step": 4674 }, { "epoch": 1.4958404436860069, "grad_norm": 0.5152391195297241, "learning_rate": 5.891659155971512e-06, "loss": 0.503, "step": 4675 }, { "epoch": 1.496160409556314, "grad_norm": 0.5337923169136047, "learning_rate": 5.8898271431525725e-06, "loss": 0.4369, "step": 4676 }, { "epoch": 1.4964803754266212, "grad_norm": 0.5717154741287231, "learning_rate": 5.887995006958031e-06, "loss": 0.4911, "step": 4677 }, { "epoch": 1.4968003412969284, "grad_norm": 0.6406721472740173, "learning_rate": 5.886162747641912e-06, "loss": 0.5007, "step": 4678 }, { "epoch": 1.4971203071672354, "grad_norm": 0.5949901938438416, "learning_rate": 5.884330365458261e-06, "loss": 0.5153, "step": 4679 }, { "epoch": 1.4974402730375427, "grad_norm": 0.5068642497062683, "learning_rate": 5.882497860661145e-06, "loss": 0.4907, "step": 4680 }, { "epoch": 1.49776023890785, "grad_norm": 0.5365095138549805, "learning_rate": 5.880665233504636e-06, "loss": 0.4697, "step": 4681 }, { "epoch": 1.498080204778157, "grad_norm": 0.5669816732406616, "learning_rate": 5.878832484242833e-06, "loss": 0.5335, "step": 4682 }, { "epoch": 1.4984001706484642, "grad_norm": 0.5329282879829407, "learning_rate": 5.87699961312985e-06, "loss": 0.4687, "step": 4683 }, { "epoch": 1.4987201365187715, "grad_norm": 0.5123113393783569, "learning_rate": 5.875166620419813e-06, "loss": 0.4799, "step": 4684 }, { "epoch": 1.4990401023890785, "grad_norm": 0.5470655560493469, "learning_rate": 5.873333506366874e-06, "loss": 0.5002, "step": 4685 }, { "epoch": 1.4993600682593857, "grad_norm": 0.5343730449676514, "learning_rate": 5.871500271225193e-06, "loss": 0.5105, "step": 4686 }, { "epoch": 1.4996800341296928, "grad_norm": 0.5782073736190796, "learning_rate": 5.869666915248951e-06, "loss": 0.5231, "step": 4687 }, { "epoch": 1.5, "grad_norm": 0.6019355058670044, "learning_rate": 5.867833438692346e-06, "loss": 0.4944, "step": 4688 }, { "epoch": 1.500319965870307, "grad_norm": 0.552795946598053, "learning_rate": 5.8659998418095865e-06, "loss": 0.421, "step": 4689 }, { "epoch": 1.5006399317406145, "grad_norm": 0.5298478603363037, "learning_rate": 5.864166124854911e-06, "loss": 0.5113, "step": 4690 }, { "epoch": 1.5009598976109215, "grad_norm": 0.5174188613891602, "learning_rate": 5.862332288082562e-06, "loss": 0.4433, "step": 4691 }, { "epoch": 1.5012798634812285, "grad_norm": 0.559489905834198, "learning_rate": 5.8604983317468036e-06, "loss": 0.4896, "step": 4692 }, { "epoch": 1.5015998293515358, "grad_norm": 0.5089955925941467, "learning_rate": 5.858664256101917e-06, "loss": 0.4669, "step": 4693 }, { "epoch": 1.501919795221843, "grad_norm": 0.5707961916923523, "learning_rate": 5.856830061402196e-06, "loss": 0.513, "step": 4694 }, { "epoch": 1.50223976109215, "grad_norm": 0.5894507765769958, "learning_rate": 5.854995747901958e-06, "loss": 0.5176, "step": 4695 }, { "epoch": 1.5025597269624573, "grad_norm": 0.5667989253997803, "learning_rate": 5.853161315855531e-06, "loss": 0.4725, "step": 4696 }, { "epoch": 1.5028796928327646, "grad_norm": 0.5133673548698425, "learning_rate": 5.85132676551726e-06, "loss": 0.4423, "step": 4697 }, { "epoch": 1.5031996587030716, "grad_norm": 0.5565043091773987, "learning_rate": 5.8494920971415105e-06, "loss": 0.5359, "step": 4698 }, { "epoch": 1.5035196245733788, "grad_norm": 0.5063019394874573, "learning_rate": 5.847657310982658e-06, "loss": 0.5081, "step": 4699 }, { "epoch": 1.503839590443686, "grad_norm": 0.58235102891922, "learning_rate": 5.845822407295101e-06, "loss": 0.5608, "step": 4700 }, { "epoch": 1.5041595563139931, "grad_norm": 0.4960373044013977, "learning_rate": 5.843987386333248e-06, "loss": 0.4317, "step": 4701 }, { "epoch": 1.5044795221843004, "grad_norm": 0.5047836303710938, "learning_rate": 5.842152248351531e-06, "loss": 0.4673, "step": 4702 }, { "epoch": 1.5047994880546076, "grad_norm": 0.5804183483123779, "learning_rate": 5.84031699360439e-06, "loss": 0.5219, "step": 4703 }, { "epoch": 1.5051194539249146, "grad_norm": 0.5353643298149109, "learning_rate": 5.838481622346286e-06, "loss": 0.5067, "step": 4704 }, { "epoch": 1.505439419795222, "grad_norm": 0.5168746113777161, "learning_rate": 5.836646134831698e-06, "loss": 0.4733, "step": 4705 }, { "epoch": 1.5057593856655291, "grad_norm": 0.5259865522384644, "learning_rate": 5.834810531315117e-06, "loss": 0.4797, "step": 4706 }, { "epoch": 1.5060793515358362, "grad_norm": 0.5396210551261902, "learning_rate": 5.832974812051051e-06, "loss": 0.463, "step": 4707 }, { "epoch": 1.5063993174061432, "grad_norm": 0.5567886233329773, "learning_rate": 5.831138977294025e-06, "loss": 0.6018, "step": 4708 }, { "epoch": 1.5067192832764507, "grad_norm": 0.48316898941993713, "learning_rate": 5.8293030272985794e-06, "loss": 0.4108, "step": 4709 }, { "epoch": 1.5070392491467577, "grad_norm": 0.5766081213951111, "learning_rate": 5.827466962319274e-06, "loss": 0.5038, "step": 4710 }, { "epoch": 1.5073592150170647, "grad_norm": 0.5991988778114319, "learning_rate": 5.825630782610676e-06, "loss": 0.477, "step": 4711 }, { "epoch": 1.507679180887372, "grad_norm": 0.5347451567649841, "learning_rate": 5.823794488427378e-06, "loss": 0.4746, "step": 4712 }, { "epoch": 1.5079991467576792, "grad_norm": 0.5310901403427124, "learning_rate": 5.821958080023982e-06, "loss": 0.5233, "step": 4713 }, { "epoch": 1.5083191126279862, "grad_norm": 0.5356836915016174, "learning_rate": 5.820121557655109e-06, "loss": 0.4868, "step": 4714 }, { "epoch": 1.5086390784982935, "grad_norm": 0.5663773417472839, "learning_rate": 5.818284921575398e-06, "loss": 0.5044, "step": 4715 }, { "epoch": 1.5089590443686007, "grad_norm": 0.5553083419799805, "learning_rate": 5.816448172039497e-06, "loss": 0.4643, "step": 4716 }, { "epoch": 1.5092790102389078, "grad_norm": 0.5697829723358154, "learning_rate": 5.814611309302074e-06, "loss": 0.5406, "step": 4717 }, { "epoch": 1.509598976109215, "grad_norm": 0.5327697992324829, "learning_rate": 5.8127743336178136e-06, "loss": 0.4588, "step": 4718 }, { "epoch": 1.5099189419795223, "grad_norm": 0.5630555748939514, "learning_rate": 5.810937245241414e-06, "loss": 0.4749, "step": 4719 }, { "epoch": 1.5102389078498293, "grad_norm": 0.5737113952636719, "learning_rate": 5.8091000444275904e-06, "loss": 0.505, "step": 4720 }, { "epoch": 1.5105588737201365, "grad_norm": 0.5312773585319519, "learning_rate": 5.807262731431069e-06, "loss": 0.4828, "step": 4721 }, { "epoch": 1.5108788395904438, "grad_norm": 0.5531858205795288, "learning_rate": 5.805425306506602e-06, "loss": 0.5074, "step": 4722 }, { "epoch": 1.5111988054607508, "grad_norm": 0.48062023520469666, "learning_rate": 5.8035877699089475e-06, "loss": 0.484, "step": 4723 }, { "epoch": 1.511518771331058, "grad_norm": 0.46315205097198486, "learning_rate": 5.801750121892881e-06, "loss": 0.4692, "step": 4724 }, { "epoch": 1.5118387372013653, "grad_norm": 0.5213401913642883, "learning_rate": 5.799912362713195e-06, "loss": 0.4811, "step": 4725 }, { "epoch": 1.5121587030716723, "grad_norm": 0.535945475101471, "learning_rate": 5.798074492624698e-06, "loss": 0.4338, "step": 4726 }, { "epoch": 1.5124786689419796, "grad_norm": 0.5230036973953247, "learning_rate": 5.796236511882214e-06, "loss": 0.4526, "step": 4727 }, { "epoch": 1.5127986348122868, "grad_norm": 0.5411539077758789, "learning_rate": 5.7943984207405776e-06, "loss": 0.5429, "step": 4728 }, { "epoch": 1.5131186006825939, "grad_norm": 0.5598258376121521, "learning_rate": 5.792560219454645e-06, "loss": 0.478, "step": 4729 }, { "epoch": 1.5134385665529009, "grad_norm": 0.5569031834602356, "learning_rate": 5.790721908279284e-06, "loss": 0.504, "step": 4730 }, { "epoch": 1.5137585324232083, "grad_norm": 0.5236627459526062, "learning_rate": 5.78888348746938e-06, "loss": 0.472, "step": 4731 }, { "epoch": 1.5140784982935154, "grad_norm": 0.5903027057647705, "learning_rate": 5.787044957279833e-06, "loss": 0.5809, "step": 4732 }, { "epoch": 1.5143984641638224, "grad_norm": 0.4684828519821167, "learning_rate": 5.785206317965554e-06, "loss": 0.4436, "step": 4733 }, { "epoch": 1.5147184300341296, "grad_norm": 0.5297166109085083, "learning_rate": 5.783367569781474e-06, "loss": 0.4934, "step": 4734 }, { "epoch": 1.515038395904437, "grad_norm": 0.5144308805465698, "learning_rate": 5.781528712982538e-06, "loss": 0.4815, "step": 4735 }, { "epoch": 1.515358361774744, "grad_norm": 0.5451584458351135, "learning_rate": 5.779689747823707e-06, "loss": 0.466, "step": 4736 }, { "epoch": 1.5156783276450512, "grad_norm": 0.509016215801239, "learning_rate": 5.777850674559954e-06, "loss": 0.4894, "step": 4737 }, { "epoch": 1.5159982935153584, "grad_norm": 0.5265939831733704, "learning_rate": 5.776011493446269e-06, "loss": 0.5545, "step": 4738 }, { "epoch": 1.5163182593856654, "grad_norm": 0.5136027932167053, "learning_rate": 5.7741722047376565e-06, "loss": 0.4837, "step": 4739 }, { "epoch": 1.5166382252559727, "grad_norm": 0.5121413469314575, "learning_rate": 5.772332808689136e-06, "loss": 0.4503, "step": 4740 }, { "epoch": 1.51695819112628, "grad_norm": 0.6478096842765808, "learning_rate": 5.770493305555743e-06, "loss": 0.5672, "step": 4741 }, { "epoch": 1.517278156996587, "grad_norm": 0.514471709728241, "learning_rate": 5.768653695592527e-06, "loss": 0.5323, "step": 4742 }, { "epoch": 1.5175981228668942, "grad_norm": 0.46245503425598145, "learning_rate": 5.766813979054552e-06, "loss": 0.4682, "step": 4743 }, { "epoch": 1.5179180887372015, "grad_norm": 0.5243746638298035, "learning_rate": 5.764974156196895e-06, "loss": 0.4612, "step": 4744 }, { "epoch": 1.5182380546075085, "grad_norm": 0.5565612316131592, "learning_rate": 5.76313422727465e-06, "loss": 0.4659, "step": 4745 }, { "epoch": 1.5185580204778157, "grad_norm": 0.48753592371940613, "learning_rate": 5.761294192542928e-06, "loss": 0.4612, "step": 4746 }, { "epoch": 1.518877986348123, "grad_norm": 0.5521674156188965, "learning_rate": 5.7594540522568495e-06, "loss": 0.4912, "step": 4747 }, { "epoch": 1.51919795221843, "grad_norm": 0.5125954151153564, "learning_rate": 5.757613806671555e-06, "loss": 0.5305, "step": 4748 }, { "epoch": 1.519517918088737, "grad_norm": 0.474223256111145, "learning_rate": 5.755773456042193e-06, "loss": 0.4656, "step": 4749 }, { "epoch": 1.5198378839590445, "grad_norm": 0.496297687292099, "learning_rate": 5.753933000623932e-06, "loss": 0.4613, "step": 4750 }, { "epoch": 1.5201578498293515, "grad_norm": 0.5795848369598389, "learning_rate": 5.752092440671952e-06, "loss": 0.4818, "step": 4751 }, { "epoch": 1.5204778156996586, "grad_norm": 0.6200262308120728, "learning_rate": 5.750251776441453e-06, "loss": 0.5102, "step": 4752 }, { "epoch": 1.520797781569966, "grad_norm": 0.5648042559623718, "learning_rate": 5.748411008187642e-06, "loss": 0.5391, "step": 4753 }, { "epoch": 1.521117747440273, "grad_norm": 0.5272075533866882, "learning_rate": 5.746570136165744e-06, "loss": 0.4633, "step": 4754 }, { "epoch": 1.52143771331058, "grad_norm": 0.5459337830543518, "learning_rate": 5.744729160630998e-06, "loss": 0.5066, "step": 4755 }, { "epoch": 1.5217576791808873, "grad_norm": 0.5689736604690552, "learning_rate": 5.7428880818386574e-06, "loss": 0.5233, "step": 4756 }, { "epoch": 1.5220776450511946, "grad_norm": 0.4891079366207123, "learning_rate": 5.741046900043992e-06, "loss": 0.439, "step": 4757 }, { "epoch": 1.5223976109215016, "grad_norm": 0.5367847084999084, "learning_rate": 5.7392056155022835e-06, "loss": 0.5036, "step": 4758 }, { "epoch": 1.5227175767918089, "grad_norm": 0.5277146100997925, "learning_rate": 5.737364228468823e-06, "loss": 0.4871, "step": 4759 }, { "epoch": 1.523037542662116, "grad_norm": 0.5709060430526733, "learning_rate": 5.73552273919893e-06, "loss": 0.5384, "step": 4760 }, { "epoch": 1.5233575085324231, "grad_norm": 0.5281832218170166, "learning_rate": 5.7336811479479215e-06, "loss": 0.469, "step": 4761 }, { "epoch": 1.5236774744027304, "grad_norm": 0.5688793659210205, "learning_rate": 5.7318394549711395e-06, "loss": 0.4974, "step": 4762 }, { "epoch": 1.5239974402730376, "grad_norm": 0.586227297782898, "learning_rate": 5.729997660523938e-06, "loss": 0.5188, "step": 4763 }, { "epoch": 1.5243174061433447, "grad_norm": 0.5579555034637451, "learning_rate": 5.72815576486168e-06, "loss": 0.5151, "step": 4764 }, { "epoch": 1.524637372013652, "grad_norm": 0.5547084808349609, "learning_rate": 5.7263137682397515e-06, "loss": 0.5007, "step": 4765 }, { "epoch": 1.5249573378839592, "grad_norm": 0.5869400501251221, "learning_rate": 5.724471670913545e-06, "loss": 0.508, "step": 4766 }, { "epoch": 1.5252773037542662, "grad_norm": 0.5429580211639404, "learning_rate": 5.722629473138469e-06, "loss": 0.4988, "step": 4767 }, { "epoch": 1.5255972696245734, "grad_norm": 0.528658390045166, "learning_rate": 5.720787175169949e-06, "loss": 0.5182, "step": 4768 }, { "epoch": 1.5259172354948807, "grad_norm": 0.5507444143295288, "learning_rate": 5.718944777263417e-06, "loss": 0.5175, "step": 4769 }, { "epoch": 1.5262372013651877, "grad_norm": 0.48885491490364075, "learning_rate": 5.717102279674329e-06, "loss": 0.4476, "step": 4770 }, { "epoch": 1.5265571672354947, "grad_norm": 0.5946596264839172, "learning_rate": 5.715259682658145e-06, "loss": 0.5351, "step": 4771 }, { "epoch": 1.5268771331058022, "grad_norm": 0.5150769948959351, "learning_rate": 5.713416986470348e-06, "loss": 0.4542, "step": 4772 }, { "epoch": 1.5271970989761092, "grad_norm": 0.5974960327148438, "learning_rate": 5.711574191366427e-06, "loss": 0.5476, "step": 4773 }, { "epoch": 1.5275170648464163, "grad_norm": 0.545831024646759, "learning_rate": 5.709731297601887e-06, "loss": 0.4823, "step": 4774 }, { "epoch": 1.5278370307167235, "grad_norm": 0.552084743976593, "learning_rate": 5.70788830543225e-06, "loss": 0.4665, "step": 4775 }, { "epoch": 1.5281569965870307, "grad_norm": 0.5725771188735962, "learning_rate": 5.706045215113047e-06, "loss": 0.4726, "step": 4776 }, { "epoch": 1.5284769624573378, "grad_norm": 0.5605619549751282, "learning_rate": 5.704202026899827e-06, "loss": 0.4549, "step": 4777 }, { "epoch": 1.528796928327645, "grad_norm": 0.5407191514968872, "learning_rate": 5.702358741048149e-06, "loss": 0.4926, "step": 4778 }, { "epoch": 1.5291168941979523, "grad_norm": 0.573394775390625, "learning_rate": 5.700515357813583e-06, "loss": 0.5661, "step": 4779 }, { "epoch": 1.5294368600682593, "grad_norm": 0.5355685949325562, "learning_rate": 5.698671877451723e-06, "loss": 0.5147, "step": 4780 }, { "epoch": 1.5297568259385665, "grad_norm": 0.5256187915802002, "learning_rate": 5.696828300218167e-06, "loss": 0.4632, "step": 4781 }, { "epoch": 1.5300767918088738, "grad_norm": 0.45056822896003723, "learning_rate": 5.694984626368527e-06, "loss": 0.4563, "step": 4782 }, { "epoch": 1.5303967576791808, "grad_norm": 0.5010212659835815, "learning_rate": 5.693140856158435e-06, "loss": 0.4869, "step": 4783 }, { "epoch": 1.530716723549488, "grad_norm": 0.4796273410320282, "learning_rate": 5.691296989843525e-06, "loss": 0.4452, "step": 4784 }, { "epoch": 1.5310366894197953, "grad_norm": 0.5340262055397034, "learning_rate": 5.6894530276794605e-06, "loss": 0.5197, "step": 4785 }, { "epoch": 1.5313566552901023, "grad_norm": 0.4710986912250519, "learning_rate": 5.6876089699219016e-06, "loss": 0.4281, "step": 4786 }, { "epoch": 1.5316766211604096, "grad_norm": 0.5724782943725586, "learning_rate": 5.685764816826532e-06, "loss": 0.5371, "step": 4787 }, { "epoch": 1.5319965870307168, "grad_norm": 0.4858018457889557, "learning_rate": 5.6839205686490474e-06, "loss": 0.4442, "step": 4788 }, { "epoch": 1.5323165529010239, "grad_norm": 0.4786800444126129, "learning_rate": 5.68207622564515e-06, "loss": 0.4393, "step": 4789 }, { "epoch": 1.5326365187713311, "grad_norm": 0.5682559609413147, "learning_rate": 5.680231788070566e-06, "loss": 0.4934, "step": 4790 }, { "epoch": 1.5329564846416384, "grad_norm": 0.5499330759048462, "learning_rate": 5.678387256181024e-06, "loss": 0.5285, "step": 4791 }, { "epoch": 1.5332764505119454, "grad_norm": 0.5110037922859192, "learning_rate": 5.6765426302322755e-06, "loss": 0.462, "step": 4792 }, { "epoch": 1.5335964163822524, "grad_norm": 0.5054475665092468, "learning_rate": 5.674697910480077e-06, "loss": 0.4424, "step": 4793 }, { "epoch": 1.5339163822525599, "grad_norm": 0.46804389357566833, "learning_rate": 5.6728530971801986e-06, "loss": 0.4438, "step": 4794 }, { "epoch": 1.534236348122867, "grad_norm": 0.5104069113731384, "learning_rate": 5.671008190588432e-06, "loss": 0.5263, "step": 4795 }, { "epoch": 1.534556313993174, "grad_norm": 0.4755001664161682, "learning_rate": 5.6691631909605705e-06, "loss": 0.4567, "step": 4796 }, { "epoch": 1.5348762798634812, "grad_norm": 0.6045056581497192, "learning_rate": 5.667318098552429e-06, "loss": 0.5137, "step": 4797 }, { "epoch": 1.5351962457337884, "grad_norm": 0.5286843776702881, "learning_rate": 5.66547291361983e-06, "loss": 0.4719, "step": 4798 }, { "epoch": 1.5355162116040955, "grad_norm": 0.5271816849708557, "learning_rate": 5.663627636418611e-06, "loss": 0.5051, "step": 4799 }, { "epoch": 1.5358361774744027, "grad_norm": 0.5123486518859863, "learning_rate": 5.661782267204622e-06, "loss": 0.4481, "step": 4800 }, { "epoch": 1.53615614334471, "grad_norm": 0.589840292930603, "learning_rate": 5.659936806233723e-06, "loss": 0.5574, "step": 4801 }, { "epoch": 1.536476109215017, "grad_norm": 0.5192856788635254, "learning_rate": 5.658091253761795e-06, "loss": 0.4439, "step": 4802 }, { "epoch": 1.5367960750853242, "grad_norm": 0.546211838722229, "learning_rate": 5.656245610044722e-06, "loss": 0.5547, "step": 4803 }, { "epoch": 1.5371160409556315, "grad_norm": 0.49008166790008545, "learning_rate": 5.654399875338406e-06, "loss": 0.4649, "step": 4804 }, { "epoch": 1.5374360068259385, "grad_norm": 0.5199187397956848, "learning_rate": 5.65255404989876e-06, "loss": 0.4941, "step": 4805 }, { "epoch": 1.5377559726962458, "grad_norm": 0.5492860078811646, "learning_rate": 5.650708133981708e-06, "loss": 0.5224, "step": 4806 }, { "epoch": 1.538075938566553, "grad_norm": 0.5844526886940002, "learning_rate": 5.648862127843193e-06, "loss": 0.5619, "step": 4807 }, { "epoch": 1.53839590443686, "grad_norm": 0.5239481925964355, "learning_rate": 5.647016031739162e-06, "loss": 0.4662, "step": 4808 }, { "epoch": 1.5387158703071673, "grad_norm": 0.49354660511016846, "learning_rate": 5.6451698459255795e-06, "loss": 0.458, "step": 4809 }, { "epoch": 1.5390358361774745, "grad_norm": 0.5162451863288879, "learning_rate": 5.6433235706584245e-06, "loss": 0.4691, "step": 4810 }, { "epoch": 1.5393558020477816, "grad_norm": 0.5093298554420471, "learning_rate": 5.641477206193678e-06, "loss": 0.4694, "step": 4811 }, { "epoch": 1.5396757679180886, "grad_norm": 0.6124956607818604, "learning_rate": 5.639630752787349e-06, "loss": 0.5314, "step": 4812 }, { "epoch": 1.539995733788396, "grad_norm": 0.5342500805854797, "learning_rate": 5.637784210695445e-06, "loss": 0.4845, "step": 4813 }, { "epoch": 1.540315699658703, "grad_norm": 0.5140454769134521, "learning_rate": 5.6359375801739925e-06, "loss": 0.504, "step": 4814 }, { "epoch": 1.54063566552901, "grad_norm": 0.5607699155807495, "learning_rate": 5.63409086147903e-06, "loss": 0.4849, "step": 4815 }, { "epoch": 1.5409556313993176, "grad_norm": 0.5525639057159424, "learning_rate": 5.632244054866607e-06, "loss": 0.5025, "step": 4816 }, { "epoch": 1.5412755972696246, "grad_norm": 0.5072596073150635, "learning_rate": 5.630397160592785e-06, "loss": 0.4693, "step": 4817 }, { "epoch": 1.5415955631399316, "grad_norm": 0.47427666187286377, "learning_rate": 5.628550178913639e-06, "loss": 0.4643, "step": 4818 }, { "epoch": 1.5419155290102389, "grad_norm": 0.5358697772026062, "learning_rate": 5.626703110085253e-06, "loss": 0.589, "step": 4819 }, { "epoch": 1.5422354948805461, "grad_norm": 0.5243690609931946, "learning_rate": 5.624855954363727e-06, "loss": 0.4273, "step": 4820 }, { "epoch": 1.5425554607508531, "grad_norm": 0.568854570388794, "learning_rate": 5.6230087120051724e-06, "loss": 0.5026, "step": 4821 }, { "epoch": 1.5428754266211604, "grad_norm": 0.6009626984596252, "learning_rate": 5.62116138326571e-06, "loss": 0.4867, "step": 4822 }, { "epoch": 1.5431953924914676, "grad_norm": 0.5987993478775024, "learning_rate": 5.6193139684014745e-06, "loss": 0.47, "step": 4823 }, { "epoch": 1.5435153583617747, "grad_norm": 0.5260308980941772, "learning_rate": 5.617466467668612e-06, "loss": 0.5007, "step": 4824 }, { "epoch": 1.543835324232082, "grad_norm": 0.5159297585487366, "learning_rate": 5.6156188813232806e-06, "loss": 0.4762, "step": 4825 }, { "epoch": 1.5441552901023892, "grad_norm": 0.4811621606349945, "learning_rate": 5.6137712096216515e-06, "loss": 0.5078, "step": 4826 }, { "epoch": 1.5444752559726962, "grad_norm": 0.4753594696521759, "learning_rate": 5.611923452819905e-06, "loss": 0.443, "step": 4827 }, { "epoch": 1.5447952218430034, "grad_norm": 0.5283935070037842, "learning_rate": 5.610075611174236e-06, "loss": 0.4821, "step": 4828 }, { "epoch": 1.5451151877133107, "grad_norm": 0.5373314023017883, "learning_rate": 5.60822768494085e-06, "loss": 0.5089, "step": 4829 }, { "epoch": 1.5454351535836177, "grad_norm": 0.5004515647888184, "learning_rate": 5.606379674375962e-06, "loss": 0.4773, "step": 4830 }, { "epoch": 1.545755119453925, "grad_norm": 0.490653395652771, "learning_rate": 5.6045315797358046e-06, "loss": 0.4468, "step": 4831 }, { "epoch": 1.5460750853242322, "grad_norm": 0.5463707447052002, "learning_rate": 5.6026834012766155e-06, "loss": 0.5, "step": 4832 }, { "epoch": 1.5463950511945392, "grad_norm": 0.4810127317905426, "learning_rate": 5.600835139254649e-06, "loss": 0.479, "step": 4833 }, { "epoch": 1.5467150170648463, "grad_norm": 0.5389829277992249, "learning_rate": 5.598986793926167e-06, "loss": 0.4864, "step": 4834 }, { "epoch": 1.5470349829351537, "grad_norm": 0.5847005844116211, "learning_rate": 5.5971383655474445e-06, "loss": 0.4849, "step": 4835 }, { "epoch": 1.5473549488054608, "grad_norm": 0.5487697720527649, "learning_rate": 5.595289854374769e-06, "loss": 0.4645, "step": 4836 }, { "epoch": 1.5476749146757678, "grad_norm": 0.5846667289733887, "learning_rate": 5.5934412606644404e-06, "loss": 0.5447, "step": 4837 }, { "epoch": 1.547994880546075, "grad_norm": 0.5189180970191956, "learning_rate": 5.591592584672767e-06, "loss": 0.398, "step": 4838 }, { "epoch": 1.5483148464163823, "grad_norm": 0.5232536196708679, "learning_rate": 5.58974382665607e-06, "loss": 0.4891, "step": 4839 }, { "epoch": 1.5486348122866893, "grad_norm": 0.5266755223274231, "learning_rate": 5.587894986870681e-06, "loss": 0.4772, "step": 4840 }, { "epoch": 1.5489547781569966, "grad_norm": 0.45549291372299194, "learning_rate": 5.586046065572945e-06, "loss": 0.424, "step": 4841 }, { "epoch": 1.5492747440273038, "grad_norm": 0.5046857595443726, "learning_rate": 5.5841970630192165e-06, "loss": 0.4693, "step": 4842 }, { "epoch": 1.5495947098976108, "grad_norm": 0.6059446930885315, "learning_rate": 5.582347979465864e-06, "loss": 0.5427, "step": 4843 }, { "epoch": 1.549914675767918, "grad_norm": 0.5254539251327515, "learning_rate": 5.580498815169263e-06, "loss": 0.4544, "step": 4844 }, { "epoch": 1.5502346416382253, "grad_norm": 0.49290019273757935, "learning_rate": 5.578649570385803e-06, "loss": 0.4965, "step": 4845 }, { "epoch": 1.5505546075085324, "grad_norm": 0.4934821128845215, "learning_rate": 5.576800245371883e-06, "loss": 0.4897, "step": 4846 }, { "epoch": 1.5508745733788396, "grad_norm": 0.5148047208786011, "learning_rate": 5.5749508403839155e-06, "loss": 0.4829, "step": 4847 }, { "epoch": 1.5511945392491469, "grad_norm": 0.48020872473716736, "learning_rate": 5.5731013556783235e-06, "loss": 0.4699, "step": 4848 }, { "epoch": 1.5515145051194539, "grad_norm": 0.5449939370155334, "learning_rate": 5.571251791511537e-06, "loss": 0.4994, "step": 4849 }, { "epoch": 1.5518344709897611, "grad_norm": 0.5520308017730713, "learning_rate": 5.569402148140004e-06, "loss": 0.5094, "step": 4850 }, { "epoch": 1.5521544368600684, "grad_norm": 0.5400788187980652, "learning_rate": 5.567552425820177e-06, "loss": 0.4565, "step": 4851 }, { "epoch": 1.5524744027303754, "grad_norm": 0.6016430258750916, "learning_rate": 5.565702624808524e-06, "loss": 0.5279, "step": 4852 }, { "epoch": 1.5527943686006824, "grad_norm": 0.49213698506355286, "learning_rate": 5.563852745361523e-06, "loss": 0.4936, "step": 4853 }, { "epoch": 1.55311433447099, "grad_norm": 0.501552164554596, "learning_rate": 5.562002787735658e-06, "loss": 0.5142, "step": 4854 }, { "epoch": 1.553434300341297, "grad_norm": 0.48596420884132385, "learning_rate": 5.560152752187432e-06, "loss": 0.4898, "step": 4855 }, { "epoch": 1.553754266211604, "grad_norm": 0.4873291850090027, "learning_rate": 5.558302638973352e-06, "loss": 0.5229, "step": 4856 }, { "epoch": 1.5540742320819114, "grad_norm": 0.5241739749908447, "learning_rate": 5.55645244834994e-06, "loss": 0.4728, "step": 4857 }, { "epoch": 1.5543941979522184, "grad_norm": 0.5372526049613953, "learning_rate": 5.554602180573727e-06, "loss": 0.5489, "step": 4858 }, { "epoch": 1.5547141638225255, "grad_norm": 0.5021660327911377, "learning_rate": 5.5527518359012535e-06, "loss": 0.4192, "step": 4859 }, { "epoch": 1.5550341296928327, "grad_norm": 0.5167604684829712, "learning_rate": 5.550901414589074e-06, "loss": 0.4931, "step": 4860 }, { "epoch": 1.55535409556314, "grad_norm": 0.5241142511367798, "learning_rate": 5.54905091689375e-06, "loss": 0.5093, "step": 4861 }, { "epoch": 1.555674061433447, "grad_norm": 0.49749425053596497, "learning_rate": 5.547200343071856e-06, "loss": 0.4647, "step": 4862 }, { "epoch": 1.5559940273037542, "grad_norm": 0.5043935179710388, "learning_rate": 5.545349693379976e-06, "loss": 0.4992, "step": 4863 }, { "epoch": 1.5563139931740615, "grad_norm": 0.5102501511573792, "learning_rate": 5.543498968074704e-06, "loss": 0.4872, "step": 4864 }, { "epoch": 1.5566339590443685, "grad_norm": 0.5491971969604492, "learning_rate": 5.541648167412648e-06, "loss": 0.5917, "step": 4865 }, { "epoch": 1.5569539249146758, "grad_norm": 0.4569743573665619, "learning_rate": 5.539797291650419e-06, "loss": 0.4059, "step": 4866 }, { "epoch": 1.557273890784983, "grad_norm": 0.48071160912513733, "learning_rate": 5.53794634104465e-06, "loss": 0.4606, "step": 4867 }, { "epoch": 1.55759385665529, "grad_norm": 0.5465747117996216, "learning_rate": 5.536095315851972e-06, "loss": 0.5211, "step": 4868 }, { "epoch": 1.5579138225255973, "grad_norm": 0.4880690276622772, "learning_rate": 5.534244216329033e-06, "loss": 0.4539, "step": 4869 }, { "epoch": 1.5582337883959045, "grad_norm": 0.5803085565567017, "learning_rate": 5.532393042732493e-06, "loss": 0.5402, "step": 4870 }, { "epoch": 1.5585537542662116, "grad_norm": 0.5311477184295654, "learning_rate": 5.530541795319013e-06, "loss": 0.4667, "step": 4871 }, { "epoch": 1.5588737201365188, "grad_norm": 0.5335061550140381, "learning_rate": 5.528690474345279e-06, "loss": 0.5134, "step": 4872 }, { "epoch": 1.559193686006826, "grad_norm": 0.5214332342147827, "learning_rate": 5.5268390800679715e-06, "loss": 0.5388, "step": 4873 }, { "epoch": 1.559513651877133, "grad_norm": 0.5451050996780396, "learning_rate": 5.524987612743793e-06, "loss": 0.4695, "step": 4874 }, { "epoch": 1.5598336177474401, "grad_norm": 0.5352328419685364, "learning_rate": 5.5231360726294515e-06, "loss": 0.5356, "step": 4875 }, { "epoch": 1.5601535836177476, "grad_norm": 0.4510428309440613, "learning_rate": 5.521284459981662e-06, "loss": 0.4489, "step": 4876 }, { "epoch": 1.5604735494880546, "grad_norm": 0.4923088848590851, "learning_rate": 5.519432775057158e-06, "loss": 0.4708, "step": 4877 }, { "epoch": 1.5607935153583616, "grad_norm": 0.4845147430896759, "learning_rate": 5.517581018112672e-06, "loss": 0.4851, "step": 4878 }, { "epoch": 1.5611134812286689, "grad_norm": 0.5657869577407837, "learning_rate": 5.515729189404956e-06, "loss": 0.5487, "step": 4879 }, { "epoch": 1.5614334470989761, "grad_norm": 0.5293117165565491, "learning_rate": 5.513877289190769e-06, "loss": 0.4913, "step": 4880 }, { "epoch": 1.5617534129692832, "grad_norm": 0.5773677825927734, "learning_rate": 5.512025317726875e-06, "loss": 0.4695, "step": 4881 }, { "epoch": 1.5620733788395904, "grad_norm": 0.5001729726791382, "learning_rate": 5.510173275270057e-06, "loss": 0.5551, "step": 4882 }, { "epoch": 1.5623933447098977, "grad_norm": 0.493765264749527, "learning_rate": 5.508321162077098e-06, "loss": 0.4966, "step": 4883 }, { "epoch": 1.5627133105802047, "grad_norm": 0.46060910820961, "learning_rate": 5.506468978404801e-06, "loss": 0.3832, "step": 4884 }, { "epoch": 1.563033276450512, "grad_norm": 0.5179235339164734, "learning_rate": 5.5046167245099705e-06, "loss": 0.5098, "step": 4885 }, { "epoch": 1.5633532423208192, "grad_norm": 0.5208465456962585, "learning_rate": 5.5027644006494215e-06, "loss": 0.5181, "step": 4886 }, { "epoch": 1.5636732081911262, "grad_norm": 0.5684758424758911, "learning_rate": 5.500912007079987e-06, "loss": 0.4807, "step": 4887 }, { "epoch": 1.5639931740614335, "grad_norm": 0.5089905858039856, "learning_rate": 5.499059544058498e-06, "loss": 0.5215, "step": 4888 }, { "epoch": 1.5643131399317407, "grad_norm": 0.4783798158168793, "learning_rate": 5.497207011841803e-06, "loss": 0.3939, "step": 4889 }, { "epoch": 1.5646331058020477, "grad_norm": 0.5981779098510742, "learning_rate": 5.4953544106867594e-06, "loss": 0.5299, "step": 4890 }, { "epoch": 1.564953071672355, "grad_norm": 0.48870930075645447, "learning_rate": 5.493501740850228e-06, "loss": 0.4722, "step": 4891 }, { "epoch": 1.5652730375426622, "grad_norm": 0.5528561472892761, "learning_rate": 5.491649002589089e-06, "loss": 0.5449, "step": 4892 }, { "epoch": 1.5655930034129693, "grad_norm": 0.5057819485664368, "learning_rate": 5.489796196160223e-06, "loss": 0.4901, "step": 4893 }, { "epoch": 1.5659129692832765, "grad_norm": 0.5546565055847168, "learning_rate": 5.4879433218205246e-06, "loss": 0.4618, "step": 4894 }, { "epoch": 1.5662329351535837, "grad_norm": 0.6191630363464355, "learning_rate": 5.4860903798269e-06, "loss": 0.55, "step": 4895 }, { "epoch": 1.5665529010238908, "grad_norm": 0.5262441039085388, "learning_rate": 5.4842373704362565e-06, "loss": 0.4451, "step": 4896 }, { "epoch": 1.5668728668941978, "grad_norm": 0.5814364552497864, "learning_rate": 5.482384293905521e-06, "loss": 0.5156, "step": 4897 }, { "epoch": 1.5671928327645053, "grad_norm": 0.552703857421875, "learning_rate": 5.480531150491622e-06, "loss": 0.5127, "step": 4898 }, { "epoch": 1.5675127986348123, "grad_norm": 0.5163992643356323, "learning_rate": 5.478677940451501e-06, "loss": 0.4666, "step": 4899 }, { "epoch": 1.5678327645051193, "grad_norm": 0.5370565056800842, "learning_rate": 5.476824664042109e-06, "loss": 0.4769, "step": 4900 }, { "epoch": 1.5681527303754266, "grad_norm": 0.5487856864929199, "learning_rate": 5.4749713215204015e-06, "loss": 0.5001, "step": 4901 }, { "epoch": 1.5684726962457338, "grad_norm": 0.547643780708313, "learning_rate": 5.473117913143351e-06, "loss": 0.5623, "step": 4902 }, { "epoch": 1.5687926621160408, "grad_norm": 0.48040950298309326, "learning_rate": 5.471264439167932e-06, "loss": 0.48, "step": 4903 }, { "epoch": 1.569112627986348, "grad_norm": 0.49863874912261963, "learning_rate": 5.4694108998511296e-06, "loss": 0.5297, "step": 4904 }, { "epoch": 1.5694325938566553, "grad_norm": 0.4937259554862976, "learning_rate": 5.467557295449944e-06, "loss": 0.4536, "step": 4905 }, { "epoch": 1.5697525597269624, "grad_norm": 0.5208419561386108, "learning_rate": 5.465703626221375e-06, "loss": 0.491, "step": 4906 }, { "epoch": 1.5700725255972696, "grad_norm": 0.5429254770278931, "learning_rate": 5.463849892422439e-06, "loss": 0.5094, "step": 4907 }, { "epoch": 1.5703924914675769, "grad_norm": 0.4844743311405182, "learning_rate": 5.461996094310157e-06, "loss": 0.4475, "step": 4908 }, { "epoch": 1.570712457337884, "grad_norm": 0.4967704117298126, "learning_rate": 5.460142232141561e-06, "loss": 0.4937, "step": 4909 }, { "epoch": 1.5710324232081911, "grad_norm": 0.5399122834205627, "learning_rate": 5.4582883061736905e-06, "loss": 0.4866, "step": 4910 }, { "epoch": 1.5713523890784984, "grad_norm": 0.5367668867111206, "learning_rate": 5.456434316663595e-06, "loss": 0.5012, "step": 4911 }, { "epoch": 1.5716723549488054, "grad_norm": 0.4544510841369629, "learning_rate": 5.454580263868333e-06, "loss": 0.4756, "step": 4912 }, { "epoch": 1.5719923208191127, "grad_norm": 0.49972185492515564, "learning_rate": 5.45272614804497e-06, "loss": 0.5177, "step": 4913 }, { "epoch": 1.57231228668942, "grad_norm": 0.4960210919380188, "learning_rate": 5.450871969450582e-06, "loss": 0.4862, "step": 4914 }, { "epoch": 1.572632252559727, "grad_norm": 0.4943525493144989, "learning_rate": 5.449017728342253e-06, "loss": 0.461, "step": 4915 }, { "epoch": 1.572952218430034, "grad_norm": 0.4947192072868347, "learning_rate": 5.447163424977076e-06, "loss": 0.4634, "step": 4916 }, { "epoch": 1.5732721843003414, "grad_norm": 0.5367088317871094, "learning_rate": 5.4453090596121525e-06, "loss": 0.5331, "step": 4917 }, { "epoch": 1.5735921501706485, "grad_norm": 0.5815829038619995, "learning_rate": 5.443454632504592e-06, "loss": 0.4941, "step": 4918 }, { "epoch": 1.5739121160409555, "grad_norm": 0.624888002872467, "learning_rate": 5.441600143911513e-06, "loss": 0.5256, "step": 4919 }, { "epoch": 1.574232081911263, "grad_norm": 0.49638548493385315, "learning_rate": 5.439745594090042e-06, "loss": 0.4947, "step": 4920 }, { "epoch": 1.57455204778157, "grad_norm": 0.48923373222351074, "learning_rate": 5.437890983297316e-06, "loss": 0.4701, "step": 4921 }, { "epoch": 1.574872013651877, "grad_norm": 0.5430967807769775, "learning_rate": 5.436036311790479e-06, "loss": 0.486, "step": 4922 }, { "epoch": 1.5751919795221843, "grad_norm": 0.5682241916656494, "learning_rate": 5.434181579826683e-06, "loss": 0.5104, "step": 4923 }, { "epoch": 1.5755119453924915, "grad_norm": 0.5226086378097534, "learning_rate": 5.432326787663088e-06, "loss": 0.492, "step": 4924 }, { "epoch": 1.5758319112627985, "grad_norm": 0.4683378040790558, "learning_rate": 5.4304719355568635e-06, "loss": 0.4571, "step": 4925 }, { "epoch": 1.5761518771331058, "grad_norm": 0.6027663946151733, "learning_rate": 5.428617023765187e-06, "loss": 0.5227, "step": 4926 }, { "epoch": 1.576471843003413, "grad_norm": 0.5830893516540527, "learning_rate": 5.426762052545245e-06, "loss": 0.4977, "step": 4927 }, { "epoch": 1.57679180887372, "grad_norm": 0.521967887878418, "learning_rate": 5.424907022154231e-06, "loss": 0.506, "step": 4928 }, { "epoch": 1.5771117747440273, "grad_norm": 0.5393236875534058, "learning_rate": 5.423051932849348e-06, "loss": 0.4989, "step": 4929 }, { "epoch": 1.5774317406143346, "grad_norm": 0.4632890820503235, "learning_rate": 5.421196784887804e-06, "loss": 0.4158, "step": 4930 }, { "epoch": 1.5777517064846416, "grad_norm": 0.6038793325424194, "learning_rate": 5.41934157852682e-06, "loss": 0.535, "step": 4931 }, { "epoch": 1.5780716723549488, "grad_norm": 0.5062716007232666, "learning_rate": 5.4174863140236215e-06, "loss": 0.4804, "step": 4932 }, { "epoch": 1.578391638225256, "grad_norm": 0.5247980952262878, "learning_rate": 5.415630991635443e-06, "loss": 0.5311, "step": 4933 }, { "epoch": 1.578711604095563, "grad_norm": 0.48763102293014526, "learning_rate": 5.413775611619529e-06, "loss": 0.4794, "step": 4934 }, { "epoch": 1.5790315699658704, "grad_norm": 0.5592321157455444, "learning_rate": 5.411920174233126e-06, "loss": 0.5148, "step": 4935 }, { "epoch": 1.5793515358361776, "grad_norm": 0.5766536593437195, "learning_rate": 5.410064679733498e-06, "loss": 0.5538, "step": 4936 }, { "epoch": 1.5796715017064846, "grad_norm": 0.5071977376937866, "learning_rate": 5.408209128377907e-06, "loss": 0.4534, "step": 4937 }, { "epoch": 1.5799914675767917, "grad_norm": 0.5171620845794678, "learning_rate": 5.40635352042363e-06, "loss": 0.5179, "step": 4938 }, { "epoch": 1.5803114334470991, "grad_norm": 0.5436701774597168, "learning_rate": 5.404497856127947e-06, "loss": 0.5014, "step": 4939 }, { "epoch": 1.5806313993174061, "grad_norm": 0.494022399187088, "learning_rate": 5.402642135748152e-06, "loss": 0.467, "step": 4940 }, { "epoch": 1.5809513651877132, "grad_norm": 0.5031909942626953, "learning_rate": 5.40078635954154e-06, "loss": 0.4756, "step": 4941 }, { "epoch": 1.5812713310580204, "grad_norm": 0.5093727707862854, "learning_rate": 5.398930527765416e-06, "loss": 0.4848, "step": 4942 }, { "epoch": 1.5815912969283277, "grad_norm": 0.5057315826416016, "learning_rate": 5.397074640677094e-06, "loss": 0.4867, "step": 4943 }, { "epoch": 1.5819112627986347, "grad_norm": 0.49545595049858093, "learning_rate": 5.395218698533897e-06, "loss": 0.5062, "step": 4944 }, { "epoch": 1.582231228668942, "grad_norm": 0.5084137916564941, "learning_rate": 5.393362701593153e-06, "loss": 0.5009, "step": 4945 }, { "epoch": 1.5825511945392492, "grad_norm": 0.5302429795265198, "learning_rate": 5.391506650112195e-06, "loss": 0.471, "step": 4946 }, { "epoch": 1.5828711604095562, "grad_norm": 0.5519916415214539, "learning_rate": 5.389650544348369e-06, "loss": 0.5387, "step": 4947 }, { "epoch": 1.5831911262798635, "grad_norm": 0.5033243298530579, "learning_rate": 5.387794384559027e-06, "loss": 0.4859, "step": 4948 }, { "epoch": 1.5835110921501707, "grad_norm": 0.462449848651886, "learning_rate": 5.3859381710015256e-06, "loss": 0.435, "step": 4949 }, { "epoch": 1.5838310580204777, "grad_norm": 0.5657418966293335, "learning_rate": 5.384081903933235e-06, "loss": 0.5097, "step": 4950 }, { "epoch": 1.584151023890785, "grad_norm": 0.5815718173980713, "learning_rate": 5.382225583611525e-06, "loss": 0.5071, "step": 4951 }, { "epoch": 1.5844709897610922, "grad_norm": 0.5068009495735168, "learning_rate": 5.380369210293777e-06, "loss": 0.5294, "step": 4952 }, { "epoch": 1.5847909556313993, "grad_norm": 0.4753865897655487, "learning_rate": 5.378512784237382e-06, "loss": 0.4522, "step": 4953 }, { "epoch": 1.5851109215017065, "grad_norm": 0.5203690528869629, "learning_rate": 5.376656305699733e-06, "loss": 0.4568, "step": 4954 }, { "epoch": 1.5854308873720138, "grad_norm": 0.5779047012329102, "learning_rate": 5.374799774938236e-06, "loss": 0.5153, "step": 4955 }, { "epoch": 1.5857508532423208, "grad_norm": 0.5613170862197876, "learning_rate": 5.372943192210297e-06, "loss": 0.5389, "step": 4956 }, { "epoch": 1.586070819112628, "grad_norm": 0.5037094950675964, "learning_rate": 5.371086557773337e-06, "loss": 0.4675, "step": 4957 }, { "epoch": 1.5863907849829353, "grad_norm": 0.5480237007141113, "learning_rate": 5.369229871884779e-06, "loss": 0.5099, "step": 4958 }, { "epoch": 1.5867107508532423, "grad_norm": 0.4700008034706116, "learning_rate": 5.367373134802055e-06, "loss": 0.4598, "step": 4959 }, { "epoch": 1.5870307167235493, "grad_norm": 0.5223121047019958, "learning_rate": 5.365516346782605e-06, "loss": 0.5131, "step": 4960 }, { "epoch": 1.5873506825938568, "grad_norm": 0.48291903734207153, "learning_rate": 5.363659508083872e-06, "loss": 0.4667, "step": 4961 }, { "epoch": 1.5876706484641638, "grad_norm": 0.5335783362388611, "learning_rate": 5.361802618963312e-06, "loss": 0.4978, "step": 4962 }, { "epoch": 1.5879906143344709, "grad_norm": 0.5844630599021912, "learning_rate": 5.359945679678383e-06, "loss": 0.5726, "step": 4963 }, { "epoch": 1.588310580204778, "grad_norm": 0.5081358551979065, "learning_rate": 5.3580886904865525e-06, "loss": 0.4611, "step": 4964 }, { "epoch": 1.5886305460750854, "grad_norm": 0.5709756016731262, "learning_rate": 5.356231651645295e-06, "loss": 0.4868, "step": 4965 }, { "epoch": 1.5889505119453924, "grad_norm": 0.5915698409080505, "learning_rate": 5.354374563412089e-06, "loss": 0.5062, "step": 4966 }, { "epoch": 1.5892704778156996, "grad_norm": 0.576073169708252, "learning_rate": 5.3525174260444255e-06, "loss": 0.4828, "step": 4967 }, { "epoch": 1.5895904436860069, "grad_norm": 0.5176089406013489, "learning_rate": 5.350660239799795e-06, "loss": 0.484, "step": 4968 }, { "epoch": 1.589910409556314, "grad_norm": 0.5273805260658264, "learning_rate": 5.348803004935701e-06, "loss": 0.4837, "step": 4969 }, { "epoch": 1.5902303754266212, "grad_norm": 0.5675475597381592, "learning_rate": 5.346945721709652e-06, "loss": 0.5104, "step": 4970 }, { "epoch": 1.5905503412969284, "grad_norm": 0.5363828539848328, "learning_rate": 5.3450883903791585e-06, "loss": 0.4595, "step": 4971 }, { "epoch": 1.5908703071672354, "grad_norm": 0.525600790977478, "learning_rate": 5.343231011201747e-06, "loss": 0.4865, "step": 4972 }, { "epoch": 1.5911902730375427, "grad_norm": 0.5501222014427185, "learning_rate": 5.341373584434941e-06, "loss": 0.4936, "step": 4973 }, { "epoch": 1.59151023890785, "grad_norm": 0.4852727949619293, "learning_rate": 5.339516110336278e-06, "loss": 0.427, "step": 4974 }, { "epoch": 1.591830204778157, "grad_norm": 0.5601401925086975, "learning_rate": 5.3376585891633e-06, "loss": 0.5627, "step": 4975 }, { "epoch": 1.5921501706484642, "grad_norm": 0.4850602447986603, "learning_rate": 5.335801021173548e-06, "loss": 0.4997, "step": 4976 }, { "epoch": 1.5924701365187715, "grad_norm": 0.5282145738601685, "learning_rate": 5.3339434066245835e-06, "loss": 0.482, "step": 4977 }, { "epoch": 1.5927901023890785, "grad_norm": 0.5373655557632446, "learning_rate": 5.3320857457739615e-06, "loss": 0.5199, "step": 4978 }, { "epoch": 1.5931100682593855, "grad_norm": 0.5853095054626465, "learning_rate": 5.330228038879253e-06, "loss": 0.5109, "step": 4979 }, { "epoch": 1.593430034129693, "grad_norm": 0.5338366031646729, "learning_rate": 5.32837028619803e-06, "loss": 0.5002, "step": 4980 }, { "epoch": 1.59375, "grad_norm": 0.5058368444442749, "learning_rate": 5.326512487987871e-06, "loss": 0.5277, "step": 4981 }, { "epoch": 1.594069965870307, "grad_norm": 0.5702630281448364, "learning_rate": 5.324654644506363e-06, "loss": 0.5024, "step": 4982 }, { "epoch": 1.5943899317406145, "grad_norm": 0.46419402956962585, "learning_rate": 5.3227967560110984e-06, "loss": 0.4306, "step": 4983 }, { "epoch": 1.5947098976109215, "grad_norm": 0.45496001839637756, "learning_rate": 5.320938822759674e-06, "loss": 0.5151, "step": 4984 }, { "epoch": 1.5950298634812285, "grad_norm": 0.5695767998695374, "learning_rate": 5.319080845009697e-06, "loss": 0.5443, "step": 4985 }, { "epoch": 1.5953498293515358, "grad_norm": 0.5018442273139954, "learning_rate": 5.317222823018775e-06, "loss": 0.4859, "step": 4986 }, { "epoch": 1.595669795221843, "grad_norm": 0.5421056151390076, "learning_rate": 5.31536475704453e-06, "loss": 0.4921, "step": 4987 }, { "epoch": 1.59598976109215, "grad_norm": 0.4972304105758667, "learning_rate": 5.31350664734458e-06, "loss": 0.5217, "step": 4988 }, { "epoch": 1.5963097269624573, "grad_norm": 0.5531447529792786, "learning_rate": 5.3116484941765564e-06, "loss": 0.4992, "step": 4989 }, { "epoch": 1.5966296928327646, "grad_norm": 0.557141125202179, "learning_rate": 5.309790297798094e-06, "loss": 0.4901, "step": 4990 }, { "epoch": 1.5969496587030716, "grad_norm": 0.600490391254425, "learning_rate": 5.307932058466835e-06, "loss": 0.5142, "step": 4991 }, { "epoch": 1.5972696245733788, "grad_norm": 0.5329923629760742, "learning_rate": 5.3060737764404254e-06, "loss": 0.5022, "step": 4992 }, { "epoch": 1.597589590443686, "grad_norm": 0.4775228500366211, "learning_rate": 5.304215451976517e-06, "loss": 0.4576, "step": 4993 }, { "epoch": 1.5979095563139931, "grad_norm": 0.5137864351272583, "learning_rate": 5.3023570853327725e-06, "loss": 0.5079, "step": 4994 }, { "epoch": 1.5982295221843004, "grad_norm": 0.48908084630966187, "learning_rate": 5.300498676766853e-06, "loss": 0.4567, "step": 4995 }, { "epoch": 1.5985494880546076, "grad_norm": 0.5525791645050049, "learning_rate": 5.29864022653643e-06, "loss": 0.5698, "step": 4996 }, { "epoch": 1.5988694539249146, "grad_norm": 0.5054503679275513, "learning_rate": 5.296781734899182e-06, "loss": 0.4754, "step": 4997 }, { "epoch": 1.599189419795222, "grad_norm": 0.5379093885421753, "learning_rate": 5.294923202112788e-06, "loss": 0.506, "step": 4998 }, { "epoch": 1.5995093856655291, "grad_norm": 0.4838557839393616, "learning_rate": 5.293064628434939e-06, "loss": 0.4752, "step": 4999 }, { "epoch": 1.5998293515358362, "grad_norm": 0.4621947407722473, "learning_rate": 5.291206014123325e-06, "loss": 0.4347, "step": 5000 }, { "epoch": 1.6001493174061432, "grad_norm": 0.5238200426101685, "learning_rate": 5.289347359435646e-06, "loss": 0.4992, "step": 5001 }, { "epoch": 1.6004692832764507, "grad_norm": 0.5577722787857056, "learning_rate": 5.28748866462961e-06, "loss": 0.5308, "step": 5002 }, { "epoch": 1.6007892491467577, "grad_norm": 0.5425005555152893, "learning_rate": 5.2856299299629235e-06, "loss": 0.5073, "step": 5003 }, { "epoch": 1.6011092150170647, "grad_norm": 0.49496665596961975, "learning_rate": 5.283771155693304e-06, "loss": 0.4808, "step": 5004 }, { "epoch": 1.601429180887372, "grad_norm": 0.5529054403305054, "learning_rate": 5.281912342078471e-06, "loss": 0.4747, "step": 5005 }, { "epoch": 1.6017491467576792, "grad_norm": 0.5010500550270081, "learning_rate": 5.280053489376154e-06, "loss": 0.425, "step": 5006 }, { "epoch": 1.6020691126279862, "grad_norm": 0.5153359770774841, "learning_rate": 5.278194597844083e-06, "loss": 0.5108, "step": 5007 }, { "epoch": 1.6023890784982935, "grad_norm": 0.5091795325279236, "learning_rate": 5.276335667739998e-06, "loss": 0.514, "step": 5008 }, { "epoch": 1.6027090443686007, "grad_norm": 0.5603525638580322, "learning_rate": 5.274476699321638e-06, "loss": 0.5229, "step": 5009 }, { "epoch": 1.6030290102389078, "grad_norm": 0.5608372092247009, "learning_rate": 5.272617692846755e-06, "loss": 0.5504, "step": 5010 }, { "epoch": 1.603348976109215, "grad_norm": 0.556414008140564, "learning_rate": 5.2707586485731e-06, "loss": 0.4701, "step": 5011 }, { "epoch": 1.6036689419795223, "grad_norm": 0.4858776330947876, "learning_rate": 5.268899566758432e-06, "loss": 0.4416, "step": 5012 }, { "epoch": 1.6039889078498293, "grad_norm": 0.5496314764022827, "learning_rate": 5.267040447660517e-06, "loss": 0.5484, "step": 5013 }, { "epoch": 1.6043088737201365, "grad_norm": 0.5210900902748108, "learning_rate": 5.265181291537122e-06, "loss": 0.5136, "step": 5014 }, { "epoch": 1.6046288395904438, "grad_norm": 0.5538973212242126, "learning_rate": 5.263322098646023e-06, "loss": 0.4832, "step": 5015 }, { "epoch": 1.6049488054607508, "grad_norm": 0.5693376064300537, "learning_rate": 5.261462869244997e-06, "loss": 0.5117, "step": 5016 }, { "epoch": 1.605268771331058, "grad_norm": 0.5388249754905701, "learning_rate": 5.25960360359183e-06, "loss": 0.5164, "step": 5017 }, { "epoch": 1.6055887372013653, "grad_norm": 0.5378933548927307, "learning_rate": 5.257744301944311e-06, "loss": 0.4877, "step": 5018 }, { "epoch": 1.6059087030716723, "grad_norm": 0.533751904964447, "learning_rate": 5.255884964560235e-06, "loss": 0.4989, "step": 5019 }, { "epoch": 1.6062286689419796, "grad_norm": 0.4835195243358612, "learning_rate": 5.2540255916974005e-06, "loss": 0.4758, "step": 5020 }, { "epoch": 1.6065486348122868, "grad_norm": 0.5001780390739441, "learning_rate": 5.252166183613612e-06, "loss": 0.511, "step": 5021 }, { "epoch": 1.6068686006825939, "grad_norm": 0.5064964890480042, "learning_rate": 5.2503067405666776e-06, "loss": 0.5093, "step": 5022 }, { "epoch": 1.6071885665529009, "grad_norm": 0.5077722668647766, "learning_rate": 5.248447262814413e-06, "loss": 0.4866, "step": 5023 }, { "epoch": 1.6075085324232083, "grad_norm": 0.5091267824172974, "learning_rate": 5.246587750614638e-06, "loss": 0.4883, "step": 5024 }, { "epoch": 1.6078284982935154, "grad_norm": 0.4769721031188965, "learning_rate": 5.244728204225175e-06, "loss": 0.4887, "step": 5025 }, { "epoch": 1.6081484641638224, "grad_norm": 0.5393322706222534, "learning_rate": 5.242868623903851e-06, "loss": 0.4896, "step": 5026 }, { "epoch": 1.6084684300341296, "grad_norm": 0.539297878742218, "learning_rate": 5.241009009908499e-06, "loss": 0.5276, "step": 5027 }, { "epoch": 1.608788395904437, "grad_norm": 0.4828234910964966, "learning_rate": 5.239149362496959e-06, "loss": 0.5081, "step": 5028 }, { "epoch": 1.609108361774744, "grad_norm": 0.47531136870384216, "learning_rate": 5.237289681927072e-06, "loss": 0.4619, "step": 5029 }, { "epoch": 1.6094283276450512, "grad_norm": 0.6127607226371765, "learning_rate": 5.235429968456686e-06, "loss": 0.5101, "step": 5030 }, { "epoch": 1.6097482935153584, "grad_norm": 0.5446188449859619, "learning_rate": 5.2335702223436505e-06, "loss": 0.5345, "step": 5031 }, { "epoch": 1.6100682593856654, "grad_norm": 0.5053405165672302, "learning_rate": 5.2317104438458235e-06, "loss": 0.4653, "step": 5032 }, { "epoch": 1.6103882252559727, "grad_norm": 0.5763628482818604, "learning_rate": 5.229850633221063e-06, "loss": 0.495, "step": 5033 }, { "epoch": 1.61070819112628, "grad_norm": 0.5162618160247803, "learning_rate": 5.2279907907272375e-06, "loss": 0.5125, "step": 5034 }, { "epoch": 1.611028156996587, "grad_norm": 0.43388307094573975, "learning_rate": 5.226130916622214e-06, "loss": 0.4038, "step": 5035 }, { "epoch": 1.6113481228668942, "grad_norm": 0.5374667048454285, "learning_rate": 5.224271011163867e-06, "loss": 0.5445, "step": 5036 }, { "epoch": 1.6116680887372015, "grad_norm": 0.5645819902420044, "learning_rate": 5.222411074610074e-06, "loss": 0.5007, "step": 5037 }, { "epoch": 1.6119880546075085, "grad_norm": 0.5632933378219604, "learning_rate": 5.2205511072187185e-06, "loss": 0.5081, "step": 5038 }, { "epoch": 1.6123080204778157, "grad_norm": 0.5653529167175293, "learning_rate": 5.218691109247686e-06, "loss": 0.5217, "step": 5039 }, { "epoch": 1.612627986348123, "grad_norm": 0.5255849361419678, "learning_rate": 5.216831080954869e-06, "loss": 0.4302, "step": 5040 }, { "epoch": 1.61294795221843, "grad_norm": 0.6479395627975464, "learning_rate": 5.214971022598162e-06, "loss": 0.571, "step": 5041 }, { "epoch": 1.613267918088737, "grad_norm": 0.5265260934829712, "learning_rate": 5.213110934435463e-06, "loss": 0.5054, "step": 5042 }, { "epoch": 1.6135878839590445, "grad_norm": 0.5972551107406616, "learning_rate": 5.211250816724676e-06, "loss": 0.4689, "step": 5043 }, { "epoch": 1.6139078498293515, "grad_norm": 0.4919164776802063, "learning_rate": 5.2093906697237104e-06, "loss": 0.4354, "step": 5044 }, { "epoch": 1.6142278156996586, "grad_norm": 0.5815161466598511, "learning_rate": 5.207530493690478e-06, "loss": 0.5371, "step": 5045 }, { "epoch": 1.614547781569966, "grad_norm": 0.5273357033729553, "learning_rate": 5.205670288882889e-06, "loss": 0.4503, "step": 5046 }, { "epoch": 1.614867747440273, "grad_norm": 0.5358985662460327, "learning_rate": 5.203810055558871e-06, "loss": 0.5407, "step": 5047 }, { "epoch": 1.61518771331058, "grad_norm": 0.5318018198013306, "learning_rate": 5.201949793976342e-06, "loss": 0.4957, "step": 5048 }, { "epoch": 1.6155076791808873, "grad_norm": 0.499167799949646, "learning_rate": 5.2000895043932296e-06, "loss": 0.4988, "step": 5049 }, { "epoch": 1.6158276450511946, "grad_norm": 0.5340614914894104, "learning_rate": 5.198229187067469e-06, "loss": 0.5203, "step": 5050 }, { "epoch": 1.6161476109215016, "grad_norm": 0.5454068183898926, "learning_rate": 5.196368842256992e-06, "loss": 0.5007, "step": 5051 }, { "epoch": 1.6164675767918089, "grad_norm": 0.49962106347084045, "learning_rate": 5.194508470219739e-06, "loss": 0.4042, "step": 5052 }, { "epoch": 1.616787542662116, "grad_norm": 0.540557861328125, "learning_rate": 5.192648071213653e-06, "loss": 0.5337, "step": 5053 }, { "epoch": 1.6171075085324231, "grad_norm": 0.49432751536369324, "learning_rate": 5.19078764549668e-06, "loss": 0.4718, "step": 5054 }, { "epoch": 1.6174274744027304, "grad_norm": 0.5162655115127563, "learning_rate": 5.18892719332677e-06, "loss": 0.4571, "step": 5055 }, { "epoch": 1.6177474402730376, "grad_norm": 0.5410633683204651, "learning_rate": 5.187066714961877e-06, "loss": 0.5203, "step": 5056 }, { "epoch": 1.6180674061433447, "grad_norm": 0.4666568636894226, "learning_rate": 5.185206210659961e-06, "loss": 0.4568, "step": 5057 }, { "epoch": 1.618387372013652, "grad_norm": 0.5138685703277588, "learning_rate": 5.183345680678979e-06, "loss": 0.4949, "step": 5058 }, { "epoch": 1.6187073378839592, "grad_norm": 0.5632014274597168, "learning_rate": 5.181485125276898e-06, "loss": 0.4725, "step": 5059 }, { "epoch": 1.6190273037542662, "grad_norm": 0.5428216457366943, "learning_rate": 5.179624544711688e-06, "loss": 0.5151, "step": 5060 }, { "epoch": 1.6193472696245734, "grad_norm": 0.47719550132751465, "learning_rate": 5.177763939241315e-06, "loss": 0.4976, "step": 5061 }, { "epoch": 1.6196672354948807, "grad_norm": 0.5554200410842896, "learning_rate": 5.175903309123761e-06, "loss": 0.4945, "step": 5062 }, { "epoch": 1.6199872013651877, "grad_norm": 0.49648791551589966, "learning_rate": 5.174042654617001e-06, "loss": 0.5075, "step": 5063 }, { "epoch": 1.6203071672354947, "grad_norm": 0.49515655636787415, "learning_rate": 5.172181975979017e-06, "loss": 0.4577, "step": 5064 }, { "epoch": 1.6206271331058022, "grad_norm": 0.4947766661643982, "learning_rate": 5.170321273467797e-06, "loss": 0.4943, "step": 5065 }, { "epoch": 1.6209470989761092, "grad_norm": 0.49154120683670044, "learning_rate": 5.1684605473413245e-06, "loss": 0.4443, "step": 5066 }, { "epoch": 1.6212670648464163, "grad_norm": 0.5223211646080017, "learning_rate": 5.166599797857595e-06, "loss": 0.5108, "step": 5067 }, { "epoch": 1.6215870307167235, "grad_norm": 0.5061191320419312, "learning_rate": 5.164739025274604e-06, "loss": 0.4437, "step": 5068 }, { "epoch": 1.6219069965870307, "grad_norm": 0.4762420356273651, "learning_rate": 5.162878229850349e-06, "loss": 0.4674, "step": 5069 }, { "epoch": 1.6222269624573378, "grad_norm": 0.5315890312194824, "learning_rate": 5.1610174118428325e-06, "loss": 0.5548, "step": 5070 }, { "epoch": 1.622546928327645, "grad_norm": 0.510001540184021, "learning_rate": 5.159156571510056e-06, "loss": 0.4548, "step": 5071 }, { "epoch": 1.6228668941979523, "grad_norm": 0.48377159237861633, "learning_rate": 5.157295709110031e-06, "loss": 0.5017, "step": 5072 }, { "epoch": 1.6231868600682593, "grad_norm": 0.5535182952880859, "learning_rate": 5.155434824900764e-06, "loss": 0.5205, "step": 5073 }, { "epoch": 1.6235068259385665, "grad_norm": 0.48844847083091736, "learning_rate": 5.153573919140275e-06, "loss": 0.4537, "step": 5074 }, { "epoch": 1.6238267918088738, "grad_norm": 0.5575363039970398, "learning_rate": 5.151712992086576e-06, "loss": 0.5385, "step": 5075 }, { "epoch": 1.6241467576791808, "grad_norm": 0.4875611960887909, "learning_rate": 5.149852043997687e-06, "loss": 0.4025, "step": 5076 }, { "epoch": 1.624466723549488, "grad_norm": 0.5373698472976685, "learning_rate": 5.1479910751316344e-06, "loss": 0.5013, "step": 5077 }, { "epoch": 1.6247866894197953, "grad_norm": 0.5397643446922302, "learning_rate": 5.146130085746439e-06, "loss": 0.4812, "step": 5078 }, { "epoch": 1.6251066552901023, "grad_norm": 0.5343145132064819, "learning_rate": 5.144269076100131e-06, "loss": 0.5401, "step": 5079 }, { "epoch": 1.6254266211604096, "grad_norm": 0.5181583166122437, "learning_rate": 5.142408046450743e-06, "loss": 0.4916, "step": 5080 }, { "epoch": 1.6257465870307168, "grad_norm": 0.5414060950279236, "learning_rate": 5.1405469970563085e-06, "loss": 0.4682, "step": 5081 }, { "epoch": 1.6260665529010239, "grad_norm": 0.5570537447929382, "learning_rate": 5.138685928174863e-06, "loss": 0.493, "step": 5082 }, { "epoch": 1.6263865187713311, "grad_norm": 0.5074068903923035, "learning_rate": 5.136824840064447e-06, "loss": 0.5192, "step": 5083 }, { "epoch": 1.6267064846416384, "grad_norm": 0.5079135298728943, "learning_rate": 5.134963732983104e-06, "loss": 0.4646, "step": 5084 }, { "epoch": 1.6270264505119454, "grad_norm": 0.5378671884536743, "learning_rate": 5.133102607188875e-06, "loss": 0.5366, "step": 5085 }, { "epoch": 1.6273464163822524, "grad_norm": 0.45548543334007263, "learning_rate": 5.131241462939809e-06, "loss": 0.4907, "step": 5086 }, { "epoch": 1.6276663822525599, "grad_norm": 0.5167208909988403, "learning_rate": 5.129380300493958e-06, "loss": 0.464, "step": 5087 }, { "epoch": 1.627986348122867, "grad_norm": 0.5381355881690979, "learning_rate": 5.127519120109371e-06, "loss": 0.486, "step": 5088 }, { "epoch": 1.628306313993174, "grad_norm": 0.4981718063354492, "learning_rate": 5.125657922044106e-06, "loss": 0.4635, "step": 5089 }, { "epoch": 1.6286262798634812, "grad_norm": 0.5355318784713745, "learning_rate": 5.123796706556217e-06, "loss": 0.4921, "step": 5090 }, { "epoch": 1.6289462457337884, "grad_norm": 0.5649579167366028, "learning_rate": 5.121935473903768e-06, "loss": 0.5134, "step": 5091 }, { "epoch": 1.6292662116040955, "grad_norm": 0.5361959338188171, "learning_rate": 5.120074224344819e-06, "loss": 0.4934, "step": 5092 }, { "epoch": 1.6295861774744027, "grad_norm": 0.510814905166626, "learning_rate": 5.118212958137432e-06, "loss": 0.4841, "step": 5093 }, { "epoch": 1.62990614334471, "grad_norm": 0.5256816148757935, "learning_rate": 5.116351675539678e-06, "loss": 0.4785, "step": 5094 }, { "epoch": 1.630226109215017, "grad_norm": 0.5437941551208496, "learning_rate": 5.1144903768096235e-06, "loss": 0.4538, "step": 5095 }, { "epoch": 1.6305460750853242, "grad_norm": 0.5067790150642395, "learning_rate": 5.112629062205341e-06, "loss": 0.5508, "step": 5096 }, { "epoch": 1.6308660409556315, "grad_norm": 0.5657281875610352, "learning_rate": 5.110767731984902e-06, "loss": 0.4965, "step": 5097 }, { "epoch": 1.6311860068259385, "grad_norm": 0.48917561769485474, "learning_rate": 5.108906386406385e-06, "loss": 0.444, "step": 5098 }, { "epoch": 1.6315059726962458, "grad_norm": 0.5413578748703003, "learning_rate": 5.107045025727867e-06, "loss": 0.529, "step": 5099 }, { "epoch": 1.631825938566553, "grad_norm": 0.5355716943740845, "learning_rate": 5.105183650207427e-06, "loss": 0.4952, "step": 5100 }, { "epoch": 1.63214590443686, "grad_norm": 0.5102506279945374, "learning_rate": 5.103322260103147e-06, "loss": 0.4945, "step": 5101 }, { "epoch": 1.6324658703071673, "grad_norm": 0.49773868918418884, "learning_rate": 5.101460855673112e-06, "loss": 0.4505, "step": 5102 }, { "epoch": 1.6327858361774745, "grad_norm": 0.5560070872306824, "learning_rate": 5.099599437175405e-06, "loss": 0.5011, "step": 5103 }, { "epoch": 1.6331058020477816, "grad_norm": 0.556323230266571, "learning_rate": 5.097738004868119e-06, "loss": 0.4982, "step": 5104 }, { "epoch": 1.6334257679180886, "grad_norm": 0.5519665479660034, "learning_rate": 5.095876559009341e-06, "loss": 0.5599, "step": 5105 }, { "epoch": 1.633745733788396, "grad_norm": 0.501521110534668, "learning_rate": 5.0940150998571614e-06, "loss": 0.5081, "step": 5106 }, { "epoch": 1.634065699658703, "grad_norm": 0.537854015827179, "learning_rate": 5.092153627669675e-06, "loss": 0.5022, "step": 5107 }, { "epoch": 1.63438566552901, "grad_norm": 0.5112746953964233, "learning_rate": 5.090292142704978e-06, "loss": 0.53, "step": 5108 }, { "epoch": 1.6347056313993176, "grad_norm": 0.429195761680603, "learning_rate": 5.0884306452211675e-06, "loss": 0.4135, "step": 5109 }, { "epoch": 1.6350255972696246, "grad_norm": 0.5935645699501038, "learning_rate": 5.08656913547634e-06, "loss": 0.5464, "step": 5110 }, { "epoch": 1.6353455631399316, "grad_norm": 0.5522560477256775, "learning_rate": 5.084707613728598e-06, "loss": 0.4994, "step": 5111 }, { "epoch": 1.6356655290102389, "grad_norm": 0.49781811237335205, "learning_rate": 5.082846080236044e-06, "loss": 0.4707, "step": 5112 }, { "epoch": 1.6359854948805461, "grad_norm": 0.5011031627655029, "learning_rate": 5.080984535256781e-06, "loss": 0.5452, "step": 5113 }, { "epoch": 1.6363054607508531, "grad_norm": 0.47108280658721924, "learning_rate": 5.0791229790489134e-06, "loss": 0.4407, "step": 5114 }, { "epoch": 1.6366254266211604, "grad_norm": 0.5003146529197693, "learning_rate": 5.077261411870553e-06, "loss": 0.5277, "step": 5115 }, { "epoch": 1.6369453924914676, "grad_norm": 0.4633661210536957, "learning_rate": 5.075399833979803e-06, "loss": 0.4814, "step": 5116 }, { "epoch": 1.6372653583617747, "grad_norm": 0.47853884100914, "learning_rate": 5.0735382456347745e-06, "loss": 0.462, "step": 5117 }, { "epoch": 1.637585324232082, "grad_norm": 0.5123675465583801, "learning_rate": 5.071676647093581e-06, "loss": 0.4885, "step": 5118 }, { "epoch": 1.6379052901023892, "grad_norm": 0.547149121761322, "learning_rate": 5.069815038614333e-06, "loss": 0.535, "step": 5119 }, { "epoch": 1.6382252559726962, "grad_norm": 0.5156007409095764, "learning_rate": 5.067953420455148e-06, "loss": 0.4916, "step": 5120 }, { "epoch": 1.6385452218430034, "grad_norm": 0.5121249556541443, "learning_rate": 5.066091792874138e-06, "loss": 0.4676, "step": 5121 }, { "epoch": 1.6388651877133107, "grad_norm": 0.5301448702812195, "learning_rate": 5.064230156129422e-06, "loss": 0.4829, "step": 5122 }, { "epoch": 1.6391851535836177, "grad_norm": 0.5886031985282898, "learning_rate": 5.062368510479116e-06, "loss": 0.4933, "step": 5123 }, { "epoch": 1.639505119453925, "grad_norm": 0.4933142066001892, "learning_rate": 5.060506856181342e-06, "loss": 0.5095, "step": 5124 }, { "epoch": 1.6398250853242322, "grad_norm": 0.4435693919658661, "learning_rate": 5.05864519349422e-06, "loss": 0.4523, "step": 5125 }, { "epoch": 1.6401450511945392, "grad_norm": 0.5381711721420288, "learning_rate": 5.056783522675871e-06, "loss": 0.5406, "step": 5126 }, { "epoch": 1.6404650170648463, "grad_norm": 0.4852531850337982, "learning_rate": 5.0549218439844185e-06, "loss": 0.4794, "step": 5127 }, { "epoch": 1.6407849829351537, "grad_norm": 0.4958217740058899, "learning_rate": 5.053060157677986e-06, "loss": 0.4866, "step": 5128 }, { "epoch": 1.6411049488054608, "grad_norm": 0.543684184551239, "learning_rate": 5.051198464014698e-06, "loss": 0.5488, "step": 5129 }, { "epoch": 1.6414249146757678, "grad_norm": 0.5541886687278748, "learning_rate": 5.049336763252683e-06, "loss": 0.4791, "step": 5130 }, { "epoch": 1.641744880546075, "grad_norm": 0.4899916648864746, "learning_rate": 5.047475055650065e-06, "loss": 0.4622, "step": 5131 }, { "epoch": 1.6420648464163823, "grad_norm": 0.4611685276031494, "learning_rate": 5.045613341464976e-06, "loss": 0.4939, "step": 5132 }, { "epoch": 1.6423848122866893, "grad_norm": 0.5908498167991638, "learning_rate": 5.043751620955539e-06, "loss": 0.558, "step": 5133 }, { "epoch": 1.6427047781569966, "grad_norm": 0.4951009154319763, "learning_rate": 5.041889894379889e-06, "loss": 0.4257, "step": 5134 }, { "epoch": 1.6430247440273038, "grad_norm": 0.5275335907936096, "learning_rate": 5.040028161996156e-06, "loss": 0.5045, "step": 5135 }, { "epoch": 1.6433447098976108, "grad_norm": 0.5004864931106567, "learning_rate": 5.038166424062466e-06, "loss": 0.4902, "step": 5136 }, { "epoch": 1.643664675767918, "grad_norm": 0.4917805790901184, "learning_rate": 5.036304680836959e-06, "loss": 0.4812, "step": 5137 }, { "epoch": 1.6439846416382253, "grad_norm": 0.5122542381286621, "learning_rate": 5.034442932577764e-06, "loss": 0.5423, "step": 5138 }, { "epoch": 1.6443046075085324, "grad_norm": 0.5459406971931458, "learning_rate": 5.032581179543014e-06, "loss": 0.4673, "step": 5139 }, { "epoch": 1.6446245733788396, "grad_norm": 0.5206505656242371, "learning_rate": 5.030719421990845e-06, "loss": 0.5056, "step": 5140 }, { "epoch": 1.6449445392491469, "grad_norm": 0.5480968356132507, "learning_rate": 5.028857660179389e-06, "loss": 0.5183, "step": 5141 }, { "epoch": 1.6452645051194539, "grad_norm": 0.6217779517173767, "learning_rate": 5.0269958943667865e-06, "loss": 0.5225, "step": 5142 }, { "epoch": 1.6455844709897611, "grad_norm": 0.4600312411785126, "learning_rate": 5.02513412481117e-06, "loss": 0.4457, "step": 5143 }, { "epoch": 1.6459044368600684, "grad_norm": 0.5307644605636597, "learning_rate": 5.023272351770675e-06, "loss": 0.4937, "step": 5144 }, { "epoch": 1.6462244027303754, "grad_norm": 0.5613411664962769, "learning_rate": 5.021410575503442e-06, "loss": 0.546, "step": 5145 }, { "epoch": 1.6465443686006824, "grad_norm": 0.5683817863464355, "learning_rate": 5.019548796267607e-06, "loss": 0.4823, "step": 5146 }, { "epoch": 1.64686433447099, "grad_norm": 0.6270329356193542, "learning_rate": 5.017687014321308e-06, "loss": 0.5589, "step": 5147 }, { "epoch": 1.647184300341297, "grad_norm": 0.4889225363731384, "learning_rate": 5.0158252299226826e-06, "loss": 0.4477, "step": 5148 }, { "epoch": 1.647504266211604, "grad_norm": 0.5191217064857483, "learning_rate": 5.0139634433298705e-06, "loss": 0.47, "step": 5149 }, { "epoch": 1.6478242320819114, "grad_norm": 0.6270520091056824, "learning_rate": 5.01210165480101e-06, "loss": 0.5093, "step": 5150 }, { "epoch": 1.6481441979522184, "grad_norm": 0.5916458964347839, "learning_rate": 5.010239864594241e-06, "loss": 0.4977, "step": 5151 }, { "epoch": 1.6484641638225255, "grad_norm": 0.5480887293815613, "learning_rate": 5.008378072967702e-06, "loss": 0.4963, "step": 5152 }, { "epoch": 1.6487841296928327, "grad_norm": 0.4946962296962738, "learning_rate": 5.006516280179532e-06, "loss": 0.4925, "step": 5153 }, { "epoch": 1.64910409556314, "grad_norm": 0.521058201789856, "learning_rate": 5.004654486487873e-06, "loss": 0.4757, "step": 5154 }, { "epoch": 1.649424061433447, "grad_norm": 0.5836396813392639, "learning_rate": 5.002792692150865e-06, "loss": 0.5098, "step": 5155 }, { "epoch": 1.6497440273037542, "grad_norm": 0.5659805536270142, "learning_rate": 5.000930897426645e-06, "loss": 0.5355, "step": 5156 }, { "epoch": 1.6500639931740615, "grad_norm": 0.4996427297592163, "learning_rate": 4.999069102573356e-06, "loss": 0.4866, "step": 5157 }, { "epoch": 1.6503839590443685, "grad_norm": 0.48106181621551514, "learning_rate": 4.997207307849136e-06, "loss": 0.4804, "step": 5158 }, { "epoch": 1.6507039249146758, "grad_norm": 0.4977264404296875, "learning_rate": 4.995345513512128e-06, "loss": 0.4477, "step": 5159 }, { "epoch": 1.651023890784983, "grad_norm": 0.5068228244781494, "learning_rate": 4.99348371982047e-06, "loss": 0.5494, "step": 5160 }, { "epoch": 1.65134385665529, "grad_norm": 0.46656063199043274, "learning_rate": 4.991621927032301e-06, "loss": 0.4418, "step": 5161 }, { "epoch": 1.6516638225255973, "grad_norm": 0.4818781912326813, "learning_rate": 4.98976013540576e-06, "loss": 0.4838, "step": 5162 }, { "epoch": 1.6519837883959045, "grad_norm": 0.5414600372314453, "learning_rate": 4.9878983451989904e-06, "loss": 0.503, "step": 5163 }, { "epoch": 1.6523037542662116, "grad_norm": 0.49550461769104004, "learning_rate": 4.98603655667013e-06, "loss": 0.4731, "step": 5164 }, { "epoch": 1.6526237201365188, "grad_norm": 0.4871700406074524, "learning_rate": 4.984174770077318e-06, "loss": 0.4822, "step": 5165 }, { "epoch": 1.652943686006826, "grad_norm": 0.47365880012512207, "learning_rate": 4.982312985678694e-06, "loss": 0.4956, "step": 5166 }, { "epoch": 1.653263651877133, "grad_norm": 0.5273645520210266, "learning_rate": 4.980451203732394e-06, "loss": 0.5123, "step": 5167 }, { "epoch": 1.6535836177474401, "grad_norm": 0.4631376564502716, "learning_rate": 4.9785894244965585e-06, "loss": 0.4321, "step": 5168 }, { "epoch": 1.6539035836177476, "grad_norm": 0.5108820199966431, "learning_rate": 4.976727648229326e-06, "loss": 0.5083, "step": 5169 }, { "epoch": 1.6542235494880546, "grad_norm": 0.542654812335968, "learning_rate": 4.974865875188832e-06, "loss": 0.4916, "step": 5170 }, { "epoch": 1.6545435153583616, "grad_norm": 0.4643901288509369, "learning_rate": 4.973004105633216e-06, "loss": 0.4344, "step": 5171 }, { "epoch": 1.6548634812286689, "grad_norm": 0.4893297255039215, "learning_rate": 4.9711423398206115e-06, "loss": 0.4908, "step": 5172 }, { "epoch": 1.6551834470989761, "grad_norm": 0.4885755181312561, "learning_rate": 4.969280578009157e-06, "loss": 0.4866, "step": 5173 }, { "epoch": 1.6555034129692832, "grad_norm": 0.5066401362419128, "learning_rate": 4.967418820456989e-06, "loss": 0.5061, "step": 5174 }, { "epoch": 1.6558233788395904, "grad_norm": 0.5579888224601746, "learning_rate": 4.965557067422238e-06, "loss": 0.4974, "step": 5175 }, { "epoch": 1.6561433447098977, "grad_norm": 0.5364543199539185, "learning_rate": 4.963695319163041e-06, "loss": 0.526, "step": 5176 }, { "epoch": 1.6564633105802047, "grad_norm": 0.5210397243499756, "learning_rate": 4.961833575937534e-06, "loss": 0.498, "step": 5177 }, { "epoch": 1.656783276450512, "grad_norm": 0.5044063329696655, "learning_rate": 4.959971838003846e-06, "loss": 0.447, "step": 5178 }, { "epoch": 1.6571032423208192, "grad_norm": 0.5081261396408081, "learning_rate": 4.9581101056201125e-06, "loss": 0.4988, "step": 5179 }, { "epoch": 1.6574232081911262, "grad_norm": 0.4985758364200592, "learning_rate": 4.956248379044462e-06, "loss": 0.4952, "step": 5180 }, { "epoch": 1.6577431740614335, "grad_norm": 0.5072740912437439, "learning_rate": 4.954386658535025e-06, "loss": 0.4663, "step": 5181 }, { "epoch": 1.6580631399317407, "grad_norm": 0.5251668095588684, "learning_rate": 4.952524944349935e-06, "loss": 0.5379, "step": 5182 }, { "epoch": 1.6583831058020477, "grad_norm": 0.4959995746612549, "learning_rate": 4.950663236747318e-06, "loss": 0.5143, "step": 5183 }, { "epoch": 1.658703071672355, "grad_norm": 0.4989515542984009, "learning_rate": 4.948801535985303e-06, "loss": 0.4442, "step": 5184 }, { "epoch": 1.6590230375426622, "grad_norm": 0.5242576599121094, "learning_rate": 4.946939842322015e-06, "loss": 0.5116, "step": 5185 }, { "epoch": 1.6593430034129693, "grad_norm": 0.5447114109992981, "learning_rate": 4.945078156015582e-06, "loss": 0.5209, "step": 5186 }, { "epoch": 1.6596629692832765, "grad_norm": 0.4931800663471222, "learning_rate": 4.94321647732413e-06, "loss": 0.4093, "step": 5187 }, { "epoch": 1.6599829351535837, "grad_norm": 0.4911198616027832, "learning_rate": 4.941354806505781e-06, "loss": 0.531, "step": 5188 }, { "epoch": 1.6603029010238908, "grad_norm": 0.47494980692863464, "learning_rate": 4.939493143818659e-06, "loss": 0.4545, "step": 5189 }, { "epoch": 1.6606228668941978, "grad_norm": 0.5940124988555908, "learning_rate": 4.937631489520886e-06, "loss": 0.544, "step": 5190 }, { "epoch": 1.6609428327645053, "grad_norm": 0.5294722318649292, "learning_rate": 4.935769843870579e-06, "loss": 0.5287, "step": 5191 }, { "epoch": 1.6612627986348123, "grad_norm": 0.4588381350040436, "learning_rate": 4.933908207125864e-06, "loss": 0.4448, "step": 5192 }, { "epoch": 1.6615827645051193, "grad_norm": 0.6200984120368958, "learning_rate": 4.932046579544853e-06, "loss": 0.561, "step": 5193 }, { "epoch": 1.6619027303754266, "grad_norm": 0.4721447229385376, "learning_rate": 4.930184961385668e-06, "loss": 0.4833, "step": 5194 }, { "epoch": 1.6622226962457338, "grad_norm": 0.5099005699157715, "learning_rate": 4.928323352906421e-06, "loss": 0.4971, "step": 5195 }, { "epoch": 1.6625426621160408, "grad_norm": 0.49571260809898376, "learning_rate": 4.9264617543652255e-06, "loss": 0.4256, "step": 5196 }, { "epoch": 1.662862627986348, "grad_norm": 0.5441554188728333, "learning_rate": 4.9246001660201986e-06, "loss": 0.5296, "step": 5197 }, { "epoch": 1.6631825938566553, "grad_norm": 0.5576848983764648, "learning_rate": 4.922738588129449e-06, "loss": 0.5088, "step": 5198 }, { "epoch": 1.6635025597269624, "grad_norm": 0.5072476863861084, "learning_rate": 4.920877020951087e-06, "loss": 0.4697, "step": 5199 }, { "epoch": 1.6638225255972696, "grad_norm": 0.5600112676620483, "learning_rate": 4.919015464743221e-06, "loss": 0.4881, "step": 5200 }, { "epoch": 1.6641424914675769, "grad_norm": 0.5764448642730713, "learning_rate": 4.917153919763956e-06, "loss": 0.4724, "step": 5201 }, { "epoch": 1.664462457337884, "grad_norm": 0.49314001202583313, "learning_rate": 4.915292386271403e-06, "loss": 0.5051, "step": 5202 }, { "epoch": 1.6647824232081911, "grad_norm": 0.5252131819725037, "learning_rate": 4.913430864523662e-06, "loss": 0.5285, "step": 5203 }, { "epoch": 1.6651023890784984, "grad_norm": 0.5187550783157349, "learning_rate": 4.911569354778836e-06, "loss": 0.4487, "step": 5204 }, { "epoch": 1.6654223549488054, "grad_norm": 0.4986821711063385, "learning_rate": 4.909707857295024e-06, "loss": 0.4566, "step": 5205 }, { "epoch": 1.6657423208191127, "grad_norm": 0.4897250831127167, "learning_rate": 4.907846372330326e-06, "loss": 0.487, "step": 5206 }, { "epoch": 1.66606228668942, "grad_norm": 0.5298472046852112, "learning_rate": 4.90598490014284e-06, "loss": 0.4788, "step": 5207 }, { "epoch": 1.666382252559727, "grad_norm": 0.5566151142120361, "learning_rate": 4.9041234409906615e-06, "loss": 0.5047, "step": 5208 }, { "epoch": 1.666702218430034, "grad_norm": 0.5224797129631042, "learning_rate": 4.902261995131883e-06, "loss": 0.4653, "step": 5209 }, { "epoch": 1.6670221843003414, "grad_norm": 0.5086098909378052, "learning_rate": 4.900400562824595e-06, "loss": 0.5158, "step": 5210 }, { "epoch": 1.6673421501706485, "grad_norm": 0.522621750831604, "learning_rate": 4.898539144326889e-06, "loss": 0.5033, "step": 5211 }, { "epoch": 1.6676621160409555, "grad_norm": 0.5157283544540405, "learning_rate": 4.8966777398968545e-06, "loss": 0.5352, "step": 5212 }, { "epoch": 1.667982081911263, "grad_norm": 0.5230862498283386, "learning_rate": 4.894816349792575e-06, "loss": 0.4778, "step": 5213 }, { "epoch": 1.66830204778157, "grad_norm": 0.5108460187911987, "learning_rate": 4.892954974272135e-06, "loss": 0.4818, "step": 5214 }, { "epoch": 1.668622013651877, "grad_norm": 0.5653759837150574, "learning_rate": 4.891093613593615e-06, "loss": 0.462, "step": 5215 }, { "epoch": 1.6689419795221843, "grad_norm": 0.5857290029525757, "learning_rate": 4.889232268015098e-06, "loss": 0.5238, "step": 5216 }, { "epoch": 1.6692619453924915, "grad_norm": 0.5446109771728516, "learning_rate": 4.887370937794661e-06, "loss": 0.5031, "step": 5217 }, { "epoch": 1.6695819112627985, "grad_norm": 0.4403999447822571, "learning_rate": 4.885509623190378e-06, "loss": 0.4663, "step": 5218 }, { "epoch": 1.6699018771331058, "grad_norm": 0.5504063367843628, "learning_rate": 4.883648324460325e-06, "loss": 0.5099, "step": 5219 }, { "epoch": 1.670221843003413, "grad_norm": 0.5568899512290955, "learning_rate": 4.881787041862569e-06, "loss": 0.4942, "step": 5220 }, { "epoch": 1.67054180887372, "grad_norm": 0.45460644364356995, "learning_rate": 4.879925775655183e-06, "loss": 0.4857, "step": 5221 }, { "epoch": 1.6708617747440273, "grad_norm": 0.48582178354263306, "learning_rate": 4.878064526096233e-06, "loss": 0.4674, "step": 5222 }, { "epoch": 1.6711817406143346, "grad_norm": 0.5432444214820862, "learning_rate": 4.876203293443783e-06, "loss": 0.5085, "step": 5223 }, { "epoch": 1.6715017064846416, "grad_norm": 0.5521763563156128, "learning_rate": 4.874342077955896e-06, "loss": 0.4981, "step": 5224 }, { "epoch": 1.6718216723549488, "grad_norm": 0.5622273683547974, "learning_rate": 4.87248087989063e-06, "loss": 0.5365, "step": 5225 }, { "epoch": 1.672141638225256, "grad_norm": 0.48718908429145813, "learning_rate": 4.870619699506044e-06, "loss": 0.5037, "step": 5226 }, { "epoch": 1.672461604095563, "grad_norm": 0.4818768799304962, "learning_rate": 4.868758537060192e-06, "loss": 0.51, "step": 5227 }, { "epoch": 1.6727815699658704, "grad_norm": 0.5165402293205261, "learning_rate": 4.866897392811127e-06, "loss": 0.4729, "step": 5228 }, { "epoch": 1.6731015358361776, "grad_norm": 0.5203539729118347, "learning_rate": 4.8650362670168995e-06, "loss": 0.5086, "step": 5229 }, { "epoch": 1.6734215017064846, "grad_norm": 0.5446471571922302, "learning_rate": 4.8631751599355535e-06, "loss": 0.4446, "step": 5230 }, { "epoch": 1.6737414675767917, "grad_norm": 0.5485162734985352, "learning_rate": 4.861314071825137e-06, "loss": 0.4937, "step": 5231 }, { "epoch": 1.6740614334470991, "grad_norm": 0.5686075091362, "learning_rate": 4.859453002943693e-06, "loss": 0.4762, "step": 5232 }, { "epoch": 1.6743813993174061, "grad_norm": 0.5025482773780823, "learning_rate": 4.857591953549258e-06, "loss": 0.5068, "step": 5233 }, { "epoch": 1.6747013651877132, "grad_norm": 0.5636971592903137, "learning_rate": 4.855730923899871e-06, "loss": 0.5251, "step": 5234 }, { "epoch": 1.6750213310580204, "grad_norm": 0.545708954334259, "learning_rate": 4.853869914253563e-06, "loss": 0.515, "step": 5235 }, { "epoch": 1.6753412969283277, "grad_norm": 0.5094165205955505, "learning_rate": 4.852008924868368e-06, "loss": 0.4409, "step": 5236 }, { "epoch": 1.6756612627986347, "grad_norm": 0.5279160141944885, "learning_rate": 4.850147956002314e-06, "loss": 0.4593, "step": 5237 }, { "epoch": 1.675981228668942, "grad_norm": 0.5463525056838989, "learning_rate": 4.8482870079134266e-06, "loss": 0.4775, "step": 5238 }, { "epoch": 1.6763011945392492, "grad_norm": 0.5538739562034607, "learning_rate": 4.846426080859728e-06, "loss": 0.4966, "step": 5239 }, { "epoch": 1.6766211604095562, "grad_norm": 0.5735815167427063, "learning_rate": 4.844565175099236e-06, "loss": 0.4424, "step": 5240 }, { "epoch": 1.6769411262798635, "grad_norm": 0.601814866065979, "learning_rate": 4.842704290889971e-06, "loss": 0.5584, "step": 5241 }, { "epoch": 1.6772610921501707, "grad_norm": 0.5215892195701599, "learning_rate": 4.840843428489947e-06, "loss": 0.415, "step": 5242 }, { "epoch": 1.6775810580204777, "grad_norm": 0.5272678732872009, "learning_rate": 4.838982588157171e-06, "loss": 0.4728, "step": 5243 }, { "epoch": 1.677901023890785, "grad_norm": 0.5651388764381409, "learning_rate": 4.837121770149654e-06, "loss": 0.505, "step": 5244 }, { "epoch": 1.6782209897610922, "grad_norm": 0.49386903643608093, "learning_rate": 4.835260974725397e-06, "loss": 0.4553, "step": 5245 }, { "epoch": 1.6785409556313993, "grad_norm": 0.4737684726715088, "learning_rate": 4.8334002021424056e-06, "loss": 0.4795, "step": 5246 }, { "epoch": 1.6788609215017065, "grad_norm": 0.5495589375495911, "learning_rate": 4.831539452658678e-06, "loss": 0.5644, "step": 5247 }, { "epoch": 1.6791808873720138, "grad_norm": 0.5037122368812561, "learning_rate": 4.829678726532207e-06, "loss": 0.4522, "step": 5248 }, { "epoch": 1.6795008532423208, "grad_norm": 0.5186344981193542, "learning_rate": 4.8278180240209835e-06, "loss": 0.5104, "step": 5249 }, { "epoch": 1.679820819112628, "grad_norm": 0.48219427466392517, "learning_rate": 4.825957345383e-06, "loss": 0.5191, "step": 5250 }, { "epoch": 1.6801407849829353, "grad_norm": 0.45844265818595886, "learning_rate": 4.82409669087624e-06, "loss": 0.4388, "step": 5251 }, { "epoch": 1.6804607508532423, "grad_norm": 0.552278995513916, "learning_rate": 4.822236060758686e-06, "loss": 0.5429, "step": 5252 }, { "epoch": 1.6807807167235493, "grad_norm": 0.5578888654708862, "learning_rate": 4.820375455288316e-06, "loss": 0.4918, "step": 5253 }, { "epoch": 1.6811006825938568, "grad_norm": 0.49955689907073975, "learning_rate": 4.818514874723103e-06, "loss": 0.4605, "step": 5254 }, { "epoch": 1.6814206484641638, "grad_norm": 0.5871486663818359, "learning_rate": 4.816654319321022e-06, "loss": 0.5048, "step": 5255 }, { "epoch": 1.6817406143344709, "grad_norm": 0.5110973119735718, "learning_rate": 4.814793789340041e-06, "loss": 0.4682, "step": 5256 }, { "epoch": 1.682060580204778, "grad_norm": 0.5397192239761353, "learning_rate": 4.812933285038125e-06, "loss": 0.5696, "step": 5257 }, { "epoch": 1.6823805460750854, "grad_norm": 0.54073566198349, "learning_rate": 4.811072806673231e-06, "loss": 0.4941, "step": 5258 }, { "epoch": 1.6827005119453924, "grad_norm": 0.4927590787410736, "learning_rate": 4.809212354503321e-06, "loss": 0.4296, "step": 5259 }, { "epoch": 1.6830204778156996, "grad_norm": 0.4850064218044281, "learning_rate": 4.807351928786348e-06, "loss": 0.4729, "step": 5260 }, { "epoch": 1.6833404436860069, "grad_norm": 0.48674237728118896, "learning_rate": 4.805491529780261e-06, "loss": 0.4501, "step": 5261 }, { "epoch": 1.683660409556314, "grad_norm": 0.4843261241912842, "learning_rate": 4.803631157743009e-06, "loss": 0.4757, "step": 5262 }, { "epoch": 1.6839803754266212, "grad_norm": 0.48974671959877014, "learning_rate": 4.801770812932533e-06, "loss": 0.4378, "step": 5263 }, { "epoch": 1.6843003412969284, "grad_norm": 0.5552740693092346, "learning_rate": 4.7999104956067696e-06, "loss": 0.524, "step": 5264 }, { "epoch": 1.6846203071672354, "grad_norm": 0.5226063132286072, "learning_rate": 4.79805020602366e-06, "loss": 0.4503, "step": 5265 }, { "epoch": 1.6849402730375427, "grad_norm": 0.5147300362586975, "learning_rate": 4.796189944441131e-06, "loss": 0.4908, "step": 5266 }, { "epoch": 1.68526023890785, "grad_norm": 0.5338008999824524, "learning_rate": 4.7943297111171115e-06, "loss": 0.4623, "step": 5267 }, { "epoch": 1.685580204778157, "grad_norm": 0.5334160923957825, "learning_rate": 4.792469506309525e-06, "loss": 0.5154, "step": 5268 }, { "epoch": 1.6859001706484642, "grad_norm": 0.5443989038467407, "learning_rate": 4.7906093302762895e-06, "loss": 0.4864, "step": 5269 }, { "epoch": 1.6862201365187715, "grad_norm": 0.5522932410240173, "learning_rate": 4.7887491832753245e-06, "loss": 0.4796, "step": 5270 }, { "epoch": 1.6865401023890785, "grad_norm": 0.5522817373275757, "learning_rate": 4.7868890655645385e-06, "loss": 0.5353, "step": 5271 }, { "epoch": 1.6868600682593855, "grad_norm": 0.5342884063720703, "learning_rate": 4.78502897740184e-06, "loss": 0.4993, "step": 5272 }, { "epoch": 1.687180034129693, "grad_norm": 0.5327131152153015, "learning_rate": 4.783168919045133e-06, "loss": 0.4668, "step": 5273 }, { "epoch": 1.6875, "grad_norm": 0.5152961611747742, "learning_rate": 4.7813088907523145e-06, "loss": 0.5055, "step": 5274 }, { "epoch": 1.687819965870307, "grad_norm": 0.5315884947776794, "learning_rate": 4.779448892781283e-06, "loss": 0.5097, "step": 5275 }, { "epoch": 1.6881399317406145, "grad_norm": 0.4565909802913666, "learning_rate": 4.7775889253899276e-06, "loss": 0.4346, "step": 5276 }, { "epoch": 1.6884598976109215, "grad_norm": 0.5067062973976135, "learning_rate": 4.7757289888361354e-06, "loss": 0.4961, "step": 5277 }, { "epoch": 1.6887798634812285, "grad_norm": 0.5119850635528564, "learning_rate": 4.773869083377788e-06, "loss": 0.5255, "step": 5278 }, { "epoch": 1.6890998293515358, "grad_norm": 0.4938701391220093, "learning_rate": 4.772009209272763e-06, "loss": 0.4959, "step": 5279 }, { "epoch": 1.689419795221843, "grad_norm": 0.5242329835891724, "learning_rate": 4.770149366778938e-06, "loss": 0.4476, "step": 5280 }, { "epoch": 1.68973976109215, "grad_norm": 0.5618206858634949, "learning_rate": 4.768289556154179e-06, "loss": 0.4743, "step": 5281 }, { "epoch": 1.6900597269624573, "grad_norm": 0.5645880103111267, "learning_rate": 4.766429777656352e-06, "loss": 0.497, "step": 5282 }, { "epoch": 1.6903796928327646, "grad_norm": 0.46402737498283386, "learning_rate": 4.764570031543316e-06, "loss": 0.5025, "step": 5283 }, { "epoch": 1.6906996587030716, "grad_norm": 0.44544854760169983, "learning_rate": 4.762710318072929e-06, "loss": 0.4349, "step": 5284 }, { "epoch": 1.6910196245733788, "grad_norm": 0.4936232566833496, "learning_rate": 4.760850637503043e-06, "loss": 0.4617, "step": 5285 }, { "epoch": 1.691339590443686, "grad_norm": 0.5291861295700073, "learning_rate": 4.758990990091502e-06, "loss": 0.5592, "step": 5286 }, { "epoch": 1.6916595563139931, "grad_norm": 0.5856315493583679, "learning_rate": 4.757131376096153e-06, "loss": 0.4911, "step": 5287 }, { "epoch": 1.6919795221843004, "grad_norm": 0.4970265030860901, "learning_rate": 4.755271795774827e-06, "loss": 0.4808, "step": 5288 }, { "epoch": 1.6922994880546076, "grad_norm": 0.48993000388145447, "learning_rate": 4.753412249385363e-06, "loss": 0.5011, "step": 5289 }, { "epoch": 1.6926194539249146, "grad_norm": 0.49721401929855347, "learning_rate": 4.751552737185588e-06, "loss": 0.4306, "step": 5290 }, { "epoch": 1.692939419795222, "grad_norm": 0.46261000633239746, "learning_rate": 4.749693259433323e-06, "loss": 0.4482, "step": 5291 }, { "epoch": 1.6932593856655291, "grad_norm": 0.5152326226234436, "learning_rate": 4.747833816386391e-06, "loss": 0.5524, "step": 5292 }, { "epoch": 1.6935793515358362, "grad_norm": 0.5263875126838684, "learning_rate": 4.745974408302602e-06, "loss": 0.5055, "step": 5293 }, { "epoch": 1.6938993174061432, "grad_norm": 0.5017513632774353, "learning_rate": 4.744115035439766e-06, "loss": 0.4955, "step": 5294 }, { "epoch": 1.6942192832764507, "grad_norm": 0.5186927914619446, "learning_rate": 4.742255698055691e-06, "loss": 0.5308, "step": 5295 }, { "epoch": 1.6945392491467577, "grad_norm": 0.5493725538253784, "learning_rate": 4.740396396408172e-06, "loss": 0.4872, "step": 5296 }, { "epoch": 1.6948592150170647, "grad_norm": 0.5045214891433716, "learning_rate": 4.738537130755006e-06, "loss": 0.4415, "step": 5297 }, { "epoch": 1.695179180887372, "grad_norm": 0.5636321306228638, "learning_rate": 4.736677901353978e-06, "loss": 0.4903, "step": 5298 }, { "epoch": 1.6954991467576792, "grad_norm": 0.5362817645072937, "learning_rate": 4.734818708462879e-06, "loss": 0.4758, "step": 5299 }, { "epoch": 1.6958191126279862, "grad_norm": 0.517052412033081, "learning_rate": 4.732959552339485e-06, "loss": 0.5091, "step": 5300 }, { "epoch": 1.6961390784982935, "grad_norm": 0.5431403517723083, "learning_rate": 4.731100433241569e-06, "loss": 0.4677, "step": 5301 }, { "epoch": 1.6964590443686007, "grad_norm": 0.5283934473991394, "learning_rate": 4.729241351426903e-06, "loss": 0.5165, "step": 5302 }, { "epoch": 1.6967790102389078, "grad_norm": 0.5136128067970276, "learning_rate": 4.727382307153246e-06, "loss": 0.4849, "step": 5303 }, { "epoch": 1.697098976109215, "grad_norm": 0.5218483209609985, "learning_rate": 4.7255233006783626e-06, "loss": 0.5034, "step": 5304 }, { "epoch": 1.6974189419795223, "grad_norm": 0.52696293592453, "learning_rate": 4.723664332260004e-06, "loss": 0.4628, "step": 5305 }, { "epoch": 1.6977389078498293, "grad_norm": 0.4817391335964203, "learning_rate": 4.721805402155919e-06, "loss": 0.4892, "step": 5306 }, { "epoch": 1.6980588737201365, "grad_norm": 0.49527376890182495, "learning_rate": 4.719946510623849e-06, "loss": 0.4826, "step": 5307 }, { "epoch": 1.6983788395904438, "grad_norm": 0.5037565231323242, "learning_rate": 4.718087657921529e-06, "loss": 0.4797, "step": 5308 }, { "epoch": 1.6986988054607508, "grad_norm": 0.5579425096511841, "learning_rate": 4.716228844306699e-06, "loss": 0.4991, "step": 5309 }, { "epoch": 1.699018771331058, "grad_norm": 0.5586383938789368, "learning_rate": 4.714370070037078e-06, "loss": 0.4815, "step": 5310 }, { "epoch": 1.6993387372013653, "grad_norm": 0.48886433243751526, "learning_rate": 4.712511335370393e-06, "loss": 0.4813, "step": 5311 }, { "epoch": 1.6996587030716723, "grad_norm": 0.5777683258056641, "learning_rate": 4.710652640564355e-06, "loss": 0.5003, "step": 5312 }, { "epoch": 1.6999786689419796, "grad_norm": 0.4665423035621643, "learning_rate": 4.708793985876676e-06, "loss": 0.4176, "step": 5313 }, { "epoch": 1.7002986348122868, "grad_norm": 0.514366626739502, "learning_rate": 4.706935371565064e-06, "loss": 0.5199, "step": 5314 }, { "epoch": 1.7006186006825939, "grad_norm": 0.5152296423912048, "learning_rate": 4.7050767978872124e-06, "loss": 0.4805, "step": 5315 }, { "epoch": 1.7009385665529009, "grad_norm": 0.5062317252159119, "learning_rate": 4.7032182651008204e-06, "loss": 0.4758, "step": 5316 }, { "epoch": 1.7012585324232083, "grad_norm": 0.5014505982398987, "learning_rate": 4.701359773463569e-06, "loss": 0.5328, "step": 5317 }, { "epoch": 1.7015784982935154, "grad_norm": 0.5042576193809509, "learning_rate": 4.699501323233148e-06, "loss": 0.4486, "step": 5318 }, { "epoch": 1.7018984641638224, "grad_norm": 0.49599841237068176, "learning_rate": 4.697642914667229e-06, "loss": 0.5328, "step": 5319 }, { "epoch": 1.7022184300341296, "grad_norm": 0.556387722492218, "learning_rate": 4.695784548023483e-06, "loss": 0.5067, "step": 5320 }, { "epoch": 1.702538395904437, "grad_norm": 0.5078489780426025, "learning_rate": 4.693926223559577e-06, "loss": 0.4924, "step": 5321 }, { "epoch": 1.702858361774744, "grad_norm": 0.540920078754425, "learning_rate": 4.692067941533166e-06, "loss": 0.5612, "step": 5322 }, { "epoch": 1.7031783276450512, "grad_norm": 0.43708547949790955, "learning_rate": 4.690209702201906e-06, "loss": 0.4392, "step": 5323 }, { "epoch": 1.7034982935153584, "grad_norm": 0.5136210918426514, "learning_rate": 4.688351505823445e-06, "loss": 0.4489, "step": 5324 }, { "epoch": 1.7038182593856654, "grad_norm": 0.5493529438972473, "learning_rate": 4.686493352655421e-06, "loss": 0.5417, "step": 5325 }, { "epoch": 1.7041382252559727, "grad_norm": 0.5002792477607727, "learning_rate": 4.684635242955473e-06, "loss": 0.5055, "step": 5326 }, { "epoch": 1.70445819112628, "grad_norm": 0.48529189825057983, "learning_rate": 4.682777176981225e-06, "loss": 0.4326, "step": 5327 }, { "epoch": 1.704778156996587, "grad_norm": 0.482902854681015, "learning_rate": 4.6809191549903045e-06, "loss": 0.4539, "step": 5328 }, { "epoch": 1.7050981228668942, "grad_norm": 0.5050659775733948, "learning_rate": 4.679061177240328e-06, "loss": 0.5154, "step": 5329 }, { "epoch": 1.7054180887372015, "grad_norm": 0.5438242554664612, "learning_rate": 4.677203243988904e-06, "loss": 0.5037, "step": 5330 }, { "epoch": 1.7057380546075085, "grad_norm": 0.4972843825817108, "learning_rate": 4.675345355493639e-06, "loss": 0.4288, "step": 5331 }, { "epoch": 1.7060580204778157, "grad_norm": 0.51446533203125, "learning_rate": 4.67348751201213e-06, "loss": 0.502, "step": 5332 }, { "epoch": 1.706377986348123, "grad_norm": 0.5037707090377808, "learning_rate": 4.671629713801971e-06, "loss": 0.5552, "step": 5333 }, { "epoch": 1.70669795221843, "grad_norm": 0.5276395082473755, "learning_rate": 4.669771961120748e-06, "loss": 0.4873, "step": 5334 }, { "epoch": 1.707017918088737, "grad_norm": 0.5363672375679016, "learning_rate": 4.667914254226039e-06, "loss": 0.4185, "step": 5335 }, { "epoch": 1.7073378839590445, "grad_norm": 0.5792633891105652, "learning_rate": 4.66605659337542e-06, "loss": 0.5128, "step": 5336 }, { "epoch": 1.7076578498293515, "grad_norm": 0.5017244219779968, "learning_rate": 4.664198978826453e-06, "loss": 0.5228, "step": 5337 }, { "epoch": 1.7079778156996586, "grad_norm": 0.5084531903266907, "learning_rate": 4.662341410836703e-06, "loss": 0.476, "step": 5338 }, { "epoch": 1.708297781569966, "grad_norm": 0.5443561673164368, "learning_rate": 4.660483889663723e-06, "loss": 0.492, "step": 5339 }, { "epoch": 1.708617747440273, "grad_norm": 0.4886680841445923, "learning_rate": 4.6586264155650594e-06, "loss": 0.4512, "step": 5340 }, { "epoch": 1.70893771331058, "grad_norm": 0.49199721217155457, "learning_rate": 4.6567689887982555e-06, "loss": 0.4372, "step": 5341 }, { "epoch": 1.7092576791808873, "grad_norm": 0.5576441884040833, "learning_rate": 4.654911609620842e-06, "loss": 0.5568, "step": 5342 }, { "epoch": 1.7095776450511946, "grad_norm": 0.5655575394630432, "learning_rate": 4.65305427829035e-06, "loss": 0.5534, "step": 5343 }, { "epoch": 1.7098976109215016, "grad_norm": 0.5352281332015991, "learning_rate": 4.651196995064301e-06, "loss": 0.462, "step": 5344 }, { "epoch": 1.7102175767918089, "grad_norm": 0.5099778175354004, "learning_rate": 4.649339760200206e-06, "loss": 0.4943, "step": 5345 }, { "epoch": 1.710537542662116, "grad_norm": 0.49300286173820496, "learning_rate": 4.647482573955577e-06, "loss": 0.4141, "step": 5346 }, { "epoch": 1.7108575085324231, "grad_norm": 0.5802779197692871, "learning_rate": 4.645625436587911e-06, "loss": 0.5397, "step": 5347 }, { "epoch": 1.7111774744027304, "grad_norm": 0.5505796074867249, "learning_rate": 4.6437683483547055e-06, "loss": 0.5048, "step": 5348 }, { "epoch": 1.7114974402730376, "grad_norm": 0.5246226191520691, "learning_rate": 4.641911309513449e-06, "loss": 0.4853, "step": 5349 }, { "epoch": 1.7118174061433447, "grad_norm": 0.475522518157959, "learning_rate": 4.640054320321618e-06, "loss": 0.4368, "step": 5350 }, { "epoch": 1.712137372013652, "grad_norm": 0.5598135590553284, "learning_rate": 4.6381973810366905e-06, "loss": 0.4944, "step": 5351 }, { "epoch": 1.7124573378839592, "grad_norm": 0.5293557643890381, "learning_rate": 4.636340491916128e-06, "loss": 0.4788, "step": 5352 }, { "epoch": 1.7127773037542662, "grad_norm": 0.4790298044681549, "learning_rate": 4.634483653217396e-06, "loss": 0.4699, "step": 5353 }, { "epoch": 1.7130972696245734, "grad_norm": 0.5051926970481873, "learning_rate": 4.632626865197946e-06, "loss": 0.4542, "step": 5354 }, { "epoch": 1.7134172354948807, "grad_norm": 0.48448479175567627, "learning_rate": 4.630770128115222e-06, "loss": 0.4639, "step": 5355 }, { "epoch": 1.7137372013651877, "grad_norm": 0.5340156555175781, "learning_rate": 4.628913442226663e-06, "loss": 0.5082, "step": 5356 }, { "epoch": 1.7140571672354947, "grad_norm": 0.5201172232627869, "learning_rate": 4.627056807789703e-06, "loss": 0.4907, "step": 5357 }, { "epoch": 1.7143771331058022, "grad_norm": 0.4603649973869324, "learning_rate": 4.625200225061765e-06, "loss": 0.4678, "step": 5358 }, { "epoch": 1.7146970989761092, "grad_norm": 0.5088764429092407, "learning_rate": 4.623343694300268e-06, "loss": 0.4958, "step": 5359 }, { "epoch": 1.7150170648464163, "grad_norm": 0.5996106863021851, "learning_rate": 4.621487215762619e-06, "loss": 0.5432, "step": 5360 }, { "epoch": 1.7153370307167235, "grad_norm": 0.4862479269504547, "learning_rate": 4.619630789706223e-06, "loss": 0.492, "step": 5361 }, { "epoch": 1.7156569965870307, "grad_norm": 0.5018873810768127, "learning_rate": 4.617774416388476e-06, "loss": 0.4865, "step": 5362 }, { "epoch": 1.7159769624573378, "grad_norm": 0.56586754322052, "learning_rate": 4.615918096066766e-06, "loss": 0.5417, "step": 5363 }, { "epoch": 1.716296928327645, "grad_norm": 0.486603707075119, "learning_rate": 4.614061828998475e-06, "loss": 0.4384, "step": 5364 }, { "epoch": 1.7166168941979523, "grad_norm": 0.485688179731369, "learning_rate": 4.612205615440975e-06, "loss": 0.502, "step": 5365 }, { "epoch": 1.7169368600682593, "grad_norm": 0.529779314994812, "learning_rate": 4.610349455651632e-06, "loss": 0.4902, "step": 5366 }, { "epoch": 1.7172568259385665, "grad_norm": 0.4693937599658966, "learning_rate": 4.608493349887807e-06, "loss": 0.4407, "step": 5367 }, { "epoch": 1.7175767918088738, "grad_norm": 0.49611836671829224, "learning_rate": 4.606637298406849e-06, "loss": 0.529, "step": 5368 }, { "epoch": 1.7178967576791808, "grad_norm": 0.5174630284309387, "learning_rate": 4.604781301466105e-06, "loss": 0.4793, "step": 5369 }, { "epoch": 1.718216723549488, "grad_norm": 0.4785979390144348, "learning_rate": 4.6029253593229065e-06, "loss": 0.4402, "step": 5370 }, { "epoch": 1.7185366894197953, "grad_norm": 0.4945693910121918, "learning_rate": 4.601069472234584e-06, "loss": 0.4642, "step": 5371 }, { "epoch": 1.7188566552901023, "grad_norm": 0.5145953893661499, "learning_rate": 4.599213640458462e-06, "loss": 0.4756, "step": 5372 }, { "epoch": 1.7191766211604096, "grad_norm": 0.5472128987312317, "learning_rate": 4.597357864251849e-06, "loss": 0.5126, "step": 5373 }, { "epoch": 1.7194965870307168, "grad_norm": 0.4303402304649353, "learning_rate": 4.5955021438720544e-06, "loss": 0.4392, "step": 5374 }, { "epoch": 1.7198165529010239, "grad_norm": 0.544348955154419, "learning_rate": 4.593646479576372e-06, "loss": 0.4915, "step": 5375 }, { "epoch": 1.7201365187713311, "grad_norm": 0.5345340967178345, "learning_rate": 4.5917908716220934e-06, "loss": 0.4787, "step": 5376 }, { "epoch": 1.7204564846416384, "grad_norm": 0.5413882732391357, "learning_rate": 4.589935320266505e-06, "loss": 0.4689, "step": 5377 }, { "epoch": 1.7207764505119454, "grad_norm": 0.5263100862503052, "learning_rate": 4.588079825766875e-06, "loss": 0.4881, "step": 5378 }, { "epoch": 1.7210964163822524, "grad_norm": 0.5405126214027405, "learning_rate": 4.586224388380474e-06, "loss": 0.4922, "step": 5379 }, { "epoch": 1.7214163822525599, "grad_norm": 0.5615954995155334, "learning_rate": 4.584369008364559e-06, "loss": 0.5068, "step": 5380 }, { "epoch": 1.721736348122867, "grad_norm": 0.5604912638664246, "learning_rate": 4.582513685976379e-06, "loss": 0.501, "step": 5381 }, { "epoch": 1.722056313993174, "grad_norm": 0.5413294434547424, "learning_rate": 4.580658421473182e-06, "loss": 0.5194, "step": 5382 }, { "epoch": 1.7223762798634812, "grad_norm": 0.5368830561637878, "learning_rate": 4.5788032151121976e-06, "loss": 0.5022, "step": 5383 }, { "epoch": 1.7226962457337884, "grad_norm": 0.5325060486793518, "learning_rate": 4.576948067150655e-06, "loss": 0.4813, "step": 5384 }, { "epoch": 1.7230162116040955, "grad_norm": 0.5181230902671814, "learning_rate": 4.575092977845771e-06, "loss": 0.4177, "step": 5385 }, { "epoch": 1.7233361774744027, "grad_norm": 0.5556210875511169, "learning_rate": 4.573237947454755e-06, "loss": 0.4621, "step": 5386 }, { "epoch": 1.72365614334471, "grad_norm": 0.5162448287010193, "learning_rate": 4.5713829762348145e-06, "loss": 0.4613, "step": 5387 }, { "epoch": 1.723976109215017, "grad_norm": 0.5468537211418152, "learning_rate": 4.569528064443138e-06, "loss": 0.5083, "step": 5388 }, { "epoch": 1.7242960750853242, "grad_norm": 0.5611181259155273, "learning_rate": 4.567673212336915e-06, "loss": 0.5377, "step": 5389 }, { "epoch": 1.7246160409556315, "grad_norm": 0.5240436792373657, "learning_rate": 4.565818420173319e-06, "loss": 0.4401, "step": 5390 }, { "epoch": 1.7249360068259385, "grad_norm": 0.5515648126602173, "learning_rate": 4.5639636882095215e-06, "loss": 0.568, "step": 5391 }, { "epoch": 1.7252559726962458, "grad_norm": 0.5274022221565247, "learning_rate": 4.5621090167026846e-06, "loss": 0.4569, "step": 5392 }, { "epoch": 1.725575938566553, "grad_norm": 0.5471010804176331, "learning_rate": 4.560254405909959e-06, "loss": 0.4855, "step": 5393 }, { "epoch": 1.72589590443686, "grad_norm": 0.5593867897987366, "learning_rate": 4.558399856088489e-06, "loss": 0.4829, "step": 5394 }, { "epoch": 1.7262158703071673, "grad_norm": 0.5176609754562378, "learning_rate": 4.5565453674954095e-06, "loss": 0.4627, "step": 5395 }, { "epoch": 1.7265358361774745, "grad_norm": 0.5561937093734741, "learning_rate": 4.554690940387848e-06, "loss": 0.5149, "step": 5396 }, { "epoch": 1.7268558020477816, "grad_norm": 0.5249950289726257, "learning_rate": 4.552836575022925e-06, "loss": 0.5108, "step": 5397 }, { "epoch": 1.7271757679180886, "grad_norm": 0.46012160181999207, "learning_rate": 4.550982271657748e-06, "loss": 0.4368, "step": 5398 }, { "epoch": 1.727495733788396, "grad_norm": 0.4988340139389038, "learning_rate": 4.54912803054942e-06, "loss": 0.4959, "step": 5399 }, { "epoch": 1.727815699658703, "grad_norm": 0.5390985608100891, "learning_rate": 4.547273851955031e-06, "loss": 0.4913, "step": 5400 }, { "epoch": 1.72813566552901, "grad_norm": 0.5166676044464111, "learning_rate": 4.545419736131668e-06, "loss": 0.46, "step": 5401 }, { "epoch": 1.7284556313993176, "grad_norm": 0.47667017579078674, "learning_rate": 4.543565683336407e-06, "loss": 0.476, "step": 5402 }, { "epoch": 1.7287755972696246, "grad_norm": 0.46250563859939575, "learning_rate": 4.541711693826311e-06, "loss": 0.5012, "step": 5403 }, { "epoch": 1.7290955631399316, "grad_norm": 0.4555782675743103, "learning_rate": 4.539857767858442e-06, "loss": 0.4517, "step": 5404 }, { "epoch": 1.7294155290102389, "grad_norm": 0.49214455485343933, "learning_rate": 4.538003905689844e-06, "loss": 0.4959, "step": 5405 }, { "epoch": 1.7297354948805461, "grad_norm": 0.5046555995941162, "learning_rate": 4.536150107577562e-06, "loss": 0.4794, "step": 5406 }, { "epoch": 1.7300554607508531, "grad_norm": 0.4979184865951538, "learning_rate": 4.534296373778627e-06, "loss": 0.4775, "step": 5407 }, { "epoch": 1.7303754266211604, "grad_norm": 0.46941980719566345, "learning_rate": 4.532442704550058e-06, "loss": 0.498, "step": 5408 }, { "epoch": 1.7306953924914676, "grad_norm": 0.41967618465423584, "learning_rate": 4.530589100148872e-06, "loss": 0.4439, "step": 5409 }, { "epoch": 1.7310153583617747, "grad_norm": 0.5239220857620239, "learning_rate": 4.528735560832071e-06, "loss": 0.5585, "step": 5410 }, { "epoch": 1.731335324232082, "grad_norm": 0.5077087879180908, "learning_rate": 4.52688208685665e-06, "loss": 0.4779, "step": 5411 }, { "epoch": 1.7316552901023892, "grad_norm": 0.5944514870643616, "learning_rate": 4.5250286784796e-06, "loss": 0.531, "step": 5412 }, { "epoch": 1.7319752559726962, "grad_norm": 0.5323827266693115, "learning_rate": 4.523175335957894e-06, "loss": 0.4265, "step": 5413 }, { "epoch": 1.7322952218430034, "grad_norm": 0.5563470721244812, "learning_rate": 4.521322059548501e-06, "loss": 0.5252, "step": 5414 }, { "epoch": 1.7326151877133107, "grad_norm": 0.5009718537330627, "learning_rate": 4.519468849508379e-06, "loss": 0.5364, "step": 5415 }, { "epoch": 1.7329351535836177, "grad_norm": 0.5379970669746399, "learning_rate": 4.51761570609448e-06, "loss": 0.4852, "step": 5416 }, { "epoch": 1.733255119453925, "grad_norm": 0.5320088267326355, "learning_rate": 4.515762629563745e-06, "loss": 0.4876, "step": 5417 }, { "epoch": 1.7335750853242322, "grad_norm": 0.5413721203804016, "learning_rate": 4.513909620173103e-06, "loss": 0.4788, "step": 5418 }, { "epoch": 1.7338950511945392, "grad_norm": 0.5124156475067139, "learning_rate": 4.512056678179477e-06, "loss": 0.4605, "step": 5419 }, { "epoch": 1.7342150170648463, "grad_norm": 0.5723578333854675, "learning_rate": 4.510203803839779e-06, "loss": 0.5124, "step": 5420 }, { "epoch": 1.7345349829351537, "grad_norm": 0.528305172920227, "learning_rate": 4.508350997410913e-06, "loss": 0.4558, "step": 5421 }, { "epoch": 1.7348549488054608, "grad_norm": 0.4989233613014221, "learning_rate": 4.506498259149774e-06, "loss": 0.4747, "step": 5422 }, { "epoch": 1.7351749146757678, "grad_norm": 0.5517019629478455, "learning_rate": 4.504645589313243e-06, "loss": 0.4991, "step": 5423 }, { "epoch": 1.735494880546075, "grad_norm": 0.5232354402542114, "learning_rate": 4.502792988158197e-06, "loss": 0.5079, "step": 5424 }, { "epoch": 1.7358148464163823, "grad_norm": 0.5111263990402222, "learning_rate": 4.500940455941504e-06, "loss": 0.4625, "step": 5425 }, { "epoch": 1.7361348122866893, "grad_norm": 0.4850997030735016, "learning_rate": 4.4990879929200145e-06, "loss": 0.4769, "step": 5426 }, { "epoch": 1.7364547781569966, "grad_norm": 0.5141841769218445, "learning_rate": 4.497235599350579e-06, "loss": 0.5197, "step": 5427 }, { "epoch": 1.7367747440273038, "grad_norm": 0.5438143610954285, "learning_rate": 4.495383275490032e-06, "loss": 0.4813, "step": 5428 }, { "epoch": 1.7370947098976108, "grad_norm": 0.5233721733093262, "learning_rate": 4.4935310215952e-06, "loss": 0.4919, "step": 5429 }, { "epoch": 1.737414675767918, "grad_norm": 0.502494752407074, "learning_rate": 4.4916788379229024e-06, "loss": 0.4678, "step": 5430 }, { "epoch": 1.7377346416382253, "grad_norm": 0.49350664019584656, "learning_rate": 4.489826724729946e-06, "loss": 0.5335, "step": 5431 }, { "epoch": 1.7380546075085324, "grad_norm": 0.49686992168426514, "learning_rate": 4.487974682273128e-06, "loss": 0.4986, "step": 5432 }, { "epoch": 1.7383745733788396, "grad_norm": 0.5533159971237183, "learning_rate": 4.486122710809234e-06, "loss": 0.513, "step": 5433 }, { "epoch": 1.7386945392491469, "grad_norm": 0.5231044292449951, "learning_rate": 4.484270810595044e-06, "loss": 0.4997, "step": 5434 }, { "epoch": 1.7390145051194539, "grad_norm": 0.520380973815918, "learning_rate": 4.4824189818873295e-06, "loss": 0.4561, "step": 5435 }, { "epoch": 1.7393344709897611, "grad_norm": 0.4811047315597534, "learning_rate": 4.480567224942845e-06, "loss": 0.4668, "step": 5436 }, { "epoch": 1.7396544368600684, "grad_norm": 0.5269007086753845, "learning_rate": 4.47871554001834e-06, "loss": 0.5144, "step": 5437 }, { "epoch": 1.7399744027303754, "grad_norm": 0.503526508808136, "learning_rate": 4.476863927370551e-06, "loss": 0.4572, "step": 5438 }, { "epoch": 1.7402943686006824, "grad_norm": 0.4947238266468048, "learning_rate": 4.475012387256208e-06, "loss": 0.5347, "step": 5439 }, { "epoch": 1.74061433447099, "grad_norm": 0.5225102305412292, "learning_rate": 4.473160919932029e-06, "loss": 0.4947, "step": 5440 }, { "epoch": 1.740934300341297, "grad_norm": 0.568291425704956, "learning_rate": 4.471309525654723e-06, "loss": 0.5045, "step": 5441 }, { "epoch": 1.741254266211604, "grad_norm": 0.4956567585468292, "learning_rate": 4.469458204680988e-06, "loss": 0.4392, "step": 5442 }, { "epoch": 1.7415742320819114, "grad_norm": 0.5219036340713501, "learning_rate": 4.46760695726751e-06, "loss": 0.4645, "step": 5443 }, { "epoch": 1.7418941979522184, "grad_norm": 0.5405728220939636, "learning_rate": 4.465755783670968e-06, "loss": 0.5228, "step": 5444 }, { "epoch": 1.7422141638225255, "grad_norm": 0.5067600011825562, "learning_rate": 4.4639046841480285e-06, "loss": 0.5074, "step": 5445 }, { "epoch": 1.7425341296928327, "grad_norm": 0.5223138332366943, "learning_rate": 4.462053658955352e-06, "loss": 0.4536, "step": 5446 }, { "epoch": 1.74285409556314, "grad_norm": 0.5472766757011414, "learning_rate": 4.460202708349582e-06, "loss": 0.4719, "step": 5447 }, { "epoch": 1.743174061433447, "grad_norm": 0.5742120742797852, "learning_rate": 4.458351832587354e-06, "loss": 0.4926, "step": 5448 }, { "epoch": 1.7434940273037542, "grad_norm": 0.548699140548706, "learning_rate": 4.456501031925297e-06, "loss": 0.5002, "step": 5449 }, { "epoch": 1.7438139931740615, "grad_norm": 0.4848234951496124, "learning_rate": 4.454650306620025e-06, "loss": 0.4795, "step": 5450 }, { "epoch": 1.7441339590443685, "grad_norm": 0.4963657855987549, "learning_rate": 4.452799656928146e-06, "loss": 0.5057, "step": 5451 }, { "epoch": 1.7444539249146758, "grad_norm": 0.49005699157714844, "learning_rate": 4.450949083106253e-06, "loss": 0.4562, "step": 5452 }, { "epoch": 1.744773890784983, "grad_norm": 0.5466716289520264, "learning_rate": 4.4490985854109285e-06, "loss": 0.5327, "step": 5453 }, { "epoch": 1.74509385665529, "grad_norm": 0.49012893438339233, "learning_rate": 4.447248164098747e-06, "loss": 0.4744, "step": 5454 }, { "epoch": 1.7454138225255973, "grad_norm": 0.4600718319416046, "learning_rate": 4.445397819426274e-06, "loss": 0.4915, "step": 5455 }, { "epoch": 1.7457337883959045, "grad_norm": 0.5044729113578796, "learning_rate": 4.4435475516500615e-06, "loss": 0.5106, "step": 5456 }, { "epoch": 1.7460537542662116, "grad_norm": 0.5379459857940674, "learning_rate": 4.441697361026649e-06, "loss": 0.5507, "step": 5457 }, { "epoch": 1.7463737201365188, "grad_norm": 0.5367534756660461, "learning_rate": 4.439847247812568e-06, "loss": 0.444, "step": 5458 }, { "epoch": 1.746693686006826, "grad_norm": 0.5278060436248779, "learning_rate": 4.437997212264343e-06, "loss": 0.5187, "step": 5459 }, { "epoch": 1.747013651877133, "grad_norm": 0.6021057367324829, "learning_rate": 4.4361472546384785e-06, "loss": 0.5194, "step": 5460 }, { "epoch": 1.7473336177474401, "grad_norm": 0.5518205761909485, "learning_rate": 4.434297375191477e-06, "loss": 0.4924, "step": 5461 }, { "epoch": 1.7476535836177476, "grad_norm": 0.5334278345108032, "learning_rate": 4.4324475741798235e-06, "loss": 0.4701, "step": 5462 }, { "epoch": 1.7479735494880546, "grad_norm": 0.538500964641571, "learning_rate": 4.430597851859996e-06, "loss": 0.47, "step": 5463 }, { "epoch": 1.7482935153583616, "grad_norm": 0.5172466039657593, "learning_rate": 4.428748208488464e-06, "loss": 0.4833, "step": 5464 }, { "epoch": 1.7486134812286689, "grad_norm": 0.4630873501300812, "learning_rate": 4.426898644321677e-06, "loss": 0.4668, "step": 5465 }, { "epoch": 1.7489334470989761, "grad_norm": 0.5098137259483337, "learning_rate": 4.425049159616085e-06, "loss": 0.4796, "step": 5466 }, { "epoch": 1.7492534129692832, "grad_norm": 0.5252134799957275, "learning_rate": 4.423199754628118e-06, "loss": 0.5101, "step": 5467 }, { "epoch": 1.7495733788395904, "grad_norm": 0.48943209648132324, "learning_rate": 4.421350429614197e-06, "loss": 0.4858, "step": 5468 }, { "epoch": 1.7498933447098977, "grad_norm": 0.4845833480358124, "learning_rate": 4.4195011848307376e-06, "loss": 0.4504, "step": 5469 }, { "epoch": 1.7502133105802047, "grad_norm": 0.5351172089576721, "learning_rate": 4.4176520205341365e-06, "loss": 0.542, "step": 5470 }, { "epoch": 1.750533276450512, "grad_norm": 0.496385395526886, "learning_rate": 4.415802936980784e-06, "loss": 0.4326, "step": 5471 }, { "epoch": 1.7508532423208192, "grad_norm": 0.5513273477554321, "learning_rate": 4.413953934427057e-06, "loss": 0.5309, "step": 5472 }, { "epoch": 1.7511732081911262, "grad_norm": 0.48049625754356384, "learning_rate": 4.41210501312932e-06, "loss": 0.4664, "step": 5473 }, { "epoch": 1.7514931740614335, "grad_norm": 0.5158618688583374, "learning_rate": 4.410256173343932e-06, "loss": 0.5327, "step": 5474 }, { "epoch": 1.7518131399317407, "grad_norm": 0.4715256094932556, "learning_rate": 4.4084074153272346e-06, "loss": 0.4593, "step": 5475 }, { "epoch": 1.7521331058020477, "grad_norm": 0.5609480738639832, "learning_rate": 4.406558739335561e-06, "loss": 0.5126, "step": 5476 }, { "epoch": 1.752453071672355, "grad_norm": 0.5760226249694824, "learning_rate": 4.4047101456252325e-06, "loss": 0.5278, "step": 5477 }, { "epoch": 1.7527730375426622, "grad_norm": 0.5125241875648499, "learning_rate": 4.402861634452556e-06, "loss": 0.4898, "step": 5478 }, { "epoch": 1.7530930034129693, "grad_norm": 0.5064290165901184, "learning_rate": 4.4010132060738356e-06, "loss": 0.4935, "step": 5479 }, { "epoch": 1.7534129692832765, "grad_norm": 0.49492013454437256, "learning_rate": 4.399164860745353e-06, "loss": 0.5455, "step": 5480 }, { "epoch": 1.7537329351535837, "grad_norm": 0.4533080756664276, "learning_rate": 4.397316598723385e-06, "loss": 0.3917, "step": 5481 }, { "epoch": 1.7540529010238908, "grad_norm": 0.48983830213546753, "learning_rate": 4.395468420264197e-06, "loss": 0.5256, "step": 5482 }, { "epoch": 1.7543728668941978, "grad_norm": 0.4834545850753784, "learning_rate": 4.393620325624038e-06, "loss": 0.4618, "step": 5483 }, { "epoch": 1.7546928327645053, "grad_norm": 0.48400387167930603, "learning_rate": 4.391772315059152e-06, "loss": 0.4857, "step": 5484 }, { "epoch": 1.7550127986348123, "grad_norm": 0.4745435416698456, "learning_rate": 4.389924388825766e-06, "loss": 0.5117, "step": 5485 }, { "epoch": 1.7553327645051193, "grad_norm": 0.48883339762687683, "learning_rate": 4.388076547180097e-06, "loss": 0.4828, "step": 5486 }, { "epoch": 1.7556527303754266, "grad_norm": 0.5395194888114929, "learning_rate": 4.386228790378351e-06, "loss": 0.4743, "step": 5487 }, { "epoch": 1.7559726962457338, "grad_norm": 0.5261681079864502, "learning_rate": 4.38438111867672e-06, "loss": 0.4712, "step": 5488 }, { "epoch": 1.7562926621160408, "grad_norm": 0.5287356972694397, "learning_rate": 4.38253353233139e-06, "loss": 0.5399, "step": 5489 }, { "epoch": 1.756612627986348, "grad_norm": 0.5015385150909424, "learning_rate": 4.380686031598527e-06, "loss": 0.464, "step": 5490 }, { "epoch": 1.7569325938566553, "grad_norm": 0.49232450127601624, "learning_rate": 4.378838616734292e-06, "loss": 0.4641, "step": 5491 }, { "epoch": 1.7572525597269624, "grad_norm": 0.49756553769111633, "learning_rate": 4.376991287994829e-06, "loss": 0.4978, "step": 5492 }, { "epoch": 1.7575725255972696, "grad_norm": 0.48109307885169983, "learning_rate": 4.375144045636273e-06, "loss": 0.4733, "step": 5493 }, { "epoch": 1.7578924914675769, "grad_norm": 0.4919877052307129, "learning_rate": 4.373296889914748e-06, "loss": 0.5169, "step": 5494 }, { "epoch": 1.758212457337884, "grad_norm": 0.46694469451904297, "learning_rate": 4.371449821086363e-06, "loss": 0.4253, "step": 5495 }, { "epoch": 1.7585324232081911, "grad_norm": 0.5384265184402466, "learning_rate": 4.369602839407217e-06, "loss": 0.4927, "step": 5496 }, { "epoch": 1.7588523890784984, "grad_norm": 0.4768705666065216, "learning_rate": 4.367755945133394e-06, "loss": 0.4495, "step": 5497 }, { "epoch": 1.7591723549488054, "grad_norm": 0.6496914625167847, "learning_rate": 4.36590913852097e-06, "loss": 0.5469, "step": 5498 }, { "epoch": 1.7594923208191127, "grad_norm": 0.47362634539604187, "learning_rate": 4.364062419826008e-06, "loss": 0.4203, "step": 5499 }, { "epoch": 1.75981228668942, "grad_norm": 0.5249969959259033, "learning_rate": 4.3622157893045565e-06, "loss": 0.5064, "step": 5500 }, { "epoch": 1.760132252559727, "grad_norm": 0.4841233789920807, "learning_rate": 4.360369247212653e-06, "loss": 0.4478, "step": 5501 }, { "epoch": 1.760452218430034, "grad_norm": 0.5113783478736877, "learning_rate": 4.358522793806322e-06, "loss": 0.5385, "step": 5502 }, { "epoch": 1.7607721843003414, "grad_norm": 0.4811757504940033, "learning_rate": 4.356676429341577e-06, "loss": 0.4544, "step": 5503 }, { "epoch": 1.7610921501706485, "grad_norm": 0.48280641436576843, "learning_rate": 4.354830154074421e-06, "loss": 0.4919, "step": 5504 }, { "epoch": 1.7614121160409555, "grad_norm": 0.4716619849205017, "learning_rate": 4.352983968260839e-06, "loss": 0.4455, "step": 5505 }, { "epoch": 1.761732081911263, "grad_norm": 0.5332163572311401, "learning_rate": 4.351137872156809e-06, "loss": 0.513, "step": 5506 }, { "epoch": 1.76205204778157, "grad_norm": 0.526752769947052, "learning_rate": 4.349291866018292e-06, "loss": 0.4645, "step": 5507 }, { "epoch": 1.762372013651877, "grad_norm": 0.5340533256530762, "learning_rate": 4.347445950101241e-06, "loss": 0.5284, "step": 5508 }, { "epoch": 1.7626919795221843, "grad_norm": 0.4629971385002136, "learning_rate": 4.345600124661596e-06, "loss": 0.4387, "step": 5509 }, { "epoch": 1.7630119453924915, "grad_norm": 0.4559417963027954, "learning_rate": 4.343754389955279e-06, "loss": 0.5109, "step": 5510 }, { "epoch": 1.7633319112627985, "grad_norm": 0.49508291482925415, "learning_rate": 4.341908746238207e-06, "loss": 0.4781, "step": 5511 }, { "epoch": 1.7636518771331058, "grad_norm": 0.4971843361854553, "learning_rate": 4.340063193766277e-06, "loss": 0.4451, "step": 5512 }, { "epoch": 1.763971843003413, "grad_norm": 0.5565189719200134, "learning_rate": 4.33821773279538e-06, "loss": 0.5169, "step": 5513 }, { "epoch": 1.76429180887372, "grad_norm": 0.47722023725509644, "learning_rate": 4.336372363581391e-06, "loss": 0.4604, "step": 5514 }, { "epoch": 1.7646117747440273, "grad_norm": 0.504019021987915, "learning_rate": 4.334527086380171e-06, "loss": 0.4904, "step": 5515 }, { "epoch": 1.7649317406143346, "grad_norm": 0.5100414156913757, "learning_rate": 4.3326819014475726e-06, "loss": 0.4528, "step": 5516 }, { "epoch": 1.7652517064846416, "grad_norm": 0.55582195520401, "learning_rate": 4.3308368090394295e-06, "loss": 0.5074, "step": 5517 }, { "epoch": 1.7655716723549488, "grad_norm": 0.530032753944397, "learning_rate": 4.328991809411569e-06, "loss": 0.4683, "step": 5518 }, { "epoch": 1.765891638225256, "grad_norm": 0.5319012403488159, "learning_rate": 4.327146902819802e-06, "loss": 0.4514, "step": 5519 }, { "epoch": 1.766211604095563, "grad_norm": 0.5432710647583008, "learning_rate": 4.325302089519926e-06, "loss": 0.4615, "step": 5520 }, { "epoch": 1.7665315699658704, "grad_norm": 0.5205046534538269, "learning_rate": 4.323457369767727e-06, "loss": 0.479, "step": 5521 }, { "epoch": 1.7668515358361776, "grad_norm": 0.6415116786956787, "learning_rate": 4.321612743818976e-06, "loss": 0.5481, "step": 5522 }, { "epoch": 1.7671715017064846, "grad_norm": 0.5402652621269226, "learning_rate": 4.319768211929435e-06, "loss": 0.4699, "step": 5523 }, { "epoch": 1.7674914675767917, "grad_norm": 0.5080769062042236, "learning_rate": 4.3179237743548516e-06, "loss": 0.5002, "step": 5524 }, { "epoch": 1.7678114334470991, "grad_norm": 0.49535703659057617, "learning_rate": 4.316079431350956e-06, "loss": 0.4754, "step": 5525 }, { "epoch": 1.7681313993174061, "grad_norm": 0.527343213558197, "learning_rate": 4.31423518317347e-06, "loss": 0.4869, "step": 5526 }, { "epoch": 1.7684513651877132, "grad_norm": 0.48264944553375244, "learning_rate": 4.3123910300781e-06, "loss": 0.4852, "step": 5527 }, { "epoch": 1.7687713310580204, "grad_norm": 0.5463948845863342, "learning_rate": 4.310546972320542e-06, "loss": 0.4725, "step": 5528 }, { "epoch": 1.7690912969283277, "grad_norm": 0.598416805267334, "learning_rate": 4.308703010156476e-06, "loss": 0.5052, "step": 5529 }, { "epoch": 1.7694112627986347, "grad_norm": 0.5130217671394348, "learning_rate": 4.306859143841569e-06, "loss": 0.4952, "step": 5530 }, { "epoch": 1.769731228668942, "grad_norm": 0.5175359845161438, "learning_rate": 4.305015373631474e-06, "loss": 0.5155, "step": 5531 }, { "epoch": 1.7700511945392492, "grad_norm": 0.5057743787765503, "learning_rate": 4.303171699781835e-06, "loss": 0.451, "step": 5532 }, { "epoch": 1.7703711604095562, "grad_norm": 0.6024937629699707, "learning_rate": 4.301328122548278e-06, "loss": 0.447, "step": 5533 }, { "epoch": 1.7706911262798635, "grad_norm": 0.5402462482452393, "learning_rate": 4.299484642186418e-06, "loss": 0.5155, "step": 5534 }, { "epoch": 1.7710110921501707, "grad_norm": 0.5450391173362732, "learning_rate": 4.297641258951855e-06, "loss": 0.5274, "step": 5535 }, { "epoch": 1.7713310580204777, "grad_norm": 0.52930748462677, "learning_rate": 4.295797973100174e-06, "loss": 0.4358, "step": 5536 }, { "epoch": 1.771651023890785, "grad_norm": 0.5490062236785889, "learning_rate": 4.293954784886954e-06, "loss": 0.52, "step": 5537 }, { "epoch": 1.7719709897610922, "grad_norm": 0.5440360307693481, "learning_rate": 4.292111694567751e-06, "loss": 0.4968, "step": 5538 }, { "epoch": 1.7722909556313993, "grad_norm": 0.5466500520706177, "learning_rate": 4.290268702398115e-06, "loss": 0.446, "step": 5539 }, { "epoch": 1.7726109215017065, "grad_norm": 0.49099576473236084, "learning_rate": 4.2884258086335755e-06, "loss": 0.5082, "step": 5540 }, { "epoch": 1.7729308873720138, "grad_norm": 0.5124669075012207, "learning_rate": 4.286583013529653e-06, "loss": 0.4779, "step": 5541 }, { "epoch": 1.7732508532423208, "grad_norm": 0.5103096961975098, "learning_rate": 4.284740317341856e-06, "loss": 0.4809, "step": 5542 }, { "epoch": 1.773570819112628, "grad_norm": 0.5411840081214905, "learning_rate": 4.282897720325673e-06, "loss": 0.4795, "step": 5543 }, { "epoch": 1.7738907849829353, "grad_norm": 0.540686309337616, "learning_rate": 4.281055222736586e-06, "loss": 0.4575, "step": 5544 }, { "epoch": 1.7742107508532423, "grad_norm": 0.5696617960929871, "learning_rate": 4.2792128248300546e-06, "loss": 0.5115, "step": 5545 }, { "epoch": 1.7745307167235493, "grad_norm": 0.5039750337600708, "learning_rate": 4.2773705268615315e-06, "loss": 0.4749, "step": 5546 }, { "epoch": 1.7748506825938568, "grad_norm": 0.522910475730896, "learning_rate": 4.275528329086457e-06, "loss": 0.4636, "step": 5547 }, { "epoch": 1.7751706484641638, "grad_norm": 0.5602678060531616, "learning_rate": 4.273686231760249e-06, "loss": 0.5559, "step": 5548 }, { "epoch": 1.7754906143344709, "grad_norm": 0.5650404691696167, "learning_rate": 4.271844235138321e-06, "loss": 0.5355, "step": 5549 }, { "epoch": 1.775810580204778, "grad_norm": 0.4923083782196045, "learning_rate": 4.270002339476065e-06, "loss": 0.4698, "step": 5550 }, { "epoch": 1.7761305460750854, "grad_norm": 0.5023771524429321, "learning_rate": 4.268160545028861e-06, "loss": 0.4528, "step": 5551 }, { "epoch": 1.7764505119453924, "grad_norm": 0.529879629611969, "learning_rate": 4.26631885205208e-06, "loss": 0.5279, "step": 5552 }, { "epoch": 1.7767704778156996, "grad_norm": 0.4917384386062622, "learning_rate": 4.264477260801072e-06, "loss": 0.4707, "step": 5553 }, { "epoch": 1.7770904436860069, "grad_norm": 0.5691450834274292, "learning_rate": 4.262635771531178e-06, "loss": 0.5206, "step": 5554 }, { "epoch": 1.777410409556314, "grad_norm": 0.5118280649185181, "learning_rate": 4.26079438449772e-06, "loss": 0.4956, "step": 5555 }, { "epoch": 1.7777303754266212, "grad_norm": 0.49388378858566284, "learning_rate": 4.258953099956008e-06, "loss": 0.4807, "step": 5556 }, { "epoch": 1.7780503412969284, "grad_norm": 0.4365639388561249, "learning_rate": 4.257111918161343e-06, "loss": 0.4138, "step": 5557 }, { "epoch": 1.7783703071672354, "grad_norm": 0.5118427872657776, "learning_rate": 4.2552708393690035e-06, "loss": 0.5282, "step": 5558 }, { "epoch": 1.7786902730375427, "grad_norm": 0.5046088695526123, "learning_rate": 4.253429863834258e-06, "loss": 0.5192, "step": 5559 }, { "epoch": 1.77901023890785, "grad_norm": 0.5091738700866699, "learning_rate": 4.25158899181236e-06, "loss": 0.5138, "step": 5560 }, { "epoch": 1.779330204778157, "grad_norm": 12.748663902282715, "learning_rate": 4.249748223558547e-06, "loss": 0.4696, "step": 5561 }, { "epoch": 1.7796501706484642, "grad_norm": 0.5930684208869934, "learning_rate": 4.247907559328049e-06, "loss": 0.5821, "step": 5562 }, { "epoch": 1.7799701365187715, "grad_norm": 0.5774797201156616, "learning_rate": 4.24606699937607e-06, "loss": 0.5031, "step": 5563 }, { "epoch": 1.7802901023890785, "grad_norm": 0.5613293051719666, "learning_rate": 4.24422654395781e-06, "loss": 0.4981, "step": 5564 }, { "epoch": 1.7806100682593855, "grad_norm": 0.5332891345024109, "learning_rate": 4.242386193328447e-06, "loss": 0.4651, "step": 5565 }, { "epoch": 1.780930034129693, "grad_norm": 0.564367949962616, "learning_rate": 4.2405459477431505e-06, "loss": 0.5202, "step": 5566 }, { "epoch": 1.78125, "grad_norm": 0.5317816734313965, "learning_rate": 4.238705807457073e-06, "loss": 0.4937, "step": 5567 }, { "epoch": 1.781569965870307, "grad_norm": 0.46839404106140137, "learning_rate": 4.236865772725351e-06, "loss": 0.4636, "step": 5568 }, { "epoch": 1.7818899317406145, "grad_norm": 0.45433491468429565, "learning_rate": 4.2350258438031075e-06, "loss": 0.4566, "step": 5569 }, { "epoch": 1.7822098976109215, "grad_norm": 0.5349555611610413, "learning_rate": 4.23318602094545e-06, "loss": 0.4865, "step": 5570 }, { "epoch": 1.7825298634812285, "grad_norm": 0.5639216303825378, "learning_rate": 4.2313463044074735e-06, "loss": 0.4934, "step": 5571 }, { "epoch": 1.7828498293515358, "grad_norm": 0.5030721426010132, "learning_rate": 4.229506694444258e-06, "loss": 0.471, "step": 5572 }, { "epoch": 1.783169795221843, "grad_norm": 0.5446768999099731, "learning_rate": 4.227667191310865e-06, "loss": 0.5168, "step": 5573 }, { "epoch": 1.78348976109215, "grad_norm": 0.5222355723381042, "learning_rate": 4.225827795262346e-06, "loss": 0.4883, "step": 5574 }, { "epoch": 1.7838097269624573, "grad_norm": 0.5582959055900574, "learning_rate": 4.223988506553732e-06, "loss": 0.4787, "step": 5575 }, { "epoch": 1.7841296928327646, "grad_norm": 0.5071340799331665, "learning_rate": 4.222149325440048e-06, "loss": 0.4875, "step": 5576 }, { "epoch": 1.7844496587030716, "grad_norm": 0.5143452882766724, "learning_rate": 4.220310252176295e-06, "loss": 0.4675, "step": 5577 }, { "epoch": 1.7847696245733788, "grad_norm": 0.4680158793926239, "learning_rate": 4.218471287017463e-06, "loss": 0.5007, "step": 5578 }, { "epoch": 1.785089590443686, "grad_norm": 0.5283663868904114, "learning_rate": 4.216632430218528e-06, "loss": 0.5296, "step": 5579 }, { "epoch": 1.7854095563139931, "grad_norm": 0.4833642244338989, "learning_rate": 4.2147936820344484e-06, "loss": 0.4334, "step": 5580 }, { "epoch": 1.7857295221843004, "grad_norm": 0.5667260885238647, "learning_rate": 4.212955042720169e-06, "loss": 0.5489, "step": 5581 }, { "epoch": 1.7860494880546076, "grad_norm": 0.5367039442062378, "learning_rate": 4.211116512530621e-06, "loss": 0.4817, "step": 5582 }, { "epoch": 1.7863694539249146, "grad_norm": 0.5276321172714233, "learning_rate": 4.2092780917207165e-06, "loss": 0.4729, "step": 5583 }, { "epoch": 1.786689419795222, "grad_norm": 0.4945152997970581, "learning_rate": 4.2074397805453575e-06, "loss": 0.4959, "step": 5584 }, { "epoch": 1.7870093856655291, "grad_norm": 0.5138298273086548, "learning_rate": 4.205601579259424e-06, "loss": 0.4802, "step": 5585 }, { "epoch": 1.7873293515358362, "grad_norm": 0.5320633053779602, "learning_rate": 4.203763488117788e-06, "loss": 0.4885, "step": 5586 }, { "epoch": 1.7876493174061432, "grad_norm": 0.5252824425697327, "learning_rate": 4.201925507375303e-06, "loss": 0.5041, "step": 5587 }, { "epoch": 1.7879692832764507, "grad_norm": 0.5192841291427612, "learning_rate": 4.200087637286807e-06, "loss": 0.5156, "step": 5588 }, { "epoch": 1.7882892491467577, "grad_norm": 0.5447317957878113, "learning_rate": 4.198249878107122e-06, "loss": 0.4994, "step": 5589 }, { "epoch": 1.7886092150170647, "grad_norm": 0.48571932315826416, "learning_rate": 4.196412230091053e-06, "loss": 0.4869, "step": 5590 }, { "epoch": 1.788929180887372, "grad_norm": 0.5020157098770142, "learning_rate": 4.194574693493398e-06, "loss": 0.4807, "step": 5591 }, { "epoch": 1.7892491467576792, "grad_norm": 0.5085739493370056, "learning_rate": 4.1927372685689315e-06, "loss": 0.4823, "step": 5592 }, { "epoch": 1.7895691126279862, "grad_norm": 0.567353367805481, "learning_rate": 4.190899955572413e-06, "loss": 0.5069, "step": 5593 }, { "epoch": 1.7898890784982935, "grad_norm": 0.5270597338676453, "learning_rate": 4.18906275475859e-06, "loss": 0.5283, "step": 5594 }, { "epoch": 1.7902090443686007, "grad_norm": 0.46188998222351074, "learning_rate": 4.187225666382187e-06, "loss": 0.4108, "step": 5595 }, { "epoch": 1.7905290102389078, "grad_norm": 0.607689380645752, "learning_rate": 4.185388690697927e-06, "loss": 0.5774, "step": 5596 }, { "epoch": 1.790848976109215, "grad_norm": 0.615105926990509, "learning_rate": 4.183551827960505e-06, "loss": 0.4953, "step": 5597 }, { "epoch": 1.7911689419795223, "grad_norm": 0.5350819230079651, "learning_rate": 4.181715078424604e-06, "loss": 0.5159, "step": 5598 }, { "epoch": 1.7914889078498293, "grad_norm": 0.5074175000190735, "learning_rate": 4.179878442344892e-06, "loss": 0.4562, "step": 5599 }, { "epoch": 1.7918088737201365, "grad_norm": 0.49207979440689087, "learning_rate": 4.178041919976018e-06, "loss": 0.4636, "step": 5600 }, { "epoch": 1.7921288395904438, "grad_norm": 0.523507297039032, "learning_rate": 4.176205511572624e-06, "loss": 0.5079, "step": 5601 }, { "epoch": 1.7924488054607508, "grad_norm": 0.5207238793373108, "learning_rate": 4.174369217389326e-06, "loss": 0.4361, "step": 5602 }, { "epoch": 1.792768771331058, "grad_norm": 0.5948736667633057, "learning_rate": 4.17253303768073e-06, "loss": 0.5183, "step": 5603 }, { "epoch": 1.7930887372013653, "grad_norm": 0.5301239490509033, "learning_rate": 4.1706969727014205e-06, "loss": 0.5135, "step": 5604 }, { "epoch": 1.7934087030716723, "grad_norm": 0.5359353423118591, "learning_rate": 4.168861022705976e-06, "loss": 0.4835, "step": 5605 }, { "epoch": 1.7937286689419796, "grad_norm": 0.5441157817840576, "learning_rate": 4.1670251879489506e-06, "loss": 0.5016, "step": 5606 }, { "epoch": 1.7940486348122868, "grad_norm": 0.5022502541542053, "learning_rate": 4.1651894686848844e-06, "loss": 0.5005, "step": 5607 }, { "epoch": 1.7943686006825939, "grad_norm": 0.5129440426826477, "learning_rate": 4.163353865168303e-06, "loss": 0.5304, "step": 5608 }, { "epoch": 1.7946885665529009, "grad_norm": 0.46050283312797546, "learning_rate": 4.161518377653714e-06, "loss": 0.4459, "step": 5609 }, { "epoch": 1.7950085324232083, "grad_norm": 0.5066482424736023, "learning_rate": 4.159683006395611e-06, "loss": 0.4895, "step": 5610 }, { "epoch": 1.7953284982935154, "grad_norm": 0.5382094979286194, "learning_rate": 4.157847751648471e-06, "loss": 0.5432, "step": 5611 }, { "epoch": 1.7956484641638224, "grad_norm": 0.4516243636608124, "learning_rate": 4.156012613666753e-06, "loss": 0.4866, "step": 5612 }, { "epoch": 1.7959684300341296, "grad_norm": 0.5349157452583313, "learning_rate": 4.154177592704902e-06, "loss": 0.5145, "step": 5613 }, { "epoch": 1.796288395904437, "grad_norm": 0.4918440580368042, "learning_rate": 4.152342689017342e-06, "loss": 0.4627, "step": 5614 }, { "epoch": 1.796608361774744, "grad_norm": 0.5058891177177429, "learning_rate": 4.15050790285849e-06, "loss": 0.4573, "step": 5615 }, { "epoch": 1.7969283276450512, "grad_norm": 0.5324491262435913, "learning_rate": 4.148673234482741e-06, "loss": 0.4959, "step": 5616 }, { "epoch": 1.7972482935153584, "grad_norm": 0.5199218988418579, "learning_rate": 4.146838684144471e-06, "loss": 0.5029, "step": 5617 }, { "epoch": 1.7975682593856654, "grad_norm": 0.5540761947631836, "learning_rate": 4.145004252098044e-06, "loss": 0.5378, "step": 5618 }, { "epoch": 1.7978882252559727, "grad_norm": 0.4679339528083801, "learning_rate": 4.143169938597805e-06, "loss": 0.4652, "step": 5619 }, { "epoch": 1.79820819112628, "grad_norm": 0.5310577154159546, "learning_rate": 4.141335743898085e-06, "loss": 0.4744, "step": 5620 }, { "epoch": 1.798528156996587, "grad_norm": 0.502953290939331, "learning_rate": 4.139501668253199e-06, "loss": 0.468, "step": 5621 }, { "epoch": 1.7988481228668942, "grad_norm": 0.4934243857860565, "learning_rate": 4.13766771191744e-06, "loss": 0.5008, "step": 5622 }, { "epoch": 1.7991680887372015, "grad_norm": 0.4875839054584503, "learning_rate": 4.135833875145091e-06, "loss": 0.5144, "step": 5623 }, { "epoch": 1.7994880546075085, "grad_norm": 0.4996601343154907, "learning_rate": 4.1340001581904135e-06, "loss": 0.463, "step": 5624 }, { "epoch": 1.7998080204778157, "grad_norm": 0.49659842252731323, "learning_rate": 4.132166561307657e-06, "loss": 0.4526, "step": 5625 }, { "epoch": 1.800127986348123, "grad_norm": 0.5389164686203003, "learning_rate": 4.130333084751051e-06, "loss": 0.4994, "step": 5626 }, { "epoch": 1.80044795221843, "grad_norm": 0.5688105821609497, "learning_rate": 4.128499728774808e-06, "loss": 0.5709, "step": 5627 }, { "epoch": 1.800767918088737, "grad_norm": 0.5026649236679077, "learning_rate": 4.126666493633128e-06, "loss": 0.4336, "step": 5628 }, { "epoch": 1.8010878839590445, "grad_norm": 0.612531304359436, "learning_rate": 4.124833379580186e-06, "loss": 0.5483, "step": 5629 }, { "epoch": 1.8014078498293515, "grad_norm": 0.554692804813385, "learning_rate": 4.1230003868701515e-06, "loss": 0.4851, "step": 5630 }, { "epoch": 1.8017278156996586, "grad_norm": 0.5775287747383118, "learning_rate": 4.121167515757168e-06, "loss": 0.5384, "step": 5631 }, { "epoch": 1.802047781569966, "grad_norm": 0.46787363290786743, "learning_rate": 4.119334766495366e-06, "loss": 0.4737, "step": 5632 }, { "epoch": 1.802367747440273, "grad_norm": 0.4719943404197693, "learning_rate": 4.117502139338858e-06, "loss": 0.4557, "step": 5633 }, { "epoch": 1.80268771331058, "grad_norm": 0.5867651700973511, "learning_rate": 4.1156696345417386e-06, "loss": 0.515, "step": 5634 }, { "epoch": 1.8030076791808873, "grad_norm": 0.47377723455429077, "learning_rate": 4.113837252358089e-06, "loss": 0.4255, "step": 5635 }, { "epoch": 1.8033276450511946, "grad_norm": 0.5013930797576904, "learning_rate": 4.112004993041972e-06, "loss": 0.4913, "step": 5636 }, { "epoch": 1.8036476109215016, "grad_norm": 0.46539056301116943, "learning_rate": 4.110172856847428e-06, "loss": 0.4621, "step": 5637 }, { "epoch": 1.8039675767918089, "grad_norm": 0.5124224424362183, "learning_rate": 4.108340844028489e-06, "loss": 0.5124, "step": 5638 }, { "epoch": 1.804287542662116, "grad_norm": 0.4970622658729553, "learning_rate": 4.106508954839166e-06, "loss": 0.489, "step": 5639 }, { "epoch": 1.8046075085324231, "grad_norm": 0.4849103093147278, "learning_rate": 4.10467718953345e-06, "loss": 0.4426, "step": 5640 }, { "epoch": 1.8049274744027304, "grad_norm": 0.4756026566028595, "learning_rate": 4.102845548365321e-06, "loss": 0.4956, "step": 5641 }, { "epoch": 1.8052474402730376, "grad_norm": 0.42802461981773376, "learning_rate": 4.101014031588734e-06, "loss": 0.4411, "step": 5642 }, { "epoch": 1.8055674061433447, "grad_norm": 0.5087606906890869, "learning_rate": 4.0991826394576315e-06, "loss": 0.4988, "step": 5643 }, { "epoch": 1.805887372013652, "grad_norm": 0.5063179731369019, "learning_rate": 4.097351372225943e-06, "loss": 0.4359, "step": 5644 }, { "epoch": 1.8062073378839592, "grad_norm": 0.5513647198677063, "learning_rate": 4.095520230147571e-06, "loss": 0.5395, "step": 5645 }, { "epoch": 1.8065273037542662, "grad_norm": 0.5584523677825928, "learning_rate": 4.093689213476408e-06, "loss": 0.5034, "step": 5646 }, { "epoch": 1.8068472696245734, "grad_norm": 0.46291977167129517, "learning_rate": 4.091858322466325e-06, "loss": 0.4515, "step": 5647 }, { "epoch": 1.8071672354948807, "grad_norm": 0.5163590312004089, "learning_rate": 4.090027557371177e-06, "loss": 0.4652, "step": 5648 }, { "epoch": 1.8074872013651877, "grad_norm": 0.5587126612663269, "learning_rate": 4.088196918444806e-06, "loss": 0.5111, "step": 5649 }, { "epoch": 1.8078071672354947, "grad_norm": 0.46405214071273804, "learning_rate": 4.086366405941027e-06, "loss": 0.3938, "step": 5650 }, { "epoch": 1.8081271331058022, "grad_norm": 0.48423051834106445, "learning_rate": 4.084536020113646e-06, "loss": 0.462, "step": 5651 }, { "epoch": 1.8084470989761092, "grad_norm": 0.5340949296951294, "learning_rate": 4.0827057612164465e-06, "loss": 0.5212, "step": 5652 }, { "epoch": 1.8087670648464163, "grad_norm": 0.5507909655570984, "learning_rate": 4.080875629503196e-06, "loss": 0.5558, "step": 5653 }, { "epoch": 1.8090870307167235, "grad_norm": 0.5128883123397827, "learning_rate": 4.079045625227646e-06, "loss": 0.4978, "step": 5654 }, { "epoch": 1.8094069965870307, "grad_norm": 0.48712173104286194, "learning_rate": 4.077215748643528e-06, "loss": 0.4804, "step": 5655 }, { "epoch": 1.8097269624573378, "grad_norm": 0.5109673142433167, "learning_rate": 4.075386000004558e-06, "loss": 0.5142, "step": 5656 }, { "epoch": 1.810046928327645, "grad_norm": 0.5423082709312439, "learning_rate": 4.073556379564429e-06, "loss": 0.4911, "step": 5657 }, { "epoch": 1.8103668941979523, "grad_norm": 0.564723551273346, "learning_rate": 4.071726887576823e-06, "loss": 0.5238, "step": 5658 }, { "epoch": 1.8106868600682593, "grad_norm": 0.47076481580734253, "learning_rate": 4.069897524295402e-06, "loss": 0.4368, "step": 5659 }, { "epoch": 1.8110068259385665, "grad_norm": 0.5020716786384583, "learning_rate": 4.068068289973807e-06, "loss": 0.4936, "step": 5660 }, { "epoch": 1.8113267918088738, "grad_norm": 0.5677059292793274, "learning_rate": 4.066239184865666e-06, "loss": 0.5229, "step": 5661 }, { "epoch": 1.8116467576791808, "grad_norm": 0.4805668592453003, "learning_rate": 4.0644102092245845e-06, "loss": 0.45, "step": 5662 }, { "epoch": 1.811966723549488, "grad_norm": 0.5515758395195007, "learning_rate": 4.062581363304152e-06, "loss": 0.5542, "step": 5663 }, { "epoch": 1.8122866894197953, "grad_norm": 0.4808933138847351, "learning_rate": 4.0607526473579434e-06, "loss": 0.4161, "step": 5664 }, { "epoch": 1.8126066552901023, "grad_norm": 0.5447078943252563, "learning_rate": 4.058924061639508e-06, "loss": 0.51, "step": 5665 }, { "epoch": 1.8129266211604096, "grad_norm": 0.5344134569168091, "learning_rate": 4.057095606402387e-06, "loss": 0.473, "step": 5666 }, { "epoch": 1.8132465870307168, "grad_norm": 0.565631628036499, "learning_rate": 4.055267281900091e-06, "loss": 0.5364, "step": 5667 }, { "epoch": 1.8135665529010239, "grad_norm": 0.5139793157577515, "learning_rate": 4.053439088386124e-06, "loss": 0.4596, "step": 5668 }, { "epoch": 1.8138865187713311, "grad_norm": 0.5374452471733093, "learning_rate": 4.051611026113967e-06, "loss": 0.5074, "step": 5669 }, { "epoch": 1.8142064846416384, "grad_norm": 0.5051067471504211, "learning_rate": 4.049783095337081e-06, "loss": 0.5007, "step": 5670 }, { "epoch": 1.8145264505119454, "grad_norm": 0.4644712805747986, "learning_rate": 4.047955296308913e-06, "loss": 0.4286, "step": 5671 }, { "epoch": 1.8148464163822524, "grad_norm": 0.5553407669067383, "learning_rate": 4.046127629282887e-06, "loss": 0.4873, "step": 5672 }, { "epoch": 1.8151663822525599, "grad_norm": 0.566815197467804, "learning_rate": 4.0443000945124135e-06, "loss": 0.5101, "step": 5673 }, { "epoch": 1.815486348122867, "grad_norm": 0.505651593208313, "learning_rate": 4.042472692250883e-06, "loss": 0.4675, "step": 5674 }, { "epoch": 1.815806313993174, "grad_norm": 0.609451413154602, "learning_rate": 4.0406454227516635e-06, "loss": 0.5255, "step": 5675 }, { "epoch": 1.8161262798634812, "grad_norm": 0.5524954795837402, "learning_rate": 4.038818286268113e-06, "loss": 0.5007, "step": 5676 }, { "epoch": 1.8164462457337884, "grad_norm": 0.5468558669090271, "learning_rate": 4.03699128305356e-06, "loss": 0.4939, "step": 5677 }, { "epoch": 1.8167662116040955, "grad_norm": 0.5983974933624268, "learning_rate": 4.035164413361325e-06, "loss": 0.4809, "step": 5678 }, { "epoch": 1.8170861774744027, "grad_norm": 0.4928704500198364, "learning_rate": 4.033337677444708e-06, "loss": 0.5119, "step": 5679 }, { "epoch": 1.81740614334471, "grad_norm": 0.4960091710090637, "learning_rate": 4.031511075556982e-06, "loss": 0.4016, "step": 5680 }, { "epoch": 1.817726109215017, "grad_norm": 0.5329631567001343, "learning_rate": 4.0296846079514114e-06, "loss": 0.5223, "step": 5681 }, { "epoch": 1.8180460750853242, "grad_norm": 0.5038538575172424, "learning_rate": 4.027858274881235e-06, "loss": 0.4667, "step": 5682 }, { "epoch": 1.8183660409556315, "grad_norm": 0.5360050201416016, "learning_rate": 4.026032076599681e-06, "loss": 0.4993, "step": 5683 }, { "epoch": 1.8186860068259385, "grad_norm": 0.4917638599872589, "learning_rate": 4.024206013359952e-06, "loss": 0.4794, "step": 5684 }, { "epoch": 1.8190059726962458, "grad_norm": 0.6024853587150574, "learning_rate": 4.02238008541523e-06, "loss": 0.5482, "step": 5685 }, { "epoch": 1.819325938566553, "grad_norm": 0.5780578255653381, "learning_rate": 4.020554293018688e-06, "loss": 0.4987, "step": 5686 }, { "epoch": 1.81964590443686, "grad_norm": 0.5435405969619751, "learning_rate": 4.018728636423469e-06, "loss": 0.452, "step": 5687 }, { "epoch": 1.8199658703071673, "grad_norm": 0.5363346338272095, "learning_rate": 4.016903115882707e-06, "loss": 0.4646, "step": 5688 }, { "epoch": 1.8202858361774745, "grad_norm": 0.5494429469108582, "learning_rate": 4.015077731649511e-06, "loss": 0.5383, "step": 5689 }, { "epoch": 1.8206058020477816, "grad_norm": 0.5325194597244263, "learning_rate": 4.013252483976972e-06, "loss": 0.5186, "step": 5690 }, { "epoch": 1.8209257679180886, "grad_norm": 0.5208801627159119, "learning_rate": 4.011427373118163e-06, "loss": 0.512, "step": 5691 }, { "epoch": 1.821245733788396, "grad_norm": 0.4978253245353699, "learning_rate": 4.009602399326137e-06, "loss": 0.4225, "step": 5692 }, { "epoch": 1.821565699658703, "grad_norm": 0.48743247985839844, "learning_rate": 4.00777756285393e-06, "loss": 0.4746, "step": 5693 }, { "epoch": 1.82188566552901, "grad_norm": 0.5696543455123901, "learning_rate": 4.005952863954559e-06, "loss": 0.5713, "step": 5694 }, { "epoch": 1.8222056313993176, "grad_norm": 0.4884602725505829, "learning_rate": 4.004128302881018e-06, "loss": 0.469, "step": 5695 }, { "epoch": 1.8225255972696246, "grad_norm": 0.5204737782478333, "learning_rate": 4.002303879886288e-06, "loss": 0.4507, "step": 5696 }, { "epoch": 1.8228455631399316, "grad_norm": 0.49271154403686523, "learning_rate": 4.000479595223322e-06, "loss": 0.4515, "step": 5697 }, { "epoch": 1.8231655290102389, "grad_norm": 0.519041895866394, "learning_rate": 3.998655449145064e-06, "loss": 0.5166, "step": 5698 }, { "epoch": 1.8234854948805461, "grad_norm": 0.5471581816673279, "learning_rate": 3.9968314419044334e-06, "loss": 0.4911, "step": 5699 }, { "epoch": 1.8238054607508531, "grad_norm": 0.6149153113365173, "learning_rate": 3.9950075737543296e-06, "loss": 0.5139, "step": 5700 }, { "epoch": 1.8241254266211604, "grad_norm": 0.5749152302742004, "learning_rate": 3.9931838449476355e-06, "loss": 0.426, "step": 5701 }, { "epoch": 1.8244453924914676, "grad_norm": 0.5508496165275574, "learning_rate": 3.991360255737211e-06, "loss": 0.5323, "step": 5702 }, { "epoch": 1.8247653583617747, "grad_norm": 0.5065162181854248, "learning_rate": 3.989536806375901e-06, "loss": 0.4695, "step": 5703 }, { "epoch": 1.825085324232082, "grad_norm": 0.5152005553245544, "learning_rate": 3.98771349711653e-06, "loss": 0.4943, "step": 5704 }, { "epoch": 1.8254052901023892, "grad_norm": 0.521493136882782, "learning_rate": 3.985890328211899e-06, "loss": 0.4214, "step": 5705 }, { "epoch": 1.8257252559726962, "grad_norm": 0.5740010142326355, "learning_rate": 3.984067299914792e-06, "loss": 0.5061, "step": 5706 }, { "epoch": 1.8260452218430034, "grad_norm": 0.5458466410636902, "learning_rate": 3.982244412477979e-06, "loss": 0.4617, "step": 5707 }, { "epoch": 1.8263651877133107, "grad_norm": 0.5201452374458313, "learning_rate": 3.980421666154202e-06, "loss": 0.4686, "step": 5708 }, { "epoch": 1.8266851535836177, "grad_norm": 0.530581533908844, "learning_rate": 3.978599061196188e-06, "loss": 0.4927, "step": 5709 }, { "epoch": 1.827005119453925, "grad_norm": 0.4961286187171936, "learning_rate": 3.976776597856643e-06, "loss": 0.4274, "step": 5710 }, { "epoch": 1.8273250853242322, "grad_norm": 0.5156742930412292, "learning_rate": 3.974954276388251e-06, "loss": 0.5662, "step": 5711 }, { "epoch": 1.8276450511945392, "grad_norm": 0.5241951942443848, "learning_rate": 3.973132097043685e-06, "loss": 0.4626, "step": 5712 }, { "epoch": 1.8279650170648463, "grad_norm": 0.5877885818481445, "learning_rate": 3.971310060075588e-06, "loss": 0.5165, "step": 5713 }, { "epoch": 1.8282849829351537, "grad_norm": 0.5485692024230957, "learning_rate": 3.969488165736589e-06, "loss": 0.4743, "step": 5714 }, { "epoch": 1.8286049488054608, "grad_norm": 0.598365843296051, "learning_rate": 3.967666414279294e-06, "loss": 0.5029, "step": 5715 }, { "epoch": 1.8289249146757678, "grad_norm": 0.5118568539619446, "learning_rate": 3.9658448059562934e-06, "loss": 0.5022, "step": 5716 }, { "epoch": 1.829244880546075, "grad_norm": 0.5245277881622314, "learning_rate": 3.9640233410201555e-06, "loss": 0.4935, "step": 5717 }, { "epoch": 1.8295648464163823, "grad_norm": 0.49036654829978943, "learning_rate": 3.962202019723427e-06, "loss": 0.4344, "step": 5718 }, { "epoch": 1.8298848122866893, "grad_norm": 0.5216358304023743, "learning_rate": 3.960380842318638e-06, "loss": 0.545, "step": 5719 }, { "epoch": 1.8302047781569966, "grad_norm": 0.5128059387207031, "learning_rate": 3.958559809058294e-06, "loss": 0.4777, "step": 5720 }, { "epoch": 1.8305247440273038, "grad_norm": 0.543674111366272, "learning_rate": 3.956738920194886e-06, "loss": 0.4862, "step": 5721 }, { "epoch": 1.8308447098976108, "grad_norm": 0.5424734950065613, "learning_rate": 3.954918175980882e-06, "loss": 0.4828, "step": 5722 }, { "epoch": 1.831164675767918, "grad_norm": 0.553587019443512, "learning_rate": 3.95309757666873e-06, "loss": 0.5023, "step": 5723 }, { "epoch": 1.8314846416382253, "grad_norm": 0.48538005352020264, "learning_rate": 3.9512771225108595e-06, "loss": 0.442, "step": 5724 }, { "epoch": 1.8318046075085324, "grad_norm": 0.5084359049797058, "learning_rate": 3.949456813759677e-06, "loss": 0.4686, "step": 5725 }, { "epoch": 1.8321245733788396, "grad_norm": 0.5093129277229309, "learning_rate": 3.947636650667569e-06, "loss": 0.5013, "step": 5726 }, { "epoch": 1.8324445392491469, "grad_norm": 0.5247644186019897, "learning_rate": 3.9458166334869085e-06, "loss": 0.5435, "step": 5727 }, { "epoch": 1.8327645051194539, "grad_norm": 0.5353206396102905, "learning_rate": 3.943996762470039e-06, "loss": 0.4698, "step": 5728 }, { "epoch": 1.8330844709897611, "grad_norm": 0.5606423020362854, "learning_rate": 3.942177037869289e-06, "loss": 0.4671, "step": 5729 }, { "epoch": 1.8334044368600684, "grad_norm": 0.5476802587509155, "learning_rate": 3.9403574599369645e-06, "loss": 0.5291, "step": 5730 }, { "epoch": 1.8337244027303754, "grad_norm": 0.5585899353027344, "learning_rate": 3.938538028925353e-06, "loss": 0.5022, "step": 5731 }, { "epoch": 1.8340443686006824, "grad_norm": 0.49875614047050476, "learning_rate": 3.936718745086721e-06, "loss": 0.5021, "step": 5732 }, { "epoch": 1.83436433447099, "grad_norm": 0.4922947883605957, "learning_rate": 3.934899608673315e-06, "loss": 0.458, "step": 5733 }, { "epoch": 1.834684300341297, "grad_norm": 0.5266348719596863, "learning_rate": 3.9330806199373595e-06, "loss": 0.4972, "step": 5734 }, { "epoch": 1.835004266211604, "grad_norm": 0.45093604922294617, "learning_rate": 3.931261779131058e-06, "loss": 0.4192, "step": 5735 }, { "epoch": 1.8353242320819114, "grad_norm": 0.5231369137763977, "learning_rate": 3.929443086506597e-06, "loss": 0.4642, "step": 5736 }, { "epoch": 1.8356441979522184, "grad_norm": 0.5847294330596924, "learning_rate": 3.9276245423161396e-06, "loss": 0.5085, "step": 5737 }, { "epoch": 1.8359641638225255, "grad_norm": 0.5863973498344421, "learning_rate": 3.925806146811832e-06, "loss": 0.5286, "step": 5738 }, { "epoch": 1.8362841296928327, "grad_norm": 0.5376518368721008, "learning_rate": 3.9239879002457935e-06, "loss": 0.4939, "step": 5739 }, { "epoch": 1.83660409556314, "grad_norm": 0.5330129861831665, "learning_rate": 3.9221698028701254e-06, "loss": 0.5003, "step": 5740 }, { "epoch": 1.836924061433447, "grad_norm": 0.5517658591270447, "learning_rate": 3.920351854936911e-06, "loss": 0.484, "step": 5741 }, { "epoch": 1.8372440273037542, "grad_norm": 0.49880000948905945, "learning_rate": 3.918534056698211e-06, "loss": 0.4407, "step": 5742 }, { "epoch": 1.8375639931740615, "grad_norm": 0.5193721652030945, "learning_rate": 3.916716408406066e-06, "loss": 0.45, "step": 5743 }, { "epoch": 1.8378839590443685, "grad_norm": 0.5570769906044006, "learning_rate": 3.914898910312495e-06, "loss": 0.5672, "step": 5744 }, { "epoch": 1.8382039249146758, "grad_norm": 0.45104196667671204, "learning_rate": 3.913081562669492e-06, "loss": 0.4154, "step": 5745 }, { "epoch": 1.838523890784983, "grad_norm": 0.5590640902519226, "learning_rate": 3.911264365729041e-06, "loss": 0.5161, "step": 5746 }, { "epoch": 1.83884385665529, "grad_norm": 0.49155643582344055, "learning_rate": 3.909447319743095e-06, "loss": 0.4712, "step": 5747 }, { "epoch": 1.8391638225255973, "grad_norm": 0.5074732303619385, "learning_rate": 3.9076304249635905e-06, "loss": 0.5029, "step": 5748 }, { "epoch": 1.8394837883959045, "grad_norm": 0.49026942253112793, "learning_rate": 3.9058136816424405e-06, "loss": 0.4676, "step": 5749 }, { "epoch": 1.8398037542662116, "grad_norm": 0.5693995952606201, "learning_rate": 3.90399709003154e-06, "loss": 0.4981, "step": 5750 }, { "epoch": 1.8401237201365188, "grad_norm": 0.6498878002166748, "learning_rate": 3.902180650382762e-06, "loss": 0.5269, "step": 5751 }, { "epoch": 1.840443686006826, "grad_norm": 0.5374650955200195, "learning_rate": 3.900364362947957e-06, "loss": 0.4717, "step": 5752 }, { "epoch": 1.840763651877133, "grad_norm": 0.4944017827510834, "learning_rate": 3.898548227978958e-06, "loss": 0.5311, "step": 5753 }, { "epoch": 1.8410836177474401, "grad_norm": 0.48711761832237244, "learning_rate": 3.89673224572757e-06, "loss": 0.4029, "step": 5754 }, { "epoch": 1.8414035836177476, "grad_norm": 0.5856059193611145, "learning_rate": 3.8949164164455825e-06, "loss": 0.517, "step": 5755 }, { "epoch": 1.8417235494880546, "grad_norm": 0.5151647329330444, "learning_rate": 3.893100740384766e-06, "loss": 0.4727, "step": 5756 }, { "epoch": 1.8420435153583616, "grad_norm": 0.5710545778274536, "learning_rate": 3.891285217796862e-06, "loss": 0.5521, "step": 5757 }, { "epoch": 1.8423634812286689, "grad_norm": 0.5346451997756958, "learning_rate": 3.889469848933598e-06, "loss": 0.4912, "step": 5758 }, { "epoch": 1.8426834470989761, "grad_norm": 0.5003721117973328, "learning_rate": 3.887654634046674e-06, "loss": 0.4756, "step": 5759 }, { "epoch": 1.8430034129692832, "grad_norm": 0.593546986579895, "learning_rate": 3.885839573387772e-06, "loss": 0.47, "step": 5760 }, { "epoch": 1.8433233788395904, "grad_norm": 0.5888879299163818, "learning_rate": 3.884024667208556e-06, "loss": 0.4706, "step": 5761 }, { "epoch": 1.8436433447098977, "grad_norm": 0.5248382687568665, "learning_rate": 3.8822099157606605e-06, "loss": 0.5013, "step": 5762 }, { "epoch": 1.8439633105802047, "grad_norm": 0.48654475808143616, "learning_rate": 3.880395319295707e-06, "loss": 0.4886, "step": 5763 }, { "epoch": 1.844283276450512, "grad_norm": 0.5649202466011047, "learning_rate": 3.8785808780652875e-06, "loss": 0.5276, "step": 5764 }, { "epoch": 1.8446032423208192, "grad_norm": 0.5007580518722534, "learning_rate": 3.876766592320977e-06, "loss": 0.4502, "step": 5765 }, { "epoch": 1.8449232081911262, "grad_norm": 0.5063193440437317, "learning_rate": 3.874952462314333e-06, "loss": 0.4951, "step": 5766 }, { "epoch": 1.8452431740614335, "grad_norm": 0.5258777737617493, "learning_rate": 3.873138488296883e-06, "loss": 0.4848, "step": 5767 }, { "epoch": 1.8455631399317407, "grad_norm": 0.5359113216400146, "learning_rate": 3.871324670520137e-06, "loss": 0.553, "step": 5768 }, { "epoch": 1.8458831058020477, "grad_norm": 0.4946892559528351, "learning_rate": 3.869511009235582e-06, "loss": 0.536, "step": 5769 }, { "epoch": 1.846203071672355, "grad_norm": 0.49602842330932617, "learning_rate": 3.867697504694685e-06, "loss": 0.4712, "step": 5770 }, { "epoch": 1.8465230375426622, "grad_norm": 0.4690818190574646, "learning_rate": 3.865884157148894e-06, "loss": 0.4463, "step": 5771 }, { "epoch": 1.8468430034129693, "grad_norm": 0.5733300447463989, "learning_rate": 3.8640709668496265e-06, "loss": 0.5591, "step": 5772 }, { "epoch": 1.8471629692832765, "grad_norm": 0.45847299695014954, "learning_rate": 3.862257934048289e-06, "loss": 0.4772, "step": 5773 }, { "epoch": 1.8474829351535837, "grad_norm": 0.5056205987930298, "learning_rate": 3.860445058996255e-06, "loss": 0.5491, "step": 5774 }, { "epoch": 1.8478029010238908, "grad_norm": 0.5060153603553772, "learning_rate": 3.858632341944884e-06, "loss": 0.4892, "step": 5775 }, { "epoch": 1.8481228668941978, "grad_norm": 0.4729257822036743, "learning_rate": 3.856819783145514e-06, "loss": 0.4729, "step": 5776 }, { "epoch": 1.8484428327645053, "grad_norm": 0.5032588243484497, "learning_rate": 3.855007382849455e-06, "loss": 0.5282, "step": 5777 }, { "epoch": 1.8487627986348123, "grad_norm": 0.5291029214859009, "learning_rate": 3.853195141308001e-06, "loss": 0.4616, "step": 5778 }, { "epoch": 1.8490827645051193, "grad_norm": 0.5785648822784424, "learning_rate": 3.851383058772418e-06, "loss": 0.5161, "step": 5779 }, { "epoch": 1.8494027303754266, "grad_norm": 0.4976552426815033, "learning_rate": 3.849571135493958e-06, "loss": 0.4668, "step": 5780 }, { "epoch": 1.8497226962457338, "grad_norm": 0.5364990234375, "learning_rate": 3.847759371723844e-06, "loss": 0.4914, "step": 5781 }, { "epoch": 1.8500426621160408, "grad_norm": 0.5276072025299072, "learning_rate": 3.845947767713278e-06, "loss": 0.4959, "step": 5782 }, { "epoch": 1.850362627986348, "grad_norm": 0.5432918071746826, "learning_rate": 3.844136323713444e-06, "loss": 0.5032, "step": 5783 }, { "epoch": 1.8506825938566553, "grad_norm": 0.5424608588218689, "learning_rate": 3.842325039975497e-06, "loss": 0.4803, "step": 5784 }, { "epoch": 1.8510025597269624, "grad_norm": 0.5053834319114685, "learning_rate": 3.840513916750577e-06, "loss": 0.4765, "step": 5785 }, { "epoch": 1.8513225255972696, "grad_norm": 0.5041948556900024, "learning_rate": 3.838702954289798e-06, "loss": 0.4887, "step": 5786 }, { "epoch": 1.8516424914675769, "grad_norm": 0.5387802720069885, "learning_rate": 3.836892152844251e-06, "loss": 0.4919, "step": 5787 }, { "epoch": 1.851962457337884, "grad_norm": 0.49143558740615845, "learning_rate": 3.835081512665007e-06, "loss": 0.437, "step": 5788 }, { "epoch": 1.8522824232081911, "grad_norm": 0.50626140832901, "learning_rate": 3.833271034003111e-06, "loss": 0.4643, "step": 5789 }, { "epoch": 1.8526023890784984, "grad_norm": 0.4904909133911133, "learning_rate": 3.83146071710959e-06, "loss": 0.4946, "step": 5790 }, { "epoch": 1.8529223549488054, "grad_norm": 0.5600085854530334, "learning_rate": 3.829650562235448e-06, "loss": 0.5578, "step": 5791 }, { "epoch": 1.8532423208191127, "grad_norm": 0.5587226152420044, "learning_rate": 3.827840569631662e-06, "loss": 0.4908, "step": 5792 }, { "epoch": 1.85356228668942, "grad_norm": 0.49955758452415466, "learning_rate": 3.826030739549193e-06, "loss": 0.4517, "step": 5793 }, { "epoch": 1.853882252559727, "grad_norm": 0.5089410543441772, "learning_rate": 3.824221072238972e-06, "loss": 0.5098, "step": 5794 }, { "epoch": 1.854202218430034, "grad_norm": 0.5281291007995605, "learning_rate": 3.8224115679519144e-06, "loss": 0.4656, "step": 5795 }, { "epoch": 1.8545221843003414, "grad_norm": 0.4958762526512146, "learning_rate": 3.820602226938911e-06, "loss": 0.5027, "step": 5796 }, { "epoch": 1.8548421501706485, "grad_norm": 0.4767692983150482, "learning_rate": 3.818793049450826e-06, "loss": 0.4702, "step": 5797 }, { "epoch": 1.8551621160409555, "grad_norm": 0.45611903071403503, "learning_rate": 3.8169840357385075e-06, "loss": 0.4344, "step": 5798 }, { "epoch": 1.855482081911263, "grad_norm": 0.5477572679519653, "learning_rate": 3.8151751860527734e-06, "loss": 0.4854, "step": 5799 }, { "epoch": 1.85580204778157, "grad_norm": 0.5560097098350525, "learning_rate": 3.813366500644426e-06, "loss": 0.5266, "step": 5800 }, { "epoch": 1.856122013651877, "grad_norm": 0.5477871894836426, "learning_rate": 3.811557979764242e-06, "loss": 0.5198, "step": 5801 }, { "epoch": 1.8564419795221843, "grad_norm": 0.5169461965560913, "learning_rate": 3.809749623662973e-06, "loss": 0.4519, "step": 5802 }, { "epoch": 1.8567619453924915, "grad_norm": 0.544291615486145, "learning_rate": 3.807941432591351e-06, "loss": 0.5343, "step": 5803 }, { "epoch": 1.8570819112627985, "grad_norm": 0.5299692153930664, "learning_rate": 3.806133406800082e-06, "loss": 0.5507, "step": 5804 }, { "epoch": 1.8574018771331058, "grad_norm": 0.5153736472129822, "learning_rate": 3.8043255465398533e-06, "loss": 0.4455, "step": 5805 }, { "epoch": 1.857721843003413, "grad_norm": 0.49252840876579285, "learning_rate": 3.802517852061327e-06, "loss": 0.5033, "step": 5806 }, { "epoch": 1.85804180887372, "grad_norm": 0.4893689453601837, "learning_rate": 3.80071032361514e-06, "loss": 0.4719, "step": 5807 }, { "epoch": 1.8583617747440273, "grad_norm": 0.5370251536369324, "learning_rate": 3.79890296145191e-06, "loss": 0.4967, "step": 5808 }, { "epoch": 1.8586817406143346, "grad_norm": 0.5664129853248596, "learning_rate": 3.797095765822228e-06, "loss": 0.514, "step": 5809 }, { "epoch": 1.8590017064846416, "grad_norm": 0.5876678824424744, "learning_rate": 3.795288736976666e-06, "loss": 0.5098, "step": 5810 }, { "epoch": 1.8593216723549488, "grad_norm": 0.5216983556747437, "learning_rate": 3.793481875165771e-06, "loss": 0.461, "step": 5811 }, { "epoch": 1.859641638225256, "grad_norm": 0.4685456156730652, "learning_rate": 3.791675180640064e-06, "loss": 0.4429, "step": 5812 }, { "epoch": 1.859961604095563, "grad_norm": 0.5082848072052002, "learning_rate": 3.789868653650046e-06, "loss": 0.5158, "step": 5813 }, { "epoch": 1.8602815699658704, "grad_norm": 0.4984305500984192, "learning_rate": 3.7880622944461975e-06, "loss": 0.4712, "step": 5814 }, { "epoch": 1.8606015358361776, "grad_norm": 0.5433828830718994, "learning_rate": 3.7862561032789692e-06, "loss": 0.5411, "step": 5815 }, { "epoch": 1.8609215017064846, "grad_norm": 0.5483769774436951, "learning_rate": 3.7844500803987925e-06, "loss": 0.42, "step": 5816 }, { "epoch": 1.8612414675767917, "grad_norm": 0.5509781241416931, "learning_rate": 3.7826442260560737e-06, "loss": 0.5298, "step": 5817 }, { "epoch": 1.8615614334470991, "grad_norm": 0.5578811168670654, "learning_rate": 3.7808385405011965e-06, "loss": 0.5022, "step": 5818 }, { "epoch": 1.8618813993174061, "grad_norm": 0.5305883884429932, "learning_rate": 3.779033023984524e-06, "loss": 0.4844, "step": 5819 }, { "epoch": 1.8622013651877132, "grad_norm": 0.500226616859436, "learning_rate": 3.7772276767563916e-06, "loss": 0.4359, "step": 5820 }, { "epoch": 1.8625213310580204, "grad_norm": 0.5373531579971313, "learning_rate": 3.775422499067113e-06, "loss": 0.5137, "step": 5821 }, { "epoch": 1.8628412969283277, "grad_norm": 0.5030234456062317, "learning_rate": 3.7736174911669775e-06, "loss": 0.4959, "step": 5822 }, { "epoch": 1.8631612627986347, "grad_norm": 0.49924901127815247, "learning_rate": 3.7718126533062504e-06, "loss": 0.5021, "step": 5823 }, { "epoch": 1.863481228668942, "grad_norm": 0.5491090416908264, "learning_rate": 3.7700079857351797e-06, "loss": 0.5029, "step": 5824 }, { "epoch": 1.8638011945392492, "grad_norm": 0.5774345993995667, "learning_rate": 3.76820348870398e-06, "loss": 0.5108, "step": 5825 }, { "epoch": 1.8641211604095562, "grad_norm": 0.5084494948387146, "learning_rate": 3.7663991624628495e-06, "loss": 0.448, "step": 5826 }, { "epoch": 1.8644411262798635, "grad_norm": 0.5092008113861084, "learning_rate": 3.7645950072619573e-06, "loss": 0.4599, "step": 5827 }, { "epoch": 1.8647610921501707, "grad_norm": 0.5794777274131775, "learning_rate": 3.762791023351454e-06, "loss": 0.542, "step": 5828 }, { "epoch": 1.8650810580204777, "grad_norm": 0.4939880967140198, "learning_rate": 3.7609872109814648e-06, "loss": 0.4244, "step": 5829 }, { "epoch": 1.865401023890785, "grad_norm": 0.4751640558242798, "learning_rate": 3.7591835704020885e-06, "loss": 0.4604, "step": 5830 }, { "epoch": 1.8657209897610922, "grad_norm": 0.4862428903579712, "learning_rate": 3.7573801018634043e-06, "loss": 0.4439, "step": 5831 }, { "epoch": 1.8660409556313993, "grad_norm": 0.5053917169570923, "learning_rate": 3.7555768056154623e-06, "loss": 0.5173, "step": 5832 }, { "epoch": 1.8663609215017065, "grad_norm": 0.47630491852760315, "learning_rate": 3.7537736819082926e-06, "loss": 0.44, "step": 5833 }, { "epoch": 1.8666808873720138, "grad_norm": 0.5234889388084412, "learning_rate": 3.751970730991903e-06, "loss": 0.5028, "step": 5834 }, { "epoch": 1.8670008532423208, "grad_norm": 0.532590389251709, "learning_rate": 3.750167953116272e-06, "loss": 0.4931, "step": 5835 }, { "epoch": 1.867320819112628, "grad_norm": 0.48530295491218567, "learning_rate": 3.748365348531359e-06, "loss": 0.4657, "step": 5836 }, { "epoch": 1.8676407849829353, "grad_norm": 0.551383376121521, "learning_rate": 3.7465629174870955e-06, "loss": 0.51, "step": 5837 }, { "epoch": 1.8679607508532423, "grad_norm": 0.4987407624721527, "learning_rate": 3.7447606602333895e-06, "loss": 0.4767, "step": 5838 }, { "epoch": 1.8682807167235493, "grad_norm": 0.49918514490127563, "learning_rate": 3.7429585770201314e-06, "loss": 0.4366, "step": 5839 }, { "epoch": 1.8686006825938568, "grad_norm": 0.5327936410903931, "learning_rate": 3.7411566680971765e-06, "loss": 0.4906, "step": 5840 }, { "epoch": 1.8689206484641638, "grad_norm": 0.49172770977020264, "learning_rate": 3.739354933714366e-06, "loss": 0.4405, "step": 5841 }, { "epoch": 1.8692406143344709, "grad_norm": 0.5563710927963257, "learning_rate": 3.7375533741215086e-06, "loss": 0.5039, "step": 5842 }, { "epoch": 1.869560580204778, "grad_norm": 0.5064439177513123, "learning_rate": 3.7357519895683948e-06, "loss": 0.4778, "step": 5843 }, { "epoch": 1.8698805460750854, "grad_norm": 0.5443150401115417, "learning_rate": 3.73395078030479e-06, "loss": 0.5132, "step": 5844 }, { "epoch": 1.8702005119453924, "grad_norm": 0.5381937623023987, "learning_rate": 3.732149746580431e-06, "loss": 0.5251, "step": 5845 }, { "epoch": 1.8705204778156996, "grad_norm": 0.5324935913085938, "learning_rate": 3.7303488886450367e-06, "loss": 0.4619, "step": 5846 }, { "epoch": 1.8708404436860069, "grad_norm": 0.49195364117622375, "learning_rate": 3.728548206748295e-06, "loss": 0.5433, "step": 5847 }, { "epoch": 1.871160409556314, "grad_norm": 0.49643033742904663, "learning_rate": 3.7267477011398735e-06, "loss": 0.426, "step": 5848 }, { "epoch": 1.8714803754266212, "grad_norm": 0.5382016897201538, "learning_rate": 3.724947372069416e-06, "loss": 0.4352, "step": 5849 }, { "epoch": 1.8718003412969284, "grad_norm": 0.5211527943611145, "learning_rate": 3.7231472197865394e-06, "loss": 0.46, "step": 5850 }, { "epoch": 1.8721203071672354, "grad_norm": 0.5157228708267212, "learning_rate": 3.721347244540837e-06, "loss": 0.5262, "step": 5851 }, { "epoch": 1.8724402730375427, "grad_norm": 0.48551246523857117, "learning_rate": 3.719547446581876e-06, "loss": 0.4761, "step": 5852 }, { "epoch": 1.87276023890785, "grad_norm": 0.47866228222846985, "learning_rate": 3.7177478261592024e-06, "loss": 0.4987, "step": 5853 }, { "epoch": 1.873080204778157, "grad_norm": 0.4556387960910797, "learning_rate": 3.7159483835223374e-06, "loss": 0.4724, "step": 5854 }, { "epoch": 1.8734001706484642, "grad_norm": 0.5026599168777466, "learning_rate": 3.7141491189207716e-06, "loss": 0.47, "step": 5855 }, { "epoch": 1.8737201365187715, "grad_norm": 0.5129727721214294, "learning_rate": 3.712350032603979e-06, "loss": 0.5134, "step": 5856 }, { "epoch": 1.8740401023890785, "grad_norm": 0.4955362379550934, "learning_rate": 3.710551124821401e-06, "loss": 0.4959, "step": 5857 }, { "epoch": 1.8743600682593855, "grad_norm": 0.4802592396736145, "learning_rate": 3.7087523958224624e-06, "loss": 0.499, "step": 5858 }, { "epoch": 1.874680034129693, "grad_norm": 0.5008774995803833, "learning_rate": 3.706953845856558e-06, "loss": 0.4934, "step": 5859 }, { "epoch": 1.875, "grad_norm": 0.5445345044136047, "learning_rate": 3.705155475173058e-06, "loss": 0.4794, "step": 5860 }, { "epoch": 1.875319965870307, "grad_norm": 0.5107492208480835, "learning_rate": 3.7033572840213093e-06, "loss": 0.4789, "step": 5861 }, { "epoch": 1.8756399317406145, "grad_norm": 0.5111907124519348, "learning_rate": 3.7015592726506318e-06, "loss": 0.4625, "step": 5862 }, { "epoch": 1.8759598976109215, "grad_norm": 0.5461314916610718, "learning_rate": 3.699761441310324e-06, "loss": 0.5222, "step": 5863 }, { "epoch": 1.8762798634812285, "grad_norm": 0.4889307916164398, "learning_rate": 3.697963790249657e-06, "loss": 0.4288, "step": 5864 }, { "epoch": 1.8765998293515358, "grad_norm": 0.4892370104789734, "learning_rate": 3.6961663197178767e-06, "loss": 0.4727, "step": 5865 }, { "epoch": 1.876919795221843, "grad_norm": 0.4976430833339691, "learning_rate": 3.6943690299642055e-06, "loss": 0.5502, "step": 5866 }, { "epoch": 1.87723976109215, "grad_norm": 0.5494089722633362, "learning_rate": 3.6925719212378364e-06, "loss": 0.4605, "step": 5867 }, { "epoch": 1.8775597269624573, "grad_norm": 0.5490610003471375, "learning_rate": 3.690774993787945e-06, "loss": 0.5349, "step": 5868 }, { "epoch": 1.8778796928327646, "grad_norm": 0.5048588514328003, "learning_rate": 3.6889782478636773e-06, "loss": 0.4975, "step": 5869 }, { "epoch": 1.8781996587030716, "grad_norm": 0.5531996488571167, "learning_rate": 3.6871816837141506e-06, "loss": 0.4456, "step": 5870 }, { "epoch": 1.8785196245733788, "grad_norm": 0.5299012660980225, "learning_rate": 3.6853853015884645e-06, "loss": 0.4258, "step": 5871 }, { "epoch": 1.878839590443686, "grad_norm": 0.5071423053741455, "learning_rate": 3.683589101735685e-06, "loss": 0.5059, "step": 5872 }, { "epoch": 1.8791595563139931, "grad_norm": 0.5356152057647705, "learning_rate": 3.6817930844048622e-06, "loss": 0.4685, "step": 5873 }, { "epoch": 1.8794795221843004, "grad_norm": 0.5635944604873657, "learning_rate": 3.6799972498450145e-06, "loss": 0.5245, "step": 5874 }, { "epoch": 1.8797994880546076, "grad_norm": 0.5151473879814148, "learning_rate": 3.678201598305135e-06, "loss": 0.4866, "step": 5875 }, { "epoch": 1.8801194539249146, "grad_norm": 0.4848635792732239, "learning_rate": 3.6764061300341956e-06, "loss": 0.4681, "step": 5876 }, { "epoch": 1.880439419795222, "grad_norm": 0.5050166845321655, "learning_rate": 3.6746108452811347e-06, "loss": 0.5087, "step": 5877 }, { "epoch": 1.8807593856655291, "grad_norm": 0.48006412386894226, "learning_rate": 3.6728157442948786e-06, "loss": 0.4778, "step": 5878 }, { "epoch": 1.8810793515358362, "grad_norm": 0.5760488510131836, "learning_rate": 3.6710208273243157e-06, "loss": 0.5432, "step": 5879 }, { "epoch": 1.8813993174061432, "grad_norm": 0.5436517000198364, "learning_rate": 3.6692260946183133e-06, "loss": 0.4493, "step": 5880 }, { "epoch": 1.8817192832764507, "grad_norm": 0.5405221581459045, "learning_rate": 3.6674315464257142e-06, "loss": 0.4942, "step": 5881 }, { "epoch": 1.8820392491467577, "grad_norm": 0.49750787019729614, "learning_rate": 3.665637182995332e-06, "loss": 0.4563, "step": 5882 }, { "epoch": 1.8823592150170647, "grad_norm": 0.539986252784729, "learning_rate": 3.663843004575962e-06, "loss": 0.5238, "step": 5883 }, { "epoch": 1.882679180887372, "grad_norm": 0.5363343358039856, "learning_rate": 3.662049011416368e-06, "loss": 0.5187, "step": 5884 }, { "epoch": 1.8829991467576792, "grad_norm": 0.5319176316261292, "learning_rate": 3.6602552037652855e-06, "loss": 0.4631, "step": 5885 }, { "epoch": 1.8833191126279862, "grad_norm": 0.4889177680015564, "learning_rate": 3.658461581871431e-06, "loss": 0.4641, "step": 5886 }, { "epoch": 1.8836390784982935, "grad_norm": 0.5030321478843689, "learning_rate": 3.6566681459834917e-06, "loss": 0.4517, "step": 5887 }, { "epoch": 1.8839590443686007, "grad_norm": 0.5236909985542297, "learning_rate": 3.6548748963501324e-06, "loss": 0.4772, "step": 5888 }, { "epoch": 1.8842790102389078, "grad_norm": 0.5497915148735046, "learning_rate": 3.6530818332199858e-06, "loss": 0.5134, "step": 5889 }, { "epoch": 1.884598976109215, "grad_norm": 0.563951849937439, "learning_rate": 3.6512889568416643e-06, "loss": 0.5396, "step": 5890 }, { "epoch": 1.8849189419795223, "grad_norm": 0.5152337551116943, "learning_rate": 3.649496267463749e-06, "loss": 0.4871, "step": 5891 }, { "epoch": 1.8852389078498293, "grad_norm": 0.44863662123680115, "learning_rate": 3.6477037653348023e-06, "loss": 0.3912, "step": 5892 }, { "epoch": 1.8855588737201365, "grad_norm": 0.5549201965332031, "learning_rate": 3.645911450703356e-06, "loss": 0.5718, "step": 5893 }, { "epoch": 1.8858788395904438, "grad_norm": 0.48657435178756714, "learning_rate": 3.6441193238179152e-06, "loss": 0.483, "step": 5894 }, { "epoch": 1.8861988054607508, "grad_norm": 0.5181356072425842, "learning_rate": 3.6423273849269624e-06, "loss": 0.4646, "step": 5895 }, { "epoch": 1.886518771331058, "grad_norm": 0.49020054936408997, "learning_rate": 3.6405356342789496e-06, "loss": 0.479, "step": 5896 }, { "epoch": 1.8868387372013653, "grad_norm": 0.42209747433662415, "learning_rate": 3.6387440721223067e-06, "loss": 0.4726, "step": 5897 }, { "epoch": 1.8871587030716723, "grad_norm": 0.48372170329093933, "learning_rate": 3.6369526987054372e-06, "loss": 0.4968, "step": 5898 }, { "epoch": 1.8874786689419796, "grad_norm": 0.5298974514007568, "learning_rate": 3.635161514276715e-06, "loss": 0.5113, "step": 5899 }, { "epoch": 1.8877986348122868, "grad_norm": 0.5127506852149963, "learning_rate": 3.6333705190844917e-06, "loss": 0.461, "step": 5900 }, { "epoch": 1.8881186006825939, "grad_norm": 0.5458589196205139, "learning_rate": 3.631579713377088e-06, "loss": 0.468, "step": 5901 }, { "epoch": 1.8884385665529009, "grad_norm": 0.5374716520309448, "learning_rate": 3.6297890974028037e-06, "loss": 0.5122, "step": 5902 }, { "epoch": 1.8887585324232083, "grad_norm": 0.45911529660224915, "learning_rate": 3.627998671409911e-06, "loss": 0.4274, "step": 5903 }, { "epoch": 1.8890784982935154, "grad_norm": 0.5461366772651672, "learning_rate": 3.626208435646652e-06, "loss": 0.5157, "step": 5904 }, { "epoch": 1.8893984641638224, "grad_norm": 0.5829647779464722, "learning_rate": 3.6244183903612473e-06, "loss": 0.5233, "step": 5905 }, { "epoch": 1.8897184300341296, "grad_norm": 0.4681430757045746, "learning_rate": 3.6226285358018854e-06, "loss": 0.4382, "step": 5906 }, { "epoch": 1.890038395904437, "grad_norm": 0.5075001120567322, "learning_rate": 3.6208388722167355e-06, "loss": 0.5129, "step": 5907 }, { "epoch": 1.890358361774744, "grad_norm": 0.515108585357666, "learning_rate": 3.6190493998539356e-06, "loss": 0.5281, "step": 5908 }, { "epoch": 1.8906783276450512, "grad_norm": 0.49774298071861267, "learning_rate": 3.6172601189615963e-06, "loss": 0.4798, "step": 5909 }, { "epoch": 1.8909982935153584, "grad_norm": 0.5024615526199341, "learning_rate": 3.615471029787807e-06, "loss": 0.4733, "step": 5910 }, { "epoch": 1.8913182593856654, "grad_norm": 0.47436338663101196, "learning_rate": 3.6136821325806227e-06, "loss": 0.4588, "step": 5911 }, { "epoch": 1.8916382252559727, "grad_norm": 0.5262418985366821, "learning_rate": 3.611893427588079e-06, "loss": 0.4737, "step": 5912 }, { "epoch": 1.89195819112628, "grad_norm": 0.5761246681213379, "learning_rate": 3.6101049150581836e-06, "loss": 0.5106, "step": 5913 }, { "epoch": 1.892278156996587, "grad_norm": 0.4981935918331146, "learning_rate": 3.608316595238912e-06, "loss": 0.4793, "step": 5914 }, { "epoch": 1.8925981228668942, "grad_norm": 0.5088338255882263, "learning_rate": 3.6065284683782197e-06, "loss": 0.522, "step": 5915 }, { "epoch": 1.8929180887372015, "grad_norm": 0.49246954917907715, "learning_rate": 3.604740534724029e-06, "loss": 0.4482, "step": 5916 }, { "epoch": 1.8932380546075085, "grad_norm": 0.5981255173683167, "learning_rate": 3.6029527945242436e-06, "loss": 0.5036, "step": 5917 }, { "epoch": 1.8935580204778157, "grad_norm": 0.6465089917182922, "learning_rate": 3.601165248026736e-06, "loss": 0.5968, "step": 5918 }, { "epoch": 1.893877986348123, "grad_norm": 0.5026783347129822, "learning_rate": 3.5993778954793467e-06, "loss": 0.409, "step": 5919 }, { "epoch": 1.89419795221843, "grad_norm": 0.501559317111969, "learning_rate": 3.597590737129897e-06, "loss": 0.4895, "step": 5920 }, { "epoch": 1.894517918088737, "grad_norm": 0.5835351347923279, "learning_rate": 3.5958037732261806e-06, "loss": 0.5273, "step": 5921 }, { "epoch": 1.8948378839590445, "grad_norm": 0.5803717970848083, "learning_rate": 3.5940170040159595e-06, "loss": 0.4824, "step": 5922 }, { "epoch": 1.8951578498293515, "grad_norm": 0.4998316764831543, "learning_rate": 3.5922304297469728e-06, "loss": 0.4551, "step": 5923 }, { "epoch": 1.8954778156996586, "grad_norm": 0.5349484086036682, "learning_rate": 3.5904440506669303e-06, "loss": 0.4978, "step": 5924 }, { "epoch": 1.895797781569966, "grad_norm": 0.5529834628105164, "learning_rate": 3.588657867023514e-06, "loss": 0.5322, "step": 5925 }, { "epoch": 1.896117747440273, "grad_norm": 0.49046170711517334, "learning_rate": 3.5868718790643847e-06, "loss": 0.4508, "step": 5926 }, { "epoch": 1.89643771331058, "grad_norm": 0.5163431763648987, "learning_rate": 3.5850860870371685e-06, "loss": 0.4882, "step": 5927 }, { "epoch": 1.8967576791808873, "grad_norm": 0.5227925777435303, "learning_rate": 3.5833004911894697e-06, "loss": 0.5048, "step": 5928 }, { "epoch": 1.8970776450511946, "grad_norm": 0.5189322829246521, "learning_rate": 3.5815150917688613e-06, "loss": 0.4615, "step": 5929 }, { "epoch": 1.8973976109215016, "grad_norm": 0.5135989189147949, "learning_rate": 3.5797298890228903e-06, "loss": 0.4782, "step": 5930 }, { "epoch": 1.8977175767918089, "grad_norm": 0.5108239650726318, "learning_rate": 3.577944883199081e-06, "loss": 0.4838, "step": 5931 }, { "epoch": 1.898037542662116, "grad_norm": 0.5215096473693848, "learning_rate": 3.5761600745449234e-06, "loss": 0.4429, "step": 5932 }, { "epoch": 1.8983575085324231, "grad_norm": 0.5620769262313843, "learning_rate": 3.574375463307885e-06, "loss": 0.4588, "step": 5933 }, { "epoch": 1.8986774744027304, "grad_norm": 0.5866304039955139, "learning_rate": 3.5725910497354037e-06, "loss": 0.5205, "step": 5934 }, { "epoch": 1.8989974402730376, "grad_norm": 0.5532835125923157, "learning_rate": 3.5708068340748893e-06, "loss": 0.5163, "step": 5935 }, { "epoch": 1.8993174061433447, "grad_norm": 0.44951871037483215, "learning_rate": 3.5690228165737283e-06, "loss": 0.4246, "step": 5936 }, { "epoch": 1.899637372013652, "grad_norm": 0.5440343618392944, "learning_rate": 3.5672389974792743e-06, "loss": 0.5024, "step": 5937 }, { "epoch": 1.8999573378839592, "grad_norm": 0.5705434679985046, "learning_rate": 3.565455377038858e-06, "loss": 0.5296, "step": 5938 }, { "epoch": 1.9002773037542662, "grad_norm": 0.556301474571228, "learning_rate": 3.563671955499779e-06, "loss": 0.4251, "step": 5939 }, { "epoch": 1.9005972696245734, "grad_norm": 0.4998564124107361, "learning_rate": 3.56188873310931e-06, "loss": 0.4588, "step": 5940 }, { "epoch": 1.9009172354948807, "grad_norm": 0.5576817393302917, "learning_rate": 3.5601057101146995e-06, "loss": 0.507, "step": 5941 }, { "epoch": 1.9012372013651877, "grad_norm": 0.4908350110054016, "learning_rate": 3.558322886763165e-06, "loss": 0.4606, "step": 5942 }, { "epoch": 1.9015571672354947, "grad_norm": 0.5293522477149963, "learning_rate": 3.5565402633018963e-06, "loss": 0.4604, "step": 5943 }, { "epoch": 1.9018771331058022, "grad_norm": 0.48738279938697815, "learning_rate": 3.5547578399780557e-06, "loss": 0.4269, "step": 5944 }, { "epoch": 1.9021970989761092, "grad_norm": 0.5112420916557312, "learning_rate": 3.5529756170387785e-06, "loss": 0.4986, "step": 5945 }, { "epoch": 1.9025170648464163, "grad_norm": 0.5296844244003296, "learning_rate": 3.5511935947311756e-06, "loss": 0.5062, "step": 5946 }, { "epoch": 1.9028370307167235, "grad_norm": 0.5048282742500305, "learning_rate": 3.549411773302322e-06, "loss": 0.4243, "step": 5947 }, { "epoch": 1.9031569965870307, "grad_norm": 0.5419057011604309, "learning_rate": 3.5476301529992717e-06, "loss": 0.5494, "step": 5948 }, { "epoch": 1.9034769624573378, "grad_norm": 0.4937518835067749, "learning_rate": 3.545848734069048e-06, "loss": 0.4584, "step": 5949 }, { "epoch": 1.903796928327645, "grad_norm": 0.5428810715675354, "learning_rate": 3.5440675167586446e-06, "loss": 0.4979, "step": 5950 }, { "epoch": 1.9041168941979523, "grad_norm": 0.5214746594429016, "learning_rate": 3.5422865013150344e-06, "loss": 0.4961, "step": 5951 }, { "epoch": 1.9044368600682593, "grad_norm": 0.5318081378936768, "learning_rate": 3.5405056879851533e-06, "loss": 0.5296, "step": 5952 }, { "epoch": 1.9047568259385665, "grad_norm": 0.488450825214386, "learning_rate": 3.5387250770159152e-06, "loss": 0.4416, "step": 5953 }, { "epoch": 1.9050767918088738, "grad_norm": 0.540209174156189, "learning_rate": 3.536944668654202e-06, "loss": 0.4512, "step": 5954 }, { "epoch": 1.9053967576791808, "grad_norm": 0.5259593725204468, "learning_rate": 3.53516446314687e-06, "loss": 0.518, "step": 5955 }, { "epoch": 1.905716723549488, "grad_norm": 0.5429878830909729, "learning_rate": 3.5333844607407497e-06, "loss": 0.5454, "step": 5956 }, { "epoch": 1.9060366894197953, "grad_norm": 0.42499151825904846, "learning_rate": 3.531604661682636e-06, "loss": 0.4153, "step": 5957 }, { "epoch": 1.9063566552901023, "grad_norm": 0.5061125755310059, "learning_rate": 3.5298250662193046e-06, "loss": 0.5393, "step": 5958 }, { "epoch": 1.9066766211604096, "grad_norm": 0.5056055188179016, "learning_rate": 3.5280456745974943e-06, "loss": 0.469, "step": 5959 }, { "epoch": 1.9069965870307168, "grad_norm": 0.5259008407592773, "learning_rate": 3.526266487063923e-06, "loss": 0.5018, "step": 5960 }, { "epoch": 1.9073165529010239, "grad_norm": 0.4866785407066345, "learning_rate": 3.524487503865277e-06, "loss": 0.4832, "step": 5961 }, { "epoch": 1.9076365187713311, "grad_norm": 0.44651931524276733, "learning_rate": 3.522708725248213e-06, "loss": 0.4442, "step": 5962 }, { "epoch": 1.9079564846416384, "grad_norm": 0.5012054443359375, "learning_rate": 3.520930151459362e-06, "loss": 0.4776, "step": 5963 }, { "epoch": 1.9082764505119454, "grad_norm": 0.4618786573410034, "learning_rate": 3.5191517827453224e-06, "loss": 0.5152, "step": 5964 }, { "epoch": 1.9085964163822524, "grad_norm": 0.5219776630401611, "learning_rate": 3.517373619352672e-06, "loss": 0.4974, "step": 5965 }, { "epoch": 1.9089163822525599, "grad_norm": 0.5021116733551025, "learning_rate": 3.5155956615279528e-06, "loss": 0.4273, "step": 5966 }, { "epoch": 1.909236348122867, "grad_norm": 0.5333878397941589, "learning_rate": 3.5138179095176805e-06, "loss": 0.5371, "step": 5967 }, { "epoch": 1.909556313993174, "grad_norm": 0.5196536779403687, "learning_rate": 3.512040363568344e-06, "loss": 0.4882, "step": 5968 }, { "epoch": 1.9098762798634812, "grad_norm": 0.5455049872398376, "learning_rate": 3.5102630239263986e-06, "loss": 0.4518, "step": 5969 }, { "epoch": 1.9101962457337884, "grad_norm": 0.5605724453926086, "learning_rate": 3.508485890838278e-06, "loss": 0.4898, "step": 5970 }, { "epoch": 1.9105162116040955, "grad_norm": 0.5000489354133606, "learning_rate": 3.506708964550385e-06, "loss": 0.5037, "step": 5971 }, { "epoch": 1.9108361774744027, "grad_norm": 0.515364408493042, "learning_rate": 3.5049322453090883e-06, "loss": 0.4956, "step": 5972 }, { "epoch": 1.91115614334471, "grad_norm": 0.5556088089942932, "learning_rate": 3.503155733360737e-06, "loss": 0.4633, "step": 5973 }, { "epoch": 1.911476109215017, "grad_norm": 0.5696067810058594, "learning_rate": 3.5013794289516406e-06, "loss": 0.4881, "step": 5974 }, { "epoch": 1.9117960750853242, "grad_norm": 0.48840129375457764, "learning_rate": 3.4996033323280907e-06, "loss": 0.536, "step": 5975 }, { "epoch": 1.9121160409556315, "grad_norm": 0.4710192084312439, "learning_rate": 3.4978274437363447e-06, "loss": 0.472, "step": 5976 }, { "epoch": 1.9124360068259385, "grad_norm": 0.47531548142433167, "learning_rate": 3.49605176342263e-06, "loss": 0.4946, "step": 5977 }, { "epoch": 1.9127559726962458, "grad_norm": 0.5320174694061279, "learning_rate": 3.4942762916331485e-06, "loss": 0.489, "step": 5978 }, { "epoch": 1.913075938566553, "grad_norm": 0.5178653001785278, "learning_rate": 3.4925010286140688e-06, "loss": 0.4951, "step": 5979 }, { "epoch": 1.91339590443686, "grad_norm": 0.45877155661582947, "learning_rate": 3.4907259746115356e-06, "loss": 0.4442, "step": 5980 }, { "epoch": 1.9137158703071673, "grad_norm": 0.5815134644508362, "learning_rate": 3.4889511298716634e-06, "loss": 0.5452, "step": 5981 }, { "epoch": 1.9140358361774745, "grad_norm": 0.5238719582557678, "learning_rate": 3.487176494640533e-06, "loss": 0.4666, "step": 5982 }, { "epoch": 1.9143558020477816, "grad_norm": 0.5438646078109741, "learning_rate": 3.485402069164203e-06, "loss": 0.5087, "step": 5983 }, { "epoch": 1.9146757679180886, "grad_norm": 0.5344850420951843, "learning_rate": 3.4836278536886964e-06, "loss": 0.4674, "step": 5984 }, { "epoch": 1.914995733788396, "grad_norm": 0.5671039819717407, "learning_rate": 3.481853848460013e-06, "loss": 0.5366, "step": 5985 }, { "epoch": 1.915315699658703, "grad_norm": 0.4894713759422302, "learning_rate": 3.4800800537241207e-06, "loss": 0.4418, "step": 5986 }, { "epoch": 1.91563566552901, "grad_norm": 0.5043044090270996, "learning_rate": 3.478306469726957e-06, "loss": 0.5091, "step": 5987 }, { "epoch": 1.9159556313993176, "grad_norm": 0.5146811008453369, "learning_rate": 3.4765330967144328e-06, "loss": 0.4145, "step": 5988 }, { "epoch": 1.9162755972696246, "grad_norm": 0.5499746203422546, "learning_rate": 3.4747599349324248e-06, "loss": 0.5641, "step": 5989 }, { "epoch": 1.9165955631399316, "grad_norm": 0.536591112613678, "learning_rate": 3.472986984626788e-06, "loss": 0.4706, "step": 5990 }, { "epoch": 1.9169155290102389, "grad_norm": 0.566261351108551, "learning_rate": 3.471214246043344e-06, "loss": 0.5222, "step": 5991 }, { "epoch": 1.9172354948805461, "grad_norm": 0.5134425759315491, "learning_rate": 3.4694417194278834e-06, "loss": 0.506, "step": 5992 }, { "epoch": 1.9175554607508531, "grad_norm": 0.4592892825603485, "learning_rate": 3.4676694050261684e-06, "loss": 0.4737, "step": 5993 }, { "epoch": 1.9178754266211604, "grad_norm": 0.5180602073669434, "learning_rate": 3.4658973030839357e-06, "loss": 0.4904, "step": 5994 }, { "epoch": 1.9181953924914676, "grad_norm": 0.5318039059638977, "learning_rate": 3.464125413846886e-06, "loss": 0.4827, "step": 5995 }, { "epoch": 1.9185153583617747, "grad_norm": 0.5263141393661499, "learning_rate": 3.4623537375606976e-06, "loss": 0.4475, "step": 5996 }, { "epoch": 1.918835324232082, "grad_norm": 0.539147138595581, "learning_rate": 3.4605822744710115e-06, "loss": 0.548, "step": 5997 }, { "epoch": 1.9191552901023892, "grad_norm": 0.48374414443969727, "learning_rate": 3.458811024823444e-06, "loss": 0.4516, "step": 5998 }, { "epoch": 1.9194752559726962, "grad_norm": 0.48777681589126587, "learning_rate": 3.4570399888635843e-06, "loss": 0.5028, "step": 5999 }, { "epoch": 1.9197952218430034, "grad_norm": 0.5044283270835876, "learning_rate": 3.455269166836985e-06, "loss": 0.4647, "step": 6000 }, { "epoch": 1.9201151877133107, "grad_norm": 0.4887102544307709, "learning_rate": 3.453498558989175e-06, "loss": 0.4405, "step": 6001 }, { "epoch": 1.9204351535836177, "grad_norm": 0.494515985250473, "learning_rate": 3.4517281655656498e-06, "loss": 0.5235, "step": 6002 }, { "epoch": 1.920755119453925, "grad_norm": 0.5164315104484558, "learning_rate": 3.449957986811876e-06, "loss": 0.479, "step": 6003 }, { "epoch": 1.9210750853242322, "grad_norm": 0.4746878147125244, "learning_rate": 3.448188022973293e-06, "loss": 0.4803, "step": 6004 }, { "epoch": 1.9213950511945392, "grad_norm": 0.5262110233306885, "learning_rate": 3.446418274295308e-06, "loss": 0.4984, "step": 6005 }, { "epoch": 1.9217150170648463, "grad_norm": 0.5051013231277466, "learning_rate": 3.4446487410232986e-06, "loss": 0.4641, "step": 6006 }, { "epoch": 1.9220349829351537, "grad_norm": 0.48198986053466797, "learning_rate": 3.4428794234026114e-06, "loss": 0.5024, "step": 6007 }, { "epoch": 1.9223549488054608, "grad_norm": 0.4754115343093872, "learning_rate": 3.4411103216785645e-06, "loss": 0.4071, "step": 6008 }, { "epoch": 1.9226749146757678, "grad_norm": 0.5091566443443298, "learning_rate": 3.439341436096449e-06, "loss": 0.5263, "step": 6009 }, { "epoch": 1.922994880546075, "grad_norm": 0.45499107241630554, "learning_rate": 3.4375727669015196e-06, "loss": 0.4544, "step": 6010 }, { "epoch": 1.9233148464163823, "grad_norm": 0.5030335187911987, "learning_rate": 3.4358043143390067e-06, "loss": 0.5287, "step": 6011 }, { "epoch": 1.9236348122866893, "grad_norm": 0.4837437570095062, "learning_rate": 3.4340360786541067e-06, "loss": 0.4994, "step": 6012 }, { "epoch": 1.9239547781569966, "grad_norm": 0.4857082664966583, "learning_rate": 3.432268060091987e-06, "loss": 0.4865, "step": 6013 }, { "epoch": 1.9242747440273038, "grad_norm": 0.4972582161426544, "learning_rate": 3.4305002588977888e-06, "loss": 0.4919, "step": 6014 }, { "epoch": 1.9245947098976108, "grad_norm": 0.5630078911781311, "learning_rate": 3.428732675316617e-06, "loss": 0.5211, "step": 6015 }, { "epoch": 1.924914675767918, "grad_norm": 0.48093336820602417, "learning_rate": 3.4269653095935514e-06, "loss": 0.4372, "step": 6016 }, { "epoch": 1.9252346416382253, "grad_norm": 0.5052854418754578, "learning_rate": 3.4251981619736363e-06, "loss": 0.4873, "step": 6017 }, { "epoch": 1.9255546075085324, "grad_norm": 0.5380025506019592, "learning_rate": 3.4234312327018904e-06, "loss": 0.5225, "step": 6018 }, { "epoch": 1.9258745733788396, "grad_norm": 0.49216553568840027, "learning_rate": 3.4216645220233023e-06, "loss": 0.4527, "step": 6019 }, { "epoch": 1.9261945392491469, "grad_norm": 0.5159412026405334, "learning_rate": 3.4198980301828256e-06, "loss": 0.4326, "step": 6020 }, { "epoch": 1.9265145051194539, "grad_norm": 0.5136810541152954, "learning_rate": 3.418131757425389e-06, "loss": 0.5023, "step": 6021 }, { "epoch": 1.9268344709897611, "grad_norm": 0.5028185248374939, "learning_rate": 3.4163657039958874e-06, "loss": 0.4665, "step": 6022 }, { "epoch": 1.9271544368600684, "grad_norm": 0.5626875162124634, "learning_rate": 3.4145998701391837e-06, "loss": 0.5032, "step": 6023 }, { "epoch": 1.9274744027303754, "grad_norm": 0.5647227168083191, "learning_rate": 3.4128342561001184e-06, "loss": 0.4871, "step": 6024 }, { "epoch": 1.9277943686006824, "grad_norm": 0.498357892036438, "learning_rate": 3.4110688621234912e-06, "loss": 0.4873, "step": 6025 }, { "epoch": 1.92811433447099, "grad_norm": 0.5906538367271423, "learning_rate": 3.4093036884540786e-06, "loss": 0.5337, "step": 6026 }, { "epoch": 1.928434300341297, "grad_norm": 0.4693874716758728, "learning_rate": 3.4075387353366206e-06, "loss": 0.4774, "step": 6027 }, { "epoch": 1.928754266211604, "grad_norm": 0.4883122742176056, "learning_rate": 3.405774003015836e-06, "loss": 0.4812, "step": 6028 }, { "epoch": 1.9290742320819114, "grad_norm": 0.5262741446495056, "learning_rate": 3.4040094917364026e-06, "loss": 0.4612, "step": 6029 }, { "epoch": 1.9293941979522184, "grad_norm": 0.5191846489906311, "learning_rate": 3.4022452017429743e-06, "loss": 0.4578, "step": 6030 }, { "epoch": 1.9297141638225255, "grad_norm": 0.5533474683761597, "learning_rate": 3.400481133280171e-06, "loss": 0.4828, "step": 6031 }, { "epoch": 1.9300341296928327, "grad_norm": 0.513857364654541, "learning_rate": 3.3987172865925787e-06, "loss": 0.4628, "step": 6032 }, { "epoch": 1.93035409556314, "grad_norm": 0.5177400708198547, "learning_rate": 3.3969536619247658e-06, "loss": 0.4707, "step": 6033 }, { "epoch": 1.930674061433447, "grad_norm": 0.5255177617073059, "learning_rate": 3.395190259521254e-06, "loss": 0.5129, "step": 6034 }, { "epoch": 1.9309940273037542, "grad_norm": 0.44974416494369507, "learning_rate": 3.3934270796265444e-06, "loss": 0.4809, "step": 6035 }, { "epoch": 1.9313139931740615, "grad_norm": 0.48602941632270813, "learning_rate": 3.391664122485103e-06, "loss": 0.5041, "step": 6036 }, { "epoch": 1.9316339590443685, "grad_norm": 0.4708314836025238, "learning_rate": 3.3899013883413635e-06, "loss": 0.4475, "step": 6037 }, { "epoch": 1.9319539249146758, "grad_norm": 0.49465641379356384, "learning_rate": 3.388138877439736e-06, "loss": 0.5045, "step": 6038 }, { "epoch": 1.932273890784983, "grad_norm": 0.525989830493927, "learning_rate": 3.3863765900245914e-06, "loss": 0.5234, "step": 6039 }, { "epoch": 1.93259385665529, "grad_norm": 0.5063377022743225, "learning_rate": 3.3846145263402744e-06, "loss": 0.4521, "step": 6040 }, { "epoch": 1.9329138225255973, "grad_norm": 0.4842839241027832, "learning_rate": 3.3828526866310962e-06, "loss": 0.4912, "step": 6041 }, { "epoch": 1.9332337883959045, "grad_norm": 0.49538734555244446, "learning_rate": 3.3810910711413376e-06, "loss": 0.5095, "step": 6042 }, { "epoch": 1.9335537542662116, "grad_norm": 0.5384979844093323, "learning_rate": 3.3793296801152508e-06, "loss": 0.4735, "step": 6043 }, { "epoch": 1.9338737201365188, "grad_norm": 0.5058707594871521, "learning_rate": 3.377568513797054e-06, "loss": 0.4972, "step": 6044 }, { "epoch": 1.934193686006826, "grad_norm": 0.5415966510772705, "learning_rate": 3.3758075724309342e-06, "loss": 0.4945, "step": 6045 }, { "epoch": 1.934513651877133, "grad_norm": 0.5162762403488159, "learning_rate": 3.374046856261048e-06, "loss": 0.4969, "step": 6046 }, { "epoch": 1.9348336177474401, "grad_norm": 0.516521692276001, "learning_rate": 3.37228636553152e-06, "loss": 0.4676, "step": 6047 }, { "epoch": 1.9351535836177476, "grad_norm": 0.4812181293964386, "learning_rate": 3.3705261004864476e-06, "loss": 0.4643, "step": 6048 }, { "epoch": 1.9354735494880546, "grad_norm": 0.5185404419898987, "learning_rate": 3.3687660613698907e-06, "loss": 0.4941, "step": 6049 }, { "epoch": 1.9357935153583616, "grad_norm": 0.558539092540741, "learning_rate": 3.367006248425883e-06, "loss": 0.5026, "step": 6050 }, { "epoch": 1.9361134812286689, "grad_norm": 0.47996214032173157, "learning_rate": 3.3652466618984226e-06, "loss": 0.4369, "step": 6051 }, { "epoch": 1.9364334470989761, "grad_norm": 0.5430442690849304, "learning_rate": 3.363487302031478e-06, "loss": 0.5251, "step": 6052 }, { "epoch": 1.9367534129692832, "grad_norm": 0.5165271162986755, "learning_rate": 3.3617281690689895e-06, "loss": 0.4804, "step": 6053 }, { "epoch": 1.9370733788395904, "grad_norm": 0.5100358128547668, "learning_rate": 3.359969263254861e-06, "loss": 0.4563, "step": 6054 }, { "epoch": 1.9373933447098977, "grad_norm": 0.5728858709335327, "learning_rate": 3.3582105848329677e-06, "loss": 0.5041, "step": 6055 }, { "epoch": 1.9377133105802047, "grad_norm": 0.5265145301818848, "learning_rate": 3.3564521340471507e-06, "loss": 0.459, "step": 6056 }, { "epoch": 1.938033276450512, "grad_norm": 0.5253676772117615, "learning_rate": 3.354693911141222e-06, "loss": 0.5262, "step": 6057 }, { "epoch": 1.9383532423208192, "grad_norm": 0.49260178208351135, "learning_rate": 3.3529359163589636e-06, "loss": 0.4921, "step": 6058 }, { "epoch": 1.9386732081911262, "grad_norm": 0.47888457775115967, "learning_rate": 3.3511781499441207e-06, "loss": 0.5452, "step": 6059 }, { "epoch": 1.9389931740614335, "grad_norm": 0.5283399224281311, "learning_rate": 3.349420612140412e-06, "loss": 0.4636, "step": 6060 }, { "epoch": 1.9393131399317407, "grad_norm": 0.5486240386962891, "learning_rate": 3.3476633031915194e-06, "loss": 0.5057, "step": 6061 }, { "epoch": 1.9396331058020477, "grad_norm": 0.4765787422657013, "learning_rate": 3.3459062233410983e-06, "loss": 0.4171, "step": 6062 }, { "epoch": 1.939953071672355, "grad_norm": 0.5192815661430359, "learning_rate": 3.34414937283277e-06, "loss": 0.4798, "step": 6063 }, { "epoch": 1.9402730375426622, "grad_norm": 0.5335144996643066, "learning_rate": 3.342392751910123e-06, "loss": 0.4964, "step": 6064 }, { "epoch": 1.9405930034129693, "grad_norm": 0.49786290526390076, "learning_rate": 3.340636360816716e-06, "loss": 0.4673, "step": 6065 }, { "epoch": 1.9409129692832765, "grad_norm": 0.4714604318141937, "learning_rate": 3.3388801997960714e-06, "loss": 0.4626, "step": 6066 }, { "epoch": 1.9412329351535837, "grad_norm": 0.4573826789855957, "learning_rate": 3.3371242690916863e-06, "loss": 0.4782, "step": 6067 }, { "epoch": 1.9415529010238908, "grad_norm": 0.4418972432613373, "learning_rate": 3.335368568947023e-06, "loss": 0.4485, "step": 6068 }, { "epoch": 1.9418728668941978, "grad_norm": 0.5344722270965576, "learning_rate": 3.3336130996055095e-06, "loss": 0.4972, "step": 6069 }, { "epoch": 1.9421928327645053, "grad_norm": 0.49242472648620605, "learning_rate": 3.3318578613105453e-06, "loss": 0.4598, "step": 6070 }, { "epoch": 1.9425127986348123, "grad_norm": 0.49834853410720825, "learning_rate": 3.3301028543054935e-06, "loss": 0.4652, "step": 6071 }, { "epoch": 1.9428327645051193, "grad_norm": 0.4919757843017578, "learning_rate": 3.328348078833691e-06, "loss": 0.4946, "step": 6072 }, { "epoch": 1.9431527303754266, "grad_norm": 0.5249433517456055, "learning_rate": 3.3265935351384386e-06, "loss": 0.4773, "step": 6073 }, { "epoch": 1.9434726962457338, "grad_norm": 0.45385199785232544, "learning_rate": 3.3248392234630056e-06, "loss": 0.4416, "step": 6074 }, { "epoch": 1.9437926621160408, "grad_norm": 0.5273494720458984, "learning_rate": 3.32308514405063e-06, "loss": 0.5478, "step": 6075 }, { "epoch": 1.944112627986348, "grad_norm": 0.4976264238357544, "learning_rate": 3.321331297144515e-06, "loss": 0.4453, "step": 6076 }, { "epoch": 1.9444325938566553, "grad_norm": 0.5137234926223755, "learning_rate": 3.3195776829878357e-06, "loss": 0.483, "step": 6077 }, { "epoch": 1.9447525597269624, "grad_norm": 0.5299399495124817, "learning_rate": 3.3178243018237334e-06, "loss": 0.5094, "step": 6078 }, { "epoch": 1.9450725255972696, "grad_norm": 0.4810827970504761, "learning_rate": 3.316071153895315e-06, "loss": 0.4915, "step": 6079 }, { "epoch": 1.9453924914675769, "grad_norm": 0.49671903252601624, "learning_rate": 3.314318239445657e-06, "loss": 0.4534, "step": 6080 }, { "epoch": 1.945712457337884, "grad_norm": 0.4928573966026306, "learning_rate": 3.312565558717801e-06, "loss": 0.5293, "step": 6081 }, { "epoch": 1.9460324232081911, "grad_norm": 0.49744442105293274, "learning_rate": 3.310813111954762e-06, "loss": 0.5188, "step": 6082 }, { "epoch": 1.9463523890784984, "grad_norm": 0.4996615946292877, "learning_rate": 3.309060899399518e-06, "loss": 0.4572, "step": 6083 }, { "epoch": 1.9466723549488054, "grad_norm": 0.48610150814056396, "learning_rate": 3.3073089212950137e-06, "loss": 0.46, "step": 6084 }, { "epoch": 1.9469923208191127, "grad_norm": 0.47651350498199463, "learning_rate": 3.3055571778841645e-06, "loss": 0.4671, "step": 6085 }, { "epoch": 1.94731228668942, "grad_norm": 0.5219941735267639, "learning_rate": 3.3038056694098485e-06, "loss": 0.4608, "step": 6086 }, { "epoch": 1.947632252559727, "grad_norm": 0.5291367769241333, "learning_rate": 3.302054396114919e-06, "loss": 0.468, "step": 6087 }, { "epoch": 1.947952218430034, "grad_norm": 0.5114938616752625, "learning_rate": 3.3003033582421905e-06, "loss": 0.4465, "step": 6088 }, { "epoch": 1.9482721843003414, "grad_norm": 0.5421054363250732, "learning_rate": 3.2985525560344458e-06, "loss": 0.5237, "step": 6089 }, { "epoch": 1.9485921501706485, "grad_norm": 0.5172793865203857, "learning_rate": 3.2968019897344367e-06, "loss": 0.4778, "step": 6090 }, { "epoch": 1.9489121160409555, "grad_norm": 0.523431122303009, "learning_rate": 3.2950516595848785e-06, "loss": 0.4951, "step": 6091 }, { "epoch": 1.949232081911263, "grad_norm": 0.5439819097518921, "learning_rate": 3.29330156582846e-06, "loss": 0.528, "step": 6092 }, { "epoch": 1.94955204778157, "grad_norm": 0.4954051673412323, "learning_rate": 3.291551708707834e-06, "loss": 0.5023, "step": 6093 }, { "epoch": 1.949872013651877, "grad_norm": 0.5080917477607727, "learning_rate": 3.289802088465617e-06, "loss": 0.4355, "step": 6094 }, { "epoch": 1.9501919795221843, "grad_norm": 0.6108106374740601, "learning_rate": 3.2880527053444e-06, "loss": 0.5668, "step": 6095 }, { "epoch": 1.9505119453924915, "grad_norm": 0.6188663840293884, "learning_rate": 3.286303559586732e-06, "loss": 0.5193, "step": 6096 }, { "epoch": 1.9508319112627985, "grad_norm": 0.5643511414527893, "learning_rate": 3.2845546514351384e-06, "loss": 0.4728, "step": 6097 }, { "epoch": 1.9511518771331058, "grad_norm": 0.5128049254417419, "learning_rate": 3.282805981132107e-06, "loss": 0.441, "step": 6098 }, { "epoch": 1.951471843003413, "grad_norm": 0.52104651927948, "learning_rate": 3.281057548920091e-06, "loss": 0.5237, "step": 6099 }, { "epoch": 1.95179180887372, "grad_norm": 0.565772533416748, "learning_rate": 3.2793093550415133e-06, "loss": 0.5355, "step": 6100 }, { "epoch": 1.9521117747440273, "grad_norm": 0.5251613259315491, "learning_rate": 3.2775613997387656e-06, "loss": 0.4681, "step": 6101 }, { "epoch": 1.9524317406143346, "grad_norm": 0.5024871826171875, "learning_rate": 3.2758136832542008e-06, "loss": 0.4825, "step": 6102 }, { "epoch": 1.9527517064846416, "grad_norm": 0.5466561317443848, "learning_rate": 3.274066205830144e-06, "loss": 0.5212, "step": 6103 }, { "epoch": 1.9530716723549488, "grad_norm": 0.5681954622268677, "learning_rate": 3.272318967708883e-06, "loss": 0.5492, "step": 6104 }, { "epoch": 1.953391638225256, "grad_norm": 0.5645036101341248, "learning_rate": 3.270571969132674e-06, "loss": 0.533, "step": 6105 }, { "epoch": 1.953711604095563, "grad_norm": 0.535334587097168, "learning_rate": 3.2688252103437456e-06, "loss": 0.4735, "step": 6106 }, { "epoch": 1.9540315699658704, "grad_norm": 0.4753856360912323, "learning_rate": 3.2670786915842824e-06, "loss": 0.4832, "step": 6107 }, { "epoch": 1.9543515358361776, "grad_norm": 0.4911159873008728, "learning_rate": 3.265332413096444e-06, "loss": 0.4495, "step": 6108 }, { "epoch": 1.9546715017064846, "grad_norm": 0.5127443075180054, "learning_rate": 3.263586375122353e-06, "loss": 0.5102, "step": 6109 }, { "epoch": 1.9549914675767917, "grad_norm": 0.4534604251384735, "learning_rate": 3.2618405779040992e-06, "loss": 0.4494, "step": 6110 }, { "epoch": 1.9553114334470991, "grad_norm": 0.4901772737503052, "learning_rate": 3.2600950216837423e-06, "loss": 0.5097, "step": 6111 }, { "epoch": 1.9556313993174061, "grad_norm": 0.48854494094848633, "learning_rate": 3.258349706703302e-06, "loss": 0.414, "step": 6112 }, { "epoch": 1.9559513651877132, "grad_norm": 0.5062629580497742, "learning_rate": 3.256604633204772e-06, "loss": 0.4779, "step": 6113 }, { "epoch": 1.9562713310580204, "grad_norm": 0.5243249535560608, "learning_rate": 3.2548598014301057e-06, "loss": 0.5221, "step": 6114 }, { "epoch": 1.9565912969283277, "grad_norm": 0.5110526084899902, "learning_rate": 3.2531152116212263e-06, "loss": 0.4385, "step": 6115 }, { "epoch": 1.9569112627986347, "grad_norm": 0.47971436381340027, "learning_rate": 3.251370864020026e-06, "loss": 0.477, "step": 6116 }, { "epoch": 1.957231228668942, "grad_norm": 0.4537786543369293, "learning_rate": 3.2496267588683583e-06, "loss": 0.4254, "step": 6117 }, { "epoch": 1.9575511945392492, "grad_norm": 0.5468919277191162, "learning_rate": 3.2478828964080476e-06, "loss": 0.4988, "step": 6118 }, { "epoch": 1.9578711604095562, "grad_norm": 0.5255555510520935, "learning_rate": 3.24613927688088e-06, "loss": 0.4974, "step": 6119 }, { "epoch": 1.9581911262798635, "grad_norm": 0.5240233540534973, "learning_rate": 3.2443959005286106e-06, "loss": 0.5151, "step": 6120 }, { "epoch": 1.9585110921501707, "grad_norm": 0.4944935142993927, "learning_rate": 3.2426527675929635e-06, "loss": 0.4733, "step": 6121 }, { "epoch": 1.9588310580204777, "grad_norm": 0.5280991792678833, "learning_rate": 3.2409098783156233e-06, "loss": 0.5099, "step": 6122 }, { "epoch": 1.959151023890785, "grad_norm": 0.5345900058746338, "learning_rate": 3.239167232938246e-06, "loss": 0.5017, "step": 6123 }, { "epoch": 1.9594709897610922, "grad_norm": 0.49158307909965515, "learning_rate": 3.237424831702449e-06, "loss": 0.4864, "step": 6124 }, { "epoch": 1.9597909556313993, "grad_norm": 0.5014450550079346, "learning_rate": 3.2356826748498182e-06, "loss": 0.4501, "step": 6125 }, { "epoch": 1.9601109215017065, "grad_norm": 0.513562023639679, "learning_rate": 3.2339407626219097e-06, "loss": 0.5151, "step": 6126 }, { "epoch": 1.9604308873720138, "grad_norm": 0.5434235334396362, "learning_rate": 3.2321990952602377e-06, "loss": 0.4818, "step": 6127 }, { "epoch": 1.9607508532423208, "grad_norm": 0.5348495841026306, "learning_rate": 3.2304576730062887e-06, "loss": 0.5835, "step": 6128 }, { "epoch": 1.961070819112628, "grad_norm": 0.5259645581245422, "learning_rate": 3.2287164961015116e-06, "loss": 0.4939, "step": 6129 }, { "epoch": 1.9613907849829353, "grad_norm": 0.5705592632293701, "learning_rate": 3.226975564787322e-06, "loss": 0.5378, "step": 6130 }, { "epoch": 1.9617107508532423, "grad_norm": 0.4692423343658447, "learning_rate": 3.225234879305105e-06, "loss": 0.4094, "step": 6131 }, { "epoch": 1.9620307167235493, "grad_norm": 0.5225099325180054, "learning_rate": 3.2234944398962066e-06, "loss": 0.481, "step": 6132 }, { "epoch": 1.9623506825938568, "grad_norm": 0.49818119406700134, "learning_rate": 3.2217542468019425e-06, "loss": 0.5051, "step": 6133 }, { "epoch": 1.9626706484641638, "grad_norm": 0.5146864652633667, "learning_rate": 3.2200143002635885e-06, "loss": 0.5107, "step": 6134 }, { "epoch": 1.9629906143344709, "grad_norm": 0.46393853425979614, "learning_rate": 3.218274600522395e-06, "loss": 0.4544, "step": 6135 }, { "epoch": 1.963310580204778, "grad_norm": 0.5225887298583984, "learning_rate": 3.2165351478195715e-06, "loss": 0.5056, "step": 6136 }, { "epoch": 1.9636305460750854, "grad_norm": 0.4513947665691376, "learning_rate": 3.2147959423962954e-06, "loss": 0.4366, "step": 6137 }, { "epoch": 1.9639505119453924, "grad_norm": 0.4780377149581909, "learning_rate": 3.2130569844937097e-06, "loss": 0.5216, "step": 6138 }, { "epoch": 1.9642704778156996, "grad_norm": 0.5713197588920593, "learning_rate": 3.2113182743529214e-06, "loss": 0.4782, "step": 6139 }, { "epoch": 1.9645904436860069, "grad_norm": 0.4937642216682434, "learning_rate": 3.2095798122150072e-06, "loss": 0.4634, "step": 6140 }, { "epoch": 1.964910409556314, "grad_norm": 0.497175931930542, "learning_rate": 3.207841598321007e-06, "loss": 0.5058, "step": 6141 }, { "epoch": 1.9652303754266212, "grad_norm": 0.5064600110054016, "learning_rate": 3.2061036329119243e-06, "loss": 0.5039, "step": 6142 }, { "epoch": 1.9655503412969284, "grad_norm": 0.47617995738983154, "learning_rate": 3.2043659162287317e-06, "loss": 0.4448, "step": 6143 }, { "epoch": 1.9658703071672354, "grad_norm": 0.4566115438938141, "learning_rate": 3.2026284485123625e-06, "loss": 0.4608, "step": 6144 }, { "epoch": 1.9661902730375427, "grad_norm": 0.5149037837982178, "learning_rate": 3.200891230003723e-06, "loss": 0.4618, "step": 6145 }, { "epoch": 1.96651023890785, "grad_norm": 0.5159351825714111, "learning_rate": 3.1991542609436797e-06, "loss": 0.4791, "step": 6146 }, { "epoch": 1.966830204778157, "grad_norm": 0.5528331995010376, "learning_rate": 3.1974175415730636e-06, "loss": 0.5118, "step": 6147 }, { "epoch": 1.9671501706484642, "grad_norm": 0.5175594687461853, "learning_rate": 3.1956810721326747e-06, "loss": 0.4688, "step": 6148 }, { "epoch": 1.9674701365187715, "grad_norm": 0.47869059443473816, "learning_rate": 3.1939448528632745e-06, "loss": 0.4518, "step": 6149 }, { "epoch": 1.9677901023890785, "grad_norm": 0.48490244150161743, "learning_rate": 3.192208884005593e-06, "loss": 0.4598, "step": 6150 }, { "epoch": 1.9681100682593855, "grad_norm": 0.5457737445831299, "learning_rate": 3.1904731658003264e-06, "loss": 0.5035, "step": 6151 }, { "epoch": 1.968430034129693, "grad_norm": 0.4673367738723755, "learning_rate": 3.1887376984881313e-06, "loss": 0.4317, "step": 6152 }, { "epoch": 1.96875, "grad_norm": 0.5494493246078491, "learning_rate": 3.187002482309634e-06, "loss": 0.5019, "step": 6153 }, { "epoch": 1.969069965870307, "grad_norm": 0.5415604710578918, "learning_rate": 3.1852675175054216e-06, "loss": 0.464, "step": 6154 }, { "epoch": 1.9693899317406145, "grad_norm": 0.5042742490768433, "learning_rate": 3.1835328043160522e-06, "loss": 0.5252, "step": 6155 }, { "epoch": 1.9697098976109215, "grad_norm": 0.4678550362586975, "learning_rate": 3.181798342982046e-06, "loss": 0.4642, "step": 6156 }, { "epoch": 1.9700298634812285, "grad_norm": 0.5744708180427551, "learning_rate": 3.180064133743885e-06, "loss": 0.5409, "step": 6157 }, { "epoch": 1.9703498293515358, "grad_norm": 0.46229907870292664, "learning_rate": 3.178330176842023e-06, "loss": 0.4953, "step": 6158 }, { "epoch": 1.970669795221843, "grad_norm": 0.5091956257820129, "learning_rate": 3.176596472516872e-06, "loss": 0.5184, "step": 6159 }, { "epoch": 1.97098976109215, "grad_norm": 0.5130820870399475, "learning_rate": 3.1748630210088137e-06, "loss": 0.4752, "step": 6160 }, { "epoch": 1.9713097269624573, "grad_norm": 0.5606593489646912, "learning_rate": 3.173129822558195e-06, "loss": 0.4496, "step": 6161 }, { "epoch": 1.9716296928327646, "grad_norm": 0.4711438715457916, "learning_rate": 3.171396877405323e-06, "loss": 0.4357, "step": 6162 }, { "epoch": 1.9719496587030716, "grad_norm": 0.5281607508659363, "learning_rate": 3.1696641857904743e-06, "loss": 0.5107, "step": 6163 }, { "epoch": 1.9722696245733788, "grad_norm": 0.494859904050827, "learning_rate": 3.1679317479538864e-06, "loss": 0.4721, "step": 6164 }, { "epoch": 1.972589590443686, "grad_norm": 0.47621551156044006, "learning_rate": 3.1661995641357666e-06, "loss": 0.5002, "step": 6165 }, { "epoch": 1.9729095563139931, "grad_norm": 0.5177987217903137, "learning_rate": 3.164467634576285e-06, "loss": 0.4769, "step": 6166 }, { "epoch": 1.9732295221843004, "grad_norm": 0.5293196439743042, "learning_rate": 3.1627359595155728e-06, "loss": 0.5277, "step": 6167 }, { "epoch": 1.9735494880546076, "grad_norm": 0.5151974558830261, "learning_rate": 3.1610045391937282e-06, "loss": 0.5385, "step": 6168 }, { "epoch": 1.9738694539249146, "grad_norm": 0.5350662469863892, "learning_rate": 3.1592733738508186e-06, "loss": 0.4679, "step": 6169 }, { "epoch": 1.974189419795222, "grad_norm": 0.5367306470870972, "learning_rate": 3.157542463726871e-06, "loss": 0.5163, "step": 6170 }, { "epoch": 1.9745093856655291, "grad_norm": 0.5237488746643066, "learning_rate": 3.1558118090618773e-06, "loss": 0.511, "step": 6171 }, { "epoch": 1.9748293515358362, "grad_norm": 0.47949159145355225, "learning_rate": 3.154081410095793e-06, "loss": 0.435, "step": 6172 }, { "epoch": 1.9751493174061432, "grad_norm": 0.5170743465423584, "learning_rate": 3.152351267068542e-06, "loss": 0.515, "step": 6173 }, { "epoch": 1.9754692832764507, "grad_norm": 0.5387949347496033, "learning_rate": 3.1506213802200113e-06, "loss": 0.4572, "step": 6174 }, { "epoch": 1.9757892491467577, "grad_norm": 0.5387278199195862, "learning_rate": 3.1488917497900523e-06, "loss": 0.562, "step": 6175 }, { "epoch": 1.9761092150170647, "grad_norm": 0.5308703184127808, "learning_rate": 3.147162376018479e-06, "loss": 0.5108, "step": 6176 }, { "epoch": 1.976429180887372, "grad_norm": 0.4508434534072876, "learning_rate": 3.1454332591450697e-06, "loss": 0.4572, "step": 6177 }, { "epoch": 1.9767491467576792, "grad_norm": 0.4709256887435913, "learning_rate": 3.1437043994095697e-06, "loss": 0.4828, "step": 6178 }, { "epoch": 1.9770691126279862, "grad_norm": 0.541800856590271, "learning_rate": 3.141975797051689e-06, "loss": 0.533, "step": 6179 }, { "epoch": 1.9773890784982935, "grad_norm": 0.4995775818824768, "learning_rate": 3.1402474523111006e-06, "loss": 0.5028, "step": 6180 }, { "epoch": 1.9777090443686007, "grad_norm": 0.4512108564376831, "learning_rate": 3.138519365427438e-06, "loss": 0.4653, "step": 6181 }, { "epoch": 1.9780290102389078, "grad_norm": 0.541901171207428, "learning_rate": 3.136791536640307e-06, "loss": 0.5351, "step": 6182 }, { "epoch": 1.978348976109215, "grad_norm": 0.47517821192741394, "learning_rate": 3.135063966189268e-06, "loss": 0.4567, "step": 6183 }, { "epoch": 1.9786689419795223, "grad_norm": 0.5447357296943665, "learning_rate": 3.1333366543138557e-06, "loss": 0.525, "step": 6184 }, { "epoch": 1.9789889078498293, "grad_norm": 0.5360710024833679, "learning_rate": 3.1316096012535625e-06, "loss": 0.4772, "step": 6185 }, { "epoch": 1.9793088737201365, "grad_norm": 0.5403434634208679, "learning_rate": 3.1298828072478443e-06, "loss": 0.5056, "step": 6186 }, { "epoch": 1.9796288395904438, "grad_norm": 0.5185534358024597, "learning_rate": 3.128156272536126e-06, "loss": 0.4904, "step": 6187 }, { "epoch": 1.9799488054607508, "grad_norm": 0.4958355128765106, "learning_rate": 3.1264299973577906e-06, "loss": 0.4628, "step": 6188 }, { "epoch": 1.980268771331058, "grad_norm": 0.4718395173549652, "learning_rate": 3.1247039819521907e-06, "loss": 0.508, "step": 6189 }, { "epoch": 1.9805887372013653, "grad_norm": 0.4683765769004822, "learning_rate": 3.122978226558641e-06, "loss": 0.4784, "step": 6190 }, { "epoch": 1.9809087030716723, "grad_norm": 0.4669650197029114, "learning_rate": 3.1212527314164177e-06, "loss": 0.4133, "step": 6191 }, { "epoch": 1.9812286689419796, "grad_norm": 0.5345683097839355, "learning_rate": 3.1195274967647638e-06, "loss": 0.5364, "step": 6192 }, { "epoch": 1.9815486348122868, "grad_norm": 0.4708634912967682, "learning_rate": 3.1178025228428826e-06, "loss": 0.4594, "step": 6193 }, { "epoch": 1.9818686006825939, "grad_norm": 0.5484588742256165, "learning_rate": 3.116077809889947e-06, "loss": 0.5032, "step": 6194 }, { "epoch": 1.9821885665529009, "grad_norm": 0.5436263084411621, "learning_rate": 3.1143533581450906e-06, "loss": 0.5252, "step": 6195 }, { "epoch": 1.9825085324232083, "grad_norm": 0.5184581875801086, "learning_rate": 3.112629167847409e-06, "loss": 0.4339, "step": 6196 }, { "epoch": 1.9828284982935154, "grad_norm": 0.5556897521018982, "learning_rate": 3.1109052392359645e-06, "loss": 0.5362, "step": 6197 }, { "epoch": 1.9831484641638224, "grad_norm": 0.44339731335639954, "learning_rate": 3.1091815725497793e-06, "loss": 0.4365, "step": 6198 }, { "epoch": 1.9834684300341296, "grad_norm": 0.5135619640350342, "learning_rate": 3.1074581680278447e-06, "loss": 0.5121, "step": 6199 }, { "epoch": 1.983788395904437, "grad_norm": 0.5377285480499268, "learning_rate": 3.1057350259091125e-06, "loss": 0.4856, "step": 6200 }, { "epoch": 1.984108361774744, "grad_norm": 0.43569591641426086, "learning_rate": 3.104012146432497e-06, "loss": 0.4141, "step": 6201 }, { "epoch": 1.9844283276450512, "grad_norm": 0.5048389434814453, "learning_rate": 3.10228952983688e-06, "loss": 0.4697, "step": 6202 }, { "epoch": 1.9847482935153584, "grad_norm": 0.506389856338501, "learning_rate": 3.1005671763611003e-06, "loss": 0.4795, "step": 6203 }, { "epoch": 1.9850682593856654, "grad_norm": 0.5272321105003357, "learning_rate": 3.0988450862439677e-06, "loss": 0.5171, "step": 6204 }, { "epoch": 1.9853882252559727, "grad_norm": 0.505204439163208, "learning_rate": 3.097123259724252e-06, "loss": 0.4919, "step": 6205 }, { "epoch": 1.98570819112628, "grad_norm": 0.5086266398429871, "learning_rate": 3.095401697040684e-06, "loss": 0.5065, "step": 6206 }, { "epoch": 1.986028156996587, "grad_norm": 0.525647759437561, "learning_rate": 3.0936803984319618e-06, "loss": 0.4548, "step": 6207 }, { "epoch": 1.9863481228668942, "grad_norm": 0.5006814002990723, "learning_rate": 3.0919593641367473e-06, "loss": 0.486, "step": 6208 }, { "epoch": 1.9866680887372015, "grad_norm": 0.47532498836517334, "learning_rate": 3.090238594393661e-06, "loss": 0.4429, "step": 6209 }, { "epoch": 1.9869880546075085, "grad_norm": 0.5055264830589294, "learning_rate": 3.088518089441293e-06, "loss": 0.4567, "step": 6210 }, { "epoch": 1.9873080204778157, "grad_norm": 0.5233423709869385, "learning_rate": 3.08679784951819e-06, "loss": 0.5191, "step": 6211 }, { "epoch": 1.987627986348123, "grad_norm": 0.4947076737880707, "learning_rate": 3.0850778748628656e-06, "loss": 0.442, "step": 6212 }, { "epoch": 1.98794795221843, "grad_norm": 0.5296012759208679, "learning_rate": 3.0833581657138e-06, "loss": 0.5251, "step": 6213 }, { "epoch": 1.988267918088737, "grad_norm": 0.4480663239955902, "learning_rate": 3.081638722309429e-06, "loss": 0.4544, "step": 6214 }, { "epoch": 1.9885878839590445, "grad_norm": 0.5297701358795166, "learning_rate": 3.0799195448881586e-06, "loss": 0.5031, "step": 6215 }, { "epoch": 1.9889078498293515, "grad_norm": 0.5122623443603516, "learning_rate": 3.078200633688352e-06, "loss": 0.4682, "step": 6216 }, { "epoch": 1.9892278156996586, "grad_norm": 0.5205539464950562, "learning_rate": 3.076481988948338e-06, "loss": 0.5313, "step": 6217 }, { "epoch": 1.989547781569966, "grad_norm": 0.5184388160705566, "learning_rate": 3.0747636109064126e-06, "loss": 0.5621, "step": 6218 }, { "epoch": 1.989867747440273, "grad_norm": 0.48881974816322327, "learning_rate": 3.073045499800828e-06, "loss": 0.4314, "step": 6219 }, { "epoch": 1.99018771331058, "grad_norm": 0.5957432389259338, "learning_rate": 3.071327655869804e-06, "loss": 0.5567, "step": 6220 }, { "epoch": 1.9905076791808873, "grad_norm": 0.5555700659751892, "learning_rate": 3.069610079351519e-06, "loss": 0.4873, "step": 6221 }, { "epoch": 1.9908276450511946, "grad_norm": 0.4987030327320099, "learning_rate": 3.067892770484119e-06, "loss": 0.4762, "step": 6222 }, { "epoch": 1.9911476109215016, "grad_norm": 0.5237892270088196, "learning_rate": 3.066175729505712e-06, "loss": 0.491, "step": 6223 }, { "epoch": 1.9914675767918089, "grad_norm": 0.46969321370124817, "learning_rate": 3.064458956654366e-06, "loss": 0.4579, "step": 6224 }, { "epoch": 1.991787542662116, "grad_norm": 0.5169256329536438, "learning_rate": 3.062742452168115e-06, "loss": 0.5175, "step": 6225 }, { "epoch": 1.9921075085324231, "grad_norm": 0.5002074837684631, "learning_rate": 3.061026216284952e-06, "loss": 0.4797, "step": 6226 }, { "epoch": 1.9924274744027304, "grad_norm": 0.49317920207977295, "learning_rate": 3.059310249242836e-06, "loss": 0.5041, "step": 6227 }, { "epoch": 1.9927474402730376, "grad_norm": 0.4308440387248993, "learning_rate": 3.057594551279691e-06, "loss": 0.4721, "step": 6228 }, { "epoch": 1.9930674061433447, "grad_norm": 0.49936333298683167, "learning_rate": 3.0558791226333974e-06, "loss": 0.527, "step": 6229 }, { "epoch": 1.993387372013652, "grad_norm": 0.534095048904419, "learning_rate": 3.054163963541803e-06, "loss": 0.4522, "step": 6230 }, { "epoch": 1.9937073378839592, "grad_norm": 0.5647724270820618, "learning_rate": 3.052449074242715e-06, "loss": 0.5229, "step": 6231 }, { "epoch": 1.9940273037542662, "grad_norm": 0.4887125492095947, "learning_rate": 3.050734454973904e-06, "loss": 0.4598, "step": 6232 }, { "epoch": 1.9943472696245734, "grad_norm": 0.5863474607467651, "learning_rate": 3.049020105973109e-06, "loss": 0.5209, "step": 6233 }, { "epoch": 1.9946672354948807, "grad_norm": 0.5121579170227051, "learning_rate": 3.0473060274780217e-06, "loss": 0.4353, "step": 6234 }, { "epoch": 1.9949872013651877, "grad_norm": 0.5075271129608154, "learning_rate": 3.0455922197263044e-06, "loss": 0.5176, "step": 6235 }, { "epoch": 1.9953071672354947, "grad_norm": 0.5416887402534485, "learning_rate": 3.0438786829555757e-06, "loss": 0.5651, "step": 6236 }, { "epoch": 1.9956271331058022, "grad_norm": 0.4627353847026825, "learning_rate": 3.04216541740342e-06, "loss": 0.4136, "step": 6237 }, { "epoch": 1.9959470989761092, "grad_norm": 0.5104613304138184, "learning_rate": 3.0404524233073864e-06, "loss": 0.4476, "step": 6238 }, { "epoch": 1.9962670648464163, "grad_norm": 0.5518894195556641, "learning_rate": 3.038739700904981e-06, "loss": 0.5122, "step": 6239 }, { "epoch": 1.9965870307167235, "grad_norm": 0.5062071681022644, "learning_rate": 3.037027250433676e-06, "loss": 0.4557, "step": 6240 }, { "epoch": 1.9969069965870307, "grad_norm": 0.5367639064788818, "learning_rate": 3.035315072130902e-06, "loss": 0.5672, "step": 6241 }, { "epoch": 1.9972269624573378, "grad_norm": 0.4900611937046051, "learning_rate": 3.033603166234059e-06, "loss": 0.3972, "step": 6242 }, { "epoch": 1.997546928327645, "grad_norm": 0.5525842905044556, "learning_rate": 3.031891532980503e-06, "loss": 0.5552, "step": 6243 }, { "epoch": 1.9978668941979523, "grad_norm": 0.5071797966957092, "learning_rate": 3.030180172607552e-06, "loss": 0.4945, "step": 6244 }, { "epoch": 1.9981868600682593, "grad_norm": 0.490200936794281, "learning_rate": 3.0284690853524918e-06, "loss": 0.481, "step": 6245 }, { "epoch": 1.9985068259385665, "grad_norm": 0.5068796277046204, "learning_rate": 3.0267582714525624e-06, "loss": 0.4779, "step": 6246 }, { "epoch": 1.9988267918088738, "grad_norm": 0.5391877889633179, "learning_rate": 3.0250477311449734e-06, "loss": 0.5154, "step": 6247 }, { "epoch": 1.9991467576791808, "grad_norm": 0.5023802518844604, "learning_rate": 3.0233374646668935e-06, "loss": 0.4056, "step": 6248 }, { "epoch": 1.999466723549488, "grad_norm": 0.49331384897232056, "learning_rate": 3.0216274722554513e-06, "loss": 0.5042, "step": 6249 }, { "epoch": 1.9997866894197953, "grad_norm": 0.4648846387863159, "learning_rate": 3.019917754147741e-06, "loss": 0.485, "step": 6250 }, { "epoch": 2.0001066552901023, "grad_norm": 1.401410460472107, "learning_rate": 3.0182083105808135e-06, "loss": 0.8008, "step": 6251 }, { "epoch": 2.0004266211604094, "grad_norm": 0.5558966398239136, "learning_rate": 3.01649914179169e-06, "loss": 0.5132, "step": 6252 }, { "epoch": 2.000746587030717, "grad_norm": 0.5161656737327576, "learning_rate": 3.014790248017347e-06, "loss": 0.4847, "step": 6253 }, { "epoch": 2.001066552901024, "grad_norm": 0.45154380798339844, "learning_rate": 3.0130816294947233e-06, "loss": 0.3873, "step": 6254 }, { "epoch": 2.001386518771331, "grad_norm": 0.4930065870285034, "learning_rate": 3.0113732864607236e-06, "loss": 0.4263, "step": 6255 }, { "epoch": 2.0017064846416384, "grad_norm": 0.5245649218559265, "learning_rate": 3.009665219152208e-06, "loss": 0.4314, "step": 6256 }, { "epoch": 2.0020264505119454, "grad_norm": 0.5030341744422913, "learning_rate": 3.007957427806004e-06, "loss": 0.4453, "step": 6257 }, { "epoch": 2.0023464163822524, "grad_norm": 0.5460236668586731, "learning_rate": 3.0062499126589018e-06, "loss": 0.4641, "step": 6258 }, { "epoch": 2.00266638225256, "grad_norm": 0.5252211689949036, "learning_rate": 3.004542673947646e-06, "loss": 0.4158, "step": 6259 }, { "epoch": 2.002986348122867, "grad_norm": 0.544998824596405, "learning_rate": 3.00283571190895e-06, "loss": 0.4475, "step": 6260 }, { "epoch": 2.003306313993174, "grad_norm": 0.5802964568138123, "learning_rate": 3.0011290267794833e-06, "loss": 0.4916, "step": 6261 }, { "epoch": 2.0036262798634814, "grad_norm": 0.5228466987609863, "learning_rate": 2.9994226187958824e-06, "loss": 0.4407, "step": 6262 }, { "epoch": 2.0039462457337884, "grad_norm": 0.5055839419364929, "learning_rate": 2.997716488194744e-06, "loss": 0.3993, "step": 6263 }, { "epoch": 2.0042662116040955, "grad_norm": 0.5381371378898621, "learning_rate": 2.9960106352126217e-06, "loss": 0.4764, "step": 6264 }, { "epoch": 2.004586177474403, "grad_norm": 0.4837213456630707, "learning_rate": 2.994305060086037e-06, "loss": 0.4731, "step": 6265 }, { "epoch": 2.00490614334471, "grad_norm": 0.5261613130569458, "learning_rate": 2.9925997630514658e-06, "loss": 0.4017, "step": 6266 }, { "epoch": 2.005226109215017, "grad_norm": 0.5161585211753845, "learning_rate": 2.990894744345354e-06, "loss": 0.4647, "step": 6267 }, { "epoch": 2.0055460750853245, "grad_norm": 0.5065637230873108, "learning_rate": 2.9891900042041043e-06, "loss": 0.4401, "step": 6268 }, { "epoch": 2.0058660409556315, "grad_norm": 0.5628411173820496, "learning_rate": 2.9874855428640783e-06, "loss": 0.4596, "step": 6269 }, { "epoch": 2.0061860068259385, "grad_norm": 0.5215651392936707, "learning_rate": 2.9857813605616036e-06, "loss": 0.4688, "step": 6270 }, { "epoch": 2.0065059726962455, "grad_norm": 0.5510452389717102, "learning_rate": 2.9840774575329644e-06, "loss": 0.4361, "step": 6271 }, { "epoch": 2.006825938566553, "grad_norm": 0.5135353207588196, "learning_rate": 2.982373834014412e-06, "loss": 0.398, "step": 6272 }, { "epoch": 2.00714590443686, "grad_norm": 0.5171473622322083, "learning_rate": 2.9806704902421557e-06, "loss": 0.4695, "step": 6273 }, { "epoch": 2.007465870307167, "grad_norm": 0.49266937375068665, "learning_rate": 2.978967426452365e-06, "loss": 0.4735, "step": 6274 }, { "epoch": 2.0077858361774745, "grad_norm": 0.5164386630058289, "learning_rate": 2.97726464288117e-06, "loss": 0.4332, "step": 6275 }, { "epoch": 2.0081058020477816, "grad_norm": 0.510276198387146, "learning_rate": 2.9755621397646682e-06, "loss": 0.4564, "step": 6276 }, { "epoch": 2.0084257679180886, "grad_norm": 0.5183811187744141, "learning_rate": 2.9738599173389104e-06, "loss": 0.4232, "step": 6277 }, { "epoch": 2.008745733788396, "grad_norm": 0.48706677556037903, "learning_rate": 2.9721579758399134e-06, "loss": 0.4091, "step": 6278 }, { "epoch": 2.009065699658703, "grad_norm": 0.465788871049881, "learning_rate": 2.9704563155036515e-06, "loss": 0.4329, "step": 6279 }, { "epoch": 2.00938566552901, "grad_norm": 0.5820158123970032, "learning_rate": 2.968754936566062e-06, "loss": 0.4563, "step": 6280 }, { "epoch": 2.0097056313993176, "grad_norm": 0.5367518067359924, "learning_rate": 2.967053839263046e-06, "loss": 0.4055, "step": 6281 }, { "epoch": 2.0100255972696246, "grad_norm": 0.5321242213249207, "learning_rate": 2.9653530238304603e-06, "loss": 0.4539, "step": 6282 }, { "epoch": 2.0103455631399316, "grad_norm": 0.5166159272193909, "learning_rate": 2.9636524905041264e-06, "loss": 0.4784, "step": 6283 }, { "epoch": 2.010665529010239, "grad_norm": 0.4449666738510132, "learning_rate": 2.961952239519823e-06, "loss": 0.4355, "step": 6284 }, { "epoch": 2.010985494880546, "grad_norm": 0.5288645029067993, "learning_rate": 2.9602522711132932e-06, "loss": 0.4802, "step": 6285 }, { "epoch": 2.011305460750853, "grad_norm": 0.5091286301612854, "learning_rate": 2.9585525855202415e-06, "loss": 0.43, "step": 6286 }, { "epoch": 2.0116254266211606, "grad_norm": 0.5545825958251953, "learning_rate": 2.9568531829763294e-06, "loss": 0.413, "step": 6287 }, { "epoch": 2.0119453924914676, "grad_norm": 0.5017845630645752, "learning_rate": 2.9551540637171815e-06, "loss": 0.4257, "step": 6288 }, { "epoch": 2.0122653583617747, "grad_norm": 0.49946045875549316, "learning_rate": 2.953455227978382e-06, "loss": 0.412, "step": 6289 }, { "epoch": 2.0125853242320817, "grad_norm": 0.5081928968429565, "learning_rate": 2.9517566759954762e-06, "loss": 0.458, "step": 6290 }, { "epoch": 2.012905290102389, "grad_norm": 0.5049950480461121, "learning_rate": 2.950058408003973e-06, "loss": 0.4547, "step": 6291 }, { "epoch": 2.013225255972696, "grad_norm": 0.5251578092575073, "learning_rate": 2.9483604242393372e-06, "loss": 0.4132, "step": 6292 }, { "epoch": 2.013545221843003, "grad_norm": 0.5312917828559875, "learning_rate": 2.946662724936998e-06, "loss": 0.4707, "step": 6293 }, { "epoch": 2.0138651877133107, "grad_norm": 0.5079889297485352, "learning_rate": 2.9449653103323405e-06, "loss": 0.468, "step": 6294 }, { "epoch": 2.0141851535836177, "grad_norm": 0.5116049647331238, "learning_rate": 2.9432681806607145e-06, "loss": 0.4347, "step": 6295 }, { "epoch": 2.0145051194539247, "grad_norm": 0.4871806800365448, "learning_rate": 2.941571336157432e-06, "loss": 0.4103, "step": 6296 }, { "epoch": 2.014825085324232, "grad_norm": 0.5868577361106873, "learning_rate": 2.939874777057758e-06, "loss": 0.4955, "step": 6297 }, { "epoch": 2.0151450511945392, "grad_norm": 0.5232118368148804, "learning_rate": 2.938178503596926e-06, "loss": 0.4416, "step": 6298 }, { "epoch": 2.0154650170648463, "grad_norm": 0.5466344356536865, "learning_rate": 2.936482516010124e-06, "loss": 0.4367, "step": 6299 }, { "epoch": 2.0157849829351537, "grad_norm": 0.5040996670722961, "learning_rate": 2.934786814532502e-06, "loss": 0.5026, "step": 6300 }, { "epoch": 2.0161049488054608, "grad_norm": 0.4805614650249481, "learning_rate": 2.9330913993991755e-06, "loss": 0.4239, "step": 6301 }, { "epoch": 2.016424914675768, "grad_norm": 0.5113465189933777, "learning_rate": 2.9313962708452116e-06, "loss": 0.4594, "step": 6302 }, { "epoch": 2.0167448805460753, "grad_norm": 0.5428706407546997, "learning_rate": 2.9297014291056446e-06, "loss": 0.4309, "step": 6303 }, { "epoch": 2.0170648464163823, "grad_norm": 0.5894550085067749, "learning_rate": 2.9280068744154645e-06, "loss": 0.5033, "step": 6304 }, { "epoch": 2.0173848122866893, "grad_norm": 0.5202749967575073, "learning_rate": 2.9263126070096222e-06, "loss": 0.45, "step": 6305 }, { "epoch": 2.017704778156997, "grad_norm": 0.47046372294425964, "learning_rate": 2.9246186271230335e-06, "loss": 0.4595, "step": 6306 }, { "epoch": 2.018024744027304, "grad_norm": 0.4563703238964081, "learning_rate": 2.9229249349905686e-06, "loss": 0.4495, "step": 6307 }, { "epoch": 2.018344709897611, "grad_norm": 0.47942081093788147, "learning_rate": 2.921231530847061e-06, "loss": 0.4429, "step": 6308 }, { "epoch": 2.0186646757679183, "grad_norm": 0.47403234243392944, "learning_rate": 2.9195384149273e-06, "loss": 0.3801, "step": 6309 }, { "epoch": 2.0189846416382253, "grad_norm": 0.4782538115978241, "learning_rate": 2.9178455874660423e-06, "loss": 0.4318, "step": 6310 }, { "epoch": 2.0193046075085324, "grad_norm": 0.517080545425415, "learning_rate": 2.9161530486979993e-06, "loss": 0.4635, "step": 6311 }, { "epoch": 2.0196245733788394, "grad_norm": 0.504300057888031, "learning_rate": 2.9144607988578433e-06, "loss": 0.453, "step": 6312 }, { "epoch": 2.019944539249147, "grad_norm": 0.45974022150039673, "learning_rate": 2.9127688381802043e-06, "loss": 0.415, "step": 6313 }, { "epoch": 2.020264505119454, "grad_norm": 0.5267173647880554, "learning_rate": 2.9110771668996772e-06, "loss": 0.4255, "step": 6314 }, { "epoch": 2.020584470989761, "grad_norm": 0.5032655000686646, "learning_rate": 2.909385785250816e-06, "loss": 0.4436, "step": 6315 }, { "epoch": 2.0209044368600684, "grad_norm": 0.4646185040473938, "learning_rate": 2.9076946934681304e-06, "loss": 0.4413, "step": 6316 }, { "epoch": 2.0212244027303754, "grad_norm": 0.4995832145214081, "learning_rate": 2.9060038917860928e-06, "loss": 0.4442, "step": 6317 }, { "epoch": 2.0215443686006824, "grad_norm": 0.5385740995407104, "learning_rate": 2.904313380439132e-06, "loss": 0.4534, "step": 6318 }, { "epoch": 2.02186433447099, "grad_norm": 0.5125446319580078, "learning_rate": 2.9026231596616435e-06, "loss": 0.428, "step": 6319 }, { "epoch": 2.022184300341297, "grad_norm": 0.5198769569396973, "learning_rate": 2.900933229687978e-06, "loss": 0.4173, "step": 6320 }, { "epoch": 2.022504266211604, "grad_norm": 0.5717306733131409, "learning_rate": 2.899243590752446e-06, "loss": 0.4562, "step": 6321 }, { "epoch": 2.0228242320819114, "grad_norm": 0.4866477847099304, "learning_rate": 2.8975542430893177e-06, "loss": 0.4519, "step": 6322 }, { "epoch": 2.0231441979522184, "grad_norm": 0.5071144700050354, "learning_rate": 2.8958651869328202e-06, "loss": 0.42, "step": 6323 }, { "epoch": 2.0234641638225255, "grad_norm": 0.45792296528816223, "learning_rate": 2.8941764225171466e-06, "loss": 0.4365, "step": 6324 }, { "epoch": 2.023784129692833, "grad_norm": 0.501045823097229, "learning_rate": 2.892487950076447e-06, "loss": 0.4774, "step": 6325 }, { "epoch": 2.02410409556314, "grad_norm": 0.4918345808982849, "learning_rate": 2.890799769844829e-06, "loss": 0.463, "step": 6326 }, { "epoch": 2.024424061433447, "grad_norm": 0.512128472328186, "learning_rate": 2.8891118820563575e-06, "loss": 0.4066, "step": 6327 }, { "epoch": 2.0247440273037545, "grad_norm": 0.5132451057434082, "learning_rate": 2.8874242869450655e-06, "loss": 0.5126, "step": 6328 }, { "epoch": 2.0250639931740615, "grad_norm": 0.4674147069454193, "learning_rate": 2.885736984744935e-06, "loss": 0.4349, "step": 6329 }, { "epoch": 2.0253839590443685, "grad_norm": 0.4867783784866333, "learning_rate": 2.8840499756899176e-06, "loss": 0.4214, "step": 6330 }, { "epoch": 2.0257039249146755, "grad_norm": 0.6133652925491333, "learning_rate": 2.8823632600139166e-06, "loss": 0.4576, "step": 6331 }, { "epoch": 2.026023890784983, "grad_norm": 0.5312769412994385, "learning_rate": 2.8806768379507953e-06, "loss": 0.4597, "step": 6332 }, { "epoch": 2.02634385665529, "grad_norm": 0.49776095151901245, "learning_rate": 2.8789907097343818e-06, "loss": 0.4206, "step": 6333 }, { "epoch": 2.026663822525597, "grad_norm": 0.49100378155708313, "learning_rate": 2.8773048755984567e-06, "loss": 0.4526, "step": 6334 }, { "epoch": 2.0269837883959045, "grad_norm": 0.5101442933082581, "learning_rate": 2.875619335776765e-06, "loss": 0.4395, "step": 6335 }, { "epoch": 2.0273037542662116, "grad_norm": 0.48366159200668335, "learning_rate": 2.8739340905030087e-06, "loss": 0.4498, "step": 6336 }, { "epoch": 2.0276237201365186, "grad_norm": 0.5114291906356812, "learning_rate": 2.8722491400108464e-06, "loss": 0.444, "step": 6337 }, { "epoch": 2.027943686006826, "grad_norm": 0.5394448637962341, "learning_rate": 2.870564484533903e-06, "loss": 0.4696, "step": 6338 }, { "epoch": 2.028263651877133, "grad_norm": 0.5102099776268005, "learning_rate": 2.8688801243057536e-06, "loss": 0.3991, "step": 6339 }, { "epoch": 2.02858361774744, "grad_norm": 0.5134938955307007, "learning_rate": 2.86719605955994e-06, "loss": 0.4225, "step": 6340 }, { "epoch": 2.0289035836177476, "grad_norm": 0.6032628417015076, "learning_rate": 2.865512290529959e-06, "loss": 0.4376, "step": 6341 }, { "epoch": 2.0292235494880546, "grad_norm": 0.5756434798240662, "learning_rate": 2.8638288174492657e-06, "loss": 0.4826, "step": 6342 }, { "epoch": 2.0295435153583616, "grad_norm": 0.516460657119751, "learning_rate": 2.8621456405512775e-06, "loss": 0.4089, "step": 6343 }, { "epoch": 2.029863481228669, "grad_norm": 0.5271695852279663, "learning_rate": 2.860462760069367e-06, "loss": 0.4177, "step": 6344 }, { "epoch": 2.030183447098976, "grad_norm": 0.5140413045883179, "learning_rate": 2.8587801762368695e-06, "loss": 0.4658, "step": 6345 }, { "epoch": 2.030503412969283, "grad_norm": 0.48984235525131226, "learning_rate": 2.8570978892870777e-06, "loss": 0.4598, "step": 6346 }, { "epoch": 2.0308233788395906, "grad_norm": 0.4675515294075012, "learning_rate": 2.855415899453239e-06, "loss": 0.4044, "step": 6347 }, { "epoch": 2.0311433447098977, "grad_norm": 0.4961652457714081, "learning_rate": 2.8537342069685657e-06, "loss": 0.4533, "step": 6348 }, { "epoch": 2.0314633105802047, "grad_norm": 0.5445696115493774, "learning_rate": 2.8520528120662296e-06, "loss": 0.4365, "step": 6349 }, { "epoch": 2.031783276450512, "grad_norm": 0.5242259502410889, "learning_rate": 2.8503717149793543e-06, "loss": 0.4994, "step": 6350 }, { "epoch": 2.032103242320819, "grad_norm": 0.5333598256111145, "learning_rate": 2.8486909159410266e-06, "loss": 0.4286, "step": 6351 }, { "epoch": 2.032423208191126, "grad_norm": 0.48089703917503357, "learning_rate": 2.847010415184289e-06, "loss": 0.4337, "step": 6352 }, { "epoch": 2.0327431740614332, "grad_norm": 0.5590294003486633, "learning_rate": 2.845330212942149e-06, "loss": 0.4655, "step": 6353 }, { "epoch": 2.0330631399317407, "grad_norm": 0.5742276906967163, "learning_rate": 2.843650309447568e-06, "loss": 0.4868, "step": 6354 }, { "epoch": 2.0333831058020477, "grad_norm": 0.504341185092926, "learning_rate": 2.8419707049334664e-06, "loss": 0.4236, "step": 6355 }, { "epoch": 2.0337030716723548, "grad_norm": 0.5049486756324768, "learning_rate": 2.8402913996327217e-06, "loss": 0.406, "step": 6356 }, { "epoch": 2.0340230375426622, "grad_norm": 0.5707356333732605, "learning_rate": 2.838612393778172e-06, "loss": 0.4955, "step": 6357 }, { "epoch": 2.0343430034129693, "grad_norm": 0.49165648221969604, "learning_rate": 2.8369336876026133e-06, "loss": 0.426, "step": 6358 }, { "epoch": 2.0346629692832763, "grad_norm": 0.5239745378494263, "learning_rate": 2.8352552813388035e-06, "loss": 0.4673, "step": 6359 }, { "epoch": 2.0349829351535837, "grad_norm": 0.5204604268074036, "learning_rate": 2.833577175219453e-06, "loss": 0.4028, "step": 6360 }, { "epoch": 2.0353029010238908, "grad_norm": 0.5317163467407227, "learning_rate": 2.831899369477233e-06, "loss": 0.4746, "step": 6361 }, { "epoch": 2.035622866894198, "grad_norm": 0.46956390142440796, "learning_rate": 2.830221864344772e-06, "loss": 0.4173, "step": 6362 }, { "epoch": 2.0359428327645053, "grad_norm": 0.5055503845214844, "learning_rate": 2.8285446600546594e-06, "loss": 0.4398, "step": 6363 }, { "epoch": 2.0362627986348123, "grad_norm": 0.5239434242248535, "learning_rate": 2.8268677568394456e-06, "loss": 0.447, "step": 6364 }, { "epoch": 2.0365827645051193, "grad_norm": 0.5025309324264526, "learning_rate": 2.8251911549316303e-06, "loss": 0.4443, "step": 6365 }, { "epoch": 2.036902730375427, "grad_norm": 0.4989284873008728, "learning_rate": 2.8235148545636776e-06, "loss": 0.4252, "step": 6366 }, { "epoch": 2.037222696245734, "grad_norm": 0.49771973490715027, "learning_rate": 2.8218388559680065e-06, "loss": 0.5184, "step": 6367 }, { "epoch": 2.037542662116041, "grad_norm": 0.47321969270706177, "learning_rate": 2.8201631593769995e-06, "loss": 0.3978, "step": 6368 }, { "epoch": 2.0378626279863483, "grad_norm": 0.5439449548721313, "learning_rate": 2.818487765022994e-06, "loss": 0.4636, "step": 6369 }, { "epoch": 2.0381825938566553, "grad_norm": 0.5097255706787109, "learning_rate": 2.816812673138285e-06, "loss": 0.4238, "step": 6370 }, { "epoch": 2.0385025597269624, "grad_norm": 0.5556920766830444, "learning_rate": 2.8151378839551248e-06, "loss": 0.4705, "step": 6371 }, { "epoch": 2.03882252559727, "grad_norm": 0.5774846076965332, "learning_rate": 2.8134633977057236e-06, "loss": 0.4399, "step": 6372 }, { "epoch": 2.039142491467577, "grad_norm": 0.5274468064308167, "learning_rate": 2.8117892146222524e-06, "loss": 0.4709, "step": 6373 }, { "epoch": 2.039462457337884, "grad_norm": 0.4775089621543884, "learning_rate": 2.8101153349368417e-06, "loss": 0.4058, "step": 6374 }, { "epoch": 2.039782423208191, "grad_norm": 0.530306875705719, "learning_rate": 2.808441758881574e-06, "loss": 0.4554, "step": 6375 }, { "epoch": 2.0401023890784984, "grad_norm": 0.4880123436450958, "learning_rate": 2.8067684866884927e-06, "loss": 0.4235, "step": 6376 }, { "epoch": 2.0404223549488054, "grad_norm": 0.5186628699302673, "learning_rate": 2.8050955185895977e-06, "loss": 0.4407, "step": 6377 }, { "epoch": 2.0407423208191124, "grad_norm": 0.5505031943321228, "learning_rate": 2.8034228548168496e-06, "loss": 0.4116, "step": 6378 }, { "epoch": 2.04106228668942, "grad_norm": 0.5448146462440491, "learning_rate": 2.801750495602167e-06, "loss": 0.4322, "step": 6379 }, { "epoch": 2.041382252559727, "grad_norm": 0.5421686768531799, "learning_rate": 2.800078441177423e-06, "loss": 0.4143, "step": 6380 }, { "epoch": 2.041702218430034, "grad_norm": 0.489848792552948, "learning_rate": 2.798406691774449e-06, "loss": 0.4667, "step": 6381 }, { "epoch": 2.0420221843003414, "grad_norm": 0.5021038055419922, "learning_rate": 2.7967352476250344e-06, "loss": 0.4193, "step": 6382 }, { "epoch": 2.0423421501706485, "grad_norm": 0.503178060054779, "learning_rate": 2.7950641089609275e-06, "loss": 0.452, "step": 6383 }, { "epoch": 2.0426621160409555, "grad_norm": 0.49223408102989197, "learning_rate": 2.7933932760138375e-06, "loss": 0.4555, "step": 6384 }, { "epoch": 2.042982081911263, "grad_norm": 0.4962237775325775, "learning_rate": 2.791722749015424e-06, "loss": 0.4129, "step": 6385 }, { "epoch": 2.04330204778157, "grad_norm": 0.5070708394050598, "learning_rate": 2.7900525281973078e-06, "loss": 0.4746, "step": 6386 }, { "epoch": 2.043622013651877, "grad_norm": 0.4765138328075409, "learning_rate": 2.788382613791066e-06, "loss": 0.4022, "step": 6387 }, { "epoch": 2.0439419795221845, "grad_norm": 0.5285912156105042, "learning_rate": 2.7867130060282345e-06, "loss": 0.471, "step": 6388 }, { "epoch": 2.0442619453924915, "grad_norm": 0.506458580493927, "learning_rate": 2.7850437051403102e-06, "loss": 0.4715, "step": 6389 }, { "epoch": 2.0445819112627985, "grad_norm": 0.5261322855949402, "learning_rate": 2.7833747113587394e-06, "loss": 0.4661, "step": 6390 }, { "epoch": 2.044901877133106, "grad_norm": 0.520087718963623, "learning_rate": 2.781706024914933e-06, "loss": 0.4166, "step": 6391 }, { "epoch": 2.045221843003413, "grad_norm": 0.5601601600646973, "learning_rate": 2.7800376460402523e-06, "loss": 0.4705, "step": 6392 }, { "epoch": 2.04554180887372, "grad_norm": 0.484840989112854, "learning_rate": 2.7783695749660228e-06, "loss": 0.4287, "step": 6393 }, { "epoch": 2.0458617747440275, "grad_norm": 0.5149590373039246, "learning_rate": 2.7767018119235263e-06, "loss": 0.4412, "step": 6394 }, { "epoch": 2.0461817406143346, "grad_norm": 0.4719947278499603, "learning_rate": 2.7750343571439987e-06, "loss": 0.4471, "step": 6395 }, { "epoch": 2.0465017064846416, "grad_norm": 0.476144015789032, "learning_rate": 2.7733672108586333e-06, "loss": 0.458, "step": 6396 }, { "epoch": 2.0468216723549486, "grad_norm": 0.5308468341827393, "learning_rate": 2.7717003732985813e-06, "loss": 0.4377, "step": 6397 }, { "epoch": 2.047141638225256, "grad_norm": 0.5079355835914612, "learning_rate": 2.770033844694954e-06, "loss": 0.4276, "step": 6398 }, { "epoch": 2.047461604095563, "grad_norm": 0.4922444224357605, "learning_rate": 2.7683676252788172e-06, "loss": 0.4352, "step": 6399 }, { "epoch": 2.04778156996587, "grad_norm": 0.5174881219863892, "learning_rate": 2.7667017152811947e-06, "loss": 0.4665, "step": 6400 }, { "epoch": 2.0481015358361776, "grad_norm": 0.5103611350059509, "learning_rate": 2.7650361149330662e-06, "loss": 0.4197, "step": 6401 }, { "epoch": 2.0484215017064846, "grad_norm": 0.5468028783798218, "learning_rate": 2.7633708244653677e-06, "loss": 0.4401, "step": 6402 }, { "epoch": 2.0487414675767917, "grad_norm": 0.5299778580665588, "learning_rate": 2.7617058441089943e-06, "loss": 0.454, "step": 6403 }, { "epoch": 2.049061433447099, "grad_norm": 0.5115585327148438, "learning_rate": 2.7600411740948007e-06, "loss": 0.4152, "step": 6404 }, { "epoch": 2.049381399317406, "grad_norm": 0.4773969054222107, "learning_rate": 2.7583768146535927e-06, "loss": 0.4427, "step": 6405 }, { "epoch": 2.049701365187713, "grad_norm": 0.4479951560497284, "learning_rate": 2.7567127660161366e-06, "loss": 0.3929, "step": 6406 }, { "epoch": 2.0500213310580206, "grad_norm": 0.4836898446083069, "learning_rate": 2.7550490284131516e-06, "loss": 0.4789, "step": 6407 }, { "epoch": 2.0503412969283277, "grad_norm": 0.5370376110076904, "learning_rate": 2.75338560207532e-06, "loss": 0.4708, "step": 6408 }, { "epoch": 2.0506612627986347, "grad_norm": 0.4753120541572571, "learning_rate": 2.7517224872332794e-06, "loss": 0.3785, "step": 6409 }, { "epoch": 2.050981228668942, "grad_norm": 0.5000243782997131, "learning_rate": 2.7500596841176207e-06, "loss": 0.4492, "step": 6410 }, { "epoch": 2.051301194539249, "grad_norm": 0.5311595797538757, "learning_rate": 2.748397192958893e-06, "loss": 0.4689, "step": 6411 }, { "epoch": 2.051621160409556, "grad_norm": 0.5121088624000549, "learning_rate": 2.746735013987602e-06, "loss": 0.4071, "step": 6412 }, { "epoch": 2.0519411262798637, "grad_norm": 0.5043737292289734, "learning_rate": 2.745073147434212e-06, "loss": 0.4012, "step": 6413 }, { "epoch": 2.0522610921501707, "grad_norm": 0.5759963393211365, "learning_rate": 2.743411593529144e-06, "loss": 0.5193, "step": 6414 }, { "epoch": 2.0525810580204777, "grad_norm": 0.48577389121055603, "learning_rate": 2.7417503525027742e-06, "loss": 0.4423, "step": 6415 }, { "epoch": 2.0529010238907848, "grad_norm": 0.5283511281013489, "learning_rate": 2.7400894245854327e-06, "loss": 0.455, "step": 6416 }, { "epoch": 2.0532209897610922, "grad_norm": 0.5407797694206238, "learning_rate": 2.7384288100074124e-06, "loss": 0.4829, "step": 6417 }, { "epoch": 2.0535409556313993, "grad_norm": 0.4855120778083801, "learning_rate": 2.736768508998957e-06, "loss": 0.3815, "step": 6418 }, { "epoch": 2.0538609215017063, "grad_norm": 0.45374947786331177, "learning_rate": 2.735108521790273e-06, "loss": 0.4237, "step": 6419 }, { "epoch": 2.0541808873720138, "grad_norm": 0.5150781869888306, "learning_rate": 2.7334488486115164e-06, "loss": 0.447, "step": 6420 }, { "epoch": 2.054500853242321, "grad_norm": 0.5354177355766296, "learning_rate": 2.7317894896928015e-06, "loss": 0.4674, "step": 6421 }, { "epoch": 2.054820819112628, "grad_norm": 0.528226912021637, "learning_rate": 2.7301304452642054e-06, "loss": 0.3772, "step": 6422 }, { "epoch": 2.0551407849829353, "grad_norm": 0.563994824886322, "learning_rate": 2.7284717155557512e-06, "loss": 0.4606, "step": 6423 }, { "epoch": 2.0554607508532423, "grad_norm": 0.5252886414527893, "learning_rate": 2.7268133007974284e-06, "loss": 0.3824, "step": 6424 }, { "epoch": 2.0557807167235493, "grad_norm": 0.5421047806739807, "learning_rate": 2.7251552012191763e-06, "loss": 0.4509, "step": 6425 }, { "epoch": 2.056100682593857, "grad_norm": 0.5202421545982361, "learning_rate": 2.72349741705089e-06, "loss": 0.4586, "step": 6426 }, { "epoch": 2.056420648464164, "grad_norm": 0.5115158557891846, "learning_rate": 2.721839948522428e-06, "loss": 0.4792, "step": 6427 }, { "epoch": 2.056740614334471, "grad_norm": 0.5127246379852295, "learning_rate": 2.720182795863595e-06, "loss": 0.4698, "step": 6428 }, { "epoch": 2.0570605802047783, "grad_norm": 0.5429564714431763, "learning_rate": 2.7185259593041633e-06, "loss": 0.4015, "step": 6429 }, { "epoch": 2.0573805460750854, "grad_norm": 0.5757423639297485, "learning_rate": 2.7168694390738515e-06, "loss": 0.4856, "step": 6430 }, { "epoch": 2.0577005119453924, "grad_norm": 0.5339115262031555, "learning_rate": 2.7152132354023365e-06, "loss": 0.4565, "step": 6431 }, { "epoch": 2.0580204778157, "grad_norm": 0.49360546469688416, "learning_rate": 2.7135573485192566e-06, "loss": 0.4268, "step": 6432 }, { "epoch": 2.058340443686007, "grad_norm": 0.5282648205757141, "learning_rate": 2.7119017786541997e-06, "loss": 0.4612, "step": 6433 }, { "epoch": 2.058660409556314, "grad_norm": 0.4971468448638916, "learning_rate": 2.7102465260367154e-06, "loss": 0.3992, "step": 6434 }, { "epoch": 2.0589803754266214, "grad_norm": 0.5643179416656494, "learning_rate": 2.7085915908963053e-06, "loss": 0.4679, "step": 6435 }, { "epoch": 2.0593003412969284, "grad_norm": 0.5144991874694824, "learning_rate": 2.706936973462425e-06, "loss": 0.3816, "step": 6436 }, { "epoch": 2.0596203071672354, "grad_norm": 0.5589467883110046, "learning_rate": 2.705282673964495e-06, "loss": 0.4387, "step": 6437 }, { "epoch": 2.0599402730375425, "grad_norm": 0.4803876578807831, "learning_rate": 2.70362869263188e-06, "loss": 0.4571, "step": 6438 }, { "epoch": 2.06026023890785, "grad_norm": 0.5414488315582275, "learning_rate": 2.701975029693912e-06, "loss": 0.4852, "step": 6439 }, { "epoch": 2.060580204778157, "grad_norm": 0.5587663054466248, "learning_rate": 2.700321685379871e-06, "loss": 0.479, "step": 6440 }, { "epoch": 2.060900170648464, "grad_norm": 0.4906553030014038, "learning_rate": 2.6986686599189914e-06, "loss": 0.4323, "step": 6441 }, { "epoch": 2.0612201365187715, "grad_norm": 0.4768359661102295, "learning_rate": 2.697015953540474e-06, "loss": 0.3904, "step": 6442 }, { "epoch": 2.0615401023890785, "grad_norm": 0.4859939217567444, "learning_rate": 2.695363566473463e-06, "loss": 0.4133, "step": 6443 }, { "epoch": 2.0618600682593855, "grad_norm": 0.49089834094047546, "learning_rate": 2.693711498947068e-06, "loss": 0.4171, "step": 6444 }, { "epoch": 2.062180034129693, "grad_norm": 0.5339997410774231, "learning_rate": 2.692059751190348e-06, "loss": 0.4976, "step": 6445 }, { "epoch": 2.0625, "grad_norm": 0.5293613076210022, "learning_rate": 2.690408323432318e-06, "loss": 0.4639, "step": 6446 }, { "epoch": 2.062819965870307, "grad_norm": 0.5344449877738953, "learning_rate": 2.688757215901955e-06, "loss": 0.4026, "step": 6447 }, { "epoch": 2.0631399317406145, "grad_norm": 0.511799693107605, "learning_rate": 2.687106428828182e-06, "loss": 0.439, "step": 6448 }, { "epoch": 2.0634598976109215, "grad_norm": 0.5103006362915039, "learning_rate": 2.6854559624398867e-06, "loss": 0.4567, "step": 6449 }, { "epoch": 2.0637798634812285, "grad_norm": 0.4594227969646454, "learning_rate": 2.6838058169659076e-06, "loss": 0.4354, "step": 6450 }, { "epoch": 2.064099829351536, "grad_norm": 0.512006402015686, "learning_rate": 2.6821559926350362e-06, "loss": 0.4884, "step": 6451 }, { "epoch": 2.064419795221843, "grad_norm": 0.49301502108573914, "learning_rate": 2.6805064896760265e-06, "loss": 0.3972, "step": 6452 }, { "epoch": 2.06473976109215, "grad_norm": 0.460877925157547, "learning_rate": 2.6788573083175806e-06, "loss": 0.4664, "step": 6453 }, { "epoch": 2.0650597269624575, "grad_norm": 0.4712378978729248, "learning_rate": 2.6772084487883633e-06, "loss": 0.4449, "step": 6454 }, { "epoch": 2.0653796928327646, "grad_norm": 0.5092011094093323, "learning_rate": 2.6755599113169866e-06, "loss": 0.4403, "step": 6455 }, { "epoch": 2.0656996587030716, "grad_norm": 0.5052369236946106, "learning_rate": 2.6739116961320266e-06, "loss": 0.4583, "step": 6456 }, { "epoch": 2.0660196245733786, "grad_norm": 0.5215187072753906, "learning_rate": 2.672263803462009e-06, "loss": 0.4734, "step": 6457 }, { "epoch": 2.066339590443686, "grad_norm": 0.47432374954223633, "learning_rate": 2.670616233535413e-06, "loss": 0.3821, "step": 6458 }, { "epoch": 2.066659556313993, "grad_norm": 0.5366644859313965, "learning_rate": 2.6689689865806806e-06, "loss": 0.4653, "step": 6459 }, { "epoch": 2.0669795221843, "grad_norm": 0.5265777707099915, "learning_rate": 2.6673220628262002e-06, "loss": 0.4624, "step": 6460 }, { "epoch": 2.0672994880546076, "grad_norm": 0.48621946573257446, "learning_rate": 2.6656754625003244e-06, "loss": 0.4698, "step": 6461 }, { "epoch": 2.0676194539249146, "grad_norm": 0.47744280099868774, "learning_rate": 2.6640291858313536e-06, "loss": 0.4359, "step": 6462 }, { "epoch": 2.0679394197952217, "grad_norm": 0.48860299587249756, "learning_rate": 2.6623832330475454e-06, "loss": 0.4267, "step": 6463 }, { "epoch": 2.068259385665529, "grad_norm": 0.4812034070491791, "learning_rate": 2.660737604377115e-06, "loss": 0.3962, "step": 6464 }, { "epoch": 2.068579351535836, "grad_norm": 0.5040679574012756, "learning_rate": 2.6590923000482285e-06, "loss": 0.4513, "step": 6465 }, { "epoch": 2.068899317406143, "grad_norm": 0.4805178642272949, "learning_rate": 2.6574473202890126e-06, "loss": 0.392, "step": 6466 }, { "epoch": 2.0692192832764507, "grad_norm": 0.5261695384979248, "learning_rate": 2.655802665327542e-06, "loss": 0.4297, "step": 6467 }, { "epoch": 2.0695392491467577, "grad_norm": 0.5518275499343872, "learning_rate": 2.6541583353918545e-06, "loss": 0.4682, "step": 6468 }, { "epoch": 2.0698592150170647, "grad_norm": 0.5347665548324585, "learning_rate": 2.6525143307099353e-06, "loss": 0.4474, "step": 6469 }, { "epoch": 2.070179180887372, "grad_norm": 0.5325558185577393, "learning_rate": 2.650870651509726e-06, "loss": 0.4325, "step": 6470 }, { "epoch": 2.070499146757679, "grad_norm": 0.5175374746322632, "learning_rate": 2.649227298019129e-06, "loss": 0.4919, "step": 6471 }, { "epoch": 2.0708191126279862, "grad_norm": 0.5054546594619751, "learning_rate": 2.6475842704659937e-06, "loss": 0.4048, "step": 6472 }, { "epoch": 2.0711390784982937, "grad_norm": 0.4748407304286957, "learning_rate": 2.645941569078131e-06, "loss": 0.4462, "step": 6473 }, { "epoch": 2.0714590443686007, "grad_norm": 0.5580139756202698, "learning_rate": 2.644299194083302e-06, "loss": 0.4826, "step": 6474 }, { "epoch": 2.0717790102389078, "grad_norm": 0.5135101079940796, "learning_rate": 2.642657145709222e-06, "loss": 0.4232, "step": 6475 }, { "epoch": 2.0720989761092152, "grad_norm": 0.5293991565704346, "learning_rate": 2.6410154241835663e-06, "loss": 0.4532, "step": 6476 }, { "epoch": 2.0724189419795223, "grad_norm": 0.49501514434814453, "learning_rate": 2.6393740297339577e-06, "loss": 0.4468, "step": 6477 }, { "epoch": 2.0727389078498293, "grad_norm": 0.45320573449134827, "learning_rate": 2.637732962587982e-06, "loss": 0.4311, "step": 6478 }, { "epoch": 2.0730588737201363, "grad_norm": 0.5398769378662109, "learning_rate": 2.6360922229731727e-06, "loss": 0.4698, "step": 6479 }, { "epoch": 2.073378839590444, "grad_norm": 0.49020859599113464, "learning_rate": 2.634451811117019e-06, "loss": 0.4481, "step": 6480 }, { "epoch": 2.073698805460751, "grad_norm": 0.5167035460472107, "learning_rate": 2.632811727246969e-06, "loss": 0.4564, "step": 6481 }, { "epoch": 2.074018771331058, "grad_norm": 0.5227987170219421, "learning_rate": 2.631171971590419e-06, "loss": 0.4182, "step": 6482 }, { "epoch": 2.0743387372013653, "grad_norm": 0.5614644289016724, "learning_rate": 2.6295325443747272e-06, "loss": 0.4358, "step": 6483 }, { "epoch": 2.0746587030716723, "grad_norm": 0.5467914938926697, "learning_rate": 2.6278934458271998e-06, "loss": 0.4996, "step": 6484 }, { "epoch": 2.0749786689419794, "grad_norm": 0.4813799262046814, "learning_rate": 2.6262546761750975e-06, "loss": 0.4124, "step": 6485 }, { "epoch": 2.075298634812287, "grad_norm": 0.5109665393829346, "learning_rate": 2.6246162356456423e-06, "loss": 0.4715, "step": 6486 }, { "epoch": 2.075618600682594, "grad_norm": 0.48429253697395325, "learning_rate": 2.6229781244660015e-06, "loss": 0.4657, "step": 6487 }, { "epoch": 2.075938566552901, "grad_norm": 0.4949950873851776, "learning_rate": 2.6213403428633055e-06, "loss": 0.4565, "step": 6488 }, { "epoch": 2.0762585324232083, "grad_norm": 0.47727730870246887, "learning_rate": 2.6197028910646304e-06, "loss": 0.3965, "step": 6489 }, { "epoch": 2.0765784982935154, "grad_norm": 0.5116509795188904, "learning_rate": 2.6180657692970157e-06, "loss": 0.4626, "step": 6490 }, { "epoch": 2.0768984641638224, "grad_norm": 0.5054216980934143, "learning_rate": 2.6164289777874468e-06, "loss": 0.4123, "step": 6491 }, { "epoch": 2.07721843003413, "grad_norm": 0.4977065324783325, "learning_rate": 2.614792516762866e-06, "loss": 0.4229, "step": 6492 }, { "epoch": 2.077538395904437, "grad_norm": 0.46970316767692566, "learning_rate": 2.613156386450174e-06, "loss": 0.4102, "step": 6493 }, { "epoch": 2.077858361774744, "grad_norm": 0.5014659762382507, "learning_rate": 2.6115205870762187e-06, "loss": 0.4661, "step": 6494 }, { "epoch": 2.0781783276450514, "grad_norm": 0.5311126112937927, "learning_rate": 2.6098851188678096e-06, "loss": 0.4616, "step": 6495 }, { "epoch": 2.0784982935153584, "grad_norm": 0.5128471255302429, "learning_rate": 2.608249982051704e-06, "loss": 0.4308, "step": 6496 }, { "epoch": 2.0788182593856654, "grad_norm": 0.5046076774597168, "learning_rate": 2.606615176854613e-06, "loss": 0.4322, "step": 6497 }, { "epoch": 2.0791382252559725, "grad_norm": 0.5225712060928345, "learning_rate": 2.6049807035032104e-06, "loss": 0.481, "step": 6498 }, { "epoch": 2.07945819112628, "grad_norm": 0.4911491870880127, "learning_rate": 2.6033465622241116e-06, "loss": 0.4292, "step": 6499 }, { "epoch": 2.079778156996587, "grad_norm": 0.49582064151763916, "learning_rate": 2.6017127532438968e-06, "loss": 0.4471, "step": 6500 }, { "epoch": 2.080098122866894, "grad_norm": 0.5208566784858704, "learning_rate": 2.6000792767890947e-06, "loss": 0.3907, "step": 6501 }, { "epoch": 2.0804180887372015, "grad_norm": 0.5576671361923218, "learning_rate": 2.5984461330861864e-06, "loss": 0.4631, "step": 6502 }, { "epoch": 2.0807380546075085, "grad_norm": 0.5146280527114868, "learning_rate": 2.596813322361612e-06, "loss": 0.4297, "step": 6503 }, { "epoch": 2.0810580204778155, "grad_norm": 0.4958689510822296, "learning_rate": 2.5951808448417603e-06, "loss": 0.432, "step": 6504 }, { "epoch": 2.081377986348123, "grad_norm": 0.4615541696548462, "learning_rate": 2.593548700752979e-06, "loss": 0.4, "step": 6505 }, { "epoch": 2.08169795221843, "grad_norm": 0.5136170983314514, "learning_rate": 2.5919168903215652e-06, "loss": 0.4242, "step": 6506 }, { "epoch": 2.082017918088737, "grad_norm": 0.5297406911849976, "learning_rate": 2.5902854137737704e-06, "loss": 0.4591, "step": 6507 }, { "epoch": 2.0823378839590445, "grad_norm": 0.5089228749275208, "learning_rate": 2.5886542713358034e-06, "loss": 0.4652, "step": 6508 }, { "epoch": 2.0826578498293515, "grad_norm": 0.5159381628036499, "learning_rate": 2.587023463233821e-06, "loss": 0.4216, "step": 6509 }, { "epoch": 2.0829778156996586, "grad_norm": 0.5108634829521179, "learning_rate": 2.5853929896939395e-06, "loss": 0.4428, "step": 6510 }, { "epoch": 2.083297781569966, "grad_norm": 0.5295661091804504, "learning_rate": 2.5837628509422253e-06, "loss": 0.5, "step": 6511 }, { "epoch": 2.083617747440273, "grad_norm": 0.5104172825813293, "learning_rate": 2.582133047204697e-06, "loss": 0.4035, "step": 6512 }, { "epoch": 2.08393771331058, "grad_norm": 0.5169571042060852, "learning_rate": 2.580503578707332e-06, "loss": 0.4415, "step": 6513 }, { "epoch": 2.0842576791808876, "grad_norm": 0.5187350511550903, "learning_rate": 2.5788744456760538e-06, "loss": 0.4642, "step": 6514 }, { "epoch": 2.0845776450511946, "grad_norm": 0.5074020624160767, "learning_rate": 2.57724564833675e-06, "loss": 0.4274, "step": 6515 }, { "epoch": 2.0848976109215016, "grad_norm": 0.4734642505645752, "learning_rate": 2.5756171869152503e-06, "loss": 0.4001, "step": 6516 }, { "epoch": 2.085217576791809, "grad_norm": 0.5345687866210938, "learning_rate": 2.573989061637343e-06, "loss": 0.4515, "step": 6517 }, { "epoch": 2.085537542662116, "grad_norm": 0.5220152139663696, "learning_rate": 2.5723612727287726e-06, "loss": 0.4412, "step": 6518 }, { "epoch": 2.085857508532423, "grad_norm": 0.48031431436538696, "learning_rate": 2.570733820415231e-06, "loss": 0.4634, "step": 6519 }, { "epoch": 2.0861774744027306, "grad_norm": 0.484093576669693, "learning_rate": 2.5691067049223683e-06, "loss": 0.3988, "step": 6520 }, { "epoch": 2.0864974402730376, "grad_norm": 0.5259190797805786, "learning_rate": 2.5674799264757867e-06, "loss": 0.4581, "step": 6521 }, { "epoch": 2.0868174061433447, "grad_norm": 0.5825842618942261, "learning_rate": 2.5658534853010368e-06, "loss": 0.4392, "step": 6522 }, { "epoch": 2.0871373720136517, "grad_norm": 0.585890531539917, "learning_rate": 2.5642273816236298e-06, "loss": 0.4769, "step": 6523 }, { "epoch": 2.087457337883959, "grad_norm": 0.5051429271697998, "learning_rate": 2.5626016156690293e-06, "loss": 0.4291, "step": 6524 }, { "epoch": 2.087777303754266, "grad_norm": 0.5297066569328308, "learning_rate": 2.560976187662646e-06, "loss": 0.4798, "step": 6525 }, { "epoch": 2.088097269624573, "grad_norm": 0.488235205411911, "learning_rate": 2.5593510978298487e-06, "loss": 0.4274, "step": 6526 }, { "epoch": 2.0884172354948807, "grad_norm": 0.46988943219184875, "learning_rate": 2.5577263463959563e-06, "loss": 0.4248, "step": 6527 }, { "epoch": 2.0887372013651877, "grad_norm": 0.5096257925033569, "learning_rate": 2.5561019335862435e-06, "loss": 0.4614, "step": 6528 }, { "epoch": 2.0890571672354947, "grad_norm": 0.48441529273986816, "learning_rate": 2.55447785962594e-06, "loss": 0.4261, "step": 6529 }, { "epoch": 2.089377133105802, "grad_norm": 0.5083130598068237, "learning_rate": 2.552854124740224e-06, "loss": 0.4786, "step": 6530 }, { "epoch": 2.0896970989761092, "grad_norm": 0.5102408528327942, "learning_rate": 2.551230729154227e-06, "loss": 0.4603, "step": 6531 }, { "epoch": 2.0900170648464163, "grad_norm": 0.48783355951309204, "learning_rate": 2.549607673093033e-06, "loss": 0.4378, "step": 6532 }, { "epoch": 2.0903370307167237, "grad_norm": 0.5020625591278076, "learning_rate": 2.547984956781683e-06, "loss": 0.4495, "step": 6533 }, { "epoch": 2.0906569965870307, "grad_norm": 0.548353910446167, "learning_rate": 2.5463625804451714e-06, "loss": 0.4467, "step": 6534 }, { "epoch": 2.0909769624573378, "grad_norm": 0.49052295088768005, "learning_rate": 2.5447405443084395e-06, "loss": 0.4383, "step": 6535 }, { "epoch": 2.0912969283276452, "grad_norm": 0.5260667204856873, "learning_rate": 2.543118848596384e-06, "loss": 0.4672, "step": 6536 }, { "epoch": 2.0916168941979523, "grad_norm": 0.41019099950790405, "learning_rate": 2.5414974935338555e-06, "loss": 0.3995, "step": 6537 }, { "epoch": 2.0919368600682593, "grad_norm": 0.5305437445640564, "learning_rate": 2.5398764793456554e-06, "loss": 0.4685, "step": 6538 }, { "epoch": 2.0922568259385668, "grad_norm": 0.5135195851325989, "learning_rate": 2.538255806256544e-06, "loss": 0.3883, "step": 6539 }, { "epoch": 2.092576791808874, "grad_norm": 0.5410434603691101, "learning_rate": 2.5366354744912267e-06, "loss": 0.445, "step": 6540 }, { "epoch": 2.092896757679181, "grad_norm": 0.5121901631355286, "learning_rate": 2.5350154842743643e-06, "loss": 0.4751, "step": 6541 }, { "epoch": 2.093216723549488, "grad_norm": 0.44693824648857117, "learning_rate": 2.5333958358305683e-06, "loss": 0.3867, "step": 6542 }, { "epoch": 2.0935366894197953, "grad_norm": 0.49896085262298584, "learning_rate": 2.531776529384407e-06, "loss": 0.4473, "step": 6543 }, { "epoch": 2.0938566552901023, "grad_norm": 0.5412548184394836, "learning_rate": 2.5301575651604014e-06, "loss": 0.4437, "step": 6544 }, { "epoch": 2.0941766211604094, "grad_norm": 0.5537082552909851, "learning_rate": 2.5285389433830206e-06, "loss": 0.4355, "step": 6545 }, { "epoch": 2.094496587030717, "grad_norm": 0.4909462630748749, "learning_rate": 2.5269206642766887e-06, "loss": 0.4478, "step": 6546 }, { "epoch": 2.094816552901024, "grad_norm": 0.514731764793396, "learning_rate": 2.5253027280657792e-06, "loss": 0.4082, "step": 6547 }, { "epoch": 2.095136518771331, "grad_norm": 0.5522803664207458, "learning_rate": 2.5236851349746242e-06, "loss": 0.5042, "step": 6548 }, { "epoch": 2.0954564846416384, "grad_norm": 0.49755656719207764, "learning_rate": 2.522067885227506e-06, "loss": 0.4522, "step": 6549 }, { "epoch": 2.0957764505119454, "grad_norm": 0.4478907883167267, "learning_rate": 2.520450979048657e-06, "loss": 0.3895, "step": 6550 }, { "epoch": 2.0960964163822524, "grad_norm": 0.5032786726951599, "learning_rate": 2.5188344166622623e-06, "loss": 0.4628, "step": 6551 }, { "epoch": 2.09641638225256, "grad_norm": 0.5694442391395569, "learning_rate": 2.5172181982924593e-06, "loss": 0.4592, "step": 6552 }, { "epoch": 2.096736348122867, "grad_norm": 0.5020767450332642, "learning_rate": 2.5156023241633394e-06, "loss": 0.4322, "step": 6553 }, { "epoch": 2.097056313993174, "grad_norm": 0.5417909622192383, "learning_rate": 2.5139867944989483e-06, "loss": 0.4904, "step": 6554 }, { "epoch": 2.0973762798634814, "grad_norm": 0.49538323283195496, "learning_rate": 2.5123716095232784e-06, "loss": 0.4396, "step": 6555 }, { "epoch": 2.0976962457337884, "grad_norm": 0.5189866423606873, "learning_rate": 2.5107567694602776e-06, "loss": 0.4718, "step": 6556 }, { "epoch": 2.0980162116040955, "grad_norm": 0.5295016765594482, "learning_rate": 2.509142274533844e-06, "loss": 0.3919, "step": 6557 }, { "epoch": 2.098336177474403, "grad_norm": 0.5462619662284851, "learning_rate": 2.50752812496783e-06, "loss": 0.5043, "step": 6558 }, { "epoch": 2.09865614334471, "grad_norm": 0.5347604751586914, "learning_rate": 2.5059143209860428e-06, "loss": 0.4307, "step": 6559 }, { "epoch": 2.098976109215017, "grad_norm": 0.49428123235702515, "learning_rate": 2.5043008628122346e-06, "loss": 0.432, "step": 6560 }, { "epoch": 2.0992960750853245, "grad_norm": 0.5311827659606934, "learning_rate": 2.502687750670114e-06, "loss": 0.4273, "step": 6561 }, { "epoch": 2.0996160409556315, "grad_norm": 0.6081832647323608, "learning_rate": 2.50107498478334e-06, "loss": 0.4983, "step": 6562 }, { "epoch": 2.0999360068259385, "grad_norm": 0.4512195885181427, "learning_rate": 2.499462565375525e-06, "loss": 0.4293, "step": 6563 }, { "epoch": 2.1002559726962455, "grad_norm": 0.493958443403244, "learning_rate": 2.497850492670236e-06, "loss": 0.4315, "step": 6564 }, { "epoch": 2.100575938566553, "grad_norm": 0.48001089692115784, "learning_rate": 2.496238766890986e-06, "loss": 0.42, "step": 6565 }, { "epoch": 2.10089590443686, "grad_norm": 0.5479942560195923, "learning_rate": 2.4946273882612432e-06, "loss": 0.4792, "step": 6566 }, { "epoch": 2.101215870307167, "grad_norm": 0.5242119431495667, "learning_rate": 2.4930163570044245e-06, "loss": 0.4022, "step": 6567 }, { "epoch": 2.1015358361774745, "grad_norm": 0.58072429895401, "learning_rate": 2.4914056733439044e-06, "loss": 0.521, "step": 6568 }, { "epoch": 2.1018558020477816, "grad_norm": 0.49397119879722595, "learning_rate": 2.4897953375030078e-06, "loss": 0.3568, "step": 6569 }, { "epoch": 2.1021757679180886, "grad_norm": 0.5828869342803955, "learning_rate": 2.4881853497050074e-06, "loss": 0.4521, "step": 6570 }, { "epoch": 2.102495733788396, "grad_norm": 0.5099934935569763, "learning_rate": 2.4865757101731295e-06, "loss": 0.4379, "step": 6571 }, { "epoch": 2.102815699658703, "grad_norm": 0.5592410564422607, "learning_rate": 2.4849664191305517e-06, "loss": 0.4023, "step": 6572 }, { "epoch": 2.10313566552901, "grad_norm": 0.5595393776893616, "learning_rate": 2.483357476800405e-06, "loss": 0.4601, "step": 6573 }, { "epoch": 2.1034556313993176, "grad_norm": 0.5343798398971558, "learning_rate": 2.4817488834057744e-06, "loss": 0.4454, "step": 6574 }, { "epoch": 2.1037755972696246, "grad_norm": 0.4935324490070343, "learning_rate": 2.4801406391696907e-06, "loss": 0.3637, "step": 6575 }, { "epoch": 2.1040955631399316, "grad_norm": 0.49617499113082886, "learning_rate": 2.4785327443151385e-06, "loss": 0.4685, "step": 6576 }, { "epoch": 2.104415529010239, "grad_norm": 0.4559192359447479, "learning_rate": 2.476925199065053e-06, "loss": 0.3727, "step": 6577 }, { "epoch": 2.104735494880546, "grad_norm": 0.5002991557121277, "learning_rate": 2.4753180036423237e-06, "loss": 0.4793, "step": 6578 }, { "epoch": 2.105055460750853, "grad_norm": 0.4963497817516327, "learning_rate": 2.473711158269792e-06, "loss": 0.4534, "step": 6579 }, { "epoch": 2.1053754266211606, "grad_norm": 0.4909219443798065, "learning_rate": 2.4721046631702478e-06, "loss": 0.4649, "step": 6580 }, { "epoch": 2.1056953924914676, "grad_norm": 0.5330654978752136, "learning_rate": 2.470498518566433e-06, "loss": 0.4395, "step": 6581 }, { "epoch": 2.1060153583617747, "grad_norm": 0.5223993062973022, "learning_rate": 2.46889272468104e-06, "loss": 0.4382, "step": 6582 }, { "epoch": 2.1063353242320817, "grad_norm": 0.49935439229011536, "learning_rate": 2.4672872817367162e-06, "loss": 0.4072, "step": 6583 }, { "epoch": 2.106655290102389, "grad_norm": 0.5430231094360352, "learning_rate": 2.465682189956059e-06, "loss": 0.4407, "step": 6584 }, { "epoch": 2.106975255972696, "grad_norm": 0.4854907691478729, "learning_rate": 2.4640774495616147e-06, "loss": 0.3854, "step": 6585 }, { "epoch": 2.107295221843003, "grad_norm": 0.4910756051540375, "learning_rate": 2.4624730607758833e-06, "loss": 0.467, "step": 6586 }, { "epoch": 2.1076151877133107, "grad_norm": 0.5427428483963013, "learning_rate": 2.4608690238213128e-06, "loss": 0.4444, "step": 6587 }, { "epoch": 2.1079351535836177, "grad_norm": 0.49487268924713135, "learning_rate": 2.459265338920307e-06, "loss": 0.3803, "step": 6588 }, { "epoch": 2.1082551194539247, "grad_norm": 0.512978196144104, "learning_rate": 2.45766200629522e-06, "loss": 0.4333, "step": 6589 }, { "epoch": 2.108575085324232, "grad_norm": 0.5519300699234009, "learning_rate": 2.456059026168355e-06, "loss": 0.4219, "step": 6590 }, { "epoch": 2.1088950511945392, "grad_norm": 0.5403918623924255, "learning_rate": 2.454456398761966e-06, "loss": 0.4367, "step": 6591 }, { "epoch": 2.1092150170648463, "grad_norm": 0.49616748094558716, "learning_rate": 2.452854124298257e-06, "loss": 0.4434, "step": 6592 }, { "epoch": 2.1095349829351537, "grad_norm": 0.512844443321228, "learning_rate": 2.451252202999389e-06, "loss": 0.4193, "step": 6593 }, { "epoch": 2.1098549488054608, "grad_norm": 0.5188925862312317, "learning_rate": 2.449650635087471e-06, "loss": 0.4732, "step": 6594 }, { "epoch": 2.110174914675768, "grad_norm": 0.507415235042572, "learning_rate": 2.4480494207845607e-06, "loss": 0.4661, "step": 6595 }, { "epoch": 2.1104948805460753, "grad_norm": 0.4596986770629883, "learning_rate": 2.4464485603126665e-06, "loss": 0.3773, "step": 6596 }, { "epoch": 2.1108148464163823, "grad_norm": 0.5360031127929688, "learning_rate": 2.4448480538937523e-06, "loss": 0.4952, "step": 6597 }, { "epoch": 2.1111348122866893, "grad_norm": 0.5112121105194092, "learning_rate": 2.4432479017497285e-06, "loss": 0.4226, "step": 6598 }, { "epoch": 2.111454778156997, "grad_norm": 0.4723941385746002, "learning_rate": 2.44164810410246e-06, "loss": 0.4349, "step": 6599 }, { "epoch": 2.111774744027304, "grad_norm": 0.5009363293647766, "learning_rate": 2.4400486611737605e-06, "loss": 0.486, "step": 6600 }, { "epoch": 2.112094709897611, "grad_norm": 0.5131105184555054, "learning_rate": 2.4384495731853918e-06, "loss": 0.4278, "step": 6601 }, { "epoch": 2.1124146757679183, "grad_norm": 0.5704697966575623, "learning_rate": 2.436850840359073e-06, "loss": 0.502, "step": 6602 }, { "epoch": 2.1127346416382253, "grad_norm": 0.48064613342285156, "learning_rate": 2.435252462916467e-06, "loss": 0.3816, "step": 6603 }, { "epoch": 2.1130546075085324, "grad_norm": 0.5179963707923889, "learning_rate": 2.433654441079194e-06, "loss": 0.4441, "step": 6604 }, { "epoch": 2.1133745733788394, "grad_norm": 0.5865185856819153, "learning_rate": 2.4320567750688206e-06, "loss": 0.4794, "step": 6605 }, { "epoch": 2.113694539249147, "grad_norm": 0.4456147253513336, "learning_rate": 2.4304594651068626e-06, "loss": 0.3302, "step": 6606 }, { "epoch": 2.114014505119454, "grad_norm": 0.5374706387519836, "learning_rate": 2.4288625114147914e-06, "loss": 0.4636, "step": 6607 }, { "epoch": 2.114334470989761, "grad_norm": 0.49862635135650635, "learning_rate": 2.427265914214027e-06, "loss": 0.4367, "step": 6608 }, { "epoch": 2.1146544368600684, "grad_norm": 0.43899568915367126, "learning_rate": 2.4256696737259393e-06, "loss": 0.4391, "step": 6609 }, { "epoch": 2.1149744027303754, "grad_norm": 0.5214098691940308, "learning_rate": 2.424073790171849e-06, "loss": 0.4664, "step": 6610 }, { "epoch": 2.1152943686006824, "grad_norm": 0.4941417872905731, "learning_rate": 2.422478263773023e-06, "loss": 0.4319, "step": 6611 }, { "epoch": 2.11561433447099, "grad_norm": 0.4795205593109131, "learning_rate": 2.420883094750688e-06, "loss": 0.431, "step": 6612 }, { "epoch": 2.115934300341297, "grad_norm": 0.5306460857391357, "learning_rate": 2.419288283326016e-06, "loss": 0.4316, "step": 6613 }, { "epoch": 2.116254266211604, "grad_norm": 0.48654627799987793, "learning_rate": 2.4176938297201285e-06, "loss": 0.3964, "step": 6614 }, { "epoch": 2.1165742320819114, "grad_norm": 0.529757022857666, "learning_rate": 2.4160997341540983e-06, "loss": 0.4257, "step": 6615 }, { "epoch": 2.1168941979522184, "grad_norm": 0.5562413930892944, "learning_rate": 2.414505996848946e-06, "loss": 0.4464, "step": 6616 }, { "epoch": 2.1172141638225255, "grad_norm": 0.4944826066493988, "learning_rate": 2.4129126180256478e-06, "loss": 0.4361, "step": 6617 }, { "epoch": 2.117534129692833, "grad_norm": 0.5483657121658325, "learning_rate": 2.4113195979051296e-06, "loss": 0.4597, "step": 6618 }, { "epoch": 2.11785409556314, "grad_norm": 0.47472381591796875, "learning_rate": 2.409726936708263e-06, "loss": 0.431, "step": 6619 }, { "epoch": 2.118174061433447, "grad_norm": 0.5364698171615601, "learning_rate": 2.408134634655873e-06, "loss": 0.4373, "step": 6620 }, { "epoch": 2.1184940273037545, "grad_norm": 0.531273365020752, "learning_rate": 2.4065426919687322e-06, "loss": 0.4414, "step": 6621 }, { "epoch": 2.1188139931740615, "grad_norm": 0.48729974031448364, "learning_rate": 2.404951108867567e-06, "loss": 0.4352, "step": 6622 }, { "epoch": 2.1191339590443685, "grad_norm": 0.5032392740249634, "learning_rate": 2.4033598855730544e-06, "loss": 0.421, "step": 6623 }, { "epoch": 2.1194539249146755, "grad_norm": 0.5721447467803955, "learning_rate": 2.4017690223058175e-06, "loss": 0.4736, "step": 6624 }, { "epoch": 2.119773890784983, "grad_norm": 0.4929790794849396, "learning_rate": 2.4001785192864314e-06, "loss": 0.4165, "step": 6625 }, { "epoch": 2.12009385665529, "grad_norm": 0.48628777265548706, "learning_rate": 2.3985883767354194e-06, "loss": 0.4266, "step": 6626 }, { "epoch": 2.120413822525597, "grad_norm": 0.48934853076934814, "learning_rate": 2.3969985948732586e-06, "loss": 0.443, "step": 6627 }, { "epoch": 2.1207337883959045, "grad_norm": 0.4963590204715729, "learning_rate": 2.395409173920376e-06, "loss": 0.4678, "step": 6628 }, { "epoch": 2.1210537542662116, "grad_norm": 0.5006743669509888, "learning_rate": 2.393820114097145e-06, "loss": 0.4004, "step": 6629 }, { "epoch": 2.1213737201365186, "grad_norm": 0.5066134929656982, "learning_rate": 2.3922314156238883e-06, "loss": 0.4578, "step": 6630 }, { "epoch": 2.121693686006826, "grad_norm": 0.5675395131111145, "learning_rate": 2.390643078720885e-06, "loss": 0.4805, "step": 6631 }, { "epoch": 2.122013651877133, "grad_norm": 0.5577268004417419, "learning_rate": 2.3890551036083564e-06, "loss": 0.5025, "step": 6632 }, { "epoch": 2.12233361774744, "grad_norm": 0.5691620707511902, "learning_rate": 2.3874674905064804e-06, "loss": 0.4362, "step": 6633 }, { "epoch": 2.1226535836177476, "grad_norm": 0.497223436832428, "learning_rate": 2.3858802396353796e-06, "loss": 0.4201, "step": 6634 }, { "epoch": 2.1229735494880546, "grad_norm": 0.5128043293952942, "learning_rate": 2.3842933512151268e-06, "loss": 0.4268, "step": 6635 }, { "epoch": 2.1232935153583616, "grad_norm": 0.5296140909194946, "learning_rate": 2.3827068254657493e-06, "loss": 0.4759, "step": 6636 }, { "epoch": 2.123613481228669, "grad_norm": 0.5045025944709778, "learning_rate": 2.381120662607217e-06, "loss": 0.4376, "step": 6637 }, { "epoch": 2.123933447098976, "grad_norm": 0.48082995414733887, "learning_rate": 2.379534862859458e-06, "loss": 0.4393, "step": 6638 }, { "epoch": 2.124253412969283, "grad_norm": 0.5115391612052917, "learning_rate": 2.3779494264423427e-06, "loss": 0.4435, "step": 6639 }, { "epoch": 2.1245733788395906, "grad_norm": 0.5318616628646851, "learning_rate": 2.3763643535756914e-06, "loss": 0.4589, "step": 6640 }, { "epoch": 2.1248933447098977, "grad_norm": 0.48749759793281555, "learning_rate": 2.374779644479281e-06, "loss": 0.4599, "step": 6641 }, { "epoch": 2.1252133105802047, "grad_norm": 0.5193473696708679, "learning_rate": 2.373195299372829e-06, "loss": 0.4457, "step": 6642 }, { "epoch": 2.125533276450512, "grad_norm": 0.4633994400501251, "learning_rate": 2.371611318476011e-06, "loss": 0.4168, "step": 6643 }, { "epoch": 2.125853242320819, "grad_norm": 0.5086625814437866, "learning_rate": 2.3700277020084454e-06, "loss": 0.4255, "step": 6644 }, { "epoch": 2.126173208191126, "grad_norm": 0.4725306034088135, "learning_rate": 2.3684444501897012e-06, "loss": 0.463, "step": 6645 }, { "epoch": 2.1264931740614337, "grad_norm": 0.49431467056274414, "learning_rate": 2.366861563239302e-06, "loss": 0.454, "step": 6646 }, { "epoch": 2.1268131399317407, "grad_norm": 0.5001236796379089, "learning_rate": 2.3652790413767125e-06, "loss": 0.4689, "step": 6647 }, { "epoch": 2.1271331058020477, "grad_norm": 0.49575915932655334, "learning_rate": 2.363696884821355e-06, "loss": 0.4317, "step": 6648 }, { "epoch": 2.1274530716723548, "grad_norm": 0.5135654807090759, "learning_rate": 2.362115093792597e-06, "loss": 0.4244, "step": 6649 }, { "epoch": 2.1277730375426622, "grad_norm": 0.5357350707054138, "learning_rate": 2.3605336685097523e-06, "loss": 0.4559, "step": 6650 }, { "epoch": 2.1280930034129693, "grad_norm": 0.4895274341106415, "learning_rate": 2.3589526091920915e-06, "loss": 0.3771, "step": 6651 }, { "epoch": 2.1284129692832763, "grad_norm": 0.5519123673439026, "learning_rate": 2.357371916058827e-06, "loss": 0.4447, "step": 6652 }, { "epoch": 2.1287329351535837, "grad_norm": 0.5529068112373352, "learning_rate": 2.3557915893291276e-06, "loss": 0.4291, "step": 6653 }, { "epoch": 2.1290529010238908, "grad_norm": 0.5326127409934998, "learning_rate": 2.354211629222105e-06, "loss": 0.4858, "step": 6654 }, { "epoch": 2.129372866894198, "grad_norm": 0.5279597640037537, "learning_rate": 2.3526320359568215e-06, "loss": 0.4218, "step": 6655 }, { "epoch": 2.1296928327645053, "grad_norm": 0.4971260726451874, "learning_rate": 2.3510528097522928e-06, "loss": 0.41, "step": 6656 }, { "epoch": 2.1300127986348123, "grad_norm": 0.4707668423652649, "learning_rate": 2.3494739508274765e-06, "loss": 0.4274, "step": 6657 }, { "epoch": 2.1303327645051193, "grad_norm": 0.5016271471977234, "learning_rate": 2.3478954594012884e-06, "loss": 0.4666, "step": 6658 }, { "epoch": 2.130652730375427, "grad_norm": 0.5279699563980103, "learning_rate": 2.3463173356925843e-06, "loss": 0.4901, "step": 6659 }, { "epoch": 2.130972696245734, "grad_norm": 0.4883582890033722, "learning_rate": 2.3447395799201733e-06, "loss": 0.4051, "step": 6660 }, { "epoch": 2.131292662116041, "grad_norm": 0.561225950717926, "learning_rate": 2.3431621923028146e-06, "loss": 0.4648, "step": 6661 }, { "epoch": 2.1316126279863483, "grad_norm": 0.5723588466644287, "learning_rate": 2.341585173059213e-06, "loss": 0.4945, "step": 6662 }, { "epoch": 2.1319325938566553, "grad_norm": 0.5918015241622925, "learning_rate": 2.340008522408027e-06, "loss": 0.4603, "step": 6663 }, { "epoch": 2.1322525597269624, "grad_norm": 0.5477120876312256, "learning_rate": 2.3384322405678576e-06, "loss": 0.4417, "step": 6664 }, { "epoch": 2.1325725255972694, "grad_norm": 0.5312170386314392, "learning_rate": 2.3368563277572624e-06, "loss": 0.4679, "step": 6665 }, { "epoch": 2.132892491467577, "grad_norm": 0.46622106432914734, "learning_rate": 2.3352807841947406e-06, "loss": 0.4656, "step": 6666 }, { "epoch": 2.133212457337884, "grad_norm": 0.4596792459487915, "learning_rate": 2.333705610098742e-06, "loss": 0.3897, "step": 6667 }, { "epoch": 2.133532423208191, "grad_norm": 0.5101672410964966, "learning_rate": 2.33213080568767e-06, "loss": 0.4683, "step": 6668 }, { "epoch": 2.1338523890784984, "grad_norm": 0.5518092513084412, "learning_rate": 2.3305563711798696e-06, "loss": 0.5045, "step": 6669 }, { "epoch": 2.1341723549488054, "grad_norm": 0.6030656695365906, "learning_rate": 2.328982306793642e-06, "loss": 0.49, "step": 6670 }, { "epoch": 2.1344923208191124, "grad_norm": 0.5064455270767212, "learning_rate": 2.32740861274723e-06, "loss": 0.4244, "step": 6671 }, { "epoch": 2.13481228668942, "grad_norm": 0.5312418937683105, "learning_rate": 2.3258352892588277e-06, "loss": 0.4291, "step": 6672 }, { "epoch": 2.135132252559727, "grad_norm": 0.5422471165657043, "learning_rate": 2.3242623365465816e-06, "loss": 0.4856, "step": 6673 }, { "epoch": 2.135452218430034, "grad_norm": 0.45353472232818604, "learning_rate": 2.322689754828579e-06, "loss": 0.4177, "step": 6674 }, { "epoch": 2.1357721843003414, "grad_norm": 0.5094054341316223, "learning_rate": 2.3211175443228655e-06, "loss": 0.4675, "step": 6675 }, { "epoch": 2.1360921501706485, "grad_norm": 0.4616287648677826, "learning_rate": 2.319545705247427e-06, "loss": 0.4171, "step": 6676 }, { "epoch": 2.1364121160409555, "grad_norm": 0.5320475101470947, "learning_rate": 2.317974237820199e-06, "loss": 0.4731, "step": 6677 }, { "epoch": 2.136732081911263, "grad_norm": 0.5035117268562317, "learning_rate": 2.3164031422590715e-06, "loss": 0.3585, "step": 6678 }, { "epoch": 2.13705204778157, "grad_norm": 0.5568263530731201, "learning_rate": 2.314832418781875e-06, "loss": 0.4652, "step": 6679 }, { "epoch": 2.137372013651877, "grad_norm": 0.5261285901069641, "learning_rate": 2.313262067606396e-06, "loss": 0.461, "step": 6680 }, { "epoch": 2.1376919795221845, "grad_norm": 0.5170149803161621, "learning_rate": 2.311692088950363e-06, "loss": 0.4082, "step": 6681 }, { "epoch": 2.1380119453924915, "grad_norm": 0.5495178699493408, "learning_rate": 2.3101224830314546e-06, "loss": 0.5008, "step": 6682 }, { "epoch": 2.1383319112627985, "grad_norm": 0.5265787243843079, "learning_rate": 2.308553250067302e-06, "loss": 0.4261, "step": 6683 }, { "epoch": 2.138651877133106, "grad_norm": 0.5479193925857544, "learning_rate": 2.3069843902754767e-06, "loss": 0.4252, "step": 6684 }, { "epoch": 2.138971843003413, "grad_norm": 0.4920346140861511, "learning_rate": 2.3054159038735076e-06, "loss": 0.4807, "step": 6685 }, { "epoch": 2.13929180887372, "grad_norm": 0.5257879495620728, "learning_rate": 2.303847791078865e-06, "loss": 0.4687, "step": 6686 }, { "epoch": 2.1396117747440275, "grad_norm": 0.46145397424697876, "learning_rate": 2.302280052108968e-06, "loss": 0.4159, "step": 6687 }, { "epoch": 2.1399317406143346, "grad_norm": 0.49745118618011475, "learning_rate": 2.3007126871811886e-06, "loss": 0.4237, "step": 6688 }, { "epoch": 2.1402517064846416, "grad_norm": 0.5214394927024841, "learning_rate": 2.299145696512841e-06, "loss": 0.458, "step": 6689 }, { "epoch": 2.1405716723549486, "grad_norm": 0.46344926953315735, "learning_rate": 2.297579080321194e-06, "loss": 0.4218, "step": 6690 }, { "epoch": 2.140891638225256, "grad_norm": 0.4784850478172302, "learning_rate": 2.2960128388234582e-06, "loss": 0.4421, "step": 6691 }, { "epoch": 2.141211604095563, "grad_norm": 0.505057692527771, "learning_rate": 2.294446972236793e-06, "loss": 0.4595, "step": 6692 }, { "epoch": 2.14153156996587, "grad_norm": 0.5089201331138611, "learning_rate": 2.292881480778312e-06, "loss": 0.4492, "step": 6693 }, { "epoch": 2.1418515358361776, "grad_norm": 0.5136494636535645, "learning_rate": 2.291316364665069e-06, "loss": 0.4154, "step": 6694 }, { "epoch": 2.1421715017064846, "grad_norm": 0.5064804553985596, "learning_rate": 2.289751624114073e-06, "loss": 0.4165, "step": 6695 }, { "epoch": 2.1424914675767917, "grad_norm": 0.4527985751628876, "learning_rate": 2.2881872593422737e-06, "loss": 0.3568, "step": 6696 }, { "epoch": 2.142811433447099, "grad_norm": 0.5171433687210083, "learning_rate": 2.286623270566572e-06, "loss": 0.4885, "step": 6697 }, { "epoch": 2.143131399317406, "grad_norm": 0.4854653775691986, "learning_rate": 2.2850596580038196e-06, "loss": 0.4323, "step": 6698 }, { "epoch": 2.143451365187713, "grad_norm": 0.46086299419403076, "learning_rate": 2.28349642187081e-06, "loss": 0.4193, "step": 6699 }, { "epoch": 2.1437713310580206, "grad_norm": 0.4680379629135132, "learning_rate": 2.281933562384292e-06, "loss": 0.4369, "step": 6700 }, { "epoch": 2.1440912969283277, "grad_norm": 0.5378003120422363, "learning_rate": 2.2803710797609545e-06, "loss": 0.5075, "step": 6701 }, { "epoch": 2.1444112627986347, "grad_norm": 0.505118191242218, "learning_rate": 2.2788089742174374e-06, "loss": 0.4213, "step": 6702 }, { "epoch": 2.144731228668942, "grad_norm": 0.5250274538993835, "learning_rate": 2.2772472459703298e-06, "loss": 0.442, "step": 6703 }, { "epoch": 2.145051194539249, "grad_norm": 0.5066376328468323, "learning_rate": 2.2756858952361687e-06, "loss": 0.4603, "step": 6704 }, { "epoch": 2.145371160409556, "grad_norm": 0.49959373474121094, "learning_rate": 2.274124922231436e-06, "loss": 0.4215, "step": 6705 }, { "epoch": 2.1456911262798632, "grad_norm": 0.48590216040611267, "learning_rate": 2.2725643271725617e-06, "loss": 0.4433, "step": 6706 }, { "epoch": 2.1460110921501707, "grad_norm": 0.4953617453575134, "learning_rate": 2.271004110275923e-06, "loss": 0.4682, "step": 6707 }, { "epoch": 2.1463310580204777, "grad_norm": 0.5015279650688171, "learning_rate": 2.2694442717578476e-06, "loss": 0.4483, "step": 6708 }, { "epoch": 2.1466510238907848, "grad_norm": 0.4883991479873657, "learning_rate": 2.2678848118346103e-06, "loss": 0.4773, "step": 6709 }, { "epoch": 2.1469709897610922, "grad_norm": 0.46480992436408997, "learning_rate": 2.2663257307224308e-06, "loss": 0.4195, "step": 6710 }, { "epoch": 2.1472909556313993, "grad_norm": 0.47466591000556946, "learning_rate": 2.264767028637477e-06, "loss": 0.4532, "step": 6711 }, { "epoch": 2.1476109215017063, "grad_norm": 0.5161939263343811, "learning_rate": 2.263208705795864e-06, "loss": 0.4606, "step": 6712 }, { "epoch": 2.1479308873720138, "grad_norm": 0.48355168104171753, "learning_rate": 2.2616507624136564e-06, "loss": 0.4097, "step": 6713 }, { "epoch": 2.148250853242321, "grad_norm": 0.5150226950645447, "learning_rate": 2.2600931987068662e-06, "loss": 0.4998, "step": 6714 }, { "epoch": 2.148570819112628, "grad_norm": 0.4854436218738556, "learning_rate": 2.258536014891451e-06, "loss": 0.4182, "step": 6715 }, { "epoch": 2.1488907849829353, "grad_norm": 0.5275769233703613, "learning_rate": 2.256979211183315e-06, "loss": 0.4387, "step": 6716 }, { "epoch": 2.1492107508532423, "grad_norm": 0.5012706518173218, "learning_rate": 2.2554227877983093e-06, "loss": 0.4187, "step": 6717 }, { "epoch": 2.1495307167235493, "grad_norm": 0.5256766676902771, "learning_rate": 2.253866744952236e-06, "loss": 0.4305, "step": 6718 }, { "epoch": 2.149850682593857, "grad_norm": 0.5074648261070251, "learning_rate": 2.252311082860844e-06, "loss": 0.4624, "step": 6719 }, { "epoch": 2.150170648464164, "grad_norm": 0.5025088787078857, "learning_rate": 2.2507558017398263e-06, "loss": 0.4458, "step": 6720 }, { "epoch": 2.150490614334471, "grad_norm": 0.47776737809181213, "learning_rate": 2.2492009018048234e-06, "loss": 0.4219, "step": 6721 }, { "epoch": 2.1508105802047783, "grad_norm": 0.48077675700187683, "learning_rate": 2.2476463832714233e-06, "loss": 0.4464, "step": 6722 }, { "epoch": 2.1511305460750854, "grad_norm": 0.5319883227348328, "learning_rate": 2.246092246355163e-06, "loss": 0.5005, "step": 6723 }, { "epoch": 2.1514505119453924, "grad_norm": 0.5291938185691833, "learning_rate": 2.2445384912715285e-06, "loss": 0.4437, "step": 6724 }, { "epoch": 2.1517704778157, "grad_norm": 0.5148395895957947, "learning_rate": 2.2429851182359465e-06, "loss": 0.4866, "step": 6725 }, { "epoch": 2.152090443686007, "grad_norm": 0.4877208471298218, "learning_rate": 2.2414321274637946e-06, "loss": 0.401, "step": 6726 }, { "epoch": 2.152410409556314, "grad_norm": 0.5249781608581543, "learning_rate": 2.239879519170395e-06, "loss": 0.4535, "step": 6727 }, { "epoch": 2.1527303754266214, "grad_norm": 0.5234302878379822, "learning_rate": 2.2383272935710205e-06, "loss": 0.4467, "step": 6728 }, { "epoch": 2.1530503412969284, "grad_norm": 0.4273567199707031, "learning_rate": 2.2367754508808914e-06, "loss": 0.3639, "step": 6729 }, { "epoch": 2.1533703071672354, "grad_norm": 0.4548000395298004, "learning_rate": 2.2352239913151696e-06, "loss": 0.4276, "step": 6730 }, { "epoch": 2.1536902730375425, "grad_norm": 0.4965096116065979, "learning_rate": 2.233672915088968e-06, "loss": 0.4642, "step": 6731 }, { "epoch": 2.15401023890785, "grad_norm": 0.5180421471595764, "learning_rate": 2.2321222224173426e-06, "loss": 0.4583, "step": 6732 }, { "epoch": 2.154330204778157, "grad_norm": 0.5392321944236755, "learning_rate": 2.2305719135153013e-06, "loss": 0.4502, "step": 6733 }, { "epoch": 2.154650170648464, "grad_norm": 0.45776090025901794, "learning_rate": 2.2290219885977976e-06, "loss": 0.4079, "step": 6734 }, { "epoch": 2.1549701365187715, "grad_norm": 0.5151669383049011, "learning_rate": 2.2274724478797284e-06, "loss": 0.4057, "step": 6735 }, { "epoch": 2.1552901023890785, "grad_norm": 0.5341548323631287, "learning_rate": 2.22592329157594e-06, "loss": 0.4673, "step": 6736 }, { "epoch": 2.1556100682593855, "grad_norm": 0.5643128156661987, "learning_rate": 2.224374519901222e-06, "loss": 0.4162, "step": 6737 }, { "epoch": 2.155930034129693, "grad_norm": 0.5588071942329407, "learning_rate": 2.2228261330703172e-06, "loss": 0.4296, "step": 6738 }, { "epoch": 2.15625, "grad_norm": 0.5966269969940186, "learning_rate": 2.2212781312979114e-06, "loss": 0.5066, "step": 6739 }, { "epoch": 2.156569965870307, "grad_norm": 0.5233896374702454, "learning_rate": 2.2197305147986352e-06, "loss": 0.3812, "step": 6740 }, { "epoch": 2.1568899317406145, "grad_norm": 0.48629361391067505, "learning_rate": 2.2181832837870683e-06, "loss": 0.4396, "step": 6741 }, { "epoch": 2.1572098976109215, "grad_norm": 0.5437921285629272, "learning_rate": 2.2166364384777337e-06, "loss": 0.4639, "step": 6742 }, { "epoch": 2.1575298634812285, "grad_norm": 0.5298428535461426, "learning_rate": 2.2150899790851056e-06, "loss": 0.4484, "step": 6743 }, { "epoch": 2.157849829351536, "grad_norm": 0.5280128121376038, "learning_rate": 2.2135439058236036e-06, "loss": 0.4625, "step": 6744 }, { "epoch": 2.158169795221843, "grad_norm": 0.5505377054214478, "learning_rate": 2.211998218907592e-06, "loss": 0.444, "step": 6745 }, { "epoch": 2.15848976109215, "grad_norm": 0.4741867184638977, "learning_rate": 2.2104529185513807e-06, "loss": 0.4072, "step": 6746 }, { "epoch": 2.1588097269624575, "grad_norm": 0.4593965411186218, "learning_rate": 2.2089080049692263e-06, "loss": 0.4331, "step": 6747 }, { "epoch": 2.1591296928327646, "grad_norm": 0.4532676339149475, "learning_rate": 2.2073634783753352e-06, "loss": 0.4536, "step": 6748 }, { "epoch": 2.1594496587030716, "grad_norm": 0.5205339193344116, "learning_rate": 2.205819338983859e-06, "loss": 0.4531, "step": 6749 }, { "epoch": 2.1597696245733786, "grad_norm": 0.501815140247345, "learning_rate": 2.2042755870088933e-06, "loss": 0.3847, "step": 6750 }, { "epoch": 2.160089590443686, "grad_norm": 0.48559510707855225, "learning_rate": 2.2027322226644803e-06, "loss": 0.4477, "step": 6751 }, { "epoch": 2.160409556313993, "grad_norm": 0.46358832716941833, "learning_rate": 2.201189246164606e-06, "loss": 0.4196, "step": 6752 }, { "epoch": 2.1607295221843, "grad_norm": 0.5158883333206177, "learning_rate": 2.199646657723214e-06, "loss": 0.4434, "step": 6753 }, { "epoch": 2.1610494880546076, "grad_norm": 0.46989548206329346, "learning_rate": 2.198104457554182e-06, "loss": 0.4095, "step": 6754 }, { "epoch": 2.1613694539249146, "grad_norm": 0.5143417119979858, "learning_rate": 2.1965626458713375e-06, "loss": 0.4407, "step": 6755 }, { "epoch": 2.1616894197952217, "grad_norm": 0.5397100448608398, "learning_rate": 2.1950212228884544e-06, "loss": 0.4951, "step": 6756 }, { "epoch": 2.162009385665529, "grad_norm": 0.4593190550804138, "learning_rate": 2.19348018881925e-06, "loss": 0.3479, "step": 6757 }, { "epoch": 2.162329351535836, "grad_norm": 0.5357738733291626, "learning_rate": 2.191939543877397e-06, "loss": 0.4345, "step": 6758 }, { "epoch": 2.162649317406143, "grad_norm": 0.47553011775016785, "learning_rate": 2.1903992882765048e-06, "loss": 0.4692, "step": 6759 }, { "epoch": 2.1629692832764507, "grad_norm": 0.49455299973487854, "learning_rate": 2.1888594222301307e-06, "loss": 0.3952, "step": 6760 }, { "epoch": 2.1632892491467577, "grad_norm": 0.48963063955307007, "learning_rate": 2.18731994595178e-06, "loss": 0.4103, "step": 6761 }, { "epoch": 2.1636092150170647, "grad_norm": 0.5070008635520935, "learning_rate": 2.1857808596548992e-06, "loss": 0.5065, "step": 6762 }, { "epoch": 2.163929180887372, "grad_norm": 0.4946386516094208, "learning_rate": 2.184242163552891e-06, "loss": 0.4266, "step": 6763 }, { "epoch": 2.164249146757679, "grad_norm": 0.5226266980171204, "learning_rate": 2.182703857859094e-06, "loss": 0.4362, "step": 6764 }, { "epoch": 2.1645691126279862, "grad_norm": 0.5144748091697693, "learning_rate": 2.181165942786797e-06, "loss": 0.4748, "step": 6765 }, { "epoch": 2.1648890784982937, "grad_norm": 0.45408496260643005, "learning_rate": 2.1796284185492333e-06, "loss": 0.388, "step": 6766 }, { "epoch": 2.1652090443686007, "grad_norm": 0.5466834306716919, "learning_rate": 2.178091285359578e-06, "loss": 0.4747, "step": 6767 }, { "epoch": 2.1655290102389078, "grad_norm": 0.4702305495738983, "learning_rate": 2.176554543430965e-06, "loss": 0.4363, "step": 6768 }, { "epoch": 2.1658489761092152, "grad_norm": 0.5069165825843811, "learning_rate": 2.175018192976461e-06, "loss": 0.4225, "step": 6769 }, { "epoch": 2.1661689419795223, "grad_norm": 0.5441505312919617, "learning_rate": 2.1734822342090835e-06, "loss": 0.4543, "step": 6770 }, { "epoch": 2.1664889078498293, "grad_norm": 0.479708731174469, "learning_rate": 2.171946667341792e-06, "loss": 0.4277, "step": 6771 }, { "epoch": 2.1668088737201368, "grad_norm": 0.47547414898872375, "learning_rate": 2.170411492587498e-06, "loss": 0.437, "step": 6772 }, { "epoch": 2.167128839590444, "grad_norm": 0.5153126120567322, "learning_rate": 2.1688767101590564e-06, "loss": 0.4622, "step": 6773 }, { "epoch": 2.167448805460751, "grad_norm": 0.5160109996795654, "learning_rate": 2.1673423202692643e-06, "loss": 0.4555, "step": 6774 }, { "epoch": 2.167768771331058, "grad_norm": 0.48195773363113403, "learning_rate": 2.165808323130868e-06, "loss": 0.472, "step": 6775 }, { "epoch": 2.1680887372013653, "grad_norm": 0.45776796340942383, "learning_rate": 2.164274718956555e-06, "loss": 0.439, "step": 6776 }, { "epoch": 2.1684087030716723, "grad_norm": 0.4817793071269989, "learning_rate": 2.1627415079589637e-06, "loss": 0.427, "step": 6777 }, { "epoch": 2.1687286689419794, "grad_norm": 0.4975053071975708, "learning_rate": 2.161208690350678e-06, "loss": 0.4498, "step": 6778 }, { "epoch": 2.169048634812287, "grad_norm": 0.460365891456604, "learning_rate": 2.159676266344222e-06, "loss": 0.378, "step": 6779 }, { "epoch": 2.169368600682594, "grad_norm": 0.512991189956665, "learning_rate": 2.1581442361520688e-06, "loss": 0.4484, "step": 6780 }, { "epoch": 2.169688566552901, "grad_norm": 0.5276342034339905, "learning_rate": 2.156612599986634e-06, "loss": 0.4575, "step": 6781 }, { "epoch": 2.1700085324232083, "grad_norm": 0.505875289440155, "learning_rate": 2.155081358060283e-06, "loss": 0.4636, "step": 6782 }, { "epoch": 2.1703284982935154, "grad_norm": 0.5176311135292053, "learning_rate": 2.153550510585326e-06, "loss": 0.466, "step": 6783 }, { "epoch": 2.1706484641638224, "grad_norm": 0.590843677520752, "learning_rate": 2.1520200577740154e-06, "loss": 0.4663, "step": 6784 }, { "epoch": 2.17096843003413, "grad_norm": 0.4701339900493622, "learning_rate": 2.1504899998385497e-06, "loss": 0.3819, "step": 6785 }, { "epoch": 2.171288395904437, "grad_norm": 0.47744128108024597, "learning_rate": 2.1489603369910724e-06, "loss": 0.4367, "step": 6786 }, { "epoch": 2.171608361774744, "grad_norm": 0.5769877433776855, "learning_rate": 2.1474310694436732e-06, "loss": 0.5047, "step": 6787 }, { "epoch": 2.1719283276450514, "grad_norm": 0.5574608445167542, "learning_rate": 2.1459021974083905e-06, "loss": 0.4447, "step": 6788 }, { "epoch": 2.1722482935153584, "grad_norm": 0.5587093830108643, "learning_rate": 2.1443737210972017e-06, "loss": 0.4411, "step": 6789 }, { "epoch": 2.1725682593856654, "grad_norm": 0.5133920907974243, "learning_rate": 2.1428456407220317e-06, "loss": 0.4533, "step": 6790 }, { "epoch": 2.1728882252559725, "grad_norm": 0.47900161147117615, "learning_rate": 2.141317956494749e-06, "loss": 0.4772, "step": 6791 }, { "epoch": 2.17320819112628, "grad_norm": 0.5402735471725464, "learning_rate": 2.1397906686271706e-06, "loss": 0.4177, "step": 6792 }, { "epoch": 2.173528156996587, "grad_norm": 0.5899126529693604, "learning_rate": 2.1382637773310587e-06, "loss": 0.4572, "step": 6793 }, { "epoch": 2.173848122866894, "grad_norm": 0.5434718728065491, "learning_rate": 2.136737282818117e-06, "loss": 0.4933, "step": 6794 }, { "epoch": 2.1741680887372015, "grad_norm": 0.4832441210746765, "learning_rate": 2.135211185299996e-06, "loss": 0.4025, "step": 6795 }, { "epoch": 2.1744880546075085, "grad_norm": 0.5162698030471802, "learning_rate": 2.1336854849882883e-06, "loss": 0.3842, "step": 6796 }, { "epoch": 2.1748080204778155, "grad_norm": 0.476591020822525, "learning_rate": 2.132160182094536e-06, "loss": 0.438, "step": 6797 }, { "epoch": 2.175127986348123, "grad_norm": 0.486438512802124, "learning_rate": 2.130635276830227e-06, "loss": 0.4443, "step": 6798 }, { "epoch": 2.17544795221843, "grad_norm": 0.5168624520301819, "learning_rate": 2.1291107694067885e-06, "loss": 0.3907, "step": 6799 }, { "epoch": 2.175767918088737, "grad_norm": 0.5580846071243286, "learning_rate": 2.1275866600355954e-06, "loss": 0.4305, "step": 6800 }, { "epoch": 2.1760878839590445, "grad_norm": 0.5284036993980408, "learning_rate": 2.1260629489279662e-06, "loss": 0.443, "step": 6801 }, { "epoch": 2.1764078498293515, "grad_norm": 0.4435729682445526, "learning_rate": 2.1245396362951664e-06, "loss": 0.4378, "step": 6802 }, { "epoch": 2.1767278156996586, "grad_norm": 0.49425411224365234, "learning_rate": 2.1230167223484072e-06, "loss": 0.4531, "step": 6803 }, { "epoch": 2.177047781569966, "grad_norm": 0.5321007370948792, "learning_rate": 2.12149420729884e-06, "loss": 0.4399, "step": 6804 }, { "epoch": 2.177367747440273, "grad_norm": 0.578448474407196, "learning_rate": 2.119972091357563e-06, "loss": 0.405, "step": 6805 }, { "epoch": 2.17768771331058, "grad_norm": 0.5024625062942505, "learning_rate": 2.118450374735621e-06, "loss": 0.469, "step": 6806 }, { "epoch": 2.1780076791808876, "grad_norm": 0.5228381156921387, "learning_rate": 2.116929057644e-06, "loss": 0.463, "step": 6807 }, { "epoch": 2.1783276450511946, "grad_norm": 0.4660845696926117, "learning_rate": 2.115408140293635e-06, "loss": 0.3963, "step": 6808 }, { "epoch": 2.1786476109215016, "grad_norm": 0.4882470369338989, "learning_rate": 2.113887622895402e-06, "loss": 0.4072, "step": 6809 }, { "epoch": 2.178967576791809, "grad_norm": 0.5116421580314636, "learning_rate": 2.11236750566012e-06, "loss": 0.4615, "step": 6810 }, { "epoch": 2.179287542662116, "grad_norm": 0.5525357723236084, "learning_rate": 2.1108477887985597e-06, "loss": 0.4693, "step": 6811 }, { "epoch": 2.179607508532423, "grad_norm": 0.491583913564682, "learning_rate": 2.109328472521427e-06, "loss": 0.3816, "step": 6812 }, { "epoch": 2.1799274744027306, "grad_norm": 0.5098915100097656, "learning_rate": 2.107809557039381e-06, "loss": 0.4588, "step": 6813 }, { "epoch": 2.1802474402730376, "grad_norm": 0.49412649869918823, "learning_rate": 2.106291042563019e-06, "loss": 0.4442, "step": 6814 }, { "epoch": 2.1805674061433447, "grad_norm": 0.49511006474494934, "learning_rate": 2.1047729293028834e-06, "loss": 0.4347, "step": 6815 }, { "epoch": 2.1808873720136517, "grad_norm": 0.5126012563705444, "learning_rate": 2.103255217469466e-06, "loss": 0.4268, "step": 6816 }, { "epoch": 2.181207337883959, "grad_norm": 0.46396976709365845, "learning_rate": 2.101737907273196e-06, "loss": 0.4187, "step": 6817 }, { "epoch": 2.181527303754266, "grad_norm": 0.5442050695419312, "learning_rate": 2.100220998924454e-06, "loss": 0.4358, "step": 6818 }, { "epoch": 2.181847269624573, "grad_norm": 0.5309011936187744, "learning_rate": 2.0987044926335583e-06, "loss": 0.4384, "step": 6819 }, { "epoch": 2.1821672354948807, "grad_norm": 0.5154093503952026, "learning_rate": 2.0971883886107737e-06, "loss": 0.3808, "step": 6820 }, { "epoch": 2.1824872013651877, "grad_norm": 0.45877769589424133, "learning_rate": 2.0956726870663123e-06, "loss": 0.3898, "step": 6821 }, { "epoch": 2.1828071672354947, "grad_norm": 0.48768162727355957, "learning_rate": 2.094157388210326e-06, "loss": 0.4781, "step": 6822 }, { "epoch": 2.183127133105802, "grad_norm": 0.4783969521522522, "learning_rate": 2.092642492252915e-06, "loss": 0.3945, "step": 6823 }, { "epoch": 2.1834470989761092, "grad_norm": 0.5190688967704773, "learning_rate": 2.09112799940412e-06, "loss": 0.4313, "step": 6824 }, { "epoch": 2.1837670648464163, "grad_norm": 0.5493019819259644, "learning_rate": 2.089613909873926e-06, "loss": 0.463, "step": 6825 }, { "epoch": 2.1840870307167237, "grad_norm": 0.4760606586933136, "learning_rate": 2.0881002238722657e-06, "loss": 0.4331, "step": 6826 }, { "epoch": 2.1844069965870307, "grad_norm": 0.49167048931121826, "learning_rate": 2.086586941609011e-06, "loss": 0.4203, "step": 6827 }, { "epoch": 2.1847269624573378, "grad_norm": 0.5753871202468872, "learning_rate": 2.0850740632939837e-06, "loss": 0.438, "step": 6828 }, { "epoch": 2.1850469283276452, "grad_norm": 0.5460748076438904, "learning_rate": 2.083561589136944e-06, "loss": 0.5011, "step": 6829 }, { "epoch": 2.1853668941979523, "grad_norm": 0.47245848178863525, "learning_rate": 2.0820495193475972e-06, "loss": 0.3838, "step": 6830 }, { "epoch": 2.1856868600682593, "grad_norm": 0.5486859679222107, "learning_rate": 2.0805378541355965e-06, "loss": 0.4317, "step": 6831 }, { "epoch": 2.1860068259385663, "grad_norm": 0.5436904430389404, "learning_rate": 2.0790265937105325e-06, "loss": 0.4212, "step": 6832 }, { "epoch": 2.186326791808874, "grad_norm": 0.5365811586380005, "learning_rate": 2.0775157382819473e-06, "loss": 0.4391, "step": 6833 }, { "epoch": 2.186646757679181, "grad_norm": 0.514493465423584, "learning_rate": 2.0760052880593213e-06, "loss": 0.4833, "step": 6834 }, { "epoch": 2.186966723549488, "grad_norm": 0.523497462272644, "learning_rate": 2.074495243252077e-06, "loss": 0.4956, "step": 6835 }, { "epoch": 2.1872866894197953, "grad_norm": 0.45324259996414185, "learning_rate": 2.072985604069589e-06, "loss": 0.3956, "step": 6836 }, { "epoch": 2.1876066552901023, "grad_norm": 0.48483917117118835, "learning_rate": 2.0714763707211664e-06, "loss": 0.4065, "step": 6837 }, { "epoch": 2.1879266211604094, "grad_norm": 0.49716633558273315, "learning_rate": 2.06996754341607e-06, "loss": 0.4386, "step": 6838 }, { "epoch": 2.188246587030717, "grad_norm": 0.510878324508667, "learning_rate": 2.0684591223634986e-06, "loss": 0.4697, "step": 6839 }, { "epoch": 2.188566552901024, "grad_norm": 0.4646270275115967, "learning_rate": 2.0669511077725945e-06, "loss": 0.3907, "step": 6840 }, { "epoch": 2.188886518771331, "grad_norm": 0.4669325053691864, "learning_rate": 2.0654434998524502e-06, "loss": 0.4183, "step": 6841 }, { "epoch": 2.1892064846416384, "grad_norm": 0.4611457884311676, "learning_rate": 2.0639362988120933e-06, "loss": 0.4072, "step": 6842 }, { "epoch": 2.1895264505119454, "grad_norm": 0.5439834594726562, "learning_rate": 2.062429504860502e-06, "loss": 0.4816, "step": 6843 }, { "epoch": 2.1898464163822524, "grad_norm": 0.4837367534637451, "learning_rate": 2.060923118206592e-06, "loss": 0.443, "step": 6844 }, { "epoch": 2.19016638225256, "grad_norm": 0.4659494161605835, "learning_rate": 2.0594171390592294e-06, "loss": 0.3955, "step": 6845 }, { "epoch": 2.190486348122867, "grad_norm": 0.5452629327774048, "learning_rate": 2.057911567627218e-06, "loss": 0.482, "step": 6846 }, { "epoch": 2.190806313993174, "grad_norm": 0.4978730082511902, "learning_rate": 2.0564064041193053e-06, "loss": 0.3921, "step": 6847 }, { "epoch": 2.1911262798634814, "grad_norm": 0.5474013686180115, "learning_rate": 2.054901648744188e-06, "loss": 0.5015, "step": 6848 }, { "epoch": 2.1914462457337884, "grad_norm": 0.5499725937843323, "learning_rate": 2.0533973017104984e-06, "loss": 0.505, "step": 6849 }, { "epoch": 2.1917662116040955, "grad_norm": 0.503440797328949, "learning_rate": 2.0518933632268197e-06, "loss": 0.3985, "step": 6850 }, { "epoch": 2.192086177474403, "grad_norm": 0.5070424675941467, "learning_rate": 2.0503898335016724e-06, "loss": 0.4199, "step": 6851 }, { "epoch": 2.19240614334471, "grad_norm": 0.5492558479309082, "learning_rate": 2.0488867127435213e-06, "loss": 0.4508, "step": 6852 }, { "epoch": 2.192726109215017, "grad_norm": 0.523219108581543, "learning_rate": 2.04738400116078e-06, "loss": 0.4231, "step": 6853 }, { "epoch": 2.1930460750853245, "grad_norm": 0.511542558670044, "learning_rate": 2.045881698961797e-06, "loss": 0.4042, "step": 6854 }, { "epoch": 2.1933660409556315, "grad_norm": 0.5142143964767456, "learning_rate": 2.044379806354872e-06, "loss": 0.4438, "step": 6855 }, { "epoch": 2.1936860068259385, "grad_norm": 0.5721973180770874, "learning_rate": 2.0428783235482423e-06, "loss": 0.4605, "step": 6856 }, { "epoch": 2.1940059726962455, "grad_norm": 0.5211235284805298, "learning_rate": 2.0413772507500895e-06, "loss": 0.4822, "step": 6857 }, { "epoch": 2.194325938566553, "grad_norm": 0.48469439148902893, "learning_rate": 2.0398765881685417e-06, "loss": 0.4353, "step": 6858 }, { "epoch": 2.19464590443686, "grad_norm": 0.4858268201351166, "learning_rate": 2.0383763360116636e-06, "loss": 0.4404, "step": 6859 }, { "epoch": 2.194965870307167, "grad_norm": 0.46958059072494507, "learning_rate": 2.036876494487472e-06, "loss": 0.461, "step": 6860 }, { "epoch": 2.1952858361774745, "grad_norm": 0.47352731227874756, "learning_rate": 2.035377063803919e-06, "loss": 0.4045, "step": 6861 }, { "epoch": 2.1956058020477816, "grad_norm": 0.5191588997840881, "learning_rate": 2.033878044168901e-06, "loss": 0.4781, "step": 6862 }, { "epoch": 2.1959257679180886, "grad_norm": 0.5520008206367493, "learning_rate": 2.032379435790262e-06, "loss": 0.4526, "step": 6863 }, { "epoch": 2.196245733788396, "grad_norm": 0.4723661541938782, "learning_rate": 2.0308812388757828e-06, "loss": 0.4322, "step": 6864 }, { "epoch": 2.196565699658703, "grad_norm": 0.5281932353973389, "learning_rate": 2.029383453633193e-06, "loss": 0.4013, "step": 6865 }, { "epoch": 2.19688566552901, "grad_norm": 0.6034097075462341, "learning_rate": 2.0278860802701616e-06, "loss": 0.4635, "step": 6866 }, { "epoch": 2.1972056313993176, "grad_norm": 0.5373727679252625, "learning_rate": 2.0263891189942993e-06, "loss": 0.4505, "step": 6867 }, { "epoch": 2.1975255972696246, "grad_norm": 0.494740754365921, "learning_rate": 2.024892570013164e-06, "loss": 0.4075, "step": 6868 }, { "epoch": 2.1978455631399316, "grad_norm": 0.4898551404476166, "learning_rate": 2.0233964335342517e-06, "loss": 0.4347, "step": 6869 }, { "epoch": 2.198165529010239, "grad_norm": 0.5141493678092957, "learning_rate": 2.0219007097650063e-06, "loss": 0.4566, "step": 6870 }, { "epoch": 2.198485494880546, "grad_norm": 0.515340268611908, "learning_rate": 2.02040539891281e-06, "loss": 0.4114, "step": 6871 }, { "epoch": 2.198805460750853, "grad_norm": 0.5628042817115784, "learning_rate": 2.0189105011849887e-06, "loss": 0.503, "step": 6872 }, { "epoch": 2.1991254266211606, "grad_norm": 0.5190582275390625, "learning_rate": 2.017416016788814e-06, "loss": 0.4681, "step": 6873 }, { "epoch": 2.1994453924914676, "grad_norm": 0.5403550863265991, "learning_rate": 2.0159219459314948e-06, "loss": 0.4603, "step": 6874 }, { "epoch": 2.1997653583617747, "grad_norm": 0.5147091150283813, "learning_rate": 2.01442828882019e-06, "loss": 0.4396, "step": 6875 }, { "epoch": 2.2000853242320817, "grad_norm": 0.475997656583786, "learning_rate": 2.0129350456619946e-06, "loss": 0.4314, "step": 6876 }, { "epoch": 2.200405290102389, "grad_norm": 0.4674367308616638, "learning_rate": 2.011442216663947e-06, "loss": 0.4133, "step": 6877 }, { "epoch": 2.200725255972696, "grad_norm": 0.4793749153614044, "learning_rate": 2.0099498020330305e-06, "loss": 0.4767, "step": 6878 }, { "epoch": 2.201045221843003, "grad_norm": 0.5369402170181274, "learning_rate": 2.0084578019761738e-06, "loss": 0.4702, "step": 6879 }, { "epoch": 2.2013651877133107, "grad_norm": 0.48916861414909363, "learning_rate": 2.006966216700241e-06, "loss": 0.3998, "step": 6880 }, { "epoch": 2.2016851535836177, "grad_norm": 0.4863379895687103, "learning_rate": 2.005475046412044e-06, "loss": 0.4966, "step": 6881 }, { "epoch": 2.2020051194539247, "grad_norm": 0.44687557220458984, "learning_rate": 2.0039842913183316e-06, "loss": 0.4415, "step": 6882 }, { "epoch": 2.202325085324232, "grad_norm": 0.4204229414463043, "learning_rate": 2.002493951625802e-06, "loss": 0.4162, "step": 6883 }, { "epoch": 2.2026450511945392, "grad_norm": 0.5047687292098999, "learning_rate": 2.0010040275410934e-06, "loss": 0.4318, "step": 6884 }, { "epoch": 2.2029650170648463, "grad_norm": 0.5496372580528259, "learning_rate": 1.9995145192707845e-06, "loss": 0.4658, "step": 6885 }, { "epoch": 2.2032849829351537, "grad_norm": 0.5273929238319397, "learning_rate": 1.998025427021397e-06, "loss": 0.4643, "step": 6886 }, { "epoch": 2.2036049488054608, "grad_norm": 0.5278632640838623, "learning_rate": 1.9965367509993942e-06, "loss": 0.3911, "step": 6887 }, { "epoch": 2.203924914675768, "grad_norm": 0.47825103998184204, "learning_rate": 1.995048491411184e-06, "loss": 0.4559, "step": 6888 }, { "epoch": 2.2042448805460753, "grad_norm": 0.4605184495449066, "learning_rate": 1.993560648463117e-06, "loss": 0.4291, "step": 6889 }, { "epoch": 2.2045648464163823, "grad_norm": 0.5016769170761108, "learning_rate": 1.992073222361483e-06, "loss": 0.4723, "step": 6890 }, { "epoch": 2.2048848122866893, "grad_norm": 0.4509531855583191, "learning_rate": 1.990586213312516e-06, "loss": 0.4172, "step": 6891 }, { "epoch": 2.205204778156997, "grad_norm": 0.4440130591392517, "learning_rate": 1.9890996215223885e-06, "loss": 0.3638, "step": 6892 }, { "epoch": 2.205524744027304, "grad_norm": 0.5383840203285217, "learning_rate": 1.98761344719722e-06, "loss": 0.4994, "step": 6893 }, { "epoch": 2.205844709897611, "grad_norm": 0.4807364046573639, "learning_rate": 1.986127690543074e-06, "loss": 0.4016, "step": 6894 }, { "epoch": 2.2061646757679183, "grad_norm": 0.5220727920532227, "learning_rate": 1.984642351765949e-06, "loss": 0.4297, "step": 6895 }, { "epoch": 2.2064846416382253, "grad_norm": 0.507565975189209, "learning_rate": 1.98315743107179e-06, "loss": 0.4154, "step": 6896 }, { "epoch": 2.2068046075085324, "grad_norm": 0.48764318227767944, "learning_rate": 1.98167292866648e-06, "loss": 0.4298, "step": 6897 }, { "epoch": 2.2071245733788394, "grad_norm": 0.495840460062027, "learning_rate": 1.9801888447558497e-06, "loss": 0.4432, "step": 6898 }, { "epoch": 2.207444539249147, "grad_norm": 0.5050677061080933, "learning_rate": 1.9787051795456714e-06, "loss": 0.4162, "step": 6899 }, { "epoch": 2.207764505119454, "grad_norm": 0.48621681332588196, "learning_rate": 1.9772219332416544e-06, "loss": 0.4683, "step": 6900 }, { "epoch": 2.208084470989761, "grad_norm": 0.45279693603515625, "learning_rate": 1.9757391060494525e-06, "loss": 0.3982, "step": 6901 }, { "epoch": 2.2084044368600684, "grad_norm": 0.5346647500991821, "learning_rate": 1.9742566981746596e-06, "loss": 0.4741, "step": 6902 }, { "epoch": 2.2087244027303754, "grad_norm": 0.5127875208854675, "learning_rate": 1.9727747098228167e-06, "loss": 0.4462, "step": 6903 }, { "epoch": 2.2090443686006824, "grad_norm": 0.49378058314323425, "learning_rate": 1.971293141199403e-06, "loss": 0.437, "step": 6904 }, { "epoch": 2.20936433447099, "grad_norm": 0.49449095129966736, "learning_rate": 1.9698119925098398e-06, "loss": 0.4527, "step": 6905 }, { "epoch": 2.209684300341297, "grad_norm": 0.49412205815315247, "learning_rate": 1.9683312639594883e-06, "loss": 0.4799, "step": 6906 }, { "epoch": 2.210004266211604, "grad_norm": 0.4692303240299225, "learning_rate": 1.966850955753653e-06, "loss": 0.3993, "step": 6907 }, { "epoch": 2.2103242320819114, "grad_norm": 0.48455101251602173, "learning_rate": 1.965371068097581e-06, "loss": 0.4132, "step": 6908 }, { "epoch": 2.2106441979522184, "grad_norm": 0.531497061252594, "learning_rate": 1.9638916011964637e-06, "loss": 0.4759, "step": 6909 }, { "epoch": 2.2109641638225255, "grad_norm": 0.5194461345672607, "learning_rate": 1.9624125552554287e-06, "loss": 0.4083, "step": 6910 }, { "epoch": 2.211284129692833, "grad_norm": 0.49568623304367065, "learning_rate": 1.960933930479545e-06, "loss": 0.4299, "step": 6911 }, { "epoch": 2.21160409556314, "grad_norm": 0.4672048091888428, "learning_rate": 1.9594557270738275e-06, "loss": 0.4346, "step": 6912 }, { "epoch": 2.211924061433447, "grad_norm": 0.509306013584137, "learning_rate": 1.9579779452432336e-06, "loss": 0.474, "step": 6913 }, { "epoch": 2.2122440273037545, "grad_norm": 0.5298650860786438, "learning_rate": 1.9565005851926574e-06, "loss": 0.4254, "step": 6914 }, { "epoch": 2.2125639931740615, "grad_norm": 0.5057072639465332, "learning_rate": 1.9550236471269358e-06, "loss": 0.4584, "step": 6915 }, { "epoch": 2.2128839590443685, "grad_norm": 0.5267782211303711, "learning_rate": 1.953547131250847e-06, "loss": 0.4478, "step": 6916 }, { "epoch": 2.2132039249146755, "grad_norm": 0.5834232568740845, "learning_rate": 1.9520710377691137e-06, "loss": 0.4841, "step": 6917 }, { "epoch": 2.213523890784983, "grad_norm": 0.5276156663894653, "learning_rate": 1.9505953668863996e-06, "loss": 0.457, "step": 6918 }, { "epoch": 2.21384385665529, "grad_norm": 0.477408766746521, "learning_rate": 1.9491201188073067e-06, "loss": 0.4318, "step": 6919 }, { "epoch": 2.214163822525597, "grad_norm": 0.5019200444221497, "learning_rate": 1.947645293736379e-06, "loss": 0.4364, "step": 6920 }, { "epoch": 2.2144837883959045, "grad_norm": 0.49255263805389404, "learning_rate": 1.9461708918781018e-06, "loss": 0.4518, "step": 6921 }, { "epoch": 2.2148037542662116, "grad_norm": 0.5077428817749023, "learning_rate": 1.944696913436905e-06, "loss": 0.4301, "step": 6922 }, { "epoch": 2.2151237201365186, "grad_norm": 0.5375750660896301, "learning_rate": 1.9432233586171576e-06, "loss": 0.4121, "step": 6923 }, { "epoch": 2.215443686006826, "grad_norm": 0.5089945197105408, "learning_rate": 1.941750227623169e-06, "loss": 0.4434, "step": 6924 }, { "epoch": 2.215763651877133, "grad_norm": 0.43878012895584106, "learning_rate": 1.940277520659191e-06, "loss": 0.4257, "step": 6925 }, { "epoch": 2.21608361774744, "grad_norm": 0.4928267300128937, "learning_rate": 1.9388052379294135e-06, "loss": 0.4155, "step": 6926 }, { "epoch": 2.2164035836177476, "grad_norm": 0.5031419992446899, "learning_rate": 1.937333379637973e-06, "loss": 0.459, "step": 6927 }, { "epoch": 2.2167235494880546, "grad_norm": 0.4916204810142517, "learning_rate": 1.935861945988945e-06, "loss": 0.3791, "step": 6928 }, { "epoch": 2.2170435153583616, "grad_norm": 0.5682798027992249, "learning_rate": 1.9343909371863452e-06, "loss": 0.427, "step": 6929 }, { "epoch": 2.217363481228669, "grad_norm": 0.5600176453590393, "learning_rate": 1.932920353434129e-06, "loss": 0.417, "step": 6930 }, { "epoch": 2.217683447098976, "grad_norm": 0.5578386187553406, "learning_rate": 1.9314501949361946e-06, "loss": 0.4224, "step": 6931 }, { "epoch": 2.218003412969283, "grad_norm": 0.5250586867332458, "learning_rate": 1.929980461896382e-06, "loss": 0.4548, "step": 6932 }, { "epoch": 2.2183233788395906, "grad_norm": 0.5524583458900452, "learning_rate": 1.928511154518473e-06, "loss": 0.439, "step": 6933 }, { "epoch": 2.2186433447098977, "grad_norm": 0.49936771392822266, "learning_rate": 1.9270422730061883e-06, "loss": 0.4048, "step": 6934 }, { "epoch": 2.2189633105802047, "grad_norm": 0.5375549793243408, "learning_rate": 1.925573817563189e-06, "loss": 0.4703, "step": 6935 }, { "epoch": 2.219283276450512, "grad_norm": 0.560430645942688, "learning_rate": 1.924105788393077e-06, "loss": 0.4663, "step": 6936 }, { "epoch": 2.219603242320819, "grad_norm": 0.5091943740844727, "learning_rate": 1.922638185699397e-06, "loss": 0.4395, "step": 6937 }, { "epoch": 2.219923208191126, "grad_norm": 0.5026044845581055, "learning_rate": 1.9211710096856373e-06, "loss": 0.4632, "step": 6938 }, { "epoch": 2.2202431740614337, "grad_norm": 0.5348517894744873, "learning_rate": 1.9197042605552206e-06, "loss": 0.4696, "step": 6939 }, { "epoch": 2.2205631399317407, "grad_norm": 0.47700411081314087, "learning_rate": 1.918237938511514e-06, "loss": 0.4222, "step": 6940 }, { "epoch": 2.2208831058020477, "grad_norm": 0.5197616219520569, "learning_rate": 1.916772043757823e-06, "loss": 0.4943, "step": 6941 }, { "epoch": 2.2212030716723548, "grad_norm": 0.48118993639945984, "learning_rate": 1.915306576497397e-06, "loss": 0.3919, "step": 6942 }, { "epoch": 2.2215230375426622, "grad_norm": 0.5554553866386414, "learning_rate": 1.9138415369334275e-06, "loss": 0.4865, "step": 6943 }, { "epoch": 2.2218430034129693, "grad_norm": 0.4820300340652466, "learning_rate": 1.912376925269041e-06, "loss": 0.4447, "step": 6944 }, { "epoch": 2.2221629692832763, "grad_norm": 0.5619463324546814, "learning_rate": 1.9109127417073087e-06, "loss": 0.486, "step": 6945 }, { "epoch": 2.2224829351535837, "grad_norm": 0.5620408654212952, "learning_rate": 1.9094489864512393e-06, "loss": 0.4141, "step": 6946 }, { "epoch": 2.2228029010238908, "grad_norm": 0.5426426529884338, "learning_rate": 1.907985659703786e-06, "loss": 0.448, "step": 6947 }, { "epoch": 2.223122866894198, "grad_norm": 0.48584941029548645, "learning_rate": 1.9065227616678422e-06, "loss": 0.4807, "step": 6948 }, { "epoch": 2.2234428327645053, "grad_norm": 0.4817311465740204, "learning_rate": 1.9050602925462392e-06, "loss": 0.4679, "step": 6949 }, { "epoch": 2.2237627986348123, "grad_norm": 0.4578678607940674, "learning_rate": 1.90359825254175e-06, "loss": 0.4147, "step": 6950 }, { "epoch": 2.2240827645051193, "grad_norm": 0.48312684893608093, "learning_rate": 1.9021366418570864e-06, "loss": 0.402, "step": 6951 }, { "epoch": 2.224402730375427, "grad_norm": 0.524979293346405, "learning_rate": 1.9006754606949041e-06, "loss": 0.5166, "step": 6952 }, { "epoch": 2.224722696245734, "grad_norm": 0.5091363787651062, "learning_rate": 1.8992147092577994e-06, "loss": 0.3952, "step": 6953 }, { "epoch": 2.225042662116041, "grad_norm": 0.5416716933250427, "learning_rate": 1.8977543877483056e-06, "loss": 0.4783, "step": 6954 }, { "epoch": 2.2253626279863483, "grad_norm": 0.5534230470657349, "learning_rate": 1.8962944963688984e-06, "loss": 0.4238, "step": 6955 }, { "epoch": 2.2256825938566553, "grad_norm": 0.4583647549152374, "learning_rate": 1.8948350353219913e-06, "loss": 0.42, "step": 6956 }, { "epoch": 2.2260025597269624, "grad_norm": 0.46035411953926086, "learning_rate": 1.893376004809942e-06, "loss": 0.439, "step": 6957 }, { "epoch": 2.2263225255972694, "grad_norm": 0.49934062361717224, "learning_rate": 1.8919174050350487e-06, "loss": 0.4576, "step": 6958 }, { "epoch": 2.226642491467577, "grad_norm": 0.5063640475273132, "learning_rate": 1.8904592361995467e-06, "loss": 0.3689, "step": 6959 }, { "epoch": 2.226962457337884, "grad_norm": 0.5481410622596741, "learning_rate": 1.8890014985056115e-06, "loss": 0.4838, "step": 6960 }, { "epoch": 2.227282423208191, "grad_norm": 0.4754042327404022, "learning_rate": 1.8875441921553606e-06, "loss": 0.4294, "step": 6961 }, { "epoch": 2.2276023890784984, "grad_norm": 0.48783764243125916, "learning_rate": 1.8860873173508509e-06, "loss": 0.4348, "step": 6962 }, { "epoch": 2.2279223549488054, "grad_norm": 0.49847304821014404, "learning_rate": 1.8846308742940834e-06, "loss": 0.4538, "step": 6963 }, { "epoch": 2.2282423208191124, "grad_norm": 0.5597853660583496, "learning_rate": 1.8831748631869923e-06, "loss": 0.4608, "step": 6964 }, { "epoch": 2.22856228668942, "grad_norm": 0.5159894227981567, "learning_rate": 1.8817192842314564e-06, "loss": 0.4063, "step": 6965 }, { "epoch": 2.228882252559727, "grad_norm": 0.482262521982193, "learning_rate": 1.8802641376292913e-06, "loss": 0.3973, "step": 6966 }, { "epoch": 2.229202218430034, "grad_norm": 0.5018547773361206, "learning_rate": 1.8788094235822563e-06, "loss": 0.4351, "step": 6967 }, { "epoch": 2.2295221843003414, "grad_norm": 0.48930323123931885, "learning_rate": 1.8773551422920522e-06, "loss": 0.4676, "step": 6968 }, { "epoch": 2.2298421501706485, "grad_norm": 0.4622410237789154, "learning_rate": 1.8759012939603133e-06, "loss": 0.3874, "step": 6969 }, { "epoch": 2.2301621160409555, "grad_norm": 0.4761062562465668, "learning_rate": 1.8744478787886188e-06, "loss": 0.4573, "step": 6970 }, { "epoch": 2.230482081911263, "grad_norm": 0.4705923795700073, "learning_rate": 1.8729948969784844e-06, "loss": 0.413, "step": 6971 }, { "epoch": 2.23080204778157, "grad_norm": 0.5186229348182678, "learning_rate": 1.8715423487313694e-06, "loss": 0.4403, "step": 6972 }, { "epoch": 2.231122013651877, "grad_norm": 0.5065740346908569, "learning_rate": 1.8700902342486732e-06, "loss": 0.4422, "step": 6973 }, { "epoch": 2.2314419795221845, "grad_norm": 0.4577891528606415, "learning_rate": 1.8686385537317314e-06, "loss": 0.3759, "step": 6974 }, { "epoch": 2.2317619453924915, "grad_norm": 0.5119481682777405, "learning_rate": 1.8671873073818208e-06, "loss": 0.4536, "step": 6975 }, { "epoch": 2.2320819112627985, "grad_norm": 0.5477125644683838, "learning_rate": 1.8657364954001578e-06, "loss": 0.4949, "step": 6976 }, { "epoch": 2.232401877133106, "grad_norm": 0.48550277948379517, "learning_rate": 1.8642861179878996e-06, "loss": 0.4123, "step": 6977 }, { "epoch": 2.232721843003413, "grad_norm": 0.5441885590553284, "learning_rate": 1.8628361753461455e-06, "loss": 0.4188, "step": 6978 }, { "epoch": 2.23304180887372, "grad_norm": 0.5613593459129333, "learning_rate": 1.86138666767593e-06, "loss": 0.5034, "step": 6979 }, { "epoch": 2.2333617747440275, "grad_norm": 0.4752897024154663, "learning_rate": 1.859937595178229e-06, "loss": 0.4031, "step": 6980 }, { "epoch": 2.2336817406143346, "grad_norm": 0.5087677836418152, "learning_rate": 1.858488958053956e-06, "loss": 0.4517, "step": 6981 }, { "epoch": 2.2340017064846416, "grad_norm": 0.4827876091003418, "learning_rate": 1.8570407565039683e-06, "loss": 0.4157, "step": 6982 }, { "epoch": 2.2343216723549486, "grad_norm": 0.5602332353591919, "learning_rate": 1.8555929907290627e-06, "loss": 0.53, "step": 6983 }, { "epoch": 2.234641638225256, "grad_norm": 0.5727624893188477, "learning_rate": 1.8541456609299712e-06, "loss": 0.4362, "step": 6984 }, { "epoch": 2.234961604095563, "grad_norm": 0.5025842189788818, "learning_rate": 1.852698767307367e-06, "loss": 0.4298, "step": 6985 }, { "epoch": 2.23528156996587, "grad_norm": 0.5044243335723877, "learning_rate": 1.8512523100618673e-06, "loss": 0.4565, "step": 6986 }, { "epoch": 2.2356015358361776, "grad_norm": 0.4869266152381897, "learning_rate": 1.8498062893940206e-06, "loss": 0.4383, "step": 6987 }, { "epoch": 2.2359215017064846, "grad_norm": 0.44381487369537354, "learning_rate": 1.8483607055043234e-06, "loss": 0.4324, "step": 6988 }, { "epoch": 2.2362414675767917, "grad_norm": 0.4683745205402374, "learning_rate": 1.8469155585932057e-06, "loss": 0.4171, "step": 6989 }, { "epoch": 2.236561433447099, "grad_norm": 0.5158386826515198, "learning_rate": 1.8454708488610372e-06, "loss": 0.4609, "step": 6990 }, { "epoch": 2.236881399317406, "grad_norm": 0.48240604996681213, "learning_rate": 1.8440265765081328e-06, "loss": 0.4259, "step": 6991 }, { "epoch": 2.237201365187713, "grad_norm": 0.4903813898563385, "learning_rate": 1.8425827417347385e-06, "loss": 0.4293, "step": 6992 }, { "epoch": 2.2375213310580206, "grad_norm": 0.5533493757247925, "learning_rate": 1.841139344741047e-06, "loss": 0.4596, "step": 6993 }, { "epoch": 2.2378412969283277, "grad_norm": 0.5921840071678162, "learning_rate": 1.8396963857271855e-06, "loss": 0.4857, "step": 6994 }, { "epoch": 2.2381612627986347, "grad_norm": 0.4598381221294403, "learning_rate": 1.8382538648932201e-06, "loss": 0.405, "step": 6995 }, { "epoch": 2.238481228668942, "grad_norm": 0.4713385999202728, "learning_rate": 1.8368117824391623e-06, "loss": 0.4308, "step": 6996 }, { "epoch": 2.238801194539249, "grad_norm": 0.47838225960731506, "learning_rate": 1.8353701385649548e-06, "loss": 0.3965, "step": 6997 }, { "epoch": 2.239121160409556, "grad_norm": 0.46495163440704346, "learning_rate": 1.8339289334704857e-06, "loss": 0.4581, "step": 6998 }, { "epoch": 2.2394411262798632, "grad_norm": 0.5104029178619385, "learning_rate": 1.832488167355579e-06, "loss": 0.4513, "step": 6999 }, { "epoch": 2.2397610921501707, "grad_norm": 0.4873708188533783, "learning_rate": 1.831047840419997e-06, "loss": 0.4124, "step": 7000 }, { "epoch": 2.2400810580204777, "grad_norm": 0.4746980369091034, "learning_rate": 1.8296079528634458e-06, "loss": 0.4306, "step": 7001 }, { "epoch": 2.2404010238907848, "grad_norm": 0.4791220426559448, "learning_rate": 1.8281685048855647e-06, "loss": 0.4564, "step": 7002 }, { "epoch": 2.2407209897610922, "grad_norm": 0.4969911575317383, "learning_rate": 1.8267294966859373e-06, "loss": 0.4386, "step": 7003 }, { "epoch": 2.2410409556313993, "grad_norm": 0.48363468050956726, "learning_rate": 1.825290928464083e-06, "loss": 0.4272, "step": 7004 }, { "epoch": 2.2413609215017063, "grad_norm": 0.5502254962921143, "learning_rate": 1.8238528004194589e-06, "loss": 0.5083, "step": 7005 }, { "epoch": 2.2416808873720138, "grad_norm": 0.5414071083068848, "learning_rate": 1.8224151127514672e-06, "loss": 0.481, "step": 7006 }, { "epoch": 2.242000853242321, "grad_norm": 0.5145581960678101, "learning_rate": 1.8209778656594406e-06, "loss": 0.4167, "step": 7007 }, { "epoch": 2.242320819112628, "grad_norm": 0.538152813911438, "learning_rate": 1.8195410593426594e-06, "loss": 0.4199, "step": 7008 }, { "epoch": 2.2426407849829353, "grad_norm": 0.5265048742294312, "learning_rate": 1.8181046940003366e-06, "loss": 0.4057, "step": 7009 }, { "epoch": 2.2429607508532423, "grad_norm": 0.49859923124313354, "learning_rate": 1.8166687698316237e-06, "loss": 0.4462, "step": 7010 }, { "epoch": 2.2432807167235493, "grad_norm": 0.4799681305885315, "learning_rate": 1.8152332870356177e-06, "loss": 0.3907, "step": 7011 }, { "epoch": 2.243600682593857, "grad_norm": 0.5121408700942993, "learning_rate": 1.8137982458113457e-06, "loss": 0.4462, "step": 7012 }, { "epoch": 2.243920648464164, "grad_norm": 0.5352764129638672, "learning_rate": 1.8123636463577821e-06, "loss": 0.4413, "step": 7013 }, { "epoch": 2.244240614334471, "grad_norm": 0.49394893646240234, "learning_rate": 1.8109294888738338e-06, "loss": 0.4361, "step": 7014 }, { "epoch": 2.2445605802047783, "grad_norm": 0.4686887562274933, "learning_rate": 1.8094957735583463e-06, "loss": 0.4071, "step": 7015 }, { "epoch": 2.2448805460750854, "grad_norm": 0.5014979243278503, "learning_rate": 1.8080625006101098e-06, "loss": 0.4334, "step": 7016 }, { "epoch": 2.2452005119453924, "grad_norm": 0.5246603488922119, "learning_rate": 1.8066296702278464e-06, "loss": 0.4769, "step": 7017 }, { "epoch": 2.2455204778157, "grad_norm": 0.5133163332939148, "learning_rate": 1.805197282610222e-06, "loss": 0.3907, "step": 7018 }, { "epoch": 2.245840443686007, "grad_norm": 0.48508474230766296, "learning_rate": 1.8037653379558368e-06, "loss": 0.4269, "step": 7019 }, { "epoch": 2.246160409556314, "grad_norm": 0.48519837856292725, "learning_rate": 1.8023338364632336e-06, "loss": 0.4751, "step": 7020 }, { "epoch": 2.2464803754266214, "grad_norm": 0.4832985997200012, "learning_rate": 1.8009027783308914e-06, "loss": 0.4336, "step": 7021 }, { "epoch": 2.2468003412969284, "grad_norm": 0.5109445452690125, "learning_rate": 1.799472163757226e-06, "loss": 0.4196, "step": 7022 }, { "epoch": 2.2471203071672354, "grad_norm": 0.5440354347229004, "learning_rate": 1.7980419929405967e-06, "loss": 0.4751, "step": 7023 }, { "epoch": 2.2474402730375425, "grad_norm": 0.7519711852073669, "learning_rate": 1.7966122660792955e-06, "loss": 0.4177, "step": 7024 }, { "epoch": 2.24776023890785, "grad_norm": 0.5572145581245422, "learning_rate": 1.795182983371559e-06, "loss": 0.4815, "step": 7025 }, { "epoch": 2.248080204778157, "grad_norm": 0.515664279460907, "learning_rate": 1.7937541450155578e-06, "loss": 0.4056, "step": 7026 }, { "epoch": 2.248400170648464, "grad_norm": 0.5235795378684998, "learning_rate": 1.7923257512093994e-06, "loss": 0.4593, "step": 7027 }, { "epoch": 2.2487201365187715, "grad_norm": 0.4442862570285797, "learning_rate": 1.7908978021511365e-06, "loss": 0.4151, "step": 7028 }, { "epoch": 2.2490401023890785, "grad_norm": 0.5126864910125732, "learning_rate": 1.7894702980387524e-06, "loss": 0.4579, "step": 7029 }, { "epoch": 2.2493600682593855, "grad_norm": 0.4956403970718384, "learning_rate": 1.788043239070175e-06, "loss": 0.4212, "step": 7030 }, { "epoch": 2.249680034129693, "grad_norm": 0.5071899890899658, "learning_rate": 1.786616625443267e-06, "loss": 0.5038, "step": 7031 }, { "epoch": 2.25, "grad_norm": 0.4895451068878174, "learning_rate": 1.7851904573558276e-06, "loss": 0.4153, "step": 7032 }, { "epoch": 2.250319965870307, "grad_norm": 0.5462383031845093, "learning_rate": 1.7837647350056008e-06, "loss": 0.4266, "step": 7033 }, { "epoch": 2.2506399317406145, "grad_norm": 0.5014021992683411, "learning_rate": 1.7823394585902615e-06, "loss": 0.4568, "step": 7034 }, { "epoch": 2.2509598976109215, "grad_norm": 0.47650790214538574, "learning_rate": 1.780914628307428e-06, "loss": 0.4126, "step": 7035 }, { "epoch": 2.2512798634812285, "grad_norm": 0.5031761527061462, "learning_rate": 1.7794902443546546e-06, "loss": 0.4157, "step": 7036 }, { "epoch": 2.251599829351536, "grad_norm": 0.5211197137832642, "learning_rate": 1.7780663069294312e-06, "loss": 0.4403, "step": 7037 }, { "epoch": 2.251919795221843, "grad_norm": 0.49045392870903015, "learning_rate": 1.776642816229192e-06, "loss": 0.4109, "step": 7038 }, { "epoch": 2.25223976109215, "grad_norm": 0.5195350646972656, "learning_rate": 1.7752197724513027e-06, "loss": 0.4583, "step": 7039 }, { "epoch": 2.252559726962457, "grad_norm": 0.468694269657135, "learning_rate": 1.7737971757930732e-06, "loss": 0.4034, "step": 7040 }, { "epoch": 2.2528796928327646, "grad_norm": 0.5023847818374634, "learning_rate": 1.7723750264517464e-06, "loss": 0.4527, "step": 7041 }, { "epoch": 2.2531996587030716, "grad_norm": 0.4959229528903961, "learning_rate": 1.7709533246245032e-06, "loss": 0.4218, "step": 7042 }, { "epoch": 2.2535196245733786, "grad_norm": 0.5296538472175598, "learning_rate": 1.7695320705084678e-06, "loss": 0.523, "step": 7043 }, { "epoch": 2.253839590443686, "grad_norm": 0.4721936583518982, "learning_rate": 1.768111264300696e-06, "loss": 0.4433, "step": 7044 }, { "epoch": 2.254159556313993, "grad_norm": 0.5049819350242615, "learning_rate": 1.7666909061981869e-06, "loss": 0.4704, "step": 7045 }, { "epoch": 2.2544795221843, "grad_norm": 0.5241411924362183, "learning_rate": 1.765270996397873e-06, "loss": 0.4218, "step": 7046 }, { "epoch": 2.2547994880546076, "grad_norm": 0.5348628163337708, "learning_rate": 1.7638515350966257e-06, "loss": 0.4897, "step": 7047 }, { "epoch": 2.2551194539249146, "grad_norm": 0.549798846244812, "learning_rate": 1.762432522491258e-06, "loss": 0.432, "step": 7048 }, { "epoch": 2.2554394197952217, "grad_norm": 0.5057252049446106, "learning_rate": 1.761013958778514e-06, "loss": 0.4452, "step": 7049 }, { "epoch": 2.255759385665529, "grad_norm": 0.47426411509513855, "learning_rate": 1.759595844155083e-06, "loss": 0.4149, "step": 7050 }, { "epoch": 2.256079351535836, "grad_norm": 0.45348483324050903, "learning_rate": 1.758178178817585e-06, "loss": 0.4352, "step": 7051 }, { "epoch": 2.256399317406143, "grad_norm": 0.4907292127609253, "learning_rate": 1.7567609629625836e-06, "loss": 0.4177, "step": 7052 }, { "epoch": 2.2567192832764507, "grad_norm": 0.5103530287742615, "learning_rate": 1.755344196786577e-06, "loss": 0.4333, "step": 7053 }, { "epoch": 2.2570392491467577, "grad_norm": 0.5013543367385864, "learning_rate": 1.7539278804859993e-06, "loss": 0.4393, "step": 7054 }, { "epoch": 2.2573592150170647, "grad_norm": 0.4777149260044098, "learning_rate": 1.752512014257227e-06, "loss": 0.4428, "step": 7055 }, { "epoch": 2.257679180887372, "grad_norm": 0.4968894124031067, "learning_rate": 1.75109659829657e-06, "loss": 0.4251, "step": 7056 }, { "epoch": 2.257999146757679, "grad_norm": 0.46732962131500244, "learning_rate": 1.7496816328002797e-06, "loss": 0.4255, "step": 7057 }, { "epoch": 2.2583191126279862, "grad_norm": 0.5094853639602661, "learning_rate": 1.7482671179645388e-06, "loss": 0.4488, "step": 7058 }, { "epoch": 2.2586390784982937, "grad_norm": 0.5172582864761353, "learning_rate": 1.746853053985476e-06, "loss": 0.4294, "step": 7059 }, { "epoch": 2.2589590443686007, "grad_norm": 0.534986138343811, "learning_rate": 1.7454394410591514e-06, "loss": 0.4278, "step": 7060 }, { "epoch": 2.2592790102389078, "grad_norm": 0.46675431728363037, "learning_rate": 1.7440262793815615e-06, "loss": 0.4329, "step": 7061 }, { "epoch": 2.2595989761092152, "grad_norm": 0.4917410612106323, "learning_rate": 1.7426135691486467e-06, "loss": 0.4416, "step": 7062 }, { "epoch": 2.2599189419795223, "grad_norm": 0.5392761826515198, "learning_rate": 1.7412013105562775e-06, "loss": 0.3825, "step": 7063 }, { "epoch": 2.2602389078498293, "grad_norm": 0.548638641834259, "learning_rate": 1.7397895038002689e-06, "loss": 0.4479, "step": 7064 }, { "epoch": 2.2605588737201368, "grad_norm": 0.5172118544578552, "learning_rate": 1.7383781490763678e-06, "loss": 0.4856, "step": 7065 }, { "epoch": 2.260878839590444, "grad_norm": 0.49654272198677063, "learning_rate": 1.7369672465802585e-06, "loss": 0.4263, "step": 7066 }, { "epoch": 2.261198805460751, "grad_norm": 0.4977187514305115, "learning_rate": 1.7355567965075675e-06, "loss": 0.4491, "step": 7067 }, { "epoch": 2.261518771331058, "grad_norm": 0.542453944683075, "learning_rate": 1.734146799053852e-06, "loss": 0.4566, "step": 7068 }, { "epoch": 2.2618387372013653, "grad_norm": 0.4947512745857239, "learning_rate": 1.7327372544146137e-06, "loss": 0.4352, "step": 7069 }, { "epoch": 2.2621587030716723, "grad_norm": 0.4490604102611542, "learning_rate": 1.7313281627852857e-06, "loss": 0.4436, "step": 7070 }, { "epoch": 2.2624786689419794, "grad_norm": 0.5164769291877747, "learning_rate": 1.7299195243612382e-06, "loss": 0.4845, "step": 7071 }, { "epoch": 2.262798634812287, "grad_norm": 0.5048587322235107, "learning_rate": 1.7285113393377845e-06, "loss": 0.4152, "step": 7072 }, { "epoch": 2.263118600682594, "grad_norm": 0.5196941494941711, "learning_rate": 1.7271036079101678e-06, "loss": 0.4732, "step": 7073 }, { "epoch": 2.263438566552901, "grad_norm": 0.5368337035179138, "learning_rate": 1.7256963302735752e-06, "loss": 0.483, "step": 7074 }, { "epoch": 2.2637585324232083, "grad_norm": 0.4829712510108948, "learning_rate": 1.7242895066231247e-06, "loss": 0.432, "step": 7075 }, { "epoch": 2.2640784982935154, "grad_norm": 0.5370932817459106, "learning_rate": 1.722883137153874e-06, "loss": 0.4103, "step": 7076 }, { "epoch": 2.2643984641638224, "grad_norm": 0.5939807295799255, "learning_rate": 1.72147722206082e-06, "loss": 0.4691, "step": 7077 }, { "epoch": 2.26471843003413, "grad_norm": 0.45269444584846497, "learning_rate": 1.7200717615388917e-06, "loss": 0.4499, "step": 7078 }, { "epoch": 2.265038395904437, "grad_norm": 0.43377363681793213, "learning_rate": 1.7186667557829612e-06, "loss": 0.4164, "step": 7079 }, { "epoch": 2.265358361774744, "grad_norm": 0.49725332856178284, "learning_rate": 1.7172622049878329e-06, "loss": 0.4945, "step": 7080 }, { "epoch": 2.265678327645051, "grad_norm": 0.5128596425056458, "learning_rate": 1.7158581093482467e-06, "loss": 0.4484, "step": 7081 }, { "epoch": 2.2659982935153584, "grad_norm": 0.47245702147483826, "learning_rate": 1.7144544690588872e-06, "loss": 0.4514, "step": 7082 }, { "epoch": 2.2663182593856654, "grad_norm": 0.427586168050766, "learning_rate": 1.713051284314366e-06, "loss": 0.3753, "step": 7083 }, { "epoch": 2.2666382252559725, "grad_norm": 0.5714744329452515, "learning_rate": 1.71164855530924e-06, "loss": 0.5082, "step": 7084 }, { "epoch": 2.26695819112628, "grad_norm": 0.5184192061424255, "learning_rate": 1.7102462822379977e-06, "loss": 0.424, "step": 7085 }, { "epoch": 2.267278156996587, "grad_norm": 0.5154229402542114, "learning_rate": 1.7088444652950647e-06, "loss": 0.4218, "step": 7086 }, { "epoch": 2.267598122866894, "grad_norm": 0.4990445077419281, "learning_rate": 1.7074431046748075e-06, "loss": 0.4072, "step": 7087 }, { "epoch": 2.2679180887372015, "grad_norm": 0.48973727226257324, "learning_rate": 1.7060422005715232e-06, "loss": 0.4353, "step": 7088 }, { "epoch": 2.2682380546075085, "grad_norm": 0.479113906621933, "learning_rate": 1.7046417531794523e-06, "loss": 0.4334, "step": 7089 }, { "epoch": 2.2685580204778155, "grad_norm": 0.5312931537628174, "learning_rate": 1.7032417626927666e-06, "loss": 0.4662, "step": 7090 }, { "epoch": 2.268877986348123, "grad_norm": 0.4754203259944916, "learning_rate": 1.7018422293055753e-06, "loss": 0.4176, "step": 7091 }, { "epoch": 2.26919795221843, "grad_norm": 0.5042561888694763, "learning_rate": 1.7004431532119264e-06, "loss": 0.4286, "step": 7092 }, { "epoch": 2.269517918088737, "grad_norm": 0.4862496554851532, "learning_rate": 1.6990445346058055e-06, "loss": 0.3956, "step": 7093 }, { "epoch": 2.2698378839590445, "grad_norm": 0.5385366678237915, "learning_rate": 1.6976463736811306e-06, "loss": 0.4688, "step": 7094 }, { "epoch": 2.2701578498293515, "grad_norm": 0.5797232389450073, "learning_rate": 1.6962486706317594e-06, "loss": 0.4677, "step": 7095 }, { "epoch": 2.2704778156996586, "grad_norm": 0.5136889219284058, "learning_rate": 1.6948514256514826e-06, "loss": 0.413, "step": 7096 }, { "epoch": 2.270797781569966, "grad_norm": 0.5192843675613403, "learning_rate": 1.6934546389340317e-06, "loss": 0.5074, "step": 7097 }, { "epoch": 2.271117747440273, "grad_norm": 0.4660189747810364, "learning_rate": 1.6920583106730749e-06, "loss": 0.3723, "step": 7098 }, { "epoch": 2.27143771331058, "grad_norm": 0.5324483513832092, "learning_rate": 1.6906624410622124e-06, "loss": 0.4553, "step": 7099 }, { "epoch": 2.2717576791808876, "grad_norm": 0.4562549293041229, "learning_rate": 1.6892670302949842e-06, "loss": 0.4392, "step": 7100 }, { "epoch": 2.2720776450511946, "grad_norm": 0.5018472671508789, "learning_rate": 1.6878720785648633e-06, "loss": 0.4063, "step": 7101 }, { "epoch": 2.2723976109215016, "grad_norm": 0.4813690483570099, "learning_rate": 1.6864775860652638e-06, "loss": 0.4224, "step": 7102 }, { "epoch": 2.272717576791809, "grad_norm": 0.5093215703964233, "learning_rate": 1.6850835529895344e-06, "loss": 0.466, "step": 7103 }, { "epoch": 2.273037542662116, "grad_norm": 0.5317978262901306, "learning_rate": 1.683689979530958e-06, "loss": 0.443, "step": 7104 }, { "epoch": 2.273357508532423, "grad_norm": 0.5211302638053894, "learning_rate": 1.682296865882756e-06, "loss": 0.4428, "step": 7105 }, { "epoch": 2.2736774744027306, "grad_norm": 0.45339956879615784, "learning_rate": 1.680904212238083e-06, "loss": 0.4145, "step": 7106 }, { "epoch": 2.2739974402730376, "grad_norm": 0.5466711521148682, "learning_rate": 1.6795120187900344e-06, "loss": 0.4522, "step": 7107 }, { "epoch": 2.2743174061433447, "grad_norm": 0.5057246088981628, "learning_rate": 1.6781202857316402e-06, "loss": 0.463, "step": 7108 }, { "epoch": 2.274637372013652, "grad_norm": 0.47965845465660095, "learning_rate": 1.676729013255865e-06, "loss": 0.4392, "step": 7109 }, { "epoch": 2.274957337883959, "grad_norm": 0.4975675940513611, "learning_rate": 1.6753382015556096e-06, "loss": 0.4446, "step": 7110 }, { "epoch": 2.275277303754266, "grad_norm": 0.50555020570755, "learning_rate": 1.673947850823711e-06, "loss": 0.4527, "step": 7111 }, { "epoch": 2.275597269624573, "grad_norm": 0.49229076504707336, "learning_rate": 1.6725579612529441e-06, "loss": 0.4666, "step": 7112 }, { "epoch": 2.2759172354948807, "grad_norm": 0.4973360598087311, "learning_rate": 1.6711685330360212e-06, "loss": 0.4604, "step": 7113 }, { "epoch": 2.2762372013651877, "grad_norm": 0.4849456548690796, "learning_rate": 1.6697795663655848e-06, "loss": 0.4274, "step": 7114 }, { "epoch": 2.2765571672354947, "grad_norm": 0.47988003492355347, "learning_rate": 1.6683910614342186e-06, "loss": 0.4254, "step": 7115 }, { "epoch": 2.276877133105802, "grad_norm": 0.5450612306594849, "learning_rate": 1.6670030184344377e-06, "loss": 0.4667, "step": 7116 }, { "epoch": 2.2771970989761092, "grad_norm": 0.502415657043457, "learning_rate": 1.665615437558698e-06, "loss": 0.4497, "step": 7117 }, { "epoch": 2.2775170648464163, "grad_norm": 0.456184983253479, "learning_rate": 1.6642283189993912e-06, "loss": 0.407, "step": 7118 }, { "epoch": 2.2778370307167237, "grad_norm": 0.5014293789863586, "learning_rate": 1.662841662948841e-06, "loss": 0.4767, "step": 7119 }, { "epoch": 2.2781569965870307, "grad_norm": 0.5107823014259338, "learning_rate": 1.6614554695993085e-06, "loss": 0.3857, "step": 7120 }, { "epoch": 2.2784769624573378, "grad_norm": 0.5226872563362122, "learning_rate": 1.6600697391429904e-06, "loss": 0.4456, "step": 7121 }, { "epoch": 2.2787969283276452, "grad_norm": 0.5985218286514282, "learning_rate": 1.6586844717720208e-06, "loss": 0.442, "step": 7122 }, { "epoch": 2.2791168941979523, "grad_norm": 0.5533532500267029, "learning_rate": 1.6572996676784704e-06, "loss": 0.4335, "step": 7123 }, { "epoch": 2.2794368600682593, "grad_norm": 0.49049973487854004, "learning_rate": 1.6559153270543421e-06, "loss": 0.4206, "step": 7124 }, { "epoch": 2.2797568259385663, "grad_norm": 0.4918212592601776, "learning_rate": 1.6545314500915771e-06, "loss": 0.3945, "step": 7125 }, { "epoch": 2.280076791808874, "grad_norm": 0.48216524720191956, "learning_rate": 1.6531480369820497e-06, "loss": 0.4967, "step": 7126 }, { "epoch": 2.280396757679181, "grad_norm": 0.5223669409751892, "learning_rate": 1.6517650879175733e-06, "loss": 0.4274, "step": 7127 }, { "epoch": 2.280716723549488, "grad_norm": 0.5535790920257568, "learning_rate": 1.6503826030898973e-06, "loss": 0.5179, "step": 7128 }, { "epoch": 2.2810366894197953, "grad_norm": 0.5251159071922302, "learning_rate": 1.6490005826907029e-06, "loss": 0.368, "step": 7129 }, { "epoch": 2.2813566552901023, "grad_norm": 0.5006959438323975, "learning_rate": 1.6476190269116087e-06, "loss": 0.401, "step": 7130 }, { "epoch": 2.2816766211604094, "grad_norm": 0.5298047661781311, "learning_rate": 1.6462379359441682e-06, "loss": 0.4692, "step": 7131 }, { "epoch": 2.281996587030717, "grad_norm": 0.50528484582901, "learning_rate": 1.6448573099798727e-06, "loss": 0.4637, "step": 7132 }, { "epoch": 2.282316552901024, "grad_norm": 0.47289609909057617, "learning_rate": 1.6434771492101487e-06, "loss": 0.4288, "step": 7133 }, { "epoch": 2.282636518771331, "grad_norm": 0.5515410900115967, "learning_rate": 1.6420974538263561e-06, "loss": 0.5173, "step": 7134 }, { "epoch": 2.2829564846416384, "grad_norm": 0.570210874080658, "learning_rate": 1.6407182240197916e-06, "loss": 0.4378, "step": 7135 }, { "epoch": 2.2832764505119454, "grad_norm": 0.4833798110485077, "learning_rate": 1.6393394599816847e-06, "loss": 0.4007, "step": 7136 }, { "epoch": 2.2835964163822524, "grad_norm": 0.47280073165893555, "learning_rate": 1.6379611619032042e-06, "loss": 0.3989, "step": 7137 }, { "epoch": 2.28391638225256, "grad_norm": 0.48720160126686096, "learning_rate": 1.6365833299754558e-06, "loss": 0.4593, "step": 7138 }, { "epoch": 2.284236348122867, "grad_norm": 0.49480339884757996, "learning_rate": 1.635205964389474e-06, "loss": 0.4236, "step": 7139 }, { "epoch": 2.284556313993174, "grad_norm": 0.4934588074684143, "learning_rate": 1.6338290653362338e-06, "loss": 0.4237, "step": 7140 }, { "epoch": 2.2848762798634814, "grad_norm": 0.49687129259109497, "learning_rate": 1.6324526330066414e-06, "loss": 0.4424, "step": 7141 }, { "epoch": 2.2851962457337884, "grad_norm": 0.5061244964599609, "learning_rate": 1.631076667591543e-06, "loss": 0.44, "step": 7142 }, { "epoch": 2.2855162116040955, "grad_norm": 0.42954832315444946, "learning_rate": 1.6297011692817189e-06, "loss": 0.3868, "step": 7143 }, { "epoch": 2.285836177474403, "grad_norm": 0.4803065061569214, "learning_rate": 1.6283261382678827e-06, "loss": 0.4496, "step": 7144 }, { "epoch": 2.28615614334471, "grad_norm": 0.5327868461608887, "learning_rate": 1.6269515747406832e-06, "loss": 0.4348, "step": 7145 }, { "epoch": 2.286476109215017, "grad_norm": 0.5120851993560791, "learning_rate": 1.6255774788907043e-06, "loss": 0.4487, "step": 7146 }, { "epoch": 2.2867960750853245, "grad_norm": 0.535893440246582, "learning_rate": 1.6242038509084672e-06, "loss": 0.4904, "step": 7147 }, { "epoch": 2.2871160409556315, "grad_norm": 0.49842455983161926, "learning_rate": 1.6228306909844294e-06, "loss": 0.4106, "step": 7148 }, { "epoch": 2.2874360068259385, "grad_norm": 0.5016635060310364, "learning_rate": 1.6214579993089784e-06, "loss": 0.4354, "step": 7149 }, { "epoch": 2.287755972696246, "grad_norm": 0.4777538478374481, "learning_rate": 1.6200857760724408e-06, "loss": 0.4368, "step": 7150 }, { "epoch": 2.288075938566553, "grad_norm": 0.5343720316886902, "learning_rate": 1.6187140214650753e-06, "loss": 0.4194, "step": 7151 }, { "epoch": 2.28839590443686, "grad_norm": 0.5467803478240967, "learning_rate": 1.6173427356770778e-06, "loss": 0.4866, "step": 7152 }, { "epoch": 2.288715870307167, "grad_norm": 0.5415265560150146, "learning_rate": 1.615971918898581e-06, "loss": 0.4156, "step": 7153 }, { "epoch": 2.2890358361774745, "grad_norm": 0.45446741580963135, "learning_rate": 1.6146015713196495e-06, "loss": 0.4037, "step": 7154 }, { "epoch": 2.2893558020477816, "grad_norm": 0.5302549004554749, "learning_rate": 1.613231693130282e-06, "loss": 0.4718, "step": 7155 }, { "epoch": 2.2896757679180886, "grad_norm": 0.5353044271469116, "learning_rate": 1.611862284520414e-06, "loss": 0.4428, "step": 7156 }, { "epoch": 2.289995733788396, "grad_norm": 0.5557809472084045, "learning_rate": 1.6104933456799154e-06, "loss": 0.4133, "step": 7157 }, { "epoch": 2.290315699658703, "grad_norm": 0.5576965808868408, "learning_rate": 1.6091248767985939e-06, "loss": 0.4685, "step": 7158 }, { "epoch": 2.29063566552901, "grad_norm": 0.4604426920413971, "learning_rate": 1.607756878066188e-06, "loss": 0.4025, "step": 7159 }, { "epoch": 2.2909556313993176, "grad_norm": 0.4758125841617584, "learning_rate": 1.6063893496723699e-06, "loss": 0.43, "step": 7160 }, { "epoch": 2.2912755972696246, "grad_norm": 0.47438958287239075, "learning_rate": 1.6050222918067527e-06, "loss": 0.4646, "step": 7161 }, { "epoch": 2.2915955631399316, "grad_norm": 0.48651546239852905, "learning_rate": 1.6036557046588775e-06, "loss": 0.4537, "step": 7162 }, { "epoch": 2.291915529010239, "grad_norm": 0.47285035252571106, "learning_rate": 1.602289588418226e-06, "loss": 0.4457, "step": 7163 }, { "epoch": 2.292235494880546, "grad_norm": 0.5200011730194092, "learning_rate": 1.6009239432742114e-06, "loss": 0.4138, "step": 7164 }, { "epoch": 2.292555460750853, "grad_norm": 0.5149220824241638, "learning_rate": 1.599558769416179e-06, "loss": 0.4855, "step": 7165 }, { "epoch": 2.29287542662116, "grad_norm": 0.5541476607322693, "learning_rate": 1.5981940670334156e-06, "loss": 0.4703, "step": 7166 }, { "epoch": 2.2931953924914676, "grad_norm": 0.5186576843261719, "learning_rate": 1.5968298363151364e-06, "loss": 0.4403, "step": 7167 }, { "epoch": 2.2935153583617747, "grad_norm": 0.5135532021522522, "learning_rate": 1.5954660774504954e-06, "loss": 0.4437, "step": 7168 }, { "epoch": 2.2938353242320817, "grad_norm": 0.5300794839859009, "learning_rate": 1.5941027906285788e-06, "loss": 0.4179, "step": 7169 }, { "epoch": 2.294155290102389, "grad_norm": 0.5552196502685547, "learning_rate": 1.592739976038406e-06, "loss": 0.4478, "step": 7170 }, { "epoch": 2.294475255972696, "grad_norm": 0.5434558987617493, "learning_rate": 1.5913776338689364e-06, "loss": 0.3948, "step": 7171 }, { "epoch": 2.294795221843003, "grad_norm": 0.5269661545753479, "learning_rate": 1.5900157643090563e-06, "loss": 0.4469, "step": 7172 }, { "epoch": 2.2951151877133107, "grad_norm": 0.5204504132270813, "learning_rate": 1.5886543675475946e-06, "loss": 0.4475, "step": 7173 }, { "epoch": 2.2954351535836177, "grad_norm": 0.5484360456466675, "learning_rate": 1.5872934437733096e-06, "loss": 0.4304, "step": 7174 }, { "epoch": 2.2957551194539247, "grad_norm": 0.5295978784561157, "learning_rate": 1.5859329931748918e-06, "loss": 0.4746, "step": 7175 }, { "epoch": 2.296075085324232, "grad_norm": 0.5483356714248657, "learning_rate": 1.5845730159409734e-06, "loss": 0.4469, "step": 7176 }, { "epoch": 2.2963950511945392, "grad_norm": 0.5087558031082153, "learning_rate": 1.583213512260114e-06, "loss": 0.4556, "step": 7177 }, { "epoch": 2.2967150170648463, "grad_norm": 0.5023752450942993, "learning_rate": 1.5818544823208126e-06, "loss": 0.4411, "step": 7178 }, { "epoch": 2.2970349829351537, "grad_norm": 0.5368869304656982, "learning_rate": 1.5804959263115e-06, "loss": 0.4111, "step": 7179 }, { "epoch": 2.2973549488054608, "grad_norm": 0.5652300715446472, "learning_rate": 1.5791378444205386e-06, "loss": 0.4498, "step": 7180 }, { "epoch": 2.297674914675768, "grad_norm": 0.5881531834602356, "learning_rate": 1.5777802368362327e-06, "loss": 0.4156, "step": 7181 }, { "epoch": 2.2979948805460753, "grad_norm": 0.495271235704422, "learning_rate": 1.576423103746812e-06, "loss": 0.4475, "step": 7182 }, { "epoch": 2.2983148464163823, "grad_norm": 0.4825277030467987, "learning_rate": 1.5750664453404485e-06, "loss": 0.4638, "step": 7183 }, { "epoch": 2.2986348122866893, "grad_norm": 0.47132453322410583, "learning_rate": 1.573710261805243e-06, "loss": 0.4329, "step": 7184 }, { "epoch": 2.298954778156997, "grad_norm": 0.46688440442085266, "learning_rate": 1.5723545533292294e-06, "loss": 0.3808, "step": 7185 }, { "epoch": 2.299274744027304, "grad_norm": 0.5292215347290039, "learning_rate": 1.5709993201003827e-06, "loss": 0.4736, "step": 7186 }, { "epoch": 2.299594709897611, "grad_norm": 0.5536664128303528, "learning_rate": 1.5696445623066031e-06, "loss": 0.4321, "step": 7187 }, { "epoch": 2.2999146757679183, "grad_norm": 0.5120112299919128, "learning_rate": 1.568290280135734e-06, "loss": 0.4162, "step": 7188 }, { "epoch": 2.3002346416382253, "grad_norm": 0.49979880452156067, "learning_rate": 1.5669364737755455e-06, "loss": 0.4444, "step": 7189 }, { "epoch": 2.3005546075085324, "grad_norm": 0.5113063454627991, "learning_rate": 1.5655831434137437e-06, "loss": 0.4755, "step": 7190 }, { "epoch": 2.30087457337884, "grad_norm": 0.4913010001182556, "learning_rate": 1.5642302892379708e-06, "loss": 0.4138, "step": 7191 }, { "epoch": 2.301194539249147, "grad_norm": 0.48320209980010986, "learning_rate": 1.5628779114358034e-06, "loss": 0.4397, "step": 7192 }, { "epoch": 2.301514505119454, "grad_norm": 0.5086919069290161, "learning_rate": 1.5615260101947482e-06, "loss": 0.4466, "step": 7193 }, { "epoch": 2.301834470989761, "grad_norm": 0.5563963055610657, "learning_rate": 1.5601745857022483e-06, "loss": 0.4394, "step": 7194 }, { "epoch": 2.3021544368600684, "grad_norm": 0.5221656560897827, "learning_rate": 1.5588236381456789e-06, "loss": 0.4436, "step": 7195 }, { "epoch": 2.3024744027303754, "grad_norm": 0.5022855997085571, "learning_rate": 1.5574731677123516e-06, "loss": 0.4231, "step": 7196 }, { "epoch": 2.3027943686006824, "grad_norm": 0.5329191088676453, "learning_rate": 1.5561231745895127e-06, "loss": 0.4163, "step": 7197 }, { "epoch": 2.30311433447099, "grad_norm": 0.5382068753242493, "learning_rate": 1.5547736589643393e-06, "loss": 0.415, "step": 7198 }, { "epoch": 2.303434300341297, "grad_norm": 0.4856245219707489, "learning_rate": 1.5534246210239406e-06, "loss": 0.4048, "step": 7199 }, { "epoch": 2.303754266211604, "grad_norm": 0.6377033591270447, "learning_rate": 1.5520760609553665e-06, "loss": 0.4741, "step": 7200 }, { "epoch": 2.3040742320819114, "grad_norm": 0.5466446280479431, "learning_rate": 1.5507279789455925e-06, "loss": 0.4119, "step": 7201 }, { "epoch": 2.3043941979522184, "grad_norm": 0.5324555039405823, "learning_rate": 1.5493803751815357e-06, "loss": 0.4644, "step": 7202 }, { "epoch": 2.3047141638225255, "grad_norm": 0.5565164685249329, "learning_rate": 1.548033249850041e-06, "loss": 0.4468, "step": 7203 }, { "epoch": 2.305034129692833, "grad_norm": 0.505616307258606, "learning_rate": 1.5466866031378874e-06, "loss": 0.445, "step": 7204 }, { "epoch": 2.30535409556314, "grad_norm": 0.5002743005752563, "learning_rate": 1.5453404352317923e-06, "loss": 0.4665, "step": 7205 }, { "epoch": 2.305674061433447, "grad_norm": 0.49529916048049927, "learning_rate": 1.5439947463184003e-06, "loss": 0.4552, "step": 7206 }, { "epoch": 2.305994027303754, "grad_norm": 0.4952756464481354, "learning_rate": 1.5426495365842958e-06, "loss": 0.3944, "step": 7207 }, { "epoch": 2.3063139931740615, "grad_norm": 0.5098868608474731, "learning_rate": 1.541304806215993e-06, "loss": 0.4404, "step": 7208 }, { "epoch": 2.3066339590443685, "grad_norm": 0.5375380516052246, "learning_rate": 1.5399605553999374e-06, "loss": 0.4544, "step": 7209 }, { "epoch": 2.3069539249146755, "grad_norm": 0.5279937982559204, "learning_rate": 1.5386167843225153e-06, "loss": 0.4714, "step": 7210 }, { "epoch": 2.307273890784983, "grad_norm": 0.4866258203983307, "learning_rate": 1.5372734931700384e-06, "loss": 0.4295, "step": 7211 }, { "epoch": 2.30759385665529, "grad_norm": 0.4863169491291046, "learning_rate": 1.5359306821287595e-06, "loss": 0.4156, "step": 7212 }, { "epoch": 2.307913822525597, "grad_norm": 0.530881404876709, "learning_rate": 1.5345883513848586e-06, "loss": 0.5004, "step": 7213 }, { "epoch": 2.3082337883959045, "grad_norm": 0.5486757755279541, "learning_rate": 1.5332465011244497e-06, "loss": 0.4835, "step": 7214 }, { "epoch": 2.3085537542662116, "grad_norm": 0.4945541322231293, "learning_rate": 1.5319051315335865e-06, "loss": 0.4081, "step": 7215 }, { "epoch": 2.3088737201365186, "grad_norm": 0.511341392993927, "learning_rate": 1.530564242798247e-06, "loss": 0.478, "step": 7216 }, { "epoch": 2.309193686006826, "grad_norm": 0.555908739566803, "learning_rate": 1.5292238351043503e-06, "loss": 0.4588, "step": 7217 }, { "epoch": 2.309513651877133, "grad_norm": 0.47775009274482727, "learning_rate": 1.5278839086377445e-06, "loss": 0.3828, "step": 7218 }, { "epoch": 2.30983361774744, "grad_norm": 0.5278932452201843, "learning_rate": 1.5265444635842109e-06, "loss": 0.4886, "step": 7219 }, { "epoch": 2.3101535836177476, "grad_norm": 0.509454607963562, "learning_rate": 1.525205500129467e-06, "loss": 0.4453, "step": 7220 }, { "epoch": 2.3104735494880546, "grad_norm": 0.4423026442527771, "learning_rate": 1.5238670184591591e-06, "loss": 0.4265, "step": 7221 }, { "epoch": 2.3107935153583616, "grad_norm": 0.48173072934150696, "learning_rate": 1.5225290187588732e-06, "loss": 0.4352, "step": 7222 }, { "epoch": 2.311113481228669, "grad_norm": 0.5268172025680542, "learning_rate": 1.521191501214122e-06, "loss": 0.3967, "step": 7223 }, { "epoch": 2.311433447098976, "grad_norm": 0.5228315591812134, "learning_rate": 1.5198544660103525e-06, "loss": 0.448, "step": 7224 }, { "epoch": 2.311753412969283, "grad_norm": 0.48992231488227844, "learning_rate": 1.5185179133329498e-06, "loss": 0.4338, "step": 7225 }, { "epoch": 2.3120733788395906, "grad_norm": 0.51529461145401, "learning_rate": 1.5171818433672254e-06, "loss": 0.427, "step": 7226 }, { "epoch": 2.3123933447098977, "grad_norm": 0.5545088052749634, "learning_rate": 1.5158462562984294e-06, "loss": 0.475, "step": 7227 }, { "epoch": 2.3127133105802047, "grad_norm": 0.49719321727752686, "learning_rate": 1.5145111523117417e-06, "loss": 0.3874, "step": 7228 }, { "epoch": 2.313033276450512, "grad_norm": 0.5087726712226868, "learning_rate": 1.513176531592274e-06, "loss": 0.4186, "step": 7229 }, { "epoch": 2.313353242320819, "grad_norm": 0.4915749430656433, "learning_rate": 1.511842394325077e-06, "loss": 0.4615, "step": 7230 }, { "epoch": 2.313673208191126, "grad_norm": 0.48190948367118835, "learning_rate": 1.5105087406951268e-06, "loss": 0.4323, "step": 7231 }, { "epoch": 2.3139931740614337, "grad_norm": 0.5035275816917419, "learning_rate": 1.5091755708873396e-06, "loss": 0.4938, "step": 7232 }, { "epoch": 2.3143131399317407, "grad_norm": 0.4648093283176422, "learning_rate": 1.5078428850865568e-06, "loss": 0.438, "step": 7233 }, { "epoch": 2.3146331058020477, "grad_norm": 0.5146045088768005, "learning_rate": 1.5065106834775606e-06, "loss": 0.441, "step": 7234 }, { "epoch": 2.3149530716723548, "grad_norm": 0.5253714919090271, "learning_rate": 1.5051789662450616e-06, "loss": 0.4765, "step": 7235 }, { "epoch": 2.3152730375426622, "grad_norm": 0.4963928461074829, "learning_rate": 1.503847733573701e-06, "loss": 0.4049, "step": 7236 }, { "epoch": 2.3155930034129693, "grad_norm": 0.5003252029418945, "learning_rate": 1.5025169856480604e-06, "loss": 0.488, "step": 7237 }, { "epoch": 2.3159129692832763, "grad_norm": 0.4468475878238678, "learning_rate": 1.5011867226526455e-06, "loss": 0.3802, "step": 7238 }, { "epoch": 2.3162329351535837, "grad_norm": 0.4494239389896393, "learning_rate": 1.4998569447719019e-06, "loss": 0.4524, "step": 7239 }, { "epoch": 2.3165529010238908, "grad_norm": 0.5028368234634399, "learning_rate": 1.4985276521902043e-06, "loss": 0.4187, "step": 7240 }, { "epoch": 2.316872866894198, "grad_norm": 0.5029499530792236, "learning_rate": 1.497198845091858e-06, "loss": 0.4275, "step": 7241 }, { "epoch": 2.3171928327645053, "grad_norm": 0.5458824634552002, "learning_rate": 1.4958705236611076e-06, "loss": 0.4246, "step": 7242 }, { "epoch": 2.3175127986348123, "grad_norm": 0.472836971282959, "learning_rate": 1.494542688082123e-06, "loss": 0.4972, "step": 7243 }, { "epoch": 2.3178327645051193, "grad_norm": 0.45238152146339417, "learning_rate": 1.4932153385390135e-06, "loss": 0.4245, "step": 7244 }, { "epoch": 2.318152730375427, "grad_norm": 0.5061943531036377, "learning_rate": 1.4918884752158158e-06, "loss": 0.4201, "step": 7245 }, { "epoch": 2.318472696245734, "grad_norm": 0.4750024974346161, "learning_rate": 1.4905620982965002e-06, "loss": 0.431, "step": 7246 }, { "epoch": 2.318792662116041, "grad_norm": 0.5249196290969849, "learning_rate": 1.489236207964973e-06, "loss": 0.4492, "step": 7247 }, { "epoch": 2.319112627986348, "grad_norm": 0.5278351306915283, "learning_rate": 1.4879108044050677e-06, "loss": 0.4652, "step": 7248 }, { "epoch": 2.3194325938566553, "grad_norm": 0.4881724417209625, "learning_rate": 1.4865858878005556e-06, "loss": 0.4085, "step": 7249 }, { "epoch": 2.3197525597269624, "grad_norm": 0.43812426924705505, "learning_rate": 1.4852614583351376e-06, "loss": 0.4386, "step": 7250 }, { "epoch": 2.3200725255972694, "grad_norm": 0.5007078647613525, "learning_rate": 1.4839375161924446e-06, "loss": 0.471, "step": 7251 }, { "epoch": 2.320392491467577, "grad_norm": 0.5291122794151306, "learning_rate": 1.482614061556047e-06, "loss": 0.4495, "step": 7252 }, { "epoch": 2.320712457337884, "grad_norm": 0.5274209380149841, "learning_rate": 1.4812910946094394e-06, "loss": 0.4757, "step": 7253 }, { "epoch": 2.321032423208191, "grad_norm": 0.4935450851917267, "learning_rate": 1.479968615536056e-06, "loss": 0.4118, "step": 7254 }, { "epoch": 2.3213523890784984, "grad_norm": 0.514780580997467, "learning_rate": 1.4786466245192594e-06, "loss": 0.4844, "step": 7255 }, { "epoch": 2.3216723549488054, "grad_norm": 0.4826414883136749, "learning_rate": 1.4773251217423424e-06, "loss": 0.4283, "step": 7256 }, { "epoch": 2.3219923208191124, "grad_norm": 0.4890921115875244, "learning_rate": 1.4760041073885367e-06, "loss": 0.3864, "step": 7257 }, { "epoch": 2.32231228668942, "grad_norm": 0.4811008870601654, "learning_rate": 1.474683581641e-06, "loss": 0.4213, "step": 7258 }, { "epoch": 2.322632252559727, "grad_norm": 0.4668603837490082, "learning_rate": 1.473363544682827e-06, "loss": 0.4649, "step": 7259 }, { "epoch": 2.322952218430034, "grad_norm": 0.4900484085083008, "learning_rate": 1.4720439966970412e-06, "loss": 0.4389, "step": 7260 }, { "epoch": 2.3232721843003414, "grad_norm": 0.5131734609603882, "learning_rate": 1.4707249378665978e-06, "loss": 0.4487, "step": 7261 }, { "epoch": 2.3235921501706485, "grad_norm": 0.4661282002925873, "learning_rate": 1.4694063683743893e-06, "loss": 0.3979, "step": 7262 }, { "epoch": 2.3239121160409555, "grad_norm": 0.5389195680618286, "learning_rate": 1.4680882884032333e-06, "loss": 0.4789, "step": 7263 }, { "epoch": 2.324232081911263, "grad_norm": 0.5405575037002563, "learning_rate": 1.466770698135887e-06, "loss": 0.5275, "step": 7264 }, { "epoch": 2.32455204778157, "grad_norm": 0.5626694560050964, "learning_rate": 1.4654535977550343e-06, "loss": 0.47, "step": 7265 }, { "epoch": 2.324872013651877, "grad_norm": 0.575118362903595, "learning_rate": 1.4641369874432904e-06, "loss": 0.4373, "step": 7266 }, { "epoch": 2.3251919795221845, "grad_norm": 0.4923705756664276, "learning_rate": 1.4628208673832072e-06, "loss": 0.4485, "step": 7267 }, { "epoch": 2.3255119453924915, "grad_norm": 0.4365101754665375, "learning_rate": 1.4615052377572675e-06, "loss": 0.418, "step": 7268 }, { "epoch": 2.3258319112627985, "grad_norm": 0.4855596721172333, "learning_rate": 1.4601900987478834e-06, "loss": 0.4342, "step": 7269 }, { "epoch": 2.326151877133106, "grad_norm": 0.5019798278808594, "learning_rate": 1.458875450537401e-06, "loss": 0.4214, "step": 7270 }, { "epoch": 2.326471843003413, "grad_norm": 0.5033852458000183, "learning_rate": 1.4575612933080952e-06, "loss": 0.4651, "step": 7271 }, { "epoch": 2.32679180887372, "grad_norm": 0.46395087242126465, "learning_rate": 1.4562476272421784e-06, "loss": 0.4352, "step": 7272 }, { "epoch": 2.3271117747440275, "grad_norm": 0.46923959255218506, "learning_rate": 1.4549344525217923e-06, "loss": 0.4099, "step": 7273 }, { "epoch": 2.3274317406143346, "grad_norm": 0.4986625909805298, "learning_rate": 1.4536217693290094e-06, "loss": 0.3984, "step": 7274 }, { "epoch": 2.3277517064846416, "grad_norm": 0.549985408782959, "learning_rate": 1.4523095778458352e-06, "loss": 0.4783, "step": 7275 }, { "epoch": 2.328071672354949, "grad_norm": 0.4645216763019562, "learning_rate": 1.450997878254204e-06, "loss": 0.4007, "step": 7276 }, { "epoch": 2.328391638225256, "grad_norm": 0.49397537112236023, "learning_rate": 1.4496866707359863e-06, "loss": 0.4575, "step": 7277 }, { "epoch": 2.328711604095563, "grad_norm": 0.5235719084739685, "learning_rate": 1.4483759554729844e-06, "loss": 0.4712, "step": 7278 }, { "epoch": 2.32903156996587, "grad_norm": 0.4658564329147339, "learning_rate": 1.447065732646929e-06, "loss": 0.414, "step": 7279 }, { "epoch": 2.3293515358361776, "grad_norm": 0.5144032835960388, "learning_rate": 1.4457560024394846e-06, "loss": 0.4855, "step": 7280 }, { "epoch": 2.3296715017064846, "grad_norm": 0.491800993680954, "learning_rate": 1.4444467650322442e-06, "loss": 0.3843, "step": 7281 }, { "epoch": 2.3299914675767917, "grad_norm": 0.5158237814903259, "learning_rate": 1.4431380206067374e-06, "loss": 0.4379, "step": 7282 }, { "epoch": 2.330311433447099, "grad_norm": 0.5110154747962952, "learning_rate": 1.4418297693444243e-06, "loss": 0.4045, "step": 7283 }, { "epoch": 2.330631399317406, "grad_norm": 0.49886924028396606, "learning_rate": 1.4405220114266949e-06, "loss": 0.4197, "step": 7284 }, { "epoch": 2.330951365187713, "grad_norm": 0.48179560899734497, "learning_rate": 1.4392147470348706e-06, "loss": 0.4358, "step": 7285 }, { "epoch": 2.3312713310580206, "grad_norm": 0.5632516741752625, "learning_rate": 1.4379079763502036e-06, "loss": 0.4751, "step": 7286 }, { "epoch": 2.3315912969283277, "grad_norm": 0.5645520687103271, "learning_rate": 1.4366016995538813e-06, "loss": 0.4382, "step": 7287 }, { "epoch": 2.3319112627986347, "grad_norm": 0.5333660840988159, "learning_rate": 1.4352959168270214e-06, "loss": 0.4607, "step": 7288 }, { "epoch": 2.332231228668942, "grad_norm": 0.48292359709739685, "learning_rate": 1.433990628350671e-06, "loss": 0.4023, "step": 7289 }, { "epoch": 2.332551194539249, "grad_norm": 0.5508813858032227, "learning_rate": 1.4326858343058104e-06, "loss": 0.5212, "step": 7290 }, { "epoch": 2.332871160409556, "grad_norm": 0.4942038059234619, "learning_rate": 1.4313815348733485e-06, "loss": 0.3886, "step": 7291 }, { "epoch": 2.3331911262798632, "grad_norm": 0.5608505606651306, "learning_rate": 1.43007773023413e-06, "loss": 0.4302, "step": 7292 }, { "epoch": 2.3335110921501707, "grad_norm": 0.5095614194869995, "learning_rate": 1.4287744205689308e-06, "loss": 0.4134, "step": 7293 }, { "epoch": 2.3338310580204777, "grad_norm": 0.535039484500885, "learning_rate": 1.4274716060584542e-06, "loss": 0.4951, "step": 7294 }, { "epoch": 2.3341510238907848, "grad_norm": 0.5037428736686707, "learning_rate": 1.4261692868833376e-06, "loss": 0.4663, "step": 7295 }, { "epoch": 2.3344709897610922, "grad_norm": 0.5374264717102051, "learning_rate": 1.424867463224147e-06, "loss": 0.4799, "step": 7296 }, { "epoch": 2.3347909556313993, "grad_norm": 0.4959246814250946, "learning_rate": 1.4235661352613832e-06, "loss": 0.4277, "step": 7297 }, { "epoch": 2.3351109215017063, "grad_norm": 0.5037208199501038, "learning_rate": 1.4222653031754795e-06, "loss": 0.4383, "step": 7298 }, { "epoch": 2.3354308873720138, "grad_norm": 0.46232280135154724, "learning_rate": 1.4209649671467956e-06, "loss": 0.3771, "step": 7299 }, { "epoch": 2.335750853242321, "grad_norm": 0.5119737386703491, "learning_rate": 1.4196651273556244e-06, "loss": 0.4818, "step": 7300 }, { "epoch": 2.336070819112628, "grad_norm": 0.49208730459213257, "learning_rate": 1.418365783982189e-06, "loss": 0.4633, "step": 7301 }, { "epoch": 2.3363907849829353, "grad_norm": 0.4549662172794342, "learning_rate": 1.417066937206647e-06, "loss": 0.4074, "step": 7302 }, { "epoch": 2.3367107508532423, "grad_norm": 0.46344906091690063, "learning_rate": 1.4157685872090854e-06, "loss": 0.4241, "step": 7303 }, { "epoch": 2.3370307167235493, "grad_norm": 0.5186562538146973, "learning_rate": 1.4144707341695213e-06, "loss": 0.4605, "step": 7304 }, { "epoch": 2.337350682593857, "grad_norm": 0.47448334097862244, "learning_rate": 1.4131733782679037e-06, "loss": 0.3952, "step": 7305 }, { "epoch": 2.337670648464164, "grad_norm": 0.568246066570282, "learning_rate": 1.4118765196841106e-06, "loss": 0.4662, "step": 7306 }, { "epoch": 2.337990614334471, "grad_norm": 0.49980390071868896, "learning_rate": 1.4105801585979546e-06, "loss": 0.4151, "step": 7307 }, { "epoch": 2.3383105802047783, "grad_norm": 0.49078819155693054, "learning_rate": 1.4092842951891788e-06, "loss": 0.4645, "step": 7308 }, { "epoch": 2.3386305460750854, "grad_norm": 0.4779077172279358, "learning_rate": 1.4079889296374561e-06, "loss": 0.4436, "step": 7309 }, { "epoch": 2.3389505119453924, "grad_norm": 0.5242723822593689, "learning_rate": 1.406694062122389e-06, "loss": 0.4812, "step": 7310 }, { "epoch": 2.3392704778157, "grad_norm": 0.48143717646598816, "learning_rate": 1.4053996928235113e-06, "loss": 0.4092, "step": 7311 }, { "epoch": 2.339590443686007, "grad_norm": 0.5340338349342346, "learning_rate": 1.4041058219202914e-06, "loss": 0.4572, "step": 7312 }, { "epoch": 2.339910409556314, "grad_norm": 0.47235822677612305, "learning_rate": 1.4028124495921263e-06, "loss": 0.4072, "step": 7313 }, { "epoch": 2.3402303754266214, "grad_norm": 0.5129916667938232, "learning_rate": 1.4015195760183426e-06, "loss": 0.4698, "step": 7314 }, { "epoch": 2.3405503412969284, "grad_norm": 0.5445955991744995, "learning_rate": 1.4002272013781988e-06, "loss": 0.4632, "step": 7315 }, { "epoch": 2.3408703071672354, "grad_norm": 0.5130749940872192, "learning_rate": 1.3989353258508825e-06, "loss": 0.4364, "step": 7316 }, { "epoch": 2.341190273037543, "grad_norm": 0.48371532559394836, "learning_rate": 1.3976439496155158e-06, "loss": 0.4288, "step": 7317 }, { "epoch": 2.34151023890785, "grad_norm": 0.5707089304924011, "learning_rate": 1.396353072851151e-06, "loss": 0.4909, "step": 7318 }, { "epoch": 2.341830204778157, "grad_norm": 0.476028710603714, "learning_rate": 1.3950626957367673e-06, "loss": 0.389, "step": 7319 }, { "epoch": 2.342150170648464, "grad_norm": 0.512357771396637, "learning_rate": 1.3937728184512777e-06, "loss": 0.4261, "step": 7320 }, { "epoch": 2.3424701365187715, "grad_norm": 0.586431622505188, "learning_rate": 1.3924834411735238e-06, "loss": 0.5006, "step": 7321 }, { "epoch": 2.3427901023890785, "grad_norm": 0.5070300698280334, "learning_rate": 1.391194564082281e-06, "loss": 0.4903, "step": 7322 }, { "epoch": 2.3431100682593855, "grad_norm": 0.486686110496521, "learning_rate": 1.3899061873562548e-06, "loss": 0.3947, "step": 7323 }, { "epoch": 2.343430034129693, "grad_norm": 0.4885390102863312, "learning_rate": 1.3886183111740786e-06, "loss": 0.4241, "step": 7324 }, { "epoch": 2.34375, "grad_norm": 0.5397626161575317, "learning_rate": 1.3873309357143183e-06, "loss": 0.4963, "step": 7325 }, { "epoch": 2.344069965870307, "grad_norm": 0.562702476978302, "learning_rate": 1.3860440611554682e-06, "loss": 0.4929, "step": 7326 }, { "epoch": 2.3443899317406145, "grad_norm": 0.5303139090538025, "learning_rate": 1.3847576876759573e-06, "loss": 0.4108, "step": 7327 }, { "epoch": 2.3447098976109215, "grad_norm": 0.5334408283233643, "learning_rate": 1.3834718154541432e-06, "loss": 0.4835, "step": 7328 }, { "epoch": 2.3450298634812285, "grad_norm": 0.440531849861145, "learning_rate": 1.3821864446683126e-06, "loss": 0.4238, "step": 7329 }, { "epoch": 2.345349829351536, "grad_norm": 0.5001742243766785, "learning_rate": 1.3809015754966843e-06, "loss": 0.4815, "step": 7330 }, { "epoch": 2.345669795221843, "grad_norm": 0.4859180748462677, "learning_rate": 1.3796172081174041e-06, "loss": 0.4336, "step": 7331 }, { "epoch": 2.34598976109215, "grad_norm": 0.5429733395576477, "learning_rate": 1.3783333427085543e-06, "loss": 0.4077, "step": 7332 }, { "epoch": 2.346309726962457, "grad_norm": 0.510171115398407, "learning_rate": 1.3770499794481446e-06, "loss": 0.4306, "step": 7333 }, { "epoch": 2.3466296928327646, "grad_norm": 0.48114025592803955, "learning_rate": 1.3757671185141136e-06, "loss": 0.4434, "step": 7334 }, { "epoch": 2.3469496587030716, "grad_norm": 0.5100584030151367, "learning_rate": 1.374484760084332e-06, "loss": 0.4625, "step": 7335 }, { "epoch": 2.3472696245733786, "grad_norm": 0.4963550269603729, "learning_rate": 1.373202904336597e-06, "loss": 0.4786, "step": 7336 }, { "epoch": 2.347589590443686, "grad_norm": 0.5266077518463135, "learning_rate": 1.3719215514486455e-06, "loss": 0.4208, "step": 7337 }, { "epoch": 2.347909556313993, "grad_norm": 0.5163435935974121, "learning_rate": 1.3706407015981355e-06, "loss": 0.4432, "step": 7338 }, { "epoch": 2.3482295221843, "grad_norm": 0.5281748175621033, "learning_rate": 1.3693603549626584e-06, "loss": 0.4732, "step": 7339 }, { "epoch": 2.3485494880546076, "grad_norm": 0.5066831707954407, "learning_rate": 1.3680805117197343e-06, "loss": 0.4656, "step": 7340 }, { "epoch": 2.3488694539249146, "grad_norm": 0.47515058517456055, "learning_rate": 1.3668011720468166e-06, "loss": 0.3988, "step": 7341 }, { "epoch": 2.3491894197952217, "grad_norm": 0.535394012928009, "learning_rate": 1.3655223361212884e-06, "loss": 0.453, "step": 7342 }, { "epoch": 2.349509385665529, "grad_norm": 0.48910632729530334, "learning_rate": 1.3642440041204614e-06, "loss": 0.4184, "step": 7343 }, { "epoch": 2.349829351535836, "grad_norm": 0.4723266065120697, "learning_rate": 1.3629661762215762e-06, "loss": 0.4477, "step": 7344 }, { "epoch": 2.350149317406143, "grad_norm": 0.48808911442756653, "learning_rate": 1.3616888526018057e-06, "loss": 0.5345, "step": 7345 }, { "epoch": 2.3504692832764507, "grad_norm": 0.4407173693180084, "learning_rate": 1.360412033438252e-06, "loss": 0.3538, "step": 7346 }, { "epoch": 2.3507892491467577, "grad_norm": 0.4907718300819397, "learning_rate": 1.35913571890795e-06, "loss": 0.4158, "step": 7347 }, { "epoch": 2.3511092150170647, "grad_norm": 0.5179418325424194, "learning_rate": 1.3578599091878608e-06, "loss": 0.4353, "step": 7348 }, { "epoch": 2.351429180887372, "grad_norm": 0.4885033071041107, "learning_rate": 1.356584604454877e-06, "loss": 0.4483, "step": 7349 }, { "epoch": 2.351749146757679, "grad_norm": 0.49781283736228943, "learning_rate": 1.3553098048858194e-06, "loss": 0.4388, "step": 7350 }, { "epoch": 2.3520691126279862, "grad_norm": 0.5235708355903625, "learning_rate": 1.3540355106574415e-06, "loss": 0.4452, "step": 7351 }, { "epoch": 2.3523890784982937, "grad_norm": 0.4559749364852905, "learning_rate": 1.3527617219464284e-06, "loss": 0.4062, "step": 7352 }, { "epoch": 2.3527090443686007, "grad_norm": 0.5198562741279602, "learning_rate": 1.3514884389293898e-06, "loss": 0.4991, "step": 7353 }, { "epoch": 2.3530290102389078, "grad_norm": 0.5171627402305603, "learning_rate": 1.3502156617828687e-06, "loss": 0.4422, "step": 7354 }, { "epoch": 2.3533489761092152, "grad_norm": 0.48058363795280457, "learning_rate": 1.3489433906833355e-06, "loss": 0.3699, "step": 7355 }, { "epoch": 2.3536689419795223, "grad_norm": 0.4736981689929962, "learning_rate": 1.347671625807193e-06, "loss": 0.4322, "step": 7356 }, { "epoch": 2.3539889078498293, "grad_norm": 0.5326634645462036, "learning_rate": 1.3464003673307751e-06, "loss": 0.4543, "step": 7357 }, { "epoch": 2.3543088737201368, "grad_norm": 0.48081734776496887, "learning_rate": 1.3451296154303423e-06, "loss": 0.4068, "step": 7358 }, { "epoch": 2.354628839590444, "grad_norm": 0.5038999319076538, "learning_rate": 1.3438593702820846e-06, "loss": 0.427, "step": 7359 }, { "epoch": 2.354948805460751, "grad_norm": 0.4914281368255615, "learning_rate": 1.3425896320621224e-06, "loss": 0.4472, "step": 7360 }, { "epoch": 2.355268771331058, "grad_norm": 0.4974600076675415, "learning_rate": 1.3413204009465075e-06, "loss": 0.4807, "step": 7361 }, { "epoch": 2.3555887372013653, "grad_norm": 0.46198078989982605, "learning_rate": 1.340051677111222e-06, "loss": 0.3904, "step": 7362 }, { "epoch": 2.3559087030716723, "grad_norm": 0.5218572616577148, "learning_rate": 1.3387834607321741e-06, "loss": 0.479, "step": 7363 }, { "epoch": 2.3562286689419794, "grad_norm": 0.5179147124290466, "learning_rate": 1.337515751985205e-06, "loss": 0.4263, "step": 7364 }, { "epoch": 2.356548634812287, "grad_norm": 0.5002532005310059, "learning_rate": 1.3362485510460803e-06, "loss": 0.4497, "step": 7365 }, { "epoch": 2.356868600682594, "grad_norm": 0.5245099663734436, "learning_rate": 1.3349818580905027e-06, "loss": 0.4351, "step": 7366 }, { "epoch": 2.357188566552901, "grad_norm": 0.5275083780288696, "learning_rate": 1.3337156732941002e-06, "loss": 0.4113, "step": 7367 }, { "epoch": 2.3575085324232083, "grad_norm": 0.5387864708900452, "learning_rate": 1.3324499968324306e-06, "loss": 0.4592, "step": 7368 }, { "epoch": 2.3578284982935154, "grad_norm": 0.50733482837677, "learning_rate": 1.3311848288809815e-06, "loss": 0.434, "step": 7369 }, { "epoch": 2.3581484641638224, "grad_norm": 0.4945586919784546, "learning_rate": 1.3299201696151676e-06, "loss": 0.4244, "step": 7370 }, { "epoch": 2.35846843003413, "grad_norm": 0.5445435643196106, "learning_rate": 1.3286560192103377e-06, "loss": 0.4904, "step": 7371 }, { "epoch": 2.358788395904437, "grad_norm": 0.46514472365379333, "learning_rate": 1.3273923778417686e-06, "loss": 0.4432, "step": 7372 }, { "epoch": 2.359108361774744, "grad_norm": 0.4684540629386902, "learning_rate": 1.3261292456846648e-06, "loss": 0.409, "step": 7373 }, { "epoch": 2.359428327645051, "grad_norm": 0.5185834169387817, "learning_rate": 1.3248666229141589e-06, "loss": 0.465, "step": 7374 }, { "epoch": 2.3597482935153584, "grad_norm": 0.5179154872894287, "learning_rate": 1.3236045097053191e-06, "loss": 0.4905, "step": 7375 }, { "epoch": 2.3600682593856654, "grad_norm": 0.4815397560596466, "learning_rate": 1.322342906233135e-06, "loss": 0.4296, "step": 7376 }, { "epoch": 2.3603882252559725, "grad_norm": 0.4623855650424957, "learning_rate": 1.3210818126725328e-06, "loss": 0.4024, "step": 7377 }, { "epoch": 2.36070819112628, "grad_norm": 0.5654313564300537, "learning_rate": 1.3198212291983636e-06, "loss": 0.5295, "step": 7378 }, { "epoch": 2.361028156996587, "grad_norm": 0.501433253288269, "learning_rate": 1.3185611559854061e-06, "loss": 0.3838, "step": 7379 }, { "epoch": 2.361348122866894, "grad_norm": 0.48429253697395325, "learning_rate": 1.3173015932083754e-06, "loss": 0.3983, "step": 7380 }, { "epoch": 2.3616680887372015, "grad_norm": 0.5181745886802673, "learning_rate": 1.316042541041907e-06, "loss": 0.4655, "step": 7381 }, { "epoch": 2.3619880546075085, "grad_norm": 0.5786999464035034, "learning_rate": 1.3147839996605743e-06, "loss": 0.4845, "step": 7382 }, { "epoch": 2.3623080204778155, "grad_norm": 0.49806299805641174, "learning_rate": 1.313525969238873e-06, "loss": 0.3984, "step": 7383 }, { "epoch": 2.362627986348123, "grad_norm": 0.5032016634941101, "learning_rate": 1.31226844995123e-06, "loss": 0.4439, "step": 7384 }, { "epoch": 2.36294795221843, "grad_norm": 0.48412343859672546, "learning_rate": 1.3110114419720043e-06, "loss": 0.3864, "step": 7385 }, { "epoch": 2.363267918088737, "grad_norm": 0.5202109217643738, "learning_rate": 1.3097549454754782e-06, "loss": 0.4555, "step": 7386 }, { "epoch": 2.3635878839590445, "grad_norm": 0.5044285655021667, "learning_rate": 1.30849896063587e-06, "loss": 0.4295, "step": 7387 }, { "epoch": 2.3639078498293515, "grad_norm": 0.5070471167564392, "learning_rate": 1.3072434876273222e-06, "loss": 0.4818, "step": 7388 }, { "epoch": 2.3642278156996586, "grad_norm": 0.5185414552688599, "learning_rate": 1.3059885266239058e-06, "loss": 0.4449, "step": 7389 }, { "epoch": 2.364547781569966, "grad_norm": 0.49383845925331116, "learning_rate": 1.3047340777996254e-06, "loss": 0.477, "step": 7390 }, { "epoch": 2.364867747440273, "grad_norm": 0.49830183386802673, "learning_rate": 1.3034801413284092e-06, "loss": 0.4316, "step": 7391 }, { "epoch": 2.36518771331058, "grad_norm": 0.479976087808609, "learning_rate": 1.3022267173841202e-06, "loss": 0.4081, "step": 7392 }, { "epoch": 2.3655076791808876, "grad_norm": 0.48571282625198364, "learning_rate": 1.3009738061405452e-06, "loss": 0.4148, "step": 7393 }, { "epoch": 2.3658276450511946, "grad_norm": 0.5023415684700012, "learning_rate": 1.2997214077714004e-06, "loss": 0.4553, "step": 7394 }, { "epoch": 2.3661476109215016, "grad_norm": 0.4909842908382416, "learning_rate": 1.2984695224503351e-06, "loss": 0.433, "step": 7395 }, { "epoch": 2.366467576791809, "grad_norm": 0.4887163043022156, "learning_rate": 1.2972181503509228e-06, "loss": 0.4577, "step": 7396 }, { "epoch": 2.366787542662116, "grad_norm": 0.46021318435668945, "learning_rate": 1.2959672916466698e-06, "loss": 0.4268, "step": 7397 }, { "epoch": 2.367107508532423, "grad_norm": 0.4698842167854309, "learning_rate": 1.2947169465110077e-06, "loss": 0.3931, "step": 7398 }, { "epoch": 2.3674274744027306, "grad_norm": 0.48359882831573486, "learning_rate": 1.2934671151172974e-06, "loss": 0.4401, "step": 7399 }, { "epoch": 2.3677474402730376, "grad_norm": 0.49636223912239075, "learning_rate": 1.2922177976388318e-06, "loss": 0.4625, "step": 7400 }, { "epoch": 2.3680674061433447, "grad_norm": 0.5095014572143555, "learning_rate": 1.2909689942488273e-06, "loss": 0.4227, "step": 7401 }, { "epoch": 2.368387372013652, "grad_norm": 0.5275278091430664, "learning_rate": 1.2897207051204358e-06, "loss": 0.4322, "step": 7402 }, { "epoch": 2.368707337883959, "grad_norm": 0.4835013449192047, "learning_rate": 1.288472930426732e-06, "loss": 0.4291, "step": 7403 }, { "epoch": 2.369027303754266, "grad_norm": 0.5402299165725708, "learning_rate": 1.28722567034072e-06, "loss": 0.4736, "step": 7404 }, { "epoch": 2.369347269624573, "grad_norm": 0.45481786131858826, "learning_rate": 1.2859789250353367e-06, "loss": 0.411, "step": 7405 }, { "epoch": 2.3696672354948807, "grad_norm": 0.470196008682251, "learning_rate": 1.2847326946834427e-06, "loss": 0.393, "step": 7406 }, { "epoch": 2.3699872013651877, "grad_norm": 0.5229184031486511, "learning_rate": 1.2834869794578308e-06, "loss": 0.5201, "step": 7407 }, { "epoch": 2.3703071672354947, "grad_norm": 0.5163376927375793, "learning_rate": 1.2822417795312193e-06, "loss": 0.4365, "step": 7408 }, { "epoch": 2.370627133105802, "grad_norm": 0.4873303472995758, "learning_rate": 1.280997095076259e-06, "loss": 0.4496, "step": 7409 }, { "epoch": 2.3709470989761092, "grad_norm": 0.505264401435852, "learning_rate": 1.279752926265525e-06, "loss": 0.4387, "step": 7410 }, { "epoch": 2.3712670648464163, "grad_norm": 0.49931851029396057, "learning_rate": 1.2785092732715225e-06, "loss": 0.4161, "step": 7411 }, { "epoch": 2.3715870307167237, "grad_norm": 0.4455562233924866, "learning_rate": 1.2772661362666877e-06, "loss": 0.4451, "step": 7412 }, { "epoch": 2.3719069965870307, "grad_norm": 0.5132851600646973, "learning_rate": 1.2760235154233801e-06, "loss": 0.4357, "step": 7413 }, { "epoch": 2.3722269624573378, "grad_norm": 0.5052698254585266, "learning_rate": 1.2747814109138939e-06, "loss": 0.4303, "step": 7414 }, { "epoch": 2.3725469283276452, "grad_norm": 0.5079882144927979, "learning_rate": 1.2735398229104463e-06, "loss": 0.4499, "step": 7415 }, { "epoch": 2.3728668941979523, "grad_norm": 0.5103276371955872, "learning_rate": 1.2722987515851836e-06, "loss": 0.4663, "step": 7416 }, { "epoch": 2.3731868600682593, "grad_norm": 0.5091108679771423, "learning_rate": 1.2710581971101855e-06, "loss": 0.3743, "step": 7417 }, { "epoch": 2.3735068259385663, "grad_norm": 0.5644721984863281, "learning_rate": 1.2698181596574528e-06, "loss": 0.4721, "step": 7418 }, { "epoch": 2.373826791808874, "grad_norm": 0.5130641460418701, "learning_rate": 1.2685786393989213e-06, "loss": 0.443, "step": 7419 }, { "epoch": 2.374146757679181, "grad_norm": 0.4706560969352722, "learning_rate": 1.2673396365064504e-06, "loss": 0.3924, "step": 7420 }, { "epoch": 2.374466723549488, "grad_norm": 0.5112258195877075, "learning_rate": 1.2661011511518278e-06, "loss": 0.4407, "step": 7421 }, { "epoch": 2.3747866894197953, "grad_norm": 0.5168745517730713, "learning_rate": 1.2648631835067742e-06, "loss": 0.4241, "step": 7422 }, { "epoch": 2.3751066552901023, "grad_norm": 0.49925437569618225, "learning_rate": 1.2636257337429319e-06, "loss": 0.4137, "step": 7423 }, { "epoch": 2.3754266211604094, "grad_norm": 0.5110605359077454, "learning_rate": 1.2623888020318787e-06, "loss": 0.4459, "step": 7424 }, { "epoch": 2.375746587030717, "grad_norm": 0.5304185152053833, "learning_rate": 1.2611523885451137e-06, "loss": 0.4655, "step": 7425 }, { "epoch": 2.376066552901024, "grad_norm": 0.5524001717567444, "learning_rate": 1.259916493454067e-06, "loss": 0.4396, "step": 7426 }, { "epoch": 2.376386518771331, "grad_norm": 0.5668018460273743, "learning_rate": 1.2586811169300994e-06, "loss": 0.4552, "step": 7427 }, { "epoch": 2.3767064846416384, "grad_norm": 0.4877117872238159, "learning_rate": 1.257446259144494e-06, "loss": 0.4348, "step": 7428 }, { "epoch": 2.3770264505119454, "grad_norm": 0.47735869884490967, "learning_rate": 1.2562119202684692e-06, "loss": 0.4409, "step": 7429 }, { "epoch": 2.3773464163822524, "grad_norm": 0.5047822594642639, "learning_rate": 1.2549781004731653e-06, "loss": 0.434, "step": 7430 }, { "epoch": 2.37766638225256, "grad_norm": 0.4749005436897278, "learning_rate": 1.253744799929652e-06, "loss": 0.3931, "step": 7431 }, { "epoch": 2.377986348122867, "grad_norm": 0.5150494575500488, "learning_rate": 1.25251201880893e-06, "loss": 0.4553, "step": 7432 }, { "epoch": 2.378306313993174, "grad_norm": 0.4489477276802063, "learning_rate": 1.2512797572819236e-06, "loss": 0.4603, "step": 7433 }, { "epoch": 2.3786262798634814, "grad_norm": 0.5023684501647949, "learning_rate": 1.2500480155194899e-06, "loss": 0.4793, "step": 7434 }, { "epoch": 2.3789462457337884, "grad_norm": 0.4785566031932831, "learning_rate": 1.2488167936924105e-06, "loss": 0.3931, "step": 7435 }, { "epoch": 2.3792662116040955, "grad_norm": 0.5234454870223999, "learning_rate": 1.2475860919713939e-06, "loss": 0.4481, "step": 7436 }, { "epoch": 2.379586177474403, "grad_norm": 0.5080173015594482, "learning_rate": 1.2463559105270817e-06, "loss": 0.4192, "step": 7437 }, { "epoch": 2.37990614334471, "grad_norm": 0.5243739485740662, "learning_rate": 1.2451262495300366e-06, "loss": 0.4371, "step": 7438 }, { "epoch": 2.380226109215017, "grad_norm": 0.5148787498474121, "learning_rate": 1.2438971091507552e-06, "loss": 0.4154, "step": 7439 }, { "epoch": 2.3805460750853245, "grad_norm": 0.4937624931335449, "learning_rate": 1.2426684895596591e-06, "loss": 0.4816, "step": 7440 }, { "epoch": 2.3808660409556315, "grad_norm": 0.4702324867248535, "learning_rate": 1.241440390927095e-06, "loss": 0.3737, "step": 7441 }, { "epoch": 2.3811860068259385, "grad_norm": 0.5017414093017578, "learning_rate": 1.2402128134233437e-06, "loss": 0.4706, "step": 7442 }, { "epoch": 2.381505972696246, "grad_norm": 0.49768519401550293, "learning_rate": 1.2389857572186082e-06, "loss": 0.425, "step": 7443 }, { "epoch": 2.381825938566553, "grad_norm": 0.47565585374832153, "learning_rate": 1.2377592224830232e-06, "loss": 0.4117, "step": 7444 }, { "epoch": 2.38214590443686, "grad_norm": 0.46606653928756714, "learning_rate": 1.2365332093866477e-06, "loss": 0.4822, "step": 7445 }, { "epoch": 2.382465870307167, "grad_norm": 0.46190446615219116, "learning_rate": 1.2353077180994693e-06, "loss": 0.4384, "step": 7446 }, { "epoch": 2.3827858361774745, "grad_norm": 0.5038273930549622, "learning_rate": 1.2340827487914048e-06, "loss": 0.4312, "step": 7447 }, { "epoch": 2.3831058020477816, "grad_norm": 0.5104163289070129, "learning_rate": 1.2328583016322992e-06, "loss": 0.4671, "step": 7448 }, { "epoch": 2.3834257679180886, "grad_norm": 0.5152691006660461, "learning_rate": 1.231634376791922e-06, "loss": 0.4744, "step": 7449 }, { "epoch": 2.383745733788396, "grad_norm": 0.5327239632606506, "learning_rate": 1.2304109744399717e-06, "loss": 0.436, "step": 7450 }, { "epoch": 2.384065699658703, "grad_norm": 0.5346876978874207, "learning_rate": 1.2291880947460732e-06, "loss": 0.4468, "step": 7451 }, { "epoch": 2.38438566552901, "grad_norm": 0.4823782444000244, "learning_rate": 1.227965737879782e-06, "loss": 0.414, "step": 7452 }, { "epoch": 2.3847056313993176, "grad_norm": 0.4793708920478821, "learning_rate": 1.2267439040105806e-06, "loss": 0.4748, "step": 7453 }, { "epoch": 2.3850255972696246, "grad_norm": 0.5064877867698669, "learning_rate": 1.2255225933078762e-06, "loss": 0.4756, "step": 7454 }, { "epoch": 2.3853455631399316, "grad_norm": 0.44026434421539307, "learning_rate": 1.224301805941005e-06, "loss": 0.3681, "step": 7455 }, { "epoch": 2.385665529010239, "grad_norm": 0.519895613193512, "learning_rate": 1.223081542079229e-06, "loss": 0.4246, "step": 7456 }, { "epoch": 2.385985494880546, "grad_norm": 0.554022490978241, "learning_rate": 1.2218618018917412e-06, "loss": 0.48, "step": 7457 }, { "epoch": 2.386305460750853, "grad_norm": 0.5087018013000488, "learning_rate": 1.2206425855476612e-06, "loss": 0.4163, "step": 7458 }, { "epoch": 2.38662542662116, "grad_norm": 0.4920301139354706, "learning_rate": 1.2194238932160335e-06, "loss": 0.4094, "step": 7459 }, { "epoch": 2.3869453924914676, "grad_norm": 0.5391239523887634, "learning_rate": 1.2182057250658307e-06, "loss": 0.4407, "step": 7460 }, { "epoch": 2.3872653583617747, "grad_norm": 0.5132863521575928, "learning_rate": 1.216988081265953e-06, "loss": 0.4178, "step": 7461 }, { "epoch": 2.3875853242320817, "grad_norm": 0.5124856233596802, "learning_rate": 1.2157709619852282e-06, "loss": 0.4094, "step": 7462 }, { "epoch": 2.387905290102389, "grad_norm": 0.5079914331436157, "learning_rate": 1.2145543673924142e-06, "loss": 0.4194, "step": 7463 }, { "epoch": 2.388225255972696, "grad_norm": 0.4989244043827057, "learning_rate": 1.213338297656191e-06, "loss": 0.4285, "step": 7464 }, { "epoch": 2.388545221843003, "grad_norm": 0.46511325240135193, "learning_rate": 1.2121227529451678e-06, "loss": 0.4163, "step": 7465 }, { "epoch": 2.3888651877133107, "grad_norm": 0.5314695239067078, "learning_rate": 1.2109077334278807e-06, "loss": 0.4577, "step": 7466 }, { "epoch": 2.3891851535836177, "grad_norm": 0.4955805838108063, "learning_rate": 1.2096932392727946e-06, "loss": 0.3961, "step": 7467 }, { "epoch": 2.3895051194539247, "grad_norm": 0.5038869380950928, "learning_rate": 1.2084792706483022e-06, "loss": 0.5004, "step": 7468 }, { "epoch": 2.389825085324232, "grad_norm": 0.5208796858787537, "learning_rate": 1.2072658277227206e-06, "loss": 0.4661, "step": 7469 }, { "epoch": 2.3901450511945392, "grad_norm": 0.5192559361457825, "learning_rate": 1.2060529106642943e-06, "loss": 0.4651, "step": 7470 }, { "epoch": 2.3904650170648463, "grad_norm": 0.482944130897522, "learning_rate": 1.204840519641195e-06, "loss": 0.4322, "step": 7471 }, { "epoch": 2.3907849829351537, "grad_norm": 0.4873718321323395, "learning_rate": 1.203628654821523e-06, "loss": 0.4704, "step": 7472 }, { "epoch": 2.3911049488054608, "grad_norm": 0.4696694016456604, "learning_rate": 1.2024173163733072e-06, "loss": 0.3924, "step": 7473 }, { "epoch": 2.391424914675768, "grad_norm": 0.5030756592750549, "learning_rate": 1.2012065044644988e-06, "loss": 0.4325, "step": 7474 }, { "epoch": 2.3917448805460753, "grad_norm": 0.5451432466506958, "learning_rate": 1.1999962192629782e-06, "loss": 0.4564, "step": 7475 }, { "epoch": 2.3920648464163823, "grad_norm": 0.5294658541679382, "learning_rate": 1.1987864609365523e-06, "loss": 0.4487, "step": 7476 }, { "epoch": 2.3923848122866893, "grad_norm": 0.4775328040122986, "learning_rate": 1.1975772296529564e-06, "loss": 0.4161, "step": 7477 }, { "epoch": 2.392704778156997, "grad_norm": 0.4947337210178375, "learning_rate": 1.1963685255798535e-06, "loss": 0.3921, "step": 7478 }, { "epoch": 2.393024744027304, "grad_norm": 0.5113786458969116, "learning_rate": 1.1951603488848307e-06, "loss": 0.4679, "step": 7479 }, { "epoch": 2.393344709897611, "grad_norm": 0.4797898828983307, "learning_rate": 1.1939526997354029e-06, "loss": 0.4468, "step": 7480 }, { "epoch": 2.3936646757679183, "grad_norm": 0.49606427550315857, "learning_rate": 1.1927455782990088e-06, "loss": 0.397, "step": 7481 }, { "epoch": 2.3939846416382253, "grad_norm": 0.5278741717338562, "learning_rate": 1.1915389847430231e-06, "loss": 0.4815, "step": 7482 }, { "epoch": 2.3943046075085324, "grad_norm": 0.505317747592926, "learning_rate": 1.1903329192347397e-06, "loss": 0.431, "step": 7483 }, { "epoch": 2.39462457337884, "grad_norm": 0.5160964131355286, "learning_rate": 1.1891273819413796e-06, "loss": 0.4359, "step": 7484 }, { "epoch": 2.394944539249147, "grad_norm": 0.4927464723587036, "learning_rate": 1.1879223730300926e-06, "loss": 0.4245, "step": 7485 }, { "epoch": 2.395264505119454, "grad_norm": 0.5302851796150208, "learning_rate": 1.1867178926679519e-06, "loss": 0.3868, "step": 7486 }, { "epoch": 2.395584470989761, "grad_norm": 0.5058532953262329, "learning_rate": 1.1855139410219657e-06, "loss": 0.4229, "step": 7487 }, { "epoch": 2.3959044368600684, "grad_norm": 0.5502296686172485, "learning_rate": 1.1843105182590609e-06, "loss": 0.4453, "step": 7488 }, { "epoch": 2.3962244027303754, "grad_norm": 0.5319601893424988, "learning_rate": 1.1831076245460926e-06, "loss": 0.4374, "step": 7489 }, { "epoch": 2.3965443686006824, "grad_norm": 0.5216536521911621, "learning_rate": 1.1819052600498444e-06, "loss": 0.4219, "step": 7490 }, { "epoch": 2.39686433447099, "grad_norm": 0.5190451741218567, "learning_rate": 1.1807034249370218e-06, "loss": 0.4338, "step": 7491 }, { "epoch": 2.397184300341297, "grad_norm": 0.4999808073043823, "learning_rate": 1.1795021193742673e-06, "loss": 0.4069, "step": 7492 }, { "epoch": 2.397504266211604, "grad_norm": 0.47495537996292114, "learning_rate": 1.1783013435281398e-06, "loss": 0.426, "step": 7493 }, { "epoch": 2.3978242320819114, "grad_norm": 0.5205498933792114, "learning_rate": 1.1771010975651287e-06, "loss": 0.4721, "step": 7494 }, { "epoch": 2.3981441979522184, "grad_norm": 0.5490537881851196, "learning_rate": 1.1759013816516486e-06, "loss": 0.4514, "step": 7495 }, { "epoch": 2.3984641638225255, "grad_norm": 0.5139085054397583, "learning_rate": 1.1747021959540418e-06, "loss": 0.4604, "step": 7496 }, { "epoch": 2.398784129692833, "grad_norm": 0.5006527304649353, "learning_rate": 1.1735035406385792e-06, "loss": 0.4862, "step": 7497 }, { "epoch": 2.39910409556314, "grad_norm": 0.5009004473686218, "learning_rate": 1.1723054158714542e-06, "loss": 0.4324, "step": 7498 }, { "epoch": 2.399424061433447, "grad_norm": 0.48826852440834045, "learning_rate": 1.1711078218187878e-06, "loss": 0.4846, "step": 7499 }, { "epoch": 2.399744027303754, "grad_norm": 0.4620589315891266, "learning_rate": 1.169910758646627e-06, "loss": 0.4054, "step": 7500 }, { "epoch": 2.4000639931740615, "grad_norm": 0.5198194980621338, "learning_rate": 1.1687142265209471e-06, "loss": 0.4111, "step": 7501 }, { "epoch": 2.4003839590443685, "grad_norm": 0.5378292202949524, "learning_rate": 1.16751822560765e-06, "loss": 0.4924, "step": 7502 }, { "epoch": 2.4007039249146755, "grad_norm": 0.4919576644897461, "learning_rate": 1.166322756072562e-06, "loss": 0.4569, "step": 7503 }, { "epoch": 2.401023890784983, "grad_norm": 0.4672337472438812, "learning_rate": 1.1651278180814356e-06, "loss": 0.3991, "step": 7504 }, { "epoch": 2.40134385665529, "grad_norm": 0.509486973285675, "learning_rate": 1.1639334117999495e-06, "loss": 0.4398, "step": 7505 }, { "epoch": 2.401663822525597, "grad_norm": 0.5020511150360107, "learning_rate": 1.1627395373937105e-06, "loss": 0.4081, "step": 7506 }, { "epoch": 2.4019837883959045, "grad_norm": 0.534724771976471, "learning_rate": 1.1615461950282526e-06, "loss": 0.4332, "step": 7507 }, { "epoch": 2.4023037542662116, "grad_norm": 0.4937615990638733, "learning_rate": 1.1603533848690323e-06, "loss": 0.4093, "step": 7508 }, { "epoch": 2.4026237201365186, "grad_norm": 0.5317237377166748, "learning_rate": 1.1591611070814345e-06, "loss": 0.4902, "step": 7509 }, { "epoch": 2.402943686006826, "grad_norm": 0.5365759134292603, "learning_rate": 1.1579693618307685e-06, "loss": 0.4281, "step": 7510 }, { "epoch": 2.403263651877133, "grad_norm": 0.4887164831161499, "learning_rate": 1.1567781492822728e-06, "loss": 0.4479, "step": 7511 }, { "epoch": 2.40358361774744, "grad_norm": 0.4960786700248718, "learning_rate": 1.155587469601111e-06, "loss": 0.4094, "step": 7512 }, { "epoch": 2.4039035836177476, "grad_norm": 0.4631226658821106, "learning_rate": 1.154397322952372e-06, "loss": 0.4193, "step": 7513 }, { "epoch": 2.4042235494880546, "grad_norm": 0.4892306923866272, "learning_rate": 1.1532077095010701e-06, "loss": 0.4171, "step": 7514 }, { "epoch": 2.4045435153583616, "grad_norm": 0.45591482520103455, "learning_rate": 1.152018629412146e-06, "loss": 0.4065, "step": 7515 }, { "epoch": 2.404863481228669, "grad_norm": 0.5342705249786377, "learning_rate": 1.1508300828504682e-06, "loss": 0.5449, "step": 7516 }, { "epoch": 2.405183447098976, "grad_norm": 0.4758126437664032, "learning_rate": 1.1496420699808313e-06, "loss": 0.4127, "step": 7517 }, { "epoch": 2.405503412969283, "grad_norm": 0.5583046078681946, "learning_rate": 1.148454590967954e-06, "loss": 0.4841, "step": 7518 }, { "epoch": 2.4058233788395906, "grad_norm": 0.5164111852645874, "learning_rate": 1.1472676459764804e-06, "loss": 0.4191, "step": 7519 }, { "epoch": 2.4061433447098977, "grad_norm": 0.46658846735954285, "learning_rate": 1.1460812351709816e-06, "loss": 0.4027, "step": 7520 }, { "epoch": 2.4064633105802047, "grad_norm": 0.5523401498794556, "learning_rate": 1.1448953587159562e-06, "loss": 0.4675, "step": 7521 }, { "epoch": 2.406783276450512, "grad_norm": 0.4960285723209381, "learning_rate": 1.143710016775828e-06, "loss": 0.4117, "step": 7522 }, { "epoch": 2.407103242320819, "grad_norm": 0.5251019597053528, "learning_rate": 1.1425252095149454e-06, "loss": 0.4264, "step": 7523 }, { "epoch": 2.407423208191126, "grad_norm": 0.5679874420166016, "learning_rate": 1.141340937097583e-06, "loss": 0.5073, "step": 7524 }, { "epoch": 2.4077431740614337, "grad_norm": 0.5471564531326294, "learning_rate": 1.1401571996879406e-06, "loss": 0.4609, "step": 7525 }, { "epoch": 2.4080631399317407, "grad_norm": 0.5527142286300659, "learning_rate": 1.1389739974501462e-06, "loss": 0.4086, "step": 7526 }, { "epoch": 2.4083831058020477, "grad_norm": 0.6052654981613159, "learning_rate": 1.137791330548253e-06, "loss": 0.4123, "step": 7527 }, { "epoch": 2.4087030716723548, "grad_norm": 0.4928595721721649, "learning_rate": 1.1366091991462386e-06, "loss": 0.5074, "step": 7528 }, { "epoch": 2.4090230375426622, "grad_norm": 0.5060864090919495, "learning_rate": 1.1354276034080059e-06, "loss": 0.4108, "step": 7529 }, { "epoch": 2.4093430034129693, "grad_norm": 0.5305314064025879, "learning_rate": 1.1342465434973844e-06, "loss": 0.4396, "step": 7530 }, { "epoch": 2.4096629692832763, "grad_norm": 0.5134252905845642, "learning_rate": 1.1330660195781295e-06, "loss": 0.4404, "step": 7531 }, { "epoch": 2.4099829351535837, "grad_norm": 0.551588237285614, "learning_rate": 1.1318860318139251e-06, "loss": 0.4645, "step": 7532 }, { "epoch": 2.4103029010238908, "grad_norm": 0.5015571117401123, "learning_rate": 1.1307065803683752e-06, "loss": 0.4487, "step": 7533 }, { "epoch": 2.410622866894198, "grad_norm": 0.4951358139514923, "learning_rate": 1.1295276654050135e-06, "loss": 0.4366, "step": 7534 }, { "epoch": 2.4109428327645053, "grad_norm": 0.5068965554237366, "learning_rate": 1.1283492870872952e-06, "loss": 0.4491, "step": 7535 }, { "epoch": 2.4112627986348123, "grad_norm": 0.5098022818565369, "learning_rate": 1.1271714455786066e-06, "loss": 0.4491, "step": 7536 }, { "epoch": 2.4115827645051193, "grad_norm": 0.5193520188331604, "learning_rate": 1.1259941410422575e-06, "loss": 0.4263, "step": 7537 }, { "epoch": 2.411902730375427, "grad_norm": 0.5024391412734985, "learning_rate": 1.1248173736414807e-06, "loss": 0.4529, "step": 7538 }, { "epoch": 2.412222696245734, "grad_norm": 0.5296918749809265, "learning_rate": 1.1236411435394378e-06, "loss": 0.4679, "step": 7539 }, { "epoch": 2.412542662116041, "grad_norm": 0.5214098691940308, "learning_rate": 1.1224654508992117e-06, "loss": 0.4602, "step": 7540 }, { "epoch": 2.412862627986348, "grad_norm": 0.510133683681488, "learning_rate": 1.1212902958838162e-06, "loss": 0.3656, "step": 7541 }, { "epoch": 2.4131825938566553, "grad_norm": 0.5108845233917236, "learning_rate": 1.1201156786561884e-06, "loss": 0.4555, "step": 7542 }, { "epoch": 2.4135025597269624, "grad_norm": 0.5690526962280273, "learning_rate": 1.1189415993791891e-06, "loss": 0.4609, "step": 7543 }, { "epoch": 2.4138225255972694, "grad_norm": 0.6002010703086853, "learning_rate": 1.1177680582156064e-06, "loss": 0.5271, "step": 7544 }, { "epoch": 2.414142491467577, "grad_norm": 0.4729880690574646, "learning_rate": 1.1165950553281518e-06, "loss": 0.3858, "step": 7545 }, { "epoch": 2.414462457337884, "grad_norm": 0.521458089351654, "learning_rate": 1.115422590879464e-06, "loss": 0.4827, "step": 7546 }, { "epoch": 2.414782423208191, "grad_norm": 0.5038489103317261, "learning_rate": 1.1142506650321088e-06, "loss": 0.4493, "step": 7547 }, { "epoch": 2.4151023890784984, "grad_norm": 0.47998735308647156, "learning_rate": 1.1130792779485739e-06, "loss": 0.4509, "step": 7548 }, { "epoch": 2.4154223549488054, "grad_norm": 0.47233617305755615, "learning_rate": 1.111908429791273e-06, "loss": 0.4703, "step": 7549 }, { "epoch": 2.4157423208191124, "grad_norm": 0.514860987663269, "learning_rate": 1.1107381207225443e-06, "loss": 0.4379, "step": 7550 }, { "epoch": 2.41606228668942, "grad_norm": 0.5110282897949219, "learning_rate": 1.109568350904654e-06, "loss": 0.4381, "step": 7551 }, { "epoch": 2.416382252559727, "grad_norm": 0.4801933765411377, "learning_rate": 1.108399120499794e-06, "loss": 0.4222, "step": 7552 }, { "epoch": 2.416702218430034, "grad_norm": 0.45939740538597107, "learning_rate": 1.107230429670077e-06, "loss": 0.4076, "step": 7553 }, { "epoch": 2.4170221843003414, "grad_norm": 0.5356783866882324, "learning_rate": 1.1060622785775433e-06, "loss": 0.4688, "step": 7554 }, { "epoch": 2.4173421501706485, "grad_norm": 0.5647779703140259, "learning_rate": 1.1048946673841598e-06, "loss": 0.4787, "step": 7555 }, { "epoch": 2.4176621160409555, "grad_norm": 0.5190406441688538, "learning_rate": 1.1037275962518162e-06, "loss": 0.4312, "step": 7556 }, { "epoch": 2.417982081911263, "grad_norm": 0.4770490825176239, "learning_rate": 1.1025610653423292e-06, "loss": 0.4366, "step": 7557 }, { "epoch": 2.41830204778157, "grad_norm": 0.4899611473083496, "learning_rate": 1.1013950748174396e-06, "loss": 0.4556, "step": 7558 }, { "epoch": 2.418622013651877, "grad_norm": 0.559202253818512, "learning_rate": 1.1002296248388116e-06, "loss": 0.4417, "step": 7559 }, { "epoch": 2.4189419795221845, "grad_norm": 0.5386381149291992, "learning_rate": 1.099064715568039e-06, "loss": 0.4158, "step": 7560 }, { "epoch": 2.4192619453924915, "grad_norm": 0.4839184582233429, "learning_rate": 1.0979003471666355e-06, "loss": 0.4706, "step": 7561 }, { "epoch": 2.4195819112627985, "grad_norm": 0.5128084421157837, "learning_rate": 1.0967365197960444e-06, "loss": 0.4509, "step": 7562 }, { "epoch": 2.419901877133106, "grad_norm": 0.4908483326435089, "learning_rate": 1.09557323361763e-06, "loss": 0.4002, "step": 7563 }, { "epoch": 2.420221843003413, "grad_norm": 0.48702266812324524, "learning_rate": 1.0944104887926833e-06, "loss": 0.4484, "step": 7564 }, { "epoch": 2.42054180887372, "grad_norm": 0.5335679054260254, "learning_rate": 1.0932482854824216e-06, "loss": 0.455, "step": 7565 }, { "epoch": 2.4208617747440275, "grad_norm": 0.5421773791313171, "learning_rate": 1.0920866238479843e-06, "loss": 0.4386, "step": 7566 }, { "epoch": 2.4211817406143346, "grad_norm": 0.5170466899871826, "learning_rate": 1.090925504050439e-06, "loss": 0.4574, "step": 7567 }, { "epoch": 2.4215017064846416, "grad_norm": 0.537121057510376, "learning_rate": 1.0897649262507753e-06, "loss": 0.4154, "step": 7568 }, { "epoch": 2.421821672354949, "grad_norm": 0.4950486123561859, "learning_rate": 1.0886048906099073e-06, "loss": 0.4175, "step": 7569 }, { "epoch": 2.422141638225256, "grad_norm": 0.46320489048957825, "learning_rate": 1.0874453972886783e-06, "loss": 0.4659, "step": 7570 }, { "epoch": 2.422461604095563, "grad_norm": 0.4615775942802429, "learning_rate": 1.08628644644785e-06, "loss": 0.4238, "step": 7571 }, { "epoch": 2.42278156996587, "grad_norm": 0.49678483605384827, "learning_rate": 1.0851280382481161e-06, "loss": 0.416, "step": 7572 }, { "epoch": 2.4231015358361776, "grad_norm": 0.5429278016090393, "learning_rate": 1.0839701728500896e-06, "loss": 0.4555, "step": 7573 }, { "epoch": 2.4234215017064846, "grad_norm": 0.5306439399719238, "learning_rate": 1.0828128504143076e-06, "loss": 0.4428, "step": 7574 }, { "epoch": 2.4237414675767917, "grad_norm": 0.5775001645088196, "learning_rate": 1.0816560711012386e-06, "loss": 0.4523, "step": 7575 }, { "epoch": 2.424061433447099, "grad_norm": 0.5291292071342468, "learning_rate": 1.080499835071267e-06, "loss": 0.4678, "step": 7576 }, { "epoch": 2.424381399317406, "grad_norm": 0.4802502989768982, "learning_rate": 1.0793441424847106e-06, "loss": 0.4286, "step": 7577 }, { "epoch": 2.424701365187713, "grad_norm": 0.48766323924064636, "learning_rate": 1.0781889935018053e-06, "loss": 0.4102, "step": 7578 }, { "epoch": 2.4250213310580206, "grad_norm": 0.5157595872879028, "learning_rate": 1.0770343882827127e-06, "loss": 0.4421, "step": 7579 }, { "epoch": 2.4253412969283277, "grad_norm": 0.5109654664993286, "learning_rate": 1.0758803269875228e-06, "loss": 0.4911, "step": 7580 }, { "epoch": 2.4256612627986347, "grad_norm": 0.462458997964859, "learning_rate": 1.0747268097762454e-06, "loss": 0.4184, "step": 7581 }, { "epoch": 2.425981228668942, "grad_norm": 0.5116490125656128, "learning_rate": 1.0735738368088188e-06, "loss": 0.4274, "step": 7582 }, { "epoch": 2.426301194539249, "grad_norm": 0.5581761002540588, "learning_rate": 1.072421408245104e-06, "loss": 0.4847, "step": 7583 }, { "epoch": 2.426621160409556, "grad_norm": 0.539322018623352, "learning_rate": 1.071269524244884e-06, "loss": 0.4175, "step": 7584 }, { "epoch": 2.4269411262798632, "grad_norm": 0.4879395365715027, "learning_rate": 1.0701181849678726e-06, "loss": 0.4472, "step": 7585 }, { "epoch": 2.4272610921501707, "grad_norm": 0.4611521363258362, "learning_rate": 1.0689673905737013e-06, "loss": 0.4399, "step": 7586 }, { "epoch": 2.4275810580204777, "grad_norm": 0.4945104420185089, "learning_rate": 1.0678171412219317e-06, "loss": 0.4799, "step": 7587 }, { "epoch": 2.4279010238907848, "grad_norm": 0.4701174199581146, "learning_rate": 1.0666674370720442e-06, "loss": 0.4129, "step": 7588 }, { "epoch": 2.4282209897610922, "grad_norm": 0.5425107479095459, "learning_rate": 1.0655182782834505e-06, "loss": 0.4462, "step": 7589 }, { "epoch": 2.4285409556313993, "grad_norm": 0.5583622455596924, "learning_rate": 1.0643696650154805e-06, "loss": 0.4759, "step": 7590 }, { "epoch": 2.4288609215017063, "grad_norm": 0.5604155659675598, "learning_rate": 1.0632215974273897e-06, "loss": 0.4477, "step": 7591 }, { "epoch": 2.4291808873720138, "grad_norm": 0.47683799266815186, "learning_rate": 1.062074075678362e-06, "loss": 0.395, "step": 7592 }, { "epoch": 2.429500853242321, "grad_norm": 0.5691552758216858, "learning_rate": 1.0609270999275e-06, "loss": 0.473, "step": 7593 }, { "epoch": 2.429820819112628, "grad_norm": 0.591694712638855, "learning_rate": 1.0597806703338354e-06, "loss": 0.4721, "step": 7594 }, { "epoch": 2.4301407849829353, "grad_norm": 0.5700772404670715, "learning_rate": 1.0586347870563213e-06, "loss": 0.439, "step": 7595 }, { "epoch": 2.4304607508532423, "grad_norm": 0.5037002563476562, "learning_rate": 1.057489450253834e-06, "loss": 0.3811, "step": 7596 }, { "epoch": 2.4307807167235493, "grad_norm": 0.5461633801460266, "learning_rate": 1.0563446600851784e-06, "loss": 0.4682, "step": 7597 }, { "epoch": 2.431100682593857, "grad_norm": 0.4951798915863037, "learning_rate": 1.0552004167090795e-06, "loss": 0.4089, "step": 7598 }, { "epoch": 2.431420648464164, "grad_norm": 0.5015349984169006, "learning_rate": 1.0540567202841894e-06, "loss": 0.4096, "step": 7599 }, { "epoch": 2.431740614334471, "grad_norm": 0.522261917591095, "learning_rate": 1.052913570969082e-06, "loss": 0.5391, "step": 7600 }, { "epoch": 2.4320605802047783, "grad_norm": 0.4680025577545166, "learning_rate": 1.0517709689222556e-06, "loss": 0.3876, "step": 7601 }, { "epoch": 2.4323805460750854, "grad_norm": 0.5641904473304749, "learning_rate": 1.0506289143021348e-06, "loss": 0.4187, "step": 7602 }, { "epoch": 2.4327005119453924, "grad_norm": 0.4958467185497284, "learning_rate": 1.0494874072670653e-06, "loss": 0.4498, "step": 7603 }, { "epoch": 2.4330204778157, "grad_norm": 0.48778653144836426, "learning_rate": 1.0483464479753208e-06, "loss": 0.4209, "step": 7604 }, { "epoch": 2.433340443686007, "grad_norm": 0.5240517854690552, "learning_rate": 1.047206036585095e-06, "loss": 0.4644, "step": 7605 }, { "epoch": 2.433660409556314, "grad_norm": 0.4938124120235443, "learning_rate": 1.0460661732545063e-06, "loss": 0.4193, "step": 7606 }, { "epoch": 2.4339803754266214, "grad_norm": 0.4557558000087738, "learning_rate": 1.0449268581416012e-06, "loss": 0.4508, "step": 7607 }, { "epoch": 2.4343003412969284, "grad_norm": 0.4963361620903015, "learning_rate": 1.043788091404343e-06, "loss": 0.4048, "step": 7608 }, { "epoch": 2.4346203071672354, "grad_norm": 0.5271517038345337, "learning_rate": 1.0426498732006273e-06, "loss": 0.479, "step": 7609 }, { "epoch": 2.434940273037543, "grad_norm": 0.4944269359111786, "learning_rate": 1.0415122036882674e-06, "loss": 0.4145, "step": 7610 }, { "epoch": 2.43526023890785, "grad_norm": 0.48690417408943176, "learning_rate": 1.0403750830250014e-06, "loss": 0.4536, "step": 7611 }, { "epoch": 2.435580204778157, "grad_norm": 0.5064367055892944, "learning_rate": 1.0392385113684949e-06, "loss": 0.4342, "step": 7612 }, { "epoch": 2.435900170648464, "grad_norm": 0.5235652923583984, "learning_rate": 1.0381024888763324e-06, "loss": 0.4116, "step": 7613 }, { "epoch": 2.4362201365187715, "grad_norm": 0.57554030418396, "learning_rate": 1.0369670157060275e-06, "loss": 0.4318, "step": 7614 }, { "epoch": 2.4365401023890785, "grad_norm": 0.6025678515434265, "learning_rate": 1.0358320920150133e-06, "loss": 0.4823, "step": 7615 }, { "epoch": 2.4368600682593855, "grad_norm": 0.4879830777645111, "learning_rate": 1.0346977179606477e-06, "loss": 0.3813, "step": 7616 }, { "epoch": 2.437180034129693, "grad_norm": 0.5200275778770447, "learning_rate": 1.0335638937002145e-06, "loss": 0.4524, "step": 7617 }, { "epoch": 2.4375, "grad_norm": 0.5162732601165771, "learning_rate": 1.0324306193909185e-06, "loss": 0.4334, "step": 7618 }, { "epoch": 2.437819965870307, "grad_norm": 0.5059532523155212, "learning_rate": 1.031297895189891e-06, "loss": 0.4214, "step": 7619 }, { "epoch": 2.4381399317406145, "grad_norm": 0.5162801742553711, "learning_rate": 1.0301657212541854e-06, "loss": 0.4186, "step": 7620 }, { "epoch": 2.4384598976109215, "grad_norm": 0.5356256365776062, "learning_rate": 1.029034097740777e-06, "loss": 0.4177, "step": 7621 }, { "epoch": 2.4387798634812285, "grad_norm": 0.50387042760849, "learning_rate": 1.0279030248065675e-06, "loss": 0.4223, "step": 7622 }, { "epoch": 2.439099829351536, "grad_norm": 0.5306486487388611, "learning_rate": 1.0267725026083842e-06, "loss": 0.4674, "step": 7623 }, { "epoch": 2.439419795221843, "grad_norm": 0.5405206680297852, "learning_rate": 1.0256425313029727e-06, "loss": 0.5039, "step": 7624 }, { "epoch": 2.43973976109215, "grad_norm": 0.5215193629264832, "learning_rate": 1.0245131110470058e-06, "loss": 0.4404, "step": 7625 }, { "epoch": 2.440059726962457, "grad_norm": 0.477377325296402, "learning_rate": 1.0233842419970773e-06, "loss": 0.4434, "step": 7626 }, { "epoch": 2.4403796928327646, "grad_norm": 0.455512672662735, "learning_rate": 1.022255924309708e-06, "loss": 0.3599, "step": 7627 }, { "epoch": 2.4406996587030716, "grad_norm": 0.4892837107181549, "learning_rate": 1.0211281581413407e-06, "loss": 0.4431, "step": 7628 }, { "epoch": 2.4410196245733786, "grad_norm": 0.492992639541626, "learning_rate": 1.0200009436483416e-06, "loss": 0.4433, "step": 7629 }, { "epoch": 2.441339590443686, "grad_norm": 0.5320823192596436, "learning_rate": 1.0188742809869994e-06, "loss": 0.4859, "step": 7630 }, { "epoch": 2.441659556313993, "grad_norm": 0.5062793493270874, "learning_rate": 1.0177481703135268e-06, "loss": 0.4247, "step": 7631 }, { "epoch": 2.4419795221843, "grad_norm": 0.48646122217178345, "learning_rate": 1.0166226117840606e-06, "loss": 0.4214, "step": 7632 }, { "epoch": 2.4422994880546076, "grad_norm": 0.5569018721580505, "learning_rate": 1.0154976055546627e-06, "loss": 0.452, "step": 7633 }, { "epoch": 2.4426194539249146, "grad_norm": 0.5125252604484558, "learning_rate": 1.0143731517813154e-06, "loss": 0.444, "step": 7634 }, { "epoch": 2.4429394197952217, "grad_norm": 0.5162290930747986, "learning_rate": 1.0132492506199244e-06, "loss": 0.4762, "step": 7635 }, { "epoch": 2.443259385665529, "grad_norm": 0.5181790590286255, "learning_rate": 1.0121259022263219e-06, "loss": 0.4177, "step": 7636 }, { "epoch": 2.443579351535836, "grad_norm": 0.4434118866920471, "learning_rate": 1.0110031067562592e-06, "loss": 0.4134, "step": 7637 }, { "epoch": 2.443899317406143, "grad_norm": 0.5199148654937744, "learning_rate": 1.0098808643654163e-06, "loss": 0.4569, "step": 7638 }, { "epoch": 2.4442192832764507, "grad_norm": 0.5765395760536194, "learning_rate": 1.0087591752093912e-06, "loss": 0.4519, "step": 7639 }, { "epoch": 2.4445392491467577, "grad_norm": 0.483059823513031, "learning_rate": 1.0076380394437073e-06, "loss": 0.438, "step": 7640 }, { "epoch": 2.4448592150170647, "grad_norm": 0.4759012758731842, "learning_rate": 1.0065174572238133e-06, "loss": 0.4232, "step": 7641 }, { "epoch": 2.445179180887372, "grad_norm": 0.5737249255180359, "learning_rate": 1.0053974287050767e-06, "loss": 0.4544, "step": 7642 }, { "epoch": 2.445499146757679, "grad_norm": 0.5525760054588318, "learning_rate": 1.0042779540427937e-06, "loss": 0.4319, "step": 7643 }, { "epoch": 2.4458191126279862, "grad_norm": 0.5611857771873474, "learning_rate": 1.0031590333921791e-06, "loss": 0.4473, "step": 7644 }, { "epoch": 2.4461390784982937, "grad_norm": 0.5297382473945618, "learning_rate": 1.0020406669083721e-06, "loss": 0.4774, "step": 7645 }, { "epoch": 2.4464590443686007, "grad_norm": 0.49018895626068115, "learning_rate": 1.0009228547464373e-06, "loss": 0.4144, "step": 7646 }, { "epoch": 2.4467790102389078, "grad_norm": 0.48025065660476685, "learning_rate": 9.99805597061358e-07, "loss": 0.4625, "step": 7647 }, { "epoch": 2.4470989761092152, "grad_norm": 0.540985643863678, "learning_rate": 9.986888940080468e-07, "loss": 0.4483, "step": 7648 }, { "epoch": 2.4474189419795223, "grad_norm": 0.4940546751022339, "learning_rate": 9.975727457413342e-07, "loss": 0.4043, "step": 7649 }, { "epoch": 2.4477389078498293, "grad_norm": 0.47084319591522217, "learning_rate": 9.964571524159738e-07, "loss": 0.4393, "step": 7650 }, { "epoch": 2.4480588737201368, "grad_norm": 0.4575006067752838, "learning_rate": 9.95342114186647e-07, "loss": 0.4384, "step": 7651 }, { "epoch": 2.448378839590444, "grad_norm": 0.5515487790107727, "learning_rate": 9.942276312079524e-07, "loss": 0.5155, "step": 7652 }, { "epoch": 2.448698805460751, "grad_norm": 0.5114273428916931, "learning_rate": 9.931137036344174e-07, "loss": 0.489, "step": 7653 }, { "epoch": 2.449018771331058, "grad_norm": 0.5337884426116943, "learning_rate": 9.920003316204874e-07, "loss": 0.4619, "step": 7654 }, { "epoch": 2.4493387372013653, "grad_norm": 0.4916205108165741, "learning_rate": 9.908875153205321e-07, "loss": 0.3839, "step": 7655 }, { "epoch": 2.4496587030716723, "grad_norm": 0.4589358866214752, "learning_rate": 9.897752548888461e-07, "loss": 0.4375, "step": 7656 }, { "epoch": 2.4499786689419794, "grad_norm": 0.45303237438201904, "learning_rate": 9.886635504796467e-07, "loss": 0.4195, "step": 7657 }, { "epoch": 2.450298634812287, "grad_norm": 0.4342546761035919, "learning_rate": 9.875524022470713e-07, "loss": 0.4156, "step": 7658 }, { "epoch": 2.450618600682594, "grad_norm": 0.480857253074646, "learning_rate": 9.86441810345183e-07, "loss": 0.4451, "step": 7659 }, { "epoch": 2.450938566552901, "grad_norm": 0.5475195050239563, "learning_rate": 9.85331774927964e-07, "loss": 0.5084, "step": 7660 }, { "epoch": 2.4512585324232083, "grad_norm": 0.48804011940956116, "learning_rate": 9.842222961493247e-07, "loss": 0.4208, "step": 7661 }, { "epoch": 2.4515784982935154, "grad_norm": 0.4797302186489105, "learning_rate": 9.831133741630966e-07, "loss": 0.4538, "step": 7662 }, { "epoch": 2.4518984641638224, "grad_norm": 0.48906010389328003, "learning_rate": 9.82005009123031e-07, "loss": 0.4558, "step": 7663 }, { "epoch": 2.45221843003413, "grad_norm": 0.522978663444519, "learning_rate": 9.808972011828055e-07, "loss": 0.4595, "step": 7664 }, { "epoch": 2.452538395904437, "grad_norm": 0.5094500780105591, "learning_rate": 9.797899504960162e-07, "loss": 0.4489, "step": 7665 }, { "epoch": 2.452858361774744, "grad_norm": 0.5380294322967529, "learning_rate": 9.78683257216187e-07, "loss": 0.4189, "step": 7666 }, { "epoch": 2.453178327645051, "grad_norm": 0.5117015242576599, "learning_rate": 9.775771214967633e-07, "loss": 0.4287, "step": 7667 }, { "epoch": 2.4534982935153584, "grad_norm": 0.4672139585018158, "learning_rate": 9.764715434911104e-07, "loss": 0.4106, "step": 7668 }, { "epoch": 2.4538182593856654, "grad_norm": 0.5067694187164307, "learning_rate": 9.753665233525188e-07, "loss": 0.469, "step": 7669 }, { "epoch": 2.4541382252559725, "grad_norm": 0.48697370290756226, "learning_rate": 9.742620612341992e-07, "loss": 0.4012, "step": 7670 }, { "epoch": 2.45445819112628, "grad_norm": 0.5106503963470459, "learning_rate": 9.731581572892878e-07, "loss": 0.4187, "step": 7671 }, { "epoch": 2.454778156996587, "grad_norm": 0.4614931046962738, "learning_rate": 9.720548116708434e-07, "loss": 0.429, "step": 7672 }, { "epoch": 2.455098122866894, "grad_norm": 0.5304867625236511, "learning_rate": 9.709520245318453e-07, "loss": 0.4584, "step": 7673 }, { "epoch": 2.4554180887372015, "grad_norm": 0.5220844149589539, "learning_rate": 9.698497960251957e-07, "loss": 0.4016, "step": 7674 }, { "epoch": 2.4557380546075085, "grad_norm": 0.4315398633480072, "learning_rate": 9.687481263037196e-07, "loss": 0.4386, "step": 7675 }, { "epoch": 2.4560580204778155, "grad_norm": 0.44316765666007996, "learning_rate": 9.67647015520165e-07, "loss": 0.4077, "step": 7676 }, { "epoch": 2.456377986348123, "grad_norm": 0.5713621973991394, "learning_rate": 9.665464638272044e-07, "loss": 0.5292, "step": 7677 }, { "epoch": 2.45669795221843, "grad_norm": 0.5227526426315308, "learning_rate": 9.654464713774286e-07, "loss": 0.4227, "step": 7678 }, { "epoch": 2.457017918088737, "grad_norm": 0.5182657837867737, "learning_rate": 9.64347038323354e-07, "loss": 0.4476, "step": 7679 }, { "epoch": 2.4573378839590445, "grad_norm": 0.4926142990589142, "learning_rate": 9.632481648174151e-07, "loss": 0.4527, "step": 7680 }, { "epoch": 2.4576578498293515, "grad_norm": 0.48976460099220276, "learning_rate": 9.621498510119753e-07, "loss": 0.4618, "step": 7681 }, { "epoch": 2.4579778156996586, "grad_norm": 0.4745182394981384, "learning_rate": 9.610520970593173e-07, "loss": 0.4175, "step": 7682 }, { "epoch": 2.458297781569966, "grad_norm": 0.5499017238616943, "learning_rate": 9.59954903111645e-07, "loss": 0.4633, "step": 7683 }, { "epoch": 2.458617747440273, "grad_norm": 0.5084142684936523, "learning_rate": 9.588582693210858e-07, "loss": 0.4562, "step": 7684 }, { "epoch": 2.45893771331058, "grad_norm": 0.4486640691757202, "learning_rate": 9.577621958396876e-07, "loss": 0.3917, "step": 7685 }, { "epoch": 2.4592576791808876, "grad_norm": 0.49820923805236816, "learning_rate": 9.566666828194243e-07, "loss": 0.448, "step": 7686 }, { "epoch": 2.4595776450511946, "grad_norm": 0.5000406503677368, "learning_rate": 9.55571730412191e-07, "loss": 0.4524, "step": 7687 }, { "epoch": 2.4598976109215016, "grad_norm": 0.48786410689353943, "learning_rate": 9.544773387698025e-07, "loss": 0.4435, "step": 7688 }, { "epoch": 2.460217576791809, "grad_norm": 0.5494711995124817, "learning_rate": 9.533835080439979e-07, "loss": 0.4356, "step": 7689 }, { "epoch": 2.460537542662116, "grad_norm": 0.5086516737937927, "learning_rate": 9.522902383864369e-07, "loss": 0.4204, "step": 7690 }, { "epoch": 2.460857508532423, "grad_norm": 0.4816153645515442, "learning_rate": 9.511975299487036e-07, "loss": 0.4525, "step": 7691 }, { "epoch": 2.4611774744027306, "grad_norm": 0.45684894919395447, "learning_rate": 9.501053828823054e-07, "loss": 0.4117, "step": 7692 }, { "epoch": 2.4614974402730376, "grad_norm": 0.45524606108665466, "learning_rate": 9.490137973386676e-07, "loss": 0.4343, "step": 7693 }, { "epoch": 2.4618174061433447, "grad_norm": 0.4852887690067291, "learning_rate": 9.479227734691404e-07, "loss": 0.4329, "step": 7694 }, { "epoch": 2.462137372013652, "grad_norm": 0.5166252851486206, "learning_rate": 9.468323114249944e-07, "loss": 0.4293, "step": 7695 }, { "epoch": 2.462457337883959, "grad_norm": 0.5214166641235352, "learning_rate": 9.457424113574238e-07, "loss": 0.4486, "step": 7696 }, { "epoch": 2.462777303754266, "grad_norm": 0.5261161923408508, "learning_rate": 9.446530734175474e-07, "loss": 0.432, "step": 7697 }, { "epoch": 2.463097269624573, "grad_norm": 0.4776790142059326, "learning_rate": 9.435642977564002e-07, "loss": 0.423, "step": 7698 }, { "epoch": 2.4634172354948807, "grad_norm": 0.5481368899345398, "learning_rate": 9.424760845249436e-07, "loss": 0.4894, "step": 7699 }, { "epoch": 2.4637372013651877, "grad_norm": 0.5501801371574402, "learning_rate": 9.413884338740581e-07, "loss": 0.4141, "step": 7700 }, { "epoch": 2.4640571672354947, "grad_norm": 0.5201511383056641, "learning_rate": 9.403013459545485e-07, "loss": 0.4394, "step": 7701 }, { "epoch": 2.464377133105802, "grad_norm": 0.5060268044471741, "learning_rate": 9.392148209171432e-07, "loss": 0.418, "step": 7702 }, { "epoch": 2.4646970989761092, "grad_norm": 0.5173326134681702, "learning_rate": 9.381288589124877e-07, "loss": 0.4517, "step": 7703 }, { "epoch": 2.4650170648464163, "grad_norm": 0.4509728252887726, "learning_rate": 9.370434600911521e-07, "loss": 0.4435, "step": 7704 }, { "epoch": 2.4653370307167237, "grad_norm": 0.5011699199676514, "learning_rate": 9.359586246036278e-07, "loss": 0.4386, "step": 7705 }, { "epoch": 2.4656569965870307, "grad_norm": 0.5216577053070068, "learning_rate": 9.34874352600329e-07, "loss": 0.4481, "step": 7706 }, { "epoch": 2.4659769624573378, "grad_norm": 0.5623379349708557, "learning_rate": 9.337906442315936e-07, "loss": 0.4597, "step": 7707 }, { "epoch": 2.4662969283276452, "grad_norm": 0.5247960090637207, "learning_rate": 9.327074996476765e-07, "loss": 0.4602, "step": 7708 }, { "epoch": 2.4666168941979523, "grad_norm": 0.49052828550338745, "learning_rate": 9.31624918998758e-07, "loss": 0.3796, "step": 7709 }, { "epoch": 2.4669368600682593, "grad_norm": 0.44440314173698425, "learning_rate": 9.305429024349371e-07, "loss": 0.404, "step": 7710 }, { "epoch": 2.4672568259385663, "grad_norm": 0.47280484437942505, "learning_rate": 9.294614501062393e-07, "loss": 0.4225, "step": 7711 }, { "epoch": 2.467576791808874, "grad_norm": 0.47758734226226807, "learning_rate": 9.283805621626091e-07, "loss": 0.436, "step": 7712 }, { "epoch": 2.467896757679181, "grad_norm": 0.5277053117752075, "learning_rate": 9.273002387539126e-07, "loss": 0.4753, "step": 7713 }, { "epoch": 2.468216723549488, "grad_norm": 0.4714113473892212, "learning_rate": 9.262204800299374e-07, "loss": 0.3981, "step": 7714 }, { "epoch": 2.4685366894197953, "grad_norm": 0.5600582957267761, "learning_rate": 9.251412861403919e-07, "loss": 0.5198, "step": 7715 }, { "epoch": 2.4688566552901023, "grad_norm": 0.48871341347694397, "learning_rate": 9.240626572349093e-07, "loss": 0.4157, "step": 7716 }, { "epoch": 2.4691766211604094, "grad_norm": 0.4396373927593231, "learning_rate": 9.229845934630443e-07, "loss": 0.414, "step": 7717 }, { "epoch": 2.469496587030717, "grad_norm": 0.5115680694580078, "learning_rate": 9.219070949742703e-07, "loss": 0.4254, "step": 7718 }, { "epoch": 2.469816552901024, "grad_norm": 0.5043185949325562, "learning_rate": 9.208301619179833e-07, "loss": 0.4315, "step": 7719 }, { "epoch": 2.470136518771331, "grad_norm": 0.5135460495948792, "learning_rate": 9.197537944435009e-07, "loss": 0.4285, "step": 7720 }, { "epoch": 2.4704564846416384, "grad_norm": 0.5104293823242188, "learning_rate": 9.186779927000633e-07, "loss": 0.4469, "step": 7721 }, { "epoch": 2.4707764505119454, "grad_norm": 0.5346876382827759, "learning_rate": 9.176027568368328e-07, "loss": 0.4353, "step": 7722 }, { "epoch": 2.4710964163822524, "grad_norm": 0.4901260435581207, "learning_rate": 9.165280870028919e-07, "loss": 0.4265, "step": 7723 }, { "epoch": 2.47141638225256, "grad_norm": 0.49625712633132935, "learning_rate": 9.154539833472442e-07, "loss": 0.4654, "step": 7724 }, { "epoch": 2.471736348122867, "grad_norm": 0.4737713932991028, "learning_rate": 9.143804460188144e-07, "loss": 0.4264, "step": 7725 }, { "epoch": 2.472056313993174, "grad_norm": 0.4819645285606384, "learning_rate": 9.133074751664506e-07, "loss": 0.4697, "step": 7726 }, { "epoch": 2.4723762798634814, "grad_norm": 0.4653226137161255, "learning_rate": 9.122350709389228e-07, "loss": 0.4025, "step": 7727 }, { "epoch": 2.4726962457337884, "grad_norm": 0.5325989723205566, "learning_rate": 9.111632334849207e-07, "loss": 0.5155, "step": 7728 }, { "epoch": 2.4730162116040955, "grad_norm": 0.4841369390487671, "learning_rate": 9.100919629530536e-07, "loss": 0.4066, "step": 7729 }, { "epoch": 2.473336177474403, "grad_norm": 0.546658992767334, "learning_rate": 9.090212594918574e-07, "loss": 0.4723, "step": 7730 }, { "epoch": 2.47365614334471, "grad_norm": 0.48808419704437256, "learning_rate": 9.07951123249784e-07, "loss": 0.4369, "step": 7731 }, { "epoch": 2.473976109215017, "grad_norm": 0.49957290291786194, "learning_rate": 9.068815543752113e-07, "loss": 0.4063, "step": 7732 }, { "epoch": 2.4742960750853245, "grad_norm": 0.5041957497596741, "learning_rate": 9.058125530164358e-07, "loss": 0.4257, "step": 7733 }, { "epoch": 2.4746160409556315, "grad_norm": 0.4725160300731659, "learning_rate": 9.047441193216733e-07, "loss": 0.4227, "step": 7734 }, { "epoch": 2.4749360068259385, "grad_norm": 0.4849986732006073, "learning_rate": 9.036762534390669e-07, "loss": 0.4623, "step": 7735 }, { "epoch": 2.475255972696246, "grad_norm": 0.4827047288417816, "learning_rate": 9.026089555166745e-07, "loss": 0.3774, "step": 7736 }, { "epoch": 2.475575938566553, "grad_norm": 0.5167376399040222, "learning_rate": 9.015422257024814e-07, "loss": 0.4669, "step": 7737 }, { "epoch": 2.47589590443686, "grad_norm": 0.4632802903652191, "learning_rate": 9.004760641443888e-07, "loss": 0.4574, "step": 7738 }, { "epoch": 2.476215870307167, "grad_norm": 0.4840168356895447, "learning_rate": 8.994104709902207e-07, "loss": 0.4308, "step": 7739 }, { "epoch": 2.4765358361774745, "grad_norm": 0.5287582278251648, "learning_rate": 8.983454463877256e-07, "loss": 0.4503, "step": 7740 }, { "epoch": 2.4768558020477816, "grad_norm": 0.4669416844844818, "learning_rate": 8.972809904845669e-07, "loss": 0.4093, "step": 7741 }, { "epoch": 2.4771757679180886, "grad_norm": 0.5467846393585205, "learning_rate": 8.962171034283362e-07, "loss": 0.4409, "step": 7742 }, { "epoch": 2.477495733788396, "grad_norm": 0.5437021851539612, "learning_rate": 8.95153785366541e-07, "loss": 0.491, "step": 7743 }, { "epoch": 2.477815699658703, "grad_norm": 0.44450342655181885, "learning_rate": 8.940910364466104e-07, "loss": 0.414, "step": 7744 }, { "epoch": 2.47813566552901, "grad_norm": 0.5120328664779663, "learning_rate": 8.930288568158985e-07, "loss": 0.4784, "step": 7745 }, { "epoch": 2.4784556313993176, "grad_norm": 0.48859596252441406, "learning_rate": 8.919672466216744e-07, "loss": 0.4486, "step": 7746 }, { "epoch": 2.4787755972696246, "grad_norm": 0.49027806520462036, "learning_rate": 8.909062060111356e-07, "loss": 0.4157, "step": 7747 }, { "epoch": 2.4790955631399316, "grad_norm": 0.4924556314945221, "learning_rate": 8.89845735131395e-07, "loss": 0.4658, "step": 7748 }, { "epoch": 2.479415529010239, "grad_norm": 0.5062253475189209, "learning_rate": 8.887858341294859e-07, "loss": 0.4544, "step": 7749 }, { "epoch": 2.479735494880546, "grad_norm": 0.4840138852596283, "learning_rate": 8.877265031523685e-07, "loss": 0.4084, "step": 7750 }, { "epoch": 2.480055460750853, "grad_norm": 0.49299880862236023, "learning_rate": 8.866677423469172e-07, "loss": 0.3985, "step": 7751 }, { "epoch": 2.48037542662116, "grad_norm": 0.5082073211669922, "learning_rate": 8.856095518599333e-07, "loss": 0.4692, "step": 7752 }, { "epoch": 2.4806953924914676, "grad_norm": 0.49071094393730164, "learning_rate": 8.845519318381346e-07, "loss": 0.475, "step": 7753 }, { "epoch": 2.4810153583617747, "grad_norm": 0.4831361174583435, "learning_rate": 8.834948824281603e-07, "loss": 0.3992, "step": 7754 }, { "epoch": 2.4813353242320817, "grad_norm": 0.49088916182518005, "learning_rate": 8.824384037765743e-07, "loss": 0.4297, "step": 7755 }, { "epoch": 2.481655290102389, "grad_norm": 0.4936610162258148, "learning_rate": 8.813824960298556e-07, "loss": 0.4683, "step": 7756 }, { "epoch": 2.481975255972696, "grad_norm": 0.5130487084388733, "learning_rate": 8.803271593344098e-07, "loss": 0.4086, "step": 7757 }, { "epoch": 2.482295221843003, "grad_norm": 0.5862554907798767, "learning_rate": 8.792723938365599e-07, "loss": 0.4342, "step": 7758 }, { "epoch": 2.4826151877133107, "grad_norm": 0.5383480191230774, "learning_rate": 8.782181996825484e-07, "loss": 0.4829, "step": 7759 }, { "epoch": 2.4829351535836177, "grad_norm": 0.4855513870716095, "learning_rate": 8.771645770185433e-07, "loss": 0.4232, "step": 7760 }, { "epoch": 2.4832551194539247, "grad_norm": 0.5087752342224121, "learning_rate": 8.761115259906289e-07, "loss": 0.4893, "step": 7761 }, { "epoch": 2.483575085324232, "grad_norm": 0.5325577855110168, "learning_rate": 8.750590467448128e-07, "loss": 0.4296, "step": 7762 }, { "epoch": 2.4838950511945392, "grad_norm": 0.5027719140052795, "learning_rate": 8.740071394270217e-07, "loss": 0.4864, "step": 7763 }, { "epoch": 2.4842150170648463, "grad_norm": 0.48960229754447937, "learning_rate": 8.729558041831055e-07, "loss": 0.4272, "step": 7764 }, { "epoch": 2.4845349829351537, "grad_norm": 0.473838210105896, "learning_rate": 8.719050411588321e-07, "loss": 0.436, "step": 7765 }, { "epoch": 2.4848549488054608, "grad_norm": 0.4734203517436981, "learning_rate": 8.708548504998898e-07, "loss": 0.4474, "step": 7766 }, { "epoch": 2.485174914675768, "grad_norm": 0.5132763981819153, "learning_rate": 8.698052323518913e-07, "loss": 0.4424, "step": 7767 }, { "epoch": 2.4854948805460753, "grad_norm": 0.5405433177947998, "learning_rate": 8.687561868603644e-07, "loss": 0.4259, "step": 7768 }, { "epoch": 2.4858148464163823, "grad_norm": 0.5086086988449097, "learning_rate": 8.677077141707635e-07, "loss": 0.4623, "step": 7769 }, { "epoch": 2.4861348122866893, "grad_norm": 0.437712162733078, "learning_rate": 8.666598144284594e-07, "loss": 0.3828, "step": 7770 }, { "epoch": 2.486454778156997, "grad_norm": 0.4647720456123352, "learning_rate": 8.656124877787437e-07, "loss": 0.3754, "step": 7771 }, { "epoch": 2.486774744027304, "grad_norm": 0.4738596975803375, "learning_rate": 8.645657343668312e-07, "loss": 0.4491, "step": 7772 }, { "epoch": 2.487094709897611, "grad_norm": 0.49070629477500916, "learning_rate": 8.635195543378532e-07, "loss": 0.4705, "step": 7773 }, { "epoch": 2.4874146757679183, "grad_norm": 0.5047940611839294, "learning_rate": 8.624739478368671e-07, "loss": 0.4486, "step": 7774 }, { "epoch": 2.4877346416382253, "grad_norm": 0.4897439479827881, "learning_rate": 8.61428915008844e-07, "loss": 0.4371, "step": 7775 }, { "epoch": 2.4880546075085324, "grad_norm": 0.48489508032798767, "learning_rate": 8.603844559986823e-07, "loss": 0.4098, "step": 7776 }, { "epoch": 2.48837457337884, "grad_norm": 0.44628262519836426, "learning_rate": 8.593405709511954e-07, "loss": 0.4284, "step": 7777 }, { "epoch": 2.488694539249147, "grad_norm": 0.5165276527404785, "learning_rate": 8.582972600111189e-07, "loss": 0.4968, "step": 7778 }, { "epoch": 2.489014505119454, "grad_norm": 0.5198566913604736, "learning_rate": 8.572545233231111e-07, "loss": 0.4104, "step": 7779 }, { "epoch": 2.489334470989761, "grad_norm": 0.5088862180709839, "learning_rate": 8.562123610317457e-07, "loss": 0.4453, "step": 7780 }, { "epoch": 2.4896544368600684, "grad_norm": 0.49031898379325867, "learning_rate": 8.551707732815234e-07, "loss": 0.4559, "step": 7781 }, { "epoch": 2.4899744027303754, "grad_norm": 0.5151739120483398, "learning_rate": 8.541297602168591e-07, "loss": 0.4375, "step": 7782 }, { "epoch": 2.4902943686006824, "grad_norm": 0.5170591473579407, "learning_rate": 8.530893219820907e-07, "loss": 0.4197, "step": 7783 }, { "epoch": 2.49061433447099, "grad_norm": 0.5250839591026306, "learning_rate": 8.520494587214773e-07, "loss": 0.46, "step": 7784 }, { "epoch": 2.490934300341297, "grad_norm": 0.47464457154273987, "learning_rate": 8.510101705791957e-07, "loss": 0.4299, "step": 7785 }, { "epoch": 2.491254266211604, "grad_norm": 0.4447919726371765, "learning_rate": 8.499714576993462e-07, "loss": 0.4009, "step": 7786 }, { "epoch": 2.4915742320819114, "grad_norm": 0.5237582325935364, "learning_rate": 8.48933320225947e-07, "loss": 0.4642, "step": 7787 }, { "epoch": 2.4918941979522184, "grad_norm": 0.5427510738372803, "learning_rate": 8.478957583029351e-07, "loss": 0.5008, "step": 7788 }, { "epoch": 2.4922141638225255, "grad_norm": 0.47621169686317444, "learning_rate": 8.468587720741728e-07, "loss": 0.38, "step": 7789 }, { "epoch": 2.492534129692833, "grad_norm": 0.517081618309021, "learning_rate": 8.458223616834371e-07, "loss": 0.4361, "step": 7790 }, { "epoch": 2.49285409556314, "grad_norm": 0.537865400314331, "learning_rate": 8.447865272744299e-07, "loss": 0.4542, "step": 7791 }, { "epoch": 2.493174061433447, "grad_norm": 0.5118927955627441, "learning_rate": 8.437512689907696e-07, "loss": 0.4356, "step": 7792 }, { "epoch": 2.493494027303754, "grad_norm": 0.4743349552154541, "learning_rate": 8.42716586975994e-07, "loss": 0.4347, "step": 7793 }, { "epoch": 2.4938139931740615, "grad_norm": 0.5667382478713989, "learning_rate": 8.416824813735669e-07, "loss": 0.511, "step": 7794 }, { "epoch": 2.4941339590443685, "grad_norm": 0.532633364200592, "learning_rate": 8.406489523268646e-07, "loss": 0.4498, "step": 7795 }, { "epoch": 2.4944539249146755, "grad_norm": 0.5290338397026062, "learning_rate": 8.396159999791903e-07, "loss": 0.4787, "step": 7796 }, { "epoch": 2.494773890784983, "grad_norm": 0.495960533618927, "learning_rate": 8.385836244737628e-07, "loss": 0.3912, "step": 7797 }, { "epoch": 2.49509385665529, "grad_norm": 0.5184521675109863, "learning_rate": 8.375518259537208e-07, "loss": 0.4055, "step": 7798 }, { "epoch": 2.495413822525597, "grad_norm": 0.5542027950286865, "learning_rate": 8.365206045621271e-07, "loss": 0.4543, "step": 7799 }, { "epoch": 2.4957337883959045, "grad_norm": 0.5616625547409058, "learning_rate": 8.354899604419586e-07, "loss": 0.4486, "step": 7800 }, { "epoch": 2.4960537542662116, "grad_norm": 0.4697994291782379, "learning_rate": 8.344598937361186e-07, "loss": 0.3706, "step": 7801 }, { "epoch": 2.4963737201365186, "grad_norm": 0.505582869052887, "learning_rate": 8.334304045874248e-07, "loss": 0.4675, "step": 7802 }, { "epoch": 2.496693686006826, "grad_norm": 0.5572977066040039, "learning_rate": 8.324014931386187e-07, "loss": 0.4193, "step": 7803 }, { "epoch": 2.497013651877133, "grad_norm": 0.4983901381492615, "learning_rate": 8.313731595323593e-07, "loss": 0.4471, "step": 7804 }, { "epoch": 2.49733361774744, "grad_norm": 0.4947405159473419, "learning_rate": 8.303454039112252e-07, "loss": 0.4262, "step": 7805 }, { "epoch": 2.4976535836177476, "grad_norm": 0.5503841042518616, "learning_rate": 8.293182264177185e-07, "loss": 0.4817, "step": 7806 }, { "epoch": 2.4979735494880546, "grad_norm": 0.47878727316856384, "learning_rate": 8.282916271942553e-07, "loss": 0.4825, "step": 7807 }, { "epoch": 2.4982935153583616, "grad_norm": 0.4644586145877838, "learning_rate": 8.272656063831785e-07, "loss": 0.4158, "step": 7808 }, { "epoch": 2.498613481228669, "grad_norm": 0.5075016021728516, "learning_rate": 8.262401641267454e-07, "loss": 0.4104, "step": 7809 }, { "epoch": 2.498933447098976, "grad_norm": 0.5721554756164551, "learning_rate": 8.252153005671332e-07, "loss": 0.4476, "step": 7810 }, { "epoch": 2.499253412969283, "grad_norm": 0.5362179279327393, "learning_rate": 8.241910158464428e-07, "loss": 0.4312, "step": 7811 }, { "epoch": 2.4995733788395906, "grad_norm": 0.4949038624763489, "learning_rate": 8.231673101066911e-07, "loss": 0.4742, "step": 7812 }, { "epoch": 2.4998933447098977, "grad_norm": 0.45464375615119934, "learning_rate": 8.221441834898175e-07, "loss": 0.4002, "step": 7813 }, { "epoch": 2.5002133105802047, "grad_norm": 0.4683190882205963, "learning_rate": 8.21121636137679e-07, "loss": 0.4823, "step": 7814 }, { "epoch": 2.500533276450512, "grad_norm": 0.4764581322669983, "learning_rate": 8.20099668192052e-07, "loss": 0.4656, "step": 7815 }, { "epoch": 2.500853242320819, "grad_norm": 0.4716465175151825, "learning_rate": 8.190782797946356e-07, "loss": 0.4191, "step": 7816 }, { "epoch": 2.501173208191126, "grad_norm": 0.5018923878669739, "learning_rate": 8.180574710870442e-07, "loss": 0.4141, "step": 7817 }, { "epoch": 2.5014931740614337, "grad_norm": 0.5134195685386658, "learning_rate": 8.170372422108164e-07, "loss": 0.4714, "step": 7818 }, { "epoch": 2.5018131399317407, "grad_norm": 0.4809771478176117, "learning_rate": 8.160175933074077e-07, "loss": 0.404, "step": 7819 }, { "epoch": 2.5021331058020477, "grad_norm": 0.5265135765075684, "learning_rate": 8.149985245181919e-07, "loss": 0.4516, "step": 7820 }, { "epoch": 2.502453071672355, "grad_norm": 0.5177709460258484, "learning_rate": 8.139800359844663e-07, "loss": 0.4256, "step": 7821 }, { "epoch": 2.5027730375426622, "grad_norm": 0.5242189764976501, "learning_rate": 8.129621278474437e-07, "loss": 0.4616, "step": 7822 }, { "epoch": 2.5030930034129693, "grad_norm": 0.4369068145751953, "learning_rate": 8.1194480024826e-07, "loss": 0.4181, "step": 7823 }, { "epoch": 2.5034129692832767, "grad_norm": 0.5571264624595642, "learning_rate": 8.109280533279685e-07, "loss": 0.4737, "step": 7824 }, { "epoch": 2.5037329351535837, "grad_norm": 0.5230574607849121, "learning_rate": 8.099118872275403e-07, "loss": 0.4011, "step": 7825 }, { "epoch": 2.5040529010238908, "grad_norm": 0.4745146632194519, "learning_rate": 8.088963020878715e-07, "loss": 0.4128, "step": 7826 }, { "epoch": 2.504372866894198, "grad_norm": 0.5156494379043579, "learning_rate": 8.078812980497708e-07, "loss": 0.4402, "step": 7827 }, { "epoch": 2.5046928327645053, "grad_norm": 0.5434547662734985, "learning_rate": 8.068668752539726e-07, "loss": 0.4597, "step": 7828 }, { "epoch": 2.5050127986348123, "grad_norm": 0.5473108291625977, "learning_rate": 8.058530338411269e-07, "loss": 0.4405, "step": 7829 }, { "epoch": 2.5053327645051193, "grad_norm": 0.5330970287322998, "learning_rate": 8.048397739518021e-07, "loss": 0.5092, "step": 7830 }, { "epoch": 2.5056527303754264, "grad_norm": 0.5310120582580566, "learning_rate": 8.03827095726491e-07, "loss": 0.4714, "step": 7831 }, { "epoch": 2.505972696245734, "grad_norm": 0.5227869749069214, "learning_rate": 8.028149993056e-07, "loss": 0.3721, "step": 7832 }, { "epoch": 2.506292662116041, "grad_norm": 0.5100311040878296, "learning_rate": 8.018034848294598e-07, "loss": 0.4337, "step": 7833 }, { "epoch": 2.506612627986348, "grad_norm": 0.5334123373031616, "learning_rate": 8.007925524383175e-07, "loss": 0.4494, "step": 7834 }, { "epoch": 2.5069325938566553, "grad_norm": 0.5083934664726257, "learning_rate": 7.997822022723378e-07, "loss": 0.4058, "step": 7835 }, { "epoch": 2.5072525597269624, "grad_norm": 0.47573235630989075, "learning_rate": 7.987724344716091e-07, "loss": 0.4076, "step": 7836 }, { "epoch": 2.5075725255972694, "grad_norm": 0.5759677886962891, "learning_rate": 7.977632491761372e-07, "loss": 0.438, "step": 7837 }, { "epoch": 2.507892491467577, "grad_norm": 1.110116958618164, "learning_rate": 7.967546465258469e-07, "loss": 0.431, "step": 7838 }, { "epoch": 2.508212457337884, "grad_norm": 0.5045497417449951, "learning_rate": 7.957466266605812e-07, "loss": 0.4415, "step": 7839 }, { "epoch": 2.508532423208191, "grad_norm": 0.48604580760002136, "learning_rate": 7.947391897201024e-07, "loss": 0.3933, "step": 7840 }, { "epoch": 2.5088523890784984, "grad_norm": 0.5094166994094849, "learning_rate": 7.937323358440935e-07, "loss": 0.4343, "step": 7841 }, { "epoch": 2.5091723549488054, "grad_norm": 0.538398265838623, "learning_rate": 7.92726065172158e-07, "loss": 0.4715, "step": 7842 }, { "epoch": 2.5094923208191124, "grad_norm": 0.5149754881858826, "learning_rate": 7.917203778438154e-07, "loss": 0.4317, "step": 7843 }, { "epoch": 2.50981228668942, "grad_norm": 0.48621267080307007, "learning_rate": 7.907152739985047e-07, "loss": 0.4721, "step": 7844 }, { "epoch": 2.510132252559727, "grad_norm": 0.4904319643974304, "learning_rate": 7.897107537755844e-07, "loss": 0.4281, "step": 7845 }, { "epoch": 2.510452218430034, "grad_norm": 0.5282472372055054, "learning_rate": 7.887068173143325e-07, "loss": 0.4568, "step": 7846 }, { "epoch": 2.5107721843003414, "grad_norm": 0.49507468938827515, "learning_rate": 7.877034647539483e-07, "loss": 0.4193, "step": 7847 }, { "epoch": 2.5110921501706485, "grad_norm": 0.4763088524341583, "learning_rate": 7.867006962335466e-07, "loss": 0.4008, "step": 7848 }, { "epoch": 2.5114121160409555, "grad_norm": 0.47316059470176697, "learning_rate": 7.856985118921618e-07, "loss": 0.4714, "step": 7849 }, { "epoch": 2.511732081911263, "grad_norm": 0.46475642919540405, "learning_rate": 7.846969118687464e-07, "loss": 0.4173, "step": 7850 }, { "epoch": 2.51205204778157, "grad_norm": 0.5356907248497009, "learning_rate": 7.836958963021762e-07, "loss": 0.4541, "step": 7851 }, { "epoch": 2.512372013651877, "grad_norm": 0.5031693577766418, "learning_rate": 7.826954653312429e-07, "loss": 0.4761, "step": 7852 }, { "epoch": 2.5126919795221845, "grad_norm": 0.49371716380119324, "learning_rate": 7.816956190946573e-07, "loss": 0.4227, "step": 7853 }, { "epoch": 2.5130119453924915, "grad_norm": 0.527224600315094, "learning_rate": 7.80696357731049e-07, "loss": 0.4879, "step": 7854 }, { "epoch": 2.5133319112627985, "grad_norm": 0.46800824999809265, "learning_rate": 7.796976813789647e-07, "loss": 0.4489, "step": 7855 }, { "epoch": 2.513651877133106, "grad_norm": 0.4291780889034271, "learning_rate": 7.786995901768746e-07, "loss": 0.399, "step": 7856 }, { "epoch": 2.513971843003413, "grad_norm": 0.5125072002410889, "learning_rate": 7.777020842631655e-07, "loss": 0.5024, "step": 7857 }, { "epoch": 2.51429180887372, "grad_norm": 0.5313035249710083, "learning_rate": 7.767051637761424e-07, "loss": 0.4516, "step": 7858 }, { "epoch": 2.5146117747440275, "grad_norm": 0.4871886968612671, "learning_rate": 7.757088288540288e-07, "loss": 0.4269, "step": 7859 }, { "epoch": 2.5149317406143346, "grad_norm": 0.5050264000892639, "learning_rate": 7.747130796349672e-07, "loss": 0.3796, "step": 7860 }, { "epoch": 2.5152517064846416, "grad_norm": 0.4975477457046509, "learning_rate": 7.737179162570196e-07, "loss": 0.4757, "step": 7861 }, { "epoch": 2.515571672354949, "grad_norm": 0.561652660369873, "learning_rate": 7.72723338858169e-07, "loss": 0.4734, "step": 7862 }, { "epoch": 2.515891638225256, "grad_norm": 0.5012123584747314, "learning_rate": 7.717293475763126e-07, "loss": 0.4151, "step": 7863 }, { "epoch": 2.516211604095563, "grad_norm": 0.5105333924293518, "learning_rate": 7.707359425492694e-07, "loss": 0.4373, "step": 7864 }, { "epoch": 2.5165315699658706, "grad_norm": 0.48380863666534424, "learning_rate": 7.69743123914774e-07, "loss": 0.4431, "step": 7865 }, { "epoch": 2.5168515358361776, "grad_norm": 0.46601566672325134, "learning_rate": 7.687508918104841e-07, "loss": 0.4391, "step": 7866 }, { "epoch": 2.5171715017064846, "grad_norm": 0.5079662203788757, "learning_rate": 7.677592463739741e-07, "loss": 0.4003, "step": 7867 }, { "epoch": 2.5174914675767917, "grad_norm": 0.4948737323284149, "learning_rate": 7.667681877427363e-07, "loss": 0.4707, "step": 7868 }, { "epoch": 2.517811433447099, "grad_norm": 0.4404037296772003, "learning_rate": 7.657777160541818e-07, "loss": 0.4013, "step": 7869 }, { "epoch": 2.518131399317406, "grad_norm": 0.47359374165534973, "learning_rate": 7.647878314456403e-07, "loss": 0.4264, "step": 7870 }, { "epoch": 2.518451365187713, "grad_norm": 0.5073278546333313, "learning_rate": 7.637985340543602e-07, "loss": 0.4524, "step": 7871 }, { "epoch": 2.51877133105802, "grad_norm": 0.44690725207328796, "learning_rate": 7.628098240175113e-07, "loss": 0.3978, "step": 7872 }, { "epoch": 2.5190912969283277, "grad_norm": 0.4936218559741974, "learning_rate": 7.618217014721779e-07, "loss": 0.4774, "step": 7873 }, { "epoch": 2.5194112627986347, "grad_norm": 0.5053908228874207, "learning_rate": 7.608341665553637e-07, "loss": 0.4424, "step": 7874 }, { "epoch": 2.5197312286689417, "grad_norm": 0.4724701941013336, "learning_rate": 7.598472194039913e-07, "loss": 0.3928, "step": 7875 }, { "epoch": 2.520051194539249, "grad_norm": 0.5237969160079956, "learning_rate": 7.58860860154903e-07, "loss": 0.4649, "step": 7876 }, { "epoch": 2.520371160409556, "grad_norm": 0.5442495346069336, "learning_rate": 7.5787508894486e-07, "loss": 0.4657, "step": 7877 }, { "epoch": 2.5206911262798632, "grad_norm": 0.4638236165046692, "learning_rate": 7.56889905910539e-07, "loss": 0.4268, "step": 7878 }, { "epoch": 2.5210110921501707, "grad_norm": 0.5017217397689819, "learning_rate": 7.559053111885373e-07, "loss": 0.4371, "step": 7879 }, { "epoch": 2.5213310580204777, "grad_norm": 0.4847548305988312, "learning_rate": 7.549213049153687e-07, "loss": 0.4173, "step": 7880 }, { "epoch": 2.5216510238907848, "grad_norm": 0.4981611669063568, "learning_rate": 7.539378872274677e-07, "loss": 0.519, "step": 7881 }, { "epoch": 2.5219709897610922, "grad_norm": 0.41899481415748596, "learning_rate": 7.529550582611883e-07, "loss": 0.4235, "step": 7882 }, { "epoch": 2.5222909556313993, "grad_norm": 0.47802966833114624, "learning_rate": 7.519728181527997e-07, "loss": 0.4436, "step": 7883 }, { "epoch": 2.5226109215017063, "grad_norm": 0.5048873424530029, "learning_rate": 7.5099116703849e-07, "loss": 0.4518, "step": 7884 }, { "epoch": 2.5229308873720138, "grad_norm": 0.43372148275375366, "learning_rate": 7.500101050543657e-07, "loss": 0.3671, "step": 7885 }, { "epoch": 2.523250853242321, "grad_norm": 0.49529823660850525, "learning_rate": 7.490296323364532e-07, "loss": 0.4827, "step": 7886 }, { "epoch": 2.523570819112628, "grad_norm": 0.49398353695869446, "learning_rate": 7.480497490206973e-07, "loss": 0.4172, "step": 7887 }, { "epoch": 2.5238907849829353, "grad_norm": 0.4616610109806061, "learning_rate": 7.470704552429592e-07, "loss": 0.4183, "step": 7888 }, { "epoch": 2.5242107508532423, "grad_norm": 0.532124936580658, "learning_rate": 7.460917511390186e-07, "loss": 0.4552, "step": 7889 }, { "epoch": 2.5245307167235493, "grad_norm": 0.4939459562301636, "learning_rate": 7.451136368445727e-07, "loss": 0.4066, "step": 7890 }, { "epoch": 2.524850682593857, "grad_norm": 0.49357783794403076, "learning_rate": 7.441361124952406e-07, "loss": 0.4387, "step": 7891 }, { "epoch": 2.525170648464164, "grad_norm": 0.48628488183021545, "learning_rate": 7.431591782265574e-07, "loss": 0.4396, "step": 7892 }, { "epoch": 2.525490614334471, "grad_norm": 0.48584863543510437, "learning_rate": 7.421828341739751e-07, "loss": 0.3749, "step": 7893 }, { "epoch": 2.5258105802047783, "grad_norm": 0.4655524492263794, "learning_rate": 7.41207080472865e-07, "loss": 0.4627, "step": 7894 }, { "epoch": 2.5261305460750854, "grad_norm": 0.540604829788208, "learning_rate": 7.402319172585154e-07, "loss": 0.4555, "step": 7895 }, { "epoch": 2.5264505119453924, "grad_norm": 0.4794887900352478, "learning_rate": 7.392573446661355e-07, "loss": 0.3782, "step": 7896 }, { "epoch": 2.5267704778157, "grad_norm": 0.4764775335788727, "learning_rate": 7.382833628308511e-07, "loss": 0.4542, "step": 7897 }, { "epoch": 2.527090443686007, "grad_norm": 0.4774472415447235, "learning_rate": 7.373099718877059e-07, "loss": 0.4158, "step": 7898 }, { "epoch": 2.527410409556314, "grad_norm": 0.4650486409664154, "learning_rate": 7.363371719716611e-07, "loss": 0.455, "step": 7899 }, { "epoch": 2.5277303754266214, "grad_norm": 0.5249002575874329, "learning_rate": 7.353649632175957e-07, "loss": 0.4391, "step": 7900 }, { "epoch": 2.5280503412969284, "grad_norm": 0.4526771605014801, "learning_rate": 7.343933457603086e-07, "loss": 0.4209, "step": 7901 }, { "epoch": 2.5283703071672354, "grad_norm": 0.4585656225681305, "learning_rate": 7.334223197345169e-07, "loss": 0.466, "step": 7902 }, { "epoch": 2.528690273037543, "grad_norm": 0.5290406942367554, "learning_rate": 7.324518852748535e-07, "loss": 0.4633, "step": 7903 }, { "epoch": 2.52901023890785, "grad_norm": 0.5417910218238831, "learning_rate": 7.314820425158687e-07, "loss": 0.466, "step": 7904 }, { "epoch": 2.529330204778157, "grad_norm": 0.5115374326705933, "learning_rate": 7.305127915920351e-07, "loss": 0.471, "step": 7905 }, { "epoch": 2.5296501706484644, "grad_norm": 0.46070459485054016, "learning_rate": 7.295441326377384e-07, "loss": 0.3598, "step": 7906 }, { "epoch": 2.5299701365187715, "grad_norm": 0.4890165328979492, "learning_rate": 7.285760657872859e-07, "loss": 0.5009, "step": 7907 }, { "epoch": 2.5302901023890785, "grad_norm": 0.4732837975025177, "learning_rate": 7.276085911749009e-07, "loss": 0.459, "step": 7908 }, { "epoch": 2.5306100682593855, "grad_norm": 0.47583815455436707, "learning_rate": 7.266417089347233e-07, "loss": 0.4212, "step": 7909 }, { "epoch": 2.530930034129693, "grad_norm": 0.48661455512046814, "learning_rate": 7.256754192008148e-07, "loss": 0.4319, "step": 7910 }, { "epoch": 2.53125, "grad_norm": 0.5234128832817078, "learning_rate": 7.247097221071509e-07, "loss": 0.4463, "step": 7911 }, { "epoch": 2.531569965870307, "grad_norm": 0.4903889000415802, "learning_rate": 7.237446177876278e-07, "loss": 0.375, "step": 7912 }, { "epoch": 2.5318899317406145, "grad_norm": 0.5294584631919861, "learning_rate": 7.227801063760581e-07, "loss": 0.4596, "step": 7913 }, { "epoch": 2.5322098976109215, "grad_norm": 0.4898788034915924, "learning_rate": 7.218161880061714e-07, "loss": 0.4073, "step": 7914 }, { "epoch": 2.5325298634812285, "grad_norm": 0.4349392056465149, "learning_rate": 7.208528628116179e-07, "loss": 0.4212, "step": 7915 }, { "epoch": 2.5328498293515356, "grad_norm": 0.4884548783302307, "learning_rate": 7.198901309259615e-07, "loss": 0.4469, "step": 7916 }, { "epoch": 2.533169795221843, "grad_norm": 0.5282070636749268, "learning_rate": 7.189279924826892e-07, "loss": 0.4284, "step": 7917 }, { "epoch": 2.53348976109215, "grad_norm": 0.5929446816444397, "learning_rate": 7.17966447615201e-07, "loss": 0.4968, "step": 7918 }, { "epoch": 2.533809726962457, "grad_norm": 0.5412628054618835, "learning_rate": 7.170054964568146e-07, "loss": 0.4404, "step": 7919 }, { "epoch": 2.5341296928327646, "grad_norm": 0.5096564292907715, "learning_rate": 7.160451391407691e-07, "loss": 0.5102, "step": 7920 }, { "epoch": 2.5344496587030716, "grad_norm": 0.46906861662864685, "learning_rate": 7.150853758002196e-07, "loss": 0.4025, "step": 7921 }, { "epoch": 2.5347696245733786, "grad_norm": 0.509339451789856, "learning_rate": 7.14126206568238e-07, "loss": 0.4543, "step": 7922 }, { "epoch": 2.535089590443686, "grad_norm": 0.44657838344573975, "learning_rate": 7.131676315778136e-07, "loss": 0.4469, "step": 7923 }, { "epoch": 2.535409556313993, "grad_norm": 0.4389392137527466, "learning_rate": 7.122096509618537e-07, "loss": 0.3871, "step": 7924 }, { "epoch": 2.5357295221843, "grad_norm": 0.48517391085624695, "learning_rate": 7.112522648531833e-07, "loss": 0.449, "step": 7925 }, { "epoch": 2.5360494880546076, "grad_norm": 0.49374955892562866, "learning_rate": 7.102954733845474e-07, "loss": 0.359, "step": 7926 }, { "epoch": 2.5363694539249146, "grad_norm": 0.5568493604660034, "learning_rate": 7.093392766886048e-07, "loss": 0.4509, "step": 7927 }, { "epoch": 2.5366894197952217, "grad_norm": 0.473632276058197, "learning_rate": 7.083836748979328e-07, "loss": 0.4456, "step": 7928 }, { "epoch": 2.537009385665529, "grad_norm": 0.4755510091781616, "learning_rate": 7.074286681450266e-07, "loss": 0.4893, "step": 7929 }, { "epoch": 2.537329351535836, "grad_norm": 0.4428599774837494, "learning_rate": 7.064742565622995e-07, "loss": 0.401, "step": 7930 }, { "epoch": 2.537649317406143, "grad_norm": 0.4869743287563324, "learning_rate": 7.055204402820831e-07, "loss": 0.4499, "step": 7931 }, { "epoch": 2.5379692832764507, "grad_norm": 0.44322407245635986, "learning_rate": 7.045672194366238e-07, "loss": 0.4087, "step": 7932 }, { "epoch": 2.5382892491467577, "grad_norm": 0.5037073493003845, "learning_rate": 7.036145941580874e-07, "loss": 0.4724, "step": 7933 }, { "epoch": 2.5386092150170647, "grad_norm": 0.5240101218223572, "learning_rate": 7.026625645785551e-07, "loss": 0.4887, "step": 7934 }, { "epoch": 2.538929180887372, "grad_norm": 0.46680012345314026, "learning_rate": 7.017111308300278e-07, "loss": 0.4437, "step": 7935 }, { "epoch": 2.539249146757679, "grad_norm": 0.45819762349128723, "learning_rate": 7.00760293044424e-07, "loss": 0.4034, "step": 7936 }, { "epoch": 2.5395691126279862, "grad_norm": 0.47192758321762085, "learning_rate": 6.998100513535782e-07, "loss": 0.4662, "step": 7937 }, { "epoch": 2.5398890784982937, "grad_norm": 0.44931477308273315, "learning_rate": 6.988604058892418e-07, "loss": 0.4233, "step": 7938 }, { "epoch": 2.5402090443686007, "grad_norm": 0.4655084013938904, "learning_rate": 6.979113567830837e-07, "loss": 0.3816, "step": 7939 }, { "epoch": 2.5405290102389078, "grad_norm": 0.5024960041046143, "learning_rate": 6.969629041666914e-07, "loss": 0.4849, "step": 7940 }, { "epoch": 2.5408489761092152, "grad_norm": 0.5373795032501221, "learning_rate": 6.9601504817157e-07, "loss": 0.427, "step": 7941 }, { "epoch": 2.5411689419795223, "grad_norm": 0.49928146600723267, "learning_rate": 6.950677889291402e-07, "loss": 0.4607, "step": 7942 }, { "epoch": 2.5414889078498293, "grad_norm": 0.5113884210586548, "learning_rate": 6.941211265707393e-07, "loss": 0.4202, "step": 7943 }, { "epoch": 2.5418088737201368, "grad_norm": 0.5413162112236023, "learning_rate": 6.93175061227625e-07, "loss": 0.4412, "step": 7944 }, { "epoch": 2.542128839590444, "grad_norm": 0.5174480080604553, "learning_rate": 6.922295930309691e-07, "loss": 0.4355, "step": 7945 }, { "epoch": 2.542448805460751, "grad_norm": 0.5377246141433716, "learning_rate": 6.912847221118635e-07, "loss": 0.4573, "step": 7946 }, { "epoch": 2.5427687713310583, "grad_norm": 0.5317425727844238, "learning_rate": 6.903404486013149e-07, "loss": 0.4932, "step": 7947 }, { "epoch": 2.5430887372013653, "grad_norm": 0.48629119992256165, "learning_rate": 6.893967726302464e-07, "loss": 0.3659, "step": 7948 }, { "epoch": 2.5434087030716723, "grad_norm": 0.4948200583457947, "learning_rate": 6.884536943295023e-07, "loss": 0.4538, "step": 7949 }, { "epoch": 2.54372866894198, "grad_norm": 0.47697150707244873, "learning_rate": 6.87511213829839e-07, "loss": 0.4669, "step": 7950 }, { "epoch": 2.544048634812287, "grad_norm": 0.5171055197715759, "learning_rate": 6.865693312619354e-07, "loss": 0.4483, "step": 7951 }, { "epoch": 2.544368600682594, "grad_norm": 0.46479135751724243, "learning_rate": 6.856280467563831e-07, "loss": 0.3814, "step": 7952 }, { "epoch": 2.544688566552901, "grad_norm": 0.5061951279640198, "learning_rate": 6.84687360443691e-07, "loss": 0.4685, "step": 7953 }, { "epoch": 2.5450085324232083, "grad_norm": 0.45566174387931824, "learning_rate": 6.837472724542893e-07, "loss": 0.4357, "step": 7954 }, { "epoch": 2.5453284982935154, "grad_norm": 0.44831565022468567, "learning_rate": 6.828077829185198e-07, "loss": 0.4058, "step": 7955 }, { "epoch": 2.5456484641638224, "grad_norm": 0.507569432258606, "learning_rate": 6.818688919666461e-07, "loss": 0.4659, "step": 7956 }, { "epoch": 2.5459684300341294, "grad_norm": 0.4822544455528259, "learning_rate": 6.809305997288451e-07, "loss": 0.4137, "step": 7957 }, { "epoch": 2.546288395904437, "grad_norm": 0.4946458339691162, "learning_rate": 6.799929063352112e-07, "loss": 0.419, "step": 7958 }, { "epoch": 2.546608361774744, "grad_norm": 0.5048835277557373, "learning_rate": 6.790558119157597e-07, "loss": 0.4896, "step": 7959 }, { "epoch": 2.546928327645051, "grad_norm": 0.4941258132457733, "learning_rate": 6.781193166004163e-07, "loss": 0.4414, "step": 7960 }, { "epoch": 2.5472482935153584, "grad_norm": 0.4280059039592743, "learning_rate": 6.771834205190303e-07, "loss": 0.4074, "step": 7961 }, { "epoch": 2.5475682593856654, "grad_norm": 0.5590117573738098, "learning_rate": 6.762481238013635e-07, "loss": 0.5315, "step": 7962 }, { "epoch": 2.5478882252559725, "grad_norm": 0.46806591749191284, "learning_rate": 6.75313426577095e-07, "loss": 0.3526, "step": 7963 }, { "epoch": 2.54820819112628, "grad_norm": 0.5423508882522583, "learning_rate": 6.743793289758238e-07, "loss": 0.4518, "step": 7964 }, { "epoch": 2.548528156996587, "grad_norm": 0.5363116264343262, "learning_rate": 6.734458311270609e-07, "loss": 0.4359, "step": 7965 }, { "epoch": 2.548848122866894, "grad_norm": 0.4925404489040375, "learning_rate": 6.725129331602398e-07, "loss": 0.4212, "step": 7966 }, { "epoch": 2.5491680887372015, "grad_norm": 0.5390374660491943, "learning_rate": 6.715806352047072e-07, "loss": 0.4659, "step": 7967 }, { "epoch": 2.5494880546075085, "grad_norm": 0.5575571060180664, "learning_rate": 6.706489373897251e-07, "loss": 0.4608, "step": 7968 }, { "epoch": 2.5498080204778155, "grad_norm": 0.45139390230178833, "learning_rate": 6.697178398444771e-07, "loss": 0.4012, "step": 7969 }, { "epoch": 2.550127986348123, "grad_norm": 0.49906036257743835, "learning_rate": 6.68787342698059e-07, "loss": 0.4613, "step": 7970 }, { "epoch": 2.55044795221843, "grad_norm": 0.5017614960670471, "learning_rate": 6.678574460794879e-07, "loss": 0.4129, "step": 7971 }, { "epoch": 2.550767918088737, "grad_norm": 0.48051193356513977, "learning_rate": 6.669281501176933e-07, "loss": 0.3836, "step": 7972 }, { "epoch": 2.5510878839590445, "grad_norm": 0.5058014392852783, "learning_rate": 6.659994549415227e-07, "loss": 0.5042, "step": 7973 }, { "epoch": 2.5514078498293515, "grad_norm": 0.4778083264827728, "learning_rate": 6.650713606797426e-07, "loss": 0.4889, "step": 7974 }, { "epoch": 2.5517278156996586, "grad_norm": 0.42727649211883545, "learning_rate": 6.641438674610318e-07, "loss": 0.3505, "step": 7975 }, { "epoch": 2.552047781569966, "grad_norm": 0.5494928359985352, "learning_rate": 6.632169754139917e-07, "loss": 0.5076, "step": 7976 }, { "epoch": 2.552367747440273, "grad_norm": 0.4783819913864136, "learning_rate": 6.622906846671339e-07, "loss": 0.4863, "step": 7977 }, { "epoch": 2.55268771331058, "grad_norm": 0.5169709324836731, "learning_rate": 6.613649953488921e-07, "loss": 0.4364, "step": 7978 }, { "epoch": 2.5530076791808876, "grad_norm": 0.475924551486969, "learning_rate": 6.604399075876133e-07, "loss": 0.406, "step": 7979 }, { "epoch": 2.5533276450511946, "grad_norm": 0.5216423869132996, "learning_rate": 6.595154215115607e-07, "loss": 0.4441, "step": 7980 }, { "epoch": 2.5536476109215016, "grad_norm": 0.516577959060669, "learning_rate": 6.585915372489177e-07, "loss": 0.4712, "step": 7981 }, { "epoch": 2.553967576791809, "grad_norm": 0.508666455745697, "learning_rate": 6.576682549277797e-07, "loss": 0.4248, "step": 7982 }, { "epoch": 2.554287542662116, "grad_norm": 0.5177512168884277, "learning_rate": 6.567455746761636e-07, "loss": 0.4926, "step": 7983 }, { "epoch": 2.554607508532423, "grad_norm": 0.4615013003349304, "learning_rate": 6.558234966219984e-07, "loss": 0.4214, "step": 7984 }, { "epoch": 2.5549274744027306, "grad_norm": 0.47998058795928955, "learning_rate": 6.549020208931306e-07, "loss": 0.4224, "step": 7985 }, { "epoch": 2.5552474402730376, "grad_norm": 0.4983801245689392, "learning_rate": 6.539811476173258e-07, "loss": 0.4008, "step": 7986 }, { "epoch": 2.5555674061433447, "grad_norm": 0.5446333885192871, "learning_rate": 6.530608769222624e-07, "loss": 0.4695, "step": 7987 }, { "epoch": 2.555887372013652, "grad_norm": 0.487994909286499, "learning_rate": 6.521412089355389e-07, "loss": 0.4271, "step": 7988 }, { "epoch": 2.556207337883959, "grad_norm": 0.4926753044128418, "learning_rate": 6.512221437846672e-07, "loss": 0.4037, "step": 7989 }, { "epoch": 2.556527303754266, "grad_norm": 0.4782487154006958, "learning_rate": 6.503036815970759e-07, "loss": 0.4544, "step": 7990 }, { "epoch": 2.5568472696245736, "grad_norm": 0.47307702898979187, "learning_rate": 6.493858225001131e-07, "loss": 0.456, "step": 7991 }, { "epoch": 2.5571672354948807, "grad_norm": 0.4964895248413086, "learning_rate": 6.484685666210389e-07, "loss": 0.4531, "step": 7992 }, { "epoch": 2.5574872013651877, "grad_norm": 0.4947316646575928, "learning_rate": 6.47551914087034e-07, "loss": 0.4128, "step": 7993 }, { "epoch": 2.5578071672354947, "grad_norm": 0.48979294300079346, "learning_rate": 6.466358650251925e-07, "loss": 0.4305, "step": 7994 }, { "epoch": 2.558127133105802, "grad_norm": 0.4775591790676117, "learning_rate": 6.457204195625244e-07, "loss": 0.4633, "step": 7995 }, { "epoch": 2.5584470989761092, "grad_norm": 0.45607683062553406, "learning_rate": 6.448055778259594e-07, "loss": 0.4346, "step": 7996 }, { "epoch": 2.5587670648464163, "grad_norm": 0.45880717039108276, "learning_rate": 6.438913399423396e-07, "loss": 0.461, "step": 7997 }, { "epoch": 2.5590870307167233, "grad_norm": 0.4997277855873108, "learning_rate": 6.429777060384268e-07, "loss": 0.3966, "step": 7998 }, { "epoch": 2.5594069965870307, "grad_norm": 0.527934193611145, "learning_rate": 6.420646762408972e-07, "loss": 0.4403, "step": 7999 }, { "epoch": 2.5597269624573378, "grad_norm": 0.5341289639472961, "learning_rate": 6.411522506763412e-07, "loss": 0.4561, "step": 8000 }, { "epoch": 2.560046928327645, "grad_norm": 0.47350558638572693, "learning_rate": 6.402404294712711e-07, "loss": 0.4201, "step": 8001 }, { "epoch": 2.5603668941979523, "grad_norm": 0.4907589554786682, "learning_rate": 6.393292127521089e-07, "loss": 0.4745, "step": 8002 }, { "epoch": 2.5606868600682593, "grad_norm": 0.4462029039859772, "learning_rate": 6.384186006451987e-07, "loss": 0.4005, "step": 8003 }, { "epoch": 2.5610068259385663, "grad_norm": 0.5270899534225464, "learning_rate": 6.375085932767966e-07, "loss": 0.4886, "step": 8004 }, { "epoch": 2.561326791808874, "grad_norm": 0.5296903252601624, "learning_rate": 6.365991907730751e-07, "loss": 0.4888, "step": 8005 }, { "epoch": 2.561646757679181, "grad_norm": 0.4909511208534241, "learning_rate": 6.35690393260126e-07, "loss": 0.4011, "step": 8006 }, { "epoch": 2.561966723549488, "grad_norm": 0.4494198262691498, "learning_rate": 6.347822008639537e-07, "loss": 0.4451, "step": 8007 }, { "epoch": 2.5622866894197953, "grad_norm": 0.5234202742576599, "learning_rate": 6.338746137104817e-07, "loss": 0.4932, "step": 8008 }, { "epoch": 2.5626066552901023, "grad_norm": 0.5346943736076355, "learning_rate": 6.329676319255468e-07, "loss": 0.4351, "step": 8009 }, { "epoch": 2.5629266211604094, "grad_norm": 0.46330562233924866, "learning_rate": 6.320612556349027e-07, "loss": 0.4454, "step": 8010 }, { "epoch": 2.563246587030717, "grad_norm": 0.4751325845718384, "learning_rate": 6.311554849642198e-07, "loss": 0.4381, "step": 8011 }, { "epoch": 2.563566552901024, "grad_norm": 0.5119494795799255, "learning_rate": 6.302503200390858e-07, "loss": 0.4228, "step": 8012 }, { "epoch": 2.563886518771331, "grad_norm": 0.5533508658409119, "learning_rate": 6.293457609850024e-07, "loss": 0.4703, "step": 8013 }, { "epoch": 2.5642064846416384, "grad_norm": 0.5752602219581604, "learning_rate": 6.284418079273869e-07, "loss": 0.441, "step": 8014 }, { "epoch": 2.5645264505119454, "grad_norm": 0.556894063949585, "learning_rate": 6.275384609915724e-07, "loss": 0.4381, "step": 8015 }, { "epoch": 2.5648464163822524, "grad_norm": 0.5390867590904236, "learning_rate": 6.266357203028111e-07, "loss": 0.4704, "step": 8016 }, { "epoch": 2.56516638225256, "grad_norm": 0.5242125988006592, "learning_rate": 6.257335859862696e-07, "loss": 0.4442, "step": 8017 }, { "epoch": 2.565486348122867, "grad_norm": 0.5321682691574097, "learning_rate": 6.248320581670281e-07, "loss": 0.4569, "step": 8018 }, { "epoch": 2.565806313993174, "grad_norm": 0.5042716264724731, "learning_rate": 6.239311369700857e-07, "loss": 0.4921, "step": 8019 }, { "epoch": 2.5661262798634814, "grad_norm": 0.5050718188285828, "learning_rate": 6.230308225203546e-07, "loss": 0.4014, "step": 8020 }, { "epoch": 2.5664462457337884, "grad_norm": 0.47061100602149963, "learning_rate": 6.221311149426657e-07, "loss": 0.3965, "step": 8021 }, { "epoch": 2.5667662116040955, "grad_norm": 0.4673781991004944, "learning_rate": 6.212320143617651e-07, "loss": 0.4455, "step": 8022 }, { "epoch": 2.567086177474403, "grad_norm": 0.45127198100090027, "learning_rate": 6.203335209023137e-07, "loss": 0.3997, "step": 8023 }, { "epoch": 2.56740614334471, "grad_norm": 0.463192880153656, "learning_rate": 6.194356346888885e-07, "loss": 0.4131, "step": 8024 }, { "epoch": 2.567726109215017, "grad_norm": 0.4891464114189148, "learning_rate": 6.185383558459818e-07, "loss": 0.4509, "step": 8025 }, { "epoch": 2.5680460750853245, "grad_norm": 0.5318314433097839, "learning_rate": 6.176416844980027e-07, "loss": 0.4168, "step": 8026 }, { "epoch": 2.5683660409556315, "grad_norm": 0.5327971577644348, "learning_rate": 6.167456207692768e-07, "loss": 0.4366, "step": 8027 }, { "epoch": 2.5686860068259385, "grad_norm": 0.4995075762271881, "learning_rate": 6.158501647840443e-07, "loss": 0.4467, "step": 8028 }, { "epoch": 2.569005972696246, "grad_norm": 0.49919721484184265, "learning_rate": 6.149553166664607e-07, "loss": 0.4718, "step": 8029 }, { "epoch": 2.569325938566553, "grad_norm": 0.48580458760261536, "learning_rate": 6.140610765405969e-07, "loss": 0.4002, "step": 8030 }, { "epoch": 2.56964590443686, "grad_norm": 0.5121810436248779, "learning_rate": 6.131674445304409e-07, "loss": 0.4802, "step": 8031 }, { "epoch": 2.5699658703071675, "grad_norm": 0.4804602563381195, "learning_rate": 6.122744207598974e-07, "loss": 0.4165, "step": 8032 }, { "epoch": 2.5702858361774745, "grad_norm": 0.5108646154403687, "learning_rate": 6.113820053527835e-07, "loss": 0.4041, "step": 8033 }, { "epoch": 2.5706058020477816, "grad_norm": 0.5154693126678467, "learning_rate": 6.104901984328348e-07, "loss": 0.4377, "step": 8034 }, { "epoch": 2.5709257679180886, "grad_norm": 0.5470184087753296, "learning_rate": 6.095990001236996e-07, "loss": 0.4881, "step": 8035 }, { "epoch": 2.571245733788396, "grad_norm": 0.47853484749794006, "learning_rate": 6.087084105489449e-07, "loss": 0.3808, "step": 8036 }, { "epoch": 2.571565699658703, "grad_norm": 0.43800899386405945, "learning_rate": 6.078184298320527e-07, "loss": 0.3837, "step": 8037 }, { "epoch": 2.57188566552901, "grad_norm": 0.47548708319664, "learning_rate": 6.069290580964194e-07, "loss": 0.4467, "step": 8038 }, { "epoch": 2.5722056313993176, "grad_norm": 0.49081090092658997, "learning_rate": 6.060402954653571e-07, "loss": 0.3955, "step": 8039 }, { "epoch": 2.5725255972696246, "grad_norm": 0.5398766398429871, "learning_rate": 6.051521420620932e-07, "loss": 0.4942, "step": 8040 }, { "epoch": 2.5728455631399316, "grad_norm": 0.47375577688217163, "learning_rate": 6.042645980097717e-07, "loss": 0.4444, "step": 8041 }, { "epoch": 2.5731655290102387, "grad_norm": 0.4815525710582733, "learning_rate": 6.033776634314525e-07, "loss": 0.3972, "step": 8042 }, { "epoch": 2.573485494880546, "grad_norm": 0.5203426480293274, "learning_rate": 6.024913384501097e-07, "loss": 0.4016, "step": 8043 }, { "epoch": 2.573805460750853, "grad_norm": 0.5383763909339905, "learning_rate": 6.016056231886336e-07, "loss": 0.4223, "step": 8044 }, { "epoch": 2.57412542662116, "grad_norm": 0.5123094916343689, "learning_rate": 6.007205177698277e-07, "loss": 0.4369, "step": 8045 }, { "epoch": 2.5744453924914676, "grad_norm": 0.4762115180492401, "learning_rate": 5.998360223164145e-07, "loss": 0.4577, "step": 8046 }, { "epoch": 2.5747653583617747, "grad_norm": 0.483458936214447, "learning_rate": 5.989521369510321e-07, "loss": 0.4668, "step": 8047 }, { "epoch": 2.5750853242320817, "grad_norm": 0.5269235372543335, "learning_rate": 5.980688617962299e-07, "loss": 0.4695, "step": 8048 }, { "epoch": 2.575405290102389, "grad_norm": 0.5068848729133606, "learning_rate": 5.971861969744758e-07, "loss": 0.4666, "step": 8049 }, { "epoch": 2.575725255972696, "grad_norm": 0.4762909710407257, "learning_rate": 5.963041426081511e-07, "loss": 0.3967, "step": 8050 }, { "epoch": 2.576045221843003, "grad_norm": 0.5501658320426941, "learning_rate": 5.954226988195544e-07, "loss": 0.502, "step": 8051 }, { "epoch": 2.5763651877133107, "grad_norm": 0.4227347671985626, "learning_rate": 5.94541865730901e-07, "loss": 0.3864, "step": 8052 }, { "epoch": 2.5766851535836177, "grad_norm": 0.48304030299186707, "learning_rate": 5.936616434643177e-07, "loss": 0.4318, "step": 8053 }, { "epoch": 2.5770051194539247, "grad_norm": 0.5073933601379395, "learning_rate": 5.927820321418482e-07, "loss": 0.4967, "step": 8054 }, { "epoch": 2.577325085324232, "grad_norm": 0.48393183946609497, "learning_rate": 5.919030318854513e-07, "loss": 0.4358, "step": 8055 }, { "epoch": 2.5776450511945392, "grad_norm": 0.48558008670806885, "learning_rate": 5.910246428170019e-07, "loss": 0.3956, "step": 8056 }, { "epoch": 2.5779650170648463, "grad_norm": 0.5119196176528931, "learning_rate": 5.901468650582908e-07, "loss": 0.4767, "step": 8057 }, { "epoch": 2.5782849829351537, "grad_norm": 0.5263835191726685, "learning_rate": 5.89269698731022e-07, "loss": 0.4206, "step": 8058 }, { "epoch": 2.5786049488054608, "grad_norm": 0.48206663131713867, "learning_rate": 5.883931439568163e-07, "loss": 0.4976, "step": 8059 }, { "epoch": 2.578924914675768, "grad_norm": 0.48202216625213623, "learning_rate": 5.875172008572055e-07, "loss": 0.4053, "step": 8060 }, { "epoch": 2.5792448805460753, "grad_norm": 0.5368192195892334, "learning_rate": 5.866418695536458e-07, "loss": 0.4666, "step": 8061 }, { "epoch": 2.5795648464163823, "grad_norm": 0.5510113835334778, "learning_rate": 5.857671501675005e-07, "loss": 0.3894, "step": 8062 }, { "epoch": 2.5798848122866893, "grad_norm": 0.594622790813446, "learning_rate": 5.8489304282005e-07, "loss": 0.5042, "step": 8063 }, { "epoch": 2.580204778156997, "grad_norm": 0.48127028346061707, "learning_rate": 5.840195476324906e-07, "loss": 0.3677, "step": 8064 }, { "epoch": 2.580524744027304, "grad_norm": 0.4667394459247589, "learning_rate": 5.831466647259315e-07, "loss": 0.4419, "step": 8065 }, { "epoch": 2.580844709897611, "grad_norm": 0.5215911865234375, "learning_rate": 5.822743942214026e-07, "loss": 0.4373, "step": 8066 }, { "epoch": 2.5811646757679183, "grad_norm": 0.4830494821071625, "learning_rate": 5.814027362398439e-07, "loss": 0.4105, "step": 8067 }, { "epoch": 2.5814846416382253, "grad_norm": 0.486611008644104, "learning_rate": 5.80531690902112e-07, "loss": 0.4652, "step": 8068 }, { "epoch": 2.5818046075085324, "grad_norm": 0.4940856397151947, "learning_rate": 5.796612583289773e-07, "loss": 0.4603, "step": 8069 }, { "epoch": 2.58212457337884, "grad_norm": 0.49336355924606323, "learning_rate": 5.787914386411248e-07, "loss": 0.4013, "step": 8070 }, { "epoch": 2.582444539249147, "grad_norm": 0.4824593663215637, "learning_rate": 5.779222319591604e-07, "loss": 0.4137, "step": 8071 }, { "epoch": 2.582764505119454, "grad_norm": 0.4997718930244446, "learning_rate": 5.770536384035985e-07, "loss": 0.4918, "step": 8072 }, { "epoch": 2.5830844709897613, "grad_norm": 0.46724408864974976, "learning_rate": 5.761856580948705e-07, "loss": 0.4103, "step": 8073 }, { "epoch": 2.5834044368600684, "grad_norm": 0.4812827408313751, "learning_rate": 5.753182911533228e-07, "loss": 0.4547, "step": 8074 }, { "epoch": 2.5837244027303754, "grad_norm": 0.5146414041519165, "learning_rate": 5.744515376992155e-07, "loss": 0.4209, "step": 8075 }, { "epoch": 2.5840443686006824, "grad_norm": 0.5635401010513306, "learning_rate": 5.735853978527284e-07, "loss": 0.4415, "step": 8076 }, { "epoch": 2.58436433447099, "grad_norm": 0.5143701434135437, "learning_rate": 5.727198717339511e-07, "loss": 0.4301, "step": 8077 }, { "epoch": 2.584684300341297, "grad_norm": 0.4612678587436676, "learning_rate": 5.718549594628892e-07, "loss": 0.4266, "step": 8078 }, { "epoch": 2.585004266211604, "grad_norm": 0.5282837748527527, "learning_rate": 5.70990661159464e-07, "loss": 0.4799, "step": 8079 }, { "epoch": 2.5853242320819114, "grad_norm": 0.4639507234096527, "learning_rate": 5.701269769435113e-07, "loss": 0.4183, "step": 8080 }, { "epoch": 2.5856441979522184, "grad_norm": 0.5629203915596008, "learning_rate": 5.692639069347838e-07, "loss": 0.3848, "step": 8081 }, { "epoch": 2.5859641638225255, "grad_norm": 0.5728508830070496, "learning_rate": 5.684014512529462e-07, "loss": 0.4767, "step": 8082 }, { "epoch": 2.5862841296928325, "grad_norm": 0.5335341691970825, "learning_rate": 5.675396100175784e-07, "loss": 0.4757, "step": 8083 }, { "epoch": 2.58660409556314, "grad_norm": 0.5024179220199585, "learning_rate": 5.666783833481749e-07, "loss": 0.4289, "step": 8084 }, { "epoch": 2.586924061433447, "grad_norm": 0.49825990200042725, "learning_rate": 5.658177713641471e-07, "loss": 0.4189, "step": 8085 }, { "epoch": 2.587244027303754, "grad_norm": 0.45117056369781494, "learning_rate": 5.649577741848211e-07, "loss": 0.3969, "step": 8086 }, { "epoch": 2.5875639931740615, "grad_norm": 0.4884481132030487, "learning_rate": 5.640983919294347e-07, "loss": 0.44, "step": 8087 }, { "epoch": 2.5878839590443685, "grad_norm": 0.5033963918685913, "learning_rate": 5.632396247171429e-07, "loss": 0.3851, "step": 8088 }, { "epoch": 2.5882039249146755, "grad_norm": 0.5034910440444946, "learning_rate": 5.62381472667014e-07, "loss": 0.4661, "step": 8089 }, { "epoch": 2.588523890784983, "grad_norm": 0.5116603970527649, "learning_rate": 5.615239358980318e-07, "loss": 0.4745, "step": 8090 }, { "epoch": 2.58884385665529, "grad_norm": 0.507566511631012, "learning_rate": 5.606670145290971e-07, "loss": 0.4319, "step": 8091 }, { "epoch": 2.589163822525597, "grad_norm": 0.4920852482318878, "learning_rate": 5.598107086790216e-07, "loss": 0.4434, "step": 8092 }, { "epoch": 2.5894837883959045, "grad_norm": 0.4991411864757538, "learning_rate": 5.589550184665327e-07, "loss": 0.4522, "step": 8093 }, { "epoch": 2.5898037542662116, "grad_norm": 0.47811561822891235, "learning_rate": 5.580999440102719e-07, "loss": 0.4392, "step": 8094 }, { "epoch": 2.5901237201365186, "grad_norm": 0.48993173241615295, "learning_rate": 5.572454854287984e-07, "loss": 0.4458, "step": 8095 }, { "epoch": 2.590443686006826, "grad_norm": 0.4690512716770172, "learning_rate": 5.563916428405836e-07, "loss": 0.3775, "step": 8096 }, { "epoch": 2.590763651877133, "grad_norm": 0.5116325616836548, "learning_rate": 5.555384163640132e-07, "loss": 0.4414, "step": 8097 }, { "epoch": 2.59108361774744, "grad_norm": 0.534573495388031, "learning_rate": 5.546858061173882e-07, "loss": 0.486, "step": 8098 }, { "epoch": 2.5914035836177476, "grad_norm": 0.5088597536087036, "learning_rate": 5.538338122189235e-07, "loss": 0.4007, "step": 8099 }, { "epoch": 2.5917235494880546, "grad_norm": 0.49535033106803894, "learning_rate": 5.52982434786749e-07, "loss": 0.4555, "step": 8100 }, { "epoch": 2.5920435153583616, "grad_norm": 0.48237720131874084, "learning_rate": 5.521316739389116e-07, "loss": 0.4199, "step": 8101 }, { "epoch": 2.592363481228669, "grad_norm": 0.4965192675590515, "learning_rate": 5.512815297933677e-07, "loss": 0.5094, "step": 8102 }, { "epoch": 2.592683447098976, "grad_norm": 0.4722084403038025, "learning_rate": 5.504320024679922e-07, "loss": 0.379, "step": 8103 }, { "epoch": 2.593003412969283, "grad_norm": 0.5180158615112305, "learning_rate": 5.495830920805711e-07, "loss": 0.4783, "step": 8104 }, { "epoch": 2.5933233788395906, "grad_norm": 0.4634142220020294, "learning_rate": 5.487347987488084e-07, "loss": 0.394, "step": 8105 }, { "epoch": 2.5936433447098977, "grad_norm": 0.5356070399284363, "learning_rate": 5.478871225903221e-07, "loss": 0.4783, "step": 8106 }, { "epoch": 2.5939633105802047, "grad_norm": 0.5005761981010437, "learning_rate": 5.47040063722642e-07, "loss": 0.4703, "step": 8107 }, { "epoch": 2.594283276450512, "grad_norm": 0.48422369360923767, "learning_rate": 5.461936222632141e-07, "loss": 0.4207, "step": 8108 }, { "epoch": 2.594603242320819, "grad_norm": 0.5311996936798096, "learning_rate": 5.453477983293975e-07, "loss": 0.4514, "step": 8109 }, { "epoch": 2.594923208191126, "grad_norm": 0.5696889758110046, "learning_rate": 5.445025920384678e-07, "loss": 0.4826, "step": 8110 }, { "epoch": 2.5952431740614337, "grad_norm": 0.5132286548614502, "learning_rate": 5.43658003507615e-07, "loss": 0.4557, "step": 8111 }, { "epoch": 2.5955631399317407, "grad_norm": 0.46289798617362976, "learning_rate": 5.428140328539411e-07, "loss": 0.3898, "step": 8112 }, { "epoch": 2.5958831058020477, "grad_norm": 0.5173072218894958, "learning_rate": 5.419706801944635e-07, "loss": 0.4478, "step": 8113 }, { "epoch": 2.596203071672355, "grad_norm": 0.5169364809989929, "learning_rate": 5.411279456461133e-07, "loss": 0.4411, "step": 8114 }, { "epoch": 2.5965230375426622, "grad_norm": 0.5261008143424988, "learning_rate": 5.402858293257379e-07, "loss": 0.4669, "step": 8115 }, { "epoch": 2.5968430034129693, "grad_norm": 0.4407719671726227, "learning_rate": 5.39444331350098e-07, "loss": 0.4142, "step": 8116 }, { "epoch": 2.5971629692832767, "grad_norm": 0.4699857831001282, "learning_rate": 5.386034518358679e-07, "loss": 0.4539, "step": 8117 }, { "epoch": 2.5974829351535837, "grad_norm": 0.4793130159378052, "learning_rate": 5.37763190899635e-07, "loss": 0.4232, "step": 8118 }, { "epoch": 2.5978029010238908, "grad_norm": 0.48637887835502625, "learning_rate": 5.369235486579055e-07, "loss": 0.469, "step": 8119 }, { "epoch": 2.598122866894198, "grad_norm": 0.4321134090423584, "learning_rate": 5.360845252270941e-07, "loss": 0.4343, "step": 8120 }, { "epoch": 2.5984428327645053, "grad_norm": 0.4490710198879242, "learning_rate": 5.352461207235344e-07, "loss": 0.4117, "step": 8121 }, { "epoch": 2.5987627986348123, "grad_norm": 0.5368947386741638, "learning_rate": 5.344083352634711e-07, "loss": 0.4582, "step": 8122 }, { "epoch": 2.5990827645051193, "grad_norm": 0.49436429142951965, "learning_rate": 5.335711689630635e-07, "loss": 0.4019, "step": 8123 }, { "epoch": 2.5994027303754264, "grad_norm": 0.505104124546051, "learning_rate": 5.327346219383872e-07, "loss": 0.4713, "step": 8124 }, { "epoch": 2.599722696245734, "grad_norm": 0.43805694580078125, "learning_rate": 5.318986943054294e-07, "loss": 0.4003, "step": 8125 }, { "epoch": 2.600042662116041, "grad_norm": 0.48276111483573914, "learning_rate": 5.310633861800934e-07, "loss": 0.4441, "step": 8126 }, { "epoch": 2.600362627986348, "grad_norm": 0.4786250591278076, "learning_rate": 5.302286976781956e-07, "loss": 0.4285, "step": 8127 }, { "epoch": 2.6006825938566553, "grad_norm": 0.49598589539527893, "learning_rate": 5.29394628915465e-07, "loss": 0.4733, "step": 8128 }, { "epoch": 2.6010025597269624, "grad_norm": 0.5020879507064819, "learning_rate": 5.285611800075479e-07, "loss": 0.4614, "step": 8129 }, { "epoch": 2.6013225255972694, "grad_norm": 0.4940076172351837, "learning_rate": 5.277283510700021e-07, "loss": 0.4899, "step": 8130 }, { "epoch": 2.601642491467577, "grad_norm": 0.49958762526512146, "learning_rate": 5.268961422183016e-07, "loss": 0.4297, "step": 8131 }, { "epoch": 2.601962457337884, "grad_norm": 0.5203652381896973, "learning_rate": 5.260645535678321e-07, "loss": 0.4593, "step": 8132 }, { "epoch": 2.602282423208191, "grad_norm": 0.47786155343055725, "learning_rate": 5.252335852338936e-07, "loss": 0.4475, "step": 8133 }, { "epoch": 2.6026023890784984, "grad_norm": 0.49558225274086, "learning_rate": 5.244032373317027e-07, "loss": 0.4031, "step": 8134 }, { "epoch": 2.6029223549488054, "grad_norm": 0.562053382396698, "learning_rate": 5.235735099763861e-07, "loss": 0.4867, "step": 8135 }, { "epoch": 2.6032423208191124, "grad_norm": 0.46849918365478516, "learning_rate": 5.227444032829887e-07, "loss": 0.4359, "step": 8136 }, { "epoch": 2.60356228668942, "grad_norm": 0.5348246693611145, "learning_rate": 5.219159173664662e-07, "loss": 0.4731, "step": 8137 }, { "epoch": 2.603882252559727, "grad_norm": 0.48375827074050903, "learning_rate": 5.210880523416884e-07, "loss": 0.4579, "step": 8138 }, { "epoch": 2.604202218430034, "grad_norm": 0.4847160577774048, "learning_rate": 5.202608083234406e-07, "loss": 0.4329, "step": 8139 }, { "epoch": 2.6045221843003414, "grad_norm": 0.4918648898601532, "learning_rate": 5.194341854264206e-07, "loss": 0.3665, "step": 8140 }, { "epoch": 2.6048421501706485, "grad_norm": 0.5149650573730469, "learning_rate": 5.186081837652418e-07, "loss": 0.5078, "step": 8141 }, { "epoch": 2.6051621160409555, "grad_norm": 0.45677924156188965, "learning_rate": 5.177828034544291e-07, "loss": 0.3819, "step": 8142 }, { "epoch": 2.605482081911263, "grad_norm": 0.475281298160553, "learning_rate": 5.169580446084226e-07, "loss": 0.3766, "step": 8143 }, { "epoch": 2.60580204778157, "grad_norm": 0.5269748568534851, "learning_rate": 5.161339073415772e-07, "loss": 0.5267, "step": 8144 }, { "epoch": 2.606122013651877, "grad_norm": 0.459114670753479, "learning_rate": 5.15310391768159e-07, "loss": 0.3896, "step": 8145 }, { "epoch": 2.6064419795221845, "grad_norm": 0.5193907022476196, "learning_rate": 5.144874980023512e-07, "loss": 0.4628, "step": 8146 }, { "epoch": 2.6067619453924915, "grad_norm": 0.5411141514778137, "learning_rate": 5.136652261582476e-07, "loss": 0.4008, "step": 8147 }, { "epoch": 2.6070819112627985, "grad_norm": 0.5180327296257019, "learning_rate": 5.128435763498568e-07, "loss": 0.4461, "step": 8148 }, { "epoch": 2.607401877133106, "grad_norm": 0.5204225778579712, "learning_rate": 5.120225486911035e-07, "loss": 0.4305, "step": 8149 }, { "epoch": 2.607721843003413, "grad_norm": 0.526789665222168, "learning_rate": 5.112021432958214e-07, "loss": 0.4612, "step": 8150 }, { "epoch": 2.60804180887372, "grad_norm": 0.47868508100509644, "learning_rate": 5.103823602777641e-07, "loss": 0.3933, "step": 8151 }, { "epoch": 2.6083617747440275, "grad_norm": 0.5638915300369263, "learning_rate": 5.095631997505923e-07, "loss": 0.4622, "step": 8152 }, { "epoch": 2.6086817406143346, "grad_norm": 0.48375749588012695, "learning_rate": 5.087446618278858e-07, "loss": 0.4052, "step": 8153 }, { "epoch": 2.6090017064846416, "grad_norm": 0.4716733396053314, "learning_rate": 5.079267466231352e-07, "loss": 0.4573, "step": 8154 }, { "epoch": 2.609321672354949, "grad_norm": 0.4670661687850952, "learning_rate": 5.07109454249744e-07, "loss": 0.4325, "step": 8155 }, { "epoch": 2.609641638225256, "grad_norm": 0.47152283787727356, "learning_rate": 5.062927848210336e-07, "loss": 0.3957, "step": 8156 }, { "epoch": 2.609961604095563, "grad_norm": 0.47465530037879944, "learning_rate": 5.054767384502335e-07, "loss": 0.4826, "step": 8157 }, { "epoch": 2.6102815699658706, "grad_norm": 0.5024846792221069, "learning_rate": 5.046613152504914e-07, "loss": 0.4282, "step": 8158 }, { "epoch": 2.6106015358361776, "grad_norm": 0.4924025237560272, "learning_rate": 5.038465153348659e-07, "loss": 0.4253, "step": 8159 }, { "epoch": 2.6109215017064846, "grad_norm": 0.49057185649871826, "learning_rate": 5.030323388163294e-07, "loss": 0.4388, "step": 8160 }, { "epoch": 2.6112414675767917, "grad_norm": 0.5156389474868774, "learning_rate": 5.022187858077698e-07, "loss": 0.4316, "step": 8161 }, { "epoch": 2.611561433447099, "grad_norm": 0.49020588397979736, "learning_rate": 5.014058564219859e-07, "loss": 0.4459, "step": 8162 }, { "epoch": 2.611881399317406, "grad_norm": 0.529273509979248, "learning_rate": 5.00593550771693e-07, "loss": 0.4647, "step": 8163 }, { "epoch": 2.612201365187713, "grad_norm": 0.5657989382743835, "learning_rate": 4.997818689695166e-07, "loss": 0.4385, "step": 8164 }, { "epoch": 2.61252133105802, "grad_norm": 0.4963492155075073, "learning_rate": 4.989708111279973e-07, "loss": 0.4139, "step": 8165 }, { "epoch": 2.6128412969283277, "grad_norm": 0.4489327073097229, "learning_rate": 4.981603773595911e-07, "loss": 0.3985, "step": 8166 }, { "epoch": 2.6131612627986347, "grad_norm": 0.49600255489349365, "learning_rate": 4.973505677766633e-07, "loss": 0.4416, "step": 8167 }, { "epoch": 2.6134812286689417, "grad_norm": 0.5063232779502869, "learning_rate": 4.965413824914966e-07, "loss": 0.4359, "step": 8168 }, { "epoch": 2.613801194539249, "grad_norm": 0.5594180822372437, "learning_rate": 4.95732821616286e-07, "loss": 0.3855, "step": 8169 }, { "epoch": 2.614121160409556, "grad_norm": 0.5148386359214783, "learning_rate": 4.949248852631367e-07, "loss": 0.4629, "step": 8170 }, { "epoch": 2.6144411262798632, "grad_norm": 0.4605209529399872, "learning_rate": 4.941175735440728e-07, "loss": 0.3764, "step": 8171 }, { "epoch": 2.6147610921501707, "grad_norm": 0.5135962963104248, "learning_rate": 4.933108865710274e-07, "loss": 0.4638, "step": 8172 }, { "epoch": 2.6150810580204777, "grad_norm": 0.5130895376205444, "learning_rate": 4.9250482445585e-07, "loss": 0.4997, "step": 8173 }, { "epoch": 2.6154010238907848, "grad_norm": 0.48560479283332825, "learning_rate": 4.916993873103015e-07, "loss": 0.3539, "step": 8174 }, { "epoch": 2.6157209897610922, "grad_norm": 0.55850750207901, "learning_rate": 4.908945752460553e-07, "loss": 0.5305, "step": 8175 }, { "epoch": 2.6160409556313993, "grad_norm": 0.5068272352218628, "learning_rate": 4.900903883747021e-07, "loss": 0.4184, "step": 8176 }, { "epoch": 2.6163609215017063, "grad_norm": 0.4663023352622986, "learning_rate": 4.892868268077411e-07, "loss": 0.3963, "step": 8177 }, { "epoch": 2.6166808873720138, "grad_norm": 0.5439704060554504, "learning_rate": 4.884838906565892e-07, "loss": 0.4643, "step": 8178 }, { "epoch": 2.617000853242321, "grad_norm": 0.5312791466712952, "learning_rate": 4.87681580032573e-07, "loss": 0.4374, "step": 8179 }, { "epoch": 2.617320819112628, "grad_norm": 0.5001805424690247, "learning_rate": 4.868798950469333e-07, "loss": 0.4276, "step": 8180 }, { "epoch": 2.6176407849829353, "grad_norm": 0.48276445269584656, "learning_rate": 4.860788358108265e-07, "loss": 0.4505, "step": 8181 }, { "epoch": 2.6179607508532423, "grad_norm": 0.5378854870796204, "learning_rate": 4.852784024353185e-07, "loss": 0.4857, "step": 8182 }, { "epoch": 2.6182807167235493, "grad_norm": 0.47105562686920166, "learning_rate": 4.844785950313924e-07, "loss": 0.4303, "step": 8183 }, { "epoch": 2.618600682593857, "grad_norm": 0.5040062069892883, "learning_rate": 4.836794137099415e-07, "loss": 0.4188, "step": 8184 }, { "epoch": 2.618920648464164, "grad_norm": 0.49857619404792786, "learning_rate": 4.828808585817724e-07, "loss": 0.4326, "step": 8185 }, { "epoch": 2.619240614334471, "grad_norm": 0.4615936875343323, "learning_rate": 4.820829297576068e-07, "loss": 0.4422, "step": 8186 }, { "epoch": 2.6195605802047783, "grad_norm": 0.48037397861480713, "learning_rate": 4.812856273480776e-07, "loss": 0.3965, "step": 8187 }, { "epoch": 2.6198805460750854, "grad_norm": 0.4728635549545288, "learning_rate": 4.804889514637334e-07, "loss": 0.4579, "step": 8188 }, { "epoch": 2.6202005119453924, "grad_norm": 0.47411414980888367, "learning_rate": 4.796929022150332e-07, "loss": 0.4767, "step": 8189 }, { "epoch": 2.6205204778157, "grad_norm": 0.4786132276058197, "learning_rate": 4.788974797123491e-07, "loss": 0.4402, "step": 8190 }, { "epoch": 2.620840443686007, "grad_norm": 0.5201287269592285, "learning_rate": 4.78102684065968e-07, "loss": 0.4799, "step": 8191 }, { "epoch": 2.621160409556314, "grad_norm": 0.44401898980140686, "learning_rate": 4.773085153860912e-07, "loss": 0.3678, "step": 8192 }, { "epoch": 2.6214803754266214, "grad_norm": 0.49735119938850403, "learning_rate": 4.765149737828295e-07, "loss": 0.4713, "step": 8193 }, { "epoch": 2.6218003412969284, "grad_norm": 0.5207501649856567, "learning_rate": 4.7572205936620874e-07, "loss": 0.4354, "step": 8194 }, { "epoch": 2.6221203071672354, "grad_norm": 0.5102244019508362, "learning_rate": 4.74929772246166e-07, "loss": 0.4234, "step": 8195 }, { "epoch": 2.622440273037543, "grad_norm": 0.49887150526046753, "learning_rate": 4.7413811253255446e-07, "loss": 0.4551, "step": 8196 }, { "epoch": 2.62276023890785, "grad_norm": 0.5083227157592773, "learning_rate": 4.733470803351392e-07, "loss": 0.4755, "step": 8197 }, { "epoch": 2.623080204778157, "grad_norm": 0.49517136812210083, "learning_rate": 4.7255667576359687e-07, "loss": 0.3996, "step": 8198 }, { "epoch": 2.6234001706484644, "grad_norm": 0.4763628840446472, "learning_rate": 4.717668989275187e-07, "loss": 0.4546, "step": 8199 }, { "epoch": 2.6237201365187715, "grad_norm": 0.4698621332645416, "learning_rate": 4.70977749936406e-07, "loss": 0.4112, "step": 8200 }, { "epoch": 2.6240401023890785, "grad_norm": 0.5033742189407349, "learning_rate": 4.701892288996768e-07, "loss": 0.4685, "step": 8201 }, { "epoch": 2.6243600682593855, "grad_norm": 0.4517287015914917, "learning_rate": 4.694013359266614e-07, "loss": 0.3929, "step": 8202 }, { "epoch": 2.624680034129693, "grad_norm": 0.5576598048210144, "learning_rate": 4.6861407112660184e-07, "loss": 0.5224, "step": 8203 }, { "epoch": 2.625, "grad_norm": 0.4753642976284027, "learning_rate": 4.6782743460865197e-07, "loss": 0.381, "step": 8204 }, { "epoch": 2.625319965870307, "grad_norm": 0.49122047424316406, "learning_rate": 4.6704142648188013e-07, "loss": 0.4727, "step": 8205 }, { "epoch": 2.6256399317406145, "grad_norm": 0.4982339143753052, "learning_rate": 4.662560468552674e-07, "loss": 0.4416, "step": 8206 }, { "epoch": 2.6259598976109215, "grad_norm": 0.5182000398635864, "learning_rate": 4.6547129583770943e-07, "loss": 0.4386, "step": 8207 }, { "epoch": 2.6262798634812285, "grad_norm": 0.5248851180076599, "learning_rate": 4.6468717353801095e-07, "loss": 0.3901, "step": 8208 }, { "epoch": 2.6265998293515356, "grad_norm": 0.5372112393379211, "learning_rate": 4.639036800648927e-07, "loss": 0.4414, "step": 8209 }, { "epoch": 2.626919795221843, "grad_norm": 0.5512729287147522, "learning_rate": 4.6312081552698454e-07, "loss": 0.4765, "step": 8210 }, { "epoch": 2.62723976109215, "grad_norm": 0.4698147177696228, "learning_rate": 4.6233858003283394e-07, "loss": 0.3375, "step": 8211 }, { "epoch": 2.627559726962457, "grad_norm": 0.520575761795044, "learning_rate": 4.6155697369089867e-07, "loss": 0.5129, "step": 8212 }, { "epoch": 2.6278796928327646, "grad_norm": 0.46919360756874084, "learning_rate": 4.607759966095493e-07, "loss": 0.4307, "step": 8213 }, { "epoch": 2.6281996587030716, "grad_norm": 0.5309707522392273, "learning_rate": 4.5999564889706806e-07, "loss": 0.5125, "step": 8214 }, { "epoch": 2.6285196245733786, "grad_norm": 0.4674423038959503, "learning_rate": 4.592159306616517e-07, "loss": 0.4261, "step": 8215 }, { "epoch": 2.628839590443686, "grad_norm": 0.5125669240951538, "learning_rate": 4.584368420114088e-07, "loss": 0.4309, "step": 8216 }, { "epoch": 2.629159556313993, "grad_norm": 0.553999662399292, "learning_rate": 4.576583830543624e-07, "loss": 0.48, "step": 8217 }, { "epoch": 2.6294795221843, "grad_norm": 0.5225322842597961, "learning_rate": 4.568805538984461e-07, "loss": 0.4011, "step": 8218 }, { "epoch": 2.6297994880546076, "grad_norm": 0.5288572907447815, "learning_rate": 4.5610335465150535e-07, "loss": 0.4341, "step": 8219 }, { "epoch": 2.6301194539249146, "grad_norm": 0.4576762914657593, "learning_rate": 4.553267854213017e-07, "loss": 0.3901, "step": 8220 }, { "epoch": 2.6304394197952217, "grad_norm": 0.46657612919807434, "learning_rate": 4.5455084631550575e-07, "loss": 0.4513, "step": 8221 }, { "epoch": 2.630759385665529, "grad_norm": 0.4978737235069275, "learning_rate": 4.537755374417041e-07, "loss": 0.3943, "step": 8222 }, { "epoch": 2.631079351535836, "grad_norm": 0.4882107973098755, "learning_rate": 4.5300085890739366e-07, "loss": 0.4666, "step": 8223 }, { "epoch": 2.631399317406143, "grad_norm": 0.5072929859161377, "learning_rate": 4.522268108199834e-07, "loss": 0.4409, "step": 8224 }, { "epoch": 2.6317192832764507, "grad_norm": 0.5407156944274902, "learning_rate": 4.5145339328679705e-07, "loss": 0.4459, "step": 8225 }, { "epoch": 2.6320392491467577, "grad_norm": 0.44135379791259766, "learning_rate": 4.506806064150704e-07, "loss": 0.4064, "step": 8226 }, { "epoch": 2.6323592150170647, "grad_norm": 0.4765286147594452, "learning_rate": 4.499084503119505e-07, "loss": 0.4493, "step": 8227 }, { "epoch": 2.632679180887372, "grad_norm": 0.5037778615951538, "learning_rate": 4.491369250844979e-07, "loss": 0.4995, "step": 8228 }, { "epoch": 2.632999146757679, "grad_norm": 0.45145121216773987, "learning_rate": 4.4836603083968476e-07, "loss": 0.4353, "step": 8229 }, { "epoch": 2.6333191126279862, "grad_norm": 0.48700815439224243, "learning_rate": 4.475957676843973e-07, "loss": 0.3971, "step": 8230 }, { "epoch": 2.6336390784982937, "grad_norm": 0.552137017250061, "learning_rate": 4.468261357254339e-07, "loss": 0.4473, "step": 8231 }, { "epoch": 2.6339590443686007, "grad_norm": 0.4934519827365875, "learning_rate": 4.4605713506950477e-07, "loss": 0.4481, "step": 8232 }, { "epoch": 2.6342790102389078, "grad_norm": 0.46769335865974426, "learning_rate": 4.452887658232319e-07, "loss": 0.3953, "step": 8233 }, { "epoch": 2.6345989761092152, "grad_norm": 0.49531757831573486, "learning_rate": 4.445210280931506e-07, "loss": 0.3934, "step": 8234 }, { "epoch": 2.6349189419795223, "grad_norm": 0.5289334654808044, "learning_rate": 4.43753921985709e-07, "loss": 0.4995, "step": 8235 }, { "epoch": 2.6352389078498293, "grad_norm": 0.4507151246070862, "learning_rate": 4.429874476072682e-07, "loss": 0.352, "step": 8236 }, { "epoch": 2.6355588737201368, "grad_norm": 0.4960966408252716, "learning_rate": 4.422216050641004e-07, "loss": 0.4657, "step": 8237 }, { "epoch": 2.635878839590444, "grad_norm": 0.5034002661705017, "learning_rate": 4.4145639446238955e-07, "loss": 0.4416, "step": 8238 }, { "epoch": 2.636198805460751, "grad_norm": 0.48766377568244934, "learning_rate": 4.406918159082335e-07, "loss": 0.4567, "step": 8239 }, { "epoch": 2.6365187713310583, "grad_norm": 0.44892147183418274, "learning_rate": 4.3992786950764145e-07, "loss": 0.3961, "step": 8240 }, { "epoch": 2.6368387372013653, "grad_norm": 0.48189565539360046, "learning_rate": 4.3916455536653744e-07, "loss": 0.4709, "step": 8241 }, { "epoch": 2.6371587030716723, "grad_norm": 0.4697919189929962, "learning_rate": 4.384018735907547e-07, "loss": 0.4415, "step": 8242 }, { "epoch": 2.63747866894198, "grad_norm": 0.48356616497039795, "learning_rate": 4.376398242860397e-07, "loss": 0.4241, "step": 8243 }, { "epoch": 2.637798634812287, "grad_norm": 0.5298052430152893, "learning_rate": 4.368784075580512e-07, "loss": 0.4637, "step": 8244 }, { "epoch": 2.638118600682594, "grad_norm": 0.4812069535255432, "learning_rate": 4.36117623512361e-07, "loss": 0.3976, "step": 8245 }, { "epoch": 2.638438566552901, "grad_norm": 0.5246396064758301, "learning_rate": 4.3535747225445345e-07, "loss": 0.4488, "step": 8246 }, { "epoch": 2.6387585324232083, "grad_norm": 0.50551438331604, "learning_rate": 4.345979538897238e-07, "loss": 0.4682, "step": 8247 }, { "epoch": 2.6390784982935154, "grad_norm": 0.527849018573761, "learning_rate": 4.3383906852348e-07, "loss": 0.4303, "step": 8248 }, { "epoch": 2.6393984641638224, "grad_norm": 0.5653607845306396, "learning_rate": 4.330808162609418e-07, "loss": 0.4698, "step": 8249 }, { "epoch": 2.6397184300341294, "grad_norm": 0.4829493761062622, "learning_rate": 4.323231972072428e-07, "loss": 0.4029, "step": 8250 }, { "epoch": 2.640038395904437, "grad_norm": 0.49808141589164734, "learning_rate": 4.315662114674285e-07, "loss": 0.3898, "step": 8251 }, { "epoch": 2.640358361774744, "grad_norm": 0.5228855013847351, "learning_rate": 4.308098591464549e-07, "loss": 0.4504, "step": 8252 }, { "epoch": 2.640678327645051, "grad_norm": 0.4826861321926117, "learning_rate": 4.300541403491909e-07, "loss": 0.3995, "step": 8253 }, { "epoch": 2.6409982935153584, "grad_norm": 0.4797203242778778, "learning_rate": 4.2929905518041714e-07, "loss": 0.4093, "step": 8254 }, { "epoch": 2.6413182593856654, "grad_norm": 0.494663268327713, "learning_rate": 4.285446037448282e-07, "loss": 0.4543, "step": 8255 }, { "epoch": 2.6416382252559725, "grad_norm": 0.4919332265853882, "learning_rate": 4.2779078614703093e-07, "loss": 0.499, "step": 8256 }, { "epoch": 2.64195819112628, "grad_norm": 0.4947248697280884, "learning_rate": 4.2703760249154124e-07, "loss": 0.4133, "step": 8257 }, { "epoch": 2.642278156996587, "grad_norm": 0.5169200301170349, "learning_rate": 4.262850528827894e-07, "loss": 0.4587, "step": 8258 }, { "epoch": 2.642598122866894, "grad_norm": 0.5032166838645935, "learning_rate": 4.255331374251159e-07, "loss": 0.4039, "step": 8259 }, { "epoch": 2.6429180887372015, "grad_norm": 0.555990993976593, "learning_rate": 4.247818562227768e-07, "loss": 0.4784, "step": 8260 }, { "epoch": 2.6432380546075085, "grad_norm": 0.5182940363883972, "learning_rate": 4.240312093799381e-07, "loss": 0.3817, "step": 8261 }, { "epoch": 2.6435580204778155, "grad_norm": 0.5078566670417786, "learning_rate": 4.2328119700067727e-07, "loss": 0.473, "step": 8262 }, { "epoch": 2.643877986348123, "grad_norm": 0.5581321716308594, "learning_rate": 4.225318191889843e-07, "loss": 0.4822, "step": 8263 }, { "epoch": 2.64419795221843, "grad_norm": 0.4526835083961487, "learning_rate": 4.2178307604876067e-07, "loss": 0.4142, "step": 8264 }, { "epoch": 2.644517918088737, "grad_norm": 0.47004154324531555, "learning_rate": 4.210349676838216e-07, "loss": 0.4693, "step": 8265 }, { "epoch": 2.6448378839590445, "grad_norm": 0.4589392840862274, "learning_rate": 4.202874941978935e-07, "loss": 0.4177, "step": 8266 }, { "epoch": 2.6451578498293515, "grad_norm": 0.4764752984046936, "learning_rate": 4.1954065569461355e-07, "loss": 0.4506, "step": 8267 }, { "epoch": 2.6454778156996586, "grad_norm": 0.5085862874984741, "learning_rate": 4.187944522775328e-07, "loss": 0.4215, "step": 8268 }, { "epoch": 2.645797781569966, "grad_norm": 0.5105277299880981, "learning_rate": 4.1804888405011125e-07, "loss": 0.4466, "step": 8269 }, { "epoch": 2.646117747440273, "grad_norm": 0.5165952444076538, "learning_rate": 4.1730395111572397e-07, "loss": 0.4488, "step": 8270 }, { "epoch": 2.64643771331058, "grad_norm": 0.5091129541397095, "learning_rate": 4.165596535776578e-07, "loss": 0.4375, "step": 8271 }, { "epoch": 2.6467576791808876, "grad_norm": 0.5199570655822754, "learning_rate": 4.1581599153911026e-07, "loss": 0.4288, "step": 8272 }, { "epoch": 2.6470776450511946, "grad_norm": 0.48100724816322327, "learning_rate": 4.150729651031898e-07, "loss": 0.4851, "step": 8273 }, { "epoch": 2.6473976109215016, "grad_norm": 0.44028621912002563, "learning_rate": 4.143305743729181e-07, "loss": 0.3605, "step": 8274 }, { "epoch": 2.647717576791809, "grad_norm": 0.5184507966041565, "learning_rate": 4.1358881945122884e-07, "loss": 0.462, "step": 8275 }, { "epoch": 2.648037542662116, "grad_norm": 0.5367111563682556, "learning_rate": 4.128477004409681e-07, "loss": 0.425, "step": 8276 }, { "epoch": 2.648357508532423, "grad_norm": 0.51981121301651, "learning_rate": 4.12107217444892e-07, "loss": 0.4634, "step": 8277 }, { "epoch": 2.6486774744027306, "grad_norm": 0.5907763242721558, "learning_rate": 4.113673705656701e-07, "loss": 0.4485, "step": 8278 }, { "epoch": 2.6489974402730376, "grad_norm": 0.47015947103500366, "learning_rate": 4.1062815990588147e-07, "loss": 0.4189, "step": 8279 }, { "epoch": 2.6493174061433447, "grad_norm": 0.5920960307121277, "learning_rate": 4.098895855680191e-07, "loss": 0.4666, "step": 8280 }, { "epoch": 2.649637372013652, "grad_norm": 0.5242739319801331, "learning_rate": 4.0915164765448936e-07, "loss": 0.4608, "step": 8281 }, { "epoch": 2.649957337883959, "grad_norm": 0.5620046854019165, "learning_rate": 4.08414346267606e-07, "loss": 0.458, "step": 8282 }, { "epoch": 2.650277303754266, "grad_norm": 0.4502480626106262, "learning_rate": 4.0767768150959785e-07, "loss": 0.4147, "step": 8283 }, { "epoch": 2.6505972696245736, "grad_norm": 0.4997509717941284, "learning_rate": 4.069416534826026e-07, "loss": 0.4684, "step": 8284 }, { "epoch": 2.6509172354948807, "grad_norm": 0.5171517729759216, "learning_rate": 4.0620626228867256e-07, "loss": 0.4625, "step": 8285 }, { "epoch": 2.6512372013651877, "grad_norm": 0.5204228162765503, "learning_rate": 4.054715080297722e-07, "loss": 0.424, "step": 8286 }, { "epoch": 2.6515571672354947, "grad_norm": 0.5608391165733337, "learning_rate": 4.0473739080777394e-07, "loss": 0.5094, "step": 8287 }, { "epoch": 2.651877133105802, "grad_norm": 0.4928218722343445, "learning_rate": 4.0400391072446533e-07, "loss": 0.401, "step": 8288 }, { "epoch": 2.6521970989761092, "grad_norm": 0.4749138653278351, "learning_rate": 4.032710678815421e-07, "loss": 0.4399, "step": 8289 }, { "epoch": 2.6525170648464163, "grad_norm": 0.4711044132709503, "learning_rate": 4.025388623806159e-07, "loss": 0.4032, "step": 8290 }, { "epoch": 2.6528370307167233, "grad_norm": 0.4894063472747803, "learning_rate": 4.0180729432320766e-07, "loss": 0.4009, "step": 8291 }, { "epoch": 2.6531569965870307, "grad_norm": 0.5121898651123047, "learning_rate": 4.010763638107501e-07, "loss": 0.4611, "step": 8292 }, { "epoch": 2.6534769624573378, "grad_norm": 0.5382512211799622, "learning_rate": 4.003460709445872e-07, "loss": 0.4387, "step": 8293 }, { "epoch": 2.653796928327645, "grad_norm": 0.5028988122940063, "learning_rate": 3.9961641582597464e-07, "loss": 0.4295, "step": 8294 }, { "epoch": 2.6541168941979523, "grad_norm": 0.5069299340248108, "learning_rate": 3.988873985560798e-07, "loss": 0.4483, "step": 8295 }, { "epoch": 2.6544368600682593, "grad_norm": 0.43606746196746826, "learning_rate": 3.9815901923598354e-07, "loss": 0.4117, "step": 8296 }, { "epoch": 2.6547568259385663, "grad_norm": 0.4850480556488037, "learning_rate": 3.9743127796667556e-07, "loss": 0.4718, "step": 8297 }, { "epoch": 2.655076791808874, "grad_norm": 0.529316246509552, "learning_rate": 3.967041748490569e-07, "loss": 0.4571, "step": 8298 }, { "epoch": 2.655396757679181, "grad_norm": 0.49332305788993835, "learning_rate": 3.9597770998394347e-07, "loss": 0.4133, "step": 8299 }, { "epoch": 2.655716723549488, "grad_norm": 0.523765504360199, "learning_rate": 3.9525188347205813e-07, "loss": 0.4775, "step": 8300 }, { "epoch": 2.6560366894197953, "grad_norm": 0.4612776041030884, "learning_rate": 3.9452669541403976e-07, "loss": 0.4597, "step": 8301 }, { "epoch": 2.6563566552901023, "grad_norm": 0.4600866436958313, "learning_rate": 3.9380214591043574e-07, "loss": 0.4735, "step": 8302 }, { "epoch": 2.6566766211604094, "grad_norm": 0.5209136605262756, "learning_rate": 3.9307823506170405e-07, "loss": 0.4345, "step": 8303 }, { "epoch": 2.656996587030717, "grad_norm": 0.5463297963142395, "learning_rate": 3.9235496296821884e-07, "loss": 0.435, "step": 8304 }, { "epoch": 2.657316552901024, "grad_norm": 0.57563316822052, "learning_rate": 3.9163232973026045e-07, "loss": 0.4648, "step": 8305 }, { "epoch": 2.657636518771331, "grad_norm": 0.5193503499031067, "learning_rate": 3.909103354480237e-07, "loss": 0.4834, "step": 8306 }, { "epoch": 2.6579564846416384, "grad_norm": 0.4927099049091339, "learning_rate": 3.9018898022161413e-07, "loss": 0.3804, "step": 8307 }, { "epoch": 2.6582764505119454, "grad_norm": 0.5175009965896606, "learning_rate": 3.894682641510478e-07, "loss": 0.4284, "step": 8308 }, { "epoch": 2.6585964163822524, "grad_norm": 0.4690502882003784, "learning_rate": 3.8874818733625363e-07, "loss": 0.4287, "step": 8309 }, { "epoch": 2.65891638225256, "grad_norm": 0.5278282165527344, "learning_rate": 3.880287498770696e-07, "loss": 0.4618, "step": 8310 }, { "epoch": 2.659236348122867, "grad_norm": 0.5248937010765076, "learning_rate": 3.873099518732493e-07, "loss": 0.4448, "step": 8311 }, { "epoch": 2.659556313993174, "grad_norm": 0.4708818793296814, "learning_rate": 3.8659179342445283e-07, "loss": 0.4113, "step": 8312 }, { "epoch": 2.6598762798634814, "grad_norm": 0.45589479804039, "learning_rate": 3.858742746302535e-07, "loss": 0.4362, "step": 8313 }, { "epoch": 2.6601962457337884, "grad_norm": 0.4845450222492218, "learning_rate": 3.851573955901383e-07, "loss": 0.3885, "step": 8314 }, { "epoch": 2.6605162116040955, "grad_norm": 0.511802077293396, "learning_rate": 3.8444115640350053e-07, "loss": 0.4661, "step": 8315 }, { "epoch": 2.660836177474403, "grad_norm": 0.5026915669441223, "learning_rate": 3.837255571696502e-07, "loss": 0.4425, "step": 8316 }, { "epoch": 2.66115614334471, "grad_norm": 0.4273529052734375, "learning_rate": 3.830105979878046e-07, "loss": 0.3899, "step": 8317 }, { "epoch": 2.661476109215017, "grad_norm": 0.5140239000320435, "learning_rate": 3.822962789570933e-07, "loss": 0.4782, "step": 8318 }, { "epoch": 2.6617960750853245, "grad_norm": 0.5597017407417297, "learning_rate": 3.815826001765594e-07, "loss": 0.4586, "step": 8319 }, { "epoch": 2.6621160409556315, "grad_norm": 0.540404200553894, "learning_rate": 3.8086956174515245e-07, "loss": 0.5083, "step": 8320 }, { "epoch": 2.6624360068259385, "grad_norm": 0.4860124886035919, "learning_rate": 3.801571637617391e-07, "loss": 0.4345, "step": 8321 }, { "epoch": 2.662755972696246, "grad_norm": 0.5035784244537354, "learning_rate": 3.79445406325093e-07, "loss": 0.416, "step": 8322 }, { "epoch": 2.663075938566553, "grad_norm": 0.5504601001739502, "learning_rate": 3.787342895338986e-07, "loss": 0.4488, "step": 8323 }, { "epoch": 2.66339590443686, "grad_norm": 0.5597847104072571, "learning_rate": 3.780238134867553e-07, "loss": 0.4273, "step": 8324 }, { "epoch": 2.6637158703071675, "grad_norm": 0.4814067482948303, "learning_rate": 3.773139782821705e-07, "loss": 0.4353, "step": 8325 }, { "epoch": 2.6640358361774745, "grad_norm": 0.520801842212677, "learning_rate": 3.766047840185638e-07, "loss": 0.4601, "step": 8326 }, { "epoch": 2.6643558020477816, "grad_norm": 0.5612490177154541, "learning_rate": 3.758962307942665e-07, "loss": 0.4002, "step": 8327 }, { "epoch": 2.6646757679180886, "grad_norm": 0.5105308890342712, "learning_rate": 3.7518831870751836e-07, "loss": 0.4121, "step": 8328 }, { "epoch": 2.664995733788396, "grad_norm": 0.5145578384399414, "learning_rate": 3.7448104785647476e-07, "loss": 0.4577, "step": 8329 }, { "epoch": 2.665315699658703, "grad_norm": 0.4928121864795685, "learning_rate": 3.737744183391978e-07, "loss": 0.4292, "step": 8330 }, { "epoch": 2.66563566552901, "grad_norm": 0.503603994846344, "learning_rate": 3.7306843025366415e-07, "loss": 0.4537, "step": 8331 }, { "epoch": 2.6659556313993176, "grad_norm": 0.5278449654579163, "learning_rate": 3.723630836977576e-07, "loss": 0.4532, "step": 8332 }, { "epoch": 2.6662755972696246, "grad_norm": 0.468374639749527, "learning_rate": 3.716583787692779e-07, "loss": 0.4234, "step": 8333 }, { "epoch": 2.6665955631399316, "grad_norm": 0.47699859738349915, "learning_rate": 3.709543155659323e-07, "loss": 0.4883, "step": 8334 }, { "epoch": 2.6669155290102387, "grad_norm": 0.456051230430603, "learning_rate": 3.702508941853383e-07, "loss": 0.3865, "step": 8335 }, { "epoch": 2.667235494880546, "grad_norm": 0.4542686343193054, "learning_rate": 3.695481147250285e-07, "loss": 0.3937, "step": 8336 }, { "epoch": 2.667555460750853, "grad_norm": 0.48175179958343506, "learning_rate": 3.6884597728244265e-07, "loss": 0.4983, "step": 8337 }, { "epoch": 2.66787542662116, "grad_norm": 0.4818973243236542, "learning_rate": 3.68144481954934e-07, "loss": 0.4271, "step": 8338 }, { "epoch": 2.6681953924914676, "grad_norm": 0.47783929109573364, "learning_rate": 3.6744362883976534e-07, "loss": 0.4392, "step": 8339 }, { "epoch": 2.6685153583617747, "grad_norm": 0.4501132667064667, "learning_rate": 3.6674341803411004e-07, "loss": 0.421, "step": 8340 }, { "epoch": 2.6688353242320817, "grad_norm": 0.5383400917053223, "learning_rate": 3.660438496350543e-07, "loss": 0.4726, "step": 8341 }, { "epoch": 2.669155290102389, "grad_norm": 0.5299282670021057, "learning_rate": 3.653449237395923e-07, "loss": 0.494, "step": 8342 }, { "epoch": 2.669475255972696, "grad_norm": 0.4743891656398773, "learning_rate": 3.646466404446336e-07, "loss": 0.4127, "step": 8343 }, { "epoch": 2.669795221843003, "grad_norm": 0.4851328432559967, "learning_rate": 3.639489998469947e-07, "loss": 0.4157, "step": 8344 }, { "epoch": 2.6701151877133107, "grad_norm": 0.49590522050857544, "learning_rate": 3.632520020434033e-07, "loss": 0.4207, "step": 8345 }, { "epoch": 2.6704351535836177, "grad_norm": 0.5184808969497681, "learning_rate": 3.6255564713050085e-07, "loss": 0.4305, "step": 8346 }, { "epoch": 2.6707551194539247, "grad_norm": 0.4785478711128235, "learning_rate": 3.618599352048363e-07, "loss": 0.477, "step": 8347 }, { "epoch": 2.671075085324232, "grad_norm": 0.47505947947502136, "learning_rate": 3.611648663628725e-07, "loss": 0.3898, "step": 8348 }, { "epoch": 2.6713950511945392, "grad_norm": 0.5582399964332581, "learning_rate": 3.604704407009801e-07, "loss": 0.4998, "step": 8349 }, { "epoch": 2.6717150170648463, "grad_norm": 0.5541906952857971, "learning_rate": 3.5977665831544217e-07, "loss": 0.4653, "step": 8350 }, { "epoch": 2.6720349829351537, "grad_norm": 0.512864351272583, "learning_rate": 3.5908351930245335e-07, "loss": 0.4074, "step": 8351 }, { "epoch": 2.6723549488054608, "grad_norm": 0.4545230269432068, "learning_rate": 3.5839102375811675e-07, "loss": 0.4042, "step": 8352 }, { "epoch": 2.672674914675768, "grad_norm": 0.4630187153816223, "learning_rate": 3.5769917177845006e-07, "loss": 0.4368, "step": 8353 }, { "epoch": 2.6729948805460753, "grad_norm": 0.550320029258728, "learning_rate": 3.570079634593776e-07, "loss": 0.4591, "step": 8354 }, { "epoch": 2.6733148464163823, "grad_norm": 0.4963036775588989, "learning_rate": 3.563173988967361e-07, "loss": 0.4149, "step": 8355 }, { "epoch": 2.6736348122866893, "grad_norm": 0.49244219064712524, "learning_rate": 3.5562747818627453e-07, "loss": 0.4326, "step": 8356 }, { "epoch": 2.673954778156997, "grad_norm": 0.4770994484424591, "learning_rate": 3.5493820142364963e-07, "loss": 0.3859, "step": 8357 }, { "epoch": 2.674274744027304, "grad_norm": 0.46881797909736633, "learning_rate": 3.5424956870443226e-07, "loss": 0.4083, "step": 8358 }, { "epoch": 2.674594709897611, "grad_norm": 0.5601597428321838, "learning_rate": 3.535615801240999e-07, "loss": 0.5077, "step": 8359 }, { "epoch": 2.6749146757679183, "grad_norm": 0.5159189105033875, "learning_rate": 3.528742357780457e-07, "loss": 0.4357, "step": 8360 }, { "epoch": 2.6752346416382253, "grad_norm": 0.4607735574245453, "learning_rate": 3.5218753576156837e-07, "loss": 0.3934, "step": 8361 }, { "epoch": 2.6755546075085324, "grad_norm": 0.5549046397209167, "learning_rate": 3.5150148016988073e-07, "loss": 0.4223, "step": 8362 }, { "epoch": 2.67587457337884, "grad_norm": 0.49229738116264343, "learning_rate": 3.508160690981055e-07, "loss": 0.4252, "step": 8363 }, { "epoch": 2.676194539249147, "grad_norm": 0.4866524636745453, "learning_rate": 3.50131302641275e-07, "loss": 0.4251, "step": 8364 }, { "epoch": 2.676514505119454, "grad_norm": 0.529787003993988, "learning_rate": 3.4944718089433337e-07, "loss": 0.435, "step": 8365 }, { "epoch": 2.6768344709897613, "grad_norm": 0.5101364850997925, "learning_rate": 3.4876370395213464e-07, "loss": 0.4214, "step": 8366 }, { "epoch": 2.6771544368600684, "grad_norm": 0.5263444185256958, "learning_rate": 3.4808087190944473e-07, "loss": 0.4555, "step": 8367 }, { "epoch": 2.6774744027303754, "grad_norm": 0.47625741362571716, "learning_rate": 3.473986848609384e-07, "loss": 0.4217, "step": 8368 }, { "epoch": 2.6777943686006824, "grad_norm": 0.5141160488128662, "learning_rate": 3.4671714290120063e-07, "loss": 0.4497, "step": 8369 }, { "epoch": 2.67811433447099, "grad_norm": 0.4759281873703003, "learning_rate": 3.4603624612472966e-07, "loss": 0.423, "step": 8370 }, { "epoch": 2.678434300341297, "grad_norm": 0.5235477685928345, "learning_rate": 3.453559946259316e-07, "loss": 0.457, "step": 8371 }, { "epoch": 2.678754266211604, "grad_norm": 0.45080694556236267, "learning_rate": 3.44676388499125e-07, "loss": 0.3974, "step": 8372 }, { "epoch": 2.6790742320819114, "grad_norm": 0.48992204666137695, "learning_rate": 3.4399742783853773e-07, "loss": 0.477, "step": 8373 }, { "epoch": 2.6793941979522184, "grad_norm": 0.4760851263999939, "learning_rate": 3.433191127383079e-07, "loss": 0.4041, "step": 8374 }, { "epoch": 2.6797141638225255, "grad_norm": 0.5457757711410522, "learning_rate": 3.426414432924857e-07, "loss": 0.487, "step": 8375 }, { "epoch": 2.6800341296928325, "grad_norm": 0.46977418661117554, "learning_rate": 3.4196441959502944e-07, "loss": 0.4664, "step": 8376 }, { "epoch": 2.68035409556314, "grad_norm": 0.43681231141090393, "learning_rate": 3.412880417398112e-07, "loss": 0.3664, "step": 8377 }, { "epoch": 2.680674061433447, "grad_norm": 0.498920202255249, "learning_rate": 3.4061230982061043e-07, "loss": 0.4783, "step": 8378 }, { "epoch": 2.680994027303754, "grad_norm": 0.49872443079948425, "learning_rate": 3.3993722393111775e-07, "loss": 0.4546, "step": 8379 }, { "epoch": 2.6813139931740615, "grad_norm": 0.4891470968723297, "learning_rate": 3.392627841649354e-07, "loss": 0.4265, "step": 8380 }, { "epoch": 2.6816339590443685, "grad_norm": 0.5192527770996094, "learning_rate": 3.385889906155748e-07, "loss": 0.4872, "step": 8381 }, { "epoch": 2.6819539249146755, "grad_norm": 0.488872766494751, "learning_rate": 3.37915843376459e-07, "loss": 0.38, "step": 8382 }, { "epoch": 2.682273890784983, "grad_norm": 0.47903430461883545, "learning_rate": 3.3724334254091997e-07, "loss": 0.4155, "step": 8383 }, { "epoch": 2.68259385665529, "grad_norm": 0.48480987548828125, "learning_rate": 3.365714882022003e-07, "loss": 0.4463, "step": 8384 }, { "epoch": 2.682913822525597, "grad_norm": 0.44706690311431885, "learning_rate": 3.35900280453455e-07, "loss": 0.4247, "step": 8385 }, { "epoch": 2.6832337883959045, "grad_norm": 0.4883236289024353, "learning_rate": 3.3522971938774575e-07, "loss": 0.4457, "step": 8386 }, { "epoch": 2.6835537542662116, "grad_norm": 0.5534349679946899, "learning_rate": 3.3455980509804865e-07, "loss": 0.4596, "step": 8387 }, { "epoch": 2.6838737201365186, "grad_norm": 0.4790193736553192, "learning_rate": 3.3389053767724723e-07, "loss": 0.3802, "step": 8388 }, { "epoch": 2.684193686006826, "grad_norm": 0.5198260545730591, "learning_rate": 3.332219172181356e-07, "loss": 0.4388, "step": 8389 }, { "epoch": 2.684513651877133, "grad_norm": 0.49348023533821106, "learning_rate": 3.325539438134201e-07, "loss": 0.3952, "step": 8390 }, { "epoch": 2.68483361774744, "grad_norm": 0.5152326226234436, "learning_rate": 3.318866175557145e-07, "loss": 0.4842, "step": 8391 }, { "epoch": 2.6851535836177476, "grad_norm": 0.49283209443092346, "learning_rate": 3.312199385375464e-07, "loss": 0.4116, "step": 8392 }, { "epoch": 2.6854735494880546, "grad_norm": 0.5174846649169922, "learning_rate": 3.3055390685135026e-07, "loss": 0.3847, "step": 8393 }, { "epoch": 2.6857935153583616, "grad_norm": 0.565240740776062, "learning_rate": 3.2988852258947155e-07, "loss": 0.4801, "step": 8394 }, { "epoch": 2.686113481228669, "grad_norm": 0.5280541777610779, "learning_rate": 3.2922378584416823e-07, "loss": 0.4187, "step": 8395 }, { "epoch": 2.686433447098976, "grad_norm": 0.5336593985557556, "learning_rate": 3.285596967076055e-07, "loss": 0.4802, "step": 8396 }, { "epoch": 2.686753412969283, "grad_norm": 0.5000171065330505, "learning_rate": 3.278962552718612e-07, "loss": 0.3496, "step": 8397 }, { "epoch": 2.6870733788395906, "grad_norm": 0.5063610672950745, "learning_rate": 3.2723346162892186e-07, "loss": 0.4361, "step": 8398 }, { "epoch": 2.6873933447098977, "grad_norm": 0.49247491359710693, "learning_rate": 3.26571315870684e-07, "loss": 0.4618, "step": 8399 }, { "epoch": 2.6877133105802047, "grad_norm": 0.5019463300704956, "learning_rate": 3.2590981808895637e-07, "loss": 0.4603, "step": 8400 }, { "epoch": 2.688033276450512, "grad_norm": 0.5265095829963684, "learning_rate": 3.252489683754545e-07, "loss": 0.4829, "step": 8401 }, { "epoch": 2.688353242320819, "grad_norm": 0.4545130431652069, "learning_rate": 3.245887668218078e-07, "loss": 0.3999, "step": 8402 }, { "epoch": 2.688673208191126, "grad_norm": 0.4803597927093506, "learning_rate": 3.239292135195532e-07, "loss": 0.4609, "step": 8403 }, { "epoch": 2.6889931740614337, "grad_norm": 0.49703601002693176, "learning_rate": 3.232703085601374e-07, "loss": 0.4472, "step": 8404 }, { "epoch": 2.6893131399317407, "grad_norm": 0.5202013254165649, "learning_rate": 3.226120520349202e-07, "loss": 0.435, "step": 8405 }, { "epoch": 2.6896331058020477, "grad_norm": 0.529909610748291, "learning_rate": 3.219544440351696e-07, "loss": 0.4801, "step": 8406 }, { "epoch": 2.689953071672355, "grad_norm": 0.511505663394928, "learning_rate": 3.2129748465206336e-07, "loss": 0.4322, "step": 8407 }, { "epoch": 2.6902730375426622, "grad_norm": 0.5222049355506897, "learning_rate": 3.2064117397668903e-07, "loss": 0.4835, "step": 8408 }, { "epoch": 2.6905930034129693, "grad_norm": 0.44696512818336487, "learning_rate": 3.1998551210004504e-07, "loss": 0.4163, "step": 8409 }, { "epoch": 2.6909129692832767, "grad_norm": 0.4778428077697754, "learning_rate": 3.193304991130397e-07, "loss": 0.4261, "step": 8410 }, { "epoch": 2.6912329351535837, "grad_norm": 0.46592646837234497, "learning_rate": 3.1867613510649264e-07, "loss": 0.4639, "step": 8411 }, { "epoch": 2.6915529010238908, "grad_norm": 0.41280320286750793, "learning_rate": 3.180224201711307e-07, "loss": 0.3912, "step": 8412 }, { "epoch": 2.691872866894198, "grad_norm": 0.48914268612861633, "learning_rate": 3.173693543975931e-07, "loss": 0.4419, "step": 8413 }, { "epoch": 2.6921928327645053, "grad_norm": 0.5670159459114075, "learning_rate": 3.167169378764273e-07, "loss": 0.4238, "step": 8414 }, { "epoch": 2.6925127986348123, "grad_norm": 0.5115007162094116, "learning_rate": 3.1606517069809216e-07, "loss": 0.4451, "step": 8415 }, { "epoch": 2.6928327645051193, "grad_norm": 0.5018688440322876, "learning_rate": 3.1541405295295645e-07, "loss": 0.4524, "step": 8416 }, { "epoch": 2.6931527303754264, "grad_norm": 0.5401268601417542, "learning_rate": 3.14763584731298e-07, "loss": 0.4099, "step": 8417 }, { "epoch": 2.693472696245734, "grad_norm": 0.5169517993927002, "learning_rate": 3.1411376612330513e-07, "loss": 0.4834, "step": 8418 }, { "epoch": 2.693792662116041, "grad_norm": 0.4915733337402344, "learning_rate": 3.134645972190753e-07, "loss": 0.4227, "step": 8419 }, { "epoch": 2.694112627986348, "grad_norm": 0.5028654336929321, "learning_rate": 3.12816078108617e-07, "loss": 0.4139, "step": 8420 }, { "epoch": 2.6944325938566553, "grad_norm": 0.5588289499282837, "learning_rate": 3.1216820888184895e-07, "loss": 0.5212, "step": 8421 }, { "epoch": 2.6947525597269624, "grad_norm": 0.5030496716499329, "learning_rate": 3.115209896285981e-07, "loss": 0.4438, "step": 8422 }, { "epoch": 2.6950725255972694, "grad_norm": 0.4874051809310913, "learning_rate": 3.1087442043860274e-07, "loss": 0.459, "step": 8423 }, { "epoch": 2.695392491467577, "grad_norm": 0.4535805583000183, "learning_rate": 3.102285014015094e-07, "loss": 0.3931, "step": 8424 }, { "epoch": 2.695712457337884, "grad_norm": 0.4591400921344757, "learning_rate": 3.095832326068765e-07, "loss": 0.415, "step": 8425 }, { "epoch": 2.696032423208191, "grad_norm": 0.4513499140739441, "learning_rate": 3.089386141441714e-07, "loss": 0.3891, "step": 8426 }, { "epoch": 2.6963523890784984, "grad_norm": 0.46307820081710815, "learning_rate": 3.082946461027708e-07, "loss": 0.4436, "step": 8427 }, { "epoch": 2.6966723549488054, "grad_norm": 0.45751821994781494, "learning_rate": 3.076513285719623e-07, "loss": 0.4138, "step": 8428 }, { "epoch": 2.6969923208191124, "grad_norm": 0.5338154435157776, "learning_rate": 3.070086616409412e-07, "loss": 0.4908, "step": 8429 }, { "epoch": 2.69731228668942, "grad_norm": 0.4924832880496979, "learning_rate": 3.063666453988151e-07, "loss": 0.4392, "step": 8430 }, { "epoch": 2.697632252559727, "grad_norm": 0.5246785879135132, "learning_rate": 3.0572527993460054e-07, "loss": 0.4646, "step": 8431 }, { "epoch": 2.697952218430034, "grad_norm": 0.4785500764846802, "learning_rate": 3.0508456533722363e-07, "loss": 0.3924, "step": 8432 }, { "epoch": 2.6982721843003414, "grad_norm": 0.4865003824234009, "learning_rate": 3.044445016955194e-07, "loss": 0.4422, "step": 8433 }, { "epoch": 2.6985921501706485, "grad_norm": 0.4638444781303406, "learning_rate": 3.038050890982336e-07, "loss": 0.4232, "step": 8434 }, { "epoch": 2.6989121160409555, "grad_norm": 0.48224329948425293, "learning_rate": 3.031663276340219e-07, "loss": 0.482, "step": 8435 }, { "epoch": 2.699232081911263, "grad_norm": 0.4687730669975281, "learning_rate": 3.0252821739144954e-07, "loss": 0.407, "step": 8436 }, { "epoch": 2.69955204778157, "grad_norm": 0.4968908131122589, "learning_rate": 3.018907584589914e-07, "loss": 0.4475, "step": 8437 }, { "epoch": 2.699872013651877, "grad_norm": 0.5187657475471497, "learning_rate": 3.012539509250317e-07, "loss": 0.4177, "step": 8438 }, { "epoch": 2.7001919795221845, "grad_norm": 0.5040234327316284, "learning_rate": 3.0061779487786325e-07, "loss": 0.4708, "step": 8439 }, { "epoch": 2.7005119453924915, "grad_norm": 0.4529518485069275, "learning_rate": 2.9998229040569147e-07, "loss": 0.4633, "step": 8440 }, { "epoch": 2.7008319112627985, "grad_norm": 0.4898926615715027, "learning_rate": 2.9934743759662985e-07, "loss": 0.4232, "step": 8441 }, { "epoch": 2.701151877133106, "grad_norm": 0.4939773976802826, "learning_rate": 2.987132365387013e-07, "loss": 0.5227, "step": 8442 }, { "epoch": 2.701471843003413, "grad_norm": 0.5070110559463501, "learning_rate": 2.980796873198383e-07, "loss": 0.4193, "step": 8443 }, { "epoch": 2.70179180887372, "grad_norm": 0.5593889355659485, "learning_rate": 2.974467900278821e-07, "loss": 0.4221, "step": 8444 }, { "epoch": 2.7021117747440275, "grad_norm": 0.511501669883728, "learning_rate": 2.96814544750586e-07, "loss": 0.4591, "step": 8445 }, { "epoch": 2.7024317406143346, "grad_norm": 0.4870849847793579, "learning_rate": 2.9618295157561205e-07, "loss": 0.4261, "step": 8446 }, { "epoch": 2.7027517064846416, "grad_norm": 0.5267578363418579, "learning_rate": 2.9555201059053073e-07, "loss": 0.4783, "step": 8447 }, { "epoch": 2.703071672354949, "grad_norm": 0.4592973589897156, "learning_rate": 2.949217218828226e-07, "loss": 0.4231, "step": 8448 }, { "epoch": 2.703391638225256, "grad_norm": 0.5345248579978943, "learning_rate": 2.942920855398773e-07, "loss": 0.4243, "step": 8449 }, { "epoch": 2.703711604095563, "grad_norm": 0.5086983442306519, "learning_rate": 2.9366310164899603e-07, "loss": 0.4987, "step": 8450 }, { "epoch": 2.7040315699658706, "grad_norm": 0.4644456207752228, "learning_rate": 2.9303477029738794e-07, "loss": 0.4356, "step": 8451 }, { "epoch": 2.7043515358361776, "grad_norm": 0.4368300139904022, "learning_rate": 2.9240709157217107e-07, "loss": 0.3591, "step": 8452 }, { "epoch": 2.7046715017064846, "grad_norm": 0.5334351658821106, "learning_rate": 2.9178006556037477e-07, "loss": 0.4868, "step": 8453 }, { "epoch": 2.7049914675767917, "grad_norm": 0.4681623876094818, "learning_rate": 2.911536923489361e-07, "loss": 0.3648, "step": 8454 }, { "epoch": 2.705311433447099, "grad_norm": 0.507057249546051, "learning_rate": 2.905279720247023e-07, "loss": 0.4935, "step": 8455 }, { "epoch": 2.705631399317406, "grad_norm": 0.4899614751338959, "learning_rate": 2.8990290467443174e-07, "loss": 0.4516, "step": 8456 }, { "epoch": 2.705951365187713, "grad_norm": 0.4832085967063904, "learning_rate": 2.8927849038479006e-07, "loss": 0.4697, "step": 8457 }, { "epoch": 2.70627133105802, "grad_norm": 0.5052345991134644, "learning_rate": 2.886547292423525e-07, "loss": 0.4208, "step": 8458 }, { "epoch": 2.7065912969283277, "grad_norm": 0.49080780148506165, "learning_rate": 2.8803162133360374e-07, "loss": 0.4259, "step": 8459 }, { "epoch": 2.7069112627986347, "grad_norm": 0.4580221474170685, "learning_rate": 2.874091667449397e-07, "loss": 0.4164, "step": 8460 }, { "epoch": 2.7072312286689417, "grad_norm": 0.44764164090156555, "learning_rate": 2.867873655626646e-07, "loss": 0.464, "step": 8461 }, { "epoch": 2.707551194539249, "grad_norm": 0.5174615383148193, "learning_rate": 2.861662178729918e-07, "loss": 0.465, "step": 8462 }, { "epoch": 2.707871160409556, "grad_norm": 0.5332177877426147, "learning_rate": 2.8554572376204347e-07, "loss": 0.4716, "step": 8463 }, { "epoch": 2.7081911262798632, "grad_norm": 0.4738543927669525, "learning_rate": 2.849258833158514e-07, "loss": 0.3845, "step": 8464 }, { "epoch": 2.7085110921501707, "grad_norm": 0.5458797812461853, "learning_rate": 2.8430669662035784e-07, "loss": 0.4715, "step": 8465 }, { "epoch": 2.7088310580204777, "grad_norm": 0.47262293100357056, "learning_rate": 2.836881637614147e-07, "loss": 0.4233, "step": 8466 }, { "epoch": 2.7091510238907848, "grad_norm": 0.4871342182159424, "learning_rate": 2.8307028482478185e-07, "loss": 0.4052, "step": 8467 }, { "epoch": 2.7094709897610922, "grad_norm": 0.5085920095443726, "learning_rate": 2.824530598961284e-07, "loss": 0.4503, "step": 8468 }, { "epoch": 2.7097909556313993, "grad_norm": 0.5077427625656128, "learning_rate": 2.8183648906103255e-07, "loss": 0.4202, "step": 8469 }, { "epoch": 2.7101109215017063, "grad_norm": 0.47658786177635193, "learning_rate": 2.812205724049838e-07, "loss": 0.4362, "step": 8470 }, { "epoch": 2.7104308873720138, "grad_norm": 0.4591486155986786, "learning_rate": 2.806053100133804e-07, "loss": 0.4146, "step": 8471 }, { "epoch": 2.710750853242321, "grad_norm": 0.5390037298202515, "learning_rate": 2.799907019715281e-07, "loss": 0.4764, "step": 8472 }, { "epoch": 2.711070819112628, "grad_norm": 0.5763192772865295, "learning_rate": 2.7937674836464256e-07, "loss": 0.5183, "step": 8473 }, { "epoch": 2.7113907849829353, "grad_norm": 0.44307756423950195, "learning_rate": 2.787634492778507e-07, "loss": 0.4003, "step": 8474 }, { "epoch": 2.7117107508532423, "grad_norm": 0.46029427647590637, "learning_rate": 2.781508047961862e-07, "loss": 0.4284, "step": 8475 }, { "epoch": 2.7120307167235493, "grad_norm": 0.5089544057846069, "learning_rate": 2.775388150045932e-07, "loss": 0.4204, "step": 8476 }, { "epoch": 2.712350682593857, "grad_norm": 0.5391183495521545, "learning_rate": 2.769274799879246e-07, "loss": 0.4842, "step": 8477 }, { "epoch": 2.712670648464164, "grad_norm": 0.4696575105190277, "learning_rate": 2.7631679983094293e-07, "loss": 0.4197, "step": 8478 }, { "epoch": 2.712990614334471, "grad_norm": 0.5044712424278259, "learning_rate": 2.757067746183195e-07, "loss": 0.4379, "step": 8479 }, { "epoch": 2.7133105802047783, "grad_norm": 0.5446345806121826, "learning_rate": 2.750974044346355e-07, "loss": 0.4504, "step": 8480 }, { "epoch": 2.7136305460750854, "grad_norm": 0.5092166662216187, "learning_rate": 2.744886893643806e-07, "loss": 0.4543, "step": 8481 }, { "epoch": 2.7139505119453924, "grad_norm": 0.48626455664634705, "learning_rate": 2.7388062949195384e-07, "loss": 0.4434, "step": 8482 }, { "epoch": 2.7142704778157, "grad_norm": 0.4884015619754791, "learning_rate": 2.7327322490166287e-07, "loss": 0.4649, "step": 8483 }, { "epoch": 2.714590443686007, "grad_norm": 0.5095328688621521, "learning_rate": 2.726664756777264e-07, "loss": 0.4626, "step": 8484 }, { "epoch": 2.714910409556314, "grad_norm": 0.4839845299720764, "learning_rate": 2.7206038190427e-07, "loss": 0.4162, "step": 8485 }, { "epoch": 2.7152303754266214, "grad_norm": 0.5074237585067749, "learning_rate": 2.7145494366532966e-07, "loss": 0.4135, "step": 8486 }, { "epoch": 2.7155503412969284, "grad_norm": 0.5321500897407532, "learning_rate": 2.7085016104484994e-07, "loss": 0.4739, "step": 8487 }, { "epoch": 2.7158703071672354, "grad_norm": 0.5675606727600098, "learning_rate": 2.702460341266838e-07, "loss": 0.4598, "step": 8488 }, { "epoch": 2.716190273037543, "grad_norm": 0.45474451780319214, "learning_rate": 2.6964256299459646e-07, "loss": 0.3731, "step": 8489 }, { "epoch": 2.71651023890785, "grad_norm": 0.5283628702163696, "learning_rate": 2.6903974773225703e-07, "loss": 0.5131, "step": 8490 }, { "epoch": 2.716830204778157, "grad_norm": 0.4831763207912445, "learning_rate": 2.684375884232493e-07, "loss": 0.3843, "step": 8491 }, { "epoch": 2.7171501706484644, "grad_norm": 0.48812925815582275, "learning_rate": 2.6783608515106195e-07, "loss": 0.4493, "step": 8492 }, { "epoch": 2.7174701365187715, "grad_norm": 0.47067147493362427, "learning_rate": 2.6723523799909387e-07, "loss": 0.4413, "step": 8493 }, { "epoch": 2.7177901023890785, "grad_norm": 0.4778099060058594, "learning_rate": 2.6663504705065404e-07, "loss": 0.4564, "step": 8494 }, { "epoch": 2.7181100682593855, "grad_norm": 0.4619276523590088, "learning_rate": 2.660355123889585e-07, "loss": 0.3582, "step": 8495 }, { "epoch": 2.718430034129693, "grad_norm": 0.5539404153823853, "learning_rate": 2.654366340971354e-07, "loss": 0.4361, "step": 8496 }, { "epoch": 2.71875, "grad_norm": 0.5206783413887024, "learning_rate": 2.648384122582187e-07, "loss": 0.4379, "step": 8497 }, { "epoch": 2.719069965870307, "grad_norm": 0.5146101117134094, "learning_rate": 2.642408469551516e-07, "loss": 0.4665, "step": 8498 }, { "epoch": 2.7193899317406145, "grad_norm": 0.4704565703868866, "learning_rate": 2.636439382707895e-07, "loss": 0.448, "step": 8499 }, { "epoch": 2.7197098976109215, "grad_norm": 0.4845392405986786, "learning_rate": 2.630476862878928e-07, "loss": 0.3829, "step": 8500 }, { "epoch": 2.7200298634812285, "grad_norm": 0.5160871148109436, "learning_rate": 2.6245209108913316e-07, "loss": 0.4631, "step": 8501 }, { "epoch": 2.7203498293515356, "grad_norm": 0.4980718791484833, "learning_rate": 2.618571527570907e-07, "loss": 0.491, "step": 8502 }, { "epoch": 2.720669795221843, "grad_norm": 0.4564629793167114, "learning_rate": 2.6126287137425376e-07, "loss": 0.3945, "step": 8503 }, { "epoch": 2.72098976109215, "grad_norm": 0.523655891418457, "learning_rate": 2.6066924702302044e-07, "loss": 0.4404, "step": 8504 }, { "epoch": 2.721309726962457, "grad_norm": 0.46495863795280457, "learning_rate": 2.6007627978569817e-07, "loss": 0.44, "step": 8505 }, { "epoch": 2.7216296928327646, "grad_norm": 0.5155764222145081, "learning_rate": 2.594839697445017e-07, "loss": 0.4447, "step": 8506 }, { "epoch": 2.7219496587030716, "grad_norm": 0.5131596326828003, "learning_rate": 2.588923169815555e-07, "loss": 0.448, "step": 8507 }, { "epoch": 2.7222696245733786, "grad_norm": 0.5179362893104553, "learning_rate": 2.5830132157889377e-07, "loss": 0.4763, "step": 8508 }, { "epoch": 2.722589590443686, "grad_norm": 0.40810224413871765, "learning_rate": 2.577109836184577e-07, "loss": 0.3922, "step": 8509 }, { "epoch": 2.722909556313993, "grad_norm": 0.48155176639556885, "learning_rate": 2.5712130318209904e-07, "loss": 0.4474, "step": 8510 }, { "epoch": 2.7232295221843, "grad_norm": 0.4342872202396393, "learning_rate": 2.5653228035157796e-07, "loss": 0.3829, "step": 8511 }, { "epoch": 2.7235494880546076, "grad_norm": 0.4778122007846832, "learning_rate": 2.5594391520856187e-07, "loss": 0.4624, "step": 8512 }, { "epoch": 2.7238694539249146, "grad_norm": 0.4993056654930115, "learning_rate": 2.5535620783463e-07, "loss": 0.446, "step": 8513 }, { "epoch": 2.7241894197952217, "grad_norm": 0.426123708486557, "learning_rate": 2.5476915831126715e-07, "loss": 0.4275, "step": 8514 }, { "epoch": 2.724509385665529, "grad_norm": 0.4558579623699188, "learning_rate": 2.5418276671986984e-07, "loss": 0.4425, "step": 8515 }, { "epoch": 2.724829351535836, "grad_norm": 0.4622271955013275, "learning_rate": 2.5359703314174147e-07, "loss": 0.4783, "step": 8516 }, { "epoch": 2.725149317406143, "grad_norm": 0.4740881025791168, "learning_rate": 2.530119576580936e-07, "loss": 0.4173, "step": 8517 }, { "epoch": 2.7254692832764507, "grad_norm": 0.4327777326107025, "learning_rate": 2.5242754035004924e-07, "loss": 0.3782, "step": 8518 }, { "epoch": 2.7257892491467577, "grad_norm": 0.470712810754776, "learning_rate": 2.518437812986374e-07, "loss": 0.4403, "step": 8519 }, { "epoch": 2.7261092150170647, "grad_norm": 0.5009084939956665, "learning_rate": 2.512606805847978e-07, "loss": 0.4344, "step": 8520 }, { "epoch": 2.726429180887372, "grad_norm": 0.5207602381706238, "learning_rate": 2.50678238289378e-07, "loss": 0.4582, "step": 8521 }, { "epoch": 2.726749146757679, "grad_norm": 0.5586817264556885, "learning_rate": 2.500964544931328e-07, "loss": 0.4734, "step": 8522 }, { "epoch": 2.7270691126279862, "grad_norm": 0.5082388520240784, "learning_rate": 2.495153292767299e-07, "loss": 0.3952, "step": 8523 }, { "epoch": 2.7273890784982937, "grad_norm": 0.46565231680870056, "learning_rate": 2.4893486272074055e-07, "loss": 0.4361, "step": 8524 }, { "epoch": 2.7277090443686007, "grad_norm": 0.46218451857566833, "learning_rate": 2.4835505490564846e-07, "loss": 0.4188, "step": 8525 }, { "epoch": 2.7280290102389078, "grad_norm": 0.483097106218338, "learning_rate": 2.477759059118451e-07, "loss": 0.3994, "step": 8526 }, { "epoch": 2.7283489761092152, "grad_norm": 0.5206490159034729, "learning_rate": 2.471974158196283e-07, "loss": 0.4603, "step": 8527 }, { "epoch": 2.7286689419795223, "grad_norm": 0.4444100558757782, "learning_rate": 2.4661958470920845e-07, "loss": 0.4267, "step": 8528 }, { "epoch": 2.7289889078498293, "grad_norm": 0.4927453398704529, "learning_rate": 2.4604241266070085e-07, "loss": 0.4268, "step": 8529 }, { "epoch": 2.7293088737201368, "grad_norm": 0.46823206543922424, "learning_rate": 2.454658997541326e-07, "loss": 0.389, "step": 8530 }, { "epoch": 2.729628839590444, "grad_norm": 0.5520978569984436, "learning_rate": 2.44890046069437e-07, "loss": 0.5481, "step": 8531 }, { "epoch": 2.729948805460751, "grad_norm": 0.4668036699295044, "learning_rate": 2.4431485168645685e-07, "loss": 0.4141, "step": 8532 }, { "epoch": 2.7302687713310583, "grad_norm": 0.49614331126213074, "learning_rate": 2.437403166849439e-07, "loss": 0.4528, "step": 8533 }, { "epoch": 2.7305887372013653, "grad_norm": 0.501336932182312, "learning_rate": 2.431664411445578e-07, "loss": 0.4259, "step": 8534 }, { "epoch": 2.7309087030716723, "grad_norm": 0.5080056190490723, "learning_rate": 2.425932251448676e-07, "loss": 0.4172, "step": 8535 }, { "epoch": 2.73122866894198, "grad_norm": 0.5129799842834473, "learning_rate": 2.420206687653498e-07, "loss": 0.4751, "step": 8536 }, { "epoch": 2.731548634812287, "grad_norm": 0.5177556872367859, "learning_rate": 2.4144877208538974e-07, "loss": 0.4168, "step": 8537 }, { "epoch": 2.731868600682594, "grad_norm": 0.4787764549255371, "learning_rate": 2.408775351842829e-07, "loss": 0.456, "step": 8538 }, { "epoch": 2.732188566552901, "grad_norm": 0.5124397277832031, "learning_rate": 2.4030695814123093e-07, "loss": 0.4248, "step": 8539 }, { "epoch": 2.7325085324232083, "grad_norm": 0.5004435181617737, "learning_rate": 2.397370410353456e-07, "loss": 0.4468, "step": 8540 }, { "epoch": 2.7328284982935154, "grad_norm": 0.4988056421279907, "learning_rate": 2.391677839456458e-07, "loss": 0.4366, "step": 8541 }, { "epoch": 2.7331484641638224, "grad_norm": 0.5729777812957764, "learning_rate": 2.385991869510601e-07, "loss": 0.4458, "step": 8542 }, { "epoch": 2.7334684300341294, "grad_norm": 0.4886725842952728, "learning_rate": 2.380312501304255e-07, "loss": 0.45, "step": 8543 }, { "epoch": 2.733788395904437, "grad_norm": 0.4933490753173828, "learning_rate": 2.3746397356248617e-07, "loss": 0.4277, "step": 8544 }, { "epoch": 2.734108361774744, "grad_norm": 0.5624293684959412, "learning_rate": 2.3689735732589757e-07, "loss": 0.4675, "step": 8545 }, { "epoch": 2.734428327645051, "grad_norm": 0.48448383808135986, "learning_rate": 2.3633140149921962e-07, "loss": 0.4117, "step": 8546 }, { "epoch": 2.7347482935153584, "grad_norm": 0.47338375449180603, "learning_rate": 2.35766106160924e-07, "loss": 0.4596, "step": 8547 }, { "epoch": 2.7350682593856654, "grad_norm": 0.43746209144592285, "learning_rate": 2.3520147138938965e-07, "loss": 0.4113, "step": 8548 }, { "epoch": 2.7353882252559725, "grad_norm": 0.44164198637008667, "learning_rate": 2.3463749726290287e-07, "loss": 0.4306, "step": 8549 }, { "epoch": 2.73570819112628, "grad_norm": 0.46448519825935364, "learning_rate": 2.3407418385966053e-07, "loss": 0.4599, "step": 8550 }, { "epoch": 2.736028156996587, "grad_norm": 0.46389082074165344, "learning_rate": 2.3351153125776515e-07, "loss": 0.4802, "step": 8551 }, { "epoch": 2.736348122866894, "grad_norm": 0.4623454511165619, "learning_rate": 2.3294953953523104e-07, "loss": 0.3895, "step": 8552 }, { "epoch": 2.7366680887372015, "grad_norm": 0.5046341419219971, "learning_rate": 2.323882087699786e-07, "loss": 0.4447, "step": 8553 }, { "epoch": 2.7369880546075085, "grad_norm": 0.47571954131126404, "learning_rate": 2.3182753903983556e-07, "loss": 0.4287, "step": 8554 }, { "epoch": 2.7373080204778155, "grad_norm": 0.5265892744064331, "learning_rate": 2.312675304225409e-07, "loss": 0.4532, "step": 8555 }, { "epoch": 2.737627986348123, "grad_norm": 0.5027657747268677, "learning_rate": 2.3070818299573972e-07, "loss": 0.4218, "step": 8556 }, { "epoch": 2.73794795221843, "grad_norm": 0.5723645091056824, "learning_rate": 2.3014949683698717e-07, "loss": 0.5078, "step": 8557 }, { "epoch": 2.738267918088737, "grad_norm": 0.519535481929779, "learning_rate": 2.2959147202374522e-07, "loss": 0.4089, "step": 8558 }, { "epoch": 2.7385878839590445, "grad_norm": 0.5125702619552612, "learning_rate": 2.290341086333836e-07, "loss": 0.4165, "step": 8559 }, { "epoch": 2.7389078498293515, "grad_norm": 0.5632970929145813, "learning_rate": 2.2847740674318385e-07, "loss": 0.5036, "step": 8560 }, { "epoch": 2.7392278156996586, "grad_norm": 0.48991167545318604, "learning_rate": 2.2792136643033035e-07, "loss": 0.4415, "step": 8561 }, { "epoch": 2.739547781569966, "grad_norm": 0.4932671785354614, "learning_rate": 2.27365987771922e-07, "loss": 0.413, "step": 8562 }, { "epoch": 2.739867747440273, "grad_norm": 0.5287392735481262, "learning_rate": 2.2681127084496047e-07, "loss": 0.4796, "step": 8563 }, { "epoch": 2.74018771331058, "grad_norm": 0.5054778456687927, "learning_rate": 2.2625721572635762e-07, "loss": 0.382, "step": 8564 }, { "epoch": 2.7405076791808876, "grad_norm": 0.4676485061645508, "learning_rate": 2.2570382249293586e-07, "loss": 0.3968, "step": 8565 }, { "epoch": 2.7408276450511946, "grad_norm": 0.47252145409584045, "learning_rate": 2.251510912214222e-07, "loss": 0.4577, "step": 8566 }, { "epoch": 2.7411476109215016, "grad_norm": 0.4993182122707367, "learning_rate": 2.2459902198845419e-07, "loss": 0.4753, "step": 8567 }, { "epoch": 2.741467576791809, "grad_norm": 0.48656511306762695, "learning_rate": 2.2404761487057726e-07, "loss": 0.4018, "step": 8568 }, { "epoch": 2.741787542662116, "grad_norm": 0.573748767375946, "learning_rate": 2.2349686994424303e-07, "loss": 0.4718, "step": 8569 }, { "epoch": 2.742107508532423, "grad_norm": 0.5051923990249634, "learning_rate": 2.2294678728581487e-07, "loss": 0.4768, "step": 8570 }, { "epoch": 2.7424274744027306, "grad_norm": 0.4523454010486603, "learning_rate": 2.2239736697156067e-07, "loss": 0.4197, "step": 8571 }, { "epoch": 2.7427474402730376, "grad_norm": 0.4958074986934662, "learning_rate": 2.2184860907766003e-07, "loss": 0.3901, "step": 8572 }, { "epoch": 2.7430674061433447, "grad_norm": 0.5585159659385681, "learning_rate": 2.2130051368019823e-07, "loss": 0.5015, "step": 8573 }, { "epoch": 2.743387372013652, "grad_norm": 0.4593925476074219, "learning_rate": 2.2075308085516835e-07, "loss": 0.4388, "step": 8574 }, { "epoch": 2.743707337883959, "grad_norm": 0.5206596255302429, "learning_rate": 2.202063106784741e-07, "loss": 0.4959, "step": 8575 }, { "epoch": 2.744027303754266, "grad_norm": 0.5175601840019226, "learning_rate": 2.196602032259243e-07, "loss": 0.4169, "step": 8576 }, { "epoch": 2.7443472696245736, "grad_norm": 0.4904636740684509, "learning_rate": 2.1911475857323893e-07, "loss": 0.4093, "step": 8577 }, { "epoch": 2.7446672354948807, "grad_norm": 0.5445221662521362, "learning_rate": 2.1856997679604363e-07, "loss": 0.4526, "step": 8578 }, { "epoch": 2.7449872013651877, "grad_norm": 0.48061031103134155, "learning_rate": 2.1802585796987243e-07, "loss": 0.4406, "step": 8579 }, { "epoch": 2.7453071672354947, "grad_norm": 0.5459372401237488, "learning_rate": 2.1748240217016936e-07, "loss": 0.4488, "step": 8580 }, { "epoch": 2.745627133105802, "grad_norm": 0.49022161960601807, "learning_rate": 2.1693960947228532e-07, "loss": 0.4235, "step": 8581 }, { "epoch": 2.7459470989761092, "grad_norm": 0.5033886432647705, "learning_rate": 2.1639747995147843e-07, "loss": 0.4521, "step": 8582 }, { "epoch": 2.7462670648464163, "grad_norm": 0.4898601472377777, "learning_rate": 2.1585601368291575e-07, "loss": 0.468, "step": 8583 }, { "epoch": 2.7465870307167233, "grad_norm": 0.4970014691352844, "learning_rate": 2.1531521074167172e-07, "loss": 0.5213, "step": 8584 }, { "epoch": 2.7469069965870307, "grad_norm": 0.44684627652168274, "learning_rate": 2.1477507120272967e-07, "loss": 0.3995, "step": 8585 }, { "epoch": 2.7472269624573378, "grad_norm": 0.5002879500389099, "learning_rate": 2.1423559514098135e-07, "loss": 0.438, "step": 8586 }, { "epoch": 2.747546928327645, "grad_norm": 0.5029903054237366, "learning_rate": 2.1369678263122583e-07, "loss": 0.405, "step": 8587 }, { "epoch": 2.7478668941979523, "grad_norm": 0.49236345291137695, "learning_rate": 2.1315863374816893e-07, "loss": 0.4375, "step": 8588 }, { "epoch": 2.7481868600682593, "grad_norm": 0.44991016387939453, "learning_rate": 2.1262114856642535e-07, "loss": 0.3779, "step": 8589 }, { "epoch": 2.7485068259385663, "grad_norm": 0.5005273818969727, "learning_rate": 2.1208432716051886e-07, "loss": 0.4834, "step": 8590 }, { "epoch": 2.748826791808874, "grad_norm": 0.4878215491771698, "learning_rate": 2.1154816960488157e-07, "loss": 0.449, "step": 8591 }, { "epoch": 2.749146757679181, "grad_norm": 0.45098960399627686, "learning_rate": 2.1101267597385066e-07, "loss": 0.3973, "step": 8592 }, { "epoch": 2.749466723549488, "grad_norm": 0.47468888759613037, "learning_rate": 2.1047784634167402e-07, "loss": 0.4342, "step": 8593 }, { "epoch": 2.7497866894197953, "grad_norm": 0.474040150642395, "learning_rate": 2.0994368078250505e-07, "loss": 0.461, "step": 8594 }, { "epoch": 2.7501066552901023, "grad_norm": 0.4642796814441681, "learning_rate": 2.094101793704073e-07, "loss": 0.4474, "step": 8595 }, { "epoch": 2.7504266211604094, "grad_norm": 0.4851500988006592, "learning_rate": 2.0887734217935218e-07, "loss": 0.4678, "step": 8596 }, { "epoch": 2.750746587030717, "grad_norm": 0.5419781804084778, "learning_rate": 2.0834516928321725e-07, "loss": 0.476, "step": 8597 }, { "epoch": 2.751066552901024, "grad_norm": 0.49836188554763794, "learning_rate": 2.0781366075578902e-07, "loss": 0.4385, "step": 8598 }, { "epoch": 2.751386518771331, "grad_norm": 0.49485135078430176, "learning_rate": 2.0728281667076133e-07, "loss": 0.4694, "step": 8599 }, { "epoch": 2.7517064846416384, "grad_norm": 0.5159654021263123, "learning_rate": 2.0675263710173697e-07, "loss": 0.452, "step": 8600 }, { "epoch": 2.7520264505119454, "grad_norm": 0.4692765474319458, "learning_rate": 2.0622312212222662e-07, "loss": 0.4189, "step": 8601 }, { "epoch": 2.7523464163822524, "grad_norm": 0.44056499004364014, "learning_rate": 2.0569427180564704e-07, "loss": 0.441, "step": 8602 }, { "epoch": 2.75266638225256, "grad_norm": 0.51723313331604, "learning_rate": 2.0516608622532518e-07, "loss": 0.4848, "step": 8603 }, { "epoch": 2.752986348122867, "grad_norm": 0.5240875482559204, "learning_rate": 2.0463856545449246e-07, "loss": 0.457, "step": 8604 }, { "epoch": 2.753306313993174, "grad_norm": 0.4982944130897522, "learning_rate": 2.0411170956629145e-07, "loss": 0.4306, "step": 8605 }, { "epoch": 2.7536262798634814, "grad_norm": 0.5124058723449707, "learning_rate": 2.0358551863377264e-07, "loss": 0.4723, "step": 8606 }, { "epoch": 2.7539462457337884, "grad_norm": 0.44902369379997253, "learning_rate": 2.030599927298915e-07, "loss": 0.4156, "step": 8607 }, { "epoch": 2.7542662116040955, "grad_norm": 0.4867861270904541, "learning_rate": 2.0253513192751374e-07, "loss": 0.4051, "step": 8608 }, { "epoch": 2.754586177474403, "grad_norm": 0.48332637548446655, "learning_rate": 2.0201093629941048e-07, "loss": 0.4343, "step": 8609 }, { "epoch": 2.75490614334471, "grad_norm": 0.5181475281715393, "learning_rate": 2.014874059182631e-07, "loss": 0.4967, "step": 8610 }, { "epoch": 2.755226109215017, "grad_norm": 0.4403003454208374, "learning_rate": 2.0096454085666016e-07, "loss": 0.321, "step": 8611 }, { "epoch": 2.7555460750853245, "grad_norm": 0.44011250138282776, "learning_rate": 2.0044234118709704e-07, "loss": 0.4645, "step": 8612 }, { "epoch": 2.7558660409556315, "grad_norm": 0.4728182554244995, "learning_rate": 1.9992080698197746e-07, "loss": 0.4957, "step": 8613 }, { "epoch": 2.7561860068259385, "grad_norm": 0.47703278064727783, "learning_rate": 1.9939993831361137e-07, "loss": 0.4213, "step": 8614 }, { "epoch": 2.756505972696246, "grad_norm": 0.4957233667373657, "learning_rate": 1.9887973525421988e-07, "loss": 0.4404, "step": 8615 }, { "epoch": 2.756825938566553, "grad_norm": 0.4639700651168823, "learning_rate": 1.983601978759292e-07, "loss": 0.4406, "step": 8616 }, { "epoch": 2.75714590443686, "grad_norm": 0.4967518448829651, "learning_rate": 1.978413262507739e-07, "loss": 0.4768, "step": 8617 }, { "epoch": 2.7574658703071675, "grad_norm": 0.5306515097618103, "learning_rate": 1.9732312045069534e-07, "loss": 0.4878, "step": 8618 }, { "epoch": 2.7577858361774745, "grad_norm": 0.4871678352355957, "learning_rate": 1.9680558054754384e-07, "loss": 0.4272, "step": 8619 }, { "epoch": 2.7581058020477816, "grad_norm": 0.448634535074234, "learning_rate": 1.9628870661307698e-07, "loss": 0.3849, "step": 8620 }, { "epoch": 2.7584257679180886, "grad_norm": 0.47452399134635925, "learning_rate": 1.957724987189602e-07, "loss": 0.4708, "step": 8621 }, { "epoch": 2.758745733788396, "grad_norm": 0.46345850825309753, "learning_rate": 1.9525695693676628e-07, "loss": 0.4261, "step": 8622 }, { "epoch": 2.759065699658703, "grad_norm": 0.5059054493904114, "learning_rate": 1.9474208133797524e-07, "loss": 0.4908, "step": 8623 }, { "epoch": 2.75938566552901, "grad_norm": 0.47566303610801697, "learning_rate": 1.9422787199397553e-07, "loss": 0.4469, "step": 8624 }, { "epoch": 2.7597056313993176, "grad_norm": 0.4674135446548462, "learning_rate": 1.9371432897606236e-07, "loss": 0.3585, "step": 8625 }, { "epoch": 2.7600255972696246, "grad_norm": 0.48827576637268066, "learning_rate": 1.9320145235544042e-07, "loss": 0.4537, "step": 8626 }, { "epoch": 2.7603455631399316, "grad_norm": 0.5700124502182007, "learning_rate": 1.926892422032195e-07, "loss": 0.4124, "step": 8627 }, { "epoch": 2.7606655290102387, "grad_norm": 0.5084264278411865, "learning_rate": 1.9217769859041834e-07, "loss": 0.4028, "step": 8628 }, { "epoch": 2.760985494880546, "grad_norm": 0.5398024320602417, "learning_rate": 1.9166682158796303e-07, "loss": 0.4654, "step": 8629 }, { "epoch": 2.761305460750853, "grad_norm": 0.5033600926399231, "learning_rate": 1.9115661126668738e-07, "loss": 0.4298, "step": 8630 }, { "epoch": 2.76162542662116, "grad_norm": 0.491950124502182, "learning_rate": 1.9064706769733378e-07, "loss": 0.4715, "step": 8631 }, { "epoch": 2.7619453924914676, "grad_norm": 0.5091946125030518, "learning_rate": 1.9013819095054954e-07, "loss": 0.4647, "step": 8632 }, { "epoch": 2.7622653583617747, "grad_norm": 0.48028603196144104, "learning_rate": 1.8962998109689156e-07, "loss": 0.4132, "step": 8633 }, { "epoch": 2.7625853242320817, "grad_norm": 0.5029857158660889, "learning_rate": 1.8912243820682296e-07, "loss": 0.486, "step": 8634 }, { "epoch": 2.762905290102389, "grad_norm": 0.4869309067726135, "learning_rate": 1.8861556235071632e-07, "loss": 0.4707, "step": 8635 }, { "epoch": 2.763225255972696, "grad_norm": 0.4437768757343292, "learning_rate": 1.881093535988504e-07, "loss": 0.3903, "step": 8636 }, { "epoch": 2.763545221843003, "grad_norm": 0.5005133748054504, "learning_rate": 1.8760381202141188e-07, "loss": 0.4381, "step": 8637 }, { "epoch": 2.7638651877133107, "grad_norm": 0.4766649901866913, "learning_rate": 1.8709893768849406e-07, "loss": 0.4338, "step": 8638 }, { "epoch": 2.7641851535836177, "grad_norm": 0.4825212061405182, "learning_rate": 1.8659473067009826e-07, "loss": 0.4023, "step": 8639 }, { "epoch": 2.7645051194539247, "grad_norm": 0.47575074434280396, "learning_rate": 1.860911910361335e-07, "loss": 0.4149, "step": 8640 }, { "epoch": 2.764825085324232, "grad_norm": 0.5040088891983032, "learning_rate": 1.855883188564167e-07, "loss": 0.4926, "step": 8641 }, { "epoch": 2.7651450511945392, "grad_norm": 0.4763101637363434, "learning_rate": 1.8508611420067157e-07, "loss": 0.406, "step": 8642 }, { "epoch": 2.7654650170648463, "grad_norm": 0.5055087804794312, "learning_rate": 1.8458457713852907e-07, "loss": 0.485, "step": 8643 }, { "epoch": 2.7657849829351537, "grad_norm": 0.527959406375885, "learning_rate": 1.8408370773952744e-07, "loss": 0.4299, "step": 8644 }, { "epoch": 2.7661049488054608, "grad_norm": 0.49691763520240784, "learning_rate": 1.8358350607311392e-07, "loss": 0.4119, "step": 8645 }, { "epoch": 2.766424914675768, "grad_norm": 0.4913119673728943, "learning_rate": 1.8308397220864193e-07, "loss": 0.406, "step": 8646 }, { "epoch": 2.7667448805460753, "grad_norm": 0.4808558225631714, "learning_rate": 1.8258510621537219e-07, "loss": 0.4341, "step": 8647 }, { "epoch": 2.7670648464163823, "grad_norm": 0.5470263361930847, "learning_rate": 1.8208690816247266e-07, "loss": 0.4325, "step": 8648 }, { "epoch": 2.7673848122866893, "grad_norm": 0.5051500201225281, "learning_rate": 1.815893781190181e-07, "loss": 0.4475, "step": 8649 }, { "epoch": 2.767704778156997, "grad_norm": 0.5044153928756714, "learning_rate": 1.810925161539945e-07, "loss": 0.4295, "step": 8650 }, { "epoch": 2.768024744027304, "grad_norm": 0.5254626870155334, "learning_rate": 1.8059632233629054e-07, "loss": 0.4999, "step": 8651 }, { "epoch": 2.768344709897611, "grad_norm": 0.479126900434494, "learning_rate": 1.801007967347046e-07, "loss": 0.4403, "step": 8652 }, { "epoch": 2.7686646757679183, "grad_norm": 0.4540811777114868, "learning_rate": 1.7960593941794113e-07, "loss": 0.4404, "step": 8653 }, { "epoch": 2.7689846416382253, "grad_norm": 0.46963462233543396, "learning_rate": 1.7911175045461305e-07, "loss": 0.4471, "step": 8654 }, { "epoch": 2.7693046075085324, "grad_norm": 0.48203012347221375, "learning_rate": 1.7861822991324107e-07, "loss": 0.4271, "step": 8655 }, { "epoch": 2.76962457337884, "grad_norm": 0.48530858755111694, "learning_rate": 1.781253778622516e-07, "loss": 0.4257, "step": 8656 }, { "epoch": 2.769944539249147, "grad_norm": 0.47368013858795166, "learning_rate": 1.776331943699794e-07, "loss": 0.4403, "step": 8657 }, { "epoch": 2.770264505119454, "grad_norm": 0.5373647809028625, "learning_rate": 1.7714167950466597e-07, "loss": 0.4392, "step": 8658 }, { "epoch": 2.7705844709897613, "grad_norm": 0.47049710154533386, "learning_rate": 1.7665083333446077e-07, "loss": 0.3905, "step": 8659 }, { "epoch": 2.7709044368600684, "grad_norm": 0.47844138741493225, "learning_rate": 1.7616065592742038e-07, "loss": 0.4395, "step": 8660 }, { "epoch": 2.7712244027303754, "grad_norm": 0.47040674090385437, "learning_rate": 1.7567114735150826e-07, "loss": 0.4368, "step": 8661 }, { "epoch": 2.7715443686006824, "grad_norm": 0.5076830983161926, "learning_rate": 1.7518230767459566e-07, "loss": 0.4793, "step": 8662 }, { "epoch": 2.77186433447099, "grad_norm": 0.4674657881259918, "learning_rate": 1.7469413696445946e-07, "loss": 0.4004, "step": 8663 }, { "epoch": 2.772184300341297, "grad_norm": 0.48944738507270813, "learning_rate": 1.7420663528878666e-07, "loss": 0.4275, "step": 8664 }, { "epoch": 2.772504266211604, "grad_norm": 0.5075875520706177, "learning_rate": 1.7371980271516987e-07, "loss": 0.4179, "step": 8665 }, { "epoch": 2.7728242320819114, "grad_norm": 0.46065860986709595, "learning_rate": 1.7323363931110836e-07, "loss": 0.458, "step": 8666 }, { "epoch": 2.7731441979522184, "grad_norm": 0.5107364654541016, "learning_rate": 1.7274814514400995e-07, "loss": 0.4784, "step": 8667 }, { "epoch": 2.7734641638225255, "grad_norm": 0.46682408452033997, "learning_rate": 1.7226332028118742e-07, "loss": 0.4453, "step": 8668 }, { "epoch": 2.7737841296928325, "grad_norm": 0.46138402819633484, "learning_rate": 1.7177916478986368e-07, "loss": 0.4547, "step": 8669 }, { "epoch": 2.77410409556314, "grad_norm": 0.47206464409828186, "learning_rate": 1.712956787371678e-07, "loss": 0.4108, "step": 8670 }, { "epoch": 2.774424061433447, "grad_norm": 0.4193723499774933, "learning_rate": 1.708128621901356e-07, "loss": 0.4219, "step": 8671 }, { "epoch": 2.774744027303754, "grad_norm": 0.5017774701118469, "learning_rate": 1.703307152157091e-07, "loss": 0.408, "step": 8672 }, { "epoch": 2.7750639931740615, "grad_norm": 0.5490996241569519, "learning_rate": 1.698492378807387e-07, "loss": 0.4885, "step": 8673 }, { "epoch": 2.7753839590443685, "grad_norm": 0.459357351064682, "learning_rate": 1.6936843025198268e-07, "loss": 0.3526, "step": 8674 }, { "epoch": 2.7757039249146755, "grad_norm": 0.5254474878311157, "learning_rate": 1.6888829239610548e-07, "loss": 0.4909, "step": 8675 }, { "epoch": 2.776023890784983, "grad_norm": 0.5045756101608276, "learning_rate": 1.6840882437967832e-07, "loss": 0.3988, "step": 8676 }, { "epoch": 2.77634385665529, "grad_norm": 0.47853589057922363, "learning_rate": 1.679300262691802e-07, "loss": 0.4152, "step": 8677 }, { "epoch": 2.776663822525597, "grad_norm": 0.4418056607246399, "learning_rate": 1.6745189813099694e-07, "loss": 0.4152, "step": 8678 }, { "epoch": 2.7769837883959045, "grad_norm": 0.5065884590148926, "learning_rate": 1.669744400314216e-07, "loss": 0.4627, "step": 8679 }, { "epoch": 2.7773037542662116, "grad_norm": 0.5764408707618713, "learning_rate": 1.6649765203665458e-07, "loss": 0.4382, "step": 8680 }, { "epoch": 2.7776237201365186, "grad_norm": 0.48257893323898315, "learning_rate": 1.6602153421280298e-07, "loss": 0.4424, "step": 8681 }, { "epoch": 2.777943686006826, "grad_norm": 0.5135520696640015, "learning_rate": 1.655460866258818e-07, "loss": 0.4749, "step": 8682 }, { "epoch": 2.778263651877133, "grad_norm": 0.4605827033519745, "learning_rate": 1.6507130934181048e-07, "loss": 0.4067, "step": 8683 }, { "epoch": 2.77858361774744, "grad_norm": 0.49601656198501587, "learning_rate": 1.6459720242641864e-07, "loss": 0.4145, "step": 8684 }, { "epoch": 2.7789035836177476, "grad_norm": 0.5408211946487427, "learning_rate": 1.6412376594544255e-07, "loss": 0.4534, "step": 8685 }, { "epoch": 2.7792235494880546, "grad_norm": 0.5210228562355042, "learning_rate": 1.6365099996452416e-07, "loss": 0.4529, "step": 8686 }, { "epoch": 2.7795435153583616, "grad_norm": 0.5135594606399536, "learning_rate": 1.6317890454921214e-07, "loss": 0.4649, "step": 8687 }, { "epoch": 2.779863481228669, "grad_norm": 0.5040867924690247, "learning_rate": 1.627074797649647e-07, "loss": 0.4136, "step": 8688 }, { "epoch": 2.780183447098976, "grad_norm": 0.5140244364738464, "learning_rate": 1.6223672567714342e-07, "loss": 0.41, "step": 8689 }, { "epoch": 2.780503412969283, "grad_norm": 0.5380143523216248, "learning_rate": 1.6176664235102114e-07, "loss": 0.4999, "step": 8690 }, { "epoch": 2.7808233788395906, "grad_norm": 0.46018654108047485, "learning_rate": 1.6129722985177453e-07, "loss": 0.4032, "step": 8691 }, { "epoch": 2.7811433447098977, "grad_norm": 0.4694654047489166, "learning_rate": 1.6082848824448715e-07, "loss": 0.4299, "step": 8692 }, { "epoch": 2.7814633105802047, "grad_norm": 0.47146186232566833, "learning_rate": 1.6036041759415255e-07, "loss": 0.4119, "step": 8693 }, { "epoch": 2.781783276450512, "grad_norm": 0.48588356375694275, "learning_rate": 1.5989301796566713e-07, "loss": 0.4093, "step": 8694 }, { "epoch": 2.782103242320819, "grad_norm": 0.4830414950847626, "learning_rate": 1.5942628942383854e-07, "loss": 0.4287, "step": 8695 }, { "epoch": 2.782423208191126, "grad_norm": 0.5008724927902222, "learning_rate": 1.589602320333783e-07, "loss": 0.422, "step": 8696 }, { "epoch": 2.7827431740614337, "grad_norm": 0.5461366772651672, "learning_rate": 1.584948458589053e-07, "loss": 0.4599, "step": 8697 }, { "epoch": 2.7830631399317407, "grad_norm": 0.557327926158905, "learning_rate": 1.580301309649468e-07, "loss": 0.4525, "step": 8698 }, { "epoch": 2.7833831058020477, "grad_norm": 0.549267590045929, "learning_rate": 1.575660874159346e-07, "loss": 0.43, "step": 8699 }, { "epoch": 2.783703071672355, "grad_norm": 0.5106770992279053, "learning_rate": 1.5710271527621112e-07, "loss": 0.4263, "step": 8700 }, { "epoch": 2.7840230375426622, "grad_norm": 0.5354407429695129, "learning_rate": 1.566400146100222e-07, "loss": 0.4589, "step": 8701 }, { "epoch": 2.7843430034129693, "grad_norm": 0.45193251967430115, "learning_rate": 1.561779854815215e-07, "loss": 0.4203, "step": 8702 }, { "epoch": 2.7846629692832767, "grad_norm": 0.4626031219959259, "learning_rate": 1.5571662795477116e-07, "loss": 0.4436, "step": 8703 }, { "epoch": 2.7849829351535837, "grad_norm": 0.5113726854324341, "learning_rate": 1.5525594209373716e-07, "loss": 0.4268, "step": 8704 }, { "epoch": 2.7853029010238908, "grad_norm": 0.5348974466323853, "learning_rate": 1.547959279622957e-07, "loss": 0.47, "step": 8705 }, { "epoch": 2.785622866894198, "grad_norm": 0.5126128792762756, "learning_rate": 1.5433658562422794e-07, "loss": 0.4555, "step": 8706 }, { "epoch": 2.7859428327645053, "grad_norm": 0.48683688044548035, "learning_rate": 1.5387791514322182e-07, "loss": 0.3778, "step": 8707 }, { "epoch": 2.7862627986348123, "grad_norm": 0.5526103973388672, "learning_rate": 1.5341991658287259e-07, "loss": 0.499, "step": 8708 }, { "epoch": 2.7865827645051193, "grad_norm": 0.49776914715766907, "learning_rate": 1.5296259000668279e-07, "loss": 0.4261, "step": 8709 }, { "epoch": 2.7869027303754264, "grad_norm": 0.5074841380119324, "learning_rate": 1.5250593547806114e-07, "loss": 0.4391, "step": 8710 }, { "epoch": 2.787222696245734, "grad_norm": 0.5573603510856628, "learning_rate": 1.520499530603231e-07, "loss": 0.4533, "step": 8711 }, { "epoch": 2.787542662116041, "grad_norm": 0.5001792311668396, "learning_rate": 1.515946428166909e-07, "loss": 0.398, "step": 8712 }, { "epoch": 2.787862627986348, "grad_norm": 0.49975937604904175, "learning_rate": 1.5114000481029456e-07, "loss": 0.4646, "step": 8713 }, { "epoch": 2.7881825938566553, "grad_norm": 0.482695609331131, "learning_rate": 1.506860391041698e-07, "loss": 0.4291, "step": 8714 }, { "epoch": 2.7885025597269624, "grad_norm": 0.49805188179016113, "learning_rate": 1.5023274576125958e-07, "loss": 0.4629, "step": 8715 }, { "epoch": 2.7888225255972694, "grad_norm": 0.5283095836639404, "learning_rate": 1.497801248444136e-07, "loss": 0.4183, "step": 8716 }, { "epoch": 2.789142491467577, "grad_norm": 0.5180644989013672, "learning_rate": 1.493281764163884e-07, "loss": 0.4163, "step": 8717 }, { "epoch": 2.789462457337884, "grad_norm": 0.48102542757987976, "learning_rate": 1.4887690053984716e-07, "loss": 0.4574, "step": 8718 }, { "epoch": 2.789782423208191, "grad_norm": 0.5150607228279114, "learning_rate": 1.4842629727735924e-07, "loss": 0.4088, "step": 8719 }, { "epoch": 2.7901023890784984, "grad_norm": 0.5212371945381165, "learning_rate": 1.4797636669140192e-07, "loss": 0.4726, "step": 8720 }, { "epoch": 2.7904223549488054, "grad_norm": 0.5373845100402832, "learning_rate": 1.4752710884435806e-07, "loss": 0.461, "step": 8721 }, { "epoch": 2.7907423208191124, "grad_norm": 0.4621168076992035, "learning_rate": 1.470785237985195e-07, "loss": 0.4207, "step": 8722 }, { "epoch": 2.79106228668942, "grad_norm": 0.48803746700286865, "learning_rate": 1.4663061161608095e-07, "loss": 0.4637, "step": 8723 }, { "epoch": 2.791382252559727, "grad_norm": 0.5152938961982727, "learning_rate": 1.4618337235914658e-07, "loss": 0.4399, "step": 8724 }, { "epoch": 2.791702218430034, "grad_norm": 0.47996997833251953, "learning_rate": 1.4573680608972796e-07, "loss": 0.4034, "step": 8725 }, { "epoch": 2.7920221843003414, "grad_norm": 0.5782660841941833, "learning_rate": 1.4529091286973994e-07, "loss": 0.5179, "step": 8726 }, { "epoch": 2.7923421501706485, "grad_norm": 0.5318395495414734, "learning_rate": 1.4484569276100756e-07, "loss": 0.4363, "step": 8727 }, { "epoch": 2.7926621160409555, "grad_norm": 0.5264575481414795, "learning_rate": 1.4440114582526089e-07, "loss": 0.5013, "step": 8728 }, { "epoch": 2.792982081911263, "grad_norm": 0.49003055691719055, "learning_rate": 1.4395727212413668e-07, "loss": 0.4107, "step": 8729 }, { "epoch": 2.79330204778157, "grad_norm": 0.4770995080471039, "learning_rate": 1.4351407171917908e-07, "loss": 0.4214, "step": 8730 }, { "epoch": 2.793622013651877, "grad_norm": 0.4444619119167328, "learning_rate": 1.4307154467183727e-07, "loss": 0.4126, "step": 8731 }, { "epoch": 2.7939419795221845, "grad_norm": 0.4984275996685028, "learning_rate": 1.426296910434688e-07, "loss": 0.4686, "step": 8732 }, { "epoch": 2.7942619453924915, "grad_norm": 0.4909839928150177, "learning_rate": 1.4218851089533802e-07, "loss": 0.4467, "step": 8733 }, { "epoch": 2.7945819112627985, "grad_norm": 0.4700436592102051, "learning_rate": 1.4174800428861324e-07, "loss": 0.4161, "step": 8734 }, { "epoch": 2.794901877133106, "grad_norm": 0.47440141439437866, "learning_rate": 1.4130817128437336e-07, "loss": 0.4766, "step": 8735 }, { "epoch": 2.795221843003413, "grad_norm": 0.49831926822662354, "learning_rate": 1.4086901194359958e-07, "loss": 0.4411, "step": 8736 }, { "epoch": 2.79554180887372, "grad_norm": 0.4541718065738678, "learning_rate": 1.4043052632718378e-07, "loss": 0.4208, "step": 8737 }, { "epoch": 2.7958617747440275, "grad_norm": 0.4604977071285248, "learning_rate": 1.3999271449592177e-07, "loss": 0.4676, "step": 8738 }, { "epoch": 2.7961817406143346, "grad_norm": 0.4659779965877533, "learning_rate": 1.395555765105161e-07, "loss": 0.387, "step": 8739 }, { "epoch": 2.7965017064846416, "grad_norm": 0.5293183326721191, "learning_rate": 1.3911911243157716e-07, "loss": 0.4838, "step": 8740 }, { "epoch": 2.796821672354949, "grad_norm": 0.5436280369758606, "learning_rate": 1.38683322319621e-07, "loss": 0.4305, "step": 8741 }, { "epoch": 2.797141638225256, "grad_norm": 0.5356222987174988, "learning_rate": 1.3824820623507095e-07, "loss": 0.4546, "step": 8742 }, { "epoch": 2.797461604095563, "grad_norm": 0.4895310699939728, "learning_rate": 1.3781376423825598e-07, "loss": 0.4064, "step": 8743 }, { "epoch": 2.7977815699658706, "grad_norm": 0.5118739008903503, "learning_rate": 1.3737999638941124e-07, "loss": 0.4892, "step": 8744 }, { "epoch": 2.7981015358361776, "grad_norm": 0.5162504315376282, "learning_rate": 1.3694690274868028e-07, "loss": 0.4285, "step": 8745 }, { "epoch": 2.7984215017064846, "grad_norm": 0.542949914932251, "learning_rate": 1.3651448337611117e-07, "loss": 0.4784, "step": 8746 }, { "epoch": 2.7987414675767917, "grad_norm": 0.5507196187973022, "learning_rate": 1.3608273833166096e-07, "loss": 0.4858, "step": 8747 }, { "epoch": 2.799061433447099, "grad_norm": 0.5387033224105835, "learning_rate": 1.3565166767519012e-07, "loss": 0.4182, "step": 8748 }, { "epoch": 2.799381399317406, "grad_norm": 0.5092856287956238, "learning_rate": 1.3522127146646691e-07, "loss": 0.4973, "step": 8749 }, { "epoch": 2.799701365187713, "grad_norm": 0.5302609801292419, "learning_rate": 1.3479154976516696e-07, "loss": 0.3835, "step": 8750 }, { "epoch": 2.80002133105802, "grad_norm": 0.4946901798248291, "learning_rate": 1.3436250263087204e-07, "loss": 0.4143, "step": 8751 }, { "epoch": 2.8003412969283277, "grad_norm": 0.46833574771881104, "learning_rate": 1.3393413012306955e-07, "loss": 0.4492, "step": 8752 }, { "epoch": 2.8006612627986347, "grad_norm": 0.42688828706741333, "learning_rate": 1.3350643230115424e-07, "loss": 0.4154, "step": 8753 }, { "epoch": 2.8009812286689417, "grad_norm": 0.5493572950363159, "learning_rate": 1.3307940922442585e-07, "loss": 0.5317, "step": 8754 }, { "epoch": 2.801301194539249, "grad_norm": 0.45775240659713745, "learning_rate": 1.3265306095209208e-07, "loss": 0.3769, "step": 8755 }, { "epoch": 2.801621160409556, "grad_norm": 0.5258458852767944, "learning_rate": 1.3222738754326836e-07, "loss": 0.4656, "step": 8756 }, { "epoch": 2.8019411262798632, "grad_norm": 0.5466823577880859, "learning_rate": 1.3180238905697306e-07, "loss": 0.4292, "step": 8757 }, { "epoch": 2.8022610921501707, "grad_norm": 0.5115330815315247, "learning_rate": 1.3137806555213295e-07, "loss": 0.4582, "step": 8758 }, { "epoch": 2.8025810580204777, "grad_norm": 0.44748345017433167, "learning_rate": 1.309544170875804e-07, "loss": 0.4004, "step": 8759 }, { "epoch": 2.8029010238907848, "grad_norm": 0.5248578190803528, "learning_rate": 1.3053144372205619e-07, "loss": 0.4329, "step": 8760 }, { "epoch": 2.8032209897610922, "grad_norm": 0.5367554426193237, "learning_rate": 1.301091455142056e-07, "loss": 0.4666, "step": 8761 }, { "epoch": 2.8035409556313993, "grad_norm": 0.5284185409545898, "learning_rate": 1.296875225225802e-07, "loss": 0.4363, "step": 8762 }, { "epoch": 2.8038609215017063, "grad_norm": 0.4763930141925812, "learning_rate": 1.292665748056393e-07, "loss": 0.3859, "step": 8763 }, { "epoch": 2.8041808873720138, "grad_norm": 0.49560821056365967, "learning_rate": 1.2884630242174734e-07, "loss": 0.431, "step": 8764 }, { "epoch": 2.804500853242321, "grad_norm": 0.4569384753704071, "learning_rate": 1.284267054291749e-07, "loss": 0.4236, "step": 8765 }, { "epoch": 2.804820819112628, "grad_norm": 0.4791736900806427, "learning_rate": 1.2800778388610157e-07, "loss": 0.4238, "step": 8766 }, { "epoch": 2.8051407849829353, "grad_norm": 0.4712306261062622, "learning_rate": 1.275895378506098e-07, "loss": 0.4277, "step": 8767 }, { "epoch": 2.8054607508532423, "grad_norm": 0.43769967555999756, "learning_rate": 1.2717196738069036e-07, "loss": 0.4321, "step": 8768 }, { "epoch": 2.8057807167235493, "grad_norm": 0.49380260705947876, "learning_rate": 1.2675507253423924e-07, "loss": 0.4275, "step": 8769 }, { "epoch": 2.806100682593857, "grad_norm": 0.5135136246681213, "learning_rate": 1.2633885336906014e-07, "loss": 0.4889, "step": 8770 }, { "epoch": 2.806420648464164, "grad_norm": 0.5384788513183594, "learning_rate": 1.2592330994286251e-07, "loss": 0.4352, "step": 8771 }, { "epoch": 2.806740614334471, "grad_norm": 0.4903396666049957, "learning_rate": 1.2550844231326188e-07, "loss": 0.4284, "step": 8772 }, { "epoch": 2.8070605802047783, "grad_norm": 0.5049971342086792, "learning_rate": 1.2509425053777947e-07, "loss": 0.4427, "step": 8773 }, { "epoch": 2.8073805460750854, "grad_norm": 0.4411507546901703, "learning_rate": 1.246807346738438e-07, "loss": 0.3914, "step": 8774 }, { "epoch": 2.8077005119453924, "grad_norm": 0.5614057779312134, "learning_rate": 1.2426789477878898e-07, "loss": 0.4433, "step": 8775 }, { "epoch": 2.8080204778157, "grad_norm": 0.46131646633148193, "learning_rate": 1.2385573090985703e-07, "loss": 0.4479, "step": 8776 }, { "epoch": 2.808340443686007, "grad_norm": 0.5110061168670654, "learning_rate": 1.234442431241939e-07, "loss": 0.4529, "step": 8777 }, { "epoch": 2.808660409556314, "grad_norm": 0.548893928527832, "learning_rate": 1.2303343147885283e-07, "loss": 0.5141, "step": 8778 }, { "epoch": 2.8089803754266214, "grad_norm": 0.5144462585449219, "learning_rate": 1.2262329603079382e-07, "loss": 0.3636, "step": 8779 }, { "epoch": 2.8093003412969284, "grad_norm": 0.5226725339889526, "learning_rate": 1.2221383683688192e-07, "loss": 0.4317, "step": 8780 }, { "epoch": 2.8096203071672354, "grad_norm": 0.4646565318107605, "learning_rate": 1.2180505395389063e-07, "loss": 0.4245, "step": 8781 }, { "epoch": 2.809940273037543, "grad_norm": 0.49531203508377075, "learning_rate": 1.2139694743849684e-07, "loss": 0.4439, "step": 8782 }, { "epoch": 2.81026023890785, "grad_norm": 0.5245881080627441, "learning_rate": 1.2098951734728525e-07, "loss": 0.4501, "step": 8783 }, { "epoch": 2.810580204778157, "grad_norm": 0.5053141117095947, "learning_rate": 1.2058276373674682e-07, "loss": 0.4572, "step": 8784 }, { "epoch": 2.8109001706484644, "grad_norm": 0.4849434792995453, "learning_rate": 1.2017668666327752e-07, "loss": 0.4485, "step": 8785 }, { "epoch": 2.8112201365187715, "grad_norm": 0.4660979211330414, "learning_rate": 1.1977128618318234e-07, "loss": 0.3886, "step": 8786 }, { "epoch": 2.8115401023890785, "grad_norm": 0.47356531023979187, "learning_rate": 1.193665623526691e-07, "loss": 0.4243, "step": 8787 }, { "epoch": 2.8118600682593855, "grad_norm": 0.46527403593063354, "learning_rate": 1.189625152278534e-07, "loss": 0.4342, "step": 8788 }, { "epoch": 2.812180034129693, "grad_norm": 0.5244778990745544, "learning_rate": 1.1855914486475661e-07, "loss": 0.4673, "step": 8789 }, { "epoch": 2.8125, "grad_norm": 0.4573207497596741, "learning_rate": 1.181564513193073e-07, "loss": 0.3831, "step": 8790 }, { "epoch": 2.812819965870307, "grad_norm": 0.448699951171875, "learning_rate": 1.177544346473386e-07, "loss": 0.3922, "step": 8791 }, { "epoch": 2.8131399317406145, "grad_norm": 0.5018029808998108, "learning_rate": 1.1735309490459201e-07, "loss": 0.4931, "step": 8792 }, { "epoch": 2.8134598976109215, "grad_norm": 0.4705091416835785, "learning_rate": 1.1695243214671192e-07, "loss": 0.3911, "step": 8793 }, { "epoch": 2.8137798634812285, "grad_norm": 0.4831168055534363, "learning_rate": 1.1655244642925113e-07, "loss": 0.4585, "step": 8794 }, { "epoch": 2.8140998293515356, "grad_norm": 0.46551117300987244, "learning_rate": 1.1615313780766967e-07, "loss": 0.4352, "step": 8795 }, { "epoch": 2.814419795221843, "grad_norm": 0.5043496489524841, "learning_rate": 1.1575450633733054e-07, "loss": 0.4108, "step": 8796 }, { "epoch": 2.81473976109215, "grad_norm": 0.5622084736824036, "learning_rate": 1.153565520735056e-07, "loss": 0.5166, "step": 8797 }, { "epoch": 2.815059726962457, "grad_norm": 0.45480015873908997, "learning_rate": 1.1495927507137073e-07, "loss": 0.3995, "step": 8798 }, { "epoch": 2.8153796928327646, "grad_norm": 0.5136435627937317, "learning_rate": 1.1456267538600797e-07, "loss": 0.4597, "step": 8799 }, { "epoch": 2.8156996587030716, "grad_norm": 0.5320642590522766, "learning_rate": 1.1416675307240887e-07, "loss": 0.4379, "step": 8800 }, { "epoch": 2.8160196245733786, "grad_norm": 0.4744340181350708, "learning_rate": 1.1377150818546734e-07, "loss": 0.4363, "step": 8801 }, { "epoch": 2.816339590443686, "grad_norm": 0.4676035940647125, "learning_rate": 1.1337694077998451e-07, "loss": 0.4385, "step": 8802 }, { "epoch": 2.816659556313993, "grad_norm": 0.4505445063114166, "learning_rate": 1.1298305091066664e-07, "loss": 0.4517, "step": 8803 }, { "epoch": 2.8169795221843, "grad_norm": 0.43211179971694946, "learning_rate": 1.1258983863212835e-07, "loss": 0.3816, "step": 8804 }, { "epoch": 2.8172994880546076, "grad_norm": 0.4516248404979706, "learning_rate": 1.121973039988894e-07, "loss": 0.4566, "step": 8805 }, { "epoch": 2.8176194539249146, "grad_norm": 0.48179879784584045, "learning_rate": 1.11805447065374e-07, "loss": 0.4273, "step": 8806 }, { "epoch": 2.8179394197952217, "grad_norm": 0.4546855390071869, "learning_rate": 1.1141426788591426e-07, "loss": 0.4601, "step": 8807 }, { "epoch": 2.818259385665529, "grad_norm": 0.5028133392333984, "learning_rate": 1.1102376651474678e-07, "loss": 0.4323, "step": 8808 }, { "epoch": 2.818579351535836, "grad_norm": 0.5004110932350159, "learning_rate": 1.1063394300601548e-07, "loss": 0.4164, "step": 8809 }, { "epoch": 2.818899317406143, "grad_norm": 0.5113457441329956, "learning_rate": 1.1024479741377048e-07, "loss": 0.4388, "step": 8810 }, { "epoch": 2.8192192832764507, "grad_norm": 0.5095594525337219, "learning_rate": 1.0985632979196692e-07, "loss": 0.4749, "step": 8811 }, { "epoch": 2.8195392491467577, "grad_norm": 0.46509113907814026, "learning_rate": 1.0946854019446673e-07, "loss": 0.429, "step": 8812 }, { "epoch": 2.8198592150170647, "grad_norm": 0.5171486735343933, "learning_rate": 1.090814286750358e-07, "loss": 0.4134, "step": 8813 }, { "epoch": 2.820179180887372, "grad_norm": 0.5246515274047852, "learning_rate": 1.0869499528734894e-07, "loss": 0.4441, "step": 8814 }, { "epoch": 2.820499146757679, "grad_norm": 0.49848657846450806, "learning_rate": 1.0830924008498612e-07, "loss": 0.4639, "step": 8815 }, { "epoch": 2.8208191126279862, "grad_norm": 0.5230051279067993, "learning_rate": 1.0792416312143172e-07, "loss": 0.3844, "step": 8816 }, { "epoch": 2.8211390784982937, "grad_norm": 0.5380415916442871, "learning_rate": 1.0753976445007696e-07, "loss": 0.4908, "step": 8817 }, { "epoch": 2.8214590443686007, "grad_norm": 0.4953198730945587, "learning_rate": 1.0715604412421976e-07, "loss": 0.4454, "step": 8818 }, { "epoch": 2.8217790102389078, "grad_norm": 0.5087031126022339, "learning_rate": 1.0677300219706311e-07, "loss": 0.3919, "step": 8819 }, { "epoch": 2.8220989761092152, "grad_norm": 0.45869845151901245, "learning_rate": 1.0639063872171674e-07, "loss": 0.4644, "step": 8820 }, { "epoch": 2.8224189419795223, "grad_norm": 0.4913416802883148, "learning_rate": 1.0600895375119602e-07, "loss": 0.4402, "step": 8821 }, { "epoch": 2.8227389078498293, "grad_norm": 0.5042554140090942, "learning_rate": 1.0562794733842086e-07, "loss": 0.4958, "step": 8822 }, { "epoch": 2.8230588737201368, "grad_norm": 0.5037972927093506, "learning_rate": 1.0524761953621843e-07, "loss": 0.414, "step": 8823 }, { "epoch": 2.823378839590444, "grad_norm": 0.4720671474933624, "learning_rate": 1.0486797039732266e-07, "loss": 0.3998, "step": 8824 }, { "epoch": 2.823698805460751, "grad_norm": 0.4891723692417145, "learning_rate": 1.0448899997437145e-07, "loss": 0.409, "step": 8825 }, { "epoch": 2.8240187713310583, "grad_norm": 0.49462926387786865, "learning_rate": 1.0411070831991055e-07, "loss": 0.4409, "step": 8826 }, { "epoch": 2.8243387372013653, "grad_norm": 0.4781778156757355, "learning_rate": 1.0373309548638966e-07, "loss": 0.4708, "step": 8827 }, { "epoch": 2.8246587030716723, "grad_norm": 0.4797002375125885, "learning_rate": 1.0335616152616468e-07, "loss": 0.4244, "step": 8828 }, { "epoch": 2.82497866894198, "grad_norm": 0.42282843589782715, "learning_rate": 1.029799064914988e-07, "loss": 0.4112, "step": 8829 }, { "epoch": 2.825298634812287, "grad_norm": 0.5273585319519043, "learning_rate": 1.0260433043456086e-07, "loss": 0.4617, "step": 8830 }, { "epoch": 2.825618600682594, "grad_norm": 0.47264957427978516, "learning_rate": 1.0222943340742365e-07, "loss": 0.4212, "step": 8831 }, { "epoch": 2.825938566552901, "grad_norm": 0.4755886197090149, "learning_rate": 1.0185521546206834e-07, "loss": 0.4078, "step": 8832 }, { "epoch": 2.8262585324232083, "grad_norm": 0.47338294982910156, "learning_rate": 1.01481676650379e-07, "loss": 0.4228, "step": 8833 }, { "epoch": 2.8265784982935154, "grad_norm": 0.4957878291606903, "learning_rate": 1.0110881702414865e-07, "loss": 0.4529, "step": 8834 }, { "epoch": 2.8268984641638224, "grad_norm": 0.5062832236289978, "learning_rate": 1.0073663663507482e-07, "loss": 0.4709, "step": 8835 }, { "epoch": 2.8272184300341294, "grad_norm": 0.483052134513855, "learning_rate": 1.0036513553476013e-07, "loss": 0.4047, "step": 8836 }, { "epoch": 2.827538395904437, "grad_norm": 0.5118340253829956, "learning_rate": 9.999431377471391e-08, "loss": 0.4398, "step": 8837 }, { "epoch": 2.827858361774744, "grad_norm": 0.5285395383834839, "learning_rate": 9.962417140635006e-08, "loss": 0.4436, "step": 8838 }, { "epoch": 2.828178327645051, "grad_norm": 0.5113656520843506, "learning_rate": 9.925470848099083e-08, "loss": 0.4113, "step": 8839 }, { "epoch": 2.8284982935153584, "grad_norm": 0.5032393336296082, "learning_rate": 9.888592504986194e-08, "loss": 0.4706, "step": 8840 }, { "epoch": 2.8288182593856654, "grad_norm": 0.5037457346916199, "learning_rate": 9.851782116409581e-08, "loss": 0.5053, "step": 8841 }, { "epoch": 2.8291382252559725, "grad_norm": 0.46051284670829773, "learning_rate": 9.81503968747305e-08, "loss": 0.4571, "step": 8842 }, { "epoch": 2.82945819112628, "grad_norm": 0.4622531533241272, "learning_rate": 9.778365223270914e-08, "loss": 0.431, "step": 8843 }, { "epoch": 2.829778156996587, "grad_norm": 0.476308673620224, "learning_rate": 9.741758728888218e-08, "loss": 0.4512, "step": 8844 }, { "epoch": 2.830098122866894, "grad_norm": 0.45065200328826904, "learning_rate": 9.705220209400457e-08, "loss": 0.4513, "step": 8845 }, { "epoch": 2.8304180887372015, "grad_norm": 0.4600454866886139, "learning_rate": 9.6687496698738e-08, "loss": 0.4176, "step": 8846 }, { "epoch": 2.8307380546075085, "grad_norm": 0.48312658071517944, "learning_rate": 9.63234711536487e-08, "loss": 0.5039, "step": 8847 }, { "epoch": 2.8310580204778155, "grad_norm": 0.5284901857376099, "learning_rate": 9.596012550920907e-08, "loss": 0.417, "step": 8848 }, { "epoch": 2.831377986348123, "grad_norm": 0.4941372275352478, "learning_rate": 9.559745981579771e-08, "loss": 0.482, "step": 8849 }, { "epoch": 2.83169795221843, "grad_norm": 0.5106128454208374, "learning_rate": 9.523547412369937e-08, "loss": 0.4064, "step": 8850 }, { "epoch": 2.832017918088737, "grad_norm": 0.5166605114936829, "learning_rate": 9.487416848310283e-08, "loss": 0.415, "step": 8851 }, { "epoch": 2.8323378839590445, "grad_norm": 0.5388782024383545, "learning_rate": 9.451354294410353e-08, "loss": 0.4282, "step": 8852 }, { "epoch": 2.8326578498293515, "grad_norm": 0.49470970034599304, "learning_rate": 9.415359755670317e-08, "loss": 0.4512, "step": 8853 }, { "epoch": 2.8329778156996586, "grad_norm": 0.4966568648815155, "learning_rate": 9.379433237080793e-08, "loss": 0.4595, "step": 8854 }, { "epoch": 2.833297781569966, "grad_norm": 0.4812455177307129, "learning_rate": 9.34357474362313e-08, "loss": 0.4621, "step": 8855 }, { "epoch": 2.833617747440273, "grad_norm": 0.5319587588310242, "learning_rate": 9.307784280269128e-08, "loss": 0.4628, "step": 8856 }, { "epoch": 2.83393771331058, "grad_norm": 0.4560822546482086, "learning_rate": 9.272061851981151e-08, "loss": 0.3885, "step": 8857 }, { "epoch": 2.8342576791808876, "grad_norm": 0.49199801683425903, "learning_rate": 9.236407463712071e-08, "loss": 0.4398, "step": 8858 }, { "epoch": 2.8345776450511946, "grad_norm": 0.4829440414905548, "learning_rate": 9.200821120405545e-08, "loss": 0.3572, "step": 8859 }, { "epoch": 2.8348976109215016, "grad_norm": 0.5515627861022949, "learning_rate": 9.165302826995626e-08, "loss": 0.5003, "step": 8860 }, { "epoch": 2.835217576791809, "grad_norm": 0.6842747330665588, "learning_rate": 9.129852588406985e-08, "loss": 0.4109, "step": 8861 }, { "epoch": 2.835537542662116, "grad_norm": 0.5266549587249756, "learning_rate": 9.094470409554746e-08, "loss": 0.4544, "step": 8862 }, { "epoch": 2.835857508532423, "grad_norm": 0.45223504304885864, "learning_rate": 9.059156295344818e-08, "loss": 0.4239, "step": 8863 }, { "epoch": 2.8361774744027306, "grad_norm": 0.5521233677864075, "learning_rate": 9.023910250673451e-08, "loss": 0.4705, "step": 8864 }, { "epoch": 2.8364974402730376, "grad_norm": 0.5143077969551086, "learning_rate": 8.988732280427625e-08, "loss": 0.4167, "step": 8865 }, { "epoch": 2.8368174061433447, "grad_norm": 0.4757063090801239, "learning_rate": 8.953622389484774e-08, "loss": 0.4431, "step": 8866 }, { "epoch": 2.837137372013652, "grad_norm": 0.47248420119285583, "learning_rate": 8.918580582712833e-08, "loss": 0.4305, "step": 8867 }, { "epoch": 2.837457337883959, "grad_norm": 0.5008695721626282, "learning_rate": 8.883606864970585e-08, "loss": 0.4174, "step": 8868 }, { "epoch": 2.837777303754266, "grad_norm": 0.5308582782745361, "learning_rate": 8.848701241106982e-08, "loss": 0.4573, "step": 8869 }, { "epoch": 2.8380972696245736, "grad_norm": 0.4937674403190613, "learning_rate": 8.813863715961879e-08, "loss": 0.4102, "step": 8870 }, { "epoch": 2.8384172354948807, "grad_norm": 0.50037682056427, "learning_rate": 8.77909429436552e-08, "loss": 0.4047, "step": 8871 }, { "epoch": 2.8387372013651877, "grad_norm": 0.577347457408905, "learning_rate": 8.744392981138605e-08, "loss": 0.5141, "step": 8872 }, { "epoch": 2.8390571672354947, "grad_norm": 0.5387556552886963, "learning_rate": 8.709759781092675e-08, "loss": 0.4247, "step": 8873 }, { "epoch": 2.839377133105802, "grad_norm": 0.485126256942749, "learning_rate": 8.675194699029499e-08, "loss": 0.4278, "step": 8874 }, { "epoch": 2.8396970989761092, "grad_norm": 0.5208066701889038, "learning_rate": 8.640697739741744e-08, "loss": 0.4033, "step": 8875 }, { "epoch": 2.8400170648464163, "grad_norm": 0.5068288445472717, "learning_rate": 8.606268908012305e-08, "loss": 0.4318, "step": 8876 }, { "epoch": 2.8403370307167233, "grad_norm": 0.470518559217453, "learning_rate": 8.571908208614866e-08, "loss": 0.458, "step": 8877 }, { "epoch": 2.8406569965870307, "grad_norm": 0.44177472591400146, "learning_rate": 8.537615646313557e-08, "loss": 0.4154, "step": 8878 }, { "epoch": 2.8409769624573378, "grad_norm": 0.4870431423187256, "learning_rate": 8.503391225863078e-08, "loss": 0.4035, "step": 8879 }, { "epoch": 2.841296928327645, "grad_norm": 0.4561026394367218, "learning_rate": 8.469234952008687e-08, "loss": 0.4108, "step": 8880 }, { "epoch": 2.8416168941979523, "grad_norm": 0.4789717495441437, "learning_rate": 8.435146829486263e-08, "loss": 0.4655, "step": 8881 }, { "epoch": 2.8419368600682593, "grad_norm": 0.48523077368736267, "learning_rate": 8.401126863021969e-08, "loss": 0.4506, "step": 8882 }, { "epoch": 2.8422568259385663, "grad_norm": 0.46199101209640503, "learning_rate": 8.367175057332976e-08, "loss": 0.4377, "step": 8883 }, { "epoch": 2.842576791808874, "grad_norm": 0.48472705483436584, "learning_rate": 8.33329141712652e-08, "loss": 0.4305, "step": 8884 }, { "epoch": 2.842896757679181, "grad_norm": 0.5093603134155273, "learning_rate": 8.299475947100733e-08, "loss": 0.4637, "step": 8885 }, { "epoch": 2.843216723549488, "grad_norm": 0.4923092722892761, "learning_rate": 8.265728651944138e-08, "loss": 0.4038, "step": 8886 }, { "epoch": 2.8435366894197953, "grad_norm": 0.48508721590042114, "learning_rate": 8.232049536335774e-08, "loss": 0.4471, "step": 8887 }, { "epoch": 2.8438566552901023, "grad_norm": 0.4965299963951111, "learning_rate": 8.198438604945457e-08, "loss": 0.4184, "step": 8888 }, { "epoch": 2.8441766211604094, "grad_norm": 0.5033362507820129, "learning_rate": 8.164895862433186e-08, "loss": 0.4325, "step": 8889 }, { "epoch": 2.844496587030717, "grad_norm": 0.5296280980110168, "learning_rate": 8.13142131344985e-08, "loss": 0.4991, "step": 8890 }, { "epoch": 2.844816552901024, "grad_norm": 0.549327552318573, "learning_rate": 8.09801496263668e-08, "loss": 0.4385, "step": 8891 }, { "epoch": 2.845136518771331, "grad_norm": 0.503280520439148, "learning_rate": 8.064676814625472e-08, "loss": 0.4513, "step": 8892 }, { "epoch": 2.8454564846416384, "grad_norm": 0.4724968373775482, "learning_rate": 8.031406874038694e-08, "loss": 0.3853, "step": 8893 }, { "epoch": 2.8457764505119454, "grad_norm": 0.5086998343467712, "learning_rate": 7.998205145489157e-08, "loss": 0.467, "step": 8894 }, { "epoch": 2.8460964163822524, "grad_norm": 0.5187538266181946, "learning_rate": 7.965071633580401e-08, "loss": 0.4731, "step": 8895 }, { "epoch": 2.84641638225256, "grad_norm": 0.511500358581543, "learning_rate": 7.932006342906361e-08, "loss": 0.4319, "step": 8896 }, { "epoch": 2.846736348122867, "grad_norm": 0.4847926199436188, "learning_rate": 7.89900927805165e-08, "loss": 0.4282, "step": 8897 }, { "epoch": 2.847056313993174, "grad_norm": 0.4554595947265625, "learning_rate": 7.866080443591273e-08, "loss": 0.4007, "step": 8898 }, { "epoch": 2.8473762798634814, "grad_norm": 0.5097051858901978, "learning_rate": 7.833219844090911e-08, "loss": 0.4787, "step": 8899 }, { "epoch": 2.8476962457337884, "grad_norm": 0.5208517909049988, "learning_rate": 7.800427484106754e-08, "loss": 0.457, "step": 8900 }, { "epoch": 2.8480162116040955, "grad_norm": 0.5112928748130798, "learning_rate": 7.767703368185386e-08, "loss": 0.4765, "step": 8901 }, { "epoch": 2.848336177474403, "grad_norm": 0.4776992201805115, "learning_rate": 7.735047500864179e-08, "loss": 0.4128, "step": 8902 }, { "epoch": 2.84865614334471, "grad_norm": 0.49549901485443115, "learning_rate": 7.702459886670788e-08, "loss": 0.5057, "step": 8903 }, { "epoch": 2.848976109215017, "grad_norm": 0.48889100551605225, "learning_rate": 7.669940530123598e-08, "loss": 0.4079, "step": 8904 }, { "epoch": 2.8492960750853245, "grad_norm": 0.5252945423126221, "learning_rate": 7.637489435731505e-08, "loss": 0.4028, "step": 8905 }, { "epoch": 2.8496160409556315, "grad_norm": 0.5165454745292664, "learning_rate": 7.60510660799374e-08, "loss": 0.4237, "step": 8906 }, { "epoch": 2.8499360068259385, "grad_norm": 0.5192331075668335, "learning_rate": 7.572792051400325e-08, "loss": 0.4722, "step": 8907 }, { "epoch": 2.850255972696246, "grad_norm": 0.4691384732723236, "learning_rate": 7.540545770431729e-08, "loss": 0.3964, "step": 8908 }, { "epoch": 2.850575938566553, "grad_norm": 0.5160424709320068, "learning_rate": 7.508367769558877e-08, "loss": 0.4559, "step": 8909 }, { "epoch": 2.85089590443686, "grad_norm": 0.4484829902648926, "learning_rate": 7.476258053243368e-08, "loss": 0.3618, "step": 8910 }, { "epoch": 2.8512158703071675, "grad_norm": 0.48174214363098145, "learning_rate": 7.44421662593714e-08, "loss": 0.4129, "step": 8911 }, { "epoch": 2.8515358361774745, "grad_norm": 0.4934282600879669, "learning_rate": 7.41224349208286e-08, "loss": 0.4777, "step": 8912 }, { "epoch": 2.8518558020477816, "grad_norm": 0.49322712421417236, "learning_rate": 7.38033865611365e-08, "loss": 0.415, "step": 8913 }, { "epoch": 2.8521757679180886, "grad_norm": 0.48833104968070984, "learning_rate": 7.348502122453083e-08, "loss": 0.4436, "step": 8914 }, { "epoch": 2.852495733788396, "grad_norm": 0.46358126401901245, "learning_rate": 7.316733895515349e-08, "loss": 0.465, "step": 8915 }, { "epoch": 2.852815699658703, "grad_norm": 0.46667855978012085, "learning_rate": 7.285033979705203e-08, "loss": 0.4088, "step": 8916 }, { "epoch": 2.85313566552901, "grad_norm": 0.5348223447799683, "learning_rate": 7.253402379417851e-08, "loss": 0.4822, "step": 8917 }, { "epoch": 2.8534556313993176, "grad_norm": 0.47619935870170593, "learning_rate": 7.221839099039063e-08, "loss": 0.4191, "step": 8918 }, { "epoch": 2.8537755972696246, "grad_norm": 0.4943060576915741, "learning_rate": 7.19034414294506e-08, "loss": 0.4274, "step": 8919 }, { "epoch": 2.8540955631399316, "grad_norm": 0.47631967067718506, "learning_rate": 7.158917515502739e-08, "loss": 0.4444, "step": 8920 }, { "epoch": 2.8544155290102387, "grad_norm": 0.5007773637771606, "learning_rate": 7.127559221069391e-08, "loss": 0.4482, "step": 8921 }, { "epoch": 2.854735494880546, "grad_norm": 0.511054515838623, "learning_rate": 7.096269263992928e-08, "loss": 0.4479, "step": 8922 }, { "epoch": 2.855055460750853, "grad_norm": 0.510521411895752, "learning_rate": 7.065047648611711e-08, "loss": 0.4566, "step": 8923 }, { "epoch": 2.85537542662116, "grad_norm": 0.5748690366744995, "learning_rate": 7.033894379254558e-08, "loss": 0.4625, "step": 8924 }, { "epoch": 2.8556953924914676, "grad_norm": 0.4568046033382416, "learning_rate": 7.002809460241123e-08, "loss": 0.3904, "step": 8925 }, { "epoch": 2.8560153583617747, "grad_norm": 0.4974222481250763, "learning_rate": 6.97179289588118e-08, "loss": 0.4638, "step": 8926 }, { "epoch": 2.8563353242320817, "grad_norm": 0.5003941655158997, "learning_rate": 6.940844690475291e-08, "loss": 0.4169, "step": 8927 }, { "epoch": 2.856655290102389, "grad_norm": 0.49940621852874756, "learning_rate": 6.909964848314465e-08, "loss": 0.4534, "step": 8928 }, { "epoch": 2.856975255972696, "grad_norm": 0.493630588054657, "learning_rate": 6.879153373680169e-08, "loss": 0.4351, "step": 8929 }, { "epoch": 2.857295221843003, "grad_norm": 0.43891242146492004, "learning_rate": 6.848410270844596e-08, "loss": 0.411, "step": 8930 }, { "epoch": 2.8576151877133107, "grad_norm": 0.4711717367172241, "learning_rate": 6.817735544070114e-08, "loss": 0.4678, "step": 8931 }, { "epoch": 2.8579351535836177, "grad_norm": 0.483100950717926, "learning_rate": 6.787129197609988e-08, "loss": 0.4362, "step": 8932 }, { "epoch": 2.8582551194539247, "grad_norm": 0.5003790855407715, "learning_rate": 6.75659123570771e-08, "loss": 0.4405, "step": 8933 }, { "epoch": 2.858575085324232, "grad_norm": 0.5048513412475586, "learning_rate": 6.726121662597396e-08, "loss": 0.4123, "step": 8934 }, { "epoch": 2.8588950511945392, "grad_norm": 0.5274044275283813, "learning_rate": 6.695720482503776e-08, "loss": 0.5795, "step": 8935 }, { "epoch": 2.8592150170648463, "grad_norm": 0.46992045640945435, "learning_rate": 6.665387699641979e-08, "loss": 0.3656, "step": 8936 }, { "epoch": 2.8595349829351537, "grad_norm": 0.5334693789482117, "learning_rate": 6.635123318217696e-08, "loss": 0.5043, "step": 8937 }, { "epoch": 2.8598549488054608, "grad_norm": 0.4710192084312439, "learning_rate": 6.60492734242707e-08, "loss": 0.4081, "step": 8938 }, { "epoch": 2.860174914675768, "grad_norm": 0.4978450834751129, "learning_rate": 6.574799776456864e-08, "loss": 0.4646, "step": 8939 }, { "epoch": 2.8604948805460753, "grad_norm": 0.4480099380016327, "learning_rate": 6.544740624484236e-08, "loss": 0.4159, "step": 8940 }, { "epoch": 2.8608148464163823, "grad_norm": 0.4901712238788605, "learning_rate": 6.514749890676963e-08, "loss": 0.4074, "step": 8941 }, { "epoch": 2.8611348122866893, "grad_norm": 0.4576753079891205, "learning_rate": 6.484827579193331e-08, "loss": 0.4081, "step": 8942 }, { "epoch": 2.861454778156997, "grad_norm": 0.48580124974250793, "learning_rate": 6.454973694182021e-08, "loss": 0.4985, "step": 8943 }, { "epoch": 2.861774744027304, "grad_norm": 0.4531460702419281, "learning_rate": 6.425188239782387e-08, "loss": 0.3939, "step": 8944 }, { "epoch": 2.862094709897611, "grad_norm": 0.507771372795105, "learning_rate": 6.395471220124183e-08, "loss": 0.4617, "step": 8945 }, { "epoch": 2.8624146757679183, "grad_norm": 0.4530040919780731, "learning_rate": 6.365822639327724e-08, "loss": 0.3676, "step": 8946 }, { "epoch": 2.8627346416382253, "grad_norm": 0.5020298361778259, "learning_rate": 6.33624250150383e-08, "loss": 0.4218, "step": 8947 }, { "epoch": 2.8630546075085324, "grad_norm": 0.5119328498840332, "learning_rate": 6.306730810753725e-08, "loss": 0.4719, "step": 8948 }, { "epoch": 2.86337457337884, "grad_norm": 0.49358975887298584, "learning_rate": 6.27728757116941e-08, "loss": 0.4098, "step": 8949 }, { "epoch": 2.863694539249147, "grad_norm": 0.47846266627311707, "learning_rate": 6.247912786833066e-08, "loss": 0.4491, "step": 8950 }, { "epoch": 2.864014505119454, "grad_norm": 0.46984216570854187, "learning_rate": 6.218606461817656e-08, "loss": 0.4107, "step": 8951 }, { "epoch": 2.8643344709897613, "grad_norm": 0.5021507143974304, "learning_rate": 6.18936860018654e-08, "loss": 0.4461, "step": 8952 }, { "epoch": 2.8646544368600684, "grad_norm": 0.5089204907417297, "learning_rate": 6.160199205993422e-08, "loss": 0.481, "step": 8953 }, { "epoch": 2.8649744027303754, "grad_norm": 0.551600992679596, "learning_rate": 6.131098283282899e-08, "loss": 0.4607, "step": 8954 }, { "epoch": 2.8652943686006824, "grad_norm": 0.5164521932601929, "learning_rate": 6.102065836089632e-08, "loss": 0.438, "step": 8955 }, { "epoch": 2.86561433447099, "grad_norm": 0.45446866750717163, "learning_rate": 6.07310186843918e-08, "loss": 0.412, "step": 8956 }, { "epoch": 2.865934300341297, "grad_norm": 0.495629221200943, "learning_rate": 6.044206384347384e-08, "loss": 0.4731, "step": 8957 }, { "epoch": 2.866254266211604, "grad_norm": 0.4896862804889679, "learning_rate": 6.015379387820541e-08, "loss": 0.4561, "step": 8958 }, { "epoch": 2.8665742320819114, "grad_norm": 0.5254265666007996, "learning_rate": 5.986620882855676e-08, "loss": 0.4918, "step": 8959 }, { "epoch": 2.8668941979522184, "grad_norm": 0.46429502964019775, "learning_rate": 5.9579308734400984e-08, "loss": 0.4501, "step": 8960 }, { "epoch": 2.8672141638225255, "grad_norm": 0.4449886083602905, "learning_rate": 5.929309363551794e-08, "loss": 0.4275, "step": 8961 }, { "epoch": 2.8675341296928325, "grad_norm": 0.5188478231430054, "learning_rate": 5.900756357159143e-08, "loss": 0.4413, "step": 8962 }, { "epoch": 2.86785409556314, "grad_norm": 0.5296725034713745, "learning_rate": 5.8722718582209796e-08, "loss": 0.4131, "step": 8963 }, { "epoch": 2.868174061433447, "grad_norm": 0.5184751152992249, "learning_rate": 5.8438558706868095e-08, "loss": 0.493, "step": 8964 }, { "epoch": 2.868494027303754, "grad_norm": 0.502002477645874, "learning_rate": 5.8155083984964835e-08, "loss": 0.4376, "step": 8965 }, { "epoch": 2.8688139931740615, "grad_norm": 0.5247063636779785, "learning_rate": 5.7872294455804666e-08, "loss": 0.4131, "step": 8966 }, { "epoch": 2.8691339590443685, "grad_norm": 0.4924825429916382, "learning_rate": 5.7590190158596235e-08, "loss": 0.486, "step": 8967 }, { "epoch": 2.8694539249146755, "grad_norm": 0.506610631942749, "learning_rate": 5.730877113245381e-08, "loss": 0.4687, "step": 8968 }, { "epoch": 2.869773890784983, "grad_norm": 0.45550239086151123, "learning_rate": 5.702803741639618e-08, "loss": 0.3963, "step": 8969 }, { "epoch": 2.87009385665529, "grad_norm": 0.4927195608615875, "learning_rate": 5.6747989049348304e-08, "loss": 0.415, "step": 8970 }, { "epoch": 2.870413822525597, "grad_norm": 0.47503048181533813, "learning_rate": 5.646862607013859e-08, "loss": 0.457, "step": 8971 }, { "epoch": 2.8707337883959045, "grad_norm": 0.5261685252189636, "learning_rate": 5.618994851750104e-08, "loss": 0.4459, "step": 8972 }, { "epoch": 2.8710537542662116, "grad_norm": 0.47380179166793823, "learning_rate": 5.5911956430074187e-08, "loss": 0.4289, "step": 8973 }, { "epoch": 2.8713737201365186, "grad_norm": 0.4902719557285309, "learning_rate": 5.563464984640221e-08, "loss": 0.4489, "step": 8974 }, { "epoch": 2.871693686006826, "grad_norm": 0.48171642422676086, "learning_rate": 5.535802880493491e-08, "loss": 0.4179, "step": 8975 }, { "epoch": 2.872013651877133, "grad_norm": 0.5338441729545593, "learning_rate": 5.508209334402548e-08, "loss": 0.4554, "step": 8976 }, { "epoch": 2.87233361774744, "grad_norm": 0.46447697281837463, "learning_rate": 5.4806843501932774e-08, "loss": 0.438, "step": 8977 }, { "epoch": 2.8726535836177476, "grad_norm": 0.5013840794563293, "learning_rate": 5.453227931682015e-08, "loss": 0.4158, "step": 8978 }, { "epoch": 2.8729735494880546, "grad_norm": 0.5281605124473572, "learning_rate": 5.42584008267566e-08, "loss": 0.4786, "step": 8979 }, { "epoch": 2.8732935153583616, "grad_norm": 0.4782954156398773, "learning_rate": 5.398520806971563e-08, "loss": 0.3993, "step": 8980 }, { "epoch": 2.873613481228669, "grad_norm": 0.5424149036407471, "learning_rate": 5.371270108357585e-08, "loss": 0.4876, "step": 8981 }, { "epoch": 2.873933447098976, "grad_norm": 0.4415556788444519, "learning_rate": 5.3440879906120344e-08, "loss": 0.4059, "step": 8982 }, { "epoch": 2.874253412969283, "grad_norm": 0.4531678259372711, "learning_rate": 5.316974457503787e-08, "loss": 0.4149, "step": 8983 }, { "epoch": 2.8745733788395906, "grad_norm": 0.543174147605896, "learning_rate": 5.2899295127921133e-08, "loss": 0.4273, "step": 8984 }, { "epoch": 2.8748933447098977, "grad_norm": 0.549419105052948, "learning_rate": 5.262953160226958e-08, "loss": 0.4826, "step": 8985 }, { "epoch": 2.8752133105802047, "grad_norm": 0.48486778140068054, "learning_rate": 5.23604540354844e-08, "loss": 0.4248, "step": 8986 }, { "epoch": 2.875533276450512, "grad_norm": 0.6450443863868713, "learning_rate": 5.2092062464875194e-08, "loss": 0.4711, "step": 8987 }, { "epoch": 2.875853242320819, "grad_norm": 0.492783784866333, "learning_rate": 5.182435692765331e-08, "loss": 0.4773, "step": 8988 }, { "epoch": 2.876173208191126, "grad_norm": 0.47193995118141174, "learning_rate": 5.155733746093683e-08, "loss": 0.4228, "step": 8989 }, { "epoch": 2.8764931740614337, "grad_norm": 0.4988788962364197, "learning_rate": 5.1291004101749476e-08, "loss": 0.5133, "step": 8990 }, { "epoch": 2.8768131399317407, "grad_norm": 0.5246462821960449, "learning_rate": 5.1025356887017815e-08, "loss": 0.4284, "step": 8991 }, { "epoch": 2.8771331058020477, "grad_norm": 0.47200262546539307, "learning_rate": 5.076039585357406e-08, "loss": 0.3902, "step": 8992 }, { "epoch": 2.877453071672355, "grad_norm": 0.5065959692001343, "learning_rate": 5.049612103815493e-08, "loss": 0.4616, "step": 8993 }, { "epoch": 2.8777730375426622, "grad_norm": 0.5456526279449463, "learning_rate": 5.023253247740334e-08, "loss": 0.4638, "step": 8994 }, { "epoch": 2.8780930034129693, "grad_norm": 0.4624581038951874, "learning_rate": 4.996963020786616e-08, "loss": 0.4235, "step": 8995 }, { "epoch": 2.8784129692832767, "grad_norm": 0.5095905065536499, "learning_rate": 4.970741426599479e-08, "loss": 0.5033, "step": 8996 }, { "epoch": 2.8787329351535837, "grad_norm": 0.47514355182647705, "learning_rate": 4.944588468814571e-08, "loss": 0.4508, "step": 8997 }, { "epoch": 2.8790529010238908, "grad_norm": 0.4647156596183777, "learning_rate": 4.91850415105799e-08, "loss": 0.4389, "step": 8998 }, { "epoch": 2.879372866894198, "grad_norm": 0.5182019472122192, "learning_rate": 4.8924884769464e-08, "loss": 0.5105, "step": 8999 }, { "epoch": 2.8796928327645053, "grad_norm": 0.4541059732437134, "learning_rate": 4.86654145008697e-08, "loss": 0.3809, "step": 9000 }, { "epoch": 2.8800127986348123, "grad_norm": 0.5203688144683838, "learning_rate": 4.840663074077212e-08, "loss": 0.4513, "step": 9001 }, { "epoch": 2.8803327645051193, "grad_norm": 0.5316109657287598, "learning_rate": 4.814853352505255e-08, "loss": 0.4178, "step": 9002 }, { "epoch": 2.8806527303754264, "grad_norm": 0.5072743892669678, "learning_rate": 4.789112288949515e-08, "loss": 0.4966, "step": 9003 }, { "epoch": 2.880972696245734, "grad_norm": 0.45236936211586, "learning_rate": 4.7634398869791375e-08, "loss": 0.4092, "step": 9004 }, { "epoch": 2.881292662116041, "grad_norm": 0.5064358711242676, "learning_rate": 4.737836150153607e-08, "loss": 0.4787, "step": 9005 }, { "epoch": 2.881612627986348, "grad_norm": 0.48052626848220825, "learning_rate": 4.7123010820229186e-08, "loss": 0.4048, "step": 9006 }, { "epoch": 2.8819325938566553, "grad_norm": 0.50408536195755, "learning_rate": 4.6868346861275725e-08, "loss": 0.4808, "step": 9007 }, { "epoch": 2.8822525597269624, "grad_norm": 0.4835711121559143, "learning_rate": 4.6614369659984116e-08, "loss": 0.4072, "step": 9008 }, { "epoch": 2.8825725255972694, "grad_norm": 0.5072152614593506, "learning_rate": 4.636107925156896e-08, "loss": 0.506, "step": 9009 }, { "epoch": 2.882892491467577, "grad_norm": 0.5044692158699036, "learning_rate": 4.610847567115051e-08, "loss": 0.464, "step": 9010 }, { "epoch": 2.883212457337884, "grad_norm": 0.4609028100967407, "learning_rate": 4.5856558953750744e-08, "loss": 0.3974, "step": 9011 }, { "epoch": 2.883532423208191, "grad_norm": 0.5189031362533569, "learning_rate": 4.5605329134299495e-08, "loss": 0.499, "step": 9012 }, { "epoch": 2.8838523890784984, "grad_norm": 0.4667864739894867, "learning_rate": 4.5354786247629456e-08, "loss": 0.3638, "step": 9013 }, { "epoch": 2.8841723549488054, "grad_norm": 0.45510849356651306, "learning_rate": 4.510493032847896e-08, "loss": 0.4153, "step": 9014 }, { "epoch": 2.8844923208191124, "grad_norm": 0.461737722158432, "learning_rate": 4.485576141149084e-08, "loss": 0.4357, "step": 9015 }, { "epoch": 2.88481228668942, "grad_norm": 0.48075640201568604, "learning_rate": 4.460727953121247e-08, "loss": 0.4205, "step": 9016 }, { "epoch": 2.885132252559727, "grad_norm": 0.39439621567726135, "learning_rate": 4.435948472209628e-08, "loss": 0.3687, "step": 9017 }, { "epoch": 2.885452218430034, "grad_norm": 0.5058373212814331, "learning_rate": 4.411237701849924e-08, "loss": 0.5451, "step": 9018 }, { "epoch": 2.8857721843003414, "grad_norm": 0.4888904392719269, "learning_rate": 4.386595645468339e-08, "loss": 0.4556, "step": 9019 }, { "epoch": 2.8860921501706485, "grad_norm": 0.5107361674308777, "learning_rate": 4.362022306481473e-08, "loss": 0.394, "step": 9020 }, { "epoch": 2.8864121160409555, "grad_norm": 0.5496082305908203, "learning_rate": 4.337517688296544e-08, "loss": 0.4273, "step": 9021 }, { "epoch": 2.886732081911263, "grad_norm": 0.5090062022209167, "learning_rate": 4.313081794311058e-08, "loss": 0.4966, "step": 9022 }, { "epoch": 2.88705204778157, "grad_norm": 0.5143817663192749, "learning_rate": 4.288714627913082e-08, "loss": 0.4048, "step": 9023 }, { "epoch": 2.887372013651877, "grad_norm": 0.5198468565940857, "learning_rate": 4.2644161924811353e-08, "loss": 0.4356, "step": 9024 }, { "epoch": 2.8876919795221845, "grad_norm": 0.5902714133262634, "learning_rate": 4.240186491384357e-08, "loss": 0.5307, "step": 9025 }, { "epoch": 2.8880119453924915, "grad_norm": 0.4560842216014862, "learning_rate": 4.216025527982115e-08, "loss": 0.387, "step": 9026 }, { "epoch": 2.8883319112627985, "grad_norm": 0.4865311086177826, "learning_rate": 4.1919333056243964e-08, "loss": 0.4501, "step": 9027 }, { "epoch": 2.888651877133106, "grad_norm": 0.5078125596046448, "learning_rate": 4.1679098276515286e-08, "loss": 0.476, "step": 9028 }, { "epoch": 2.888971843003413, "grad_norm": 0.5110453963279724, "learning_rate": 4.143955097394459e-08, "loss": 0.4446, "step": 9029 }, { "epoch": 2.88929180887372, "grad_norm": 0.503882884979248, "learning_rate": 4.120069118174586e-08, "loss": 0.4279, "step": 9030 }, { "epoch": 2.8896117747440275, "grad_norm": 0.534324049949646, "learning_rate": 4.0962518933037046e-08, "loss": 0.4493, "step": 9031 }, { "epoch": 2.8899317406143346, "grad_norm": 0.49789708852767944, "learning_rate": 4.0725034260840626e-08, "loss": 0.3742, "step": 9032 }, { "epoch": 2.8902517064846416, "grad_norm": 0.5121422410011292, "learning_rate": 4.04882371980847e-08, "loss": 0.4662, "step": 9033 }, { "epoch": 2.890571672354949, "grad_norm": 0.4967876970767975, "learning_rate": 4.0252127777600233e-08, "loss": 0.4169, "step": 9034 }, { "epoch": 2.890891638225256, "grad_norm": 0.5471508502960205, "learning_rate": 4.001670603212604e-08, "loss": 0.4557, "step": 9035 }, { "epoch": 2.891211604095563, "grad_norm": 0.46366459131240845, "learning_rate": 3.978197199430212e-08, "loss": 0.4757, "step": 9036 }, { "epoch": 2.8915315699658706, "grad_norm": 0.5181254148483276, "learning_rate": 3.9547925696675206e-08, "loss": 0.4907, "step": 9037 }, { "epoch": 2.8918515358361776, "grad_norm": 0.5185445547103882, "learning_rate": 3.931456717169546e-08, "loss": 0.4245, "step": 9038 }, { "epoch": 2.8921715017064846, "grad_norm": 0.4751928448677063, "learning_rate": 3.908189645171867e-08, "loss": 0.3828, "step": 9039 }, { "epoch": 2.8924914675767917, "grad_norm": 0.5134016871452332, "learning_rate": 3.8849913569005694e-08, "loss": 0.4076, "step": 9040 }, { "epoch": 2.892811433447099, "grad_norm": 0.4868691563606262, "learning_rate": 3.86186185557208e-08, "loss": 0.4647, "step": 9041 }, { "epoch": 2.893131399317406, "grad_norm": 0.48324650526046753, "learning_rate": 3.838801144393278e-08, "loss": 0.4142, "step": 9042 }, { "epoch": 2.893451365187713, "grad_norm": 0.46404677629470825, "learning_rate": 3.8158092265616065e-08, "loss": 0.4704, "step": 9043 }, { "epoch": 2.89377133105802, "grad_norm": 0.46939048171043396, "learning_rate": 3.792886105264848e-08, "loss": 0.3937, "step": 9044 }, { "epoch": 2.8940912969283277, "grad_norm": 0.4731648564338684, "learning_rate": 3.7700317836814605e-08, "loss": 0.4177, "step": 9045 }, { "epoch": 2.8944112627986347, "grad_norm": 0.4886084496974945, "learning_rate": 3.7472462649801314e-08, "loss": 0.474, "step": 9046 }, { "epoch": 2.8947312286689417, "grad_norm": 0.5067484378814697, "learning_rate": 3.7245295523201106e-08, "loss": 0.434, "step": 9047 }, { "epoch": 2.895051194539249, "grad_norm": 0.4630763530731201, "learning_rate": 3.701881648851102e-08, "loss": 0.4451, "step": 9048 }, { "epoch": 2.895371160409556, "grad_norm": 0.509672999382019, "learning_rate": 3.679302557713205e-08, "loss": 0.4372, "step": 9049 }, { "epoch": 2.8956911262798632, "grad_norm": 0.4773665964603424, "learning_rate": 3.656792282037136e-08, "loss": 0.4372, "step": 9050 }, { "epoch": 2.8960110921501707, "grad_norm": 0.5030218958854675, "learning_rate": 3.634350824943955e-08, "loss": 0.389, "step": 9051 }, { "epoch": 2.8963310580204777, "grad_norm": 0.48134681582450867, "learning_rate": 3.6119781895451734e-08, "loss": 0.4991, "step": 9052 }, { "epoch": 2.8966510238907848, "grad_norm": 0.43066349625587463, "learning_rate": 3.5896743789427535e-08, "loss": 0.3744, "step": 9053 }, { "epoch": 2.8969709897610922, "grad_norm": 0.4765835702419281, "learning_rate": 3.567439396229222e-08, "loss": 0.4312, "step": 9054 }, { "epoch": 2.8972909556313993, "grad_norm": 0.476105272769928, "learning_rate": 3.545273244487391e-08, "loss": 0.4339, "step": 9055 }, { "epoch": 2.8976109215017063, "grad_norm": 0.5043574571609497, "learning_rate": 3.523175926790745e-08, "loss": 0.4558, "step": 9056 }, { "epoch": 2.8979308873720138, "grad_norm": 0.5222077369689941, "learning_rate": 3.501147446202946e-08, "loss": 0.4618, "step": 9057 }, { "epoch": 2.898250853242321, "grad_norm": 0.4806877076625824, "learning_rate": 3.479187805778439e-08, "loss": 0.3512, "step": 9058 }, { "epoch": 2.898570819112628, "grad_norm": 0.4741007685661316, "learning_rate": 3.4572970085617883e-08, "loss": 0.4531, "step": 9059 }, { "epoch": 2.8988907849829353, "grad_norm": 0.5301935076713562, "learning_rate": 3.4354750575883446e-08, "loss": 0.4383, "step": 9060 }, { "epoch": 2.8992107508532423, "grad_norm": 0.5026822090148926, "learning_rate": 3.413721955883631e-08, "loss": 0.4406, "step": 9061 }, { "epoch": 2.8995307167235493, "grad_norm": 0.4861147701740265, "learning_rate": 3.3920377064637355e-08, "loss": 0.4568, "step": 9062 }, { "epoch": 2.899850682593857, "grad_norm": 0.5339757800102234, "learning_rate": 3.370422312335309e-08, "loss": 0.4428, "step": 9063 }, { "epoch": 2.900170648464164, "grad_norm": 0.5031761527061462, "learning_rate": 3.3488757764952306e-08, "loss": 0.3909, "step": 9064 }, { "epoch": 2.900490614334471, "grad_norm": 0.5004369616508484, "learning_rate": 3.3273981019310566e-08, "loss": 0.4481, "step": 9065 }, { "epoch": 2.9008105802047783, "grad_norm": 0.46735748648643494, "learning_rate": 3.305989291620626e-08, "loss": 0.406, "step": 9066 }, { "epoch": 2.9011305460750854, "grad_norm": 0.550030529499054, "learning_rate": 3.2846493485323426e-08, "loss": 0.4686, "step": 9067 }, { "epoch": 2.9014505119453924, "grad_norm": 0.4990123510360718, "learning_rate": 3.263378275624951e-08, "loss": 0.4335, "step": 9068 }, { "epoch": 2.9017704778157, "grad_norm": 0.43858641386032104, "learning_rate": 3.2421760758477585e-08, "loss": 0.4257, "step": 9069 }, { "epoch": 2.902090443686007, "grad_norm": 0.4837779104709625, "learning_rate": 3.221042752140524e-08, "loss": 0.4663, "step": 9070 }, { "epoch": 2.902410409556314, "grad_norm": 0.4876711368560791, "learning_rate": 3.1999783074332934e-08, "loss": 0.4159, "step": 9071 }, { "epoch": 2.9027303754266214, "grad_norm": 0.5029987692832947, "learning_rate": 3.1789827446467304e-08, "loss": 0.4138, "step": 9072 }, { "epoch": 2.9030503412969284, "grad_norm": 0.4941578507423401, "learning_rate": 3.15805606669195e-08, "loss": 0.449, "step": 9073 }, { "epoch": 2.9033703071672354, "grad_norm": 0.4757187068462372, "learning_rate": 3.137198276470355e-08, "loss": 0.4441, "step": 9074 }, { "epoch": 2.903690273037543, "grad_norm": 0.509201169013977, "learning_rate": 3.116409376874019e-08, "loss": 0.4959, "step": 9075 }, { "epoch": 2.90401023890785, "grad_norm": 0.4921058714389801, "learning_rate": 3.095689370785249e-08, "loss": 0.4105, "step": 9076 }, { "epoch": 2.904330204778157, "grad_norm": 0.48168763518333435, "learning_rate": 3.0750382610769124e-08, "loss": 0.4493, "step": 9077 }, { "epoch": 2.9046501706484644, "grad_norm": 0.49827325344085693, "learning_rate": 3.0544560506123863e-08, "loss": 0.3925, "step": 9078 }, { "epoch": 2.9049701365187715, "grad_norm": 0.48839378356933594, "learning_rate": 3.0339427422453327e-08, "loss": 0.3944, "step": 9079 }, { "epoch": 2.9052901023890785, "grad_norm": 0.4565756916999817, "learning_rate": 3.013498338820031e-08, "loss": 0.4225, "step": 9080 }, { "epoch": 2.9056100682593855, "grad_norm": 0.5143803954124451, "learning_rate": 2.993122843171048e-08, "loss": 0.4555, "step": 9081 }, { "epoch": 2.905930034129693, "grad_norm": 0.5379270911216736, "learning_rate": 2.9728162581234566e-08, "loss": 0.4531, "step": 9082 }, { "epoch": 2.90625, "grad_norm": 0.5080519914627075, "learning_rate": 2.9525785864928934e-08, "loss": 0.4265, "step": 9083 }, { "epoch": 2.906569965870307, "grad_norm": 0.4725300669670105, "learning_rate": 2.9324098310852257e-08, "loss": 0.4317, "step": 9084 }, { "epoch": 2.9068899317406145, "grad_norm": 0.4862784743309021, "learning_rate": 2.912309994696938e-08, "loss": 0.4418, "step": 9085 }, { "epoch": 2.9072098976109215, "grad_norm": 0.4593566060066223, "learning_rate": 2.8922790801149126e-08, "loss": 0.4044, "step": 9086 }, { "epoch": 2.9075298634812285, "grad_norm": 0.4919830858707428, "learning_rate": 2.8723170901163722e-08, "loss": 0.4559, "step": 9087 }, { "epoch": 2.9078498293515356, "grad_norm": 0.5015954375267029, "learning_rate": 2.8524240274691582e-08, "loss": 0.464, "step": 9088 }, { "epoch": 2.908169795221843, "grad_norm": 0.5168560147285461, "learning_rate": 2.8325998949314536e-08, "loss": 0.4154, "step": 9089 }, { "epoch": 2.90848976109215, "grad_norm": 0.5248748660087585, "learning_rate": 2.8128446952518373e-08, "loss": 0.4314, "step": 9090 }, { "epoch": 2.908809726962457, "grad_norm": 0.4684077799320221, "learning_rate": 2.7931584311694514e-08, "loss": 0.4436, "step": 9091 }, { "epoch": 2.9091296928327646, "grad_norm": 0.5102325677871704, "learning_rate": 2.773541105413835e-08, "loss": 0.4127, "step": 9092 }, { "epoch": 2.9094496587030716, "grad_norm": 0.5072308778762817, "learning_rate": 2.7539927207048678e-08, "loss": 0.4392, "step": 9093 }, { "epoch": 2.9097696245733786, "grad_norm": 0.4549092650413513, "learning_rate": 2.7345132797530482e-08, "loss": 0.4767, "step": 9094 }, { "epoch": 2.910089590443686, "grad_norm": 0.4574567377567291, "learning_rate": 2.7151027852591606e-08, "loss": 0.411, "step": 9095 }, { "epoch": 2.910409556313993, "grad_norm": 0.527850866317749, "learning_rate": 2.6957612399145517e-08, "loss": 0.4501, "step": 9096 }, { "epoch": 2.9107295221843, "grad_norm": 0.5241904258728027, "learning_rate": 2.6764886464009097e-08, "loss": 0.4183, "step": 9097 }, { "epoch": 2.9110494880546076, "grad_norm": 0.5417770147323608, "learning_rate": 2.6572850073903755e-08, "loss": 0.4734, "step": 9098 }, { "epoch": 2.9113694539249146, "grad_norm": 0.48998844623565674, "learning_rate": 2.638150325545652e-08, "loss": 0.4398, "step": 9099 }, { "epoch": 2.9116894197952217, "grad_norm": 0.5481809973716736, "learning_rate": 2.6190846035196727e-08, "loss": 0.4864, "step": 9100 }, { "epoch": 2.912009385665529, "grad_norm": 0.5044158101081848, "learning_rate": 2.6000878439559896e-08, "loss": 0.4075, "step": 9101 }, { "epoch": 2.912329351535836, "grad_norm": 0.5413654446601868, "learning_rate": 2.5811600494885512e-08, "loss": 0.4452, "step": 9102 }, { "epoch": 2.912649317406143, "grad_norm": 0.4841751456260681, "learning_rate": 2.5623012227416476e-08, "loss": 0.4071, "step": 9103 }, { "epoch": 2.9129692832764507, "grad_norm": 0.5086470246315002, "learning_rate": 2.5435113663301315e-08, "loss": 0.4196, "step": 9104 }, { "epoch": 2.9132892491467577, "grad_norm": 0.4594883620738983, "learning_rate": 2.5247904828591963e-08, "loss": 0.4427, "step": 9105 }, { "epoch": 2.9136092150170647, "grad_norm": 0.4493332803249359, "learning_rate": 2.5061385749245437e-08, "loss": 0.4014, "step": 9106 }, { "epoch": 2.913929180887372, "grad_norm": 0.46458470821380615, "learning_rate": 2.4875556451123274e-08, "loss": 0.4496, "step": 9107 }, { "epoch": 2.914249146757679, "grad_norm": 0.5013710856437683, "learning_rate": 2.469041695998986e-08, "loss": 0.411, "step": 9108 }, { "epoch": 2.9145691126279862, "grad_norm": 0.48893997073173523, "learning_rate": 2.4505967301515778e-08, "loss": 0.4391, "step": 9109 }, { "epoch": 2.9148890784982937, "grad_norm": 0.49383464455604553, "learning_rate": 2.4322207501274453e-08, "loss": 0.4164, "step": 9110 }, { "epoch": 2.9152090443686007, "grad_norm": 0.44160425662994385, "learning_rate": 2.413913758474551e-08, "loss": 0.4516, "step": 9111 }, { "epoch": 2.9155290102389078, "grad_norm": 0.47670525312423706, "learning_rate": 2.3956757577311418e-08, "loss": 0.4217, "step": 9112 }, { "epoch": 2.9158489761092152, "grad_norm": 0.49880895018577576, "learning_rate": 2.377506750425862e-08, "loss": 0.4099, "step": 9113 }, { "epoch": 2.9161689419795223, "grad_norm": 0.49023520946502686, "learning_rate": 2.359406739077974e-08, "loss": 0.4096, "step": 9114 }, { "epoch": 2.9164889078498293, "grad_norm": 0.5317866206169128, "learning_rate": 2.341375726197026e-08, "loss": 0.4396, "step": 9115 }, { "epoch": 2.9168088737201368, "grad_norm": 0.4866684675216675, "learning_rate": 2.3234137142830182e-08, "loss": 0.4448, "step": 9116 }, { "epoch": 2.917128839590444, "grad_norm": 0.49172016978263855, "learning_rate": 2.3055207058264028e-08, "loss": 0.435, "step": 9117 }, { "epoch": 2.917448805460751, "grad_norm": 0.4982343018054962, "learning_rate": 2.2876967033080845e-08, "loss": 0.5196, "step": 9118 }, { "epoch": 2.9177687713310583, "grad_norm": 0.4825814366340637, "learning_rate": 2.2699417091993635e-08, "loss": 0.4358, "step": 9119 }, { "epoch": 2.9180887372013653, "grad_norm": 0.5194673538208008, "learning_rate": 2.2522557259620492e-08, "loss": 0.4736, "step": 9120 }, { "epoch": 2.9184087030716723, "grad_norm": 0.4512815773487091, "learning_rate": 2.234638756048235e-08, "loss": 0.4362, "step": 9121 }, { "epoch": 2.91872866894198, "grad_norm": 0.46083351969718933, "learning_rate": 2.217090801900634e-08, "loss": 0.4274, "step": 9122 }, { "epoch": 2.919048634812287, "grad_norm": 0.42566928267478943, "learning_rate": 2.1996118659521894e-08, "loss": 0.4126, "step": 9123 }, { "epoch": 2.919368600682594, "grad_norm": 0.4711596369743347, "learning_rate": 2.1822019506264612e-08, "loss": 0.463, "step": 9124 }, { "epoch": 2.919688566552901, "grad_norm": 0.49391430616378784, "learning_rate": 2.164861058337353e-08, "loss": 0.4164, "step": 9125 }, { "epoch": 2.9200085324232083, "grad_norm": 0.547099232673645, "learning_rate": 2.1475891914891078e-08, "loss": 0.4504, "step": 9126 }, { "epoch": 2.9203284982935154, "grad_norm": 0.4892997741699219, "learning_rate": 2.1303863524765878e-08, "loss": 0.4054, "step": 9127 }, { "epoch": 2.9206484641638224, "grad_norm": 0.5020419955253601, "learning_rate": 2.1132525436849406e-08, "loss": 0.4643, "step": 9128 }, { "epoch": 2.9209684300341294, "grad_norm": 0.5101315975189209, "learning_rate": 2.0961877674898213e-08, "loss": 0.4345, "step": 9129 }, { "epoch": 2.921288395904437, "grad_norm": 0.44621843099594116, "learning_rate": 2.0791920262572262e-08, "loss": 0.4093, "step": 9130 }, { "epoch": 2.921608361774744, "grad_norm": 0.457491934299469, "learning_rate": 2.062265322343715e-08, "loss": 0.4225, "step": 9131 }, { "epoch": 2.921928327645051, "grad_norm": 0.5055102705955505, "learning_rate": 2.0454076580961323e-08, "loss": 0.4948, "step": 9132 }, { "epoch": 2.9222482935153584, "grad_norm": 0.5012072920799255, "learning_rate": 2.0286190358517754e-08, "loss": 0.4174, "step": 9133 }, { "epoch": 2.9225682593856654, "grad_norm": 0.4871751368045807, "learning_rate": 2.0118994579385042e-08, "loss": 0.3989, "step": 9134 }, { "epoch": 2.9228882252559725, "grad_norm": 0.49165430665016174, "learning_rate": 1.9952489266744647e-08, "loss": 0.4385, "step": 9135 }, { "epoch": 2.92320819112628, "grad_norm": 0.4882758855819702, "learning_rate": 1.9786674443683096e-08, "loss": 0.444, "step": 9136 }, { "epoch": 2.923528156996587, "grad_norm": 0.45996996760368347, "learning_rate": 1.9621550133190337e-08, "loss": 0.4277, "step": 9137 }, { "epoch": 2.923848122866894, "grad_norm": 0.46109628677368164, "learning_rate": 1.945711635816083e-08, "loss": 0.3773, "step": 9138 }, { "epoch": 2.9241680887372015, "grad_norm": 0.5478599667549133, "learning_rate": 1.9293373141394124e-08, "loss": 0.4842, "step": 9139 }, { "epoch": 2.9244880546075085, "grad_norm": 0.49445387721061707, "learning_rate": 1.9130320505593158e-08, "loss": 0.4225, "step": 9140 }, { "epoch": 2.9248080204778155, "grad_norm": 0.5079352259635925, "learning_rate": 1.896795847336541e-08, "loss": 0.4762, "step": 9141 }, { "epoch": 2.925127986348123, "grad_norm": 0.4407789409160614, "learning_rate": 1.8806287067222873e-08, "loss": 0.3709, "step": 9142 }, { "epoch": 2.92544795221843, "grad_norm": 0.47105005383491516, "learning_rate": 1.8645306309580948e-08, "loss": 0.4343, "step": 9143 }, { "epoch": 2.925767918088737, "grad_norm": 0.49983924627304077, "learning_rate": 1.8485016222759555e-08, "loss": 0.4664, "step": 9144 }, { "epoch": 2.9260878839590445, "grad_norm": 0.48586347699165344, "learning_rate": 1.83254168289837e-08, "loss": 0.4526, "step": 9145 }, { "epoch": 2.9264078498293515, "grad_norm": 0.5456278324127197, "learning_rate": 1.8166508150382344e-08, "loss": 0.4782, "step": 9146 }, { "epoch": 2.9267278156996586, "grad_norm": 0.4486168622970581, "learning_rate": 1.8008290208987868e-08, "loss": 0.417, "step": 9147 }, { "epoch": 2.927047781569966, "grad_norm": 0.5022203326225281, "learning_rate": 1.785076302673716e-08, "loss": 0.4686, "step": 9148 }, { "epoch": 2.927367747440273, "grad_norm": 0.4830387532711029, "learning_rate": 1.7693926625471646e-08, "loss": 0.4596, "step": 9149 }, { "epoch": 2.92768771331058, "grad_norm": 0.4588986039161682, "learning_rate": 1.7537781026937816e-08, "loss": 0.4222, "step": 9150 }, { "epoch": 2.9280076791808876, "grad_norm": 0.49632179737091064, "learning_rate": 1.738232625278391e-08, "loss": 0.4845, "step": 9151 }, { "epoch": 2.9283276450511946, "grad_norm": 0.4504736065864563, "learning_rate": 1.7227562324565463e-08, "loss": 0.38, "step": 9152 }, { "epoch": 2.9286476109215016, "grad_norm": 0.49249497056007385, "learning_rate": 1.7073489263739197e-08, "loss": 0.4617, "step": 9153 }, { "epoch": 2.928967576791809, "grad_norm": 0.46786049008369446, "learning_rate": 1.6920107091668582e-08, "loss": 0.4591, "step": 9154 }, { "epoch": 2.929287542662116, "grad_norm": 0.46979671716690063, "learning_rate": 1.676741582961994e-08, "loss": 0.4544, "step": 9155 }, { "epoch": 2.929607508532423, "grad_norm": 0.5029367208480835, "learning_rate": 1.6615415498764108e-08, "loss": 0.4066, "step": 9156 }, { "epoch": 2.9299274744027306, "grad_norm": 0.5062602162361145, "learning_rate": 1.6464106120175905e-08, "loss": 0.4635, "step": 9157 }, { "epoch": 2.9302474402730376, "grad_norm": 0.491169810295105, "learning_rate": 1.6313487714834654e-08, "loss": 0.4324, "step": 9158 }, { "epoch": 2.9305674061433447, "grad_norm": 0.5217692255973816, "learning_rate": 1.616356030362365e-08, "loss": 0.4255, "step": 9159 }, { "epoch": 2.930887372013652, "grad_norm": 0.4856349527835846, "learning_rate": 1.6014323907331265e-08, "loss": 0.4283, "step": 9160 }, { "epoch": 2.931207337883959, "grad_norm": 0.5454949140548706, "learning_rate": 1.5865778546648723e-08, "loss": 0.4642, "step": 9161 }, { "epoch": 2.931527303754266, "grad_norm": 0.49683502316474915, "learning_rate": 1.571792424217178e-08, "loss": 0.4336, "step": 9162 }, { "epoch": 2.9318472696245736, "grad_norm": 0.5096780061721802, "learning_rate": 1.557076101440125e-08, "loss": 0.4651, "step": 9163 }, { "epoch": 2.9321672354948807, "grad_norm": 0.47015881538391113, "learning_rate": 1.5424288883740257e-08, "loss": 0.4176, "step": 9164 }, { "epoch": 2.9324872013651877, "grad_norm": 0.5371462106704712, "learning_rate": 1.5278507870499226e-08, "loss": 0.4615, "step": 9165 }, { "epoch": 2.9328071672354947, "grad_norm": 0.5015416741371155, "learning_rate": 1.513341799488921e-08, "loss": 0.4304, "step": 9166 }, { "epoch": 2.933127133105802, "grad_norm": 0.49367377161979675, "learning_rate": 1.4989019277028004e-08, "loss": 0.3683, "step": 9167 }, { "epoch": 2.9334470989761092, "grad_norm": 0.4831542372703552, "learning_rate": 1.4845311736936263e-08, "loss": 0.4894, "step": 9168 }, { "epoch": 2.9337670648464163, "grad_norm": 0.4677984416484833, "learning_rate": 1.4702295394539157e-08, "loss": 0.3889, "step": 9169 }, { "epoch": 2.9340870307167233, "grad_norm": 0.5112711787223816, "learning_rate": 1.4559970269666379e-08, "loss": 0.4102, "step": 9170 }, { "epoch": 2.9344069965870307, "grad_norm": 0.5255852937698364, "learning_rate": 1.441833638205159e-08, "loss": 0.4651, "step": 9171 }, { "epoch": 2.9347269624573378, "grad_norm": 0.48539504408836365, "learning_rate": 1.4277393751332414e-08, "loss": 0.4518, "step": 9172 }, { "epoch": 2.935046928327645, "grad_norm": 0.4464510679244995, "learning_rate": 1.4137142397049885e-08, "loss": 0.3935, "step": 9173 }, { "epoch": 2.9353668941979523, "grad_norm": 0.4925348162651062, "learning_rate": 1.3997582338651228e-08, "loss": 0.456, "step": 9174 }, { "epoch": 2.9356868600682593, "grad_norm": 0.48311126232147217, "learning_rate": 1.3858713595485961e-08, "loss": 0.4418, "step": 9175 }, { "epoch": 2.9360068259385663, "grad_norm": 0.46080172061920166, "learning_rate": 1.3720536186808686e-08, "loss": 0.4259, "step": 9176 }, { "epoch": 2.936326791808874, "grad_norm": 0.43515244126319885, "learning_rate": 1.3583050131777408e-08, "loss": 0.3937, "step": 9177 }, { "epoch": 2.936646757679181, "grad_norm": 0.5481618046760559, "learning_rate": 1.3446255449455215e-08, "loss": 0.5069, "step": 9178 }, { "epoch": 2.936966723549488, "grad_norm": 0.4946882426738739, "learning_rate": 1.3310152158808598e-08, "loss": 0.4196, "step": 9179 }, { "epoch": 2.9372866894197953, "grad_norm": 0.5199326872825623, "learning_rate": 1.3174740278708575e-08, "loss": 0.4455, "step": 9180 }, { "epoch": 2.9376066552901023, "grad_norm": 0.5129525065422058, "learning_rate": 1.3040019827930128e-08, "loss": 0.4764, "step": 9181 }, { "epoch": 2.9379266211604094, "grad_norm": 0.506719708442688, "learning_rate": 1.2905990825152204e-08, "loss": 0.4782, "step": 9182 }, { "epoch": 2.938246587030717, "grad_norm": 0.5135883688926697, "learning_rate": 1.277265328895827e-08, "loss": 0.4102, "step": 9183 }, { "epoch": 2.938566552901024, "grad_norm": 0.4967620074748993, "learning_rate": 1.264000723783576e-08, "loss": 0.4614, "step": 9184 }, { "epoch": 2.938886518771331, "grad_norm": 0.4385746419429779, "learning_rate": 1.250805269017663e-08, "loss": 0.4591, "step": 9185 }, { "epoch": 2.9392064846416384, "grad_norm": 0.5029006004333496, "learning_rate": 1.237678966427569e-08, "loss": 0.3995, "step": 9186 }, { "epoch": 2.9395264505119454, "grad_norm": 0.5193878412246704, "learning_rate": 1.224621817833338e-08, "loss": 0.4779, "step": 9187 }, { "epoch": 2.9398464163822524, "grad_norm": 0.4947200417518616, "learning_rate": 1.2116338250452997e-08, "loss": 0.4176, "step": 9188 }, { "epoch": 2.94016638225256, "grad_norm": 0.4940420091152191, "learning_rate": 1.1987149898642914e-08, "loss": 0.4224, "step": 9189 }, { "epoch": 2.940486348122867, "grad_norm": 0.5115923881530762, "learning_rate": 1.1858653140815468e-08, "loss": 0.4605, "step": 9190 }, { "epoch": 2.940806313993174, "grad_norm": 0.5231035351753235, "learning_rate": 1.173084799478641e-08, "loss": 0.4506, "step": 9191 }, { "epoch": 2.9411262798634814, "grad_norm": 0.4822664260864258, "learning_rate": 1.1603734478276562e-08, "loss": 0.3945, "step": 9192 }, { "epoch": 2.9414462457337884, "grad_norm": 0.4457946717739105, "learning_rate": 1.1477312608910162e-08, "loss": 0.4186, "step": 9193 }, { "epoch": 2.9417662116040955, "grad_norm": 0.5013504028320312, "learning_rate": 1.135158240421541e-08, "loss": 0.495, "step": 9194 }, { "epoch": 2.942086177474403, "grad_norm": 0.5230039954185486, "learning_rate": 1.1226543881625584e-08, "loss": 0.4535, "step": 9195 }, { "epoch": 2.94240614334471, "grad_norm": 0.5289256572723389, "learning_rate": 1.1102197058476816e-08, "loss": 0.438, "step": 9196 }, { "epoch": 2.942726109215017, "grad_norm": 0.5156910419464111, "learning_rate": 1.0978541952010313e-08, "loss": 0.4271, "step": 9197 }, { "epoch": 2.9430460750853245, "grad_norm": 0.4751654863357544, "learning_rate": 1.0855578579370696e-08, "loss": 0.414, "step": 9198 }, { "epoch": 2.9433660409556315, "grad_norm": 0.5000825524330139, "learning_rate": 1.0733306957607104e-08, "loss": 0.4947, "step": 9199 }, { "epoch": 2.9436860068259385, "grad_norm": 0.5216525793075562, "learning_rate": 1.0611727103673196e-08, "loss": 0.3871, "step": 9200 }, { "epoch": 2.944005972696246, "grad_norm": 0.4903882145881653, "learning_rate": 1.0490839034425493e-08, "loss": 0.4292, "step": 9201 }, { "epoch": 2.944325938566553, "grad_norm": 0.49286291003227234, "learning_rate": 1.0370642766625583e-08, "loss": 0.424, "step": 9202 }, { "epoch": 2.94464590443686, "grad_norm": 0.46055853366851807, "learning_rate": 1.025113831693847e-08, "loss": 0.4237, "step": 9203 }, { "epoch": 2.9449658703071675, "grad_norm": 0.5028175115585327, "learning_rate": 1.0132325701934232e-08, "loss": 0.4496, "step": 9204 }, { "epoch": 2.9452858361774745, "grad_norm": 0.4903533160686493, "learning_rate": 1.0014204938085803e-08, "loss": 0.4153, "step": 9205 }, { "epoch": 2.9456058020477816, "grad_norm": 0.4870935082435608, "learning_rate": 9.89677604177064e-09, "loss": 0.42, "step": 9206 }, { "epoch": 2.9459257679180886, "grad_norm": 0.5625002980232239, "learning_rate": 9.780039029271271e-09, "loss": 0.4543, "step": 9207 }, { "epoch": 2.946245733788396, "grad_norm": 0.46420183777809143, "learning_rate": 9.663993916772529e-09, "loss": 0.4816, "step": 9208 }, { "epoch": 2.946565699658703, "grad_norm": 0.46914464235305786, "learning_rate": 9.548640720364321e-09, "loss": 0.4378, "step": 9209 }, { "epoch": 2.94688566552901, "grad_norm": 0.4514414370059967, "learning_rate": 9.43397945604163e-09, "loss": 0.3825, "step": 9210 }, { "epoch": 2.9472056313993176, "grad_norm": 0.4555622637271881, "learning_rate": 9.320010139700631e-09, "loss": 0.4416, "step": 9211 }, { "epoch": 2.9475255972696246, "grad_norm": 0.5080385804176331, "learning_rate": 9.20673278714479e-09, "loss": 0.4571, "step": 9212 }, { "epoch": 2.9478455631399316, "grad_norm": 0.5202873945236206, "learning_rate": 9.094147414078769e-09, "loss": 0.4595, "step": 9213 }, { "epoch": 2.9481655290102387, "grad_norm": 0.4716443121433258, "learning_rate": 8.982254036113968e-09, "loss": 0.399, "step": 9214 }, { "epoch": 2.948485494880546, "grad_norm": 0.5085718035697937, "learning_rate": 8.871052668763535e-09, "loss": 0.44, "step": 9215 }, { "epoch": 2.948805460750853, "grad_norm": 0.47173622250556946, "learning_rate": 8.760543327446802e-09, "loss": 0.4069, "step": 9216 }, { "epoch": 2.94912542662116, "grad_norm": 0.4833856225013733, "learning_rate": 8.650726027484846e-09, "loss": 0.392, "step": 9217 }, { "epoch": 2.9494453924914676, "grad_norm": 0.45571327209472656, "learning_rate": 8.541600784104376e-09, "loss": 0.4697, "step": 9218 }, { "epoch": 2.9497653583617747, "grad_norm": 0.4360343813896179, "learning_rate": 8.433167612436066e-09, "loss": 0.3868, "step": 9219 }, { "epoch": 2.9500853242320817, "grad_norm": 0.4683946371078491, "learning_rate": 8.325426527514003e-09, "loss": 0.4701, "step": 9220 }, { "epoch": 2.950405290102389, "grad_norm": 0.5024915933609009, "learning_rate": 8.21837754427679e-09, "loss": 0.4405, "step": 9221 }, { "epoch": 2.950725255972696, "grad_norm": 0.4808758795261383, "learning_rate": 8.112020677567001e-09, "loss": 0.4219, "step": 9222 }, { "epoch": 2.951045221843003, "grad_norm": 0.47212451696395874, "learning_rate": 8.006355942130616e-09, "loss": 0.4211, "step": 9223 }, { "epoch": 2.9513651877133107, "grad_norm": 0.5444512367248535, "learning_rate": 7.901383352618142e-09, "loss": 0.5328, "step": 9224 }, { "epoch": 2.9516851535836177, "grad_norm": 0.4807533323764801, "learning_rate": 7.797102923585154e-09, "loss": 0.3795, "step": 9225 }, { "epoch": 2.9520051194539247, "grad_norm": 0.5102534890174866, "learning_rate": 7.693514669489533e-09, "loss": 0.4764, "step": 9226 }, { "epoch": 2.952325085324232, "grad_norm": 0.4702845811843872, "learning_rate": 7.590618604693678e-09, "loss": 0.4287, "step": 9227 }, { "epoch": 2.9526450511945392, "grad_norm": 0.5033854842185974, "learning_rate": 7.488414743464512e-09, "loss": 0.457, "step": 9228 }, { "epoch": 2.9529650170648463, "grad_norm": 0.5184718370437622, "learning_rate": 7.38690309997292e-09, "loss": 0.4333, "step": 9229 }, { "epoch": 2.9532849829351537, "grad_norm": 0.5364457368850708, "learning_rate": 7.286083688293755e-09, "loss": 0.4411, "step": 9230 }, { "epoch": 2.9536049488054608, "grad_norm": 0.47790831327438354, "learning_rate": 7.1859565224052796e-09, "loss": 0.4661, "step": 9231 }, { "epoch": 2.953924914675768, "grad_norm": 0.44603395462036133, "learning_rate": 7.0865216161902785e-09, "loss": 0.4012, "step": 9232 }, { "epoch": 2.9542448805460753, "grad_norm": 0.5121586918830872, "learning_rate": 6.9877789834355e-09, "loss": 0.4345, "step": 9233 }, { "epoch": 2.9545648464163823, "grad_norm": 0.4695791006088257, "learning_rate": 6.889728637832216e-09, "loss": 0.3884, "step": 9234 }, { "epoch": 2.9548848122866893, "grad_norm": 0.5383626818656921, "learning_rate": 6.792370592975106e-09, "loss": 0.4676, "step": 9235 }, { "epoch": 2.955204778156997, "grad_norm": 0.48278743028640747, "learning_rate": 6.6957048623628175e-09, "loss": 0.4897, "step": 9236 }, { "epoch": 2.955524744027304, "grad_norm": 0.47618865966796875, "learning_rate": 6.599731459397962e-09, "loss": 0.3991, "step": 9237 }, { "epoch": 2.955844709897611, "grad_norm": 0.5008996725082397, "learning_rate": 6.504450397387119e-09, "loss": 0.4679, "step": 9238 }, { "epoch": 2.9561646757679183, "grad_norm": 0.4668422043323517, "learning_rate": 6.40986168954194e-09, "loss": 0.3855, "step": 9239 }, { "epoch": 2.9564846416382253, "grad_norm": 0.510735034942627, "learning_rate": 6.315965348976938e-09, "loss": 0.4149, "step": 9240 }, { "epoch": 2.9568046075085324, "grad_norm": 0.5300773978233337, "learning_rate": 6.222761388711141e-09, "loss": 0.4485, "step": 9241 }, { "epoch": 2.95712457337884, "grad_norm": 0.4763416051864624, "learning_rate": 6.130249821666434e-09, "loss": 0.4383, "step": 9242 }, { "epoch": 2.957444539249147, "grad_norm": 0.5363229513168335, "learning_rate": 6.038430660670891e-09, "loss": 0.4253, "step": 9243 }, { "epoch": 2.957764505119454, "grad_norm": 0.5187003016471863, "learning_rate": 5.947303918454328e-09, "loss": 0.4287, "step": 9244 }, { "epoch": 2.9580844709897613, "grad_norm": 0.49342259764671326, "learning_rate": 5.856869607652749e-09, "loss": 0.3999, "step": 9245 }, { "epoch": 2.9584044368600684, "grad_norm": 0.5214436650276184, "learning_rate": 5.767127740803902e-09, "loss": 0.4429, "step": 9246 }, { "epoch": 2.9587244027303754, "grad_norm": 0.4750804007053375, "learning_rate": 5.678078330351166e-09, "loss": 0.3914, "step": 9247 }, { "epoch": 2.9590443686006824, "grad_norm": 0.4475878179073334, "learning_rate": 5.589721388640778e-09, "loss": 0.393, "step": 9248 }, { "epoch": 2.95936433447099, "grad_norm": 0.4907412827014923, "learning_rate": 5.5020569279246035e-09, "loss": 0.4505, "step": 9249 }, { "epoch": 2.959684300341297, "grad_norm": 0.4824308454990387, "learning_rate": 5.415084960356254e-09, "loss": 0.4407, "step": 9250 }, { "epoch": 2.960004266211604, "grad_norm": 0.4802906811237335, "learning_rate": 5.3288054979949714e-09, "loss": 0.4897, "step": 9251 }, { "epoch": 2.9603242320819114, "grad_norm": 0.48446446657180786, "learning_rate": 5.243218552803964e-09, "loss": 0.4279, "step": 9252 }, { "epoch": 2.9606441979522184, "grad_norm": 0.4675389528274536, "learning_rate": 5.1583241366492955e-09, "loss": 0.3847, "step": 9253 }, { "epoch": 2.9609641638225255, "grad_norm": 0.5105575919151306, "learning_rate": 5.074122261301551e-09, "loss": 0.4312, "step": 9254 }, { "epoch": 2.9612841296928325, "grad_norm": 0.4578179121017456, "learning_rate": 4.99061293843639e-09, "loss": 0.3716, "step": 9255 }, { "epoch": 2.96160409556314, "grad_norm": 0.5118416547775269, "learning_rate": 4.907796179631774e-09, "loss": 0.484, "step": 9256 }, { "epoch": 2.961924061433447, "grad_norm": 0.47899672389030457, "learning_rate": 4.825671996370184e-09, "loss": 0.4052, "step": 9257 }, { "epoch": 2.962244027303754, "grad_norm": 0.5589344501495361, "learning_rate": 4.744240400038624e-09, "loss": 0.4194, "step": 9258 }, { "epoch": 2.9625639931740615, "grad_norm": 0.49536803364753723, "learning_rate": 4.663501401927506e-09, "loss": 0.4636, "step": 9259 }, { "epoch": 2.9628839590443685, "grad_norm": 0.49311965703964233, "learning_rate": 4.583455013231764e-09, "loss": 0.4318, "step": 9260 }, { "epoch": 2.9632039249146755, "grad_norm": 0.5307648777961731, "learning_rate": 4.504101245049741e-09, "loss": 0.3942, "step": 9261 }, { "epoch": 2.963523890784983, "grad_norm": 0.5528405904769897, "learning_rate": 4.425440108383195e-09, "loss": 0.4615, "step": 9262 }, { "epoch": 2.96384385665529, "grad_norm": 0.5247182846069336, "learning_rate": 4.347471614139509e-09, "loss": 0.4698, "step": 9263 }, { "epoch": 2.964163822525597, "grad_norm": 0.5076631307601929, "learning_rate": 4.270195773128927e-09, "loss": 0.4439, "step": 9264 }, { "epoch": 2.9644837883959045, "grad_norm": 0.4680469036102295, "learning_rate": 4.193612596065655e-09, "loss": 0.4456, "step": 9265 }, { "epoch": 2.9648037542662116, "grad_norm": 0.4061358571052551, "learning_rate": 4.117722093567867e-09, "loss": 0.3656, "step": 9266 }, { "epoch": 2.9651237201365186, "grad_norm": 0.4813680350780487, "learning_rate": 4.042524276158255e-09, "loss": 0.4453, "step": 9267 }, { "epoch": 2.965443686006826, "grad_norm": 0.5288780927658081, "learning_rate": 3.968019154262926e-09, "loss": 0.4411, "step": 9268 }, { "epoch": 2.965763651877133, "grad_norm": 0.5130483508110046, "learning_rate": 3.894206738212502e-09, "loss": 0.438, "step": 9269 }, { "epoch": 2.96608361774744, "grad_norm": 0.5286841988563538, "learning_rate": 3.8210870382404674e-09, "loss": 0.4828, "step": 9270 }, { "epoch": 2.9664035836177476, "grad_norm": 0.4753878116607666, "learning_rate": 3.748660064484821e-09, "loss": 0.4217, "step": 9271 }, { "epoch": 2.9667235494880546, "grad_norm": 0.4546852707862854, "learning_rate": 3.6769258269886422e-09, "loss": 0.4621, "step": 9272 }, { "epoch": 2.9670435153583616, "grad_norm": 0.5022838711738586, "learning_rate": 3.6058843356973074e-09, "loss": 0.4346, "step": 9273 }, { "epoch": 2.967363481228669, "grad_norm": 0.48522311449050903, "learning_rate": 3.5355356004607154e-09, "loss": 0.3815, "step": 9274 }, { "epoch": 2.967683447098976, "grad_norm": 0.5357267260551453, "learning_rate": 3.4658796310332866e-09, "loss": 0.4844, "step": 9275 }, { "epoch": 2.968003412969283, "grad_norm": 0.4758678674697876, "learning_rate": 3.3969164370722953e-09, "loss": 0.3583, "step": 9276 }, { "epoch": 2.9683233788395906, "grad_norm": 0.46168282628059387, "learning_rate": 3.3286460281400922e-09, "loss": 0.4112, "step": 9277 }, { "epoch": 2.9686433447098977, "grad_norm": 0.5056720972061157, "learning_rate": 3.2610684137024397e-09, "loss": 0.5097, "step": 9278 }, { "epoch": 2.9689633105802047, "grad_norm": 0.5224137306213379, "learning_rate": 3.194183603128509e-09, "loss": 0.4443, "step": 9279 }, { "epoch": 2.969283276450512, "grad_norm": 0.5291537046432495, "learning_rate": 3.1279916056925484e-09, "loss": 0.4944, "step": 9280 }, { "epoch": 2.969603242320819, "grad_norm": 0.4236122965812683, "learning_rate": 3.062492430572217e-09, "loss": 0.3447, "step": 9281 }, { "epoch": 2.969923208191126, "grad_norm": 0.4741148054599762, "learning_rate": 2.9976860868485837e-09, "loss": 0.4604, "step": 9282 }, { "epoch": 2.9702431740614337, "grad_norm": 0.4553922414779663, "learning_rate": 2.933572583506683e-09, "loss": 0.4211, "step": 9283 }, { "epoch": 2.9705631399317407, "grad_norm": 0.5211372971534729, "learning_rate": 2.8701519294371815e-09, "loss": 0.4789, "step": 9284 }, { "epoch": 2.9708831058020477, "grad_norm": 0.49699050188064575, "learning_rate": 2.8074241334330454e-09, "loss": 0.4508, "step": 9285 }, { "epoch": 2.971203071672355, "grad_norm": 0.4980183243751526, "learning_rate": 2.7453892041906516e-09, "loss": 0.41, "step": 9286 }, { "epoch": 2.9715230375426622, "grad_norm": 0.4982837736606598, "learning_rate": 2.684047150312563e-09, "loss": 0.4297, "step": 9287 }, { "epoch": 2.9718430034129693, "grad_norm": 0.5232399106025696, "learning_rate": 2.6233979803025334e-09, "loss": 0.4717, "step": 9288 }, { "epoch": 2.9721629692832767, "grad_norm": 0.48403775691986084, "learning_rate": 2.5634417025710565e-09, "loss": 0.4464, "step": 9289 }, { "epoch": 2.9724829351535837, "grad_norm": 0.5008255243301392, "learning_rate": 2.504178325429818e-09, "loss": 0.3797, "step": 9290 }, { "epoch": 2.9728029010238908, "grad_norm": 0.5325677394866943, "learning_rate": 2.445607857096688e-09, "loss": 0.4651, "step": 9291 }, { "epoch": 2.973122866894198, "grad_norm": 0.48184674978256226, "learning_rate": 2.3877303056923928e-09, "loss": 0.4226, "step": 9292 }, { "epoch": 2.9734428327645053, "grad_norm": 0.46005287766456604, "learning_rate": 2.330545679241625e-09, "loss": 0.4657, "step": 9293 }, { "epoch": 2.9737627986348123, "grad_norm": 0.45725423097610474, "learning_rate": 2.2740539856724866e-09, "loss": 0.4548, "step": 9294 }, { "epoch": 2.9740827645051193, "grad_norm": 0.49043959379196167, "learning_rate": 2.218255232818711e-09, "loss": 0.4154, "step": 9295 }, { "epoch": 2.9744027303754264, "grad_norm": 0.428393691778183, "learning_rate": 2.1631494284163334e-09, "loss": 0.3719, "step": 9296 }, { "epoch": 2.974722696245734, "grad_norm": 0.5096278786659241, "learning_rate": 2.1087365801053526e-09, "loss": 0.5253, "step": 9297 }, { "epoch": 2.975042662116041, "grad_norm": 0.5088695287704468, "learning_rate": 2.055016695430845e-09, "loss": 0.4528, "step": 9298 }, { "epoch": 2.975362627986348, "grad_norm": 0.4937836825847626, "learning_rate": 2.001989781841296e-09, "loss": 0.4428, "step": 9299 }, { "epoch": 2.9756825938566553, "grad_norm": 0.4787329435348511, "learning_rate": 1.949655846688048e-09, "loss": 0.4407, "step": 9300 }, { "epoch": 2.9760025597269624, "grad_norm": 0.47262081503868103, "learning_rate": 1.898014897228628e-09, "loss": 0.4247, "step": 9301 }, { "epoch": 2.9763225255972694, "grad_norm": 0.501465916633606, "learning_rate": 1.8470669406217556e-09, "loss": 0.419, "step": 9302 }, { "epoch": 2.976642491467577, "grad_norm": 0.49505382776260376, "learning_rate": 1.7968119839317787e-09, "loss": 0.4641, "step": 9303 }, { "epoch": 2.976962457337884, "grad_norm": 0.5140721797943115, "learning_rate": 1.7472500341270126e-09, "loss": 0.4348, "step": 9304 }, { "epoch": 2.977282423208191, "grad_norm": 0.5133354663848877, "learning_rate": 1.6983810980791826e-09, "loss": 0.4573, "step": 9305 }, { "epoch": 2.9776023890784984, "grad_norm": 0.4683056175708771, "learning_rate": 1.6502051825639798e-09, "loss": 0.3955, "step": 9306 }, { "epoch": 2.9779223549488054, "grad_norm": 0.5300466418266296, "learning_rate": 1.602722294261061e-09, "loss": 0.4995, "step": 9307 }, { "epoch": 2.9782423208191124, "grad_norm": 0.5027908086776733, "learning_rate": 1.5559324397534937e-09, "loss": 0.4312, "step": 9308 }, { "epoch": 2.97856228668942, "grad_norm": 0.5182973146438599, "learning_rate": 1.5098356255294211e-09, "loss": 0.4583, "step": 9309 }, { "epoch": 2.978882252559727, "grad_norm": 0.49473249912261963, "learning_rate": 1.4644318579798422e-09, "loss": 0.3811, "step": 9310 }, { "epoch": 2.979202218430034, "grad_norm": 0.5003598928451538, "learning_rate": 1.4197211434002767e-09, "loss": 0.4496, "step": 9311 }, { "epoch": 2.9795221843003414, "grad_norm": 0.5183418989181519, "learning_rate": 1.3757034879896547e-09, "loss": 0.4161, "step": 9312 }, { "epoch": 2.9798421501706485, "grad_norm": 0.49360767006874084, "learning_rate": 1.3323788978514275e-09, "loss": 0.4789, "step": 9313 }, { "epoch": 2.9801621160409555, "grad_norm": 0.5218793153762817, "learning_rate": 1.2897473789924563e-09, "loss": 0.4396, "step": 9314 }, { "epoch": 2.980482081911263, "grad_norm": 0.5581367611885071, "learning_rate": 1.2478089373235691e-09, "loss": 0.4587, "step": 9315 }, { "epoch": 2.98080204778157, "grad_norm": 0.5236092805862427, "learning_rate": 1.2065635786595586e-09, "loss": 0.4835, "step": 9316 }, { "epoch": 2.981122013651877, "grad_norm": 0.5084599256515503, "learning_rate": 1.1660113087191839e-09, "loss": 0.4404, "step": 9317 }, { "epoch": 2.9814419795221845, "grad_norm": 0.49674293398857117, "learning_rate": 1.126152133125169e-09, "loss": 0.4192, "step": 9318 }, { "epoch": 2.9817619453924915, "grad_norm": 0.5019124746322632, "learning_rate": 1.0869860574036494e-09, "loss": 0.4674, "step": 9319 }, { "epoch": 2.9820819112627985, "grad_norm": 0.513853132724762, "learning_rate": 1.0485130869858362e-09, "loss": 0.4452, "step": 9320 }, { "epoch": 2.982401877133106, "grad_norm": 0.5213266015052795, "learning_rate": 1.0107332272052406e-09, "loss": 0.4618, "step": 9321 }, { "epoch": 2.982721843003413, "grad_norm": 0.46743345260620117, "learning_rate": 9.7364648330045e-10, "loss": 0.4253, "step": 9322 }, { "epoch": 2.98304180887372, "grad_norm": 0.5130221247673035, "learning_rate": 9.372528604134623e-10, "loss": 0.4654, "step": 9323 }, { "epoch": 2.9833617747440275, "grad_norm": 0.5008715391159058, "learning_rate": 9.015523635907963e-10, "loss": 0.4023, "step": 9324 }, { "epoch": 2.9836817406143346, "grad_norm": 0.44110965728759766, "learning_rate": 8.665449977818263e-10, "loss": 0.414, "step": 9325 }, { "epoch": 2.9840017064846416, "grad_norm": 0.5047010183334351, "learning_rate": 8.322307678398922e-10, "loss": 0.4989, "step": 9326 }, { "epoch": 2.984321672354949, "grad_norm": 0.481261283159256, "learning_rate": 7.98609678523965e-10, "loss": 0.4233, "step": 9327 }, { "epoch": 2.984641638225256, "grad_norm": 0.4948277771472931, "learning_rate": 7.656817344947609e-10, "loss": 0.4815, "step": 9328 }, { "epoch": 2.984961604095563, "grad_norm": 0.44666704535484314, "learning_rate": 7.334469403180721e-10, "loss": 0.3975, "step": 9329 }, { "epoch": 2.9852815699658706, "grad_norm": 0.5035291910171509, "learning_rate": 7.019053004631016e-10, "loss": 0.4343, "step": 9330 }, { "epoch": 2.9856015358361776, "grad_norm": 0.5192816853523254, "learning_rate": 6.710568193035727e-10, "loss": 0.4249, "step": 9331 }, { "epoch": 2.9859215017064846, "grad_norm": 0.5104855298995972, "learning_rate": 6.409015011160646e-10, "loss": 0.4427, "step": 9332 }, { "epoch": 2.9862414675767917, "grad_norm": 0.5305399298667908, "learning_rate": 6.114393500822324e-10, "loss": 0.4319, "step": 9333 }, { "epoch": 2.986561433447099, "grad_norm": 0.5200382471084595, "learning_rate": 5.826703702865866e-10, "loss": 0.4428, "step": 9334 }, { "epoch": 2.986881399317406, "grad_norm": 0.5008326768875122, "learning_rate": 5.545945657181584e-10, "loss": 0.4537, "step": 9335 }, { "epoch": 2.987201365187713, "grad_norm": 0.5317647457122803, "learning_rate": 5.272119402693898e-10, "loss": 0.4286, "step": 9336 }, { "epoch": 2.98752133105802, "grad_norm": 0.5395078659057617, "learning_rate": 5.005224977372436e-10, "loss": 0.428, "step": 9337 }, { "epoch": 2.9878412969283277, "grad_norm": 0.4737189710140228, "learning_rate": 4.745262418226481e-10, "loss": 0.4336, "step": 9338 }, { "epoch": 2.9881612627986347, "grad_norm": 0.4658548831939697, "learning_rate": 4.492231761293875e-10, "loss": 0.3766, "step": 9339 }, { "epoch": 2.9884812286689417, "grad_norm": 0.5189458727836609, "learning_rate": 4.246133041657663e-10, "loss": 0.4898, "step": 9340 }, { "epoch": 2.988801194539249, "grad_norm": 0.5214295387268066, "learning_rate": 4.006966293440551e-10, "loss": 0.4769, "step": 9341 }, { "epoch": 2.989121160409556, "grad_norm": 0.5181043744087219, "learning_rate": 3.7747315498049e-10, "loss": 0.3411, "step": 9342 }, { "epoch": 2.9894411262798632, "grad_norm": 0.5489748120307922, "learning_rate": 3.5494288429527293e-10, "loss": 0.4787, "step": 9343 }, { "epoch": 2.9897610921501707, "grad_norm": 0.540492594242096, "learning_rate": 3.3310582041146124e-10, "loss": 0.4249, "step": 9344 }, { "epoch": 2.9900810580204777, "grad_norm": 0.5374926328659058, "learning_rate": 3.1196196635774335e-10, "loss": 0.4918, "step": 9345 }, { "epoch": 2.9904010238907848, "grad_norm": 0.4693905711174011, "learning_rate": 2.91511325065108e-10, "loss": 0.4418, "step": 9346 }, { "epoch": 2.9907209897610922, "grad_norm": 0.43871355056762695, "learning_rate": 2.7175389936906496e-10, "loss": 0.4497, "step": 9347 }, { "epoch": 2.9910409556313993, "grad_norm": 0.5393214821815491, "learning_rate": 2.526896920090893e-10, "loss": 0.4206, "step": 9348 }, { "epoch": 2.9913609215017063, "grad_norm": 0.5345500111579895, "learning_rate": 2.3431870562917735e-10, "loss": 0.4384, "step": 9349 }, { "epoch": 2.9916808873720138, "grad_norm": 0.4962906539440155, "learning_rate": 2.166409427750704e-10, "loss": 0.4434, "step": 9350 }, { "epoch": 2.992000853242321, "grad_norm": 0.48774805665016174, "learning_rate": 1.9965640589925118e-10, "loss": 0.4356, "step": 9351 }, { "epoch": 2.992320819112628, "grad_norm": 0.5116692781448364, "learning_rate": 1.8336509735539243e-10, "loss": 0.4268, "step": 9352 }, { "epoch": 2.9926407849829353, "grad_norm": 0.5087159276008606, "learning_rate": 1.6776701940335317e-10, "loss": 0.4679, "step": 9353 }, { "epoch": 2.9929607508532423, "grad_norm": 0.5321253538131714, "learning_rate": 1.528621742047376e-10, "loss": 0.4676, "step": 9354 }, { "epoch": 2.9932807167235493, "grad_norm": 0.5010133981704712, "learning_rate": 1.386505638273361e-10, "loss": 0.4319, "step": 9355 }, { "epoch": 2.993600682593857, "grad_norm": 0.48460671305656433, "learning_rate": 1.2513219024068435e-10, "loss": 0.3796, "step": 9356 }, { "epoch": 2.993920648464164, "grad_norm": 0.4713209569454193, "learning_rate": 1.1230705531994901e-10, "loss": 0.4135, "step": 9357 }, { "epoch": 2.994240614334471, "grad_norm": 0.4603041410446167, "learning_rate": 1.0017516084204204e-10, "loss": 0.455, "step": 9358 }, { "epoch": 2.9945605802047783, "grad_norm": 0.4950495660305023, "learning_rate": 8.873650849061666e-11, "loss": 0.4907, "step": 9359 }, { "epoch": 2.9948805460750854, "grad_norm": 0.47134140133857727, "learning_rate": 7.799109985051623e-11, "loss": 0.4352, "step": 9360 }, { "epoch": 2.9952005119453924, "grad_norm": 0.5164924263954163, "learning_rate": 6.793893641221516e-11, "loss": 0.4315, "step": 9361 }, { "epoch": 2.9955204778157, "grad_norm": 0.5457006096839905, "learning_rate": 5.858001956904335e-11, "loss": 0.444, "step": 9362 }, { "epoch": 2.995840443686007, "grad_norm": 0.511311411857605, "learning_rate": 4.9914350618851527e-11, "loss": 0.4658, "step": 9363 }, { "epoch": 2.996160409556314, "grad_norm": 0.49430567026138306, "learning_rate": 4.1941930763456094e-11, "loss": 0.4161, "step": 9364 }, { "epoch": 2.9964803754266214, "grad_norm": 0.5116695761680603, "learning_rate": 3.466276110752897e-11, "loss": 0.4619, "step": 9365 }, { "epoch": 2.9968003412969284, "grad_norm": 0.49776217341423035, "learning_rate": 2.8076842660818003e-11, "loss": 0.4505, "step": 9366 }, { "epoch": 2.9971203071672354, "grad_norm": 0.478668212890625, "learning_rate": 2.2184176336481623e-11, "loss": 0.427, "step": 9367 }, { "epoch": 2.997440273037543, "grad_norm": 0.4620792269706726, "learning_rate": 1.698476295108886e-11, "loss": 0.4259, "step": 9368 }, { "epoch": 2.99776023890785, "grad_norm": 0.5169647336006165, "learning_rate": 1.2478603226284691e-11, "loss": 0.4261, "step": 9369 }, { "epoch": 2.998080204778157, "grad_norm": 0.5284416079521179, "learning_rate": 8.66569778656956e-12, "loss": 0.5411, "step": 9370 }, { "epoch": 2.9984001706484644, "grad_norm": 0.409438818693161, "learning_rate": 5.5460471598545174e-12, "loss": 0.3569, "step": 9371 }, { "epoch": 2.9987201365187715, "grad_norm": 0.5102793574333191, "learning_rate": 3.119651779681654e-12, "loss": 0.4702, "step": 9372 }, { "epoch": 2.9990401023890785, "grad_norm": 0.5210458040237427, "learning_rate": 1.3865119824485462e-12, "loss": 0.4577, "step": 9373 }, { "epoch": 2.9993600682593855, "grad_norm": 0.508808434009552, "learning_rate": 3.466280074082562e-13, "loss": 0.4223, "step": 9374 }, { "epoch": 2.999680034129693, "grad_norm": 0.5104720592498779, "learning_rate": 0.0, "loss": 0.4301, "step": 9375 }, { "epoch": 2.999680034129693, "step": 9375, "total_flos": 5611685260214272.0, "train_loss": 0.5075681580543518, "train_runtime": 141329.7996, "train_samples_per_second": 6.368, "train_steps_per_second": 0.066 } ], "logging_steps": 1.0, "max_steps": 9375, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5611685260214272.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }