{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9936034115138592, "eval_steps": 500, "global_step": 936, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0031982942430703624, "grad_norm": 7.622738078733055, "learning_rate": 1.0638297872340426e-07, "loss": 1.1909, "step": 1 }, { "epoch": 0.006396588486140725, "grad_norm": 7.574176961529125, "learning_rate": 2.1276595744680852e-07, "loss": 1.1838, "step": 2 }, { "epoch": 0.009594882729211088, "grad_norm": 7.4636690419508005, "learning_rate": 3.1914893617021275e-07, "loss": 1.1971, "step": 3 }, { "epoch": 0.01279317697228145, "grad_norm": 7.513735532368614, "learning_rate": 4.2553191489361704e-07, "loss": 1.1795, "step": 4 }, { "epoch": 0.015991471215351813, "grad_norm": 7.333440889375827, "learning_rate": 5.319148936170213e-07, "loss": 1.1448, "step": 5 }, { "epoch": 0.019189765458422176, "grad_norm": 7.3639120181383575, "learning_rate": 6.382978723404255e-07, "loss": 1.1893, "step": 6 }, { "epoch": 0.022388059701492536, "grad_norm": 7.438753810381023, "learning_rate": 7.446808510638298e-07, "loss": 1.2042, "step": 7 }, { "epoch": 0.0255863539445629, "grad_norm": 6.96598869339596, "learning_rate": 8.510638297872341e-07, "loss": 1.1527, "step": 8 }, { "epoch": 0.028784648187633263, "grad_norm": 6.606115410466868, "learning_rate": 9.574468085106384e-07, "loss": 1.1691, "step": 9 }, { "epoch": 0.031982942430703626, "grad_norm": 6.428618826670432, "learning_rate": 1.0638297872340427e-06, "loss": 1.1713, "step": 10 }, { "epoch": 0.035181236673773986, "grad_norm": 5.30277143517507, "learning_rate": 1.170212765957447e-06, "loss": 1.1162, "step": 11 }, { "epoch": 0.03837953091684435, "grad_norm": 5.216323252768823, "learning_rate": 1.276595744680851e-06, "loss": 1.1101, "step": 12 }, { "epoch": 0.04157782515991471, "grad_norm": 5.091458168528318, "learning_rate": 1.3829787234042555e-06, "loss": 1.1071, "step": 13 }, { "epoch": 0.04477611940298507, "grad_norm": 3.3709915569255746, "learning_rate": 1.4893617021276596e-06, "loss": 1.0666, "step": 14 }, { "epoch": 0.04797441364605544, "grad_norm": 3.3278073990556303, "learning_rate": 1.595744680851064e-06, "loss": 1.0571, "step": 15 }, { "epoch": 0.0511727078891258, "grad_norm": 3.1457570812535685, "learning_rate": 1.7021276595744682e-06, "loss": 1.0419, "step": 16 }, { "epoch": 0.054371002132196165, "grad_norm": 3.0524886673021165, "learning_rate": 1.8085106382978727e-06, "loss": 1.0643, "step": 17 }, { "epoch": 0.057569296375266525, "grad_norm": 2.9940425825944055, "learning_rate": 1.9148936170212767e-06, "loss": 1.0184, "step": 18 }, { "epoch": 0.060767590618336885, "grad_norm": 3.353851239083325, "learning_rate": 2.021276595744681e-06, "loss": 1.0121, "step": 19 }, { "epoch": 0.06396588486140725, "grad_norm": 3.7889395105244073, "learning_rate": 2.1276595744680853e-06, "loss": 0.9889, "step": 20 }, { "epoch": 0.06716417910447761, "grad_norm": 3.8783603137474665, "learning_rate": 2.2340425531914894e-06, "loss": 0.996, "step": 21 }, { "epoch": 0.07036247334754797, "grad_norm": 3.5352669616023475, "learning_rate": 2.340425531914894e-06, "loss": 0.9742, "step": 22 }, { "epoch": 0.07356076759061833, "grad_norm": 3.2857152101090605, "learning_rate": 2.446808510638298e-06, "loss": 0.953, "step": 23 }, { "epoch": 0.0767590618336887, "grad_norm": 2.7570947843045857, "learning_rate": 2.553191489361702e-06, "loss": 0.9587, "step": 24 }, { "epoch": 0.07995735607675906, "grad_norm": 1.9278968128889937, "learning_rate": 2.6595744680851065e-06, "loss": 0.9329, "step": 25 }, { "epoch": 0.08315565031982942, "grad_norm": 1.78597612548795, "learning_rate": 2.765957446808511e-06, "loss": 0.9337, "step": 26 }, { "epoch": 0.08635394456289978, "grad_norm": 1.9761748056190767, "learning_rate": 2.8723404255319155e-06, "loss": 0.9235, "step": 27 }, { "epoch": 0.08955223880597014, "grad_norm": 1.9959330898097134, "learning_rate": 2.978723404255319e-06, "loss": 0.9095, "step": 28 }, { "epoch": 0.09275053304904052, "grad_norm": 1.6188409417305705, "learning_rate": 3.0851063829787237e-06, "loss": 0.8924, "step": 29 }, { "epoch": 0.09594882729211088, "grad_norm": 1.5343331514431935, "learning_rate": 3.191489361702128e-06, "loss": 0.8759, "step": 30 }, { "epoch": 0.09914712153518124, "grad_norm": 1.431667627920631, "learning_rate": 3.297872340425532e-06, "loss": 0.8694, "step": 31 }, { "epoch": 0.1023454157782516, "grad_norm": 1.2543200154352037, "learning_rate": 3.4042553191489363e-06, "loss": 0.8512, "step": 32 }, { "epoch": 0.10554371002132196, "grad_norm": 1.0539218969611874, "learning_rate": 3.510638297872341e-06, "loss": 0.8458, "step": 33 }, { "epoch": 0.10874200426439233, "grad_norm": 1.1170568972274193, "learning_rate": 3.6170212765957453e-06, "loss": 0.8899, "step": 34 }, { "epoch": 0.11194029850746269, "grad_norm": 1.2202744700302341, "learning_rate": 3.723404255319149e-06, "loss": 0.8393, "step": 35 }, { "epoch": 0.11513859275053305, "grad_norm": 1.0830349710322051, "learning_rate": 3.8297872340425535e-06, "loss": 0.871, "step": 36 }, { "epoch": 0.11833688699360341, "grad_norm": 0.9052734617810034, "learning_rate": 3.936170212765958e-06, "loss": 0.8202, "step": 37 }, { "epoch": 0.12153518123667377, "grad_norm": 0.8993984549295001, "learning_rate": 4.042553191489362e-06, "loss": 0.801, "step": 38 }, { "epoch": 0.12473347547974413, "grad_norm": 0.9725408899239824, "learning_rate": 4.148936170212766e-06, "loss": 0.8701, "step": 39 }, { "epoch": 0.1279317697228145, "grad_norm": 0.943110853985239, "learning_rate": 4.255319148936171e-06, "loss": 0.8114, "step": 40 }, { "epoch": 0.13113006396588486, "grad_norm": 0.9352658574729733, "learning_rate": 4.361702127659575e-06, "loss": 0.8401, "step": 41 }, { "epoch": 0.13432835820895522, "grad_norm": 1.0170142598852345, "learning_rate": 4.468085106382979e-06, "loss": 0.8198, "step": 42 }, { "epoch": 0.13752665245202558, "grad_norm": 0.9009283876561643, "learning_rate": 4.574468085106383e-06, "loss": 0.8055, "step": 43 }, { "epoch": 0.14072494669509594, "grad_norm": 0.8173324407768994, "learning_rate": 4.680851063829788e-06, "loss": 0.8042, "step": 44 }, { "epoch": 0.1439232409381663, "grad_norm": 0.8483547771193565, "learning_rate": 4.787234042553192e-06, "loss": 0.7773, "step": 45 }, { "epoch": 0.14712153518123666, "grad_norm": 0.7692184274159833, "learning_rate": 4.893617021276596e-06, "loss": 0.7997, "step": 46 }, { "epoch": 0.15031982942430705, "grad_norm": 0.81536863431857, "learning_rate": 5e-06, "loss": 0.7892, "step": 47 }, { "epoch": 0.1535181236673774, "grad_norm": 0.8158497316472542, "learning_rate": 5.106382978723404e-06, "loss": 0.7978, "step": 48 }, { "epoch": 0.15671641791044777, "grad_norm": 0.9305475572050518, "learning_rate": 5.212765957446809e-06, "loss": 0.8087, "step": 49 }, { "epoch": 0.15991471215351813, "grad_norm": 0.8236050831503633, "learning_rate": 5.319148936170213e-06, "loss": 0.7607, "step": 50 }, { "epoch": 0.1631130063965885, "grad_norm": 0.7846154878148738, "learning_rate": 5.425531914893617e-06, "loss": 0.8175, "step": 51 }, { "epoch": 0.16631130063965885, "grad_norm": 0.9528828885498349, "learning_rate": 5.531914893617022e-06, "loss": 0.791, "step": 52 }, { "epoch": 0.1695095948827292, "grad_norm": 0.7710572192032797, "learning_rate": 5.638297872340426e-06, "loss": 0.7986, "step": 53 }, { "epoch": 0.17270788912579957, "grad_norm": 0.8318885111022261, "learning_rate": 5.744680851063831e-06, "loss": 0.7732, "step": 54 }, { "epoch": 0.17590618336886993, "grad_norm": 0.8923328928381091, "learning_rate": 5.851063829787235e-06, "loss": 0.7506, "step": 55 }, { "epoch": 0.1791044776119403, "grad_norm": 0.8476738749734117, "learning_rate": 5.957446808510638e-06, "loss": 0.8041, "step": 56 }, { "epoch": 0.18230277185501065, "grad_norm": 0.6991963183642709, "learning_rate": 6.063829787234044e-06, "loss": 0.743, "step": 57 }, { "epoch": 0.18550106609808104, "grad_norm": 0.9527151297958727, "learning_rate": 6.170212765957447e-06, "loss": 0.7783, "step": 58 }, { "epoch": 0.1886993603411514, "grad_norm": 0.8889593346162057, "learning_rate": 6.276595744680851e-06, "loss": 0.7437, "step": 59 }, { "epoch": 0.19189765458422176, "grad_norm": 0.6699184888475535, "learning_rate": 6.382978723404256e-06, "loss": 0.7714, "step": 60 }, { "epoch": 0.19509594882729211, "grad_norm": 0.7705563645271767, "learning_rate": 6.48936170212766e-06, "loss": 0.7744, "step": 61 }, { "epoch": 0.19829424307036247, "grad_norm": 0.8167415683795367, "learning_rate": 6.595744680851064e-06, "loss": 0.7561, "step": 62 }, { "epoch": 0.20149253731343283, "grad_norm": 0.81962712226257, "learning_rate": 6.702127659574469e-06, "loss": 0.7891, "step": 63 }, { "epoch": 0.2046908315565032, "grad_norm": 0.6986640424503381, "learning_rate": 6.808510638297873e-06, "loss": 0.7572, "step": 64 }, { "epoch": 0.20788912579957355, "grad_norm": 0.756369209825303, "learning_rate": 6.914893617021278e-06, "loss": 0.7458, "step": 65 }, { "epoch": 0.21108742004264391, "grad_norm": 0.7270975619851376, "learning_rate": 7.021276595744682e-06, "loss": 0.7116, "step": 66 }, { "epoch": 0.21428571428571427, "grad_norm": 0.7077904574727415, "learning_rate": 7.127659574468085e-06, "loss": 0.741, "step": 67 }, { "epoch": 0.21748400852878466, "grad_norm": 0.7367201939644344, "learning_rate": 7.234042553191491e-06, "loss": 0.738, "step": 68 }, { "epoch": 0.22068230277185502, "grad_norm": 0.7461379677761271, "learning_rate": 7.340425531914894e-06, "loss": 0.7402, "step": 69 }, { "epoch": 0.22388059701492538, "grad_norm": 0.7635469236526181, "learning_rate": 7.446808510638298e-06, "loss": 0.7539, "step": 70 }, { "epoch": 0.22707889125799574, "grad_norm": 0.8662247901922688, "learning_rate": 7.553191489361703e-06, "loss": 0.7515, "step": 71 }, { "epoch": 0.2302771855010661, "grad_norm": 0.7957789030344851, "learning_rate": 7.659574468085107e-06, "loss": 0.7416, "step": 72 }, { "epoch": 0.23347547974413646, "grad_norm": 0.7423199681203461, "learning_rate": 7.765957446808511e-06, "loss": 0.7629, "step": 73 }, { "epoch": 0.23667377398720682, "grad_norm": 0.9173283664095695, "learning_rate": 7.872340425531916e-06, "loss": 0.7452, "step": 74 }, { "epoch": 0.23987206823027718, "grad_norm": 0.78368199600868, "learning_rate": 7.97872340425532e-06, "loss": 0.7268, "step": 75 }, { "epoch": 0.24307036247334754, "grad_norm": 0.8361765532940366, "learning_rate": 8.085106382978723e-06, "loss": 0.7584, "step": 76 }, { "epoch": 0.2462686567164179, "grad_norm": 0.809289414930847, "learning_rate": 8.191489361702128e-06, "loss": 0.7323, "step": 77 }, { "epoch": 0.24946695095948826, "grad_norm": 0.7351470133306395, "learning_rate": 8.297872340425532e-06, "loss": 0.7363, "step": 78 }, { "epoch": 0.2526652452025586, "grad_norm": 0.8056016102088669, "learning_rate": 8.404255319148937e-06, "loss": 0.723, "step": 79 }, { "epoch": 0.255863539445629, "grad_norm": 0.7345894297808727, "learning_rate": 8.510638297872341e-06, "loss": 0.769, "step": 80 }, { "epoch": 0.25906183368869934, "grad_norm": 0.7863593332470072, "learning_rate": 8.617021276595746e-06, "loss": 0.728, "step": 81 }, { "epoch": 0.2622601279317697, "grad_norm": 0.7627382259597176, "learning_rate": 8.72340425531915e-06, "loss": 0.7283, "step": 82 }, { "epoch": 0.26545842217484006, "grad_norm": 0.7824566533257207, "learning_rate": 8.829787234042555e-06, "loss": 0.7529, "step": 83 }, { "epoch": 0.26865671641791045, "grad_norm": 0.7231171750656687, "learning_rate": 8.936170212765958e-06, "loss": 0.7335, "step": 84 }, { "epoch": 0.27185501066098083, "grad_norm": 0.9121689099240826, "learning_rate": 9.042553191489362e-06, "loss": 0.756, "step": 85 }, { "epoch": 0.27505330490405117, "grad_norm": 0.7192041394994152, "learning_rate": 9.148936170212767e-06, "loss": 0.7008, "step": 86 }, { "epoch": 0.27825159914712155, "grad_norm": 0.926261732519985, "learning_rate": 9.255319148936171e-06, "loss": 0.7169, "step": 87 }, { "epoch": 0.2814498933901919, "grad_norm": 0.7106373516758131, "learning_rate": 9.361702127659576e-06, "loss": 0.7019, "step": 88 }, { "epoch": 0.2846481876332623, "grad_norm": 0.8650610615775703, "learning_rate": 9.46808510638298e-06, "loss": 0.7579, "step": 89 }, { "epoch": 0.2878464818763326, "grad_norm": 0.8537551564881841, "learning_rate": 9.574468085106385e-06, "loss": 0.7151, "step": 90 }, { "epoch": 0.291044776119403, "grad_norm": 0.9816090964885784, "learning_rate": 9.680851063829787e-06, "loss": 0.7361, "step": 91 }, { "epoch": 0.2942430703624733, "grad_norm": 0.7896344914068675, "learning_rate": 9.787234042553192e-06, "loss": 0.7241, "step": 92 }, { "epoch": 0.2974413646055437, "grad_norm": 0.8361537117215864, "learning_rate": 9.893617021276596e-06, "loss": 0.7063, "step": 93 }, { "epoch": 0.3006396588486141, "grad_norm": 0.8817821003662552, "learning_rate": 1e-05, "loss": 0.711, "step": 94 }, { "epoch": 0.30383795309168443, "grad_norm": 1.1654827546629623, "learning_rate": 9.999965197129365e-06, "loss": 0.7493, "step": 95 }, { "epoch": 0.3070362473347548, "grad_norm": 0.6905754899045732, "learning_rate": 9.999860789001947e-06, "loss": 0.7189, "step": 96 }, { "epoch": 0.31023454157782515, "grad_norm": 1.0193923266199096, "learning_rate": 9.999686777071233e-06, "loss": 0.748, "step": 97 }, { "epoch": 0.31343283582089554, "grad_norm": 0.8631527647207492, "learning_rate": 9.999443163759669e-06, "loss": 0.7297, "step": 98 }, { "epoch": 0.31663113006396587, "grad_norm": 0.7959272086157606, "learning_rate": 9.999129952458628e-06, "loss": 0.6892, "step": 99 }, { "epoch": 0.31982942430703626, "grad_norm": 1.019986780308216, "learning_rate": 9.998747147528375e-06, "loss": 0.6945, "step": 100 }, { "epoch": 0.3230277185501066, "grad_norm": 0.9796997640078874, "learning_rate": 9.998294754297992e-06, "loss": 0.7683, "step": 101 }, { "epoch": 0.326226012793177, "grad_norm": 0.9025864669755898, "learning_rate": 9.997772779065312e-06, "loss": 0.707, "step": 102 }, { "epoch": 0.3294243070362473, "grad_norm": 0.9406972010353718, "learning_rate": 9.997181229096831e-06, "loss": 0.7148, "step": 103 }, { "epoch": 0.3326226012793177, "grad_norm": 0.8270166326880105, "learning_rate": 9.996520112627602e-06, "loss": 0.7217, "step": 104 }, { "epoch": 0.3358208955223881, "grad_norm": 0.8843538366343328, "learning_rate": 9.995789438861128e-06, "loss": 0.7195, "step": 105 }, { "epoch": 0.3390191897654584, "grad_norm": 0.8701458072422218, "learning_rate": 9.994989217969224e-06, "loss": 0.7175, "step": 106 }, { "epoch": 0.3422174840085288, "grad_norm": 0.7418076755999989, "learning_rate": 9.994119461091885e-06, "loss": 0.7158, "step": 107 }, { "epoch": 0.34541577825159914, "grad_norm": 0.8703976870763576, "learning_rate": 9.993180180337126e-06, "loss": 0.6813, "step": 108 }, { "epoch": 0.3486140724946695, "grad_norm": 0.9178054118198938, "learning_rate": 9.992171388780814e-06, "loss": 0.712, "step": 109 }, { "epoch": 0.35181236673773986, "grad_norm": 0.8589547671449834, "learning_rate": 9.991093100466482e-06, "loss": 0.7499, "step": 110 }, { "epoch": 0.35501066098081024, "grad_norm": 0.8726951716749137, "learning_rate": 9.989945330405146e-06, "loss": 0.6837, "step": 111 }, { "epoch": 0.3582089552238806, "grad_norm": 0.927713037073185, "learning_rate": 9.988728094575082e-06, "loss": 0.7342, "step": 112 }, { "epoch": 0.36140724946695096, "grad_norm": 0.7505991388636275, "learning_rate": 9.98744140992161e-06, "loss": 0.7094, "step": 113 }, { "epoch": 0.3646055437100213, "grad_norm": 0.894268442242807, "learning_rate": 9.986085294356858e-06, "loss": 0.6777, "step": 114 }, { "epoch": 0.3678038379530917, "grad_norm": 0.8024161706862449, "learning_rate": 9.98465976675951e-06, "loss": 0.7261, "step": 115 }, { "epoch": 0.37100213219616207, "grad_norm": 0.9360632386011705, "learning_rate": 9.983164846974549e-06, "loss": 0.7111, "step": 116 }, { "epoch": 0.3742004264392324, "grad_norm": 0.850822368769409, "learning_rate": 9.981600555812975e-06, "loss": 0.6618, "step": 117 }, { "epoch": 0.3773987206823028, "grad_norm": 0.8087687979800354, "learning_rate": 9.979966915051517e-06, "loss": 0.7151, "step": 118 }, { "epoch": 0.3805970149253731, "grad_norm": 1.019238061277482, "learning_rate": 9.978263947432331e-06, "loss": 0.7419, "step": 119 }, { "epoch": 0.3837953091684435, "grad_norm": 0.6679781593773177, "learning_rate": 9.976491676662679e-06, "loss": 0.6984, "step": 120 }, { "epoch": 0.38699360341151384, "grad_norm": 0.8025666983281575, "learning_rate": 9.974650127414609e-06, "loss": 0.7041, "step": 121 }, { "epoch": 0.39019189765458423, "grad_norm": 0.6935416167824833, "learning_rate": 9.972739325324596e-06, "loss": 0.6972, "step": 122 }, { "epoch": 0.39339019189765456, "grad_norm": 0.8012590306354953, "learning_rate": 9.970759296993205e-06, "loss": 0.6719, "step": 123 }, { "epoch": 0.39658848614072495, "grad_norm": 0.7351333900689639, "learning_rate": 9.968710069984699e-06, "loss": 0.6897, "step": 124 }, { "epoch": 0.3997867803837953, "grad_norm": 0.8515682416300615, "learning_rate": 9.966591672826674e-06, "loss": 0.6911, "step": 125 }, { "epoch": 0.40298507462686567, "grad_norm": 0.7745318085363246, "learning_rate": 9.964404135009649e-06, "loss": 0.6768, "step": 126 }, { "epoch": 0.40618336886993606, "grad_norm": 0.7565019041685611, "learning_rate": 9.962147486986664e-06, "loss": 0.7018, "step": 127 }, { "epoch": 0.4093816631130064, "grad_norm": 0.675609674290225, "learning_rate": 9.959821760172849e-06, "loss": 0.6978, "step": 128 }, { "epoch": 0.4125799573560768, "grad_norm": 0.8186758848215644, "learning_rate": 9.957426986944994e-06, "loss": 0.6931, "step": 129 }, { "epoch": 0.4157782515991471, "grad_norm": 0.7207234783292092, "learning_rate": 9.95496320064109e-06, "loss": 0.7082, "step": 130 }, { "epoch": 0.4189765458422175, "grad_norm": 0.7354378059056297, "learning_rate": 9.952430435559873e-06, "loss": 0.7363, "step": 131 }, { "epoch": 0.42217484008528783, "grad_norm": 0.8258670771003666, "learning_rate": 9.94982872696034e-06, "loss": 0.71, "step": 132 }, { "epoch": 0.4253731343283582, "grad_norm": 0.7436932431094343, "learning_rate": 9.947158111061263e-06, "loss": 0.7276, "step": 133 }, { "epoch": 0.42857142857142855, "grad_norm": 0.6981052362583808, "learning_rate": 9.94441862504068e-06, "loss": 0.6774, "step": 134 }, { "epoch": 0.43176972281449894, "grad_norm": 0.8840543705470293, "learning_rate": 9.941610307035385e-06, "loss": 0.7257, "step": 135 }, { "epoch": 0.4349680170575693, "grad_norm": 0.6694638600717385, "learning_rate": 9.938733196140386e-06, "loss": 0.7029, "step": 136 }, { "epoch": 0.43816631130063965, "grad_norm": 0.8441420481093084, "learning_rate": 9.935787332408375e-06, "loss": 0.7274, "step": 137 }, { "epoch": 0.44136460554371004, "grad_norm": 0.8505776286992377, "learning_rate": 9.932772756849152e-06, "loss": 0.6935, "step": 138 }, { "epoch": 0.4445628997867804, "grad_norm": 0.9226301253646606, "learning_rate": 9.929689511429075e-06, "loss": 0.7264, "step": 139 }, { "epoch": 0.44776119402985076, "grad_norm": 0.7579619597097683, "learning_rate": 9.926537639070457e-06, "loss": 0.7076, "step": 140 }, { "epoch": 0.4509594882729211, "grad_norm": 0.8076488456889905, "learning_rate": 9.923317183650985e-06, "loss": 0.7003, "step": 141 }, { "epoch": 0.4541577825159915, "grad_norm": 0.8485727454100752, "learning_rate": 9.92002819000309e-06, "loss": 0.7211, "step": 142 }, { "epoch": 0.4573560767590618, "grad_norm": 0.751313097685783, "learning_rate": 9.916670703913345e-06, "loss": 0.6859, "step": 143 }, { "epoch": 0.4605543710021322, "grad_norm": 0.7996244442440978, "learning_rate": 9.913244772121811e-06, "loss": 0.7048, "step": 144 }, { "epoch": 0.46375266524520253, "grad_norm": 0.7586742602393676, "learning_rate": 9.90975044232139e-06, "loss": 0.7232, "step": 145 }, { "epoch": 0.4669509594882729, "grad_norm": 0.7333771413773464, "learning_rate": 9.90618776315717e-06, "loss": 0.6973, "step": 146 }, { "epoch": 0.4701492537313433, "grad_norm": 0.7462538908947492, "learning_rate": 9.902556784225729e-06, "loss": 0.6956, "step": 147 }, { "epoch": 0.47334754797441364, "grad_norm": 0.7953075557133956, "learning_rate": 9.898857556074469e-06, "loss": 0.7225, "step": 148 }, { "epoch": 0.47654584221748403, "grad_norm": 0.7060064737270609, "learning_rate": 9.895090130200889e-06, "loss": 0.6814, "step": 149 }, { "epoch": 0.47974413646055436, "grad_norm": 0.8120142107026623, "learning_rate": 9.891254559051886e-06, "loss": 0.6731, "step": 150 }, { "epoch": 0.48294243070362475, "grad_norm": 0.9453211383140191, "learning_rate": 9.887350896023015e-06, "loss": 0.6725, "step": 151 }, { "epoch": 0.4861407249466951, "grad_norm": 0.753978978180176, "learning_rate": 9.883379195457747e-06, "loss": 0.6931, "step": 152 }, { "epoch": 0.48933901918976547, "grad_norm": 0.9761958786101086, "learning_rate": 9.879339512646714e-06, "loss": 0.6964, "step": 153 }, { "epoch": 0.4925373134328358, "grad_norm": 0.8092682567275233, "learning_rate": 9.875231903826936e-06, "loss": 0.7033, "step": 154 }, { "epoch": 0.4957356076759062, "grad_norm": 0.8755474532977587, "learning_rate": 9.871056426181052e-06, "loss": 0.697, "step": 155 }, { "epoch": 0.4989339019189765, "grad_norm": 0.7311549338301705, "learning_rate": 9.8668131378365e-06, "loss": 0.6999, "step": 156 }, { "epoch": 0.502132196162047, "grad_norm": 0.7037792105922891, "learning_rate": 9.862502097864726e-06, "loss": 0.6792, "step": 157 }, { "epoch": 0.5053304904051172, "grad_norm": 0.7713310350403229, "learning_rate": 9.858123366280358e-06, "loss": 0.7028, "step": 158 }, { "epoch": 0.5085287846481876, "grad_norm": 0.7608743370139368, "learning_rate": 9.853677004040368e-06, "loss": 0.7262, "step": 159 }, { "epoch": 0.511727078891258, "grad_norm": 0.7523293064381485, "learning_rate": 9.849163073043223e-06, "loss": 0.6968, "step": 160 }, { "epoch": 0.5149253731343284, "grad_norm": 0.7701188802305937, "learning_rate": 9.844581636128025e-06, "loss": 0.6878, "step": 161 }, { "epoch": 0.5181236673773987, "grad_norm": 0.6850118378660099, "learning_rate": 9.83993275707364e-06, "loss": 0.6843, "step": 162 }, { "epoch": 0.5213219616204691, "grad_norm": 0.7633903463897005, "learning_rate": 9.835216500597797e-06, "loss": 0.6625, "step": 163 }, { "epoch": 0.5245202558635395, "grad_norm": 0.7032490531444039, "learning_rate": 9.830432932356207e-06, "loss": 0.6787, "step": 164 }, { "epoch": 0.5277185501066098, "grad_norm": 0.8377452657391034, "learning_rate": 9.82558211894163e-06, "loss": 0.6485, "step": 165 }, { "epoch": 0.5309168443496801, "grad_norm": 0.6618917554232225, "learning_rate": 9.820664127882958e-06, "loss": 0.6789, "step": 166 }, { "epoch": 0.5341151385927505, "grad_norm": 0.8178971073023995, "learning_rate": 9.815679027644273e-06, "loss": 0.7136, "step": 167 }, { "epoch": 0.5373134328358209, "grad_norm": 0.7226926940415138, "learning_rate": 9.8106268876239e-06, "loss": 0.6973, "step": 168 }, { "epoch": 0.5405117270788913, "grad_norm": 0.8219860052512269, "learning_rate": 9.805507778153423e-06, "loss": 0.6833, "step": 169 }, { "epoch": 0.5437100213219617, "grad_norm": 0.8090438198755149, "learning_rate": 9.800321770496726e-06, "loss": 0.6683, "step": 170 }, { "epoch": 0.5469083155650319, "grad_norm": 0.8076219868171068, "learning_rate": 9.79506893684899e-06, "loss": 0.6742, "step": 171 }, { "epoch": 0.5501066098081023, "grad_norm": 0.7650394887530159, "learning_rate": 9.789749350335693e-06, "loss": 0.6809, "step": 172 }, { "epoch": 0.5533049040511727, "grad_norm": 0.6909569362578575, "learning_rate": 9.784363085011587e-06, "loss": 0.6987, "step": 173 }, { "epoch": 0.5565031982942431, "grad_norm": 0.7544714952212014, "learning_rate": 9.778910215859666e-06, "loss": 0.7286, "step": 174 }, { "epoch": 0.5597014925373134, "grad_norm": 0.6910806053162777, "learning_rate": 9.773390818790136e-06, "loss": 0.7004, "step": 175 }, { "epoch": 0.5628997867803838, "grad_norm": 0.728635086997373, "learning_rate": 9.767804970639338e-06, "loss": 0.7107, "step": 176 }, { "epoch": 0.5660980810234542, "grad_norm": 0.7164830915823774, "learning_rate": 9.762152749168693e-06, "loss": 0.6659, "step": 177 }, { "epoch": 0.5692963752665245, "grad_norm": 0.7185222557170264, "learning_rate": 9.756434233063616e-06, "loss": 0.7204, "step": 178 }, { "epoch": 0.5724946695095949, "grad_norm": 0.6376198071084583, "learning_rate": 9.750649501932414e-06, "loss": 0.6851, "step": 179 }, { "epoch": 0.5756929637526652, "grad_norm": 0.7372459805811605, "learning_rate": 9.744798636305189e-06, "loss": 0.6655, "step": 180 }, { "epoch": 0.5788912579957356, "grad_norm": 0.7192794699200221, "learning_rate": 9.738881717632709e-06, "loss": 0.6981, "step": 181 }, { "epoch": 0.582089552238806, "grad_norm": 0.8372267929782506, "learning_rate": 9.732898828285273e-06, "loss": 0.6609, "step": 182 }, { "epoch": 0.5852878464818764, "grad_norm": 0.6890860653831234, "learning_rate": 9.726850051551575e-06, "loss": 0.6849, "step": 183 }, { "epoch": 0.5884861407249466, "grad_norm": 1.0979021153485466, "learning_rate": 9.72073547163753e-06, "loss": 0.7055, "step": 184 }, { "epoch": 0.591684434968017, "grad_norm": 0.718530118293507, "learning_rate": 9.714555173665112e-06, "loss": 0.6793, "step": 185 }, { "epoch": 0.5948827292110874, "grad_norm": 0.9568746124941592, "learning_rate": 9.708309243671167e-06, "loss": 0.6805, "step": 186 }, { "epoch": 0.5980810234541578, "grad_norm": 0.9190645493855029, "learning_rate": 9.701997768606209e-06, "loss": 0.6578, "step": 187 }, { "epoch": 0.6012793176972282, "grad_norm": 0.8452449705256939, "learning_rate": 9.695620836333219e-06, "loss": 0.6752, "step": 188 }, { "epoch": 0.6044776119402985, "grad_norm": 0.912469389514112, "learning_rate": 9.68917853562642e-06, "loss": 0.677, "step": 189 }, { "epoch": 0.6076759061833689, "grad_norm": 0.8380041206744561, "learning_rate": 9.68267095617003e-06, "loss": 0.6985, "step": 190 }, { "epoch": 0.6108742004264393, "grad_norm": 0.863868290234638, "learning_rate": 9.676098188557032e-06, "loss": 0.7313, "step": 191 }, { "epoch": 0.6140724946695096, "grad_norm": 0.8174466263379166, "learning_rate": 9.669460324287899e-06, "loss": 0.7053, "step": 192 }, { "epoch": 0.6172707889125799, "grad_norm": 0.8909312106674119, "learning_rate": 9.662757455769317e-06, "loss": 0.682, "step": 193 }, { "epoch": 0.6204690831556503, "grad_norm": 0.7302375395625422, "learning_rate": 9.655989676312918e-06, "loss": 0.6594, "step": 194 }, { "epoch": 0.6236673773987207, "grad_norm": 0.8735527453330153, "learning_rate": 9.649157080133962e-06, "loss": 0.6674, "step": 195 }, { "epoch": 0.6268656716417911, "grad_norm": 0.7064746459902383, "learning_rate": 9.642259762350034e-06, "loss": 0.7017, "step": 196 }, { "epoch": 0.6300639658848614, "grad_norm": 0.8655024689230618, "learning_rate": 9.635297818979715e-06, "loss": 0.6661, "step": 197 }, { "epoch": 0.6332622601279317, "grad_norm": 0.728341775629865, "learning_rate": 9.628271346941252e-06, "loss": 0.6925, "step": 198 }, { "epoch": 0.6364605543710021, "grad_norm": 0.6968228679265215, "learning_rate": 9.621180444051206e-06, "loss": 0.685, "step": 199 }, { "epoch": 0.6396588486140725, "grad_norm": 0.8921095235064438, "learning_rate": 9.614025209023084e-06, "loss": 0.7016, "step": 200 }, { "epoch": 0.6428571428571429, "grad_norm": 0.7770028881285537, "learning_rate": 9.606805741465977e-06, "loss": 0.7077, "step": 201 }, { "epoch": 0.6460554371002132, "grad_norm": 0.7270295337970796, "learning_rate": 9.59952214188316e-06, "loss": 0.6779, "step": 202 }, { "epoch": 0.6492537313432836, "grad_norm": 0.8302887272339062, "learning_rate": 9.592174511670704e-06, "loss": 0.6722, "step": 203 }, { "epoch": 0.652452025586354, "grad_norm": 0.8688719467779429, "learning_rate": 9.58476295311606e-06, "loss": 0.7104, "step": 204 }, { "epoch": 0.6556503198294243, "grad_norm": 0.6323690135784534, "learning_rate": 9.577287569396632e-06, "loss": 0.6927, "step": 205 }, { "epoch": 0.6588486140724946, "grad_norm": 0.8941706514509578, "learning_rate": 9.569748464578343e-06, "loss": 0.6926, "step": 206 }, { "epoch": 0.662046908315565, "grad_norm": 0.7211897394095068, "learning_rate": 9.562145743614193e-06, "loss": 0.7006, "step": 207 }, { "epoch": 0.6652452025586354, "grad_norm": 0.655783741036758, "learning_rate": 9.554479512342785e-06, "loss": 0.7108, "step": 208 }, { "epoch": 0.6684434968017058, "grad_norm": 0.7467548640108442, "learning_rate": 9.54674987748686e-06, "loss": 0.6649, "step": 209 }, { "epoch": 0.6716417910447762, "grad_norm": 0.7171171346221873, "learning_rate": 9.538956946651816e-06, "loss": 0.6531, "step": 210 }, { "epoch": 0.6748400852878464, "grad_norm": 0.7120324790582493, "learning_rate": 9.531100828324191e-06, "loss": 0.6795, "step": 211 }, { "epoch": 0.6780383795309168, "grad_norm": 0.675292328763824, "learning_rate": 9.52318163187018e-06, "loss": 0.6883, "step": 212 }, { "epoch": 0.6812366737739872, "grad_norm": 0.9706242893642346, "learning_rate": 9.515199467534086e-06, "loss": 0.7105, "step": 213 }, { "epoch": 0.6844349680170576, "grad_norm": 0.6339606603859417, "learning_rate": 9.507154446436806e-06, "loss": 0.7006, "step": 214 }, { "epoch": 0.6876332622601279, "grad_norm": 0.8270660302761437, "learning_rate": 9.499046680574267e-06, "loss": 0.6937, "step": 215 }, { "epoch": 0.6908315565031983, "grad_norm": 0.8339411710564606, "learning_rate": 9.490876282815884e-06, "loss": 0.7089, "step": 216 }, { "epoch": 0.6940298507462687, "grad_norm": 0.7095006207653165, "learning_rate": 9.482643366902972e-06, "loss": 0.6774, "step": 217 }, { "epoch": 0.697228144989339, "grad_norm": 0.8701992961050132, "learning_rate": 9.474348047447177e-06, "loss": 0.6987, "step": 218 }, { "epoch": 0.7004264392324094, "grad_norm": 0.8555157892062962, "learning_rate": 9.465990439928868e-06, "loss": 0.6888, "step": 219 }, { "epoch": 0.7036247334754797, "grad_norm": 0.7482188301136834, "learning_rate": 9.457570660695542e-06, "loss": 0.6924, "step": 220 }, { "epoch": 0.7068230277185501, "grad_norm": 0.6679769126751529, "learning_rate": 9.449088826960187e-06, "loss": 0.6814, "step": 221 }, { "epoch": 0.7100213219616205, "grad_norm": 0.6481228296820869, "learning_rate": 9.440545056799677e-06, "loss": 0.6927, "step": 222 }, { "epoch": 0.7132196162046909, "grad_norm": 0.6138493570700504, "learning_rate": 9.431939469153096e-06, "loss": 0.6848, "step": 223 }, { "epoch": 0.7164179104477612, "grad_norm": 0.6857392440178816, "learning_rate": 9.423272183820109e-06, "loss": 0.7016, "step": 224 }, { "epoch": 0.7196162046908315, "grad_norm": 0.6565822633668216, "learning_rate": 9.41454332145928e-06, "loss": 0.684, "step": 225 }, { "epoch": 0.7228144989339019, "grad_norm": 0.7686352189035123, "learning_rate": 9.405753003586396e-06, "loss": 0.662, "step": 226 }, { "epoch": 0.7260127931769723, "grad_norm": 0.6889440928778451, "learning_rate": 9.396901352572771e-06, "loss": 0.6757, "step": 227 }, { "epoch": 0.7292110874200426, "grad_norm": 0.6644883797889718, "learning_rate": 9.387988491643558e-06, "loss": 0.69, "step": 228 }, { "epoch": 0.732409381663113, "grad_norm": 0.7626373390965248, "learning_rate": 9.379014544876011e-06, "loss": 0.7187, "step": 229 }, { "epoch": 0.7356076759061834, "grad_norm": 0.704507663933334, "learning_rate": 9.369979637197774e-06, "loss": 0.6585, "step": 230 }, { "epoch": 0.7388059701492538, "grad_norm": 0.7336157000029004, "learning_rate": 9.360883894385137e-06, "loss": 0.7127, "step": 231 }, { "epoch": 0.7420042643923241, "grad_norm": 0.6546636167664962, "learning_rate": 9.351727443061284e-06, "loss": 0.704, "step": 232 }, { "epoch": 0.7452025586353944, "grad_norm": 0.7043313515356551, "learning_rate": 9.342510410694529e-06, "loss": 0.6852, "step": 233 }, { "epoch": 0.7484008528784648, "grad_norm": 0.9011726584289667, "learning_rate": 9.33323292559655e-06, "loss": 0.7081, "step": 234 }, { "epoch": 0.7515991471215352, "grad_norm": 0.7292470054254178, "learning_rate": 9.323895116920591e-06, "loss": 0.669, "step": 235 }, { "epoch": 0.7547974413646056, "grad_norm": 0.8382635859763398, "learning_rate": 9.31449711465967e-06, "loss": 0.6763, "step": 236 }, { "epoch": 0.7579957356076759, "grad_norm": 0.7483080444786336, "learning_rate": 9.305039049644772e-06, "loss": 0.6567, "step": 237 }, { "epoch": 0.7611940298507462, "grad_norm": 0.7261935584991153, "learning_rate": 9.29552105354302e-06, "loss": 0.6714, "step": 238 }, { "epoch": 0.7643923240938166, "grad_norm": 0.8645856822885789, "learning_rate": 9.28594325885585e-06, "loss": 0.6625, "step": 239 }, { "epoch": 0.767590618336887, "grad_norm": 0.7568040003927768, "learning_rate": 9.27630579891716e-06, "loss": 0.6191, "step": 240 }, { "epoch": 0.7707889125799574, "grad_norm": 0.7959983248684005, "learning_rate": 9.266608807891459e-06, "loss": 0.72, "step": 241 }, { "epoch": 0.7739872068230277, "grad_norm": 0.6158464477516697, "learning_rate": 9.256852420771999e-06, "loss": 0.6984, "step": 242 }, { "epoch": 0.7771855010660981, "grad_norm": 0.6475971949577826, "learning_rate": 9.24703677337889e-06, "loss": 0.6707, "step": 243 }, { "epoch": 0.7803837953091685, "grad_norm": 0.7514772530157263, "learning_rate": 9.237162002357214e-06, "loss": 0.7025, "step": 244 }, { "epoch": 0.7835820895522388, "grad_norm": 0.7153934193506174, "learning_rate": 9.227228245175127e-06, "loss": 0.7084, "step": 245 }, { "epoch": 0.7867803837953091, "grad_norm": 0.6510532625278841, "learning_rate": 9.217235640121927e-06, "loss": 0.7032, "step": 246 }, { "epoch": 0.7899786780383795, "grad_norm": 0.648491391649832, "learning_rate": 9.207184326306155e-06, "loss": 0.703, "step": 247 }, { "epoch": 0.7931769722814499, "grad_norm": 0.8034299471010478, "learning_rate": 9.197074443653643e-06, "loss": 0.7118, "step": 248 }, { "epoch": 0.7963752665245203, "grad_norm": 0.7868890537454554, "learning_rate": 9.186906132905563e-06, "loss": 0.6907, "step": 249 }, { "epoch": 0.7995735607675906, "grad_norm": 0.7573924459698659, "learning_rate": 9.176679535616477e-06, "loss": 0.7133, "step": 250 }, { "epoch": 0.802771855010661, "grad_norm": 0.7039169561600006, "learning_rate": 9.166394794152363e-06, "loss": 0.6749, "step": 251 }, { "epoch": 0.8059701492537313, "grad_norm": 0.8423685923463412, "learning_rate": 9.156052051688633e-06, "loss": 0.7378, "step": 252 }, { "epoch": 0.8091684434968017, "grad_norm": 0.6131291481922342, "learning_rate": 9.145651452208133e-06, "loss": 0.6415, "step": 253 }, { "epoch": 0.8123667377398721, "grad_norm": 0.7195983825244481, "learning_rate": 9.135193140499155e-06, "loss": 0.6613, "step": 254 }, { "epoch": 0.8155650319829424, "grad_norm": 0.8577511387252152, "learning_rate": 9.124677262153405e-06, "loss": 0.721, "step": 255 }, { "epoch": 0.8187633262260128, "grad_norm": 0.7278841089447043, "learning_rate": 9.114103963563986e-06, "loss": 0.6738, "step": 256 }, { "epoch": 0.8219616204690832, "grad_norm": 0.6856569850641855, "learning_rate": 9.103473391923354e-06, "loss": 0.6588, "step": 257 }, { "epoch": 0.8251599147121536, "grad_norm": 0.9021119379041687, "learning_rate": 9.092785695221271e-06, "loss": 0.7182, "step": 258 }, { "epoch": 0.8283582089552238, "grad_norm": 0.8796683162834094, "learning_rate": 9.08204102224275e-06, "loss": 0.6816, "step": 259 }, { "epoch": 0.8315565031982942, "grad_norm": 0.847520407573671, "learning_rate": 9.071239522565978e-06, "loss": 0.6508, "step": 260 }, { "epoch": 0.8347547974413646, "grad_norm": 0.9049944252517715, "learning_rate": 9.06038134656023e-06, "loss": 0.6925, "step": 261 }, { "epoch": 0.837953091684435, "grad_norm": 0.8571499479329611, "learning_rate": 9.049466645383785e-06, "loss": 0.681, "step": 262 }, { "epoch": 0.8411513859275054, "grad_norm": 0.745522815593341, "learning_rate": 9.038495570981814e-06, "loss": 0.6525, "step": 263 }, { "epoch": 0.8443496801705757, "grad_norm": 0.947292350163867, "learning_rate": 9.027468276084274e-06, "loss": 0.6696, "step": 264 }, { "epoch": 0.847547974413646, "grad_norm": 0.8373012552761833, "learning_rate": 9.016384914203771e-06, "loss": 0.6782, "step": 265 }, { "epoch": 0.8507462686567164, "grad_norm": 0.7211699745049627, "learning_rate": 9.00524563963343e-06, "loss": 0.7015, "step": 266 }, { "epoch": 0.8539445628997868, "grad_norm": 0.973744577195585, "learning_rate": 8.99405060744474e-06, "loss": 0.6598, "step": 267 }, { "epoch": 0.8571428571428571, "grad_norm": 0.7600474078060623, "learning_rate": 8.982799973485407e-06, "loss": 0.6853, "step": 268 }, { "epoch": 0.8603411513859275, "grad_norm": 0.7568774174674711, "learning_rate": 8.971493894377174e-06, "loss": 0.6963, "step": 269 }, { "epoch": 0.8635394456289979, "grad_norm": 1.2096518035065966, "learning_rate": 8.960132527513642e-06, "loss": 0.6901, "step": 270 }, { "epoch": 0.8667377398720683, "grad_norm": 0.8943559219904526, "learning_rate": 8.94871603105809e-06, "loss": 0.6782, "step": 271 }, { "epoch": 0.8699360341151386, "grad_norm": 0.7461346678625066, "learning_rate": 8.937244563941248e-06, "loss": 0.7018, "step": 272 }, { "epoch": 0.8731343283582089, "grad_norm": 1.0000505503295154, "learning_rate": 8.925718285859118e-06, "loss": 0.6875, "step": 273 }, { "epoch": 0.8763326226012793, "grad_norm": 0.7858176236082158, "learning_rate": 8.914137357270723e-06, "loss": 0.7121, "step": 274 }, { "epoch": 0.8795309168443497, "grad_norm": 0.7063872884648656, "learning_rate": 8.902501939395887e-06, "loss": 0.6695, "step": 275 }, { "epoch": 0.8827292110874201, "grad_norm": 0.8044416558628715, "learning_rate": 8.890812194212987e-06, "loss": 0.7076, "step": 276 }, { "epoch": 0.8859275053304904, "grad_norm": 0.6393740328761567, "learning_rate": 8.879068284456702e-06, "loss": 0.6556, "step": 277 }, { "epoch": 0.8891257995735607, "grad_norm": 0.857231208160294, "learning_rate": 8.867270373615735e-06, "loss": 0.7148, "step": 278 }, { "epoch": 0.8923240938166311, "grad_norm": 0.6337872608316819, "learning_rate": 8.855418625930556e-06, "loss": 0.7365, "step": 279 }, { "epoch": 0.8955223880597015, "grad_norm": 0.6706185344174747, "learning_rate": 8.8435132063911e-06, "loss": 0.6699, "step": 280 }, { "epoch": 0.8987206823027718, "grad_norm": 0.7688016478702769, "learning_rate": 8.83155428073448e-06, "loss": 0.6695, "step": 281 }, { "epoch": 0.9019189765458422, "grad_norm": 0.6859750413913875, "learning_rate": 8.81954201544267e-06, "loss": 0.7105, "step": 282 }, { "epoch": 0.9051172707889126, "grad_norm": 0.736049871487562, "learning_rate": 8.8074765777402e-06, "loss": 0.6566, "step": 283 }, { "epoch": 0.908315565031983, "grad_norm": 0.6566455673708244, "learning_rate": 8.79535813559181e-06, "loss": 0.7087, "step": 284 }, { "epoch": 0.9115138592750534, "grad_norm": 0.6678755774132027, "learning_rate": 8.783186857700137e-06, "loss": 0.6803, "step": 285 }, { "epoch": 0.9147121535181236, "grad_norm": 0.6358862703989032, "learning_rate": 8.77096291350334e-06, "loss": 0.6477, "step": 286 }, { "epoch": 0.917910447761194, "grad_norm": 0.621135300076204, "learning_rate": 8.75868647317276e-06, "loss": 0.6853, "step": 287 }, { "epoch": 0.9211087420042644, "grad_norm": 0.8269166477046754, "learning_rate": 8.746357707610544e-06, "loss": 0.7068, "step": 288 }, { "epoch": 0.9243070362473348, "grad_norm": 0.754354903003167, "learning_rate": 8.733976788447265e-06, "loss": 0.6775, "step": 289 }, { "epoch": 0.9275053304904051, "grad_norm": 0.6328527198398567, "learning_rate": 8.721543888039534e-06, "loss": 0.6738, "step": 290 }, { "epoch": 0.9307036247334755, "grad_norm": 0.7230763976135067, "learning_rate": 8.709059179467598e-06, "loss": 0.6448, "step": 291 }, { "epoch": 0.9339019189765458, "grad_norm": 0.6877836548216805, "learning_rate": 8.69652283653294e-06, "loss": 0.6798, "step": 292 }, { "epoch": 0.9371002132196162, "grad_norm": 0.7564591211606494, "learning_rate": 8.683935033755848e-06, "loss": 0.6953, "step": 293 }, { "epoch": 0.9402985074626866, "grad_norm": 0.668332745907526, "learning_rate": 8.671295946372989e-06, "loss": 0.681, "step": 294 }, { "epoch": 0.9434968017057569, "grad_norm": 0.7159527856011911, "learning_rate": 8.658605750334972e-06, "loss": 0.6988, "step": 295 }, { "epoch": 0.9466950959488273, "grad_norm": 0.6848922808640394, "learning_rate": 8.6458646223039e-06, "loss": 0.641, "step": 296 }, { "epoch": 0.9498933901918977, "grad_norm": 0.7548334675392807, "learning_rate": 8.6330727396509e-06, "loss": 0.6713, "step": 297 }, { "epoch": 0.9530916844349681, "grad_norm": 0.5861356337358591, "learning_rate": 8.620230280453672e-06, "loss": 0.691, "step": 298 }, { "epoch": 0.9562899786780383, "grad_norm": 0.7949755015478441, "learning_rate": 8.607337423493996e-06, "loss": 0.6808, "step": 299 }, { "epoch": 0.9594882729211087, "grad_norm": 0.6765008542247006, "learning_rate": 8.594394348255239e-06, "loss": 0.6898, "step": 300 }, { "epoch": 0.9626865671641791, "grad_norm": 0.5806572889436659, "learning_rate": 8.581401234919873e-06, "loss": 0.6492, "step": 301 }, { "epoch": 0.9658848614072495, "grad_norm": 0.7469524089876209, "learning_rate": 8.568358264366958e-06, "loss": 0.6821, "step": 302 }, { "epoch": 0.9690831556503199, "grad_norm": 0.7249833696216659, "learning_rate": 8.555265618169615e-06, "loss": 0.6522, "step": 303 }, { "epoch": 0.9722814498933902, "grad_norm": 0.6218765386847785, "learning_rate": 8.542123478592518e-06, "loss": 0.6843, "step": 304 }, { "epoch": 0.9754797441364605, "grad_norm": 0.7551061996851178, "learning_rate": 8.528932028589337e-06, "loss": 0.6728, "step": 305 }, { "epoch": 0.9786780383795309, "grad_norm": 0.674338314023206, "learning_rate": 8.515691451800206e-06, "loss": 0.6736, "step": 306 }, { "epoch": 0.9818763326226013, "grad_norm": 0.7849066941451676, "learning_rate": 8.502401932549154e-06, "loss": 0.6878, "step": 307 }, { "epoch": 0.9850746268656716, "grad_norm": 0.7498368721683543, "learning_rate": 8.489063655841552e-06, "loss": 0.6903, "step": 308 }, { "epoch": 0.988272921108742, "grad_norm": 0.665126766182557, "learning_rate": 8.475676807361526e-06, "loss": 0.6552, "step": 309 }, { "epoch": 0.9914712153518124, "grad_norm": 0.637731666313635, "learning_rate": 8.462241573469378e-06, "loss": 0.6769, "step": 310 }, { "epoch": 0.9946695095948828, "grad_norm": 0.6112924731775418, "learning_rate": 8.448758141198991e-06, "loss": 0.677, "step": 311 }, { "epoch": 0.997867803837953, "grad_norm": 0.7383946127305492, "learning_rate": 8.435226698255228e-06, "loss": 0.6746, "step": 312 }, { "epoch": 1.0010660980810235, "grad_norm": 0.626462069677627, "learning_rate": 8.421647433011306e-06, "loss": 0.6566, "step": 313 }, { "epoch": 1.004264392324094, "grad_norm": 0.7510688924872979, "learning_rate": 8.408020534506195e-06, "loss": 0.6281, "step": 314 }, { "epoch": 1.007462686567164, "grad_norm": 0.7595637013734888, "learning_rate": 8.394346192441967e-06, "loss": 0.6484, "step": 315 }, { "epoch": 1.0106609808102345, "grad_norm": 0.8191267012599663, "learning_rate": 8.380624597181165e-06, "loss": 0.6269, "step": 316 }, { "epoch": 1.0138592750533049, "grad_norm": 0.6920756752948478, "learning_rate": 8.366855939744152e-06, "loss": 0.6407, "step": 317 }, { "epoch": 1.0170575692963753, "grad_norm": 0.8113121478207607, "learning_rate": 8.353040411806449e-06, "loss": 0.6379, "step": 318 }, { "epoch": 1.0202558635394456, "grad_norm": 0.6038342001520437, "learning_rate": 8.339178205696067e-06, "loss": 0.6219, "step": 319 }, { "epoch": 1.023454157782516, "grad_norm": 0.6525041753408835, "learning_rate": 8.325269514390835e-06, "loss": 0.6451, "step": 320 }, { "epoch": 1.0266524520255864, "grad_norm": 0.7876363756773992, "learning_rate": 8.311314531515707e-06, "loss": 0.6556, "step": 321 }, { "epoch": 1.0298507462686568, "grad_norm": 0.6729418158706656, "learning_rate": 8.297313451340064e-06, "loss": 0.6278, "step": 322 }, { "epoch": 1.033049040511727, "grad_norm": 0.7323640919082973, "learning_rate": 8.283266468775024e-06, "loss": 0.635, "step": 323 }, { "epoch": 1.0362473347547974, "grad_norm": 0.7456586009899282, "learning_rate": 8.269173779370712e-06, "loss": 0.6301, "step": 324 }, { "epoch": 1.0394456289978677, "grad_norm": 0.6472864535960438, "learning_rate": 8.255035579313545e-06, "loss": 0.6514, "step": 325 }, { "epoch": 1.0426439232409381, "grad_norm": 0.7261880506431115, "learning_rate": 8.240852065423507e-06, "loss": 0.6255, "step": 326 }, { "epoch": 1.0458422174840085, "grad_norm": 0.7277147979365199, "learning_rate": 8.226623435151389e-06, "loss": 0.5958, "step": 327 }, { "epoch": 1.049040511727079, "grad_norm": 0.6624168018368254, "learning_rate": 8.21234988657607e-06, "loss": 0.6179, "step": 328 }, { "epoch": 1.0522388059701493, "grad_norm": 0.7453935828883017, "learning_rate": 8.198031618401733e-06, "loss": 0.6108, "step": 329 }, { "epoch": 1.0554371002132197, "grad_norm": 0.7264796554563445, "learning_rate": 8.183668829955111e-06, "loss": 0.6238, "step": 330 }, { "epoch": 1.05863539445629, "grad_norm": 0.7682551654401545, "learning_rate": 8.169261721182715e-06, "loss": 0.6191, "step": 331 }, { "epoch": 1.0618336886993602, "grad_norm": 0.6228417217082994, "learning_rate": 8.154810492648038e-06, "loss": 0.6316, "step": 332 }, { "epoch": 1.0650319829424306, "grad_norm": 0.7327288572841154, "learning_rate": 8.140315345528778e-06, "loss": 0.6586, "step": 333 }, { "epoch": 1.068230277185501, "grad_norm": 0.6964883436081012, "learning_rate": 8.125776481614025e-06, "loss": 0.6414, "step": 334 }, { "epoch": 1.0714285714285714, "grad_norm": 0.6054633221728152, "learning_rate": 8.111194103301461e-06, "loss": 0.5893, "step": 335 }, { "epoch": 1.0746268656716418, "grad_norm": 0.6959078284441366, "learning_rate": 8.096568413594533e-06, "loss": 0.6209, "step": 336 }, { "epoch": 1.0778251599147122, "grad_norm": 0.6687442847144212, "learning_rate": 8.081899616099638e-06, "loss": 0.6395, "step": 337 }, { "epoch": 1.0810234541577826, "grad_norm": 0.6672132910162282, "learning_rate": 8.067187915023283e-06, "loss": 0.6453, "step": 338 }, { "epoch": 1.084221748400853, "grad_norm": 0.7011871375952131, "learning_rate": 8.052433515169235e-06, "loss": 0.6578, "step": 339 }, { "epoch": 1.0874200426439233, "grad_norm": 0.6341715440450284, "learning_rate": 8.037636621935686e-06, "loss": 0.5993, "step": 340 }, { "epoch": 1.0906183368869935, "grad_norm": 0.7942056782846189, "learning_rate": 8.022797441312376e-06, "loss": 0.6377, "step": 341 }, { "epoch": 1.0938166311300639, "grad_norm": 0.6703609675609049, "learning_rate": 8.007916179877742e-06, "loss": 0.6262, "step": 342 }, { "epoch": 1.0970149253731343, "grad_norm": 0.6839776213487608, "learning_rate": 7.99299304479603e-06, "loss": 0.6244, "step": 343 }, { "epoch": 1.1002132196162047, "grad_norm": 0.7336708760979732, "learning_rate": 7.978028243814416e-06, "loss": 0.6607, "step": 344 }, { "epoch": 1.103411513859275, "grad_norm": 0.6632430703502449, "learning_rate": 7.96302198526011e-06, "loss": 0.6556, "step": 345 }, { "epoch": 1.1066098081023454, "grad_norm": 0.695919842154835, "learning_rate": 7.947974478037468e-06, "loss": 0.6194, "step": 346 }, { "epoch": 1.1098081023454158, "grad_norm": 0.654985383307004, "learning_rate": 7.932885931625063e-06, "loss": 0.6015, "step": 347 }, { "epoch": 1.1130063965884862, "grad_norm": 0.7284161828955869, "learning_rate": 7.917756556072792e-06, "loss": 0.6238, "step": 348 }, { "epoch": 1.1162046908315566, "grad_norm": 0.7633948849076417, "learning_rate": 7.902586561998928e-06, "loss": 0.6829, "step": 349 }, { "epoch": 1.1194029850746268, "grad_norm": 0.7808526782166596, "learning_rate": 7.887376160587214e-06, "loss": 0.6286, "step": 350 }, { "epoch": 1.1226012793176972, "grad_norm": 0.7168210695937199, "learning_rate": 7.8721255635839e-06, "loss": 0.655, "step": 351 }, { "epoch": 1.1257995735607675, "grad_norm": 0.7093615169947094, "learning_rate": 7.85683498329481e-06, "loss": 0.6082, "step": 352 }, { "epoch": 1.128997867803838, "grad_norm": 0.7319031414944822, "learning_rate": 7.841504632582378e-06, "loss": 0.6046, "step": 353 }, { "epoch": 1.1321961620469083, "grad_norm": 0.7551618483349399, "learning_rate": 7.826134724862687e-06, "loss": 0.6229, "step": 354 }, { "epoch": 1.1353944562899787, "grad_norm": 0.5826096940443887, "learning_rate": 7.810725474102504e-06, "loss": 0.6491, "step": 355 }, { "epoch": 1.138592750533049, "grad_norm": 0.6417225168262856, "learning_rate": 7.795277094816292e-06, "loss": 0.6261, "step": 356 }, { "epoch": 1.1417910447761195, "grad_norm": 0.6906660028796798, "learning_rate": 7.779789802063229e-06, "loss": 0.6716, "step": 357 }, { "epoch": 1.1449893390191899, "grad_norm": 0.6599388631575054, "learning_rate": 7.764263811444214e-06, "loss": 0.6223, "step": 358 }, { "epoch": 1.14818763326226, "grad_norm": 0.6878046102310988, "learning_rate": 7.748699339098864e-06, "loss": 0.6535, "step": 359 }, { "epoch": 1.1513859275053304, "grad_norm": 0.6983975410121682, "learning_rate": 7.733096601702508e-06, "loss": 0.6175, "step": 360 }, { "epoch": 1.1545842217484008, "grad_norm": 0.7101392030170859, "learning_rate": 7.717455816463161e-06, "loss": 0.6064, "step": 361 }, { "epoch": 1.1577825159914712, "grad_norm": 0.7138468401070005, "learning_rate": 7.70177720111852e-06, "loss": 0.6318, "step": 362 }, { "epoch": 1.1609808102345416, "grad_norm": 0.8020962425971306, "learning_rate": 7.68606097393291e-06, "loss": 0.5551, "step": 363 }, { "epoch": 1.164179104477612, "grad_norm": 0.5838755573325491, "learning_rate": 7.67030735369426e-06, "loss": 0.6681, "step": 364 }, { "epoch": 1.1673773987206824, "grad_norm": 0.7126521508489613, "learning_rate": 7.654516559711053e-06, "loss": 0.6146, "step": 365 }, { "epoch": 1.1705756929637527, "grad_norm": 0.786856054891589, "learning_rate": 7.638688811809274e-06, "loss": 0.592, "step": 366 }, { "epoch": 1.1737739872068231, "grad_norm": 0.6150312500511127, "learning_rate": 7.622824330329345e-06, "loss": 0.6454, "step": 367 }, { "epoch": 1.1769722814498933, "grad_norm": 0.6533779011208625, "learning_rate": 7.6069233361230696e-06, "loss": 0.5972, "step": 368 }, { "epoch": 1.1801705756929637, "grad_norm": 0.7040587378616768, "learning_rate": 7.590986050550542e-06, "loss": 0.6506, "step": 369 }, { "epoch": 1.183368869936034, "grad_norm": 0.6498966757491511, "learning_rate": 7.575012695477076e-06, "loss": 0.6686, "step": 370 }, { "epoch": 1.1865671641791045, "grad_norm": 0.686323024280706, "learning_rate": 7.55900349327012e-06, "loss": 0.6156, "step": 371 }, { "epoch": 1.1897654584221748, "grad_norm": 0.6423904326022212, "learning_rate": 7.542958666796149e-06, "loss": 0.6074, "step": 372 }, { "epoch": 1.1929637526652452, "grad_norm": 0.6178701938146715, "learning_rate": 7.526878439417572e-06, "loss": 0.6224, "step": 373 }, { "epoch": 1.1961620469083156, "grad_norm": 0.6496299199143347, "learning_rate": 7.510763034989616e-06, "loss": 0.6122, "step": 374 }, { "epoch": 1.199360341151386, "grad_norm": 0.6906674571020524, "learning_rate": 7.494612677857218e-06, "loss": 0.5982, "step": 375 }, { "epoch": 1.2025586353944564, "grad_norm": 0.7648845383563382, "learning_rate": 7.478427592851894e-06, "loss": 0.6388, "step": 376 }, { "epoch": 1.2057569296375266, "grad_norm": 0.696212830007392, "learning_rate": 7.462208005288609e-06, "loss": 0.6195, "step": 377 }, { "epoch": 1.208955223880597, "grad_norm": 0.8576480726323676, "learning_rate": 7.44595414096265e-06, "loss": 0.62, "step": 378 }, { "epoch": 1.2121535181236673, "grad_norm": 0.6976210286882664, "learning_rate": 7.429666226146468e-06, "loss": 0.6416, "step": 379 }, { "epoch": 1.2153518123667377, "grad_norm": 0.720435239108808, "learning_rate": 7.413344487586542e-06, "loss": 0.663, "step": 380 }, { "epoch": 1.2185501066098081, "grad_norm": 0.6855025363671802, "learning_rate": 7.396989152500215e-06, "loss": 0.6244, "step": 381 }, { "epoch": 1.2217484008528785, "grad_norm": 0.7874191554629412, "learning_rate": 7.380600448572532e-06, "loss": 0.669, "step": 382 }, { "epoch": 1.224946695095949, "grad_norm": 0.7434853121491252, "learning_rate": 7.364178603953066e-06, "loss": 0.6448, "step": 383 }, { "epoch": 1.2281449893390193, "grad_norm": 0.61487183758893, "learning_rate": 7.347723847252756e-06, "loss": 0.6406, "step": 384 }, { "epoch": 1.2313432835820897, "grad_norm": 0.6917610834136999, "learning_rate": 7.331236407540704e-06, "loss": 0.6311, "step": 385 }, { "epoch": 1.2345415778251598, "grad_norm": 0.7293794438462118, "learning_rate": 7.314716514341007e-06, "loss": 0.6496, "step": 386 }, { "epoch": 1.2377398720682302, "grad_norm": 0.6943306944663377, "learning_rate": 7.298164397629545e-06, "loss": 0.695, "step": 387 }, { "epoch": 1.2409381663113006, "grad_norm": 0.727606577778182, "learning_rate": 7.28158028783079e-06, "loss": 0.6472, "step": 388 }, { "epoch": 1.244136460554371, "grad_norm": 0.6612780887089154, "learning_rate": 7.2649644158145925e-06, "loss": 0.6581, "step": 389 }, { "epoch": 1.2473347547974414, "grad_norm": 0.6796926109344666, "learning_rate": 7.248317012892969e-06, "loss": 0.6455, "step": 390 }, { "epoch": 1.2505330490405118, "grad_norm": 0.7347393619920833, "learning_rate": 7.231638310816888e-06, "loss": 0.6135, "step": 391 }, { "epoch": 1.2537313432835822, "grad_norm": 0.6507765447015836, "learning_rate": 7.214928541773027e-06, "loss": 0.6528, "step": 392 }, { "epoch": 1.2569296375266523, "grad_norm": 0.6378971746133411, "learning_rate": 7.198187938380565e-06, "loss": 0.64, "step": 393 }, { "epoch": 1.260127931769723, "grad_norm": 0.6973995690775606, "learning_rate": 7.1814167336879195e-06, "loss": 0.6372, "step": 394 }, { "epoch": 1.263326226012793, "grad_norm": 0.6611835574299032, "learning_rate": 7.164615161169518e-06, "loss": 0.641, "step": 395 }, { "epoch": 1.2665245202558635, "grad_norm": 0.5638998333261266, "learning_rate": 7.147783454722545e-06, "loss": 0.6191, "step": 396 }, { "epoch": 1.2697228144989339, "grad_norm": 0.5840735647196549, "learning_rate": 7.130921848663678e-06, "loss": 0.6461, "step": 397 }, { "epoch": 1.2729211087420043, "grad_norm": 0.6067150183156588, "learning_rate": 7.1140305777258355e-06, "loss": 0.6534, "step": 398 }, { "epoch": 1.2761194029850746, "grad_norm": 0.5533289520902701, "learning_rate": 7.097109877054906e-06, "loss": 0.6251, "step": 399 }, { "epoch": 1.279317697228145, "grad_norm": 0.6286453359485137, "learning_rate": 7.080159982206471e-06, "loss": 0.6192, "step": 400 }, { "epoch": 1.2825159914712154, "grad_norm": 0.5518718217422934, "learning_rate": 7.06318112914253e-06, "loss": 0.6014, "step": 401 }, { "epoch": 1.2857142857142856, "grad_norm": 0.5849867238933993, "learning_rate": 7.046173554228213e-06, "loss": 0.6131, "step": 402 }, { "epoch": 1.2889125799573562, "grad_norm": 0.6401106522610778, "learning_rate": 7.029137494228491e-06, "loss": 0.6638, "step": 403 }, { "epoch": 1.2921108742004264, "grad_norm": 0.571567700082923, "learning_rate": 7.012073186304885e-06, "loss": 0.6347, "step": 404 }, { "epoch": 1.2953091684434968, "grad_norm": 0.5618388361367395, "learning_rate": 6.994980868012151e-06, "loss": 0.6374, "step": 405 }, { "epoch": 1.2985074626865671, "grad_norm": 0.5861018343702368, "learning_rate": 6.9778607772949894e-06, "loss": 0.6427, "step": 406 }, { "epoch": 1.3017057569296375, "grad_norm": 0.5732316822425729, "learning_rate": 6.9607131524847175e-06, "loss": 0.6153, "step": 407 }, { "epoch": 1.304904051172708, "grad_norm": 0.6579936163217828, "learning_rate": 6.943538232295965e-06, "loss": 0.6822, "step": 408 }, { "epoch": 1.3081023454157783, "grad_norm": 0.6429931495365935, "learning_rate": 6.926336255823341e-06, "loss": 0.6703, "step": 409 }, { "epoch": 1.3113006396588487, "grad_norm": 0.6970899348524371, "learning_rate": 6.909107462538113e-06, "loss": 0.6118, "step": 410 }, { "epoch": 1.3144989339019189, "grad_norm": 0.6680718435346977, "learning_rate": 6.891852092284863e-06, "loss": 0.6485, "step": 411 }, { "epoch": 1.3176972281449895, "grad_norm": 0.6600380168580939, "learning_rate": 6.874570385278161e-06, "loss": 0.6717, "step": 412 }, { "epoch": 1.3208955223880596, "grad_norm": 0.802744507326931, "learning_rate": 6.857262582099209e-06, "loss": 0.6225, "step": 413 }, { "epoch": 1.32409381663113, "grad_norm": 0.5662862687056852, "learning_rate": 6.839928923692505e-06, "loss": 0.6327, "step": 414 }, { "epoch": 1.3272921108742004, "grad_norm": 0.7774960950807415, "learning_rate": 6.822569651362475e-06, "loss": 0.629, "step": 415 }, { "epoch": 1.3304904051172708, "grad_norm": 0.7049572238409981, "learning_rate": 6.805185006770125e-06, "loss": 0.6693, "step": 416 }, { "epoch": 1.3336886993603412, "grad_norm": 0.583155396015153, "learning_rate": 6.787775231929666e-06, "loss": 0.6237, "step": 417 }, { "epoch": 1.3368869936034116, "grad_norm": 0.7111519870233683, "learning_rate": 6.7703405692051585e-06, "loss": 0.6394, "step": 418 }, { "epoch": 1.340085287846482, "grad_norm": 0.5653594640288497, "learning_rate": 6.752881261307125e-06, "loss": 0.6155, "step": 419 }, { "epoch": 1.3432835820895521, "grad_norm": 0.5953314550960257, "learning_rate": 6.735397551289179e-06, "loss": 0.6768, "step": 420 }, { "epoch": 1.3464818763326227, "grad_norm": 0.6173871359335188, "learning_rate": 6.717889682544641e-06, "loss": 0.6524, "step": 421 }, { "epoch": 1.349680170575693, "grad_norm": 0.5904374139677531, "learning_rate": 6.700357898803146e-06, "loss": 0.6171, "step": 422 }, { "epoch": 1.3528784648187633, "grad_norm": 0.6013270780619727, "learning_rate": 6.6828024441272554e-06, "loss": 0.6331, "step": 423 }, { "epoch": 1.3560767590618337, "grad_norm": 0.7234400477846492, "learning_rate": 6.665223562909058e-06, "loss": 0.6048, "step": 424 }, { "epoch": 1.359275053304904, "grad_norm": 0.5513590207931245, "learning_rate": 6.647621499866762e-06, "loss": 0.6315, "step": 425 }, { "epoch": 1.3624733475479744, "grad_norm": 0.6666248996294633, "learning_rate": 6.629996500041299e-06, "loss": 0.6611, "step": 426 }, { "epoch": 1.3656716417910448, "grad_norm": 0.7473392889269375, "learning_rate": 6.612348808792904e-06, "loss": 0.6509, "step": 427 }, { "epoch": 1.3688699360341152, "grad_norm": 0.6178536139476686, "learning_rate": 6.5946786717977026e-06, "loss": 0.5846, "step": 428 }, { "epoch": 1.3720682302771854, "grad_norm": 0.7286227764621949, "learning_rate": 6.576986335044292e-06, "loss": 0.6225, "step": 429 }, { "epoch": 1.375266524520256, "grad_norm": 0.6891105977017067, "learning_rate": 6.5592720448303174e-06, "loss": 0.6247, "step": 430 }, { "epoch": 1.3784648187633262, "grad_norm": 0.548934245482671, "learning_rate": 6.541536047759034e-06, "loss": 0.6446, "step": 431 }, { "epoch": 1.3816631130063965, "grad_norm": 0.6602568691180308, "learning_rate": 6.523778590735892e-06, "loss": 0.6262, "step": 432 }, { "epoch": 1.384861407249467, "grad_norm": 0.6490608935621192, "learning_rate": 6.5059999209650795e-06, "loss": 0.6272, "step": 433 }, { "epoch": 1.3880597014925373, "grad_norm": 0.6052733094339158, "learning_rate": 6.488200285946094e-06, "loss": 0.6625, "step": 434 }, { "epoch": 1.3912579957356077, "grad_norm": 0.7850931175623906, "learning_rate": 6.470379933470296e-06, "loss": 0.6289, "step": 435 }, { "epoch": 1.394456289978678, "grad_norm": 0.6381779692800353, "learning_rate": 6.452539111617454e-06, "loss": 0.6403, "step": 436 }, { "epoch": 1.3976545842217485, "grad_norm": 0.6118616893209208, "learning_rate": 6.434678068752293e-06, "loss": 0.6407, "step": 437 }, { "epoch": 1.4008528784648187, "grad_norm": 0.6834821963612697, "learning_rate": 6.416797053521039e-06, "loss": 0.6274, "step": 438 }, { "epoch": 1.4040511727078893, "grad_norm": 0.6399343872362031, "learning_rate": 6.398896314847954e-06, "loss": 0.5884, "step": 439 }, { "epoch": 1.4072494669509594, "grad_norm": 0.650012682249965, "learning_rate": 6.380976101931879e-06, "loss": 0.5957, "step": 440 }, { "epoch": 1.4104477611940298, "grad_norm": 0.7005173466154051, "learning_rate": 6.363036664242751e-06, "loss": 0.6632, "step": 441 }, { "epoch": 1.4136460554371002, "grad_norm": 0.6793406210914823, "learning_rate": 6.345078251518144e-06, "loss": 0.6597, "step": 442 }, { "epoch": 1.4168443496801706, "grad_norm": 0.6296952630265025, "learning_rate": 6.327101113759783e-06, "loss": 0.6719, "step": 443 }, { "epoch": 1.420042643923241, "grad_norm": 0.6793002098632206, "learning_rate": 6.3091055012300675e-06, "loss": 0.6683, "step": 444 }, { "epoch": 1.4232409381663114, "grad_norm": 0.632818749841171, "learning_rate": 6.291091664448589e-06, "loss": 0.636, "step": 445 }, { "epoch": 1.4264392324093818, "grad_norm": 0.6812898757622275, "learning_rate": 6.273059854188636e-06, "loss": 0.6437, "step": 446 }, { "epoch": 1.429637526652452, "grad_norm": 0.6587981529111426, "learning_rate": 6.25501032147372e-06, "loss": 0.683, "step": 447 }, { "epoch": 1.4328358208955223, "grad_norm": 0.6112506718049283, "learning_rate": 6.236943317574054e-06, "loss": 0.6291, "step": 448 }, { "epoch": 1.4360341151385927, "grad_norm": 0.6816015509553757, "learning_rate": 6.218859094003082e-06, "loss": 0.6389, "step": 449 }, { "epoch": 1.439232409381663, "grad_norm": 0.5678562801569867, "learning_rate": 6.200757902513962e-06, "loss": 0.655, "step": 450 }, { "epoch": 1.4424307036247335, "grad_norm": 0.5811458853598366, "learning_rate": 6.182639995096061e-06, "loss": 0.6762, "step": 451 }, { "epoch": 1.4456289978678039, "grad_norm": 0.7324330200735513, "learning_rate": 6.164505623971458e-06, "loss": 0.6298, "step": 452 }, { "epoch": 1.4488272921108742, "grad_norm": 0.5725759197205573, "learning_rate": 6.146355041591419e-06, "loss": 0.6573, "step": 453 }, { "epoch": 1.4520255863539446, "grad_norm": 0.6093969811734463, "learning_rate": 6.128188500632892e-06, "loss": 0.6226, "step": 454 }, { "epoch": 1.455223880597015, "grad_norm": 0.5642498697251256, "learning_rate": 6.11000625399499e-06, "loss": 0.6047, "step": 455 }, { "epoch": 1.4584221748400852, "grad_norm": 0.5740303890176889, "learning_rate": 6.091808554795462e-06, "loss": 0.6318, "step": 456 }, { "epoch": 1.4616204690831556, "grad_norm": 0.6288307609988142, "learning_rate": 6.073595656367175e-06, "loss": 0.647, "step": 457 }, { "epoch": 1.464818763326226, "grad_norm": 0.5866496634459744, "learning_rate": 6.055367812254592e-06, "loss": 0.6098, "step": 458 }, { "epoch": 1.4680170575692963, "grad_norm": 0.5931796642802533, "learning_rate": 6.037125276210229e-06, "loss": 0.6682, "step": 459 }, { "epoch": 1.4712153518123667, "grad_norm": 0.5952005451682014, "learning_rate": 6.0188683021911394e-06, "loss": 0.6415, "step": 460 }, { "epoch": 1.4744136460554371, "grad_norm": 0.5818037194226144, "learning_rate": 6.000597144355361e-06, "loss": 0.6301, "step": 461 }, { "epoch": 1.4776119402985075, "grad_norm": 0.5536985639025898, "learning_rate": 5.982312057058392e-06, "loss": 0.6189, "step": 462 }, { "epoch": 1.480810234541578, "grad_norm": 0.5574476870925947, "learning_rate": 5.964013294849646e-06, "loss": 0.6513, "step": 463 }, { "epoch": 1.4840085287846483, "grad_norm": 0.6007921638296236, "learning_rate": 5.9457011124689025e-06, "loss": 0.6562, "step": 464 }, { "epoch": 1.4872068230277184, "grad_norm": 0.6484774122884718, "learning_rate": 5.927375764842766e-06, "loss": 0.6415, "step": 465 }, { "epoch": 1.4904051172707888, "grad_norm": 0.5367744850334513, "learning_rate": 5.9090375070811215e-06, "loss": 0.6324, "step": 466 }, { "epoch": 1.4936034115138592, "grad_norm": 0.6330458661332963, "learning_rate": 5.890686594473571e-06, "loss": 0.6773, "step": 467 }, { "epoch": 1.4968017057569296, "grad_norm": 0.6239236549424172, "learning_rate": 5.872323282485889e-06, "loss": 0.62, "step": 468 }, { "epoch": 1.5, "grad_norm": 0.7714169311784487, "learning_rate": 5.853947826756465e-06, "loss": 0.635, "step": 469 }, { "epoch": 1.5031982942430704, "grad_norm": 0.6052673944971001, "learning_rate": 5.835560483092743e-06, "loss": 0.6212, "step": 470 }, { "epoch": 1.5063965884861408, "grad_norm": 0.6500878797817035, "learning_rate": 5.8171615074676615e-06, "loss": 0.6235, "step": 471 }, { "epoch": 1.509594882729211, "grad_norm": 0.5806431215575556, "learning_rate": 5.798751156016085e-06, "loss": 0.618, "step": 472 }, { "epoch": 1.5127931769722816, "grad_norm": 0.6174497537734422, "learning_rate": 5.780329685031247e-06, "loss": 0.6218, "step": 473 }, { "epoch": 1.5159914712153517, "grad_norm": 0.6376322903314644, "learning_rate": 5.7618973509611755e-06, "loss": 0.6133, "step": 474 }, { "epoch": 1.5191897654584223, "grad_norm": 0.5842117132744907, "learning_rate": 5.743454410405126e-06, "loss": 0.6135, "step": 475 }, { "epoch": 1.5223880597014925, "grad_norm": 0.7067488456038941, "learning_rate": 5.72500112011001e-06, "loss": 0.6172, "step": 476 }, { "epoch": 1.5255863539445629, "grad_norm": 0.563785993728458, "learning_rate": 5.706537736966814e-06, "loss": 0.598, "step": 477 }, { "epoch": 1.5287846481876333, "grad_norm": 0.6261356329525944, "learning_rate": 5.688064518007036e-06, "loss": 0.6362, "step": 478 }, { "epoch": 1.5319829424307037, "grad_norm": 0.7762542900583126, "learning_rate": 5.669581720399094e-06, "loss": 0.6108, "step": 479 }, { "epoch": 1.535181236673774, "grad_norm": 0.5837610551252048, "learning_rate": 5.651089601444752e-06, "loss": 0.6516, "step": 480 }, { "epoch": 1.5383795309168442, "grad_norm": 0.5716180642723976, "learning_rate": 5.632588418575542e-06, "loss": 0.6345, "step": 481 }, { "epoch": 1.5415778251599148, "grad_norm": 0.5946826841288883, "learning_rate": 5.614078429349172e-06, "loss": 0.5934, "step": 482 }, { "epoch": 1.544776119402985, "grad_norm": 0.6512015058357723, "learning_rate": 5.5955598914459465e-06, "loss": 0.6457, "step": 483 }, { "epoch": 1.5479744136460556, "grad_norm": 0.5539302739317081, "learning_rate": 5.577033062665179e-06, "loss": 0.6234, "step": 484 }, { "epoch": 1.5511727078891258, "grad_norm": 0.5788744632809979, "learning_rate": 5.558498200921597e-06, "loss": 0.5965, "step": 485 }, { "epoch": 1.5543710021321961, "grad_norm": 0.5723934422902696, "learning_rate": 5.53995556424176e-06, "loss": 0.6148, "step": 486 }, { "epoch": 1.5575692963752665, "grad_norm": 0.6217220728990693, "learning_rate": 5.521405410760462e-06, "loss": 0.6306, "step": 487 }, { "epoch": 1.560767590618337, "grad_norm": 0.5798449514904306, "learning_rate": 5.50284799871714e-06, "loss": 0.6523, "step": 488 }, { "epoch": 1.5639658848614073, "grad_norm": 0.5884980910713539, "learning_rate": 5.484283586452279e-06, "loss": 0.6409, "step": 489 }, { "epoch": 1.5671641791044775, "grad_norm": 0.5451220864593269, "learning_rate": 5.465712432403812e-06, "loss": 0.64, "step": 490 }, { "epoch": 1.570362473347548, "grad_norm": 0.6378194747314097, "learning_rate": 5.447134795103531e-06, "loss": 0.6369, "step": 491 }, { "epoch": 1.5735607675906182, "grad_norm": 0.6422794734982141, "learning_rate": 5.428550933173476e-06, "loss": 0.6315, "step": 492 }, { "epoch": 1.5767590618336889, "grad_norm": 0.6165632358558669, "learning_rate": 5.409961105322347e-06, "loss": 0.606, "step": 493 }, { "epoch": 1.579957356076759, "grad_norm": 0.5655676781911688, "learning_rate": 5.391365570341893e-06, "loss": 0.6454, "step": 494 }, { "epoch": 1.5831556503198294, "grad_norm": 0.5938708749925583, "learning_rate": 5.372764587103309e-06, "loss": 0.5788, "step": 495 }, { "epoch": 1.5863539445628998, "grad_norm": 0.6088992599619968, "learning_rate": 5.3541584145536475e-06, "loss": 0.6072, "step": 496 }, { "epoch": 1.5895522388059702, "grad_norm": 0.5784074732476567, "learning_rate": 5.335547311712188e-06, "loss": 0.6217, "step": 497 }, { "epoch": 1.5927505330490406, "grad_norm": 0.6355703007033081, "learning_rate": 5.3169315376668566e-06, "loss": 0.6003, "step": 498 }, { "epoch": 1.5959488272921107, "grad_norm": 0.5969121792808912, "learning_rate": 5.2983113515706045e-06, "loss": 0.6141, "step": 499 }, { "epoch": 1.5991471215351813, "grad_norm": 0.5840423887981219, "learning_rate": 5.279687012637798e-06, "loss": 0.6454, "step": 500 }, { "epoch": 1.6023454157782515, "grad_norm": 0.632640436994868, "learning_rate": 5.2610587801406256e-06, "loss": 0.619, "step": 501 }, { "epoch": 1.6055437100213221, "grad_norm": 0.6215694598661943, "learning_rate": 5.242426913405471e-06, "loss": 0.612, "step": 502 }, { "epoch": 1.6087420042643923, "grad_norm": 0.6102315942715076, "learning_rate": 5.223791671809314e-06, "loss": 0.6246, "step": 503 }, { "epoch": 1.6119402985074627, "grad_norm": 0.6343359621017866, "learning_rate": 5.2051533147761155e-06, "loss": 0.6361, "step": 504 }, { "epoch": 1.615138592750533, "grad_norm": 0.6647629539626027, "learning_rate": 5.186512101773206e-06, "loss": 0.6227, "step": 505 }, { "epoch": 1.6183368869936035, "grad_norm": 0.6372800082536703, "learning_rate": 5.167868292307679e-06, "loss": 0.6478, "step": 506 }, { "epoch": 1.6215351812366738, "grad_norm": 0.5768049939693273, "learning_rate": 5.149222145922765e-06, "loss": 0.6459, "step": 507 }, { "epoch": 1.624733475479744, "grad_norm": 0.625505130556027, "learning_rate": 5.130573922194236e-06, "loss": 0.6587, "step": 508 }, { "epoch": 1.6279317697228146, "grad_norm": 0.6164713574466578, "learning_rate": 5.111923880726779e-06, "loss": 0.631, "step": 509 }, { "epoch": 1.6311300639658848, "grad_norm": 0.6064673645339017, "learning_rate": 5.093272281150383e-06, "loss": 0.6425, "step": 510 }, { "epoch": 1.6343283582089554, "grad_norm": 0.6081052704243078, "learning_rate": 5.074619383116733e-06, "loss": 0.5795, "step": 511 }, { "epoch": 1.6375266524520256, "grad_norm": 0.6412746235926382, "learning_rate": 5.05596544629559e-06, "loss": 0.6341, "step": 512 }, { "epoch": 1.640724946695096, "grad_norm": 0.6297169827461854, "learning_rate": 5.03731073037117e-06, "loss": 0.6413, "step": 513 }, { "epoch": 1.6439232409381663, "grad_norm": 0.5638619759271226, "learning_rate": 5.018655495038542e-06, "loss": 0.642, "step": 514 }, { "epoch": 1.6471215351812367, "grad_norm": 0.556408547120617, "learning_rate": 5e-06, "loss": 0.6713, "step": 515 }, { "epoch": 1.650319829424307, "grad_norm": 0.6515376681950392, "learning_rate": 4.981344504961459e-06, "loss": 0.6701, "step": 516 }, { "epoch": 1.6535181236673773, "grad_norm": 0.5492041340566489, "learning_rate": 4.962689269628832e-06, "loss": 0.6173, "step": 517 }, { "epoch": 1.6567164179104479, "grad_norm": 0.5686773096619224, "learning_rate": 4.944034553704412e-06, "loss": 0.6461, "step": 518 }, { "epoch": 1.659914712153518, "grad_norm": 0.6435263865701101, "learning_rate": 4.9253806168832685e-06, "loss": 0.6123, "step": 519 }, { "epoch": 1.6631130063965884, "grad_norm": 0.5691664452687136, "learning_rate": 4.906727718849619e-06, "loss": 0.6238, "step": 520 }, { "epoch": 1.6663113006396588, "grad_norm": 0.567539010985681, "learning_rate": 4.888076119273223e-06, "loss": 0.6184, "step": 521 }, { "epoch": 1.6695095948827292, "grad_norm": 0.6144632127897085, "learning_rate": 4.8694260778057655e-06, "loss": 0.5979, "step": 522 }, { "epoch": 1.6727078891257996, "grad_norm": 0.5911520451028203, "learning_rate": 4.850777854077235e-06, "loss": 0.6469, "step": 523 }, { "epoch": 1.67590618336887, "grad_norm": 0.5454031941868988, "learning_rate": 4.832131707692322e-06, "loss": 0.6122, "step": 524 }, { "epoch": 1.6791044776119404, "grad_norm": 0.5981940676142077, "learning_rate": 4.813487898226794e-06, "loss": 0.6427, "step": 525 }, { "epoch": 1.6823027718550105, "grad_norm": 0.5547199384817155, "learning_rate": 4.7948466852238844e-06, "loss": 0.6278, "step": 526 }, { "epoch": 1.6855010660980811, "grad_norm": 0.5468385314356093, "learning_rate": 4.7762083281906864e-06, "loss": 0.594, "step": 527 }, { "epoch": 1.6886993603411513, "grad_norm": 0.5456445809907, "learning_rate": 4.757573086594529e-06, "loss": 0.5884, "step": 528 }, { "epoch": 1.6918976545842217, "grad_norm": 0.6646249117314513, "learning_rate": 4.738941219859375e-06, "loss": 0.6306, "step": 529 }, { "epoch": 1.695095948827292, "grad_norm": 0.6292949980039241, "learning_rate": 4.720312987362204e-06, "loss": 0.6619, "step": 530 }, { "epoch": 1.6982942430703625, "grad_norm": 0.6322989923958425, "learning_rate": 4.701688648429399e-06, "loss": 0.6266, "step": 531 }, { "epoch": 1.7014925373134329, "grad_norm": 0.5305082924051039, "learning_rate": 4.683068462333144e-06, "loss": 0.6342, "step": 532 }, { "epoch": 1.7046908315565032, "grad_norm": 0.550929491236182, "learning_rate": 4.6644526882878145e-06, "loss": 0.6392, "step": 533 }, { "epoch": 1.7078891257995736, "grad_norm": 0.6232015245825243, "learning_rate": 4.645841585446356e-06, "loss": 0.6064, "step": 534 }, { "epoch": 1.7110874200426438, "grad_norm": 0.5434413619942932, "learning_rate": 4.6272354128966924e-06, "loss": 0.6501, "step": 535 }, { "epoch": 1.7142857142857144, "grad_norm": 0.5191107189759604, "learning_rate": 4.6086344296581095e-06, "loss": 0.6448, "step": 536 }, { "epoch": 1.7174840085287846, "grad_norm": 0.5434292262769143, "learning_rate": 4.590038894677653e-06, "loss": 0.6074, "step": 537 }, { "epoch": 1.720682302771855, "grad_norm": 0.4655865353721599, "learning_rate": 4.5714490668265245e-06, "loss": 0.6236, "step": 538 }, { "epoch": 1.7238805970149254, "grad_norm": 0.550629570360553, "learning_rate": 4.55286520489647e-06, "loss": 0.6397, "step": 539 }, { "epoch": 1.7270788912579957, "grad_norm": 0.6004187629555796, "learning_rate": 4.534287567596189e-06, "loss": 0.6385, "step": 540 }, { "epoch": 1.7302771855010661, "grad_norm": 0.5176720613714986, "learning_rate": 4.515716413547722e-06, "loss": 0.6429, "step": 541 }, { "epoch": 1.7334754797441365, "grad_norm": 0.5115763781808625, "learning_rate": 4.497152001282861e-06, "loss": 0.6382, "step": 542 }, { "epoch": 1.736673773987207, "grad_norm": 0.5562269449292555, "learning_rate": 4.478594589239539e-06, "loss": 0.6236, "step": 543 }, { "epoch": 1.739872068230277, "grad_norm": 0.5670914639063015, "learning_rate": 4.460044435758241e-06, "loss": 0.6441, "step": 544 }, { "epoch": 1.7430703624733477, "grad_norm": 0.5489747231617312, "learning_rate": 4.441501799078405e-06, "loss": 0.613, "step": 545 }, { "epoch": 1.7462686567164178, "grad_norm": 0.5424103892867449, "learning_rate": 4.4229669373348225e-06, "loss": 0.6253, "step": 546 }, { "epoch": 1.7494669509594882, "grad_norm": 0.5562973456307682, "learning_rate": 4.404440108554055e-06, "loss": 0.6168, "step": 547 }, { "epoch": 1.7526652452025586, "grad_norm": 0.533299067462362, "learning_rate": 4.3859215706508295e-06, "loss": 0.6485, "step": 548 }, { "epoch": 1.755863539445629, "grad_norm": 0.5172786400673339, "learning_rate": 4.3674115814244595e-06, "loss": 0.6323, "step": 549 }, { "epoch": 1.7590618336886994, "grad_norm": 0.5085177296011396, "learning_rate": 4.348910398555249e-06, "loss": 0.5899, "step": 550 }, { "epoch": 1.7622601279317696, "grad_norm": 0.47810924329720716, "learning_rate": 4.330418279600907e-06, "loss": 0.6379, "step": 551 }, { "epoch": 1.7654584221748402, "grad_norm": 0.5321362592336698, "learning_rate": 4.311935481992965e-06, "loss": 0.6151, "step": 552 }, { "epoch": 1.7686567164179103, "grad_norm": 0.5446514716404816, "learning_rate": 4.2934622630331855e-06, "loss": 0.6079, "step": 553 }, { "epoch": 1.771855010660981, "grad_norm": 0.5237047324414964, "learning_rate": 4.274998879889991e-06, "loss": 0.6427, "step": 554 }, { "epoch": 1.775053304904051, "grad_norm": 0.5365142222456774, "learning_rate": 4.2565455895948745e-06, "loss": 0.582, "step": 555 }, { "epoch": 1.7782515991471215, "grad_norm": 0.5308651285729982, "learning_rate": 4.238102649038825e-06, "loss": 0.6575, "step": 556 }, { "epoch": 1.7814498933901919, "grad_norm": 0.5717647379657843, "learning_rate": 4.219670314968754e-06, "loss": 0.62, "step": 557 }, { "epoch": 1.7846481876332623, "grad_norm": 0.5050163481740146, "learning_rate": 4.2012488439839185e-06, "loss": 0.6636, "step": 558 }, { "epoch": 1.7878464818763327, "grad_norm": 0.49227167259915716, "learning_rate": 4.182838492532342e-06, "loss": 0.5945, "step": 559 }, { "epoch": 1.7910447761194028, "grad_norm": 0.5302015513388415, "learning_rate": 4.164439516907258e-06, "loss": 0.6106, "step": 560 }, { "epoch": 1.7942430703624734, "grad_norm": 0.64011360289916, "learning_rate": 4.146052173243538e-06, "loss": 0.6593, "step": 561 }, { "epoch": 1.7974413646055436, "grad_norm": 0.601687625944857, "learning_rate": 4.127676717514114e-06, "loss": 0.6092, "step": 562 }, { "epoch": 1.8006396588486142, "grad_norm": 0.524585479629224, "learning_rate": 4.109313405526433e-06, "loss": 0.6446, "step": 563 }, { "epoch": 1.8038379530916844, "grad_norm": 0.6161148649503772, "learning_rate": 4.090962492918881e-06, "loss": 0.6233, "step": 564 }, { "epoch": 1.8070362473347548, "grad_norm": 0.6062559672756939, "learning_rate": 4.072624235157234e-06, "loss": 0.6623, "step": 565 }, { "epoch": 1.8102345415778252, "grad_norm": 0.5985336173857585, "learning_rate": 4.054298887531099e-06, "loss": 0.582, "step": 566 }, { "epoch": 1.8134328358208955, "grad_norm": 0.5632646612772408, "learning_rate": 4.035986705150355e-06, "loss": 0.6159, "step": 567 }, { "epoch": 1.816631130063966, "grad_norm": 0.7068557752731836, "learning_rate": 4.017687942941609e-06, "loss": 0.6536, "step": 568 }, { "epoch": 1.819829424307036, "grad_norm": 0.5730952893121638, "learning_rate": 3.9994028556446404e-06, "loss": 0.6241, "step": 569 }, { "epoch": 1.8230277185501067, "grad_norm": 0.5175887707114529, "learning_rate": 3.981131697808862e-06, "loss": 0.6627, "step": 570 }, { "epoch": 1.8262260127931769, "grad_norm": 0.5796987142687964, "learning_rate": 3.9628747237897715e-06, "loss": 0.6536, "step": 571 }, { "epoch": 1.8294243070362475, "grad_norm": 0.5271014946968099, "learning_rate": 3.94463218774541e-06, "loss": 0.6467, "step": 572 }, { "epoch": 1.8326226012793176, "grad_norm": 0.5704417597925869, "learning_rate": 3.926404343632826e-06, "loss": 0.6831, "step": 573 }, { "epoch": 1.835820895522388, "grad_norm": 0.5292627931377012, "learning_rate": 3.90819144520454e-06, "loss": 0.6408, "step": 574 }, { "epoch": 1.8390191897654584, "grad_norm": 0.5646881580541779, "learning_rate": 3.889993746005011e-06, "loss": 0.6348, "step": 575 }, { "epoch": 1.8422174840085288, "grad_norm": 0.5401410573662442, "learning_rate": 3.8718114993671086e-06, "loss": 0.5917, "step": 576 }, { "epoch": 1.8454157782515992, "grad_norm": 0.5033737411697443, "learning_rate": 3.853644958408582e-06, "loss": 0.6454, "step": 577 }, { "epoch": 1.8486140724946694, "grad_norm": 0.5317406038132683, "learning_rate": 3.835494376028544e-06, "loss": 0.579, "step": 578 }, { "epoch": 1.85181236673774, "grad_norm": 0.5247944872571422, "learning_rate": 3.817360004903939e-06, "loss": 0.6443, "step": 579 }, { "epoch": 1.8550106609808101, "grad_norm": 0.5255266393479685, "learning_rate": 3.799242097486038e-06, "loss": 0.6548, "step": 580 }, { "epoch": 1.8582089552238807, "grad_norm": 0.5562031814775014, "learning_rate": 3.7811409059969177e-06, "loss": 0.6368, "step": 581 }, { "epoch": 1.861407249466951, "grad_norm": 0.5490844678533507, "learning_rate": 3.7630566824259456e-06, "loss": 0.6001, "step": 582 }, { "epoch": 1.8646055437100213, "grad_norm": 0.5257106510728633, "learning_rate": 3.7449896785262817e-06, "loss": 0.6092, "step": 583 }, { "epoch": 1.8678038379530917, "grad_norm": 0.5410953577384838, "learning_rate": 3.726940145811363e-06, "loss": 0.643, "step": 584 }, { "epoch": 1.871002132196162, "grad_norm": 0.5205935779573823, "learning_rate": 3.708908335551412e-06, "loss": 0.6609, "step": 585 }, { "epoch": 1.8742004264392325, "grad_norm": 0.5715057818226926, "learning_rate": 3.6908944987699346e-06, "loss": 0.6191, "step": 586 }, { "epoch": 1.8773987206823026, "grad_norm": 0.5410333169459866, "learning_rate": 3.67289888624022e-06, "loss": 0.6169, "step": 587 }, { "epoch": 1.8805970149253732, "grad_norm": 0.5321189795175738, "learning_rate": 3.6549217484818576e-06, "loss": 0.5957, "step": 588 }, { "epoch": 1.8837953091684434, "grad_norm": 0.6038607670908029, "learning_rate": 3.6369633357572514e-06, "loss": 0.6112, "step": 589 }, { "epoch": 1.886993603411514, "grad_norm": 0.5382547649391735, "learning_rate": 3.6190238980681235e-06, "loss": 0.6446, "step": 590 }, { "epoch": 1.8901918976545842, "grad_norm": 0.509428705908829, "learning_rate": 3.6011036851520465e-06, "loss": 0.6245, "step": 591 }, { "epoch": 1.8933901918976546, "grad_norm": 0.5331030357626905, "learning_rate": 3.583202946478963e-06, "loss": 0.6667, "step": 592 }, { "epoch": 1.896588486140725, "grad_norm": 0.5957179693199653, "learning_rate": 3.5653219312477085e-06, "loss": 0.6164, "step": 593 }, { "epoch": 1.8997867803837953, "grad_norm": 0.540978235449404, "learning_rate": 3.5474608883825475e-06, "loss": 0.6814, "step": 594 }, { "epoch": 1.9029850746268657, "grad_norm": 0.5609593877190009, "learning_rate": 3.529620066529704e-06, "loss": 0.6337, "step": 595 }, { "epoch": 1.906183368869936, "grad_norm": 0.5759309632996505, "learning_rate": 3.5117997140539073e-06, "loss": 0.637, "step": 596 }, { "epoch": 1.9093816631130065, "grad_norm": 0.5913831985711084, "learning_rate": 3.4940000790349226e-06, "loss": 0.6288, "step": 597 }, { "epoch": 1.9125799573560767, "grad_norm": 0.5075699761691839, "learning_rate": 3.47622140926411e-06, "loss": 0.6173, "step": 598 }, { "epoch": 1.9157782515991473, "grad_norm": 0.6653450535384141, "learning_rate": 3.458463952240967e-06, "loss": 0.6523, "step": 599 }, { "epoch": 1.9189765458422174, "grad_norm": 0.628552652910662, "learning_rate": 3.4407279551696846e-06, "loss": 0.6358, "step": 600 }, { "epoch": 1.9221748400852878, "grad_norm": 0.5333242599508652, "learning_rate": 3.4230136649557087e-06, "loss": 0.6288, "step": 601 }, { "epoch": 1.9253731343283582, "grad_norm": 0.47146898112937263, "learning_rate": 3.4053213282022983e-06, "loss": 0.6596, "step": 602 }, { "epoch": 1.9285714285714286, "grad_norm": 0.6621044867075212, "learning_rate": 3.387651191207097e-06, "loss": 0.5934, "step": 603 }, { "epoch": 1.931769722814499, "grad_norm": 0.5923983943553847, "learning_rate": 3.370003499958703e-06, "loss": 0.6332, "step": 604 }, { "epoch": 1.9349680170575692, "grad_norm": 0.5262532242135438, "learning_rate": 3.352378500133239e-06, "loss": 0.6164, "step": 605 }, { "epoch": 1.9381663113006398, "grad_norm": 0.6174684357894091, "learning_rate": 3.334776437090944e-06, "loss": 0.6276, "step": 606 }, { "epoch": 1.94136460554371, "grad_norm": 0.6066479008889557, "learning_rate": 3.317197555872745e-06, "loss": 0.6061, "step": 607 }, { "epoch": 1.9445628997867805, "grad_norm": 0.5282593114168354, "learning_rate": 3.2996421011968546e-06, "loss": 0.5957, "step": 608 }, { "epoch": 1.9477611940298507, "grad_norm": 0.5428512832531854, "learning_rate": 3.28211031745536e-06, "loss": 0.6274, "step": 609 }, { "epoch": 1.950959488272921, "grad_norm": 0.563963200457514, "learning_rate": 3.264602448710822e-06, "loss": 0.5977, "step": 610 }, { "epoch": 1.9541577825159915, "grad_norm": 0.506718113639374, "learning_rate": 3.2471187386928766e-06, "loss": 0.5975, "step": 611 }, { "epoch": 1.9573560767590619, "grad_norm": 0.5077307814187035, "learning_rate": 3.2296594307948428e-06, "loss": 0.6401, "step": 612 }, { "epoch": 1.9605543710021323, "grad_norm": 0.5128850695519592, "learning_rate": 3.212224768070334e-06, "loss": 0.6381, "step": 613 }, { "epoch": 1.9637526652452024, "grad_norm": 0.5083765825134322, "learning_rate": 3.194814993229878e-06, "loss": 0.62, "step": 614 }, { "epoch": 1.966950959488273, "grad_norm": 0.47099407882432265, "learning_rate": 3.177430348637527e-06, "loss": 0.6125, "step": 615 }, { "epoch": 1.9701492537313432, "grad_norm": 0.5530258695290763, "learning_rate": 3.1600710763074972e-06, "loss": 0.6356, "step": 616 }, { "epoch": 1.9733475479744138, "grad_norm": 0.5347670899530949, "learning_rate": 3.142737417900793e-06, "loss": 0.6611, "step": 617 }, { "epoch": 1.976545842217484, "grad_norm": 0.5188746097889174, "learning_rate": 3.125429614721842e-06, "loss": 0.6185, "step": 618 }, { "epoch": 1.9797441364605544, "grad_norm": 0.47423520536928904, "learning_rate": 3.1081479077151387e-06, "loss": 0.6207, "step": 619 }, { "epoch": 1.9829424307036247, "grad_norm": 0.5191715465473807, "learning_rate": 3.090892537461889e-06, "loss": 0.604, "step": 620 }, { "epoch": 1.9861407249466951, "grad_norm": 0.6101222943327476, "learning_rate": 3.0736637441766594e-06, "loss": 0.6234, "step": 621 }, { "epoch": 1.9893390191897655, "grad_norm": 0.5351791226481845, "learning_rate": 3.056461767704037e-06, "loss": 0.6411, "step": 622 }, { "epoch": 1.9925373134328357, "grad_norm": 0.5193133031919753, "learning_rate": 3.039286847515284e-06, "loss": 0.6175, "step": 623 }, { "epoch": 1.9957356076759063, "grad_norm": 0.5316524549516811, "learning_rate": 3.0221392227050126e-06, "loss": 0.6408, "step": 624 }, { "epoch": 1.9989339019189765, "grad_norm": 0.5093874593326244, "learning_rate": 3.00501913198785e-06, "loss": 0.6261, "step": 625 }, { "epoch": 2.002132196162047, "grad_norm": 0.6163744912473614, "learning_rate": 2.9879268136951163e-06, "loss": 0.5951, "step": 626 }, { "epoch": 2.0053304904051172, "grad_norm": 0.5783808952696761, "learning_rate": 2.970862505771509e-06, "loss": 0.5622, "step": 627 }, { "epoch": 2.008528784648188, "grad_norm": 0.49273447076344123, "learning_rate": 2.953826445771788e-06, "loss": 0.5673, "step": 628 }, { "epoch": 2.011727078891258, "grad_norm": 0.5859845266864249, "learning_rate": 2.9368188708574706e-06, "loss": 0.6246, "step": 629 }, { "epoch": 2.014925373134328, "grad_norm": 0.5611571675265489, "learning_rate": 2.9198400177935303e-06, "loss": 0.6277, "step": 630 }, { "epoch": 2.018123667377399, "grad_norm": 0.5276964961204649, "learning_rate": 2.902890122945096e-06, "loss": 0.5577, "step": 631 }, { "epoch": 2.021321961620469, "grad_norm": 0.525544540323068, "learning_rate": 2.8859694222741653e-06, "loss": 0.6051, "step": 632 }, { "epoch": 2.0245202558635396, "grad_norm": 0.5673683781918863, "learning_rate": 2.869078151336323e-06, "loss": 0.5808, "step": 633 }, { "epoch": 2.0277185501066097, "grad_norm": 0.5589784105831517, "learning_rate": 2.852216545277456e-06, "loss": 0.6106, "step": 634 }, { "epoch": 2.0309168443496803, "grad_norm": 0.5030496312589189, "learning_rate": 2.835384838830481e-06, "loss": 0.5541, "step": 635 }, { "epoch": 2.0341151385927505, "grad_norm": 0.6708437250927518, "learning_rate": 2.8185832663120817e-06, "loss": 0.6271, "step": 636 }, { "epoch": 2.0373134328358207, "grad_norm": 0.5763870321226472, "learning_rate": 2.8018120616194356e-06, "loss": 0.5507, "step": 637 }, { "epoch": 2.0405117270788913, "grad_norm": 0.5752129249821096, "learning_rate": 2.785071458226972e-06, "loss": 0.5978, "step": 638 }, { "epoch": 2.0437100213219614, "grad_norm": 0.5054941801603793, "learning_rate": 2.768361689183113e-06, "loss": 0.6203, "step": 639 }, { "epoch": 2.046908315565032, "grad_norm": 0.5481782008730065, "learning_rate": 2.7516829871070295e-06, "loss": 0.6093, "step": 640 }, { "epoch": 2.050106609808102, "grad_norm": 0.5323921761861825, "learning_rate": 2.735035584185409e-06, "loss": 0.6007, "step": 641 }, { "epoch": 2.053304904051173, "grad_norm": 0.5293554299631367, "learning_rate": 2.718419712169213e-06, "loss": 0.5884, "step": 642 }, { "epoch": 2.056503198294243, "grad_norm": 0.553435984362884, "learning_rate": 2.7018356023704574e-06, "loss": 0.5326, "step": 643 }, { "epoch": 2.0597014925373136, "grad_norm": 0.5139320287446054, "learning_rate": 2.685283485658995e-06, "loss": 0.5639, "step": 644 }, { "epoch": 2.0628997867803838, "grad_norm": 0.49577644020424133, "learning_rate": 2.668763592459297e-06, "loss": 0.5571, "step": 645 }, { "epoch": 2.066098081023454, "grad_norm": 0.5089676714103052, "learning_rate": 2.6522761527472464e-06, "loss": 0.6163, "step": 646 }, { "epoch": 2.0692963752665245, "grad_norm": 0.5593194742250533, "learning_rate": 2.6358213960469357e-06, "loss": 0.5808, "step": 647 }, { "epoch": 2.0724946695095947, "grad_norm": 0.5130651147305723, "learning_rate": 2.6193995514274705e-06, "loss": 0.5953, "step": 648 }, { "epoch": 2.0756929637526653, "grad_norm": 0.5609831284876498, "learning_rate": 2.6030108474997854e-06, "loss": 0.5841, "step": 649 }, { "epoch": 2.0788912579957355, "grad_norm": 0.47312462257776017, "learning_rate": 2.586655512413458e-06, "loss": 0.5967, "step": 650 }, { "epoch": 2.082089552238806, "grad_norm": 0.5001915039170618, "learning_rate": 2.5703337738535324e-06, "loss": 0.5485, "step": 651 }, { "epoch": 2.0852878464818763, "grad_norm": 0.5785964658330373, "learning_rate": 2.554045859037353e-06, "loss": 0.566, "step": 652 }, { "epoch": 2.088486140724947, "grad_norm": 0.5912116290504553, "learning_rate": 2.5377919947113917e-06, "loss": 0.5856, "step": 653 }, { "epoch": 2.091684434968017, "grad_norm": 0.5209304706365177, "learning_rate": 2.521572407148107e-06, "loss": 0.5556, "step": 654 }, { "epoch": 2.094882729211087, "grad_norm": 0.5101895622941909, "learning_rate": 2.505387322142782e-06, "loss": 0.5776, "step": 655 }, { "epoch": 2.098081023454158, "grad_norm": 0.4880680240587064, "learning_rate": 2.4892369650103837e-06, "loss": 0.6056, "step": 656 }, { "epoch": 2.101279317697228, "grad_norm": 0.5514381593453722, "learning_rate": 2.4731215605824304e-06, "loss": 0.6006, "step": 657 }, { "epoch": 2.1044776119402986, "grad_norm": 0.5751639987148548, "learning_rate": 2.4570413332038523e-06, "loss": 0.5665, "step": 658 }, { "epoch": 2.1076759061833688, "grad_norm": 0.5804924762978739, "learning_rate": 2.440996506729881e-06, "loss": 0.5973, "step": 659 }, { "epoch": 2.1108742004264394, "grad_norm": 0.4918030532858259, "learning_rate": 2.4249873045229244e-06, "loss": 0.57, "step": 660 }, { "epoch": 2.1140724946695095, "grad_norm": 0.5243890793969488, "learning_rate": 2.4090139494494596e-06, "loss": 0.5871, "step": 661 }, { "epoch": 2.11727078891258, "grad_norm": 0.4806747676363039, "learning_rate": 2.3930766638769325e-06, "loss": 0.5598, "step": 662 }, { "epoch": 2.1204690831556503, "grad_norm": 0.5175543377119891, "learning_rate": 2.3771756696706553e-06, "loss": 0.6051, "step": 663 }, { "epoch": 2.1236673773987205, "grad_norm": 0.5246309451300202, "learning_rate": 2.3613111881907273e-06, "loss": 0.598, "step": 664 }, { "epoch": 2.126865671641791, "grad_norm": 0.526456441833946, "learning_rate": 2.345483440288947e-06, "loss": 0.547, "step": 665 }, { "epoch": 2.1300639658848612, "grad_norm": 0.47184799877761335, "learning_rate": 2.3296926463057396e-06, "loss": 0.6138, "step": 666 }, { "epoch": 2.133262260127932, "grad_norm": 0.48133254553344923, "learning_rate": 2.313939026067091e-06, "loss": 0.583, "step": 667 }, { "epoch": 2.136460554371002, "grad_norm": 0.5064732476569918, "learning_rate": 2.29822279888148e-06, "loss": 0.5454, "step": 668 }, { "epoch": 2.1396588486140726, "grad_norm": 0.5781888574823206, "learning_rate": 2.2825441835368377e-06, "loss": 0.5712, "step": 669 }, { "epoch": 2.142857142857143, "grad_norm": 0.5978766513888759, "learning_rate": 2.2669033982974946e-06, "loss": 0.5695, "step": 670 }, { "epoch": 2.1460554371002134, "grad_norm": 0.4555096407611019, "learning_rate": 2.2513006609011365e-06, "loss": 0.5555, "step": 671 }, { "epoch": 2.1492537313432836, "grad_norm": 0.4676695942649127, "learning_rate": 2.235736188555787e-06, "loss": 0.5735, "step": 672 }, { "epoch": 2.1524520255863537, "grad_norm": 0.4763343046497065, "learning_rate": 2.2202101979367735e-06, "loss": 0.5955, "step": 673 }, { "epoch": 2.1556503198294243, "grad_norm": 0.4657805538089298, "learning_rate": 2.2047229051837107e-06, "loss": 0.6076, "step": 674 }, { "epoch": 2.1588486140724945, "grad_norm": 0.4838763721788115, "learning_rate": 2.189274525897498e-06, "loss": 0.616, "step": 675 }, { "epoch": 2.162046908315565, "grad_norm": 0.5013682671987199, "learning_rate": 2.173865275137314e-06, "loss": 0.5809, "step": 676 }, { "epoch": 2.1652452025586353, "grad_norm": 0.4730326559560223, "learning_rate": 2.158495367417625e-06, "loss": 0.5758, "step": 677 }, { "epoch": 2.168443496801706, "grad_norm": 0.49708415063251904, "learning_rate": 2.143165016705192e-06, "loss": 0.5859, "step": 678 }, { "epoch": 2.171641791044776, "grad_norm": 0.4952662090916961, "learning_rate": 2.1278744364161007e-06, "loss": 0.5575, "step": 679 }, { "epoch": 2.1748400852878467, "grad_norm": 0.46833442631122313, "learning_rate": 2.1126238394127868e-06, "loss": 0.5974, "step": 680 }, { "epoch": 2.178038379530917, "grad_norm": 0.4714670060045907, "learning_rate": 2.0974134380010726e-06, "loss": 0.5751, "step": 681 }, { "epoch": 2.181236673773987, "grad_norm": 0.4772805496941911, "learning_rate": 2.082243443927212e-06, "loss": 0.5996, "step": 682 }, { "epoch": 2.1844349680170576, "grad_norm": 0.4606017759107387, "learning_rate": 2.0671140683749386e-06, "loss": 0.5817, "step": 683 }, { "epoch": 2.1876332622601278, "grad_norm": 0.4774965126191024, "learning_rate": 2.052025521962534e-06, "loss": 0.5961, "step": 684 }, { "epoch": 2.1908315565031984, "grad_norm": 0.49760796326925766, "learning_rate": 2.03697801473989e-06, "loss": 0.5788, "step": 685 }, { "epoch": 2.1940298507462686, "grad_norm": 0.4871186232939493, "learning_rate": 2.0219717561855857e-06, "loss": 0.5879, "step": 686 }, { "epoch": 2.197228144989339, "grad_norm": 0.464109823915951, "learning_rate": 2.0070069552039722e-06, "loss": 0.6154, "step": 687 }, { "epoch": 2.2004264392324093, "grad_norm": 0.4856649377645587, "learning_rate": 1.992083820122259e-06, "loss": 0.5788, "step": 688 }, { "epoch": 2.20362473347548, "grad_norm": 0.46996237636688565, "learning_rate": 1.9772025586876252e-06, "loss": 0.5953, "step": 689 }, { "epoch": 2.20682302771855, "grad_norm": 0.46859069921217344, "learning_rate": 1.962363378064316e-06, "loss": 0.6083, "step": 690 }, { "epoch": 2.2100213219616203, "grad_norm": 0.5168988401256138, "learning_rate": 1.947566484830765e-06, "loss": 0.5814, "step": 691 }, { "epoch": 2.213219616204691, "grad_norm": 0.5406960885215742, "learning_rate": 1.9328120849767198e-06, "loss": 0.5561, "step": 692 }, { "epoch": 2.216417910447761, "grad_norm": 0.44334448566836304, "learning_rate": 1.9181003839003627e-06, "loss": 0.5927, "step": 693 }, { "epoch": 2.2196162046908317, "grad_norm": 0.4557583490316895, "learning_rate": 1.9034315864054682e-06, "loss": 0.5862, "step": 694 }, { "epoch": 2.222814498933902, "grad_norm": 0.5092922861926251, "learning_rate": 1.8888058966985407e-06, "loss": 0.5668, "step": 695 }, { "epoch": 2.2260127931769724, "grad_norm": 0.5004818455277625, "learning_rate": 1.8742235183859747e-06, "loss": 0.5968, "step": 696 }, { "epoch": 2.2292110874200426, "grad_norm": 0.4949014258712852, "learning_rate": 1.8596846544712233e-06, "loss": 0.622, "step": 697 }, { "epoch": 2.232409381663113, "grad_norm": 0.47178385456983946, "learning_rate": 1.8451895073519643e-06, "loss": 0.5725, "step": 698 }, { "epoch": 2.2356076759061834, "grad_norm": 0.4709195571648972, "learning_rate": 1.8307382788172877e-06, "loss": 0.5991, "step": 699 }, { "epoch": 2.2388059701492535, "grad_norm": 0.4806780947771992, "learning_rate": 1.8163311700448899e-06, "loss": 0.5976, "step": 700 }, { "epoch": 2.242004264392324, "grad_norm": 0.46974071200274314, "learning_rate": 1.8019683815982691e-06, "loss": 0.6091, "step": 701 }, { "epoch": 2.2452025586353943, "grad_norm": 0.45361098939294514, "learning_rate": 1.7876501134239316e-06, "loss": 0.5933, "step": 702 }, { "epoch": 2.248400852878465, "grad_norm": 0.4850717880695125, "learning_rate": 1.7733765648486134e-06, "loss": 0.5597, "step": 703 }, { "epoch": 2.251599147121535, "grad_norm": 0.4692790998035614, "learning_rate": 1.7591479345764972e-06, "loss": 0.6007, "step": 704 }, { "epoch": 2.2547974413646057, "grad_norm": 0.4906873053697701, "learning_rate": 1.7449644206864564e-06, "loss": 0.552, "step": 705 }, { "epoch": 2.257995735607676, "grad_norm": 0.481979324803139, "learning_rate": 1.7308262206292898e-06, "loss": 0.6125, "step": 706 }, { "epoch": 2.2611940298507465, "grad_norm": 0.5016963903077071, "learning_rate": 1.7167335312249766e-06, "loss": 0.6017, "step": 707 }, { "epoch": 2.2643923240938166, "grad_norm": 0.47149916005013454, "learning_rate": 1.7026865486599375e-06, "loss": 0.5869, "step": 708 }, { "epoch": 2.267590618336887, "grad_norm": 0.507514553246801, "learning_rate": 1.6886854684842962e-06, "loss": 0.5926, "step": 709 }, { "epoch": 2.2707889125799574, "grad_norm": 0.49251780946608525, "learning_rate": 1.6747304856091662e-06, "loss": 0.5874, "step": 710 }, { "epoch": 2.2739872068230276, "grad_norm": 0.47492882189499497, "learning_rate": 1.660821794303934e-06, "loss": 0.6, "step": 711 }, { "epoch": 2.277185501066098, "grad_norm": 0.4403913104531341, "learning_rate": 1.6469595881935523e-06, "loss": 0.5673, "step": 712 }, { "epoch": 2.2803837953091683, "grad_norm": 0.5039434479895408, "learning_rate": 1.6331440602558501e-06, "loss": 0.5908, "step": 713 }, { "epoch": 2.283582089552239, "grad_norm": 0.4893793347672981, "learning_rate": 1.6193754028188363e-06, "loss": 0.5757, "step": 714 }, { "epoch": 2.286780383795309, "grad_norm": 0.4833544188283998, "learning_rate": 1.6056538075580342e-06, "loss": 0.6034, "step": 715 }, { "epoch": 2.2899786780383797, "grad_norm": 0.46660784623966395, "learning_rate": 1.591979465493806e-06, "loss": 0.554, "step": 716 }, { "epoch": 2.29317697228145, "grad_norm": 0.4941030437078769, "learning_rate": 1.5783525669886934e-06, "loss": 0.5615, "step": 717 }, { "epoch": 2.29637526652452, "grad_norm": 0.467606067398601, "learning_rate": 1.5647733017447741e-06, "loss": 0.608, "step": 718 }, { "epoch": 2.2995735607675907, "grad_norm": 0.47853133402269155, "learning_rate": 1.5512418588010086e-06, "loss": 0.5894, "step": 719 }, { "epoch": 2.302771855010661, "grad_norm": 0.4972054054110685, "learning_rate": 1.5377584265306222e-06, "loss": 0.5282, "step": 720 }, { "epoch": 2.3059701492537314, "grad_norm": 0.4727141579634826, "learning_rate": 1.5243231926384744e-06, "loss": 0.6045, "step": 721 }, { "epoch": 2.3091684434968016, "grad_norm": 0.46328977747110367, "learning_rate": 1.510936344158448e-06, "loss": 0.6181, "step": 722 }, { "epoch": 2.3123667377398722, "grad_norm": 0.46878962646890604, "learning_rate": 1.4975980674508472e-06, "loss": 0.5932, "step": 723 }, { "epoch": 2.3155650319829424, "grad_norm": 0.46930491695796345, "learning_rate": 1.484308548199796e-06, "loss": 0.5943, "step": 724 }, { "epoch": 2.318763326226013, "grad_norm": 0.4819028591025632, "learning_rate": 1.4710679714106635e-06, "loss": 0.6015, "step": 725 }, { "epoch": 2.321961620469083, "grad_norm": 0.4490778160016328, "learning_rate": 1.4578765214074842e-06, "loss": 0.5885, "step": 726 }, { "epoch": 2.3251599147121533, "grad_norm": 0.439589868568974, "learning_rate": 1.444734381830386e-06, "loss": 0.5752, "step": 727 }, { "epoch": 2.328358208955224, "grad_norm": 0.4595533358265758, "learning_rate": 1.4316417356330441e-06, "loss": 0.5933, "step": 728 }, { "epoch": 2.331556503198294, "grad_norm": 0.46808692540880764, "learning_rate": 1.4185987650801286e-06, "loss": 0.5593, "step": 729 }, { "epoch": 2.3347547974413647, "grad_norm": 0.4422620648261011, "learning_rate": 1.4056056517447637e-06, "loss": 0.5725, "step": 730 }, { "epoch": 2.337953091684435, "grad_norm": 0.47954403115117483, "learning_rate": 1.392662576506007e-06, "loss": 0.5761, "step": 731 }, { "epoch": 2.3411513859275055, "grad_norm": 0.4546489872200875, "learning_rate": 1.3797697195463278e-06, "loss": 0.5871, "step": 732 }, { "epoch": 2.3443496801705757, "grad_norm": 0.4500895428397602, "learning_rate": 1.3669272603491002e-06, "loss": 0.5644, "step": 733 }, { "epoch": 2.3475479744136463, "grad_norm": 0.4805002284443759, "learning_rate": 1.3541353776961035e-06, "loss": 0.5714, "step": 734 }, { "epoch": 2.3507462686567164, "grad_norm": 0.450604856228356, "learning_rate": 1.3413942496650301e-06, "loss": 0.5569, "step": 735 }, { "epoch": 2.3539445628997866, "grad_norm": 0.44449130255853053, "learning_rate": 1.3287040536270135e-06, "loss": 0.5673, "step": 736 }, { "epoch": 2.357142857142857, "grad_norm": 0.4311651330481667, "learning_rate": 1.3160649662441532e-06, "loss": 0.5649, "step": 737 }, { "epoch": 2.3603411513859274, "grad_norm": 0.5481566773492595, "learning_rate": 1.30347716346706e-06, "loss": 0.5805, "step": 738 }, { "epoch": 2.363539445628998, "grad_norm": 0.4694122862360678, "learning_rate": 1.290940820532403e-06, "loss": 0.5862, "step": 739 }, { "epoch": 2.366737739872068, "grad_norm": 0.4808027009258574, "learning_rate": 1.2784561119604683e-06, "loss": 0.5844, "step": 740 }, { "epoch": 2.3699360341151388, "grad_norm": 0.4741021908385686, "learning_rate": 1.266023211552736e-06, "loss": 0.5927, "step": 741 }, { "epoch": 2.373134328358209, "grad_norm": 0.5070129440283249, "learning_rate": 1.2536422923894565e-06, "loss": 0.5989, "step": 742 }, { "epoch": 2.3763326226012795, "grad_norm": 0.49982577127543343, "learning_rate": 1.2413135268272403e-06, "loss": 0.6086, "step": 743 }, { "epoch": 2.3795309168443497, "grad_norm": 0.5171582506065449, "learning_rate": 1.2290370864966623e-06, "loss": 0.5935, "step": 744 }, { "epoch": 2.38272921108742, "grad_norm": 0.5058849568215932, "learning_rate": 1.2168131422998653e-06, "loss": 0.5788, "step": 745 }, { "epoch": 2.3859275053304905, "grad_norm": 0.49576030363447515, "learning_rate": 1.2046418644081904e-06, "loss": 0.6084, "step": 746 }, { "epoch": 2.3891257995735606, "grad_norm": 0.4737787057177936, "learning_rate": 1.192523422259802e-06, "loss": 0.5667, "step": 747 }, { "epoch": 2.3923240938166312, "grad_norm": 0.4628087834904147, "learning_rate": 1.1804579845573288e-06, "loss": 0.6369, "step": 748 }, { "epoch": 2.3955223880597014, "grad_norm": 0.44818958645228585, "learning_rate": 1.1684457192655207e-06, "loss": 0.5785, "step": 749 }, { "epoch": 2.398720682302772, "grad_norm": 0.44830220311328195, "learning_rate": 1.156486793608899e-06, "loss": 0.5932, "step": 750 }, { "epoch": 2.401918976545842, "grad_norm": 0.5032388277593316, "learning_rate": 1.144581374069444e-06, "loss": 0.6266, "step": 751 }, { "epoch": 2.405117270788913, "grad_norm": 0.5016006802472114, "learning_rate": 1.1327296263842653e-06, "loss": 0.5956, "step": 752 }, { "epoch": 2.408315565031983, "grad_norm": 0.4643916700878555, "learning_rate": 1.120931715543299e-06, "loss": 0.5694, "step": 753 }, { "epoch": 2.411513859275053, "grad_norm": 0.44920509828052607, "learning_rate": 1.1091878057870137e-06, "loss": 0.6153, "step": 754 }, { "epoch": 2.4147121535181237, "grad_norm": 0.45686584478530823, "learning_rate": 1.0974980606041152e-06, "loss": 0.5805, "step": 755 }, { "epoch": 2.417910447761194, "grad_norm": 0.46553788908049154, "learning_rate": 1.0858626427292796e-06, "loss": 0.599, "step": 756 }, { "epoch": 2.4211087420042645, "grad_norm": 0.5153605958479428, "learning_rate": 1.074281714140884e-06, "loss": 0.6158, "step": 757 }, { "epoch": 2.4243070362473347, "grad_norm": 0.5316883972134206, "learning_rate": 1.0627554360587533e-06, "loss": 0.6284, "step": 758 }, { "epoch": 2.4275053304904053, "grad_norm": 0.4612691902886173, "learning_rate": 1.0512839689419124e-06, "loss": 0.629, "step": 759 }, { "epoch": 2.4307036247334755, "grad_norm": 0.45681038140195945, "learning_rate": 1.0398674724863584e-06, "loss": 0.5693, "step": 760 }, { "epoch": 2.433901918976546, "grad_norm": 0.48533257603226454, "learning_rate": 1.0285061056228273e-06, "loss": 0.5905, "step": 761 }, { "epoch": 2.4371002132196162, "grad_norm": 0.4746224830922052, "learning_rate": 1.0172000265145938e-06, "loss": 0.6055, "step": 762 }, { "epoch": 2.4402985074626864, "grad_norm": 0.5086026458920127, "learning_rate": 1.0059493925552604e-06, "loss": 0.5744, "step": 763 }, { "epoch": 2.443496801705757, "grad_norm": 0.4514225973264305, "learning_rate": 9.947543603665711e-07, "loss": 0.5964, "step": 764 }, { "epoch": 2.446695095948827, "grad_norm": 0.475954682216827, "learning_rate": 9.836150857962296e-07, "loss": 0.6195, "step": 765 }, { "epoch": 2.449893390191898, "grad_norm": 0.516019394227879, "learning_rate": 9.72531723915726e-07, "loss": 0.5606, "step": 766 }, { "epoch": 2.453091684434968, "grad_norm": 0.4863472899308536, "learning_rate": 9.615044290181863e-07, "loss": 0.5711, "step": 767 }, { "epoch": 2.4562899786780386, "grad_norm": 0.433161350310435, "learning_rate": 9.505333546162171e-07, "loss": 0.61, "step": 768 }, { "epoch": 2.4594882729211087, "grad_norm": 0.4744290062548913, "learning_rate": 9.396186534397711e-07, "loss": 0.5657, "step": 769 }, { "epoch": 2.4626865671641793, "grad_norm": 0.4438296651802267, "learning_rate": 9.287604774340236e-07, "loss": 0.6051, "step": 770 }, { "epoch": 2.4658848614072495, "grad_norm": 0.4527860416616376, "learning_rate": 9.179589777572496e-07, "loss": 0.5531, "step": 771 }, { "epoch": 2.4690831556503197, "grad_norm": 0.4863675854140087, "learning_rate": 9.07214304778729e-07, "loss": 0.5948, "step": 772 }, { "epoch": 2.4722814498933903, "grad_norm": 0.45267687004556684, "learning_rate": 8.965266080766471e-07, "loss": 0.5954, "step": 773 }, { "epoch": 2.4754797441364604, "grad_norm": 0.4577039509757653, "learning_rate": 8.858960364360142e-07, "loss": 0.5749, "step": 774 }, { "epoch": 2.478678038379531, "grad_norm": 0.42098442226808497, "learning_rate": 8.753227378465956e-07, "loss": 0.5503, "step": 775 }, { "epoch": 2.481876332622601, "grad_norm": 0.4593978666976355, "learning_rate": 8.648068595008458e-07, "loss": 0.6078, "step": 776 }, { "epoch": 2.485074626865672, "grad_norm": 0.4859190992732641, "learning_rate": 8.543485477918672e-07, "loss": 0.5579, "step": 777 }, { "epoch": 2.488272921108742, "grad_norm": 0.43344193599128084, "learning_rate": 8.439479483113683e-07, "loss": 0.578, "step": 778 }, { "epoch": 2.4914712153518126, "grad_norm": 0.46299605001742955, "learning_rate": 8.336052058476374e-07, "loss": 0.5936, "step": 779 }, { "epoch": 2.4946695095948828, "grad_norm": 0.48867807045497547, "learning_rate": 8.233204643835235e-07, "loss": 0.5904, "step": 780 }, { "epoch": 2.497867803837953, "grad_norm": 0.49771350857070334, "learning_rate": 8.130938670944377e-07, "loss": 0.6007, "step": 781 }, { "epoch": 2.5010660980810235, "grad_norm": 0.4577972103614504, "learning_rate": 8.029255563463589e-07, "loss": 0.5895, "step": 782 }, { "epoch": 2.5042643923240937, "grad_norm": 0.47989050108275627, "learning_rate": 7.928156736938458e-07, "loss": 0.5949, "step": 783 }, { "epoch": 2.5074626865671643, "grad_norm": 0.4371318457371128, "learning_rate": 7.827643598780748e-07, "loss": 0.5528, "step": 784 }, { "epoch": 2.5106609808102345, "grad_norm": 0.46786863175003857, "learning_rate": 7.72771754824877e-07, "loss": 0.6073, "step": 785 }, { "epoch": 2.5138592750533046, "grad_norm": 0.44519658517690325, "learning_rate": 7.628379976427868e-07, "loss": 0.6135, "step": 786 }, { "epoch": 2.5170575692963753, "grad_norm": 0.44205578681704094, "learning_rate": 7.529632266211112e-07, "loss": 0.5942, "step": 787 }, { "epoch": 2.520255863539446, "grad_norm": 0.4875271600407386, "learning_rate": 7.431475792280018e-07, "loss": 0.5807, "step": 788 }, { "epoch": 2.523454157782516, "grad_norm": 0.4710353166878217, "learning_rate": 7.333911921085418e-07, "loss": 0.6056, "step": 789 }, { "epoch": 2.526652452025586, "grad_norm": 0.48054394292285285, "learning_rate": 7.23694201082843e-07, "loss": 0.5591, "step": 790 }, { "epoch": 2.529850746268657, "grad_norm": 0.4868072363712072, "learning_rate": 7.140567411441529e-07, "loss": 0.5832, "step": 791 }, { "epoch": 2.533049040511727, "grad_norm": 0.4412560252403492, "learning_rate": 7.044789464569817e-07, "loss": 0.5937, "step": 792 }, { "epoch": 2.5362473347547976, "grad_norm": 0.4511415013667377, "learning_rate": 6.94960950355229e-07, "loss": 0.6048, "step": 793 }, { "epoch": 2.5394456289978677, "grad_norm": 0.4454888744137283, "learning_rate": 6.855028853403295e-07, "loss": 0.6137, "step": 794 }, { "epoch": 2.542643923240938, "grad_norm": 0.5107375323074698, "learning_rate": 6.761048830794098e-07, "loss": 0.6191, "step": 795 }, { "epoch": 2.5458422174840085, "grad_norm": 0.46547987418755626, "learning_rate": 6.667670744034498e-07, "loss": 0.5814, "step": 796 }, { "epoch": 2.549040511727079, "grad_norm": 0.4703517207855011, "learning_rate": 6.574895893054711e-07, "loss": 0.5778, "step": 797 }, { "epoch": 2.5522388059701493, "grad_norm": 0.4746163195915809, "learning_rate": 6.482725569387171e-07, "loss": 0.5706, "step": 798 }, { "epoch": 2.5554371002132195, "grad_norm": 0.470907057244819, "learning_rate": 6.391161056148637e-07, "loss": 0.5868, "step": 799 }, { "epoch": 2.55863539445629, "grad_norm": 0.4623234041016843, "learning_rate": 6.300203628022272e-07, "loss": 0.5983, "step": 800 }, { "epoch": 2.5618336886993602, "grad_norm": 0.42562483606080814, "learning_rate": 6.209854551239902e-07, "loss": 0.5927, "step": 801 }, { "epoch": 2.565031982942431, "grad_norm": 0.5128197818499242, "learning_rate": 6.120115083564432e-07, "loss": 0.5741, "step": 802 }, { "epoch": 2.568230277185501, "grad_norm": 0.47049712813520317, "learning_rate": 6.030986474272288e-07, "loss": 0.584, "step": 803 }, { "epoch": 2.571428571428571, "grad_norm": 0.44693735376914034, "learning_rate": 5.942469964136055e-07, "loss": 0.5903, "step": 804 }, { "epoch": 2.574626865671642, "grad_norm": 0.42861119739729087, "learning_rate": 5.854566785407212e-07, "loss": 0.597, "step": 805 }, { "epoch": 2.5778251599147124, "grad_norm": 0.44886124383973774, "learning_rate": 5.767278161798912e-07, "loss": 0.5721, "step": 806 }, { "epoch": 2.5810234541577826, "grad_norm": 0.4202452468317384, "learning_rate": 5.680605308469045e-07, "loss": 0.5828, "step": 807 }, { "epoch": 2.5842217484008527, "grad_norm": 0.4235020132956652, "learning_rate": 5.594549432003244e-07, "loss": 0.5846, "step": 808 }, { "epoch": 2.5874200426439233, "grad_norm": 0.4603963689104671, "learning_rate": 5.509111730398125e-07, "loss": 0.5753, "step": 809 }, { "epoch": 2.5906183368869935, "grad_norm": 0.47640520792175567, "learning_rate": 5.42429339304461e-07, "loss": 0.5813, "step": 810 }, { "epoch": 2.593816631130064, "grad_norm": 0.4964363064529672, "learning_rate": 5.340095600711343e-07, "loss": 0.5483, "step": 811 }, { "epoch": 2.5970149253731343, "grad_norm": 0.4188032394181886, "learning_rate": 5.256519525528254e-07, "loss": 0.6258, "step": 812 }, { "epoch": 2.6002132196162044, "grad_norm": 0.4418308069828843, "learning_rate": 5.173566330970286e-07, "loss": 0.5858, "step": 813 }, { "epoch": 2.603411513859275, "grad_norm": 0.4729321718897582, "learning_rate": 5.091237171841173e-07, "loss": 0.56, "step": 814 }, { "epoch": 2.6066098081023457, "grad_norm": 0.4393226839327768, "learning_rate": 5.009533194257332e-07, "loss": 0.5494, "step": 815 }, { "epoch": 2.609808102345416, "grad_norm": 0.44734542269470695, "learning_rate": 4.92845553563196e-07, "loss": 0.626, "step": 816 }, { "epoch": 2.613006396588486, "grad_norm": 0.4501393477711489, "learning_rate": 4.848005324659144e-07, "loss": 0.5966, "step": 817 }, { "epoch": 2.6162046908315566, "grad_norm": 0.4524819111331228, "learning_rate": 4.768183681298211e-07, "loss": 0.5816, "step": 818 }, { "epoch": 2.6194029850746268, "grad_norm": 0.44348410421643397, "learning_rate": 4.6889917167580903e-07, "loss": 0.5752, "step": 819 }, { "epoch": 2.6226012793176974, "grad_norm": 0.46496313577314546, "learning_rate": 4.6104305334818577e-07, "loss": 0.5894, "step": 820 }, { "epoch": 2.6257995735607675, "grad_norm": 0.45530128557907246, "learning_rate": 4.532501225131408e-07, "loss": 0.628, "step": 821 }, { "epoch": 2.6289978678038377, "grad_norm": 0.4557231787890459, "learning_rate": 4.455204876572172e-07, "loss": 0.5963, "step": 822 }, { "epoch": 2.6321961620469083, "grad_norm": 0.413730075405882, "learning_rate": 4.3785425638580847e-07, "loss": 0.5895, "step": 823 }, { "epoch": 2.635394456289979, "grad_norm": 0.46478458824267105, "learning_rate": 4.3025153542165744e-07, "loss": 0.5628, "step": 824 }, { "epoch": 2.638592750533049, "grad_norm": 0.4660432242750011, "learning_rate": 4.2271243060336976e-07, "loss": 0.5843, "step": 825 }, { "epoch": 2.6417910447761193, "grad_norm": 0.45857321057691036, "learning_rate": 4.1523704688394176e-07, "loss": 0.569, "step": 826 }, { "epoch": 2.64498933901919, "grad_norm": 0.454757225375927, "learning_rate": 4.0782548832929646e-07, "loss": 0.5631, "step": 827 }, { "epoch": 2.64818763326226, "grad_norm": 0.4219569542186656, "learning_rate": 4.0047785811684116e-07, "loss": 0.5923, "step": 828 }, { "epoch": 2.6513859275053306, "grad_norm": 0.407295453366224, "learning_rate": 3.931942585340243e-07, "loss": 0.6345, "step": 829 }, { "epoch": 2.654584221748401, "grad_norm": 0.4422205956943115, "learning_rate": 3.8597479097691626e-07, "loss": 0.6016, "step": 830 }, { "epoch": 2.657782515991471, "grad_norm": 0.4322372658598502, "learning_rate": 3.788195559487956e-07, "loss": 0.5717, "step": 831 }, { "epoch": 2.6609808102345416, "grad_norm": 0.446074402964079, "learning_rate": 3.717286530587483e-07, "loss": 0.5933, "step": 832 }, { "epoch": 2.664179104477612, "grad_norm": 0.44149384007245823, "learning_rate": 3.6470218102028607e-07, "loss": 0.5935, "step": 833 }, { "epoch": 2.6673773987206824, "grad_norm": 0.48606195110523104, "learning_rate": 3.577402376499672e-07, "loss": 0.5632, "step": 834 }, { "epoch": 2.6705756929637525, "grad_norm": 0.45172402428831737, "learning_rate": 3.508429198660379e-07, "loss": 0.6271, "step": 835 }, { "epoch": 2.673773987206823, "grad_norm": 0.4783802058190356, "learning_rate": 3.440103236870823e-07, "loss": 0.5614, "step": 836 }, { "epoch": 2.6769722814498933, "grad_norm": 0.4125951168982884, "learning_rate": 3.372425442306837e-07, "loss": 0.6403, "step": 837 }, { "epoch": 2.680170575692964, "grad_norm": 0.45804236331506576, "learning_rate": 3.3053967571210375e-07, "loss": 0.6235, "step": 838 }, { "epoch": 2.683368869936034, "grad_norm": 0.4481909177996369, "learning_rate": 3.2390181144296815e-07, "loss": 0.6079, "step": 839 }, { "epoch": 2.6865671641791042, "grad_norm": 0.44229637332098864, "learning_rate": 3.1732904382996975e-07, "loss": 0.6046, "step": 840 }, { "epoch": 2.689765458422175, "grad_norm": 0.456497771982713, "learning_rate": 3.108214643735813e-07, "loss": 0.6003, "step": 841 }, { "epoch": 2.6929637526652455, "grad_norm": 0.43916739323489756, "learning_rate": 3.04379163666782e-07, "loss": 0.596, "step": 842 }, { "epoch": 2.6961620469083156, "grad_norm": 0.4833880509304977, "learning_rate": 2.98002231393793e-07, "loss": 0.602, "step": 843 }, { "epoch": 2.699360341151386, "grad_norm": 0.44190751175823845, "learning_rate": 2.916907563288357e-07, "loss": 0.6089, "step": 844 }, { "epoch": 2.7025586353944564, "grad_norm": 0.4409073668802216, "learning_rate": 2.854448263348891e-07, "loss": 0.6085, "step": 845 }, { "epoch": 2.7057569296375266, "grad_norm": 0.44376266623437516, "learning_rate": 2.792645283624712e-07, "loss": 0.5517, "step": 846 }, { "epoch": 2.708955223880597, "grad_norm": 0.44030030505529505, "learning_rate": 2.7314994844842623e-07, "loss": 0.5704, "step": 847 }, { "epoch": 2.7121535181236673, "grad_norm": 0.4811909570823317, "learning_rate": 2.671011717147276e-07, "loss": 0.5887, "step": 848 }, { "epoch": 2.7153518123667375, "grad_norm": 0.46435424416208204, "learning_rate": 2.611182823672931e-07, "loss": 0.5753, "step": 849 }, { "epoch": 2.718550106609808, "grad_norm": 0.45062900256380783, "learning_rate": 2.5520136369481194e-07, "loss": 0.5956, "step": 850 }, { "epoch": 2.7217484008528787, "grad_norm": 0.4373848417355698, "learning_rate": 2.493504980675865e-07, "loss": 0.5725, "step": 851 }, { "epoch": 2.724946695095949, "grad_norm": 0.41456459520431993, "learning_rate": 2.4356576693638555e-07, "loss": 0.5989, "step": 852 }, { "epoch": 2.728144989339019, "grad_norm": 0.45641172841307615, "learning_rate": 2.3784725083130678e-07, "loss": 0.5394, "step": 853 }, { "epoch": 2.7313432835820897, "grad_norm": 0.41453418124512204, "learning_rate": 2.3219502936066228e-07, "loss": 0.5822, "step": 854 }, { "epoch": 2.73454157782516, "grad_norm": 0.47178126843622886, "learning_rate": 2.266091812098642e-07, "loss": 0.6153, "step": 855 }, { "epoch": 2.7377398720682304, "grad_norm": 0.415357561340448, "learning_rate": 2.210897841403331e-07, "loss": 0.6319, "step": 856 }, { "epoch": 2.7409381663113006, "grad_norm": 0.429713391850993, "learning_rate": 2.1563691498841465e-07, "loss": 0.6076, "step": 857 }, { "epoch": 2.7441364605543708, "grad_norm": 0.4364411910317568, "learning_rate": 2.1025064966430697e-07, "loss": 0.5809, "step": 858 }, { "epoch": 2.7473347547974414, "grad_norm": 0.4438187825256016, "learning_rate": 2.0493106315100987e-07, "loss": 0.6077, "step": 859 }, { "epoch": 2.750533049040512, "grad_norm": 0.4396721972544317, "learning_rate": 1.9967822950327453e-07, "loss": 0.5649, "step": 860 }, { "epoch": 2.753731343283582, "grad_norm": 0.4087723994643433, "learning_rate": 1.944922218465778e-07, "loss": 0.5983, "step": 861 }, { "epoch": 2.7569296375266523, "grad_norm": 0.45839789417365656, "learning_rate": 1.8937311237610168e-07, "loss": 0.5666, "step": 862 }, { "epoch": 2.760127931769723, "grad_norm": 0.4623951503889097, "learning_rate": 1.8432097235572655e-07, "loss": 0.5814, "step": 863 }, { "epoch": 2.763326226012793, "grad_norm": 0.42397071107319473, "learning_rate": 1.793358721170435e-07, "loss": 0.6227, "step": 864 }, { "epoch": 2.7665245202558637, "grad_norm": 0.46632194080900585, "learning_rate": 1.7441788105837133e-07, "loss": 0.5687, "step": 865 }, { "epoch": 2.769722814498934, "grad_norm": 0.42383172618691023, "learning_rate": 1.6956706764379438e-07, "loss": 0.5571, "step": 866 }, { "epoch": 2.772921108742004, "grad_norm": 0.42444537413289984, "learning_rate": 1.6478349940220294e-07, "loss": 0.5672, "step": 867 }, { "epoch": 2.7761194029850746, "grad_norm": 0.442713023638936, "learning_rate": 1.6006724292636166e-07, "loss": 0.6251, "step": 868 }, { "epoch": 2.7793176972281453, "grad_norm": 0.44882872466351875, "learning_rate": 1.5541836387197528e-07, "loss": 0.6309, "step": 869 }, { "epoch": 2.7825159914712154, "grad_norm": 0.4426652587385554, "learning_rate": 1.508369269567783e-07, "loss": 0.6038, "step": 870 }, { "epoch": 2.7857142857142856, "grad_norm": 0.43159736313225455, "learning_rate": 1.4632299595963294e-07, "loss": 0.6136, "step": 871 }, { "epoch": 2.788912579957356, "grad_norm": 0.4328150604393648, "learning_rate": 1.418766337196431e-07, "loss": 0.5821, "step": 872 }, { "epoch": 2.7921108742004264, "grad_norm": 0.4510225598844259, "learning_rate": 1.374979021352757e-07, "loss": 0.544, "step": 873 }, { "epoch": 2.795309168443497, "grad_norm": 0.42414299491852264, "learning_rate": 1.3318686216350241e-07, "loss": 0.5917, "step": 874 }, { "epoch": 2.798507462686567, "grad_norm": 0.4491433631888564, "learning_rate": 1.2894357381894984e-07, "loss": 0.5709, "step": 875 }, { "epoch": 2.8017057569296373, "grad_norm": 0.435024650739532, "learning_rate": 1.2476809617306408e-07, "loss": 0.5768, "step": 876 }, { "epoch": 2.804904051172708, "grad_norm": 0.4376084724071413, "learning_rate": 1.206604873532885e-07, "loss": 0.5747, "step": 877 }, { "epoch": 2.8081023454157785, "grad_norm": 0.4430099826070007, "learning_rate": 1.166208045422551e-07, "loss": 0.5671, "step": 878 }, { "epoch": 2.8113006396588487, "grad_norm": 0.45821945748201703, "learning_rate": 1.1264910397698614e-07, "loss": 0.5801, "step": 879 }, { "epoch": 2.814498933901919, "grad_norm": 0.42968896616005114, "learning_rate": 1.0874544094811424e-07, "loss": 0.609, "step": 880 }, { "epoch": 2.8176972281449895, "grad_norm": 0.4392112009989005, "learning_rate": 1.0490986979911189e-07, "loss": 0.6004, "step": 881 }, { "epoch": 2.8208955223880596, "grad_norm": 0.4678179782402967, "learning_rate": 1.0114244392553318e-07, "loss": 0.5799, "step": 882 }, { "epoch": 2.8240938166311302, "grad_norm": 0.4649676388307849, "learning_rate": 9.744321577427218e-08, "loss": 0.5694, "step": 883 }, { "epoch": 2.8272921108742004, "grad_norm": 0.4410645863970879, "learning_rate": 9.381223684283291e-08, "loss": 0.5792, "step": 884 }, { "epoch": 2.8304904051172706, "grad_norm": 0.41344134456514686, "learning_rate": 9.024955767861054e-08, "loss": 0.6179, "step": 885 }, { "epoch": 2.833688699360341, "grad_norm": 0.45207918900612454, "learning_rate": 8.675522787819023e-08, "loss": 0.5765, "step": 886 }, { "epoch": 2.836886993603412, "grad_norm": 0.4815854246181784, "learning_rate": 8.332929608665553e-08, "loss": 0.5622, "step": 887 }, { "epoch": 2.840085287846482, "grad_norm": 0.4104529223314507, "learning_rate": 7.997180999691101e-08, "loss": 0.6061, "step": 888 }, { "epoch": 2.843283582089552, "grad_norm": 0.43863504264752207, "learning_rate": 7.668281634901686e-08, "loss": 0.6205, "step": 889 }, { "epoch": 2.8464818763326227, "grad_norm": 0.42107787916551426, "learning_rate": 7.346236092954318e-08, "loss": 0.6112, "step": 890 }, { "epoch": 2.849680170575693, "grad_norm": 0.45281531695898647, "learning_rate": 7.031048857092604e-08, "loss": 0.5897, "step": 891 }, { "epoch": 2.8528784648187635, "grad_norm": 0.4461369389804025, "learning_rate": 6.722724315084805e-08, "loss": 0.5908, "step": 892 }, { "epoch": 2.8560767590618337, "grad_norm": 0.4523838631460964, "learning_rate": 6.421266759162659e-08, "loss": 0.6059, "step": 893 }, { "epoch": 2.859275053304904, "grad_norm": 0.4201076758194753, "learning_rate": 6.12668038596137e-08, "loss": 0.6017, "step": 894 }, { "epoch": 2.8624733475479744, "grad_norm": 0.4576782992482849, "learning_rate": 5.838969296461605e-08, "loss": 0.5595, "step": 895 }, { "epoch": 2.8656716417910446, "grad_norm": 0.452789058187118, "learning_rate": 5.5581374959320366e-08, "loss": 0.6069, "step": 896 }, { "epoch": 2.868869936034115, "grad_norm": 0.4502806418223326, "learning_rate": 5.2841888938738314e-08, "loss": 0.595, "step": 897 }, { "epoch": 2.8720682302771854, "grad_norm": 0.4291345074640387, "learning_rate": 5.017127303966085e-08, "loss": 0.5737, "step": 898 }, { "epoch": 2.875266524520256, "grad_norm": 0.4275943368139046, "learning_rate": 4.7569564440128055e-08, "loss": 0.5802, "step": 899 }, { "epoch": 2.878464818763326, "grad_norm": 0.4783054811845979, "learning_rate": 4.50367993589107e-08, "loss": 0.5523, "step": 900 }, { "epoch": 2.8816631130063968, "grad_norm": 0.4207771911644098, "learning_rate": 4.257301305500672e-08, "loss": 0.6295, "step": 901 }, { "epoch": 2.884861407249467, "grad_norm": 0.4709474401474473, "learning_rate": 4.0178239827151077e-08, "loss": 0.5824, "step": 902 }, { "epoch": 2.888059701492537, "grad_norm": 0.47155045759162584, "learning_rate": 3.785251301333726e-08, "loss": 0.594, "step": 903 }, { "epoch": 2.8912579957356077, "grad_norm": 0.4456059748610983, "learning_rate": 3.559586499035206e-08, "loss": 0.5662, "step": 904 }, { "epoch": 2.894456289978678, "grad_norm": 0.4575365472232499, "learning_rate": 3.340832717332765e-08, "loss": 0.596, "step": 905 }, { "epoch": 2.8976545842217485, "grad_norm": 0.45897611846221986, "learning_rate": 3.128993001530245e-08, "loss": 0.5467, "step": 906 }, { "epoch": 2.9008528784648187, "grad_norm": 0.46086480691517584, "learning_rate": 2.9240703006797044e-08, "loss": 0.591, "step": 907 }, { "epoch": 2.9040511727078893, "grad_norm": 0.4578137617838968, "learning_rate": 2.7260674675404498e-08, "loss": 0.5807, "step": 908 }, { "epoch": 2.9072494669509594, "grad_norm": 0.4744681575485193, "learning_rate": 2.5349872585392898e-08, "loss": 0.5882, "step": 909 }, { "epoch": 2.91044776119403, "grad_norm": 0.43778348787470217, "learning_rate": 2.3508323337321225e-08, "loss": 0.555, "step": 910 }, { "epoch": 2.9136460554371, "grad_norm": 0.43656355895512294, "learning_rate": 2.1736052567670195e-08, "loss": 0.5777, "step": 911 }, { "epoch": 2.9168443496801704, "grad_norm": 0.4664832575605358, "learning_rate": 2.0033084948483104e-08, "loss": 0.5442, "step": 912 }, { "epoch": 2.920042643923241, "grad_norm": 0.438143790231801, "learning_rate": 1.8399444187024995e-08, "loss": 0.5745, "step": 913 }, { "epoch": 2.923240938166311, "grad_norm": 0.42188051822462064, "learning_rate": 1.6835153025451246e-08, "loss": 0.6018, "step": 914 }, { "epoch": 2.9264392324093818, "grad_norm": 0.41865517546309167, "learning_rate": 1.534023324049061e-08, "loss": 0.5891, "step": 915 }, { "epoch": 2.929637526652452, "grad_norm": 0.41837714838468193, "learning_rate": 1.3914705643143788e-08, "loss": 0.6085, "step": 916 }, { "epoch": 2.9328358208955225, "grad_norm": 0.465979632617469, "learning_rate": 1.2558590078390886e-08, "loss": 0.5841, "step": 917 }, { "epoch": 2.9360341151385927, "grad_norm": 0.45654976609627784, "learning_rate": 1.1271905424918294e-08, "loss": 0.6035, "step": 918 }, { "epoch": 2.9392324093816633, "grad_norm": 0.45629479117279054, "learning_rate": 1.0054669594853905e-08, "loss": 0.5988, "step": 919 }, { "epoch": 2.9424307036247335, "grad_norm": 0.44536389990804115, "learning_rate": 8.906899533517866e-09, "loss": 0.6208, "step": 920 }, { "epoch": 2.9456289978678036, "grad_norm": 0.4513455406012691, "learning_rate": 7.828611219187765e-09, "loss": 0.6199, "step": 921 }, { "epoch": 2.9488272921108742, "grad_norm": 0.43738097436145257, "learning_rate": 6.819819662874372e-09, "loss": 0.5942, "step": 922 }, { "epoch": 2.9520255863539444, "grad_norm": 0.45522291904606826, "learning_rate": 5.88053890811513e-09, "loss": 0.6035, "step": 923 }, { "epoch": 2.955223880597015, "grad_norm": 0.4521268106702668, "learning_rate": 5.0107820307770945e-09, "loss": 0.6053, "step": 924 }, { "epoch": 2.958422174840085, "grad_norm": 0.4313674140155561, "learning_rate": 4.210561138873193e-09, "loss": 0.58, "step": 925 }, { "epoch": 2.961620469083156, "grad_norm": 0.4389755845527305, "learning_rate": 3.4798873723984604e-09, "loss": 0.6198, "step": 926 }, { "epoch": 2.964818763326226, "grad_norm": 0.4358918838142038, "learning_rate": 2.818770903170176e-09, "loss": 0.6201, "step": 927 }, { "epoch": 2.9680170575692966, "grad_norm": 0.446788281214219, "learning_rate": 2.2272209346885233e-09, "loss": 0.5517, "step": 928 }, { "epoch": 2.9712153518123667, "grad_norm": 0.4627514418027653, "learning_rate": 1.7052457020089175e-09, "loss": 0.602, "step": 929 }, { "epoch": 2.974413646055437, "grad_norm": 0.46487447660834663, "learning_rate": 1.2528524716259872e-09, "loss": 0.5523, "step": 930 }, { "epoch": 2.9776119402985075, "grad_norm": 0.4291177591260577, "learning_rate": 8.700475413719877e-10, "loss": 0.6038, "step": 931 }, { "epoch": 2.9808102345415777, "grad_norm": 0.4440151991197288, "learning_rate": 5.568362403318706e-10, "loss": 0.5722, "step": 932 }, { "epoch": 2.9840085287846483, "grad_norm": 0.41546760769750524, "learning_rate": 3.132229287666766e-10, "loss": 0.6158, "step": 933 }, { "epoch": 2.9872068230277184, "grad_norm": 0.47012397918132254, "learning_rate": 1.3921099805302985e-10, "loss": 0.612, "step": 934 }, { "epoch": 2.990405117270789, "grad_norm": 0.4452241770705769, "learning_rate": 3.480287063706289e-11, "loss": 0.5822, "step": 935 }, { "epoch": 2.9936034115138592, "grad_norm": 0.4503894897532014, "learning_rate": 0.0, "loss": 0.5513, "step": 936 }, { "epoch": 2.9936034115138592, "step": 936, "total_flos": 9.051650706620744e+17, "train_loss": 0.0, "train_runtime": 4.9645, "train_samples_per_second": 18128.643, "train_steps_per_second": 188.538 } ], "logging_steps": 1, "max_steps": 936, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9.051650706620744e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }