{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.994059405940594, "eval_steps": 500, "global_step": 378, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.015841584158415842, "grad_norm": 9.02661650060573, "learning_rate": 3.3333333333333333e-06, "loss": 1.9636, "step": 2 }, { "epoch": 0.031683168316831684, "grad_norm": 6.243135269339089, "learning_rate": 6.666666666666667e-06, "loss": 1.9061, "step": 4 }, { "epoch": 0.047524752475247525, "grad_norm": 5.193323051829328, "learning_rate": 1e-05, "loss": 1.6579, "step": 6 }, { "epoch": 0.06336633663366337, "grad_norm": 5.161379651013384, "learning_rate": 1.3333333333333333e-05, "loss": 1.4226, "step": 8 }, { "epoch": 0.07920792079207921, "grad_norm": 3.1813181834190916, "learning_rate": 1.6666666666666667e-05, "loss": 1.2698, "step": 10 }, { "epoch": 0.09504950495049505, "grad_norm": 3.948535232635907, "learning_rate": 2e-05, "loss": 1.1848, "step": 12 }, { "epoch": 0.11089108910891089, "grad_norm": 2.6512379356069657, "learning_rate": 1.999852647705027e-05, "loss": 1.2354, "step": 14 }, { "epoch": 0.12673267326732673, "grad_norm": 3.032346402915996, "learning_rate": 1.9994106342455053e-05, "loss": 1.161, "step": 16 }, { "epoch": 0.14257425742574256, "grad_norm": 2.1212186467865726, "learning_rate": 1.9986740898848306e-05, "loss": 1.1529, "step": 18 }, { "epoch": 0.15841584158415842, "grad_norm": 2.135193559072075, "learning_rate": 1.9976432316860065e-05, "loss": 1.0521, "step": 20 }, { "epoch": 0.17425742574257425, "grad_norm": 2.1958764723179547, "learning_rate": 1.9963183634476757e-05, "loss": 1.0078, "step": 22 }, { "epoch": 0.1900990099009901, "grad_norm": 2.2450349863339, "learning_rate": 1.9946998756145894e-05, "loss": 0.9967, "step": 24 }, { "epoch": 0.20594059405940593, "grad_norm": 2.509752007569588, "learning_rate": 1.99278824516254e-05, "loss": 1.0667, "step": 26 }, { "epoch": 0.22178217821782178, "grad_norm": 2.201210356151739, "learning_rate": 1.990584035457797e-05, "loss": 1.0156, "step": 28 }, { "epoch": 0.2376237623762376, "grad_norm": 2.1135504562735723, "learning_rate": 1.9880878960910772e-05, "loss": 0.9989, "step": 30 }, { "epoch": 0.25346534653465347, "grad_norm": 2.0708467368043655, "learning_rate": 1.985300562686109e-05, "loss": 0.9244, "step": 32 }, { "epoch": 0.2693069306930693, "grad_norm": 1.9881447087099158, "learning_rate": 1.982222856682841e-05, "loss": 0.912, "step": 34 }, { "epoch": 0.2851485148514851, "grad_norm": 2.1711952522187965, "learning_rate": 1.978855685095358e-05, "loss": 0.9678, "step": 36 }, { "epoch": 0.300990099009901, "grad_norm": 1.8520892574995997, "learning_rate": 1.9752000402445824e-05, "loss": 0.9405, "step": 38 }, { "epoch": 0.31683168316831684, "grad_norm": 2.330675399944507, "learning_rate": 1.9712569994658315e-05, "loss": 1.0241, "step": 40 }, { "epoch": 0.3326732673267327, "grad_norm": 1.9912264091727097, "learning_rate": 1.9670277247913205e-05, "loss": 0.9204, "step": 42 }, { "epoch": 0.3485148514851485, "grad_norm": 2.123562111523868, "learning_rate": 1.9625134626077084e-05, "loss": 0.9179, "step": 44 }, { "epoch": 0.36435643564356435, "grad_norm": 1.9790301716804515, "learning_rate": 1.9577155432887805e-05, "loss": 0.898, "step": 46 }, { "epoch": 0.3801980198019802, "grad_norm": 2.008624349315442, "learning_rate": 1.9526353808033827e-05, "loss": 0.9364, "step": 48 }, { "epoch": 0.39603960396039606, "grad_norm": 1.888190643885973, "learning_rate": 1.947274472298717e-05, "loss": 0.8907, "step": 50 }, { "epoch": 0.41188118811881186, "grad_norm": 2.015566919384652, "learning_rate": 1.941634397659126e-05, "loss": 0.9009, "step": 52 }, { "epoch": 0.4277227722772277, "grad_norm": 1.9768310339741355, "learning_rate": 1.9357168190404937e-05, "loss": 0.8975, "step": 54 }, { "epoch": 0.44356435643564357, "grad_norm": 2.133682758220573, "learning_rate": 1.9295234803804005e-05, "loss": 0.9143, "step": 56 }, { "epoch": 0.4594059405940594, "grad_norm": 1.9749224153125167, "learning_rate": 1.9230562068841764e-05, "loss": 0.9031, "step": 58 }, { "epoch": 0.4752475247524752, "grad_norm": 1.8623068380901364, "learning_rate": 1.916316904487005e-05, "loss": 0.8818, "step": 60 }, { "epoch": 0.4910891089108911, "grad_norm": 1.8392963872778836, "learning_rate": 1.909307559292236e-05, "loss": 0.8816, "step": 62 }, { "epoch": 0.5069306930693069, "grad_norm": 1.8853129700358418, "learning_rate": 1.9020302369860708e-05, "loss": 0.9097, "step": 64 }, { "epoch": 0.5227722772277228, "grad_norm": 2.137270128251829, "learning_rate": 1.8944870822287957e-05, "loss": 0.9296, "step": 66 }, { "epoch": 0.5386138613861386, "grad_norm": 1.7161417851549035, "learning_rate": 1.8866803180227403e-05, "loss": 0.8762, "step": 68 }, { "epoch": 0.5544554455445545, "grad_norm": 2.4673647833737093, "learning_rate": 1.8786122450571485e-05, "loss": 0.8757, "step": 70 }, { "epoch": 0.5702970297029702, "grad_norm": 1.9999583978057465, "learning_rate": 1.8702852410301556e-05, "loss": 0.9574, "step": 72 }, { "epoch": 0.5861386138613861, "grad_norm": 1.908768010347383, "learning_rate": 1.861701759948068e-05, "loss": 0.864, "step": 74 }, { "epoch": 0.601980198019802, "grad_norm": 1.9712540583735692, "learning_rate": 1.85286433140216e-05, "loss": 0.9225, "step": 76 }, { "epoch": 0.6178217821782178, "grad_norm": 1.7764975660995597, "learning_rate": 1.8437755598231857e-05, "loss": 0.9098, "step": 78 }, { "epoch": 0.6336633663366337, "grad_norm": 1.8462120353772475, "learning_rate": 1.8344381237138473e-05, "loss": 0.9059, "step": 80 }, { "epoch": 0.6495049504950495, "grad_norm": 1.7841160221457657, "learning_rate": 1.8248547748594246e-05, "loss": 0.8664, "step": 82 }, { "epoch": 0.6653465346534654, "grad_norm": 1.8561876013842784, "learning_rate": 1.8150283375168112e-05, "loss": 0.8621, "step": 84 }, { "epoch": 0.6811881188118812, "grad_norm": 2.203573345708881, "learning_rate": 1.8049617075821962e-05, "loss": 0.8719, "step": 86 }, { "epoch": 0.697029702970297, "grad_norm": 1.7796878917747134, "learning_rate": 1.794657851737625e-05, "loss": 0.8964, "step": 88 }, { "epoch": 0.7128712871287128, "grad_norm": 1.838149293961002, "learning_rate": 1.7841198065767107e-05, "loss": 0.898, "step": 90 }, { "epoch": 0.7287128712871287, "grad_norm": 1.903112272121249, "learning_rate": 1.77335067770973e-05, "loss": 0.8394, "step": 92 }, { "epoch": 0.7445544554455445, "grad_norm": 1.968601238697486, "learning_rate": 1.7623536388483902e-05, "loss": 0.8807, "step": 94 }, { "epoch": 0.7603960396039604, "grad_norm": 1.975435786088584, "learning_rate": 1.7511319308705198e-05, "loss": 0.829, "step": 96 }, { "epoch": 0.7762376237623763, "grad_norm": 2.008008309376045, "learning_rate": 1.7396888608649673e-05, "loss": 0.8512, "step": 98 }, { "epoch": 0.7920792079207921, "grad_norm": 1.8389509044704804, "learning_rate": 1.7280278011569848e-05, "loss": 0.8546, "step": 100 }, { "epoch": 0.807920792079208, "grad_norm": 1.842123818268232, "learning_rate": 1.7161521883143936e-05, "loss": 0.8344, "step": 102 }, { "epoch": 0.8237623762376237, "grad_norm": 1.7379472653976658, "learning_rate": 1.7040655221348057e-05, "loss": 0.8621, "step": 104 }, { "epoch": 0.8396039603960396, "grad_norm": 1.8262243602907722, "learning_rate": 1.6917713646142222e-05, "loss": 0.8454, "step": 106 }, { "epoch": 0.8554455445544554, "grad_norm": 1.656358917650264, "learning_rate": 1.679273338897293e-05, "loss": 0.8517, "step": 108 }, { "epoch": 0.8712871287128713, "grad_norm": 1.741005790971336, "learning_rate": 1.6665751282095634e-05, "loss": 0.8372, "step": 110 }, { "epoch": 0.8871287128712871, "grad_norm": 1.7417421499425474, "learning_rate": 1.653680474772006e-05, "loss": 0.8436, "step": 112 }, { "epoch": 0.902970297029703, "grad_norm": 1.8414378471720574, "learning_rate": 1.6405931786981753e-05, "loss": 0.84, "step": 114 }, { "epoch": 0.9188118811881189, "grad_norm": 1.710424834244549, "learning_rate": 1.6273170968742942e-05, "loss": 0.8252, "step": 116 }, { "epoch": 0.9346534653465347, "grad_norm": 1.7602252189051653, "learning_rate": 1.613856141822612e-05, "loss": 0.8332, "step": 118 }, { "epoch": 0.9504950495049505, "grad_norm": 1.7838059979323873, "learning_rate": 1.6002142805483686e-05, "loss": 0.8228, "step": 120 }, { "epoch": 0.9663366336633663, "grad_norm": 1.7374417300201197, "learning_rate": 1.586395533370696e-05, "loss": 0.8231, "step": 122 }, { "epoch": 0.9821782178217822, "grad_norm": 1.6853967583594676, "learning_rate": 1.572403972737815e-05, "loss": 0.8337, "step": 124 }, { "epoch": 0.998019801980198, "grad_norm": 1.887105041037784, "learning_rate": 1.5582437220268648e-05, "loss": 0.8572, "step": 126 }, { "epoch": 1.0138613861386139, "grad_norm": 1.6531514365483646, "learning_rate": 1.5439189543287247e-05, "loss": 0.6016, "step": 128 }, { "epoch": 1.0297029702970297, "grad_norm": 1.736528180579246, "learning_rate": 1.529433891218185e-05, "loss": 0.5991, "step": 130 }, { "epoch": 1.0455445544554456, "grad_norm": 1.7551648317268331, "learning_rate": 1.5147928015098309e-05, "loss": 0.5635, "step": 132 }, { "epoch": 1.0613861386138614, "grad_norm": 1.8298936253985936, "learning_rate": 1.5000000000000002e-05, "loss": 0.565, "step": 134 }, { "epoch": 1.0772277227722773, "grad_norm": 1.7317255025108698, "learning_rate": 1.4850598461951963e-05, "loss": 0.5592, "step": 136 }, { "epoch": 1.0930693069306932, "grad_norm": 1.748739410420368, "learning_rate": 1.4699767430273202e-05, "loss": 0.574, "step": 138 }, { "epoch": 1.108910891089109, "grad_norm": 1.7869424929925428, "learning_rate": 1.454755135556106e-05, "loss": 0.5741, "step": 140 }, { "epoch": 1.1247524752475249, "grad_norm": 1.8524853975885442, "learning_rate": 1.4393995096591415e-05, "loss": 0.5454, "step": 142 }, { "epoch": 1.1405940594059407, "grad_norm": 1.8345725351471323, "learning_rate": 1.423914390709861e-05, "loss": 0.547, "step": 144 }, { "epoch": 1.1564356435643564, "grad_norm": 1.8060990165115691, "learning_rate": 1.4083043422438936e-05, "loss": 0.5953, "step": 146 }, { "epoch": 1.1722772277227722, "grad_norm": 1.6416315510212192, "learning_rate": 1.3925739646141721e-05, "loss": 0.5412, "step": 148 }, { "epoch": 1.188118811881188, "grad_norm": 1.664162470996372, "learning_rate": 1.3767278936351853e-05, "loss": 0.5293, "step": 150 }, { "epoch": 1.203960396039604, "grad_norm": 1.6113229932234892, "learning_rate": 1.3607707992167836e-05, "loss": 0.5276, "step": 152 }, { "epoch": 1.2198019801980198, "grad_norm": 1.799039577469552, "learning_rate": 1.3447073839879339e-05, "loss": 0.5569, "step": 154 }, { "epoch": 1.2356435643564356, "grad_norm": 1.6604650735978537, "learning_rate": 1.3285423819108349e-05, "loss": 0.5435, "step": 156 }, { "epoch": 1.2514851485148515, "grad_norm": 1.7636186077229687, "learning_rate": 1.3122805568857948e-05, "loss": 0.5837, "step": 158 }, { "epoch": 1.2673267326732673, "grad_norm": 1.6232817216358415, "learning_rate": 1.2959267013472894e-05, "loss": 0.546, "step": 160 }, { "epoch": 1.2831683168316832, "grad_norm": 1.6172712656831925, "learning_rate": 1.2794856348516095e-05, "loss": 0.5574, "step": 162 }, { "epoch": 1.299009900990099, "grad_norm": 1.8102400872571573, "learning_rate": 1.2629622026565147e-05, "loss": 0.5796, "step": 164 }, { "epoch": 1.314851485148515, "grad_norm": 1.7802623079191595, "learning_rate": 1.2463612742933148e-05, "loss": 0.5442, "step": 166 }, { "epoch": 1.3306930693069308, "grad_norm": 1.7022382654116954, "learning_rate": 1.2296877421317958e-05, "loss": 0.585, "step": 168 }, { "epoch": 1.3465346534653464, "grad_norm": 1.6401839574055097, "learning_rate": 1.2129465199384158e-05, "loss": 0.5397, "step": 170 }, { "epoch": 1.3623762376237623, "grad_norm": 1.7101624547819703, "learning_rate": 1.196142541428197e-05, "loss": 0.5745, "step": 172 }, { "epoch": 1.378217821782178, "grad_norm": 1.6535320779874634, "learning_rate": 1.1792807588107358e-05, "loss": 0.5357, "step": 174 }, { "epoch": 1.394059405940594, "grad_norm": 1.6245529400611858, "learning_rate": 1.1623661413307638e-05, "loss": 0.5543, "step": 176 }, { "epoch": 1.4099009900990098, "grad_norm": 1.6629345686182546, "learning_rate": 1.14540367380369e-05, "loss": 0.5237, "step": 178 }, { "epoch": 1.4257425742574257, "grad_norm": 1.68909608575907, "learning_rate": 1.1283983551465512e-05, "loss": 0.5746, "step": 180 }, { "epoch": 1.4415841584158415, "grad_norm": 1.7585724609639268, "learning_rate": 1.1113551969048088e-05, "loss": 0.542, "step": 182 }, { "epoch": 1.4574257425742574, "grad_norm": 1.598454179166171, "learning_rate": 1.0942792217754245e-05, "loss": 0.5217, "step": 184 }, { "epoch": 1.4732673267326732, "grad_norm": 1.720537069197336, "learning_rate": 1.0771754621266466e-05, "loss": 0.5567, "step": 186 }, { "epoch": 1.489108910891089, "grad_norm": 1.6147204766507826, "learning_rate": 1.0600489585149485e-05, "loss": 0.5509, "step": 188 }, { "epoch": 1.504950495049505, "grad_norm": 1.7292799453559213, "learning_rate": 1.0429047581995547e-05, "loss": 0.5479, "step": 190 }, { "epoch": 1.5207920792079208, "grad_norm": 1.7407208786866153, "learning_rate": 1.0257479136549889e-05, "loss": 0.5539, "step": 192 }, { "epoch": 1.5366336633663367, "grad_norm": 1.6829968026665005, "learning_rate": 1.0085834810820871e-05, "loss": 0.5236, "step": 194 }, { "epoch": 1.5524752475247525, "grad_norm": 1.5077533197060748, "learning_rate": 9.914165189179132e-06, "loss": 0.5492, "step": 196 }, { "epoch": 1.5683168316831684, "grad_norm": 1.619225808673778, "learning_rate": 9.742520863450116e-06, "loss": 0.5435, "step": 198 }, { "epoch": 1.5841584158415842, "grad_norm": 1.6946922694133721, "learning_rate": 9.570952418004455e-06, "loss": 0.5628, "step": 200 }, { "epoch": 1.6, "grad_norm": 1.7264082675727195, "learning_rate": 9.399510414850518e-06, "loss": 0.5405, "step": 202 }, { "epoch": 1.615841584158416, "grad_norm": 1.668467187331713, "learning_rate": 9.228245378733537e-06, "loss": 0.5503, "step": 204 }, { "epoch": 1.6316831683168318, "grad_norm": 1.6228977289298525, "learning_rate": 9.057207782245756e-06, "loss": 0.552, "step": 206 }, { "epoch": 1.6475247524752477, "grad_norm": 1.6720968963932732, "learning_rate": 8.886448030951912e-06, "loss": 0.5266, "step": 208 }, { "epoch": 1.6633663366336635, "grad_norm": 1.6117561664476598, "learning_rate": 8.71601644853449e-06, "loss": 0.5439, "step": 210 }, { "epoch": 1.6792079207920794, "grad_norm": 1.7309748707062824, "learning_rate": 8.545963261963102e-06, "loss": 0.5518, "step": 212 }, { "epoch": 1.695049504950495, "grad_norm": 1.727253391380127, "learning_rate": 8.376338586692367e-06, "loss": 0.5689, "step": 214 }, { "epoch": 1.7108910891089109, "grad_norm": 1.6465767825228266, "learning_rate": 8.207192411892645e-06, "loss": 0.5263, "step": 216 }, { "epoch": 1.7267326732673267, "grad_norm": 1.7061559526686028, "learning_rate": 8.038574585718032e-06, "loss": 0.561, "step": 218 }, { "epoch": 1.7425742574257426, "grad_norm": 1.641091211390163, "learning_rate": 7.870534800615845e-06, "loss": 0.5279, "step": 220 }, { "epoch": 1.7584158415841584, "grad_norm": 1.703349596429662, "learning_rate": 7.703122578682047e-06, "loss": 0.5496, "step": 222 }, { "epoch": 1.7742574257425743, "grad_norm": 1.6149104022531628, "learning_rate": 7.536387257066854e-06, "loss": 0.5592, "step": 224 }, { "epoch": 1.7900990099009901, "grad_norm": 1.5562454704981712, "learning_rate": 7.370377973434854e-06, "loss": 0.5439, "step": 226 }, { "epoch": 1.805940594059406, "grad_norm": 1.8575887871163193, "learning_rate": 7.2051436514839064e-06, "loss": 0.5488, "step": 228 }, { "epoch": 1.8217821782178216, "grad_norm": 1.584715339525317, "learning_rate": 7.040732986527108e-06, "loss": 0.5308, "step": 230 }, { "epoch": 1.8376237623762375, "grad_norm": 1.501398258461206, "learning_rate": 6.877194431142055e-06, "loss": 0.5239, "step": 232 }, { "epoch": 1.8534653465346533, "grad_norm": 1.6132499536201124, "learning_rate": 6.714576180891653e-06, "loss": 0.5522, "step": 234 }, { "epoch": 1.8693069306930692, "grad_norm": 1.6315843930884224, "learning_rate": 6.552926160120663e-06, "loss": 0.5077, "step": 236 }, { "epoch": 1.885148514851485, "grad_norm": 1.7272400167492237, "learning_rate": 6.3922920078321685e-06, "loss": 0.5633, "step": 238 }, { "epoch": 1.900990099009901, "grad_norm": 1.6689075327553573, "learning_rate": 6.232721063648148e-06, "loss": 0.5055, "step": 240 }, { "epoch": 1.9168316831683168, "grad_norm": 1.578818325825261, "learning_rate": 6.074260353858283e-06, "loss": 0.5389, "step": 242 }, { "epoch": 1.9326732673267326, "grad_norm": 1.7205801010677875, "learning_rate": 5.916956577561066e-06, "loss": 0.5562, "step": 244 }, { "epoch": 1.9485148514851485, "grad_norm": 1.6477307933447243, "learning_rate": 5.760856092901394e-06, "loss": 0.5193, "step": 246 }, { "epoch": 1.9643564356435643, "grad_norm": 1.5685239373769044, "learning_rate": 5.6060049034085815e-06, "loss": 0.5512, "step": 248 }, { "epoch": 1.9801980198019802, "grad_norm": 1.6936425396642782, "learning_rate": 5.4524486444389455e-06, "loss": 0.5584, "step": 250 }, { "epoch": 1.996039603960396, "grad_norm": 1.7086317273357263, "learning_rate": 5.300232569726805e-06, "loss": 0.5644, "step": 252 }, { "epoch": 2.011881188118812, "grad_norm": 1.5710974953396362, "learning_rate": 5.14940153804804e-06, "loss": 0.3551, "step": 254 }, { "epoch": 2.0277227722772277, "grad_norm": 1.3779991120357036, "learning_rate": 5.000000000000003e-06, "loss": 0.3204, "step": 256 }, { "epoch": 2.0435643564356436, "grad_norm": 1.3532509589808863, "learning_rate": 4.852071984901696e-06, "loss": 0.3381, "step": 258 }, { "epoch": 2.0594059405940595, "grad_norm": 1.5655956293278384, "learning_rate": 4.705661087818149e-06, "loss": 0.3114, "step": 260 }, { "epoch": 2.0752475247524753, "grad_norm": 1.8047288146207374, "learning_rate": 4.560810456712754e-06, "loss": 0.3167, "step": 262 }, { "epoch": 2.091089108910891, "grad_norm": 1.5969295392705383, "learning_rate": 4.417562779731355e-06, "loss": 0.2968, "step": 264 }, { "epoch": 2.106930693069307, "grad_norm": 1.5043862632074243, "learning_rate": 4.275960272621852e-06, "loss": 0.2873, "step": 266 }, { "epoch": 2.122772277227723, "grad_norm": 1.4566588428064557, "learning_rate": 4.1360446662930445e-06, "loss": 0.2906, "step": 268 }, { "epoch": 2.1386138613861387, "grad_norm": 1.4148675515775622, "learning_rate": 3.997857194516319e-06, "loss": 0.2902, "step": 270 }, { "epoch": 2.1544554455445546, "grad_norm": 1.4039039705486787, "learning_rate": 3.86143858177388e-06, "loss": 0.3055, "step": 272 }, { "epoch": 2.1702970297029704, "grad_norm": 1.4590478348079479, "learning_rate": 3.7268290312570622e-06, "loss": 0.2887, "step": 274 }, { "epoch": 2.1861386138613863, "grad_norm": 1.4904023110394045, "learning_rate": 3.594068213018249e-06, "loss": 0.2838, "step": 276 }, { "epoch": 2.201980198019802, "grad_norm": 1.4223228096195222, "learning_rate": 3.4631952522799396e-06, "loss": 0.2816, "step": 278 }, { "epoch": 2.217821782178218, "grad_norm": 1.4645691354886492, "learning_rate": 3.334248717904368e-06, "loss": 0.304, "step": 280 }, { "epoch": 2.233663366336634, "grad_norm": 1.446048636808659, "learning_rate": 3.207266611027069e-06, "loss": 0.3098, "step": 282 }, { "epoch": 2.2495049504950497, "grad_norm": 1.4585412000734919, "learning_rate": 3.082286353857782e-06, "loss": 0.2881, "step": 284 }, { "epoch": 2.2653465346534656, "grad_norm": 1.4508547813683121, "learning_rate": 2.9593447786519424e-06, "loss": 0.2961, "step": 286 }, { "epoch": 2.2811881188118814, "grad_norm": 1.4305009349332931, "learning_rate": 2.8384781168560693e-06, "loss": 0.3015, "step": 288 }, { "epoch": 2.297029702970297, "grad_norm": 1.3742813840340038, "learning_rate": 2.719721988430153e-06, "loss": 0.2826, "step": 290 }, { "epoch": 2.3128712871287127, "grad_norm": 1.4476247412504035, "learning_rate": 2.6031113913503337e-06, "loss": 0.2768, "step": 292 }, { "epoch": 2.3287128712871286, "grad_norm": 1.3974509182841102, "learning_rate": 2.4886806912948034e-06, "loss": 0.2669, "step": 294 }, { "epoch": 2.3445544554455444, "grad_norm": 1.465709495723541, "learning_rate": 2.376463611516098e-06, "loss": 0.2952, "step": 296 }, { "epoch": 2.3603960396039603, "grad_norm": 1.4155568971087533, "learning_rate": 2.2664932229027025e-06, "loss": 0.2863, "step": 298 }, { "epoch": 2.376237623762376, "grad_norm": 1.4443007558100314, "learning_rate": 2.158801934232897e-06, "loss": 0.286, "step": 300 }, { "epoch": 2.392079207920792, "grad_norm": 1.3965088698230241, "learning_rate": 2.0534214826237486e-06, "loss": 0.2783, "step": 302 }, { "epoch": 2.407920792079208, "grad_norm": 1.4407127199864882, "learning_rate": 1.9503829241780416e-06, "loss": 0.2852, "step": 304 }, { "epoch": 2.4237623762376237, "grad_norm": 1.4265187454796464, "learning_rate": 1.8497166248318876e-06, "loss": 0.2945, "step": 306 }, { "epoch": 2.4396039603960396, "grad_norm": 1.5393606880372226, "learning_rate": 1.7514522514057552e-06, "loss": 0.2906, "step": 308 }, { "epoch": 2.4554455445544554, "grad_norm": 1.509234066590099, "learning_rate": 1.6556187628615273e-06, "loss": 0.2803, "step": 310 }, { "epoch": 2.4712871287128713, "grad_norm": 1.4716057289372597, "learning_rate": 1.5622444017681438e-06, "loss": 0.2834, "step": 312 }, { "epoch": 2.487128712871287, "grad_norm": 1.4208546193648086, "learning_rate": 1.4713566859784045e-06, "loss": 0.2807, "step": 314 }, { "epoch": 2.502970297029703, "grad_norm": 1.4813272843019751, "learning_rate": 1.3829824005193183e-06, "loss": 0.2606, "step": 316 }, { "epoch": 2.518811881188119, "grad_norm": 1.4317127596411277, "learning_rate": 1.2971475896984475e-06, "loss": 0.2872, "step": 318 }, { "epoch": 2.5346534653465347, "grad_norm": 1.4190350255812814, "learning_rate": 1.2138775494285181e-06, "loss": 0.2849, "step": 320 }, { "epoch": 2.5504950495049505, "grad_norm": 1.3786354752806598, "learning_rate": 1.1331968197725985e-06, "loss": 0.2748, "step": 322 }, { "epoch": 2.5663366336633664, "grad_norm": 1.3687349820385863, "learning_rate": 1.0551291777120465e-06, "loss": 0.2637, "step": 324 }, { "epoch": 2.5821782178217823, "grad_norm": 1.5296662542963093, "learning_rate": 9.796976301392935e-07, "loss": 0.2763, "step": 326 }, { "epoch": 2.598019801980198, "grad_norm": 1.438465874033189, "learning_rate": 9.069244070776428e-07, "loss": 0.3135, "step": 328 }, { "epoch": 2.613861386138614, "grad_norm": 1.423122514861154, "learning_rate": 8.368309551299536e-07, "loss": 0.2867, "step": 330 }, { "epoch": 2.62970297029703, "grad_norm": 1.479495787675627, "learning_rate": 7.694379311582401e-07, "loss": 0.284, "step": 332 }, { "epoch": 2.6455445544554457, "grad_norm": 1.4083216396155038, "learning_rate": 7.047651961959978e-07, "loss": 0.2815, "step": 334 }, { "epoch": 2.6613861386138615, "grad_norm": 1.3782207183520463, "learning_rate": 6.428318095950648e-07, "loss": 0.2746, "step": 336 }, { "epoch": 2.6772277227722774, "grad_norm": 1.4227248154028977, "learning_rate": 5.836560234087418e-07, "loss": 0.2646, "step": 338 }, { "epoch": 2.693069306930693, "grad_norm": 1.4574818091415243, "learning_rate": 5.272552770128314e-07, "loss": 0.2725, "step": 340 }, { "epoch": 2.7089108910891087, "grad_norm": 1.3601121324659562, "learning_rate": 4.73646191966175e-07, "loss": 0.2609, "step": 342 }, { "epoch": 2.7247524752475245, "grad_norm": 1.4223626310719977, "learning_rate": 4.2284456711219723e-07, "loss": 0.2949, "step": 344 }, { "epoch": 2.7405940594059404, "grad_norm": 1.3931852574014099, "learning_rate": 3.748653739229191e-07, "loss": 0.2818, "step": 346 }, { "epoch": 2.756435643564356, "grad_norm": 1.4108773420640486, "learning_rate": 3.2972275208679625e-07, "loss": 0.2888, "step": 348 }, { "epoch": 2.772277227722772, "grad_norm": 1.3496089808630627, "learning_rate": 2.8743000534168673e-07, "loss": 0.2651, "step": 350 }, { "epoch": 2.788118811881188, "grad_norm": 1.4057990507827347, "learning_rate": 2.479995975541749e-07, "loss": 0.2895, "step": 352 }, { "epoch": 2.803960396039604, "grad_norm": 1.4102665625188138, "learning_rate": 2.1144314904642194e-07, "loss": 0.2609, "step": 354 }, { "epoch": 2.8198019801980196, "grad_norm": 1.3292831585550606, "learning_rate": 1.7777143317159407e-07, "loss": 0.2717, "step": 356 }, { "epoch": 2.8356435643564355, "grad_norm": 1.3815722866503821, "learning_rate": 1.4699437313891007e-07, "loss": 0.2941, "step": 358 }, { "epoch": 2.8514851485148514, "grad_norm": 1.4033894606032193, "learning_rate": 1.1912103908922945e-07, "loss": 0.2747, "step": 360 }, { "epoch": 2.867326732673267, "grad_norm": 1.463974792809012, "learning_rate": 9.415964542203059e-08, "loss": 0.2759, "step": 362 }, { "epoch": 2.883168316831683, "grad_norm": 1.3556200240858571, "learning_rate": 7.21175483745995e-08, "loss": 0.2728, "step": 364 }, { "epoch": 2.899009900990099, "grad_norm": 1.4407004492072946, "learning_rate": 5.300124385410943e-08, "loss": 0.255, "step": 366 }, { "epoch": 2.9148514851485148, "grad_norm": 1.3953279840487336, "learning_rate": 3.681636552324452e-08, "loss": 0.2831, "step": 368 }, { "epoch": 2.9306930693069306, "grad_norm": 1.3987888464947156, "learning_rate": 2.3567683139936736e-08, "loss": 0.2607, "step": 370 }, { "epoch": 2.9465346534653465, "grad_norm": 1.4508777672512516, "learning_rate": 1.325910115169471e-08, "loss": 0.2599, "step": 372 }, { "epoch": 2.9623762376237623, "grad_norm": 1.3800670482596784, "learning_rate": 5.8936575449475284e-09, "loss": 0.2661, "step": 374 }, { "epoch": 2.978217821782178, "grad_norm": 1.394978304799224, "learning_rate": 1.47352294973091e-09, "loss": 0.286, "step": 376 }, { "epoch": 2.994059405940594, "grad_norm": 1.4850311019997162, "learning_rate": 0.0, "loss": 0.285, "step": 378 }, { "epoch": 2.994059405940594, "step": 378, "total_flos": 1243685281333248.0, "train_loss": 0.6043043333701986, "train_runtime": 10201.0989, "train_samples_per_second": 1.188, "train_steps_per_second": 0.037 } ], "logging_steps": 2, "max_steps": 378, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 32, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1243685281333248.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }