{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.008237986270022883, "eval_steps": 50, "global_step": 450, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 1.8306636155606407e-05, "grad_norm": 6.099522590637207, "learning_rate": 0.0, "loss": 0.6392, "step": 1 }, { "epoch": 3.6613272311212814e-05, "grad_norm": 8.014727592468262, "learning_rate": 2.0000000000000002e-07, "loss": 0.8375, "step": 2 }, { "epoch": 5.491990846681922e-05, "grad_norm": 10.58983039855957, "learning_rate": 4.0000000000000003e-07, "loss": 0.5021, "step": 3 }, { "epoch": 7.322654462242563e-05, "grad_norm": 6.642515659332275, "learning_rate": 6.000000000000001e-07, "loss": 0.7624, "step": 4 }, { "epoch": 9.153318077803204e-05, "grad_norm": 8.307601928710938, "learning_rate": 8.000000000000001e-07, "loss": 1.005, "step": 5 }, { "epoch": 0.00010983981693363844, "grad_norm": 11.282310485839844, "learning_rate": 1.0000000000000002e-06, "loss": 0.939, "step": 6 }, { "epoch": 0.00012814645308924485, "grad_norm": 9.1802339553833, "learning_rate": 1.2000000000000002e-06, "loss": 1.0419, "step": 7 }, { "epoch": 0.00014645308924485126, "grad_norm": 2.651777744293213, "learning_rate": 1.4000000000000001e-06, "loss": 0.4708, "step": 8 }, { "epoch": 0.00016475972540045766, "grad_norm": 7.600271701812744, "learning_rate": 1.6000000000000001e-06, "loss": 0.6082, "step": 9 }, { "epoch": 0.00018306636155606407, "grad_norm": 6.370179653167725, "learning_rate": 1.8000000000000001e-06, "loss": 0.8817, "step": 10 }, { "epoch": 0.00020137299771167048, "grad_norm": 7.1832499504089355, "learning_rate": 2.0000000000000003e-06, "loss": 0.5443, "step": 11 }, { "epoch": 0.00021967963386727689, "grad_norm": 6.7331223487854, "learning_rate": 2.2e-06, "loss": 0.5602, "step": 12 }, { "epoch": 0.0002379862700228833, "grad_norm": 8.009416580200195, "learning_rate": 2.4000000000000003e-06, "loss": 0.741, "step": 13 }, { "epoch": 0.0002562929061784897, "grad_norm": 9.607633590698242, "learning_rate": 2.6e-06, "loss": 0.7022, "step": 14 }, { "epoch": 0.00027459954233409613, "grad_norm": 10.585896492004395, "learning_rate": 2.8000000000000003e-06, "loss": 1.1929, "step": 15 }, { "epoch": 0.0002929061784897025, "grad_norm": 3.4262940883636475, "learning_rate": 3e-06, "loss": 0.5217, "step": 16 }, { "epoch": 0.00031121281464530895, "grad_norm": 11.94228458404541, "learning_rate": 3.2000000000000003e-06, "loss": 1.0208, "step": 17 }, { "epoch": 0.00032951945080091533, "grad_norm": 8.004399299621582, "learning_rate": 3.4000000000000005e-06, "loss": 0.5694, "step": 18 }, { "epoch": 0.00034782608695652176, "grad_norm": 7.853118896484375, "learning_rate": 3.6000000000000003e-06, "loss": 0.5695, "step": 19 }, { "epoch": 0.00036613272311212814, "grad_norm": 6.011293411254883, "learning_rate": 3.8000000000000005e-06, "loss": 0.4659, "step": 20 }, { "epoch": 0.0003844393592677346, "grad_norm": 10.169012069702148, "learning_rate": 4.000000000000001e-06, "loss": 0.8618, "step": 21 }, { "epoch": 0.00040274599542334096, "grad_norm": 3.0373430252075195, "learning_rate": 4.2000000000000004e-06, "loss": 0.2469, "step": 22 }, { "epoch": 0.0004210526315789474, "grad_norm": 7.847787380218506, "learning_rate": 4.4e-06, "loss": 0.738, "step": 23 }, { "epoch": 0.00043935926773455377, "grad_norm": 6.6885809898376465, "learning_rate": 4.600000000000001e-06, "loss": 0.6311, "step": 24 }, { "epoch": 0.0004576659038901602, "grad_norm": 7.829857349395752, "learning_rate": 4.800000000000001e-06, "loss": 0.5967, "step": 25 }, { "epoch": 0.0004759725400457666, "grad_norm": 15.481698989868164, "learning_rate": 5e-06, "loss": 0.9285, "step": 26 }, { "epoch": 0.000494279176201373, "grad_norm": 23.23982810974121, "learning_rate": 5.2e-06, "loss": 0.6175, "step": 27 }, { "epoch": 0.0005125858123569794, "grad_norm": 9.985934257507324, "learning_rate": 5.400000000000001e-06, "loss": 0.4765, "step": 28 }, { "epoch": 0.0005308924485125858, "grad_norm": 17.83190155029297, "learning_rate": 5.600000000000001e-06, "loss": 1.1398, "step": 29 }, { "epoch": 0.0005491990846681923, "grad_norm": 7.259341239929199, "learning_rate": 5.8e-06, "loss": 0.3864, "step": 30 }, { "epoch": 0.0005675057208237986, "grad_norm": 8.247320175170898, "learning_rate": 6e-06, "loss": 0.5853, "step": 31 }, { "epoch": 0.000585812356979405, "grad_norm": 8.801642417907715, "learning_rate": 6.200000000000001e-06, "loss": 1.197, "step": 32 }, { "epoch": 0.0006041189931350115, "grad_norm": 8.88306713104248, "learning_rate": 6.4000000000000006e-06, "loss": 0.3122, "step": 33 }, { "epoch": 0.0006224256292906179, "grad_norm": 10.055649757385254, "learning_rate": 6.600000000000001e-06, "loss": 0.5566, "step": 34 }, { "epoch": 0.0006407322654462242, "grad_norm": 12.692378044128418, "learning_rate": 6.800000000000001e-06, "loss": 1.0306, "step": 35 }, { "epoch": 0.0006590389016018307, "grad_norm": 6.696789741516113, "learning_rate": 7e-06, "loss": 0.3655, "step": 36 }, { "epoch": 0.0006773455377574371, "grad_norm": 5.720799922943115, "learning_rate": 7.2000000000000005e-06, "loss": 0.4338, "step": 37 }, { "epoch": 0.0006956521739130435, "grad_norm": 6.0295610427856445, "learning_rate": 7.4e-06, "loss": 0.5826, "step": 38 }, { "epoch": 0.0007139588100686499, "grad_norm": 3.975026845932007, "learning_rate": 7.600000000000001e-06, "loss": 0.5669, "step": 39 }, { "epoch": 0.0007322654462242563, "grad_norm": 8.472329139709473, "learning_rate": 7.800000000000002e-06, "loss": 0.8793, "step": 40 }, { "epoch": 0.0007505720823798627, "grad_norm": 4.2498321533203125, "learning_rate": 8.000000000000001e-06, "loss": 0.3409, "step": 41 }, { "epoch": 0.0007688787185354692, "grad_norm": 8.282734870910645, "learning_rate": 8.2e-06, "loss": 0.549, "step": 42 }, { "epoch": 0.0007871853546910755, "grad_norm": 7.966742992401123, "learning_rate": 8.400000000000001e-06, "loss": 0.7406, "step": 43 }, { "epoch": 0.0008054919908466819, "grad_norm": 17.36241340637207, "learning_rate": 8.6e-06, "loss": 1.0215, "step": 44 }, { "epoch": 0.0008237986270022883, "grad_norm": 8.946808815002441, "learning_rate": 8.8e-06, "loss": 0.5102, "step": 45 }, { "epoch": 0.0008421052631578948, "grad_norm": 9.421518325805664, "learning_rate": 9e-06, "loss": 1.1237, "step": 46 }, { "epoch": 0.0008604118993135011, "grad_norm": 11.310033798217773, "learning_rate": 9.200000000000002e-06, "loss": 0.7825, "step": 47 }, { "epoch": 0.0008787185354691075, "grad_norm": 11.267560005187988, "learning_rate": 9.4e-06, "loss": 1.0458, "step": 48 }, { "epoch": 0.000897025171624714, "grad_norm": 2.9605565071105957, "learning_rate": 9.600000000000001e-06, "loss": 0.2229, "step": 49 }, { "epoch": 0.0009153318077803204, "grad_norm": 5.507591724395752, "learning_rate": 9.800000000000001e-06, "loss": 0.3601, "step": 50 }, { "epoch": 0.0009336384439359267, "grad_norm": 7.83518123626709, "learning_rate": 1e-05, "loss": 0.5938, "step": 51 }, { "epoch": 0.0009519450800915332, "grad_norm": 6.110620975494385, "learning_rate": 1.02e-05, "loss": 0.8618, "step": 52 }, { "epoch": 0.0009702517162471396, "grad_norm": 7.460414409637451, "learning_rate": 1.04e-05, "loss": 0.8692, "step": 53 }, { "epoch": 0.000988558352402746, "grad_norm": 7.539251804351807, "learning_rate": 1.0600000000000002e-05, "loss": 0.7236, "step": 54 }, { "epoch": 0.0010068649885583525, "grad_norm": 5.357115268707275, "learning_rate": 1.0800000000000002e-05, "loss": 0.3101, "step": 55 }, { "epoch": 0.0010251716247139588, "grad_norm": 3.6166787147521973, "learning_rate": 1.1000000000000001e-05, "loss": 0.4426, "step": 56 }, { "epoch": 0.0010434782608695651, "grad_norm": 5.8816680908203125, "learning_rate": 1.1200000000000001e-05, "loss": 0.4829, "step": 57 }, { "epoch": 0.0010617848970251717, "grad_norm": 8.535571098327637, "learning_rate": 1.14e-05, "loss": 0.9369, "step": 58 }, { "epoch": 0.001080091533180778, "grad_norm": 5.3663482666015625, "learning_rate": 1.16e-05, "loss": 0.3664, "step": 59 }, { "epoch": 0.0010983981693363845, "grad_norm": 10.66060733795166, "learning_rate": 1.18e-05, "loss": 1.2902, "step": 60 }, { "epoch": 0.0011167048054919909, "grad_norm": 2.3751587867736816, "learning_rate": 1.2e-05, "loss": 0.2074, "step": 61 }, { "epoch": 0.0011350114416475972, "grad_norm": 6.494527816772461, "learning_rate": 1.22e-05, "loss": 0.5742, "step": 62 }, { "epoch": 0.0011533180778032037, "grad_norm": 1.9130464792251587, "learning_rate": 1.2400000000000002e-05, "loss": 0.1833, "step": 63 }, { "epoch": 0.00117162471395881, "grad_norm": 7.697911262512207, "learning_rate": 1.2600000000000001e-05, "loss": 1.0019, "step": 64 }, { "epoch": 0.0011899313501144164, "grad_norm": 16.3773250579834, "learning_rate": 1.2800000000000001e-05, "loss": 1.5617, "step": 65 }, { "epoch": 0.001208237986270023, "grad_norm": 6.1076836585998535, "learning_rate": 1.3000000000000001e-05, "loss": 0.5233, "step": 66 }, { "epoch": 0.0012265446224256293, "grad_norm": 1.6926074028015137, "learning_rate": 1.3200000000000002e-05, "loss": 0.2422, "step": 67 }, { "epoch": 0.0012448512585812358, "grad_norm": 3.413689374923706, "learning_rate": 1.3400000000000002e-05, "loss": 0.2152, "step": 68 }, { "epoch": 0.0012631578947368421, "grad_norm": 9.58433723449707, "learning_rate": 1.3600000000000002e-05, "loss": 1.0936, "step": 69 }, { "epoch": 0.0012814645308924484, "grad_norm": 7.524669647216797, "learning_rate": 1.38e-05, "loss": 1.1815, "step": 70 }, { "epoch": 0.001299771167048055, "grad_norm": 6.692052841186523, "learning_rate": 1.4e-05, "loss": 0.5309, "step": 71 }, { "epoch": 0.0013180778032036613, "grad_norm": 14.243474006652832, "learning_rate": 1.4200000000000001e-05, "loss": 0.8553, "step": 72 }, { "epoch": 0.0013363844393592676, "grad_norm": 10.684292793273926, "learning_rate": 1.4400000000000001e-05, "loss": 0.9399, "step": 73 }, { "epoch": 0.0013546910755148742, "grad_norm": 5.949723243713379, "learning_rate": 1.46e-05, "loss": 0.7573, "step": 74 }, { "epoch": 0.0013729977116704805, "grad_norm": 6.119026184082031, "learning_rate": 1.48e-05, "loss": 0.5557, "step": 75 }, { "epoch": 0.001391304347826087, "grad_norm": 6.306831359863281, "learning_rate": 1.5000000000000002e-05, "loss": 0.6165, "step": 76 }, { "epoch": 0.0014096109839816934, "grad_norm": 5.815463066101074, "learning_rate": 1.5200000000000002e-05, "loss": 0.6491, "step": 77 }, { "epoch": 0.0014279176201372997, "grad_norm": 4.443011283874512, "learning_rate": 1.54e-05, "loss": 0.4984, "step": 78 }, { "epoch": 0.0014462242562929062, "grad_norm": 4.615481376647949, "learning_rate": 1.5600000000000003e-05, "loss": 0.3598, "step": 79 }, { "epoch": 0.0014645308924485126, "grad_norm": 6.703000545501709, "learning_rate": 1.58e-05, "loss": 0.926, "step": 80 }, { "epoch": 0.001482837528604119, "grad_norm": 3.9417295455932617, "learning_rate": 1.6000000000000003e-05, "loss": 0.4523, "step": 81 }, { "epoch": 0.0015011441647597254, "grad_norm": 5.64249849319458, "learning_rate": 1.62e-05, "loss": 0.4319, "step": 82 }, { "epoch": 0.0015194508009153318, "grad_norm": 4.152187347412109, "learning_rate": 1.64e-05, "loss": 0.3675, "step": 83 }, { "epoch": 0.0015377574370709383, "grad_norm": 5.285708427429199, "learning_rate": 1.66e-05, "loss": 0.383, "step": 84 }, { "epoch": 0.0015560640732265446, "grad_norm": 9.757657051086426, "learning_rate": 1.6800000000000002e-05, "loss": 0.8236, "step": 85 }, { "epoch": 0.001574370709382151, "grad_norm": 3.1572093963623047, "learning_rate": 1.7e-05, "loss": 0.556, "step": 86 }, { "epoch": 0.0015926773455377575, "grad_norm": 4.673479080200195, "learning_rate": 1.72e-05, "loss": 0.3168, "step": 87 }, { "epoch": 0.0016109839816933638, "grad_norm": 4.761971950531006, "learning_rate": 1.7400000000000003e-05, "loss": 0.7314, "step": 88 }, { "epoch": 0.0016292906178489702, "grad_norm": 4.868711471557617, "learning_rate": 1.76e-05, "loss": 0.5399, "step": 89 }, { "epoch": 0.0016475972540045767, "grad_norm": 11.782761573791504, "learning_rate": 1.7800000000000002e-05, "loss": 0.7641, "step": 90 }, { "epoch": 0.001665903890160183, "grad_norm": 7.96957540512085, "learning_rate": 1.8e-05, "loss": 0.7079, "step": 91 }, { "epoch": 0.0016842105263157896, "grad_norm": 2.9694902896881104, "learning_rate": 1.8200000000000002e-05, "loss": 0.2075, "step": 92 }, { "epoch": 0.001702517162471396, "grad_norm": 8.391504287719727, "learning_rate": 1.8400000000000003e-05, "loss": 0.7879, "step": 93 }, { "epoch": 0.0017208237986270022, "grad_norm": 5.025172710418701, "learning_rate": 1.86e-05, "loss": 0.7385, "step": 94 }, { "epoch": 0.0017391304347826088, "grad_norm": 9.963128089904785, "learning_rate": 1.88e-05, "loss": 0.8, "step": 95 }, { "epoch": 0.001757437070938215, "grad_norm": 5.368072032928467, "learning_rate": 1.9e-05, "loss": 0.6446, "step": 96 }, { "epoch": 0.0017757437070938214, "grad_norm": 3.758234739303589, "learning_rate": 1.9200000000000003e-05, "loss": 0.4276, "step": 97 }, { "epoch": 0.001794050343249428, "grad_norm": 5.807685852050781, "learning_rate": 1.94e-05, "loss": 0.4856, "step": 98 }, { "epoch": 0.0018123569794050343, "grad_norm": 4.966123104095459, "learning_rate": 1.9600000000000002e-05, "loss": 0.3153, "step": 99 }, { "epoch": 0.0018306636155606408, "grad_norm": 6.432656764984131, "learning_rate": 1.98e-05, "loss": 0.4553, "step": 100 }, { "epoch": 0.0018489702517162471, "grad_norm": 6.9228410720825195, "learning_rate": 2e-05, "loss": 0.375, "step": 101 }, { "epoch": 0.0018672768878718535, "grad_norm": 5.237428665161133, "learning_rate": 1.9999939076577906e-05, "loss": 0.484, "step": 102 }, { "epoch": 0.00188558352402746, "grad_norm": 5.173490524291992, "learning_rate": 1.9999756307053947e-05, "loss": 0.4135, "step": 103 }, { "epoch": 0.0019038901601830663, "grad_norm": 4.331842422485352, "learning_rate": 1.9999451693655125e-05, "loss": 0.3857, "step": 104 }, { "epoch": 0.0019221967963386727, "grad_norm": 6.129421710968018, "learning_rate": 1.9999025240093045e-05, "loss": 0.5126, "step": 105 }, { "epoch": 0.0019405034324942792, "grad_norm": 4.619616985321045, "learning_rate": 1.9998476951563914e-05, "loss": 0.477, "step": 106 }, { "epoch": 0.0019588100686498858, "grad_norm": 3.936915397644043, "learning_rate": 1.9997806834748455e-05, "loss": 0.3516, "step": 107 }, { "epoch": 0.001977116704805492, "grad_norm": 8.568585395812988, "learning_rate": 1.9997014897811834e-05, "loss": 0.7743, "step": 108 }, { "epoch": 0.0019954233409610984, "grad_norm": 4.892576694488525, "learning_rate": 1.9996101150403543e-05, "loss": 0.4471, "step": 109 }, { "epoch": 0.002013729977116705, "grad_norm": 10.018097877502441, "learning_rate": 1.9995065603657317e-05, "loss": 1.1247, "step": 110 }, { "epoch": 0.002032036613272311, "grad_norm": 5.875444412231445, "learning_rate": 1.999390827019096e-05, "loss": 0.7445, "step": 111 }, { "epoch": 0.0020503432494279176, "grad_norm": 7.044773101806641, "learning_rate": 1.999262916410621e-05, "loss": 0.4433, "step": 112 }, { "epoch": 0.002068649885583524, "grad_norm": 4.854395866394043, "learning_rate": 1.9991228300988586e-05, "loss": 0.4177, "step": 113 }, { "epoch": 0.0020869565217391303, "grad_norm": 3.947049379348755, "learning_rate": 1.998970569790715e-05, "loss": 0.3204, "step": 114 }, { "epoch": 0.002105263157894737, "grad_norm": 2.9266178607940674, "learning_rate": 1.9988061373414342e-05, "loss": 0.3421, "step": 115 }, { "epoch": 0.0021235697940503433, "grad_norm": 14.068440437316895, "learning_rate": 1.9986295347545738e-05, "loss": 1.2803, "step": 116 }, { "epoch": 0.0021418764302059494, "grad_norm": 2.746870279312134, "learning_rate": 1.9984407641819812e-05, "loss": 0.1913, "step": 117 }, { "epoch": 0.002160183066361556, "grad_norm": 5.603977680206299, "learning_rate": 1.9982398279237657e-05, "loss": 0.535, "step": 118 }, { "epoch": 0.0021784897025171625, "grad_norm": 4.298122882843018, "learning_rate": 1.9980267284282718e-05, "loss": 0.3521, "step": 119 }, { "epoch": 0.002196796338672769, "grad_norm": 3.4252305030822754, "learning_rate": 1.9978014682920503e-05, "loss": 0.3127, "step": 120 }, { "epoch": 0.002215102974828375, "grad_norm": 4.943688869476318, "learning_rate": 1.9975640502598243e-05, "loss": 0.2819, "step": 121 }, { "epoch": 0.0022334096109839817, "grad_norm": 6.345489978790283, "learning_rate": 1.997314477224458e-05, "loss": 0.8686, "step": 122 }, { "epoch": 0.0022517162471395883, "grad_norm": 6.48477840423584, "learning_rate": 1.9970527522269204e-05, "loss": 0.4396, "step": 123 }, { "epoch": 0.0022700228832951944, "grad_norm": 12.649872779846191, "learning_rate": 1.9967788784562474e-05, "loss": 0.4447, "step": 124 }, { "epoch": 0.002288329519450801, "grad_norm": 3.368851900100708, "learning_rate": 1.9964928592495046e-05, "loss": 0.213, "step": 125 }, { "epoch": 0.0023066361556064075, "grad_norm": 5.273118019104004, "learning_rate": 1.9961946980917457e-05, "loss": 0.6374, "step": 126 }, { "epoch": 0.0023249427917620136, "grad_norm": 8.816353797912598, "learning_rate": 1.9958843986159705e-05, "loss": 0.8738, "step": 127 }, { "epoch": 0.00234324942791762, "grad_norm": 7.0079345703125, "learning_rate": 1.99556196460308e-05, "loss": 0.7057, "step": 128 }, { "epoch": 0.0023615560640732267, "grad_norm": 2.5678298473358154, "learning_rate": 1.9952273999818312e-05, "loss": 0.182, "step": 129 }, { "epoch": 0.0023798627002288328, "grad_norm": 5.133856773376465, "learning_rate": 1.9948807088287884e-05, "loss": 0.3598, "step": 130 }, { "epoch": 0.0023981693363844393, "grad_norm": 9.948471069335938, "learning_rate": 1.9945218953682736e-05, "loss": 0.4483, "step": 131 }, { "epoch": 0.002416475972540046, "grad_norm": 5.9176459312438965, "learning_rate": 1.9941509639723155e-05, "loss": 0.8298, "step": 132 }, { "epoch": 0.002434782608695652, "grad_norm": 2.4215924739837646, "learning_rate": 1.9937679191605964e-05, "loss": 0.1617, "step": 133 }, { "epoch": 0.0024530892448512585, "grad_norm": 11.955750465393066, "learning_rate": 1.9933727656003964e-05, "loss": 0.609, "step": 134 }, { "epoch": 0.002471395881006865, "grad_norm": 5.947329998016357, "learning_rate": 1.992965508106537e-05, "loss": 0.5718, "step": 135 }, { "epoch": 0.0024897025171624716, "grad_norm": 5.794270038604736, "learning_rate": 1.9925461516413224e-05, "loss": 0.5392, "step": 136 }, { "epoch": 0.0025080091533180777, "grad_norm": 5.274944305419922, "learning_rate": 1.9921147013144782e-05, "loss": 0.5278, "step": 137 }, { "epoch": 0.0025263157894736842, "grad_norm": 5.6011962890625, "learning_rate": 1.9916711623830904e-05, "loss": 0.4404, "step": 138 }, { "epoch": 0.0025446224256292908, "grad_norm": 7.330393314361572, "learning_rate": 1.991215540251542e-05, "loss": 0.6554, "step": 139 }, { "epoch": 0.002562929061784897, "grad_norm": 8.020462036132812, "learning_rate": 1.9907478404714438e-05, "loss": 0.7649, "step": 140 }, { "epoch": 0.0025812356979405034, "grad_norm": 3.3086674213409424, "learning_rate": 1.9902680687415704e-05, "loss": 0.3638, "step": 141 }, { "epoch": 0.00259954233409611, "grad_norm": 7.1618781089782715, "learning_rate": 1.989776230907789e-05, "loss": 0.4597, "step": 142 }, { "epoch": 0.002617848970251716, "grad_norm": 5.93783712387085, "learning_rate": 1.9892723329629885e-05, "loss": 0.4452, "step": 143 }, { "epoch": 0.0026361556064073226, "grad_norm": 5.123695373535156, "learning_rate": 1.988756381047006e-05, "loss": 0.5444, "step": 144 }, { "epoch": 0.002654462242562929, "grad_norm": 7.8915181159973145, "learning_rate": 1.988228381446553e-05, "loss": 0.758, "step": 145 }, { "epoch": 0.0026727688787185353, "grad_norm": 7.822751998901367, "learning_rate": 1.9876883405951378e-05, "loss": 0.6956, "step": 146 }, { "epoch": 0.002691075514874142, "grad_norm": 2.519596815109253, "learning_rate": 1.987136265072988e-05, "loss": 0.238, "step": 147 }, { "epoch": 0.0027093821510297484, "grad_norm": 3.658250093460083, "learning_rate": 1.9865721616069695e-05, "loss": 0.3844, "step": 148 }, { "epoch": 0.0027276887871853545, "grad_norm": 5.0784430503845215, "learning_rate": 1.985996037070505e-05, "loss": 0.5564, "step": 149 }, { "epoch": 0.002745995423340961, "grad_norm": 6.504764556884766, "learning_rate": 1.9854078984834904e-05, "loss": 0.4866, "step": 150 }, { "epoch": 0.0027643020594965676, "grad_norm": 4.09471321105957, "learning_rate": 1.9848077530122083e-05, "loss": 0.2998, "step": 151 }, { "epoch": 0.002782608695652174, "grad_norm": 5.1176886558532715, "learning_rate": 1.984195607969242e-05, "loss": 0.2826, "step": 152 }, { "epoch": 0.00280091533180778, "grad_norm": 12.725276947021484, "learning_rate": 1.983571470813386e-05, "loss": 0.7915, "step": 153 }, { "epoch": 0.0028192219679633868, "grad_norm": 7.101748943328857, "learning_rate": 1.9829353491495545e-05, "loss": 0.7455, "step": 154 }, { "epoch": 0.0028375286041189933, "grad_norm": 6.615306854248047, "learning_rate": 1.982287250728689e-05, "loss": 0.8178, "step": 155 }, { "epoch": 0.0028558352402745994, "grad_norm": 6.7726240158081055, "learning_rate": 1.9816271834476642e-05, "loss": 0.3764, "step": 156 }, { "epoch": 0.002874141876430206, "grad_norm": 4.513782978057861, "learning_rate": 1.9809551553491918e-05, "loss": 0.3002, "step": 157 }, { "epoch": 0.0028924485125858125, "grad_norm": 6.409885883331299, "learning_rate": 1.9802711746217222e-05, "loss": 0.4299, "step": 158 }, { "epoch": 0.0029107551487414186, "grad_norm": 1.4180457592010498, "learning_rate": 1.979575249599344e-05, "loss": 0.0436, "step": 159 }, { "epoch": 0.002929061784897025, "grad_norm": 7.368013858795166, "learning_rate": 1.9788673887616852e-05, "loss": 0.4066, "step": 160 }, { "epoch": 0.0029473684210526317, "grad_norm": 3.97849178314209, "learning_rate": 1.9781476007338058e-05, "loss": 0.2171, "step": 161 }, { "epoch": 0.002965675057208238, "grad_norm": 8.33909797668457, "learning_rate": 1.9774158942860962e-05, "loss": 0.5762, "step": 162 }, { "epoch": 0.0029839816933638443, "grad_norm": 1.4704389572143555, "learning_rate": 1.9766722783341682e-05, "loss": 0.1528, "step": 163 }, { "epoch": 0.003002288329519451, "grad_norm": 6.321743965148926, "learning_rate": 1.9759167619387474e-05, "loss": 0.4852, "step": 164 }, { "epoch": 0.0030205949656750574, "grad_norm": 8.366913795471191, "learning_rate": 1.9751493543055634e-05, "loss": 0.4964, "step": 165 }, { "epoch": 0.0030389016018306635, "grad_norm": 6.21965217590332, "learning_rate": 1.9743700647852356e-05, "loss": 0.527, "step": 166 }, { "epoch": 0.00305720823798627, "grad_norm": 9.058350563049316, "learning_rate": 1.9735789028731603e-05, "loss": 0.9455, "step": 167 }, { "epoch": 0.0030755148741418766, "grad_norm": 7.52754020690918, "learning_rate": 1.972775878209397e-05, "loss": 0.618, "step": 168 }, { "epoch": 0.0030938215102974827, "grad_norm": 8.560585975646973, "learning_rate": 1.9719610005785466e-05, "loss": 0.476, "step": 169 }, { "epoch": 0.0031121281464530893, "grad_norm": 7.208701133728027, "learning_rate": 1.971134279909636e-05, "loss": 0.7781, "step": 170 }, { "epoch": 0.003130434782608696, "grad_norm": 5.757314205169678, "learning_rate": 1.9702957262759964e-05, "loss": 0.518, "step": 171 }, { "epoch": 0.003148741418764302, "grad_norm": 2.8194165229797363, "learning_rate": 1.9694453498951392e-05, "loss": 0.2128, "step": 172 }, { "epoch": 0.0031670480549199085, "grad_norm": 3.0178167819976807, "learning_rate": 1.9685831611286312e-05, "loss": 0.3055, "step": 173 }, { "epoch": 0.003185354691075515, "grad_norm": 1.6432427167892456, "learning_rate": 1.9677091704819714e-05, "loss": 0.216, "step": 174 }, { "epoch": 0.003203661327231121, "grad_norm": 8.396689414978027, "learning_rate": 1.9668233886044597e-05, "loss": 0.6197, "step": 175 }, { "epoch": 0.0032219679633867277, "grad_norm": 7.577538967132568, "learning_rate": 1.9659258262890683e-05, "loss": 0.7145, "step": 176 }, { "epoch": 0.003240274599542334, "grad_norm": 4.660325527191162, "learning_rate": 1.9650164944723116e-05, "loss": 0.3487, "step": 177 }, { "epoch": 0.0032585812356979403, "grad_norm": 2.1696557998657227, "learning_rate": 1.96409540423411e-05, "loss": 0.3491, "step": 178 }, { "epoch": 0.003276887871853547, "grad_norm": 3.31132435798645, "learning_rate": 1.9631625667976584e-05, "loss": 0.2511, "step": 179 }, { "epoch": 0.0032951945080091534, "grad_norm": 5.26754903793335, "learning_rate": 1.9622179935292855e-05, "loss": 0.4235, "step": 180 }, { "epoch": 0.00331350114416476, "grad_norm": 4.9605865478515625, "learning_rate": 1.961261695938319e-05, "loss": 0.2065, "step": 181 }, { "epoch": 0.003331807780320366, "grad_norm": 5.259721755981445, "learning_rate": 1.9602936856769432e-05, "loss": 0.3804, "step": 182 }, { "epoch": 0.0033501144164759726, "grad_norm": 3.692326784133911, "learning_rate": 1.9593139745400575e-05, "loss": 0.3388, "step": 183 }, { "epoch": 0.003368421052631579, "grad_norm": 4.228832721710205, "learning_rate": 1.9583225744651334e-05, "loss": 0.2778, "step": 184 }, { "epoch": 0.0033867276887871852, "grad_norm": 8.879986763000488, "learning_rate": 1.9573194975320672e-05, "loss": 0.7031, "step": 185 }, { "epoch": 0.003405034324942792, "grad_norm": 7.6145853996276855, "learning_rate": 1.9563047559630356e-05, "loss": 0.6261, "step": 186 }, { "epoch": 0.0034233409610983983, "grad_norm": 10.076028823852539, "learning_rate": 1.9552783621223437e-05, "loss": 0.6617, "step": 187 }, { "epoch": 0.0034416475972540044, "grad_norm": 4.281607151031494, "learning_rate": 1.954240328516277e-05, "loss": 0.2965, "step": 188 }, { "epoch": 0.003459954233409611, "grad_norm": 3.990032911300659, "learning_rate": 1.9531906677929472e-05, "loss": 0.2232, "step": 189 }, { "epoch": 0.0034782608695652175, "grad_norm": 10.36355209350586, "learning_rate": 1.9521293927421388e-05, "loss": 0.3606, "step": 190 }, { "epoch": 0.0034965675057208236, "grad_norm": 9.396668434143066, "learning_rate": 1.9510565162951538e-05, "loss": 0.6809, "step": 191 }, { "epoch": 0.00351487414187643, "grad_norm": 4.969096660614014, "learning_rate": 1.9499720515246524e-05, "loss": 0.3711, "step": 192 }, { "epoch": 0.0035331807780320367, "grad_norm": 2.855027914047241, "learning_rate": 1.9488760116444966e-05, "loss": 0.3334, "step": 193 }, { "epoch": 0.003551487414187643, "grad_norm": 9.800653457641602, "learning_rate": 1.947768410009586e-05, "loss": 0.8046, "step": 194 }, { "epoch": 0.0035697940503432494, "grad_norm": 6.4518632888793945, "learning_rate": 1.9466492601156964e-05, "loss": 0.7623, "step": 195 }, { "epoch": 0.003588100686498856, "grad_norm": 9.151044845581055, "learning_rate": 1.945518575599317e-05, "loss": 0.4479, "step": 196 }, { "epoch": 0.0036064073226544625, "grad_norm": 6.074557304382324, "learning_rate": 1.944376370237481e-05, "loss": 0.3635, "step": 197 }, { "epoch": 0.0036247139588100686, "grad_norm": 6.840311050415039, "learning_rate": 1.943222657947601e-05, "loss": 0.3051, "step": 198 }, { "epoch": 0.003643020594965675, "grad_norm": 3.7585532665252686, "learning_rate": 1.942057452787297e-05, "loss": 0.3266, "step": 199 }, { "epoch": 0.0036613272311212816, "grad_norm": 4.411327838897705, "learning_rate": 1.9408807689542257e-05, "loss": 0.2551, "step": 200 }, { "epoch": 0.0036796338672768878, "grad_norm": 3.8587820529937744, "learning_rate": 1.9396926207859085e-05, "loss": 0.3105, "step": 201 }, { "epoch": 0.0036979405034324943, "grad_norm": 4.376249313354492, "learning_rate": 1.938493022759556e-05, "loss": 0.2243, "step": 202 }, { "epoch": 0.003716247139588101, "grad_norm": 5.54730224609375, "learning_rate": 1.937281989491892e-05, "loss": 0.4158, "step": 203 }, { "epoch": 0.003734553775743707, "grad_norm": 2.6035118103027344, "learning_rate": 1.9360595357389735e-05, "loss": 0.1788, "step": 204 }, { "epoch": 0.0037528604118993135, "grad_norm": 10.728706359863281, "learning_rate": 1.9348256763960146e-05, "loss": 1.0367, "step": 205 }, { "epoch": 0.00377116704805492, "grad_norm": 5.158708572387695, "learning_rate": 1.9335804264972018e-05, "loss": 0.4394, "step": 206 }, { "epoch": 0.003789473684210526, "grad_norm": 9.297528266906738, "learning_rate": 1.9323238012155125e-05, "loss": 0.4115, "step": 207 }, { "epoch": 0.0038077803203661327, "grad_norm": 11.042555809020996, "learning_rate": 1.9310558158625286e-05, "loss": 0.7565, "step": 208 }, { "epoch": 0.0038260869565217392, "grad_norm": 6.377527236938477, "learning_rate": 1.9297764858882516e-05, "loss": 0.5614, "step": 209 }, { "epoch": 0.0038443935926773453, "grad_norm": 3.4690988063812256, "learning_rate": 1.9284858268809135e-05, "loss": 0.1462, "step": 210 }, { "epoch": 0.003862700228832952, "grad_norm": 8.21645450592041, "learning_rate": 1.9271838545667876e-05, "loss": 0.6482, "step": 211 }, { "epoch": 0.0038810068649885584, "grad_norm": 13.70019817352295, "learning_rate": 1.925870584809995e-05, "loss": 1.3146, "step": 212 }, { "epoch": 0.003899313501144165, "grad_norm": 6.1762375831604, "learning_rate": 1.9245460336123136e-05, "loss": 0.2371, "step": 213 }, { "epoch": 0.0039176201372997715, "grad_norm": 3.4190640449523926, "learning_rate": 1.923210217112981e-05, "loss": 0.1756, "step": 214 }, { "epoch": 0.003935926773455377, "grad_norm": 4.4201250076293945, "learning_rate": 1.9218631515885007e-05, "loss": 0.3832, "step": 215 }, { "epoch": 0.003954233409610984, "grad_norm": 4.916990756988525, "learning_rate": 1.9205048534524405e-05, "loss": 0.55, "step": 216 }, { "epoch": 0.00397254004576659, "grad_norm": 5.253897666931152, "learning_rate": 1.9191353392552346e-05, "loss": 0.4081, "step": 217 }, { "epoch": 0.003990846681922197, "grad_norm": 9.138065338134766, "learning_rate": 1.9177546256839814e-05, "loss": 0.4096, "step": 218 }, { "epoch": 0.004009153318077803, "grad_norm": 5.5543437004089355, "learning_rate": 1.9163627295622397e-05, "loss": 0.3725, "step": 219 }, { "epoch": 0.00402745995423341, "grad_norm": 6.35120153427124, "learning_rate": 1.914959667849825e-05, "loss": 0.6631, "step": 220 }, { "epoch": 0.0040457665903890164, "grad_norm": 6.352338790893555, "learning_rate": 1.913545457642601e-05, "loss": 0.6608, "step": 221 }, { "epoch": 0.004064073226544622, "grad_norm": 5.439535140991211, "learning_rate": 1.9121201161722732e-05, "loss": 0.3461, "step": 222 }, { "epoch": 0.004082379862700229, "grad_norm": 6.05610990524292, "learning_rate": 1.910683660806177e-05, "loss": 0.7047, "step": 223 }, { "epoch": 0.004100686498855835, "grad_norm": 16.13218879699707, "learning_rate": 1.9092361090470688e-05, "loss": 0.9417, "step": 224 }, { "epoch": 0.004118993135011442, "grad_norm": 11.939602851867676, "learning_rate": 1.907777478532909e-05, "loss": 0.4711, "step": 225 }, { "epoch": 0.004137299771167048, "grad_norm": 9.328332901000977, "learning_rate": 1.9063077870366504e-05, "loss": 0.4825, "step": 226 }, { "epoch": 0.004155606407322655, "grad_norm": 3.4537312984466553, "learning_rate": 1.9048270524660197e-05, "loss": 0.2696, "step": 227 }, { "epoch": 0.0041739130434782605, "grad_norm": 8.465746879577637, "learning_rate": 1.903335292863301e-05, "loss": 0.3368, "step": 228 }, { "epoch": 0.004192219679633867, "grad_norm": 2.80598783493042, "learning_rate": 1.901832526405114e-05, "loss": 0.2279, "step": 229 }, { "epoch": 0.004210526315789474, "grad_norm": 9.187653541564941, "learning_rate": 1.9003187714021936e-05, "loss": 0.5421, "step": 230 }, { "epoch": 0.00422883295194508, "grad_norm": 2.9173386096954346, "learning_rate": 1.8987940462991673e-05, "loss": 0.1865, "step": 231 }, { "epoch": 0.004247139588100687, "grad_norm": 8.808391571044922, "learning_rate": 1.8972583696743284e-05, "loss": 0.6127, "step": 232 }, { "epoch": 0.004265446224256293, "grad_norm": 3.902742862701416, "learning_rate": 1.895711760239413e-05, "loss": 0.4918, "step": 233 }, { "epoch": 0.004283752860411899, "grad_norm": 7.3988237380981445, "learning_rate": 1.8941542368393683e-05, "loss": 0.4754, "step": 234 }, { "epoch": 0.0043020594965675054, "grad_norm": 3.434345006942749, "learning_rate": 1.892585818452126e-05, "loss": 0.2698, "step": 235 }, { "epoch": 0.004320366132723112, "grad_norm": 5.183997631072998, "learning_rate": 1.891006524188368e-05, "loss": 0.2622, "step": 236 }, { "epoch": 0.0043386727688787185, "grad_norm": 4.6859235763549805, "learning_rate": 1.889416373291298e-05, "loss": 0.2881, "step": 237 }, { "epoch": 0.004356979405034325, "grad_norm": 3.172283887863159, "learning_rate": 1.8878153851364013e-05, "loss": 0.2605, "step": 238 }, { "epoch": 0.004375286041189932, "grad_norm": 11.702736854553223, "learning_rate": 1.8862035792312148e-05, "loss": 0.6882, "step": 239 }, { "epoch": 0.004393592677345538, "grad_norm": 6.005770683288574, "learning_rate": 1.884580975215084e-05, "loss": 0.4795, "step": 240 }, { "epoch": 0.004411899313501144, "grad_norm": 3.117210865020752, "learning_rate": 1.8829475928589272e-05, "loss": 0.3978, "step": 241 }, { "epoch": 0.00443020594965675, "grad_norm": 6.215647220611572, "learning_rate": 1.8813034520649923e-05, "loss": 0.4373, "step": 242 }, { "epoch": 0.004448512585812357, "grad_norm": 5.335672378540039, "learning_rate": 1.879648572866617e-05, "loss": 0.2903, "step": 243 }, { "epoch": 0.0044668192219679635, "grad_norm": 1.4431347846984863, "learning_rate": 1.8779829754279806e-05, "loss": 0.1387, "step": 244 }, { "epoch": 0.00448512585812357, "grad_norm": 8.997931480407715, "learning_rate": 1.8763066800438638e-05, "loss": 0.7452, "step": 245 }, { "epoch": 0.0045034324942791765, "grad_norm": 7.0658721923828125, "learning_rate": 1.874619707139396e-05, "loss": 0.4082, "step": 246 }, { "epoch": 0.004521739130434782, "grad_norm": 16.99602508544922, "learning_rate": 1.8729220772698096e-05, "loss": 0.5462, "step": 247 }, { "epoch": 0.004540045766590389, "grad_norm": 4.291457653045654, "learning_rate": 1.8712138111201898e-05, "loss": 0.2849, "step": 248 }, { "epoch": 0.004558352402745995, "grad_norm": 1.2436845302581787, "learning_rate": 1.869494929505219e-05, "loss": 0.1299, "step": 249 }, { "epoch": 0.004576659038901602, "grad_norm": 13.032325744628906, "learning_rate": 1.8677654533689287e-05, "loss": 1.1319, "step": 250 }, { "epoch": 0.004594965675057208, "grad_norm": 4.819525718688965, "learning_rate": 1.866025403784439e-05, "loss": 0.2317, "step": 251 }, { "epoch": 0.004613272311212815, "grad_norm": 4.058289051055908, "learning_rate": 1.864274801953705e-05, "loss": 0.1868, "step": 252 }, { "epoch": 0.0046315789473684215, "grad_norm": 5.123549461364746, "learning_rate": 1.8625136692072577e-05, "loss": 0.2682, "step": 253 }, { "epoch": 0.004649885583524027, "grad_norm": 4.105663299560547, "learning_rate": 1.860742027003944e-05, "loss": 0.2852, "step": 254 }, { "epoch": 0.004668192219679634, "grad_norm": 8.609756469726562, "learning_rate": 1.8589598969306646e-05, "loss": 0.638, "step": 255 }, { "epoch": 0.00468649885583524, "grad_norm": 10.366744041442871, "learning_rate": 1.8571673007021124e-05, "loss": 0.7038, "step": 256 }, { "epoch": 0.004704805491990847, "grad_norm": 3.733368158340454, "learning_rate": 1.855364260160507e-05, "loss": 0.3285, "step": 257 }, { "epoch": 0.004723112128146453, "grad_norm": 8.309182167053223, "learning_rate": 1.8535507972753275e-05, "loss": 0.9593, "step": 258 }, { "epoch": 0.00474141876430206, "grad_norm": 9.302151679992676, "learning_rate": 1.851726934143048e-05, "loss": 0.5591, "step": 259 }, { "epoch": 0.0047597254004576655, "grad_norm": 1.1159979104995728, "learning_rate": 1.849892692986864e-05, "loss": 0.1096, "step": 260 }, { "epoch": 0.004778032036613272, "grad_norm": 5.174903392791748, "learning_rate": 1.848048096156426e-05, "loss": 0.2641, "step": 261 }, { "epoch": 0.004796338672768879, "grad_norm": 4.628619194030762, "learning_rate": 1.8461931661275642e-05, "loss": 0.4448, "step": 262 }, { "epoch": 0.004814645308924485, "grad_norm": 2.9997923374176025, "learning_rate": 1.8443279255020153e-05, "loss": 0.2183, "step": 263 }, { "epoch": 0.004832951945080092, "grad_norm": 9.500200271606445, "learning_rate": 1.842452397007148e-05, "loss": 0.4821, "step": 264 }, { "epoch": 0.004851258581235698, "grad_norm": 7.443029403686523, "learning_rate": 1.8405666034956842e-05, "loss": 0.6391, "step": 265 }, { "epoch": 0.004869565217391304, "grad_norm": 13.987396240234375, "learning_rate": 1.8386705679454243e-05, "loss": 0.8478, "step": 266 }, { "epoch": 0.0048878718535469105, "grad_norm": 3.692591667175293, "learning_rate": 1.836764313458962e-05, "loss": 0.2808, "step": 267 }, { "epoch": 0.004906178489702517, "grad_norm": 6.317406177520752, "learning_rate": 1.8348478632634067e-05, "loss": 0.3462, "step": 268 }, { "epoch": 0.0049244851258581235, "grad_norm": 10.023675918579102, "learning_rate": 1.8329212407100996e-05, "loss": 1.0451, "step": 269 }, { "epoch": 0.00494279176201373, "grad_norm": 5.884584426879883, "learning_rate": 1.8309844692743283e-05, "loss": 0.5844, "step": 270 }, { "epoch": 0.004961098398169337, "grad_norm": 2.740469217300415, "learning_rate": 1.8290375725550417e-05, "loss": 0.1969, "step": 271 }, { "epoch": 0.004979405034324943, "grad_norm": 6.099292755126953, "learning_rate": 1.827080574274562e-05, "loss": 0.726, "step": 272 }, { "epoch": 0.004997711670480549, "grad_norm": 6.810953140258789, "learning_rate": 1.8251134982782952e-05, "loss": 0.5525, "step": 273 }, { "epoch": 0.005016018306636155, "grad_norm": 17.33843231201172, "learning_rate": 1.8231363685344422e-05, "loss": 0.9222, "step": 274 }, { "epoch": 0.005034324942791762, "grad_norm": 6.817041397094727, "learning_rate": 1.821149209133704e-05, "loss": 0.4998, "step": 275 }, { "epoch": 0.0050526315789473685, "grad_norm": 9.165572166442871, "learning_rate": 1.819152044288992e-05, "loss": 0.6976, "step": 276 }, { "epoch": 0.005070938215102975, "grad_norm": 4.791163444519043, "learning_rate": 1.8171448983351284e-05, "loss": 0.4825, "step": 277 }, { "epoch": 0.0050892448512585816, "grad_norm": 2.2928450107574463, "learning_rate": 1.815127795728554e-05, "loss": 0.1951, "step": 278 }, { "epoch": 0.005107551487414187, "grad_norm": 4.821455955505371, "learning_rate": 1.8131007610470278e-05, "loss": 0.2852, "step": 279 }, { "epoch": 0.005125858123569794, "grad_norm": 6.22464656829834, "learning_rate": 1.8110638189893267e-05, "loss": 0.3738, "step": 280 }, { "epoch": 0.0051441647597254, "grad_norm": 4.849123001098633, "learning_rate": 1.8090169943749477e-05, "loss": 0.296, "step": 281 }, { "epoch": 0.005162471395881007, "grad_norm": 6.145839691162109, "learning_rate": 1.806960312143802e-05, "loss": 0.3723, "step": 282 }, { "epoch": 0.005180778032036613, "grad_norm": 4.0611724853515625, "learning_rate": 1.804893797355914e-05, "loss": 0.4265, "step": 283 }, { "epoch": 0.00519908466819222, "grad_norm": 3.252166271209717, "learning_rate": 1.8028174751911147e-05, "loss": 0.3467, "step": 284 }, { "epoch": 0.0052173913043478265, "grad_norm": 6.532531261444092, "learning_rate": 1.8007313709487334e-05, "loss": 0.3968, "step": 285 }, { "epoch": 0.005235697940503432, "grad_norm": 8.545121192932129, "learning_rate": 1.798635510047293e-05, "loss": 0.5772, "step": 286 }, { "epoch": 0.005254004576659039, "grad_norm": 3.646812677383423, "learning_rate": 1.7965299180241963e-05, "loss": 0.1868, "step": 287 }, { "epoch": 0.005272311212814645, "grad_norm": 6.056759357452393, "learning_rate": 1.7944146205354182e-05, "loss": 0.615, "step": 288 }, { "epoch": 0.005290617848970252, "grad_norm": 9.61752986907959, "learning_rate": 1.792289643355191e-05, "loss": 0.6434, "step": 289 }, { "epoch": 0.005308924485125858, "grad_norm": 7.096751689910889, "learning_rate": 1.7901550123756906e-05, "loss": 0.3842, "step": 290 }, { "epoch": 0.005327231121281465, "grad_norm": 3.8110721111297607, "learning_rate": 1.788010753606722e-05, "loss": 0.2328, "step": 291 }, { "epoch": 0.0053455377574370706, "grad_norm": 5.858436584472656, "learning_rate": 1.785856893175402e-05, "loss": 0.431, "step": 292 }, { "epoch": 0.005363844393592677, "grad_norm": 4.623099327087402, "learning_rate": 1.78369345732584e-05, "loss": 0.337, "step": 293 }, { "epoch": 0.005382151029748284, "grad_norm": 5.802640914916992, "learning_rate": 1.781520472418819e-05, "loss": 0.4879, "step": 294 }, { "epoch": 0.00540045766590389, "grad_norm": 5.544589042663574, "learning_rate": 1.7793379649314743e-05, "loss": 0.4911, "step": 295 }, { "epoch": 0.005418764302059497, "grad_norm": 6.768428802490234, "learning_rate": 1.777145961456971e-05, "loss": 0.5283, "step": 296 }, { "epoch": 0.005437070938215103, "grad_norm": 4.4601731300354, "learning_rate": 1.7749444887041797e-05, "loss": 0.4052, "step": 297 }, { "epoch": 0.005455377574370709, "grad_norm": 16.009952545166016, "learning_rate": 1.7727335734973512e-05, "loss": 0.4814, "step": 298 }, { "epoch": 0.0054736842105263155, "grad_norm": 10.445368766784668, "learning_rate": 1.7705132427757895e-05, "loss": 0.8704, "step": 299 }, { "epoch": 0.005491990846681922, "grad_norm": 3.6240663528442383, "learning_rate": 1.7682835235935236e-05, "loss": 0.1805, "step": 300 }, { "epoch": 0.005510297482837529, "grad_norm": 6.040288925170898, "learning_rate": 1.766044443118978e-05, "loss": 0.536, "step": 301 }, { "epoch": 0.005528604118993135, "grad_norm": 5.494821548461914, "learning_rate": 1.7637960286346423e-05, "loss": 0.3142, "step": 302 }, { "epoch": 0.005546910755148742, "grad_norm": 4.8304219245910645, "learning_rate": 1.761538307536737e-05, "loss": 0.444, "step": 303 }, { "epoch": 0.005565217391304348, "grad_norm": 8.82863712310791, "learning_rate": 1.759271307334881e-05, "loss": 0.7043, "step": 304 }, { "epoch": 0.005583524027459954, "grad_norm": 11.274286270141602, "learning_rate": 1.7569950556517566e-05, "loss": 1.0306, "step": 305 }, { "epoch": 0.00560183066361556, "grad_norm": 7.698188304901123, "learning_rate": 1.7547095802227723e-05, "loss": 0.5291, "step": 306 }, { "epoch": 0.005620137299771167, "grad_norm": 6.544892311096191, "learning_rate": 1.7524149088957244e-05, "loss": 0.679, "step": 307 }, { "epoch": 0.0056384439359267735, "grad_norm": 13.220139503479004, "learning_rate": 1.7501110696304598e-05, "loss": 0.7442, "step": 308 }, { "epoch": 0.00565675057208238, "grad_norm": 4.368415355682373, "learning_rate": 1.747798090498532e-05, "loss": 0.3106, "step": 309 }, { "epoch": 0.005675057208237987, "grad_norm": 5.565270900726318, "learning_rate": 1.7454759996828622e-05, "loss": 0.2814, "step": 310 }, { "epoch": 0.005693363844393592, "grad_norm": 6.844931125640869, "learning_rate": 1.7431448254773943e-05, "loss": 0.5184, "step": 311 }, { "epoch": 0.005711670480549199, "grad_norm": 4.695760726928711, "learning_rate": 1.74080459628675e-05, "loss": 0.463, "step": 312 }, { "epoch": 0.005729977116704805, "grad_norm": 7.724005222320557, "learning_rate": 1.7384553406258842e-05, "loss": 0.468, "step": 313 }, { "epoch": 0.005748283752860412, "grad_norm": 3.6156113147735596, "learning_rate": 1.7360970871197347e-05, "loss": 0.3026, "step": 314 }, { "epoch": 0.0057665903890160184, "grad_norm": 1.1042062044143677, "learning_rate": 1.7337298645028764e-05, "loss": 0.0863, "step": 315 }, { "epoch": 0.005784897025171625, "grad_norm": 3.604750633239746, "learning_rate": 1.7313537016191706e-05, "loss": 0.1948, "step": 316 }, { "epoch": 0.0058032036613272315, "grad_norm": 3.3758010864257812, "learning_rate": 1.7289686274214116e-05, "loss": 0.2802, "step": 317 }, { "epoch": 0.005821510297482837, "grad_norm": 12.799773216247559, "learning_rate": 1.7265746709709762e-05, "loss": 0.9552, "step": 318 }, { "epoch": 0.005839816933638444, "grad_norm": 2.934678077697754, "learning_rate": 1.7241718614374678e-05, "loss": 0.2117, "step": 319 }, { "epoch": 0.00585812356979405, "grad_norm": 2.4731688499450684, "learning_rate": 1.7217602280983622e-05, "loss": 0.1622, "step": 320 }, { "epoch": 0.005876430205949657, "grad_norm": 2.6618807315826416, "learning_rate": 1.7193398003386514e-05, "loss": 0.211, "step": 321 }, { "epoch": 0.005894736842105263, "grad_norm": 5.7936906814575195, "learning_rate": 1.716910607650483e-05, "loss": 0.6325, "step": 322 }, { "epoch": 0.00591304347826087, "grad_norm": 5.763343334197998, "learning_rate": 1.7144726796328034e-05, "loss": 0.6117, "step": 323 }, { "epoch": 0.005931350114416476, "grad_norm": 6.462403774261475, "learning_rate": 1.712026045990997e-05, "loss": 0.3717, "step": 324 }, { "epoch": 0.005949656750572082, "grad_norm": 2.8275930881500244, "learning_rate": 1.709570736536521e-05, "loss": 0.1894, "step": 325 }, { "epoch": 0.005967963386727689, "grad_norm": 7.836248874664307, "learning_rate": 1.7071067811865477e-05, "loss": 0.4134, "step": 326 }, { "epoch": 0.005986270022883295, "grad_norm": 4.337584972381592, "learning_rate": 1.7046342099635948e-05, "loss": 0.3016, "step": 327 }, { "epoch": 0.006004576659038902, "grad_norm": 4.750609874725342, "learning_rate": 1.7021530529951627e-05, "loss": 0.7295, "step": 328 }, { "epoch": 0.006022883295194508, "grad_norm": 4.368948459625244, "learning_rate": 1.6996633405133656e-05, "loss": 0.4981, "step": 329 }, { "epoch": 0.006041189931350115, "grad_norm": 5.648835182189941, "learning_rate": 1.697165102854565e-05, "loss": 0.3509, "step": 330 }, { "epoch": 0.0060594965675057205, "grad_norm": 6.33278226852417, "learning_rate": 1.6946583704589973e-05, "loss": 0.553, "step": 331 }, { "epoch": 0.006077803203661327, "grad_norm": 3.5381555557250977, "learning_rate": 1.692143173870407e-05, "loss": 0.3945, "step": 332 }, { "epoch": 0.006096109839816934, "grad_norm": 6.314048767089844, "learning_rate": 1.68961954373567e-05, "loss": 0.4074, "step": 333 }, { "epoch": 0.00611441647597254, "grad_norm": 10.963640213012695, "learning_rate": 1.6870875108044233e-05, "loss": 0.4685, "step": 334 }, { "epoch": 0.006132723112128147, "grad_norm": 4.838613510131836, "learning_rate": 1.684547105928689e-05, "loss": 0.3112, "step": 335 }, { "epoch": 0.006151029748283753, "grad_norm": 6.644429683685303, "learning_rate": 1.6819983600624986e-05, "loss": 0.4721, "step": 336 }, { "epoch": 0.006169336384439359, "grad_norm": 7.574588775634766, "learning_rate": 1.6794413042615168e-05, "loss": 0.4311, "step": 337 }, { "epoch": 0.0061876430205949655, "grad_norm": 16.03948402404785, "learning_rate": 1.6768759696826608e-05, "loss": 1.1899, "step": 338 }, { "epoch": 0.006205949656750572, "grad_norm": 8.45483112335205, "learning_rate": 1.6743023875837233e-05, "loss": 0.4542, "step": 339 }, { "epoch": 0.0062242562929061785, "grad_norm": 5.345456600189209, "learning_rate": 1.6717205893229904e-05, "loss": 0.514, "step": 340 }, { "epoch": 0.006242562929061785, "grad_norm": 6.069270610809326, "learning_rate": 1.6691306063588583e-05, "loss": 0.675, "step": 341 }, { "epoch": 0.006260869565217392, "grad_norm": 8.289877891540527, "learning_rate": 1.6665324702494524e-05, "loss": 0.4616, "step": 342 }, { "epoch": 0.006279176201372997, "grad_norm": 8.072705268859863, "learning_rate": 1.6639262126522417e-05, "loss": 0.4349, "step": 343 }, { "epoch": 0.006297482837528604, "grad_norm": 8.641850471496582, "learning_rate": 1.661311865323652e-05, "loss": 0.4676, "step": 344 }, { "epoch": 0.00631578947368421, "grad_norm": 4.457225799560547, "learning_rate": 1.6586894601186804e-05, "loss": 0.4408, "step": 345 }, { "epoch": 0.006334096109839817, "grad_norm": 4.834903240203857, "learning_rate": 1.6560590289905074e-05, "loss": 0.3358, "step": 346 }, { "epoch": 0.0063524027459954235, "grad_norm": 11.663195610046387, "learning_rate": 1.6534206039901057e-05, "loss": 0.6072, "step": 347 }, { "epoch": 0.00637070938215103, "grad_norm": 3.625476360321045, "learning_rate": 1.650774217265851e-05, "loss": 0.2633, "step": 348 }, { "epoch": 0.0063890160183066366, "grad_norm": 6.559308052062988, "learning_rate": 1.6481199010631312e-05, "loss": 0.2867, "step": 349 }, { "epoch": 0.006407322654462242, "grad_norm": 11.27363395690918, "learning_rate": 1.645457687723951e-05, "loss": 0.8009, "step": 350 }, { "epoch": 0.006425629290617849, "grad_norm": 6.976966381072998, "learning_rate": 1.6427876096865394e-05, "loss": 0.5221, "step": 351 }, { "epoch": 0.006443935926773455, "grad_norm": 5.93279504776001, "learning_rate": 1.6401096994849558e-05, "loss": 0.3511, "step": 352 }, { "epoch": 0.006462242562929062, "grad_norm": 5.895767688751221, "learning_rate": 1.63742398974869e-05, "loss": 0.3957, "step": 353 }, { "epoch": 0.006480549199084668, "grad_norm": 4.896230220794678, "learning_rate": 1.6347305132022677e-05, "loss": 0.4095, "step": 354 }, { "epoch": 0.006498855835240275, "grad_norm": 4.811487674713135, "learning_rate": 1.632029302664851e-05, "loss": 0.3359, "step": 355 }, { "epoch": 0.006517162471395881, "grad_norm": 6.440311908721924, "learning_rate": 1.6293203910498375e-05, "loss": 0.4722, "step": 356 }, { "epoch": 0.006535469107551487, "grad_norm": 4.4008965492248535, "learning_rate": 1.6266038113644605e-05, "loss": 0.325, "step": 357 }, { "epoch": 0.006553775743707094, "grad_norm": 5.9258713722229, "learning_rate": 1.6238795967093865e-05, "loss": 0.4317, "step": 358 }, { "epoch": 0.0065720823798627, "grad_norm": 9.05823040008545, "learning_rate": 1.6211477802783105e-05, "loss": 0.6237, "step": 359 }, { "epoch": 0.006590389016018307, "grad_norm": 12.546273231506348, "learning_rate": 1.6184083953575543e-05, "loss": 0.4178, "step": 360 }, { "epoch": 0.006608695652173913, "grad_norm": 12.221634864807129, "learning_rate": 1.6156614753256583e-05, "loss": 0.6527, "step": 361 }, { "epoch": 0.00662700228832952, "grad_norm": 8.335714340209961, "learning_rate": 1.6129070536529767e-05, "loss": 0.5695, "step": 362 }, { "epoch": 0.0066453089244851255, "grad_norm": 9.451141357421875, "learning_rate": 1.610145163901268e-05, "loss": 0.4343, "step": 363 }, { "epoch": 0.006663615560640732, "grad_norm": 6.33415412902832, "learning_rate": 1.607375839723287e-05, "loss": 0.3475, "step": 364 }, { "epoch": 0.006681922196796339, "grad_norm": 8.798182487487793, "learning_rate": 1.6045991148623752e-05, "loss": 0.4564, "step": 365 }, { "epoch": 0.006700228832951945, "grad_norm": 7.0115647315979, "learning_rate": 1.6018150231520486e-05, "loss": 0.4848, "step": 366 }, { "epoch": 0.006718535469107552, "grad_norm": 11.819734573364258, "learning_rate": 1.599023598515586e-05, "loss": 0.9038, "step": 367 }, { "epoch": 0.006736842105263158, "grad_norm": 7.8828325271606445, "learning_rate": 1.5962248749656158e-05, "loss": 0.252, "step": 368 }, { "epoch": 0.006755148741418764, "grad_norm": 1.1683067083358765, "learning_rate": 1.5934188866037017e-05, "loss": 0.0875, "step": 369 }, { "epoch": 0.0067734553775743705, "grad_norm": 4.919974327087402, "learning_rate": 1.5906056676199256e-05, "loss": 0.1626, "step": 370 }, { "epoch": 0.006791762013729977, "grad_norm": 5.247345447540283, "learning_rate": 1.5877852522924733e-05, "loss": 0.2482, "step": 371 }, { "epoch": 0.006810068649885584, "grad_norm": 2.816721200942993, "learning_rate": 1.584957674987216e-05, "loss": 0.1829, "step": 372 }, { "epoch": 0.00682837528604119, "grad_norm": 15.067553520202637, "learning_rate": 1.5821229701572897e-05, "loss": 0.9891, "step": 373 }, { "epoch": 0.006846681922196797, "grad_norm": 12.05018424987793, "learning_rate": 1.5792811723426787e-05, "loss": 0.5695, "step": 374 }, { "epoch": 0.006864988558352402, "grad_norm": 7.035960674285889, "learning_rate": 1.5764323161697933e-05, "loss": 0.362, "step": 375 }, { "epoch": 0.006883295194508009, "grad_norm": 10.163368225097656, "learning_rate": 1.573576436351046e-05, "loss": 0.5263, "step": 376 }, { "epoch": 0.006901601830663615, "grad_norm": 8.28557014465332, "learning_rate": 1.570713567684432e-05, "loss": 0.4129, "step": 377 }, { "epoch": 0.006919908466819222, "grad_norm": 12.570793151855469, "learning_rate": 1.5678437450531014e-05, "loss": 0.7611, "step": 378 }, { "epoch": 0.0069382151029748285, "grad_norm": 6.368137836456299, "learning_rate": 1.564967003424938e-05, "loss": 0.5362, "step": 379 }, { "epoch": 0.006956521739130435, "grad_norm": 3.3329522609710693, "learning_rate": 1.5620833778521306e-05, "loss": 0.3633, "step": 380 }, { "epoch": 0.006974828375286042, "grad_norm": 5.921566009521484, "learning_rate": 1.5591929034707468e-05, "loss": 0.45, "step": 381 }, { "epoch": 0.006993135011441647, "grad_norm": 4.826172828674316, "learning_rate": 1.556295615500305e-05, "loss": 0.3278, "step": 382 }, { "epoch": 0.007011441647597254, "grad_norm": 6.323006629943848, "learning_rate": 1.553391549243344e-05, "loss": 0.3469, "step": 383 }, { "epoch": 0.00702974828375286, "grad_norm": 5.0740203857421875, "learning_rate": 1.5504807400849957e-05, "loss": 0.3713, "step": 384 }, { "epoch": 0.007048054919908467, "grad_norm": 4.266338348388672, "learning_rate": 1.5475632234925505e-05, "loss": 0.2165, "step": 385 }, { "epoch": 0.007066361556064073, "grad_norm": 4.6646504402160645, "learning_rate": 1.5446390350150272e-05, "loss": 0.4328, "step": 386 }, { "epoch": 0.00708466819221968, "grad_norm": 4.261294841766357, "learning_rate": 1.54170821028274e-05, "loss": 0.4952, "step": 387 }, { "epoch": 0.007102974828375286, "grad_norm": 4.2962846755981445, "learning_rate": 1.5387707850068633e-05, "loss": 0.3523, "step": 388 }, { "epoch": 0.007121281464530892, "grad_norm": 3.559885263442993, "learning_rate": 1.5358267949789968e-05, "loss": 0.1836, "step": 389 }, { "epoch": 0.007139588100686499, "grad_norm": 4.17678689956665, "learning_rate": 1.53287627607073e-05, "loss": 0.1978, "step": 390 }, { "epoch": 0.007157894736842105, "grad_norm": 9.370595932006836, "learning_rate": 1.529919264233205e-05, "loss": 0.5643, "step": 391 }, { "epoch": 0.007176201372997712, "grad_norm": 4.1542463302612305, "learning_rate": 1.5269557954966777e-05, "loss": 0.3046, "step": 392 }, { "epoch": 0.007194508009153318, "grad_norm": 7.581055164337158, "learning_rate": 1.5239859059700794e-05, "loss": 0.4282, "step": 393 }, { "epoch": 0.007212814645308925, "grad_norm": 2.4097397327423096, "learning_rate": 1.5210096318405768e-05, "loss": 0.154, "step": 394 }, { "epoch": 0.007231121281464531, "grad_norm": 3.8569397926330566, "learning_rate": 1.5180270093731305e-05, "loss": 0.0843, "step": 395 }, { "epoch": 0.007249427917620137, "grad_norm": 7.704616069793701, "learning_rate": 1.5150380749100545e-05, "loss": 0.5627, "step": 396 }, { "epoch": 0.007267734553775744, "grad_norm": 4.946331024169922, "learning_rate": 1.5120428648705716e-05, "loss": 0.2496, "step": 397 }, { "epoch": 0.00728604118993135, "grad_norm": 3.8404288291931152, "learning_rate": 1.5090414157503715e-05, "loss": 0.1916, "step": 398 }, { "epoch": 0.007304347826086957, "grad_norm": 1.226649522781372, "learning_rate": 1.5060337641211637e-05, "loss": 0.1359, "step": 399 }, { "epoch": 0.007322654462242563, "grad_norm": 10.208996772766113, "learning_rate": 1.5030199466302354e-05, "loss": 0.3555, "step": 400 }, { "epoch": 0.007340961098398169, "grad_norm": 7.164267539978027, "learning_rate": 1.5000000000000002e-05, "loss": 0.4576, "step": 401 }, { "epoch": 0.0073592677345537755, "grad_norm": 6.956635475158691, "learning_rate": 1.4969739610275556e-05, "loss": 0.3394, "step": 402 }, { "epoch": 0.007377574370709382, "grad_norm": 11.293649673461914, "learning_rate": 1.493941866584231e-05, "loss": 0.6647, "step": 403 }, { "epoch": 0.007395881006864989, "grad_norm": 9.281755447387695, "learning_rate": 1.490903753615141e-05, "loss": 0.4833, "step": 404 }, { "epoch": 0.007414187643020595, "grad_norm": 3.4506609439849854, "learning_rate": 1.4878596591387329e-05, "loss": 0.239, "step": 405 }, { "epoch": 0.007432494279176202, "grad_norm": 1.7865498065948486, "learning_rate": 1.4848096202463373e-05, "loss": 0.135, "step": 406 }, { "epoch": 0.007450800915331807, "grad_norm": 2.3500795364379883, "learning_rate": 1.4817536741017153e-05, "loss": 0.2015, "step": 407 }, { "epoch": 0.007469107551487414, "grad_norm": 8.29547119140625, "learning_rate": 1.478691857940607e-05, "loss": 0.4149, "step": 408 }, { "epoch": 0.0074874141876430204, "grad_norm": 2.8842177391052246, "learning_rate": 1.4756242090702756e-05, "loss": 0.1729, "step": 409 }, { "epoch": 0.007505720823798627, "grad_norm": 8.263029098510742, "learning_rate": 1.4725507648690542e-05, "loss": 0.5903, "step": 410 }, { "epoch": 0.0075240274599542335, "grad_norm": 4.133641719818115, "learning_rate": 1.469471562785891e-05, "loss": 0.3041, "step": 411 }, { "epoch": 0.00754233409610984, "grad_norm": 9.040873527526855, "learning_rate": 1.4663866403398915e-05, "loss": 0.6155, "step": 412 }, { "epoch": 0.007560640732265447, "grad_norm": 5.259375095367432, "learning_rate": 1.463296035119862e-05, "loss": 0.357, "step": 413 }, { "epoch": 0.007578947368421052, "grad_norm": 5.846144676208496, "learning_rate": 1.4601997847838518e-05, "loss": 0.2454, "step": 414 }, { "epoch": 0.007597254004576659, "grad_norm": 5.604888439178467, "learning_rate": 1.4570979270586944e-05, "loss": 0.4757, "step": 415 }, { "epoch": 0.007615560640732265, "grad_norm": 6.676849365234375, "learning_rate": 1.4539904997395468e-05, "loss": 0.2546, "step": 416 }, { "epoch": 0.007633867276887872, "grad_norm": 7.3254523277282715, "learning_rate": 1.4508775406894308e-05, "loss": 0.4365, "step": 417 }, { "epoch": 0.0076521739130434785, "grad_norm": 7.06563663482666, "learning_rate": 1.4477590878387697e-05, "loss": 0.3348, "step": 418 }, { "epoch": 0.007670480549199085, "grad_norm": 4.191524982452393, "learning_rate": 1.4446351791849276e-05, "loss": 0.2775, "step": 419 }, { "epoch": 0.007688787185354691, "grad_norm": 6.883635520935059, "learning_rate": 1.4415058527917454e-05, "loss": 0.336, "step": 420 }, { "epoch": 0.007707093821510297, "grad_norm": 2.3200838565826416, "learning_rate": 1.4383711467890776e-05, "loss": 0.1396, "step": 421 }, { "epoch": 0.007725400457665904, "grad_norm": 4.974113464355469, "learning_rate": 1.4352310993723277e-05, "loss": 0.2149, "step": 422 }, { "epoch": 0.00774370709382151, "grad_norm": 7.065567970275879, "learning_rate": 1.4320857488019826e-05, "loss": 0.3439, "step": 423 }, { "epoch": 0.007762013729977117, "grad_norm": 6.593064308166504, "learning_rate": 1.4289351334031461e-05, "loss": 0.327, "step": 424 }, { "epoch": 0.007780320366132723, "grad_norm": 4.018970012664795, "learning_rate": 1.4257792915650728e-05, "loss": 0.2449, "step": 425 }, { "epoch": 0.00779862700228833, "grad_norm": 1.6839569807052612, "learning_rate": 1.4226182617406996e-05, "loss": 0.1382, "step": 426 }, { "epoch": 0.007816933638443936, "grad_norm": 3.03248929977417, "learning_rate": 1.4194520824461773e-05, "loss": 0.1527, "step": 427 }, { "epoch": 0.007835240274599543, "grad_norm": 10.530253410339355, "learning_rate": 1.4162807922604014e-05, "loss": 0.4916, "step": 428 }, { "epoch": 0.00785354691075515, "grad_norm": 9.175287246704102, "learning_rate": 1.413104429824542e-05, "loss": 0.5146, "step": 429 }, { "epoch": 0.007871853546910754, "grad_norm": 9.110077857971191, "learning_rate": 1.4099230338415728e-05, "loss": 0.2713, "step": 430 }, { "epoch": 0.007890160183066361, "grad_norm": 6.078294277191162, "learning_rate": 1.4067366430758004e-05, "loss": 0.297, "step": 431 }, { "epoch": 0.007908466819221967, "grad_norm": 9.00320816040039, "learning_rate": 1.4035452963523903e-05, "loss": 0.219, "step": 432 }, { "epoch": 0.007926773455377574, "grad_norm": 11.155050277709961, "learning_rate": 1.4003490325568953e-05, "loss": 0.5544, "step": 433 }, { "epoch": 0.00794508009153318, "grad_norm": 9.077802658081055, "learning_rate": 1.3971478906347806e-05, "loss": 0.3477, "step": 434 }, { "epoch": 0.007963386727688787, "grad_norm": 5.8080668449401855, "learning_rate": 1.3939419095909513e-05, "loss": 0.275, "step": 435 }, { "epoch": 0.007981693363844394, "grad_norm": 19.085351943969727, "learning_rate": 1.3907311284892737e-05, "loss": 0.3296, "step": 436 }, { "epoch": 0.008, "grad_norm": 12.508645057678223, "learning_rate": 1.3875155864521031e-05, "loss": 0.5693, "step": 437 }, { "epoch": 0.008018306636155607, "grad_norm": 5.50254487991333, "learning_rate": 1.3842953226598036e-05, "loss": 0.1062, "step": 438 }, { "epoch": 0.008036613272311213, "grad_norm": 14.789398193359375, "learning_rate": 1.3810703763502744e-05, "loss": 0.9844, "step": 439 }, { "epoch": 0.00805491990846682, "grad_norm": 19.38977813720703, "learning_rate": 1.3778407868184674e-05, "loss": 0.6566, "step": 440 }, { "epoch": 0.008073226544622426, "grad_norm": 13.897294998168945, "learning_rate": 1.3746065934159123e-05, "loss": 0.7898, "step": 441 }, { "epoch": 0.008091533180778033, "grad_norm": 9.941490173339844, "learning_rate": 1.371367835550235e-05, "loss": 0.4879, "step": 442 }, { "epoch": 0.008109839816933638, "grad_norm": 2.901772975921631, "learning_rate": 1.3681245526846782e-05, "loss": 0.2148, "step": 443 }, { "epoch": 0.008128146453089244, "grad_norm": 4.679471492767334, "learning_rate": 1.3648767843376196e-05, "loss": 0.2317, "step": 444 }, { "epoch": 0.00814645308924485, "grad_norm": 3.4032607078552246, "learning_rate": 1.3616245700820922e-05, "loss": 0.2151, "step": 445 }, { "epoch": 0.008164759725400457, "grad_norm": 2.5919456481933594, "learning_rate": 1.3583679495453e-05, "loss": 0.1284, "step": 446 }, { "epoch": 0.008183066361556064, "grad_norm": 6.18490743637085, "learning_rate": 1.3551069624081372e-05, "loss": 0.3029, "step": 447 }, { "epoch": 0.00820137299771167, "grad_norm": 10.599230766296387, "learning_rate": 1.3518416484047018e-05, "loss": 0.5019, "step": 448 }, { "epoch": 0.008219679633867277, "grad_norm": 8.500578880310059, "learning_rate": 1.3485720473218153e-05, "loss": 0.4647, "step": 449 }, { "epoch": 0.008237986270022883, "grad_norm": 2.737617254257202, "learning_rate": 1.3452981989985347e-05, "loss": 0.1593, "step": 450 } ], "logging_steps": 1, "max_steps": 1000, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }