{ "best_metric": null, "best_model_checkpoint": null, "epoch": 6.0, "eval_steps": 200, "global_step": 210, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02857142857142857, "grad_norm": 19.810293087197298, "learning_rate": 5e-08, "loss": 1.6319, "step": 1 }, { "epoch": 0.05714285714285714, "grad_norm": 20.10065322014341, "learning_rate": 1e-07, "loss": 1.5294, "step": 2 }, { "epoch": 0.08571428571428572, "grad_norm": 22.139565739974667, "learning_rate": 1.5e-07, "loss": 1.568, "step": 3 }, { "epoch": 0.11428571428571428, "grad_norm": 17.607582690653754, "learning_rate": 2e-07, "loss": 1.6628, "step": 4 }, { "epoch": 0.14285714285714285, "grad_norm": 24.910794234048186, "learning_rate": 2.5e-07, "loss": 1.5738, "step": 5 }, { "epoch": 0.17142857142857143, "grad_norm": 18.006572442480767, "learning_rate": 3e-07, "loss": 1.5898, "step": 6 }, { "epoch": 0.2, "grad_norm": 22.152647510768777, "learning_rate": 3.5e-07, "loss": 1.5891, "step": 7 }, { "epoch": 0.22857142857142856, "grad_norm": 17.044462125735237, "learning_rate": 4e-07, "loss": 1.5277, "step": 8 }, { "epoch": 0.2571428571428571, "grad_norm": 19.26164281649356, "learning_rate": 4.5e-07, "loss": 1.5504, "step": 9 }, { "epoch": 0.2857142857142857, "grad_norm": 20.409183726640773, "learning_rate": 5e-07, "loss": 1.5509, "step": 10 }, { "epoch": 0.3142857142857143, "grad_norm": 16.865224915691606, "learning_rate": 5.5e-07, "loss": 1.5752, "step": 11 }, { "epoch": 0.34285714285714286, "grad_norm": 15.722428651072727, "learning_rate": 6e-07, "loss": 1.6208, "step": 12 }, { "epoch": 0.37142857142857144, "grad_norm": 13.23419740428603, "learning_rate": 6.5e-07, "loss": 1.515, "step": 13 }, { "epoch": 0.4, "grad_norm": 12.45173447148941, "learning_rate": 7e-07, "loss": 1.4752, "step": 14 }, { "epoch": 0.42857142857142855, "grad_norm": 10.684936055938318, "learning_rate": 7.5e-07, "loss": 1.5073, "step": 15 }, { "epoch": 0.45714285714285713, "grad_norm": 8.612999241840305, "learning_rate": 8e-07, "loss": 1.4206, "step": 16 }, { "epoch": 0.4857142857142857, "grad_norm": 9.64409059662565, "learning_rate": 8.499999999999999e-07, "loss": 1.3659, "step": 17 }, { "epoch": 0.5142857142857142, "grad_norm": 9.53721910636901, "learning_rate": 9e-07, "loss": 1.4999, "step": 18 }, { "epoch": 0.5428571428571428, "grad_norm": 15.44618942298652, "learning_rate": 9.499999999999999e-07, "loss": 1.452, "step": 19 }, { "epoch": 0.5714285714285714, "grad_norm": 7.797878397007685, "learning_rate": 1e-06, "loss": 1.4059, "step": 20 }, { "epoch": 0.6, "grad_norm": 8.611751272514192, "learning_rate": 9.999316524962345e-07, "loss": 1.3416, "step": 21 }, { "epoch": 0.6285714285714286, "grad_norm": 8.60172100284345, "learning_rate": 9.99726628670463e-07, "loss": 1.3334, "step": 22 }, { "epoch": 0.6571428571428571, "grad_norm": 7.707986778608113, "learning_rate": 9.993849845741523e-07, "loss": 1.3032, "step": 23 }, { "epoch": 0.6857142857142857, "grad_norm": 7.692312413668571, "learning_rate": 9.989068136093872e-07, "loss": 1.3742, "step": 24 }, { "epoch": 0.7142857142857143, "grad_norm": 6.364211690724886, "learning_rate": 9.982922465033348e-07, "loss": 1.2641, "step": 25 }, { "epoch": 0.7428571428571429, "grad_norm": 6.85599706491363, "learning_rate": 9.975414512725056e-07, "loss": 1.2624, "step": 26 }, { "epoch": 0.7714285714285715, "grad_norm": 6.401247246252304, "learning_rate": 9.966546331768192e-07, "loss": 1.258, "step": 27 }, { "epoch": 0.8, "grad_norm": 9.111986647021705, "learning_rate": 9.956320346634875e-07, "loss": 1.3228, "step": 28 }, { "epoch": 0.8285714285714286, "grad_norm": 6.326533367398195, "learning_rate": 9.944739353007341e-07, "loss": 1.2034, "step": 29 }, { "epoch": 0.8571428571428571, "grad_norm": 6.561308869602955, "learning_rate": 9.931806517013612e-07, "loss": 1.1795, "step": 30 }, { "epoch": 0.8857142857142857, "grad_norm": 5.9273022974036795, "learning_rate": 9.917525374361911e-07, "loss": 1.2288, "step": 31 }, { "epoch": 0.9142857142857143, "grad_norm": 5.348130355740025, "learning_rate": 9.901899829374047e-07, "loss": 1.1764, "step": 32 }, { "epoch": 0.9428571428571428, "grad_norm": 5.274551340001603, "learning_rate": 9.884934153917996e-07, "loss": 1.1402, "step": 33 }, { "epoch": 0.9714285714285714, "grad_norm": 5.681514700875701, "learning_rate": 9.866632986240029e-07, "loss": 1.1604, "step": 34 }, { "epoch": 1.0, "grad_norm": 5.77664764652199, "learning_rate": 9.847001329696652e-07, "loss": 1.1567, "step": 35 }, { "epoch": 1.0285714285714285, "grad_norm": 5.836094968508217, "learning_rate": 9.826044551386742e-07, "loss": 1.084, "step": 36 }, { "epoch": 1.0571428571428572, "grad_norm": 5.827344070494953, "learning_rate": 9.803768380684242e-07, "loss": 1.1305, "step": 37 }, { "epoch": 1.0857142857142856, "grad_norm": 5.1906563321747115, "learning_rate": 9.780178907671788e-07, "loss": 1.0574, "step": 38 }, { "epoch": 1.1142857142857143, "grad_norm": 5.562551195695788, "learning_rate": 9.755282581475767e-07, "loss": 1.1391, "step": 39 }, { "epoch": 1.1428571428571428, "grad_norm": 5.717325403158778, "learning_rate": 9.729086208503173e-07, "loss": 1.1267, "step": 40 }, { "epoch": 1.1714285714285715, "grad_norm": 5.969025543565558, "learning_rate": 9.701596950580807e-07, "loss": 1.1055, "step": 41 }, { "epoch": 1.2, "grad_norm": 4.274011328484305, "learning_rate": 9.672822322997304e-07, "loss": 1.0392, "step": 42 }, { "epoch": 1.2285714285714286, "grad_norm": 4.4561906748259545, "learning_rate": 9.642770192448535e-07, "loss": 1.0393, "step": 43 }, { "epoch": 1.2571428571428571, "grad_norm": 4.618703002376858, "learning_rate": 9.611448774886923e-07, "loss": 1.0795, "step": 44 }, { "epoch": 1.2857142857142856, "grad_norm": 4.473713036590507, "learning_rate": 9.578866633275286e-07, "loss": 1.0194, "step": 45 }, { "epoch": 1.3142857142857143, "grad_norm": 4.3150195602071415, "learning_rate": 9.545032675245813e-07, "loss": 1.0409, "step": 46 }, { "epoch": 1.342857142857143, "grad_norm": 4.847108292314838, "learning_rate": 9.509956150664795e-07, "loss": 1.0475, "step": 47 }, { "epoch": 1.3714285714285714, "grad_norm": 4.716522743464805, "learning_rate": 9.473646649103817e-07, "loss": 1.046, "step": 48 }, { "epoch": 1.4, "grad_norm": 4.942648131785751, "learning_rate": 9.436114097218058e-07, "loss": 0.9876, "step": 49 }, { "epoch": 1.4285714285714286, "grad_norm": 5.1753554721234565, "learning_rate": 9.397368756032444e-07, "loss": 1.0394, "step": 50 }, { "epoch": 1.457142857142857, "grad_norm": 4.996816148990504, "learning_rate": 9.357421218136386e-07, "loss": 0.9647, "step": 51 }, { "epoch": 1.4857142857142858, "grad_norm": 5.982295572473842, "learning_rate": 9.316282404787869e-07, "loss": 1.0986, "step": 52 }, { "epoch": 1.5142857142857142, "grad_norm": 4.701714221899932, "learning_rate": 9.273963562927694e-07, "loss": 1.0493, "step": 53 }, { "epoch": 1.5428571428571427, "grad_norm": 4.514923820929648, "learning_rate": 9.230476262104676e-07, "loss": 0.9847, "step": 54 }, { "epoch": 1.5714285714285714, "grad_norm": 4.701919910281847, "learning_rate": 9.185832391312642e-07, "loss": 0.9794, "step": 55 }, { "epoch": 1.6, "grad_norm": 5.175054800986126, "learning_rate": 9.1400441557401e-07, "loss": 1.0607, "step": 56 }, { "epoch": 1.6285714285714286, "grad_norm": 4.080487196055978, "learning_rate": 9.093124073433462e-07, "loss": 0.9856, "step": 57 }, { "epoch": 1.657142857142857, "grad_norm": 4.168373937047882, "learning_rate": 9.045084971874737e-07, "loss": 0.9107, "step": 58 }, { "epoch": 1.6857142857142857, "grad_norm": 4.223886586576581, "learning_rate": 8.995939984474623e-07, "loss": 0.9733, "step": 59 }, { "epoch": 1.7142857142857144, "grad_norm": 4.513527637157093, "learning_rate": 8.945702546981968e-07, "loss": 0.9476, "step": 60 }, { "epoch": 1.7428571428571429, "grad_norm": 6.610164639785705, "learning_rate": 8.894386393810562e-07, "loss": 0.9161, "step": 61 }, { "epoch": 1.7714285714285714, "grad_norm": 4.703651247017542, "learning_rate": 8.842005554284295e-07, "loss": 1.0058, "step": 62 }, { "epoch": 1.8, "grad_norm": 4.843849342547474, "learning_rate": 8.788574348801674e-07, "loss": 0.8815, "step": 63 }, { "epoch": 1.8285714285714287, "grad_norm": 4.850275003883815, "learning_rate": 8.734107384920769e-07, "loss": 0.9945, "step": 64 }, { "epoch": 1.8571428571428572, "grad_norm": 5.453604273607254, "learning_rate": 8.678619553365658e-07, "loss": 0.9822, "step": 65 }, { "epoch": 1.8857142857142857, "grad_norm": 5.3760203289286155, "learning_rate": 8.622126023955445e-07, "loss": 0.9781, "step": 66 }, { "epoch": 1.9142857142857141, "grad_norm": 3.7890108184255062, "learning_rate": 8.564642241456986e-07, "loss": 0.9188, "step": 67 }, { "epoch": 1.9428571428571428, "grad_norm": 4.482132592462384, "learning_rate": 8.506183921362442e-07, "loss": 0.9193, "step": 68 }, { "epoch": 1.9714285714285715, "grad_norm": 5.311797532408616, "learning_rate": 8.446767045592829e-07, "loss": 0.9563, "step": 69 }, { "epoch": 2.0, "grad_norm": 4.950363360704343, "learning_rate": 8.386407858128706e-07, "loss": 0.9172, "step": 70 }, { "epoch": 2.0285714285714285, "grad_norm": 4.784370125705732, "learning_rate": 8.325122860569241e-07, "loss": 0.8843, "step": 71 }, { "epoch": 2.057142857142857, "grad_norm": 3.63025112886302, "learning_rate": 8.262928807620843e-07, "loss": 0.8547, "step": 72 }, { "epoch": 2.085714285714286, "grad_norm": 5.218576028144485, "learning_rate": 8.199842702516582e-07, "loss": 0.9414, "step": 73 }, { "epoch": 2.1142857142857143, "grad_norm": 4.677814392084542, "learning_rate": 8.135881792367685e-07, "loss": 0.8748, "step": 74 }, { "epoch": 2.142857142857143, "grad_norm": 4.400225941517684, "learning_rate": 8.071063563448339e-07, "loss": 0.8646, "step": 75 }, { "epoch": 2.1714285714285713, "grad_norm": 4.629375190268374, "learning_rate": 8.005405736415125e-07, "loss": 0.8744, "step": 76 }, { "epoch": 2.2, "grad_norm": 4.1365979610935, "learning_rate": 7.938926261462365e-07, "loss": 0.8735, "step": 77 }, { "epoch": 2.2285714285714286, "grad_norm": 4.796843279900597, "learning_rate": 7.871643313414718e-07, "loss": 0.9161, "step": 78 }, { "epoch": 2.257142857142857, "grad_norm": 4.297808554492733, "learning_rate": 7.803575286758363e-07, "loss": 0.8119, "step": 79 }, { "epoch": 2.2857142857142856, "grad_norm": 4.229241471681201, "learning_rate": 7.734740790612136e-07, "loss": 0.8698, "step": 80 }, { "epoch": 2.314285714285714, "grad_norm": 5.474712766868538, "learning_rate": 7.665158643639969e-07, "loss": 0.8673, "step": 81 }, { "epoch": 2.342857142857143, "grad_norm": 5.8097182446336495, "learning_rate": 7.594847868906076e-07, "loss": 0.8561, "step": 82 }, { "epoch": 2.3714285714285714, "grad_norm": 6.090812187892629, "learning_rate": 7.523827688674219e-07, "loss": 0.9032, "step": 83 }, { "epoch": 2.4, "grad_norm": 4.2456911319822614, "learning_rate": 7.452117519152541e-07, "loss": 0.8431, "step": 84 }, { "epoch": 2.4285714285714284, "grad_norm": 4.562925662849317, "learning_rate": 7.379736965185368e-07, "loss": 0.844, "step": 85 }, { "epoch": 2.4571428571428573, "grad_norm": 4.517660029971984, "learning_rate": 7.306705814893439e-07, "loss": 0.8071, "step": 86 }, { "epoch": 2.4857142857142858, "grad_norm": 4.970820782182107, "learning_rate": 7.233044034264033e-07, "loss": 0.8726, "step": 87 }, { "epoch": 2.5142857142857142, "grad_norm": 4.817397128446438, "learning_rate": 7.158771761692464e-07, "loss": 0.8651, "step": 88 }, { "epoch": 2.5428571428571427, "grad_norm": 4.521311016914525, "learning_rate": 7.083909302476452e-07, "loss": 0.8023, "step": 89 }, { "epoch": 2.571428571428571, "grad_norm": 5.57347106069374, "learning_rate": 7.008477123264847e-07, "loss": 0.8107, "step": 90 }, { "epoch": 2.6, "grad_norm": 4.958424900327067, "learning_rate": 6.932495846462261e-07, "loss": 0.8321, "step": 91 }, { "epoch": 2.6285714285714286, "grad_norm": 4.34206205323412, "learning_rate": 6.855986244591103e-07, "loss": 0.8003, "step": 92 }, { "epoch": 2.657142857142857, "grad_norm": 4.417137471491474, "learning_rate": 6.778969234612583e-07, "loss": 0.81, "step": 93 }, { "epoch": 2.685714285714286, "grad_norm": 4.714391490425693, "learning_rate": 6.701465872208216e-07, "loss": 0.8362, "step": 94 }, { "epoch": 2.7142857142857144, "grad_norm": 3.835031010544762, "learning_rate": 6.623497346023417e-07, "loss": 0.7408, "step": 95 }, { "epoch": 2.742857142857143, "grad_norm": 4.43328902038276, "learning_rate": 6.545084971874736e-07, "loss": 0.803, "step": 96 }, { "epoch": 2.7714285714285714, "grad_norm": 5.573988325671216, "learning_rate": 6.466250186922324e-07, "loss": 0.8499, "step": 97 }, { "epoch": 2.8, "grad_norm": 5.080352906047553, "learning_rate": 6.387014543809223e-07, "loss": 0.8536, "step": 98 }, { "epoch": 2.8285714285714287, "grad_norm": 4.433727646568338, "learning_rate": 6.307399704769098e-07, "loss": 0.8167, "step": 99 }, { "epoch": 2.857142857142857, "grad_norm": 4.72983530688845, "learning_rate": 6.227427435703995e-07, "loss": 0.8247, "step": 100 }, { "epoch": 2.8857142857142857, "grad_norm": 4.031675844503338, "learning_rate": 6.147119600233758e-07, "loss": 0.7923, "step": 101 }, { "epoch": 2.914285714285714, "grad_norm": 4.657562682714027, "learning_rate": 6.066498153718734e-07, "loss": 0.8259, "step": 102 }, { "epoch": 2.942857142857143, "grad_norm": 3.742914618609364, "learning_rate": 5.985585137257401e-07, "loss": 0.7416, "step": 103 }, { "epoch": 2.9714285714285715, "grad_norm": 5.392956655820792, "learning_rate": 5.90440267166055e-07, "loss": 0.86, "step": 104 }, { "epoch": 3.0, "grad_norm": 4.370818695376769, "learning_rate": 5.82297295140367e-07, "loss": 0.7756, "step": 105 }, { "epoch": 3.0285714285714285, "grad_norm": 4.1080059715561426, "learning_rate": 5.741318238559209e-07, "loss": 0.7417, "step": 106 }, { "epoch": 3.057142857142857, "grad_norm": 5.242835158751313, "learning_rate": 5.659460856710345e-07, "loss": 0.7413, "step": 107 }, { "epoch": 3.085714285714286, "grad_norm": 4.213242126183361, "learning_rate": 5.577423184847931e-07, "loss": 0.774, "step": 108 }, { "epoch": 3.1142857142857143, "grad_norm": 5.14212832829309, "learning_rate": 5.495227651252315e-07, "loss": 0.7596, "step": 109 }, { "epoch": 3.142857142857143, "grad_norm": 4.536657005455878, "learning_rate": 5.412896727361662e-07, "loss": 0.7494, "step": 110 }, { "epoch": 3.1714285714285713, "grad_norm": 5.180517603140962, "learning_rate": 5.330452921628497e-07, "loss": 0.8146, "step": 111 }, { "epoch": 3.2, "grad_norm": 4.499166275543811, "learning_rate": 5.247918773366111e-07, "loss": 0.7437, "step": 112 }, { "epoch": 3.2285714285714286, "grad_norm": 4.78892229810655, "learning_rate": 5.165316846586541e-07, "loss": 0.732, "step": 113 }, { "epoch": 3.257142857142857, "grad_norm": 4.643947668361775, "learning_rate": 5.082669723831793e-07, "loss": 0.769, "step": 114 }, { "epoch": 3.2857142857142856, "grad_norm": 4.79036988016491, "learning_rate": 5e-07, "loss": 0.7586, "step": 115 }, { "epoch": 3.314285714285714, "grad_norm": 5.1592671986124845, "learning_rate": 4.917330276168208e-07, "loss": 0.7513, "step": 116 }, { "epoch": 3.342857142857143, "grad_norm": 3.944922328543468, "learning_rate": 4.834683153413459e-07, "loss": 0.7184, "step": 117 }, { "epoch": 3.3714285714285714, "grad_norm": 4.429646542430914, "learning_rate": 4.752081226633888e-07, "loss": 0.6937, "step": 118 }, { "epoch": 3.4, "grad_norm": 5.016602862879482, "learning_rate": 4.669547078371503e-07, "loss": 0.7638, "step": 119 }, { "epoch": 3.4285714285714284, "grad_norm": 4.443230989038584, "learning_rate": 4.5871032726383385e-07, "loss": 0.7029, "step": 120 }, { "epoch": 3.4571428571428573, "grad_norm": 4.244778886390162, "learning_rate": 4.5047723487476864e-07, "loss": 0.7298, "step": 121 }, { "epoch": 3.4857142857142858, "grad_norm": 5.068619137512623, "learning_rate": 4.4225768151520694e-07, "loss": 0.7716, "step": 122 }, { "epoch": 3.5142857142857142, "grad_norm": 4.652036232274711, "learning_rate": 4.340539143289655e-07, "loss": 0.7276, "step": 123 }, { "epoch": 3.5428571428571427, "grad_norm": 4.95389292196556, "learning_rate": 4.258681761440789e-07, "loss": 0.7345, "step": 124 }, { "epoch": 3.571428571428571, "grad_norm": 4.275659541196405, "learning_rate": 4.1770270485963294e-07, "loss": 0.7323, "step": 125 }, { "epoch": 3.6, "grad_norm": 4.607417501587952, "learning_rate": 4.095597328339452e-07, "loss": 0.7606, "step": 126 }, { "epoch": 3.6285714285714286, "grad_norm": 4.794919375541705, "learning_rate": 4.0144148627425986e-07, "loss": 0.7149, "step": 127 }, { "epoch": 3.657142857142857, "grad_norm": 4.102606927748694, "learning_rate": 3.9335018462812664e-07, "loss": 0.6956, "step": 128 }, { "epoch": 3.685714285714286, "grad_norm": 4.243593175311764, "learning_rate": 3.8528803997662423e-07, "loss": 0.6894, "step": 129 }, { "epoch": 3.7142857142857144, "grad_norm": 4.76702832294717, "learning_rate": 3.772572564296004e-07, "loss": 0.7188, "step": 130 }, { "epoch": 3.742857142857143, "grad_norm": 4.651784836564843, "learning_rate": 3.692600295230901e-07, "loss": 0.7059, "step": 131 }, { "epoch": 3.7714285714285714, "grad_norm": 4.977499267065962, "learning_rate": 3.612985456190778e-07, "loss": 0.7022, "step": 132 }, { "epoch": 3.8, "grad_norm": 4.629091116356692, "learning_rate": 3.533749813077677e-07, "loss": 0.692, "step": 133 }, { "epoch": 3.8285714285714287, "grad_norm": 5.180269168283943, "learning_rate": 3.454915028125263e-07, "loss": 0.7245, "step": 134 }, { "epoch": 3.857142857142857, "grad_norm": 4.804587126213361, "learning_rate": 3.3765026539765827e-07, "loss": 0.6965, "step": 135 }, { "epoch": 3.8857142857142857, "grad_norm": 5.24711927553455, "learning_rate": 3.2985341277917846e-07, "loss": 0.7271, "step": 136 }, { "epoch": 3.914285714285714, "grad_norm": 4.306919498942911, "learning_rate": 3.221030765387417e-07, "loss": 0.7287, "step": 137 }, { "epoch": 3.942857142857143, "grad_norm": 5.1155454022030815, "learning_rate": 3.1440137554088953e-07, "loss": 0.7441, "step": 138 }, { "epoch": 3.9714285714285715, "grad_norm": 4.184043691688485, "learning_rate": 3.06750415353774e-07, "loss": 0.7349, "step": 139 }, { "epoch": 4.0, "grad_norm": 4.262350945325998, "learning_rate": 2.9915228767351535e-07, "loss": 0.7377, "step": 140 }, { "epoch": 4.0285714285714285, "grad_norm": 4.103707140569048, "learning_rate": 2.916090697523549e-07, "loss": 0.6864, "step": 141 }, { "epoch": 4.057142857142857, "grad_norm": 4.9358404348712765, "learning_rate": 2.841228238307536e-07, "loss": 0.7073, "step": 142 }, { "epoch": 4.085714285714285, "grad_norm": 4.224580170912438, "learning_rate": 2.7669559657359673e-07, "loss": 0.6669, "step": 143 }, { "epoch": 4.114285714285714, "grad_norm": 4.571605759259632, "learning_rate": 2.6932941851065615e-07, "loss": 0.6673, "step": 144 }, { "epoch": 4.142857142857143, "grad_norm": 4.161244011003443, "learning_rate": 2.620263034814632e-07, "loss": 0.6644, "step": 145 }, { "epoch": 4.171428571428572, "grad_norm": 4.128269627797568, "learning_rate": 2.547882480847461e-07, "loss": 0.6568, "step": 146 }, { "epoch": 4.2, "grad_norm": 4.562112299679241, "learning_rate": 2.476172311325783e-07, "loss": 0.7048, "step": 147 }, { "epoch": 4.228571428571429, "grad_norm": 5.336142167342819, "learning_rate": 2.4051521310939254e-07, "loss": 0.7082, "step": 148 }, { "epoch": 4.257142857142857, "grad_norm": 4.796197499964798, "learning_rate": 2.3348413563600323e-07, "loss": 0.6497, "step": 149 }, { "epoch": 4.285714285714286, "grad_norm": 4.825647232190606, "learning_rate": 2.2652592093878665e-07, "loss": 0.6851, "step": 150 }, { "epoch": 4.314285714285714, "grad_norm": 4.462461397938811, "learning_rate": 2.1964247132416368e-07, "loss": 0.6974, "step": 151 }, { "epoch": 4.3428571428571425, "grad_norm": 4.829646877353178, "learning_rate": 2.128356686585282e-07, "loss": 0.6439, "step": 152 }, { "epoch": 4.371428571428572, "grad_norm": 5.043886293339906, "learning_rate": 2.0610737385376348e-07, "loss": 0.7215, "step": 153 }, { "epoch": 4.4, "grad_norm": 4.598743427591961, "learning_rate": 1.9945942635848745e-07, "loss": 0.644, "step": 154 }, { "epoch": 4.428571428571429, "grad_norm": 3.8253679810994012, "learning_rate": 1.9289364365516607e-07, "loss": 0.6397, "step": 155 }, { "epoch": 4.457142857142857, "grad_norm": 4.807293265176188, "learning_rate": 1.8641182076323148e-07, "loss": 0.689, "step": 156 }, { "epoch": 4.485714285714286, "grad_norm": 3.8062822348294363, "learning_rate": 1.8001572974834168e-07, "loss": 0.6466, "step": 157 }, { "epoch": 4.514285714285714, "grad_norm": 4.883594185819761, "learning_rate": 1.7370711923791564e-07, "loss": 0.6691, "step": 158 }, { "epoch": 4.542857142857143, "grad_norm": 5.143361513426633, "learning_rate": 1.674877139430758e-07, "loss": 0.6432, "step": 159 }, { "epoch": 4.571428571428571, "grad_norm": 4.99547680018691, "learning_rate": 1.6135921418712955e-07, "loss": 0.6685, "step": 160 }, { "epoch": 4.6, "grad_norm": 4.6838462077186795, "learning_rate": 1.553232954407171e-07, "loss": 0.657, "step": 161 }, { "epoch": 4.628571428571428, "grad_norm": 4.841965011015589, "learning_rate": 1.493816078637557e-07, "loss": 0.6671, "step": 162 }, { "epoch": 4.6571428571428575, "grad_norm": 5.2138374952108535, "learning_rate": 1.435357758543015e-07, "loss": 0.7393, "step": 163 }, { "epoch": 4.685714285714286, "grad_norm": 4.7664344007331225, "learning_rate": 1.3778739760445552e-07, "loss": 0.6904, "step": 164 }, { "epoch": 4.714285714285714, "grad_norm": 4.757741131293643, "learning_rate": 1.321380446634342e-07, "loss": 0.7093, "step": 165 }, { "epoch": 4.742857142857143, "grad_norm": 5.064771039141748, "learning_rate": 1.2658926150792322e-07, "loss": 0.6588, "step": 166 }, { "epoch": 4.771428571428571, "grad_norm": 4.745894552114562, "learning_rate": 1.2114256511983274e-07, "loss": 0.6914, "step": 167 }, { "epoch": 4.8, "grad_norm": 4.5842953923274425, "learning_rate": 1.1579944457157059e-07, "loss": 0.6819, "step": 168 }, { "epoch": 4.828571428571428, "grad_norm": 6.02007293004415, "learning_rate": 1.1056136061894384e-07, "loss": 0.7107, "step": 169 }, { "epoch": 4.857142857142857, "grad_norm": 4.04792450108683, "learning_rate": 1.0542974530180327e-07, "loss": 0.6577, "step": 170 }, { "epoch": 4.885714285714286, "grad_norm": 4.334397719277511, "learning_rate": 1.0040600155253764e-07, "loss": 0.6926, "step": 171 }, { "epoch": 4.914285714285715, "grad_norm": 4.763783176385905, "learning_rate": 9.549150281252632e-08, "loss": 0.6824, "step": 172 }, { "epoch": 4.942857142857143, "grad_norm": 3.8987128530052075, "learning_rate": 9.068759265665382e-08, "loss": 0.6293, "step": 173 }, { "epoch": 4.9714285714285715, "grad_norm": 5.260545837340523, "learning_rate": 8.599558442598998e-08, "loss": 0.7605, "step": 174 }, { "epoch": 5.0, "grad_norm": 4.28845521901137, "learning_rate": 8.141676086873573e-08, "loss": 0.6685, "step": 175 }, { "epoch": 5.0285714285714285, "grad_norm": 4.317069591639633, "learning_rate": 7.695237378953224e-08, "loss": 0.6178, "step": 176 }, { "epoch": 5.057142857142857, "grad_norm": 4.673363595181425, "learning_rate": 7.260364370723043e-08, "loss": 0.6338, "step": 177 }, { "epoch": 5.085714285714285, "grad_norm": 4.179297774765176, "learning_rate": 6.837175952121304e-08, "loss": 0.656, "step": 178 }, { "epoch": 5.114285714285714, "grad_norm": 4.2740125604739685, "learning_rate": 6.42578781863613e-08, "loss": 0.7062, "step": 179 }, { "epoch": 5.142857142857143, "grad_norm": 3.7179175942036378, "learning_rate": 6.026312439675551e-08, "loss": 0.6132, "step": 180 }, { "epoch": 5.171428571428572, "grad_norm": 4.586691811463999, "learning_rate": 5.638859027819409e-08, "loss": 0.6846, "step": 181 }, { "epoch": 5.2, "grad_norm": 4.74824620606461, "learning_rate": 5.263533508961826e-08, "loss": 0.695, "step": 182 }, { "epoch": 5.228571428571429, "grad_norm": 4.552865468215117, "learning_rate": 4.9004384933520547e-08, "loss": 0.6363, "step": 183 }, { "epoch": 5.257142857142857, "grad_norm": 3.868803646466012, "learning_rate": 4.549673247541874e-08, "loss": 0.6197, "step": 184 }, { "epoch": 5.285714285714286, "grad_norm": 4.257422013394126, "learning_rate": 4.2113336672471245e-08, "loss": 0.6325, "step": 185 }, { "epoch": 5.314285714285714, "grad_norm": 4.138411901026596, "learning_rate": 3.8855122511307626e-08, "loss": 0.6074, "step": 186 }, { "epoch": 5.3428571428571425, "grad_norm": 4.386100060157392, "learning_rate": 3.572298075514652e-08, "loss": 0.6298, "step": 187 }, { "epoch": 5.371428571428572, "grad_norm": 3.899133464905973, "learning_rate": 3.271776770026963e-08, "loss": 0.5991, "step": 188 }, { "epoch": 5.4, "grad_norm": 5.227219263192165, "learning_rate": 2.9840304941919416e-08, "loss": 0.6917, "step": 189 }, { "epoch": 5.428571428571429, "grad_norm": 4.194540185559828, "learning_rate": 2.7091379149682682e-08, "loss": 0.6592, "step": 190 }, { "epoch": 5.457142857142857, "grad_norm": 4.103130224713111, "learning_rate": 2.4471741852423233e-08, "loss": 0.6467, "step": 191 }, { "epoch": 5.485714285714286, "grad_norm": 4.401362210384649, "learning_rate": 2.1982109232821176e-08, "loss": 0.6471, "step": 192 }, { "epoch": 5.514285714285714, "grad_norm": 4.4025758572699365, "learning_rate": 1.9623161931575926e-08, "loss": 0.6537, "step": 193 }, { "epoch": 5.542857142857143, "grad_norm": 4.940644969360135, "learning_rate": 1.7395544861325718e-08, "loss": 0.6647, "step": 194 }, { "epoch": 5.571428571428571, "grad_norm": 4.562635610658606, "learning_rate": 1.5299867030334813e-08, "loss": 0.6422, "step": 195 }, { "epoch": 5.6, "grad_norm": 4.58494040795585, "learning_rate": 1.3336701375997127e-08, "loss": 0.7141, "step": 196 }, { "epoch": 5.628571428571428, "grad_norm": 4.240912508697285, "learning_rate": 1.1506584608200364e-08, "loss": 0.6423, "step": 197 }, { "epoch": 5.6571428571428575, "grad_norm": 4.474353548561071, "learning_rate": 9.810017062595321e-09, "loss": 0.6362, "step": 198 }, { "epoch": 5.685714285714286, "grad_norm": 5.263243471264244, "learning_rate": 8.247462563808816e-09, "loss": 0.6436, "step": 199 }, { "epoch": 5.714285714285714, "grad_norm": 5.629514926131208, "learning_rate": 6.819348298638839e-09, "loss": 0.6434, "step": 200 }, { "epoch": 5.714285714285714, "eval_loss": 1.0673894882202148, "eval_runtime": 1.3433, "eval_samples_per_second": 17.122, "eval_steps_per_second": 0.744, "step": 200 }, { "epoch": 5.742857142857143, "grad_norm": 4.388069390318575, "learning_rate": 5.526064699265753e-09, "loss": 0.6946, "step": 201 }, { "epoch": 5.771428571428571, "grad_norm": 3.85292435393061, "learning_rate": 4.367965336512403e-09, "loss": 0.6113, "step": 202 }, { "epoch": 5.8, "grad_norm": 4.3104727339441755, "learning_rate": 3.3453668231809283e-09, "loss": 0.6493, "step": 203 }, { "epoch": 5.828571428571428, "grad_norm": 4.21000584494849, "learning_rate": 2.458548727494292e-09, "loss": 0.653, "step": 204 }, { "epoch": 5.857142857142857, "grad_norm": 5.114411845255924, "learning_rate": 1.7077534966650765e-09, "loss": 0.6527, "step": 205 }, { "epoch": 5.885714285714286, "grad_norm": 4.988848182549562, "learning_rate": 1.0931863906127325e-09, "loss": 0.6935, "step": 206 }, { "epoch": 5.914285714285715, "grad_norm": 4.005152725886777, "learning_rate": 6.150154258476314e-10, "loss": 0.6559, "step": 207 }, { "epoch": 5.942857142857143, "grad_norm": 6.395756261520298, "learning_rate": 2.733713295369755e-10, "loss": 0.693, "step": 208 }, { "epoch": 5.9714285714285715, "grad_norm": 5.465002272924693, "learning_rate": 6.834750376549791e-11, "loss": 0.7231, "step": 209 }, { "epoch": 6.0, "grad_norm": 4.273940345648053, "learning_rate": 0.0, "loss": 0.6288, "step": 210 }, { "epoch": 6.0, "step": 210, "total_flos": 18587454013440.0, "train_loss": 0.8877790099098569, "train_runtime": 1031.8496, "train_samples_per_second": 12.932, "train_steps_per_second": 0.204 } ], "logging_steps": 1.0, "max_steps": 210, "num_input_tokens_seen": 0, "num_train_epochs": 6, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 18587454013440.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }