{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.0, "eval_steps": 146, "global_step": 1456, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005494505494505495, "grad_norm": Infinity, "learning_rate": 0.0, "loss": 2.0983, "step": 2 }, { "epoch": 0.01098901098901099, "grad_norm": 5.516454696655273, "learning_rate": 1.36986301369863e-07, "loss": 2.0688, "step": 4 }, { "epoch": 0.016483516483516484, "grad_norm": 5.619696617126465, "learning_rate": 2.73972602739726e-07, "loss": 2.0901, "step": 6 }, { "epoch": 0.02197802197802198, "grad_norm": 5.722362995147705, "learning_rate": 4.1095890410958903e-07, "loss": 2.1, "step": 8 }, { "epoch": 0.027472527472527472, "grad_norm": 5.279537200927734, "learning_rate": 5.47945205479452e-07, "loss": 2.0519, "step": 10 }, { "epoch": 0.03296703296703297, "grad_norm": 4.9255170822143555, "learning_rate": 6.849315068493151e-07, "loss": 2.0328, "step": 12 }, { "epoch": 0.038461538461538464, "grad_norm": 4.8295369148254395, "learning_rate": 8.219178082191781e-07, "loss": 1.9894, "step": 14 }, { "epoch": 0.04395604395604396, "grad_norm": 6.103366374969482, "learning_rate": 9.589041095890411e-07, "loss": 1.9078, "step": 16 }, { "epoch": 0.04945054945054945, "grad_norm": 5.239104747772217, "learning_rate": 1.095890410958904e-06, "loss": 1.8896, "step": 18 }, { "epoch": 0.054945054945054944, "grad_norm": 4.494051933288574, "learning_rate": 1.2328767123287673e-06, "loss": 1.6875, "step": 20 }, { "epoch": 0.06043956043956044, "grad_norm": 4.799252510070801, "learning_rate": 1.3698630136986302e-06, "loss": 1.6452, "step": 22 }, { "epoch": 0.06593406593406594, "grad_norm": 4.425687789916992, "learning_rate": 1.5068493150684932e-06, "loss": 1.5281, "step": 24 }, { "epoch": 0.07142857142857142, "grad_norm": 2.157963514328003, "learning_rate": 1.6438356164383561e-06, "loss": 1.3483, "step": 26 }, { "epoch": 0.07692307692307693, "grad_norm": 1.2025972604751587, "learning_rate": 1.7808219178082193e-06, "loss": 1.3064, "step": 28 }, { "epoch": 0.08241758241758242, "grad_norm": 0.9520919322967529, "learning_rate": 1.9178082191780823e-06, "loss": 1.2681, "step": 30 }, { "epoch": 0.08791208791208792, "grad_norm": 0.6353013515472412, "learning_rate": 2.0547945205479454e-06, "loss": 1.2484, "step": 32 }, { "epoch": 0.09340659340659341, "grad_norm": 0.46617600321769714, "learning_rate": 2.191780821917808e-06, "loss": 1.24, "step": 34 }, { "epoch": 0.0989010989010989, "grad_norm": 0.44788283109664917, "learning_rate": 2.3287671232876713e-06, "loss": 1.2342, "step": 36 }, { "epoch": 0.1043956043956044, "grad_norm": 0.3921670615673065, "learning_rate": 2.4657534246575345e-06, "loss": 1.217, "step": 38 }, { "epoch": 0.10989010989010989, "grad_norm": 0.3610630929470062, "learning_rate": 2.6027397260273973e-06, "loss": 1.225, "step": 40 }, { "epoch": 0.11538461538461539, "grad_norm": 0.3526926040649414, "learning_rate": 2.7397260273972604e-06, "loss": 1.2088, "step": 42 }, { "epoch": 0.12087912087912088, "grad_norm": 0.3451712727546692, "learning_rate": 2.876712328767123e-06, "loss": 1.2198, "step": 44 }, { "epoch": 0.12637362637362637, "grad_norm": 0.33328160643577576, "learning_rate": 3.0136986301369864e-06, "loss": 1.2173, "step": 46 }, { "epoch": 0.13186813186813187, "grad_norm": 0.3273652195930481, "learning_rate": 3.1506849315068495e-06, "loss": 1.2088, "step": 48 }, { "epoch": 0.13736263736263737, "grad_norm": 0.3195970952510834, "learning_rate": 3.2876712328767123e-06, "loss": 1.2054, "step": 50 }, { "epoch": 0.14285714285714285, "grad_norm": 0.2997814416885376, "learning_rate": 3.4246575342465754e-06, "loss": 1.2046, "step": 52 }, { "epoch": 0.14835164835164835, "grad_norm": 0.2609250247478485, "learning_rate": 3.5616438356164386e-06, "loss": 1.2058, "step": 54 }, { "epoch": 0.15384615384615385, "grad_norm": 0.21140357851982117, "learning_rate": 3.6986301369863014e-06, "loss": 1.2001, "step": 56 }, { "epoch": 0.15934065934065933, "grad_norm": 0.16791734099388123, "learning_rate": 3.8356164383561645e-06, "loss": 1.2017, "step": 58 }, { "epoch": 0.16483516483516483, "grad_norm": 0.13620080053806305, "learning_rate": 3.972602739726027e-06, "loss": 1.1951, "step": 60 }, { "epoch": 0.17032967032967034, "grad_norm": 0.12290852516889572, "learning_rate": 4.109589041095891e-06, "loss": 1.2006, "step": 62 }, { "epoch": 0.17582417582417584, "grad_norm": 0.1265764832496643, "learning_rate": 4.246575342465754e-06, "loss": 1.2009, "step": 64 }, { "epoch": 0.1813186813186813, "grad_norm": 0.10449671745300293, "learning_rate": 4.383561643835616e-06, "loss": 1.1948, "step": 66 }, { "epoch": 0.18681318681318682, "grad_norm": 0.09835775941610336, "learning_rate": 4.52054794520548e-06, "loss": 1.186, "step": 68 }, { "epoch": 0.19230769230769232, "grad_norm": 0.10505539923906326, "learning_rate": 4.657534246575343e-06, "loss": 1.1913, "step": 70 }, { "epoch": 0.1978021978021978, "grad_norm": 0.13944971561431885, "learning_rate": 4.7945205479452054e-06, "loss": 1.1911, "step": 72 }, { "epoch": 0.2032967032967033, "grad_norm": 0.1453215479850769, "learning_rate": 4.931506849315069e-06, "loss": 1.1969, "step": 74 }, { "epoch": 0.2087912087912088, "grad_norm": 0.12393910437822342, "learning_rate": 5.068493150684932e-06, "loss": 1.199, "step": 76 }, { "epoch": 0.21428571428571427, "grad_norm": 0.10112921893596649, "learning_rate": 5.2054794520547945e-06, "loss": 1.1914, "step": 78 }, { "epoch": 0.21978021978021978, "grad_norm": 0.1048625037074089, "learning_rate": 5.342465753424658e-06, "loss": 1.1937, "step": 80 }, { "epoch": 0.22527472527472528, "grad_norm": 0.1040705144405365, "learning_rate": 5.479452054794521e-06, "loss": 1.1889, "step": 82 }, { "epoch": 0.23076923076923078, "grad_norm": 0.14360405504703522, "learning_rate": 5.6164383561643845e-06, "loss": 1.1946, "step": 84 }, { "epoch": 0.23626373626373626, "grad_norm": 0.12936367094516754, "learning_rate": 5.753424657534246e-06, "loss": 1.1882, "step": 86 }, { "epoch": 0.24175824175824176, "grad_norm": 0.08769886940717697, "learning_rate": 5.89041095890411e-06, "loss": 1.1873, "step": 88 }, { "epoch": 0.24725274725274726, "grad_norm": 0.1197541207075119, "learning_rate": 6.027397260273973e-06, "loss": 1.1863, "step": 90 }, { "epoch": 0.25274725274725274, "grad_norm": 0.12938760221004486, "learning_rate": 6.164383561643836e-06, "loss": 1.188, "step": 92 }, { "epoch": 0.25824175824175827, "grad_norm": 0.14707419276237488, "learning_rate": 6.301369863013699e-06, "loss": 1.1886, "step": 94 }, { "epoch": 0.26373626373626374, "grad_norm": 0.12229153513908386, "learning_rate": 6.438356164383563e-06, "loss": 1.1908, "step": 96 }, { "epoch": 0.2692307692307692, "grad_norm": 0.1133919209241867, "learning_rate": 6.5753424657534245e-06, "loss": 1.1825, "step": 98 }, { "epoch": 0.27472527472527475, "grad_norm": 0.12165709584951401, "learning_rate": 6.712328767123288e-06, "loss": 1.1807, "step": 100 }, { "epoch": 0.2802197802197802, "grad_norm": 0.12554234266281128, "learning_rate": 6.849315068493151e-06, "loss": 1.1831, "step": 102 }, { "epoch": 0.2857142857142857, "grad_norm": 0.11460216343402863, "learning_rate": 6.9863013698630145e-06, "loss": 1.1865, "step": 104 }, { "epoch": 0.29120879120879123, "grad_norm": 0.1516668051481247, "learning_rate": 7.123287671232877e-06, "loss": 1.1871, "step": 106 }, { "epoch": 0.2967032967032967, "grad_norm": 0.15025198459625244, "learning_rate": 7.260273972602741e-06, "loss": 1.1844, "step": 108 }, { "epoch": 0.3021978021978022, "grad_norm": 0.13930697739124298, "learning_rate": 7.397260273972603e-06, "loss": 1.1803, "step": 110 }, { "epoch": 0.3076923076923077, "grad_norm": 0.14465640485286713, "learning_rate": 7.534246575342466e-06, "loss": 1.1895, "step": 112 }, { "epoch": 0.3131868131868132, "grad_norm": 0.12900669872760773, "learning_rate": 7.671232876712329e-06, "loss": 1.1874, "step": 114 }, { "epoch": 0.31868131868131866, "grad_norm": 0.12059827893972397, "learning_rate": 7.808219178082192e-06, "loss": 1.184, "step": 116 }, { "epoch": 0.3241758241758242, "grad_norm": 0.11196708679199219, "learning_rate": 7.945205479452055e-06, "loss": 1.1876, "step": 118 }, { "epoch": 0.32967032967032966, "grad_norm": 0.11420655995607376, "learning_rate": 8.082191780821919e-06, "loss": 1.1799, "step": 120 }, { "epoch": 0.33516483516483514, "grad_norm": 0.10233011841773987, "learning_rate": 8.219178082191782e-06, "loss": 1.1836, "step": 122 }, { "epoch": 0.34065934065934067, "grad_norm": 0.13560251891613007, "learning_rate": 8.356164383561644e-06, "loss": 1.1825, "step": 124 }, { "epoch": 0.34615384615384615, "grad_norm": 0.1417393684387207, "learning_rate": 8.493150684931507e-06, "loss": 1.181, "step": 126 }, { "epoch": 0.3516483516483517, "grad_norm": 0.12368721514940262, "learning_rate": 8.63013698630137e-06, "loss": 1.1857, "step": 128 }, { "epoch": 0.35714285714285715, "grad_norm": 0.14341862499713898, "learning_rate": 8.767123287671233e-06, "loss": 1.1839, "step": 130 }, { "epoch": 0.3626373626373626, "grad_norm": 0.14925454556941986, "learning_rate": 8.904109589041097e-06, "loss": 1.1792, "step": 132 }, { "epoch": 0.36813186813186816, "grad_norm": 0.1684715300798416, "learning_rate": 9.04109589041096e-06, "loss": 1.1806, "step": 134 }, { "epoch": 0.37362637362637363, "grad_norm": 0.10095636546611786, "learning_rate": 9.178082191780823e-06, "loss": 1.176, "step": 136 }, { "epoch": 0.3791208791208791, "grad_norm": 0.11856956034898758, "learning_rate": 9.315068493150685e-06, "loss": 1.1849, "step": 138 }, { "epoch": 0.38461538461538464, "grad_norm": 0.19565819203853607, "learning_rate": 9.452054794520548e-06, "loss": 1.1803, "step": 140 }, { "epoch": 0.3901098901098901, "grad_norm": 0.15888364613056183, "learning_rate": 9.589041095890411e-06, "loss": 1.1756, "step": 142 }, { "epoch": 0.3956043956043956, "grad_norm": 0.16350317001342773, "learning_rate": 9.726027397260275e-06, "loss": 1.1777, "step": 144 }, { "epoch": 0.4010989010989011, "grad_norm": 0.13230757415294647, "learning_rate": 9.863013698630138e-06, "loss": 1.1843, "step": 146 }, { "epoch": 0.4010989010989011, "eval_loss": 1.1792316436767578, "eval_runtime": 299.2609, "eval_samples_per_second": 17.296, "eval_steps_per_second": 0.271, "step": 146 }, { "epoch": 0.4065934065934066, "grad_norm": 0.1627519279718399, "learning_rate": 1e-05, "loss": 1.174, "step": 148 }, { "epoch": 0.41208791208791207, "grad_norm": 0.13225838541984558, "learning_rate": 9.999942488284598e-06, "loss": 1.1779, "step": 150 }, { "epoch": 0.4175824175824176, "grad_norm": 0.15863117575645447, "learning_rate": 9.999769954461425e-06, "loss": 1.1767, "step": 152 }, { "epoch": 0.4230769230769231, "grad_norm": 0.15866652131080627, "learning_rate": 9.999482402499569e-06, "loss": 1.1758, "step": 154 }, { "epoch": 0.42857142857142855, "grad_norm": 0.1270761787891388, "learning_rate": 9.999079839014074e-06, "loss": 1.1756, "step": 156 }, { "epoch": 0.4340659340659341, "grad_norm": 0.10430735349655151, "learning_rate": 9.998562273265786e-06, "loss": 1.1728, "step": 158 }, { "epoch": 0.43956043956043955, "grad_norm": 0.13678255677223206, "learning_rate": 9.997929717161142e-06, "loss": 1.1774, "step": 160 }, { "epoch": 0.44505494505494503, "grad_norm": 0.13817106187343597, "learning_rate": 9.997182185251896e-06, "loss": 1.1804, "step": 162 }, { "epoch": 0.45054945054945056, "grad_norm": 0.15956689417362213, "learning_rate": 9.996319694734787e-06, "loss": 1.1802, "step": 164 }, { "epoch": 0.45604395604395603, "grad_norm": 0.2157265841960907, "learning_rate": 9.995342265451138e-06, "loss": 1.1803, "step": 166 }, { "epoch": 0.46153846153846156, "grad_norm": 0.17677265405654907, "learning_rate": 9.994249919886402e-06, "loss": 1.1764, "step": 168 }, { "epoch": 0.46703296703296704, "grad_norm": 0.16796888411045074, "learning_rate": 9.993042683169647e-06, "loss": 1.1702, "step": 170 }, { "epoch": 0.4725274725274725, "grad_norm": 0.16481636464595795, "learning_rate": 9.991720583072975e-06, "loss": 1.1844, "step": 172 }, { "epoch": 0.47802197802197804, "grad_norm": 0.19735674560070038, "learning_rate": 9.990283650010883e-06, "loss": 1.1791, "step": 174 }, { "epoch": 0.4835164835164835, "grad_norm": 0.15785232186317444, "learning_rate": 9.988731917039564e-06, "loss": 1.1783, "step": 176 }, { "epoch": 0.489010989010989, "grad_norm": 0.17257508635520935, "learning_rate": 9.98706541985615e-06, "loss": 1.1769, "step": 178 }, { "epoch": 0.4945054945054945, "grad_norm": 0.1660294085741043, "learning_rate": 9.985284196797884e-06, "loss": 1.181, "step": 180 }, { "epoch": 0.5, "grad_norm": 0.1156650260090828, "learning_rate": 9.983388288841246e-06, "loss": 1.1865, "step": 182 }, { "epoch": 0.5054945054945055, "grad_norm": 0.0954098179936409, "learning_rate": 9.981377739601002e-06, "loss": 1.1787, "step": 184 }, { "epoch": 0.510989010989011, "grad_norm": 0.0957237258553505, "learning_rate": 9.979252595329204e-06, "loss": 1.1731, "step": 186 }, { "epoch": 0.5164835164835165, "grad_norm": 0.11179114878177643, "learning_rate": 9.977012904914133e-06, "loss": 1.166, "step": 188 }, { "epoch": 0.521978021978022, "grad_norm": 0.12088882923126221, "learning_rate": 9.974658719879163e-06, "loss": 1.1706, "step": 190 }, { "epoch": 0.5274725274725275, "grad_norm": 0.1067051887512207, "learning_rate": 9.972190094381578e-06, "loss": 1.178, "step": 192 }, { "epoch": 0.532967032967033, "grad_norm": 0.09633362293243408, "learning_rate": 9.96960708521134e-06, "loss": 1.1763, "step": 194 }, { "epoch": 0.5384615384615384, "grad_norm": 0.11866485327482224, "learning_rate": 9.966909751789758e-06, "loss": 1.1777, "step": 196 }, { "epoch": 0.5439560439560439, "grad_norm": 0.11669634282588959, "learning_rate": 9.964098156168143e-06, "loss": 1.1753, "step": 198 }, { "epoch": 0.5494505494505495, "grad_norm": 0.1488696038722992, "learning_rate": 9.96117236302637e-06, "loss": 1.1656, "step": 200 }, { "epoch": 0.554945054945055, "grad_norm": 0.1773654669523239, "learning_rate": 9.958132439671392e-06, "loss": 1.1809, "step": 202 }, { "epoch": 0.5604395604395604, "grad_norm": 0.17689326405525208, "learning_rate": 9.954978456035695e-06, "loss": 1.1722, "step": 204 }, { "epoch": 0.5659340659340659, "grad_norm": 0.13418439030647278, "learning_rate": 9.951710484675677e-06, "loss": 1.167, "step": 206 }, { "epoch": 0.5714285714285714, "grad_norm": 0.14982537925243378, "learning_rate": 9.948328600769996e-06, "loss": 1.1713, "step": 208 }, { "epoch": 0.5769230769230769, "grad_norm": 0.12595325708389282, "learning_rate": 9.94483288211783e-06, "loss": 1.1677, "step": 210 }, { "epoch": 0.5824175824175825, "grad_norm": 0.10920244455337524, "learning_rate": 9.941223409137088e-06, "loss": 1.1727, "step": 212 }, { "epoch": 0.5879120879120879, "grad_norm": 0.10856951773166656, "learning_rate": 9.937500264862567e-06, "loss": 1.1739, "step": 214 }, { "epoch": 0.5934065934065934, "grad_norm": 0.13141585886478424, "learning_rate": 9.933663534944029e-06, "loss": 1.176, "step": 216 }, { "epoch": 0.5989010989010989, "grad_norm": 0.13502047955989838, "learning_rate": 9.929713307644245e-06, "loss": 1.178, "step": 218 }, { "epoch": 0.6043956043956044, "grad_norm": 0.1468881517648697, "learning_rate": 9.925649673836949e-06, "loss": 1.1819, "step": 220 }, { "epoch": 0.6098901098901099, "grad_norm": 0.13739052414894104, "learning_rate": 9.921472727004765e-06, "loss": 1.1784, "step": 222 }, { "epoch": 0.6153846153846154, "grad_norm": 0.13080830872058868, "learning_rate": 9.917182563237045e-06, "loss": 1.1762, "step": 224 }, { "epoch": 0.6208791208791209, "grad_norm": 0.11544028669595718, "learning_rate": 9.912779281227656e-06, "loss": 1.1733, "step": 226 }, { "epoch": 0.6263736263736264, "grad_norm": 0.14270079135894775, "learning_rate": 9.908262982272724e-06, "loss": 1.1812, "step": 228 }, { "epoch": 0.6318681318681318, "grad_norm": 0.11610660701990128, "learning_rate": 9.903633770268286e-06, "loss": 1.1803, "step": 230 }, { "epoch": 0.6373626373626373, "grad_norm": 0.10949360579252243, "learning_rate": 9.89889175170791e-06, "loss": 1.1776, "step": 232 }, { "epoch": 0.6428571428571429, "grad_norm": 0.11754804104566574, "learning_rate": 9.894037035680246e-06, "loss": 1.169, "step": 234 }, { "epoch": 0.6483516483516484, "grad_norm": 0.1328520029783249, "learning_rate": 9.889069733866515e-06, "loss": 1.1663, "step": 236 }, { "epoch": 0.6538461538461539, "grad_norm": 0.12216369807720184, "learning_rate": 9.883989960537934e-06, "loss": 1.1779, "step": 238 }, { "epoch": 0.6593406593406593, "grad_norm": 0.11219753324985504, "learning_rate": 9.878797832553093e-06, "loss": 1.1718, "step": 240 }, { "epoch": 0.6648351648351648, "grad_norm": 0.11396931111812592, "learning_rate": 9.873493469355271e-06, "loss": 1.175, "step": 242 }, { "epoch": 0.6703296703296703, "grad_norm": 0.104461669921875, "learning_rate": 9.868076992969672e-06, "loss": 1.1761, "step": 244 }, { "epoch": 0.6758241758241759, "grad_norm": 0.11892971396446228, "learning_rate": 9.862548528000644e-06, "loss": 1.1713, "step": 246 }, { "epoch": 0.6813186813186813, "grad_norm": 0.10305490344762802, "learning_rate": 9.85690820162878e-06, "loss": 1.1751, "step": 248 }, { "epoch": 0.6868131868131868, "grad_norm": 0.11825086921453476, "learning_rate": 9.851156143608025e-06, "loss": 1.1749, "step": 250 }, { "epoch": 0.6923076923076923, "grad_norm": 0.09655775129795074, "learning_rate": 9.845292486262664e-06, "loss": 1.1733, "step": 252 }, { "epoch": 0.6978021978021978, "grad_norm": 0.09013015031814575, "learning_rate": 9.839317364484295e-06, "loss": 1.1736, "step": 254 }, { "epoch": 0.7032967032967034, "grad_norm": 0.12731756269931793, "learning_rate": 9.83323091572872e-06, "loss": 1.1735, "step": 256 }, { "epoch": 0.7087912087912088, "grad_norm": 0.11326012015342712, "learning_rate": 9.827033280012783e-06, "loss": 1.1678, "step": 258 }, { "epoch": 0.7142857142857143, "grad_norm": 0.09636470675468445, "learning_rate": 9.820724599911147e-06, "loss": 1.1735, "step": 260 }, { "epoch": 0.7197802197802198, "grad_norm": 0.12517008185386658, "learning_rate": 9.81430502055302e-06, "loss": 1.1839, "step": 262 }, { "epoch": 0.7252747252747253, "grad_norm": 0.14088405668735504, "learning_rate": 9.807774689618806e-06, "loss": 1.1734, "step": 264 }, { "epoch": 0.7307692307692307, "grad_norm": 0.09721764922142029, "learning_rate": 9.801133757336726e-06, "loss": 1.1681, "step": 266 }, { "epoch": 0.7362637362637363, "grad_norm": 0.08106404542922974, "learning_rate": 9.794382376479334e-06, "loss": 1.1684, "step": 268 }, { "epoch": 0.7417582417582418, "grad_norm": 0.1210760697722435, "learning_rate": 9.787520702360035e-06, "loss": 1.1656, "step": 270 }, { "epoch": 0.7472527472527473, "grad_norm": 0.12102904915809631, "learning_rate": 9.780548892829486e-06, "loss": 1.1743, "step": 272 }, { "epoch": 0.7527472527472527, "grad_norm": 0.11343714594841003, "learning_rate": 9.773467108271978e-06, "loss": 1.1756, "step": 274 }, { "epoch": 0.7582417582417582, "grad_norm": 0.11999528110027313, "learning_rate": 9.766275511601742e-06, "loss": 1.1782, "step": 276 }, { "epoch": 0.7637362637362637, "grad_norm": 0.148450568318367, "learning_rate": 9.7589742682592e-06, "loss": 1.1796, "step": 278 }, { "epoch": 0.7692307692307693, "grad_norm": 0.09977131336927414, "learning_rate": 9.751563546207167e-06, "loss": 1.1686, "step": 280 }, { "epoch": 0.7747252747252747, "grad_norm": 0.12250227481126785, "learning_rate": 9.744043515926975e-06, "loss": 1.1673, "step": 282 }, { "epoch": 0.7802197802197802, "grad_norm": 0.12362032383680344, "learning_rate": 9.736414350414564e-06, "loss": 1.1707, "step": 284 }, { "epoch": 0.7857142857142857, "grad_norm": 0.09825081378221512, "learning_rate": 9.72867622517649e-06, "loss": 1.1719, "step": 286 }, { "epoch": 0.7912087912087912, "grad_norm": 0.13811303675174713, "learning_rate": 9.720829318225897e-06, "loss": 1.1711, "step": 288 }, { "epoch": 0.7967032967032966, "grad_norm": 0.14676132798194885, "learning_rate": 9.712873810078415e-06, "loss": 1.1728, "step": 290 }, { "epoch": 0.8021978021978022, "grad_norm": 0.09035024791955948, "learning_rate": 9.704809883748012e-06, "loss": 1.1776, "step": 292 }, { "epoch": 0.8021978021978022, "eval_loss": 1.1732887029647827, "eval_runtime": 298.7065, "eval_samples_per_second": 17.328, "eval_steps_per_second": 0.271, "step": 292 }, { "epoch": 0.8076923076923077, "grad_norm": 0.10824787616729736, "learning_rate": 9.696637724742785e-06, "loss": 1.1735, "step": 294 }, { "epoch": 0.8131868131868132, "grad_norm": 0.10512899607419968, "learning_rate": 9.688357521060685e-06, "loss": 1.1702, "step": 296 }, { "epoch": 0.8186813186813187, "grad_norm": 0.12007107585668564, "learning_rate": 9.6799694631852e-06, "loss": 1.1687, "step": 298 }, { "epoch": 0.8241758241758241, "grad_norm": 0.1161542534828186, "learning_rate": 9.67147374408097e-06, "loss": 1.1726, "step": 300 }, { "epoch": 0.8296703296703297, "grad_norm": 0.16156896948814392, "learning_rate": 9.662870559189344e-06, "loss": 1.1639, "step": 302 }, { "epoch": 0.8351648351648352, "grad_norm": 0.12350093573331833, "learning_rate": 9.654160106423891e-06, "loss": 1.1711, "step": 304 }, { "epoch": 0.8406593406593407, "grad_norm": 0.1292959749698639, "learning_rate": 9.645342586165845e-06, "loss": 1.1677, "step": 306 }, { "epoch": 0.8461538461538461, "grad_norm": 0.09656315296888351, "learning_rate": 9.63641820125949e-06, "loss": 1.1773, "step": 308 }, { "epoch": 0.8516483516483516, "grad_norm": 0.13036847114562988, "learning_rate": 9.627387157007502e-06, "loss": 1.1819, "step": 310 }, { "epoch": 0.8571428571428571, "grad_norm": 0.12335841357707977, "learning_rate": 9.618249661166218e-06, "loss": 1.1688, "step": 312 }, { "epoch": 0.8626373626373627, "grad_norm": 0.13838279247283936, "learning_rate": 9.609005923940865e-06, "loss": 1.1665, "step": 314 }, { "epoch": 0.8681318681318682, "grad_norm": 0.13689468801021576, "learning_rate": 9.599656157980715e-06, "loss": 1.1696, "step": 316 }, { "epoch": 0.8736263736263736, "grad_norm": 0.13239113986492157, "learning_rate": 9.590200578374198e-06, "loss": 1.1686, "step": 318 }, { "epoch": 0.8791208791208791, "grad_norm": 0.09575112909078598, "learning_rate": 9.580639402643957e-06, "loss": 1.1673, "step": 320 }, { "epoch": 0.8846153846153846, "grad_norm": 0.11522715538740158, "learning_rate": 9.570972850741839e-06, "loss": 1.1685, "step": 322 }, { "epoch": 0.8901098901098901, "grad_norm": 0.14126405119895935, "learning_rate": 9.561201145043835e-06, "loss": 1.1788, "step": 324 }, { "epoch": 0.8956043956043956, "grad_norm": 0.17362023890018463, "learning_rate": 9.551324510344972e-06, "loss": 1.1746, "step": 326 }, { "epoch": 0.9010989010989011, "grad_norm": 0.1431788206100464, "learning_rate": 9.541343173854128e-06, "loss": 1.1719, "step": 328 }, { "epoch": 0.9065934065934066, "grad_norm": 0.11747945845127106, "learning_rate": 9.531257365188818e-06, "loss": 1.1763, "step": 330 }, { "epoch": 0.9120879120879121, "grad_norm": 0.10873839259147644, "learning_rate": 9.521067316369903e-06, "loss": 1.173, "step": 332 }, { "epoch": 0.9175824175824175, "grad_norm": 0.10153867304325104, "learning_rate": 9.510773261816261e-06, "loss": 1.1681, "step": 334 }, { "epoch": 0.9230769230769231, "grad_norm": 0.07133303582668304, "learning_rate": 9.500375438339384e-06, "loss": 1.1643, "step": 336 }, { "epoch": 0.9285714285714286, "grad_norm": 0.09183746576309204, "learning_rate": 9.48987408513794e-06, "loss": 1.1788, "step": 338 }, { "epoch": 0.9340659340659341, "grad_norm": 0.10790558159351349, "learning_rate": 9.47926944379226e-06, "loss": 1.1817, "step": 340 }, { "epoch": 0.9395604395604396, "grad_norm": 0.09112295508384705, "learning_rate": 9.468561758258795e-06, "loss": 1.1674, "step": 342 }, { "epoch": 0.945054945054945, "grad_norm": 0.08475514501333237, "learning_rate": 9.457751274864486e-06, "loss": 1.1706, "step": 344 }, { "epoch": 0.9505494505494505, "grad_norm": 0.10173246264457703, "learning_rate": 9.446838242301113e-06, "loss": 1.1644, "step": 346 }, { "epoch": 0.9560439560439561, "grad_norm": 0.10165045410394669, "learning_rate": 9.435822911619564e-06, "loss": 1.1694, "step": 348 }, { "epoch": 0.9615384615384616, "grad_norm": 0.10506884753704071, "learning_rate": 9.424705536224065e-06, "loss": 1.1708, "step": 350 }, { "epoch": 0.967032967032967, "grad_norm": 0.10648410022258759, "learning_rate": 9.41348637186635e-06, "loss": 1.1712, "step": 352 }, { "epoch": 0.9725274725274725, "grad_norm": 0.11728779226541519, "learning_rate": 9.40216567663977e-06, "loss": 1.1838, "step": 354 }, { "epoch": 0.978021978021978, "grad_norm": 0.12955018877983093, "learning_rate": 9.390743710973366e-06, "loss": 1.1773, "step": 356 }, { "epoch": 0.9835164835164835, "grad_norm": 0.14313377439975739, "learning_rate": 9.379220737625877e-06, "loss": 1.1693, "step": 358 }, { "epoch": 0.989010989010989, "grad_norm": 0.17154935002326965, "learning_rate": 9.367597021679686e-06, "loss": 1.1674, "step": 360 }, { "epoch": 0.9945054945054945, "grad_norm": 0.13390299677848816, "learning_rate": 9.35587283053473e-06, "loss": 1.1636, "step": 362 }, { "epoch": 1.0, "grad_norm": 0.13049638271331787, "learning_rate": 9.344048433902351e-06, "loss": 1.1678, "step": 364 }, { "epoch": 1.0054945054945055, "grad_norm": 0.11013516038656235, "learning_rate": 9.332124103799075e-06, "loss": 1.1627, "step": 366 }, { "epoch": 1.010989010989011, "grad_norm": 0.11757113039493561, "learning_rate": 9.320100114540382e-06, "loss": 1.1679, "step": 368 }, { "epoch": 1.0164835164835164, "grad_norm": 0.11947564780712128, "learning_rate": 9.307976742734366e-06, "loss": 1.1767, "step": 370 }, { "epoch": 1.021978021978022, "grad_norm": 0.1045905202627182, "learning_rate": 9.295754267275393e-06, "loss": 1.1694, "step": 372 }, { "epoch": 1.0274725274725274, "grad_norm": 0.10813715308904648, "learning_rate": 9.283432969337672e-06, "loss": 1.1752, "step": 374 }, { "epoch": 1.032967032967033, "grad_norm": 0.11131364107131958, "learning_rate": 9.271013132368799e-06, "loss": 1.1714, "step": 376 }, { "epoch": 1.0384615384615385, "grad_norm": 0.09973058849573135, "learning_rate": 9.258495042083222e-06, "loss": 1.1696, "step": 378 }, { "epoch": 1.043956043956044, "grad_norm": 0.10416287928819656, "learning_rate": 9.245878986455684e-06, "loss": 1.1641, "step": 380 }, { "epoch": 1.0494505494505495, "grad_norm": 0.08854486793279648, "learning_rate": 9.23316525571458e-06, "loss": 1.1649, "step": 382 }, { "epoch": 1.054945054945055, "grad_norm": 0.1364666223526001, "learning_rate": 9.2203541423353e-06, "loss": 1.1769, "step": 384 }, { "epoch": 1.0604395604395604, "grad_norm": 0.14936769008636475, "learning_rate": 9.207445941033483e-06, "loss": 1.1686, "step": 386 }, { "epoch": 1.065934065934066, "grad_norm": 0.18906323611736298, "learning_rate": 9.19444094875825e-06, "loss": 1.1692, "step": 388 }, { "epoch": 1.0714285714285714, "grad_norm": 0.44419270753860474, "learning_rate": 9.18133946468537e-06, "loss": 1.1746, "step": 390 }, { "epoch": 1.0769230769230769, "grad_norm": 0.1541990488767624, "learning_rate": 9.16814179021037e-06, "loss": 1.1735, "step": 392 }, { "epoch": 1.0824175824175823, "grad_norm": 0.10936509072780609, "learning_rate": 9.154848228941607e-06, "loss": 1.1734, "step": 394 }, { "epoch": 1.0879120879120878, "grad_norm": 0.0885339230298996, "learning_rate": 9.14145908669329e-06, "loss": 1.1702, "step": 396 }, { "epoch": 1.0934065934065935, "grad_norm": 0.07241977006196976, "learning_rate": 9.127974671478432e-06, "loss": 1.1714, "step": 398 }, { "epoch": 1.098901098901099, "grad_norm": 0.08955956995487213, "learning_rate": 9.114395293501775e-06, "loss": 1.1712, "step": 400 }, { "epoch": 1.1043956043956045, "grad_norm": 0.11948683112859726, "learning_rate": 9.100721265152644e-06, "loss": 1.1799, "step": 402 }, { "epoch": 1.10989010989011, "grad_norm": 0.11843076348304749, "learning_rate": 9.086952900997774e-06, "loss": 1.1716, "step": 404 }, { "epoch": 1.1153846153846154, "grad_norm": 0.08606066554784775, "learning_rate": 9.073090517774057e-06, "loss": 1.1777, "step": 406 }, { "epoch": 1.120879120879121, "grad_norm": 0.08759862184524536, "learning_rate": 9.059134434381274e-06, "loss": 1.1698, "step": 408 }, { "epoch": 1.1263736263736264, "grad_norm": 0.10086913406848907, "learning_rate": 9.045084971874738e-06, "loss": 1.1641, "step": 410 }, { "epoch": 1.1318681318681318, "grad_norm": 0.11529091000556946, "learning_rate": 9.030942453457928e-06, "loss": 1.1655, "step": 412 }, { "epoch": 1.1373626373626373, "grad_norm": 0.0781669095158577, "learning_rate": 9.01670720447504e-06, "loss": 1.1678, "step": 414 }, { "epoch": 1.1428571428571428, "grad_norm": 0.09989538788795471, "learning_rate": 9.00237955240351e-06, "loss": 1.1689, "step": 416 }, { "epoch": 1.1483516483516483, "grad_norm": 0.12073778361082077, "learning_rate": 8.987959826846479e-06, "loss": 1.1747, "step": 418 }, { "epoch": 1.1538461538461537, "grad_norm": 0.12605726718902588, "learning_rate": 8.973448359525207e-06, "loss": 1.1725, "step": 420 }, { "epoch": 1.1593406593406592, "grad_norm": 0.12503278255462646, "learning_rate": 8.958845484271443e-06, "loss": 1.1709, "step": 422 }, { "epoch": 1.164835164835165, "grad_norm": 0.10487185418605804, "learning_rate": 8.944151537019752e-06, "loss": 1.1566, "step": 424 }, { "epoch": 1.1703296703296704, "grad_norm": 0.09222570061683655, "learning_rate": 8.929366855799777e-06, "loss": 1.1808, "step": 426 }, { "epoch": 1.1758241758241759, "grad_norm": 0.10268251597881317, "learning_rate": 8.914491780728471e-06, "loss": 1.1782, "step": 428 }, { "epoch": 1.1813186813186813, "grad_norm": 0.08701249957084656, "learning_rate": 8.899526654002268e-06, "loss": 1.1716, "step": 430 }, { "epoch": 1.1868131868131868, "grad_norm": 0.11233299225568771, "learning_rate": 8.88447181988921e-06, "loss": 1.1699, "step": 432 }, { "epoch": 1.1923076923076923, "grad_norm": 0.09972134232521057, "learning_rate": 8.869327624721033e-06, "loss": 1.1687, "step": 434 }, { "epoch": 1.1978021978021978, "grad_norm": 0.10448331385850906, "learning_rate": 8.854094416885192e-06, "loss": 1.1656, "step": 436 }, { "epoch": 1.2032967032967032, "grad_norm": 0.11365893483161926, "learning_rate": 8.838772546816857e-06, "loss": 1.1764, "step": 438 }, { "epoch": 1.2032967032967032, "eval_loss": 1.1709927320480347, "eval_runtime": 299.4182, "eval_samples_per_second": 17.287, "eval_steps_per_second": 0.271, "step": 438 }, { "epoch": 1.2087912087912087, "grad_norm": 0.1238979920744896, "learning_rate": 8.823362366990833e-06, "loss": 1.1718, "step": 440 }, { "epoch": 1.2142857142857142, "grad_norm": 0.12039055675268173, "learning_rate": 8.807864231913475e-06, "loss": 1.1686, "step": 442 }, { "epoch": 1.2197802197802199, "grad_norm": 0.09191784262657166, "learning_rate": 8.792278498114517e-06, "loss": 1.1702, "step": 444 }, { "epoch": 1.2252747252747254, "grad_norm": 0.08927666395902634, "learning_rate": 8.77660552413887e-06, "loss": 1.1761, "step": 446 }, { "epoch": 1.2307692307692308, "grad_norm": 0.08614211529493332, "learning_rate": 8.760845670538387e-06, "loss": 1.1743, "step": 448 }, { "epoch": 1.2362637362637363, "grad_norm": 0.10282223671674728, "learning_rate": 8.744999299863549e-06, "loss": 1.1705, "step": 450 }, { "epoch": 1.2417582417582418, "grad_norm": 0.10722998529672623, "learning_rate": 8.729066776655144e-06, "loss": 1.169, "step": 452 }, { "epoch": 1.2472527472527473, "grad_norm": 0.09755399823188782, "learning_rate": 8.713048467435865e-06, "loss": 1.1732, "step": 454 }, { "epoch": 1.2527472527472527, "grad_norm": 0.09675360471010208, "learning_rate": 8.696944740701891e-06, "loss": 1.1696, "step": 456 }, { "epoch": 1.2582417582417582, "grad_norm": 0.08405736833810806, "learning_rate": 8.6807559669144e-06, "loss": 1.1678, "step": 458 }, { "epoch": 1.2637362637362637, "grad_norm": 0.08441252261400223, "learning_rate": 8.664482518491053e-06, "loss": 1.1684, "step": 460 }, { "epoch": 1.2692307692307692, "grad_norm": 0.11046714335680008, "learning_rate": 8.648124769797424e-06, "loss": 1.1747, "step": 462 }, { "epoch": 1.2747252747252746, "grad_norm": 0.10532383620738983, "learning_rate": 8.631683097138386e-06, "loss": 1.1728, "step": 464 }, { "epoch": 1.2802197802197801, "grad_norm": 0.09521500766277313, "learning_rate": 8.615157878749462e-06, "loss": 1.169, "step": 466 }, { "epoch": 1.2857142857142856, "grad_norm": 0.09845519065856934, "learning_rate": 8.598549494788111e-06, "loss": 1.1667, "step": 468 }, { "epoch": 1.2912087912087913, "grad_norm": 0.10502909123897552, "learning_rate": 8.581858327324996e-06, "loss": 1.1709, "step": 470 }, { "epoch": 1.2967032967032968, "grad_norm": 0.0786222293972969, "learning_rate": 8.565084760335188e-06, "loss": 1.165, "step": 472 }, { "epoch": 1.3021978021978022, "grad_norm": 0.10599285364151001, "learning_rate": 8.548229179689325e-06, "loss": 1.1718, "step": 474 }, { "epoch": 1.3076923076923077, "grad_norm": 0.11965076625347137, "learning_rate": 8.531291973144755e-06, "loss": 1.1683, "step": 476 }, { "epoch": 1.3131868131868132, "grad_norm": 0.10626693069934845, "learning_rate": 8.5142735303366e-06, "loss": 1.1676, "step": 478 }, { "epoch": 1.3186813186813187, "grad_norm": 0.10863650590181351, "learning_rate": 8.497174242768792e-06, "loss": 1.17, "step": 480 }, { "epoch": 1.3241758241758241, "grad_norm": 0.11272372305393219, "learning_rate": 8.479994503805079e-06, "loss": 1.1674, "step": 482 }, { "epoch": 1.3296703296703296, "grad_norm": 0.1116040050983429, "learning_rate": 8.462734708659959e-06, "loss": 1.1711, "step": 484 }, { "epoch": 1.335164835164835, "grad_norm": 0.1068933829665184, "learning_rate": 8.445395254389605e-06, "loss": 1.179, "step": 486 }, { "epoch": 1.3406593406593408, "grad_norm": 0.11034112423658371, "learning_rate": 8.427976539882725e-06, "loss": 1.1617, "step": 488 }, { "epoch": 1.3461538461538463, "grad_norm": 0.11086605489253998, "learning_rate": 8.410478965851371e-06, "loss": 1.171, "step": 490 }, { "epoch": 1.3516483516483517, "grad_norm": 0.1053035780787468, "learning_rate": 8.39290293482175e-06, "loss": 1.1696, "step": 492 }, { "epoch": 1.3571428571428572, "grad_norm": 0.11563800275325775, "learning_rate": 8.375248851124937e-06, "loss": 1.1817, "step": 494 }, { "epoch": 1.3626373626373627, "grad_norm": 0.09770724177360535, "learning_rate": 8.357517120887586e-06, "loss": 1.1726, "step": 496 }, { "epoch": 1.3681318681318682, "grad_norm": 0.10480888187885284, "learning_rate": 8.339708152022586e-06, "loss": 1.1716, "step": 498 }, { "epoch": 1.3736263736263736, "grad_norm": 0.0983305424451828, "learning_rate": 8.321822354219677e-06, "loss": 1.1708, "step": 500 }, { "epoch": 1.379120879120879, "grad_norm": 0.11572114378213882, "learning_rate": 8.303860138936027e-06, "loss": 1.1669, "step": 502 }, { "epoch": 1.3846153846153846, "grad_norm": 0.15876273810863495, "learning_rate": 8.285821919386758e-06, "loss": 1.172, "step": 504 }, { "epoch": 1.39010989010989, "grad_norm": 0.09302208572626114, "learning_rate": 8.267708110535449e-06, "loss": 1.1689, "step": 506 }, { "epoch": 1.3956043956043955, "grad_norm": 0.0982600748538971, "learning_rate": 8.24951912908459e-06, "loss": 1.1683, "step": 508 }, { "epoch": 1.401098901098901, "grad_norm": 0.1328704059123993, "learning_rate": 8.231255393465993e-06, "loss": 1.1686, "step": 510 }, { "epoch": 1.4065934065934065, "grad_norm": 0.1403021216392517, "learning_rate": 8.21291732383116e-06, "loss": 1.1654, "step": 512 }, { "epoch": 1.412087912087912, "grad_norm": 0.10671450942754745, "learning_rate": 8.19450534204163e-06, "loss": 1.1671, "step": 514 }, { "epoch": 1.4175824175824177, "grad_norm": 0.10965994000434875, "learning_rate": 8.176019871659263e-06, "loss": 1.1791, "step": 516 }, { "epoch": 1.4230769230769231, "grad_norm": 0.09535852819681168, "learning_rate": 8.157461337936506e-06, "loss": 1.1654, "step": 518 }, { "epoch": 1.4285714285714286, "grad_norm": 0.09910845011472702, "learning_rate": 8.138830167806601e-06, "loss": 1.1612, "step": 520 }, { "epoch": 1.434065934065934, "grad_norm": 0.08044712990522385, "learning_rate": 8.120126789873775e-06, "loss": 1.169, "step": 522 }, { "epoch": 1.4395604395604396, "grad_norm": 0.11208353191614151, "learning_rate": 8.10135163440336e-06, "loss": 1.1676, "step": 524 }, { "epoch": 1.445054945054945, "grad_norm": 0.09039778262376785, "learning_rate": 8.08250513331192e-06, "loss": 1.1647, "step": 526 }, { "epoch": 1.4505494505494505, "grad_norm": 0.08799196779727936, "learning_rate": 8.063587720157298e-06, "loss": 1.1594, "step": 528 }, { "epoch": 1.456043956043956, "grad_norm": 0.09703148156404495, "learning_rate": 8.044599830128643e-06, "loss": 1.1751, "step": 530 }, { "epoch": 1.4615384615384617, "grad_norm": 0.10245784372091293, "learning_rate": 8.02554190003641e-06, "loss": 1.1746, "step": 532 }, { "epoch": 1.4670329670329672, "grad_norm": 0.09949609637260437, "learning_rate": 8.006414368302297e-06, "loss": 1.1696, "step": 534 }, { "epoch": 1.4725274725274726, "grad_norm": 0.11576993763446808, "learning_rate": 7.98721767494917e-06, "loss": 1.1662, "step": 536 }, { "epoch": 1.478021978021978, "grad_norm": 0.09676701575517654, "learning_rate": 7.967952261590936e-06, "loss": 1.1652, "step": 538 }, { "epoch": 1.4835164835164836, "grad_norm": 0.10620467364788055, "learning_rate": 7.94861857142238e-06, "loss": 1.1659, "step": 540 }, { "epoch": 1.489010989010989, "grad_norm": 0.09917207062244415, "learning_rate": 7.929217049208977e-06, "loss": 1.175, "step": 542 }, { "epoch": 1.4945054945054945, "grad_norm": 0.131545290350914, "learning_rate": 7.90974814127666e-06, "loss": 1.1619, "step": 544 }, { "epoch": 1.5, "grad_norm": 0.16173675656318665, "learning_rate": 7.890212295501542e-06, "loss": 1.1718, "step": 546 }, { "epoch": 1.5054945054945055, "grad_norm": 0.11190006136894226, "learning_rate": 7.870609961299627e-06, "loss": 1.1707, "step": 548 }, { "epoch": 1.510989010989011, "grad_norm": 0.0947834774851799, "learning_rate": 7.850941589616458e-06, "loss": 1.1703, "step": 550 }, { "epoch": 1.5164835164835164, "grad_norm": 0.12736476957798004, "learning_rate": 7.831207632916757e-06, "loss": 1.1743, "step": 552 }, { "epoch": 1.521978021978022, "grad_norm": 0.17213790118694305, "learning_rate": 7.811408545174001e-06, "loss": 1.1763, "step": 554 }, { "epoch": 1.5274725274725274, "grad_norm": 0.26195263862609863, "learning_rate": 7.791544781859993e-06, "loss": 1.1741, "step": 556 }, { "epoch": 1.5329670329670328, "grad_norm": 0.1335473209619522, "learning_rate": 7.771616799934372e-06, "loss": 1.169, "step": 558 }, { "epoch": 1.5384615384615383, "grad_norm": 0.11248943954706192, "learning_rate": 7.751625057834107e-06, "loss": 1.1611, "step": 560 }, { "epoch": 1.5439560439560438, "grad_norm": 0.12497013807296753, "learning_rate": 7.731570015462953e-06, "loss": 1.1657, "step": 562 }, { "epoch": 1.5494505494505495, "grad_norm": 0.10039670765399933, "learning_rate": 7.711452134180865e-06, "loss": 1.1689, "step": 564 }, { "epoch": 1.554945054945055, "grad_norm": 0.09104856103658676, "learning_rate": 7.691271876793387e-06, "loss": 1.1691, "step": 566 }, { "epoch": 1.5604395604395604, "grad_norm": 0.10671708732843399, "learning_rate": 7.67102970754101e-06, "loss": 1.1622, "step": 568 }, { "epoch": 1.565934065934066, "grad_norm": 0.10622277110815048, "learning_rate": 7.65072609208848e-06, "loss": 1.1716, "step": 570 }, { "epoch": 1.5714285714285714, "grad_norm": 0.09694940596818924, "learning_rate": 7.630361497514104e-06, "loss": 1.1681, "step": 572 }, { "epoch": 1.5769230769230769, "grad_norm": 0.11600632965564728, "learning_rate": 7.6099363922989845e-06, "loss": 1.169, "step": 574 }, { "epoch": 1.5824175824175826, "grad_norm": 0.10654427111148834, "learning_rate": 7.5894512463162595e-06, "loss": 1.1711, "step": 576 }, { "epoch": 1.587912087912088, "grad_norm": 0.08246386051177979, "learning_rate": 7.568906530820281e-06, "loss": 1.1661, "step": 578 }, { "epoch": 1.5934065934065935, "grad_norm": 0.09424301981925964, "learning_rate": 7.5483027184357825e-06, "loss": 1.166, "step": 580 }, { "epoch": 1.598901098901099, "grad_norm": 0.08908534795045853, "learning_rate": 7.527640283147003e-06, "loss": 1.1698, "step": 582 }, { "epoch": 1.6043956043956045, "grad_norm": 0.10692695528268814, "learning_rate": 7.50691970028678e-06, "loss": 1.1701, "step": 584 }, { "epoch": 1.6043956043956045, "eval_loss": 1.1695584058761597, "eval_runtime": 299.2734, "eval_samples_per_second": 17.295, "eval_steps_per_second": 0.271, "step": 584 }, { "epoch": 1.60989010989011, "grad_norm": 0.09625241905450821, "learning_rate": 7.486141446525619e-06, "loss": 1.1734, "step": 586 }, { "epoch": 1.6153846153846154, "grad_norm": 0.09392037242650986, "learning_rate": 7.465305999860728e-06, "loss": 1.1698, "step": 588 }, { "epoch": 1.620879120879121, "grad_norm": 0.08401469886302948, "learning_rate": 7.444413839605017e-06, "loss": 1.1659, "step": 590 }, { "epoch": 1.6263736263736264, "grad_norm": 0.11073325574398041, "learning_rate": 7.423465446376079e-06, "loss": 1.166, "step": 592 }, { "epoch": 1.6318681318681318, "grad_norm": 0.09533084183931351, "learning_rate": 7.402461302085121e-06, "loss": 1.1734, "step": 594 }, { "epoch": 1.6373626373626373, "grad_norm": 0.08371090888977051, "learning_rate": 7.381401889925894e-06, "loss": 1.167, "step": 596 }, { "epoch": 1.6428571428571428, "grad_norm": 0.0968126654624939, "learning_rate": 7.360287694363566e-06, "loss": 1.1679, "step": 598 }, { "epoch": 1.6483516483516483, "grad_norm": 0.09040292352437973, "learning_rate": 7.3391192011235764e-06, "loss": 1.1754, "step": 600 }, { "epoch": 1.6538461538461537, "grad_norm": 0.09428130835294724, "learning_rate": 7.317896897180472e-06, "loss": 1.1713, "step": 602 }, { "epoch": 1.6593406593406592, "grad_norm": 0.08703873306512833, "learning_rate": 7.296621270746691e-06, "loss": 1.1688, "step": 604 }, { "epoch": 1.6648351648351647, "grad_norm": 0.09342585504055023, "learning_rate": 7.275292811261346e-06, "loss": 1.1651, "step": 606 }, { "epoch": 1.6703296703296702, "grad_norm": 0.11370475590229034, "learning_rate": 7.253912009378953e-06, "loss": 1.1721, "step": 608 }, { "epoch": 1.6758241758241759, "grad_norm": 0.0810871347784996, "learning_rate": 7.2324793569581474e-06, "loss": 1.174, "step": 610 }, { "epoch": 1.6813186813186813, "grad_norm": 0.09331346303224564, "learning_rate": 7.210995347050372e-06, "loss": 1.1691, "step": 612 }, { "epoch": 1.6868131868131868, "grad_norm": 0.1314326524734497, "learning_rate": 7.189460473888535e-06, "loss": 1.1672, "step": 614 }, { "epoch": 1.6923076923076923, "grad_norm": 0.12525762617588043, "learning_rate": 7.167875232875632e-06, "loss": 1.1645, "step": 616 }, { "epoch": 1.6978021978021978, "grad_norm": 0.13776274025440216, "learning_rate": 7.146240120573358e-06, "loss": 1.1693, "step": 618 }, { "epoch": 1.7032967032967035, "grad_norm": 0.1336820125579834, "learning_rate": 7.124555634690684e-06, "loss": 1.1649, "step": 620 }, { "epoch": 1.708791208791209, "grad_norm": 0.12508928775787354, "learning_rate": 7.1028222740724e-06, "loss": 1.169, "step": 622 }, { "epoch": 1.7142857142857144, "grad_norm": 0.10577461123466492, "learning_rate": 7.081040538687649e-06, "loss": 1.1575, "step": 624 }, { "epoch": 1.7197802197802199, "grad_norm": 0.11875636875629425, "learning_rate": 7.059210929618416e-06, "loss": 1.1666, "step": 626 }, { "epoch": 1.7252747252747254, "grad_norm": 0.10840770602226257, "learning_rate": 7.037333949048005e-06, "loss": 1.1664, "step": 628 }, { "epoch": 1.7307692307692308, "grad_norm": 0.11352021992206573, "learning_rate": 7.0154101002494914e-06, "loss": 1.1643, "step": 630 }, { "epoch": 1.7362637362637363, "grad_norm": 0.08095169812440872, "learning_rate": 6.993439887574133e-06, "loss": 1.1691, "step": 632 }, { "epoch": 1.7417582417582418, "grad_norm": 0.09175344556570053, "learning_rate": 6.971423816439782e-06, "loss": 1.1664, "step": 634 }, { "epoch": 1.7472527472527473, "grad_norm": 0.07321044057607651, "learning_rate": 6.949362393319239e-06, "loss": 1.1649, "step": 636 }, { "epoch": 1.7527472527472527, "grad_norm": 0.10566896945238113, "learning_rate": 6.927256125728624e-06, "loss": 1.167, "step": 638 }, { "epoch": 1.7582417582417582, "grad_norm": 0.10230179876089096, "learning_rate": 6.905105522215684e-06, "loss": 1.1569, "step": 640 }, { "epoch": 1.7637362637362637, "grad_norm": 0.09293391555547714, "learning_rate": 6.8829110923481e-06, "loss": 1.1677, "step": 642 }, { "epoch": 1.7692307692307692, "grad_norm": 0.08087108284235, "learning_rate": 6.8606733467017675e-06, "loss": 1.1691, "step": 644 }, { "epoch": 1.7747252747252746, "grad_norm": 0.08191601186990738, "learning_rate": 6.838392796849042e-06, "loss": 1.1683, "step": 646 }, { "epoch": 1.7802197802197801, "grad_norm": 0.08746813237667084, "learning_rate": 6.816069955346986e-06, "loss": 1.1643, "step": 648 }, { "epoch": 1.7857142857142856, "grad_norm": 0.10125511139631271, "learning_rate": 6.7937053357255585e-06, "loss": 1.1716, "step": 650 }, { "epoch": 1.791208791208791, "grad_norm": 0.08800845593214035, "learning_rate": 6.771299452475818e-06, "loss": 1.1721, "step": 652 }, { "epoch": 1.7967032967032965, "grad_norm": 0.09477917104959488, "learning_rate": 6.748852821038075e-06, "loss": 1.1649, "step": 654 }, { "epoch": 1.8021978021978022, "grad_norm": 0.10864371061325073, "learning_rate": 6.7263659577900375e-06, "loss": 1.1621, "step": 656 }, { "epoch": 1.8076923076923077, "grad_norm": 0.07818944752216339, "learning_rate": 6.703839380034945e-06, "loss": 1.1667, "step": 658 }, { "epoch": 1.8131868131868132, "grad_norm": 0.0869673416018486, "learning_rate": 6.681273605989643e-06, "loss": 1.1692, "step": 660 }, { "epoch": 1.8186813186813187, "grad_norm": 0.10159242898225784, "learning_rate": 6.6586691547726855e-06, "loss": 1.1708, "step": 662 }, { "epoch": 1.8241758241758241, "grad_norm": 0.20149534940719604, "learning_rate": 6.636026546392374e-06, "loss": 1.175, "step": 664 }, { "epoch": 1.8296703296703298, "grad_norm": 0.35707324743270874, "learning_rate": 6.613346301734813e-06, "loss": 1.1556, "step": 666 }, { "epoch": 1.8351648351648353, "grad_norm": 0.10949152708053589, "learning_rate": 6.590628942551909e-06, "loss": 1.1696, "step": 668 }, { "epoch": 1.8406593406593408, "grad_norm": 0.07581349462270737, "learning_rate": 6.567874991449383e-06, "loss": 1.1701, "step": 670 }, { "epoch": 1.8461538461538463, "grad_norm": 0.11389974504709244, "learning_rate": 6.545084971874738e-06, "loss": 1.1721, "step": 672 }, { "epoch": 1.8516483516483517, "grad_norm": 0.10378382354974747, "learning_rate": 6.522259408105223e-06, "loss": 1.1674, "step": 674 }, { "epoch": 1.8571428571428572, "grad_norm": 0.09723298251628876, "learning_rate": 6.499398825235767e-06, "loss": 1.1696, "step": 676 }, { "epoch": 1.8626373626373627, "grad_norm": 0.10554076731204987, "learning_rate": 6.476503749166903e-06, "loss": 1.1674, "step": 678 }, { "epoch": 1.8681318681318682, "grad_norm": 0.09231211990118027, "learning_rate": 6.453574706592676e-06, "loss": 1.1652, "step": 680 }, { "epoch": 1.8736263736263736, "grad_norm": 0.09046723693609238, "learning_rate": 6.4306122249885105e-06, "loss": 1.1716, "step": 682 }, { "epoch": 1.879120879120879, "grad_norm": 0.09619590640068054, "learning_rate": 6.407616832599091e-06, "loss": 1.1699, "step": 684 }, { "epoch": 1.8846153846153846, "grad_norm": 0.08878965675830841, "learning_rate": 6.384589058426201e-06, "loss": 1.1701, "step": 686 }, { "epoch": 1.89010989010989, "grad_norm": 0.0882963240146637, "learning_rate": 6.36152943221656e-06, "loss": 1.1757, "step": 688 }, { "epoch": 1.8956043956043955, "grad_norm": 0.10285267233848572, "learning_rate": 6.338438484449632e-06, "loss": 1.1612, "step": 690 }, { "epoch": 1.901098901098901, "grad_norm": 0.08342345803976059, "learning_rate": 6.31531674632542e-06, "loss": 1.1696, "step": 692 }, { "epoch": 1.9065934065934065, "grad_norm": 0.09322002530097961, "learning_rate": 6.292164749752256e-06, "loss": 1.1733, "step": 694 }, { "epoch": 1.912087912087912, "grad_norm": 0.10594601929187775, "learning_rate": 6.268983027334557e-06, "loss": 1.1675, "step": 696 }, { "epoch": 1.9175824175824174, "grad_norm": 0.10415042191743851, "learning_rate": 6.245772112360568e-06, "loss": 1.1706, "step": 698 }, { "epoch": 1.9230769230769231, "grad_norm": 0.08474233746528625, "learning_rate": 6.222532538790107e-06, "loss": 1.1717, "step": 700 }, { "epoch": 1.9285714285714286, "grad_norm": 0.09690573066473007, "learning_rate": 6.199264841242267e-06, "loss": 1.1579, "step": 702 }, { "epoch": 1.934065934065934, "grad_norm": 0.09174709767103195, "learning_rate": 6.17596955498313e-06, "loss": 1.1707, "step": 704 }, { "epoch": 1.9395604395604396, "grad_norm": 0.09256250411272049, "learning_rate": 6.1526472159134454e-06, "loss": 1.1675, "step": 706 }, { "epoch": 1.945054945054945, "grad_norm": 0.115385040640831, "learning_rate": 6.129298360556304e-06, "loss": 1.1666, "step": 708 }, { "epoch": 1.9505494505494505, "grad_norm": 0.09870624542236328, "learning_rate": 6.105923526044794e-06, "loss": 1.1712, "step": 710 }, { "epoch": 1.9560439560439562, "grad_norm": 0.09220367670059204, "learning_rate": 6.08252325010965e-06, "loss": 1.1702, "step": 712 }, { "epoch": 1.9615384615384617, "grad_norm": 0.08229216188192368, "learning_rate": 6.059098071066874e-06, "loss": 1.1641, "step": 714 }, { "epoch": 1.9670329670329672, "grad_norm": 0.08601760119199753, "learning_rate": 6.035648527805359e-06, "loss": 1.1658, "step": 716 }, { "epoch": 1.9725274725274726, "grad_norm": 0.09920581430196762, "learning_rate": 6.012175159774488e-06, "loss": 1.1627, "step": 718 }, { "epoch": 1.978021978021978, "grad_norm": 0.09022627770900726, "learning_rate": 5.988678506971726e-06, "loss": 1.1723, "step": 720 }, { "epoch": 1.9835164835164836, "grad_norm": 0.08053141087293625, "learning_rate": 5.965159109930196e-06, "loss": 1.1785, "step": 722 }, { "epoch": 1.989010989010989, "grad_norm": 0.07670550793409348, "learning_rate": 5.941617509706247e-06, "loss": 1.1606, "step": 724 }, { "epoch": 1.9945054945054945, "grad_norm": 0.0779750868678093, "learning_rate": 5.9180542478670025e-06, "loss": 1.1624, "step": 726 }, { "epoch": 2.0, "grad_norm": 0.0955984815955162, "learning_rate": 5.894469866477905e-06, "loss": 1.1647, "step": 728 }, { "epoch": 2.0054945054945055, "grad_norm": 0.08561510592699051, "learning_rate": 5.87086490809025e-06, "loss": 1.1675, "step": 730 }, { "epoch": 2.0054945054945055, "eval_loss": 1.168563961982727, "eval_runtime": 298.1262, "eval_samples_per_second": 17.362, "eval_steps_per_second": 0.272, "step": 730 }, { "epoch": 2.010989010989011, "grad_norm": 0.08887989073991776, "learning_rate": 5.847239915728695e-06, "loss": 1.1711, "step": 732 }, { "epoch": 2.0164835164835164, "grad_norm": 0.1793489307165146, "learning_rate": 5.823595432878775e-06, "loss": 1.1689, "step": 734 }, { "epoch": 2.021978021978022, "grad_norm": 0.14795880019664764, "learning_rate": 5.799932003474398e-06, "loss": 1.1692, "step": 736 }, { "epoch": 2.0274725274725274, "grad_norm": 0.08862569183111191, "learning_rate": 5.776250171885329e-06, "loss": 1.1653, "step": 738 }, { "epoch": 2.032967032967033, "grad_norm": 0.10424422472715378, "learning_rate": 5.752550482904674e-06, "loss": 1.1651, "step": 740 }, { "epoch": 2.0384615384615383, "grad_norm": 0.0930488184094429, "learning_rate": 5.728833481736339e-06, "loss": 1.1666, "step": 742 }, { "epoch": 2.043956043956044, "grad_norm": 0.1007128432393074, "learning_rate": 5.705099713982491e-06, "loss": 1.1737, "step": 744 }, { "epoch": 2.0494505494505493, "grad_norm": 0.0997442975640297, "learning_rate": 5.6813497256310124e-06, "loss": 1.1701, "step": 746 }, { "epoch": 2.0549450549450547, "grad_norm": 0.09021608531475067, "learning_rate": 5.6575840630429295e-06, "loss": 1.1677, "step": 748 }, { "epoch": 2.0604395604395602, "grad_norm": 0.08044509589672089, "learning_rate": 5.633803272939851e-06, "loss": 1.1696, "step": 750 }, { "epoch": 2.065934065934066, "grad_norm": 0.08909393101930618, "learning_rate": 5.610007902391387e-06, "loss": 1.1616, "step": 752 }, { "epoch": 2.0714285714285716, "grad_norm": 0.08932027965784073, "learning_rate": 5.586198498802577e-06, "loss": 1.1623, "step": 754 }, { "epoch": 2.076923076923077, "grad_norm": 0.1681545227766037, "learning_rate": 5.562375609901273e-06, "loss": 1.1654, "step": 756 }, { "epoch": 2.0824175824175826, "grad_norm": 0.1894630342721939, "learning_rate": 5.538539783725556e-06, "loss": 1.1681, "step": 758 }, { "epoch": 2.087912087912088, "grad_norm": 0.11349444836378098, "learning_rate": 5.51469156861113e-06, "loss": 1.1688, "step": 760 }, { "epoch": 2.0934065934065935, "grad_norm": 0.08764316141605377, "learning_rate": 5.490831513178698e-06, "loss": 1.1718, "step": 762 }, { "epoch": 2.098901098901099, "grad_norm": 0.08182983100414276, "learning_rate": 5.466960166321348e-06, "loss": 1.1698, "step": 764 }, { "epoch": 2.1043956043956045, "grad_norm": 0.08446374535560608, "learning_rate": 5.44307807719192e-06, "loss": 1.1605, "step": 766 }, { "epoch": 2.10989010989011, "grad_norm": 0.06986381858587265, "learning_rate": 5.4191857951903825e-06, "loss": 1.1661, "step": 768 }, { "epoch": 2.1153846153846154, "grad_norm": 0.09340260177850723, "learning_rate": 5.395283869951184e-06, "loss": 1.1685, "step": 770 }, { "epoch": 2.120879120879121, "grad_norm": 0.10494975745677948, "learning_rate": 5.371372851330612e-06, "loss": 1.1748, "step": 772 }, { "epoch": 2.1263736263736264, "grad_norm": 0.09624402970075607, "learning_rate": 5.347453289394146e-06, "loss": 1.1683, "step": 774 }, { "epoch": 2.131868131868132, "grad_norm": 0.09158363193273544, "learning_rate": 5.3235257344037996e-06, "loss": 1.1693, "step": 776 }, { "epoch": 2.1373626373626373, "grad_norm": 0.07997617870569229, "learning_rate": 5.29959073680547e-06, "loss": 1.1671, "step": 778 }, { "epoch": 2.142857142857143, "grad_norm": 0.08307652175426483, "learning_rate": 5.275648847216263e-06, "loss": 1.1648, "step": 780 }, { "epoch": 2.1483516483516483, "grad_norm": 0.09365742653608322, "learning_rate": 5.251700616411836e-06, "loss": 1.1738, "step": 782 }, { "epoch": 2.1538461538461537, "grad_norm": 0.09766220301389694, "learning_rate": 5.22774659531372e-06, "loss": 1.1649, "step": 784 }, { "epoch": 2.159340659340659, "grad_norm": 0.10807793587446213, "learning_rate": 5.203787334976655e-06, "loss": 1.1728, "step": 786 }, { "epoch": 2.1648351648351647, "grad_norm": 0.07477546483278275, "learning_rate": 5.179823386575908e-06, "loss": 1.1701, "step": 788 }, { "epoch": 2.17032967032967, "grad_norm": 0.08150489628314972, "learning_rate": 5.155855301394585e-06, "loss": 1.1672, "step": 790 }, { "epoch": 2.1758241758241756, "grad_norm": 0.11139123886823654, "learning_rate": 5.131883630810966e-06, "loss": 1.172, "step": 792 }, { "epoch": 2.181318681318681, "grad_norm": 0.09428671002388, "learning_rate": 5.107908926285813e-06, "loss": 1.1662, "step": 794 }, { "epoch": 2.186813186813187, "grad_norm": 0.11421363055706024, "learning_rate": 5.083931739349675e-06, "loss": 1.1666, "step": 796 }, { "epoch": 2.1923076923076925, "grad_norm": 0.08080089837312698, "learning_rate": 5.059952621590216e-06, "loss": 1.1681, "step": 798 }, { "epoch": 2.197802197802198, "grad_norm": 0.09634223580360413, "learning_rate": 5.035972124639511e-06, "loss": 1.1654, "step": 800 }, { "epoch": 2.2032967032967035, "grad_norm": 0.08151613175868988, "learning_rate": 5.011990800161369e-06, "loss": 1.1668, "step": 802 }, { "epoch": 2.208791208791209, "grad_norm": 0.09357167035341263, "learning_rate": 4.988009199838632e-06, "loss": 1.1611, "step": 804 }, { "epoch": 2.2142857142857144, "grad_norm": 0.08290430158376694, "learning_rate": 4.96402787536049e-06, "loss": 1.1562, "step": 806 }, { "epoch": 2.21978021978022, "grad_norm": 0.09985602647066116, "learning_rate": 4.940047378409786e-06, "loss": 1.1716, "step": 808 }, { "epoch": 2.2252747252747254, "grad_norm": 0.08841745555400848, "learning_rate": 4.9160682606503255e-06, "loss": 1.1723, "step": 810 }, { "epoch": 2.230769230769231, "grad_norm": 0.08120245486497879, "learning_rate": 4.892091073714189e-06, "loss": 1.1642, "step": 812 }, { "epoch": 2.2362637362637363, "grad_norm": 0.08893705904483795, "learning_rate": 4.868116369189033e-06, "loss": 1.1614, "step": 814 }, { "epoch": 2.241758241758242, "grad_norm": 0.16162922978401184, "learning_rate": 4.844144698605418e-06, "loss": 1.1682, "step": 816 }, { "epoch": 2.2472527472527473, "grad_norm": 0.08067131042480469, "learning_rate": 4.820176613424095e-06, "loss": 1.1676, "step": 818 }, { "epoch": 2.2527472527472527, "grad_norm": 0.11158166825771332, "learning_rate": 4.796212665023345e-06, "loss": 1.1731, "step": 820 }, { "epoch": 2.258241758241758, "grad_norm": 0.08565322309732437, "learning_rate": 4.7722534046862805e-06, "loss": 1.1624, "step": 822 }, { "epoch": 2.2637362637362637, "grad_norm": 0.09021101146936417, "learning_rate": 4.748299383588167e-06, "loss": 1.1649, "step": 824 }, { "epoch": 2.269230769230769, "grad_norm": 0.09416112303733826, "learning_rate": 4.7243511527837374e-06, "loss": 1.1643, "step": 826 }, { "epoch": 2.2747252747252746, "grad_norm": 0.10305804759263992, "learning_rate": 4.7004092631945315e-06, "loss": 1.1709, "step": 828 }, { "epoch": 2.28021978021978, "grad_norm": 0.11213865131139755, "learning_rate": 4.6764742655962e-06, "loss": 1.1684, "step": 830 }, { "epoch": 2.2857142857142856, "grad_norm": 0.10108992457389832, "learning_rate": 4.652546710605857e-06, "loss": 1.1619, "step": 832 }, { "epoch": 2.291208791208791, "grad_norm": 0.08697344362735748, "learning_rate": 4.628627148669391e-06, "loss": 1.1686, "step": 834 }, { "epoch": 2.2967032967032965, "grad_norm": 0.10428237169981003, "learning_rate": 4.604716130048818e-06, "loss": 1.1668, "step": 836 }, { "epoch": 2.302197802197802, "grad_norm": 0.08793749660253525, "learning_rate": 4.580814204809618e-06, "loss": 1.1713, "step": 838 }, { "epoch": 2.3076923076923075, "grad_norm": 0.11877471953630447, "learning_rate": 4.5569219228080805e-06, "loss": 1.162, "step": 840 }, { "epoch": 2.313186813186813, "grad_norm": 0.08249559253454208, "learning_rate": 4.5330398336786526e-06, "loss": 1.1678, "step": 842 }, { "epoch": 2.3186813186813184, "grad_norm": 0.11602500081062317, "learning_rate": 4.509168486821304e-06, "loss": 1.164, "step": 844 }, { "epoch": 2.3241758241758244, "grad_norm": 0.11740805953741074, "learning_rate": 4.48530843138887e-06, "loss": 1.1643, "step": 846 }, { "epoch": 2.32967032967033, "grad_norm": 0.10009398311376572, "learning_rate": 4.4614602162744455e-06, "loss": 1.1629, "step": 848 }, { "epoch": 2.3351648351648353, "grad_norm": 0.07898547500371933, "learning_rate": 4.4376243900987296e-06, "loss": 1.1568, "step": 850 }, { "epoch": 2.340659340659341, "grad_norm": 0.08736059069633484, "learning_rate": 4.413801501197424e-06, "loss": 1.1703, "step": 852 }, { "epoch": 2.3461538461538463, "grad_norm": 0.08960308879613876, "learning_rate": 4.389992097608613e-06, "loss": 1.1632, "step": 854 }, { "epoch": 2.3516483516483517, "grad_norm": 0.10276441276073456, "learning_rate": 4.366196727060152e-06, "loss": 1.1634, "step": 856 }, { "epoch": 2.357142857142857, "grad_norm": 0.07904151827096939, "learning_rate": 4.342415936957073e-06, "loss": 1.1709, "step": 858 }, { "epoch": 2.3626373626373627, "grad_norm": 0.07296542823314667, "learning_rate": 4.318650274368989e-06, "loss": 1.1672, "step": 860 }, { "epoch": 2.368131868131868, "grad_norm": 0.08600784838199615, "learning_rate": 4.294900286017509e-06, "loss": 1.1612, "step": 862 }, { "epoch": 2.3736263736263736, "grad_norm": 0.08018805086612701, "learning_rate": 4.271166518263662e-06, "loss": 1.1722, "step": 864 }, { "epoch": 2.379120879120879, "grad_norm": 0.07395070046186447, "learning_rate": 4.247449517095329e-06, "loss": 1.1711, "step": 866 }, { "epoch": 2.3846153846153846, "grad_norm": 0.07601243257522583, "learning_rate": 4.223749828114672e-06, "loss": 1.1771, "step": 868 }, { "epoch": 2.39010989010989, "grad_norm": 0.09423944354057312, "learning_rate": 4.2000679965256045e-06, "loss": 1.1603, "step": 870 }, { "epoch": 2.3956043956043955, "grad_norm": 0.08968156576156616, "learning_rate": 4.176404567121225e-06, "loss": 1.1608, "step": 872 }, { "epoch": 2.401098901098901, "grad_norm": 0.10625026375055313, "learning_rate": 4.152760084271305e-06, "loss": 1.1574, "step": 874 }, { "epoch": 2.4065934065934065, "grad_norm": 0.18610498309135437, "learning_rate": 4.129135091909752e-06, "loss": 1.1747, "step": 876 }, { "epoch": 2.4065934065934065, "eval_loss": 1.1678768396377563, "eval_runtime": 299.619, "eval_samples_per_second": 17.275, "eval_steps_per_second": 0.27, "step": 876 }, { "epoch": 2.412087912087912, "grad_norm": 0.15964192152023315, "learning_rate": 4.105530133522096e-06, "loss": 1.1678, "step": 878 }, { "epoch": 2.4175824175824174, "grad_norm": 0.11698123812675476, "learning_rate": 4.081945752133e-06, "loss": 1.1678, "step": 880 }, { "epoch": 2.423076923076923, "grad_norm": 0.16632910072803497, "learning_rate": 4.058382490293755e-06, "loss": 1.1722, "step": 882 }, { "epoch": 2.4285714285714284, "grad_norm": 0.17577840387821198, "learning_rate": 4.034840890069805e-06, "loss": 1.1643, "step": 884 }, { "epoch": 2.4340659340659343, "grad_norm": 0.17302070558071136, "learning_rate": 4.0113214930282765e-06, "loss": 1.1641, "step": 886 }, { "epoch": 2.4395604395604398, "grad_norm": 0.2563433349132538, "learning_rate": 3.987824840225512e-06, "loss": 1.1678, "step": 888 }, { "epoch": 2.4450549450549453, "grad_norm": 0.19088977575302124, "learning_rate": 3.964351472194642e-06, "loss": 1.174, "step": 890 }, { "epoch": 2.4505494505494507, "grad_norm": 0.22029054164886475, "learning_rate": 3.940901928933127e-06, "loss": 1.1641, "step": 892 }, { "epoch": 2.456043956043956, "grad_norm": 0.26073768734931946, "learning_rate": 3.917476749890351e-06, "loss": 1.1594, "step": 894 }, { "epoch": 2.4615384615384617, "grad_norm": 0.14887195825576782, "learning_rate": 3.894076473955207e-06, "loss": 1.1696, "step": 896 }, { "epoch": 2.467032967032967, "grad_norm": 0.09114201366901398, "learning_rate": 3.8707016394436985e-06, "loss": 1.1686, "step": 898 }, { "epoch": 2.4725274725274726, "grad_norm": 0.12883096933364868, "learning_rate": 3.847352784086556e-06, "loss": 1.1614, "step": 900 }, { "epoch": 2.478021978021978, "grad_norm": 0.10688427090644836, "learning_rate": 3.8240304450168716e-06, "loss": 1.1583, "step": 902 }, { "epoch": 2.4835164835164836, "grad_norm": 0.08104757964611053, "learning_rate": 3.8007351587577342e-06, "loss": 1.1713, "step": 904 }, { "epoch": 2.489010989010989, "grad_norm": 0.11496758460998535, "learning_rate": 3.777467461209895e-06, "loss": 1.1662, "step": 906 }, { "epoch": 2.4945054945054945, "grad_norm": 0.08005277067422867, "learning_rate": 3.754227887639434e-06, "loss": 1.1683, "step": 908 }, { "epoch": 2.5, "grad_norm": 0.08510690182447433, "learning_rate": 3.7310169726654444e-06, "loss": 1.1603, "step": 910 }, { "epoch": 2.5054945054945055, "grad_norm": 0.08159387856721878, "learning_rate": 3.707835250247745e-06, "loss": 1.1703, "step": 912 }, { "epoch": 2.510989010989011, "grad_norm": 0.08646809309720993, "learning_rate": 3.684683253674583e-06, "loss": 1.1744, "step": 914 }, { "epoch": 2.5164835164835164, "grad_norm": 0.10196442157030106, "learning_rate": 3.6615615155503703e-06, "loss": 1.1607, "step": 916 }, { "epoch": 2.521978021978022, "grad_norm": 0.07023598998785019, "learning_rate": 3.638470567783442e-06, "loss": 1.17, "step": 918 }, { "epoch": 2.5274725274725274, "grad_norm": 0.09794013947248459, "learning_rate": 3.615410941573799e-06, "loss": 1.1713, "step": 920 }, { "epoch": 2.532967032967033, "grad_norm": 0.07678301632404327, "learning_rate": 3.59238316740091e-06, "loss": 1.1673, "step": 922 }, { "epoch": 2.5384615384615383, "grad_norm": 0.09355033934116364, "learning_rate": 3.5693877750114903e-06, "loss": 1.1639, "step": 924 }, { "epoch": 2.543956043956044, "grad_norm": 0.08433817327022552, "learning_rate": 3.546425293407324e-06, "loss": 1.1614, "step": 926 }, { "epoch": 2.5494505494505493, "grad_norm": 0.08144976943731308, "learning_rate": 3.523496250833098e-06, "loss": 1.1599, "step": 928 }, { "epoch": 2.5549450549450547, "grad_norm": 0.0796588808298111, "learning_rate": 3.5006011747642366e-06, "loss": 1.1667, "step": 930 }, { "epoch": 2.5604395604395602, "grad_norm": 0.08032579720020294, "learning_rate": 3.4777405918947795e-06, "loss": 1.1612, "step": 932 }, { "epoch": 2.5659340659340657, "grad_norm": 0.08575133234262466, "learning_rate": 3.4549150281252635e-06, "loss": 1.1695, "step": 934 }, { "epoch": 2.571428571428571, "grad_norm": 0.08064404129981995, "learning_rate": 3.4321250085506174e-06, "loss": 1.1698, "step": 936 }, { "epoch": 2.5769230769230766, "grad_norm": 0.07455819100141525, "learning_rate": 3.4093710574480926e-06, "loss": 1.1643, "step": 938 }, { "epoch": 2.5824175824175826, "grad_norm": 0.08538255095481873, "learning_rate": 3.386653698265189e-06, "loss": 1.1593, "step": 940 }, { "epoch": 2.587912087912088, "grad_norm": 0.08684508502483368, "learning_rate": 3.3639734536076263e-06, "loss": 1.1651, "step": 942 }, { "epoch": 2.5934065934065935, "grad_norm": 0.08089859038591385, "learning_rate": 3.341330845227316e-06, "loss": 1.165, "step": 944 }, { "epoch": 2.598901098901099, "grad_norm": 0.07942517846822739, "learning_rate": 3.3187263940103587e-06, "loss": 1.1685, "step": 946 }, { "epoch": 2.6043956043956045, "grad_norm": 0.08427475392818451, "learning_rate": 3.296160619965056e-06, "loss": 1.1547, "step": 948 }, { "epoch": 2.60989010989011, "grad_norm": 0.08193778246641159, "learning_rate": 3.2736340422099633e-06, "loss": 1.1651, "step": 950 }, { "epoch": 2.6153846153846154, "grad_norm": 0.0785454586148262, "learning_rate": 3.2511471789619274e-06, "loss": 1.169, "step": 952 }, { "epoch": 2.620879120879121, "grad_norm": 0.08144666999578476, "learning_rate": 3.228700547524184e-06, "loss": 1.1656, "step": 954 }, { "epoch": 2.6263736263736264, "grad_norm": 0.08215656876564026, "learning_rate": 3.206294664274443e-06, "loss": 1.163, "step": 956 }, { "epoch": 2.631868131868132, "grad_norm": 0.07643819600343704, "learning_rate": 3.183930044653014e-06, "loss": 1.1609, "step": 958 }, { "epoch": 2.6373626373626373, "grad_norm": 0.06851556152105331, "learning_rate": 3.1616072031509594e-06, "loss": 1.168, "step": 960 }, { "epoch": 2.642857142857143, "grad_norm": 0.07472985982894897, "learning_rate": 3.139326653298236e-06, "loss": 1.1672, "step": 962 }, { "epoch": 2.6483516483516483, "grad_norm": 0.08239593356847763, "learning_rate": 3.117088907651902e-06, "loss": 1.1668, "step": 964 }, { "epoch": 2.6538461538461537, "grad_norm": 0.07118227332830429, "learning_rate": 3.094894477784318e-06, "loss": 1.163, "step": 966 }, { "epoch": 2.659340659340659, "grad_norm": 0.08010434359312057, "learning_rate": 3.0727438742713766e-06, "loss": 1.1674, "step": 968 }, { "epoch": 2.6648351648351647, "grad_norm": 0.06395678967237473, "learning_rate": 3.0506376066807632e-06, "loss": 1.1709, "step": 970 }, { "epoch": 2.67032967032967, "grad_norm": 0.07488574087619781, "learning_rate": 3.028576183560221e-06, "loss": 1.1616, "step": 972 }, { "epoch": 2.675824175824176, "grad_norm": 0.0717063769698143, "learning_rate": 3.006560112425867e-06, "loss": 1.159, "step": 974 }, { "epoch": 2.6813186813186816, "grad_norm": 0.0665455088019371, "learning_rate": 2.9845898997505102e-06, "loss": 1.1717, "step": 976 }, { "epoch": 2.686813186813187, "grad_norm": 0.07839926332235336, "learning_rate": 2.962666050951997e-06, "loss": 1.1635, "step": 978 }, { "epoch": 2.6923076923076925, "grad_norm": 0.08876995742321014, "learning_rate": 2.940789070381587e-06, "loss": 1.1634, "step": 980 }, { "epoch": 2.697802197802198, "grad_norm": 0.09154446423053741, "learning_rate": 2.918959461312353e-06, "loss": 1.1651, "step": 982 }, { "epoch": 2.7032967032967035, "grad_norm": 0.08424372225999832, "learning_rate": 2.897177725927599e-06, "loss": 1.153, "step": 984 }, { "epoch": 2.708791208791209, "grad_norm": 0.11759259551763535, "learning_rate": 2.8754443653093186e-06, "loss": 1.1662, "step": 986 }, { "epoch": 2.7142857142857144, "grad_norm": 0.08856125921010971, "learning_rate": 2.853759879426644e-06, "loss": 1.1644, "step": 988 }, { "epoch": 2.71978021978022, "grad_norm": 0.08224651217460632, "learning_rate": 2.8321247671243695e-06, "loss": 1.1628, "step": 990 }, { "epoch": 2.7252747252747254, "grad_norm": 0.11355423927307129, "learning_rate": 2.8105395261114666e-06, "loss": 1.1663, "step": 992 }, { "epoch": 2.730769230769231, "grad_norm": 0.06467260420322418, "learning_rate": 2.7890046529496284e-06, "loss": 1.1736, "step": 994 }, { "epoch": 2.7362637362637363, "grad_norm": 0.08972840011119843, "learning_rate": 2.7675206430418542e-06, "loss": 1.1692, "step": 996 }, { "epoch": 2.741758241758242, "grad_norm": 0.09491213411092758, "learning_rate": 2.7460879906210485e-06, "loss": 1.1707, "step": 998 }, { "epoch": 2.7472527472527473, "grad_norm": 0.07063695043325424, "learning_rate": 2.7247071887386544e-06, "loss": 1.1605, "step": 1000 }, { "epoch": 2.7527472527472527, "grad_norm": 0.07039818912744522, "learning_rate": 2.70337872925331e-06, "loss": 1.1648, "step": 1002 }, { "epoch": 2.758241758241758, "grad_norm": 0.07515694946050644, "learning_rate": 2.68210310281953e-06, "loss": 1.1563, "step": 1004 }, { "epoch": 2.7637362637362637, "grad_norm": 0.06818992644548416, "learning_rate": 2.6608807988764252e-06, "loss": 1.1735, "step": 1006 }, { "epoch": 2.769230769230769, "grad_norm": 0.0827600434422493, "learning_rate": 2.6397123056364364e-06, "loss": 1.1698, "step": 1008 }, { "epoch": 2.7747252747252746, "grad_norm": 0.10222798585891724, "learning_rate": 2.618598110074105e-06, "loss": 1.1689, "step": 1010 }, { "epoch": 2.78021978021978, "grad_norm": 0.0823441818356514, "learning_rate": 2.5975386979148792e-06, "loss": 1.166, "step": 1012 }, { "epoch": 2.7857142857142856, "grad_norm": 0.0735846534371376, "learning_rate": 2.576534553623925e-06, "loss": 1.1592, "step": 1014 }, { "epoch": 2.791208791208791, "grad_norm": 0.07179060578346252, "learning_rate": 2.5555861603949832e-06, "loss": 1.1755, "step": 1016 }, { "epoch": 2.7967032967032965, "grad_norm": 0.08374089747667313, "learning_rate": 2.534694000139273e-06, "loss": 1.1757, "step": 1018 }, { "epoch": 2.802197802197802, "grad_norm": 0.07065007835626602, "learning_rate": 2.513858553474382e-06, "loss": 1.1698, "step": 1020 }, { "epoch": 2.8076923076923075, "grad_norm": 0.08045931160449982, "learning_rate": 2.4930802997132213e-06, "loss": 1.1582, "step": 1022 }, { "epoch": 2.8076923076923075, "eval_loss": 1.1672074794769287, "eval_runtime": 299.934, "eval_samples_per_second": 17.257, "eval_steps_per_second": 0.27, "step": 1022 }, { "epoch": 2.813186813186813, "grad_norm": 0.07823072373867035, "learning_rate": 2.4723597168529984e-06, "loss": 1.1596, "step": 1024 }, { "epoch": 2.8186813186813184, "grad_norm": 0.08508284389972687, "learning_rate": 2.4516972815642166e-06, "loss": 1.1723, "step": 1026 }, { "epoch": 2.824175824175824, "grad_norm": 0.07922125607728958, "learning_rate": 2.4310934691797207e-06, "loss": 1.1617, "step": 1028 }, { "epoch": 2.82967032967033, "grad_norm": 0.0768875777721405, "learning_rate": 2.410548753683743e-06, "loss": 1.1664, "step": 1030 }, { "epoch": 2.8351648351648353, "grad_norm": 0.07595735043287277, "learning_rate": 2.390063607701016e-06, "loss": 1.1731, "step": 1032 }, { "epoch": 2.840659340659341, "grad_norm": 0.08126144856214523, "learning_rate": 2.369638502485897e-06, "loss": 1.163, "step": 1034 }, { "epoch": 2.8461538461538463, "grad_norm": 0.06850147992372513, "learning_rate": 2.3492739079115214e-06, "loss": 1.1685, "step": 1036 }, { "epoch": 2.8516483516483517, "grad_norm": 0.08522983640432358, "learning_rate": 2.3289702924589914e-06, "loss": 1.1686, "step": 1038 }, { "epoch": 2.857142857142857, "grad_norm": 0.08155310899019241, "learning_rate": 2.3087281232066134e-06, "loss": 1.1653, "step": 1040 }, { "epoch": 2.8626373626373627, "grad_norm": 0.07040958106517792, "learning_rate": 2.2885478658191364e-06, "loss": 1.1557, "step": 1042 }, { "epoch": 2.868131868131868, "grad_norm": 0.0749848261475563, "learning_rate": 2.268429984537048e-06, "loss": 1.1724, "step": 1044 }, { "epoch": 2.8736263736263736, "grad_norm": 0.06682237237691879, "learning_rate": 2.248374942165894e-06, "loss": 1.158, "step": 1046 }, { "epoch": 2.879120879120879, "grad_norm": 0.0770927369594574, "learning_rate": 2.2283832000656304e-06, "loss": 1.1676, "step": 1048 }, { "epoch": 2.8846153846153846, "grad_norm": 0.08294253051280975, "learning_rate": 2.2084552181400087e-06, "loss": 1.1654, "step": 1050 }, { "epoch": 2.89010989010989, "grad_norm": 0.08596468716859818, "learning_rate": 2.188591454826e-06, "loss": 1.1689, "step": 1052 }, { "epoch": 2.8956043956043955, "grad_norm": 0.08016978204250336, "learning_rate": 2.168792367083243e-06, "loss": 1.1614, "step": 1054 }, { "epoch": 2.901098901098901, "grad_norm": 0.0838996022939682, "learning_rate": 2.1490584103835433e-06, "loss": 1.1658, "step": 1056 }, { "epoch": 2.9065934065934065, "grad_norm": 0.07020293921232224, "learning_rate": 2.1293900387003742e-06, "loss": 1.1594, "step": 1058 }, { "epoch": 2.912087912087912, "grad_norm": 0.07320253551006317, "learning_rate": 2.109787704498459e-06, "loss": 1.1635, "step": 1060 }, { "epoch": 2.9175824175824174, "grad_norm": 0.0821409672498703, "learning_rate": 2.0902518587233418e-06, "loss": 1.1628, "step": 1062 }, { "epoch": 2.9230769230769234, "grad_norm": 0.07211437821388245, "learning_rate": 2.0707829507910237e-06, "loss": 1.1689, "step": 1064 }, { "epoch": 2.928571428571429, "grad_norm": 0.08253902941942215, "learning_rate": 2.051381428577622e-06, "loss": 1.1654, "step": 1066 }, { "epoch": 2.9340659340659343, "grad_norm": 0.080472432076931, "learning_rate": 2.0320477384090665e-06, "loss": 1.1707, "step": 1068 }, { "epoch": 2.9395604395604398, "grad_norm": 0.08331170678138733, "learning_rate": 2.012782325050831e-06, "loss": 1.1729, "step": 1070 }, { "epoch": 2.9450549450549453, "grad_norm": 0.07666671276092529, "learning_rate": 1.9935856316977044e-06, "loss": 1.1715, "step": 1072 }, { "epoch": 2.9505494505494507, "grad_norm": 0.09519699960947037, "learning_rate": 1.9744580999635902e-06, "loss": 1.1629, "step": 1074 }, { "epoch": 2.956043956043956, "grad_norm": 0.0872369259595871, "learning_rate": 1.9554001698713572e-06, "loss": 1.1683, "step": 1076 }, { "epoch": 2.9615384615384617, "grad_norm": 0.07113870233297348, "learning_rate": 1.936412279842705e-06, "loss": 1.1637, "step": 1078 }, { "epoch": 2.967032967032967, "grad_norm": 0.08545450866222382, "learning_rate": 1.9174948666880805e-06, "loss": 1.1627, "step": 1080 }, { "epoch": 2.9725274725274726, "grad_norm": 0.08711759746074677, "learning_rate": 1.8986483655966408e-06, "loss": 1.1582, "step": 1082 }, { "epoch": 2.978021978021978, "grad_norm": 0.08870179951190948, "learning_rate": 1.879873210126229e-06, "loss": 1.167, "step": 1084 }, { "epoch": 2.9835164835164836, "grad_norm": 0.1006346121430397, "learning_rate": 1.8611698321933991e-06, "loss": 1.1724, "step": 1086 }, { "epoch": 2.989010989010989, "grad_norm": 0.07408854365348816, "learning_rate": 1.8425386620634961e-06, "loss": 1.1705, "step": 1088 }, { "epoch": 2.9945054945054945, "grad_norm": 0.09146919846534729, "learning_rate": 1.8239801283407393e-06, "loss": 1.16, "step": 1090 }, { "epoch": 3.0, "grad_norm": 0.0709662064909935, "learning_rate": 1.8054946579583732e-06, "loss": 1.1636, "step": 1092 }, { "epoch": 3.0054945054945055, "grad_norm": 0.10273056477308273, "learning_rate": 1.787082676168842e-06, "loss": 1.1647, "step": 1094 }, { "epoch": 3.010989010989011, "grad_norm": 0.07023092359304428, "learning_rate": 1.7687446065340074e-06, "loss": 1.162, "step": 1096 }, { "epoch": 3.0164835164835164, "grad_norm": 0.08073507994413376, "learning_rate": 1.7504808709154104e-06, "loss": 1.1697, "step": 1098 }, { "epoch": 3.021978021978022, "grad_norm": 0.07398983091115952, "learning_rate": 1.7322918894645525e-06, "loss": 1.1637, "step": 1100 }, { "epoch": 3.0274725274725274, "grad_norm": 0.08476796001195908, "learning_rate": 1.7141780806132429e-06, "loss": 1.161, "step": 1102 }, { "epoch": 3.032967032967033, "grad_norm": 0.07763465493917465, "learning_rate": 1.696139861063974e-06, "loss": 1.1685, "step": 1104 }, { "epoch": 3.0384615384615383, "grad_norm": 0.08244740962982178, "learning_rate": 1.6781776457803227e-06, "loss": 1.1683, "step": 1106 }, { "epoch": 3.043956043956044, "grad_norm": 0.09578699618577957, "learning_rate": 1.660291847977415e-06, "loss": 1.1677, "step": 1108 }, { "epoch": 3.0494505494505493, "grad_norm": 0.07818014919757843, "learning_rate": 1.6424828791124159e-06, "loss": 1.166, "step": 1110 }, { "epoch": 3.0549450549450547, "grad_norm": 0.07996879518032074, "learning_rate": 1.624751148875065e-06, "loss": 1.1732, "step": 1112 }, { "epoch": 3.0604395604395602, "grad_norm": 0.06753943115472794, "learning_rate": 1.6070970651782514e-06, "loss": 1.1628, "step": 1114 }, { "epoch": 3.065934065934066, "grad_norm": 0.06478522717952728, "learning_rate": 1.5895210341486279e-06, "loss": 1.1607, "step": 1116 }, { "epoch": 3.0714285714285716, "grad_norm": 0.07904413342475891, "learning_rate": 1.5720234601172767e-06, "loss": 1.1655, "step": 1118 }, { "epoch": 3.076923076923077, "grad_norm": 0.07776004076004028, "learning_rate": 1.5546047456103964e-06, "loss": 1.1677, "step": 1120 }, { "epoch": 3.0824175824175826, "grad_norm": 0.07198958098888397, "learning_rate": 1.537265291340042e-06, "loss": 1.1636, "step": 1122 }, { "epoch": 3.087912087912088, "grad_norm": 0.07310649752616882, "learning_rate": 1.5200054961949233e-06, "loss": 1.1601, "step": 1124 }, { "epoch": 3.0934065934065935, "grad_norm": 0.07591935992240906, "learning_rate": 1.5028257572312105e-06, "loss": 1.1649, "step": 1126 }, { "epoch": 3.098901098901099, "grad_norm": 0.07674799114465714, "learning_rate": 1.485726469663401e-06, "loss": 1.1704, "step": 1128 }, { "epoch": 3.1043956043956045, "grad_norm": 0.09393850713968277, "learning_rate": 1.468708026855245e-06, "loss": 1.1692, "step": 1130 }, { "epoch": 3.10989010989011, "grad_norm": 0.07599301636219025, "learning_rate": 1.4517708203106763e-06, "loss": 1.1645, "step": 1132 }, { "epoch": 3.1153846153846154, "grad_norm": 0.08221649378538132, "learning_rate": 1.4349152396648153e-06, "loss": 1.1631, "step": 1134 }, { "epoch": 3.120879120879121, "grad_norm": 0.07295921444892883, "learning_rate": 1.4181416726750052e-06, "loss": 1.1612, "step": 1136 }, { "epoch": 3.1263736263736264, "grad_norm": 0.079580157995224, "learning_rate": 1.4014505052118893e-06, "loss": 1.1599, "step": 1138 }, { "epoch": 3.131868131868132, "grad_norm": 0.09730138629674911, "learning_rate": 1.3848421212505404e-06, "loss": 1.1632, "step": 1140 }, { "epoch": 3.1373626373626373, "grad_norm": 0.09287888556718826, "learning_rate": 1.3683169028616155e-06, "loss": 1.1602, "step": 1142 }, { "epoch": 3.142857142857143, "grad_norm": 0.07959942519664764, "learning_rate": 1.3518752302025773e-06, "loss": 1.1629, "step": 1144 }, { "epoch": 3.1483516483516483, "grad_norm": 0.07988713681697845, "learning_rate": 1.3355174815089477e-06, "loss": 1.1641, "step": 1146 }, { "epoch": 3.1538461538461537, "grad_norm": 0.06947878748178482, "learning_rate": 1.3192440330856005e-06, "loss": 1.1614, "step": 1148 }, { "epoch": 3.159340659340659, "grad_norm": 0.06804593652486801, "learning_rate": 1.30305525929811e-06, "loss": 1.1666, "step": 1150 }, { "epoch": 3.1648351648351647, "grad_norm": 0.07112333178520203, "learning_rate": 1.2869515325641357e-06, "loss": 1.1592, "step": 1152 }, { "epoch": 3.17032967032967, "grad_norm": 0.07313236594200134, "learning_rate": 1.2709332233448573e-06, "loss": 1.1686, "step": 1154 }, { "epoch": 3.1758241758241756, "grad_norm": 0.07543834298849106, "learning_rate": 1.2550007001364518e-06, "loss": 1.1626, "step": 1156 }, { "epoch": 3.181318681318681, "grad_norm": 0.07730558514595032, "learning_rate": 1.239154329461615e-06, "loss": 1.164, "step": 1158 }, { "epoch": 3.186813186813187, "grad_norm": 0.07528182864189148, "learning_rate": 1.223394475861131e-06, "loss": 1.1621, "step": 1160 }, { "epoch": 3.1923076923076925, "grad_norm": 0.07378531992435455, "learning_rate": 1.207721501885486e-06, "loss": 1.162, "step": 1162 }, { "epoch": 3.197802197802198, "grad_norm": 0.07334302365779877, "learning_rate": 1.1921357680865258e-06, "loss": 1.1642, "step": 1164 }, { "epoch": 3.2032967032967035, "grad_norm": 0.07697130739688873, "learning_rate": 1.1766376330091684e-06, "loss": 1.1647, "step": 1166 }, { "epoch": 3.208791208791209, "grad_norm": 0.0726918876171112, "learning_rate": 1.1612274531831463e-06, "loss": 1.1719, "step": 1168 }, { "epoch": 3.208791208791209, "eval_loss": 1.1669209003448486, "eval_runtime": 299.9103, "eval_samples_per_second": 17.258, "eval_steps_per_second": 0.27, "step": 1168 }, { "epoch": 3.2142857142857144, "grad_norm": 0.07847239077091217, "learning_rate": 1.1459055831148074e-06, "loss": 1.1651, "step": 1170 }, { "epoch": 3.21978021978022, "grad_norm": 0.06886722892522812, "learning_rate": 1.1306723752789672e-06, "loss": 1.1648, "step": 1172 }, { "epoch": 3.2252747252747254, "grad_norm": 0.0860794335603714, "learning_rate": 1.1155281801107897e-06, "loss": 1.168, "step": 1174 }, { "epoch": 3.230769230769231, "grad_norm": 0.07369523495435715, "learning_rate": 1.1004733459977325e-06, "loss": 1.1669, "step": 1176 }, { "epoch": 3.2362637362637363, "grad_norm": 0.07247929275035858, "learning_rate": 1.0855082192715294e-06, "loss": 1.168, "step": 1178 }, { "epoch": 3.241758241758242, "grad_norm": 0.07475174218416214, "learning_rate": 1.0706331442002226e-06, "loss": 1.1622, "step": 1180 }, { "epoch": 3.2472527472527473, "grad_norm": 0.07838430255651474, "learning_rate": 1.0558484629802502e-06, "loss": 1.171, "step": 1182 }, { "epoch": 3.2527472527472527, "grad_norm": 0.0696970596909523, "learning_rate": 1.041154515728559e-06, "loss": 1.1621, "step": 1184 }, { "epoch": 3.258241758241758, "grad_norm": 0.07396616786718369, "learning_rate": 1.0265516404747943e-06, "loss": 1.1641, "step": 1186 }, { "epoch": 3.2637362637362637, "grad_norm": 0.08104917407035828, "learning_rate": 1.0120401731535213e-06, "loss": 1.1663, "step": 1188 }, { "epoch": 3.269230769230769, "grad_norm": 0.0944487676024437, "learning_rate": 9.976204475964907e-07, "loss": 1.1618, "step": 1190 }, { "epoch": 3.2747252747252746, "grad_norm": 0.07517724484205246, "learning_rate": 9.832927955249605e-07, "loss": 1.1726, "step": 1192 }, { "epoch": 3.28021978021978, "grad_norm": 0.08082670718431473, "learning_rate": 9.690575465420733e-07, "loss": 1.1689, "step": 1194 }, { "epoch": 3.2857142857142856, "grad_norm": 0.07103955000638962, "learning_rate": 9.549150281252633e-07, "loss": 1.1724, "step": 1196 }, { "epoch": 3.291208791208791, "grad_norm": 0.08387453854084015, "learning_rate": 9.408655656187282e-07, "loss": 1.1598, "step": 1198 }, { "epoch": 3.2967032967032965, "grad_norm": 0.07405807077884674, "learning_rate": 9.269094822259439e-07, "loss": 1.1648, "step": 1200 }, { "epoch": 3.302197802197802, "grad_norm": 0.07046142220497131, "learning_rate": 9.130470990022283e-07, "loss": 1.1707, "step": 1202 }, { "epoch": 3.3076923076923075, "grad_norm": 0.07296803593635559, "learning_rate": 8.992787348473575e-07, "loss": 1.1642, "step": 1204 }, { "epoch": 3.313186813186813, "grad_norm": 0.0734080895781517, "learning_rate": 8.856047064982276e-07, "loss": 1.1558, "step": 1206 }, { "epoch": 3.3186813186813184, "grad_norm": 0.07395216077566147, "learning_rate": 8.720253285215685e-07, "loss": 1.1721, "step": 1208 }, { "epoch": 3.3241758241758244, "grad_norm": 0.07748089730739594, "learning_rate": 8.585409133067119e-07, "loss": 1.1653, "step": 1210 }, { "epoch": 3.32967032967033, "grad_norm": 0.07397377490997314, "learning_rate": 8.451517710583934e-07, "loss": 1.1623, "step": 1212 }, { "epoch": 3.3351648351648353, "grad_norm": 0.07272295653820038, "learning_rate": 8.318582097896316e-07, "loss": 1.1643, "step": 1214 }, { "epoch": 3.340659340659341, "grad_norm": 0.07338716834783554, "learning_rate": 8.18660535314631e-07, "loss": 1.1612, "step": 1216 }, { "epoch": 3.3461538461538463, "grad_norm": 0.06233609840273857, "learning_rate": 8.055590512417499e-07, "loss": 1.1642, "step": 1218 }, { "epoch": 3.3516483516483517, "grad_norm": 0.07398121803998947, "learning_rate": 7.925540589665187e-07, "loss": 1.1719, "step": 1220 }, { "epoch": 3.357142857142857, "grad_norm": 0.0798347070813179, "learning_rate": 7.796458576647015e-07, "loss": 1.1669, "step": 1222 }, { "epoch": 3.3626373626373627, "grad_norm": 0.06861409544944763, "learning_rate": 7.668347442854218e-07, "loss": 1.1698, "step": 1224 }, { "epoch": 3.368131868131868, "grad_norm": 0.06391950696706772, "learning_rate": 7.541210135443188e-07, "loss": 1.166, "step": 1226 }, { "epoch": 3.3736263736263736, "grad_norm": 0.07030785083770752, "learning_rate": 7.415049579167783e-07, "loss": 1.166, "step": 1228 }, { "epoch": 3.379120879120879, "grad_norm": 0.07540637254714966, "learning_rate": 7.289868676312023e-07, "loss": 1.1643, "step": 1230 }, { "epoch": 3.3846153846153846, "grad_norm": 0.06893002241849899, "learning_rate": 7.165670306623296e-07, "loss": 1.163, "step": 1232 }, { "epoch": 3.39010989010989, "grad_norm": 0.07210598886013031, "learning_rate": 7.042457327246088e-07, "loss": 1.1648, "step": 1234 }, { "epoch": 3.3956043956043955, "grad_norm": 0.076958067715168, "learning_rate": 6.920232572656349e-07, "loss": 1.1635, "step": 1236 }, { "epoch": 3.401098901098901, "grad_norm": 0.07737283408641815, "learning_rate": 6.79899885459619e-07, "loss": 1.161, "step": 1238 }, { "epoch": 3.4065934065934065, "grad_norm": 0.07089677453041077, "learning_rate": 6.678758962009241e-07, "loss": 1.1595, "step": 1240 }, { "epoch": 3.412087912087912, "grad_norm": 0.06980301439762115, "learning_rate": 6.559515660976506e-07, "loss": 1.1646, "step": 1242 }, { "epoch": 3.4175824175824174, "grad_norm": 0.0859365463256836, "learning_rate": 6.441271694652701e-07, "loss": 1.1653, "step": 1244 }, { "epoch": 3.423076923076923, "grad_norm": 0.0688340812921524, "learning_rate": 6.32402978320315e-07, "loss": 1.171, "step": 1246 }, { "epoch": 3.4285714285714284, "grad_norm": 0.07327734678983688, "learning_rate": 6.207792623741249e-07, "loss": 1.1664, "step": 1248 }, { "epoch": 3.4340659340659343, "grad_norm": 0.06622574478387833, "learning_rate": 6.092562890266341e-07, "loss": 1.162, "step": 1250 }, { "epoch": 3.4395604395604398, "grad_norm": 0.06825845688581467, "learning_rate": 5.97834323360233e-07, "loss": 1.1601, "step": 1252 }, { "epoch": 3.4450549450549453, "grad_norm": 0.0715174600481987, "learning_rate": 5.86513628133652e-07, "loss": 1.1653, "step": 1254 }, { "epoch": 3.4505494505494507, "grad_norm": 0.06850449740886688, "learning_rate": 5.75294463775935e-07, "loss": 1.1625, "step": 1256 }, { "epoch": 3.456043956043956, "grad_norm": 0.061297595500946045, "learning_rate": 5.641770883804365e-07, "loss": 1.168, "step": 1258 }, { "epoch": 3.4615384615384617, "grad_norm": 0.07319982349872589, "learning_rate": 5.531617576988879e-07, "loss": 1.1693, "step": 1260 }, { "epoch": 3.467032967032967, "grad_norm": 0.07959607243537903, "learning_rate": 5.422487251355146e-07, "loss": 1.1625, "step": 1262 }, { "epoch": 3.4725274725274726, "grad_norm": 0.06803479790687561, "learning_rate": 5.314382417412062e-07, "loss": 1.162, "step": 1264 }, { "epoch": 3.478021978021978, "grad_norm": 0.07737120240926743, "learning_rate": 5.207305562077403e-07, "loss": 1.1705, "step": 1266 }, { "epoch": 3.4835164835164836, "grad_norm": 0.06910370290279388, "learning_rate": 5.101259148620618e-07, "loss": 1.1619, "step": 1268 }, { "epoch": 3.489010989010989, "grad_norm": 0.06398583203554153, "learning_rate": 4.99624561660616e-07, "loss": 1.1659, "step": 1270 }, { "epoch": 3.4945054945054945, "grad_norm": 0.06311172246932983, "learning_rate": 4.892267381837396e-07, "loss": 1.1595, "step": 1272 }, { "epoch": 3.5, "grad_norm": 0.06687135994434357, "learning_rate": 4.789326836300983e-07, "loss": 1.1639, "step": 1274 }, { "epoch": 3.5054945054945055, "grad_norm": 0.06371094286441803, "learning_rate": 4.687426348111834e-07, "loss": 1.1644, "step": 1276 }, { "epoch": 3.510989010989011, "grad_norm": 0.0776761993765831, "learning_rate": 4.586568261458729e-07, "loss": 1.1745, "step": 1278 }, { "epoch": 3.5164835164835164, "grad_norm": 0.06816533952951431, "learning_rate": 4.486754896550288e-07, "loss": 1.1636, "step": 1280 }, { "epoch": 3.521978021978022, "grad_norm": 0.07598984241485596, "learning_rate": 4.3879885495616505e-07, "loss": 1.1657, "step": 1282 }, { "epoch": 3.5274725274725274, "grad_norm": 0.06942006200551987, "learning_rate": 4.290271492581627e-07, "loss": 1.1638, "step": 1284 }, { "epoch": 3.532967032967033, "grad_norm": 0.06620027124881744, "learning_rate": 4.1936059735604497e-07, "loss": 1.1698, "step": 1286 }, { "epoch": 3.5384615384615383, "grad_norm": 0.07451992481946945, "learning_rate": 4.0979942162580387e-07, "loss": 1.1663, "step": 1288 }, { "epoch": 3.543956043956044, "grad_norm": 0.07096972316503525, "learning_rate": 4.003438420192873e-07, "loss": 1.1627, "step": 1290 }, { "epoch": 3.5494505494505493, "grad_norm": 0.0655079111456871, "learning_rate": 3.9099407605913576e-07, "loss": 1.1645, "step": 1292 }, { "epoch": 3.5549450549450547, "grad_norm": 0.06677290052175522, "learning_rate": 3.8175033883378233e-07, "loss": 1.1642, "step": 1294 }, { "epoch": 3.5604395604395602, "grad_norm": 0.06589250266551971, "learning_rate": 3.7261284299249967e-07, "loss": 1.1661, "step": 1296 }, { "epoch": 3.5659340659340657, "grad_norm": 0.06369265913963318, "learning_rate": 3.63581798740511e-07, "loss": 1.1591, "step": 1298 }, { "epoch": 3.571428571428571, "grad_norm": 0.06584078073501587, "learning_rate": 3.5465741383415684e-07, "loss": 1.1669, "step": 1300 }, { "epoch": 3.5769230769230766, "grad_norm": 0.07220979779958725, "learning_rate": 3.4583989357611037e-07, "loss": 1.1637, "step": 1302 }, { "epoch": 3.5824175824175826, "grad_norm": 0.07124118506908417, "learning_rate": 3.371294408106585e-07, "loss": 1.1685, "step": 1304 }, { "epoch": 3.587912087912088, "grad_norm": 0.0632823258638382, "learning_rate": 3.285262559190322e-07, "loss": 1.1646, "step": 1306 }, { "epoch": 3.5934065934065935, "grad_norm": 0.07331771403551102, "learning_rate": 3.20030536814801e-07, "loss": 1.1639, "step": 1308 }, { "epoch": 3.598901098901099, "grad_norm": 0.07538831979036331, "learning_rate": 3.1164247893931575e-07, "loss": 1.1642, "step": 1310 }, { "epoch": 3.6043956043956045, "grad_norm": 0.06919172406196594, "learning_rate": 3.033622752572157e-07, "loss": 1.165, "step": 1312 }, { "epoch": 3.60989010989011, "grad_norm": 0.0683453157544136, "learning_rate": 2.951901162519877e-07, "loss": 1.1657, "step": 1314 }, { "epoch": 3.60989010989011, "eval_loss": 1.1666840314865112, "eval_runtime": 300.2612, "eval_samples_per_second": 17.238, "eval_steps_per_second": 0.27, "step": 1314 }, { "epoch": 3.6153846153846154, "grad_norm": 0.07349937409162521, "learning_rate": 2.8712618992158656e-07, "loss": 1.167, "step": 1316 }, { "epoch": 3.620879120879121, "grad_norm": 0.0776103138923645, "learning_rate": 2.791706817741041e-07, "loss": 1.1628, "step": 1318 }, { "epoch": 3.6263736263736264, "grad_norm": 0.06213715299963951, "learning_rate": 2.7132377482351037e-07, "loss": 1.1623, "step": 1320 }, { "epoch": 3.631868131868132, "grad_norm": 0.06590726226568222, "learning_rate": 2.635856495854372e-07, "loss": 1.1675, "step": 1322 }, { "epoch": 3.6373626373626373, "grad_norm": 0.06559966504573822, "learning_rate": 2.5595648407302496e-07, "loss": 1.1692, "step": 1324 }, { "epoch": 3.642857142857143, "grad_norm": 0.0632384866476059, "learning_rate": 2.484364537928341e-07, "loss": 1.1632, "step": 1326 }, { "epoch": 3.6483516483516483, "grad_norm": 0.07011737674474716, "learning_rate": 2.41025731740801e-07, "loss": 1.1626, "step": 1328 }, { "epoch": 3.6538461538461537, "grad_norm": 0.06871969997882843, "learning_rate": 2.3372448839825978e-07, "loss": 1.1597, "step": 1330 }, { "epoch": 3.659340659340659, "grad_norm": 0.06817732751369476, "learning_rate": 2.2653289172802295e-07, "loss": 1.1714, "step": 1332 }, { "epoch": 3.6648351648351647, "grad_norm": 0.06385420262813568, "learning_rate": 2.194511071705141e-07, "loss": 1.1747, "step": 1334 }, { "epoch": 3.67032967032967, "grad_norm": 0.06753863394260406, "learning_rate": 2.1247929763996534e-07, "loss": 1.1602, "step": 1336 }, { "epoch": 3.675824175824176, "grad_norm": 0.0678897351026535, "learning_rate": 2.0561762352066638e-07, "loss": 1.1685, "step": 1338 }, { "epoch": 3.6813186813186816, "grad_norm": 0.06596899032592773, "learning_rate": 1.988662426632765e-07, "loss": 1.1657, "step": 1340 }, { "epoch": 3.686813186813187, "grad_norm": 0.0690767765045166, "learning_rate": 1.922253103811944e-07, "loss": 1.167, "step": 1342 }, { "epoch": 3.6923076923076925, "grad_norm": 0.07675404101610184, "learning_rate": 1.85694979446982e-07, "loss": 1.1661, "step": 1344 }, { "epoch": 3.697802197802198, "grad_norm": 0.06820071488618851, "learning_rate": 1.7927540008885414e-07, "loss": 1.1635, "step": 1346 }, { "epoch": 3.7032967032967035, "grad_norm": 0.07236689329147339, "learning_rate": 1.729667199872187e-07, "loss": 1.1581, "step": 1348 }, { "epoch": 3.708791208791209, "grad_norm": 0.06689772009849548, "learning_rate": 1.6676908427128103e-07, "loss": 1.1617, "step": 1350 }, { "epoch": 3.7142857142857144, "grad_norm": 0.06681732088327408, "learning_rate": 1.6068263551570596e-07, "loss": 1.1732, "step": 1352 }, { "epoch": 3.71978021978022, "grad_norm": 0.07003802061080933, "learning_rate": 1.5470751373733773e-07, "loss": 1.1631, "step": 1354 }, { "epoch": 3.7252747252747254, "grad_norm": 0.06234800070524216, "learning_rate": 1.488438563919764e-07, "loss": 1.1696, "step": 1356 }, { "epoch": 3.730769230769231, "grad_norm": 0.0727643072605133, "learning_rate": 1.4309179837122045e-07, "loss": 1.1654, "step": 1358 }, { "epoch": 3.7362637362637363, "grad_norm": 0.0751447007060051, "learning_rate": 1.374514719993575e-07, "loss": 1.1647, "step": 1360 }, { "epoch": 3.741758241758242, "grad_norm": 0.061782509088516235, "learning_rate": 1.3192300703032733e-07, "loss": 1.1672, "step": 1362 }, { "epoch": 3.7472527472527473, "grad_norm": 0.06823485344648361, "learning_rate": 1.2650653064473106e-07, "loss": 1.1682, "step": 1364 }, { "epoch": 3.7527472527472527, "grad_norm": 0.061277277767658234, "learning_rate": 1.2120216744690716e-07, "loss": 1.1671, "step": 1366 }, { "epoch": 3.758241758241758, "grad_norm": 0.06526022404432297, "learning_rate": 1.1601003946206723e-07, "loss": 1.1643, "step": 1368 }, { "epoch": 3.7637362637362637, "grad_norm": 0.06104410067200661, "learning_rate": 1.1093026613348601e-07, "loss": 1.1574, "step": 1370 }, { "epoch": 3.769230769230769, "grad_norm": 0.07134535163640976, "learning_rate": 1.0596296431975406e-07, "loss": 1.1614, "step": 1372 }, { "epoch": 3.7747252747252746, "grad_norm": 0.06213730573654175, "learning_rate": 1.0110824829209164e-07, "loss": 1.1657, "step": 1374 }, { "epoch": 3.78021978021978, "grad_norm": 0.06847725808620453, "learning_rate": 9.636622973171583e-08, "loss": 1.1663, "step": 1376 }, { "epoch": 3.7857142857142856, "grad_norm": 0.07414616644382477, "learning_rate": 9.17370177272775e-08, "loss": 1.1668, "step": 1378 }, { "epoch": 3.791208791208791, "grad_norm": 0.06581508368253708, "learning_rate": 8.72207187723445e-08, "loss": 1.1693, "step": 1380 }, { "epoch": 3.7967032967032965, "grad_norm": 0.06397808343172073, "learning_rate": 8.281743676295639e-08, "loss": 1.159, "step": 1382 }, { "epoch": 3.802197802197802, "grad_norm": 0.06523909419775009, "learning_rate": 7.852727299523577e-08, "loss": 1.1562, "step": 1384 }, { "epoch": 3.8076923076923075, "grad_norm": 0.06779211014509201, "learning_rate": 7.435032616305238e-08, "loss": 1.1727, "step": 1386 }, { "epoch": 3.813186813186813, "grad_norm": 0.06538354605436325, "learning_rate": 7.028669235575714e-08, "loss": 1.16, "step": 1388 }, { "epoch": 3.8186813186813184, "grad_norm": 0.06357073783874512, "learning_rate": 6.633646505597113e-08, "loss": 1.1637, "step": 1390 }, { "epoch": 3.824175824175824, "grad_norm": 0.0635320246219635, "learning_rate": 6.249973513743345e-08, "loss": 1.1639, "step": 1392 }, { "epoch": 3.82967032967033, "grad_norm": 0.06830117851495743, "learning_rate": 5.8776590862911764e-08, "loss": 1.1585, "step": 1394 }, { "epoch": 3.8351648351648353, "grad_norm": 0.06444601714611053, "learning_rate": 5.5167117882171104e-08, "loss": 1.1655, "step": 1396 }, { "epoch": 3.840659340659341, "grad_norm": 0.06375352293252945, "learning_rate": 5.167139923000553e-08, "loss": 1.167, "step": 1398 }, { "epoch": 3.8461538461538463, "grad_norm": 0.06707257032394409, "learning_rate": 4.828951532432457e-08, "loss": 1.171, "step": 1400 }, { "epoch": 3.8516483516483517, "grad_norm": 0.06371629983186722, "learning_rate": 4.5021543964306466e-08, "loss": 1.1648, "step": 1402 }, { "epoch": 3.857142857142857, "grad_norm": 0.0662907212972641, "learning_rate": 4.186756032860728e-08, "loss": 1.1609, "step": 1404 }, { "epoch": 3.8626373626373627, "grad_norm": 0.06402067840099335, "learning_rate": 3.8827636973630126e-08, "loss": 1.1653, "step": 1406 }, { "epoch": 3.868131868131868, "grad_norm": 0.07026324421167374, "learning_rate": 3.590184383185758e-08, "loss": 1.1583, "step": 1408 }, { "epoch": 3.8736263736263736, "grad_norm": 0.06799634546041489, "learning_rate": 3.309024821024354e-08, "loss": 1.1688, "step": 1410 }, { "epoch": 3.879120879120879, "grad_norm": 0.06444913148880005, "learning_rate": 3.039291478866169e-08, "loss": 1.1616, "step": 1412 }, { "epoch": 3.8846153846153846, "grad_norm": 0.06899133324623108, "learning_rate": 2.7809905618422227e-08, "loss": 1.1598, "step": 1414 }, { "epoch": 3.89010989010989, "grad_norm": 0.06785506010055542, "learning_rate": 2.534128012083914e-08, "loss": 1.1532, "step": 1416 }, { "epoch": 3.8956043956043955, "grad_norm": 0.06701900064945221, "learning_rate": 2.298709508586794e-08, "loss": 1.1654, "step": 1418 }, { "epoch": 3.901098901098901, "grad_norm": 0.06285455822944641, "learning_rate": 2.074740467079672e-08, "loss": 1.157, "step": 1420 }, { "epoch": 3.9065934065934065, "grad_norm": 0.06540726125240326, "learning_rate": 1.862226039899995e-08, "loss": 1.1608, "step": 1422 }, { "epoch": 3.912087912087912, "grad_norm": 0.07258068770170212, "learning_rate": 1.661171115875493e-08, "loss": 1.1696, "step": 1424 }, { "epoch": 3.9175824175824174, "grad_norm": 0.06025758385658264, "learning_rate": 1.4715803202116075e-08, "loss": 1.1644, "step": 1426 }, { "epoch": 3.9230769230769234, "grad_norm": 0.07302309572696686, "learning_rate": 1.2934580143851294e-08, "loss": 1.1621, "step": 1428 }, { "epoch": 3.928571428571429, "grad_norm": 0.05992519110441208, "learning_rate": 1.1268082960436688e-08, "loss": 1.1689, "step": 1430 }, { "epoch": 3.9340659340659343, "grad_norm": 0.06481453776359558, "learning_rate": 9.716349989118412e-09, "loss": 1.166, "step": 1432 }, { "epoch": 3.9395604395604398, "grad_norm": 0.07549932599067688, "learning_rate": 8.279416927026163e-09, "loss": 1.159, "step": 1434 }, { "epoch": 3.9450549450549453, "grad_norm": 0.06642387062311172, "learning_rate": 6.9573168303532775e-09, "loss": 1.1602, "step": 1436 }, { "epoch": 3.9505494505494507, "grad_norm": 0.0685223713517189, "learning_rate": 5.750080113598455e-09, "loss": 1.1659, "step": 1438 }, { "epoch": 3.956043956043956, "grad_norm": 0.06598740071058273, "learning_rate": 4.65773454886298e-09, "loss": 1.1607, "step": 1440 }, { "epoch": 3.9615384615384617, "grad_norm": 0.06653392314910889, "learning_rate": 3.6803052652134572e-09, "loss": 1.1545, "step": 1442 }, { "epoch": 3.967032967032967, "grad_norm": 0.06884676963090897, "learning_rate": 2.817814748104497e-09, "loss": 1.1557, "step": 1444 }, { "epoch": 3.9725274725274726, "grad_norm": 0.06552927196025848, "learning_rate": 2.070282838859683e-09, "loss": 1.1591, "step": 1446 }, { "epoch": 3.978021978021978, "grad_norm": 0.06410259753465652, "learning_rate": 1.4377267342158274e-09, "loss": 1.169, "step": 1448 }, { "epoch": 3.9835164835164836, "grad_norm": 0.06548061966896057, "learning_rate": 9.201609859271765e-10, "loss": 1.1665, "step": 1450 }, { "epoch": 3.989010989010989, "grad_norm": 0.06399548798799515, "learning_rate": 5.17597500432343e-10, "loss": 1.1717, "step": 1452 }, { "epoch": 3.9945054945054945, "grad_norm": 0.06421104073524475, "learning_rate": 2.3004553857675082e-10, "loss": 1.1725, "step": 1454 }, { "epoch": 4.0, "grad_norm": 0.06589093804359436, "learning_rate": 5.751171540391287e-11, "loss": 1.1709, "step": 1456 } ], "logging_steps": 2, "max_steps": 1456, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 364, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.7105900771564585e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }