diff --git "a/trainer_state.json" "b/trainer_state.json" deleted file mode 100644--- "a/trainer_state.json" +++ /dev/null @@ -1,21523 +0,0 @@ -{ - "best_global_step": 5900, - "best_metric": 1.0105234384536743, - "best_model_checkpoint": "./outputs/checkpoint-5900", - "epoch": 2.176326530612245, - "eval_steps": 100, - "global_step": 6000, - "is_hyper_param_search": false, - "is_local_process_zero": true, - "is_world_process_zero": true, - "log_history": [ - { - "epoch": 0.0007256235827664399, - "grad_norm": 0.88671875, - "learning_rate": 2.0000000000000003e-06, - "loss": 1.5116, - "step": 2 - }, - { - "epoch": 0.0014512471655328798, - "grad_norm": 1.8046875, - "learning_rate": 6e-06, - "loss": 1.7699, - "step": 4 - }, - { - "epoch": 0.0021768707482993197, - "grad_norm": 0.7890625, - "learning_rate": 1e-05, - "loss": 1.4788, - "step": 6 - }, - { - "epoch": 0.0029024943310657597, - "grad_norm": 1.3671875, - "learning_rate": 1.4000000000000001e-05, - "loss": 1.6688, - "step": 8 - }, - { - "epoch": 0.0036281179138321997, - "grad_norm": 1.7421875, - "learning_rate": 1.8e-05, - "loss": 1.798, - "step": 10 - }, - { - "epoch": 0.004353741496598639, - "grad_norm": 1.171875, - "learning_rate": 2.2000000000000003e-05, - "loss": 1.6686, - "step": 12 - }, - { - "epoch": 0.005079365079365079, - "grad_norm": 1.1484375, - "learning_rate": 2.6000000000000002e-05, - "loss": 1.8021, - "step": 14 - }, - { - "epoch": 0.005804988662131519, - "grad_norm": 0.7890625, - "learning_rate": 3e-05, - "loss": 1.6768, - "step": 16 - }, - { - "epoch": 0.006530612244897959, - "grad_norm": 0.83203125, - "learning_rate": 3.4000000000000007e-05, - "loss": 1.4667, - "step": 18 - }, - { - "epoch": 0.0072562358276643995, - "grad_norm": 0.98828125, - "learning_rate": 3.8e-05, - "loss": 1.6074, - "step": 20 - }, - { - "epoch": 0.00798185941043084, - "grad_norm": 0.90234375, - "learning_rate": 4.2e-05, - "loss": 1.6353, - "step": 22 - }, - { - "epoch": 0.008707482993197279, - "grad_norm": 0.92578125, - "learning_rate": 4.600000000000001e-05, - "loss": 1.5452, - "step": 24 - }, - { - "epoch": 0.00943310657596372, - "grad_norm": 0.87109375, - "learning_rate": 5e-05, - "loss": 1.6205, - "step": 26 - }, - { - "epoch": 0.010158730158730159, - "grad_norm": 0.61328125, - "learning_rate": 5.4000000000000005e-05, - "loss": 1.4568, - "step": 28 - }, - { - "epoch": 0.010884353741496598, - "grad_norm": 0.69921875, - "learning_rate": 5.8e-05, - "loss": 1.4901, - "step": 30 - }, - { - "epoch": 0.011609977324263039, - "grad_norm": 0.98046875, - "learning_rate": 6.2e-05, - "loss": 1.422, - "step": 32 - }, - { - "epoch": 0.012335600907029478, - "grad_norm": 0.984375, - "learning_rate": 6.6e-05, - "loss": 1.406, - "step": 34 - }, - { - "epoch": 0.013061224489795919, - "grad_norm": 0.69921875, - "learning_rate": 7e-05, - "loss": 1.2822, - "step": 36 - }, - { - "epoch": 0.013786848072562358, - "grad_norm": 0.86328125, - "learning_rate": 7.4e-05, - "loss": 1.3677, - "step": 38 - }, - { - "epoch": 0.014512471655328799, - "grad_norm": 0.78125, - "learning_rate": 7.800000000000001e-05, - "loss": 1.3597, - "step": 40 - }, - { - "epoch": 0.015238095238095238, - "grad_norm": 0.828125, - "learning_rate": 8.2e-05, - "loss": 1.263, - "step": 42 - }, - { - "epoch": 0.01596371882086168, - "grad_norm": 0.75390625, - "learning_rate": 8.6e-05, - "loss": 1.314, - "step": 44 - }, - { - "epoch": 0.016689342403628116, - "grad_norm": 0.46484375, - "learning_rate": 9e-05, - "loss": 1.2784, - "step": 46 - }, - { - "epoch": 0.017414965986394557, - "grad_norm": 0.65625, - "learning_rate": 9.4e-05, - "loss": 1.1945, - "step": 48 - }, - { - "epoch": 0.018140589569160998, - "grad_norm": 0.5703125, - "learning_rate": 9.8e-05, - "loss": 1.2637, - "step": 50 - }, - { - "epoch": 0.01886621315192744, - "grad_norm": 0.625, - "learning_rate": 0.00010200000000000001, - "loss": 1.1792, - "step": 52 - }, - { - "epoch": 0.019591836734693877, - "grad_norm": 0.349609375, - "learning_rate": 0.00010600000000000002, - "loss": 1.1901, - "step": 54 - }, - { - "epoch": 0.020317460317460317, - "grad_norm": 0.392578125, - "learning_rate": 0.00011000000000000002, - "loss": 1.1808, - "step": 56 - }, - { - "epoch": 0.02104308390022676, - "grad_norm": 0.34765625, - "learning_rate": 0.00011399999999999999, - "loss": 1.134, - "step": 58 - }, - { - "epoch": 0.021768707482993196, - "grad_norm": 0.326171875, - "learning_rate": 0.000118, - "loss": 1.1071, - "step": 60 - }, - { - "epoch": 0.022494331065759637, - "grad_norm": 0.255859375, - "learning_rate": 0.000122, - "loss": 1.1066, - "step": 62 - }, - { - "epoch": 0.023219954648526078, - "grad_norm": 0.30859375, - "learning_rate": 0.000126, - "loss": 1.1922, - "step": 64 - }, - { - "epoch": 0.02394557823129252, - "grad_norm": 0.2578125, - "learning_rate": 0.00013000000000000002, - "loss": 1.1028, - "step": 66 - }, - { - "epoch": 0.024671201814058956, - "grad_norm": 0.369140625, - "learning_rate": 0.000134, - "loss": 1.0819, - "step": 68 - }, - { - "epoch": 0.025396825396825397, - "grad_norm": 0.275390625, - "learning_rate": 0.000138, - "loss": 1.1744, - "step": 70 - }, - { - "epoch": 0.026122448979591838, - "grad_norm": 0.26171875, - "learning_rate": 0.000142, - "loss": 1.0784, - "step": 72 - }, - { - "epoch": 0.026848072562358275, - "grad_norm": 0.21875, - "learning_rate": 0.000146, - "loss": 1.0396, - "step": 74 - }, - { - "epoch": 0.027573696145124716, - "grad_norm": 0.2333984375, - "learning_rate": 0.00015000000000000001, - "loss": 1.2014, - "step": 76 - }, - { - "epoch": 0.028299319727891157, - "grad_norm": 0.23828125, - "learning_rate": 0.000154, - "loss": 1.1024, - "step": 78 - }, - { - "epoch": 0.029024943310657598, - "grad_norm": 0.28125, - "learning_rate": 0.00015800000000000002, - "loss": 0.9752, - "step": 80 - }, - { - "epoch": 0.029750566893424035, - "grad_norm": 0.2373046875, - "learning_rate": 0.000162, - "loss": 1.0902, - "step": 82 - }, - { - "epoch": 0.030476190476190476, - "grad_norm": 0.19921875, - "learning_rate": 0.000166, - "loss": 1.0058, - "step": 84 - }, - { - "epoch": 0.031201814058956917, - "grad_norm": 0.2197265625, - "learning_rate": 0.00017, - "loss": 1.0682, - "step": 86 - }, - { - "epoch": 0.03192743764172336, - "grad_norm": 0.2119140625, - "learning_rate": 0.000174, - "loss": 1.0874, - "step": 88 - }, - { - "epoch": 0.0326530612244898, - "grad_norm": 0.458984375, - "learning_rate": 0.00017800000000000002, - "loss": 1.0739, - "step": 90 - }, - { - "epoch": 0.03337868480725623, - "grad_norm": 0.212890625, - "learning_rate": 0.000182, - "loss": 1.0259, - "step": 92 - }, - { - "epoch": 0.034104308390022674, - "grad_norm": 0.2470703125, - "learning_rate": 0.00018600000000000002, - "loss": 1.0279, - "step": 94 - }, - { - "epoch": 0.034829931972789115, - "grad_norm": 0.2001953125, - "learning_rate": 0.00019, - "loss": 1.0263, - "step": 96 - }, - { - "epoch": 0.035555555555555556, - "grad_norm": 0.251953125, - "learning_rate": 0.000194, - "loss": 1.0206, - "step": 98 - }, - { - "epoch": 0.036281179138321996, - "grad_norm": 0.2353515625, - "learning_rate": 0.00019800000000000002, - "loss": 1.1275, - "step": 100 - }, - { - "epoch": 0.036281179138321996, - "eval_loss": 1.0672887563705444, - "eval_runtime": 99.6091, - "eval_samples_per_second": 18.071, - "eval_steps_per_second": 0.572, - "step": 100 - }, - { - "epoch": 0.03700680272108844, - "grad_norm": 0.267578125, - "learning_rate": 0.0001999755231917758, - "loss": 1.1141, - "step": 102 - }, - { - "epoch": 0.03773242630385488, - "grad_norm": 0.2080078125, - "learning_rate": 0.00019992656957532737, - "loss": 1.0543, - "step": 104 - }, - { - "epoch": 0.03845804988662131, - "grad_norm": 0.20703125, - "learning_rate": 0.00019987761595887898, - "loss": 1.0389, - "step": 106 - }, - { - "epoch": 0.03918367346938775, - "grad_norm": 0.1982421875, - "learning_rate": 0.00019982866234243056, - "loss": 1.0699, - "step": 108 - }, - { - "epoch": 0.039909297052154194, - "grad_norm": 0.20703125, - "learning_rate": 0.00019977970872598214, - "loss": 1.0756, - "step": 110 - }, - { - "epoch": 0.040634920634920635, - "grad_norm": 0.1845703125, - "learning_rate": 0.00019973075510953373, - "loss": 1.0309, - "step": 112 - }, - { - "epoch": 0.041360544217687076, - "grad_norm": 0.1884765625, - "learning_rate": 0.0001996818014930853, - "loss": 0.9821, - "step": 114 - }, - { - "epoch": 0.04208616780045352, - "grad_norm": 0.2353515625, - "learning_rate": 0.0001996328478766369, - "loss": 1.0897, - "step": 116 - }, - { - "epoch": 0.04281179138321996, - "grad_norm": 0.2080078125, - "learning_rate": 0.0001995838942601885, - "loss": 1.0576, - "step": 118 - }, - { - "epoch": 0.04353741496598639, - "grad_norm": 0.2001953125, - "learning_rate": 0.00019953494064374008, - "loss": 1.0482, - "step": 120 - }, - { - "epoch": 0.04426303854875283, - "grad_norm": 0.259765625, - "learning_rate": 0.00019948598702729164, - "loss": 1.0018, - "step": 122 - }, - { - "epoch": 0.04498866213151927, - "grad_norm": 0.1865234375, - "learning_rate": 0.00019943703341084322, - "loss": 1.0637, - "step": 124 - }, - { - "epoch": 0.045714285714285714, - "grad_norm": 0.1806640625, - "learning_rate": 0.00019938807979439483, - "loss": 1.0471, - "step": 126 - }, - { - "epoch": 0.046439909297052155, - "grad_norm": 0.1875, - "learning_rate": 0.0001993391261779464, - "loss": 0.9958, - "step": 128 - }, - { - "epoch": 0.047165532879818596, - "grad_norm": 0.201171875, - "learning_rate": 0.000199290172561498, - "loss": 1.0396, - "step": 130 - }, - { - "epoch": 0.04789115646258504, - "grad_norm": 0.2333984375, - "learning_rate": 0.00019924121894504958, - "loss": 1.0407, - "step": 132 - }, - { - "epoch": 0.04861678004535147, - "grad_norm": 0.201171875, - "learning_rate": 0.00019919226532860116, - "loss": 1.043, - "step": 134 - }, - { - "epoch": 0.04934240362811791, - "grad_norm": 0.193359375, - "learning_rate": 0.00019914331171215274, - "loss": 1.0859, - "step": 136 - }, - { - "epoch": 0.05006802721088435, - "grad_norm": 0.181640625, - "learning_rate": 0.00019909435809570435, - "loss": 1.1251, - "step": 138 - }, - { - "epoch": 0.050793650793650794, - "grad_norm": 0.2431640625, - "learning_rate": 0.0001990454044792559, - "loss": 1.0666, - "step": 140 - }, - { - "epoch": 0.051519274376417235, - "grad_norm": 0.203125, - "learning_rate": 0.0001989964508628075, - "loss": 0.9962, - "step": 142 - }, - { - "epoch": 0.052244897959183675, - "grad_norm": 0.1923828125, - "learning_rate": 0.00019894749724635907, - "loss": 1.1262, - "step": 144 - }, - { - "epoch": 0.052970521541950116, - "grad_norm": 0.1875, - "learning_rate": 0.00019889854362991068, - "loss": 1.0174, - "step": 146 - }, - { - "epoch": 0.05369614512471655, - "grad_norm": 0.181640625, - "learning_rate": 0.00019884959001346226, - "loss": 1.0509, - "step": 148 - }, - { - "epoch": 0.05442176870748299, - "grad_norm": 0.21875, - "learning_rate": 0.00019880063639701382, - "loss": 1.05, - "step": 150 - }, - { - "epoch": 0.05514739229024943, - "grad_norm": 0.2109375, - "learning_rate": 0.00019875168278056543, - "loss": 1.134, - "step": 152 - }, - { - "epoch": 0.05587301587301587, - "grad_norm": 0.1650390625, - "learning_rate": 0.000198702729164117, - "loss": 0.9511, - "step": 154 - }, - { - "epoch": 0.056598639455782314, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001986537755476686, - "loss": 0.9989, - "step": 156 - }, - { - "epoch": 0.057324263038548755, - "grad_norm": 0.162109375, - "learning_rate": 0.0001986048219312202, - "loss": 0.9466, - "step": 158 - }, - { - "epoch": 0.058049886621315196, - "grad_norm": 0.162109375, - "learning_rate": 0.00019855586831477175, - "loss": 0.9605, - "step": 160 - }, - { - "epoch": 0.05877551020408163, - "grad_norm": 0.2001953125, - "learning_rate": 0.00019850691469832334, - "loss": 1.0511, - "step": 162 - }, - { - "epoch": 0.05950113378684807, - "grad_norm": 0.205078125, - "learning_rate": 0.00019845796108187492, - "loss": 1.0575, - "step": 164 - }, - { - "epoch": 0.06022675736961451, - "grad_norm": 0.189453125, - "learning_rate": 0.00019840900746542653, - "loss": 1.0211, - "step": 166 - }, - { - "epoch": 0.06095238095238095, - "grad_norm": 0.1923828125, - "learning_rate": 0.0001983600538489781, - "loss": 1.0435, - "step": 168 - }, - { - "epoch": 0.06167800453514739, - "grad_norm": 0.2021484375, - "learning_rate": 0.00019831110023252967, - "loss": 0.9651, - "step": 170 - }, - { - "epoch": 0.062403628117913834, - "grad_norm": 0.17578125, - "learning_rate": 0.00019826214661608128, - "loss": 0.9683, - "step": 172 - }, - { - "epoch": 0.06312925170068028, - "grad_norm": 0.1796875, - "learning_rate": 0.00019821319299963286, - "loss": 1.0064, - "step": 174 - }, - { - "epoch": 0.06385487528344672, - "grad_norm": 0.1875, - "learning_rate": 0.00019816423938318444, - "loss": 1.0669, - "step": 176 - }, - { - "epoch": 0.06458049886621316, - "grad_norm": 0.1796875, - "learning_rate": 0.00019811528576673605, - "loss": 1.0931, - "step": 178 - }, - { - "epoch": 0.0653061224489796, - "grad_norm": 0.19921875, - "learning_rate": 0.0001980663321502876, - "loss": 1.04, - "step": 180 - }, - { - "epoch": 0.06603174603174604, - "grad_norm": 0.1884765625, - "learning_rate": 0.0001980173785338392, - "loss": 0.9917, - "step": 182 - }, - { - "epoch": 0.06675736961451247, - "grad_norm": 0.166015625, - "learning_rate": 0.00019796842491739077, - "loss": 0.9985, - "step": 184 - }, - { - "epoch": 0.0674829931972789, - "grad_norm": 0.1728515625, - "learning_rate": 0.00019791947130094238, - "loss": 0.9511, - "step": 186 - }, - { - "epoch": 0.06820861678004535, - "grad_norm": 0.169921875, - "learning_rate": 0.00019787051768449396, - "loss": 0.9866, - "step": 188 - }, - { - "epoch": 0.06893424036281179, - "grad_norm": 0.1982421875, - "learning_rate": 0.00019782156406804552, - "loss": 1.0879, - "step": 190 - }, - { - "epoch": 0.06965986394557823, - "grad_norm": 0.181640625, - "learning_rate": 0.00019777261045159713, - "loss": 1.006, - "step": 192 - }, - { - "epoch": 0.07038548752834467, - "grad_norm": 0.1787109375, - "learning_rate": 0.0001977236568351487, - "loss": 1.0737, - "step": 194 - }, - { - "epoch": 0.07111111111111111, - "grad_norm": 0.203125, - "learning_rate": 0.0001976747032187003, - "loss": 0.9719, - "step": 196 - }, - { - "epoch": 0.07183673469387755, - "grad_norm": 0.21484375, - "learning_rate": 0.0001976257496022519, - "loss": 1.0539, - "step": 198 - }, - { - "epoch": 0.07256235827664399, - "grad_norm": 0.19921875, - "learning_rate": 0.00019757679598580345, - "loss": 1.0863, - "step": 200 - }, - { - "epoch": 0.07256235827664399, - "eval_loss": 1.0431190729141235, - "eval_runtime": 99.2841, - "eval_samples_per_second": 18.13, - "eval_steps_per_second": 0.574, - "step": 200 - }, - { - "epoch": 0.07328798185941043, - "grad_norm": 0.2021484375, - "learning_rate": 0.00019752784236935504, - "loss": 1.1022, - "step": 202 - }, - { - "epoch": 0.07401360544217687, - "grad_norm": 0.177734375, - "learning_rate": 0.00019747888875290662, - "loss": 0.9781, - "step": 204 - }, - { - "epoch": 0.07473922902494332, - "grad_norm": 0.21875, - "learning_rate": 0.00019742993513645823, - "loss": 1.0735, - "step": 206 - }, - { - "epoch": 0.07546485260770976, - "grad_norm": 0.1865234375, - "learning_rate": 0.0001973809815200098, - "loss": 0.957, - "step": 208 - }, - { - "epoch": 0.0761904761904762, - "grad_norm": 0.158203125, - "learning_rate": 0.00019733202790356137, - "loss": 1.0229, - "step": 210 - }, - { - "epoch": 0.07691609977324262, - "grad_norm": 0.1962890625, - "learning_rate": 0.00019728307428711298, - "loss": 0.9876, - "step": 212 - }, - { - "epoch": 0.07764172335600907, - "grad_norm": 0.1884765625, - "learning_rate": 0.00019723412067066456, - "loss": 1.0047, - "step": 214 - }, - { - "epoch": 0.0783673469387755, - "grad_norm": 0.1875, - "learning_rate": 0.00019718516705421614, - "loss": 0.9995, - "step": 216 - }, - { - "epoch": 0.07909297052154195, - "grad_norm": 0.1923828125, - "learning_rate": 0.00019713621343776772, - "loss": 1.0447, - "step": 218 - }, - { - "epoch": 0.07981859410430839, - "grad_norm": 0.193359375, - "learning_rate": 0.0001970872598213193, - "loss": 1.0969, - "step": 220 - }, - { - "epoch": 0.08054421768707483, - "grad_norm": 0.203125, - "learning_rate": 0.0001970383062048709, - "loss": 0.9948, - "step": 222 - }, - { - "epoch": 0.08126984126984127, - "grad_norm": 0.1591796875, - "learning_rate": 0.00019698935258842247, - "loss": 0.9642, - "step": 224 - }, - { - "epoch": 0.08199546485260771, - "grad_norm": 0.19140625, - "learning_rate": 0.00019694039897197408, - "loss": 1.0414, - "step": 226 - }, - { - "epoch": 0.08272108843537415, - "grad_norm": 0.1943359375, - "learning_rate": 0.00019689144535552563, - "loss": 0.9293, - "step": 228 - }, - { - "epoch": 0.08344671201814059, - "grad_norm": 0.2021484375, - "learning_rate": 0.00019684249173907722, - "loss": 0.9743, - "step": 230 - }, - { - "epoch": 0.08417233560090703, - "grad_norm": 0.189453125, - "learning_rate": 0.00019679353812262883, - "loss": 1.0206, - "step": 232 - }, - { - "epoch": 0.08489795918367347, - "grad_norm": 0.1787109375, - "learning_rate": 0.0001967445845061804, - "loss": 0.9956, - "step": 234 - }, - { - "epoch": 0.08562358276643992, - "grad_norm": 0.185546875, - "learning_rate": 0.000196695630889732, - "loss": 1.0204, - "step": 236 - }, - { - "epoch": 0.08634920634920636, - "grad_norm": 0.177734375, - "learning_rate": 0.00019664667727328357, - "loss": 0.9644, - "step": 238 - }, - { - "epoch": 0.08707482993197278, - "grad_norm": 0.1748046875, - "learning_rate": 0.00019659772365683515, - "loss": 0.9574, - "step": 240 - }, - { - "epoch": 0.08780045351473922, - "grad_norm": 0.181640625, - "learning_rate": 0.00019654877004038674, - "loss": 1.0699, - "step": 242 - }, - { - "epoch": 0.08852607709750566, - "grad_norm": 0.193359375, - "learning_rate": 0.00019649981642393832, - "loss": 1.0396, - "step": 244 - }, - { - "epoch": 0.0892517006802721, - "grad_norm": 0.1611328125, - "learning_rate": 0.00019645086280748993, - "loss": 0.9685, - "step": 246 - }, - { - "epoch": 0.08997732426303855, - "grad_norm": 0.162109375, - "learning_rate": 0.00019640190919104148, - "loss": 0.9963, - "step": 248 - }, - { - "epoch": 0.09070294784580499, - "grad_norm": 0.181640625, - "learning_rate": 0.00019635295557459307, - "loss": 1.0085, - "step": 250 - }, - { - "epoch": 0.09142857142857143, - "grad_norm": 0.216796875, - "learning_rate": 0.00019630400195814468, - "loss": 1.0444, - "step": 252 - }, - { - "epoch": 0.09215419501133787, - "grad_norm": 0.1591796875, - "learning_rate": 0.00019625504834169626, - "loss": 0.9959, - "step": 254 - }, - { - "epoch": 0.09287981859410431, - "grad_norm": 0.189453125, - "learning_rate": 0.00019620609472524784, - "loss": 1.0705, - "step": 256 - }, - { - "epoch": 0.09360544217687075, - "grad_norm": 0.1796875, - "learning_rate": 0.00019615714110879942, - "loss": 1.0762, - "step": 258 - }, - { - "epoch": 0.09433106575963719, - "grad_norm": 0.1826171875, - "learning_rate": 0.000196108187492351, - "loss": 1.0227, - "step": 260 - }, - { - "epoch": 0.09505668934240363, - "grad_norm": 0.2080078125, - "learning_rate": 0.00019605923387590259, - "loss": 1.0503, - "step": 262 - }, - { - "epoch": 0.09578231292517007, - "grad_norm": 0.1796875, - "learning_rate": 0.00019601028025945417, - "loss": 1.0013, - "step": 264 - }, - { - "epoch": 0.09650793650793651, - "grad_norm": 0.1689453125, - "learning_rate": 0.00019596132664300578, - "loss": 1.0844, - "step": 266 - }, - { - "epoch": 0.09723356009070294, - "grad_norm": 0.1767578125, - "learning_rate": 0.00019591237302655733, - "loss": 0.9824, - "step": 268 - }, - { - "epoch": 0.09795918367346938, - "grad_norm": 0.1533203125, - "learning_rate": 0.00019586341941010892, - "loss": 0.9688, - "step": 270 - }, - { - "epoch": 0.09868480725623582, - "grad_norm": 0.169921875, - "learning_rate": 0.00019581446579366052, - "loss": 0.923, - "step": 272 - }, - { - "epoch": 0.09941043083900226, - "grad_norm": 0.2197265625, - "learning_rate": 0.0001957655121772121, - "loss": 0.9817, - "step": 274 - }, - { - "epoch": 0.1001360544217687, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001957165585607637, - "loss": 0.9691, - "step": 276 - }, - { - "epoch": 0.10086167800453515, - "grad_norm": 0.1806640625, - "learning_rate": 0.00019566760494431527, - "loss": 1.019, - "step": 278 - }, - { - "epoch": 0.10158730158730159, - "grad_norm": 0.1767578125, - "learning_rate": 0.00019561865132786685, - "loss": 1.008, - "step": 280 - }, - { - "epoch": 0.10231292517006803, - "grad_norm": 0.162109375, - "learning_rate": 0.00019556969771141844, - "loss": 0.9477, - "step": 282 - }, - { - "epoch": 0.10303854875283447, - "grad_norm": 0.1748046875, - "learning_rate": 0.00019552074409497002, - "loss": 0.962, - "step": 284 - }, - { - "epoch": 0.10376417233560091, - "grad_norm": 0.169921875, - "learning_rate": 0.0001954717904785216, - "loss": 0.9685, - "step": 286 - }, - { - "epoch": 0.10448979591836735, - "grad_norm": 0.1650390625, - "learning_rate": 0.00019542283686207318, - "loss": 1.015, - "step": 288 - }, - { - "epoch": 0.10521541950113379, - "grad_norm": 0.166015625, - "learning_rate": 0.00019537388324562477, - "loss": 0.9902, - "step": 290 - }, - { - "epoch": 0.10594104308390023, - "grad_norm": 0.16796875, - "learning_rate": 0.00019532492962917637, - "loss": 1.0309, - "step": 292 - }, - { - "epoch": 0.10666666666666667, - "grad_norm": 0.1953125, - "learning_rate": 0.00019527597601272796, - "loss": 1.0179, - "step": 294 - }, - { - "epoch": 0.1073922902494331, - "grad_norm": 0.16796875, - "learning_rate": 0.0001952270223962795, - "loss": 1.0208, - "step": 296 - }, - { - "epoch": 0.10811791383219954, - "grad_norm": 0.16796875, - "learning_rate": 0.00019517806877983112, - "loss": 1.0252, - "step": 298 - }, - { - "epoch": 0.10884353741496598, - "grad_norm": 0.171875, - "learning_rate": 0.0001951291151633827, - "loss": 1.0185, - "step": 300 - }, - { - "epoch": 0.10884353741496598, - "eval_loss": 1.036349892616272, - "eval_runtime": 91.1673, - "eval_samples_per_second": 19.744, - "eval_steps_per_second": 0.625, - "step": 300 - }, - { - "epoch": 0.10956916099773242, - "grad_norm": 0.177734375, - "learning_rate": 0.00019508016154693429, - "loss": 0.9886, - "step": 302 - }, - { - "epoch": 0.11029478458049886, - "grad_norm": 0.1689453125, - "learning_rate": 0.00019503120793048587, - "loss": 1.0495, - "step": 304 - }, - { - "epoch": 0.1110204081632653, - "grad_norm": 0.1533203125, - "learning_rate": 0.00019498225431403745, - "loss": 1.0057, - "step": 306 - }, - { - "epoch": 0.11174603174603175, - "grad_norm": 0.1728515625, - "learning_rate": 0.00019493330069758903, - "loss": 1.042, - "step": 308 - }, - { - "epoch": 0.11247165532879819, - "grad_norm": 0.1767578125, - "learning_rate": 0.00019488434708114062, - "loss": 1.0619, - "step": 310 - }, - { - "epoch": 0.11319727891156463, - "grad_norm": 0.150390625, - "learning_rate": 0.00019483539346469222, - "loss": 1.0055, - "step": 312 - }, - { - "epoch": 0.11392290249433107, - "grad_norm": 0.1826171875, - "learning_rate": 0.0001947864398482438, - "loss": 0.9609, - "step": 314 - }, - { - "epoch": 0.11464852607709751, - "grad_norm": 0.1689453125, - "learning_rate": 0.00019473748623179536, - "loss": 1.0478, - "step": 316 - }, - { - "epoch": 0.11537414965986395, - "grad_norm": 0.1826171875, - "learning_rate": 0.00019468853261534697, - "loss": 1.0008, - "step": 318 - }, - { - "epoch": 0.11609977324263039, - "grad_norm": 0.15625, - "learning_rate": 0.00019463957899889855, - "loss": 0.9464, - "step": 320 - }, - { - "epoch": 0.11682539682539683, - "grad_norm": 0.16015625, - "learning_rate": 0.00019459062538245014, - "loss": 1.0368, - "step": 322 - }, - { - "epoch": 0.11755102040816326, - "grad_norm": 0.171875, - "learning_rate": 0.00019454167176600172, - "loss": 1.0134, - "step": 324 - }, - { - "epoch": 0.1182766439909297, - "grad_norm": 0.15625, - "learning_rate": 0.0001944927181495533, - "loss": 0.9533, - "step": 326 - }, - { - "epoch": 0.11900226757369614, - "grad_norm": 0.1689453125, - "learning_rate": 0.00019444376453310488, - "loss": 0.9526, - "step": 328 - }, - { - "epoch": 0.11972789115646258, - "grad_norm": 0.18359375, - "learning_rate": 0.00019439481091665647, - "loss": 0.9875, - "step": 330 - }, - { - "epoch": 0.12045351473922902, - "grad_norm": 0.1787109375, - "learning_rate": 0.00019434585730020807, - "loss": 1.0052, - "step": 332 - }, - { - "epoch": 0.12117913832199546, - "grad_norm": 0.1943359375, - "learning_rate": 0.00019429690368375966, - "loss": 1.0672, - "step": 334 - }, - { - "epoch": 0.1219047619047619, - "grad_norm": 0.171875, - "learning_rate": 0.0001942479500673112, - "loss": 0.9938, - "step": 336 - }, - { - "epoch": 0.12263038548752835, - "grad_norm": 0.1669921875, - "learning_rate": 0.00019419899645086282, - "loss": 0.9938, - "step": 338 - }, - { - "epoch": 0.12335600907029479, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001941500428344144, - "loss": 0.9733, - "step": 340 - }, - { - "epoch": 0.12408163265306123, - "grad_norm": 0.166015625, - "learning_rate": 0.00019410108921796599, - "loss": 0.9917, - "step": 342 - }, - { - "epoch": 0.12480725623582767, - "grad_norm": 0.2041015625, - "learning_rate": 0.00019405213560151757, - "loss": 0.9594, - "step": 344 - }, - { - "epoch": 0.1255328798185941, - "grad_norm": 0.177734375, - "learning_rate": 0.00019400318198506915, - "loss": 1.0255, - "step": 346 - }, - { - "epoch": 0.12625850340136055, - "grad_norm": 0.2294921875, - "learning_rate": 0.00019395422836862073, - "loss": 1.0643, - "step": 348 - }, - { - "epoch": 0.12698412698412698, - "grad_norm": 0.1669921875, - "learning_rate": 0.00019390527475217232, - "loss": 0.9907, - "step": 350 - }, - { - "epoch": 0.12770975056689343, - "grad_norm": 0.1640625, - "learning_rate": 0.00019385632113572392, - "loss": 1.0455, - "step": 352 - }, - { - "epoch": 0.12843537414965986, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001938073675192755, - "loss": 0.9612, - "step": 354 - }, - { - "epoch": 0.1291609977324263, - "grad_norm": 0.146484375, - "learning_rate": 0.00019375841390282706, - "loss": 0.9235, - "step": 356 - }, - { - "epoch": 0.12988662131519274, - "grad_norm": 0.193359375, - "learning_rate": 0.00019370946028637867, - "loss": 1.009, - "step": 358 - }, - { - "epoch": 0.1306122448979592, - "grad_norm": 0.1728515625, - "learning_rate": 0.00019366050666993025, - "loss": 1.0014, - "step": 360 - }, - { - "epoch": 0.13133786848072562, - "grad_norm": 0.173828125, - "learning_rate": 0.00019361155305348184, - "loss": 0.9666, - "step": 362 - }, - { - "epoch": 0.13206349206349208, - "grad_norm": 0.1728515625, - "learning_rate": 0.00019356259943703342, - "loss": 1.0091, - "step": 364 - }, - { - "epoch": 0.1327891156462585, - "grad_norm": 0.166015625, - "learning_rate": 0.000193513645820585, - "loss": 1.0238, - "step": 366 - }, - { - "epoch": 0.13351473922902493, - "grad_norm": 0.1611328125, - "learning_rate": 0.00019346469220413658, - "loss": 1.0511, - "step": 368 - }, - { - "epoch": 0.13424036281179139, - "grad_norm": 0.1513671875, - "learning_rate": 0.00019341573858768817, - "loss": 0.986, - "step": 370 - }, - { - "epoch": 0.1349659863945578, - "grad_norm": 0.16796875, - "learning_rate": 0.00019336678497123977, - "loss": 1.0094, - "step": 372 - }, - { - "epoch": 0.13569160997732427, - "grad_norm": 0.15234375, - "learning_rate": 0.00019331783135479133, - "loss": 1.0596, - "step": 374 - }, - { - "epoch": 0.1364172335600907, - "grad_norm": 0.171875, - "learning_rate": 0.0001932688777383429, - "loss": 0.9845, - "step": 376 - }, - { - "epoch": 0.13714285714285715, - "grad_norm": 0.158203125, - "learning_rate": 0.00019321992412189452, - "loss": 0.9855, - "step": 378 - }, - { - "epoch": 0.13786848072562358, - "grad_norm": 0.1806640625, - "learning_rate": 0.0001931709705054461, - "loss": 1.039, - "step": 380 - }, - { - "epoch": 0.13859410430839003, - "grad_norm": 0.1728515625, - "learning_rate": 0.00019312201688899769, - "loss": 1.0108, - "step": 382 - }, - { - "epoch": 0.13931972789115646, - "grad_norm": 0.18359375, - "learning_rate": 0.00019307306327254927, - "loss": 0.977, - "step": 384 - }, - { - "epoch": 0.1400453514739229, - "grad_norm": 0.17578125, - "learning_rate": 0.00019302410965610085, - "loss": 0.9754, - "step": 386 - }, - { - "epoch": 0.14077097505668934, - "grad_norm": 0.1728515625, - "learning_rate": 0.00019297515603965243, - "loss": 1.1274, - "step": 388 - }, - { - "epoch": 0.1414965986394558, - "grad_norm": 0.1689453125, - "learning_rate": 0.00019292620242320401, - "loss": 1.0126, - "step": 390 - }, - { - "epoch": 0.14222222222222222, - "grad_norm": 0.1572265625, - "learning_rate": 0.00019287724880675562, - "loss": 1.0009, - "step": 392 - }, - { - "epoch": 0.14294784580498865, - "grad_norm": 0.1806640625, - "learning_rate": 0.00019282829519030718, - "loss": 0.9729, - "step": 394 - }, - { - "epoch": 0.1436734693877551, - "grad_norm": 0.171875, - "learning_rate": 0.00019277934157385876, - "loss": 1.0157, - "step": 396 - }, - { - "epoch": 0.14439909297052153, - "grad_norm": 0.1689453125, - "learning_rate": 0.00019273038795741037, - "loss": 0.9222, - "step": 398 - }, - { - "epoch": 0.14512471655328799, - "grad_norm": 0.1669921875, - "learning_rate": 0.00019268143434096195, - "loss": 0.9869, - "step": 400 - }, - { - "epoch": 0.14512471655328799, - "eval_loss": 1.033414363861084, - "eval_runtime": 94.5288, - "eval_samples_per_second": 19.042, - "eval_steps_per_second": 0.603, - "step": 400 - }, - { - "epoch": 0.1458503401360544, - "grad_norm": 0.173828125, - "learning_rate": 0.00019263248072451354, - "loss": 1.0265, - "step": 402 - }, - { - "epoch": 0.14657596371882087, - "grad_norm": 0.1748046875, - "learning_rate": 0.00019258352710806512, - "loss": 1.0718, - "step": 404 - }, - { - "epoch": 0.1473015873015873, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001925345734916167, - "loss": 1.0709, - "step": 406 - }, - { - "epoch": 0.14802721088435375, - "grad_norm": 0.1689453125, - "learning_rate": 0.00019248561987516828, - "loss": 1.0135, - "step": 408 - }, - { - "epoch": 0.14875283446712018, - "grad_norm": 0.169921875, - "learning_rate": 0.00019243666625871986, - "loss": 0.9483, - "step": 410 - }, - { - "epoch": 0.14947845804988663, - "grad_norm": 0.16015625, - "learning_rate": 0.00019238771264227147, - "loss": 1.011, - "step": 412 - }, - { - "epoch": 0.15020408163265306, - "grad_norm": 0.158203125, - "learning_rate": 0.00019233875902582303, - "loss": 1.0434, - "step": 414 - }, - { - "epoch": 0.1509297052154195, - "grad_norm": 0.185546875, - "learning_rate": 0.0001922898054093746, - "loss": 1.0594, - "step": 416 - }, - { - "epoch": 0.15165532879818594, - "grad_norm": 0.18359375, - "learning_rate": 0.00019224085179292622, - "loss": 0.998, - "step": 418 - }, - { - "epoch": 0.1523809523809524, - "grad_norm": 0.173828125, - "learning_rate": 0.0001921918981764778, - "loss": 1.0772, - "step": 420 - }, - { - "epoch": 0.15310657596371882, - "grad_norm": 0.1435546875, - "learning_rate": 0.00019214294456002939, - "loss": 0.9089, - "step": 422 - }, - { - "epoch": 0.15383219954648525, - "grad_norm": 0.203125, - "learning_rate": 0.00019209399094358097, - "loss": 1.0361, - "step": 424 - }, - { - "epoch": 0.1545578231292517, - "grad_norm": 0.1865234375, - "learning_rate": 0.00019204503732713255, - "loss": 1.0325, - "step": 426 - }, - { - "epoch": 0.15528344671201813, - "grad_norm": 0.1552734375, - "learning_rate": 0.00019199608371068413, - "loss": 1.0237, - "step": 428 - }, - { - "epoch": 0.15600907029478459, - "grad_norm": 0.166015625, - "learning_rate": 0.00019194713009423571, - "loss": 0.9513, - "step": 430 - }, - { - "epoch": 0.156734693877551, - "grad_norm": 0.15625, - "learning_rate": 0.0001918981764777873, - "loss": 0.9614, - "step": 432 - }, - { - "epoch": 0.15746031746031747, - "grad_norm": 0.1552734375, - "learning_rate": 0.00019184922286133888, - "loss": 0.9507, - "step": 434 - }, - { - "epoch": 0.1581859410430839, - "grad_norm": 0.1533203125, - "learning_rate": 0.00019180026924489046, - "loss": 0.9842, - "step": 436 - }, - { - "epoch": 0.15891156462585035, - "grad_norm": 0.146484375, - "learning_rate": 0.00019175131562844207, - "loss": 1.0938, - "step": 438 - }, - { - "epoch": 0.15963718820861678, - "grad_norm": 0.171875, - "learning_rate": 0.00019170236201199365, - "loss": 0.9905, - "step": 440 - }, - { - "epoch": 0.16036281179138323, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001916534083955452, - "loss": 1.049, - "step": 442 - }, - { - "epoch": 0.16108843537414966, - "grad_norm": 0.15234375, - "learning_rate": 0.00019160445477909682, - "loss": 0.9727, - "step": 444 - }, - { - "epoch": 0.1618140589569161, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001915555011626484, - "loss": 1.0255, - "step": 446 - }, - { - "epoch": 0.16253968253968254, - "grad_norm": 0.2197265625, - "learning_rate": 0.00019150654754619998, - "loss": 0.9454, - "step": 448 - }, - { - "epoch": 0.16326530612244897, - "grad_norm": 0.1806640625, - "learning_rate": 0.00019145759392975156, - "loss": 1.0037, - "step": 450 - }, - { - "epoch": 0.16399092970521542, - "grad_norm": 0.1533203125, - "learning_rate": 0.00019140864031330315, - "loss": 0.9791, - "step": 452 - }, - { - "epoch": 0.16471655328798185, - "grad_norm": 0.1455078125, - "learning_rate": 0.00019135968669685473, - "loss": 1.0033, - "step": 454 - }, - { - "epoch": 0.1654421768707483, - "grad_norm": 0.1689453125, - "learning_rate": 0.0001913107330804063, - "loss": 1.0124, - "step": 456 - }, - { - "epoch": 0.16616780045351473, - "grad_norm": 0.173828125, - "learning_rate": 0.00019126177946395792, - "loss": 1.0341, - "step": 458 - }, - { - "epoch": 0.16689342403628118, - "grad_norm": 0.15234375, - "learning_rate": 0.0001912128258475095, - "loss": 0.999, - "step": 460 - }, - { - "epoch": 0.1676190476190476, - "grad_norm": 0.166015625, - "learning_rate": 0.00019116387223106106, - "loss": 1.0343, - "step": 462 - }, - { - "epoch": 0.16834467120181407, - "grad_norm": 0.1767578125, - "learning_rate": 0.00019111491861461267, - "loss": 1.1115, - "step": 464 - }, - { - "epoch": 0.1690702947845805, - "grad_norm": 0.1640625, - "learning_rate": 0.00019106596499816425, - "loss": 1.0324, - "step": 466 - }, - { - "epoch": 0.16979591836734695, - "grad_norm": 0.171875, - "learning_rate": 0.00019101701138171583, - "loss": 1.0122, - "step": 468 - }, - { - "epoch": 0.17052154195011338, - "grad_norm": 0.18359375, - "learning_rate": 0.00019096805776526741, - "loss": 1.075, - "step": 470 - }, - { - "epoch": 0.17124716553287983, - "grad_norm": 0.1748046875, - "learning_rate": 0.000190919104148819, - "loss": 1.117, - "step": 472 - }, - { - "epoch": 0.17197278911564626, - "grad_norm": 0.1826171875, - "learning_rate": 0.00019087015053237058, - "loss": 0.9855, - "step": 474 - }, - { - "epoch": 0.1726984126984127, - "grad_norm": 0.1533203125, - "learning_rate": 0.00019082119691592216, - "loss": 0.9953, - "step": 476 - }, - { - "epoch": 0.17342403628117914, - "grad_norm": 0.1953125, - "learning_rate": 0.00019077224329947377, - "loss": 1.0023, - "step": 478 - }, - { - "epoch": 0.17414965986394557, - "grad_norm": 0.16796875, - "learning_rate": 0.00019072328968302535, - "loss": 1.0291, - "step": 480 - }, - { - "epoch": 0.17487528344671202, - "grad_norm": 0.146484375, - "learning_rate": 0.0001906743360665769, - "loss": 0.9518, - "step": 482 - }, - { - "epoch": 0.17560090702947845, - "grad_norm": 0.162109375, - "learning_rate": 0.00019062538245012852, - "loss": 1.0413, - "step": 484 - }, - { - "epoch": 0.1763265306122449, - "grad_norm": 0.1484375, - "learning_rate": 0.0001905764288336801, - "loss": 1.0636, - "step": 486 - }, - { - "epoch": 0.17705215419501133, - "grad_norm": 0.1630859375, - "learning_rate": 0.00019052747521723168, - "loss": 1.041, - "step": 488 - }, - { - "epoch": 0.17777777777777778, - "grad_norm": 0.1689453125, - "learning_rate": 0.00019047852160078326, - "loss": 1.0237, - "step": 490 - }, - { - "epoch": 0.1785034013605442, - "grad_norm": 0.1630859375, - "learning_rate": 0.00019042956798433485, - "loss": 0.9516, - "step": 492 - }, - { - "epoch": 0.17922902494331067, - "grad_norm": 0.1708984375, - "learning_rate": 0.00019038061436788643, - "loss": 1.0947, - "step": 494 - }, - { - "epoch": 0.1799546485260771, - "grad_norm": 0.1708984375, - "learning_rate": 0.000190331660751438, - "loss": 1.1305, - "step": 496 - }, - { - "epoch": 0.18068027210884355, - "grad_norm": 0.140625, - "learning_rate": 0.00019028270713498962, - "loss": 1.0146, - "step": 498 - }, - { - "epoch": 0.18140589569160998, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001902337535185412, - "loss": 0.9402, - "step": 500 - }, - { - "epoch": 0.18140589569160998, - "eval_loss": 1.030946969985962, - "eval_runtime": 94.9838, - "eval_samples_per_second": 18.951, - "eval_steps_per_second": 0.6, - "step": 500 - }, - { - "epoch": 0.18213151927437643, - "grad_norm": 0.177734375, - "learning_rate": 0.00019018479990209276, - "loss": 0.9476, - "step": 502 - }, - { - "epoch": 0.18285714285714286, - "grad_norm": 0.1669921875, - "learning_rate": 0.00019013584628564437, - "loss": 1.1158, - "step": 504 - }, - { - "epoch": 0.18358276643990928, - "grad_norm": 0.1640625, - "learning_rate": 0.00019008689266919595, - "loss": 1.0539, - "step": 506 - }, - { - "epoch": 0.18430839002267574, - "grad_norm": 0.1572265625, - "learning_rate": 0.00019003793905274753, - "loss": 1.0797, - "step": 508 - }, - { - "epoch": 0.18503401360544217, - "grad_norm": 0.1533203125, - "learning_rate": 0.00018998898543629911, - "loss": 0.9487, - "step": 510 - }, - { - "epoch": 0.18575963718820862, - "grad_norm": 0.14453125, - "learning_rate": 0.0001899400318198507, - "loss": 1.0215, - "step": 512 - }, - { - "epoch": 0.18648526077097505, - "grad_norm": 0.1708984375, - "learning_rate": 0.00018989107820340228, - "loss": 1.0063, - "step": 514 - }, - { - "epoch": 0.1872108843537415, - "grad_norm": 0.1748046875, - "learning_rate": 0.00018984212458695386, - "loss": 1.108, - "step": 516 - }, - { - "epoch": 0.18793650793650793, - "grad_norm": 0.1552734375, - "learning_rate": 0.00018979317097050547, - "loss": 1.0557, - "step": 518 - }, - { - "epoch": 0.18866213151927438, - "grad_norm": 0.169921875, - "learning_rate": 0.00018974421735405703, - "loss": 0.977, - "step": 520 - }, - { - "epoch": 0.1893877551020408, - "grad_norm": 0.1640625, - "learning_rate": 0.0001896952637376086, - "loss": 1.0308, - "step": 522 - }, - { - "epoch": 0.19011337868480727, - "grad_norm": 0.1767578125, - "learning_rate": 0.00018964631012116022, - "loss": 1.0537, - "step": 524 - }, - { - "epoch": 0.1908390022675737, - "grad_norm": 0.189453125, - "learning_rate": 0.0001895973565047118, - "loss": 1.0358, - "step": 526 - }, - { - "epoch": 0.19156462585034015, - "grad_norm": 0.1435546875, - "learning_rate": 0.00018954840288826338, - "loss": 1.0018, - "step": 528 - }, - { - "epoch": 0.19229024943310657, - "grad_norm": 0.1787109375, - "learning_rate": 0.00018949944927181496, - "loss": 1.0255, - "step": 530 - }, - { - "epoch": 0.19301587301587303, - "grad_norm": 0.140625, - "learning_rate": 0.00018945049565536655, - "loss": 0.9378, - "step": 532 - }, - { - "epoch": 0.19374149659863946, - "grad_norm": 0.1728515625, - "learning_rate": 0.00018940154203891813, - "loss": 1.0373, - "step": 534 - }, - { - "epoch": 0.19446712018140588, - "grad_norm": 0.1640625, - "learning_rate": 0.0001893525884224697, - "loss": 1.0026, - "step": 536 - }, - { - "epoch": 0.19519274376417234, - "grad_norm": 0.1826171875, - "learning_rate": 0.00018930363480602132, - "loss": 1.0897, - "step": 538 - }, - { - "epoch": 0.19591836734693877, - "grad_norm": 0.15234375, - "learning_rate": 0.00018925468118957288, - "loss": 0.9296, - "step": 540 - }, - { - "epoch": 0.19664399092970522, - "grad_norm": 0.1806640625, - "learning_rate": 0.00018920572757312446, - "loss": 1.0485, - "step": 542 - }, - { - "epoch": 0.19736961451247165, - "grad_norm": 0.1875, - "learning_rate": 0.00018915677395667607, - "loss": 0.9125, - "step": 544 - }, - { - "epoch": 0.1980952380952381, - "grad_norm": 0.150390625, - "learning_rate": 0.00018910782034022765, - "loss": 1.0227, - "step": 546 - }, - { - "epoch": 0.19882086167800453, - "grad_norm": 0.158203125, - "learning_rate": 0.00018905886672377923, - "loss": 0.9335, - "step": 548 - }, - { - "epoch": 0.19954648526077098, - "grad_norm": 0.1787109375, - "learning_rate": 0.00018900991310733081, - "loss": 1.0549, - "step": 550 - }, - { - "epoch": 0.2002721088435374, - "grad_norm": 0.1962890625, - "learning_rate": 0.0001889609594908824, - "loss": 0.9836, - "step": 552 - }, - { - "epoch": 0.20099773242630387, - "grad_norm": 0.15234375, - "learning_rate": 0.00018891200587443398, - "loss": 1.0058, - "step": 554 - }, - { - "epoch": 0.2017233560090703, - "grad_norm": 0.154296875, - "learning_rate": 0.00018886305225798556, - "loss": 1.0077, - "step": 556 - }, - { - "epoch": 0.20244897959183675, - "grad_norm": 0.150390625, - "learning_rate": 0.00018881409864153717, - "loss": 0.9855, - "step": 558 - }, - { - "epoch": 0.20317460317460317, - "grad_norm": 0.1611328125, - "learning_rate": 0.00018876514502508873, - "loss": 0.9711, - "step": 560 - }, - { - "epoch": 0.2039002267573696, - "grad_norm": 0.142578125, - "learning_rate": 0.0001887161914086403, - "loss": 0.9793, - "step": 562 - }, - { - "epoch": 0.20462585034013606, - "grad_norm": 0.171875, - "learning_rate": 0.00018866723779219192, - "loss": 1.0408, - "step": 564 - }, - { - "epoch": 0.20535147392290248, - "grad_norm": 0.1787109375, - "learning_rate": 0.0001886182841757435, - "loss": 1.0226, - "step": 566 - }, - { - "epoch": 0.20607709750566894, - "grad_norm": 0.169921875, - "learning_rate": 0.00018856933055929508, - "loss": 0.9565, - "step": 568 - }, - { - "epoch": 0.20680272108843537, - "grad_norm": 0.1611328125, - "learning_rate": 0.00018852037694284666, - "loss": 1.0553, - "step": 570 - }, - { - "epoch": 0.20752834467120182, - "grad_norm": 0.162109375, - "learning_rate": 0.00018847142332639825, - "loss": 1.0416, - "step": 572 - }, - { - "epoch": 0.20825396825396825, - "grad_norm": 0.181640625, - "learning_rate": 0.00018842246970994983, - "loss": 1.0078, - "step": 574 - }, - { - "epoch": 0.2089795918367347, - "grad_norm": 0.154296875, - "learning_rate": 0.0001883735160935014, - "loss": 0.9935, - "step": 576 - }, - { - "epoch": 0.20970521541950113, - "grad_norm": 0.1708984375, - "learning_rate": 0.000188324562477053, - "loss": 1.0732, - "step": 578 - }, - { - "epoch": 0.21043083900226758, - "grad_norm": 0.1728515625, - "learning_rate": 0.00018827560886060458, - "loss": 1.013, - "step": 580 - }, - { - "epoch": 0.211156462585034, - "grad_norm": 0.1494140625, - "learning_rate": 0.00018822665524415616, - "loss": 1.0011, - "step": 582 - }, - { - "epoch": 0.21188208616780047, - "grad_norm": 0.1435546875, - "learning_rate": 0.00018817770162770777, - "loss": 1.0494, - "step": 584 - }, - { - "epoch": 0.2126077097505669, - "grad_norm": 0.1513671875, - "learning_rate": 0.00018812874801125935, - "loss": 0.9439, - "step": 586 - }, - { - "epoch": 0.21333333333333335, - "grad_norm": 0.169921875, - "learning_rate": 0.0001880797943948109, - "loss": 1.0595, - "step": 588 - }, - { - "epoch": 0.21405895691609977, - "grad_norm": 0.171875, - "learning_rate": 0.00018803084077836251, - "loss": 0.9564, - "step": 590 - }, - { - "epoch": 0.2147845804988662, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001879818871619141, - "loss": 1.1101, - "step": 592 - }, - { - "epoch": 0.21551020408163266, - "grad_norm": 0.1513671875, - "learning_rate": 0.00018793293354546568, - "loss": 0.9792, - "step": 594 - }, - { - "epoch": 0.21623582766439908, - "grad_norm": 0.1396484375, - "learning_rate": 0.00018788397992901726, - "loss": 0.9356, - "step": 596 - }, - { - "epoch": 0.21696145124716554, - "grad_norm": 0.1318359375, - "learning_rate": 0.00018783502631256884, - "loss": 0.9553, - "step": 598 - }, - { - "epoch": 0.21768707482993196, - "grad_norm": 0.1650390625, - "learning_rate": 0.00018778607269612043, - "loss": 0.9454, - "step": 600 - }, - { - "epoch": 0.21768707482993196, - "eval_loss": 1.0292859077453613, - "eval_runtime": 95.3967, - "eval_samples_per_second": 18.869, - "eval_steps_per_second": 0.598, - "step": 600 - }, - { - "epoch": 0.21841269841269842, - "grad_norm": 0.1572265625, - "learning_rate": 0.000187737119079672, - "loss": 1.0517, - "step": 602 - }, - { - "epoch": 0.21913832199546485, - "grad_norm": 0.1484375, - "learning_rate": 0.00018768816546322362, - "loss": 1.0417, - "step": 604 - }, - { - "epoch": 0.2198639455782313, - "grad_norm": 0.1484375, - "learning_rate": 0.0001876392118467752, - "loss": 0.9824, - "step": 606 - }, - { - "epoch": 0.22058956916099773, - "grad_norm": 0.158203125, - "learning_rate": 0.00018759025823032675, - "loss": 1.0043, - "step": 608 - }, - { - "epoch": 0.22131519274376418, - "grad_norm": 0.1611328125, - "learning_rate": 0.00018754130461387836, - "loss": 0.94, - "step": 610 - }, - { - "epoch": 0.2220408163265306, - "grad_norm": 0.185546875, - "learning_rate": 0.00018749235099742995, - "loss": 0.9731, - "step": 612 - }, - { - "epoch": 0.22276643990929706, - "grad_norm": 0.162109375, - "learning_rate": 0.00018744339738098153, - "loss": 1.0113, - "step": 614 - }, - { - "epoch": 0.2234920634920635, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001873944437645331, - "loss": 1.0831, - "step": 616 - }, - { - "epoch": 0.22421768707482992, - "grad_norm": 0.1630859375, - "learning_rate": 0.0001873454901480847, - "loss": 1.0206, - "step": 618 - }, - { - "epoch": 0.22494331065759637, - "grad_norm": 0.15625, - "learning_rate": 0.00018729653653163628, - "loss": 1.0225, - "step": 620 - }, - { - "epoch": 0.2256689342403628, - "grad_norm": 0.16015625, - "learning_rate": 0.00018724758291518786, - "loss": 0.9866, - "step": 622 - }, - { - "epoch": 0.22639455782312926, - "grad_norm": 0.154296875, - "learning_rate": 0.00018719862929873947, - "loss": 1.0209, - "step": 624 - }, - { - "epoch": 0.22712018140589568, - "grad_norm": 0.173828125, - "learning_rate": 0.00018714967568229105, - "loss": 0.9636, - "step": 626 - }, - { - "epoch": 0.22784580498866214, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001871007220658426, - "loss": 0.9683, - "step": 628 - }, - { - "epoch": 0.22857142857142856, - "grad_norm": 0.2099609375, - "learning_rate": 0.00018705176844939421, - "loss": 0.9994, - "step": 630 - }, - { - "epoch": 0.22929705215419502, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001870028148329458, - "loss": 1.0061, - "step": 632 - }, - { - "epoch": 0.23002267573696145, - "grad_norm": 0.1591796875, - "learning_rate": 0.00018695386121649738, - "loss": 0.9647, - "step": 634 - }, - { - "epoch": 0.2307482993197279, - "grad_norm": 0.181640625, - "learning_rate": 0.00018690490760004896, - "loss": 1.0194, - "step": 636 - }, - { - "epoch": 0.23147392290249433, - "grad_norm": 0.171875, - "learning_rate": 0.00018685595398360054, - "loss": 1.0266, - "step": 638 - }, - { - "epoch": 0.23219954648526078, - "grad_norm": 0.16015625, - "learning_rate": 0.00018680700036715213, - "loss": 1.011, - "step": 640 - }, - { - "epoch": 0.2329251700680272, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001867580467507037, - "loss": 0.9558, - "step": 642 - }, - { - "epoch": 0.23365079365079366, - "grad_norm": 0.15625, - "learning_rate": 0.00018670909313425532, - "loss": 1.0053, - "step": 644 - }, - { - "epoch": 0.2343764172335601, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001866601395178069, - "loss": 1.054, - "step": 646 - }, - { - "epoch": 0.23510204081632652, - "grad_norm": 0.158203125, - "learning_rate": 0.00018661118590135845, - "loss": 0.9628, - "step": 648 - }, - { - "epoch": 0.23582766439909297, - "grad_norm": 0.1435546875, - "learning_rate": 0.00018656223228491006, - "loss": 1.0158, - "step": 650 - }, - { - "epoch": 0.2365532879818594, - "grad_norm": 0.15234375, - "learning_rate": 0.00018651327866846165, - "loss": 0.9922, - "step": 652 - }, - { - "epoch": 0.23727891156462586, - "grad_norm": 0.154296875, - "learning_rate": 0.00018646432505201323, - "loss": 0.9373, - "step": 654 - }, - { - "epoch": 0.23800453514739228, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001864153714355648, - "loss": 0.993, - "step": 656 - }, - { - "epoch": 0.23873015873015874, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001863664178191164, - "loss": 1.0862, - "step": 658 - }, - { - "epoch": 0.23945578231292516, - "grad_norm": 0.1611328125, - "learning_rate": 0.00018631746420266798, - "loss": 0.9666, - "step": 660 - }, - { - "epoch": 0.24018140589569162, - "grad_norm": 0.1484375, - "learning_rate": 0.00018626851058621956, - "loss": 0.9805, - "step": 662 - }, - { - "epoch": 0.24090702947845805, - "grad_norm": 0.15625, - "learning_rate": 0.00018621955696977117, - "loss": 0.9948, - "step": 664 - }, - { - "epoch": 0.2416326530612245, - "grad_norm": 0.1435546875, - "learning_rate": 0.00018617060335332272, - "loss": 1.0119, - "step": 666 - }, - { - "epoch": 0.24235827664399093, - "grad_norm": 0.154296875, - "learning_rate": 0.0001861216497368743, - "loss": 1.041, - "step": 668 - }, - { - "epoch": 0.24308390022675738, - "grad_norm": 0.1728515625, - "learning_rate": 0.00018607269612042591, - "loss": 0.9779, - "step": 670 - }, - { - "epoch": 0.2438095238095238, - "grad_norm": 0.16796875, - "learning_rate": 0.0001860237425039775, - "loss": 1.0314, - "step": 672 - }, - { - "epoch": 0.24453514739229024, - "grad_norm": 0.1650390625, - "learning_rate": 0.00018597478888752908, - "loss": 1.0456, - "step": 674 - }, - { - "epoch": 0.2452607709750567, - "grad_norm": 0.1396484375, - "learning_rate": 0.00018592583527108066, - "loss": 1.0908, - "step": 676 - }, - { - "epoch": 0.24598639455782312, - "grad_norm": 0.2216796875, - "learning_rate": 0.00018587688165463224, - "loss": 0.9864, - "step": 678 - }, - { - "epoch": 0.24671201814058957, - "grad_norm": 0.25390625, - "learning_rate": 0.00018582792803818383, - "loss": 1.0069, - "step": 680 - }, - { - "epoch": 0.247437641723356, - "grad_norm": 0.1787109375, - "learning_rate": 0.0001857789744217354, - "loss": 1.0085, - "step": 682 - }, - { - "epoch": 0.24816326530612245, - "grad_norm": 0.158203125, - "learning_rate": 0.00018573002080528702, - "loss": 0.9846, - "step": 684 - }, - { - "epoch": 0.24888888888888888, - "grad_norm": 0.1845703125, - "learning_rate": 0.00018568106718883857, - "loss": 1.0301, - "step": 686 - }, - { - "epoch": 0.24961451247165534, - "grad_norm": 0.1865234375, - "learning_rate": 0.00018563211357239015, - "loss": 1.031, - "step": 688 - }, - { - "epoch": 0.2503401360544218, - "grad_norm": 0.1669921875, - "learning_rate": 0.00018558315995594176, - "loss": 1.043, - "step": 690 - }, - { - "epoch": 0.2510657596371882, - "grad_norm": 0.15625, - "learning_rate": 0.00018553420633949335, - "loss": 0.9125, - "step": 692 - }, - { - "epoch": 0.25179138321995465, - "grad_norm": 0.158203125, - "learning_rate": 0.00018548525272304493, - "loss": 1.0135, - "step": 694 - }, - { - "epoch": 0.2525170068027211, - "grad_norm": 0.15625, - "learning_rate": 0.0001854362991065965, - "loss": 0.9936, - "step": 696 - }, - { - "epoch": 0.25324263038548755, - "grad_norm": 0.142578125, - "learning_rate": 0.0001853873454901481, - "loss": 0.9794, - "step": 698 - }, - { - "epoch": 0.25396825396825395, - "grad_norm": 0.1640625, - "learning_rate": 0.00018533839187369968, - "loss": 1.0237, - "step": 700 - }, - { - "epoch": 0.25396825396825395, - "eval_loss": 1.0280758142471313, - "eval_runtime": 92.7311, - "eval_samples_per_second": 19.411, - "eval_steps_per_second": 0.615, - "step": 700 - }, - { - "epoch": 0.2546938775510204, - "grad_norm": 0.150390625, - "learning_rate": 0.00018528943825725126, - "loss": 1.069, - "step": 702 - }, - { - "epoch": 0.25541950113378686, - "grad_norm": 0.1591796875, - "learning_rate": 0.00018524048464080287, - "loss": 0.902, - "step": 704 - }, - { - "epoch": 0.25614512471655326, - "grad_norm": 0.171875, - "learning_rate": 0.00018519153102435442, - "loss": 1.0793, - "step": 706 - }, - { - "epoch": 0.2568707482993197, - "grad_norm": 0.1572265625, - "learning_rate": 0.000185142577407906, - "loss": 1.0871, - "step": 708 - }, - { - "epoch": 0.2575963718820862, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001850936237914576, - "loss": 0.9444, - "step": 710 - }, - { - "epoch": 0.2583219954648526, - "grad_norm": 0.169921875, - "learning_rate": 0.0001850446701750092, - "loss": 1.045, - "step": 712 - }, - { - "epoch": 0.259047619047619, - "grad_norm": 0.1513671875, - "learning_rate": 0.00018499571655856078, - "loss": 1.0274, - "step": 714 - }, - { - "epoch": 0.2597732426303855, - "grad_norm": 0.1611328125, - "learning_rate": 0.00018494676294211236, - "loss": 1.0513, - "step": 716 - }, - { - "epoch": 0.26049886621315194, - "grad_norm": 0.16015625, - "learning_rate": 0.00018489780932566394, - "loss": 0.9794, - "step": 718 - }, - { - "epoch": 0.2612244897959184, - "grad_norm": 0.1533203125, - "learning_rate": 0.00018484885570921552, - "loss": 0.9927, - "step": 720 - }, - { - "epoch": 0.2619501133786848, - "grad_norm": 0.162109375, - "learning_rate": 0.0001847999020927671, - "loss": 0.9424, - "step": 722 - }, - { - "epoch": 0.26267573696145124, - "grad_norm": 0.1689453125, - "learning_rate": 0.0001847509484763187, - "loss": 0.9492, - "step": 724 - }, - { - "epoch": 0.2634013605442177, - "grad_norm": 0.1708984375, - "learning_rate": 0.00018470199485987027, - "loss": 1.0054, - "step": 726 - }, - { - "epoch": 0.26412698412698415, - "grad_norm": 0.287109375, - "learning_rate": 0.00018465304124342185, - "loss": 1.0401, - "step": 728 - }, - { - "epoch": 0.26485260770975055, - "grad_norm": 0.140625, - "learning_rate": 0.00018460408762697346, - "loss": 0.9785, - "step": 730 - }, - { - "epoch": 0.265578231292517, - "grad_norm": 0.169921875, - "learning_rate": 0.00018455513401052505, - "loss": 1.0867, - "step": 732 - }, - { - "epoch": 0.26630385487528346, - "grad_norm": 0.2001953125, - "learning_rate": 0.0001845061803940766, - "loss": 0.9375, - "step": 734 - }, - { - "epoch": 0.26702947845804986, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001844572267776282, - "loss": 0.9576, - "step": 736 - }, - { - "epoch": 0.2677551020408163, - "grad_norm": 0.1630859375, - "learning_rate": 0.0001844082731611798, - "loss": 1.0314, - "step": 738 - }, - { - "epoch": 0.26848072562358277, - "grad_norm": 0.142578125, - "learning_rate": 0.00018435931954473137, - "loss": 0.9826, - "step": 740 - }, - { - "epoch": 0.2692063492063492, - "grad_norm": 0.201171875, - "learning_rate": 0.00018431036592828296, - "loss": 1.0339, - "step": 742 - }, - { - "epoch": 0.2699319727891156, - "grad_norm": 0.13671875, - "learning_rate": 0.00018426141231183454, - "loss": 1.0327, - "step": 744 - }, - { - "epoch": 0.2706575963718821, - "grad_norm": 0.150390625, - "learning_rate": 0.00018421245869538612, - "loss": 1.029, - "step": 746 - }, - { - "epoch": 0.27138321995464854, - "grad_norm": 0.15625, - "learning_rate": 0.0001841635050789377, - "loss": 1.0364, - "step": 748 - }, - { - "epoch": 0.272108843537415, - "grad_norm": 0.166015625, - "learning_rate": 0.0001841145514624893, - "loss": 0.925, - "step": 750 - }, - { - "epoch": 0.2728344671201814, - "grad_norm": 0.16796875, - "learning_rate": 0.0001840655978460409, - "loss": 0.9998, - "step": 752 - }, - { - "epoch": 0.27356009070294784, - "grad_norm": 0.1826171875, - "learning_rate": 0.00018401664422959245, - "loss": 1.0414, - "step": 754 - }, - { - "epoch": 0.2742857142857143, - "grad_norm": 0.1494140625, - "learning_rate": 0.00018396769061314406, - "loss": 1.0585, - "step": 756 - }, - { - "epoch": 0.2750113378684807, - "grad_norm": 0.1533203125, - "learning_rate": 0.00018391873699669564, - "loss": 0.945, - "step": 758 - }, - { - "epoch": 0.27573696145124715, - "grad_norm": 0.1826171875, - "learning_rate": 0.00018386978338024722, - "loss": 1.0171, - "step": 760 - }, - { - "epoch": 0.2764625850340136, - "grad_norm": 0.154296875, - "learning_rate": 0.0001838208297637988, - "loss": 1.0176, - "step": 762 - }, - { - "epoch": 0.27718820861678006, - "grad_norm": 0.1484375, - "learning_rate": 0.0001837718761473504, - "loss": 0.9225, - "step": 764 - }, - { - "epoch": 0.27791383219954646, - "grad_norm": 0.146484375, - "learning_rate": 0.00018372292253090197, - "loss": 1.0185, - "step": 766 - }, - { - "epoch": 0.2786394557823129, - "grad_norm": 0.150390625, - "learning_rate": 0.00018367396891445355, - "loss": 1.0516, - "step": 768 - }, - { - "epoch": 0.27936507936507937, - "grad_norm": 0.1474609375, - "learning_rate": 0.00018362501529800516, - "loss": 0.972, - "step": 770 - }, - { - "epoch": 0.2800907029478458, - "grad_norm": 0.1552734375, - "learning_rate": 0.00018357606168155675, - "loss": 1.0903, - "step": 772 - }, - { - "epoch": 0.2808163265306122, - "grad_norm": 0.1484375, - "learning_rate": 0.0001835271080651083, - "loss": 1.0122, - "step": 774 - }, - { - "epoch": 0.2815419501133787, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001834781544486599, - "loss": 1.0724, - "step": 776 - }, - { - "epoch": 0.28226757369614514, - "grad_norm": 0.1640625, - "learning_rate": 0.0001834292008322115, - "loss": 1.02, - "step": 778 - }, - { - "epoch": 0.2829931972789116, - "grad_norm": 0.1357421875, - "learning_rate": 0.00018338024721576307, - "loss": 0.9654, - "step": 780 - }, - { - "epoch": 0.283718820861678, - "grad_norm": 0.173828125, - "learning_rate": 0.00018333129359931466, - "loss": 1.0622, - "step": 782 - }, - { - "epoch": 0.28444444444444444, - "grad_norm": 0.16015625, - "learning_rate": 0.00018328233998286624, - "loss": 0.9866, - "step": 784 - }, - { - "epoch": 0.2851700680272109, - "grad_norm": 0.1611328125, - "learning_rate": 0.00018323338636641782, - "loss": 1.048, - "step": 786 - }, - { - "epoch": 0.2858956916099773, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001831844327499694, - "loss": 0.9802, - "step": 788 - }, - { - "epoch": 0.28662131519274375, - "grad_norm": 0.17578125, - "learning_rate": 0.000183135479133521, - "loss": 0.9756, - "step": 790 - }, - { - "epoch": 0.2873469387755102, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001830865255170726, - "loss": 1.0143, - "step": 792 - }, - { - "epoch": 0.28807256235827666, - "grad_norm": 0.1787109375, - "learning_rate": 0.00018303757190062415, - "loss": 1.0164, - "step": 794 - }, - { - "epoch": 0.28879818594104306, - "grad_norm": 0.169921875, - "learning_rate": 0.00018298861828417576, - "loss": 0.9813, - "step": 796 - }, - { - "epoch": 0.2895238095238095, - "grad_norm": 0.166015625, - "learning_rate": 0.00018293966466772734, - "loss": 1.0568, - "step": 798 - }, - { - "epoch": 0.29024943310657597, - "grad_norm": 0.1474609375, - "learning_rate": 0.00018289071105127892, - "loss": 0.9557, - "step": 800 - }, - { - "epoch": 0.29024943310657597, - "eval_loss": 1.0269455909729004, - "eval_runtime": 94.2392, - "eval_samples_per_second": 19.1, - "eval_steps_per_second": 0.605, - "step": 800 - }, - { - "epoch": 0.2909750566893424, - "grad_norm": 0.162109375, - "learning_rate": 0.0001828417574348305, - "loss": 1.0491, - "step": 802 - }, - { - "epoch": 0.2917006802721088, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001827928038183821, - "loss": 0.9705, - "step": 804 - }, - { - "epoch": 0.2924263038548753, - "grad_norm": 0.15234375, - "learning_rate": 0.00018274385020193367, - "loss": 1.0566, - "step": 806 - }, - { - "epoch": 0.29315192743764173, - "grad_norm": 0.1611328125, - "learning_rate": 0.00018269489658548525, - "loss": 0.9782, - "step": 808 - }, - { - "epoch": 0.2938775510204082, - "grad_norm": 0.154296875, - "learning_rate": 0.00018264594296903686, - "loss": 0.9967, - "step": 810 - }, - { - "epoch": 0.2946031746031746, - "grad_norm": 0.1630859375, - "learning_rate": 0.00018259698935258842, - "loss": 1.0371, - "step": 812 - }, - { - "epoch": 0.29532879818594104, - "grad_norm": 0.158203125, - "learning_rate": 0.00018254803573614, - "loss": 0.9803, - "step": 814 - }, - { - "epoch": 0.2960544217687075, - "grad_norm": 0.158203125, - "learning_rate": 0.0001824990821196916, - "loss": 0.9969, - "step": 816 - }, - { - "epoch": 0.2967800453514739, - "grad_norm": 0.154296875, - "learning_rate": 0.0001824501285032432, - "loss": 1.0483, - "step": 818 - }, - { - "epoch": 0.29750566893424035, - "grad_norm": 0.1669921875, - "learning_rate": 0.00018240117488679477, - "loss": 1.0164, - "step": 820 - }, - { - "epoch": 0.2982312925170068, - "grad_norm": 0.1630859375, - "learning_rate": 0.00018235222127034636, - "loss": 0.9407, - "step": 822 - }, - { - "epoch": 0.29895691609977326, - "grad_norm": 0.158203125, - "learning_rate": 0.00018230326765389794, - "loss": 1.0397, - "step": 824 - }, - { - "epoch": 0.29968253968253966, - "grad_norm": 0.138671875, - "learning_rate": 0.00018225431403744952, - "loss": 0.9518, - "step": 826 - }, - { - "epoch": 0.3004081632653061, - "grad_norm": 0.154296875, - "learning_rate": 0.0001822053604210011, - "loss": 1.0551, - "step": 828 - }, - { - "epoch": 0.30113378684807257, - "grad_norm": 0.16796875, - "learning_rate": 0.0001821564068045527, - "loss": 0.9826, - "step": 830 - }, - { - "epoch": 0.301859410430839, - "grad_norm": 0.14453125, - "learning_rate": 0.00018210745318810427, - "loss": 0.945, - "step": 832 - }, - { - "epoch": 0.3025850340136054, - "grad_norm": 0.1875, - "learning_rate": 0.00018205849957165585, - "loss": 0.9464, - "step": 834 - }, - { - "epoch": 0.3033106575963719, - "grad_norm": 0.1611328125, - "learning_rate": 0.00018200954595520746, - "loss": 0.982, - "step": 836 - }, - { - "epoch": 0.30403628117913833, - "grad_norm": 0.173828125, - "learning_rate": 0.00018196059233875904, - "loss": 1.0529, - "step": 838 - }, - { - "epoch": 0.3047619047619048, - "grad_norm": 0.162109375, - "learning_rate": 0.00018191163872231062, - "loss": 0.9982, - "step": 840 - }, - { - "epoch": 0.3054875283446712, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001818626851058622, - "loss": 0.946, - "step": 842 - }, - { - "epoch": 0.30621315192743764, - "grad_norm": 0.1396484375, - "learning_rate": 0.0001818137314894138, - "loss": 1.0053, - "step": 844 - }, - { - "epoch": 0.3069387755102041, - "grad_norm": 0.1669921875, - "learning_rate": 0.00018176477787296537, - "loss": 0.9307, - "step": 846 - }, - { - "epoch": 0.3076643990929705, - "grad_norm": 0.162109375, - "learning_rate": 0.00018171582425651695, - "loss": 0.9697, - "step": 848 - }, - { - "epoch": 0.30839002267573695, - "grad_norm": 0.146484375, - "learning_rate": 0.00018166687064006856, - "loss": 0.9617, - "step": 850 - }, - { - "epoch": 0.3091156462585034, - "grad_norm": 0.15234375, - "learning_rate": 0.00018161791702362012, - "loss": 0.9534, - "step": 852 - }, - { - "epoch": 0.30984126984126986, - "grad_norm": 0.15234375, - "learning_rate": 0.0001815689634071717, - "loss": 1.0296, - "step": 854 - }, - { - "epoch": 0.31056689342403626, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001815200097907233, - "loss": 1.0481, - "step": 856 - }, - { - "epoch": 0.3112925170068027, - "grad_norm": 0.154296875, - "learning_rate": 0.0001814710561742749, - "loss": 1.0118, - "step": 858 - }, - { - "epoch": 0.31201814058956917, - "grad_norm": 0.1533203125, - "learning_rate": 0.00018142210255782647, - "loss": 1.0305, - "step": 860 - }, - { - "epoch": 0.3127437641723356, - "grad_norm": 0.1591796875, - "learning_rate": 0.00018137314894137806, - "loss": 1.0597, - "step": 862 - }, - { - "epoch": 0.313469387755102, - "grad_norm": 0.1591796875, - "learning_rate": 0.00018132419532492964, - "loss": 1.0554, - "step": 864 - }, - { - "epoch": 0.3141950113378685, - "grad_norm": 0.134765625, - "learning_rate": 0.00018127524170848122, - "loss": 0.9902, - "step": 866 - }, - { - "epoch": 0.31492063492063493, - "grad_norm": 0.1435546875, - "learning_rate": 0.0001812262880920328, - "loss": 0.9705, - "step": 868 - }, - { - "epoch": 0.31564625850340133, - "grad_norm": 0.173828125, - "learning_rate": 0.00018117733447558439, - "loss": 1.0712, - "step": 870 - }, - { - "epoch": 0.3163718820861678, - "grad_norm": 0.14453125, - "learning_rate": 0.00018112838085913597, - "loss": 1.0368, - "step": 872 - }, - { - "epoch": 0.31709750566893424, - "grad_norm": 0.14453125, - "learning_rate": 0.00018107942724268755, - "loss": 0.9667, - "step": 874 - }, - { - "epoch": 0.3178231292517007, - "grad_norm": 0.17578125, - "learning_rate": 0.00018103047362623916, - "loss": 0.9888, - "step": 876 - }, - { - "epoch": 0.3185487528344671, - "grad_norm": 0.1689453125, - "learning_rate": 0.00018098152000979074, - "loss": 0.9979, - "step": 878 - }, - { - "epoch": 0.31927437641723355, - "grad_norm": 0.154296875, - "learning_rate": 0.0001809325663933423, - "loss": 0.9947, - "step": 880 - }, - { - "epoch": 0.32, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001808836127768939, - "loss": 1.0054, - "step": 882 - }, - { - "epoch": 0.32072562358276646, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001808346591604455, - "loss": 1.0014, - "step": 884 - }, - { - "epoch": 0.32145124716553286, - "grad_norm": 0.2021484375, - "learning_rate": 0.00018078570554399707, - "loss": 0.9482, - "step": 886 - }, - { - "epoch": 0.3221768707482993, - "grad_norm": 0.146484375, - "learning_rate": 0.00018073675192754865, - "loss": 1.0337, - "step": 888 - }, - { - "epoch": 0.32290249433106577, - "grad_norm": 0.150390625, - "learning_rate": 0.00018068779831110024, - "loss": 0.9075, - "step": 890 - }, - { - "epoch": 0.3236281179138322, - "grad_norm": 0.1650390625, - "learning_rate": 0.00018063884469465182, - "loss": 0.9731, - "step": 892 - }, - { - "epoch": 0.3243537414965986, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001805898910782034, - "loss": 1.0288, - "step": 894 - }, - { - "epoch": 0.3250793650793651, - "grad_norm": 0.1611328125, - "learning_rate": 0.000180540937461755, - "loss": 0.9984, - "step": 896 - }, - { - "epoch": 0.32580498866213153, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001804919838453066, - "loss": 0.9527, - "step": 898 - }, - { - "epoch": 0.32653061224489793, - "grad_norm": 0.150390625, - "learning_rate": 0.00018044303022885815, - "loss": 0.9399, - "step": 900 - }, - { - "epoch": 0.32653061224489793, - "eval_loss": 1.0258883237838745, - "eval_runtime": 95.9536, - "eval_samples_per_second": 18.759, - "eval_steps_per_second": 0.594, - "step": 900 - }, - { - "epoch": 0.3272562358276644, - "grad_norm": 0.1630859375, - "learning_rate": 0.00018039407661240976, - "loss": 0.9891, - "step": 902 - }, - { - "epoch": 0.32798185941043084, - "grad_norm": 0.16015625, - "learning_rate": 0.00018034512299596134, - "loss": 0.9403, - "step": 904 - }, - { - "epoch": 0.3287074829931973, - "grad_norm": 0.1787109375, - "learning_rate": 0.00018029616937951292, - "loss": 1.0682, - "step": 906 - }, - { - "epoch": 0.3294331065759637, - "grad_norm": 0.1435546875, - "learning_rate": 0.0001802472157630645, - "loss": 1.0294, - "step": 908 - }, - { - "epoch": 0.33015873015873015, - "grad_norm": 0.1435546875, - "learning_rate": 0.00018019826214661609, - "loss": 0.9923, - "step": 910 - }, - { - "epoch": 0.3308843537414966, - "grad_norm": 0.1533203125, - "learning_rate": 0.00018014930853016767, - "loss": 0.9585, - "step": 912 - }, - { - "epoch": 0.33160997732426306, - "grad_norm": 0.2041015625, - "learning_rate": 0.00018010035491371925, - "loss": 1.0097, - "step": 914 - }, - { - "epoch": 0.33233560090702946, - "grad_norm": 0.1533203125, - "learning_rate": 0.00018005140129727086, - "loss": 0.9935, - "step": 916 - }, - { - "epoch": 0.3330612244897959, - "grad_norm": 0.14453125, - "learning_rate": 0.00018000244768082244, - "loss": 0.9699, - "step": 918 - }, - { - "epoch": 0.33378684807256237, - "grad_norm": 0.146484375, - "learning_rate": 0.000179953494064374, - "loss": 1.0012, - "step": 920 - }, - { - "epoch": 0.3345124716553288, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001799045404479256, - "loss": 0.9984, - "step": 922 - }, - { - "epoch": 0.3352380952380952, - "grad_norm": 0.154296875, - "learning_rate": 0.0001798555868314772, - "loss": 0.978, - "step": 924 - }, - { - "epoch": 0.3359637188208617, - "grad_norm": 0.146484375, - "learning_rate": 0.00017980663321502877, - "loss": 1.0188, - "step": 926 - }, - { - "epoch": 0.33668934240362813, - "grad_norm": 0.169921875, - "learning_rate": 0.00017975767959858035, - "loss": 0.9777, - "step": 928 - }, - { - "epoch": 0.33741496598639453, - "grad_norm": 0.1416015625, - "learning_rate": 0.00017970872598213194, - "loss": 1.042, - "step": 930 - }, - { - "epoch": 0.338140589569161, - "grad_norm": 0.1455078125, - "learning_rate": 0.00017965977236568352, - "loss": 1.0074, - "step": 932 - }, - { - "epoch": 0.33886621315192744, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001796108187492351, - "loss": 1.0228, - "step": 934 - }, - { - "epoch": 0.3395918367346939, - "grad_norm": 0.169921875, - "learning_rate": 0.0001795618651327867, - "loss": 1.0906, - "step": 936 - }, - { - "epoch": 0.3403174603174603, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001795129115163383, - "loss": 1.0117, - "step": 938 - }, - { - "epoch": 0.34104308390022675, - "grad_norm": 0.185546875, - "learning_rate": 0.00017946395789988985, - "loss": 1.0274, - "step": 940 - }, - { - "epoch": 0.3417687074829932, - "grad_norm": 0.15234375, - "learning_rate": 0.00017941500428344146, - "loss": 1.0271, - "step": 942 - }, - { - "epoch": 0.34249433106575966, - "grad_norm": 0.181640625, - "learning_rate": 0.00017936605066699304, - "loss": 1.0064, - "step": 944 - }, - { - "epoch": 0.34321995464852606, - "grad_norm": 0.13671875, - "learning_rate": 0.00017931709705054462, - "loss": 0.9696, - "step": 946 - }, - { - "epoch": 0.3439455782312925, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001792681434340962, - "loss": 1.0467, - "step": 948 - }, - { - "epoch": 0.34467120181405897, - "grad_norm": 0.1748046875, - "learning_rate": 0.00017921918981764779, - "loss": 1.0357, - "step": 950 - }, - { - "epoch": 0.3453968253968254, - "grad_norm": 0.1357421875, - "learning_rate": 0.00017917023620119937, - "loss": 0.9136, - "step": 952 - }, - { - "epoch": 0.3461224489795918, - "grad_norm": 0.1630859375, - "learning_rate": 0.00017912128258475095, - "loss": 1.036, - "step": 954 - }, - { - "epoch": 0.3468480725623583, - "grad_norm": 0.1484375, - "learning_rate": 0.00017907232896830256, - "loss": 0.9946, - "step": 956 - }, - { - "epoch": 0.34757369614512473, - "grad_norm": 0.146484375, - "learning_rate": 0.00017902337535185411, - "loss": 0.9912, - "step": 958 - }, - { - "epoch": 0.34829931972789113, - "grad_norm": 0.138671875, - "learning_rate": 0.0001789744217354057, - "loss": 0.9606, - "step": 960 - }, - { - "epoch": 0.3490249433106576, - "grad_norm": 0.1396484375, - "learning_rate": 0.0001789254681189573, - "loss": 0.9763, - "step": 962 - }, - { - "epoch": 0.34975056689342404, - "grad_norm": 0.1337890625, - "learning_rate": 0.0001788765145025089, - "loss": 0.9389, - "step": 964 - }, - { - "epoch": 0.3504761904761905, - "grad_norm": 0.14453125, - "learning_rate": 0.00017882756088606047, - "loss": 0.9526, - "step": 966 - }, - { - "epoch": 0.3512018140589569, - "grad_norm": 0.1728515625, - "learning_rate": 0.00017877860726961205, - "loss": 0.9523, - "step": 968 - }, - { - "epoch": 0.35192743764172335, - "grad_norm": 0.1533203125, - "learning_rate": 0.00017872965365316364, - "loss": 1.0858, - "step": 970 - }, - { - "epoch": 0.3526530612244898, - "grad_norm": 0.1552734375, - "learning_rate": 0.00017868070003671522, - "loss": 0.9991, - "step": 972 - }, - { - "epoch": 0.35337868480725626, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001786317464202668, - "loss": 0.9704, - "step": 974 - }, - { - "epoch": 0.35410430839002266, - "grad_norm": 0.138671875, - "learning_rate": 0.0001785827928038184, - "loss": 0.9472, - "step": 976 - }, - { - "epoch": 0.3548299319727891, - "grad_norm": 0.140625, - "learning_rate": 0.00017853383918736996, - "loss": 1.0397, - "step": 978 - }, - { - "epoch": 0.35555555555555557, - "grad_norm": 0.154296875, - "learning_rate": 0.00017848488557092155, - "loss": 1.0412, - "step": 980 - }, - { - "epoch": 0.35628117913832197, - "grad_norm": 0.1591796875, - "learning_rate": 0.00017843593195447316, - "loss": 1.07, - "step": 982 - }, - { - "epoch": 0.3570068027210884, - "grad_norm": 0.15625, - "learning_rate": 0.00017838697833802474, - "loss": 0.947, - "step": 984 - }, - { - "epoch": 0.3577324263038549, - "grad_norm": 0.1494140625, - "learning_rate": 0.00017833802472157632, - "loss": 0.9768, - "step": 986 - }, - { - "epoch": 0.35845804988662133, - "grad_norm": 0.154296875, - "learning_rate": 0.0001782890711051279, - "loss": 1.0229, - "step": 988 - }, - { - "epoch": 0.35918367346938773, - "grad_norm": 0.1611328125, - "learning_rate": 0.00017824011748867949, - "loss": 1.0429, - "step": 990 - }, - { - "epoch": 0.3599092970521542, - "grad_norm": 0.154296875, - "learning_rate": 0.00017819116387223107, - "loss": 0.9842, - "step": 992 - }, - { - "epoch": 0.36063492063492064, - "grad_norm": 0.1416015625, - "learning_rate": 0.00017814221025578265, - "loss": 0.981, - "step": 994 - }, - { - "epoch": 0.3613605442176871, - "grad_norm": 0.1494140625, - "learning_rate": 0.00017809325663933426, - "loss": 1.0028, - "step": 996 - }, - { - "epoch": 0.3620861678004535, - "grad_norm": 0.1572265625, - "learning_rate": 0.00017804430302288581, - "loss": 1.061, - "step": 998 - }, - { - "epoch": 0.36281179138321995, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001779953494064374, - "loss": 1.0082, - "step": 1000 - }, - { - "epoch": 0.36281179138321995, - "eval_loss": 1.0250780582427979, - "eval_runtime": 92.3729, - "eval_samples_per_second": 19.486, - "eval_steps_per_second": 0.617, - "step": 1000 - }, - { - "epoch": 0.3635374149659864, - "grad_norm": 0.158203125, - "learning_rate": 0.000177946395789989, - "loss": 0.9744, - "step": 1002 - }, - { - "epoch": 0.36426303854875286, - "grad_norm": 0.1669921875, - "learning_rate": 0.0001778974421735406, - "loss": 1.0106, - "step": 1004 - }, - { - "epoch": 0.36498866213151926, - "grad_norm": 0.1455078125, - "learning_rate": 0.00017784848855709217, - "loss": 0.8815, - "step": 1006 - }, - { - "epoch": 0.3657142857142857, - "grad_norm": 0.1435546875, - "learning_rate": 0.00017779953494064375, - "loss": 0.9473, - "step": 1008 - }, - { - "epoch": 0.36643990929705217, - "grad_norm": 0.15625, - "learning_rate": 0.00017775058132419534, - "loss": 0.9883, - "step": 1010 - }, - { - "epoch": 0.36716553287981857, - "grad_norm": 0.12890625, - "learning_rate": 0.00017770162770774692, - "loss": 1.0715, - "step": 1012 - }, - { - "epoch": 0.367891156462585, - "grad_norm": 0.1630859375, - "learning_rate": 0.0001776526740912985, - "loss": 0.9768, - "step": 1014 - }, - { - "epoch": 0.3686167800453515, - "grad_norm": 0.1484375, - "learning_rate": 0.00017760372047485008, - "loss": 0.9155, - "step": 1016 - }, - { - "epoch": 0.36934240362811793, - "grad_norm": 0.154296875, - "learning_rate": 0.00017755476685840166, - "loss": 1.051, - "step": 1018 - }, - { - "epoch": 0.37006802721088433, - "grad_norm": 0.14453125, - "learning_rate": 0.00017750581324195325, - "loss": 1.0361, - "step": 1020 - }, - { - "epoch": 0.3707936507936508, - "grad_norm": 0.138671875, - "learning_rate": 0.00017745685962550486, - "loss": 1.024, - "step": 1022 - }, - { - "epoch": 0.37151927437641724, - "grad_norm": 0.1484375, - "learning_rate": 0.00017740790600905644, - "loss": 1.0117, - "step": 1024 - }, - { - "epoch": 0.3722448979591837, - "grad_norm": 0.1650390625, - "learning_rate": 0.000177358952392608, - "loss": 0.992, - "step": 1026 - }, - { - "epoch": 0.3729705215419501, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001773099987761596, - "loss": 1.0872, - "step": 1028 - }, - { - "epoch": 0.37369614512471655, - "grad_norm": 0.1435546875, - "learning_rate": 0.00017726104515971118, - "loss": 0.9335, - "step": 1030 - }, - { - "epoch": 0.374421768707483, - "grad_norm": 0.1494140625, - "learning_rate": 0.00017721209154326277, - "loss": 1.017, - "step": 1032 - }, - { - "epoch": 0.37514739229024946, - "grad_norm": 0.1455078125, - "learning_rate": 0.00017716313792681435, - "loss": 0.9909, - "step": 1034 - }, - { - "epoch": 0.37587301587301586, - "grad_norm": 0.1484375, - "learning_rate": 0.00017711418431036593, - "loss": 1.0097, - "step": 1036 - }, - { - "epoch": 0.3765986394557823, - "grad_norm": 0.1455078125, - "learning_rate": 0.00017706523069391751, - "loss": 0.9652, - "step": 1038 - }, - { - "epoch": 0.37732426303854877, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001770162770774691, - "loss": 0.9893, - "step": 1040 - }, - { - "epoch": 0.37804988662131517, - "grad_norm": 0.13671875, - "learning_rate": 0.0001769673234610207, - "loss": 1.0294, - "step": 1042 - }, - { - "epoch": 0.3787755102040816, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001769183698445723, - "loss": 1.0322, - "step": 1044 - }, - { - "epoch": 0.3795011337868481, - "grad_norm": 0.146484375, - "learning_rate": 0.00017686941622812384, - "loss": 0.908, - "step": 1046 - }, - { - "epoch": 0.38022675736961453, - "grad_norm": 0.169921875, - "learning_rate": 0.00017682046261167545, - "loss": 1.0802, - "step": 1048 - }, - { - "epoch": 0.38095238095238093, - "grad_norm": 0.1748046875, - "learning_rate": 0.00017677150899522703, - "loss": 1.1756, - "step": 1050 - }, - { - "epoch": 0.3816780045351474, - "grad_norm": 0.15234375, - "learning_rate": 0.00017672255537877862, - "loss": 0.9534, - "step": 1052 - }, - { - "epoch": 0.38240362811791384, - "grad_norm": 0.177734375, - "learning_rate": 0.0001766736017623302, - "loss": 0.9512, - "step": 1054 - }, - { - "epoch": 0.3831292517006803, - "grad_norm": 0.1513671875, - "learning_rate": 0.00017662464814588178, - "loss": 1.0466, - "step": 1056 - }, - { - "epoch": 0.3838548752834467, - "grad_norm": 0.173828125, - "learning_rate": 0.00017657569452943336, - "loss": 1.0233, - "step": 1058 - }, - { - "epoch": 0.38458049886621315, - "grad_norm": 0.1591796875, - "learning_rate": 0.00017652674091298495, - "loss": 1.0718, - "step": 1060 - }, - { - "epoch": 0.3853061224489796, - "grad_norm": 0.1689453125, - "learning_rate": 0.00017647778729653656, - "loss": 1.0447, - "step": 1062 - }, - { - "epoch": 0.38603174603174606, - "grad_norm": 0.16796875, - "learning_rate": 0.00017642883368008814, - "loss": 1.0104, - "step": 1064 - }, - { - "epoch": 0.38675736961451246, - "grad_norm": 0.1357421875, - "learning_rate": 0.0001763798800636397, - "loss": 0.9774, - "step": 1066 - }, - { - "epoch": 0.3874829931972789, - "grad_norm": 0.158203125, - "learning_rate": 0.0001763309264471913, - "loss": 0.9544, - "step": 1068 - }, - { - "epoch": 0.38820861678004537, - "grad_norm": 0.1884765625, - "learning_rate": 0.00017628197283074288, - "loss": 1.0006, - "step": 1070 - }, - { - "epoch": 0.38893424036281177, - "grad_norm": 0.138671875, - "learning_rate": 0.00017623301921429447, - "loss": 0.9301, - "step": 1072 - }, - { - "epoch": 0.3896598639455782, - "grad_norm": 0.1533203125, - "learning_rate": 0.00017618406559784605, - "loss": 1.0312, - "step": 1074 - }, - { - "epoch": 0.3903854875283447, - "grad_norm": 0.138671875, - "learning_rate": 0.00017613511198139763, - "loss": 0.9978, - "step": 1076 - }, - { - "epoch": 0.39111111111111113, - "grad_norm": 0.158203125, - "learning_rate": 0.00017608615836494921, - "loss": 0.9708, - "step": 1078 - }, - { - "epoch": 0.39183673469387753, - "grad_norm": 0.1376953125, - "learning_rate": 0.0001760372047485008, - "loss": 0.9779, - "step": 1080 - }, - { - "epoch": 0.392562358276644, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001759882511320524, - "loss": 0.9583, - "step": 1082 - }, - { - "epoch": 0.39328798185941044, - "grad_norm": 0.1591796875, - "learning_rate": 0.000175939297515604, - "loss": 1.0361, - "step": 1084 - }, - { - "epoch": 0.3940136054421769, - "grad_norm": 0.1650390625, - "learning_rate": 0.00017589034389915554, - "loss": 1.061, - "step": 1086 - }, - { - "epoch": 0.3947392290249433, - "grad_norm": 0.177734375, - "learning_rate": 0.00017584139028270715, - "loss": 1.0351, - "step": 1088 - }, - { - "epoch": 0.39546485260770975, - "grad_norm": 0.1533203125, - "learning_rate": 0.00017579243666625873, - "loss": 0.9836, - "step": 1090 - }, - { - "epoch": 0.3961904761904762, - "grad_norm": 0.1376953125, - "learning_rate": 0.00017574348304981032, - "loss": 0.9366, - "step": 1092 - }, - { - "epoch": 0.3969160997732426, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001756945294333619, - "loss": 0.9386, - "step": 1094 - }, - { - "epoch": 0.39764172335600906, - "grad_norm": 0.134765625, - "learning_rate": 0.00017564557581691348, - "loss": 0.9872, - "step": 1096 - }, - { - "epoch": 0.3983673469387755, - "grad_norm": 0.1572265625, - "learning_rate": 0.00017559662220046506, - "loss": 0.962, - "step": 1098 - }, - { - "epoch": 0.39909297052154197, - "grad_norm": 0.1611328125, - "learning_rate": 0.00017554766858401665, - "loss": 1.0224, - "step": 1100 - }, - { - "epoch": 0.39909297052154197, - "eval_loss": 1.024340033531189, - "eval_runtime": 96.4085, - "eval_samples_per_second": 18.671, - "eval_steps_per_second": 0.591, - "step": 1100 - }, - { - "epoch": 0.39981859410430837, - "grad_norm": 0.146484375, - "learning_rate": 0.00017549871496756826, - "loss": 0.9655, - "step": 1102 - }, - { - "epoch": 0.4005442176870748, - "grad_norm": 0.158203125, - "learning_rate": 0.0001754497613511198, - "loss": 0.9732, - "step": 1104 - }, - { - "epoch": 0.4012698412698413, - "grad_norm": 0.1630859375, - "learning_rate": 0.0001754008077346714, - "loss": 1.0525, - "step": 1106 - }, - { - "epoch": 0.40199546485260773, - "grad_norm": 0.1591796875, - "learning_rate": 0.000175351854118223, - "loss": 0.9244, - "step": 1108 - }, - { - "epoch": 0.40272108843537413, - "grad_norm": 0.162109375, - "learning_rate": 0.00017530290050177458, - "loss": 0.9976, - "step": 1110 - }, - { - "epoch": 0.4034467120181406, - "grad_norm": 0.1533203125, - "learning_rate": 0.00017525394688532617, - "loss": 0.9802, - "step": 1112 - }, - { - "epoch": 0.40417233560090704, - "grad_norm": 0.158203125, - "learning_rate": 0.00017520499326887775, - "loss": 0.9932, - "step": 1114 - }, - { - "epoch": 0.4048979591836735, - "grad_norm": 0.15234375, - "learning_rate": 0.00017515603965242933, - "loss": 1.0189, - "step": 1116 - }, - { - "epoch": 0.4056235827664399, - "grad_norm": 0.142578125, - "learning_rate": 0.00017510708603598091, - "loss": 0.908, - "step": 1118 - }, - { - "epoch": 0.40634920634920635, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001750581324195325, - "loss": 0.9909, - "step": 1120 - }, - { - "epoch": 0.4070748299319728, - "grad_norm": 0.166015625, - "learning_rate": 0.0001750091788030841, - "loss": 1.0264, - "step": 1122 - }, - { - "epoch": 0.4078004535147392, - "grad_norm": 0.1865234375, - "learning_rate": 0.00017496022518663566, - "loss": 1.0606, - "step": 1124 - }, - { - "epoch": 0.40852607709750566, - "grad_norm": 0.1552734375, - "learning_rate": 0.00017491127157018724, - "loss": 0.9924, - "step": 1126 - }, - { - "epoch": 0.4092517006802721, - "grad_norm": 0.14453125, - "learning_rate": 0.00017486231795373885, - "loss": 0.9948, - "step": 1128 - }, - { - "epoch": 0.40997732426303857, - "grad_norm": 0.134765625, - "learning_rate": 0.00017481336433729043, - "loss": 0.9702, - "step": 1130 - }, - { - "epoch": 0.41070294784580497, - "grad_norm": 0.15234375, - "learning_rate": 0.00017476441072084202, - "loss": 1.0459, - "step": 1132 - }, - { - "epoch": 0.4114285714285714, - "grad_norm": 0.150390625, - "learning_rate": 0.0001747154571043936, - "loss": 0.9733, - "step": 1134 - }, - { - "epoch": 0.4121541950113379, - "grad_norm": 0.14453125, - "learning_rate": 0.00017466650348794518, - "loss": 0.9991, - "step": 1136 - }, - { - "epoch": 0.41287981859410433, - "grad_norm": 0.1630859375, - "learning_rate": 0.00017461754987149676, - "loss": 0.9936, - "step": 1138 - }, - { - "epoch": 0.41360544217687073, - "grad_norm": 0.166015625, - "learning_rate": 0.00017456859625504835, - "loss": 0.9779, - "step": 1140 - }, - { - "epoch": 0.4143310657596372, - "grad_norm": 0.1669921875, - "learning_rate": 0.00017451964263859996, - "loss": 0.996, - "step": 1142 - }, - { - "epoch": 0.41505668934240364, - "grad_norm": 0.140625, - "learning_rate": 0.0001744706890221515, - "loss": 1.0079, - "step": 1144 - }, - { - "epoch": 0.4157823129251701, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001744217354057031, - "loss": 0.9805, - "step": 1146 - }, - { - "epoch": 0.4165079365079365, - "grad_norm": 0.146484375, - "learning_rate": 0.0001743727817892547, - "loss": 0.9657, - "step": 1148 - }, - { - "epoch": 0.41723356009070295, - "grad_norm": 0.1416015625, - "learning_rate": 0.00017432382817280628, - "loss": 1.0011, - "step": 1150 - }, - { - "epoch": 0.4179591836734694, - "grad_norm": 0.14453125, - "learning_rate": 0.00017427487455635787, - "loss": 1.0128, - "step": 1152 - }, - { - "epoch": 0.4186848072562358, - "grad_norm": 0.154296875, - "learning_rate": 0.00017422592093990945, - "loss": 0.97, - "step": 1154 - }, - { - "epoch": 0.41941043083900226, - "grad_norm": 0.1611328125, - "learning_rate": 0.00017417696732346103, - "loss": 1.0179, - "step": 1156 - }, - { - "epoch": 0.4201360544217687, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001741280137070126, - "loss": 1.0459, - "step": 1158 - }, - { - "epoch": 0.42086167800453517, - "grad_norm": 0.1845703125, - "learning_rate": 0.0001740790600905642, - "loss": 1.0039, - "step": 1160 - }, - { - "epoch": 0.42158730158730157, - "grad_norm": 0.1484375, - "learning_rate": 0.00017403010647411578, - "loss": 1.0498, - "step": 1162 - }, - { - "epoch": 0.422312925170068, - "grad_norm": 0.146484375, - "learning_rate": 0.00017398115285766736, - "loss": 1.0111, - "step": 1164 - }, - { - "epoch": 0.4230385487528345, - "grad_norm": 0.1513671875, - "learning_rate": 0.00017393219924121894, - "loss": 0.9945, - "step": 1166 - }, - { - "epoch": 0.42376417233560093, - "grad_norm": 0.1474609375, - "learning_rate": 0.00017388324562477055, - "loss": 0.9439, - "step": 1168 - }, - { - "epoch": 0.42448979591836733, - "grad_norm": 0.146484375, - "learning_rate": 0.00017383429200832213, - "loss": 0.9019, - "step": 1170 - }, - { - "epoch": 0.4252154195011338, - "grad_norm": 0.1787109375, - "learning_rate": 0.0001737853383918737, - "loss": 0.9996, - "step": 1172 - }, - { - "epoch": 0.42594104308390024, - "grad_norm": 0.171875, - "learning_rate": 0.0001737363847754253, - "loss": 1.0023, - "step": 1174 - }, - { - "epoch": 0.4266666666666667, - "grad_norm": 0.181640625, - "learning_rate": 0.00017368743115897688, - "loss": 0.9721, - "step": 1176 - }, - { - "epoch": 0.4273922902494331, - "grad_norm": 0.154296875, - "learning_rate": 0.00017363847754252846, - "loss": 1.0203, - "step": 1178 - }, - { - "epoch": 0.42811791383219955, - "grad_norm": 0.16015625, - "learning_rate": 0.00017358952392608005, - "loss": 0.9972, - "step": 1180 - }, - { - "epoch": 0.428843537414966, - "grad_norm": 0.1728515625, - "learning_rate": 0.00017354057030963163, - "loss": 0.9853, - "step": 1182 - }, - { - "epoch": 0.4295691609977324, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001734916166931832, - "loss": 0.9784, - "step": 1184 - }, - { - "epoch": 0.43029478458049886, - "grad_norm": 0.166015625, - "learning_rate": 0.0001734426630767348, - "loss": 1.0201, - "step": 1186 - }, - { - "epoch": 0.4310204081632653, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001733937094602864, - "loss": 1.0091, - "step": 1188 - }, - { - "epoch": 0.43174603174603177, - "grad_norm": 0.158203125, - "learning_rate": 0.00017334475584383798, - "loss": 1.0313, - "step": 1190 - }, - { - "epoch": 0.43247165532879817, - "grad_norm": 0.1474609375, - "learning_rate": 0.00017329580222738954, - "loss": 1.0487, - "step": 1192 - }, - { - "epoch": 0.4331972789115646, - "grad_norm": 0.1455078125, - "learning_rate": 0.00017324684861094115, - "loss": 1.0059, - "step": 1194 - }, - { - "epoch": 0.4339229024943311, - "grad_norm": 0.1484375, - "learning_rate": 0.00017319789499449273, - "loss": 0.9841, - "step": 1196 - }, - { - "epoch": 0.43464852607709753, - "grad_norm": 0.1484375, - "learning_rate": 0.0001731489413780443, - "loss": 0.9891, - "step": 1198 - }, - { - "epoch": 0.43537414965986393, - "grad_norm": 0.154296875, - "learning_rate": 0.0001730999877615959, - "loss": 1.0482, - "step": 1200 - }, - { - "epoch": 0.43537414965986393, - "eval_loss": 1.023897647857666, - "eval_runtime": 91.99, - "eval_samples_per_second": 19.567, - "eval_steps_per_second": 0.62, - "step": 1200 - }, - { - "epoch": 0.4360997732426304, - "grad_norm": 0.1630859375, - "learning_rate": 0.00017305103414514748, - "loss": 0.9969, - "step": 1202 - }, - { - "epoch": 0.43682539682539684, - "grad_norm": 0.1435546875, - "learning_rate": 0.00017300208052869906, - "loss": 0.9979, - "step": 1204 - }, - { - "epoch": 0.43755102040816324, - "grad_norm": 0.158203125, - "learning_rate": 0.00017295312691225064, - "loss": 0.9974, - "step": 1206 - }, - { - "epoch": 0.4382766439909297, - "grad_norm": 0.16015625, - "learning_rate": 0.00017290417329580225, - "loss": 1.0932, - "step": 1208 - }, - { - "epoch": 0.43900226757369615, - "grad_norm": 0.1474609375, - "learning_rate": 0.00017285521967935383, - "loss": 0.9845, - "step": 1210 - }, - { - "epoch": 0.4397278911564626, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001728062660629054, - "loss": 1.0405, - "step": 1212 - }, - { - "epoch": 0.440453514739229, - "grad_norm": 0.1396484375, - "learning_rate": 0.000172757312446457, - "loss": 0.9646, - "step": 1214 - }, - { - "epoch": 0.44117913832199546, - "grad_norm": 0.13671875, - "learning_rate": 0.00017270835883000858, - "loss": 0.9535, - "step": 1216 - }, - { - "epoch": 0.4419047619047619, - "grad_norm": 0.166015625, - "learning_rate": 0.00017265940521356016, - "loss": 1.0106, - "step": 1218 - }, - { - "epoch": 0.44263038548752837, - "grad_norm": 0.1787109375, - "learning_rate": 0.00017261045159711175, - "loss": 0.9388, - "step": 1220 - }, - { - "epoch": 0.44335600907029477, - "grad_norm": 0.142578125, - "learning_rate": 0.00017256149798066333, - "loss": 0.8825, - "step": 1222 - }, - { - "epoch": 0.4440816326530612, - "grad_norm": 0.150390625, - "learning_rate": 0.0001725125443642149, - "loss": 0.98, - "step": 1224 - }, - { - "epoch": 0.4448072562358277, - "grad_norm": 0.1484375, - "learning_rate": 0.0001724635907477665, - "loss": 0.996, - "step": 1226 - }, - { - "epoch": 0.44553287981859413, - "grad_norm": 0.146484375, - "learning_rate": 0.0001724146371313181, - "loss": 0.9907, - "step": 1228 - }, - { - "epoch": 0.44625850340136053, - "grad_norm": 0.154296875, - "learning_rate": 0.00017236568351486968, - "loss": 1.0013, - "step": 1230 - }, - { - "epoch": 0.446984126984127, - "grad_norm": 0.1748046875, - "learning_rate": 0.00017231672989842124, - "loss": 0.9736, - "step": 1232 - }, - { - "epoch": 0.44770975056689344, - "grad_norm": 0.1484375, - "learning_rate": 0.00017226777628197285, - "loss": 1.0385, - "step": 1234 - }, - { - "epoch": 0.44843537414965984, - "grad_norm": 0.1748046875, - "learning_rate": 0.00017221882266552443, - "loss": 0.9977, - "step": 1236 - }, - { - "epoch": 0.4491609977324263, - "grad_norm": 0.166015625, - "learning_rate": 0.000172169869049076, - "loss": 1.0035, - "step": 1238 - }, - { - "epoch": 0.44988662131519275, - "grad_norm": 0.162109375, - "learning_rate": 0.0001721209154326276, - "loss": 1.0349, - "step": 1240 - }, - { - "epoch": 0.4506122448979592, - "grad_norm": 0.1650390625, - "learning_rate": 0.00017207196181617918, - "loss": 1.0325, - "step": 1242 - }, - { - "epoch": 0.4513378684807256, - "grad_norm": 0.1796875, - "learning_rate": 0.00017202300819973076, - "loss": 1.0148, - "step": 1244 - }, - { - "epoch": 0.45206349206349206, - "grad_norm": 0.1533203125, - "learning_rate": 0.00017197405458328234, - "loss": 0.997, - "step": 1246 - }, - { - "epoch": 0.4527891156462585, - "grad_norm": 0.1962890625, - "learning_rate": 0.00017192510096683395, - "loss": 0.913, - "step": 1248 - }, - { - "epoch": 0.45351473922902497, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001718761473503855, - "loss": 0.9991, - "step": 1250 - }, - { - "epoch": 0.45424036281179136, - "grad_norm": 0.146484375, - "learning_rate": 0.0001718271937339371, - "loss": 1.0035, - "step": 1252 - }, - { - "epoch": 0.4549659863945578, - "grad_norm": 0.130859375, - "learning_rate": 0.0001717782401174887, - "loss": 0.9321, - "step": 1254 - }, - { - "epoch": 0.4556916099773243, - "grad_norm": 0.142578125, - "learning_rate": 0.00017172928650104028, - "loss": 0.9722, - "step": 1256 - }, - { - "epoch": 0.45641723356009073, - "grad_norm": 0.1533203125, - "learning_rate": 0.00017168033288459186, - "loss": 1.0485, - "step": 1258 - }, - { - "epoch": 0.45714285714285713, - "grad_norm": 0.14453125, - "learning_rate": 0.00017163137926814345, - "loss": 0.953, - "step": 1260 - }, - { - "epoch": 0.4578684807256236, - "grad_norm": 0.1884765625, - "learning_rate": 0.00017158242565169503, - "loss": 1.0478, - "step": 1262 - }, - { - "epoch": 0.45859410430839004, - "grad_norm": 0.1337890625, - "learning_rate": 0.0001715334720352466, - "loss": 0.8863, - "step": 1264 - }, - { - "epoch": 0.45931972789115644, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001714845184187982, - "loss": 0.9771, - "step": 1266 - }, - { - "epoch": 0.4600453514739229, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001714355648023498, - "loss": 0.9525, - "step": 1268 - }, - { - "epoch": 0.46077097505668935, - "grad_norm": 0.1533203125, - "learning_rate": 0.00017138661118590136, - "loss": 1.0151, - "step": 1270 - }, - { - "epoch": 0.4614965986394558, - "grad_norm": 0.16015625, - "learning_rate": 0.00017133765756945294, - "loss": 1.0558, - "step": 1272 - }, - { - "epoch": 0.4622222222222222, - "grad_norm": 0.138671875, - "learning_rate": 0.00017128870395300455, - "loss": 1.0136, - "step": 1274 - }, - { - "epoch": 0.46294784580498866, - "grad_norm": 0.1474609375, - "learning_rate": 0.00017123975033655613, - "loss": 0.9975, - "step": 1276 - }, - { - "epoch": 0.4636734693877551, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001711907967201077, - "loss": 1.028, - "step": 1278 - }, - { - "epoch": 0.46439909297052157, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001711418431036593, - "loss": 0.9967, - "step": 1280 - }, - { - "epoch": 0.46512471655328796, - "grad_norm": 0.1650390625, - "learning_rate": 0.00017109288948721088, - "loss": 0.939, - "step": 1282 - }, - { - "epoch": 0.4658503401360544, - "grad_norm": 0.15234375, - "learning_rate": 0.00017104393587076246, - "loss": 1.011, - "step": 1284 - }, - { - "epoch": 0.4665759637188209, - "grad_norm": 0.146484375, - "learning_rate": 0.00017099498225431404, - "loss": 1.0439, - "step": 1286 - }, - { - "epoch": 0.46730158730158733, - "grad_norm": 0.1748046875, - "learning_rate": 0.00017094602863786565, - "loss": 1.0081, - "step": 1288 - }, - { - "epoch": 0.46802721088435373, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001708970750214172, - "loss": 0.9553, - "step": 1290 - }, - { - "epoch": 0.4687528344671202, - "grad_norm": 0.1640625, - "learning_rate": 0.0001708481214049688, - "loss": 0.9824, - "step": 1292 - }, - { - "epoch": 0.46947845804988664, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001707991677885204, - "loss": 1.045, - "step": 1294 - }, - { - "epoch": 0.47020408163265304, - "grad_norm": 0.1611328125, - "learning_rate": 0.00017075021417207198, - "loss": 1.0157, - "step": 1296 - }, - { - "epoch": 0.4709297052154195, - "grad_norm": 0.185546875, - "learning_rate": 0.00017070126055562356, - "loss": 0.9727, - "step": 1298 - }, - { - "epoch": 0.47165532879818595, - "grad_norm": 0.1650390625, - "learning_rate": 0.00017065230693917515, - "loss": 0.9486, - "step": 1300 - }, - { - "epoch": 0.47165532879818595, - "eval_loss": 1.0229827165603638, - "eval_runtime": 91.9183, - "eval_samples_per_second": 19.583, - "eval_steps_per_second": 0.62, - "step": 1300 - }, - { - "epoch": 0.4723809523809524, - "grad_norm": 0.1455078125, - "learning_rate": 0.00017060335332272673, - "loss": 1.0029, - "step": 1302 - }, - { - "epoch": 0.4731065759637188, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001705543997062783, - "loss": 1.0371, - "step": 1304 - }, - { - "epoch": 0.47383219954648526, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001705054460898299, - "loss": 0.9764, - "step": 1306 - }, - { - "epoch": 0.4745578231292517, - "grad_norm": 0.169921875, - "learning_rate": 0.00017045649247338147, - "loss": 0.9555, - "step": 1308 - }, - { - "epoch": 0.47528344671201816, - "grad_norm": 0.171875, - "learning_rate": 0.00017040753885693306, - "loss": 1.0557, - "step": 1310 - }, - { - "epoch": 0.47600907029478456, - "grad_norm": 0.18359375, - "learning_rate": 0.00017035858524048464, - "loss": 0.9758, - "step": 1312 - }, - { - "epoch": 0.476734693877551, - "grad_norm": 0.1611328125, - "learning_rate": 0.00017030963162403625, - "loss": 1.0147, - "step": 1314 - }, - { - "epoch": 0.4774603174603175, - "grad_norm": 0.1630859375, - "learning_rate": 0.00017026067800758783, - "loss": 1.0386, - "step": 1316 - }, - { - "epoch": 0.4781859410430839, - "grad_norm": 0.1435546875, - "learning_rate": 0.00017021172439113939, - "loss": 0.9744, - "step": 1318 - }, - { - "epoch": 0.47891156462585033, - "grad_norm": 0.1611328125, - "learning_rate": 0.000170162770774691, - "loss": 0.9863, - "step": 1320 - }, - { - "epoch": 0.4796371882086168, - "grad_norm": 0.171875, - "learning_rate": 0.00017011381715824258, - "loss": 1.0313, - "step": 1322 - }, - { - "epoch": 0.48036281179138324, - "grad_norm": 0.1484375, - "learning_rate": 0.00017006486354179416, - "loss": 0.9102, - "step": 1324 - }, - { - "epoch": 0.48108843537414964, - "grad_norm": 0.1650390625, - "learning_rate": 0.00017001590992534574, - "loss": 1.0352, - "step": 1326 - }, - { - "epoch": 0.4818140589569161, - "grad_norm": 0.140625, - "learning_rate": 0.00016996695630889732, - "loss": 1.011, - "step": 1328 - }, - { - "epoch": 0.48253968253968255, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001699180026924489, - "loss": 0.9854, - "step": 1330 - }, - { - "epoch": 0.483265306122449, - "grad_norm": 0.15234375, - "learning_rate": 0.0001698690490760005, - "loss": 0.9816, - "step": 1332 - }, - { - "epoch": 0.4839909297052154, - "grad_norm": 0.140625, - "learning_rate": 0.0001698200954595521, - "loss": 0.9738, - "step": 1334 - }, - { - "epoch": 0.48471655328798186, - "grad_norm": 0.19140625, - "learning_rate": 0.00016977114184310368, - "loss": 1.0348, - "step": 1336 - }, - { - "epoch": 0.4854421768707483, - "grad_norm": 0.1669921875, - "learning_rate": 0.00016972218822665524, - "loss": 1.0062, - "step": 1338 - }, - { - "epoch": 0.48616780045351476, - "grad_norm": 0.1357421875, - "learning_rate": 0.00016967323461020684, - "loss": 0.9665, - "step": 1340 - }, - { - "epoch": 0.48689342403628116, - "grad_norm": 0.140625, - "learning_rate": 0.00016962428099375843, - "loss": 0.978, - "step": 1342 - }, - { - "epoch": 0.4876190476190476, - "grad_norm": 0.1484375, - "learning_rate": 0.00016957532737731, - "loss": 1.0053, - "step": 1344 - }, - { - "epoch": 0.4883446712018141, - "grad_norm": 0.162109375, - "learning_rate": 0.0001695263737608616, - "loss": 1.0164, - "step": 1346 - }, - { - "epoch": 0.4890702947845805, - "grad_norm": 0.154296875, - "learning_rate": 0.00016947742014441317, - "loss": 1.0096, - "step": 1348 - }, - { - "epoch": 0.4897959183673469, - "grad_norm": 0.146484375, - "learning_rate": 0.00016942846652796476, - "loss": 0.9065, - "step": 1350 - }, - { - "epoch": 0.4905215419501134, - "grad_norm": 0.1494140625, - "learning_rate": 0.00016937951291151634, - "loss": 0.9667, - "step": 1352 - }, - { - "epoch": 0.49124716553287984, - "grad_norm": 0.15234375, - "learning_rate": 0.00016933055929506795, - "loss": 1.0211, - "step": 1354 - }, - { - "epoch": 0.49197278911564624, - "grad_norm": 0.1640625, - "learning_rate": 0.00016928160567861953, - "loss": 1.1053, - "step": 1356 - }, - { - "epoch": 0.4926984126984127, - "grad_norm": 0.1455078125, - "learning_rate": 0.00016923265206217109, - "loss": 1.0523, - "step": 1358 - }, - { - "epoch": 0.49342403628117915, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001691836984457227, - "loss": 0.9851, - "step": 1360 - }, - { - "epoch": 0.4941496598639456, - "grad_norm": 0.1611328125, - "learning_rate": 0.00016913474482927428, - "loss": 1.0227, - "step": 1362 - }, - { - "epoch": 0.494875283446712, - "grad_norm": 0.1474609375, - "learning_rate": 0.00016908579121282586, - "loss": 0.9792, - "step": 1364 - }, - { - "epoch": 0.49560090702947845, - "grad_norm": 0.15234375, - "learning_rate": 0.00016903683759637744, - "loss": 0.9671, - "step": 1366 - }, - { - "epoch": 0.4963265306122449, - "grad_norm": 0.154296875, - "learning_rate": 0.00016898788397992902, - "loss": 0.9982, - "step": 1368 - }, - { - "epoch": 0.49705215419501136, - "grad_norm": 0.1484375, - "learning_rate": 0.0001689389303634806, - "loss": 0.9745, - "step": 1370 - }, - { - "epoch": 0.49777777777777776, - "grad_norm": 0.166015625, - "learning_rate": 0.0001688899767470322, - "loss": 1.0174, - "step": 1372 - }, - { - "epoch": 0.4985034013605442, - "grad_norm": 0.150390625, - "learning_rate": 0.0001688410231305838, - "loss": 1.0175, - "step": 1374 - }, - { - "epoch": 0.4992290249433107, - "grad_norm": 0.1708984375, - "learning_rate": 0.00016879206951413538, - "loss": 0.9615, - "step": 1376 - }, - { - "epoch": 0.4999546485260771, - "grad_norm": 0.1572265625, - "learning_rate": 0.00016874311589768694, - "loss": 0.9905, - "step": 1378 - }, - { - "epoch": 0.5006802721088436, - "grad_norm": 0.2158203125, - "learning_rate": 0.00016869416228123854, - "loss": 0.9977, - "step": 1380 - }, - { - "epoch": 0.50140589569161, - "grad_norm": 0.1552734375, - "learning_rate": 0.00016864520866479013, - "loss": 0.9374, - "step": 1382 - }, - { - "epoch": 0.5021315192743764, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001685962550483417, - "loss": 1.0049, - "step": 1384 - }, - { - "epoch": 0.5028571428571429, - "grad_norm": 0.158203125, - "learning_rate": 0.0001685473014318933, - "loss": 0.9839, - "step": 1386 - }, - { - "epoch": 0.5035827664399093, - "grad_norm": 0.162109375, - "learning_rate": 0.00016849834781544487, - "loss": 0.9599, - "step": 1388 - }, - { - "epoch": 0.5043083900226757, - "grad_norm": 0.1767578125, - "learning_rate": 0.00016844939419899646, - "loss": 0.9604, - "step": 1390 - }, - { - "epoch": 0.5050340136054422, - "grad_norm": 0.142578125, - "learning_rate": 0.00016840044058254804, - "loss": 0.9192, - "step": 1392 - }, - { - "epoch": 0.5057596371882086, - "grad_norm": 0.142578125, - "learning_rate": 0.00016835148696609965, - "loss": 0.9461, - "step": 1394 - }, - { - "epoch": 0.5064852607709751, - "grad_norm": 0.17578125, - "learning_rate": 0.0001683025333496512, - "loss": 0.9682, - "step": 1396 - }, - { - "epoch": 0.5072108843537415, - "grad_norm": 0.1455078125, - "learning_rate": 0.00016825357973320279, - "loss": 1.0531, - "step": 1398 - }, - { - "epoch": 0.5079365079365079, - "grad_norm": 0.1845703125, - "learning_rate": 0.0001682046261167544, - "loss": 0.998, - "step": 1400 - }, - { - "epoch": 0.5079365079365079, - "eval_loss": 1.0224273204803467, - "eval_runtime": 90.8291, - "eval_samples_per_second": 19.817, - "eval_steps_per_second": 0.628, - "step": 1400 - }, - { - "epoch": 0.5086621315192744, - "grad_norm": 0.16796875, - "learning_rate": 0.00016815567250030598, - "loss": 1.0047, - "step": 1402 - }, - { - "epoch": 0.5093877551020408, - "grad_norm": 0.1591796875, - "learning_rate": 0.00016810671888385756, - "loss": 0.9845, - "step": 1404 - }, - { - "epoch": 0.5101133786848072, - "grad_norm": 0.1669921875, - "learning_rate": 0.00016805776526740914, - "loss": 0.9928, - "step": 1406 - }, - { - "epoch": 0.5108390022675737, - "grad_norm": 0.1552734375, - "learning_rate": 0.00016800881165096072, - "loss": 0.971, - "step": 1408 - }, - { - "epoch": 0.5115646258503401, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001679598580345123, - "loss": 0.9549, - "step": 1410 - }, - { - "epoch": 0.5122902494331065, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001679109044180639, - "loss": 1.056, - "step": 1412 - }, - { - "epoch": 0.513015873015873, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001678619508016155, - "loss": 1.037, - "step": 1414 - }, - { - "epoch": 0.5137414965986394, - "grad_norm": 0.150390625, - "learning_rate": 0.00016781299718516705, - "loss": 1.0098, - "step": 1416 - }, - { - "epoch": 0.514467120181406, - "grad_norm": 0.1416015625, - "learning_rate": 0.00016776404356871864, - "loss": 0.9179, - "step": 1418 - }, - { - "epoch": 0.5151927437641723, - "grad_norm": 0.154296875, - "learning_rate": 0.00016771508995227024, - "loss": 1.0747, - "step": 1420 - }, - { - "epoch": 0.5159183673469387, - "grad_norm": 0.142578125, - "learning_rate": 0.00016766613633582183, - "loss": 0.9267, - "step": 1422 - }, - { - "epoch": 0.5166439909297053, - "grad_norm": 0.146484375, - "learning_rate": 0.0001676171827193734, - "loss": 0.9201, - "step": 1424 - }, - { - "epoch": 0.5173696145124717, - "grad_norm": 0.15234375, - "learning_rate": 0.000167568229102925, - "loss": 1.0046, - "step": 1426 - }, - { - "epoch": 0.518095238095238, - "grad_norm": 0.158203125, - "learning_rate": 0.00016751927548647657, - "loss": 1.0547, - "step": 1428 - }, - { - "epoch": 0.5188208616780046, - "grad_norm": 0.158203125, - "learning_rate": 0.00016747032187002816, - "loss": 0.9742, - "step": 1430 - }, - { - "epoch": 0.519546485260771, - "grad_norm": 0.162109375, - "learning_rate": 0.00016742136825357974, - "loss": 1.0289, - "step": 1432 - }, - { - "epoch": 0.5202721088435374, - "grad_norm": 0.1435546875, - "learning_rate": 0.00016737241463713135, - "loss": 0.9975, - "step": 1434 - }, - { - "epoch": 0.5209977324263039, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001673234610206829, - "loss": 1.0441, - "step": 1436 - }, - { - "epoch": 0.5217233560090703, - "grad_norm": 0.1669921875, - "learning_rate": 0.00016727450740423449, - "loss": 1.0447, - "step": 1438 - }, - { - "epoch": 0.5224489795918368, - "grad_norm": 0.1845703125, - "learning_rate": 0.0001672255537877861, - "loss": 1.083, - "step": 1440 - }, - { - "epoch": 0.5231746031746032, - "grad_norm": 0.1416015625, - "learning_rate": 0.00016717660017133768, - "loss": 0.9575, - "step": 1442 - }, - { - "epoch": 0.5239002267573696, - "grad_norm": 0.16796875, - "learning_rate": 0.00016712764655488926, - "loss": 1.0298, - "step": 1444 - }, - { - "epoch": 0.5246258503401361, - "grad_norm": 0.166015625, - "learning_rate": 0.00016707869293844084, - "loss": 1.0205, - "step": 1446 - }, - { - "epoch": 0.5253514739229025, - "grad_norm": 0.171875, - "learning_rate": 0.00016702973932199242, - "loss": 0.9916, - "step": 1448 - }, - { - "epoch": 0.5260770975056689, - "grad_norm": 0.1572265625, - "learning_rate": 0.000166980785705544, - "loss": 1.1116, - "step": 1450 - }, - { - "epoch": 0.5268027210884354, - "grad_norm": 0.44140625, - "learning_rate": 0.0001669318320890956, - "loss": 1.072, - "step": 1452 - }, - { - "epoch": 0.5275283446712018, - "grad_norm": 0.1943359375, - "learning_rate": 0.00016688287847264717, - "loss": 0.9555, - "step": 1454 - }, - { - "epoch": 0.5282539682539683, - "grad_norm": 0.2177734375, - "learning_rate": 0.00016683392485619875, - "loss": 0.9393, - "step": 1456 - }, - { - "epoch": 0.5289795918367347, - "grad_norm": 0.162109375, - "learning_rate": 0.00016678497123975033, - "loss": 0.9854, - "step": 1458 - }, - { - "epoch": 0.5297052154195011, - "grad_norm": 0.1708984375, - "learning_rate": 0.00016673601762330194, - "loss": 1.0133, - "step": 1460 - }, - { - "epoch": 0.5304308390022676, - "grad_norm": 0.177734375, - "learning_rate": 0.00016668706400685353, - "loss": 1.0171, - "step": 1462 - }, - { - "epoch": 0.531156462585034, - "grad_norm": 0.1767578125, - "learning_rate": 0.00016663811039040508, - "loss": 0.9813, - "step": 1464 - }, - { - "epoch": 0.5318820861678004, - "grad_norm": 0.166015625, - "learning_rate": 0.0001665891567739567, - "loss": 1.0612, - "step": 1466 - }, - { - "epoch": 0.5326077097505669, - "grad_norm": 0.1669921875, - "learning_rate": 0.00016654020315750827, - "loss": 0.9901, - "step": 1468 - }, - { - "epoch": 0.5333333333333333, - "grad_norm": 0.1513671875, - "learning_rate": 0.00016649124954105986, - "loss": 1.0232, - "step": 1470 - }, - { - "epoch": 0.5340589569160997, - "grad_norm": 0.17578125, - "learning_rate": 0.00016644229592461144, - "loss": 1.0505, - "step": 1472 - }, - { - "epoch": 0.5347845804988662, - "grad_norm": 0.1494140625, - "learning_rate": 0.00016639334230816302, - "loss": 0.9528, - "step": 1474 - }, - { - "epoch": 0.5355102040816326, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001663443886917146, - "loss": 1.0136, - "step": 1476 - }, - { - "epoch": 0.5362358276643991, - "grad_norm": 0.1337890625, - "learning_rate": 0.00016629543507526618, - "loss": 0.9827, - "step": 1478 - }, - { - "epoch": 0.5369614512471655, - "grad_norm": 0.16796875, - "learning_rate": 0.0001662464814588178, - "loss": 0.9911, - "step": 1480 - }, - { - "epoch": 0.5376870748299319, - "grad_norm": 0.16015625, - "learning_rate": 0.00016619752784236938, - "loss": 0.9756, - "step": 1482 - }, - { - "epoch": 0.5384126984126985, - "grad_norm": 0.162109375, - "learning_rate": 0.00016614857422592093, - "loss": 0.9821, - "step": 1484 - }, - { - "epoch": 0.5391383219954649, - "grad_norm": 0.1591796875, - "learning_rate": 0.00016609962060947254, - "loss": 0.9637, - "step": 1486 - }, - { - "epoch": 0.5398639455782313, - "grad_norm": 0.1591796875, - "learning_rate": 0.00016605066699302412, - "loss": 1.0195, - "step": 1488 - }, - { - "epoch": 0.5405895691609978, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001660017133765757, - "loss": 1.0409, - "step": 1490 - }, - { - "epoch": 0.5413151927437642, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001659527597601273, - "loss": 1.0361, - "step": 1492 - }, - { - "epoch": 0.5420408163265306, - "grad_norm": 0.15625, - "learning_rate": 0.00016590380614367887, - "loss": 1.0197, - "step": 1494 - }, - { - "epoch": 0.5427664399092971, - "grad_norm": 0.177734375, - "learning_rate": 0.00016585485252723045, - "loss": 0.9975, - "step": 1496 - }, - { - "epoch": 0.5434920634920635, - "grad_norm": 0.154296875, - "learning_rate": 0.00016580589891078203, - "loss": 0.9776, - "step": 1498 - }, - { - "epoch": 0.54421768707483, - "grad_norm": 0.1474609375, - "learning_rate": 0.00016575694529433364, - "loss": 0.9366, - "step": 1500 - }, - { - "epoch": 0.54421768707483, - "eval_loss": 1.0217424631118774, - "eval_runtime": 92.5726, - "eval_samples_per_second": 19.444, - "eval_steps_per_second": 0.616, - "step": 1500 - }, - { - "epoch": 0.5449433106575964, - "grad_norm": 0.1689453125, - "learning_rate": 0.00016570799167788523, - "loss": 1.0161, - "step": 1502 - }, - { - "epoch": 0.5456689342403628, - "grad_norm": 0.1640625, - "learning_rate": 0.00016565903806143678, - "loss": 1.07, - "step": 1504 - }, - { - "epoch": 0.5463945578231293, - "grad_norm": 0.138671875, - "learning_rate": 0.0001656100844449884, - "loss": 0.8992, - "step": 1506 - }, - { - "epoch": 0.5471201814058957, - "grad_norm": 0.1396484375, - "learning_rate": 0.00016556113082853997, - "loss": 1.123, - "step": 1508 - }, - { - "epoch": 0.5478458049886621, - "grad_norm": 0.1533203125, - "learning_rate": 0.00016551217721209156, - "loss": 1.1036, - "step": 1510 - }, - { - "epoch": 0.5485714285714286, - "grad_norm": 0.1630859375, - "learning_rate": 0.00016546322359564314, - "loss": 0.8912, - "step": 1512 - }, - { - "epoch": 0.549297052154195, - "grad_norm": 0.1552734375, - "learning_rate": 0.00016541426997919472, - "loss": 0.9678, - "step": 1514 - }, - { - "epoch": 0.5500226757369614, - "grad_norm": 0.15234375, - "learning_rate": 0.0001653653163627463, - "loss": 0.9695, - "step": 1516 - }, - { - "epoch": 0.5507482993197279, - "grad_norm": 0.1474609375, - "learning_rate": 0.00016531636274629788, - "loss": 0.9806, - "step": 1518 - }, - { - "epoch": 0.5514739229024943, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001652674091298495, - "loss": 1.0622, - "step": 1520 - }, - { - "epoch": 0.5521995464852608, - "grad_norm": 0.1435546875, - "learning_rate": 0.00016521845551340108, - "loss": 0.9587, - "step": 1522 - }, - { - "epoch": 0.5529251700680272, - "grad_norm": 0.1728515625, - "learning_rate": 0.00016516950189695263, - "loss": 0.9624, - "step": 1524 - }, - { - "epoch": 0.5536507936507936, - "grad_norm": 0.1435546875, - "learning_rate": 0.00016512054828050424, - "loss": 0.9404, - "step": 1526 - }, - { - "epoch": 0.5543764172335601, - "grad_norm": 0.1650390625, - "learning_rate": 0.00016507159466405582, - "loss": 1.0409, - "step": 1528 - }, - { - "epoch": 0.5551020408163265, - "grad_norm": 0.15625, - "learning_rate": 0.0001650226410476074, - "loss": 1.0547, - "step": 1530 - }, - { - "epoch": 0.5558276643990929, - "grad_norm": 0.150390625, - "learning_rate": 0.00016497368743115896, - "loss": 0.9696, - "step": 1532 - }, - { - "epoch": 0.5565532879818594, - "grad_norm": 0.1435546875, - "learning_rate": 0.00016492473381471057, - "loss": 1.032, - "step": 1534 - }, - { - "epoch": 0.5572789115646258, - "grad_norm": 0.1484375, - "learning_rate": 0.00016487578019826215, - "loss": 1.0726, - "step": 1536 - }, - { - "epoch": 0.5580045351473923, - "grad_norm": 0.1552734375, - "learning_rate": 0.00016482682658181373, - "loss": 1.0213, - "step": 1538 - }, - { - "epoch": 0.5587301587301587, - "grad_norm": 0.1337890625, - "learning_rate": 0.00016477787296536534, - "loss": 0.9804, - "step": 1540 - }, - { - "epoch": 0.5594557823129251, - "grad_norm": 0.162109375, - "learning_rate": 0.0001647289193489169, - "loss": 1.0306, - "step": 1542 - }, - { - "epoch": 0.5601814058956917, - "grad_norm": 0.1650390625, - "learning_rate": 0.00016467996573246848, - "loss": 1.0537, - "step": 1544 - }, - { - "epoch": 0.560907029478458, - "grad_norm": 0.146484375, - "learning_rate": 0.0001646310121160201, - "loss": 0.928, - "step": 1546 - }, - { - "epoch": 0.5616326530612245, - "grad_norm": 0.1572265625, - "learning_rate": 0.00016458205849957167, - "loss": 1.0409, - "step": 1548 - }, - { - "epoch": 0.562358276643991, - "grad_norm": 0.154296875, - "learning_rate": 0.00016453310488312326, - "loss": 0.966, - "step": 1550 - }, - { - "epoch": 0.5630839002267574, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001644841512666748, - "loss": 1.0074, - "step": 1552 - }, - { - "epoch": 0.5638095238095238, - "grad_norm": 0.166015625, - "learning_rate": 0.00016443519765022642, - "loss": 0.9354, - "step": 1554 - }, - { - "epoch": 0.5645351473922903, - "grad_norm": 0.154296875, - "learning_rate": 0.000164386244033778, - "loss": 1.0507, - "step": 1556 - }, - { - "epoch": 0.5652607709750567, - "grad_norm": 0.166015625, - "learning_rate": 0.00016433729041732958, - "loss": 1.0014, - "step": 1558 - }, - { - "epoch": 0.5659863945578232, - "grad_norm": 0.1357421875, - "learning_rate": 0.0001642883368008812, - "loss": 0.9745, - "step": 1560 - }, - { - "epoch": 0.5667120181405896, - "grad_norm": 0.1533203125, - "learning_rate": 0.00016423938318443275, - "loss": 0.9293, - "step": 1562 - }, - { - "epoch": 0.567437641723356, - "grad_norm": 0.1611328125, - "learning_rate": 0.00016419042956798433, - "loss": 1.1054, - "step": 1564 - }, - { - "epoch": 0.5681632653061225, - "grad_norm": 0.150390625, - "learning_rate": 0.00016414147595153594, - "loss": 0.9644, - "step": 1566 - }, - { - "epoch": 0.5688888888888889, - "grad_norm": 0.171875, - "learning_rate": 0.00016409252233508752, - "loss": 1.0018, - "step": 1568 - }, - { - "epoch": 0.5696145124716553, - "grad_norm": 0.19140625, - "learning_rate": 0.0001640435687186391, - "loss": 1.0597, - "step": 1570 - }, - { - "epoch": 0.5703401360544218, - "grad_norm": 0.1650390625, - "learning_rate": 0.00016399461510219066, - "loss": 1.0581, - "step": 1572 - }, - { - "epoch": 0.5710657596371882, - "grad_norm": 0.1416015625, - "learning_rate": 0.00016394566148574227, - "loss": 1.0389, - "step": 1574 - }, - { - "epoch": 0.5717913832199546, - "grad_norm": 0.1640625, - "learning_rate": 0.00016389670786929385, - "loss": 0.9944, - "step": 1576 - }, - { - "epoch": 0.5725170068027211, - "grad_norm": 0.1435546875, - "learning_rate": 0.00016384775425284543, - "loss": 1.0125, - "step": 1578 - }, - { - "epoch": 0.5732426303854875, - "grad_norm": 0.169921875, - "learning_rate": 0.00016379880063639704, - "loss": 1.0272, - "step": 1580 - }, - { - "epoch": 0.573968253968254, - "grad_norm": 0.16796875, - "learning_rate": 0.0001637498470199486, - "loss": 1.0639, - "step": 1582 - }, - { - "epoch": 0.5746938775510204, - "grad_norm": 0.1650390625, - "learning_rate": 0.00016370089340350018, - "loss": 0.9577, - "step": 1584 - }, - { - "epoch": 0.5754195011337868, - "grad_norm": 0.154296875, - "learning_rate": 0.0001636519397870518, - "loss": 1.0295, - "step": 1586 - }, - { - "epoch": 0.5761451247165533, - "grad_norm": 0.1640625, - "learning_rate": 0.00016360298617060337, - "loss": 1.0028, - "step": 1588 - }, - { - "epoch": 0.5768707482993197, - "grad_norm": 0.1669921875, - "learning_rate": 0.00016355403255415496, - "loss": 1.0054, - "step": 1590 - }, - { - "epoch": 0.5775963718820861, - "grad_norm": 0.162109375, - "learning_rate": 0.00016350507893770654, - "loss": 0.9994, - "step": 1592 - }, - { - "epoch": 0.5783219954648526, - "grad_norm": 0.1435546875, - "learning_rate": 0.00016345612532125812, - "loss": 0.932, - "step": 1594 - }, - { - "epoch": 0.579047619047619, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001634071717048097, - "loss": 0.9427, - "step": 1596 - }, - { - "epoch": 0.5797732426303854, - "grad_norm": 0.1455078125, - "learning_rate": 0.00016335821808836128, - "loss": 0.9373, - "step": 1598 - }, - { - "epoch": 0.5804988662131519, - "grad_norm": 0.16015625, - "learning_rate": 0.00016330926447191287, - "loss": 1.0353, - "step": 1600 - }, - { - "epoch": 0.5804988662131519, - "eval_loss": 1.0210474729537964, - "eval_runtime": 93.6809, - "eval_samples_per_second": 19.214, - "eval_steps_per_second": 0.608, - "step": 1600 - }, - { - "epoch": 0.5812244897959183, - "grad_norm": 0.1650390625, - "learning_rate": 0.00016326031085546445, - "loss": 0.9768, - "step": 1602 - }, - { - "epoch": 0.5819501133786849, - "grad_norm": 0.1748046875, - "learning_rate": 0.00016321135723901603, - "loss": 0.9428, - "step": 1604 - }, - { - "epoch": 0.5826757369614513, - "grad_norm": 0.1640625, - "learning_rate": 0.00016316240362256764, - "loss": 1.0214, - "step": 1606 - }, - { - "epoch": 0.5834013605442177, - "grad_norm": 0.1513671875, - "learning_rate": 0.00016311345000611922, - "loss": 0.92, - "step": 1608 - }, - { - "epoch": 0.5841269841269842, - "grad_norm": 0.18359375, - "learning_rate": 0.00016306449638967078, - "loss": 1.0035, - "step": 1610 - }, - { - "epoch": 0.5848526077097506, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001630155427732224, - "loss": 1.0347, - "step": 1612 - }, - { - "epoch": 0.585578231292517, - "grad_norm": 0.185546875, - "learning_rate": 0.00016296658915677397, - "loss": 1.0052, - "step": 1614 - }, - { - "epoch": 0.5863038548752835, - "grad_norm": 0.1435546875, - "learning_rate": 0.00016291763554032555, - "loss": 0.9912, - "step": 1616 - }, - { - "epoch": 0.5870294784580499, - "grad_norm": 0.1669921875, - "learning_rate": 0.00016286868192387713, - "loss": 0.9186, - "step": 1618 - }, - { - "epoch": 0.5877551020408164, - "grad_norm": 0.15234375, - "learning_rate": 0.00016281972830742872, - "loss": 1.0071, - "step": 1620 - }, - { - "epoch": 0.5884807256235828, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001627707746909803, - "loss": 0.9302, - "step": 1622 - }, - { - "epoch": 0.5892063492063492, - "grad_norm": 0.1640625, - "learning_rate": 0.00016272182107453188, - "loss": 1.0591, - "step": 1624 - }, - { - "epoch": 0.5899319727891157, - "grad_norm": 0.173828125, - "learning_rate": 0.0001626728674580835, - "loss": 0.9493, - "step": 1626 - }, - { - "epoch": 0.5906575963718821, - "grad_norm": 0.1904296875, - "learning_rate": 0.00016262391384163507, - "loss": 1.0083, - "step": 1628 - }, - { - "epoch": 0.5913832199546485, - "grad_norm": 0.16796875, - "learning_rate": 0.00016257496022518663, - "loss": 1.0483, - "step": 1630 - }, - { - "epoch": 0.592108843537415, - "grad_norm": 0.1533203125, - "learning_rate": 0.00016252600660873824, - "loss": 1.0442, - "step": 1632 - }, - { - "epoch": 0.5928344671201814, - "grad_norm": 0.1513671875, - "learning_rate": 0.00016247705299228982, - "loss": 1.0041, - "step": 1634 - }, - { - "epoch": 0.5935600907029478, - "grad_norm": 0.1396484375, - "learning_rate": 0.0001624280993758414, - "loss": 0.8814, - "step": 1636 - }, - { - "epoch": 0.5942857142857143, - "grad_norm": 0.154296875, - "learning_rate": 0.00016237914575939298, - "loss": 0.9601, - "step": 1638 - }, - { - "epoch": 0.5950113378684807, - "grad_norm": 0.15234375, - "learning_rate": 0.00016233019214294457, - "loss": 1.0641, - "step": 1640 - }, - { - "epoch": 0.5957369614512472, - "grad_norm": 0.1611328125, - "learning_rate": 0.00016228123852649615, - "loss": 0.928, - "step": 1642 - }, - { - "epoch": 0.5964625850340136, - "grad_norm": 0.1484375, - "learning_rate": 0.00016223228491004773, - "loss": 0.9936, - "step": 1644 - }, - { - "epoch": 0.59718820861678, - "grad_norm": 0.177734375, - "learning_rate": 0.00016218333129359934, - "loss": 1.0311, - "step": 1646 - }, - { - "epoch": 0.5979138321995465, - "grad_norm": 0.134765625, - "learning_rate": 0.00016213437767715092, - "loss": 1.0138, - "step": 1648 - }, - { - "epoch": 0.5986394557823129, - "grad_norm": 0.1337890625, - "learning_rate": 0.00016208542406070248, - "loss": 0.9511, - "step": 1650 - }, - { - "epoch": 0.5993650793650793, - "grad_norm": 0.142578125, - "learning_rate": 0.0001620364704442541, - "loss": 0.9711, - "step": 1652 - }, - { - "epoch": 0.6000907029478458, - "grad_norm": 0.1904296875, - "learning_rate": 0.00016198751682780567, - "loss": 0.9674, - "step": 1654 - }, - { - "epoch": 0.6008163265306122, - "grad_norm": 0.1455078125, - "learning_rate": 0.00016193856321135725, - "loss": 1.0381, - "step": 1656 - }, - { - "epoch": 0.6015419501133786, - "grad_norm": 0.1591796875, - "learning_rate": 0.00016188960959490883, - "loss": 0.9523, - "step": 1658 - }, - { - "epoch": 0.6022675736961451, - "grad_norm": 0.166015625, - "learning_rate": 0.00016184065597846042, - "loss": 0.9676, - "step": 1660 - }, - { - "epoch": 0.6029931972789115, - "grad_norm": 0.17578125, - "learning_rate": 0.000161791702362012, - "loss": 0.9896, - "step": 1662 - }, - { - "epoch": 0.603718820861678, - "grad_norm": 0.16015625, - "learning_rate": 0.00016174274874556358, - "loss": 0.9777, - "step": 1664 - }, - { - "epoch": 0.6044444444444445, - "grad_norm": 0.15625, - "learning_rate": 0.0001616937951291152, - "loss": 0.9839, - "step": 1666 - }, - { - "epoch": 0.6051700680272109, - "grad_norm": 0.1396484375, - "learning_rate": 0.00016164484151266677, - "loss": 0.9354, - "step": 1668 - }, - { - "epoch": 0.6058956916099774, - "grad_norm": 0.173828125, - "learning_rate": 0.00016159588789621833, - "loss": 0.9798, - "step": 1670 - }, - { - "epoch": 0.6066213151927438, - "grad_norm": 0.171875, - "learning_rate": 0.00016154693427976994, - "loss": 0.99, - "step": 1672 - }, - { - "epoch": 0.6073469387755102, - "grad_norm": 0.169921875, - "learning_rate": 0.00016149798066332152, - "loss": 1.0575, - "step": 1674 - }, - { - "epoch": 0.6080725623582767, - "grad_norm": 0.1904296875, - "learning_rate": 0.0001614490270468731, - "loss": 1.0201, - "step": 1676 - }, - { - "epoch": 0.6087981859410431, - "grad_norm": 0.162109375, - "learning_rate": 0.00016140007343042466, - "loss": 0.9888, - "step": 1678 - }, - { - "epoch": 0.6095238095238096, - "grad_norm": 0.1650390625, - "learning_rate": 0.00016135111981397627, - "loss": 0.9927, - "step": 1680 - }, - { - "epoch": 0.610249433106576, - "grad_norm": 0.1455078125, - "learning_rate": 0.00016130216619752785, - "loss": 1.0186, - "step": 1682 - }, - { - "epoch": 0.6109750566893424, - "grad_norm": 0.162109375, - "learning_rate": 0.00016125321258107943, - "loss": 1.0128, - "step": 1684 - }, - { - "epoch": 0.6117006802721089, - "grad_norm": 0.166015625, - "learning_rate": 0.00016120425896463104, - "loss": 1.0212, - "step": 1686 - }, - { - "epoch": 0.6124263038548753, - "grad_norm": 0.15625, - "learning_rate": 0.0001611553053481826, - "loss": 0.9624, - "step": 1688 - }, - { - "epoch": 0.6131519274376417, - "grad_norm": 0.1416015625, - "learning_rate": 0.00016110635173173418, - "loss": 1.0341, - "step": 1690 - }, - { - "epoch": 0.6138775510204082, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001610573981152858, - "loss": 0.9915, - "step": 1692 - }, - { - "epoch": 0.6146031746031746, - "grad_norm": 0.1591796875, - "learning_rate": 0.00016100844449883737, - "loss": 1.0157, - "step": 1694 - }, - { - "epoch": 0.615328798185941, - "grad_norm": 0.1474609375, - "learning_rate": 0.00016095949088238895, - "loss": 1.0553, - "step": 1696 - }, - { - "epoch": 0.6160544217687075, - "grad_norm": 0.185546875, - "learning_rate": 0.0001609105372659405, - "loss": 0.9294, - "step": 1698 - }, - { - "epoch": 0.6167800453514739, - "grad_norm": 0.1748046875, - "learning_rate": 0.00016086158364949212, - "loss": 0.9272, - "step": 1700 - }, - { - "epoch": 0.6167800453514739, - "eval_loss": 1.0209547281265259, - "eval_runtime": 91.8934, - "eval_samples_per_second": 19.588, - "eval_steps_per_second": 0.62, - "step": 1700 - }, - { - "epoch": 0.6175056689342404, - "grad_norm": 0.158203125, - "learning_rate": 0.0001608126300330437, - "loss": 0.976, - "step": 1702 - }, - { - "epoch": 0.6182312925170068, - "grad_norm": 0.1435546875, - "learning_rate": 0.00016076367641659528, - "loss": 0.9733, - "step": 1704 - }, - { - "epoch": 0.6189569160997732, - "grad_norm": 0.150390625, - "learning_rate": 0.0001607147228001469, - "loss": 1.0233, - "step": 1706 - }, - { - "epoch": 0.6196825396825397, - "grad_norm": 0.158203125, - "learning_rate": 0.00016066576918369845, - "loss": 0.9805, - "step": 1708 - }, - { - "epoch": 0.6204081632653061, - "grad_norm": 0.146484375, - "learning_rate": 0.00016061681556725003, - "loss": 0.9321, - "step": 1710 - }, - { - "epoch": 0.6211337868480725, - "grad_norm": 0.1875, - "learning_rate": 0.00016056786195080164, - "loss": 1.0678, - "step": 1712 - }, - { - "epoch": 0.621859410430839, - "grad_norm": 0.1533203125, - "learning_rate": 0.00016051890833435322, - "loss": 1.0146, - "step": 1714 - }, - { - "epoch": 0.6225850340136054, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001604699547179048, - "loss": 0.8791, - "step": 1716 - }, - { - "epoch": 0.6233106575963718, - "grad_norm": 0.150390625, - "learning_rate": 0.00016042100110145636, - "loss": 0.9568, - "step": 1718 - }, - { - "epoch": 0.6240362811791383, - "grad_norm": 0.1484375, - "learning_rate": 0.00016037204748500797, - "loss": 0.9577, - "step": 1720 - }, - { - "epoch": 0.6247619047619047, - "grad_norm": 0.140625, - "learning_rate": 0.00016032309386855955, - "loss": 0.9506, - "step": 1722 - }, - { - "epoch": 0.6254875283446713, - "grad_norm": 0.1455078125, - "learning_rate": 0.00016027414025211113, - "loss": 0.9523, - "step": 1724 - }, - { - "epoch": 0.6262131519274377, - "grad_norm": 0.1484375, - "learning_rate": 0.00016022518663566274, - "loss": 0.997, - "step": 1726 - }, - { - "epoch": 0.626938775510204, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001601762330192143, - "loss": 0.9419, - "step": 1728 - }, - { - "epoch": 0.6276643990929706, - "grad_norm": 0.1669921875, - "learning_rate": 0.00016012727940276588, - "loss": 0.9835, - "step": 1730 - }, - { - "epoch": 0.628390022675737, - "grad_norm": 0.16015625, - "learning_rate": 0.0001600783257863175, - "loss": 0.9472, - "step": 1732 - }, - { - "epoch": 0.6291156462585034, - "grad_norm": 0.177734375, - "learning_rate": 0.00016002937216986907, - "loss": 1.0306, - "step": 1734 - }, - { - "epoch": 0.6298412698412699, - "grad_norm": 0.166015625, - "learning_rate": 0.00015998041855342065, - "loss": 1.0021, - "step": 1736 - }, - { - "epoch": 0.6305668934240363, - "grad_norm": 0.1376953125, - "learning_rate": 0.0001599314649369722, - "loss": 0.9455, - "step": 1738 - }, - { - "epoch": 0.6312925170068027, - "grad_norm": 0.173828125, - "learning_rate": 0.00015988251132052382, - "loss": 1.0614, - "step": 1740 - }, - { - "epoch": 0.6320181405895692, - "grad_norm": 0.15234375, - "learning_rate": 0.0001598335577040754, - "loss": 0.9934, - "step": 1742 - }, - { - "epoch": 0.6327437641723356, - "grad_norm": 0.1435546875, - "learning_rate": 0.00015978460408762698, - "loss": 1.0372, - "step": 1744 - }, - { - "epoch": 0.6334693877551021, - "grad_norm": 0.14453125, - "learning_rate": 0.00015973565047117856, - "loss": 1.0382, - "step": 1746 - }, - { - "epoch": 0.6341950113378685, - "grad_norm": 0.1650390625, - "learning_rate": 0.00015968669685473015, - "loss": 1.001, - "step": 1748 - }, - { - "epoch": 0.6349206349206349, - "grad_norm": 0.1455078125, - "learning_rate": 0.00015963774323828173, - "loss": 0.9544, - "step": 1750 - }, - { - "epoch": 0.6356462585034014, - "grad_norm": 0.140625, - "learning_rate": 0.00015958878962183334, - "loss": 0.9517, - "step": 1752 - }, - { - "epoch": 0.6363718820861678, - "grad_norm": 0.169921875, - "learning_rate": 0.00015953983600538492, - "loss": 0.9706, - "step": 1754 - }, - { - "epoch": 0.6370975056689342, - "grad_norm": 0.1376953125, - "learning_rate": 0.00015949088238893647, - "loss": 1.0153, - "step": 1756 - }, - { - "epoch": 0.6378231292517007, - "grad_norm": 0.146484375, - "learning_rate": 0.00015944192877248806, - "loss": 0.9559, - "step": 1758 - }, - { - "epoch": 0.6385487528344671, - "grad_norm": 0.15234375, - "learning_rate": 0.00015939297515603967, - "loss": 0.9877, - "step": 1760 - }, - { - "epoch": 0.6392743764172336, - "grad_norm": 0.16015625, - "learning_rate": 0.00015934402153959125, - "loss": 0.9975, - "step": 1762 - }, - { - "epoch": 0.64, - "grad_norm": 0.1640625, - "learning_rate": 0.00015929506792314283, - "loss": 0.9666, - "step": 1764 - }, - { - "epoch": 0.6407256235827664, - "grad_norm": 0.15234375, - "learning_rate": 0.0001592461143066944, - "loss": 1.028, - "step": 1766 - }, - { - "epoch": 0.6414512471655329, - "grad_norm": 0.1533203125, - "learning_rate": 0.000159197160690246, - "loss": 1.0233, - "step": 1768 - }, - { - "epoch": 0.6421768707482993, - "grad_norm": 0.1513671875, - "learning_rate": 0.00015914820707379758, - "loss": 1.017, - "step": 1770 - }, - { - "epoch": 0.6429024943310657, - "grad_norm": 0.14453125, - "learning_rate": 0.0001590992534573492, - "loss": 1.0094, - "step": 1772 - }, - { - "epoch": 0.6436281179138322, - "grad_norm": 0.1689453125, - "learning_rate": 0.00015905029984090077, - "loss": 0.9779, - "step": 1774 - }, - { - "epoch": 0.6443537414965986, - "grad_norm": 0.1806640625, - "learning_rate": 0.00015900134622445232, - "loss": 0.9917, - "step": 1776 - }, - { - "epoch": 0.645079365079365, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001589523926080039, - "loss": 0.9526, - "step": 1778 - }, - { - "epoch": 0.6458049886621315, - "grad_norm": 0.158203125, - "learning_rate": 0.00015890343899155552, - "loss": 1.0131, - "step": 1780 - }, - { - "epoch": 0.6465306122448979, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001588544853751071, - "loss": 1.053, - "step": 1782 - }, - { - "epoch": 0.6472562358276645, - "grad_norm": 0.1484375, - "learning_rate": 0.00015880553175865868, - "loss": 0.9761, - "step": 1784 - }, - { - "epoch": 0.6479818594104308, - "grad_norm": 0.1640625, - "learning_rate": 0.00015875657814221026, - "loss": 0.9547, - "step": 1786 - }, - { - "epoch": 0.6487074829931972, - "grad_norm": 0.1533203125, - "learning_rate": 0.00015870762452576184, - "loss": 1.0081, - "step": 1788 - }, - { - "epoch": 0.6494331065759638, - "grad_norm": 0.1826171875, - "learning_rate": 0.00015865867090931343, - "loss": 1.007, - "step": 1790 - }, - { - "epoch": 0.6501587301587302, - "grad_norm": 0.1513671875, - "learning_rate": 0.00015860971729286504, - "loss": 1.0005, - "step": 1792 - }, - { - "epoch": 0.6508843537414966, - "grad_norm": 0.1298828125, - "learning_rate": 0.00015856076367641662, - "loss": 0.9823, - "step": 1794 - }, - { - "epoch": 0.6516099773242631, - "grad_norm": 0.1396484375, - "learning_rate": 0.00015851181005996817, - "loss": 0.9324, - "step": 1796 - }, - { - "epoch": 0.6523356009070295, - "grad_norm": 0.1640625, - "learning_rate": 0.00015846285644351976, - "loss": 1.0323, - "step": 1798 - }, - { - "epoch": 0.6530612244897959, - "grad_norm": 0.1552734375, - "learning_rate": 0.00015841390282707137, - "loss": 0.942, - "step": 1800 - }, - { - "epoch": 0.6530612244897959, - "eval_loss": 1.0201084613800049, - "eval_runtime": 94.9174, - "eval_samples_per_second": 18.964, - "eval_steps_per_second": 0.601, - "step": 1800 - }, - { - "epoch": 0.6537868480725624, - "grad_norm": 0.162109375, - "learning_rate": 0.00015836494921062295, - "loss": 0.9913, - "step": 1802 - }, - { - "epoch": 0.6545124716553288, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015831599559417453, - "loss": 1.098, - "step": 1804 - }, - { - "epoch": 0.6552380952380953, - "grad_norm": 0.173828125, - "learning_rate": 0.0001582670419777261, - "loss": 1.0758, - "step": 1806 - }, - { - "epoch": 0.6559637188208617, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001582180883612777, - "loss": 0.9549, - "step": 1808 - }, - { - "epoch": 0.6566893424036281, - "grad_norm": 0.171875, - "learning_rate": 0.00015816913474482928, - "loss": 1.0207, - "step": 1810 - }, - { - "epoch": 0.6574149659863946, - "grad_norm": 0.158203125, - "learning_rate": 0.0001581201811283809, - "loss": 0.9206, - "step": 1812 - }, - { - "epoch": 0.658140589569161, - "grad_norm": 0.1572265625, - "learning_rate": 0.00015807122751193247, - "loss": 0.9818, - "step": 1814 - }, - { - "epoch": 0.6588662131519274, - "grad_norm": 0.1552734375, - "learning_rate": 0.00015802227389548402, - "loss": 1.0962, - "step": 1816 - }, - { - "epoch": 0.6595918367346939, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001579733202790356, - "loss": 0.9518, - "step": 1818 - }, - { - "epoch": 0.6603174603174603, - "grad_norm": 0.154296875, - "learning_rate": 0.00015792436666258722, - "loss": 0.9924, - "step": 1820 - }, - { - "epoch": 0.6610430839002267, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001578754130461388, - "loss": 0.9963, - "step": 1822 - }, - { - "epoch": 0.6617687074829932, - "grad_norm": 0.140625, - "learning_rate": 0.00015782645942969035, - "loss": 0.983, - "step": 1824 - }, - { - "epoch": 0.6624943310657596, - "grad_norm": 0.162109375, - "learning_rate": 0.00015777750581324196, - "loss": 1.0437, - "step": 1826 - }, - { - "epoch": 0.6632199546485261, - "grad_norm": 0.1689453125, - "learning_rate": 0.00015772855219679354, - "loss": 1.0547, - "step": 1828 - }, - { - "epoch": 0.6639455782312925, - "grad_norm": 0.1572265625, - "learning_rate": 0.00015767959858034513, - "loss": 0.9898, - "step": 1830 - }, - { - "epoch": 0.6646712018140589, - "grad_norm": 0.1435546875, - "learning_rate": 0.00015763064496389674, - "loss": 1.0343, - "step": 1832 - }, - { - "epoch": 0.6653968253968254, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001575816913474483, - "loss": 1.0069, - "step": 1834 - }, - { - "epoch": 0.6661224489795918, - "grad_norm": 0.1611328125, - "learning_rate": 0.00015753273773099987, - "loss": 1.032, - "step": 1836 - }, - { - "epoch": 0.6668480725623582, - "grad_norm": 0.15625, - "learning_rate": 0.00015748378411455146, - "loss": 0.9829, - "step": 1838 - }, - { - "epoch": 0.6675736961451247, - "grad_norm": 0.14453125, - "learning_rate": 0.00015743483049810307, - "loss": 1.0174, - "step": 1840 - }, - { - "epoch": 0.6682993197278911, - "grad_norm": 0.142578125, - "learning_rate": 0.00015738587688165465, - "loss": 1.0235, - "step": 1842 - }, - { - "epoch": 0.6690249433106576, - "grad_norm": 0.154296875, - "learning_rate": 0.0001573369232652062, - "loss": 0.9988, - "step": 1844 - }, - { - "epoch": 0.669750566893424, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001572879696487578, - "loss": 0.9146, - "step": 1846 - }, - { - "epoch": 0.6704761904761904, - "grad_norm": 0.158203125, - "learning_rate": 0.0001572390160323094, - "loss": 1.0098, - "step": 1848 - }, - { - "epoch": 0.671201814058957, - "grad_norm": 0.16796875, - "learning_rate": 0.00015719006241586098, - "loss": 0.9822, - "step": 1850 - }, - { - "epoch": 0.6719274376417234, - "grad_norm": 0.1484375, - "learning_rate": 0.00015714110879941259, - "loss": 1.0375, - "step": 1852 - }, - { - "epoch": 0.6726530612244898, - "grad_norm": 0.171875, - "learning_rate": 0.00015709215518296414, - "loss": 1.0528, - "step": 1854 - }, - { - "epoch": 0.6733786848072563, - "grad_norm": 0.1484375, - "learning_rate": 0.00015704320156651572, - "loss": 0.9579, - "step": 1856 - }, - { - "epoch": 0.6741043083900227, - "grad_norm": 0.16796875, - "learning_rate": 0.0001569942479500673, - "loss": 1.0379, - "step": 1858 - }, - { - "epoch": 0.6748299319727891, - "grad_norm": 0.1611328125, - "learning_rate": 0.00015694529433361892, - "loss": 1.0233, - "step": 1860 - }, - { - "epoch": 0.6755555555555556, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001568963407171705, - "loss": 0.9767, - "step": 1862 - }, - { - "epoch": 0.676281179138322, - "grad_norm": 0.173828125, - "learning_rate": 0.00015684738710072205, - "loss": 0.9701, - "step": 1864 - }, - { - "epoch": 0.6770068027210885, - "grad_norm": 0.162109375, - "learning_rate": 0.00015679843348427366, - "loss": 0.9249, - "step": 1866 - }, - { - "epoch": 0.6777324263038549, - "grad_norm": 0.150390625, - "learning_rate": 0.00015674947986782524, - "loss": 0.9929, - "step": 1868 - }, - { - "epoch": 0.6784580498866213, - "grad_norm": 0.1328125, - "learning_rate": 0.00015670052625137683, - "loss": 0.937, - "step": 1870 - }, - { - "epoch": 0.6791836734693878, - "grad_norm": 0.1455078125, - "learning_rate": 0.00015665157263492844, - "loss": 0.9915, - "step": 1872 - }, - { - "epoch": 0.6799092970521542, - "grad_norm": 0.1533203125, - "learning_rate": 0.00015660261901848, - "loss": 1.0443, - "step": 1874 - }, - { - "epoch": 0.6806349206349206, - "grad_norm": 0.14453125, - "learning_rate": 0.00015655366540203157, - "loss": 1.051, - "step": 1876 - }, - { - "epoch": 0.6813605442176871, - "grad_norm": 0.1337890625, - "learning_rate": 0.00015650471178558316, - "loss": 0.9451, - "step": 1878 - }, - { - "epoch": 0.6820861678004535, - "grad_norm": 0.1591796875, - "learning_rate": 0.00015645575816913477, - "loss": 0.9998, - "step": 1880 - }, - { - "epoch": 0.6828117913832199, - "grad_norm": 0.150390625, - "learning_rate": 0.00015640680455268635, - "loss": 0.9396, - "step": 1882 - }, - { - "epoch": 0.6835374149659864, - "grad_norm": 0.158203125, - "learning_rate": 0.0001563578509362379, - "loss": 1.0181, - "step": 1884 - }, - { - "epoch": 0.6842630385487528, - "grad_norm": 0.158203125, - "learning_rate": 0.0001563088973197895, - "loss": 1.0061, - "step": 1886 - }, - { - "epoch": 0.6849886621315193, - "grad_norm": 0.1484375, - "learning_rate": 0.0001562599437033411, - "loss": 0.9953, - "step": 1888 - }, - { - "epoch": 0.6857142857142857, - "grad_norm": 0.15625, - "learning_rate": 0.00015621099008689268, - "loss": 1.0486, - "step": 1890 - }, - { - "epoch": 0.6864399092970521, - "grad_norm": 0.1806640625, - "learning_rate": 0.00015616203647044426, - "loss": 0.9057, - "step": 1892 - }, - { - "epoch": 0.6871655328798186, - "grad_norm": 0.158203125, - "learning_rate": 0.00015611308285399584, - "loss": 0.9876, - "step": 1894 - }, - { - "epoch": 0.687891156462585, - "grad_norm": 0.130859375, - "learning_rate": 0.00015606412923754742, - "loss": 0.9982, - "step": 1896 - }, - { - "epoch": 0.6886167800453514, - "grad_norm": 0.173828125, - "learning_rate": 0.000156015175621099, - "loss": 0.9909, - "step": 1898 - }, - { - "epoch": 0.6893424036281179, - "grad_norm": 0.1533203125, - "learning_rate": 0.00015596622200465062, - "loss": 0.9472, - "step": 1900 - }, - { - "epoch": 0.6893424036281179, - "eval_loss": 1.019601821899414, - "eval_runtime": 96.3922, - "eval_samples_per_second": 18.674, - "eval_steps_per_second": 0.591, - "step": 1900 - }, - { - "epoch": 0.6900680272108843, - "grad_norm": 0.1572265625, - "learning_rate": 0.00015591726838820217, - "loss": 0.9808, - "step": 1902 - }, - { - "epoch": 0.6907936507936508, - "grad_norm": 0.1474609375, - "learning_rate": 0.00015586831477175375, - "loss": 1.0067, - "step": 1904 - }, - { - "epoch": 0.6915192743764172, - "grad_norm": 0.1640625, - "learning_rate": 0.00015581936115530536, - "loss": 0.941, - "step": 1906 - }, - { - "epoch": 0.6922448979591836, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015577040753885694, - "loss": 0.944, - "step": 1908 - }, - { - "epoch": 0.6929705215419502, - "grad_norm": 0.15625, - "learning_rate": 0.00015572145392240853, - "loss": 0.9849, - "step": 1910 - }, - { - "epoch": 0.6936961451247166, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001556725003059601, - "loss": 0.9713, - "step": 1912 - }, - { - "epoch": 0.694421768707483, - "grad_norm": 0.16015625, - "learning_rate": 0.0001556235466895117, - "loss": 1.0218, - "step": 1914 - }, - { - "epoch": 0.6951473922902495, - "grad_norm": 0.1640625, - "learning_rate": 0.00015557459307306327, - "loss": 0.9998, - "step": 1916 - }, - { - "epoch": 0.6958730158730159, - "grad_norm": 0.1435546875, - "learning_rate": 0.00015552563945661486, - "loss": 1.0357, - "step": 1918 - }, - { - "epoch": 0.6965986394557823, - "grad_norm": 0.1533203125, - "learning_rate": 0.00015547668584016647, - "loss": 0.9338, - "step": 1920 - }, - { - "epoch": 0.6973242630385488, - "grad_norm": 0.14453125, - "learning_rate": 0.00015542773222371802, - "loss": 1.0691, - "step": 1922 - }, - { - "epoch": 0.6980498866213152, - "grad_norm": 0.146484375, - "learning_rate": 0.0001553787786072696, - "loss": 0.994, - "step": 1924 - }, - { - "epoch": 0.6987755102040817, - "grad_norm": 0.150390625, - "learning_rate": 0.0001553298249908212, - "loss": 0.9554, - "step": 1926 - }, - { - "epoch": 0.6995011337868481, - "grad_norm": 0.19921875, - "learning_rate": 0.0001552808713743728, - "loss": 1.0239, - "step": 1928 - }, - { - "epoch": 0.7002267573696145, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015523191775792438, - "loss": 1.0417, - "step": 1930 - }, - { - "epoch": 0.700952380952381, - "grad_norm": 0.1474609375, - "learning_rate": 0.00015518296414147596, - "loss": 0.9733, - "step": 1932 - }, - { - "epoch": 0.7016780045351474, - "grad_norm": 0.17578125, - "learning_rate": 0.00015513401052502754, - "loss": 1.0096, - "step": 1934 - }, - { - "epoch": 0.7024036281179138, - "grad_norm": 0.1767578125, - "learning_rate": 0.00015508505690857912, - "loss": 0.9957, - "step": 1936 - }, - { - "epoch": 0.7031292517006803, - "grad_norm": 0.1787109375, - "learning_rate": 0.0001550361032921307, - "loss": 0.9878, - "step": 1938 - }, - { - "epoch": 0.7038548752834467, - "grad_norm": 0.1591796875, - "learning_rate": 0.00015498714967568232, - "loss": 0.9317, - "step": 1940 - }, - { - "epoch": 0.7045804988662131, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015493819605923387, - "loss": 1.0407, - "step": 1942 - }, - { - "epoch": 0.7053061224489796, - "grad_norm": 0.15625, - "learning_rate": 0.00015488924244278545, - "loss": 1.0233, - "step": 1944 - }, - { - "epoch": 0.706031746031746, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015484028882633706, - "loss": 0.9254, - "step": 1946 - }, - { - "epoch": 0.7067573696145125, - "grad_norm": 0.220703125, - "learning_rate": 0.00015479133520988864, - "loss": 0.9546, - "step": 1948 - }, - { - "epoch": 0.7074829931972789, - "grad_norm": 0.208984375, - "learning_rate": 0.00015474238159344023, - "loss": 0.9609, - "step": 1950 - }, - { - "epoch": 0.7082086167800453, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001546934279769918, - "loss": 1.0335, - "step": 1952 - }, - { - "epoch": 0.7089342403628118, - "grad_norm": 0.158203125, - "learning_rate": 0.0001546444743605434, - "loss": 0.952, - "step": 1954 - }, - { - "epoch": 0.7096598639455782, - "grad_norm": 0.16015625, - "learning_rate": 0.00015459552074409497, - "loss": 0.9557, - "step": 1956 - }, - { - "epoch": 0.7103854875283446, - "grad_norm": 0.142578125, - "learning_rate": 0.00015454656712764656, - "loss": 0.967, - "step": 1958 - }, - { - "epoch": 0.7111111111111111, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015449761351119816, - "loss": 1.0578, - "step": 1960 - }, - { - "epoch": 0.7118367346938775, - "grad_norm": 0.134765625, - "learning_rate": 0.00015444865989474972, - "loss": 1.0424, - "step": 1962 - }, - { - "epoch": 0.7125623582766439, - "grad_norm": 0.162109375, - "learning_rate": 0.0001543997062783013, - "loss": 1.0075, - "step": 1964 - }, - { - "epoch": 0.7132879818594104, - "grad_norm": 0.162109375, - "learning_rate": 0.0001543507526618529, - "loss": 0.9869, - "step": 1966 - }, - { - "epoch": 0.7140136054421768, - "grad_norm": 0.140625, - "learning_rate": 0.0001543017990454045, - "loss": 0.9439, - "step": 1968 - }, - { - "epoch": 0.7147392290249434, - "grad_norm": 0.1669921875, - "learning_rate": 0.00015425284542895605, - "loss": 1.004, - "step": 1970 - }, - { - "epoch": 0.7154648526077098, - "grad_norm": 0.169921875, - "learning_rate": 0.00015420389181250766, - "loss": 0.9863, - "step": 1972 - }, - { - "epoch": 0.7161904761904762, - "grad_norm": 0.13671875, - "learning_rate": 0.00015415493819605924, - "loss": 0.9668, - "step": 1974 - }, - { - "epoch": 0.7169160997732427, - "grad_norm": 0.2373046875, - "learning_rate": 0.00015410598457961082, - "loss": 0.9949, - "step": 1976 - }, - { - "epoch": 0.7176417233560091, - "grad_norm": 0.1640625, - "learning_rate": 0.0001540570309631624, - "loss": 0.959, - "step": 1978 - }, - { - "epoch": 0.7183673469387755, - "grad_norm": 0.16796875, - "learning_rate": 0.000154008077346714, - "loss": 1.0601, - "step": 1980 - }, - { - "epoch": 0.719092970521542, - "grad_norm": 0.1552734375, - "learning_rate": 0.00015395912373026557, - "loss": 1.0484, - "step": 1982 - }, - { - "epoch": 0.7198185941043084, - "grad_norm": 0.1435546875, - "learning_rate": 0.00015391017011381715, - "loss": 0.9602, - "step": 1984 - }, - { - "epoch": 0.7205442176870749, - "grad_norm": 0.158203125, - "learning_rate": 0.00015386121649736876, - "loss": 0.9339, - "step": 1986 - }, - { - "epoch": 0.7212698412698413, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015381226288092034, - "loss": 1.0522, - "step": 1988 - }, - { - "epoch": 0.7219954648526077, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001537633092644719, - "loss": 0.9165, - "step": 1990 - }, - { - "epoch": 0.7227210884353742, - "grad_norm": 0.146484375, - "learning_rate": 0.0001537143556480235, - "loss": 0.9953, - "step": 1992 - }, - { - "epoch": 0.7234467120181406, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001536654020315751, - "loss": 0.9712, - "step": 1994 - }, - { - "epoch": 0.724172335600907, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015361644841512667, - "loss": 1.0144, - "step": 1996 - }, - { - "epoch": 0.7248979591836735, - "grad_norm": 0.1767578125, - "learning_rate": 0.00015356749479867826, - "loss": 1.0188, - "step": 1998 - }, - { - "epoch": 0.7256235827664399, - "grad_norm": 0.1376953125, - "learning_rate": 0.00015351854118222984, - "loss": 1.0385, - "step": 2000 - }, - { - "epoch": 0.7256235827664399, - "eval_loss": 1.0189679861068726, - "eval_runtime": 100.5257, - "eval_samples_per_second": 17.906, - "eval_steps_per_second": 0.567, - "step": 2000 - }, - { - "epoch": 0.7263492063492063, - "grad_norm": 0.1689453125, - "learning_rate": 0.00015346958756578142, - "loss": 0.9979, - "step": 2002 - }, - { - "epoch": 0.7270748299319728, - "grad_norm": 0.146484375, - "learning_rate": 0.000153420633949333, - "loss": 0.9475, - "step": 2004 - }, - { - "epoch": 0.7278004535147392, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001533716803328846, - "loss": 1.0247, - "step": 2006 - }, - { - "epoch": 0.7285260770975057, - "grad_norm": 0.2890625, - "learning_rate": 0.0001533227267164362, - "loss": 1.0989, - "step": 2008 - }, - { - "epoch": 0.7292517006802721, - "grad_norm": 0.1552734375, - "learning_rate": 0.00015327377309998775, - "loss": 0.9775, - "step": 2010 - }, - { - "epoch": 0.7299773242630385, - "grad_norm": 0.154296875, - "learning_rate": 0.00015322481948353936, - "loss": 1.0218, - "step": 2012 - }, - { - "epoch": 0.730702947845805, - "grad_norm": 0.130859375, - "learning_rate": 0.00015317586586709094, - "loss": 0.943, - "step": 2014 - }, - { - "epoch": 0.7314285714285714, - "grad_norm": 0.15234375, - "learning_rate": 0.00015312691225064252, - "loss": 0.9455, - "step": 2016 - }, - { - "epoch": 0.7321541950113378, - "grad_norm": 0.177734375, - "learning_rate": 0.0001530779586341941, - "loss": 0.9995, - "step": 2018 - }, - { - "epoch": 0.7328798185941043, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001530290050177457, - "loss": 1.0052, - "step": 2020 - }, - { - "epoch": 0.7336054421768707, - "grad_norm": 0.166015625, - "learning_rate": 0.00015298005140129727, - "loss": 1.0352, - "step": 2022 - }, - { - "epoch": 0.7343310657596371, - "grad_norm": 0.1484375, - "learning_rate": 0.00015293109778484885, - "loss": 1.0545, - "step": 2024 - }, - { - "epoch": 0.7350566893424036, - "grad_norm": 0.150390625, - "learning_rate": 0.00015288214416840046, - "loss": 1.0652, - "step": 2026 - }, - { - "epoch": 0.73578231292517, - "grad_norm": 0.1669921875, - "learning_rate": 0.00015283319055195204, - "loss": 1.0658, - "step": 2028 - }, - { - "epoch": 0.7365079365079366, - "grad_norm": 0.162109375, - "learning_rate": 0.0001527842369355036, - "loss": 1.0105, - "step": 2030 - }, - { - "epoch": 0.737233560090703, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001527352833190552, - "loss": 1.0369, - "step": 2032 - }, - { - "epoch": 0.7379591836734694, - "grad_norm": 0.16015625, - "learning_rate": 0.0001526863297026068, - "loss": 1.0044, - "step": 2034 - }, - { - "epoch": 0.7386848072562359, - "grad_norm": 0.154296875, - "learning_rate": 0.00015263737608615837, - "loss": 0.9738, - "step": 2036 - }, - { - "epoch": 0.7394104308390023, - "grad_norm": 0.154296875, - "learning_rate": 0.00015258842246970996, - "loss": 1.0011, - "step": 2038 - }, - { - "epoch": 0.7401360544217687, - "grad_norm": 0.1328125, - "learning_rate": 0.00015253946885326154, - "loss": 0.8985, - "step": 2040 - }, - { - "epoch": 0.7408616780045352, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015249051523681312, - "loss": 1.0394, - "step": 2042 - }, - { - "epoch": 0.7415873015873016, - "grad_norm": 0.15625, - "learning_rate": 0.0001524415616203647, - "loss": 1.02, - "step": 2044 - }, - { - "epoch": 0.742312925170068, - "grad_norm": 0.138671875, - "learning_rate": 0.0001523926080039163, - "loss": 0.9851, - "step": 2046 - }, - { - "epoch": 0.7430385487528345, - "grad_norm": 0.1630859375, - "learning_rate": 0.00015234365438746787, - "loss": 1.0116, - "step": 2048 - }, - { - "epoch": 0.7437641723356009, - "grad_norm": 0.16015625, - "learning_rate": 0.00015229470077101945, - "loss": 1.041, - "step": 2050 - }, - { - "epoch": 0.7444897959183674, - "grad_norm": 0.1650390625, - "learning_rate": 0.00015224574715457106, - "loss": 0.9993, - "step": 2052 - }, - { - "epoch": 0.7452154195011338, - "grad_norm": 0.138671875, - "learning_rate": 0.00015219679353812264, - "loss": 0.9596, - "step": 2054 - }, - { - "epoch": 0.7459410430839002, - "grad_norm": 0.150390625, - "learning_rate": 0.00015214783992167422, - "loss": 1.0042, - "step": 2056 - }, - { - "epoch": 0.7466666666666667, - "grad_norm": 0.158203125, - "learning_rate": 0.0001520988863052258, - "loss": 0.9341, - "step": 2058 - }, - { - "epoch": 0.7473922902494331, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001520499326887774, - "loss": 0.9923, - "step": 2060 - }, - { - "epoch": 0.7481179138321995, - "grad_norm": 0.1435546875, - "learning_rate": 0.00015200097907232897, - "loss": 1.0059, - "step": 2062 - }, - { - "epoch": 0.748843537414966, - "grad_norm": 0.1455078125, - "learning_rate": 0.00015195202545588055, - "loss": 0.9428, - "step": 2064 - }, - { - "epoch": 0.7495691609977324, - "grad_norm": 0.1591796875, - "learning_rate": 0.00015190307183943216, - "loss": 0.9996, - "step": 2066 - }, - { - "epoch": 0.7502947845804989, - "grad_norm": 0.1474609375, - "learning_rate": 0.00015185411822298372, - "loss": 0.9503, - "step": 2068 - }, - { - "epoch": 0.7510204081632653, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001518051646065353, - "loss": 1.0125, - "step": 2070 - }, - { - "epoch": 0.7517460317460317, - "grad_norm": 0.169921875, - "learning_rate": 0.0001517562109900869, - "loss": 0.9977, - "step": 2072 - }, - { - "epoch": 0.7524716553287982, - "grad_norm": 0.146484375, - "learning_rate": 0.0001517072573736385, - "loss": 1.0281, - "step": 2074 - }, - { - "epoch": 0.7531972789115646, - "grad_norm": 0.1435546875, - "learning_rate": 0.00015165830375719007, - "loss": 0.9786, - "step": 2076 - }, - { - "epoch": 0.753922902494331, - "grad_norm": 0.1611328125, - "learning_rate": 0.00015160935014074166, - "loss": 0.9559, - "step": 2078 - }, - { - "epoch": 0.7546485260770975, - "grad_norm": 0.138671875, - "learning_rate": 0.00015156039652429324, - "loss": 0.9641, - "step": 2080 - }, - { - "epoch": 0.7553741496598639, - "grad_norm": 0.1474609375, - "learning_rate": 0.00015151144290784482, - "loss": 0.9441, - "step": 2082 - }, - { - "epoch": 0.7560997732426303, - "grad_norm": 0.1484375, - "learning_rate": 0.0001514624892913964, - "loss": 0.999, - "step": 2084 - }, - { - "epoch": 0.7568253968253968, - "grad_norm": 0.2021484375, - "learning_rate": 0.000151413535674948, - "loss": 1.0128, - "step": 2086 - }, - { - "epoch": 0.7575510204081632, - "grad_norm": 0.1640625, - "learning_rate": 0.00015136458205849957, - "loss": 1.0748, - "step": 2088 - }, - { - "epoch": 0.7582766439909298, - "grad_norm": 0.1328125, - "learning_rate": 0.00015131562844205115, - "loss": 0.8835, - "step": 2090 - }, - { - "epoch": 0.7590022675736962, - "grad_norm": 0.150390625, - "learning_rate": 0.00015126667482560276, - "loss": 0.9798, - "step": 2092 - }, - { - "epoch": 0.7597278911564626, - "grad_norm": 0.173828125, - "learning_rate": 0.00015121772120915434, - "loss": 0.961, - "step": 2094 - }, - { - "epoch": 0.7604535147392291, - "grad_norm": 0.1416015625, - "learning_rate": 0.00015116876759270592, - "loss": 0.9972, - "step": 2096 - }, - { - "epoch": 0.7611791383219955, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001511198139762575, - "loss": 0.9814, - "step": 2098 - }, - { - "epoch": 0.7619047619047619, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001510708603598091, - "loss": 0.9646, - "step": 2100 - }, - { - "epoch": 0.7619047619047619, - "eval_loss": 1.0186907052993774, - "eval_runtime": 95.5861, - "eval_samples_per_second": 18.831, - "eval_steps_per_second": 0.596, - "step": 2100 - }, - { - "epoch": 0.7626303854875284, - "grad_norm": 0.1689453125, - "learning_rate": 0.00015102190674336067, - "loss": 1.0291, - "step": 2102 - }, - { - "epoch": 0.7633560090702948, - "grad_norm": 0.1650390625, - "learning_rate": 0.00015097295312691225, - "loss": 0.944, - "step": 2104 - }, - { - "epoch": 0.7640816326530612, - "grad_norm": 0.1533203125, - "learning_rate": 0.00015092399951046386, - "loss": 0.9971, - "step": 2106 - }, - { - "epoch": 0.7648072562358277, - "grad_norm": 0.1494140625, - "learning_rate": 0.00015087504589401542, - "loss": 0.9816, - "step": 2108 - }, - { - "epoch": 0.7655328798185941, - "grad_norm": 0.158203125, - "learning_rate": 0.000150826092277567, - "loss": 1.0105, - "step": 2110 - }, - { - "epoch": 0.7662585034013606, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001507771386611186, - "loss": 1.0059, - "step": 2112 - }, - { - "epoch": 0.766984126984127, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001507281850446702, - "loss": 0.9797, - "step": 2114 - }, - { - "epoch": 0.7677097505668934, - "grad_norm": 0.1474609375, - "learning_rate": 0.00015067923142822175, - "loss": 0.9623, - "step": 2116 - }, - { - "epoch": 0.7684353741496599, - "grad_norm": 0.17578125, - "learning_rate": 0.00015063027781177335, - "loss": 1.0352, - "step": 2118 - }, - { - "epoch": 0.7691609977324263, - "grad_norm": 0.1689453125, - "learning_rate": 0.00015058132419532494, - "loss": 1.0687, - "step": 2120 - }, - { - "epoch": 0.7698866213151927, - "grad_norm": 0.13671875, - "learning_rate": 0.00015053237057887652, - "loss": 0.9323, - "step": 2122 - }, - { - "epoch": 0.7706122448979592, - "grad_norm": 0.1640625, - "learning_rate": 0.0001504834169624281, - "loss": 0.9436, - "step": 2124 - }, - { - "epoch": 0.7713378684807256, - "grad_norm": 0.18359375, - "learning_rate": 0.00015043446334597968, - "loss": 1.0228, - "step": 2126 - }, - { - "epoch": 0.7720634920634921, - "grad_norm": 0.150390625, - "learning_rate": 0.00015038550972953127, - "loss": 0.9885, - "step": 2128 - }, - { - "epoch": 0.7727891156462585, - "grad_norm": 0.173828125, - "learning_rate": 0.00015033655611308285, - "loss": 0.9654, - "step": 2130 - }, - { - "epoch": 0.7735147392290249, - "grad_norm": 0.3046875, - "learning_rate": 0.00015028760249663446, - "loss": 1.1141, - "step": 2132 - }, - { - "epoch": 0.7742403628117914, - "grad_norm": 0.140625, - "learning_rate": 0.00015023864888018604, - "loss": 0.9624, - "step": 2134 - }, - { - "epoch": 0.7749659863945578, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001501896952637376, - "loss": 0.9589, - "step": 2136 - }, - { - "epoch": 0.7756916099773242, - "grad_norm": 0.1826171875, - "learning_rate": 0.0001501407416472892, - "loss": 1.0223, - "step": 2138 - }, - { - "epoch": 0.7764172335600907, - "grad_norm": 0.1640625, - "learning_rate": 0.0001500917880308408, - "loss": 0.9818, - "step": 2140 - }, - { - "epoch": 0.7771428571428571, - "grad_norm": 0.14453125, - "learning_rate": 0.00015004283441439237, - "loss": 0.9146, - "step": 2142 - }, - { - "epoch": 0.7778684807256235, - "grad_norm": 0.1357421875, - "learning_rate": 0.00014999388079794395, - "loss": 0.9904, - "step": 2144 - }, - { - "epoch": 0.77859410430839, - "grad_norm": 0.1708984375, - "learning_rate": 0.00014994492718149553, - "loss": 0.9883, - "step": 2146 - }, - { - "epoch": 0.7793197278911564, - "grad_norm": 0.1748046875, - "learning_rate": 0.00014989597356504712, - "loss": 1.0015, - "step": 2148 - }, - { - "epoch": 0.780045351473923, - "grad_norm": 0.158203125, - "learning_rate": 0.0001498470199485987, - "loss": 1.0113, - "step": 2150 - }, - { - "epoch": 0.7807709750566894, - "grad_norm": 0.1416015625, - "learning_rate": 0.0001497980663321503, - "loss": 0.9982, - "step": 2152 - }, - { - "epoch": 0.7814965986394558, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001497491127157019, - "loss": 1.0084, - "step": 2154 - }, - { - "epoch": 0.7822222222222223, - "grad_norm": 0.146484375, - "learning_rate": 0.00014970015909925345, - "loss": 0.9613, - "step": 2156 - }, - { - "epoch": 0.7829478458049887, - "grad_norm": 0.14453125, - "learning_rate": 0.00014965120548280505, - "loss": 1.0032, - "step": 2158 - }, - { - "epoch": 0.7836734693877551, - "grad_norm": 0.1455078125, - "learning_rate": 0.00014960225186635664, - "loss": 0.979, - "step": 2160 - }, - { - "epoch": 0.7843990929705216, - "grad_norm": 0.1591796875, - "learning_rate": 0.00014955329824990822, - "loss": 0.9816, - "step": 2162 - }, - { - "epoch": 0.785124716553288, - "grad_norm": 0.1689453125, - "learning_rate": 0.0001495043446334598, - "loss": 0.9398, - "step": 2164 - }, - { - "epoch": 0.7858503401360544, - "grad_norm": 0.150390625, - "learning_rate": 0.00014945539101701138, - "loss": 1.0336, - "step": 2166 - }, - { - "epoch": 0.7865759637188209, - "grad_norm": 0.150390625, - "learning_rate": 0.00014940643740056297, - "loss": 0.9288, - "step": 2168 - }, - { - "epoch": 0.7873015873015873, - "grad_norm": 0.1435546875, - "learning_rate": 0.00014935748378411455, - "loss": 1.0003, - "step": 2170 - }, - { - "epoch": 0.7880272108843538, - "grad_norm": 0.138671875, - "learning_rate": 0.00014930853016766616, - "loss": 0.9418, - "step": 2172 - }, - { - "epoch": 0.7887528344671202, - "grad_norm": 0.158203125, - "learning_rate": 0.00014925957655121774, - "loss": 1.0262, - "step": 2174 - }, - { - "epoch": 0.7894784580498866, - "grad_norm": 0.146484375, - "learning_rate": 0.0001492106229347693, - "loss": 1.0496, - "step": 2176 - }, - { - "epoch": 0.7902040816326531, - "grad_norm": 0.154296875, - "learning_rate": 0.0001491616693183209, - "loss": 0.937, - "step": 2178 - }, - { - "epoch": 0.7909297052154195, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001491127157018725, - "loss": 0.9422, - "step": 2180 - }, - { - "epoch": 0.7916553287981859, - "grad_norm": 0.1728515625, - "learning_rate": 0.00014906376208542407, - "loss": 0.9921, - "step": 2182 - }, - { - "epoch": 0.7923809523809524, - "grad_norm": 0.173828125, - "learning_rate": 0.00014901480846897565, - "loss": 0.9662, - "step": 2184 - }, - { - "epoch": 0.7931065759637188, - "grad_norm": 0.1826171875, - "learning_rate": 0.00014896585485252723, - "loss": 1.0241, - "step": 2186 - }, - { - "epoch": 0.7938321995464852, - "grad_norm": 0.16015625, - "learning_rate": 0.00014891690123607882, - "loss": 1.0144, - "step": 2188 - }, - { - "epoch": 0.7945578231292517, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001488679476196304, - "loss": 0.9861, - "step": 2190 - }, - { - "epoch": 0.7952834467120181, - "grad_norm": 0.150390625, - "learning_rate": 0.000148818994003182, - "loss": 0.9249, - "step": 2192 - }, - { - "epoch": 0.7960090702947846, - "grad_norm": 0.1435546875, - "learning_rate": 0.00014877004038673356, - "loss": 0.9389, - "step": 2194 - }, - { - "epoch": 0.796734693877551, - "grad_norm": 0.1552734375, - "learning_rate": 0.00014872108677028515, - "loss": 0.9589, - "step": 2196 - }, - { - "epoch": 0.7974603174603174, - "grad_norm": 0.1572265625, - "learning_rate": 0.00014867213315383675, - "loss": 1.0548, - "step": 2198 - }, - { - "epoch": 0.7981859410430839, - "grad_norm": 0.189453125, - "learning_rate": 0.00014862317953738834, - "loss": 1.0526, - "step": 2200 - }, - { - "epoch": 0.7981859410430839, - "eval_loss": 1.0182089805603027, - "eval_runtime": 95.9846, - "eval_samples_per_second": 18.753, - "eval_steps_per_second": 0.594, - "step": 2200 - }, - { - "epoch": 0.7989115646258503, - "grad_norm": 0.1572265625, - "learning_rate": 0.00014857422592093992, - "loss": 1.0364, - "step": 2202 - }, - { - "epoch": 0.7996371882086167, - "grad_norm": 0.166015625, - "learning_rate": 0.0001485252723044915, - "loss": 0.9324, - "step": 2204 - }, - { - "epoch": 0.8003628117913832, - "grad_norm": 0.1630859375, - "learning_rate": 0.00014847631868804308, - "loss": 0.936, - "step": 2206 - }, - { - "epoch": 0.8010884353741496, - "grad_norm": 0.1728515625, - "learning_rate": 0.00014842736507159467, - "loss": 1.0559, - "step": 2208 - }, - { - "epoch": 0.8018140589569162, - "grad_norm": 0.1484375, - "learning_rate": 0.00014837841145514625, - "loss": 1.0164, - "step": 2210 - }, - { - "epoch": 0.8025396825396826, - "grad_norm": 0.1640625, - "learning_rate": 0.00014832945783869786, - "loss": 1.0198, - "step": 2212 - }, - { - "epoch": 0.803265306122449, - "grad_norm": 0.1796875, - "learning_rate": 0.0001482805042222494, - "loss": 1.0224, - "step": 2214 - }, - { - "epoch": 0.8039909297052155, - "grad_norm": 0.1396484375, - "learning_rate": 0.000148231550605801, - "loss": 0.9483, - "step": 2216 - }, - { - "epoch": 0.8047165532879819, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001481825969893526, - "loss": 1.0488, - "step": 2218 - }, - { - "epoch": 0.8054421768707483, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001481336433729042, - "loss": 0.9822, - "step": 2220 - }, - { - "epoch": 0.8061678004535148, - "grad_norm": 0.1572265625, - "learning_rate": 0.00014808468975645577, - "loss": 0.969, - "step": 2222 - }, - { - "epoch": 0.8068934240362812, - "grad_norm": 0.1669921875, - "learning_rate": 0.00014803573614000735, - "loss": 0.9803, - "step": 2224 - }, - { - "epoch": 0.8076190476190476, - "grad_norm": 0.1708984375, - "learning_rate": 0.00014798678252355893, - "loss": 0.9859, - "step": 2226 - }, - { - "epoch": 0.8083446712018141, - "grad_norm": 0.142578125, - "learning_rate": 0.00014793782890711052, - "loss": 1.0081, - "step": 2228 - }, - { - "epoch": 0.8090702947845805, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001478888752906621, - "loss": 0.9469, - "step": 2230 - }, - { - "epoch": 0.809795918367347, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001478399216742137, - "loss": 1.001, - "step": 2232 - }, - { - "epoch": 0.8105215419501134, - "grad_norm": 0.1640625, - "learning_rate": 0.00014779096805776526, - "loss": 1.0429, - "step": 2234 - }, - { - "epoch": 0.8112471655328798, - "grad_norm": 0.1640625, - "learning_rate": 0.00014774201444131684, - "loss": 0.9483, - "step": 2236 - }, - { - "epoch": 0.8119727891156463, - "grad_norm": 0.142578125, - "learning_rate": 0.00014769306082486845, - "loss": 1.0031, - "step": 2238 - }, - { - "epoch": 0.8126984126984127, - "grad_norm": 0.1513671875, - "learning_rate": 0.00014764410720842004, - "loss": 0.9813, - "step": 2240 - }, - { - "epoch": 0.8134240362811791, - "grad_norm": 0.15625, - "learning_rate": 0.00014759515359197162, - "loss": 0.986, - "step": 2242 - }, - { - "epoch": 0.8141496598639456, - "grad_norm": 0.146484375, - "learning_rate": 0.0001475461999755232, - "loss": 1.0328, - "step": 2244 - }, - { - "epoch": 0.814875283446712, - "grad_norm": 0.1552734375, - "learning_rate": 0.00014749724635907478, - "loss": 1.032, - "step": 2246 - }, - { - "epoch": 0.8156009070294784, - "grad_norm": 0.1376953125, - "learning_rate": 0.00014744829274262637, - "loss": 0.9657, - "step": 2248 - }, - { - "epoch": 0.8163265306122449, - "grad_norm": 0.162109375, - "learning_rate": 0.00014739933912617795, - "loss": 0.9797, - "step": 2250 - }, - { - "epoch": 0.8170521541950113, - "grad_norm": 0.1826171875, - "learning_rate": 0.00014735038550972956, - "loss": 0.9919, - "step": 2252 - }, - { - "epoch": 0.8177777777777778, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001473014318932811, - "loss": 1.0706, - "step": 2254 - }, - { - "epoch": 0.8185034013605442, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001472524782768327, - "loss": 0.9914, - "step": 2256 - }, - { - "epoch": 0.8192290249433106, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001472035246603843, - "loss": 1.0764, - "step": 2258 - }, - { - "epoch": 0.8199546485260771, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001471545710439359, - "loss": 0.973, - "step": 2260 - }, - { - "epoch": 0.8206802721088435, - "grad_norm": 0.15234375, - "learning_rate": 0.00014710561742748744, - "loss": 0.9797, - "step": 2262 - }, - { - "epoch": 0.8214058956916099, - "grad_norm": 0.150390625, - "learning_rate": 0.00014705666381103905, - "loss": 0.9685, - "step": 2264 - }, - { - "epoch": 0.8221315192743764, - "grad_norm": 0.1591796875, - "learning_rate": 0.00014700771019459063, - "loss": 1.0145, - "step": 2266 - }, - { - "epoch": 0.8228571428571428, - "grad_norm": 0.1787109375, - "learning_rate": 0.00014695875657814222, - "loss": 0.9714, - "step": 2268 - }, - { - "epoch": 0.8235827664399092, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001469098029616938, - "loss": 0.9954, - "step": 2270 - }, - { - "epoch": 0.8243083900226758, - "grad_norm": 0.1728515625, - "learning_rate": 0.00014686084934524538, - "loss": 1.1002, - "step": 2272 - }, - { - "epoch": 0.8250340136054422, - "grad_norm": 0.1474609375, - "learning_rate": 0.00014681189572879696, - "loss": 0.9418, - "step": 2274 - }, - { - "epoch": 0.8257596371882087, - "grad_norm": 0.1357421875, - "learning_rate": 0.00014676294211234854, - "loss": 1.044, - "step": 2276 - }, - { - "epoch": 0.8264852607709751, - "grad_norm": 0.1669921875, - "learning_rate": 0.00014671398849590015, - "loss": 1.0346, - "step": 2278 - }, - { - "epoch": 0.8272108843537415, - "grad_norm": 0.1904296875, - "learning_rate": 0.00014666503487945174, - "loss": 1.027, - "step": 2280 - }, - { - "epoch": 0.827936507936508, - "grad_norm": 0.154296875, - "learning_rate": 0.0001466160812630033, - "loss": 0.9886, - "step": 2282 - }, - { - "epoch": 0.8286621315192744, - "grad_norm": 0.1884765625, - "learning_rate": 0.0001465671276465549, - "loss": 1.0345, - "step": 2284 - }, - { - "epoch": 0.8293877551020408, - "grad_norm": 0.177734375, - "learning_rate": 0.00014651817403010648, - "loss": 1.0252, - "step": 2286 - }, - { - "epoch": 0.8301133786848073, - "grad_norm": 0.1865234375, - "learning_rate": 0.00014646922041365807, - "loss": 1.0475, - "step": 2288 - }, - { - "epoch": 0.8308390022675737, - "grad_norm": 0.1806640625, - "learning_rate": 0.00014642026679720965, - "loss": 1.0286, - "step": 2290 - }, - { - "epoch": 0.8315646258503402, - "grad_norm": 0.1591796875, - "learning_rate": 0.00014637131318076123, - "loss": 0.9599, - "step": 2292 - }, - { - "epoch": 0.8322902494331066, - "grad_norm": 0.166015625, - "learning_rate": 0.0001463223595643128, - "loss": 1.0277, - "step": 2294 - }, - { - "epoch": 0.833015873015873, - "grad_norm": 0.15234375, - "learning_rate": 0.0001462734059478644, - "loss": 0.9699, - "step": 2296 - }, - { - "epoch": 0.8337414965986395, - "grad_norm": 0.193359375, - "learning_rate": 0.000146224452331416, - "loss": 0.9157, - "step": 2298 - }, - { - "epoch": 0.8344671201814059, - "grad_norm": 0.169921875, - "learning_rate": 0.00014617549871496759, - "loss": 1.0078, - "step": 2300 - }, - { - "epoch": 0.8344671201814059, - "eval_loss": 1.018009901046753, - "eval_runtime": 97.4465, - "eval_samples_per_second": 18.472, - "eval_steps_per_second": 0.585, - "step": 2300 - }, - { - "epoch": 0.8351927437641723, - "grad_norm": 0.1376953125, - "learning_rate": 0.00014612654509851914, - "loss": 0.9975, - "step": 2302 - }, - { - "epoch": 0.8359183673469388, - "grad_norm": 0.1552734375, - "learning_rate": 0.00014607759148207075, - "loss": 1.0193, - "step": 2304 - }, - { - "epoch": 0.8366439909297052, - "grad_norm": 0.166015625, - "learning_rate": 0.00014602863786562233, - "loss": 0.9847, - "step": 2306 - }, - { - "epoch": 0.8373696145124716, - "grad_norm": 0.1630859375, - "learning_rate": 0.00014597968424917392, - "loss": 1.0365, - "step": 2308 - }, - { - "epoch": 0.8380952380952381, - "grad_norm": 0.1484375, - "learning_rate": 0.0001459307306327255, - "loss": 0.9204, - "step": 2310 - }, - { - "epoch": 0.8388208616780045, - "grad_norm": 0.15234375, - "learning_rate": 0.00014588177701627708, - "loss": 0.973, - "step": 2312 - }, - { - "epoch": 0.839546485260771, - "grad_norm": 0.1640625, - "learning_rate": 0.00014583282339982866, - "loss": 1.0135, - "step": 2314 - }, - { - "epoch": 0.8402721088435374, - "grad_norm": 0.150390625, - "learning_rate": 0.00014578386978338024, - "loss": 0.9772, - "step": 2316 - }, - { - "epoch": 0.8409977324263038, - "grad_norm": 0.1484375, - "learning_rate": 0.00014573491616693185, - "loss": 0.9281, - "step": 2318 - }, - { - "epoch": 0.8417233560090703, - "grad_norm": 0.1572265625, - "learning_rate": 0.00014568596255048344, - "loss": 1.0221, - "step": 2320 - }, - { - "epoch": 0.8424489795918367, - "grad_norm": 0.1826171875, - "learning_rate": 0.000145637008934035, - "loss": 0.972, - "step": 2322 - }, - { - "epoch": 0.8431746031746031, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001455880553175866, - "loss": 1.0392, - "step": 2324 - }, - { - "epoch": 0.8439002267573696, - "grad_norm": 0.1845703125, - "learning_rate": 0.00014553910170113818, - "loss": 1.0201, - "step": 2326 - }, - { - "epoch": 0.844625850340136, - "grad_norm": 0.158203125, - "learning_rate": 0.00014549014808468977, - "loss": 0.9544, - "step": 2328 - }, - { - "epoch": 0.8453514739229024, - "grad_norm": 0.1533203125, - "learning_rate": 0.00014544119446824135, - "loss": 0.9848, - "step": 2330 - }, - { - "epoch": 0.846077097505669, - "grad_norm": 0.166015625, - "learning_rate": 0.00014539224085179293, - "loss": 0.9847, - "step": 2332 - }, - { - "epoch": 0.8468027210884354, - "grad_norm": 0.16796875, - "learning_rate": 0.0001453432872353445, - "loss": 1.0878, - "step": 2334 - }, - { - "epoch": 0.8475283446712019, - "grad_norm": 0.162109375, - "learning_rate": 0.0001452943336188961, - "loss": 0.9742, - "step": 2336 - }, - { - "epoch": 0.8482539682539683, - "grad_norm": 0.1484375, - "learning_rate": 0.0001452453800024477, - "loss": 1.06, - "step": 2338 - }, - { - "epoch": 0.8489795918367347, - "grad_norm": 0.1435546875, - "learning_rate": 0.00014519642638599926, - "loss": 0.9542, - "step": 2340 - }, - { - "epoch": 0.8497052154195012, - "grad_norm": 0.1611328125, - "learning_rate": 0.00014514747276955084, - "loss": 0.9095, - "step": 2342 - }, - { - "epoch": 0.8504308390022676, - "grad_norm": 0.1689453125, - "learning_rate": 0.00014509851915310245, - "loss": 1.0327, - "step": 2344 - }, - { - "epoch": 0.851156462585034, - "grad_norm": 0.14453125, - "learning_rate": 0.00014504956553665403, - "loss": 0.932, - "step": 2346 - }, - { - "epoch": 0.8518820861678005, - "grad_norm": 0.1494140625, - "learning_rate": 0.00014500061192020562, - "loss": 0.9619, - "step": 2348 - }, - { - "epoch": 0.8526077097505669, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001449516583037572, - "loss": 0.9473, - "step": 2350 - }, - { - "epoch": 0.8533333333333334, - "grad_norm": 0.1611328125, - "learning_rate": 0.00014490270468730878, - "loss": 1.0192, - "step": 2352 - }, - { - "epoch": 0.8540589569160998, - "grad_norm": 0.1484375, - "learning_rate": 0.00014485375107086036, - "loss": 0.9686, - "step": 2354 - }, - { - "epoch": 0.8547845804988662, - "grad_norm": 0.1611328125, - "learning_rate": 0.00014480479745441194, - "loss": 1.0355, - "step": 2356 - }, - { - "epoch": 0.8555102040816327, - "grad_norm": 0.169921875, - "learning_rate": 0.00014475584383796355, - "loss": 0.9756, - "step": 2358 - }, - { - "epoch": 0.8562358276643991, - "grad_norm": 0.1669921875, - "learning_rate": 0.0001447068902215151, - "loss": 1.0697, - "step": 2360 - }, - { - "epoch": 0.8569614512471655, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001446579366050667, - "loss": 0.9959, - "step": 2362 - }, - { - "epoch": 0.857687074829932, - "grad_norm": 0.1689453125, - "learning_rate": 0.0001446089829886183, - "loss": 0.984, - "step": 2364 - }, - { - "epoch": 0.8584126984126984, - "grad_norm": 0.16015625, - "learning_rate": 0.00014456002937216988, - "loss": 0.9598, - "step": 2366 - }, - { - "epoch": 0.8591383219954648, - "grad_norm": 0.166015625, - "learning_rate": 0.00014451107575572147, - "loss": 1.0026, - "step": 2368 - }, - { - "epoch": 0.8598639455782313, - "grad_norm": 0.138671875, - "learning_rate": 0.00014446212213927305, - "loss": 0.9883, - "step": 2370 - }, - { - "epoch": 0.8605895691609977, - "grad_norm": 0.1611328125, - "learning_rate": 0.00014441316852282463, - "loss": 1.0565, - "step": 2372 - }, - { - "epoch": 0.8613151927437642, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001443642149063762, - "loss": 0.9496, - "step": 2374 - }, - { - "epoch": 0.8620408163265306, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001443152612899278, - "loss": 0.961, - "step": 2376 - }, - { - "epoch": 0.862766439909297, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001442663076734794, - "loss": 0.9518, - "step": 2378 - }, - { - "epoch": 0.8634920634920635, - "grad_norm": 0.1455078125, - "learning_rate": 0.00014421735405703096, - "loss": 0.9264, - "step": 2380 - }, - { - "epoch": 0.8642176870748299, - "grad_norm": 0.1533203125, - "learning_rate": 0.00014416840044058254, - "loss": 1.0389, - "step": 2382 - }, - { - "epoch": 0.8649433106575963, - "grad_norm": 0.130859375, - "learning_rate": 0.00014411944682413415, - "loss": 1.0283, - "step": 2384 - }, - { - "epoch": 0.8656689342403628, - "grad_norm": 0.193359375, - "learning_rate": 0.00014407049320768573, - "loss": 1.0456, - "step": 2386 - }, - { - "epoch": 0.8663945578231292, - "grad_norm": 0.1484375, - "learning_rate": 0.00014402153959123732, - "loss": 1.007, - "step": 2388 - }, - { - "epoch": 0.8671201814058956, - "grad_norm": 0.169921875, - "learning_rate": 0.0001439725859747889, - "loss": 1.0643, - "step": 2390 - }, - { - "epoch": 0.8678458049886622, - "grad_norm": 0.14453125, - "learning_rate": 0.00014392363235834048, - "loss": 1.0052, - "step": 2392 - }, - { - "epoch": 0.8685714285714285, - "grad_norm": 0.1533203125, - "learning_rate": 0.00014387467874189206, - "loss": 0.9164, - "step": 2394 - }, - { - "epoch": 0.8692970521541951, - "grad_norm": 0.181640625, - "learning_rate": 0.00014382572512544364, - "loss": 0.9513, - "step": 2396 - }, - { - "epoch": 0.8700226757369615, - "grad_norm": 0.1669921875, - "learning_rate": 0.00014377677150899525, - "loss": 0.9772, - "step": 2398 - }, - { - "epoch": 0.8707482993197279, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001437278178925468, - "loss": 1.0452, - "step": 2400 - }, - { - "epoch": 0.8707482993197279, - "eval_loss": 1.0175837278366089, - "eval_runtime": 93.47, - "eval_samples_per_second": 19.258, - "eval_steps_per_second": 0.61, - "step": 2400 - }, - { - "epoch": 0.8714739229024944, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001436788642760984, - "loss": 0.9611, - "step": 2402 - }, - { - "epoch": 0.8721995464852608, - "grad_norm": 0.1455078125, - "learning_rate": 0.00014362991065965, - "loss": 0.9569, - "step": 2404 - }, - { - "epoch": 0.8729251700680272, - "grad_norm": 0.1513671875, - "learning_rate": 0.00014358095704320158, - "loss": 1.0252, - "step": 2406 - }, - { - "epoch": 0.8736507936507937, - "grad_norm": 0.1630859375, - "learning_rate": 0.00014353200342675314, - "loss": 1.0249, - "step": 2408 - }, - { - "epoch": 0.8743764172335601, - "grad_norm": 0.1728515625, - "learning_rate": 0.00014348304981030475, - "loss": 1.034, - "step": 2410 - }, - { - "epoch": 0.8751020408163265, - "grad_norm": 0.162109375, - "learning_rate": 0.00014343409619385633, - "loss": 0.9909, - "step": 2412 - }, - { - "epoch": 0.875827664399093, - "grad_norm": 0.150390625, - "learning_rate": 0.0001433851425774079, - "loss": 0.9933, - "step": 2414 - }, - { - "epoch": 0.8765532879818594, - "grad_norm": 0.21875, - "learning_rate": 0.0001433361889609595, - "loss": 0.8944, - "step": 2416 - }, - { - "epoch": 0.8772789115646259, - "grad_norm": 0.154296875, - "learning_rate": 0.00014328723534451108, - "loss": 1.0102, - "step": 2418 - }, - { - "epoch": 0.8780045351473923, - "grad_norm": 0.1474609375, - "learning_rate": 0.00014323828172806266, - "loss": 0.959, - "step": 2420 - }, - { - "epoch": 0.8787301587301587, - "grad_norm": 0.1328125, - "learning_rate": 0.00014318932811161424, - "loss": 0.9532, - "step": 2422 - }, - { - "epoch": 0.8794557823129252, - "grad_norm": 0.14453125, - "learning_rate": 0.00014314037449516585, - "loss": 0.9811, - "step": 2424 - }, - { - "epoch": 0.8801814058956916, - "grad_norm": 0.154296875, - "learning_rate": 0.00014309142087871743, - "loss": 0.9697, - "step": 2426 - }, - { - "epoch": 0.880907029478458, - "grad_norm": 0.18359375, - "learning_rate": 0.000143042467262269, - "loss": 1.0135, - "step": 2428 - }, - { - "epoch": 0.8816326530612245, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001429935136458206, - "loss": 1.0377, - "step": 2430 - }, - { - "epoch": 0.8823582766439909, - "grad_norm": 0.1474609375, - "learning_rate": 0.00014294456002937218, - "loss": 0.9046, - "step": 2432 - }, - { - "epoch": 0.8830839002267574, - "grad_norm": 0.1630859375, - "learning_rate": 0.00014289560641292376, - "loss": 0.9787, - "step": 2434 - }, - { - "epoch": 0.8838095238095238, - "grad_norm": 0.15625, - "learning_rate": 0.00014284665279647534, - "loss": 1.0511, - "step": 2436 - }, - { - "epoch": 0.8845351473922902, - "grad_norm": 0.15625, - "learning_rate": 0.00014279769918002693, - "loss": 0.9777, - "step": 2438 - }, - { - "epoch": 0.8852607709750567, - "grad_norm": 0.1669921875, - "learning_rate": 0.0001427487455635785, - "loss": 0.9929, - "step": 2440 - }, - { - "epoch": 0.8859863945578231, - "grad_norm": 0.1484375, - "learning_rate": 0.0001426997919471301, - "loss": 1.077, - "step": 2442 - }, - { - "epoch": 0.8867120181405895, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001426508383306817, - "loss": 1.0122, - "step": 2444 - }, - { - "epoch": 0.887437641723356, - "grad_norm": 0.13671875, - "learning_rate": 0.00014260188471423328, - "loss": 0.9913, - "step": 2446 - }, - { - "epoch": 0.8881632653061224, - "grad_norm": 0.15234375, - "learning_rate": 0.00014255293109778484, - "loss": 0.9736, - "step": 2448 - }, - { - "epoch": 0.8888888888888888, - "grad_norm": 0.162109375, - "learning_rate": 0.00014250397748133645, - "loss": 0.9983, - "step": 2450 - }, - { - "epoch": 0.8896145124716553, - "grad_norm": 0.1826171875, - "learning_rate": 0.00014245502386488803, - "loss": 1.033, - "step": 2452 - }, - { - "epoch": 0.8903401360544217, - "grad_norm": 0.154296875, - "learning_rate": 0.0001424060702484396, - "loss": 1.0804, - "step": 2454 - }, - { - "epoch": 0.8910657596371883, - "grad_norm": 0.17578125, - "learning_rate": 0.0001423571166319912, - "loss": 0.9584, - "step": 2456 - }, - { - "epoch": 0.8917913832199547, - "grad_norm": 0.1728515625, - "learning_rate": 0.00014230816301554278, - "loss": 1.0151, - "step": 2458 - }, - { - "epoch": 0.8925170068027211, - "grad_norm": 0.169921875, - "learning_rate": 0.00014225920939909436, - "loss": 1.0216, - "step": 2460 - }, - { - "epoch": 0.8932426303854876, - "grad_norm": 0.1494140625, - "learning_rate": 0.00014221025578264594, - "loss": 1.0113, - "step": 2462 - }, - { - "epoch": 0.893968253968254, - "grad_norm": 0.134765625, - "learning_rate": 0.00014216130216619755, - "loss": 1.0047, - "step": 2464 - }, - { - "epoch": 0.8946938775510204, - "grad_norm": 0.162109375, - "learning_rate": 0.00014211234854974913, - "loss": 0.9717, - "step": 2466 - }, - { - "epoch": 0.8954195011337869, - "grad_norm": 0.1484375, - "learning_rate": 0.0001420633949333007, - "loss": 1.0086, - "step": 2468 - }, - { - "epoch": 0.8961451247165533, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001420144413168523, - "loss": 0.9738, - "step": 2470 - }, - { - "epoch": 0.8968707482993197, - "grad_norm": 0.1552734375, - "learning_rate": 0.00014196548770040388, - "loss": 1.0254, - "step": 2472 - }, - { - "epoch": 0.8975963718820862, - "grad_norm": 0.154296875, - "learning_rate": 0.00014191653408395546, - "loss": 0.9363, - "step": 2474 - }, - { - "epoch": 0.8983219954648526, - "grad_norm": 0.1591796875, - "learning_rate": 0.00014186758046750704, - "loss": 1.0181, - "step": 2476 - }, - { - "epoch": 0.8990476190476191, - "grad_norm": 0.1650390625, - "learning_rate": 0.00014181862685105863, - "loss": 0.9986, - "step": 2478 - }, - { - "epoch": 0.8997732426303855, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001417696732346102, - "loss": 1.0503, - "step": 2480 - }, - { - "epoch": 0.9004988662131519, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001417207196181618, - "loss": 1.0875, - "step": 2482 - }, - { - "epoch": 0.9012244897959184, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001416717660017134, - "loss": 1.0045, - "step": 2484 - }, - { - "epoch": 0.9019501133786848, - "grad_norm": 0.150390625, - "learning_rate": 0.00014162281238526496, - "loss": 1.0325, - "step": 2486 - }, - { - "epoch": 0.9026757369614512, - "grad_norm": 0.1650390625, - "learning_rate": 0.00014157385876881654, - "loss": 0.9717, - "step": 2488 - }, - { - "epoch": 0.9034013605442177, - "grad_norm": 0.1396484375, - "learning_rate": 0.00014152490515236815, - "loss": 1.0202, - "step": 2490 - }, - { - "epoch": 0.9041269841269841, - "grad_norm": 0.1708984375, - "learning_rate": 0.00014147595153591973, - "loss": 0.9743, - "step": 2492 - }, - { - "epoch": 0.9048526077097505, - "grad_norm": 0.162109375, - "learning_rate": 0.0001414269979194713, - "loss": 0.9956, - "step": 2494 - }, - { - "epoch": 0.905578231292517, - "grad_norm": 0.16796875, - "learning_rate": 0.0001413780443030229, - "loss": 0.9782, - "step": 2496 - }, - { - "epoch": 0.9063038548752834, - "grad_norm": 0.16796875, - "learning_rate": 0.00014132909068657448, - "loss": 1.0001, - "step": 2498 - }, - { - "epoch": 0.9070294784580499, - "grad_norm": 0.1396484375, - "learning_rate": 0.00014128013707012606, - "loss": 0.9439, - "step": 2500 - }, - { - "epoch": 0.9070294784580499, - "eval_loss": 1.0171304941177368, - "eval_runtime": 93.0998, - "eval_samples_per_second": 19.334, - "eval_steps_per_second": 0.612, - "step": 2500 - }, - { - "epoch": 0.9077551020408163, - "grad_norm": 0.181640625, - "learning_rate": 0.00014123118345367764, - "loss": 1.0005, - "step": 2502 - }, - { - "epoch": 0.9084807256235827, - "grad_norm": 0.1591796875, - "learning_rate": 0.00014118222983722925, - "loss": 0.9762, - "step": 2504 - }, - { - "epoch": 0.9092063492063492, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001411332762207808, - "loss": 1.0189, - "step": 2506 - }, - { - "epoch": 0.9099319727891156, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001410843226043324, - "loss": 1.0226, - "step": 2508 - }, - { - "epoch": 0.910657596371882, - "grad_norm": 0.1630859375, - "learning_rate": 0.000141035368987884, - "loss": 1.008, - "step": 2510 - }, - { - "epoch": 0.9113832199546485, - "grad_norm": 0.1572265625, - "learning_rate": 0.00014098641537143558, - "loss": 0.9309, - "step": 2512 - }, - { - "epoch": 0.912108843537415, - "grad_norm": 0.15625, - "learning_rate": 0.00014093746175498716, - "loss": 0.9321, - "step": 2514 - }, - { - "epoch": 0.9128344671201815, - "grad_norm": 0.1689453125, - "learning_rate": 0.00014088850813853874, - "loss": 1.0389, - "step": 2516 - }, - { - "epoch": 0.9135600907029479, - "grad_norm": 0.166015625, - "learning_rate": 0.00014083955452209033, - "loss": 0.9912, - "step": 2518 - }, - { - "epoch": 0.9142857142857143, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001407906009056419, - "loss": 1.0086, - "step": 2520 - }, - { - "epoch": 0.9150113378684808, - "grad_norm": 0.1630859375, - "learning_rate": 0.0001407416472891935, - "loss": 1.0779, - "step": 2522 - }, - { - "epoch": 0.9157369614512472, - "grad_norm": 0.1669921875, - "learning_rate": 0.0001406926936727451, - "loss": 1.0576, - "step": 2524 - }, - { - "epoch": 0.9164625850340136, - "grad_norm": 0.1533203125, - "learning_rate": 0.00014064374005629666, - "loss": 1.0101, - "step": 2526 - }, - { - "epoch": 0.9171882086167801, - "grad_norm": 0.162109375, - "learning_rate": 0.00014059478643984824, - "loss": 1.0382, - "step": 2528 - }, - { - "epoch": 0.9179138321995465, - "grad_norm": 0.1748046875, - "learning_rate": 0.00014054583282339985, - "loss": 1.0193, - "step": 2530 - }, - { - "epoch": 0.9186394557823129, - "grad_norm": 0.1484375, - "learning_rate": 0.00014049687920695143, - "loss": 1.0049, - "step": 2532 - }, - { - "epoch": 0.9193650793650794, - "grad_norm": 0.162109375, - "learning_rate": 0.000140447925590503, - "loss": 0.9785, - "step": 2534 - }, - { - "epoch": 0.9200907029478458, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001403989719740546, - "loss": 0.9122, - "step": 2536 - }, - { - "epoch": 0.9208163265306123, - "grad_norm": 0.146484375, - "learning_rate": 0.00014035001835760618, - "loss": 0.9331, - "step": 2538 - }, - { - "epoch": 0.9215419501133787, - "grad_norm": 0.14453125, - "learning_rate": 0.00014030106474115776, - "loss": 1.0167, - "step": 2540 - }, - { - "epoch": 0.9222675736961451, - "grad_norm": 0.154296875, - "learning_rate": 0.00014025211112470934, - "loss": 0.9788, - "step": 2542 - }, - { - "epoch": 0.9229931972789116, - "grad_norm": 0.169921875, - "learning_rate": 0.00014020315750826095, - "loss": 1.0077, - "step": 2544 - }, - { - "epoch": 0.923718820861678, - "grad_norm": 0.154296875, - "learning_rate": 0.0001401542038918125, - "loss": 1.0014, - "step": 2546 - }, - { - "epoch": 0.9244444444444444, - "grad_norm": 0.15234375, - "learning_rate": 0.0001401052502753641, - "loss": 0.9479, - "step": 2548 - }, - { - "epoch": 0.9251700680272109, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001400562966589157, - "loss": 1.0187, - "step": 2550 - }, - { - "epoch": 0.9258956916099773, - "grad_norm": 0.1494140625, - "learning_rate": 0.00014000734304246728, - "loss": 0.9687, - "step": 2552 - }, - { - "epoch": 0.9266213151927437, - "grad_norm": 0.162109375, - "learning_rate": 0.00013995838942601883, - "loss": 0.9767, - "step": 2554 - }, - { - "epoch": 0.9273469387755102, - "grad_norm": 0.146484375, - "learning_rate": 0.00013990943580957044, - "loss": 1.0263, - "step": 2556 - }, - { - "epoch": 0.9280725623582766, - "grad_norm": 0.1591796875, - "learning_rate": 0.00013986048219312203, - "loss": 0.979, - "step": 2558 - }, - { - "epoch": 0.9287981859410431, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001398115285766736, - "loss": 1.0121, - "step": 2560 - }, - { - "epoch": 0.9295238095238095, - "grad_norm": 0.169921875, - "learning_rate": 0.0001397625749602252, - "loss": 0.9887, - "step": 2562 - }, - { - "epoch": 0.9302494331065759, - "grad_norm": 0.1484375, - "learning_rate": 0.00013971362134377677, - "loss": 0.949, - "step": 2564 - }, - { - "epoch": 0.9309750566893424, - "grad_norm": 0.171875, - "learning_rate": 0.00013966466772732835, - "loss": 0.9792, - "step": 2566 - }, - { - "epoch": 0.9317006802721088, - "grad_norm": 0.1708984375, - "learning_rate": 0.00013961571411087994, - "loss": 0.9836, - "step": 2568 - }, - { - "epoch": 0.9324263038548752, - "grad_norm": 0.1513671875, - "learning_rate": 0.00013956676049443155, - "loss": 0.9845, - "step": 2570 - }, - { - "epoch": 0.9331519274376417, - "grad_norm": 0.1533203125, - "learning_rate": 0.00013951780687798313, - "loss": 0.9387, - "step": 2572 - }, - { - "epoch": 0.9338775510204081, - "grad_norm": 0.14453125, - "learning_rate": 0.00013946885326153468, - "loss": 1.0167, - "step": 2574 - }, - { - "epoch": 0.9346031746031747, - "grad_norm": 0.1435546875, - "learning_rate": 0.0001394198996450863, - "loss": 1.0307, - "step": 2576 - }, - { - "epoch": 0.9353287981859411, - "grad_norm": 0.173828125, - "learning_rate": 0.00013937094602863788, - "loss": 1.0411, - "step": 2578 - }, - { - "epoch": 0.9360544217687075, - "grad_norm": 0.15234375, - "learning_rate": 0.00013932199241218946, - "loss": 0.9697, - "step": 2580 - }, - { - "epoch": 0.936780045351474, - "grad_norm": 0.142578125, - "learning_rate": 0.00013927303879574104, - "loss": 0.9579, - "step": 2582 - }, - { - "epoch": 0.9375056689342404, - "grad_norm": 0.1435546875, - "learning_rate": 0.00013922408517929262, - "loss": 0.9208, - "step": 2584 - }, - { - "epoch": 0.9382312925170068, - "grad_norm": 0.158203125, - "learning_rate": 0.0001391751315628442, - "loss": 1.0053, - "step": 2586 - }, - { - "epoch": 0.9389569160997733, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001391261779463958, - "loss": 0.9873, - "step": 2588 - }, - { - "epoch": 0.9396825396825397, - "grad_norm": 0.140625, - "learning_rate": 0.0001390772243299474, - "loss": 0.9857, - "step": 2590 - }, - { - "epoch": 0.9404081632653061, - "grad_norm": 0.1484375, - "learning_rate": 0.00013902827071349898, - "loss": 1.0317, - "step": 2592 - }, - { - "epoch": 0.9411337868480726, - "grad_norm": 0.1455078125, - "learning_rate": 0.00013897931709705053, - "loss": 0.939, - "step": 2594 - }, - { - "epoch": 0.941859410430839, - "grad_norm": 0.1708984375, - "learning_rate": 0.00013893036348060214, - "loss": 1.006, - "step": 2596 - }, - { - "epoch": 0.9425850340136055, - "grad_norm": 0.158203125, - "learning_rate": 0.00013888140986415373, - "loss": 1.083, - "step": 2598 - }, - { - "epoch": 0.9433106575963719, - "grad_norm": 0.146484375, - "learning_rate": 0.0001388324562477053, - "loss": 1.0546, - "step": 2600 - }, - { - "epoch": 0.9433106575963719, - "eval_loss": 1.016438603401184, - "eval_runtime": 97.3894, - "eval_samples_per_second": 18.482, - "eval_steps_per_second": 0.585, - "step": 2600 - }, - { - "epoch": 0.9440362811791383, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001387835026312569, - "loss": 0.9987, - "step": 2602 - }, - { - "epoch": 0.9447619047619048, - "grad_norm": 0.158203125, - "learning_rate": 0.00013873454901480847, - "loss": 0.9945, - "step": 2604 - }, - { - "epoch": 0.9454875283446712, - "grad_norm": 0.15625, - "learning_rate": 0.00013868559539836005, - "loss": 1.0031, - "step": 2606 - }, - { - "epoch": 0.9462131519274376, - "grad_norm": 0.1552734375, - "learning_rate": 0.00013863664178191164, - "loss": 1.0133, - "step": 2608 - }, - { - "epoch": 0.9469387755102041, - "grad_norm": 0.142578125, - "learning_rate": 0.00013858768816546325, - "loss": 1.0281, - "step": 2610 - }, - { - "epoch": 0.9476643990929705, - "grad_norm": 0.1689453125, - "learning_rate": 0.00013853873454901483, - "loss": 1.0062, - "step": 2612 - }, - { - "epoch": 0.9483900226757369, - "grad_norm": 0.150390625, - "learning_rate": 0.00013848978093256638, - "loss": 0.9394, - "step": 2614 - }, - { - "epoch": 0.9491156462585034, - "grad_norm": 0.1591796875, - "learning_rate": 0.000138440827316118, - "loss": 0.978, - "step": 2616 - }, - { - "epoch": 0.9498412698412698, - "grad_norm": 0.1513671875, - "learning_rate": 0.00013839187369966958, - "loss": 0.9216, - "step": 2618 - }, - { - "epoch": 0.9505668934240363, - "grad_norm": 0.140625, - "learning_rate": 0.00013834292008322116, - "loss": 0.9806, - "step": 2620 - }, - { - "epoch": 0.9512925170068027, - "grad_norm": 0.16796875, - "learning_rate": 0.00013829396646677274, - "loss": 1.0237, - "step": 2622 - }, - { - "epoch": 0.9520181405895691, - "grad_norm": 0.1455078125, - "learning_rate": 0.00013824501285032432, - "loss": 0.9156, - "step": 2624 - }, - { - "epoch": 0.9527437641723356, - "grad_norm": 0.140625, - "learning_rate": 0.0001381960592338759, - "loss": 0.9625, - "step": 2626 - }, - { - "epoch": 0.953469387755102, - "grad_norm": 0.169921875, - "learning_rate": 0.0001381471056174275, - "loss": 1.0259, - "step": 2628 - }, - { - "epoch": 0.9541950113378684, - "grad_norm": 0.140625, - "learning_rate": 0.0001380981520009791, - "loss": 0.921, - "step": 2630 - }, - { - "epoch": 0.954920634920635, - "grad_norm": 0.1533203125, - "learning_rate": 0.00013804919838453065, - "loss": 0.9542, - "step": 2632 - }, - { - "epoch": 0.9556462585034013, - "grad_norm": 0.1806640625, - "learning_rate": 0.00013800024476808223, - "loss": 1.0174, - "step": 2634 - }, - { - "epoch": 0.9563718820861677, - "grad_norm": 0.169921875, - "learning_rate": 0.00013795129115163384, - "loss": 1.0429, - "step": 2636 - }, - { - "epoch": 0.9570975056689343, - "grad_norm": 0.15625, - "learning_rate": 0.00013790233753518543, - "loss": 0.9952, - "step": 2638 - }, - { - "epoch": 0.9578231292517007, - "grad_norm": 0.1337890625, - "learning_rate": 0.000137853383918737, - "loss": 0.957, - "step": 2640 - }, - { - "epoch": 0.9585487528344672, - "grad_norm": 0.1884765625, - "learning_rate": 0.0001378044303022886, - "loss": 1.0364, - "step": 2642 - }, - { - "epoch": 0.9592743764172336, - "grad_norm": 0.150390625, - "learning_rate": 0.00013775547668584017, - "loss": 0.9722, - "step": 2644 - }, - { - "epoch": 0.96, - "grad_norm": 0.130859375, - "learning_rate": 0.00013770652306939175, - "loss": 0.8664, - "step": 2646 - }, - { - "epoch": 0.9607256235827665, - "grad_norm": 0.154296875, - "learning_rate": 0.00013765756945294334, - "loss": 1.0287, - "step": 2648 - }, - { - "epoch": 0.9614512471655329, - "grad_norm": 0.1630859375, - "learning_rate": 0.00013760861583649495, - "loss": 0.9959, - "step": 2650 - }, - { - "epoch": 0.9621768707482993, - "grad_norm": 0.15625, - "learning_rate": 0.0001375596622200465, - "loss": 0.9511, - "step": 2652 - }, - { - "epoch": 0.9629024943310658, - "grad_norm": 0.150390625, - "learning_rate": 0.00013751070860359808, - "loss": 0.9145, - "step": 2654 - }, - { - "epoch": 0.9636281179138322, - "grad_norm": 0.15234375, - "learning_rate": 0.0001374617549871497, - "loss": 1.0201, - "step": 2656 - }, - { - "epoch": 0.9643537414965987, - "grad_norm": 0.1494140625, - "learning_rate": 0.00013741280137070128, - "loss": 0.9204, - "step": 2658 - }, - { - "epoch": 0.9650793650793651, - "grad_norm": 0.1787109375, - "learning_rate": 0.00013736384775425286, - "loss": 0.9881, - "step": 2660 - }, - { - "epoch": 0.9658049886621315, - "grad_norm": 0.14453125, - "learning_rate": 0.00013731489413780444, - "loss": 1.0235, - "step": 2662 - }, - { - "epoch": 0.966530612244898, - "grad_norm": 0.142578125, - "learning_rate": 0.00013726594052135602, - "loss": 0.9197, - "step": 2664 - }, - { - "epoch": 0.9672562358276644, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001372169869049076, - "loss": 0.957, - "step": 2666 - }, - { - "epoch": 0.9679818594104308, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001371680332884592, - "loss": 1.0025, - "step": 2668 - }, - { - "epoch": 0.9687074829931973, - "grad_norm": 0.1357421875, - "learning_rate": 0.0001371190796720108, - "loss": 0.9623, - "step": 2670 - }, - { - "epoch": 0.9694331065759637, - "grad_norm": 0.1513671875, - "learning_rate": 0.00013707012605556235, - "loss": 0.9854, - "step": 2672 - }, - { - "epoch": 0.9701587301587301, - "grad_norm": 0.1494140625, - "learning_rate": 0.00013702117243911393, - "loss": 0.9908, - "step": 2674 - }, - { - "epoch": 0.9708843537414966, - "grad_norm": 0.1552734375, - "learning_rate": 0.00013697221882266554, - "loss": 0.9643, - "step": 2676 - }, - { - "epoch": 0.971609977324263, - "grad_norm": 0.140625, - "learning_rate": 0.00013692326520621713, - "loss": 1.0681, - "step": 2678 - }, - { - "epoch": 0.9723356009070295, - "grad_norm": 0.16015625, - "learning_rate": 0.0001368743115897687, - "loss": 0.9858, - "step": 2680 - }, - { - "epoch": 0.9730612244897959, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001368253579733203, - "loss": 0.9974, - "step": 2682 - }, - { - "epoch": 0.9737868480725623, - "grad_norm": 0.150390625, - "learning_rate": 0.00013677640435687187, - "loss": 0.9672, - "step": 2684 - }, - { - "epoch": 0.9745124716553288, - "grad_norm": 0.166015625, - "learning_rate": 0.00013672745074042345, - "loss": 0.9644, - "step": 2686 - }, - { - "epoch": 0.9752380952380952, - "grad_norm": 0.142578125, - "learning_rate": 0.00013667849712397504, - "loss": 0.9687, - "step": 2688 - }, - { - "epoch": 0.9759637188208616, - "grad_norm": 0.1435546875, - "learning_rate": 0.00013662954350752665, - "loss": 0.9365, - "step": 2690 - }, - { - "epoch": 0.9766893424036281, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001365805898910782, - "loss": 1.0098, - "step": 2692 - }, - { - "epoch": 0.9774149659863945, - "grad_norm": 0.1767578125, - "learning_rate": 0.00013653163627462978, - "loss": 1.0377, - "step": 2694 - }, - { - "epoch": 0.978140589569161, - "grad_norm": 0.1513671875, - "learning_rate": 0.0001364826826581814, - "loss": 0.9091, - "step": 2696 - }, - { - "epoch": 0.9788662131519275, - "grad_norm": 0.1435546875, - "learning_rate": 0.00013643372904173298, - "loss": 1.0475, - "step": 2698 - }, - { - "epoch": 0.9795918367346939, - "grad_norm": 0.1474609375, - "learning_rate": 0.00013638477542528453, - "loss": 0.9671, - "step": 2700 - }, - { - "epoch": 0.9795918367346939, - "eval_loss": 1.0159815549850464, - "eval_runtime": 97.2766, - "eval_samples_per_second": 18.504, - "eval_steps_per_second": 0.586, - "step": 2700 - }, - { - "epoch": 0.9803174603174604, - "grad_norm": 0.1494140625, - "learning_rate": 0.00013633582180883614, - "loss": 0.9728, - "step": 2702 - }, - { - "epoch": 0.9810430839002268, - "grad_norm": 0.1640625, - "learning_rate": 0.00013628686819238772, - "loss": 1.0255, - "step": 2704 - }, - { - "epoch": 0.9817687074829932, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001362379145759393, - "loss": 1.0246, - "step": 2706 - }, - { - "epoch": 0.9824943310657597, - "grad_norm": 0.1640625, - "learning_rate": 0.00013618896095949089, - "loss": 0.9583, - "step": 2708 - }, - { - "epoch": 0.9832199546485261, - "grad_norm": 0.158203125, - "learning_rate": 0.00013614000734304247, - "loss": 0.9655, - "step": 2710 - }, - { - "epoch": 0.9839455782312925, - "grad_norm": 0.15234375, - "learning_rate": 0.00013609105372659405, - "loss": 0.9796, - "step": 2712 - }, - { - "epoch": 0.984671201814059, - "grad_norm": 0.14453125, - "learning_rate": 0.00013604210011014563, - "loss": 0.9351, - "step": 2714 - }, - { - "epoch": 0.9853968253968254, - "grad_norm": 0.201171875, - "learning_rate": 0.00013599314649369724, - "loss": 1.0364, - "step": 2716 - }, - { - "epoch": 0.9861224489795918, - "grad_norm": 0.1357421875, - "learning_rate": 0.00013594419287724882, - "loss": 1.031, - "step": 2718 - }, - { - "epoch": 0.9868480725623583, - "grad_norm": 0.13671875, - "learning_rate": 0.00013589523926080038, - "loss": 0.9863, - "step": 2720 - }, - { - "epoch": 0.9875736961451247, - "grad_norm": 0.1640625, - "learning_rate": 0.000135846285644352, - "loss": 0.9365, - "step": 2722 - }, - { - "epoch": 0.9882993197278912, - "grad_norm": 0.177734375, - "learning_rate": 0.00013579733202790357, - "loss": 1.0187, - "step": 2724 - }, - { - "epoch": 0.9890249433106576, - "grad_norm": 0.138671875, - "learning_rate": 0.00013574837841145515, - "loss": 1.0717, - "step": 2726 - }, - { - "epoch": 0.989750566893424, - "grad_norm": 0.162109375, - "learning_rate": 0.00013569942479500674, - "loss": 1.0162, - "step": 2728 - }, - { - "epoch": 0.9904761904761905, - "grad_norm": 0.181640625, - "learning_rate": 0.00013565047117855832, - "loss": 1.0003, - "step": 2730 - }, - { - "epoch": 0.9912018140589569, - "grad_norm": 0.14453125, - "learning_rate": 0.0001356015175621099, - "loss": 0.9634, - "step": 2732 - }, - { - "epoch": 0.9919274376417233, - "grad_norm": 0.13671875, - "learning_rate": 0.00013555256394566148, - "loss": 0.9042, - "step": 2734 - }, - { - "epoch": 0.9926530612244898, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001355036103292131, - "loss": 1.0227, - "step": 2736 - }, - { - "epoch": 0.9933786848072562, - "grad_norm": 0.154296875, - "learning_rate": 0.00013545465671276467, - "loss": 0.9887, - "step": 2738 - }, - { - "epoch": 0.9941043083900227, - "grad_norm": 0.1748046875, - "learning_rate": 0.00013540570309631623, - "loss": 0.9897, - "step": 2740 - }, - { - "epoch": 0.9948299319727891, - "grad_norm": 0.1650390625, - "learning_rate": 0.00013535674947986784, - "loss": 0.9965, - "step": 2742 - }, - { - "epoch": 0.9955555555555555, - "grad_norm": 0.1630859375, - "learning_rate": 0.00013530779586341942, - "loss": 0.9902, - "step": 2744 - }, - { - "epoch": 0.996281179138322, - "grad_norm": 0.1416015625, - "learning_rate": 0.000135258842246971, - "loss": 1.0522, - "step": 2746 - }, - { - "epoch": 0.9970068027210884, - "grad_norm": 0.158203125, - "learning_rate": 0.00013520988863052259, - "loss": 0.9595, - "step": 2748 - }, - { - "epoch": 0.9977324263038548, - "grad_norm": 0.1494140625, - "learning_rate": 0.00013516093501407417, - "loss": 0.9819, - "step": 2750 - }, - { - "epoch": 0.9984580498866213, - "grad_norm": 0.185546875, - "learning_rate": 0.00013511198139762575, - "loss": 1.1001, - "step": 2752 - }, - { - "epoch": 0.9991836734693877, - "grad_norm": 0.150390625, - "learning_rate": 0.00013506302778117733, - "loss": 0.9502, - "step": 2754 - }, - { - "epoch": 0.9999092970521541, - "grad_norm": 0.15625, - "learning_rate": 0.00013501407416472894, - "loss": 0.9582, - "step": 2756 - }, - { - "epoch": 1.0003628117913832, - "grad_norm": 0.1494140625, - "learning_rate": 0.00013496512054828052, - "loss": 0.9314, - "step": 2758 - }, - { - "epoch": 1.0010884353741496, - "grad_norm": 0.162109375, - "learning_rate": 0.00013491616693183208, - "loss": 0.9523, - "step": 2760 - }, - { - "epoch": 1.001814058956916, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001348672133153837, - "loss": 1.0069, - "step": 2762 - }, - { - "epoch": 1.0025396825396826, - "grad_norm": 0.1611328125, - "learning_rate": 0.00013481825969893527, - "loss": 0.9828, - "step": 2764 - }, - { - "epoch": 1.003265306122449, - "grad_norm": 0.169921875, - "learning_rate": 0.00013476930608248685, - "loss": 1.0188, - "step": 2766 - }, - { - "epoch": 1.0039909297052154, - "grad_norm": 0.1611328125, - "learning_rate": 0.00013472035246603844, - "loss": 0.9474, - "step": 2768 - }, - { - "epoch": 1.0047165532879818, - "grad_norm": 0.1455078125, - "learning_rate": 0.00013467139884959002, - "loss": 0.9772, - "step": 2770 - }, - { - "epoch": 1.0054421768707482, - "grad_norm": 0.154296875, - "learning_rate": 0.0001346224452331416, - "loss": 1.0774, - "step": 2772 - }, - { - "epoch": 1.0061678004535148, - "grad_norm": 0.1435546875, - "learning_rate": 0.00013457349161669318, - "loss": 0.9454, - "step": 2774 - }, - { - "epoch": 1.0068934240362812, - "grad_norm": 0.158203125, - "learning_rate": 0.0001345245380002448, - "loss": 0.9687, - "step": 2776 - }, - { - "epoch": 1.0076190476190476, - "grad_norm": 0.1484375, - "learning_rate": 0.00013447558438379635, - "loss": 1.0119, - "step": 2778 - }, - { - "epoch": 1.008344671201814, - "grad_norm": 0.1689453125, - "learning_rate": 0.00013442663076734793, - "loss": 1.0194, - "step": 2780 - }, - { - "epoch": 1.0090702947845804, - "grad_norm": 0.1396484375, - "learning_rate": 0.00013437767715089954, - "loss": 0.9619, - "step": 2782 - }, - { - "epoch": 1.0097959183673468, - "grad_norm": 0.1669921875, - "learning_rate": 0.00013432872353445112, - "loss": 0.9702, - "step": 2784 - }, - { - "epoch": 1.0105215419501135, - "grad_norm": 0.1376953125, - "learning_rate": 0.0001342797699180027, - "loss": 0.9513, - "step": 2786 - }, - { - "epoch": 1.0112471655328799, - "grad_norm": 0.1494140625, - "learning_rate": 0.00013423081630155429, - "loss": 1.0061, - "step": 2788 - }, - { - "epoch": 1.0119727891156463, - "grad_norm": 0.1357421875, - "learning_rate": 0.00013418186268510587, - "loss": 0.9964, - "step": 2790 - }, - { - "epoch": 1.0126984126984127, - "grad_norm": 0.1640625, - "learning_rate": 0.00013413290906865745, - "loss": 0.9677, - "step": 2792 - }, - { - "epoch": 1.013424036281179, - "grad_norm": 0.1494140625, - "learning_rate": 0.00013408395545220903, - "loss": 0.9397, - "step": 2794 - }, - { - "epoch": 1.0141496598639457, - "grad_norm": 0.1474609375, - "learning_rate": 0.00013403500183576064, - "loss": 0.9438, - "step": 2796 - }, - { - "epoch": 1.014875283446712, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001339860482193122, - "loss": 0.9445, - "step": 2798 - }, - { - "epoch": 1.0156009070294785, - "grad_norm": 0.1591796875, - "learning_rate": 0.00013393709460286378, - "loss": 0.9987, - "step": 2800 - }, - { - "epoch": 1.0156009070294785, - "eval_loss": 1.0157527923583984, - "eval_runtime": 93.9867, - "eval_samples_per_second": 19.152, - "eval_steps_per_second": 0.606, - "step": 2800 - }, - { - "epoch": 1.0163265306122449, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001338881409864154, - "loss": 1.0336, - "step": 2802 - }, - { - "epoch": 1.0170521541950113, - "grad_norm": 0.162109375, - "learning_rate": 0.00013383918736996697, - "loss": 1.0242, - "step": 2804 - }, - { - "epoch": 1.0177777777777777, - "grad_norm": 0.154296875, - "learning_rate": 0.00013379023375351855, - "loss": 0.9604, - "step": 2806 - }, - { - "epoch": 1.0185034013605443, - "grad_norm": 0.146484375, - "learning_rate": 0.00013374128013707014, - "loss": 1.0008, - "step": 2808 - }, - { - "epoch": 1.0192290249433107, - "grad_norm": 0.162109375, - "learning_rate": 0.00013369232652062172, - "loss": 0.9853, - "step": 2810 - }, - { - "epoch": 1.019954648526077, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001336433729041733, - "loss": 1.1074, - "step": 2812 - }, - { - "epoch": 1.0206802721088435, - "grad_norm": 0.1435546875, - "learning_rate": 0.00013359441928772488, - "loss": 1.0543, - "step": 2814 - }, - { - "epoch": 1.0214058956916099, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001335454656712765, - "loss": 0.9545, - "step": 2816 - }, - { - "epoch": 1.0221315192743765, - "grad_norm": 0.162109375, - "learning_rate": 0.00013349651205482805, - "loss": 0.9777, - "step": 2818 - }, - { - "epoch": 1.022857142857143, - "grad_norm": 0.150390625, - "learning_rate": 0.00013344755843837963, - "loss": 1.002, - "step": 2820 - }, - { - "epoch": 1.0235827664399093, - "grad_norm": 0.158203125, - "learning_rate": 0.00013339860482193124, - "loss": 0.9998, - "step": 2822 - }, - { - "epoch": 1.0243083900226757, - "grad_norm": 0.1669921875, - "learning_rate": 0.00013334965120548282, - "loss": 0.9978, - "step": 2824 - }, - { - "epoch": 1.025034013605442, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001333006975890344, - "loss": 1.0295, - "step": 2826 - }, - { - "epoch": 1.0257596371882087, - "grad_norm": 0.1533203125, - "learning_rate": 0.00013325174397258599, - "loss": 1.0145, - "step": 2828 - }, - { - "epoch": 1.0264852607709751, - "grad_norm": 0.1708984375, - "learning_rate": 0.00013320279035613757, - "loss": 0.9537, - "step": 2830 - }, - { - "epoch": 1.0272108843537415, - "grad_norm": 0.1455078125, - "learning_rate": 0.00013315383673968915, - "loss": 1.091, - "step": 2832 - }, - { - "epoch": 1.027936507936508, - "grad_norm": 0.150390625, - "learning_rate": 0.00013310488312324073, - "loss": 1.0578, - "step": 2834 - }, - { - "epoch": 1.0286621315192743, - "grad_norm": 0.15234375, - "learning_rate": 0.00013305592950679234, - "loss": 0.9262, - "step": 2836 - }, - { - "epoch": 1.0293877551020407, - "grad_norm": 0.1640625, - "learning_rate": 0.0001330069758903439, - "loss": 0.9779, - "step": 2838 - }, - { - "epoch": 1.0301133786848073, - "grad_norm": 0.1806640625, - "learning_rate": 0.00013295802227389548, - "loss": 0.9869, - "step": 2840 - }, - { - "epoch": 1.0308390022675737, - "grad_norm": 0.15625, - "learning_rate": 0.0001329090686574471, - "loss": 0.9595, - "step": 2842 - }, - { - "epoch": 1.0315646258503401, - "grad_norm": 0.197265625, - "learning_rate": 0.00013286011504099867, - "loss": 1.0107, - "step": 2844 - }, - { - "epoch": 1.0322902494331065, - "grad_norm": 0.1572265625, - "learning_rate": 0.00013281116142455023, - "loss": 1.0039, - "step": 2846 - }, - { - "epoch": 1.033015873015873, - "grad_norm": 0.1728515625, - "learning_rate": 0.00013276220780810184, - "loss": 0.9353, - "step": 2848 - }, - { - "epoch": 1.0337414965986396, - "grad_norm": 0.162109375, - "learning_rate": 0.00013271325419165342, - "loss": 0.9496, - "step": 2850 - }, - { - "epoch": 1.034467120181406, - "grad_norm": 0.1416015625, - "learning_rate": 0.000132664300575205, - "loss": 1.0133, - "step": 2852 - }, - { - "epoch": 1.0351927437641724, - "grad_norm": 0.1455078125, - "learning_rate": 0.00013261534695875658, - "loss": 0.9736, - "step": 2854 - }, - { - "epoch": 1.0359183673469388, - "grad_norm": 0.1650390625, - "learning_rate": 0.00013256639334230816, - "loss": 1.0096, - "step": 2856 - }, - { - "epoch": 1.0366439909297052, - "grad_norm": 0.1767578125, - "learning_rate": 0.00013251743972585975, - "loss": 1.0358, - "step": 2858 - }, - { - "epoch": 1.0373696145124716, - "grad_norm": 0.1650390625, - "learning_rate": 0.00013246848610941133, - "loss": 1.0608, - "step": 2860 - }, - { - "epoch": 1.0380952380952382, - "grad_norm": 0.16796875, - "learning_rate": 0.00013241953249296294, - "loss": 0.9904, - "step": 2862 - }, - { - "epoch": 1.0388208616780046, - "grad_norm": 0.154296875, - "learning_rate": 0.00013237057887651452, - "loss": 0.9738, - "step": 2864 - }, - { - "epoch": 1.039546485260771, - "grad_norm": 0.185546875, - "learning_rate": 0.00013232162526006608, - "loss": 1.0697, - "step": 2866 - }, - { - "epoch": 1.0402721088435374, - "grad_norm": 0.146484375, - "learning_rate": 0.00013227267164361769, - "loss": 0.9685, - "step": 2868 - }, - { - "epoch": 1.0409977324263038, - "grad_norm": 0.1669921875, - "learning_rate": 0.00013222371802716927, - "loss": 1.0041, - "step": 2870 - }, - { - "epoch": 1.0417233560090704, - "grad_norm": 0.1513671875, - "learning_rate": 0.00013217476441072085, - "loss": 0.9809, - "step": 2872 - }, - { - "epoch": 1.0424489795918368, - "grad_norm": 0.158203125, - "learning_rate": 0.00013212581079427243, - "loss": 0.9908, - "step": 2874 - }, - { - "epoch": 1.0431746031746032, - "grad_norm": 0.16015625, - "learning_rate": 0.00013207685717782401, - "loss": 1.032, - "step": 2876 - }, - { - "epoch": 1.0439002267573696, - "grad_norm": 0.146484375, - "learning_rate": 0.0001320279035613756, - "loss": 0.9646, - "step": 2878 - }, - { - "epoch": 1.044625850340136, - "grad_norm": 0.166015625, - "learning_rate": 0.00013197894994492718, - "loss": 0.9965, - "step": 2880 - }, - { - "epoch": 1.0453514739229024, - "grad_norm": 0.1640625, - "learning_rate": 0.0001319299963284788, - "loss": 0.9963, - "step": 2882 - }, - { - "epoch": 1.046077097505669, - "grad_norm": 0.1650390625, - "learning_rate": 0.00013188104271203037, - "loss": 1.0057, - "step": 2884 - }, - { - "epoch": 1.0468027210884354, - "grad_norm": 0.1572265625, - "learning_rate": 0.00013183208909558193, - "loss": 0.9545, - "step": 2886 - }, - { - "epoch": 1.0475283446712018, - "grad_norm": 0.1640625, - "learning_rate": 0.00013178313547913354, - "loss": 0.9863, - "step": 2888 - }, - { - "epoch": 1.0482539682539682, - "grad_norm": 0.15234375, - "learning_rate": 0.00013173418186268512, - "loss": 1.0419, - "step": 2890 - }, - { - "epoch": 1.0489795918367346, - "grad_norm": 0.1796875, - "learning_rate": 0.0001316852282462367, - "loss": 1.013, - "step": 2892 - }, - { - "epoch": 1.0497052154195012, - "grad_norm": 0.1533203125, - "learning_rate": 0.00013163627462978828, - "loss": 1.096, - "step": 2894 - }, - { - "epoch": 1.0504308390022676, - "grad_norm": 0.146484375, - "learning_rate": 0.00013158732101333986, - "loss": 1.0023, - "step": 2896 - }, - { - "epoch": 1.051156462585034, - "grad_norm": 0.1533203125, - "learning_rate": 0.00013153836739689145, - "loss": 0.9743, - "step": 2898 - }, - { - "epoch": 1.0518820861678004, - "grad_norm": 0.1669921875, - "learning_rate": 0.00013148941378044303, - "loss": 1.0481, - "step": 2900 - }, - { - "epoch": 1.0518820861678004, - "eval_loss": 1.0157631635665894, - "eval_runtime": 98.609, - "eval_samples_per_second": 18.254, - "eval_steps_per_second": 0.578, - "step": 2900 - }, - { - "epoch": 1.0526077097505668, - "grad_norm": 0.15625, - "learning_rate": 0.00013144046016399464, - "loss": 1.0343, - "step": 2902 - }, - { - "epoch": 1.0533333333333332, - "grad_norm": 0.1513671875, - "learning_rate": 0.00013139150654754622, - "loss": 0.9971, - "step": 2904 - }, - { - "epoch": 1.0540589569160999, - "grad_norm": 0.1689453125, - "learning_rate": 0.00013134255293109778, - "loss": 0.9762, - "step": 2906 - }, - { - "epoch": 1.0547845804988663, - "grad_norm": 0.154296875, - "learning_rate": 0.00013129359931464939, - "loss": 1.1036, - "step": 2908 - }, - { - "epoch": 1.0555102040816327, - "grad_norm": 0.169921875, - "learning_rate": 0.00013124464569820097, - "loss": 0.9814, - "step": 2910 - }, - { - "epoch": 1.056235827664399, - "grad_norm": 0.1533203125, - "learning_rate": 0.00013119569208175255, - "loss": 0.9185, - "step": 2912 - }, - { - "epoch": 1.0569614512471655, - "grad_norm": 0.1669921875, - "learning_rate": 0.00013114673846530413, - "loss": 0.9613, - "step": 2914 - }, - { - "epoch": 1.057687074829932, - "grad_norm": 0.14453125, - "learning_rate": 0.00013109778484885571, - "loss": 0.9276, - "step": 2916 - }, - { - "epoch": 1.0584126984126985, - "grad_norm": 0.1416015625, - "learning_rate": 0.0001310488312324073, - "loss": 0.9268, - "step": 2918 - }, - { - "epoch": 1.0591383219954649, - "grad_norm": 0.1533203125, - "learning_rate": 0.00013099987761595888, - "loss": 1.0498, - "step": 2920 - }, - { - "epoch": 1.0598639455782313, - "grad_norm": 0.158203125, - "learning_rate": 0.0001309509239995105, - "loss": 1.0223, - "step": 2922 - }, - { - "epoch": 1.0605895691609977, - "grad_norm": 0.150390625, - "learning_rate": 0.00013090197038306204, - "loss": 0.9899, - "step": 2924 - }, - { - "epoch": 1.061315192743764, - "grad_norm": 0.1533203125, - "learning_rate": 0.00013085301676661363, - "loss": 0.9639, - "step": 2926 - }, - { - "epoch": 1.0620408163265307, - "grad_norm": 0.1513671875, - "learning_rate": 0.00013080406315016524, - "loss": 0.9369, - "step": 2928 - }, - { - "epoch": 1.062766439909297, - "grad_norm": 0.1455078125, - "learning_rate": 0.00013075510953371682, - "loss": 0.93, - "step": 2930 - }, - { - "epoch": 1.0634920634920635, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001307061559172684, - "loss": 0.993, - "step": 2932 - }, - { - "epoch": 1.0642176870748299, - "grad_norm": 0.181640625, - "learning_rate": 0.00013065720230081998, - "loss": 1.0847, - "step": 2934 - }, - { - "epoch": 1.0649433106575963, - "grad_norm": 0.1552734375, - "learning_rate": 0.00013060824868437156, - "loss": 1.003, - "step": 2936 - }, - { - "epoch": 1.065668934240363, - "grad_norm": 0.166015625, - "learning_rate": 0.00013055929506792315, - "loss": 1.0098, - "step": 2938 - }, - { - "epoch": 1.0663945578231293, - "grad_norm": 0.1513671875, - "learning_rate": 0.00013051034145147473, - "loss": 0.9617, - "step": 2940 - }, - { - "epoch": 1.0671201814058957, - "grad_norm": 0.15234375, - "learning_rate": 0.00013046138783502634, - "loss": 0.9282, - "step": 2942 - }, - { - "epoch": 1.067845804988662, - "grad_norm": 0.1416015625, - "learning_rate": 0.0001304124342185779, - "loss": 0.9462, - "step": 2944 - }, - { - "epoch": 1.0685714285714285, - "grad_norm": 0.1650390625, - "learning_rate": 0.00013036348060212948, - "loss": 1.0117, - "step": 2946 - }, - { - "epoch": 1.069297052154195, - "grad_norm": 0.15625, - "learning_rate": 0.00013031452698568109, - "loss": 0.9634, - "step": 2948 - }, - { - "epoch": 1.0700226757369615, - "grad_norm": 0.171875, - "learning_rate": 0.00013026557336923267, - "loss": 1.0243, - "step": 2950 - }, - { - "epoch": 1.070748299319728, - "grad_norm": 0.142578125, - "learning_rate": 0.00013021661975278425, - "loss": 0.9924, - "step": 2952 - }, - { - "epoch": 1.0714739229024943, - "grad_norm": 0.177734375, - "learning_rate": 0.00013016766613633583, - "loss": 0.9808, - "step": 2954 - }, - { - "epoch": 1.0721995464852607, - "grad_norm": 0.181640625, - "learning_rate": 0.00013011871251988741, - "loss": 1.0503, - "step": 2956 - }, - { - "epoch": 1.0729251700680271, - "grad_norm": 0.1328125, - "learning_rate": 0.000130069758903439, - "loss": 1.042, - "step": 2958 - }, - { - "epoch": 1.0736507936507937, - "grad_norm": 0.154296875, - "learning_rate": 0.00013002080528699058, - "loss": 1.0484, - "step": 2960 - }, - { - "epoch": 1.0743764172335601, - "grad_norm": 0.14453125, - "learning_rate": 0.0001299718516705422, - "loss": 0.925, - "step": 2962 - }, - { - "epoch": 1.0751020408163265, - "grad_norm": 0.154296875, - "learning_rate": 0.00012992289805409374, - "loss": 1.0459, - "step": 2964 - }, - { - "epoch": 1.075827664399093, - "grad_norm": 0.1845703125, - "learning_rate": 0.00012987394443764533, - "loss": 0.9794, - "step": 2966 - }, - { - "epoch": 1.0765532879818593, - "grad_norm": 0.1884765625, - "learning_rate": 0.00012982499082119694, - "loss": 0.887, - "step": 2968 - }, - { - "epoch": 1.0772789115646257, - "grad_norm": 0.15234375, - "learning_rate": 0.00012977603720474852, - "loss": 1.0116, - "step": 2970 - }, - { - "epoch": 1.0780045351473924, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001297270835883001, - "loss": 0.9688, - "step": 2972 - }, - { - "epoch": 1.0787301587301588, - "grad_norm": 0.1591796875, - "learning_rate": 0.00012967812997185168, - "loss": 1.024, - "step": 2974 - }, - { - "epoch": 1.0794557823129252, - "grad_norm": 0.154296875, - "learning_rate": 0.00012962917635540326, - "loss": 1.0757, - "step": 2976 - }, - { - "epoch": 1.0801814058956916, - "grad_norm": 0.1728515625, - "learning_rate": 0.00012958022273895485, - "loss": 0.9665, - "step": 2978 - }, - { - "epoch": 1.080907029478458, - "grad_norm": 0.18359375, - "learning_rate": 0.00012953126912250643, - "loss": 0.9317, - "step": 2980 - }, - { - "epoch": 1.0816326530612246, - "grad_norm": 0.1748046875, - "learning_rate": 0.00012948231550605804, - "loss": 0.9634, - "step": 2982 - }, - { - "epoch": 1.082358276643991, - "grad_norm": 0.154296875, - "learning_rate": 0.0001294333618896096, - "loss": 0.9887, - "step": 2984 - }, - { - "epoch": 1.0830839002267574, - "grad_norm": 0.15625, - "learning_rate": 0.00012938440827316118, - "loss": 0.8876, - "step": 2986 - }, - { - "epoch": 1.0838095238095238, - "grad_norm": 0.169921875, - "learning_rate": 0.00012933545465671279, - "loss": 0.9634, - "step": 2988 - }, - { - "epoch": 1.0845351473922902, - "grad_norm": 0.15625, - "learning_rate": 0.00012928650104026437, - "loss": 0.9834, - "step": 2990 - }, - { - "epoch": 1.0852607709750566, - "grad_norm": 0.154296875, - "learning_rate": 0.00012923754742381592, - "loss": 0.99, - "step": 2992 - }, - { - "epoch": 1.0859863945578232, - "grad_norm": 0.142578125, - "learning_rate": 0.00012918859380736753, - "loss": 0.9375, - "step": 2994 - }, - { - "epoch": 1.0867120181405896, - "grad_norm": 0.154296875, - "learning_rate": 0.00012913964019091911, - "loss": 0.9757, - "step": 2996 - }, - { - "epoch": 1.087437641723356, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001290906865744707, - "loss": 0.9934, - "step": 2998 - }, - { - "epoch": 1.0881632653061224, - "grad_norm": 0.1591796875, - "learning_rate": 0.00012904173295802228, - "loss": 0.9745, - "step": 3000 - }, - { - "epoch": 1.0881632653061224, - "eval_loss": 1.0154409408569336, - "eval_runtime": 92.0958, - "eval_samples_per_second": 19.545, - "eval_steps_per_second": 0.619, - "step": 3000 - }, - { - "epoch": 1.0888888888888888, - "grad_norm": 0.140625, - "learning_rate": 0.00012899277934157386, - "loss": 0.9854, - "step": 3002 - }, - { - "epoch": 1.0896145124716554, - "grad_norm": 0.19140625, - "learning_rate": 0.00012894382572512544, - "loss": 0.9796, - "step": 3004 - }, - { - "epoch": 1.0903401360544218, - "grad_norm": 0.166015625, - "learning_rate": 0.00012889487210867703, - "loss": 0.9706, - "step": 3006 - }, - { - "epoch": 1.0910657596371882, - "grad_norm": 0.15625, - "learning_rate": 0.00012884591849222864, - "loss": 0.9691, - "step": 3008 - }, - { - "epoch": 1.0917913832199546, - "grad_norm": 0.1728515625, - "learning_rate": 0.00012879696487578022, - "loss": 1.0203, - "step": 3010 - }, - { - "epoch": 1.092517006802721, - "grad_norm": 0.1865234375, - "learning_rate": 0.00012874801125933177, - "loss": 0.9979, - "step": 3012 - }, - { - "epoch": 1.0932426303854874, - "grad_norm": 0.166015625, - "learning_rate": 0.00012869905764288338, - "loss": 1.0656, - "step": 3014 - }, - { - "epoch": 1.093968253968254, - "grad_norm": 0.1591796875, - "learning_rate": 0.00012865010402643496, - "loss": 0.9869, - "step": 3016 - }, - { - "epoch": 1.0946938775510204, - "grad_norm": 0.162109375, - "learning_rate": 0.00012860115040998655, - "loss": 1.0094, - "step": 3018 - }, - { - "epoch": 1.0954195011337868, - "grad_norm": 0.1748046875, - "learning_rate": 0.00012855219679353813, - "loss": 1.0243, - "step": 3020 - }, - { - "epoch": 1.0961451247165532, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001285032431770897, - "loss": 0.9527, - "step": 3022 - }, - { - "epoch": 1.0968707482993196, - "grad_norm": 0.150390625, - "learning_rate": 0.0001284542895606413, - "loss": 1.0238, - "step": 3024 - }, - { - "epoch": 1.0975963718820863, - "grad_norm": 0.1708984375, - "learning_rate": 0.00012840533594419288, - "loss": 0.9843, - "step": 3026 - }, - { - "epoch": 1.0983219954648527, - "grad_norm": 0.15625, - "learning_rate": 0.00012835638232774448, - "loss": 0.9041, - "step": 3028 - }, - { - "epoch": 1.099047619047619, - "grad_norm": 0.1533203125, - "learning_rate": 0.00012830742871129607, - "loss": 0.9152, - "step": 3030 - }, - { - "epoch": 1.0997732426303855, - "grad_norm": 0.1669921875, - "learning_rate": 0.00012825847509484762, - "loss": 0.9771, - "step": 3032 - }, - { - "epoch": 1.1004988662131518, - "grad_norm": 0.1572265625, - "learning_rate": 0.00012820952147839923, - "loss": 0.952, - "step": 3034 - }, - { - "epoch": 1.1012244897959185, - "grad_norm": 0.17578125, - "learning_rate": 0.00012816056786195081, - "loss": 0.9888, - "step": 3036 - }, - { - "epoch": 1.1019501133786849, - "grad_norm": 0.1640625, - "learning_rate": 0.0001281116142455024, - "loss": 0.9741, - "step": 3038 - }, - { - "epoch": 1.1026757369614513, - "grad_norm": 0.1689453125, - "learning_rate": 0.00012806266062905398, - "loss": 0.9094, - "step": 3040 - }, - { - "epoch": 1.1034013605442177, - "grad_norm": 0.1728515625, - "learning_rate": 0.00012801370701260556, - "loss": 0.9386, - "step": 3042 - }, - { - "epoch": 1.104126984126984, - "grad_norm": 0.15234375, - "learning_rate": 0.00012796475339615714, - "loss": 0.9503, - "step": 3044 - }, - { - "epoch": 1.1048526077097505, - "grad_norm": 0.1455078125, - "learning_rate": 0.00012791579977970873, - "loss": 0.9595, - "step": 3046 - }, - { - "epoch": 1.105578231292517, - "grad_norm": 0.1513671875, - "learning_rate": 0.00012786684616326033, - "loss": 0.9771, - "step": 3048 - }, - { - "epoch": 1.1063038548752835, - "grad_norm": 0.17578125, - "learning_rate": 0.00012781789254681192, - "loss": 1.054, - "step": 3050 - }, - { - "epoch": 1.1070294784580499, - "grad_norm": 0.173828125, - "learning_rate": 0.00012776893893036347, - "loss": 0.9665, - "step": 3052 - }, - { - "epoch": 1.1077551020408163, - "grad_norm": 0.138671875, - "learning_rate": 0.00012771998531391508, - "loss": 1.0094, - "step": 3054 - }, - { - "epoch": 1.1084807256235827, - "grad_norm": 0.1650390625, - "learning_rate": 0.00012767103169746666, - "loss": 0.9462, - "step": 3056 - }, - { - "epoch": 1.1092063492063493, - "grad_norm": 0.1572265625, - "learning_rate": 0.00012762207808101825, - "loss": 1.0243, - "step": 3058 - }, - { - "epoch": 1.1099319727891157, - "grad_norm": 0.1513671875, - "learning_rate": 0.00012757312446456983, - "loss": 0.9708, - "step": 3060 - }, - { - "epoch": 1.110657596371882, - "grad_norm": 0.15625, - "learning_rate": 0.0001275241708481214, - "loss": 0.9986, - "step": 3062 - }, - { - "epoch": 1.1113832199546485, - "grad_norm": 0.1513671875, - "learning_rate": 0.000127475217231673, - "loss": 0.9911, - "step": 3064 - }, - { - "epoch": 1.112108843537415, - "grad_norm": 0.1787109375, - "learning_rate": 0.00012742626361522458, - "loss": 1.0371, - "step": 3066 - }, - { - "epoch": 1.1128344671201813, - "grad_norm": 0.14453125, - "learning_rate": 0.00012737730999877618, - "loss": 1.0138, - "step": 3068 - }, - { - "epoch": 1.113560090702948, - "grad_norm": 0.154296875, - "learning_rate": 0.00012732835638232774, - "loss": 0.9498, - "step": 3070 - }, - { - "epoch": 1.1142857142857143, - "grad_norm": 0.15625, - "learning_rate": 0.00012727940276587932, - "loss": 1.0003, - "step": 3072 - }, - { - "epoch": 1.1150113378684807, - "grad_norm": 0.1396484375, - "learning_rate": 0.00012723044914943093, - "loss": 1.0018, - "step": 3074 - }, - { - "epoch": 1.1157369614512471, - "grad_norm": 0.140625, - "learning_rate": 0.00012718149553298251, - "loss": 0.9182, - "step": 3076 - }, - { - "epoch": 1.1164625850340135, - "grad_norm": 0.166015625, - "learning_rate": 0.0001271325419165341, - "loss": 0.9443, - "step": 3078 - }, - { - "epoch": 1.1171882086167801, - "grad_norm": 0.171875, - "learning_rate": 0.00012708358830008568, - "loss": 1.0236, - "step": 3080 - }, - { - "epoch": 1.1179138321995465, - "grad_norm": 0.162109375, - "learning_rate": 0.00012703463468363726, - "loss": 0.977, - "step": 3082 - }, - { - "epoch": 1.118639455782313, - "grad_norm": 0.15234375, - "learning_rate": 0.00012698568106718884, - "loss": 0.9371, - "step": 3084 - }, - { - "epoch": 1.1193650793650793, - "grad_norm": 0.1767578125, - "learning_rate": 0.00012693672745074043, - "loss": 0.9309, - "step": 3086 - }, - { - "epoch": 1.1200907029478457, - "grad_norm": 0.1669921875, - "learning_rate": 0.00012688777383429203, - "loss": 0.9547, - "step": 3088 - }, - { - "epoch": 1.1208163265306124, - "grad_norm": 0.23828125, - "learning_rate": 0.0001268388202178436, - "loss": 1.0593, - "step": 3090 - }, - { - "epoch": 1.1215419501133788, - "grad_norm": 0.1748046875, - "learning_rate": 0.00012678986660139517, - "loss": 0.9916, - "step": 3092 - }, - { - "epoch": 1.1222675736961452, - "grad_norm": 0.171875, - "learning_rate": 0.00012674091298494678, - "loss": 0.9708, - "step": 3094 - }, - { - "epoch": 1.1229931972789116, - "grad_norm": 0.1904296875, - "learning_rate": 0.00012669195936849836, - "loss": 0.9498, - "step": 3096 - }, - { - "epoch": 1.123718820861678, - "grad_norm": 0.1728515625, - "learning_rate": 0.00012664300575204995, - "loss": 0.9388, - "step": 3098 - }, - { - "epoch": 1.1244444444444444, - "grad_norm": 0.1494140625, - "learning_rate": 0.00012659405213560153, - "loss": 0.9372, - "step": 3100 - }, - { - "epoch": 1.1244444444444444, - "eval_loss": 1.0154296159744263, - "eval_runtime": 89.182, - "eval_samples_per_second": 20.183, - "eval_steps_per_second": 0.639, - "step": 3100 - }, - { - "epoch": 1.125170068027211, - "grad_norm": 0.19140625, - "learning_rate": 0.0001265450985191531, - "loss": 0.9911, - "step": 3102 - }, - { - "epoch": 1.1258956916099774, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001264961449027047, - "loss": 1.0097, - "step": 3104 - }, - { - "epoch": 1.1266213151927438, - "grad_norm": 0.162109375, - "learning_rate": 0.00012644719128625628, - "loss": 0.9942, - "step": 3106 - }, - { - "epoch": 1.1273469387755102, - "grad_norm": 0.162109375, - "learning_rate": 0.00012639823766980788, - "loss": 0.9891, - "step": 3108 - }, - { - "epoch": 1.1280725623582766, - "grad_norm": 0.1689453125, - "learning_rate": 0.00012634928405335944, - "loss": 1.0132, - "step": 3110 - }, - { - "epoch": 1.1287981859410432, - "grad_norm": 0.1689453125, - "learning_rate": 0.00012630033043691102, - "loss": 0.9039, - "step": 3112 - }, - { - "epoch": 1.1295238095238096, - "grad_norm": 0.1767578125, - "learning_rate": 0.00012625137682046263, - "loss": 0.9398, - "step": 3114 - }, - { - "epoch": 1.130249433106576, - "grad_norm": 0.1923828125, - "learning_rate": 0.00012620242320401421, - "loss": 0.9643, - "step": 3116 - }, - { - "epoch": 1.1309750566893424, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001261534695875658, - "loss": 0.9252, - "step": 3118 - }, - { - "epoch": 1.1317006802721088, - "grad_norm": 0.17578125, - "learning_rate": 0.00012610451597111738, - "loss": 0.9849, - "step": 3120 - }, - { - "epoch": 1.1324263038548752, - "grad_norm": 0.162109375, - "learning_rate": 0.00012605556235466896, - "loss": 1.004, - "step": 3122 - }, - { - "epoch": 1.1331519274376418, - "grad_norm": 0.1845703125, - "learning_rate": 0.00012600660873822054, - "loss": 1.0351, - "step": 3124 - }, - { - "epoch": 1.1338775510204082, - "grad_norm": 0.1533203125, - "learning_rate": 0.00012595765512177213, - "loss": 0.9665, - "step": 3126 - }, - { - "epoch": 1.1346031746031746, - "grad_norm": 0.1552734375, - "learning_rate": 0.00012590870150532373, - "loss": 1.0028, - "step": 3128 - }, - { - "epoch": 1.135328798185941, - "grad_norm": 0.1689453125, - "learning_rate": 0.0001258597478888753, - "loss": 0.9811, - "step": 3130 - }, - { - "epoch": 1.1360544217687074, - "grad_norm": 0.1650390625, - "learning_rate": 0.00012581079427242687, - "loss": 1.01, - "step": 3132 - }, - { - "epoch": 1.136780045351474, - "grad_norm": 0.15625, - "learning_rate": 0.00012576184065597848, - "loss": 0.9534, - "step": 3134 - }, - { - "epoch": 1.1375056689342404, - "grad_norm": 0.1630859375, - "learning_rate": 0.00012571288703953006, - "loss": 0.9672, - "step": 3136 - }, - { - "epoch": 1.1382312925170068, - "grad_norm": 0.150390625, - "learning_rate": 0.00012566393342308162, - "loss": 1.0314, - "step": 3138 - }, - { - "epoch": 1.1389569160997732, - "grad_norm": 0.1904296875, - "learning_rate": 0.00012561497980663323, - "loss": 0.96, - "step": 3140 - }, - { - "epoch": 1.1396825396825396, - "grad_norm": 0.1630859375, - "learning_rate": 0.0001255660261901848, - "loss": 0.9722, - "step": 3142 - }, - { - "epoch": 1.140408163265306, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001255170725737364, - "loss": 1.063, - "step": 3144 - }, - { - "epoch": 1.1411337868480727, - "grad_norm": 0.1796875, - "learning_rate": 0.00012546811895728798, - "loss": 0.987, - "step": 3146 - }, - { - "epoch": 1.141859410430839, - "grad_norm": 0.1552734375, - "learning_rate": 0.00012541916534083956, - "loss": 0.9693, - "step": 3148 - }, - { - "epoch": 1.1425850340136054, - "grad_norm": 0.140625, - "learning_rate": 0.00012537021172439114, - "loss": 0.9687, - "step": 3150 - }, - { - "epoch": 1.1433106575963718, - "grad_norm": 0.1533203125, - "learning_rate": 0.00012532125810794272, - "loss": 0.9822, - "step": 3152 - }, - { - "epoch": 1.1440362811791382, - "grad_norm": 0.162109375, - "learning_rate": 0.00012527230449149433, - "loss": 0.9652, - "step": 3154 - }, - { - "epoch": 1.1447619047619049, - "grad_norm": 0.158203125, - "learning_rate": 0.0001252233508750459, - "loss": 0.8838, - "step": 3156 - }, - { - "epoch": 1.1454875283446713, - "grad_norm": 0.1611328125, - "learning_rate": 0.00012517439725859747, - "loss": 0.9707, - "step": 3158 - }, - { - "epoch": 1.1462131519274377, - "grad_norm": 0.16015625, - "learning_rate": 0.00012512544364214908, - "loss": 1.0355, - "step": 3160 - }, - { - "epoch": 1.146938775510204, - "grad_norm": 0.15625, - "learning_rate": 0.00012507649002570066, - "loss": 1.0814, - "step": 3162 - }, - { - "epoch": 1.1476643990929705, - "grad_norm": 0.15234375, - "learning_rate": 0.00012502753640925224, - "loss": 1.005, - "step": 3164 - }, - { - "epoch": 1.1483900226757369, - "grad_norm": 0.166015625, - "learning_rate": 0.00012497858279280382, - "loss": 0.9581, - "step": 3166 - }, - { - "epoch": 1.1491156462585035, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001249296291763554, - "loss": 1.0175, - "step": 3168 - }, - { - "epoch": 1.1498412698412699, - "grad_norm": 0.16796875, - "learning_rate": 0.000124880675559907, - "loss": 0.9524, - "step": 3170 - }, - { - "epoch": 1.1505668934240363, - "grad_norm": 0.158203125, - "learning_rate": 0.00012483172194345857, - "loss": 1.0104, - "step": 3172 - }, - { - "epoch": 1.1512925170068027, - "grad_norm": 0.169921875, - "learning_rate": 0.00012478276832701018, - "loss": 0.9864, - "step": 3174 - }, - { - "epoch": 1.152018140589569, - "grad_norm": 0.1572265625, - "learning_rate": 0.00012473381471056176, - "loss": 1.0231, - "step": 3176 - }, - { - "epoch": 1.1527437641723357, - "grad_norm": 0.162109375, - "learning_rate": 0.00012468486109411332, - "loss": 0.9632, - "step": 3178 - }, - { - "epoch": 1.153469387755102, - "grad_norm": 0.1806640625, - "learning_rate": 0.00012463590747766493, - "loss": 0.9549, - "step": 3180 - }, - { - "epoch": 1.1541950113378685, - "grad_norm": 0.150390625, - "learning_rate": 0.0001245869538612165, - "loss": 1.0362, - "step": 3182 - }, - { - "epoch": 1.154920634920635, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001245380002447681, - "loss": 0.9991, - "step": 3184 - }, - { - "epoch": 1.1556462585034013, - "grad_norm": 0.1396484375, - "learning_rate": 0.00012448904662831967, - "loss": 0.9353, - "step": 3186 - }, - { - "epoch": 1.1563718820861677, - "grad_norm": 0.1494140625, - "learning_rate": 0.00012444009301187126, - "loss": 0.9651, - "step": 3188 - }, - { - "epoch": 1.1570975056689343, - "grad_norm": 0.1484375, - "learning_rate": 0.00012439113939542284, - "loss": 0.9336, - "step": 3190 - }, - { - "epoch": 1.1578231292517007, - "grad_norm": 0.166015625, - "learning_rate": 0.00012434218577897442, - "loss": 0.9326, - "step": 3192 - }, - { - "epoch": 1.1585487528344671, - "grad_norm": 0.1767578125, - "learning_rate": 0.00012429323216252603, - "loss": 1.0304, - "step": 3194 - }, - { - "epoch": 1.1592743764172335, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001242442785460776, - "loss": 0.9833, - "step": 3196 - }, - { - "epoch": 1.16, - "grad_norm": 0.1533203125, - "learning_rate": 0.00012419532492962917, - "loss": 1.0228, - "step": 3198 - }, - { - "epoch": 1.1607256235827665, - "grad_norm": 0.16015625, - "learning_rate": 0.00012414637131318078, - "loss": 0.988, - "step": 3200 - }, - { - "epoch": 1.1607256235827665, - "eval_loss": 1.014972448348999, - "eval_runtime": 94.878, - "eval_samples_per_second": 18.972, - "eval_steps_per_second": 0.601, - "step": 3200 - }, - { - "epoch": 1.161451247165533, - "grad_norm": 0.1552734375, - "learning_rate": 0.00012409741769673236, - "loss": 0.9443, - "step": 3202 - }, - { - "epoch": 1.1621768707482993, - "grad_norm": 0.171875, - "learning_rate": 0.00012404846408028394, - "loss": 1.0144, - "step": 3204 - }, - { - "epoch": 1.1629024943310657, - "grad_norm": 0.166015625, - "learning_rate": 0.0001239995104638355, - "loss": 0.8949, - "step": 3206 - }, - { - "epoch": 1.1636281179138321, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001239505568473871, - "loss": 0.9979, - "step": 3208 - }, - { - "epoch": 1.1643537414965985, - "grad_norm": 0.173828125, - "learning_rate": 0.0001239016032309387, - "loss": 0.985, - "step": 3210 - }, - { - "epoch": 1.1650793650793652, - "grad_norm": 0.1689453125, - "learning_rate": 0.00012385264961449027, - "loss": 1.018, - "step": 3212 - }, - { - "epoch": 1.1658049886621316, - "grad_norm": 0.1904296875, - "learning_rate": 0.00012380369599804188, - "loss": 0.9736, - "step": 3214 - }, - { - "epoch": 1.166530612244898, - "grad_norm": 0.1728515625, - "learning_rate": 0.00012375474238159344, - "loss": 0.9888, - "step": 3216 - }, - { - "epoch": 1.1672562358276644, - "grad_norm": 0.166015625, - "learning_rate": 0.00012370578876514502, - "loss": 0.9882, - "step": 3218 - }, - { - "epoch": 1.1679818594104308, - "grad_norm": 0.162109375, - "learning_rate": 0.00012365683514869663, - "loss": 1.0883, - "step": 3220 - }, - { - "epoch": 1.1687074829931974, - "grad_norm": 0.185546875, - "learning_rate": 0.0001236078815322482, - "loss": 1.0099, - "step": 3222 - }, - { - "epoch": 1.1694331065759638, - "grad_norm": 0.203125, - "learning_rate": 0.0001235589279157998, - "loss": 0.991, - "step": 3224 - }, - { - "epoch": 1.1701587301587302, - "grad_norm": 0.1533203125, - "learning_rate": 0.00012350997429935135, - "loss": 0.9971, - "step": 3226 - }, - { - "epoch": 1.1708843537414966, - "grad_norm": 0.181640625, - "learning_rate": 0.00012346102068290296, - "loss": 1.0944, - "step": 3228 - }, - { - "epoch": 1.171609977324263, - "grad_norm": 0.1533203125, - "learning_rate": 0.00012341206706645454, - "loss": 1.0144, - "step": 3230 - }, - { - "epoch": 1.1723356009070294, - "grad_norm": 0.158203125, - "learning_rate": 0.00012336311345000612, - "loss": 0.9596, - "step": 3232 - }, - { - "epoch": 1.173061224489796, - "grad_norm": 0.16796875, - "learning_rate": 0.00012331415983355773, - "loss": 1.0307, - "step": 3234 - }, - { - "epoch": 1.1737868480725624, - "grad_norm": 0.1787109375, - "learning_rate": 0.00012326520621710929, - "loss": 0.9639, - "step": 3236 - }, - { - "epoch": 1.1745124716553288, - "grad_norm": 0.1591796875, - "learning_rate": 0.00012321625260066087, - "loss": 0.9618, - "step": 3238 - }, - { - "epoch": 1.1752380952380952, - "grad_norm": 0.1796875, - "learning_rate": 0.00012316729898421248, - "loss": 0.9739, - "step": 3240 - }, - { - "epoch": 1.1759637188208616, - "grad_norm": 0.173828125, - "learning_rate": 0.00012311834536776406, - "loss": 1.0316, - "step": 3242 - }, - { - "epoch": 1.1766893424036282, - "grad_norm": 0.15625, - "learning_rate": 0.00012306939175131564, - "loss": 0.9901, - "step": 3244 - }, - { - "epoch": 1.1774149659863946, - "grad_norm": 0.173828125, - "learning_rate": 0.0001230204381348672, - "loss": 1.0439, - "step": 3246 - }, - { - "epoch": 1.178140589569161, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001229714845184188, - "loss": 0.9763, - "step": 3248 - }, - { - "epoch": 1.1788662131519274, - "grad_norm": 0.1767578125, - "learning_rate": 0.0001229225309019704, - "loss": 0.9723, - "step": 3250 - }, - { - "epoch": 1.1795918367346938, - "grad_norm": 0.1552734375, - "learning_rate": 0.00012287357728552197, - "loss": 1.1122, - "step": 3252 - }, - { - "epoch": 1.1803174603174602, - "grad_norm": 0.18359375, - "learning_rate": 0.00012282462366907358, - "loss": 1.0547, - "step": 3254 - }, - { - "epoch": 1.1810430839002268, - "grad_norm": 0.142578125, - "learning_rate": 0.00012277567005262514, - "loss": 0.9226, - "step": 3256 - }, - { - "epoch": 1.1817687074829932, - "grad_norm": 0.1396484375, - "learning_rate": 0.00012272671643617672, - "loss": 0.9439, - "step": 3258 - }, - { - "epoch": 1.1824943310657596, - "grad_norm": 0.15234375, - "learning_rate": 0.00012267776281972833, - "loss": 1.0307, - "step": 3260 - }, - { - "epoch": 1.183219954648526, - "grad_norm": 0.1669921875, - "learning_rate": 0.0001226288092032799, - "loss": 1.0502, - "step": 3262 - }, - { - "epoch": 1.1839455782312924, - "grad_norm": 0.1923828125, - "learning_rate": 0.0001225798555868315, - "loss": 1.0492, - "step": 3264 - }, - { - "epoch": 1.184671201814059, - "grad_norm": 0.1533203125, - "learning_rate": 0.00012253090197038305, - "loss": 0.962, - "step": 3266 - }, - { - "epoch": 1.1853968253968254, - "grad_norm": 0.166015625, - "learning_rate": 0.00012248194835393466, - "loss": 0.9904, - "step": 3268 - }, - { - "epoch": 1.1861224489795918, - "grad_norm": 0.146484375, - "learning_rate": 0.00012243299473748624, - "loss": 0.9677, - "step": 3270 - }, - { - "epoch": 1.1868480725623582, - "grad_norm": 0.1611328125, - "learning_rate": 0.00012238404112103782, - "loss": 1.0684, - "step": 3272 - }, - { - "epoch": 1.1875736961451246, - "grad_norm": 0.169921875, - "learning_rate": 0.00012233508750458943, - "loss": 1.0299, - "step": 3274 - }, - { - "epoch": 1.188299319727891, - "grad_norm": 0.1572265625, - "learning_rate": 0.00012228613388814099, - "loss": 0.98, - "step": 3276 - }, - { - "epoch": 1.1890249433106577, - "grad_norm": 0.1591796875, - "learning_rate": 0.00012223718027169257, - "loss": 1.0087, - "step": 3278 - }, - { - "epoch": 1.189750566893424, - "grad_norm": 0.166015625, - "learning_rate": 0.00012218822665524418, - "loss": 1.0092, - "step": 3280 - }, - { - "epoch": 1.1904761904761905, - "grad_norm": 0.162109375, - "learning_rate": 0.00012213927303879576, - "loss": 1.0301, - "step": 3282 - }, - { - "epoch": 1.1912018140589569, - "grad_norm": 0.154296875, - "learning_rate": 0.00012209031942234731, - "loss": 1.0265, - "step": 3284 - }, - { - "epoch": 1.1919274376417233, - "grad_norm": 0.25, - "learning_rate": 0.00012204136580589891, - "loss": 0.9792, - "step": 3286 - }, - { - "epoch": 1.1926530612244899, - "grad_norm": 0.1640625, - "learning_rate": 0.0001219924121894505, - "loss": 0.9287, - "step": 3288 - }, - { - "epoch": 1.1933786848072563, - "grad_norm": 0.1650390625, - "learning_rate": 0.00012194345857300209, - "loss": 0.9637, - "step": 3290 - }, - { - "epoch": 1.1941043083900227, - "grad_norm": 0.1640625, - "learning_rate": 0.00012189450495655368, - "loss": 0.984, - "step": 3292 - }, - { - "epoch": 1.194829931972789, - "grad_norm": 0.1708984375, - "learning_rate": 0.00012184555134010524, - "loss": 1.0349, - "step": 3294 - }, - { - "epoch": 1.1955555555555555, - "grad_norm": 0.1357421875, - "learning_rate": 0.00012179659772365684, - "loss": 0.9786, - "step": 3296 - }, - { - "epoch": 1.1962811791383219, - "grad_norm": 0.19921875, - "learning_rate": 0.00012174764410720843, - "loss": 0.9701, - "step": 3298 - }, - { - "epoch": 1.1970068027210885, - "grad_norm": 0.1572265625, - "learning_rate": 0.00012169869049076001, - "loss": 1.0021, - "step": 3300 - }, - { - "epoch": 1.1970068027210885, - "eval_loss": 1.014991044998169, - "eval_runtime": 91.949, - "eval_samples_per_second": 19.576, - "eval_steps_per_second": 0.62, - "step": 3300 - }, - { - "epoch": 1.197732426303855, - "grad_norm": 0.1533203125, - "learning_rate": 0.00012164973687431161, - "loss": 0.8863, - "step": 3302 - }, - { - "epoch": 1.1984580498866213, - "grad_norm": 0.1650390625, - "learning_rate": 0.00012160078325786316, - "loss": 1.0103, - "step": 3304 - }, - { - "epoch": 1.1991836734693877, - "grad_norm": 0.1640625, - "learning_rate": 0.00012155182964141476, - "loss": 0.9499, - "step": 3306 - }, - { - "epoch": 1.199909297052154, - "grad_norm": 0.169921875, - "learning_rate": 0.00012150287602496636, - "loss": 1.0125, - "step": 3308 - }, - { - "epoch": 1.2006349206349207, - "grad_norm": 0.16796875, - "learning_rate": 0.00012145392240851794, - "loss": 0.9703, - "step": 3310 - }, - { - "epoch": 1.2013605442176871, - "grad_norm": 0.1767578125, - "learning_rate": 0.00012140496879206953, - "loss": 0.9431, - "step": 3312 - }, - { - "epoch": 1.2020861678004535, - "grad_norm": 0.1474609375, - "learning_rate": 0.00012135601517562109, - "loss": 0.9832, - "step": 3314 - }, - { - "epoch": 1.20281179138322, - "grad_norm": 0.1611328125, - "learning_rate": 0.00012130706155917269, - "loss": 0.9843, - "step": 3316 - }, - { - "epoch": 1.2035374149659863, - "grad_norm": 0.1494140625, - "learning_rate": 0.00012125810794272428, - "loss": 0.9591, - "step": 3318 - }, - { - "epoch": 1.2042630385487527, - "grad_norm": 0.150390625, - "learning_rate": 0.00012120915432627586, - "loss": 0.9599, - "step": 3320 - }, - { - "epoch": 1.2049886621315193, - "grad_norm": 0.158203125, - "learning_rate": 0.00012116020070982746, - "loss": 1.0021, - "step": 3322 - }, - { - "epoch": 1.2057142857142857, - "grad_norm": 0.166015625, - "learning_rate": 0.00012111124709337901, - "loss": 1.0136, - "step": 3324 - }, - { - "epoch": 1.2064399092970521, - "grad_norm": 0.205078125, - "learning_rate": 0.00012106229347693061, - "loss": 0.9422, - "step": 3326 - }, - { - "epoch": 1.2071655328798185, - "grad_norm": 0.1806640625, - "learning_rate": 0.0001210133398604822, - "loss": 1.0347, - "step": 3328 - }, - { - "epoch": 1.2078911564625852, - "grad_norm": 0.1591796875, - "learning_rate": 0.00012096438624403379, - "loss": 1.0006, - "step": 3330 - }, - { - "epoch": 1.2086167800453516, - "grad_norm": 0.158203125, - "learning_rate": 0.00012091543262758538, - "loss": 0.9784, - "step": 3332 - }, - { - "epoch": 1.209342403628118, - "grad_norm": 0.166015625, - "learning_rate": 0.00012086647901113694, - "loss": 1.0127, - "step": 3334 - }, - { - "epoch": 1.2100680272108844, - "grad_norm": 0.15234375, - "learning_rate": 0.00012081752539468854, - "loss": 1.0538, - "step": 3336 - }, - { - "epoch": 1.2107936507936508, - "grad_norm": 0.1591796875, - "learning_rate": 0.00012076857177824013, - "loss": 0.9858, - "step": 3338 - }, - { - "epoch": 1.2115192743764172, - "grad_norm": 0.1806640625, - "learning_rate": 0.00012071961816179171, - "loss": 0.978, - "step": 3340 - }, - { - "epoch": 1.2122448979591836, - "grad_norm": 0.1669921875, - "learning_rate": 0.00012067066454534331, - "loss": 0.9754, - "step": 3342 - }, - { - "epoch": 1.2129705215419502, - "grad_norm": 0.1728515625, - "learning_rate": 0.00012062171092889486, - "loss": 0.9934, - "step": 3344 - }, - { - "epoch": 1.2136961451247166, - "grad_norm": 0.1689453125, - "learning_rate": 0.00012057275731244646, - "loss": 0.9599, - "step": 3346 - }, - { - "epoch": 1.214421768707483, - "grad_norm": 0.1611328125, - "learning_rate": 0.00012052380369599806, - "loss": 1.0248, - "step": 3348 - }, - { - "epoch": 1.2151473922902494, - "grad_norm": 0.1611328125, - "learning_rate": 0.00012047485007954964, - "loss": 1.0095, - "step": 3350 - }, - { - "epoch": 1.215873015873016, - "grad_norm": 0.154296875, - "learning_rate": 0.00012042589646310121, - "loss": 0.9968, - "step": 3352 - }, - { - "epoch": 1.2165986394557824, - "grad_norm": 0.16796875, - "learning_rate": 0.00012037694284665279, - "loss": 0.9093, - "step": 3354 - }, - { - "epoch": 1.2173242630385488, - "grad_norm": 0.142578125, - "learning_rate": 0.00012032798923020439, - "loss": 0.9017, - "step": 3356 - }, - { - "epoch": 1.2180498866213152, - "grad_norm": 0.1640625, - "learning_rate": 0.00012027903561375598, - "loss": 0.9775, - "step": 3358 - }, - { - "epoch": 1.2187755102040816, - "grad_norm": 0.169921875, - "learning_rate": 0.00012023008199730756, - "loss": 1.0245, - "step": 3360 - }, - { - "epoch": 1.219501133786848, - "grad_norm": 0.158203125, - "learning_rate": 0.00012018112838085913, - "loss": 1.0026, - "step": 3362 - }, - { - "epoch": 1.2202267573696144, - "grad_norm": 0.1748046875, - "learning_rate": 0.00012013217476441071, - "loss": 0.991, - "step": 3364 - }, - { - "epoch": 1.220952380952381, - "grad_norm": 0.1787109375, - "learning_rate": 0.00012008322114796231, - "loss": 0.9144, - "step": 3366 - }, - { - "epoch": 1.2216780045351474, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001200342675315139, - "loss": 0.9246, - "step": 3368 - }, - { - "epoch": 1.2224036281179138, - "grad_norm": 0.1630859375, - "learning_rate": 0.00011998531391506549, - "loss": 0.991, - "step": 3370 - }, - { - "epoch": 1.2231292517006802, - "grad_norm": 0.16015625, - "learning_rate": 0.00011993636029861706, - "loss": 1.0354, - "step": 3372 - }, - { - "epoch": 1.2238548752834468, - "grad_norm": 0.1513671875, - "learning_rate": 0.00011988740668216864, - "loss": 1.0143, - "step": 3374 - }, - { - "epoch": 1.2245804988662132, - "grad_norm": 0.259765625, - "learning_rate": 0.00011983845306572024, - "loss": 1.0311, - "step": 3376 - }, - { - "epoch": 1.2253061224489796, - "grad_norm": 0.1572265625, - "learning_rate": 0.00011978949944927183, - "loss": 0.9852, - "step": 3378 - }, - { - "epoch": 1.226031746031746, - "grad_norm": 0.18359375, - "learning_rate": 0.00011974054583282341, - "loss": 0.9433, - "step": 3380 - }, - { - "epoch": 1.2267573696145124, - "grad_norm": 0.1552734375, - "learning_rate": 0.00011969159221637498, - "loss": 0.9716, - "step": 3382 - }, - { - "epoch": 1.2274829931972788, - "grad_norm": 0.15234375, - "learning_rate": 0.00011964263859992656, - "loss": 0.9286, - "step": 3384 - }, - { - "epoch": 1.2282086167800454, - "grad_norm": 0.142578125, - "learning_rate": 0.00011959368498347816, - "loss": 0.9936, - "step": 3386 - }, - { - "epoch": 1.2289342403628118, - "grad_norm": 0.16015625, - "learning_rate": 0.00011954473136702976, - "loss": 0.9766, - "step": 3388 - }, - { - "epoch": 1.2296598639455782, - "grad_norm": 0.1630859375, - "learning_rate": 0.00011949577775058134, - "loss": 0.9575, - "step": 3390 - }, - { - "epoch": 1.2303854875283446, - "grad_norm": 0.1650390625, - "learning_rate": 0.00011944682413413291, - "loss": 0.9791, - "step": 3392 - }, - { - "epoch": 1.231111111111111, - "grad_norm": 0.1806640625, - "learning_rate": 0.00011939787051768449, - "loss": 1.0045, - "step": 3394 - }, - { - "epoch": 1.2318367346938777, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011934891690123609, - "loss": 1.0042, - "step": 3396 - }, - { - "epoch": 1.232562358276644, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011929996328478768, - "loss": 0.9742, - "step": 3398 - }, - { - "epoch": 1.2332879818594105, - "grad_norm": 0.1904296875, - "learning_rate": 0.00011925100966833926, - "loss": 0.9748, - "step": 3400 - }, - { - "epoch": 1.2332879818594105, - "eval_loss": 1.0148086547851562, - "eval_runtime": 94.1179, - "eval_samples_per_second": 19.125, - "eval_steps_per_second": 0.606, - "step": 3400 - }, - { - "epoch": 1.2340136054421769, - "grad_norm": 0.150390625, - "learning_rate": 0.00011920205605189083, - "loss": 0.9734, - "step": 3402 - }, - { - "epoch": 1.2347392290249433, - "grad_norm": 0.150390625, - "learning_rate": 0.00011915310243544241, - "loss": 0.9926, - "step": 3404 - }, - { - "epoch": 1.2354648526077097, - "grad_norm": 0.162109375, - "learning_rate": 0.00011910414881899401, - "loss": 0.9998, - "step": 3406 - }, - { - "epoch": 1.2361904761904763, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001190551952025456, - "loss": 0.9564, - "step": 3408 - }, - { - "epoch": 1.2369160997732427, - "grad_norm": 0.1572265625, - "learning_rate": 0.00011900624158609719, - "loss": 0.9322, - "step": 3410 - }, - { - "epoch": 1.237641723356009, - "grad_norm": 0.1494140625, - "learning_rate": 0.00011895728796964876, - "loss": 0.9525, - "step": 3412 - }, - { - "epoch": 1.2383673469387755, - "grad_norm": 0.1669921875, - "learning_rate": 0.00011890833435320034, - "loss": 1.0165, - "step": 3414 - }, - { - "epoch": 1.2390929705215419, - "grad_norm": 0.15625, - "learning_rate": 0.00011885938073675194, - "loss": 0.9926, - "step": 3416 - }, - { - "epoch": 1.2398185941043085, - "grad_norm": 0.1611328125, - "learning_rate": 0.00011881042712030353, - "loss": 0.9599, - "step": 3418 - }, - { - "epoch": 1.240544217687075, - "grad_norm": 0.1669921875, - "learning_rate": 0.00011876147350385511, - "loss": 0.9723, - "step": 3420 - }, - { - "epoch": 1.2412698412698413, - "grad_norm": 0.16796875, - "learning_rate": 0.00011871251988740668, - "loss": 1.0077, - "step": 3422 - }, - { - "epoch": 1.2419954648526077, - "grad_norm": 0.1533203125, - "learning_rate": 0.00011866356627095826, - "loss": 0.9708, - "step": 3424 - }, - { - "epoch": 1.242721088435374, - "grad_norm": 0.17578125, - "learning_rate": 0.00011861461265450986, - "loss": 1.0205, - "step": 3426 - }, - { - "epoch": 1.2434467120181405, - "grad_norm": 0.166015625, - "learning_rate": 0.00011856565903806146, - "loss": 0.9795, - "step": 3428 - }, - { - "epoch": 1.2441723356009071, - "grad_norm": 0.1552734375, - "learning_rate": 0.00011851670542161301, - "loss": 1.0165, - "step": 3430 - }, - { - "epoch": 1.2448979591836735, - "grad_norm": 0.1708984375, - "learning_rate": 0.00011846775180516461, - "loss": 1.0622, - "step": 3432 - }, - { - "epoch": 1.24562358276644, - "grad_norm": 0.169921875, - "learning_rate": 0.00011841879818871619, - "loss": 1.0099, - "step": 3434 - }, - { - "epoch": 1.2463492063492063, - "grad_norm": 0.1455078125, - "learning_rate": 0.00011836984457226779, - "loss": 0.9006, - "step": 3436 - }, - { - "epoch": 1.2470748299319727, - "grad_norm": 0.16015625, - "learning_rate": 0.00011832089095581938, - "loss": 1.0503, - "step": 3438 - }, - { - "epoch": 1.2478004535147393, - "grad_norm": 0.158203125, - "learning_rate": 0.00011827193733937094, - "loss": 0.9561, - "step": 3440 - }, - { - "epoch": 1.2485260770975057, - "grad_norm": 0.15234375, - "learning_rate": 0.00011822298372292253, - "loss": 0.9966, - "step": 3442 - }, - { - "epoch": 1.2492517006802721, - "grad_norm": 0.1552734375, - "learning_rate": 0.00011817403010647411, - "loss": 1.0036, - "step": 3444 - }, - { - "epoch": 1.2499773242630385, - "grad_norm": 0.1669921875, - "learning_rate": 0.00011812507649002571, - "loss": 1.0254, - "step": 3446 - }, - { - "epoch": 1.250702947845805, - "grad_norm": 0.1396484375, - "learning_rate": 0.0001180761228735773, - "loss": 0.9544, - "step": 3448 - }, - { - "epoch": 1.2514285714285713, - "grad_norm": 0.17578125, - "learning_rate": 0.00011802716925712886, - "loss": 0.9625, - "step": 3450 - }, - { - "epoch": 1.2521541950113377, - "grad_norm": 0.142578125, - "learning_rate": 0.00011797821564068046, - "loss": 0.9375, - "step": 3452 - }, - { - "epoch": 1.2528798185941044, - "grad_norm": 0.16796875, - "learning_rate": 0.00011792926202423204, - "loss": 1.0041, - "step": 3454 - }, - { - "epoch": 1.2536054421768708, - "grad_norm": 0.1455078125, - "learning_rate": 0.00011788030840778364, - "loss": 0.9661, - "step": 3456 - }, - { - "epoch": 1.2543310657596372, - "grad_norm": 0.1748046875, - "learning_rate": 0.00011783135479133523, - "loss": 1.0311, - "step": 3458 - }, - { - "epoch": 1.2550566893424036, - "grad_norm": 0.146484375, - "learning_rate": 0.00011778240117488679, - "loss": 0.9824, - "step": 3460 - }, - { - "epoch": 1.2557823129251702, - "grad_norm": 0.16796875, - "learning_rate": 0.00011773344755843838, - "loss": 0.9351, - "step": 3462 - }, - { - "epoch": 1.2565079365079366, - "grad_norm": 0.181640625, - "learning_rate": 0.00011768449394198996, - "loss": 1.0692, - "step": 3464 - }, - { - "epoch": 1.257233560090703, - "grad_norm": 0.16796875, - "learning_rate": 0.00011763554032554156, - "loss": 1.0367, - "step": 3466 - }, - { - "epoch": 1.2579591836734694, - "grad_norm": 0.162109375, - "learning_rate": 0.00011758658670909316, - "loss": 0.9531, - "step": 3468 - }, - { - "epoch": 1.2586848072562358, - "grad_norm": 0.16015625, - "learning_rate": 0.00011753763309264471, - "loss": 1.0415, - "step": 3470 - }, - { - "epoch": 1.2594104308390022, - "grad_norm": 0.16015625, - "learning_rate": 0.00011748867947619631, - "loss": 0.9577, - "step": 3472 - }, - { - "epoch": 1.2601360544217686, - "grad_norm": 0.1552734375, - "learning_rate": 0.00011743972585974789, - "loss": 1.0151, - "step": 3474 - }, - { - "epoch": 1.2608616780045352, - "grad_norm": 0.158203125, - "learning_rate": 0.00011739077224329948, - "loss": 0.9575, - "step": 3476 - }, - { - "epoch": 1.2615873015873016, - "grad_norm": 0.150390625, - "learning_rate": 0.00011734181862685108, - "loss": 0.9676, - "step": 3478 - }, - { - "epoch": 1.262312925170068, - "grad_norm": 0.1611328125, - "learning_rate": 0.00011729286501040264, - "loss": 0.907, - "step": 3480 - }, - { - "epoch": 1.2630385487528344, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011724391139395423, - "loss": 0.9871, - "step": 3482 - }, - { - "epoch": 1.263764172335601, - "grad_norm": 0.16015625, - "learning_rate": 0.00011719495777750581, - "loss": 0.9981, - "step": 3484 - }, - { - "epoch": 1.2644897959183674, - "grad_norm": 0.1552734375, - "learning_rate": 0.00011714600416105741, - "loss": 0.993, - "step": 3486 - }, - { - "epoch": 1.2652154195011338, - "grad_norm": 0.1787109375, - "learning_rate": 0.000117097050544609, - "loss": 1.0024, - "step": 3488 - }, - { - "epoch": 1.2659410430839002, - "grad_norm": 0.1748046875, - "learning_rate": 0.00011704809692816056, - "loss": 0.9617, - "step": 3490 - }, - { - "epoch": 1.2666666666666666, - "grad_norm": 0.1640625, - "learning_rate": 0.00011699914331171216, - "loss": 0.9186, - "step": 3492 - }, - { - "epoch": 1.267392290249433, - "grad_norm": 0.162109375, - "learning_rate": 0.00011695018969526374, - "loss": 0.9795, - "step": 3494 - }, - { - "epoch": 1.2681179138321996, - "grad_norm": 0.1669921875, - "learning_rate": 0.00011690123607881533, - "loss": 1.0009, - "step": 3496 - }, - { - "epoch": 1.268843537414966, - "grad_norm": 0.171875, - "learning_rate": 0.0001168522824623669, - "loss": 1.001, - "step": 3498 - }, - { - "epoch": 1.2695691609977324, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011680332884591849, - "loss": 1.0089, - "step": 3500 - }, - { - "epoch": 1.2695691609977324, - "eval_loss": 1.0145820379257202, - "eval_runtime": 87.8413, - "eval_samples_per_second": 20.491, - "eval_steps_per_second": 0.649, - "step": 3500 - }, - { - "epoch": 1.2702947845804988, - "grad_norm": 0.1767578125, - "learning_rate": 0.00011675437522947008, - "loss": 1.0022, - "step": 3502 - }, - { - "epoch": 1.2710204081632652, - "grad_norm": 0.154296875, - "learning_rate": 0.00011670542161302166, - "loss": 1.0038, - "step": 3504 - }, - { - "epoch": 1.2717460317460318, - "grad_norm": 0.158203125, - "learning_rate": 0.00011665646799657326, - "loss": 0.9343, - "step": 3506 - }, - { - "epoch": 1.2724716553287982, - "grad_norm": 0.16015625, - "learning_rate": 0.00011660751438012483, - "loss": 1.0078, - "step": 3508 - }, - { - "epoch": 1.2731972789115646, - "grad_norm": 0.154296875, - "learning_rate": 0.00011655856076367641, - "loss": 0.9831, - "step": 3510 - }, - { - "epoch": 1.273922902494331, - "grad_norm": 0.1640625, - "learning_rate": 0.000116509607147228, - "loss": 1.0197, - "step": 3512 - }, - { - "epoch": 1.2746485260770974, - "grad_norm": 0.162109375, - "learning_rate": 0.00011646065353077959, - "loss": 1.0065, - "step": 3514 - }, - { - "epoch": 1.2753741496598638, - "grad_norm": 0.16015625, - "learning_rate": 0.00011641169991433118, - "loss": 0.9438, - "step": 3516 - }, - { - "epoch": 1.2760997732426305, - "grad_norm": 0.1708984375, - "learning_rate": 0.00011636274629788275, - "loss": 1.0004, - "step": 3518 - }, - { - "epoch": 1.2768253968253969, - "grad_norm": 0.1708984375, - "learning_rate": 0.00011631379268143434, - "loss": 0.9518, - "step": 3520 - }, - { - "epoch": 1.2775510204081633, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011626483906498593, - "loss": 0.9643, - "step": 3522 - }, - { - "epoch": 1.2782766439909297, - "grad_norm": 0.1552734375, - "learning_rate": 0.00011621588544853751, - "loss": 0.977, - "step": 3524 - }, - { - "epoch": 1.2790022675736963, - "grad_norm": 0.1708984375, - "learning_rate": 0.00011616693183208911, - "loss": 1.0163, - "step": 3526 - }, - { - "epoch": 1.2797278911564627, - "grad_norm": 0.2109375, - "learning_rate": 0.00011611797821564068, - "loss": 0.9853, - "step": 3528 - }, - { - "epoch": 1.280453514739229, - "grad_norm": 0.1611328125, - "learning_rate": 0.00011606902459919226, - "loss": 1.0327, - "step": 3530 - }, - { - "epoch": 1.2811791383219955, - "grad_norm": 0.1708984375, - "learning_rate": 0.00011602007098274386, - "loss": 1.039, - "step": 3532 - }, - { - "epoch": 1.2819047619047619, - "grad_norm": 0.1640625, - "learning_rate": 0.00011597111736629544, - "loss": 1.0176, - "step": 3534 - }, - { - "epoch": 1.2826303854875283, - "grad_norm": 0.1533203125, - "learning_rate": 0.00011592216374984703, - "loss": 0.9863, - "step": 3536 - }, - { - "epoch": 1.2833560090702947, - "grad_norm": 0.1640625, - "learning_rate": 0.0001158732101333986, - "loss": 0.9882, - "step": 3538 - }, - { - "epoch": 1.2840816326530613, - "grad_norm": 0.1630859375, - "learning_rate": 0.00011582425651695019, - "loss": 1.0326, - "step": 3540 - }, - { - "epoch": 1.2848072562358277, - "grad_norm": 0.171875, - "learning_rate": 0.00011577530290050178, - "loss": 0.9556, - "step": 3542 - }, - { - "epoch": 1.285532879818594, - "grad_norm": 0.1484375, - "learning_rate": 0.00011572634928405336, - "loss": 0.9931, - "step": 3544 - }, - { - "epoch": 1.2862585034013605, - "grad_norm": 0.16796875, - "learning_rate": 0.00011567739566760496, - "loss": 0.9452, - "step": 3546 - }, - { - "epoch": 1.2869841269841271, - "grad_norm": 0.1669921875, - "learning_rate": 0.00011562844205115653, - "loss": 0.9842, - "step": 3548 - }, - { - "epoch": 1.2877097505668935, - "grad_norm": 0.197265625, - "learning_rate": 0.00011557948843470811, - "loss": 1.0206, - "step": 3550 - }, - { - "epoch": 1.28843537414966, - "grad_norm": 0.134765625, - "learning_rate": 0.0001155305348182597, - "loss": 0.909, - "step": 3552 - }, - { - "epoch": 1.2891609977324263, - "grad_norm": 0.1640625, - "learning_rate": 0.00011548158120181129, - "loss": 1.0186, - "step": 3554 - }, - { - "epoch": 1.2898866213151927, - "grad_norm": 0.1396484375, - "learning_rate": 0.00011543262758536288, - "loss": 1.0197, - "step": 3556 - }, - { - "epoch": 1.2906122448979591, - "grad_norm": 0.1669921875, - "learning_rate": 0.00011538367396891445, - "loss": 0.9857, - "step": 3558 - }, - { - "epoch": 1.2913378684807255, - "grad_norm": 0.15625, - "learning_rate": 0.00011533472035246604, - "loss": 0.9722, - "step": 3560 - }, - { - "epoch": 1.2920634920634921, - "grad_norm": 0.1650390625, - "learning_rate": 0.00011528576673601763, - "loss": 1.0129, - "step": 3562 - }, - { - "epoch": 1.2927891156462585, - "grad_norm": 0.1650390625, - "learning_rate": 0.00011523681311956921, - "loss": 0.9774, - "step": 3564 - }, - { - "epoch": 1.293514739229025, - "grad_norm": 0.1767578125, - "learning_rate": 0.00011518785950312081, - "loss": 1.0214, - "step": 3566 - }, - { - "epoch": 1.2942403628117913, - "grad_norm": 0.166015625, - "learning_rate": 0.00011513890588667238, - "loss": 0.9524, - "step": 3568 - }, - { - "epoch": 1.294965986394558, - "grad_norm": 0.1611328125, - "learning_rate": 0.00011508995227022396, - "loss": 0.9879, - "step": 3570 - }, - { - "epoch": 1.2956916099773244, - "grad_norm": 0.1552734375, - "learning_rate": 0.00011504099865377556, - "loss": 0.946, - "step": 3572 - }, - { - "epoch": 1.2964172335600908, - "grad_norm": 0.17578125, - "learning_rate": 0.00011499204503732714, - "loss": 1.0078, - "step": 3574 - }, - { - "epoch": 1.2971428571428572, - "grad_norm": 0.162109375, - "learning_rate": 0.00011494309142087871, - "loss": 1.0155, - "step": 3576 - }, - { - "epoch": 1.2978684807256236, - "grad_norm": 0.1435546875, - "learning_rate": 0.0001148941378044303, - "loss": 0.9912, - "step": 3578 - }, - { - "epoch": 1.29859410430839, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011484518418798189, - "loss": 0.9606, - "step": 3580 - }, - { - "epoch": 1.2993197278911564, - "grad_norm": 0.1650390625, - "learning_rate": 0.00011479623057153348, - "loss": 0.9986, - "step": 3582 - }, - { - "epoch": 1.300045351473923, - "grad_norm": 0.1796875, - "learning_rate": 0.00011474727695508506, - "loss": 0.9684, - "step": 3584 - }, - { - "epoch": 1.3007709750566894, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011469832333863663, - "loss": 0.9675, - "step": 3586 - }, - { - "epoch": 1.3014965986394558, - "grad_norm": 0.1533203125, - "learning_rate": 0.00011464936972218823, - "loss": 0.971, - "step": 3588 - }, - { - "epoch": 1.3022222222222222, - "grad_norm": 0.1591796875, - "learning_rate": 0.00011460041610573981, - "loss": 0.9238, - "step": 3590 - }, - { - "epoch": 1.3029478458049888, - "grad_norm": 0.1630859375, - "learning_rate": 0.0001145514624892914, - "loss": 0.9196, - "step": 3592 - }, - { - "epoch": 1.3036734693877552, - "grad_norm": 0.1552734375, - "learning_rate": 0.00011450250887284299, - "loss": 0.9861, - "step": 3594 - }, - { - "epoch": 1.3043990929705216, - "grad_norm": 0.185546875, - "learning_rate": 0.00011445355525639456, - "loss": 1.0305, - "step": 3596 - }, - { - "epoch": 1.305124716553288, - "grad_norm": 0.1513671875, - "learning_rate": 0.00011440460163994615, - "loss": 0.9368, - "step": 3598 - }, - { - "epoch": 1.3058503401360544, - "grad_norm": 0.15625, - "learning_rate": 0.00011435564802349774, - "loss": 0.9788, - "step": 3600 - }, - { - "epoch": 1.3058503401360544, - "eval_loss": 1.014317274093628, - "eval_runtime": 87.2865, - "eval_samples_per_second": 20.622, - "eval_steps_per_second": 0.653, - "step": 3600 - }, - { - "epoch": 1.3065759637188208, - "grad_norm": 0.15234375, - "learning_rate": 0.00011430669440704933, - "loss": 0.9431, - "step": 3602 - }, - { - "epoch": 1.3073015873015872, - "grad_norm": 0.154296875, - "learning_rate": 0.00011425774079060091, - "loss": 0.9195, - "step": 3604 - }, - { - "epoch": 1.3080272108843538, - "grad_norm": 0.1806640625, - "learning_rate": 0.00011420878717415248, - "loss": 0.9597, - "step": 3606 - }, - { - "epoch": 1.3087528344671202, - "grad_norm": 0.1533203125, - "learning_rate": 0.00011415983355770408, - "loss": 0.9325, - "step": 3608 - }, - { - "epoch": 1.3094784580498866, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011411087994125566, - "loss": 0.9972, - "step": 3610 - }, - { - "epoch": 1.310204081632653, - "grad_norm": 0.1591796875, - "learning_rate": 0.00011406192632480726, - "loss": 1.0023, - "step": 3612 - }, - { - "epoch": 1.3109297052154196, - "grad_norm": 0.1767578125, - "learning_rate": 0.00011401297270835884, - "loss": 0.9872, - "step": 3614 - }, - { - "epoch": 1.311655328798186, - "grad_norm": 0.150390625, - "learning_rate": 0.00011396401909191041, - "loss": 1.0391, - "step": 3616 - }, - { - "epoch": 1.3123809523809524, - "grad_norm": 0.1572265625, - "learning_rate": 0.000113915065475462, - "loss": 0.9912, - "step": 3618 - }, - { - "epoch": 1.3131065759637188, - "grad_norm": 0.1640625, - "learning_rate": 0.00011386611185901359, - "loss": 1.0563, - "step": 3620 - }, - { - "epoch": 1.3138321995464852, - "grad_norm": 0.1640625, - "learning_rate": 0.00011381715824256518, - "loss": 0.9985, - "step": 3622 - }, - { - "epoch": 1.3145578231292516, - "grad_norm": 0.1591796875, - "learning_rate": 0.00011376820462611676, - "loss": 0.9683, - "step": 3624 - }, - { - "epoch": 1.315283446712018, - "grad_norm": 0.17578125, - "learning_rate": 0.00011371925100966833, - "loss": 0.9682, - "step": 3626 - }, - { - "epoch": 1.3160090702947846, - "grad_norm": 0.16015625, - "learning_rate": 0.00011367029739321993, - "loss": 1.0611, - "step": 3628 - }, - { - "epoch": 1.316734693877551, - "grad_norm": 0.166015625, - "learning_rate": 0.00011362134377677151, - "loss": 0.9994, - "step": 3630 - }, - { - "epoch": 1.3174603174603174, - "grad_norm": 0.140625, - "learning_rate": 0.0001135723901603231, - "loss": 0.9085, - "step": 3632 - }, - { - "epoch": 1.3181859410430838, - "grad_norm": 0.16796875, - "learning_rate": 0.00011352343654387469, - "loss": 0.9891, - "step": 3634 - }, - { - "epoch": 1.3189115646258505, - "grad_norm": 0.1591796875, - "learning_rate": 0.00011347448292742626, - "loss": 1.0091, - "step": 3636 - }, - { - "epoch": 1.3196371882086169, - "grad_norm": 0.1650390625, - "learning_rate": 0.00011342552931097785, - "loss": 0.9623, - "step": 3638 - }, - { - "epoch": 1.3203628117913833, - "grad_norm": 0.1650390625, - "learning_rate": 0.00011337657569452944, - "loss": 1.0065, - "step": 3640 - }, - { - "epoch": 1.3210884353741497, - "grad_norm": 0.1533203125, - "learning_rate": 0.00011332762207808103, - "loss": 0.9271, - "step": 3642 - }, - { - "epoch": 1.321814058956916, - "grad_norm": 0.17578125, - "learning_rate": 0.0001132786684616326, - "loss": 0.9958, - "step": 3644 - }, - { - "epoch": 1.3225396825396825, - "grad_norm": 0.15625, - "learning_rate": 0.00011322971484518418, - "loss": 0.9128, - "step": 3646 - }, - { - "epoch": 1.3232653061224489, - "grad_norm": 0.158203125, - "learning_rate": 0.00011318076122873578, - "loss": 0.9164, - "step": 3648 - }, - { - "epoch": 1.3239909297052155, - "grad_norm": 0.1708984375, - "learning_rate": 0.00011313180761228736, - "loss": 1.0168, - "step": 3650 - }, - { - "epoch": 1.3247165532879819, - "grad_norm": 0.162109375, - "learning_rate": 0.00011308285399583896, - "loss": 0.9683, - "step": 3652 - }, - { - "epoch": 1.3254421768707483, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011303390037939052, - "loss": 0.971, - "step": 3654 - }, - { - "epoch": 1.3261678004535147, - "grad_norm": 0.14453125, - "learning_rate": 0.00011298494676294211, - "loss": 0.9244, - "step": 3656 - }, - { - "epoch": 1.3268934240362813, - "grad_norm": 0.2060546875, - "learning_rate": 0.0001129359931464937, - "loss": 0.9584, - "step": 3658 - }, - { - "epoch": 1.3276190476190477, - "grad_norm": 0.16015625, - "learning_rate": 0.00011288703953004529, - "loss": 0.882, - "step": 3660 - }, - { - "epoch": 1.328344671201814, - "grad_norm": 0.1708984375, - "learning_rate": 0.00011283808591359688, - "loss": 0.9848, - "step": 3662 - }, - { - "epoch": 1.3290702947845805, - "grad_norm": 0.1650390625, - "learning_rate": 0.00011278913229714845, - "loss": 1.1064, - "step": 3664 - }, - { - "epoch": 1.329795918367347, - "grad_norm": 0.1533203125, - "learning_rate": 0.00011274017868070003, - "loss": 0.9595, - "step": 3666 - }, - { - "epoch": 1.3305215419501133, - "grad_norm": 0.1455078125, - "learning_rate": 0.00011269122506425163, - "loss": 0.9298, - "step": 3668 - }, - { - "epoch": 1.3312471655328797, - "grad_norm": 0.185546875, - "learning_rate": 0.00011264227144780321, - "loss": 1.0159, - "step": 3670 - }, - { - "epoch": 1.3319727891156463, - "grad_norm": 0.14453125, - "learning_rate": 0.0001125933178313548, - "loss": 0.9367, - "step": 3672 - }, - { - "epoch": 1.3326984126984127, - "grad_norm": 0.166015625, - "learning_rate": 0.00011254436421490637, - "loss": 0.9847, - "step": 3674 - }, - { - "epoch": 1.3334240362811791, - "grad_norm": 0.1376953125, - "learning_rate": 0.00011249541059845796, - "loss": 0.9506, - "step": 3676 - }, - { - "epoch": 1.3341496598639455, - "grad_norm": 0.154296875, - "learning_rate": 0.00011244645698200955, - "loss": 1.0063, - "step": 3678 - }, - { - "epoch": 1.3348752834467121, - "grad_norm": 0.1376953125, - "learning_rate": 0.00011239750336556114, - "loss": 0.9438, - "step": 3680 - }, - { - "epoch": 1.3356009070294785, - "grad_norm": 0.1767578125, - "learning_rate": 0.00011234854974911273, - "loss": 0.968, - "step": 3682 - }, - { - "epoch": 1.336326530612245, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001122995961326643, - "loss": 0.9975, - "step": 3684 - }, - { - "epoch": 1.3370521541950113, - "grad_norm": 0.171875, - "learning_rate": 0.00011225064251621588, - "loss": 1.0121, - "step": 3686 - }, - { - "epoch": 1.3377777777777777, - "grad_norm": 0.15234375, - "learning_rate": 0.00011220168889976748, - "loss": 0.9627, - "step": 3688 - }, - { - "epoch": 1.3385034013605441, - "grad_norm": 0.1572265625, - "learning_rate": 0.00011215273528331906, - "loss": 1.0203, - "step": 3690 - }, - { - "epoch": 1.3392290249433105, - "grad_norm": 0.185546875, - "learning_rate": 0.00011210378166687066, - "loss": 1.0594, - "step": 3692 - }, - { - "epoch": 1.3399546485260772, - "grad_norm": 0.1513671875, - "learning_rate": 0.00011205482805042222, - "loss": 0.9768, - "step": 3694 - }, - { - "epoch": 1.3406802721088436, - "grad_norm": 0.1484375, - "learning_rate": 0.00011200587443397381, - "loss": 1.027, - "step": 3696 - }, - { - "epoch": 1.34140589569161, - "grad_norm": 0.162109375, - "learning_rate": 0.0001119569208175254, - "loss": 1.0098, - "step": 3698 - }, - { - "epoch": 1.3421315192743764, - "grad_norm": 0.1572265625, - "learning_rate": 0.00011190796720107698, - "loss": 0.975, - "step": 3700 - }, - { - "epoch": 1.3421315192743764, - "eval_loss": 1.0139893293380737, - "eval_runtime": 87.2842, - "eval_samples_per_second": 20.622, - "eval_steps_per_second": 0.653, - "step": 3700 - }, - { - "epoch": 1.342857142857143, - "grad_norm": 0.1787109375, - "learning_rate": 0.00011185901358462858, - "loss": 0.9606, - "step": 3702 - }, - { - "epoch": 1.3435827664399094, - "grad_norm": 0.1826171875, - "learning_rate": 0.00011181005996818015, - "loss": 1.0558, - "step": 3704 - }, - { - "epoch": 1.3443083900226758, - "grad_norm": 0.1640625, - "learning_rate": 0.00011176110635173173, - "loss": 0.9632, - "step": 3706 - }, - { - "epoch": 1.3450340136054422, - "grad_norm": 0.1591796875, - "learning_rate": 0.00011171215273528333, - "loss": 1.0016, - "step": 3708 - }, - { - "epoch": 1.3457596371882086, - "grad_norm": 0.158203125, - "learning_rate": 0.00011166319911883491, - "loss": 0.9078, - "step": 3710 - }, - { - "epoch": 1.346485260770975, - "grad_norm": 0.1669921875, - "learning_rate": 0.0001116142455023865, - "loss": 1.0287, - "step": 3712 - }, - { - "epoch": 1.3472108843537414, - "grad_norm": 0.15625, - "learning_rate": 0.00011156529188593807, - "loss": 0.9694, - "step": 3714 - }, - { - "epoch": 1.347936507936508, - "grad_norm": 0.2001953125, - "learning_rate": 0.00011151633826948966, - "loss": 0.9519, - "step": 3716 - }, - { - "epoch": 1.3486621315192744, - "grad_norm": 0.158203125, - "learning_rate": 0.00011146738465304125, - "loss": 0.9624, - "step": 3718 - }, - { - "epoch": 1.3493877551020408, - "grad_norm": 0.154296875, - "learning_rate": 0.00011141843103659283, - "loss": 0.9302, - "step": 3720 - }, - { - "epoch": 1.3501133786848072, - "grad_norm": 0.150390625, - "learning_rate": 0.0001113694774201444, - "loss": 0.9119, - "step": 3722 - }, - { - "epoch": 1.3508390022675738, - "grad_norm": 0.1455078125, - "learning_rate": 0.000111320523803696, - "loss": 0.9269, - "step": 3724 - }, - { - "epoch": 1.3515646258503402, - "grad_norm": 0.1767578125, - "learning_rate": 0.00011127157018724758, - "loss": 1.0261, - "step": 3726 - }, - { - "epoch": 1.3522902494331066, - "grad_norm": 0.1591796875, - "learning_rate": 0.00011122261657079918, - "loss": 1.0123, - "step": 3728 - }, - { - "epoch": 1.353015873015873, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011117366295435076, - "loss": 1.0042, - "step": 3730 - }, - { - "epoch": 1.3537414965986394, - "grad_norm": 0.169921875, - "learning_rate": 0.00011112470933790233, - "loss": 0.9707, - "step": 3732 - }, - { - "epoch": 1.3544671201814058, - "grad_norm": 0.1806640625, - "learning_rate": 0.00011107575572145392, - "loss": 0.9504, - "step": 3734 - }, - { - "epoch": 1.3551927437641722, - "grad_norm": 0.1796875, - "learning_rate": 0.0001110268021050055, - "loss": 0.9999, - "step": 3736 - }, - { - "epoch": 1.3559183673469388, - "grad_norm": 0.189453125, - "learning_rate": 0.0001109778484885571, - "loss": 1.0107, - "step": 3738 - }, - { - "epoch": 1.3566439909297052, - "grad_norm": 0.16015625, - "learning_rate": 0.00011092889487210868, - "loss": 0.9823, - "step": 3740 - }, - { - "epoch": 1.3573696145124716, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011087994125566025, - "loss": 0.9654, - "step": 3742 - }, - { - "epoch": 1.358095238095238, - "grad_norm": 0.2099609375, - "learning_rate": 0.00011083098763921185, - "loss": 0.983, - "step": 3744 - }, - { - "epoch": 1.3588208616780046, - "grad_norm": 0.158203125, - "learning_rate": 0.00011078203402276343, - "loss": 1.0468, - "step": 3746 - }, - { - "epoch": 1.359546485260771, - "grad_norm": 0.1669921875, - "learning_rate": 0.00011073308040631503, - "loss": 0.9342, - "step": 3748 - }, - { - "epoch": 1.3602721088435374, - "grad_norm": 0.15625, - "learning_rate": 0.00011068412678986661, - "loss": 0.9798, - "step": 3750 - }, - { - "epoch": 1.3609977324263038, - "grad_norm": 0.1787109375, - "learning_rate": 0.00011063517317341818, - "loss": 0.9705, - "step": 3752 - }, - { - "epoch": 1.3617233560090702, - "grad_norm": 0.1884765625, - "learning_rate": 0.00011058621955696977, - "loss": 1.0327, - "step": 3754 - }, - { - "epoch": 1.3624489795918366, - "grad_norm": 0.1708984375, - "learning_rate": 0.00011053726594052136, - "loss": 1.0191, - "step": 3756 - }, - { - "epoch": 1.363174603174603, - "grad_norm": 0.1689453125, - "learning_rate": 0.00011048831232407295, - "loss": 0.9705, - "step": 3758 - }, - { - "epoch": 1.3639002267573697, - "grad_norm": 0.1611328125, - "learning_rate": 0.00011043935870762453, - "loss": 0.9572, - "step": 3760 - }, - { - "epoch": 1.364625850340136, - "grad_norm": 0.15234375, - "learning_rate": 0.0001103904050911761, - "loss": 0.9237, - "step": 3762 - }, - { - "epoch": 1.3653514739229025, - "grad_norm": 0.169921875, - "learning_rate": 0.0001103414514747277, - "loss": 0.8971, - "step": 3764 - }, - { - "epoch": 1.3660770975056689, - "grad_norm": 0.154296875, - "learning_rate": 0.00011029249785827928, - "loss": 0.9106, - "step": 3766 - }, - { - "epoch": 1.3668027210884355, - "grad_norm": 0.216796875, - "learning_rate": 0.00011024354424183088, - "loss": 1.0726, - "step": 3768 - }, - { - "epoch": 1.3675283446712019, - "grad_norm": 0.1474609375, - "learning_rate": 0.00011019459062538246, - "loss": 0.946, - "step": 3770 - }, - { - "epoch": 1.3682539682539683, - "grad_norm": 0.1513671875, - "learning_rate": 0.00011014563700893403, - "loss": 0.936, - "step": 3772 - }, - { - "epoch": 1.3689795918367347, - "grad_norm": 0.166015625, - "learning_rate": 0.00011009668339248562, - "loss": 0.9972, - "step": 3774 - }, - { - "epoch": 1.369705215419501, - "grad_norm": 0.171875, - "learning_rate": 0.0001100477297760372, - "loss": 1.0203, - "step": 3776 - }, - { - "epoch": 1.3704308390022675, - "grad_norm": 0.1826171875, - "learning_rate": 0.0001099987761595888, - "loss": 1.016, - "step": 3778 - }, - { - "epoch": 1.3711564625850339, - "grad_norm": 0.18359375, - "learning_rate": 0.00010994982254314038, - "loss": 0.9996, - "step": 3780 - }, - { - "epoch": 1.3718820861678005, - "grad_norm": 0.16796875, - "learning_rate": 0.00010990086892669195, - "loss": 0.9701, - "step": 3782 - }, - { - "epoch": 1.372607709750567, - "grad_norm": 0.1630859375, - "learning_rate": 0.00010985191531024355, - "loss": 0.9935, - "step": 3784 - }, - { - "epoch": 1.3733333333333333, - "grad_norm": 0.1396484375, - "learning_rate": 0.00010980296169379513, - "loss": 0.9978, - "step": 3786 - }, - { - "epoch": 1.3740589569160997, - "grad_norm": 0.16015625, - "learning_rate": 0.00010975400807734673, - "loss": 0.9602, - "step": 3788 - }, - { - "epoch": 1.3747845804988663, - "grad_norm": 0.150390625, - "learning_rate": 0.0001097050544608983, - "loss": 0.9894, - "step": 3790 - }, - { - "epoch": 1.3755102040816327, - "grad_norm": 0.154296875, - "learning_rate": 0.00010965610084444988, - "loss": 0.9851, - "step": 3792 - }, - { - "epoch": 1.3762358276643991, - "grad_norm": 0.1669921875, - "learning_rate": 0.00010960714722800147, - "loss": 1.0069, - "step": 3794 - }, - { - "epoch": 1.3769614512471655, - "grad_norm": 0.158203125, - "learning_rate": 0.00010955819361155306, - "loss": 0.9355, - "step": 3796 - }, - { - "epoch": 1.377687074829932, - "grad_norm": 0.1884765625, - "learning_rate": 0.00010950923999510465, - "loss": 1.0524, - "step": 3798 - }, - { - "epoch": 1.3784126984126983, - "grad_norm": 0.169921875, - "learning_rate": 0.00010946028637865622, - "loss": 0.9241, - "step": 3800 - }, - { - "epoch": 1.3784126984126983, - "eval_loss": 1.0139381885528564, - "eval_runtime": 88.7285, - "eval_samples_per_second": 20.287, - "eval_steps_per_second": 0.642, - "step": 3800 - }, - { - "epoch": 1.379138321995465, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001094113327622078, - "loss": 0.9646, - "step": 3802 - }, - { - "epoch": 1.3798639455782313, - "grad_norm": 0.169921875, - "learning_rate": 0.0001093623791457594, - "loss": 0.9912, - "step": 3804 - }, - { - "epoch": 1.3805895691609977, - "grad_norm": 0.166015625, - "learning_rate": 0.00010931342552931098, - "loss": 1.1027, - "step": 3806 - }, - { - "epoch": 1.3813151927437641, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010926447191286258, - "loss": 1.0653, - "step": 3808 - }, - { - "epoch": 1.3820408163265305, - "grad_norm": 0.1484375, - "learning_rate": 0.00010921551829641415, - "loss": 0.8741, - "step": 3810 - }, - { - "epoch": 1.3827664399092972, - "grad_norm": 0.16015625, - "learning_rate": 0.00010916656467996573, - "loss": 1.0081, - "step": 3812 - }, - { - "epoch": 1.3834920634920636, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010911761106351732, - "loss": 1.0412, - "step": 3814 - }, - { - "epoch": 1.38421768707483, - "grad_norm": 0.158203125, - "learning_rate": 0.0001090686574470689, - "loss": 1.0026, - "step": 3816 - }, - { - "epoch": 1.3849433106575963, - "grad_norm": 0.177734375, - "learning_rate": 0.0001090197038306205, - "loss": 1.0484, - "step": 3818 - }, - { - "epoch": 1.3856689342403627, - "grad_norm": 0.1572265625, - "learning_rate": 0.00010897075021417207, - "loss": 1.0069, - "step": 3820 - }, - { - "epoch": 1.3863945578231291, - "grad_norm": 0.1650390625, - "learning_rate": 0.00010892179659772365, - "loss": 0.9699, - "step": 3822 - }, - { - "epoch": 1.3871201814058958, - "grad_norm": 0.1572265625, - "learning_rate": 0.00010887284298127525, - "loss": 0.9712, - "step": 3824 - }, - { - "epoch": 1.3878458049886622, - "grad_norm": 0.154296875, - "learning_rate": 0.00010882388936482683, - "loss": 0.9634, - "step": 3826 - }, - { - "epoch": 1.3885714285714286, - "grad_norm": 0.1416015625, - "learning_rate": 0.00010877493574837843, - "loss": 0.9306, - "step": 3828 - }, - { - "epoch": 1.389297052154195, - "grad_norm": 0.18359375, - "learning_rate": 0.00010872598213193, - "loss": 0.9489, - "step": 3830 - }, - { - "epoch": 1.3900226757369616, - "grad_norm": 0.15625, - "learning_rate": 0.00010867702851548158, - "loss": 1.0097, - "step": 3832 - }, - { - "epoch": 1.390748299319728, - "grad_norm": 0.1513671875, - "learning_rate": 0.00010862807489903317, - "loss": 0.9603, - "step": 3834 - }, - { - "epoch": 1.3914739229024944, - "grad_norm": 0.1513671875, - "learning_rate": 0.00010857912128258476, - "loss": 0.944, - "step": 3836 - }, - { - "epoch": 1.3921995464852608, - "grad_norm": 0.1650390625, - "learning_rate": 0.00010853016766613635, - "loss": 0.9558, - "step": 3838 - }, - { - "epoch": 1.3929251700680272, - "grad_norm": 0.181640625, - "learning_rate": 0.00010848121404968792, - "loss": 0.98, - "step": 3840 - }, - { - "epoch": 1.3936507936507936, - "grad_norm": 0.158203125, - "learning_rate": 0.0001084322604332395, - "loss": 1.0133, - "step": 3842 - }, - { - "epoch": 1.39437641723356, - "grad_norm": 0.1708984375, - "learning_rate": 0.0001083833068167911, - "loss": 1.0192, - "step": 3844 - }, - { - "epoch": 1.3951020408163266, - "grad_norm": 0.1650390625, - "learning_rate": 0.00010833435320034268, - "loss": 1.0125, - "step": 3846 - }, - { - "epoch": 1.395827664399093, - "grad_norm": 0.158203125, - "learning_rate": 0.00010828539958389428, - "loss": 0.9468, - "step": 3848 - }, - { - "epoch": 1.3965532879818594, - "grad_norm": 0.1884765625, - "learning_rate": 0.00010823644596744585, - "loss": 0.9093, - "step": 3850 - }, - { - "epoch": 1.3972789115646258, - "grad_norm": 0.171875, - "learning_rate": 0.00010818749235099743, - "loss": 0.9601, - "step": 3852 - }, - { - "epoch": 1.3980045351473924, - "grad_norm": 0.1513671875, - "learning_rate": 0.00010813853873454902, - "loss": 0.9524, - "step": 3854 - }, - { - "epoch": 1.3987301587301588, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001080895851181006, - "loss": 0.9915, - "step": 3856 - }, - { - "epoch": 1.3994557823129252, - "grad_norm": 0.162109375, - "learning_rate": 0.0001080406315016522, - "loss": 0.9875, - "step": 3858 - }, - { - "epoch": 1.4001814058956916, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010799167788520377, - "loss": 0.9353, - "step": 3860 - }, - { - "epoch": 1.400907029478458, - "grad_norm": 0.1796875, - "learning_rate": 0.00010794272426875535, - "loss": 0.974, - "step": 3862 - }, - { - "epoch": 1.4016326530612244, - "grad_norm": 0.16796875, - "learning_rate": 0.00010789377065230695, - "loss": 0.9781, - "step": 3864 - }, - { - "epoch": 1.4023582766439908, - "grad_norm": 0.181640625, - "learning_rate": 0.00010784481703585853, - "loss": 0.9633, - "step": 3866 - }, - { - "epoch": 1.4030839002267574, - "grad_norm": 0.1455078125, - "learning_rate": 0.0001077958634194101, - "loss": 0.909, - "step": 3868 - }, - { - "epoch": 1.4038095238095238, - "grad_norm": 0.1806640625, - "learning_rate": 0.0001077469098029617, - "loss": 1.0195, - "step": 3870 - }, - { - "epoch": 1.4045351473922902, - "grad_norm": 0.1748046875, - "learning_rate": 0.00010769795618651328, - "loss": 1.0009, - "step": 3872 - }, - { - "epoch": 1.4052607709750566, - "grad_norm": 0.1630859375, - "learning_rate": 0.00010764900257006487, - "loss": 0.941, - "step": 3874 - }, - { - "epoch": 1.4059863945578233, - "grad_norm": 0.146484375, - "learning_rate": 0.00010760004895361646, - "loss": 0.8787, - "step": 3876 - }, - { - "epoch": 1.4067120181405897, - "grad_norm": 0.15234375, - "learning_rate": 0.00010755109533716802, - "loss": 1.0049, - "step": 3878 - }, - { - "epoch": 1.407437641723356, - "grad_norm": 0.1669921875, - "learning_rate": 0.00010750214172071962, - "loss": 1.0413, - "step": 3880 - }, - { - "epoch": 1.4081632653061225, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001074531881042712, - "loss": 1.0002, - "step": 3882 - }, - { - "epoch": 1.4088888888888889, - "grad_norm": 0.19140625, - "learning_rate": 0.0001074042344878228, - "loss": 1.0016, - "step": 3884 - }, - { - "epoch": 1.4096145124716553, - "grad_norm": 0.1884765625, - "learning_rate": 0.00010735528087137438, - "loss": 0.9948, - "step": 3886 - }, - { - "epoch": 1.4103401360544217, - "grad_norm": 0.173828125, - "learning_rate": 0.00010730632725492595, - "loss": 0.9512, - "step": 3888 - }, - { - "epoch": 1.4110657596371883, - "grad_norm": 0.16796875, - "learning_rate": 0.00010725737363847755, - "loss": 0.9985, - "step": 3890 - }, - { - "epoch": 1.4117913832199547, - "grad_norm": 0.166015625, - "learning_rate": 0.00010720842002202913, - "loss": 0.9329, - "step": 3892 - }, - { - "epoch": 1.412517006802721, - "grad_norm": 0.1650390625, - "learning_rate": 0.00010715946640558072, - "loss": 0.9697, - "step": 3894 - }, - { - "epoch": 1.4132426303854875, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001071105127891323, - "loss": 0.9691, - "step": 3896 - }, - { - "epoch": 1.413968253968254, - "grad_norm": 0.1474609375, - "learning_rate": 0.00010706155917268387, - "loss": 0.9706, - "step": 3898 - }, - { - "epoch": 1.4146938775510205, - "grad_norm": 0.171875, - "learning_rate": 0.00010701260555623547, - "loss": 1.0023, - "step": 3900 - }, - { - "epoch": 1.4146938775510205, - "eval_loss": 1.013636589050293, - "eval_runtime": 92.7582, - "eval_samples_per_second": 19.405, - "eval_steps_per_second": 0.615, - "step": 3900 - }, - { - "epoch": 1.415419501133787, - "grad_norm": 0.154296875, - "learning_rate": 0.00010696365193978705, - "loss": 0.9634, - "step": 3902 - }, - { - "epoch": 1.4161451247165533, - "grad_norm": 0.1611328125, - "learning_rate": 0.00010691469832333865, - "loss": 0.9808, - "step": 3904 - }, - { - "epoch": 1.4168707482993197, - "grad_norm": 0.169921875, - "learning_rate": 0.00010686574470689023, - "loss": 0.9637, - "step": 3906 - }, - { - "epoch": 1.417596371882086, - "grad_norm": 0.1474609375, - "learning_rate": 0.0001068167910904418, - "loss": 0.9661, - "step": 3908 - }, - { - "epoch": 1.4183219954648525, - "grad_norm": 0.1650390625, - "learning_rate": 0.0001067678374739934, - "loss": 0.9844, - "step": 3910 - }, - { - "epoch": 1.4190476190476191, - "grad_norm": 0.150390625, - "learning_rate": 0.00010671888385754498, - "loss": 0.9415, - "step": 3912 - }, - { - "epoch": 1.4197732426303855, - "grad_norm": 0.1552734375, - "learning_rate": 0.00010666993024109657, - "loss": 1.0491, - "step": 3914 - }, - { - "epoch": 1.420498866213152, - "grad_norm": 0.1845703125, - "learning_rate": 0.00010662097662464816, - "loss": 1.007, - "step": 3916 - }, - { - "epoch": 1.4212244897959183, - "grad_norm": 0.154296875, - "learning_rate": 0.00010657202300819972, - "loss": 0.9802, - "step": 3918 - }, - { - "epoch": 1.421950113378685, - "grad_norm": 0.1611328125, - "learning_rate": 0.00010652306939175132, - "loss": 1.0043, - "step": 3920 - }, - { - "epoch": 1.4226757369614513, - "grad_norm": 0.1611328125, - "learning_rate": 0.0001064741157753029, - "loss": 1.0034, - "step": 3922 - }, - { - "epoch": 1.4234013605442177, - "grad_norm": 0.1806640625, - "learning_rate": 0.0001064251621588545, - "loss": 0.9387, - "step": 3924 - }, - { - "epoch": 1.4241269841269841, - "grad_norm": 0.1650390625, - "learning_rate": 0.00010637620854240608, - "loss": 0.9047, - "step": 3926 - }, - { - "epoch": 1.4248526077097505, - "grad_norm": 0.1552734375, - "learning_rate": 0.00010632725492595765, - "loss": 0.9024, - "step": 3928 - }, - { - "epoch": 1.425578231292517, - "grad_norm": 0.1552734375, - "learning_rate": 0.00010627830130950925, - "loss": 1.024, - "step": 3930 - }, - { - "epoch": 1.4263038548752833, - "grad_norm": 0.169921875, - "learning_rate": 0.00010622934769306083, - "loss": 0.9421, - "step": 3932 - }, - { - "epoch": 1.42702947845805, - "grad_norm": 0.1640625, - "learning_rate": 0.00010618039407661242, - "loss": 1.0215, - "step": 3934 - }, - { - "epoch": 1.4277551020408163, - "grad_norm": 0.1650390625, - "learning_rate": 0.00010613144046016399, - "loss": 0.9815, - "step": 3936 - }, - { - "epoch": 1.4284807256235827, - "grad_norm": 0.1494140625, - "learning_rate": 0.00010608248684371557, - "loss": 0.9606, - "step": 3938 - }, - { - "epoch": 1.4292063492063491, - "grad_norm": 0.158203125, - "learning_rate": 0.00010603353322726717, - "loss": 0.9918, - "step": 3940 - }, - { - "epoch": 1.4299319727891158, - "grad_norm": 0.1630859375, - "learning_rate": 0.00010598457961081875, - "loss": 0.9695, - "step": 3942 - }, - { - "epoch": 1.4306575963718822, - "grad_norm": 0.177734375, - "learning_rate": 0.00010593562599437035, - "loss": 0.9732, - "step": 3944 - }, - { - "epoch": 1.4313832199546486, - "grad_norm": 0.1787109375, - "learning_rate": 0.00010588667237792192, - "loss": 1.0474, - "step": 3946 - }, - { - "epoch": 1.432108843537415, - "grad_norm": 0.1728515625, - "learning_rate": 0.0001058377187614735, - "loss": 1.0052, - "step": 3948 - }, - { - "epoch": 1.4328344671201814, - "grad_norm": 0.1533203125, - "learning_rate": 0.0001057887651450251, - "loss": 0.9909, - "step": 3950 - }, - { - "epoch": 1.4335600907029478, - "grad_norm": 0.2109375, - "learning_rate": 0.00010573981152857668, - "loss": 0.9774, - "step": 3952 - }, - { - "epoch": 1.4342857142857142, - "grad_norm": 0.1787109375, - "learning_rate": 0.00010569085791212827, - "loss": 0.9736, - "step": 3954 - }, - { - "epoch": 1.4350113378684808, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010564190429567984, - "loss": 1.0008, - "step": 3956 - }, - { - "epoch": 1.4357369614512472, - "grad_norm": 0.1669921875, - "learning_rate": 0.00010559295067923142, - "loss": 0.9877, - "step": 3958 - }, - { - "epoch": 1.4364625850340136, - "grad_norm": 0.171875, - "learning_rate": 0.00010554399706278302, - "loss": 0.9768, - "step": 3960 - }, - { - "epoch": 1.43718820861678, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001054950434463346, - "loss": 0.9943, - "step": 3962 - }, - { - "epoch": 1.4379138321995466, - "grad_norm": 0.1865234375, - "learning_rate": 0.0001054460898298862, - "loss": 1.0235, - "step": 3964 - }, - { - "epoch": 1.438639455782313, - "grad_norm": 0.1767578125, - "learning_rate": 0.00010539713621343777, - "loss": 1.0468, - "step": 3966 - }, - { - "epoch": 1.4393650793650794, - "grad_norm": 0.16015625, - "learning_rate": 0.00010534818259698935, - "loss": 0.9501, - "step": 3968 - }, - { - "epoch": 1.4400907029478458, - "grad_norm": 0.1572265625, - "learning_rate": 0.00010529922898054095, - "loss": 1.0001, - "step": 3970 - }, - { - "epoch": 1.4408163265306122, - "grad_norm": 0.1865234375, - "learning_rate": 0.00010525027536409253, - "loss": 0.9847, - "step": 3972 - }, - { - "epoch": 1.4415419501133786, - "grad_norm": 0.1552734375, - "learning_rate": 0.00010520132174764412, - "loss": 0.9209, - "step": 3974 - }, - { - "epoch": 1.442267573696145, - "grad_norm": 0.1748046875, - "learning_rate": 0.00010515236813119569, - "loss": 0.969, - "step": 3976 - }, - { - "epoch": 1.4429931972789116, - "grad_norm": 0.140625, - "learning_rate": 0.00010510341451474727, - "loss": 0.8542, - "step": 3978 - }, - { - "epoch": 1.443718820861678, - "grad_norm": 0.1650390625, - "learning_rate": 0.00010505446089829887, - "loss": 0.9088, - "step": 3980 - }, - { - "epoch": 1.4444444444444444, - "grad_norm": 0.1875, - "learning_rate": 0.00010500550728185045, - "loss": 0.9864, - "step": 3982 - }, - { - "epoch": 1.4451700680272108, - "grad_norm": 0.1806640625, - "learning_rate": 0.00010495655366540205, - "loss": 1.0233, - "step": 3984 - }, - { - "epoch": 1.4458956916099774, - "grad_norm": 0.15234375, - "learning_rate": 0.00010490760004895362, - "loss": 0.9392, - "step": 3986 - }, - { - "epoch": 1.4466213151927438, - "grad_norm": 0.1748046875, - "learning_rate": 0.0001048586464325052, - "loss": 0.9399, - "step": 3988 - }, - { - "epoch": 1.4473469387755102, - "grad_norm": 0.171875, - "learning_rate": 0.0001048096928160568, - "loss": 1.0332, - "step": 3990 - }, - { - "epoch": 1.4480725623582766, - "grad_norm": 0.2060546875, - "learning_rate": 0.00010476073919960838, - "loss": 1.0093, - "step": 3992 - }, - { - "epoch": 1.448798185941043, - "grad_norm": 0.1806640625, - "learning_rate": 0.00010471178558315997, - "loss": 0.9698, - "step": 3994 - }, - { - "epoch": 1.4495238095238094, - "grad_norm": 0.162109375, - "learning_rate": 0.00010466283196671154, - "loss": 0.9719, - "step": 3996 - }, - { - "epoch": 1.4502494331065758, - "grad_norm": 0.1787109375, - "learning_rate": 0.00010461387835026312, - "loss": 1.0843, - "step": 3998 - }, - { - "epoch": 1.4509750566893425, - "grad_norm": 0.1748046875, - "learning_rate": 0.00010456492473381472, - "loss": 0.9405, - "step": 4000 - }, - { - "epoch": 1.4509750566893425, - "eval_loss": 1.0138306617736816, - "eval_runtime": 90.2991, - "eval_samples_per_second": 19.934, - "eval_steps_per_second": 0.631, - "step": 4000 - }, - { - "epoch": 1.4517006802721089, - "grad_norm": 0.16015625, - "learning_rate": 0.0001045159711173663, - "loss": 0.9971, - "step": 4002 - }, - { - "epoch": 1.4524263038548753, - "grad_norm": 0.1552734375, - "learning_rate": 0.0001044670175009179, - "loss": 0.9311, - "step": 4004 - }, - { - "epoch": 1.4531519274376417, - "grad_norm": 0.1513671875, - "learning_rate": 0.00010441806388446947, - "loss": 0.9054, - "step": 4006 - }, - { - "epoch": 1.4538775510204083, - "grad_norm": 0.16015625, - "learning_rate": 0.00010436911026802105, - "loss": 1.0426, - "step": 4008 - }, - { - "epoch": 1.4546031746031747, - "grad_norm": 0.1728515625, - "learning_rate": 0.00010432015665157264, - "loss": 0.9125, - "step": 4010 - }, - { - "epoch": 1.455328798185941, - "grad_norm": 0.16796875, - "learning_rate": 0.00010427120303512423, - "loss": 0.9421, - "step": 4012 - }, - { - "epoch": 1.4560544217687075, - "grad_norm": 0.1591796875, - "learning_rate": 0.0001042222494186758, - "loss": 0.9698, - "step": 4014 - }, - { - "epoch": 1.4567800453514739, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010417329580222739, - "loss": 1.0262, - "step": 4016 - }, - { - "epoch": 1.4575056689342403, - "grad_norm": 0.166015625, - "learning_rate": 0.00010412434218577897, - "loss": 0.9472, - "step": 4018 - }, - { - "epoch": 1.4582312925170067, - "grad_norm": 0.142578125, - "learning_rate": 0.00010407538856933057, - "loss": 0.981, - "step": 4020 - }, - { - "epoch": 1.4589569160997733, - "grad_norm": 0.1591796875, - "learning_rate": 0.00010402643495288215, - "loss": 0.9012, - "step": 4022 - }, - { - "epoch": 1.4596825396825397, - "grad_norm": 0.158203125, - "learning_rate": 0.00010397748133643372, - "loss": 0.9633, - "step": 4024 - }, - { - "epoch": 1.460408163265306, - "grad_norm": 0.1669921875, - "learning_rate": 0.00010392852771998532, - "loss": 0.9586, - "step": 4026 - }, - { - "epoch": 1.4611337868480725, - "grad_norm": 0.193359375, - "learning_rate": 0.0001038795741035369, - "loss": 1.0336, - "step": 4028 - }, - { - "epoch": 1.4618594104308391, - "grad_norm": 0.1669921875, - "learning_rate": 0.0001038306204870885, - "loss": 0.9896, - "step": 4030 - }, - { - "epoch": 1.4625850340136055, - "grad_norm": 0.185546875, - "learning_rate": 0.00010378166687064008, - "loss": 1.0079, - "step": 4032 - }, - { - "epoch": 1.463310657596372, - "grad_norm": 0.166015625, - "learning_rate": 0.00010373271325419165, - "loss": 0.9626, - "step": 4034 - }, - { - "epoch": 1.4640362811791383, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010368375963774324, - "loss": 0.9859, - "step": 4036 - }, - { - "epoch": 1.4647619047619047, - "grad_norm": 0.1484375, - "learning_rate": 0.00010363480602129482, - "loss": 0.9595, - "step": 4038 - }, - { - "epoch": 1.465487528344671, - "grad_norm": 0.1650390625, - "learning_rate": 0.00010358585240484642, - "loss": 0.9691, - "step": 4040 - }, - { - "epoch": 1.4662131519274375, - "grad_norm": 0.15234375, - "learning_rate": 0.000103536898788398, - "loss": 0.983, - "step": 4042 - }, - { - "epoch": 1.4669387755102041, - "grad_norm": 0.171875, - "learning_rate": 0.00010348794517194957, - "loss": 1.0172, - "step": 4044 - }, - { - "epoch": 1.4676643990929705, - "grad_norm": 0.1728515625, - "learning_rate": 0.00010343899155550117, - "loss": 0.9945, - "step": 4046 - }, - { - "epoch": 1.468390022675737, - "grad_norm": 0.1708984375, - "learning_rate": 0.00010339003793905275, - "loss": 0.9569, - "step": 4048 - }, - { - "epoch": 1.4691156462585033, - "grad_norm": 0.1591796875, - "learning_rate": 0.00010334108432260434, - "loss": 0.9445, - "step": 4050 - }, - { - "epoch": 1.46984126984127, - "grad_norm": 0.18359375, - "learning_rate": 0.00010329213070615593, - "loss": 1.0193, - "step": 4052 - }, - { - "epoch": 1.4705668934240363, - "grad_norm": 0.1806640625, - "learning_rate": 0.0001032431770897075, - "loss": 1.0227, - "step": 4054 - }, - { - "epoch": 1.4712925170068027, - "grad_norm": 0.171875, - "learning_rate": 0.00010319422347325909, - "loss": 0.9067, - "step": 4056 - }, - { - "epoch": 1.4720181405895691, - "grad_norm": 0.1630859375, - "learning_rate": 0.00010314526985681067, - "loss": 0.9907, - "step": 4058 - }, - { - "epoch": 1.4727437641723355, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010309631624036227, - "loss": 0.963, - "step": 4060 - }, - { - "epoch": 1.473469387755102, - "grad_norm": 0.15234375, - "learning_rate": 0.00010304736262391385, - "loss": 0.9769, - "step": 4062 - }, - { - "epoch": 1.4741950113378683, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010299840900746542, - "loss": 0.9689, - "step": 4064 - }, - { - "epoch": 1.474920634920635, - "grad_norm": 0.1796875, - "learning_rate": 0.00010294945539101702, - "loss": 0.9922, - "step": 4066 - }, - { - "epoch": 1.4756462585034014, - "grad_norm": 0.1572265625, - "learning_rate": 0.0001029005017745686, - "loss": 0.9429, - "step": 4068 - }, - { - "epoch": 1.4763718820861678, - "grad_norm": 0.1884765625, - "learning_rate": 0.0001028515481581202, - "loss": 1.0386, - "step": 4070 - }, - { - "epoch": 1.4770975056689342, - "grad_norm": 0.1630859375, - "learning_rate": 0.00010280259454167178, - "loss": 0.9666, - "step": 4072 - }, - { - "epoch": 1.4778231292517008, - "grad_norm": 0.146484375, - "learning_rate": 0.00010275364092522335, - "loss": 1.0141, - "step": 4074 - }, - { - "epoch": 1.4785487528344672, - "grad_norm": 0.1591796875, - "learning_rate": 0.00010270468730877494, - "loss": 0.9448, - "step": 4076 - }, - { - "epoch": 1.4792743764172336, - "grad_norm": 0.1484375, - "learning_rate": 0.00010265573369232652, - "loss": 1.0185, - "step": 4078 - }, - { - "epoch": 1.48, - "grad_norm": 0.1591796875, - "learning_rate": 0.00010260678007587812, - "loss": 0.9499, - "step": 4080 - }, - { - "epoch": 1.4807256235827664, - "grad_norm": 0.1708984375, - "learning_rate": 0.00010255782645942969, - "loss": 0.9646, - "step": 4082 - }, - { - "epoch": 1.4814512471655328, - "grad_norm": 0.1708984375, - "learning_rate": 0.00010250887284298127, - "loss": 1.0299, - "step": 4084 - }, - { - "epoch": 1.4821768707482994, - "grad_norm": 0.1611328125, - "learning_rate": 0.00010245991922653287, - "loss": 0.9877, - "step": 4086 - }, - { - "epoch": 1.4829024943310658, - "grad_norm": 0.173828125, - "learning_rate": 0.00010241096561008445, - "loss": 0.9352, - "step": 4088 - }, - { - "epoch": 1.4836281179138322, - "grad_norm": 0.158203125, - "learning_rate": 0.00010236201199363604, - "loss": 0.9672, - "step": 4090 - }, - { - "epoch": 1.4843537414965986, - "grad_norm": 0.1767578125, - "learning_rate": 0.00010231305837718761, - "loss": 0.9205, - "step": 4092 - }, - { - "epoch": 1.485079365079365, - "grad_norm": 0.1640625, - "learning_rate": 0.0001022641047607392, - "loss": 1.0447, - "step": 4094 - }, - { - "epoch": 1.4858049886621316, - "grad_norm": 0.1611328125, - "learning_rate": 0.00010221515114429079, - "loss": 0.9848, - "step": 4096 - }, - { - "epoch": 1.486530612244898, - "grad_norm": 0.1875, - "learning_rate": 0.00010216619752784237, - "loss": 0.9774, - "step": 4098 - }, - { - "epoch": 1.4872562358276644, - "grad_norm": 0.1708984375, - "learning_rate": 0.00010211724391139397, - "loss": 0.9378, - "step": 4100 - }, - { - "epoch": 1.4872562358276644, - "eval_loss": 1.0134267807006836, - "eval_runtime": 90.0252, - "eval_samples_per_second": 19.994, - "eval_steps_per_second": 0.633, - "step": 4100 - }, - { - "epoch": 1.4879818594104308, - "grad_norm": 0.16796875, - "learning_rate": 0.00010206829029494554, - "loss": 0.9782, - "step": 4102 - }, - { - "epoch": 1.4887074829931972, - "grad_norm": 0.1708984375, - "learning_rate": 0.00010201933667849712, - "loss": 1.0532, - "step": 4104 - }, - { - "epoch": 1.4894331065759636, - "grad_norm": 0.1513671875, - "learning_rate": 0.00010197038306204872, - "loss": 0.9129, - "step": 4106 - }, - { - "epoch": 1.4901587301587302, - "grad_norm": 0.1494140625, - "learning_rate": 0.0001019214294456003, - "loss": 0.9277, - "step": 4108 - }, - { - "epoch": 1.4908843537414966, - "grad_norm": 0.17578125, - "learning_rate": 0.0001018724758291519, - "loss": 0.9731, - "step": 4110 - }, - { - "epoch": 1.491609977324263, - "grad_norm": 0.158203125, - "learning_rate": 0.00010182352221270346, - "loss": 0.9863, - "step": 4112 - }, - { - "epoch": 1.4923356009070294, - "grad_norm": 0.15625, - "learning_rate": 0.00010177456859625505, - "loss": 1.1322, - "step": 4114 - }, - { - "epoch": 1.4930612244897958, - "grad_norm": 0.181640625, - "learning_rate": 0.00010172561497980664, - "loss": 0.9763, - "step": 4116 - }, - { - "epoch": 1.4937868480725625, - "grad_norm": 0.173828125, - "learning_rate": 0.00010167666136335822, - "loss": 0.978, - "step": 4118 - }, - { - "epoch": 1.4945124716553289, - "grad_norm": 0.1748046875, - "learning_rate": 0.00010162770774690982, - "loss": 1.0198, - "step": 4120 - }, - { - "epoch": 1.4952380952380953, - "grad_norm": 0.162109375, - "learning_rate": 0.00010157875413046139, - "loss": 0.8959, - "step": 4122 - }, - { - "epoch": 1.4959637188208617, - "grad_norm": 0.173828125, - "learning_rate": 0.00010152980051401297, - "loss": 0.9886, - "step": 4124 - }, - { - "epoch": 1.496689342403628, - "grad_norm": 0.1513671875, - "learning_rate": 0.00010148084689756457, - "loss": 0.9965, - "step": 4126 - }, - { - "epoch": 1.4974149659863945, - "grad_norm": 0.1689453125, - "learning_rate": 0.00010143189328111615, - "loss": 0.961, - "step": 4128 - }, - { - "epoch": 1.498140589569161, - "grad_norm": 0.1455078125, - "learning_rate": 0.00010138293966466774, - "loss": 0.9631, - "step": 4130 - }, - { - "epoch": 1.4988662131519275, - "grad_norm": 0.208984375, - "learning_rate": 0.00010133398604821931, - "loss": 1.0086, - "step": 4132 - }, - { - "epoch": 1.4995918367346939, - "grad_norm": 0.154296875, - "learning_rate": 0.0001012850324317709, - "loss": 0.9727, - "step": 4134 - }, - { - "epoch": 1.5003174603174603, - "grad_norm": 0.150390625, - "learning_rate": 0.00010123607881532249, - "loss": 0.9614, - "step": 4136 - }, - { - "epoch": 1.501043083900227, - "grad_norm": 0.1708984375, - "learning_rate": 0.00010118712519887407, - "loss": 0.9524, - "step": 4138 - }, - { - "epoch": 1.5017687074829933, - "grad_norm": 0.171875, - "learning_rate": 0.00010113817158242567, - "loss": 0.9623, - "step": 4140 - }, - { - "epoch": 1.5024943310657597, - "grad_norm": 0.1728515625, - "learning_rate": 0.00010108921796597724, - "loss": 0.9599, - "step": 4142 - }, - { - "epoch": 1.503219954648526, - "grad_norm": 0.162109375, - "learning_rate": 0.00010104026434952882, - "loss": 0.9384, - "step": 4144 - }, - { - "epoch": 1.5039455782312925, - "grad_norm": 0.1591796875, - "learning_rate": 0.00010099131073308042, - "loss": 0.9673, - "step": 4146 - }, - { - "epoch": 1.504671201814059, - "grad_norm": 0.16015625, - "learning_rate": 0.000100942357116632, - "loss": 0.9375, - "step": 4148 - }, - { - "epoch": 1.5053968253968253, - "grad_norm": 0.181640625, - "learning_rate": 0.0001008934035001836, - "loss": 0.884, - "step": 4150 - }, - { - "epoch": 1.5061224489795917, - "grad_norm": 0.1513671875, - "learning_rate": 0.00010084444988373516, - "loss": 0.9254, - "step": 4152 - }, - { - "epoch": 1.5068480725623583, - "grad_norm": 0.203125, - "learning_rate": 0.00010079549626728675, - "loss": 1.0317, - "step": 4154 - }, - { - "epoch": 1.5075736961451247, - "grad_norm": 0.17578125, - "learning_rate": 0.00010074654265083834, - "loss": 0.9853, - "step": 4156 - }, - { - "epoch": 1.508299319727891, - "grad_norm": 0.166015625, - "learning_rate": 0.00010069758903438992, - "loss": 0.9551, - "step": 4158 - }, - { - "epoch": 1.5090249433106577, - "grad_norm": 0.1640625, - "learning_rate": 0.00010064863541794149, - "loss": 1.0439, - "step": 4160 - }, - { - "epoch": 1.5097505668934241, - "grad_norm": 0.1875, - "learning_rate": 0.00010059968180149309, - "loss": 0.9087, - "step": 4162 - }, - { - "epoch": 1.5104761904761905, - "grad_norm": 0.15234375, - "learning_rate": 0.00010055072818504467, - "loss": 0.9663, - "step": 4164 - }, - { - "epoch": 1.511201814058957, - "grad_norm": 0.162109375, - "learning_rate": 0.00010050177456859627, - "loss": 0.9691, - "step": 4166 - }, - { - "epoch": 1.5119274376417233, - "grad_norm": 0.171875, - "learning_rate": 0.00010045282095214785, - "loss": 1.0104, - "step": 4168 - }, - { - "epoch": 1.5126530612244897, - "grad_norm": 0.169921875, - "learning_rate": 0.00010040386733569942, - "loss": 1.022, - "step": 4170 - }, - { - "epoch": 1.5133786848072561, - "grad_norm": 0.1396484375, - "learning_rate": 0.00010035491371925101, - "loss": 1.0086, - "step": 4172 - }, - { - "epoch": 1.5141043083900225, - "grad_norm": 0.166015625, - "learning_rate": 0.0001003059601028026, - "loss": 0.9908, - "step": 4174 - }, - { - "epoch": 1.5148299319727891, - "grad_norm": 0.22265625, - "learning_rate": 0.00010025700648635419, - "loss": 0.9026, - "step": 4176 - }, - { - "epoch": 1.5155555555555555, - "grad_norm": 0.1630859375, - "learning_rate": 0.00010020805286990577, - "loss": 1.0208, - "step": 4178 - }, - { - "epoch": 1.516281179138322, - "grad_norm": 0.1591796875, - "learning_rate": 0.00010015909925345734, - "loss": 0.9172, - "step": 4180 - }, - { - "epoch": 1.5170068027210886, - "grad_norm": 0.1708984375, - "learning_rate": 0.00010011014563700894, - "loss": 0.9348, - "step": 4182 - }, - { - "epoch": 1.517732426303855, - "grad_norm": 0.1591796875, - "learning_rate": 0.00010006119202056052, - "loss": 0.9899, - "step": 4184 - }, - { - "epoch": 1.5184580498866214, - "grad_norm": 0.1953125, - "learning_rate": 0.00010001223840411212, - "loss": 1.0367, - "step": 4186 - }, - { - "epoch": 1.5191836734693878, - "grad_norm": 0.1669921875, - "learning_rate": 9.996328478766368e-05, - "loss": 0.9138, - "step": 4188 - }, - { - "epoch": 1.5199092970521542, - "grad_norm": 0.1689453125, - "learning_rate": 9.991433117121528e-05, - "loss": 1.0288, - "step": 4190 - }, - { - "epoch": 1.5206349206349206, - "grad_norm": 0.158203125, - "learning_rate": 9.986537755476686e-05, - "loss": 0.9677, - "step": 4192 - }, - { - "epoch": 1.521360544217687, - "grad_norm": 0.1962890625, - "learning_rate": 9.981642393831845e-05, - "loss": 1.0642, - "step": 4194 - }, - { - "epoch": 1.5220861678004534, - "grad_norm": 0.16796875, - "learning_rate": 9.976747032187004e-05, - "loss": 1.0481, - "step": 4196 - }, - { - "epoch": 1.52281179138322, - "grad_norm": 0.1669921875, - "learning_rate": 9.971851670542161e-05, - "loss": 1.0118, - "step": 4198 - }, - { - "epoch": 1.5235374149659864, - "grad_norm": 0.1650390625, - "learning_rate": 9.96695630889732e-05, - "loss": 0.9331, - "step": 4200 - }, - { - "epoch": 1.5235374149659864, - "eval_loss": 1.013152003288269, - "eval_runtime": 89.7624, - "eval_samples_per_second": 20.053, - "eval_steps_per_second": 0.635, - "step": 4200 - }, - { - "epoch": 1.5242630385487528, - "grad_norm": 0.16796875, - "learning_rate": 9.962060947252479e-05, - "loss": 1.0175, - "step": 4202 - }, - { - "epoch": 1.5249886621315194, - "grad_norm": 0.16796875, - "learning_rate": 9.957165585607637e-05, - "loss": 1.0108, - "step": 4204 - }, - { - "epoch": 1.5257142857142858, - "grad_norm": 0.1572265625, - "learning_rate": 9.952270223962795e-05, - "loss": 1.0153, - "step": 4206 - }, - { - "epoch": 1.5264399092970522, - "grad_norm": 0.1767578125, - "learning_rate": 9.947374862317953e-05, - "loss": 1.0924, - "step": 4208 - }, - { - "epoch": 1.5271655328798186, - "grad_norm": 0.162109375, - "learning_rate": 9.942479500673113e-05, - "loss": 0.9743, - "step": 4210 - }, - { - "epoch": 1.527891156462585, - "grad_norm": 0.177734375, - "learning_rate": 9.937584139028271e-05, - "loss": 0.9311, - "step": 4212 - }, - { - "epoch": 1.5286167800453514, - "grad_norm": 0.158203125, - "learning_rate": 9.93268877738343e-05, - "loss": 0.9674, - "step": 4214 - }, - { - "epoch": 1.5293424036281178, - "grad_norm": 0.1611328125, - "learning_rate": 9.927793415738588e-05, - "loss": 0.9447, - "step": 4216 - }, - { - "epoch": 1.5300680272108842, - "grad_norm": 0.1533203125, - "learning_rate": 9.922898054093746e-05, - "loss": 1.0299, - "step": 4218 - }, - { - "epoch": 1.5307936507936508, - "grad_norm": 0.166015625, - "learning_rate": 9.918002692448906e-05, - "loss": 0.954, - "step": 4220 - }, - { - "epoch": 1.5315192743764172, - "grad_norm": 0.171875, - "learning_rate": 9.913107330804064e-05, - "loss": 1.012, - "step": 4222 - }, - { - "epoch": 1.5322448979591838, - "grad_norm": 0.1748046875, - "learning_rate": 9.908211969159222e-05, - "loss": 1.0334, - "step": 4224 - }, - { - "epoch": 1.5329705215419502, - "grad_norm": 0.1474609375, - "learning_rate": 9.90331660751438e-05, - "loss": 0.9732, - "step": 4226 - }, - { - "epoch": 1.5336961451247166, - "grad_norm": 0.15234375, - "learning_rate": 9.898421245869538e-05, - "loss": 0.9949, - "step": 4228 - }, - { - "epoch": 1.534421768707483, - "grad_norm": 0.1806640625, - "learning_rate": 9.893525884224698e-05, - "loss": 1.0049, - "step": 4230 - }, - { - "epoch": 1.5351473922902494, - "grad_norm": 0.158203125, - "learning_rate": 9.888630522579856e-05, - "loss": 0.9727, - "step": 4232 - }, - { - "epoch": 1.5358730158730158, - "grad_norm": 0.1630859375, - "learning_rate": 9.883735160935014e-05, - "loss": 0.9889, - "step": 4234 - }, - { - "epoch": 1.5365986394557822, - "grad_norm": 0.1669921875, - "learning_rate": 9.878839799290173e-05, - "loss": 0.9699, - "step": 4236 - }, - { - "epoch": 1.5373242630385486, - "grad_norm": 0.1650390625, - "learning_rate": 9.873944437645331e-05, - "loss": 0.9659, - "step": 4238 - }, - { - "epoch": 1.538049886621315, - "grad_norm": 0.1533203125, - "learning_rate": 9.86904907600049e-05, - "loss": 0.9404, - "step": 4240 - }, - { - "epoch": 1.5387755102040817, - "grad_norm": 0.17578125, - "learning_rate": 9.864153714355649e-05, - "loss": 0.9822, - "step": 4242 - }, - { - "epoch": 1.539501133786848, - "grad_norm": 0.154296875, - "learning_rate": 9.859258352710807e-05, - "loss": 0.9127, - "step": 4244 - }, - { - "epoch": 1.5402267573696147, - "grad_norm": 0.1630859375, - "learning_rate": 9.854362991065965e-05, - "loss": 1.0608, - "step": 4246 - }, - { - "epoch": 1.540952380952381, - "grad_norm": 0.1689453125, - "learning_rate": 9.849467629421123e-05, - "loss": 1.0184, - "step": 4248 - }, - { - "epoch": 1.5416780045351475, - "grad_norm": 0.1767578125, - "learning_rate": 9.844572267776282e-05, - "loss": 1.0275, - "step": 4250 - }, - { - "epoch": 1.5424036281179139, - "grad_norm": 0.1630859375, - "learning_rate": 9.839676906131441e-05, - "loss": 0.9631, - "step": 4252 - }, - { - "epoch": 1.5431292517006803, - "grad_norm": 0.1611328125, - "learning_rate": 9.8347815444866e-05, - "loss": 0.9986, - "step": 4254 - }, - { - "epoch": 1.5438548752834467, - "grad_norm": 0.16015625, - "learning_rate": 9.829886182841758e-05, - "loss": 0.9216, - "step": 4256 - }, - { - "epoch": 1.544580498866213, - "grad_norm": 0.1669921875, - "learning_rate": 9.824990821196916e-05, - "loss": 0.939, - "step": 4258 - }, - { - "epoch": 1.5453061224489795, - "grad_norm": 0.1630859375, - "learning_rate": 9.820095459552074e-05, - "loss": 1.049, - "step": 4260 - }, - { - "epoch": 1.5460317460317459, - "grad_norm": 0.1435546875, - "learning_rate": 9.815200097907234e-05, - "loss": 0.9569, - "step": 4262 - }, - { - "epoch": 1.5467573696145125, - "grad_norm": 0.17578125, - "learning_rate": 9.810304736262392e-05, - "loss": 1.0119, - "step": 4264 - }, - { - "epoch": 1.5474829931972789, - "grad_norm": 0.1875, - "learning_rate": 9.80540937461755e-05, - "loss": 1.1073, - "step": 4266 - }, - { - "epoch": 1.5482086167800455, - "grad_norm": 0.1494140625, - "learning_rate": 9.800514012972708e-05, - "loss": 0.9696, - "step": 4268 - }, - { - "epoch": 1.548934240362812, - "grad_norm": 0.154296875, - "learning_rate": 9.795618651327867e-05, - "loss": 0.8951, - "step": 4270 - }, - { - "epoch": 1.5496598639455783, - "grad_norm": 0.193359375, - "learning_rate": 9.790723289683026e-05, - "loss": 0.94, - "step": 4272 - }, - { - "epoch": 1.5503854875283447, - "grad_norm": 0.1708984375, - "learning_rate": 9.785827928038184e-05, - "loss": 1.0004, - "step": 4274 - }, - { - "epoch": 1.551111111111111, - "grad_norm": 0.1591796875, - "learning_rate": 9.780932566393343e-05, - "loss": 0.9606, - "step": 4276 - }, - { - "epoch": 1.5518367346938775, - "grad_norm": 0.162109375, - "learning_rate": 9.776037204748501e-05, - "loss": 1.0272, - "step": 4278 - }, - { - "epoch": 1.552562358276644, - "grad_norm": 0.1767578125, - "learning_rate": 9.771141843103659e-05, - "loss": 0.9806, - "step": 4280 - }, - { - "epoch": 1.5532879818594103, - "grad_norm": 0.1484375, - "learning_rate": 9.766246481458819e-05, - "loss": 0.8835, - "step": 4282 - }, - { - "epoch": 1.5540136054421767, - "grad_norm": 0.16015625, - "learning_rate": 9.761351119813976e-05, - "loss": 0.9617, - "step": 4284 - }, - { - "epoch": 1.5547392290249433, - "grad_norm": 0.154296875, - "learning_rate": 9.756455758169135e-05, - "loss": 1.047, - "step": 4286 - }, - { - "epoch": 1.5554648526077097, - "grad_norm": 0.1591796875, - "learning_rate": 9.751560396524293e-05, - "loss": 0.9978, - "step": 4288 - }, - { - "epoch": 1.5561904761904763, - "grad_norm": 0.1689453125, - "learning_rate": 9.746665034879452e-05, - "loss": 0.9766, - "step": 4290 - }, - { - "epoch": 1.5569160997732427, - "grad_norm": 0.173828125, - "learning_rate": 9.741769673234611e-05, - "loss": 1.0446, - "step": 4292 - }, - { - "epoch": 1.5576417233560091, - "grad_norm": 0.177734375, - "learning_rate": 9.736874311589768e-05, - "loss": 1.0301, - "step": 4294 - }, - { - "epoch": 1.5583673469387755, - "grad_norm": 0.1767578125, - "learning_rate": 9.731978949944928e-05, - "loss": 1.0035, - "step": 4296 - }, - { - "epoch": 1.559092970521542, - "grad_norm": 0.1962890625, - "learning_rate": 9.727083588300086e-05, - "loss": 0.9809, - "step": 4298 - }, - { - "epoch": 1.5598185941043083, - "grad_norm": 0.1728515625, - "learning_rate": 9.722188226655244e-05, - "loss": 0.9959, - "step": 4300 - }, - { - "epoch": 1.5598185941043083, - "eval_loss": 1.0130585432052612, - "eval_runtime": 87.5971, - "eval_samples_per_second": 20.549, - "eval_steps_per_second": 0.651, - "step": 4300 - }, - { - "epoch": 1.5605442176870747, - "grad_norm": 0.15234375, - "learning_rate": 9.717292865010404e-05, - "loss": 1.0011, - "step": 4302 - }, - { - "epoch": 1.5612698412698411, - "grad_norm": 0.1767578125, - "learning_rate": 9.71239750336556e-05, - "loss": 0.934, - "step": 4304 - }, - { - "epoch": 1.5619954648526078, - "grad_norm": 0.14453125, - "learning_rate": 9.70750214172072e-05, - "loss": 0.9783, - "step": 4306 - }, - { - "epoch": 1.5627210884353742, - "grad_norm": 0.1796875, - "learning_rate": 9.702606780075878e-05, - "loss": 0.995, - "step": 4308 - }, - { - "epoch": 1.5634467120181406, - "grad_norm": 0.162109375, - "learning_rate": 9.697711418431037e-05, - "loss": 0.932, - "step": 4310 - }, - { - "epoch": 1.5641723356009072, - "grad_norm": 0.154296875, - "learning_rate": 9.692816056786196e-05, - "loss": 1.0029, - "step": 4312 - }, - { - "epoch": 1.5648979591836736, - "grad_norm": 0.1708984375, - "learning_rate": 9.687920695141353e-05, - "loss": 0.9971, - "step": 4314 - }, - { - "epoch": 1.56562358276644, - "grad_norm": 0.1845703125, - "learning_rate": 9.683025333496513e-05, - "loss": 0.9928, - "step": 4316 - }, - { - "epoch": 1.5663492063492064, - "grad_norm": 0.16015625, - "learning_rate": 9.678129971851671e-05, - "loss": 0.9279, - "step": 4318 - }, - { - "epoch": 1.5670748299319728, - "grad_norm": 0.17578125, - "learning_rate": 9.673234610206829e-05, - "loss": 0.937, - "step": 4320 - }, - { - "epoch": 1.5678004535147392, - "grad_norm": 0.1611328125, - "learning_rate": 9.668339248561989e-05, - "loss": 1.1021, - "step": 4322 - }, - { - "epoch": 1.5685260770975056, - "grad_norm": 0.1513671875, - "learning_rate": 9.663443886917146e-05, - "loss": 0.938, - "step": 4324 - }, - { - "epoch": 1.569251700680272, - "grad_norm": 0.177734375, - "learning_rate": 9.658548525272305e-05, - "loss": 0.992, - "step": 4326 - }, - { - "epoch": 1.5699773242630386, - "grad_norm": 0.1787109375, - "learning_rate": 9.653653163627463e-05, - "loss": 1.0522, - "step": 4328 - }, - { - "epoch": 1.570702947845805, - "grad_norm": 0.16015625, - "learning_rate": 9.648757801982622e-05, - "loss": 1.0242, - "step": 4330 - }, - { - "epoch": 1.5714285714285714, - "grad_norm": 0.154296875, - "learning_rate": 9.643862440337781e-05, - "loss": 0.9797, - "step": 4332 - }, - { - "epoch": 1.572154195011338, - "grad_norm": 0.185546875, - "learning_rate": 9.638967078692938e-05, - "loss": 1.0288, - "step": 4334 - }, - { - "epoch": 1.5728798185941044, - "grad_norm": 0.162109375, - "learning_rate": 9.634071717048098e-05, - "loss": 0.9567, - "step": 4336 - }, - { - "epoch": 1.5736054421768708, - "grad_norm": 0.1591796875, - "learning_rate": 9.629176355403256e-05, - "loss": 1.0486, - "step": 4338 - }, - { - "epoch": 1.5743310657596372, - "grad_norm": 0.1611328125, - "learning_rate": 9.624280993758414e-05, - "loss": 0.9999, - "step": 4340 - }, - { - "epoch": 1.5750566893424036, - "grad_norm": 0.1904296875, - "learning_rate": 9.619385632113574e-05, - "loss": 0.9742, - "step": 4342 - }, - { - "epoch": 1.57578231292517, - "grad_norm": 0.154296875, - "learning_rate": 9.61449027046873e-05, - "loss": 0.9984, - "step": 4344 - }, - { - "epoch": 1.5765079365079364, - "grad_norm": 0.15625, - "learning_rate": 9.60959490882389e-05, - "loss": 0.987, - "step": 4346 - }, - { - "epoch": 1.5772335600907028, - "grad_norm": 0.158203125, - "learning_rate": 9.604699547179048e-05, - "loss": 1.0147, - "step": 4348 - }, - { - "epoch": 1.5779591836734694, - "grad_norm": 0.169921875, - "learning_rate": 9.599804185534207e-05, - "loss": 0.9514, - "step": 4350 - }, - { - "epoch": 1.5786848072562358, - "grad_norm": 0.150390625, - "learning_rate": 9.594908823889365e-05, - "loss": 0.8686, - "step": 4352 - }, - { - "epoch": 1.5794104308390022, - "grad_norm": 0.173828125, - "learning_rate": 9.590013462244523e-05, - "loss": 0.9818, - "step": 4354 - }, - { - "epoch": 1.5801360544217689, - "grad_norm": 0.1572265625, - "learning_rate": 9.585118100599683e-05, - "loss": 0.9206, - "step": 4356 - }, - { - "epoch": 1.5808616780045353, - "grad_norm": 0.1630859375, - "learning_rate": 9.580222738954841e-05, - "loss": 1.0003, - "step": 4358 - }, - { - "epoch": 1.5815873015873017, - "grad_norm": 0.1533203125, - "learning_rate": 9.575327377309999e-05, - "loss": 0.9899, - "step": 4360 - }, - { - "epoch": 1.582312925170068, - "grad_norm": 0.1650390625, - "learning_rate": 9.570432015665157e-05, - "loss": 0.9102, - "step": 4362 - }, - { - "epoch": 1.5830385487528345, - "grad_norm": 0.1484375, - "learning_rate": 9.565536654020316e-05, - "loss": 0.9594, - "step": 4364 - }, - { - "epoch": 1.5837641723356009, - "grad_norm": 0.21875, - "learning_rate": 9.560641292375475e-05, - "loss": 0.9382, - "step": 4366 - }, - { - "epoch": 1.5844897959183672, - "grad_norm": 0.158203125, - "learning_rate": 9.555745930730633e-05, - "loss": 1.0231, - "step": 4368 - }, - { - "epoch": 1.5852154195011336, - "grad_norm": 0.1689453125, - "learning_rate": 9.550850569085792e-05, - "loss": 0.9838, - "step": 4370 - }, - { - "epoch": 1.5859410430839003, - "grad_norm": 0.17578125, - "learning_rate": 9.54595520744095e-05, - "loss": 1.0085, - "step": 4372 - }, - { - "epoch": 1.5866666666666667, - "grad_norm": 0.1640625, - "learning_rate": 9.541059845796108e-05, - "loss": 0.9167, - "step": 4374 - }, - { - "epoch": 1.587392290249433, - "grad_norm": 0.181640625, - "learning_rate": 9.536164484151268e-05, - "loss": 0.9559, - "step": 4376 - }, - { - "epoch": 1.5881179138321997, - "grad_norm": 0.1689453125, - "learning_rate": 9.531269122506426e-05, - "loss": 0.9368, - "step": 4378 - }, - { - "epoch": 1.588843537414966, - "grad_norm": 0.1767578125, - "learning_rate": 9.526373760861584e-05, - "loss": 0.9922, - "step": 4380 - }, - { - "epoch": 1.5895691609977325, - "grad_norm": 0.1640625, - "learning_rate": 9.521478399216742e-05, - "loss": 0.9523, - "step": 4382 - }, - { - "epoch": 1.5902947845804989, - "grad_norm": 0.17578125, - "learning_rate": 9.5165830375719e-05, - "loss": 0.9917, - "step": 4384 - }, - { - "epoch": 1.5910204081632653, - "grad_norm": 0.1787109375, - "learning_rate": 9.51168767592706e-05, - "loss": 0.9945, - "step": 4386 - }, - { - "epoch": 1.5917460317460317, - "grad_norm": 0.205078125, - "learning_rate": 9.506792314282218e-05, - "loss": 1.0831, - "step": 4388 - }, - { - "epoch": 1.592471655328798, - "grad_norm": 0.1982421875, - "learning_rate": 9.501896952637377e-05, - "loss": 1.0329, - "step": 4390 - }, - { - "epoch": 1.5931972789115645, - "grad_norm": 0.171875, - "learning_rate": 9.497001590992535e-05, - "loss": 0.8841, - "step": 4392 - }, - { - "epoch": 1.593922902494331, - "grad_norm": 0.1650390625, - "learning_rate": 9.492106229347693e-05, - "loss": 0.8609, - "step": 4394 - }, - { - "epoch": 1.5946485260770975, - "grad_norm": 0.16796875, - "learning_rate": 9.487210867702851e-05, - "loss": 0.9959, - "step": 4396 - }, - { - "epoch": 1.595374149659864, - "grad_norm": 0.1669921875, - "learning_rate": 9.482315506058011e-05, - "loss": 0.9974, - "step": 4398 - }, - { - "epoch": 1.5960997732426305, - "grad_norm": 0.1728515625, - "learning_rate": 9.477420144413169e-05, - "loss": 0.9548, - "step": 4400 - }, - { - "epoch": 1.5960997732426305, - "eval_loss": 1.0129448175430298, - "eval_runtime": 87.6312, - "eval_samples_per_second": 20.541, - "eval_steps_per_second": 0.65, - "step": 4400 - }, - { - "epoch": 1.596825396825397, - "grad_norm": 0.166015625, - "learning_rate": 9.472524782768327e-05, - "loss": 1.0032, - "step": 4402 - }, - { - "epoch": 1.5975510204081633, - "grad_norm": 0.1787109375, - "learning_rate": 9.467629421123486e-05, - "loss": 0.9867, - "step": 4404 - }, - { - "epoch": 1.5982766439909297, - "grad_norm": 0.162109375, - "learning_rate": 9.462734059478644e-05, - "loss": 1.0104, - "step": 4406 - }, - { - "epoch": 1.5990022675736961, - "grad_norm": 0.177734375, - "learning_rate": 9.457838697833803e-05, - "loss": 0.881, - "step": 4408 - }, - { - "epoch": 1.5997278911564625, - "grad_norm": 0.1650390625, - "learning_rate": 9.452943336188962e-05, - "loss": 0.9992, - "step": 4410 - }, - { - "epoch": 1.600453514739229, - "grad_norm": 0.2236328125, - "learning_rate": 9.44804797454412e-05, - "loss": 0.9666, - "step": 4412 - }, - { - "epoch": 1.6011791383219953, - "grad_norm": 0.1748046875, - "learning_rate": 9.443152612899278e-05, - "loss": 1.0162, - "step": 4414 - }, - { - "epoch": 1.601904761904762, - "grad_norm": 0.1494140625, - "learning_rate": 9.438257251254436e-05, - "loss": 0.8902, - "step": 4416 - }, - { - "epoch": 1.6026303854875283, - "grad_norm": 0.1650390625, - "learning_rate": 9.433361889609596e-05, - "loss": 0.9424, - "step": 4418 - }, - { - "epoch": 1.6033560090702947, - "grad_norm": 0.1767578125, - "learning_rate": 9.428466527964754e-05, - "loss": 0.9927, - "step": 4420 - }, - { - "epoch": 1.6040816326530614, - "grad_norm": 0.177734375, - "learning_rate": 9.423571166319912e-05, - "loss": 0.9972, - "step": 4422 - }, - { - "epoch": 1.6048072562358278, - "grad_norm": 0.19921875, - "learning_rate": 9.41867580467507e-05, - "loss": 0.938, - "step": 4424 - }, - { - "epoch": 1.6055328798185942, - "grad_norm": 0.1552734375, - "learning_rate": 9.413780443030229e-05, - "loss": 0.8708, - "step": 4426 - }, - { - "epoch": 1.6062585034013606, - "grad_norm": 0.1796875, - "learning_rate": 9.408885081385388e-05, - "loss": 1.0519, - "step": 4428 - }, - { - "epoch": 1.606984126984127, - "grad_norm": 0.1748046875, - "learning_rate": 9.403989719740545e-05, - "loss": 0.9837, - "step": 4430 - }, - { - "epoch": 1.6077097505668934, - "grad_norm": 0.1611328125, - "learning_rate": 9.399094358095705e-05, - "loss": 1.0007, - "step": 4432 - }, - { - "epoch": 1.6084353741496598, - "grad_norm": 0.1767578125, - "learning_rate": 9.394198996450863e-05, - "loss": 0.9906, - "step": 4434 - }, - { - "epoch": 1.6091609977324262, - "grad_norm": 0.1533203125, - "learning_rate": 9.389303634806021e-05, - "loss": 0.9736, - "step": 4436 - }, - { - "epoch": 1.6098866213151928, - "grad_norm": 0.193359375, - "learning_rate": 9.384408273161181e-05, - "loss": 1.0087, - "step": 4438 - }, - { - "epoch": 1.6106122448979592, - "grad_norm": 0.1455078125, - "learning_rate": 9.379512911516338e-05, - "loss": 0.9447, - "step": 4440 - }, - { - "epoch": 1.6113378684807256, - "grad_norm": 0.1708984375, - "learning_rate": 9.374617549871497e-05, - "loss": 1.0428, - "step": 4442 - }, - { - "epoch": 1.6120634920634922, - "grad_norm": 0.154296875, - "learning_rate": 9.369722188226656e-05, - "loss": 0.9664, - "step": 4444 - }, - { - "epoch": 1.6127891156462586, - "grad_norm": 0.173828125, - "learning_rate": 9.364826826581814e-05, - "loss": 0.9886, - "step": 4446 - }, - { - "epoch": 1.613514739229025, - "grad_norm": 0.19921875, - "learning_rate": 9.359931464936973e-05, - "loss": 1.0257, - "step": 4448 - }, - { - "epoch": 1.6142403628117914, - "grad_norm": 0.15234375, - "learning_rate": 9.35503610329213e-05, - "loss": 0.9166, - "step": 4450 - }, - { - "epoch": 1.6149659863945578, - "grad_norm": 0.201171875, - "learning_rate": 9.35014074164729e-05, - "loss": 1.0876, - "step": 4452 - }, - { - "epoch": 1.6156916099773242, - "grad_norm": 0.1533203125, - "learning_rate": 9.345245380002448e-05, - "loss": 0.9517, - "step": 4454 - }, - { - "epoch": 1.6164172335600906, - "grad_norm": 0.154296875, - "learning_rate": 9.340350018357606e-05, - "loss": 0.9019, - "step": 4456 - }, - { - "epoch": 1.617142857142857, - "grad_norm": 0.1533203125, - "learning_rate": 9.335454656712766e-05, - "loss": 0.9497, - "step": 4458 - }, - { - "epoch": 1.6178684807256236, - "grad_norm": 0.1572265625, - "learning_rate": 9.330559295067923e-05, - "loss": 0.9503, - "step": 4460 - }, - { - "epoch": 1.61859410430839, - "grad_norm": 0.1689453125, - "learning_rate": 9.325663933423082e-05, - "loss": 0.9898, - "step": 4462 - }, - { - "epoch": 1.6193197278911564, - "grad_norm": 0.15234375, - "learning_rate": 9.32076857177824e-05, - "loss": 0.9641, - "step": 4464 - }, - { - "epoch": 1.620045351473923, - "grad_norm": 0.1865234375, - "learning_rate": 9.315873210133399e-05, - "loss": 0.963, - "step": 4466 - }, - { - "epoch": 1.6207709750566894, - "grad_norm": 0.1826171875, - "learning_rate": 9.310977848488558e-05, - "loss": 0.9416, - "step": 4468 - }, - { - "epoch": 1.6214965986394558, - "grad_norm": 0.162109375, - "learning_rate": 9.306082486843715e-05, - "loss": 1.0424, - "step": 4470 - }, - { - "epoch": 1.6222222222222222, - "grad_norm": 0.1474609375, - "learning_rate": 9.301187125198875e-05, - "loss": 0.9659, - "step": 4472 - }, - { - "epoch": 1.6229478458049886, - "grad_norm": 0.1513671875, - "learning_rate": 9.296291763554033e-05, - "loss": 0.8367, - "step": 4474 - }, - { - "epoch": 1.623673469387755, - "grad_norm": 0.1455078125, - "learning_rate": 9.291396401909191e-05, - "loss": 0.9668, - "step": 4476 - }, - { - "epoch": 1.6243990929705214, - "grad_norm": 0.1416015625, - "learning_rate": 9.286501040264351e-05, - "loss": 0.9655, - "step": 4478 - }, - { - "epoch": 1.6251247165532878, - "grad_norm": 0.189453125, - "learning_rate": 9.281605678619508e-05, - "loss": 0.9554, - "step": 4480 - }, - { - "epoch": 1.6258503401360545, - "grad_norm": 0.158203125, - "learning_rate": 9.276710316974667e-05, - "loss": 0.9243, - "step": 4482 - }, - { - "epoch": 1.6265759637188208, - "grad_norm": 0.1572265625, - "learning_rate": 9.271814955329826e-05, - "loss": 0.9596, - "step": 4484 - }, - { - "epoch": 1.6273015873015872, - "grad_norm": 0.15625, - "learning_rate": 9.266919593684984e-05, - "loss": 0.9765, - "step": 4486 - }, - { - "epoch": 1.6280272108843539, - "grad_norm": 0.16015625, - "learning_rate": 9.262024232040143e-05, - "loss": 0.9482, - "step": 4488 - }, - { - "epoch": 1.6287528344671203, - "grad_norm": 0.1689453125, - "learning_rate": 9.2571288703953e-05, - "loss": 0.9621, - "step": 4490 - }, - { - "epoch": 1.6294784580498867, - "grad_norm": 0.18359375, - "learning_rate": 9.25223350875046e-05, - "loss": 1.0181, - "step": 4492 - }, - { - "epoch": 1.630204081632653, - "grad_norm": 0.1669921875, - "learning_rate": 9.247338147105618e-05, - "loss": 0.9446, - "step": 4494 - }, - { - "epoch": 1.6309297052154195, - "grad_norm": 0.162109375, - "learning_rate": 9.242442785460776e-05, - "loss": 0.9495, - "step": 4496 - }, - { - "epoch": 1.6316553287981859, - "grad_norm": 0.169921875, - "learning_rate": 9.237547423815934e-05, - "loss": 1.0283, - "step": 4498 - }, - { - "epoch": 1.6323809523809523, - "grad_norm": 0.1826171875, - "learning_rate": 9.232652062171093e-05, - "loss": 1.0349, - "step": 4500 - }, - { - "epoch": 1.6323809523809523, - "eval_loss": 1.0125901699066162, - "eval_runtime": 86.0097, - "eval_samples_per_second": 20.928, - "eval_steps_per_second": 0.663, - "step": 4500 - }, - { - "epoch": 1.6331065759637187, - "grad_norm": 0.169921875, - "learning_rate": 9.227756700526252e-05, - "loss": 1.0119, - "step": 4502 - }, - { - "epoch": 1.6338321995464853, - "grad_norm": 0.162109375, - "learning_rate": 9.22286133888141e-05, - "loss": 0.9775, - "step": 4504 - }, - { - "epoch": 1.6345578231292517, - "grad_norm": 0.177734375, - "learning_rate": 9.217965977236569e-05, - "loss": 0.9544, - "step": 4506 - }, - { - "epoch": 1.6352834467120183, - "grad_norm": 0.16796875, - "learning_rate": 9.213070615591727e-05, - "loss": 0.941, - "step": 4508 - }, - { - "epoch": 1.6360090702947847, - "grad_norm": 0.1669921875, - "learning_rate": 9.208175253946885e-05, - "loss": 0.9603, - "step": 4510 - }, - { - "epoch": 1.636734693877551, - "grad_norm": 0.154296875, - "learning_rate": 9.203279892302045e-05, - "loss": 0.9505, - "step": 4512 - }, - { - "epoch": 1.6374603174603175, - "grad_norm": 0.1689453125, - "learning_rate": 9.198384530657203e-05, - "loss": 1.0291, - "step": 4514 - }, - { - "epoch": 1.638185941043084, - "grad_norm": 0.1923828125, - "learning_rate": 9.193489169012361e-05, - "loss": 0.8945, - "step": 4516 - }, - { - "epoch": 1.6389115646258503, - "grad_norm": 0.166015625, - "learning_rate": 9.18859380736752e-05, - "loss": 1.0053, - "step": 4518 - }, - { - "epoch": 1.6396371882086167, - "grad_norm": 0.17578125, - "learning_rate": 9.183698445722678e-05, - "loss": 0.9705, - "step": 4520 - }, - { - "epoch": 1.640362811791383, - "grad_norm": 0.1611328125, - "learning_rate": 9.178803084077837e-05, - "loss": 0.9802, - "step": 4522 - }, - { - "epoch": 1.6410884353741495, - "grad_norm": 0.1611328125, - "learning_rate": 9.173907722432996e-05, - "loss": 0.9674, - "step": 4524 - }, - { - "epoch": 1.6418140589569161, - "grad_norm": 0.1708984375, - "learning_rate": 9.169012360788154e-05, - "loss": 1.0555, - "step": 4526 - }, - { - "epoch": 1.6425396825396825, - "grad_norm": 0.1689453125, - "learning_rate": 9.164116999143312e-05, - "loss": 0.9737, - "step": 4528 - }, - { - "epoch": 1.6432653061224491, - "grad_norm": 0.1845703125, - "learning_rate": 9.15922163749847e-05, - "loss": 0.9983, - "step": 4530 - }, - { - "epoch": 1.6439909297052155, - "grad_norm": 0.1552734375, - "learning_rate": 9.15432627585363e-05, - "loss": 0.9307, - "step": 4532 - }, - { - "epoch": 1.644716553287982, - "grad_norm": 0.1552734375, - "learning_rate": 9.149430914208788e-05, - "loss": 0.9447, - "step": 4534 - }, - { - "epoch": 1.6454421768707483, - "grad_norm": 0.15234375, - "learning_rate": 9.144535552563946e-05, - "loss": 0.9611, - "step": 4536 - }, - { - "epoch": 1.6461678004535147, - "grad_norm": 0.1787109375, - "learning_rate": 9.139640190919104e-05, - "loss": 1.0478, - "step": 4538 - }, - { - "epoch": 1.6468934240362811, - "grad_norm": 0.17578125, - "learning_rate": 9.134744829274263e-05, - "loss": 1.002, - "step": 4540 - }, - { - "epoch": 1.6476190476190475, - "grad_norm": 0.1533203125, - "learning_rate": 9.129849467629421e-05, - "loss": 0.9249, - "step": 4542 - }, - { - "epoch": 1.648344671201814, - "grad_norm": 0.1826171875, - "learning_rate": 9.12495410598458e-05, - "loss": 0.9697, - "step": 4544 - }, - { - "epoch": 1.6490702947845803, - "grad_norm": 0.1728515625, - "learning_rate": 9.120058744339739e-05, - "loss": 0.9693, - "step": 4546 - }, - { - "epoch": 1.649795918367347, - "grad_norm": 0.1806640625, - "learning_rate": 9.115163382694897e-05, - "loss": 0.9989, - "step": 4548 - }, - { - "epoch": 1.6505215419501134, - "grad_norm": 0.1650390625, - "learning_rate": 9.110268021050055e-05, - "loss": 1.0239, - "step": 4550 - }, - { - "epoch": 1.65124716553288, - "grad_norm": 0.1533203125, - "learning_rate": 9.105372659405213e-05, - "loss": 0.9383, - "step": 4552 - }, - { - "epoch": 1.6519727891156464, - "grad_norm": 0.1796875, - "learning_rate": 9.100477297760373e-05, - "loss": 0.9475, - "step": 4554 - }, - { - "epoch": 1.6526984126984128, - "grad_norm": 0.1474609375, - "learning_rate": 9.095581936115531e-05, - "loss": 0.9904, - "step": 4556 - }, - { - "epoch": 1.6534240362811792, - "grad_norm": 0.16796875, - "learning_rate": 9.09068657447069e-05, - "loss": 0.9174, - "step": 4558 - }, - { - "epoch": 1.6541496598639456, - "grad_norm": 0.166015625, - "learning_rate": 9.085791212825848e-05, - "loss": 1.0161, - "step": 4560 - }, - { - "epoch": 1.654875283446712, - "grad_norm": 0.16015625, - "learning_rate": 9.080895851181006e-05, - "loss": 1.1105, - "step": 4562 - }, - { - "epoch": 1.6556009070294784, - "grad_norm": 0.1669921875, - "learning_rate": 9.076000489536165e-05, - "loss": 1.0257, - "step": 4564 - }, - { - "epoch": 1.6563265306122448, - "grad_norm": 0.1513671875, - "learning_rate": 9.071105127891324e-05, - "loss": 0.9142, - "step": 4566 - }, - { - "epoch": 1.6570521541950112, - "grad_norm": 0.1826171875, - "learning_rate": 9.066209766246482e-05, - "loss": 0.9849, - "step": 4568 - }, - { - "epoch": 1.6577777777777778, - "grad_norm": 0.1787109375, - "learning_rate": 9.06131440460164e-05, - "loss": 0.9778, - "step": 4570 - }, - { - "epoch": 1.6585034013605442, - "grad_norm": 0.1787109375, - "learning_rate": 9.056419042956798e-05, - "loss": 0.9673, - "step": 4572 - }, - { - "epoch": 1.6592290249433108, - "grad_norm": 0.16015625, - "learning_rate": 9.051523681311958e-05, - "loss": 1.0083, - "step": 4574 - }, - { - "epoch": 1.6599546485260772, - "grad_norm": 0.1689453125, - "learning_rate": 9.046628319667115e-05, - "loss": 0.9852, - "step": 4576 - }, - { - "epoch": 1.6606802721088436, - "grad_norm": 0.19921875, - "learning_rate": 9.041732958022274e-05, - "loss": 0.9384, - "step": 4578 - }, - { - "epoch": 1.66140589569161, - "grad_norm": 0.166015625, - "learning_rate": 9.036837596377433e-05, - "loss": 1.0178, - "step": 4580 - }, - { - "epoch": 1.6621315192743764, - "grad_norm": 0.166015625, - "learning_rate": 9.031942234732591e-05, - "loss": 0.9881, - "step": 4582 - }, - { - "epoch": 1.6628571428571428, - "grad_norm": 0.1845703125, - "learning_rate": 9.02704687308775e-05, - "loss": 1.0091, - "step": 4584 - }, - { - "epoch": 1.6635827664399092, - "grad_norm": 0.1767578125, - "learning_rate": 9.022151511442907e-05, - "loss": 1.0278, - "step": 4586 - }, - { - "epoch": 1.6643083900226756, - "grad_norm": 0.1640625, - "learning_rate": 9.017256149798067e-05, - "loss": 0.9655, - "step": 4588 - }, - { - "epoch": 1.6650340136054422, - "grad_norm": 0.15625, - "learning_rate": 9.012360788153225e-05, - "loss": 0.9821, - "step": 4590 - }, - { - "epoch": 1.6657596371882086, - "grad_norm": 0.1572265625, - "learning_rate": 9.007465426508383e-05, - "loss": 1.0376, - "step": 4592 - }, - { - "epoch": 1.666485260770975, - "grad_norm": 0.150390625, - "learning_rate": 9.002570064863543e-05, - "loss": 0.9686, - "step": 4594 - }, - { - "epoch": 1.6672108843537417, - "grad_norm": 0.1787109375, - "learning_rate": 8.9976747032187e-05, - "loss": 1.0581, - "step": 4596 - }, - { - "epoch": 1.667936507936508, - "grad_norm": 0.201171875, - "learning_rate": 8.99277934157386e-05, - "loss": 0.9976, - "step": 4598 - }, - { - "epoch": 1.6686621315192744, - "grad_norm": 0.1806640625, - "learning_rate": 8.987883979929018e-05, - "loss": 0.9585, - "step": 4600 - }, - { - "epoch": 1.6686621315192744, - "eval_loss": 1.0123869180679321, - "eval_runtime": 90.9583, - "eval_samples_per_second": 19.789, - "eval_steps_per_second": 0.627, - "step": 4600 - }, - { - "epoch": 1.6693877551020408, - "grad_norm": 0.15234375, - "learning_rate": 8.982988618284176e-05, - "loss": 0.9106, - "step": 4602 - }, - { - "epoch": 1.6701133786848072, - "grad_norm": 0.1611328125, - "learning_rate": 8.978093256639335e-05, - "loss": 0.9744, - "step": 4604 - }, - { - "epoch": 1.6708390022675736, - "grad_norm": 0.1875, - "learning_rate": 8.973197894994492e-05, - "loss": 0.9738, - "step": 4606 - }, - { - "epoch": 1.67156462585034, - "grad_norm": 0.158203125, - "learning_rate": 8.968302533349652e-05, - "loss": 0.9827, - "step": 4608 - }, - { - "epoch": 1.6722902494331064, - "grad_norm": 0.220703125, - "learning_rate": 8.96340717170481e-05, - "loss": 1.0603, - "step": 4610 - }, - { - "epoch": 1.673015873015873, - "grad_norm": 0.162109375, - "learning_rate": 8.958511810059968e-05, - "loss": 0.9742, - "step": 4612 - }, - { - "epoch": 1.6737414965986395, - "grad_norm": 0.1591796875, - "learning_rate": 8.953616448415128e-05, - "loss": 0.9745, - "step": 4614 - }, - { - "epoch": 1.6744671201814059, - "grad_norm": 0.1806640625, - "learning_rate": 8.948721086770285e-05, - "loss": 1.039, - "step": 4616 - }, - { - "epoch": 1.6751927437641725, - "grad_norm": 0.1748046875, - "learning_rate": 8.943825725125444e-05, - "loss": 0.917, - "step": 4618 - }, - { - "epoch": 1.6759183673469389, - "grad_norm": 0.1728515625, - "learning_rate": 8.938930363480603e-05, - "loss": 1.0145, - "step": 4620 - }, - { - "epoch": 1.6766439909297053, - "grad_norm": 0.1552734375, - "learning_rate": 8.934035001835761e-05, - "loss": 0.8931, - "step": 4622 - }, - { - "epoch": 1.6773696145124717, - "grad_norm": 0.169921875, - "learning_rate": 8.92913964019092e-05, - "loss": 0.9811, - "step": 4624 - }, - { - "epoch": 1.678095238095238, - "grad_norm": 0.185546875, - "learning_rate": 8.924244278546077e-05, - "loss": 0.9463, - "step": 4626 - }, - { - "epoch": 1.6788208616780045, - "grad_norm": 0.1552734375, - "learning_rate": 8.919348916901237e-05, - "loss": 0.9064, - "step": 4628 - }, - { - "epoch": 1.6795464852607709, - "grad_norm": 0.177734375, - "learning_rate": 8.914453555256395e-05, - "loss": 1.0477, - "step": 4630 - }, - { - "epoch": 1.6802721088435373, - "grad_norm": 0.201171875, - "learning_rate": 8.909558193611553e-05, - "loss": 1.0279, - "step": 4632 - }, - { - "epoch": 1.680997732426304, - "grad_norm": 0.1640625, - "learning_rate": 8.904662831966713e-05, - "loss": 0.9726, - "step": 4634 - }, - { - "epoch": 1.6817233560090703, - "grad_norm": 0.1484375, - "learning_rate": 8.89976747032187e-05, - "loss": 0.9378, - "step": 4636 - }, - { - "epoch": 1.6824489795918367, - "grad_norm": 0.177734375, - "learning_rate": 8.89487210867703e-05, - "loss": 0.9877, - "step": 4638 - }, - { - "epoch": 1.6831746031746033, - "grad_norm": 0.1494140625, - "learning_rate": 8.889976747032188e-05, - "loss": 0.9537, - "step": 4640 - }, - { - "epoch": 1.6839002267573697, - "grad_norm": 0.181640625, - "learning_rate": 8.885081385387346e-05, - "loss": 0.9996, - "step": 4642 - }, - { - "epoch": 1.6846258503401361, - "grad_norm": 0.1611328125, - "learning_rate": 8.880186023742504e-05, - "loss": 0.9682, - "step": 4644 - }, - { - "epoch": 1.6853514739229025, - "grad_norm": 0.1533203125, - "learning_rate": 8.875290662097662e-05, - "loss": 0.995, - "step": 4646 - }, - { - "epoch": 1.686077097505669, - "grad_norm": 0.1455078125, - "learning_rate": 8.870395300452822e-05, - "loss": 0.9594, - "step": 4648 - }, - { - "epoch": 1.6868027210884353, - "grad_norm": 0.16796875, - "learning_rate": 8.86549993880798e-05, - "loss": 0.9323, - "step": 4650 - }, - { - "epoch": 1.6875283446712017, - "grad_norm": 0.171875, - "learning_rate": 8.860604577163138e-05, - "loss": 1.0264, - "step": 4652 - }, - { - "epoch": 1.6882539682539681, - "grad_norm": 0.1552734375, - "learning_rate": 8.855709215518297e-05, - "loss": 0.9587, - "step": 4654 - }, - { - "epoch": 1.6889795918367347, - "grad_norm": 0.16796875, - "learning_rate": 8.850813853873455e-05, - "loss": 0.9402, - "step": 4656 - }, - { - "epoch": 1.6897052154195011, - "grad_norm": 0.1708984375, - "learning_rate": 8.845918492228614e-05, - "loss": 0.9673, - "step": 4658 - }, - { - "epoch": 1.6904308390022675, - "grad_norm": 0.1611328125, - "learning_rate": 8.841023130583773e-05, - "loss": 0.9555, - "step": 4660 - }, - { - "epoch": 1.6911564625850342, - "grad_norm": 0.1650390625, - "learning_rate": 8.836127768938931e-05, - "loss": 0.9361, - "step": 4662 - }, - { - "epoch": 1.6918820861678006, - "grad_norm": 0.1533203125, - "learning_rate": 8.831232407294089e-05, - "loss": 0.9679, - "step": 4664 - }, - { - "epoch": 1.692607709750567, - "grad_norm": 0.166015625, - "learning_rate": 8.826337045649247e-05, - "loss": 0.9455, - "step": 4666 - }, - { - "epoch": 1.6933333333333334, - "grad_norm": 0.1669921875, - "learning_rate": 8.821441684004407e-05, - "loss": 0.9773, - "step": 4668 - }, - { - "epoch": 1.6940589569160998, - "grad_norm": 0.1591796875, - "learning_rate": 8.816546322359565e-05, - "loss": 0.95, - "step": 4670 - }, - { - "epoch": 1.6947845804988662, - "grad_norm": 0.150390625, - "learning_rate": 8.811650960714723e-05, - "loss": 0.9737, - "step": 4672 - }, - { - "epoch": 1.6955102040816326, - "grad_norm": 0.1708984375, - "learning_rate": 8.806755599069882e-05, - "loss": 1.0247, - "step": 4674 - }, - { - "epoch": 1.696235827664399, - "grad_norm": 0.1474609375, - "learning_rate": 8.80186023742504e-05, - "loss": 0.9881, - "step": 4676 - }, - { - "epoch": 1.6969614512471656, - "grad_norm": 0.173828125, - "learning_rate": 8.7969648757802e-05, - "loss": 1.0309, - "step": 4678 - }, - { - "epoch": 1.697687074829932, - "grad_norm": 0.1689453125, - "learning_rate": 8.792069514135358e-05, - "loss": 0.9641, - "step": 4680 - }, - { - "epoch": 1.6984126984126984, - "grad_norm": 0.1640625, - "learning_rate": 8.787174152490516e-05, - "loss": 0.9768, - "step": 4682 - }, - { - "epoch": 1.699138321995465, - "grad_norm": 0.1650390625, - "learning_rate": 8.782278790845674e-05, - "loss": 0.9526, - "step": 4684 - }, - { - "epoch": 1.6998639455782314, - "grad_norm": 0.1611328125, - "learning_rate": 8.777383429200832e-05, - "loss": 1.0107, - "step": 4686 - }, - { - "epoch": 1.7005895691609978, - "grad_norm": 0.17578125, - "learning_rate": 8.77248806755599e-05, - "loss": 1.0372, - "step": 4688 - }, - { - "epoch": 1.7013151927437642, - "grad_norm": 0.16015625, - "learning_rate": 8.76759270591115e-05, - "loss": 0.949, - "step": 4690 - }, - { - "epoch": 1.7020408163265306, - "grad_norm": 0.1552734375, - "learning_rate": 8.762697344266308e-05, - "loss": 0.9846, - "step": 4692 - }, - { - "epoch": 1.702766439909297, - "grad_norm": 0.1904296875, - "learning_rate": 8.757801982621467e-05, - "loss": 0.9488, - "step": 4694 - }, - { - "epoch": 1.7034920634920634, - "grad_norm": 0.1455078125, - "learning_rate": 8.752906620976625e-05, - "loss": 0.9439, - "step": 4696 - }, - { - "epoch": 1.7042176870748298, - "grad_norm": 0.1669921875, - "learning_rate": 8.748011259331783e-05, - "loss": 0.9623, - "step": 4698 - }, - { - "epoch": 1.7049433106575964, - "grad_norm": 0.1728515625, - "learning_rate": 8.743115897686943e-05, - "loss": 1.0633, - "step": 4700 - }, - { - "epoch": 1.7049433106575964, - "eval_loss": 1.0122075080871582, - "eval_runtime": 84.9355, - "eval_samples_per_second": 21.193, - "eval_steps_per_second": 0.671, - "step": 4700 - }, - { - "epoch": 1.7056689342403628, - "grad_norm": 0.1591796875, - "learning_rate": 8.738220536042101e-05, - "loss": 0.9102, - "step": 4702 - }, - { - "epoch": 1.7063945578231292, - "grad_norm": 0.1552734375, - "learning_rate": 8.733325174397259e-05, - "loss": 0.9428, - "step": 4704 - }, - { - "epoch": 1.7071201814058958, - "grad_norm": 0.15625, - "learning_rate": 8.728429812752417e-05, - "loss": 0.9601, - "step": 4706 - }, - { - "epoch": 1.7078458049886622, - "grad_norm": 0.16015625, - "learning_rate": 8.723534451107576e-05, - "loss": 0.9997, - "step": 4708 - }, - { - "epoch": 1.7085714285714286, - "grad_norm": 0.1806640625, - "learning_rate": 8.718639089462735e-05, - "loss": 0.9719, - "step": 4710 - }, - { - "epoch": 1.709297052154195, - "grad_norm": 0.162109375, - "learning_rate": 8.713743727817893e-05, - "loss": 0.938, - "step": 4712 - }, - { - "epoch": 1.7100226757369614, - "grad_norm": 0.16796875, - "learning_rate": 8.708848366173052e-05, - "loss": 0.9769, - "step": 4714 - }, - { - "epoch": 1.7107482993197278, - "grad_norm": 0.18359375, - "learning_rate": 8.70395300452821e-05, - "loss": 0.9675, - "step": 4716 - }, - { - "epoch": 1.7114739229024942, - "grad_norm": 0.1865234375, - "learning_rate": 8.699057642883368e-05, - "loss": 0.9881, - "step": 4718 - }, - { - "epoch": 1.7121995464852606, - "grad_norm": 0.1640625, - "learning_rate": 8.694162281238528e-05, - "loss": 1.0473, - "step": 4720 - }, - { - "epoch": 1.7129251700680272, - "grad_norm": 0.15234375, - "learning_rate": 8.689266919593684e-05, - "loss": 0.9488, - "step": 4722 - }, - { - "epoch": 1.7136507936507936, - "grad_norm": 0.1796875, - "learning_rate": 8.684371557948844e-05, - "loss": 1.0057, - "step": 4724 - }, - { - "epoch": 1.71437641723356, - "grad_norm": 0.1689453125, - "learning_rate": 8.679476196304002e-05, - "loss": 0.9249, - "step": 4726 - }, - { - "epoch": 1.7151020408163267, - "grad_norm": 0.1943359375, - "learning_rate": 8.67458083465916e-05, - "loss": 0.9699, - "step": 4728 - }, - { - "epoch": 1.715827664399093, - "grad_norm": 0.1787109375, - "learning_rate": 8.66968547301432e-05, - "loss": 1.001, - "step": 4730 - }, - { - "epoch": 1.7165532879818595, - "grad_norm": 0.1708984375, - "learning_rate": 8.664790111369477e-05, - "loss": 0.901, - "step": 4732 - }, - { - "epoch": 1.7172789115646259, - "grad_norm": 0.173828125, - "learning_rate": 8.659894749724637e-05, - "loss": 1.026, - "step": 4734 - }, - { - "epoch": 1.7180045351473923, - "grad_norm": 0.173828125, - "learning_rate": 8.654999388079795e-05, - "loss": 0.9686, - "step": 4736 - }, - { - "epoch": 1.7187301587301587, - "grad_norm": 0.203125, - "learning_rate": 8.650104026434953e-05, - "loss": 1.0217, - "step": 4738 - }, - { - "epoch": 1.719455782312925, - "grad_norm": 0.1669921875, - "learning_rate": 8.645208664790113e-05, - "loss": 1.0212, - "step": 4740 - }, - { - "epoch": 1.7201814058956915, - "grad_norm": 0.162109375, - "learning_rate": 8.64031330314527e-05, - "loss": 0.9317, - "step": 4742 - }, - { - "epoch": 1.720907029478458, - "grad_norm": 0.158203125, - "learning_rate": 8.635417941500429e-05, - "loss": 0.9294, - "step": 4744 - }, - { - "epoch": 1.7216326530612245, - "grad_norm": 0.1474609375, - "learning_rate": 8.630522579855587e-05, - "loss": 0.9838, - "step": 4746 - }, - { - "epoch": 1.7223582766439909, - "grad_norm": 0.1796875, - "learning_rate": 8.625627218210746e-05, - "loss": 0.9376, - "step": 4748 - }, - { - "epoch": 1.7230839002267575, - "grad_norm": 0.177734375, - "learning_rate": 8.620731856565905e-05, - "loss": 1.0101, - "step": 4750 - }, - { - "epoch": 1.723809523809524, - "grad_norm": 0.16796875, - "learning_rate": 8.615836494921062e-05, - "loss": 0.9394, - "step": 4752 - }, - { - "epoch": 1.7245351473922903, - "grad_norm": 0.1689453125, - "learning_rate": 8.610941133276222e-05, - "loss": 1.0197, - "step": 4754 - }, - { - "epoch": 1.7252607709750567, - "grad_norm": 0.166015625, - "learning_rate": 8.60604577163138e-05, - "loss": 1.0194, - "step": 4756 - }, - { - "epoch": 1.725986394557823, - "grad_norm": 0.1865234375, - "learning_rate": 8.601150409986538e-05, - "loss": 1.0242, - "step": 4758 - }, - { - "epoch": 1.7267120181405895, - "grad_norm": 0.185546875, - "learning_rate": 8.596255048341698e-05, - "loss": 0.9163, - "step": 4760 - }, - { - "epoch": 1.727437641723356, - "grad_norm": 0.1826171875, - "learning_rate": 8.591359686696854e-05, - "loss": 0.9777, - "step": 4762 - }, - { - "epoch": 1.7281632653061223, - "grad_norm": 0.162109375, - "learning_rate": 8.586464325052014e-05, - "loss": 0.9872, - "step": 4764 - }, - { - "epoch": 1.728888888888889, - "grad_norm": 0.162109375, - "learning_rate": 8.581568963407172e-05, - "loss": 1.059, - "step": 4766 - }, - { - "epoch": 1.7296145124716553, - "grad_norm": 0.1845703125, - "learning_rate": 8.57667360176233e-05, - "loss": 1.0383, - "step": 4768 - }, - { - "epoch": 1.7303401360544217, - "grad_norm": 0.158203125, - "learning_rate": 8.57177824011749e-05, - "loss": 0.9363, - "step": 4770 - }, - { - "epoch": 1.7310657596371883, - "grad_norm": 0.154296875, - "learning_rate": 8.566882878472647e-05, - "loss": 0.8923, - "step": 4772 - }, - { - "epoch": 1.7317913832199547, - "grad_norm": 0.1640625, - "learning_rate": 8.561987516827807e-05, - "loss": 0.9878, - "step": 4774 - }, - { - "epoch": 1.7325170068027211, - "grad_norm": 0.1875, - "learning_rate": 8.557092155182965e-05, - "loss": 1.0398, - "step": 4776 - }, - { - "epoch": 1.7332426303854875, - "grad_norm": 0.19140625, - "learning_rate": 8.552196793538123e-05, - "loss": 0.9416, - "step": 4778 - }, - { - "epoch": 1.733968253968254, - "grad_norm": 0.1953125, - "learning_rate": 8.547301431893283e-05, - "loss": 1.051, - "step": 4780 - }, - { - "epoch": 1.7346938775510203, - "grad_norm": 0.212890625, - "learning_rate": 8.54240607024844e-05, - "loss": 1.0494, - "step": 4782 - }, - { - "epoch": 1.7354195011337867, - "grad_norm": 0.1748046875, - "learning_rate": 8.537510708603599e-05, - "loss": 1.0535, - "step": 4784 - }, - { - "epoch": 1.7361451247165531, - "grad_norm": 0.173828125, - "learning_rate": 8.532615346958757e-05, - "loss": 1.0028, - "step": 4786 - }, - { - "epoch": 1.7368707482993198, - "grad_norm": 0.189453125, - "learning_rate": 8.527719985313915e-05, - "loss": 1.0047, - "step": 4788 - }, - { - "epoch": 1.7375963718820862, - "grad_norm": 0.166015625, - "learning_rate": 8.522824623669074e-05, - "loss": 0.9975, - "step": 4790 - }, - { - "epoch": 1.7383219954648526, - "grad_norm": 0.1650390625, - "learning_rate": 8.517929262024232e-05, - "loss": 1.0143, - "step": 4792 - }, - { - "epoch": 1.7390476190476192, - "grad_norm": 0.1826171875, - "learning_rate": 8.513033900379392e-05, - "loss": 0.948, - "step": 4794 - }, - { - "epoch": 1.7397732426303856, - "grad_norm": 0.1923828125, - "learning_rate": 8.50813853873455e-05, - "loss": 0.9424, - "step": 4796 - }, - { - "epoch": 1.740498866213152, - "grad_norm": 0.1962890625, - "learning_rate": 8.503243177089708e-05, - "loss": 0.9775, - "step": 4798 - }, - { - "epoch": 1.7412244897959184, - "grad_norm": 0.19140625, - "learning_rate": 8.498347815444866e-05, - "loss": 0.9664, - "step": 4800 - }, - { - "epoch": 1.7412244897959184, - "eval_loss": 1.012211561203003, - "eval_runtime": 102.4092, - "eval_samples_per_second": 17.577, - "eval_steps_per_second": 0.557, - "step": 4800 - }, - { - "epoch": 1.7419501133786848, - "grad_norm": 0.18359375, - "learning_rate": 8.493452453800024e-05, - "loss": 0.9895, - "step": 4802 - }, - { - "epoch": 1.7426757369614512, - "grad_norm": 0.1484375, - "learning_rate": 8.488557092155184e-05, - "loss": 0.995, - "step": 4804 - }, - { - "epoch": 1.7434013605442176, - "grad_norm": 0.1669921875, - "learning_rate": 8.483661730510342e-05, - "loss": 0.9932, - "step": 4806 - }, - { - "epoch": 1.744126984126984, - "grad_norm": 0.1484375, - "learning_rate": 8.4787663688655e-05, - "loss": 1.005, - "step": 4808 - }, - { - "epoch": 1.7448526077097506, - "grad_norm": 0.1728515625, - "learning_rate": 8.473871007220659e-05, - "loss": 0.9672, - "step": 4810 - }, - { - "epoch": 1.745578231292517, - "grad_norm": 0.1435546875, - "learning_rate": 8.468975645575817e-05, - "loss": 0.9399, - "step": 4812 - }, - { - "epoch": 1.7463038548752836, - "grad_norm": 0.15625, - "learning_rate": 8.464080283930977e-05, - "loss": 0.9642, - "step": 4814 - }, - { - "epoch": 1.74702947845805, - "grad_norm": 0.1748046875, - "learning_rate": 8.459184922286135e-05, - "loss": 0.9698, - "step": 4816 - }, - { - "epoch": 1.7477551020408164, - "grad_norm": 0.1611328125, - "learning_rate": 8.454289560641293e-05, - "loss": 0.981, - "step": 4818 - }, - { - "epoch": 1.7484807256235828, - "grad_norm": 0.169921875, - "learning_rate": 8.449394198996451e-05, - "loss": 0.9702, - "step": 4820 - }, - { - "epoch": 1.7492063492063492, - "grad_norm": 0.150390625, - "learning_rate": 8.44449883735161e-05, - "loss": 0.9315, - "step": 4822 - }, - { - "epoch": 1.7499319727891156, - "grad_norm": 0.1669921875, - "learning_rate": 8.439603475706769e-05, - "loss": 0.9465, - "step": 4824 - }, - { - "epoch": 1.750657596371882, - "grad_norm": 0.1572265625, - "learning_rate": 8.434708114061927e-05, - "loss": 1.013, - "step": 4826 - }, - { - "epoch": 1.7513832199546484, - "grad_norm": 0.1806640625, - "learning_rate": 8.429812752417085e-05, - "loss": 0.9432, - "step": 4828 - }, - { - "epoch": 1.7521088435374148, - "grad_norm": 0.181640625, - "learning_rate": 8.424917390772244e-05, - "loss": 1.0176, - "step": 4830 - }, - { - "epoch": 1.7528344671201814, - "grad_norm": 0.171875, - "learning_rate": 8.420022029127402e-05, - "loss": 1.0079, - "step": 4832 - }, - { - "epoch": 1.7535600907029478, - "grad_norm": 0.1474609375, - "learning_rate": 8.41512666748256e-05, - "loss": 0.9487, - "step": 4834 - }, - { - "epoch": 1.7542857142857144, - "grad_norm": 0.1611328125, - "learning_rate": 8.41023130583772e-05, - "loss": 0.945, - "step": 4836 - }, - { - "epoch": 1.7550113378684808, - "grad_norm": 0.166015625, - "learning_rate": 8.405335944192878e-05, - "loss": 0.9486, - "step": 4838 - }, - { - "epoch": 1.7557369614512472, - "grad_norm": 0.1796875, - "learning_rate": 8.400440582548036e-05, - "loss": 0.9777, - "step": 4840 - }, - { - "epoch": 1.7564625850340136, - "grad_norm": 0.16015625, - "learning_rate": 8.395545220903194e-05, - "loss": 0.9026, - "step": 4842 - }, - { - "epoch": 1.75718820861678, - "grad_norm": 0.1669921875, - "learning_rate": 8.390649859258353e-05, - "loss": 1.082, - "step": 4844 - }, - { - "epoch": 1.7579138321995464, - "grad_norm": 0.1669921875, - "learning_rate": 8.385754497613512e-05, - "loss": 0.9822, - "step": 4846 - }, - { - "epoch": 1.7586394557823128, - "grad_norm": 0.162109375, - "learning_rate": 8.38085913596867e-05, - "loss": 0.9299, - "step": 4848 - }, - { - "epoch": 1.7593650793650792, - "grad_norm": 0.1640625, - "learning_rate": 8.375963774323829e-05, - "loss": 0.9228, - "step": 4850 - }, - { - "epoch": 1.7600907029478456, - "grad_norm": 0.1767578125, - "learning_rate": 8.371068412678987e-05, - "loss": 0.985, - "step": 4852 - }, - { - "epoch": 1.7608163265306123, - "grad_norm": 0.181640625, - "learning_rate": 8.366173051034145e-05, - "loss": 0.9734, - "step": 4854 - }, - { - "epoch": 1.7615419501133787, - "grad_norm": 0.1640625, - "learning_rate": 8.361277689389305e-05, - "loss": 0.9293, - "step": 4856 - }, - { - "epoch": 1.7622675736961453, - "grad_norm": 0.158203125, - "learning_rate": 8.356382327744463e-05, - "loss": 0.99, - "step": 4858 - }, - { - "epoch": 1.7629931972789117, - "grad_norm": 0.1640625, - "learning_rate": 8.351486966099621e-05, - "loss": 0.9639, - "step": 4860 - }, - { - "epoch": 1.763718820861678, - "grad_norm": 0.177734375, - "learning_rate": 8.34659160445478e-05, - "loss": 1.0291, - "step": 4862 - }, - { - "epoch": 1.7644444444444445, - "grad_norm": 0.1611328125, - "learning_rate": 8.341696242809938e-05, - "loss": 0.9175, - "step": 4864 - }, - { - "epoch": 1.7651700680272109, - "grad_norm": 0.17578125, - "learning_rate": 8.336800881165097e-05, - "loss": 1.0067, - "step": 4866 - }, - { - "epoch": 1.7658956916099773, - "grad_norm": 0.1669921875, - "learning_rate": 8.331905519520254e-05, - "loss": 0.9996, - "step": 4868 - }, - { - "epoch": 1.7666213151927437, - "grad_norm": 0.16796875, - "learning_rate": 8.327010157875414e-05, - "loss": 0.9372, - "step": 4870 - }, - { - "epoch": 1.76734693877551, - "grad_norm": 0.1572265625, - "learning_rate": 8.322114796230572e-05, - "loss": 0.9316, - "step": 4872 - }, - { - "epoch": 1.7680725623582765, - "grad_norm": 0.15234375, - "learning_rate": 8.31721943458573e-05, - "loss": 1.0023, - "step": 4874 - }, - { - "epoch": 1.768798185941043, - "grad_norm": 0.177734375, - "learning_rate": 8.31232407294089e-05, - "loss": 1.0662, - "step": 4876 - }, - { - "epoch": 1.7695238095238095, - "grad_norm": 0.1796875, - "learning_rate": 8.307428711296047e-05, - "loss": 0.9817, - "step": 4878 - }, - { - "epoch": 1.7702494331065761, - "grad_norm": 0.162109375, - "learning_rate": 8.302533349651206e-05, - "loss": 0.9366, - "step": 4880 - }, - { - "epoch": 1.7709750566893425, - "grad_norm": 0.1826171875, - "learning_rate": 8.297637988006364e-05, - "loss": 0.956, - "step": 4882 - }, - { - "epoch": 1.771700680272109, - "grad_norm": 0.140625, - "learning_rate": 8.292742626361523e-05, - "loss": 0.9326, - "step": 4884 - }, - { - "epoch": 1.7724263038548753, - "grad_norm": 0.1630859375, - "learning_rate": 8.287847264716682e-05, - "loss": 0.9756, - "step": 4886 - }, - { - "epoch": 1.7731519274376417, - "grad_norm": 0.173828125, - "learning_rate": 8.282951903071839e-05, - "loss": 1.0435, - "step": 4888 - }, - { - "epoch": 1.7738775510204081, - "grad_norm": 0.1689453125, - "learning_rate": 8.278056541426999e-05, - "loss": 1.0294, - "step": 4890 - }, - { - "epoch": 1.7746031746031745, - "grad_norm": 0.154296875, - "learning_rate": 8.273161179782157e-05, - "loss": 0.9823, - "step": 4892 - }, - { - "epoch": 1.775328798185941, - "grad_norm": 0.162109375, - "learning_rate": 8.268265818137315e-05, - "loss": 0.9292, - "step": 4894 - }, - { - "epoch": 1.7760544217687075, - "grad_norm": 0.1728515625, - "learning_rate": 8.263370456492475e-05, - "loss": 0.9869, - "step": 4896 - }, - { - "epoch": 1.776780045351474, - "grad_norm": 0.16015625, - "learning_rate": 8.258475094847632e-05, - "loss": 0.9646, - "step": 4898 - }, - { - "epoch": 1.7775056689342403, - "grad_norm": 0.1875, - "learning_rate": 8.253579733202791e-05, - "loss": 0.9385, - "step": 4900 - }, - { - "epoch": 1.7775056689342403, - "eval_loss": 1.0119349956512451, - "eval_runtime": 86.664, - "eval_samples_per_second": 20.77, - "eval_steps_per_second": 0.658, - "step": 4900 - }, - { - "epoch": 1.778231292517007, - "grad_norm": 0.1513671875, - "learning_rate": 8.248684371557948e-05, - "loss": 0.9462, - "step": 4902 - }, - { - "epoch": 1.7789569160997734, - "grad_norm": 0.1533203125, - "learning_rate": 8.243789009913108e-05, - "loss": 1.0036, - "step": 4904 - }, - { - "epoch": 1.7796825396825398, - "grad_norm": 0.166015625, - "learning_rate": 8.238893648268267e-05, - "loss": 0.9715, - "step": 4906 - }, - { - "epoch": 1.7804081632653062, - "grad_norm": 0.1591796875, - "learning_rate": 8.233998286623424e-05, - "loss": 1.0437, - "step": 4908 - }, - { - "epoch": 1.7811337868480726, - "grad_norm": 0.1728515625, - "learning_rate": 8.229102924978584e-05, - "loss": 0.9423, - "step": 4910 - }, - { - "epoch": 1.781859410430839, - "grad_norm": 0.1630859375, - "learning_rate": 8.22420756333374e-05, - "loss": 0.9602, - "step": 4912 - }, - { - "epoch": 1.7825850340136054, - "grad_norm": 0.16015625, - "learning_rate": 8.2193122016889e-05, - "loss": 0.9675, - "step": 4914 - }, - { - "epoch": 1.7833106575963718, - "grad_norm": 0.1748046875, - "learning_rate": 8.21441684004406e-05, - "loss": 0.9956, - "step": 4916 - }, - { - "epoch": 1.7840362811791384, - "grad_norm": 0.1396484375, - "learning_rate": 8.209521478399217e-05, - "loss": 0.9582, - "step": 4918 - }, - { - "epoch": 1.7847619047619048, - "grad_norm": 0.1708984375, - "learning_rate": 8.204626116754376e-05, - "loss": 0.9657, - "step": 4920 - }, - { - "epoch": 1.7854875283446712, - "grad_norm": 0.169921875, - "learning_rate": 8.199730755109533e-05, - "loss": 0.9727, - "step": 4922 - }, - { - "epoch": 1.7862131519274378, - "grad_norm": 0.185546875, - "learning_rate": 8.194835393464693e-05, - "loss": 0.9675, - "step": 4924 - }, - { - "epoch": 1.7869387755102042, - "grad_norm": 0.1806640625, - "learning_rate": 8.189940031819852e-05, - "loss": 0.9393, - "step": 4926 - }, - { - "epoch": 1.7876643990929706, - "grad_norm": 0.16015625, - "learning_rate": 8.185044670175009e-05, - "loss": 0.9847, - "step": 4928 - }, - { - "epoch": 1.788390022675737, - "grad_norm": 0.1728515625, - "learning_rate": 8.180149308530169e-05, - "loss": 0.9224, - "step": 4930 - }, - { - "epoch": 1.7891156462585034, - "grad_norm": 0.162109375, - "learning_rate": 8.175253946885327e-05, - "loss": 1.0141, - "step": 4932 - }, - { - "epoch": 1.7898412698412698, - "grad_norm": 0.173828125, - "learning_rate": 8.170358585240485e-05, - "loss": 1.0221, - "step": 4934 - }, - { - "epoch": 1.7905668934240362, - "grad_norm": 0.1630859375, - "learning_rate": 8.165463223595643e-05, - "loss": 0.9043, - "step": 4936 - }, - { - "epoch": 1.7912925170068026, - "grad_norm": 0.173828125, - "learning_rate": 8.160567861950802e-05, - "loss": 0.9836, - "step": 4938 - }, - { - "epoch": 1.7920181405895692, - "grad_norm": 0.1640625, - "learning_rate": 8.155672500305961e-05, - "loss": 0.9173, - "step": 4940 - }, - { - "epoch": 1.7927437641723356, - "grad_norm": 0.16796875, - "learning_rate": 8.15077713866112e-05, - "loss": 0.9854, - "step": 4942 - }, - { - "epoch": 1.793469387755102, - "grad_norm": 0.1748046875, - "learning_rate": 8.145881777016278e-05, - "loss": 0.9927, - "step": 4944 - }, - { - "epoch": 1.7941950113378686, - "grad_norm": 0.16796875, - "learning_rate": 8.140986415371436e-05, - "loss": 0.9674, - "step": 4946 - }, - { - "epoch": 1.794920634920635, - "grad_norm": 0.1572265625, - "learning_rate": 8.136091053726594e-05, - "loss": 0.9844, - "step": 4948 - }, - { - "epoch": 1.7956462585034014, - "grad_norm": 0.1630859375, - "learning_rate": 8.131195692081754e-05, - "loss": 0.8923, - "step": 4950 - }, - { - "epoch": 1.7963718820861678, - "grad_norm": 0.154296875, - "learning_rate": 8.126300330436912e-05, - "loss": 0.9434, - "step": 4952 - }, - { - "epoch": 1.7970975056689342, - "grad_norm": 0.185546875, - "learning_rate": 8.12140496879207e-05, - "loss": 1.0141, - "step": 4954 - }, - { - "epoch": 1.7978231292517006, - "grad_norm": 0.1689453125, - "learning_rate": 8.116509607147228e-05, - "loss": 0.9718, - "step": 4956 - }, - { - "epoch": 1.798548752834467, - "grad_norm": 0.1884765625, - "learning_rate": 8.111614245502387e-05, - "loss": 1.1004, - "step": 4958 - }, - { - "epoch": 1.7992743764172334, - "grad_norm": 0.169921875, - "learning_rate": 8.106718883857546e-05, - "loss": 0.9479, - "step": 4960 - }, - { - "epoch": 1.8, - "grad_norm": 0.154296875, - "learning_rate": 8.101823522212704e-05, - "loss": 0.9463, - "step": 4962 - }, - { - "epoch": 1.8007256235827664, - "grad_norm": 0.171875, - "learning_rate": 8.096928160567863e-05, - "loss": 0.9659, - "step": 4964 - }, - { - "epoch": 1.8014512471655328, - "grad_norm": 0.1611328125, - "learning_rate": 8.092032798923021e-05, - "loss": 1.0237, - "step": 4966 - }, - { - "epoch": 1.8021768707482995, - "grad_norm": 0.17578125, - "learning_rate": 8.087137437278179e-05, - "loss": 1.0192, - "step": 4968 - }, - { - "epoch": 1.8029024943310659, - "grad_norm": 0.1640625, - "learning_rate": 8.082242075633339e-05, - "loss": 0.9587, - "step": 4970 - }, - { - "epoch": 1.8036281179138323, - "grad_norm": 0.16796875, - "learning_rate": 8.077346713988497e-05, - "loss": 1.0048, - "step": 4972 - }, - { - "epoch": 1.8043537414965987, - "grad_norm": 0.18359375, - "learning_rate": 8.072451352343655e-05, - "loss": 0.9607, - "step": 4974 - }, - { - "epoch": 1.805079365079365, - "grad_norm": 0.1640625, - "learning_rate": 8.067555990698813e-05, - "loss": 1.0168, - "step": 4976 - }, - { - "epoch": 1.8058049886621315, - "grad_norm": 0.1728515625, - "learning_rate": 8.062660629053972e-05, - "loss": 0.9538, - "step": 4978 - }, - { - "epoch": 1.8065306122448979, - "grad_norm": 0.1796875, - "learning_rate": 8.05776526740913e-05, - "loss": 0.9468, - "step": 4980 - }, - { - "epoch": 1.8072562358276643, - "grad_norm": 0.1728515625, - "learning_rate": 8.05286990576429e-05, - "loss": 0.995, - "step": 4982 - }, - { - "epoch": 1.8079818594104309, - "grad_norm": 0.1787109375, - "learning_rate": 8.047974544119448e-05, - "loss": 0.954, - "step": 4984 - }, - { - "epoch": 1.8087074829931973, - "grad_norm": 0.1484375, - "learning_rate": 8.043079182474606e-05, - "loss": 0.957, - "step": 4986 - }, - { - "epoch": 1.8094331065759637, - "grad_norm": 0.1806640625, - "learning_rate": 8.038183820829764e-05, - "loss": 0.9804, - "step": 4988 - }, - { - "epoch": 1.8101587301587303, - "grad_norm": 0.1904296875, - "learning_rate": 8.033288459184922e-05, - "loss": 0.9968, - "step": 4990 - }, - { - "epoch": 1.8108843537414967, - "grad_norm": 0.1640625, - "learning_rate": 8.028393097540082e-05, - "loss": 0.9741, - "step": 4992 - }, - { - "epoch": 1.811609977324263, - "grad_norm": 0.1982421875, - "learning_rate": 8.02349773589524e-05, - "loss": 0.9552, - "step": 4994 - }, - { - "epoch": 1.8123356009070295, - "grad_norm": 0.16015625, - "learning_rate": 8.018602374250398e-05, - "loss": 1.0086, - "step": 4996 - }, - { - "epoch": 1.813061224489796, - "grad_norm": 0.1572265625, - "learning_rate": 8.013707012605557e-05, - "loss": 0.9693, - "step": 4998 - }, - { - "epoch": 1.8137868480725623, - "grad_norm": 0.1767578125, - "learning_rate": 8.008811650960715e-05, - "loss": 1.0028, - "step": 5000 - }, - { - "epoch": 1.8137868480725623, - "eval_loss": 1.0117406845092773, - "eval_runtime": 82.6702, - "eval_samples_per_second": 21.773, - "eval_steps_per_second": 0.689, - "step": 5000 - }, - { - "epoch": 1.8145124716553287, - "grad_norm": 0.15234375, - "learning_rate": 8.003916289315874e-05, - "loss": 0.968, - "step": 5002 - }, - { - "epoch": 1.815238095238095, - "grad_norm": 0.1474609375, - "learning_rate": 7.999020927671033e-05, - "loss": 1.0029, - "step": 5004 - }, - { - "epoch": 1.8159637188208617, - "grad_norm": 0.1826171875, - "learning_rate": 7.994125566026191e-05, - "loss": 0.9718, - "step": 5006 - }, - { - "epoch": 1.8166893424036281, - "grad_norm": 0.1552734375, - "learning_rate": 7.989230204381349e-05, - "loss": 0.9443, - "step": 5008 - }, - { - "epoch": 1.8174149659863945, - "grad_norm": 0.1962890625, - "learning_rate": 7.984334842736507e-05, - "loss": 1.054, - "step": 5010 - }, - { - "epoch": 1.8181405895691611, - "grad_norm": 0.171875, - "learning_rate": 7.979439481091667e-05, - "loss": 1.0275, - "step": 5012 - }, - { - "epoch": 1.8188662131519275, - "grad_norm": 0.181640625, - "learning_rate": 7.974544119446824e-05, - "loss": 1.0257, - "step": 5014 - }, - { - "epoch": 1.819591836734694, - "grad_norm": 0.16015625, - "learning_rate": 7.969648757801983e-05, - "loss": 0.9781, - "step": 5016 - }, - { - "epoch": 1.8203174603174603, - "grad_norm": 0.1787109375, - "learning_rate": 7.964753396157142e-05, - "loss": 0.9828, - "step": 5018 - }, - { - "epoch": 1.8210430839002267, - "grad_norm": 0.1845703125, - "learning_rate": 7.9598580345123e-05, - "loss": 0.9198, - "step": 5020 - }, - { - "epoch": 1.8217687074829931, - "grad_norm": 0.173828125, - "learning_rate": 7.95496267286746e-05, - "loss": 0.9939, - "step": 5022 - }, - { - "epoch": 1.8224943310657595, - "grad_norm": 0.1650390625, - "learning_rate": 7.950067311222616e-05, - "loss": 0.9844, - "step": 5024 - }, - { - "epoch": 1.823219954648526, - "grad_norm": 0.1875, - "learning_rate": 7.945171949577776e-05, - "loss": 0.9858, - "step": 5026 - }, - { - "epoch": 1.8239455782312926, - "grad_norm": 0.1826171875, - "learning_rate": 7.940276587932934e-05, - "loss": 0.9681, - "step": 5028 - }, - { - "epoch": 1.824671201814059, - "grad_norm": 0.1650390625, - "learning_rate": 7.935381226288092e-05, - "loss": 1.0324, - "step": 5030 - }, - { - "epoch": 1.8253968253968254, - "grad_norm": 0.216796875, - "learning_rate": 7.930485864643252e-05, - "loss": 1.0411, - "step": 5032 - }, - { - "epoch": 1.826122448979592, - "grad_norm": 0.2080078125, - "learning_rate": 7.925590502998409e-05, - "loss": 0.9887, - "step": 5034 - }, - { - "epoch": 1.8268480725623584, - "grad_norm": 0.15625, - "learning_rate": 7.920695141353568e-05, - "loss": 0.985, - "step": 5036 - }, - { - "epoch": 1.8275736961451248, - "grad_norm": 0.1923828125, - "learning_rate": 7.915799779708727e-05, - "loss": 0.9819, - "step": 5038 - }, - { - "epoch": 1.8282993197278912, - "grad_norm": 0.1640625, - "learning_rate": 7.910904418063885e-05, - "loss": 0.9861, - "step": 5040 - }, - { - "epoch": 1.8290249433106576, - "grad_norm": 0.255859375, - "learning_rate": 7.906009056419044e-05, - "loss": 1.0396, - "step": 5042 - }, - { - "epoch": 1.829750566893424, - "grad_norm": 0.1826171875, - "learning_rate": 7.901113694774201e-05, - "loss": 1.0312, - "step": 5044 - }, - { - "epoch": 1.8304761904761904, - "grad_norm": 0.1630859375, - "learning_rate": 7.896218333129361e-05, - "loss": 1.0328, - "step": 5046 - }, - { - "epoch": 1.8312018140589568, - "grad_norm": 0.169921875, - "learning_rate": 7.891322971484518e-05, - "loss": 0.9466, - "step": 5048 - }, - { - "epoch": 1.8319274376417234, - "grad_norm": 0.1796875, - "learning_rate": 7.886427609839677e-05, - "loss": 0.9832, - "step": 5050 - }, - { - "epoch": 1.8326530612244898, - "grad_norm": 0.2060546875, - "learning_rate": 7.881532248194837e-05, - "loss": 0.9828, - "step": 5052 - }, - { - "epoch": 1.8333786848072562, - "grad_norm": 0.15234375, - "learning_rate": 7.876636886549994e-05, - "loss": 0.9619, - "step": 5054 - }, - { - "epoch": 1.8341043083900228, - "grad_norm": 0.15625, - "learning_rate": 7.871741524905153e-05, - "loss": 0.8759, - "step": 5056 - }, - { - "epoch": 1.8348299319727892, - "grad_norm": 0.1572265625, - "learning_rate": 7.86684616326031e-05, - "loss": 1.0589, - "step": 5058 - }, - { - "epoch": 1.8355555555555556, - "grad_norm": 0.1767578125, - "learning_rate": 7.86195080161547e-05, - "loss": 1.0069, - "step": 5060 - }, - { - "epoch": 1.836281179138322, - "grad_norm": 0.17578125, - "learning_rate": 7.857055439970629e-05, - "loss": 0.9726, - "step": 5062 - }, - { - "epoch": 1.8370068027210884, - "grad_norm": 0.173828125, - "learning_rate": 7.852160078325786e-05, - "loss": 0.9907, - "step": 5064 - }, - { - "epoch": 1.8377324263038548, - "grad_norm": 0.1611328125, - "learning_rate": 7.847264716680946e-05, - "loss": 0.9354, - "step": 5066 - }, - { - "epoch": 1.8384580498866212, - "grad_norm": 0.1591796875, - "learning_rate": 7.842369355036103e-05, - "loss": 0.9583, - "step": 5068 - }, - { - "epoch": 1.8391836734693876, - "grad_norm": 0.15234375, - "learning_rate": 7.837473993391262e-05, - "loss": 0.9899, - "step": 5070 - }, - { - "epoch": 1.8399092970521542, - "grad_norm": 0.1630859375, - "learning_rate": 7.832578631746422e-05, - "loss": 1.0061, - "step": 5072 - }, - { - "epoch": 1.8406349206349206, - "grad_norm": 0.1396484375, - "learning_rate": 7.827683270101579e-05, - "loss": 0.8678, - "step": 5074 - }, - { - "epoch": 1.841360544217687, - "grad_norm": 0.1806640625, - "learning_rate": 7.822787908456738e-05, - "loss": 0.9876, - "step": 5076 - }, - { - "epoch": 1.8420861678004536, - "grad_norm": 0.158203125, - "learning_rate": 7.817892546811895e-05, - "loss": 0.9555, - "step": 5078 - }, - { - "epoch": 1.84281179138322, - "grad_norm": 0.169921875, - "learning_rate": 7.812997185167055e-05, - "loss": 1.0482, - "step": 5080 - }, - { - "epoch": 1.8435374149659864, - "grad_norm": 0.1708984375, - "learning_rate": 7.808101823522213e-05, - "loss": 0.9854, - "step": 5082 - }, - { - "epoch": 1.8442630385487528, - "grad_norm": 0.173828125, - "learning_rate": 7.803206461877371e-05, - "loss": 0.9896, - "step": 5084 - }, - { - "epoch": 1.8449886621315192, - "grad_norm": 0.1650390625, - "learning_rate": 7.798311100232531e-05, - "loss": 0.9083, - "step": 5086 - }, - { - "epoch": 1.8457142857142856, - "grad_norm": 0.1875, - "learning_rate": 7.793415738587688e-05, - "loss": 0.9821, - "step": 5088 - }, - { - "epoch": 1.846439909297052, - "grad_norm": 0.1708984375, - "learning_rate": 7.788520376942847e-05, - "loss": 1.0252, - "step": 5090 - }, - { - "epoch": 1.8471655328798184, - "grad_norm": 0.1640625, - "learning_rate": 7.783625015298005e-05, - "loss": 1.0168, - "step": 5092 - }, - { - "epoch": 1.847891156462585, - "grad_norm": 0.18359375, - "learning_rate": 7.778729653653164e-05, - "loss": 1.0525, - "step": 5094 - }, - { - "epoch": 1.8486167800453515, - "grad_norm": 0.16796875, - "learning_rate": 7.773834292008323e-05, - "loss": 0.946, - "step": 5096 - }, - { - "epoch": 1.8493424036281179, - "grad_norm": 0.2001953125, - "learning_rate": 7.76893893036348e-05, - "loss": 0.9013, - "step": 5098 - }, - { - "epoch": 1.8500680272108845, - "grad_norm": 0.1875, - "learning_rate": 7.76404356871864e-05, - "loss": 0.9918, - "step": 5100 - }, - { - "epoch": 1.8500680272108845, - "eval_loss": 1.0116716623306274, - "eval_runtime": 85.8661, - "eval_samples_per_second": 20.963, - "eval_steps_per_second": 0.664, - "step": 5100 - }, - { - "epoch": 1.8507936507936509, - "grad_norm": 0.1513671875, - "learning_rate": 7.759148207073798e-05, - "loss": 0.988, - "step": 5102 - }, - { - "epoch": 1.8515192743764173, - "grad_norm": 0.1904296875, - "learning_rate": 7.754252845428956e-05, - "loss": 0.9003, - "step": 5104 - }, - { - "epoch": 1.8522448979591837, - "grad_norm": 0.1494140625, - "learning_rate": 7.749357483784116e-05, - "loss": 0.9411, - "step": 5106 - }, - { - "epoch": 1.85297052154195, - "grad_norm": 0.16015625, - "learning_rate": 7.744462122139273e-05, - "loss": 0.9319, - "step": 5108 - }, - { - "epoch": 1.8536961451247165, - "grad_norm": 0.189453125, - "learning_rate": 7.739566760494432e-05, - "loss": 1.046, - "step": 5110 - }, - { - "epoch": 1.8544217687074829, - "grad_norm": 0.14453125, - "learning_rate": 7.73467139884959e-05, - "loss": 0.8999, - "step": 5112 - }, - { - "epoch": 1.8551473922902493, - "grad_norm": 0.169921875, - "learning_rate": 7.729776037204749e-05, - "loss": 1.0543, - "step": 5114 - }, - { - "epoch": 1.855873015873016, - "grad_norm": 0.1689453125, - "learning_rate": 7.724880675559908e-05, - "loss": 1.0118, - "step": 5116 - }, - { - "epoch": 1.8565986394557823, - "grad_norm": 0.1650390625, - "learning_rate": 7.719985313915065e-05, - "loss": 1.0171, - "step": 5118 - }, - { - "epoch": 1.857324263038549, - "grad_norm": 0.158203125, - "learning_rate": 7.715089952270225e-05, - "loss": 0.9521, - "step": 5120 - }, - { - "epoch": 1.8580498866213153, - "grad_norm": 0.154296875, - "learning_rate": 7.710194590625383e-05, - "loss": 0.9663, - "step": 5122 - }, - { - "epoch": 1.8587755102040817, - "grad_norm": 0.1826171875, - "learning_rate": 7.705299228980541e-05, - "loss": 0.9721, - "step": 5124 - }, - { - "epoch": 1.8595011337868481, - "grad_norm": 0.17578125, - "learning_rate": 7.7004038673357e-05, - "loss": 1.0055, - "step": 5126 - }, - { - "epoch": 1.8602267573696145, - "grad_norm": 0.1650390625, - "learning_rate": 7.695508505690858e-05, - "loss": 0.9677, - "step": 5128 - }, - { - "epoch": 1.860952380952381, - "grad_norm": 0.16796875, - "learning_rate": 7.690613144046017e-05, - "loss": 0.9718, - "step": 5130 - }, - { - "epoch": 1.8616780045351473, - "grad_norm": 0.1865234375, - "learning_rate": 7.685717782401175e-05, - "loss": 0.9995, - "step": 5132 - }, - { - "epoch": 1.8624036281179137, - "grad_norm": 0.154296875, - "learning_rate": 7.680822420756334e-05, - "loss": 0.9169, - "step": 5134 - }, - { - "epoch": 1.8631292517006801, - "grad_norm": 0.166015625, - "learning_rate": 7.675927059111492e-05, - "loss": 0.9085, - "step": 5136 - }, - { - "epoch": 1.8638548752834467, - "grad_norm": 0.1796875, - "learning_rate": 7.67103169746665e-05, - "loss": 0.9904, - "step": 5138 - }, - { - "epoch": 1.8645804988662131, - "grad_norm": 0.1708984375, - "learning_rate": 7.66613633582181e-05, - "loss": 1.0268, - "step": 5140 - }, - { - "epoch": 1.8653061224489798, - "grad_norm": 0.1533203125, - "learning_rate": 7.661240974176968e-05, - "loss": 1.0152, - "step": 5142 - }, - { - "epoch": 1.8660317460317462, - "grad_norm": 0.1708984375, - "learning_rate": 7.656345612532126e-05, - "loss": 1.0148, - "step": 5144 - }, - { - "epoch": 1.8667573696145126, - "grad_norm": 0.16015625, - "learning_rate": 7.651450250887284e-05, - "loss": 0.9883, - "step": 5146 - }, - { - "epoch": 1.867482993197279, - "grad_norm": 0.1669921875, - "learning_rate": 7.646554889242443e-05, - "loss": 1.0816, - "step": 5148 - }, - { - "epoch": 1.8682086167800454, - "grad_norm": 0.17578125, - "learning_rate": 7.641659527597602e-05, - "loss": 0.9106, - "step": 5150 - }, - { - "epoch": 1.8689342403628117, - "grad_norm": 0.177734375, - "learning_rate": 7.63676416595276e-05, - "loss": 0.9509, - "step": 5152 - }, - { - "epoch": 1.8696598639455781, - "grad_norm": 0.1689453125, - "learning_rate": 7.631868804307919e-05, - "loss": 0.8867, - "step": 5154 - }, - { - "epoch": 1.8703854875283445, - "grad_norm": 0.1953125, - "learning_rate": 7.626973442663077e-05, - "loss": 1.0658, - "step": 5156 - }, - { - "epoch": 1.871111111111111, - "grad_norm": 0.1396484375, - "learning_rate": 7.622078081018235e-05, - "loss": 0.9254, - "step": 5158 - }, - { - "epoch": 1.8718367346938776, - "grad_norm": 0.15234375, - "learning_rate": 7.617182719373393e-05, - "loss": 0.9928, - "step": 5160 - }, - { - "epoch": 1.872562358276644, - "grad_norm": 0.16796875, - "learning_rate": 7.612287357728553e-05, - "loss": 0.9523, - "step": 5162 - }, - { - "epoch": 1.8732879818594106, - "grad_norm": 0.1826171875, - "learning_rate": 7.607391996083711e-05, - "loss": 1.0135, - "step": 5164 - }, - { - "epoch": 1.874013605442177, - "grad_norm": 0.1708984375, - "learning_rate": 7.60249663443887e-05, - "loss": 1.0446, - "step": 5166 - }, - { - "epoch": 1.8747392290249434, - "grad_norm": 0.1787109375, - "learning_rate": 7.597601272794028e-05, - "loss": 0.9577, - "step": 5168 - }, - { - "epoch": 1.8754648526077098, - "grad_norm": 0.16796875, - "learning_rate": 7.592705911149186e-05, - "loss": 1.0376, - "step": 5170 - }, - { - "epoch": 1.8761904761904762, - "grad_norm": 0.15625, - "learning_rate": 7.587810549504345e-05, - "loss": 0.9016, - "step": 5172 - }, - { - "epoch": 1.8769160997732426, - "grad_norm": 0.1689453125, - "learning_rate": 7.582915187859504e-05, - "loss": 0.9314, - "step": 5174 - }, - { - "epoch": 1.877641723356009, - "grad_norm": 0.1728515625, - "learning_rate": 7.578019826214662e-05, - "loss": 0.9535, - "step": 5176 - }, - { - "epoch": 1.8783673469387754, - "grad_norm": 0.166015625, - "learning_rate": 7.57312446456982e-05, - "loss": 1.0041, - "step": 5178 - }, - { - "epoch": 1.879092970521542, - "grad_norm": 0.1611328125, - "learning_rate": 7.568229102924978e-05, - "loss": 0.9118, - "step": 5180 - }, - { - "epoch": 1.8798185941043084, - "grad_norm": 0.1650390625, - "learning_rate": 7.563333741280138e-05, - "loss": 0.9734, - "step": 5182 - }, - { - "epoch": 1.8805442176870748, - "grad_norm": 0.162109375, - "learning_rate": 7.558438379635296e-05, - "loss": 0.941, - "step": 5184 - }, - { - "epoch": 1.8812698412698414, - "grad_norm": 0.173828125, - "learning_rate": 7.553543017990454e-05, - "loss": 1.0394, - "step": 5186 - }, - { - "epoch": 1.8819954648526078, - "grad_norm": 0.154296875, - "learning_rate": 7.548647656345613e-05, - "loss": 0.9472, - "step": 5188 - }, - { - "epoch": 1.8827210884353742, - "grad_norm": 0.1787109375, - "learning_rate": 7.543752294700771e-05, - "loss": 0.9045, - "step": 5190 - }, - { - "epoch": 1.8834467120181406, - "grad_norm": 0.18359375, - "learning_rate": 7.53885693305593e-05, - "loss": 1.0148, - "step": 5192 - }, - { - "epoch": 1.884172335600907, - "grad_norm": 0.1572265625, - "learning_rate": 7.533961571411087e-05, - "loss": 0.9824, - "step": 5194 - }, - { - "epoch": 1.8848979591836734, - "grad_norm": 0.1689453125, - "learning_rate": 7.529066209766247e-05, - "loss": 1.0019, - "step": 5196 - }, - { - "epoch": 1.8856235827664398, - "grad_norm": 0.216796875, - "learning_rate": 7.524170848121405e-05, - "loss": 1.0268, - "step": 5198 - }, - { - "epoch": 1.8863492063492062, - "grad_norm": 0.166015625, - "learning_rate": 7.519275486476563e-05, - "loss": 1.0249, - "step": 5200 - }, - { - "epoch": 1.8863492063492062, - "eval_loss": 1.011323094367981, - "eval_runtime": 85.9575, - "eval_samples_per_second": 20.941, - "eval_steps_per_second": 0.663, - "step": 5200 - }, - { - "epoch": 1.8870748299319728, - "grad_norm": 0.181640625, - "learning_rate": 7.514380124831723e-05, - "loss": 0.9658, - "step": 5202 - }, - { - "epoch": 1.8878004535147392, - "grad_norm": 0.16796875, - "learning_rate": 7.50948476318688e-05, - "loss": 1.0172, - "step": 5204 - }, - { - "epoch": 1.8885260770975056, - "grad_norm": 0.1875, - "learning_rate": 7.50458940154204e-05, - "loss": 0.9279, - "step": 5206 - }, - { - "epoch": 1.8892517006802723, - "grad_norm": 0.162109375, - "learning_rate": 7.499694039897198e-05, - "loss": 0.9781, - "step": 5208 - }, - { - "epoch": 1.8899773242630387, - "grad_norm": 0.1845703125, - "learning_rate": 7.494798678252356e-05, - "loss": 1.0788, - "step": 5210 - }, - { - "epoch": 1.890702947845805, - "grad_norm": 0.13671875, - "learning_rate": 7.489903316607515e-05, - "loss": 0.9595, - "step": 5212 - }, - { - "epoch": 1.8914285714285715, - "grad_norm": 0.158203125, - "learning_rate": 7.485007954962672e-05, - "loss": 0.9927, - "step": 5214 - }, - { - "epoch": 1.8921541950113379, - "grad_norm": 0.1748046875, - "learning_rate": 7.480112593317832e-05, - "loss": 1.0254, - "step": 5216 - }, - { - "epoch": 1.8928798185941043, - "grad_norm": 0.1923828125, - "learning_rate": 7.47521723167299e-05, - "loss": 1.0419, - "step": 5218 - }, - { - "epoch": 1.8936054421768707, - "grad_norm": 0.205078125, - "learning_rate": 7.470321870028148e-05, - "loss": 0.9921, - "step": 5220 - }, - { - "epoch": 1.894331065759637, - "grad_norm": 0.17578125, - "learning_rate": 7.465426508383308e-05, - "loss": 0.9195, - "step": 5222 - }, - { - "epoch": 1.8950566893424037, - "grad_norm": 0.1748046875, - "learning_rate": 7.460531146738465e-05, - "loss": 1.0073, - "step": 5224 - }, - { - "epoch": 1.89578231292517, - "grad_norm": 0.19921875, - "learning_rate": 7.455635785093624e-05, - "loss": 1.0088, - "step": 5226 - }, - { - "epoch": 1.8965079365079365, - "grad_norm": 0.1689453125, - "learning_rate": 7.450740423448783e-05, - "loss": 0.938, - "step": 5228 - }, - { - "epoch": 1.897233560090703, - "grad_norm": 0.146484375, - "learning_rate": 7.445845061803941e-05, - "loss": 0.9745, - "step": 5230 - }, - { - "epoch": 1.8979591836734695, - "grad_norm": 0.16796875, - "learning_rate": 7.4409497001591e-05, - "loss": 0.9514, - "step": 5232 - }, - { - "epoch": 1.898684807256236, - "grad_norm": 0.173828125, - "learning_rate": 7.436054338514257e-05, - "loss": 0.9912, - "step": 5234 - }, - { - "epoch": 1.8994104308390023, - "grad_norm": 0.1640625, - "learning_rate": 7.431158976869417e-05, - "loss": 0.9794, - "step": 5236 - }, - { - "epoch": 1.9001360544217687, - "grad_norm": 0.2099609375, - "learning_rate": 7.426263615224575e-05, - "loss": 1.1444, - "step": 5238 - }, - { - "epoch": 1.900861678004535, - "grad_norm": 0.1591796875, - "learning_rate": 7.421368253579733e-05, - "loss": 0.9846, - "step": 5240 - }, - { - "epoch": 1.9015873015873015, - "grad_norm": 0.18359375, - "learning_rate": 7.416472891934893e-05, - "loss": 0.9898, - "step": 5242 - }, - { - "epoch": 1.902312925170068, - "grad_norm": 0.1455078125, - "learning_rate": 7.41157753029005e-05, - "loss": 0.9763, - "step": 5244 - }, - { - "epoch": 1.9030385487528345, - "grad_norm": 0.171875, - "learning_rate": 7.40668216864521e-05, - "loss": 0.9776, - "step": 5246 - }, - { - "epoch": 1.903764172335601, - "grad_norm": 0.17578125, - "learning_rate": 7.401786807000368e-05, - "loss": 0.9866, - "step": 5248 - }, - { - "epoch": 1.9044897959183673, - "grad_norm": 0.166015625, - "learning_rate": 7.396891445355526e-05, - "loss": 0.9892, - "step": 5250 - }, - { - "epoch": 1.905215419501134, - "grad_norm": 0.15625, - "learning_rate": 7.391996083710685e-05, - "loss": 0.9734, - "step": 5252 - }, - { - "epoch": 1.9059410430839003, - "grad_norm": 0.1953125, - "learning_rate": 7.387100722065842e-05, - "loss": 1.0002, - "step": 5254 - }, - { - "epoch": 1.9066666666666667, - "grad_norm": 0.1533203125, - "learning_rate": 7.382205360421002e-05, - "loss": 0.9108, - "step": 5256 - }, - { - "epoch": 1.9073922902494331, - "grad_norm": 0.1806640625, - "learning_rate": 7.37730999877616e-05, - "loss": 0.9652, - "step": 5258 - }, - { - "epoch": 1.9081179138321995, - "grad_norm": 0.2041015625, - "learning_rate": 7.372414637131318e-05, - "loss": 0.9549, - "step": 5260 - }, - { - "epoch": 1.908843537414966, - "grad_norm": 0.16796875, - "learning_rate": 7.367519275486478e-05, - "loss": 1.0341, - "step": 5262 - }, - { - "epoch": 1.9095691609977323, - "grad_norm": 0.1767578125, - "learning_rate": 7.362623913841635e-05, - "loss": 0.9873, - "step": 5264 - }, - { - "epoch": 1.9102947845804987, - "grad_norm": 0.173828125, - "learning_rate": 7.357728552196794e-05, - "loss": 0.9916, - "step": 5266 - }, - { - "epoch": 1.9110204081632653, - "grad_norm": 0.162109375, - "learning_rate": 7.352833190551953e-05, - "loss": 0.96, - "step": 5268 - }, - { - "epoch": 1.9117460317460317, - "grad_norm": 0.1708984375, - "learning_rate": 7.347937828907111e-05, - "loss": 0.9318, - "step": 5270 - }, - { - "epoch": 1.9124716553287981, - "grad_norm": 0.1611328125, - "learning_rate": 7.343042467262269e-05, - "loss": 0.9651, - "step": 5272 - }, - { - "epoch": 1.9131972789115648, - "grad_norm": 0.1669921875, - "learning_rate": 7.338147105617427e-05, - "loss": 0.9995, - "step": 5274 - }, - { - "epoch": 1.9139229024943312, - "grad_norm": 0.1494140625, - "learning_rate": 7.333251743972587e-05, - "loss": 1.0009, - "step": 5276 - }, - { - "epoch": 1.9146485260770976, - "grad_norm": 0.171875, - "learning_rate": 7.328356382327745e-05, - "loss": 1.0127, - "step": 5278 - }, - { - "epoch": 1.915374149659864, - "grad_norm": 0.1416015625, - "learning_rate": 7.323461020682903e-05, - "loss": 0.9979, - "step": 5280 - }, - { - "epoch": 1.9160997732426304, - "grad_norm": 0.2001953125, - "learning_rate": 7.318565659038062e-05, - "loss": 1.057, - "step": 5282 - }, - { - "epoch": 1.9168253968253968, - "grad_norm": 0.16015625, - "learning_rate": 7.31367029739322e-05, - "loss": 1.0084, - "step": 5284 - }, - { - "epoch": 1.9175510204081632, - "grad_norm": 0.1787109375, - "learning_rate": 7.308774935748379e-05, - "loss": 1.0373, - "step": 5286 - }, - { - "epoch": 1.9182766439909296, - "grad_norm": 0.1923828125, - "learning_rate": 7.303879574103538e-05, - "loss": 1.0019, - "step": 5288 - }, - { - "epoch": 1.9190022675736962, - "grad_norm": 0.18359375, - "learning_rate": 7.298984212458696e-05, - "loss": 0.9535, - "step": 5290 - }, - { - "epoch": 1.9197278911564626, - "grad_norm": 0.169921875, - "learning_rate": 7.294088850813854e-05, - "loss": 0.9318, - "step": 5292 - }, - { - "epoch": 1.920453514739229, - "grad_norm": 0.1484375, - "learning_rate": 7.289193489169012e-05, - "loss": 0.9384, - "step": 5294 - }, - { - "epoch": 1.9211791383219956, - "grad_norm": 0.177734375, - "learning_rate": 7.284298127524172e-05, - "loss": 0.9661, - "step": 5296 - }, - { - "epoch": 1.921904761904762, - "grad_norm": 0.166015625, - "learning_rate": 7.27940276587933e-05, - "loss": 0.9253, - "step": 5298 - }, - { - "epoch": 1.9226303854875284, - "grad_norm": 0.1513671875, - "learning_rate": 7.274507404234488e-05, - "loss": 0.9973, - "step": 5300 - }, - { - "epoch": 1.9226303854875284, - "eval_loss": 1.0112332105636597, - "eval_runtime": 86.071, - "eval_samples_per_second": 20.913, - "eval_steps_per_second": 0.662, - "step": 5300 - }, - { - "epoch": 1.9233560090702948, - "grad_norm": 0.1796875, - "learning_rate": 7.269612042589646e-05, - "loss": 1.0532, - "step": 5302 - }, - { - "epoch": 1.9240816326530612, - "grad_norm": 0.1669921875, - "learning_rate": 7.264716680944805e-05, - "loss": 0.9237, - "step": 5304 - }, - { - "epoch": 1.9248072562358276, - "grad_norm": 0.15625, - "learning_rate": 7.259821319299963e-05, - "loss": 0.9588, - "step": 5306 - }, - { - "epoch": 1.925532879818594, - "grad_norm": 0.16796875, - "learning_rate": 7.254925957655123e-05, - "loss": 1.0311, - "step": 5308 - }, - { - "epoch": 1.9262585034013604, - "grad_norm": 0.15625, - "learning_rate": 7.250030596010281e-05, - "loss": 0.901, - "step": 5310 - }, - { - "epoch": 1.926984126984127, - "grad_norm": 0.18359375, - "learning_rate": 7.245135234365439e-05, - "loss": 1.0205, - "step": 5312 - }, - { - "epoch": 1.9277097505668934, - "grad_norm": 0.1533203125, - "learning_rate": 7.240239872720597e-05, - "loss": 0.958, - "step": 5314 - }, - { - "epoch": 1.9284353741496598, - "grad_norm": 0.17578125, - "learning_rate": 7.235344511075755e-05, - "loss": 0.996, - "step": 5316 - }, - { - "epoch": 1.9291609977324264, - "grad_norm": 0.162109375, - "learning_rate": 7.230449149430915e-05, - "loss": 0.9822, - "step": 5318 - }, - { - "epoch": 1.9298866213151928, - "grad_norm": 0.1484375, - "learning_rate": 7.225553787786073e-05, - "loss": 0.9584, - "step": 5320 - }, - { - "epoch": 1.9306122448979592, - "grad_norm": 0.1611328125, - "learning_rate": 7.220658426141231e-05, - "loss": 0.9161, - "step": 5322 - }, - { - "epoch": 1.9313378684807256, - "grad_norm": 0.158203125, - "learning_rate": 7.21576306449639e-05, - "loss": 0.9766, - "step": 5324 - }, - { - "epoch": 1.932063492063492, - "grad_norm": 0.166015625, - "learning_rate": 7.210867702851548e-05, - "loss": 1.0074, - "step": 5326 - }, - { - "epoch": 1.9327891156462584, - "grad_norm": 0.166015625, - "learning_rate": 7.205972341206708e-05, - "loss": 0.9183, - "step": 5328 - }, - { - "epoch": 1.9335147392290248, - "grad_norm": 0.2021484375, - "learning_rate": 7.201076979561866e-05, - "loss": 1.01, - "step": 5330 - }, - { - "epoch": 1.9342403628117912, - "grad_norm": 0.18359375, - "learning_rate": 7.196181617917024e-05, - "loss": 0.969, - "step": 5332 - }, - { - "epoch": 1.9349659863945579, - "grad_norm": 0.177734375, - "learning_rate": 7.191286256272182e-05, - "loss": 1.0358, - "step": 5334 - }, - { - "epoch": 1.9356916099773243, - "grad_norm": 0.1865234375, - "learning_rate": 7.18639089462734e-05, - "loss": 1.0049, - "step": 5336 - }, - { - "epoch": 1.9364172335600907, - "grad_norm": 0.1611328125, - "learning_rate": 7.1814955329825e-05, - "loss": 0.9693, - "step": 5338 - }, - { - "epoch": 1.9371428571428573, - "grad_norm": 0.14453125, - "learning_rate": 7.176600171337657e-05, - "loss": 0.9089, - "step": 5340 - }, - { - "epoch": 1.9378684807256237, - "grad_norm": 0.18359375, - "learning_rate": 7.171704809692816e-05, - "loss": 0.9832, - "step": 5342 - }, - { - "epoch": 1.93859410430839, - "grad_norm": 0.162109375, - "learning_rate": 7.166809448047975e-05, - "loss": 0.9502, - "step": 5344 - }, - { - "epoch": 1.9393197278911565, - "grad_norm": 0.15625, - "learning_rate": 7.161914086403133e-05, - "loss": 0.9835, - "step": 5346 - }, - { - "epoch": 1.9400453514739229, - "grad_norm": 0.1689453125, - "learning_rate": 7.157018724758293e-05, - "loss": 0.9826, - "step": 5348 - }, - { - "epoch": 1.9407709750566893, - "grad_norm": 0.166015625, - "learning_rate": 7.15212336311345e-05, - "loss": 0.9625, - "step": 5350 - }, - { - "epoch": 1.9414965986394557, - "grad_norm": 0.1728515625, - "learning_rate": 7.147228001468609e-05, - "loss": 0.9491, - "step": 5352 - }, - { - "epoch": 1.942222222222222, - "grad_norm": 0.1845703125, - "learning_rate": 7.142332639823767e-05, - "loss": 1.0431, - "step": 5354 - }, - { - "epoch": 1.9429478458049887, - "grad_norm": 0.15625, - "learning_rate": 7.137437278178925e-05, - "loss": 1.0504, - "step": 5356 - }, - { - "epoch": 1.943673469387755, - "grad_norm": 0.1865234375, - "learning_rate": 7.132541916534085e-05, - "loss": 1.0378, - "step": 5358 - }, - { - "epoch": 1.9443990929705215, - "grad_norm": 0.1572265625, - "learning_rate": 7.127646554889242e-05, - "loss": 0.949, - "step": 5360 - }, - { - "epoch": 1.9451247165532881, - "grad_norm": 0.1640625, - "learning_rate": 7.122751193244401e-05, - "loss": 0.9973, - "step": 5362 - }, - { - "epoch": 1.9458503401360545, - "grad_norm": 0.1630859375, - "learning_rate": 7.11785583159956e-05, - "loss": 0.996, - "step": 5364 - }, - { - "epoch": 1.946575963718821, - "grad_norm": 0.171875, - "learning_rate": 7.112960469954718e-05, - "loss": 1.0218, - "step": 5366 - }, - { - "epoch": 1.9473015873015873, - "grad_norm": 0.1669921875, - "learning_rate": 7.108065108309878e-05, - "loss": 1.0081, - "step": 5368 - }, - { - "epoch": 1.9480272108843537, - "grad_norm": 0.1611328125, - "learning_rate": 7.103169746665034e-05, - "loss": 0.9438, - "step": 5370 - }, - { - "epoch": 1.94875283446712, - "grad_norm": 0.2109375, - "learning_rate": 7.098274385020194e-05, - "loss": 0.9731, - "step": 5372 - }, - { - "epoch": 1.9494784580498865, - "grad_norm": 0.1611328125, - "learning_rate": 7.093379023375352e-05, - "loss": 0.9133, - "step": 5374 - }, - { - "epoch": 1.950204081632653, - "grad_norm": 0.1767578125, - "learning_rate": 7.08848366173051e-05, - "loss": 0.946, - "step": 5376 - }, - { - "epoch": 1.9509297052154195, - "grad_norm": 0.19140625, - "learning_rate": 7.08358830008567e-05, - "loss": 0.9632, - "step": 5378 - }, - { - "epoch": 1.951655328798186, - "grad_norm": 0.1650390625, - "learning_rate": 7.078692938440827e-05, - "loss": 0.9542, - "step": 5380 - }, - { - "epoch": 1.9523809523809523, - "grad_norm": 0.1455078125, - "learning_rate": 7.073797576795986e-05, - "loss": 0.9328, - "step": 5382 - }, - { - "epoch": 1.953106575963719, - "grad_norm": 0.1728515625, - "learning_rate": 7.068902215151145e-05, - "loss": 0.9566, - "step": 5384 - }, - { - "epoch": 1.9538321995464853, - "grad_norm": 0.1865234375, - "learning_rate": 7.064006853506303e-05, - "loss": 0.9863, - "step": 5386 - }, - { - "epoch": 1.9545578231292517, - "grad_norm": 0.1494140625, - "learning_rate": 7.059111491861462e-05, - "loss": 0.9328, - "step": 5388 - }, - { - "epoch": 1.9552834467120181, - "grad_norm": 0.169921875, - "learning_rate": 7.05421613021662e-05, - "loss": 0.9978, - "step": 5390 - }, - { - "epoch": 1.9560090702947845, - "grad_norm": 0.177734375, - "learning_rate": 7.049320768571779e-05, - "loss": 1.0063, - "step": 5392 - }, - { - "epoch": 1.956734693877551, - "grad_norm": 0.1669921875, - "learning_rate": 7.044425406926937e-05, - "loss": 0.9898, - "step": 5394 - }, - { - "epoch": 1.9574603174603173, - "grad_norm": 0.166015625, - "learning_rate": 7.039530045282095e-05, - "loss": 0.9487, - "step": 5396 - }, - { - "epoch": 1.9581859410430837, - "grad_norm": 0.162109375, - "learning_rate": 7.034634683637255e-05, - "loss": 0.9949, - "step": 5398 - }, - { - "epoch": 1.9589115646258504, - "grad_norm": 0.1494140625, - "learning_rate": 7.029739321992412e-05, - "loss": 0.9803, - "step": 5400 - }, - { - "epoch": 1.9589115646258504, - "eval_loss": 1.011096715927124, - "eval_runtime": 82.3524, - "eval_samples_per_second": 21.857, - "eval_steps_per_second": 0.692, - "step": 5400 - }, - { - "epoch": 1.9596371882086168, - "grad_norm": 0.1298828125, - "learning_rate": 7.024843960347571e-05, - "loss": 0.9387, - "step": 5402 - }, - { - "epoch": 1.9603628117913834, - "grad_norm": 0.1630859375, - "learning_rate": 7.01994859870273e-05, - "loss": 0.9136, - "step": 5404 - }, - { - "epoch": 1.9610884353741498, - "grad_norm": 0.1748046875, - "learning_rate": 7.015053237057888e-05, - "loss": 0.9976, - "step": 5406 - }, - { - "epoch": 1.9618140589569162, - "grad_norm": 0.150390625, - "learning_rate": 7.010157875413047e-05, - "loss": 0.9514, - "step": 5408 - }, - { - "epoch": 1.9625396825396826, - "grad_norm": 0.1591796875, - "learning_rate": 7.005262513768204e-05, - "loss": 0.9502, - "step": 5410 - }, - { - "epoch": 1.963265306122449, - "grad_norm": 0.16015625, - "learning_rate": 7.000367152123364e-05, - "loss": 0.9255, - "step": 5412 - }, - { - "epoch": 1.9639909297052154, - "grad_norm": 0.1748046875, - "learning_rate": 6.995471790478522e-05, - "loss": 0.9789, - "step": 5414 - }, - { - "epoch": 1.9647165532879818, - "grad_norm": 0.15234375, - "learning_rate": 6.99057642883368e-05, - "loss": 0.9001, - "step": 5416 - }, - { - "epoch": 1.9654421768707482, - "grad_norm": 0.150390625, - "learning_rate": 6.985681067188839e-05, - "loss": 1.0297, - "step": 5418 - }, - { - "epoch": 1.9661678004535146, - "grad_norm": 0.1748046875, - "learning_rate": 6.980785705543997e-05, - "loss": 0.9494, - "step": 5420 - }, - { - "epoch": 1.9668934240362812, - "grad_norm": 0.1591796875, - "learning_rate": 6.975890343899156e-05, - "loss": 0.9232, - "step": 5422 - }, - { - "epoch": 1.9676190476190476, - "grad_norm": 0.158203125, - "learning_rate": 6.970994982254315e-05, - "loss": 0.9913, - "step": 5424 - }, - { - "epoch": 1.9683446712018142, - "grad_norm": 0.1494140625, - "learning_rate": 6.966099620609473e-05, - "loss": 0.9597, - "step": 5426 - }, - { - "epoch": 1.9690702947845806, - "grad_norm": 0.1650390625, - "learning_rate": 6.961204258964631e-05, - "loss": 0.9634, - "step": 5428 - }, - { - "epoch": 1.969795918367347, - "grad_norm": 0.1943359375, - "learning_rate": 6.95630889731979e-05, - "loss": 1.027, - "step": 5430 - }, - { - "epoch": 1.9705215419501134, - "grad_norm": 0.17578125, - "learning_rate": 6.951413535674949e-05, - "loss": 0.886, - "step": 5432 - }, - { - "epoch": 1.9712471655328798, - "grad_norm": 0.1650390625, - "learning_rate": 6.946518174030107e-05, - "loss": 1.0277, - "step": 5434 - }, - { - "epoch": 1.9719727891156462, - "grad_norm": 0.166015625, - "learning_rate": 6.941622812385265e-05, - "loss": 1.018, - "step": 5436 - }, - { - "epoch": 1.9726984126984126, - "grad_norm": 0.17578125, - "learning_rate": 6.936727450740424e-05, - "loss": 0.9642, - "step": 5438 - }, - { - "epoch": 1.973424036281179, - "grad_norm": 0.1748046875, - "learning_rate": 6.931832089095582e-05, - "loss": 0.9943, - "step": 5440 - }, - { - "epoch": 1.9741496598639454, - "grad_norm": 0.166015625, - "learning_rate": 6.926936727450741e-05, - "loss": 0.9534, - "step": 5442 - }, - { - "epoch": 1.974875283446712, - "grad_norm": 0.1845703125, - "learning_rate": 6.9220413658059e-05, - "loss": 0.9578, - "step": 5444 - }, - { - "epoch": 1.9756009070294784, - "grad_norm": 0.1787109375, - "learning_rate": 6.917146004161058e-05, - "loss": 0.9278, - "step": 5446 - }, - { - "epoch": 1.976326530612245, - "grad_norm": 0.1962890625, - "learning_rate": 6.912250642516216e-05, - "loss": 1.0084, - "step": 5448 - }, - { - "epoch": 1.9770521541950115, - "grad_norm": 0.171875, - "learning_rate": 6.907355280871374e-05, - "loss": 0.9462, - "step": 5450 - }, - { - "epoch": 1.9777777777777779, - "grad_norm": 0.1611328125, - "learning_rate": 6.902459919226533e-05, - "loss": 0.9877, - "step": 5452 - }, - { - "epoch": 1.9785034013605443, - "grad_norm": 0.169921875, - "learning_rate": 6.897564557581692e-05, - "loss": 0.9353, - "step": 5454 - }, - { - "epoch": 1.9792290249433107, - "grad_norm": 0.15625, - "learning_rate": 6.89266919593685e-05, - "loss": 1.0031, - "step": 5456 - }, - { - "epoch": 1.979954648526077, - "grad_norm": 0.171875, - "learning_rate": 6.887773834292009e-05, - "loss": 0.9874, - "step": 5458 - }, - { - "epoch": 1.9806802721088435, - "grad_norm": 0.1767578125, - "learning_rate": 6.882878472647167e-05, - "loss": 0.9689, - "step": 5460 - }, - { - "epoch": 1.9814058956916099, - "grad_norm": 0.1630859375, - "learning_rate": 6.877983111002325e-05, - "loss": 0.9945, - "step": 5462 - }, - { - "epoch": 1.9821315192743763, - "grad_norm": 0.1748046875, - "learning_rate": 6.873087749357485e-05, - "loss": 0.9695, - "step": 5464 - }, - { - "epoch": 1.9828571428571429, - "grad_norm": 0.1591796875, - "learning_rate": 6.868192387712643e-05, - "loss": 0.9294, - "step": 5466 - }, - { - "epoch": 1.9835827664399093, - "grad_norm": 0.16015625, - "learning_rate": 6.863297026067801e-05, - "loss": 1.0133, - "step": 5468 - }, - { - "epoch": 1.984308390022676, - "grad_norm": 0.171875, - "learning_rate": 6.85840166442296e-05, - "loss": 0.9074, - "step": 5470 - }, - { - "epoch": 1.9850340136054423, - "grad_norm": 0.1982421875, - "learning_rate": 6.853506302778118e-05, - "loss": 0.9792, - "step": 5472 - }, - { - "epoch": 1.9857596371882087, - "grad_norm": 0.1845703125, - "learning_rate": 6.848610941133277e-05, - "loss": 1.0607, - "step": 5474 - }, - { - "epoch": 1.986485260770975, - "grad_norm": 0.1572265625, - "learning_rate": 6.843715579488435e-05, - "loss": 0.9649, - "step": 5476 - }, - { - "epoch": 1.9872108843537415, - "grad_norm": 0.1650390625, - "learning_rate": 6.838820217843594e-05, - "loss": 0.9928, - "step": 5478 - }, - { - "epoch": 1.987936507936508, - "grad_norm": 0.2119140625, - "learning_rate": 6.833924856198752e-05, - "loss": 0.9035, - "step": 5480 - }, - { - "epoch": 1.9886621315192743, - "grad_norm": 0.1875, - "learning_rate": 6.82902949455391e-05, - "loss": 1.0789, - "step": 5482 - }, - { - "epoch": 1.9893877551020407, - "grad_norm": 0.181640625, - "learning_rate": 6.82413413290907e-05, - "loss": 0.9973, - "step": 5484 - }, - { - "epoch": 1.9901133786848073, - "grad_norm": 0.1669921875, - "learning_rate": 6.819238771264227e-05, - "loss": 1.0186, - "step": 5486 - }, - { - "epoch": 1.9908390022675737, - "grad_norm": 0.1630859375, - "learning_rate": 6.814343409619386e-05, - "loss": 0.9575, - "step": 5488 - }, - { - "epoch": 1.99156462585034, - "grad_norm": 0.1513671875, - "learning_rate": 6.809448047974544e-05, - "loss": 0.9181, - "step": 5490 - }, - { - "epoch": 1.9922902494331067, - "grad_norm": 0.1826171875, - "learning_rate": 6.804552686329703e-05, - "loss": 0.9937, - "step": 5492 - }, - { - "epoch": 1.9930158730158731, - "grad_norm": 0.1923828125, - "learning_rate": 6.799657324684862e-05, - "loss": 0.9643, - "step": 5494 - }, - { - "epoch": 1.9937414965986395, - "grad_norm": 0.154296875, - "learning_rate": 6.794761963040019e-05, - "loss": 0.9727, - "step": 5496 - }, - { - "epoch": 1.994467120181406, - "grad_norm": 0.1552734375, - "learning_rate": 6.789866601395179e-05, - "loss": 0.9577, - "step": 5498 - }, - { - "epoch": 1.9951927437641723, - "grad_norm": 0.1611328125, - "learning_rate": 6.784971239750337e-05, - "loss": 0.9759, - "step": 5500 - }, - { - "epoch": 1.9951927437641723, - "eval_loss": 1.0108295679092407, - "eval_runtime": 88.2979, - "eval_samples_per_second": 20.386, - "eval_steps_per_second": 0.646, - "step": 5500 - }, - { - "epoch": 1.9959183673469387, - "grad_norm": 0.1748046875, - "learning_rate": 6.780075878105495e-05, - "loss": 1.0464, - "step": 5502 - }, - { - "epoch": 1.9966439909297051, - "grad_norm": 0.177734375, - "learning_rate": 6.775180516460655e-05, - "loss": 1.0008, - "step": 5504 - }, - { - "epoch": 1.9973696145124715, - "grad_norm": 0.19140625, - "learning_rate": 6.770285154815812e-05, - "loss": 0.9546, - "step": 5506 - }, - { - "epoch": 1.9980952380952381, - "grad_norm": 0.1611328125, - "learning_rate": 6.765389793170971e-05, - "loss": 0.9799, - "step": 5508 - }, - { - "epoch": 1.9988208616780045, - "grad_norm": 0.1650390625, - "learning_rate": 6.760494431526129e-05, - "loss": 1.0274, - "step": 5510 - }, - { - "epoch": 1.999546485260771, - "grad_norm": 0.16015625, - "learning_rate": 6.755599069881288e-05, - "loss": 0.9454, - "step": 5512 - }, - { - "epoch": 2.0, - "grad_norm": 0.3046875, - "learning_rate": 6.750703708236447e-05, - "loss": 0.9132, - "step": 5514 - }, - { - "epoch": 2.0007256235827664, - "grad_norm": 0.19140625, - "learning_rate": 6.745808346591604e-05, - "loss": 0.9337, - "step": 5516 - }, - { - "epoch": 2.001451247165533, - "grad_norm": 0.150390625, - "learning_rate": 6.740912984946764e-05, - "loss": 0.9514, - "step": 5518 - }, - { - "epoch": 2.002176870748299, - "grad_norm": 0.1806640625, - "learning_rate": 6.736017623301922e-05, - "loss": 1.0014, - "step": 5520 - }, - { - "epoch": 2.0029024943310656, - "grad_norm": 0.1796875, - "learning_rate": 6.73112226165708e-05, - "loss": 0.9509, - "step": 5522 - }, - { - "epoch": 2.003628117913832, - "grad_norm": 0.154296875, - "learning_rate": 6.72622690001224e-05, - "loss": 1.0404, - "step": 5524 - }, - { - "epoch": 2.004353741496599, - "grad_norm": 0.16796875, - "learning_rate": 6.721331538367396e-05, - "loss": 0.9281, - "step": 5526 - }, - { - "epoch": 2.0050793650793652, - "grad_norm": 0.16796875, - "learning_rate": 6.716436176722556e-05, - "loss": 0.9993, - "step": 5528 - }, - { - "epoch": 2.0058049886621316, - "grad_norm": 0.162109375, - "learning_rate": 6.711540815077714e-05, - "loss": 1.0137, - "step": 5530 - }, - { - "epoch": 2.006530612244898, - "grad_norm": 0.1708984375, - "learning_rate": 6.706645453432873e-05, - "loss": 0.9299, - "step": 5532 - }, - { - "epoch": 2.0072562358276644, - "grad_norm": 0.1689453125, - "learning_rate": 6.701750091788032e-05, - "loss": 0.9612, - "step": 5534 - }, - { - "epoch": 2.007981859410431, - "grad_norm": 0.171875, - "learning_rate": 6.696854730143189e-05, - "loss": 1.0255, - "step": 5536 - }, - { - "epoch": 2.0087074829931972, - "grad_norm": 0.158203125, - "learning_rate": 6.691959368498349e-05, - "loss": 0.9482, - "step": 5538 - }, - { - "epoch": 2.0094331065759636, - "grad_norm": 0.1630859375, - "learning_rate": 6.687064006853507e-05, - "loss": 0.9571, - "step": 5540 - }, - { - "epoch": 2.01015873015873, - "grad_norm": 0.1572265625, - "learning_rate": 6.682168645208665e-05, - "loss": 0.9913, - "step": 5542 - }, - { - "epoch": 2.0108843537414964, - "grad_norm": 0.169921875, - "learning_rate": 6.677273283563825e-05, - "loss": 0.9695, - "step": 5544 - }, - { - "epoch": 2.011609977324263, - "grad_norm": 0.1591796875, - "learning_rate": 6.672377921918981e-05, - "loss": 0.9805, - "step": 5546 - }, - { - "epoch": 2.0123356009070297, - "grad_norm": 0.1689453125, - "learning_rate": 6.667482560274141e-05, - "loss": 0.9468, - "step": 5548 - }, - { - "epoch": 2.013061224489796, - "grad_norm": 0.1484375, - "learning_rate": 6.662587198629299e-05, - "loss": 0.9305, - "step": 5550 - }, - { - "epoch": 2.0137868480725625, - "grad_norm": 0.1494140625, - "learning_rate": 6.657691836984458e-05, - "loss": 0.956, - "step": 5552 - }, - { - "epoch": 2.014512471655329, - "grad_norm": 0.1689453125, - "learning_rate": 6.652796475339617e-05, - "loss": 0.9536, - "step": 5554 - }, - { - "epoch": 2.0152380952380953, - "grad_norm": 0.173828125, - "learning_rate": 6.647901113694774e-05, - "loss": 0.9555, - "step": 5556 - }, - { - "epoch": 2.0159637188208617, - "grad_norm": 0.150390625, - "learning_rate": 6.643005752049934e-05, - "loss": 0.9548, - "step": 5558 - }, - { - "epoch": 2.016689342403628, - "grad_norm": 0.1689453125, - "learning_rate": 6.638110390405092e-05, - "loss": 1.0376, - "step": 5560 - }, - { - "epoch": 2.0174149659863945, - "grad_norm": 0.173828125, - "learning_rate": 6.63321502876025e-05, - "loss": 0.9717, - "step": 5562 - }, - { - "epoch": 2.018140589569161, - "grad_norm": 0.1865234375, - "learning_rate": 6.628319667115408e-05, - "loss": 1.0095, - "step": 5564 - }, - { - "epoch": 2.0188662131519273, - "grad_norm": 0.1533203125, - "learning_rate": 6.623424305470566e-05, - "loss": 0.9583, - "step": 5566 - }, - { - "epoch": 2.0195918367346937, - "grad_norm": 0.1640625, - "learning_rate": 6.618528943825726e-05, - "loss": 1.0386, - "step": 5568 - }, - { - "epoch": 2.0203174603174605, - "grad_norm": 0.18359375, - "learning_rate": 6.613633582180884e-05, - "loss": 1.0481, - "step": 5570 - }, - { - "epoch": 2.021043083900227, - "grad_norm": 0.15625, - "learning_rate": 6.608738220536043e-05, - "loss": 0.9812, - "step": 5572 - }, - { - "epoch": 2.0217687074829933, - "grad_norm": 0.16796875, - "learning_rate": 6.603842858891201e-05, - "loss": 0.9529, - "step": 5574 - }, - { - "epoch": 2.0224943310657597, - "grad_norm": 0.1611328125, - "learning_rate": 6.598947497246359e-05, - "loss": 0.9702, - "step": 5576 - }, - { - "epoch": 2.023219954648526, - "grad_norm": 0.1796875, - "learning_rate": 6.594052135601519e-05, - "loss": 1.0463, - "step": 5578 - }, - { - "epoch": 2.0239455782312925, - "grad_norm": 0.1552734375, - "learning_rate": 6.589156773956677e-05, - "loss": 0.9735, - "step": 5580 - }, - { - "epoch": 2.024671201814059, - "grad_norm": 0.1591796875, - "learning_rate": 6.584261412311835e-05, - "loss": 0.9631, - "step": 5582 - }, - { - "epoch": 2.0253968253968253, - "grad_norm": 0.1494140625, - "learning_rate": 6.579366050666993e-05, - "loss": 1.0417, - "step": 5584 - }, - { - "epoch": 2.0261224489795917, - "grad_norm": 0.171875, - "learning_rate": 6.574470689022151e-05, - "loss": 0.9679, - "step": 5586 - }, - { - "epoch": 2.026848072562358, - "grad_norm": 0.15625, - "learning_rate": 6.569575327377311e-05, - "loss": 0.9433, - "step": 5588 - }, - { - "epoch": 2.0275736961451245, - "grad_norm": 0.1650390625, - "learning_rate": 6.564679965732469e-05, - "loss": 1.1191, - "step": 5590 - }, - { - "epoch": 2.0282993197278913, - "grad_norm": 0.166015625, - "learning_rate": 6.559784604087628e-05, - "loss": 1.0056, - "step": 5592 - }, - { - "epoch": 2.0290249433106577, - "grad_norm": 0.140625, - "learning_rate": 6.554889242442786e-05, - "loss": 0.8881, - "step": 5594 - }, - { - "epoch": 2.029750566893424, - "grad_norm": 0.15234375, - "learning_rate": 6.549993880797944e-05, - "loss": 1.005, - "step": 5596 - }, - { - "epoch": 2.0304761904761905, - "grad_norm": 0.16015625, - "learning_rate": 6.545098519153102e-05, - "loss": 0.9251, - "step": 5598 - }, - { - "epoch": 2.031201814058957, - "grad_norm": 0.1630859375, - "learning_rate": 6.540203157508262e-05, - "loss": 0.9907, - "step": 5600 - }, - { - "epoch": 2.031201814058957, - "eval_loss": 1.0105645656585693, - "eval_runtime": 88.0197, - "eval_samples_per_second": 20.45, - "eval_steps_per_second": 0.648, - "step": 5600 - }, - { - "epoch": 2.0319274376417233, - "grad_norm": 0.1669921875, - "learning_rate": 6.53530779586342e-05, - "loss": 1.0097, - "step": 5602 - }, - { - "epoch": 2.0326530612244897, - "grad_norm": 0.1806640625, - "learning_rate": 6.530412434218578e-05, - "loss": 0.9781, - "step": 5604 - }, - { - "epoch": 2.033378684807256, - "grad_norm": 0.1552734375, - "learning_rate": 6.525517072573736e-05, - "loss": 0.9432, - "step": 5606 - }, - { - "epoch": 2.0341043083900225, - "grad_norm": 0.1875, - "learning_rate": 6.520621710928895e-05, - "loss": 0.9518, - "step": 5608 - }, - { - "epoch": 2.034829931972789, - "grad_norm": 0.1650390625, - "learning_rate": 6.515726349284054e-05, - "loss": 0.9617, - "step": 5610 - }, - { - "epoch": 2.0355555555555553, - "grad_norm": 0.181640625, - "learning_rate": 6.510830987639212e-05, - "loss": 0.954, - "step": 5612 - }, - { - "epoch": 2.036281179138322, - "grad_norm": 0.173828125, - "learning_rate": 6.505935625994371e-05, - "loss": 1.0424, - "step": 5614 - }, - { - "epoch": 2.0370068027210886, - "grad_norm": 0.251953125, - "learning_rate": 6.501040264349529e-05, - "loss": 1.0404, - "step": 5616 - }, - { - "epoch": 2.037732426303855, - "grad_norm": 0.181640625, - "learning_rate": 6.496144902704687e-05, - "loss": 0.989, - "step": 5618 - }, - { - "epoch": 2.0384580498866214, - "grad_norm": 0.1787109375, - "learning_rate": 6.491249541059847e-05, - "loss": 0.9792, - "step": 5620 - }, - { - "epoch": 2.039183673469388, - "grad_norm": 0.169921875, - "learning_rate": 6.486354179415005e-05, - "loss": 1.0176, - "step": 5622 - }, - { - "epoch": 2.039909297052154, - "grad_norm": 0.18359375, - "learning_rate": 6.481458817770163e-05, - "loss": 0.9985, - "step": 5624 - }, - { - "epoch": 2.0406349206349206, - "grad_norm": 0.1689453125, - "learning_rate": 6.476563456125321e-05, - "loss": 0.9782, - "step": 5626 - }, - { - "epoch": 2.041360544217687, - "grad_norm": 0.158203125, - "learning_rate": 6.47166809448048e-05, - "loss": 0.9238, - "step": 5628 - }, - { - "epoch": 2.0420861678004534, - "grad_norm": 0.162109375, - "learning_rate": 6.466772732835639e-05, - "loss": 1.0303, - "step": 5630 - }, - { - "epoch": 2.0428117913832198, - "grad_norm": 0.1767578125, - "learning_rate": 6.461877371190796e-05, - "loss": 1.0011, - "step": 5632 - }, - { - "epoch": 2.043537414965986, - "grad_norm": 0.162109375, - "learning_rate": 6.456982009545956e-05, - "loss": 0.981, - "step": 5634 - }, - { - "epoch": 2.044263038548753, - "grad_norm": 0.19140625, - "learning_rate": 6.452086647901114e-05, - "loss": 0.9464, - "step": 5636 - }, - { - "epoch": 2.0449886621315194, - "grad_norm": 0.1669921875, - "learning_rate": 6.447191286256272e-05, - "loss": 1.0066, - "step": 5638 - }, - { - "epoch": 2.045714285714286, - "grad_norm": 0.16015625, - "learning_rate": 6.442295924611432e-05, - "loss": 0.9959, - "step": 5640 - }, - { - "epoch": 2.046439909297052, - "grad_norm": 0.16796875, - "learning_rate": 6.437400562966589e-05, - "loss": 0.942, - "step": 5642 - }, - { - "epoch": 2.0471655328798186, - "grad_norm": 0.1630859375, - "learning_rate": 6.432505201321748e-05, - "loss": 0.9856, - "step": 5644 - }, - { - "epoch": 2.047891156462585, - "grad_norm": 0.1962890625, - "learning_rate": 6.427609839676906e-05, - "loss": 0.9828, - "step": 5646 - }, - { - "epoch": 2.0486167800453514, - "grad_norm": 0.1728515625, - "learning_rate": 6.422714478032065e-05, - "loss": 0.9923, - "step": 5648 - }, - { - "epoch": 2.049342403628118, - "grad_norm": 0.1748046875, - "learning_rate": 6.417819116387224e-05, - "loss": 1.0303, - "step": 5650 - }, - { - "epoch": 2.050068027210884, - "grad_norm": 0.1689453125, - "learning_rate": 6.412923754742381e-05, - "loss": 1.0702, - "step": 5652 - }, - { - "epoch": 2.0507936507936506, - "grad_norm": 0.21875, - "learning_rate": 6.408028393097541e-05, - "loss": 1.0109, - "step": 5654 - }, - { - "epoch": 2.0515192743764175, - "grad_norm": 0.1826171875, - "learning_rate": 6.403133031452699e-05, - "loss": 0.9457, - "step": 5656 - }, - { - "epoch": 2.052244897959184, - "grad_norm": 0.1630859375, - "learning_rate": 6.398237669807857e-05, - "loss": 1.0746, - "step": 5658 - }, - { - "epoch": 2.0529705215419503, - "grad_norm": 0.16796875, - "learning_rate": 6.393342308163017e-05, - "loss": 0.9641, - "step": 5660 - }, - { - "epoch": 2.0536961451247167, - "grad_norm": 0.1669921875, - "learning_rate": 6.388446946518174e-05, - "loss": 1.0007, - "step": 5662 - }, - { - "epoch": 2.054421768707483, - "grad_norm": 0.203125, - "learning_rate": 6.383551584873333e-05, - "loss": 0.9972, - "step": 5664 - }, - { - "epoch": 2.0551473922902495, - "grad_norm": 0.1787109375, - "learning_rate": 6.378656223228491e-05, - "loss": 1.0813, - "step": 5666 - }, - { - "epoch": 2.055873015873016, - "grad_norm": 0.154296875, - "learning_rate": 6.37376086158365e-05, - "loss": 0.9023, - "step": 5668 - }, - { - "epoch": 2.0565986394557823, - "grad_norm": 0.154296875, - "learning_rate": 6.368865499938809e-05, - "loss": 0.9559, - "step": 5670 - }, - { - "epoch": 2.0573242630385487, - "grad_norm": 0.150390625, - "learning_rate": 6.363970138293966e-05, - "loss": 0.9024, - "step": 5672 - }, - { - "epoch": 2.058049886621315, - "grad_norm": 0.1494140625, - "learning_rate": 6.359074776649126e-05, - "loss": 0.9186, - "step": 5674 - }, - { - "epoch": 2.0587755102040814, - "grad_norm": 0.1767578125, - "learning_rate": 6.354179415004284e-05, - "loss": 1.0022, - "step": 5676 - }, - { - "epoch": 2.0595011337868483, - "grad_norm": 0.1865234375, - "learning_rate": 6.349284053359442e-05, - "loss": 1.0076, - "step": 5678 - }, - { - "epoch": 2.0602267573696147, - "grad_norm": 0.166015625, - "learning_rate": 6.344388691714602e-05, - "loss": 0.9731, - "step": 5680 - }, - { - "epoch": 2.060952380952381, - "grad_norm": 0.166015625, - "learning_rate": 6.339493330069759e-05, - "loss": 0.9956, - "step": 5682 - }, - { - "epoch": 2.0616780045351475, - "grad_norm": 0.18359375, - "learning_rate": 6.334597968424918e-05, - "loss": 0.918, - "step": 5684 - }, - { - "epoch": 2.062403628117914, - "grad_norm": 0.1689453125, - "learning_rate": 6.329702606780076e-05, - "loss": 0.9229, - "step": 5686 - }, - { - "epoch": 2.0631292517006803, - "grad_norm": 0.15234375, - "learning_rate": 6.324807245135235e-05, - "loss": 0.9616, - "step": 5688 - }, - { - "epoch": 2.0638548752834467, - "grad_norm": 0.177734375, - "learning_rate": 6.319911883490394e-05, - "loss": 1.0174, - "step": 5690 - }, - { - "epoch": 2.064580498866213, - "grad_norm": 0.1533203125, - "learning_rate": 6.315016521845551e-05, - "loss": 1.0449, - "step": 5692 - }, - { - "epoch": 2.0653061224489795, - "grad_norm": 0.1787109375, - "learning_rate": 6.310121160200711e-05, - "loss": 0.9916, - "step": 5694 - }, - { - "epoch": 2.066031746031746, - "grad_norm": 0.166015625, - "learning_rate": 6.305225798555869e-05, - "loss": 0.9486, - "step": 5696 - }, - { - "epoch": 2.0667573696145123, - "grad_norm": 0.1494140625, - "learning_rate": 6.300330436911027e-05, - "loss": 0.9569, - "step": 5698 - }, - { - "epoch": 2.067482993197279, - "grad_norm": 0.1640625, - "learning_rate": 6.295435075266187e-05, - "loss": 0.9016, - "step": 5700 - }, - { - "epoch": 2.067482993197279, - "eval_loss": 1.0106786489486694, - "eval_runtime": 89.0366, - "eval_samples_per_second": 20.216, - "eval_steps_per_second": 0.64, - "step": 5700 - }, - { - "epoch": 2.0682086167800455, - "grad_norm": 0.1640625, - "learning_rate": 6.290539713621344e-05, - "loss": 0.9435, - "step": 5702 - }, - { - "epoch": 2.068934240362812, - "grad_norm": 0.1904296875, - "learning_rate": 6.285644351976503e-05, - "loss": 1.0447, - "step": 5704 - }, - { - "epoch": 2.0696598639455783, - "grad_norm": 0.1689453125, - "learning_rate": 6.280748990331661e-05, - "loss": 0.9554, - "step": 5706 - }, - { - "epoch": 2.0703854875283447, - "grad_norm": 0.162109375, - "learning_rate": 6.27585362868682e-05, - "loss": 1.03, - "step": 5708 - }, - { - "epoch": 2.071111111111111, - "grad_norm": 0.18359375, - "learning_rate": 6.270958267041978e-05, - "loss": 0.9301, - "step": 5710 - }, - { - "epoch": 2.0718367346938775, - "grad_norm": 0.1787109375, - "learning_rate": 6.266062905397136e-05, - "loss": 1.0046, - "step": 5712 - }, - { - "epoch": 2.072562358276644, - "grad_norm": 0.185546875, - "learning_rate": 6.261167543752296e-05, - "loss": 1.0349, - "step": 5714 - }, - { - "epoch": 2.0732879818594103, - "grad_norm": 0.171875, - "learning_rate": 6.256272182107454e-05, - "loss": 1.0612, - "step": 5716 - }, - { - "epoch": 2.0740136054421767, - "grad_norm": 0.1591796875, - "learning_rate": 6.251376820462612e-05, - "loss": 0.9338, - "step": 5718 - }, - { - "epoch": 2.074739229024943, - "grad_norm": 0.1826171875, - "learning_rate": 6.24648145881777e-05, - "loss": 1.0275, - "step": 5720 - }, - { - "epoch": 2.07546485260771, - "grad_norm": 0.1552734375, - "learning_rate": 6.241586097172929e-05, - "loss": 0.9196, - "step": 5722 - }, - { - "epoch": 2.0761904761904764, - "grad_norm": 0.150390625, - "learning_rate": 6.236690735528088e-05, - "loss": 0.9807, - "step": 5724 - }, - { - "epoch": 2.0769160997732428, - "grad_norm": 0.171875, - "learning_rate": 6.231795373883246e-05, - "loss": 0.9343, - "step": 5726 - }, - { - "epoch": 2.077641723356009, - "grad_norm": 0.181640625, - "learning_rate": 6.226900012238405e-05, - "loss": 0.9618, - "step": 5728 - }, - { - "epoch": 2.0783673469387756, - "grad_norm": 0.18359375, - "learning_rate": 6.222004650593563e-05, - "loss": 0.9586, - "step": 5730 - }, - { - "epoch": 2.079092970521542, - "grad_norm": 0.1640625, - "learning_rate": 6.217109288948721e-05, - "loss": 1.0001, - "step": 5732 - }, - { - "epoch": 2.0798185941043084, - "grad_norm": 0.18359375, - "learning_rate": 6.21221392730388e-05, - "loss": 1.0526, - "step": 5734 - }, - { - "epoch": 2.0805442176870748, - "grad_norm": 0.1806640625, - "learning_rate": 6.207318565659039e-05, - "loss": 0.9504, - "step": 5736 - }, - { - "epoch": 2.081269841269841, - "grad_norm": 0.150390625, - "learning_rate": 6.202423204014197e-05, - "loss": 0.923, - "step": 5738 - }, - { - "epoch": 2.0819954648526076, - "grad_norm": 0.1845703125, - "learning_rate": 6.197527842369355e-05, - "loss": 0.9957, - "step": 5740 - }, - { - "epoch": 2.082721088435374, - "grad_norm": 0.1728515625, - "learning_rate": 6.192632480724514e-05, - "loss": 0.8884, - "step": 5742 - }, - { - "epoch": 2.083446712018141, - "grad_norm": 0.162109375, - "learning_rate": 6.187737119079672e-05, - "loss": 0.9332, - "step": 5744 - }, - { - "epoch": 2.084172335600907, - "grad_norm": 0.1904296875, - "learning_rate": 6.182841757434831e-05, - "loss": 0.9743, - "step": 5746 - }, - { - "epoch": 2.0848979591836736, - "grad_norm": 0.1669921875, - "learning_rate": 6.17794639578999e-05, - "loss": 0.9578, - "step": 5748 - }, - { - "epoch": 2.08562358276644, - "grad_norm": 0.1865234375, - "learning_rate": 6.173051034145148e-05, - "loss": 0.9793, - "step": 5750 - }, - { - "epoch": 2.0863492063492064, - "grad_norm": 0.162109375, - "learning_rate": 6.168155672500306e-05, - "loss": 0.9264, - "step": 5752 - }, - { - "epoch": 2.087074829931973, - "grad_norm": 0.171875, - "learning_rate": 6.163260310855464e-05, - "loss": 0.92, - "step": 5754 - }, - { - "epoch": 2.087800453514739, - "grad_norm": 0.1806640625, - "learning_rate": 6.158364949210624e-05, - "loss": 1.0329, - "step": 5756 - }, - { - "epoch": 2.0885260770975056, - "grad_norm": 0.1796875, - "learning_rate": 6.153469587565782e-05, - "loss": 0.9942, - "step": 5758 - }, - { - "epoch": 2.089251700680272, - "grad_norm": 0.1494140625, - "learning_rate": 6.14857422592094e-05, - "loss": 0.9291, - "step": 5760 - }, - { - "epoch": 2.0899773242630384, - "grad_norm": 0.150390625, - "learning_rate": 6.143678864276099e-05, - "loss": 0.9504, - "step": 5762 - }, - { - "epoch": 2.090702947845805, - "grad_norm": 0.1728515625, - "learning_rate": 6.138783502631257e-05, - "loss": 0.9671, - "step": 5764 - }, - { - "epoch": 2.0914285714285716, - "grad_norm": 0.20703125, - "learning_rate": 6.133888140986416e-05, - "loss": 1.0002, - "step": 5766 - }, - { - "epoch": 2.092154195011338, - "grad_norm": 0.1591796875, - "learning_rate": 6.128992779341575e-05, - "loss": 0.9597, - "step": 5768 - }, - { - "epoch": 2.0928798185941044, - "grad_norm": 0.1884765625, - "learning_rate": 6.124097417696733e-05, - "loss": 1.0265, - "step": 5770 - }, - { - "epoch": 2.093605442176871, - "grad_norm": 0.1796875, - "learning_rate": 6.119202056051891e-05, - "loss": 1.035, - "step": 5772 - }, - { - "epoch": 2.0943310657596372, - "grad_norm": 0.166015625, - "learning_rate": 6.114306694407049e-05, - "loss": 0.9861, - "step": 5774 - }, - { - "epoch": 2.0950566893424036, - "grad_norm": 0.173828125, - "learning_rate": 6.109411332762209e-05, - "loss": 1.011, - "step": 5776 - }, - { - "epoch": 2.09578231292517, - "grad_norm": 0.162109375, - "learning_rate": 6.104515971117366e-05, - "loss": 0.9615, - "step": 5778 - }, - { - "epoch": 2.0965079365079364, - "grad_norm": 0.1650390625, - "learning_rate": 6.099620609472525e-05, - "loss": 1.0444, - "step": 5780 - }, - { - "epoch": 2.097233560090703, - "grad_norm": 0.17578125, - "learning_rate": 6.094725247827684e-05, - "loss": 0.9458, - "step": 5782 - }, - { - "epoch": 2.0979591836734692, - "grad_norm": 0.1533203125, - "learning_rate": 6.089829886182842e-05, - "loss": 0.9296, - "step": 5784 - }, - { - "epoch": 2.0986848072562356, - "grad_norm": 0.1630859375, - "learning_rate": 6.084934524538001e-05, - "loss": 0.8853, - "step": 5786 - }, - { - "epoch": 2.0994104308390025, - "grad_norm": 0.2177734375, - "learning_rate": 6.080039162893158e-05, - "loss": 0.9348, - "step": 5788 - }, - { - "epoch": 2.100136054421769, - "grad_norm": 0.1611328125, - "learning_rate": 6.075143801248318e-05, - "loss": 0.934, - "step": 5790 - }, - { - "epoch": 2.1008616780045353, - "grad_norm": 0.1875, - "learning_rate": 6.070248439603477e-05, - "loss": 0.9812, - "step": 5792 - }, - { - "epoch": 2.1015873015873017, - "grad_norm": 0.1806640625, - "learning_rate": 6.065353077958634e-05, - "loss": 0.9687, - "step": 5794 - }, - { - "epoch": 2.102312925170068, - "grad_norm": 0.150390625, - "learning_rate": 6.060457716313793e-05, - "loss": 0.9097, - "step": 5796 - }, - { - "epoch": 2.1030385487528345, - "grad_norm": 0.166015625, - "learning_rate": 6.055562354668951e-05, - "loss": 0.9202, - "step": 5798 - }, - { - "epoch": 2.103764172335601, - "grad_norm": 0.17578125, - "learning_rate": 6.05066699302411e-05, - "loss": 0.9264, - "step": 5800 - }, - { - "epoch": 2.103764172335601, - "eval_loss": 1.010640025138855, - "eval_runtime": 86.5952, - "eval_samples_per_second": 20.786, - "eval_steps_per_second": 0.658, - "step": 5800 - }, - { - "epoch": 2.1044897959183673, - "grad_norm": 0.1689453125, - "learning_rate": 6.045771631379269e-05, - "loss": 0.9798, - "step": 5802 - }, - { - "epoch": 2.1052154195011337, - "grad_norm": 0.1640625, - "learning_rate": 6.040876269734427e-05, - "loss": 0.953, - "step": 5804 - }, - { - "epoch": 2.1059410430839, - "grad_norm": 0.169921875, - "learning_rate": 6.035980908089586e-05, - "loss": 0.9906, - "step": 5806 - }, - { - "epoch": 2.1066666666666665, - "grad_norm": 0.197265625, - "learning_rate": 6.031085546444743e-05, - "loss": 0.9762, - "step": 5808 - }, - { - "epoch": 2.1073922902494333, - "grad_norm": 0.166015625, - "learning_rate": 6.026190184799903e-05, - "loss": 0.9851, - "step": 5810 - }, - { - "epoch": 2.1081179138321997, - "grad_norm": 0.171875, - "learning_rate": 6.0212948231550604e-05, - "loss": 0.9887, - "step": 5812 - }, - { - "epoch": 2.108843537414966, - "grad_norm": 0.169921875, - "learning_rate": 6.016399461510219e-05, - "loss": 0.9786, - "step": 5814 - }, - { - "epoch": 2.1095691609977325, - "grad_norm": 0.1650390625, - "learning_rate": 6.011504099865378e-05, - "loss": 0.952, - "step": 5816 - }, - { - "epoch": 2.110294784580499, - "grad_norm": 0.1728515625, - "learning_rate": 6.006608738220536e-05, - "loss": 1.0122, - "step": 5818 - }, - { - "epoch": 2.1110204081632653, - "grad_norm": 0.15625, - "learning_rate": 6.001713376575695e-05, - "loss": 0.9715, - "step": 5820 - }, - { - "epoch": 2.1117460317460317, - "grad_norm": 0.1806640625, - "learning_rate": 5.996818014930853e-05, - "loss": 1.005, - "step": 5822 - }, - { - "epoch": 2.112471655328798, - "grad_norm": 0.1875, - "learning_rate": 5.991922653286012e-05, - "loss": 1.021, - "step": 5824 - }, - { - "epoch": 2.1131972789115645, - "grad_norm": 0.1591796875, - "learning_rate": 5.987027291641171e-05, - "loss": 0.9739, - "step": 5826 - }, - { - "epoch": 2.113922902494331, - "grad_norm": 0.173828125, - "learning_rate": 5.982131929996328e-05, - "loss": 0.9251, - "step": 5828 - }, - { - "epoch": 2.1146485260770973, - "grad_norm": 0.1640625, - "learning_rate": 5.977236568351488e-05, - "loss": 1.0101, - "step": 5830 - }, - { - "epoch": 2.115374149659864, - "grad_norm": 0.1728515625, - "learning_rate": 5.9723412067066454e-05, - "loss": 0.9652, - "step": 5832 - }, - { - "epoch": 2.1160997732426305, - "grad_norm": 0.16015625, - "learning_rate": 5.967445845061804e-05, - "loss": 0.9139, - "step": 5834 - }, - { - "epoch": 2.116825396825397, - "grad_norm": 0.158203125, - "learning_rate": 5.962550483416963e-05, - "loss": 1.0025, - "step": 5836 - }, - { - "epoch": 2.1175510204081633, - "grad_norm": 0.1708984375, - "learning_rate": 5.957655121772121e-05, - "loss": 0.973, - "step": 5838 - }, - { - "epoch": 2.1182766439909297, - "grad_norm": 0.1484375, - "learning_rate": 5.95275976012728e-05, - "loss": 0.918, - "step": 5840 - }, - { - "epoch": 2.119002267573696, - "grad_norm": 0.1572265625, - "learning_rate": 5.947864398482438e-05, - "loss": 0.9193, - "step": 5842 - }, - { - "epoch": 2.1197278911564625, - "grad_norm": 0.177734375, - "learning_rate": 5.942969036837597e-05, - "loss": 0.949, - "step": 5844 - }, - { - "epoch": 2.120453514739229, - "grad_norm": 0.1796875, - "learning_rate": 5.938073675192756e-05, - "loss": 0.9686, - "step": 5846 - }, - { - "epoch": 2.1211791383219953, - "grad_norm": 0.1796875, - "learning_rate": 5.933178313547913e-05, - "loss": 1.0197, - "step": 5848 - }, - { - "epoch": 2.1219047619047617, - "grad_norm": 0.1708984375, - "learning_rate": 5.928282951903073e-05, - "loss": 0.9537, - "step": 5850 - }, - { - "epoch": 2.122630385487528, - "grad_norm": 0.1669921875, - "learning_rate": 5.9233875902582304e-05, - "loss": 0.9571, - "step": 5852 - }, - { - "epoch": 2.123356009070295, - "grad_norm": 0.1748046875, - "learning_rate": 5.918492228613389e-05, - "loss": 0.9315, - "step": 5854 - }, - { - "epoch": 2.1240816326530614, - "grad_norm": 0.1689453125, - "learning_rate": 5.913596866968547e-05, - "loss": 0.9558, - "step": 5856 - }, - { - "epoch": 2.124807256235828, - "grad_norm": 0.16015625, - "learning_rate": 5.908701505323706e-05, - "loss": 0.9256, - "step": 5858 - }, - { - "epoch": 2.125532879818594, - "grad_norm": 0.21875, - "learning_rate": 5.903806143678865e-05, - "loss": 0.9814, - "step": 5860 - }, - { - "epoch": 2.1262585034013606, - "grad_norm": 0.1982421875, - "learning_rate": 5.898910782034023e-05, - "loss": 1.0241, - "step": 5862 - }, - { - "epoch": 2.126984126984127, - "grad_norm": 0.16796875, - "learning_rate": 5.894015420389182e-05, - "loss": 0.9556, - "step": 5864 - }, - { - "epoch": 2.1277097505668934, - "grad_norm": 0.1748046875, - "learning_rate": 5.889120058744339e-05, - "loss": 1.0065, - "step": 5866 - }, - { - "epoch": 2.1284353741496598, - "grad_norm": 0.1708984375, - "learning_rate": 5.884224697099498e-05, - "loss": 0.9229, - "step": 5868 - }, - { - "epoch": 2.129160997732426, - "grad_norm": 0.1591796875, - "learning_rate": 5.879329335454658e-05, - "loss": 0.8913, - "step": 5870 - }, - { - "epoch": 2.1298866213151926, - "grad_norm": 0.193359375, - "learning_rate": 5.8744339738098153e-05, - "loss": 0.9706, - "step": 5872 - }, - { - "epoch": 2.130612244897959, - "grad_norm": 0.1796875, - "learning_rate": 5.869538612164974e-05, - "loss": 0.9607, - "step": 5874 - }, - { - "epoch": 2.131337868480726, - "grad_norm": 0.1728515625, - "learning_rate": 5.864643250520132e-05, - "loss": 0.9314, - "step": 5876 - }, - { - "epoch": 2.132063492063492, - "grad_norm": 0.173828125, - "learning_rate": 5.859747888875291e-05, - "loss": 0.97, - "step": 5878 - }, - { - "epoch": 2.1327891156462586, - "grad_norm": 0.1650390625, - "learning_rate": 5.85485252723045e-05, - "loss": 0.9887, - "step": 5880 - }, - { - "epoch": 2.133514739229025, - "grad_norm": 0.162109375, - "learning_rate": 5.849957165585608e-05, - "loss": 1.013, - "step": 5882 - }, - { - "epoch": 2.1342403628117914, - "grad_norm": 0.15625, - "learning_rate": 5.845061803940767e-05, - "loss": 0.9536, - "step": 5884 - }, - { - "epoch": 2.134965986394558, - "grad_norm": 0.166015625, - "learning_rate": 5.840166442295924e-05, - "loss": 0.9715, - "step": 5886 - }, - { - "epoch": 2.135691609977324, - "grad_norm": 0.15234375, - "learning_rate": 5.835271080651083e-05, - "loss": 1.0258, - "step": 5888 - }, - { - "epoch": 2.1364172335600906, - "grad_norm": 0.1689453125, - "learning_rate": 5.8303757190062414e-05, - "loss": 0.9495, - "step": 5890 - }, - { - "epoch": 2.137142857142857, - "grad_norm": 0.166015625, - "learning_rate": 5.8254803573614e-05, - "loss": 0.9514, - "step": 5892 - }, - { - "epoch": 2.1378684807256234, - "grad_norm": 0.181640625, - "learning_rate": 5.820584995716559e-05, - "loss": 1.001, - "step": 5894 - }, - { - "epoch": 2.13859410430839, - "grad_norm": 0.177734375, - "learning_rate": 5.815689634071717e-05, - "loss": 0.9775, - "step": 5896 - }, - { - "epoch": 2.1393197278911567, - "grad_norm": 0.1826171875, - "learning_rate": 5.810794272426876e-05, - "loss": 0.9368, - "step": 5898 - }, - { - "epoch": 2.140045351473923, - "grad_norm": 0.1650390625, - "learning_rate": 5.805898910782034e-05, - "loss": 0.9397, - "step": 5900 - }, - { - "epoch": 2.140045351473923, - "eval_loss": 1.0105234384536743, - "eval_runtime": 86.09, - "eval_samples_per_second": 20.908, - "eval_steps_per_second": 0.662, - "step": 5900 - }, - { - "epoch": 2.1407709750566895, - "grad_norm": 0.1787109375, - "learning_rate": 5.801003549137193e-05, - "loss": 1.0924, - "step": 5902 - }, - { - "epoch": 2.141496598639456, - "grad_norm": 0.1796875, - "learning_rate": 5.796108187492352e-05, - "loss": 0.9779, - "step": 5904 - }, - { - "epoch": 2.1422222222222222, - "grad_norm": 0.169921875, - "learning_rate": 5.791212825847509e-05, - "loss": 0.9686, - "step": 5906 - }, - { - "epoch": 2.1429478458049886, - "grad_norm": 0.1728515625, - "learning_rate": 5.786317464202668e-05, - "loss": 0.9359, - "step": 5908 - }, - { - "epoch": 2.143673469387755, - "grad_norm": 0.185546875, - "learning_rate": 5.7814221025578264e-05, - "loss": 0.982, - "step": 5910 - }, - { - "epoch": 2.1443990929705214, - "grad_norm": 0.1787109375, - "learning_rate": 5.776526740912985e-05, - "loss": 0.887, - "step": 5912 - }, - { - "epoch": 2.145124716553288, - "grad_norm": 0.1611328125, - "learning_rate": 5.771631379268144e-05, - "loss": 0.9527, - "step": 5914 - }, - { - "epoch": 2.1458503401360542, - "grad_norm": 0.1875, - "learning_rate": 5.766736017623302e-05, - "loss": 0.9901, - "step": 5916 - }, - { - "epoch": 2.1465759637188206, - "grad_norm": 0.169921875, - "learning_rate": 5.761840655978461e-05, - "loss": 1.0351, - "step": 5918 - }, - { - "epoch": 2.1473015873015875, - "grad_norm": 0.16796875, - "learning_rate": 5.756945294333619e-05, - "loss": 1.0367, - "step": 5920 - }, - { - "epoch": 2.148027210884354, - "grad_norm": 0.1796875, - "learning_rate": 5.752049932688778e-05, - "loss": 0.9757, - "step": 5922 - }, - { - "epoch": 2.1487528344671203, - "grad_norm": 0.1708984375, - "learning_rate": 5.7471545710439354e-05, - "loss": 0.9131, - "step": 5924 - }, - { - "epoch": 2.1494784580498867, - "grad_norm": 0.1650390625, - "learning_rate": 5.742259209399094e-05, - "loss": 0.9754, - "step": 5926 - }, - { - "epoch": 2.150204081632653, - "grad_norm": 0.1640625, - "learning_rate": 5.737363847754253e-05, - "loss": 1.0091, - "step": 5928 - }, - { - "epoch": 2.1509297052154195, - "grad_norm": 0.197265625, - "learning_rate": 5.7324684861094114e-05, - "loss": 1.0215, - "step": 5930 - }, - { - "epoch": 2.151655328798186, - "grad_norm": 0.1826171875, - "learning_rate": 5.72757312446457e-05, - "loss": 0.9583, - "step": 5932 - }, - { - "epoch": 2.1523809523809523, - "grad_norm": 0.1796875, - "learning_rate": 5.722677762819728e-05, - "loss": 1.0407, - "step": 5934 - }, - { - "epoch": 2.1531065759637187, - "grad_norm": 0.1484375, - "learning_rate": 5.717782401174887e-05, - "loss": 0.8784, - "step": 5936 - }, - { - "epoch": 2.153832199546485, - "grad_norm": 0.1962890625, - "learning_rate": 5.712887039530046e-05, - "loss": 0.9976, - "step": 5938 - }, - { - "epoch": 2.1545578231292515, - "grad_norm": 0.1923828125, - "learning_rate": 5.707991677885204e-05, - "loss": 0.9972, - "step": 5940 - }, - { - "epoch": 2.1552834467120183, - "grad_norm": 0.1650390625, - "learning_rate": 5.703096316240363e-05, - "loss": 0.9872, - "step": 5942 - }, - { - "epoch": 2.1560090702947847, - "grad_norm": 0.169921875, - "learning_rate": 5.6982009545955204e-05, - "loss": 0.9176, - "step": 5944 - }, - { - "epoch": 2.156734693877551, - "grad_norm": 0.1591796875, - "learning_rate": 5.693305592950679e-05, - "loss": 0.9286, - "step": 5946 - }, - { - "epoch": 2.1574603174603175, - "grad_norm": 0.16015625, - "learning_rate": 5.688410231305838e-05, - "loss": 0.9204, - "step": 5948 - }, - { - "epoch": 2.158185941043084, - "grad_norm": 0.1591796875, - "learning_rate": 5.6835148696609964e-05, - "loss": 0.9486, - "step": 5950 - }, - { - "epoch": 2.1589115646258503, - "grad_norm": 0.158203125, - "learning_rate": 5.678619508016155e-05, - "loss": 1.0603, - "step": 5952 - }, - { - "epoch": 2.1596371882086167, - "grad_norm": 0.1748046875, - "learning_rate": 5.673724146371313e-05, - "loss": 0.9545, - "step": 5954 - }, - { - "epoch": 2.160362811791383, - "grad_norm": 0.1884765625, - "learning_rate": 5.668828784726472e-05, - "loss": 1.0123, - "step": 5956 - }, - { - "epoch": 2.1610884353741495, - "grad_norm": 0.150390625, - "learning_rate": 5.66393342308163e-05, - "loss": 0.9411, - "step": 5958 - }, - { - "epoch": 2.161814058956916, - "grad_norm": 0.15625, - "learning_rate": 5.659038061436789e-05, - "loss": 0.9927, - "step": 5960 - }, - { - "epoch": 2.1625396825396823, - "grad_norm": 0.166015625, - "learning_rate": 5.654142699791948e-05, - "loss": 0.9126, - "step": 5962 - }, - { - "epoch": 2.163265306122449, - "grad_norm": 0.193359375, - "learning_rate": 5.6492473381471054e-05, - "loss": 0.9676, - "step": 5964 - }, - { - "epoch": 2.1639909297052156, - "grad_norm": 0.1630859375, - "learning_rate": 5.644351976502264e-05, - "loss": 0.9483, - "step": 5966 - }, - { - "epoch": 2.164716553287982, - "grad_norm": 0.1474609375, - "learning_rate": 5.6394566148574225e-05, - "loss": 0.9706, - "step": 5968 - }, - { - "epoch": 2.1654421768707484, - "grad_norm": 0.162109375, - "learning_rate": 5.6345612532125814e-05, - "loss": 0.978, - "step": 5970 - }, - { - "epoch": 2.1661678004535148, - "grad_norm": 0.19140625, - "learning_rate": 5.62966589156774e-05, - "loss": 0.9953, - "step": 5972 - }, - { - "epoch": 2.166893424036281, - "grad_norm": 0.1650390625, - "learning_rate": 5.624770529922898e-05, - "loss": 0.9638, - "step": 5974 - }, - { - "epoch": 2.1676190476190476, - "grad_norm": 0.1787109375, - "learning_rate": 5.619875168278057e-05, - "loss": 1.0009, - "step": 5976 - }, - { - "epoch": 2.168344671201814, - "grad_norm": 0.1875, - "learning_rate": 5.614979806633215e-05, - "loss": 1.0743, - "step": 5978 - }, - { - "epoch": 2.1690702947845804, - "grad_norm": 0.1640625, - "learning_rate": 5.610084444988374e-05, - "loss": 0.9951, - "step": 5980 - }, - { - "epoch": 2.1697959183673468, - "grad_norm": 0.1826171875, - "learning_rate": 5.605189083343533e-05, - "loss": 0.9734, - "step": 5982 - }, - { - "epoch": 2.170521541950113, - "grad_norm": 0.18359375, - "learning_rate": 5.6002937216986903e-05, - "loss": 1.0389, - "step": 5984 - }, - { - "epoch": 2.17124716553288, - "grad_norm": 0.1845703125, - "learning_rate": 5.595398360053849e-05, - "loss": 1.0822, - "step": 5986 - }, - { - "epoch": 2.1719727891156464, - "grad_norm": 0.1767578125, - "learning_rate": 5.5905029984090075e-05, - "loss": 0.9547, - "step": 5988 - }, - { - "epoch": 2.172698412698413, - "grad_norm": 0.15625, - "learning_rate": 5.5856076367641664e-05, - "loss": 0.9614, - "step": 5990 - }, - { - "epoch": 2.173424036281179, - "grad_norm": 0.177734375, - "learning_rate": 5.580712275119325e-05, - "loss": 0.9657, - "step": 5992 - }, - { - "epoch": 2.1741496598639456, - "grad_norm": 0.177734375, - "learning_rate": 5.575816913474483e-05, - "loss": 0.9888, - "step": 5994 - }, - { - "epoch": 2.174875283446712, - "grad_norm": 0.1572265625, - "learning_rate": 5.570921551829642e-05, - "loss": 0.9204, - "step": 5996 - }, - { - "epoch": 2.1756009070294784, - "grad_norm": 0.1748046875, - "learning_rate": 5.5660261901848e-05, - "loss": 1.0059, - "step": 5998 - }, - { - "epoch": 2.176326530612245, - "grad_norm": 0.1640625, - "learning_rate": 5.561130828539959e-05, - "loss": 1.0289, - "step": 6000 - }, - { - "epoch": 2.176326530612245, - "eval_loss": 1.0105350017547607, - "eval_runtime": 89.1474, - "eval_samples_per_second": 20.191, - "eval_steps_per_second": 0.639, - "step": 6000 - } - ], - "logging_steps": 2, - "max_steps": 8271, - "num_input_tokens_seen": 0, - "num_train_epochs": 3, - "save_steps": 100, - "stateful_callbacks": { - "EarlyStoppingCallback": { - "args": { - "early_stopping_patience": 3, - "early_stopping_threshold": 0.0 - }, - "attributes": { - "early_stopping_patience_counter": 1 - } - }, - "TrainerControl": { - "args": { - "should_epoch_stop": false, - "should_evaluate": false, - "should_log": false, - "should_save": true, - "should_training_stop": false - }, - "attributes": {} - } - }, - "total_flos": 1.0187451388598944e+19, - "train_batch_size": 2, - "trial_name": null, - "trial_params": null -}