{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 3099, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0048408568316592035, "grad_norm": 178.71523526371087, "learning_rate": 6.451612903225807e-07, "loss": 15.8295, "num_input_tokens_seen": 1971904, "step": 5, "train_runtime": 965.9569, "train_tokens_per_second": 2041.4 }, { "epoch": 0.009681713663318407, "grad_norm": 157.57466219746485, "learning_rate": 1.4516129032258066e-06, "loss": 15.3582, "num_input_tokens_seen": 4038576, "step": 10, "train_runtime": 1957.3451, "train_tokens_per_second": 2063.293 }, { "epoch": 0.014522570494977611, "grad_norm": 156.50950938238927, "learning_rate": 2.2580645161290324e-06, "loss": 12.8373, "num_input_tokens_seen": 5891616, "step": 15, "train_runtime": 2767.3168, "train_tokens_per_second": 2128.999 }, { "epoch": 0.019363427326636814, "grad_norm": 114.22743265072562, "learning_rate": 3.064516129032258e-06, "loss": 9.4205, "num_input_tokens_seen": 7849472, "step": 20, "train_runtime": 3522.7348, "train_tokens_per_second": 2228.232 }, { "epoch": 0.024204284158296018, "grad_norm": 135.40188397844682, "learning_rate": 3.870967741935484e-06, "loss": 7.6571, "num_input_tokens_seen": 9689648, "step": 25, "train_runtime": 3820.9748, "train_tokens_per_second": 2535.91 }, { "epoch": 0.029045140989955223, "grad_norm": 210.55574450271726, "learning_rate": 4.67741935483871e-06, "loss": 6.7016, "num_input_tokens_seen": 11652144, "step": 30, "train_runtime": 4196.5025, "train_tokens_per_second": 2776.632 }, { "epoch": 0.03388599782161442, "grad_norm": 208.41807492778292, "learning_rate": 5.483870967741936e-06, "loss": 5.8834, "num_input_tokens_seen": 13544176, "step": 35, "train_runtime": 4483.8305, "train_tokens_per_second": 3020.671 }, { "epoch": 0.03872685465327363, "grad_norm": 202.64781308250403, "learning_rate": 6.290322580645161e-06, "loss": 5.0909, "num_input_tokens_seen": 15315344, "step": 40, "train_runtime": 4739.2398, "train_tokens_per_second": 3231.604 }, { "epoch": 0.04356771148493283, "grad_norm": 208.02120960809935, "learning_rate": 7.096774193548387e-06, "loss": 4.3397, "num_input_tokens_seen": 17094640, "step": 45, "train_runtime": 4992.3854, "train_tokens_per_second": 3424.143 }, { "epoch": 0.048408568316592036, "grad_norm": 200.84606442004377, "learning_rate": 7.903225806451613e-06, "loss": 3.5344, "num_input_tokens_seen": 18990000, "step": 50, "train_runtime": 5321.5443, "train_tokens_per_second": 3568.513 }, { "epoch": 0.05324942514825124, "grad_norm": 180.4972165263306, "learning_rate": 8.70967741935484e-06, "loss": 2.5445, "num_input_tokens_seen": 20854912, "step": 55, "train_runtime": 5860.7159, "train_tokens_per_second": 3558.424 }, { "epoch": 0.058090281979910445, "grad_norm": 142.268336355056, "learning_rate": 9.516129032258064e-06, "loss": 1.5114, "num_input_tokens_seen": 22734480, "step": 60, "train_runtime": 6689.9051, "train_tokens_per_second": 3398.326 }, { "epoch": 0.06293113881156964, "grad_norm": 57.87038800361589, "learning_rate": 1.0322580645161291e-05, "loss": 0.681, "num_input_tokens_seen": 24705984, "step": 65, "train_runtime": 7695.5578, "train_tokens_per_second": 3210.421 }, { "epoch": 0.06777199564322885, "grad_norm": 11.382633199665975, "learning_rate": 1.1129032258064517e-05, "loss": 0.3083, "num_input_tokens_seen": 26623312, "step": 70, "train_runtime": 8523.5587, "train_tokens_per_second": 3123.497 }, { "epoch": 0.07261285247488805, "grad_norm": 1.7740955778794014, "learning_rate": 1.1935483870967743e-05, "loss": 0.2031, "num_input_tokens_seen": 28560464, "step": 75, "train_runtime": 9449.2243, "train_tokens_per_second": 3022.519 }, { "epoch": 0.07745370930654726, "grad_norm": 2.7185275952966643, "learning_rate": 1.2741935483870968e-05, "loss": 0.1776, "num_input_tokens_seen": 30353664, "step": 80, "train_runtime": 10242.2399, "train_tokens_per_second": 2963.577 }, { "epoch": 0.08229456613820646, "grad_norm": 11.173889555440166, "learning_rate": 1.3548387096774195e-05, "loss": 0.1556, "num_input_tokens_seen": 32276544, "step": 85, "train_runtime": 11127.6455, "train_tokens_per_second": 2900.573 }, { "epoch": 0.08713542296986566, "grad_norm": 0.43982142593021933, "learning_rate": 1.4354838709677421e-05, "loss": 0.1393, "num_input_tokens_seen": 34137648, "step": 90, "train_runtime": 11810.6178, "train_tokens_per_second": 2890.42 }, { "epoch": 0.09197627980152487, "grad_norm": 0.41292625320897214, "learning_rate": 1.5161290322580646e-05, "loss": 0.1302, "num_input_tokens_seen": 35998352, "step": 95, "train_runtime": 12094.7097, "train_tokens_per_second": 2976.372 }, { "epoch": 0.09681713663318407, "grad_norm": 0.2858556913615977, "learning_rate": 1.596774193548387e-05, "loss": 0.1201, "num_input_tokens_seen": 37869184, "step": 100, "train_runtime": 12394.8839, "train_tokens_per_second": 3055.227 }, { "epoch": 0.10165799346484328, "grad_norm": 0.30648690802872136, "learning_rate": 1.6774193548387098e-05, "loss": 0.1209, "num_input_tokens_seen": 39694272, "step": 105, "train_runtime": 12703.266, "train_tokens_per_second": 3124.73 }, { "epoch": 0.10649885029650248, "grad_norm": 0.2626276819380433, "learning_rate": 1.7580645161290325e-05, "loss": 0.1209, "num_input_tokens_seen": 41530000, "step": 110, "train_runtime": 12989.8073, "train_tokens_per_second": 3197.122 }, { "epoch": 0.11133970712816169, "grad_norm": 0.25596054422612646, "learning_rate": 1.838709677419355e-05, "loss": 0.1105, "num_input_tokens_seen": 43396048, "step": 115, "train_runtime": 13306.8176, "train_tokens_per_second": 3261.189 }, { "epoch": 0.11618056395982089, "grad_norm": 0.233381326668112, "learning_rate": 1.9193548387096774e-05, "loss": 0.1074, "num_input_tokens_seen": 45343968, "step": 120, "train_runtime": 14110.9269, "train_tokens_per_second": 3213.394 }, { "epoch": 0.1210214207914801, "grad_norm": 0.21251277353342254, "learning_rate": 2e-05, "loss": 0.1037, "num_input_tokens_seen": 47322080, "step": 125, "train_runtime": 15041.1565, "train_tokens_per_second": 3146.173 }, { "epoch": 0.12586227762313928, "grad_norm": 0.2127294082203629, "learning_rate": 2.0806451612903226e-05, "loss": 0.1062, "num_input_tokens_seen": 49256592, "step": 130, "train_runtime": 16039.8532, "train_tokens_per_second": 3070.888 }, { "epoch": 0.1307031344547985, "grad_norm": 0.37877204299340766, "learning_rate": 2.1612903225806454e-05, "loss": 0.1022, "num_input_tokens_seen": 51154720, "step": 135, "train_runtime": 16882.3542, "train_tokens_per_second": 3030.07 }, { "epoch": 0.1355439912864577, "grad_norm": 0.23637714784511715, "learning_rate": 2.2419354838709678e-05, "loss": 0.0953, "num_input_tokens_seen": 53264704, "step": 140, "train_runtime": 18022.3173, "train_tokens_per_second": 2955.486 }, { "epoch": 0.1403848481181169, "grad_norm": 0.25592037593094186, "learning_rate": 2.3225806451612906e-05, "loss": 0.1071, "num_input_tokens_seen": 55099456, "step": 145, "train_runtime": 18802.202, "train_tokens_per_second": 2930.479 }, { "epoch": 0.1452257049497761, "grad_norm": 0.24437969699285353, "learning_rate": 2.403225806451613e-05, "loss": 0.1, "num_input_tokens_seen": 57033824, "step": 150, "train_runtime": 19705.9937, "train_tokens_per_second": 2894.237 }, { "epoch": 0.15006656178143532, "grad_norm": 0.21535509944605138, "learning_rate": 2.4838709677419354e-05, "loss": 0.1024, "num_input_tokens_seen": 58876656, "step": 155, "train_runtime": 20511.3799, "train_tokens_per_second": 2870.439 }, { "epoch": 0.1549074186130945, "grad_norm": 0.2173650300325744, "learning_rate": 2.5645161290322582e-05, "loss": 0.0978, "num_input_tokens_seen": 60877568, "step": 160, "train_runtime": 21475.5689, "train_tokens_per_second": 2834.736 }, { "epoch": 0.15974827544475373, "grad_norm": 0.22756324955154553, "learning_rate": 2.645161290322581e-05, "loss": 0.1081, "num_input_tokens_seen": 62756608, "step": 165, "train_runtime": 22319.9154, "train_tokens_per_second": 2811.687 }, { "epoch": 0.16458913227641292, "grad_norm": 0.17479824979429293, "learning_rate": 2.7258064516129034e-05, "loss": 0.0934, "num_input_tokens_seen": 64688544, "step": 170, "train_runtime": 23223.6296, "train_tokens_per_second": 2785.462 }, { "epoch": 0.16942998910807214, "grad_norm": 0.18928070472014152, "learning_rate": 2.806451612903226e-05, "loss": 0.0956, "num_input_tokens_seen": 66503488, "step": 175, "train_runtime": 23940.3182, "train_tokens_per_second": 2777.887 }, { "epoch": 0.17427084593973133, "grad_norm": 0.19529886904869875, "learning_rate": 2.8870967741935483e-05, "loss": 0.0945, "num_input_tokens_seen": 68426064, "step": 180, "train_runtime": 24825.8993, "train_tokens_per_second": 2756.237 }, { "epoch": 0.17911170277139055, "grad_norm": 0.19453179688267283, "learning_rate": 2.967741935483871e-05, "loss": 0.0957, "num_input_tokens_seen": 70300224, "step": 185, "train_runtime": 25652.738, "train_tokens_per_second": 2740.457 }, { "epoch": 0.18395255960304974, "grad_norm": 0.21743752742438213, "learning_rate": 3.0483870967741935e-05, "loss": 0.1048, "num_input_tokens_seen": 72153584, "step": 190, "train_runtime": 26532.9803, "train_tokens_per_second": 2719.392 }, { "epoch": 0.18879341643470895, "grad_norm": 0.17245846000709703, "learning_rate": 3.1290322580645166e-05, "loss": 0.1038, "num_input_tokens_seen": 73989040, "step": 195, "train_runtime": 27354.3265, "train_tokens_per_second": 2704.839 }, { "epoch": 0.19363427326636815, "grad_norm": 0.18251260903855057, "learning_rate": 3.2096774193548393e-05, "loss": 0.106, "num_input_tokens_seen": 75844448, "step": 200, "train_runtime": 28180.8862, "train_tokens_per_second": 2691.344 }, { "epoch": 0.19847513009802736, "grad_norm": 0.17893448735702916, "learning_rate": 3.2903225806451614e-05, "loss": 0.0979, "num_input_tokens_seen": 77627520, "step": 205, "train_runtime": 28955.3576, "train_tokens_per_second": 2680.938 }, { "epoch": 0.20331598692968655, "grad_norm": 0.20201575024600307, "learning_rate": 3.370967741935484e-05, "loss": 0.0953, "num_input_tokens_seen": 79416736, "step": 210, "train_runtime": 29709.3856, "train_tokens_per_second": 2673.119 }, { "epoch": 0.20815684376134574, "grad_norm": 0.1752100891136059, "learning_rate": 3.451612903225806e-05, "loss": 0.0936, "num_input_tokens_seen": 81375920, "step": 215, "train_runtime": 30576.7229, "train_tokens_per_second": 2661.368 }, { "epoch": 0.21299770059300496, "grad_norm": 0.1691448592145883, "learning_rate": 3.532258064516129e-05, "loss": 0.0954, "num_input_tokens_seen": 83230304, "step": 220, "train_runtime": 31349.3154, "train_tokens_per_second": 2654.932 }, { "epoch": 0.21783855742466415, "grad_norm": 0.19940515973456982, "learning_rate": 3.612903225806452e-05, "loss": 0.095, "num_input_tokens_seen": 85271952, "step": 225, "train_runtime": 32261.937, "train_tokens_per_second": 2643.113 }, { "epoch": 0.22267941425632337, "grad_norm": 0.20741829321407812, "learning_rate": 3.6935483870967746e-05, "loss": 0.0898, "num_input_tokens_seen": 87088336, "step": 230, "train_runtime": 33023.2573, "train_tokens_per_second": 2637.182 }, { "epoch": 0.22752027108798256, "grad_norm": 0.19208547326415293, "learning_rate": 3.7741935483870974e-05, "loss": 0.1018, "num_input_tokens_seen": 88804864, "step": 235, "train_runtime": 33715.7377, "train_tokens_per_second": 2633.929 }, { "epoch": 0.23236112791964178, "grad_norm": 0.1778620676104918, "learning_rate": 3.8548387096774195e-05, "loss": 0.1003, "num_input_tokens_seen": 90678512, "step": 240, "train_runtime": 34461.3908, "train_tokens_per_second": 2631.307 }, { "epoch": 0.23720198475130097, "grad_norm": 0.19195813050583843, "learning_rate": 3.935483870967742e-05, "loss": 0.1012, "num_input_tokens_seen": 92516880, "step": 245, "train_runtime": 35260.5603, "train_tokens_per_second": 2623.806 }, { "epoch": 0.2420428415829602, "grad_norm": 0.1935559412076288, "learning_rate": 4.016129032258065e-05, "loss": 0.0944, "num_input_tokens_seen": 94387008, "step": 250, "train_runtime": 36080.7203, "train_tokens_per_second": 2615.996 }, { "epoch": 0.24688369841461938, "grad_norm": 0.1871264431371236, "learning_rate": 4.096774193548387e-05, "loss": 0.0952, "num_input_tokens_seen": 96291040, "step": 255, "train_runtime": 36929.8595, "train_tokens_per_second": 2607.403 }, { "epoch": 0.25172455524627857, "grad_norm": 0.16780208844120814, "learning_rate": 4.17741935483871e-05, "loss": 0.1029, "num_input_tokens_seen": 98122448, "step": 260, "train_runtime": 37715.1617, "train_tokens_per_second": 2601.671 }, { "epoch": 0.2565654120779378, "grad_norm": 0.15596752998410582, "learning_rate": 4.258064516129032e-05, "loss": 0.0955, "num_input_tokens_seen": 100020384, "step": 265, "train_runtime": 38487.5175, "train_tokens_per_second": 2598.775 }, { "epoch": 0.261406268909597, "grad_norm": 0.1887794985586671, "learning_rate": 4.3387096774193554e-05, "loss": 0.0966, "num_input_tokens_seen": 101811184, "step": 270, "train_runtime": 39201.4963, "train_tokens_per_second": 2597.125 }, { "epoch": 0.2662471257412562, "grad_norm": 0.21763986935352733, "learning_rate": 4.4193548387096775e-05, "loss": 0.0977, "num_input_tokens_seen": 103544960, "step": 275, "train_runtime": 39839.3963, "train_tokens_per_second": 2599.059 }, { "epoch": 0.2710879825729154, "grad_norm": 0.1422251923467383, "learning_rate": 4.5e-05, "loss": 0.0941, "num_input_tokens_seen": 105431200, "step": 280, "train_runtime": 40502.3645, "train_tokens_per_second": 2603.088 }, { "epoch": 0.27592883940457463, "grad_norm": 0.19060031607841202, "learning_rate": 4.580645161290323e-05, "loss": 0.0973, "num_input_tokens_seen": 107368704, "step": 285, "train_runtime": 40864.2265, "train_tokens_per_second": 2627.45 }, { "epoch": 0.2807696962362338, "grad_norm": 0.15547346024027353, "learning_rate": 4.661290322580645e-05, "loss": 0.0949, "num_input_tokens_seen": 109257200, "step": 290, "train_runtime": 41191.3176, "train_tokens_per_second": 2652.433 }, { "epoch": 0.285610553067893, "grad_norm": 0.18137966749300305, "learning_rate": 4.741935483870968e-05, "loss": 0.0953, "num_input_tokens_seen": 111150704, "step": 295, "train_runtime": 41496.46, "train_tokens_per_second": 2678.559 }, { "epoch": 0.2904514098995522, "grad_norm": 0.1726911372012115, "learning_rate": 4.822580645161291e-05, "loss": 0.097, "num_input_tokens_seen": 113034512, "step": 300, "train_runtime": 41825.6854, "train_tokens_per_second": 2702.514 }, { "epoch": 0.29529226673121145, "grad_norm": 0.14521468889499597, "learning_rate": 4.903225806451613e-05, "loss": 0.0897, "num_input_tokens_seen": 114937056, "step": 305, "train_runtime": 42162.0241, "train_tokens_per_second": 2726.08 }, { "epoch": 0.30013312356287064, "grad_norm": 0.15231521973585768, "learning_rate": 4.983870967741936e-05, "loss": 0.0887, "num_input_tokens_seen": 116955552, "step": 310, "train_runtime": 42539.7887, "train_tokens_per_second": 2749.321 }, { "epoch": 0.30497398039452983, "grad_norm": 0.1534116557864518, "learning_rate": 4.999974623485097e-05, "loss": 0.0981, "num_input_tokens_seen": 118830688, "step": 315, "train_runtime": 42871.2685, "train_tokens_per_second": 2771.802 }, { "epoch": 0.309814837226189, "grad_norm": 0.1686940018305674, "learning_rate": 4.999871532276243e-05, "loss": 0.0974, "num_input_tokens_seen": 120691504, "step": 320, "train_runtime": 43165.3478, "train_tokens_per_second": 2796.028 }, { "epoch": 0.3146556940578482, "grad_norm": 0.16433905364856274, "learning_rate": 4.999689143608857e-05, "loss": 0.0948, "num_input_tokens_seen": 122560384, "step": 325, "train_runtime": 43458.1159, "train_tokens_per_second": 2820.196 }, { "epoch": 0.31949655088950746, "grad_norm": 0.14174604664203966, "learning_rate": 4.999427463268419e-05, "loss": 0.0901, "num_input_tokens_seen": 124493200, "step": 330, "train_runtime": 43797.1282, "train_tokens_per_second": 2842.497 }, { "epoch": 0.32433740772116665, "grad_norm": 0.1695048202968491, "learning_rate": 4.999086499555591e-05, "loss": 0.09, "num_input_tokens_seen": 126353344, "step": 335, "train_runtime": 44105.2533, "train_tokens_per_second": 2864.814 }, { "epoch": 0.32917826455282584, "grad_norm": 0.17629027656795873, "learning_rate": 4.9986662632859504e-05, "loss": 0.0943, "num_input_tokens_seen": 128177040, "step": 340, "train_runtime": 44411.4535, "train_tokens_per_second": 2886.126 }, { "epoch": 0.33401912138448503, "grad_norm": 0.16367775596487932, "learning_rate": 4.9981667677896506e-05, "loss": 0.0921, "num_input_tokens_seen": 129970848, "step": 345, "train_runtime": 44670.9184, "train_tokens_per_second": 2909.518 }, { "epoch": 0.3388599782161443, "grad_norm": 0.14429236107461654, "learning_rate": 4.997588028910993e-05, "loss": 0.0933, "num_input_tokens_seen": 131934512, "step": 350, "train_runtime": 45023.4447, "train_tokens_per_second": 2930.351 }, { "epoch": 0.34370083504780347, "grad_norm": 0.14127408264764416, "learning_rate": 4.996930065007932e-05, "loss": 0.0919, "num_input_tokens_seen": 133828000, "step": 355, "train_runtime": 45336.3241, "train_tokens_per_second": 2951.893 }, { "epoch": 0.34854169187946266, "grad_norm": 0.1480021003298484, "learning_rate": 4.996192896951485e-05, "loss": 0.0894, "num_input_tokens_seen": 135655808, "step": 360, "train_runtime": 45602.4522, "train_tokens_per_second": 2974.748 }, { "epoch": 0.35338254871112185, "grad_norm": 0.16542232457655492, "learning_rate": 4.995376548125075e-05, "loss": 0.089, "num_input_tokens_seen": 137547952, "step": 365, "train_runtime": 45922.7124, "train_tokens_per_second": 2995.205 }, { "epoch": 0.3582234055427811, "grad_norm": 0.15136452715359272, "learning_rate": 4.994481044423788e-05, "loss": 0.0963, "num_input_tokens_seen": 139408624, "step": 370, "train_runtime": 46232.1379, "train_tokens_per_second": 3015.405 }, { "epoch": 0.3630642623744403, "grad_norm": 0.1569038436076336, "learning_rate": 4.99350641425355e-05, "loss": 0.098, "num_input_tokens_seen": 141146720, "step": 375, "train_runtime": 46495.1626, "train_tokens_per_second": 3035.729 }, { "epoch": 0.3679051192060995, "grad_norm": 0.15583043362247376, "learning_rate": 4.99245268853023e-05, "loss": 0.0927, "num_input_tokens_seen": 142875840, "step": 380, "train_runtime": 46725.0644, "train_tokens_per_second": 3057.799 }, { "epoch": 0.37274597603775866, "grad_norm": 0.17120656256693706, "learning_rate": 4.9913199006786515e-05, "loss": 0.0881, "num_input_tokens_seen": 144878720, "step": 385, "train_runtime": 47082.6128, "train_tokens_per_second": 3077.117 }, { "epoch": 0.3775868328694179, "grad_norm": 0.16312830261625755, "learning_rate": 4.9901080866315416e-05, "loss": 0.1036, "num_input_tokens_seen": 146604080, "step": 390, "train_runtime": 47321.9205, "train_tokens_per_second": 3098.016 }, { "epoch": 0.3824276897010771, "grad_norm": 0.11584886103538528, "learning_rate": 4.988817284828386e-05, "loss": 0.0879, "num_input_tokens_seen": 148505760, "step": 395, "train_runtime": 47625.9364, "train_tokens_per_second": 3118.17 }, { "epoch": 0.3872685465327363, "grad_norm": 0.1146886026314572, "learning_rate": 4.9874475362142076e-05, "loss": 0.0893, "num_input_tokens_seen": 150426000, "step": 400, "train_runtime": 47956.2272, "train_tokens_per_second": 3136.735 }, { "epoch": 0.3921094033643955, "grad_norm": 0.12910217232559762, "learning_rate": 4.9859988842382744e-05, "loss": 0.0909, "num_input_tokens_seen": 152243872, "step": 405, "train_runtime": 48242.6162, "train_tokens_per_second": 3155.796 }, { "epoch": 0.3969502601960547, "grad_norm": 0.15108152268102795, "learning_rate": 4.984471374852715e-05, "loss": 0.0947, "num_input_tokens_seen": 154052240, "step": 410, "train_runtime": 48527.6459, "train_tokens_per_second": 3174.525 }, { "epoch": 0.4017911170277139, "grad_norm": 0.14679850444524517, "learning_rate": 4.9828650565110644e-05, "loss": 0.0907, "num_input_tokens_seen": 156020992, "step": 415, "train_runtime": 48888.6576, "train_tokens_per_second": 3191.354 }, { "epoch": 0.4066319738593731, "grad_norm": 0.1327811810612439, "learning_rate": 4.981179980166726e-05, "loss": 0.09, "num_input_tokens_seen": 157936240, "step": 420, "train_runtime": 49230.7981, "train_tokens_per_second": 3208.078 }, { "epoch": 0.4114728306910323, "grad_norm": 0.14604735675352598, "learning_rate": 4.979416199271354e-05, "loss": 0.0903, "num_input_tokens_seen": 159733040, "step": 425, "train_runtime": 49491.6444, "train_tokens_per_second": 3227.475 }, { "epoch": 0.4163136875226915, "grad_norm": 0.14435246296113152, "learning_rate": 4.977573769773162e-05, "loss": 0.093, "num_input_tokens_seen": 161662848, "step": 430, "train_runtime": 49847.5614, "train_tokens_per_second": 3243.145 }, { "epoch": 0.42115454435435073, "grad_norm": 0.14601625994718231, "learning_rate": 4.975652750115143e-05, "loss": 0.0953, "num_input_tokens_seen": 163510512, "step": 435, "train_runtime": 50144.9198, "train_tokens_per_second": 3260.759 }, { "epoch": 0.4259954011860099, "grad_norm": 0.13966377448924783, "learning_rate": 4.973653201233219e-05, "loss": 0.0925, "num_input_tokens_seen": 165482800, "step": 440, "train_runtime": 50546.5604, "train_tokens_per_second": 3273.869 }, { "epoch": 0.4308362580176691, "grad_norm": 0.1417296980128132, "learning_rate": 4.971575186554307e-05, "loss": 0.0893, "num_input_tokens_seen": 167426896, "step": 445, "train_runtime": 50915.2604, "train_tokens_per_second": 3288.344 }, { "epoch": 0.4356771148493283, "grad_norm": 0.13565426401676745, "learning_rate": 4.969418771994309e-05, "loss": 0.0943, "num_input_tokens_seen": 169233824, "step": 450, "train_runtime": 51185.0637, "train_tokens_per_second": 3306.313 }, { "epoch": 0.44051797168098755, "grad_norm": 0.1454259863391591, "learning_rate": 4.967184025956015e-05, "loss": 0.0938, "num_input_tokens_seen": 171055232, "step": 455, "train_runtime": 51458.1503, "train_tokens_per_second": 3324.162 }, { "epoch": 0.44535882851264674, "grad_norm": 0.13730748863216408, "learning_rate": 4.9648710193269424e-05, "loss": 0.0915, "num_input_tokens_seen": 172893920, "step": 460, "train_runtime": 51744.3108, "train_tokens_per_second": 3341.313 }, { "epoch": 0.45019968534430593, "grad_norm": 0.12243741977778125, "learning_rate": 4.962479825477079e-05, "loss": 0.0849, "num_input_tokens_seen": 174782016, "step": 465, "train_runtime": 52060.903, "train_tokens_per_second": 3357.261 }, { "epoch": 0.4550405421759651, "grad_norm": 0.11854640411216295, "learning_rate": 4.9600105202565605e-05, "loss": 0.0886, "num_input_tokens_seen": 176671136, "step": 470, "train_runtime": 52381.9286, "train_tokens_per_second": 3372.75 }, { "epoch": 0.45988139900762437, "grad_norm": 0.14968248763796943, "learning_rate": 4.9574631819932646e-05, "loss": 0.0924, "num_input_tokens_seen": 178524848, "step": 475, "train_runtime": 52697.5978, "train_tokens_per_second": 3387.723 }, { "epoch": 0.46472225583928356, "grad_norm": 0.1435093288248635, "learning_rate": 4.9548378914903225e-05, "loss": 0.0872, "num_input_tokens_seen": 180401664, "step": 480, "train_runtime": 53014.3662, "train_tokens_per_second": 3402.883 }, { "epoch": 0.46956311267094275, "grad_norm": 0.15828157632384718, "learning_rate": 4.95213473202356e-05, "loss": 0.0948, "num_input_tokens_seen": 182234608, "step": 485, "train_runtime": 53294.1919, "train_tokens_per_second": 3419.408 }, { "epoch": 0.47440396950260194, "grad_norm": 0.11560281909563656, "learning_rate": 4.9493537893388544e-05, "loss": 0.0884, "num_input_tokens_seen": 184157296, "step": 490, "train_runtime": 53645.7615, "train_tokens_per_second": 3432.84 }, { "epoch": 0.4792448263342612, "grad_norm": 0.14531956629532314, "learning_rate": 4.9464951516494116e-05, "loss": 0.0916, "num_input_tokens_seen": 185978048, "step": 495, "train_runtime": 53924.698, "train_tokens_per_second": 3448.847 }, { "epoch": 0.4840856831659204, "grad_norm": 0.11828670163218845, "learning_rate": 4.943558909632975e-05, "loss": 0.0877, "num_input_tokens_seen": 187867312, "step": 500, "train_runtime": 54232.0578, "train_tokens_per_second": 3464.138 }, { "epoch": 0.48892653999757957, "grad_norm": 0.1315673057494712, "learning_rate": 4.94054515642894e-05, "loss": 0.0962, "num_input_tokens_seen": 189760400, "step": 505, "train_runtime": 54567.8241, "train_tokens_per_second": 3477.515 }, { "epoch": 0.49376739682923876, "grad_norm": 0.13676145087673666, "learning_rate": 4.937453987635408e-05, "loss": 0.084, "num_input_tokens_seen": 191773264, "step": 510, "train_runtime": 54988.129, "train_tokens_per_second": 3487.539 }, { "epoch": 0.498608253660898, "grad_norm": 0.1274509635019909, "learning_rate": 4.934285501306152e-05, "loss": 0.0868, "num_input_tokens_seen": 193750176, "step": 515, "train_runtime": 55385.4065, "train_tokens_per_second": 3498.217 }, { "epoch": 0.5034491104925571, "grad_norm": 0.12982283309243445, "learning_rate": 4.931039797947498e-05, "loss": 0.0895, "num_input_tokens_seen": 195664256, "step": 520, "train_runtime": 55687.4295, "train_tokens_per_second": 3513.616 }, { "epoch": 0.5082899673242164, "grad_norm": 0.13281239797421768, "learning_rate": 4.9277169805151496e-05, "loss": 0.0875, "num_input_tokens_seen": 197577536, "step": 525, "train_runtime": 56037.147, "train_tokens_per_second": 3525.831 }, { "epoch": 0.5131308241558756, "grad_norm": 0.127865250498207, "learning_rate": 4.924317154410915e-05, "loss": 0.0907, "num_input_tokens_seen": 199441664, "step": 530, "train_runtime": 56386.1137, "train_tokens_per_second": 3537.071 }, { "epoch": 0.5179716809875348, "grad_norm": 0.12535903140665275, "learning_rate": 4.9208404274793615e-05, "loss": 0.0851, "num_input_tokens_seen": 201395760, "step": 535, "train_runtime": 56725.1327, "train_tokens_per_second": 3550.38 }, { "epoch": 0.522812537819194, "grad_norm": 0.11973273742173074, "learning_rate": 4.917286910004402e-05, "loss": 0.0859, "num_input_tokens_seen": 203289168, "step": 540, "train_runtime": 57048.2676, "train_tokens_per_second": 3563.459 }, { "epoch": 0.5276533946508531, "grad_norm": 0.13617614252406723, "learning_rate": 4.9136567147057886e-05, "loss": 0.0932, "num_input_tokens_seen": 205086960, "step": 545, "train_runtime": 57320.4179, "train_tokens_per_second": 3577.904 }, { "epoch": 0.5324942514825124, "grad_norm": 0.12228835012992155, "learning_rate": 4.9099499567355465e-05, "loss": 0.0877, "num_input_tokens_seen": 207038560, "step": 550, "train_runtime": 57676.6952, "train_tokens_per_second": 3589.64 }, { "epoch": 0.5373351083141716, "grad_norm": 0.11616581415998756, "learning_rate": 4.9061667536743096e-05, "loss": 0.0849, "num_input_tokens_seen": 208994656, "step": 555, "train_runtime": 58017.1366, "train_tokens_per_second": 3602.292 }, { "epoch": 0.5421759651458308, "grad_norm": 0.1227132194885445, "learning_rate": 4.9023072255276e-05, "loss": 0.0924, "num_input_tokens_seen": 210860592, "step": 560, "train_runtime": 58355.6811, "train_tokens_per_second": 3613.369 }, { "epoch": 0.54701682197749, "grad_norm": 0.1331754681110147, "learning_rate": 4.89837149472202e-05, "loss": 0.0852, "num_input_tokens_seen": 212822400, "step": 565, "train_runtime": 58727.1386, "train_tokens_per_second": 3623.919 }, { "epoch": 0.5518576788091493, "grad_norm": 0.11555719706146601, "learning_rate": 4.894359686101363e-05, "loss": 0.0894, "num_input_tokens_seen": 214750096, "step": 570, "train_runtime": 59082.0703, "train_tokens_per_second": 3634.776 }, { "epoch": 0.5566985356408084, "grad_norm": 0.12553710069358454, "learning_rate": 4.890271926922661e-05, "loss": 0.0859, "num_input_tokens_seen": 216744256, "step": 575, "train_runtime": 59467.6278, "train_tokens_per_second": 3644.744 }, { "epoch": 0.5615393924724676, "grad_norm": 0.12296197236193043, "learning_rate": 4.886108346852142e-05, "loss": 0.0843, "num_input_tokens_seen": 218618416, "step": 580, "train_runtime": 59760.4082, "train_tokens_per_second": 3658.248 }, { "epoch": 0.5663802493041268, "grad_norm": 0.13282536365892234, "learning_rate": 4.881869077961119e-05, "loss": 0.0881, "num_input_tokens_seen": 220473312, "step": 585, "train_runtime": 60072.1334, "train_tokens_per_second": 3670.143 }, { "epoch": 0.571221106135786, "grad_norm": 0.1192404447399879, "learning_rate": 4.8775542547218035e-05, "loss": 0.0838, "num_input_tokens_seen": 222434400, "step": 590, "train_runtime": 60432.6507, "train_tokens_per_second": 3680.699 }, { "epoch": 0.5760619629674453, "grad_norm": 0.12137471825767925, "learning_rate": 4.873164014003035e-05, "loss": 0.0823, "num_input_tokens_seen": 224344864, "step": 595, "train_runtime": 60761.8507, "train_tokens_per_second": 3692.199 }, { "epoch": 0.5809028197991044, "grad_norm": 0.10630144681377657, "learning_rate": 4.868698495065942e-05, "loss": 0.0908, "num_input_tokens_seen": 226325424, "step": 600, "train_runtime": 61173.0901, "train_tokens_per_second": 3699.755 }, { "epoch": 0.5857436766307637, "grad_norm": 0.12903663659157497, "learning_rate": 4.8641578395595244e-05, "loss": 0.0841, "num_input_tokens_seen": 228223680, "step": 605, "train_runtime": 61472.6009, "train_tokens_per_second": 3712.608 }, { "epoch": 0.5905845334624229, "grad_norm": 0.1120445766713908, "learning_rate": 4.8595421915161606e-05, "loss": 0.09, "num_input_tokens_seen": 230051616, "step": 610, "train_runtime": 61773.5624, "train_tokens_per_second": 3724.111 }, { "epoch": 0.595425390294082, "grad_norm": 0.1403828596714754, "learning_rate": 4.854851697347038e-05, "loss": 0.0953, "num_input_tokens_seen": 231784640, "step": 615, "train_runtime": 62017.5435, "train_tokens_per_second": 3737.404 }, { "epoch": 0.6002662471257413, "grad_norm": 0.12026744008912688, "learning_rate": 4.8500865058375084e-05, "loss": 0.0818, "num_input_tokens_seen": 233624880, "step": 620, "train_runtime": 62301.1537, "train_tokens_per_second": 3749.929 }, { "epoch": 0.6051071039574004, "grad_norm": 0.13228231302304957, "learning_rate": 4.845246768142371e-05, "loss": 0.0919, "num_input_tokens_seen": 235580064, "step": 625, "train_runtime": 62644.2313, "train_tokens_per_second": 3760.603 }, { "epoch": 0.6099479607890597, "grad_norm": 0.11162603952629033, "learning_rate": 4.840332637781072e-05, "loss": 0.0806, "num_input_tokens_seen": 237579456, "step": 630, "train_runtime": 63013.8952, "train_tokens_per_second": 3770.271 }, { "epoch": 0.6147888176207189, "grad_norm": 0.13202350019993933, "learning_rate": 4.835344270632844e-05, "loss": 0.0894, "num_input_tokens_seen": 239379872, "step": 635, "train_runtime": 63306.9343, "train_tokens_per_second": 3781.258 }, { "epoch": 0.619629674452378, "grad_norm": 0.1262730134357675, "learning_rate": 4.830281824931751e-05, "loss": 0.0917, "num_input_tokens_seen": 241343936, "step": 640, "train_runtime": 63660.7557, "train_tokens_per_second": 3791.094 }, { "epoch": 0.6244705312840373, "grad_norm": 0.12514465647043127, "learning_rate": 4.825145461261677e-05, "loss": 0.0949, "num_input_tokens_seen": 243126480, "step": 645, "train_runtime": 63956.927, "train_tokens_per_second": 3801.41 }, { "epoch": 0.6293113881156964, "grad_norm": 0.11879787426434715, "learning_rate": 4.8199353425512276e-05, "loss": 0.0836, "num_input_tokens_seen": 245036080, "step": 650, "train_runtime": 64282.9428, "train_tokens_per_second": 3811.837 }, { "epoch": 0.6341522449473557, "grad_norm": 0.12807215290275195, "learning_rate": 4.814651634068565e-05, "loss": 0.0919, "num_input_tokens_seen": 246836544, "step": 655, "train_runtime": 64554.9622, "train_tokens_per_second": 3823.665 }, { "epoch": 0.6389931017790149, "grad_norm": 0.13318265083798486, "learning_rate": 4.809294503416164e-05, "loss": 0.0885, "num_input_tokens_seen": 248715200, "step": 660, "train_runtime": 64862.1325, "train_tokens_per_second": 3834.521 }, { "epoch": 0.643833958610674, "grad_norm": 0.11706198751336716, "learning_rate": 4.8038641205254945e-05, "loss": 0.0845, "num_input_tokens_seen": 250584112, "step": 665, "train_runtime": 65162.2083, "train_tokens_per_second": 3845.544 }, { "epoch": 0.6486748154423333, "grad_norm": 0.10849730614741822, "learning_rate": 4.7983606576516335e-05, "loss": 0.0834, "num_input_tokens_seen": 252446000, "step": 670, "train_runtime": 65447.5256, "train_tokens_per_second": 3857.228 }, { "epoch": 0.6535156722739925, "grad_norm": 0.11644681119826476, "learning_rate": 4.792784289367799e-05, "loss": 0.0889, "num_input_tokens_seen": 254336512, "step": 675, "train_runtime": 65807.5992, "train_tokens_per_second": 3864.85 }, { "epoch": 0.6583565291056517, "grad_norm": 0.12648798160840266, "learning_rate": 4.787135192559814e-05, "loss": 0.081, "num_input_tokens_seen": 256248240, "step": 680, "train_runtime": 66118.3408, "train_tokens_per_second": 3875.6 }, { "epoch": 0.6631973859373109, "grad_norm": 0.13179948151989662, "learning_rate": 4.781413546420494e-05, "loss": 0.0925, "num_input_tokens_seen": 258018928, "step": 685, "train_runtime": 66397.3901, "train_tokens_per_second": 3885.98 }, { "epoch": 0.6680382427689701, "grad_norm": 0.11355476378512105, "learning_rate": 4.775619532443965e-05, "loss": 0.081, "num_input_tokens_seen": 260040928, "step": 690, "train_runtime": 66781.9966, "train_tokens_per_second": 3893.878 }, { "epoch": 0.6728790996006293, "grad_norm": 0.1102920346549403, "learning_rate": 4.7697533344199054e-05, "loss": 0.081, "num_input_tokens_seen": 261986240, "step": 695, "train_runtime": 67112.4268, "train_tokens_per_second": 3903.692 }, { "epoch": 0.6777199564322886, "grad_norm": 0.11923784097224148, "learning_rate": 4.763815138427713e-05, "loss": 0.0834, "num_input_tokens_seen": 263808880, "step": 700, "train_runtime": 67389.9808, "train_tokens_per_second": 3914.66 }, { "epoch": 0.6825608132639477, "grad_norm": 0.13058866115819348, "learning_rate": 4.757805132830609e-05, "loss": 0.0858, "num_input_tokens_seen": 265662032, "step": 705, "train_runtime": 67694.3239, "train_tokens_per_second": 3924.436 }, { "epoch": 0.6874016700956069, "grad_norm": 0.10903443202471683, "learning_rate": 4.751723508269659e-05, "loss": 0.084, "num_input_tokens_seen": 267501344, "step": 710, "train_runtime": 68025.3342, "train_tokens_per_second": 3932.378 }, { "epoch": 0.6922425269272662, "grad_norm": 0.1065038057537401, "learning_rate": 4.745570457657722e-05, "loss": 0.0843, "num_input_tokens_seen": 269394176, "step": 715, "train_runtime": 68343.6923, "train_tokens_per_second": 3941.756 }, { "epoch": 0.6970833837589253, "grad_norm": 0.11228653074537234, "learning_rate": 4.7393461761733395e-05, "loss": 0.0911, "num_input_tokens_seen": 271288208, "step": 720, "train_runtime": 68705.6867, "train_tokens_per_second": 3948.555 }, { "epoch": 0.7019242405905846, "grad_norm": 0.10423741587634047, "learning_rate": 4.733050861254538e-05, "loss": 0.0759, "num_input_tokens_seen": 273149952, "step": 725, "train_runtime": 69004.9126, "train_tokens_per_second": 3958.413 }, { "epoch": 0.7067650974222437, "grad_norm": 0.12067320290718164, "learning_rate": 4.7266847125925686e-05, "loss": 0.0874, "num_input_tokens_seen": 275017888, "step": 730, "train_runtime": 69310.7545, "train_tokens_per_second": 3967.896 }, { "epoch": 0.7116059542539029, "grad_norm": 0.126476130474256, "learning_rate": 4.720247932125572e-05, "loss": 0.0877, "num_input_tokens_seen": 276821728, "step": 735, "train_runtime": 69607.9909, "train_tokens_per_second": 3976.867 }, { "epoch": 0.7164468110855622, "grad_norm": 0.14675215045590576, "learning_rate": 4.713740724032173e-05, "loss": 0.0854, "num_input_tokens_seen": 278648688, "step": 740, "train_runtime": 69890.2265, "train_tokens_per_second": 3986.948 }, { "epoch": 0.7212876679172213, "grad_norm": 0.10393986794688335, "learning_rate": 4.7071632947250056e-05, "loss": 0.0856, "num_input_tokens_seen": 280573744, "step": 745, "train_runtime": 70220.8361, "train_tokens_per_second": 3995.591 }, { "epoch": 0.7261285247488806, "grad_norm": 0.12245429208403327, "learning_rate": 4.700515852844161e-05, "loss": 0.088, "num_input_tokens_seen": 282386080, "step": 750, "train_runtime": 70532.5051, "train_tokens_per_second": 4003.63 }, { "epoch": 0.7309693815805397, "grad_norm": 0.15490079892082556, "learning_rate": 4.6937986092505734e-05, "loss": 0.0853, "num_input_tokens_seen": 284163392, "step": 755, "train_runtime": 70777.4141, "train_tokens_per_second": 4014.888 }, { "epoch": 0.735810238412199, "grad_norm": 0.1355824935484372, "learning_rate": 4.687011777019331e-05, "loss": 0.0897, "num_input_tokens_seen": 285979472, "step": 760, "train_runtime": 71065.8783, "train_tokens_per_second": 4024.146 }, { "epoch": 0.7406510952438582, "grad_norm": 0.10890370807480093, "learning_rate": 4.680155571432915e-05, "loss": 0.0866, "num_input_tokens_seen": 287803904, "step": 765, "train_runtime": 71342.8364, "train_tokens_per_second": 4034.097 }, { "epoch": 0.7454919520755173, "grad_norm": 0.10818940985882454, "learning_rate": 4.673230209974372e-05, "loss": 0.084, "num_input_tokens_seen": 289770544, "step": 770, "train_runtime": 71692.4983, "train_tokens_per_second": 4041.853 }, { "epoch": 0.7503328089071766, "grad_norm": 0.11393295664454559, "learning_rate": 4.666235912320416e-05, "loss": 0.0847, "num_input_tokens_seen": 291639680, "step": 775, "train_runtime": 71991.759, "train_tokens_per_second": 4051.015 }, { "epoch": 0.7551736657388358, "grad_norm": 0.10791548067564018, "learning_rate": 4.6591729003344604e-05, "loss": 0.0855, "num_input_tokens_seen": 293493632, "step": 780, "train_runtime": 72290.0262, "train_tokens_per_second": 4059.946 }, { "epoch": 0.760014522570495, "grad_norm": 0.12476206555821662, "learning_rate": 4.652041398059577e-05, "loss": 0.0874, "num_input_tokens_seen": 295416672, "step": 785, "train_runtime": 72613.7692, "train_tokens_per_second": 4068.329 }, { "epoch": 0.7648553794021542, "grad_norm": 0.11581041520648833, "learning_rate": 4.644841631711393e-05, "loss": 0.0917, "num_input_tokens_seen": 297230848, "step": 790, "train_runtime": 72898.9964, "train_tokens_per_second": 4077.297 }, { "epoch": 0.7696962362338133, "grad_norm": 0.1207889618919906, "learning_rate": 4.637573829670913e-05, "loss": 0.0862, "num_input_tokens_seen": 299220432, "step": 795, "train_runtime": 73281.0367, "train_tokens_per_second": 4083.19 }, { "epoch": 0.7745370930654726, "grad_norm": 0.13046576053931733, "learning_rate": 4.630238222477277e-05, "loss": 0.0876, "num_input_tokens_seen": 301103984, "step": 800, "train_runtime": 73600.0892, "train_tokens_per_second": 4091.082 }, { "epoch": 0.7793779498971318, "grad_norm": 0.11722404450994561, "learning_rate": 4.622835042820445e-05, "loss": 0.0832, "num_input_tokens_seen": 303004240, "step": 805, "train_runtime": 73903.033, "train_tokens_per_second": 4100.024 }, { "epoch": 0.784218806728791, "grad_norm": 0.11528092874791906, "learning_rate": 4.615364525533817e-05, "loss": 0.0875, "num_input_tokens_seen": 304776752, "step": 810, "train_runtime": 74158.7012, "train_tokens_per_second": 4109.791 }, { "epoch": 0.7890596635604502, "grad_norm": 0.1118105371174961, "learning_rate": 4.6078269075867844e-05, "loss": 0.0838, "num_input_tokens_seen": 306564704, "step": 815, "train_runtime": 74444.0338, "train_tokens_per_second": 4118.056 }, { "epoch": 0.7939005203921095, "grad_norm": 0.12405349561980368, "learning_rate": 4.600222428077212e-05, "loss": 0.0831, "num_input_tokens_seen": 308443424, "step": 820, "train_runtime": 74777.6015, "train_tokens_per_second": 4124.81 }, { "epoch": 0.7987413772237686, "grad_norm": 0.10166638613764127, "learning_rate": 4.592551328223854e-05, "loss": 0.0861, "num_input_tokens_seen": 310326496, "step": 825, "train_runtime": 75075.0708, "train_tokens_per_second": 4133.549 }, { "epoch": 0.8035822340554278, "grad_norm": 0.1304318839178686, "learning_rate": 4.584813851358705e-05, "loss": 0.0868, "num_input_tokens_seen": 312324656, "step": 830, "train_runtime": 75496.3515, "train_tokens_per_second": 4136.95 }, { "epoch": 0.808423090887087, "grad_norm": 0.11074066557269138, "learning_rate": 4.577010242919277e-05, "loss": 0.0823, "num_input_tokens_seen": 314264240, "step": 835, "train_runtime": 75859.0005, "train_tokens_per_second": 4142.742 }, { "epoch": 0.8132639477187462, "grad_norm": 0.1163799358603999, "learning_rate": 4.569140750440817e-05, "loss": 0.0823, "num_input_tokens_seen": 316091424, "step": 840, "train_runtime": 76138.2624, "train_tokens_per_second": 4151.545 }, { "epoch": 0.8181048045504055, "grad_norm": 0.11515834663049564, "learning_rate": 4.561205623548453e-05, "loss": 0.0816, "num_input_tokens_seen": 318022304, "step": 845, "train_runtime": 76454.3449, "train_tokens_per_second": 4159.637 }, { "epoch": 0.8229456613820646, "grad_norm": 0.10167337175748853, "learning_rate": 4.5532051139492784e-05, "loss": 0.0814, "num_input_tokens_seen": 319889920, "step": 850, "train_runtime": 76782.5683, "train_tokens_per_second": 4166.179 }, { "epoch": 0.8277865182137238, "grad_norm": 0.10009892012288468, "learning_rate": 4.545139475424366e-05, "loss": 0.0855, "num_input_tokens_seen": 321805040, "step": 855, "train_runtime": 77138.4208, "train_tokens_per_second": 4171.787 }, { "epoch": 0.832627375045383, "grad_norm": 0.10222131699800209, "learning_rate": 4.537008963820717e-05, "loss": 0.0859, "num_input_tokens_seen": 323627168, "step": 860, "train_runtime": 77426.607, "train_tokens_per_second": 4179.793 }, { "epoch": 0.8374682318770422, "grad_norm": 0.10643170835147328, "learning_rate": 4.5288138370431464e-05, "loss": 0.0866, "num_input_tokens_seen": 325450784, "step": 865, "train_runtime": 77714.3766, "train_tokens_per_second": 4187.781 }, { "epoch": 0.8423090887087015, "grad_norm": 0.1002014887379627, "learning_rate": 4.520554355046105e-05, "loss": 0.0794, "num_input_tokens_seen": 327331824, "step": 870, "train_runtime": 78022.0251, "train_tokens_per_second": 4195.377 }, { "epoch": 0.8471499455403606, "grad_norm": 0.12980276451699957, "learning_rate": 4.512230779825427e-05, "loss": 0.0853, "num_input_tokens_seen": 329167072, "step": 875, "train_runtime": 78294.26, "train_tokens_per_second": 4204.23 }, { "epoch": 0.8519908023720199, "grad_norm": 0.12209635825164061, "learning_rate": 4.503843375410024e-05, "loss": 0.0862, "num_input_tokens_seen": 331011792, "step": 880, "train_runtime": 78589.3442, "train_tokens_per_second": 4211.917 }, { "epoch": 0.8568316592036791, "grad_norm": 0.12814219418396117, "learning_rate": 4.49539240785351e-05, "loss": 0.0862, "num_input_tokens_seen": 332759152, "step": 885, "train_runtime": 78840.0264, "train_tokens_per_second": 4220.688 }, { "epoch": 0.8616725160353382, "grad_norm": 0.10182567961465826, "learning_rate": 4.4868781452257604e-05, "loss": 0.0854, "num_input_tokens_seen": 334548544, "step": 890, "train_runtime": 79107.638, "train_tokens_per_second": 4229.03 }, { "epoch": 0.8665133728669975, "grad_norm": 0.1059374934901486, "learning_rate": 4.478300857604407e-05, "loss": 0.081, "num_input_tokens_seen": 336480736, "step": 895, "train_runtime": 79448.3723, "train_tokens_per_second": 4235.212 }, { "epoch": 0.8713542296986566, "grad_norm": 0.11153870028757291, "learning_rate": 4.469660817066277e-05, "loss": 0.0799, "num_input_tokens_seen": 338427280, "step": 900, "train_runtime": 79789.1685, "train_tokens_per_second": 4241.519 }, { "epoch": 0.8761950865303159, "grad_norm": 0.13350987905341574, "learning_rate": 4.460958297678756e-05, "loss": 0.0828, "num_input_tokens_seen": 340350320, "step": 905, "train_runtime": 80140.7635, "train_tokens_per_second": 4246.906 }, { "epoch": 0.8810359433619751, "grad_norm": 0.10558711020322371, "learning_rate": 4.452193575491096e-05, "loss": 0.0815, "num_input_tokens_seen": 342226288, "step": 910, "train_runtime": 80448.6307, "train_tokens_per_second": 4253.973 }, { "epoch": 0.8858768001936342, "grad_norm": 0.1139445337081212, "learning_rate": 4.443366928525663e-05, "loss": 0.0813, "num_input_tokens_seen": 344202544, "step": 915, "train_runtime": 80804.3894, "train_tokens_per_second": 4259.701 }, { "epoch": 0.8907176570252935, "grad_norm": 0.12328404863523318, "learning_rate": 4.4344786367691125e-05, "loss": 0.0889, "num_input_tokens_seen": 346094384, "step": 920, "train_runtime": 81120.3268, "train_tokens_per_second": 4266.432 }, { "epoch": 0.8955585138569527, "grad_norm": 0.0973559450919178, "learning_rate": 4.425528982163512e-05, "loss": 0.0806, "num_input_tokens_seen": 348009840, "step": 925, "train_runtime": 81458.3407, "train_tokens_per_second": 4272.243 }, { "epoch": 0.9003993706886119, "grad_norm": 0.09888756400918433, "learning_rate": 4.416518248597395e-05, "loss": 0.0816, "num_input_tokens_seen": 349923616, "step": 930, "train_runtime": 81791.2966, "train_tokens_per_second": 4278.25 }, { "epoch": 0.9052402275202711, "grad_norm": 0.10273344580681865, "learning_rate": 4.4074467218967594e-05, "loss": 0.0808, "num_input_tokens_seen": 351982800, "step": 935, "train_runtime": 82224.6329, "train_tokens_per_second": 4280.746 }, { "epoch": 0.9100810843519302, "grad_norm": 0.1137919757065847, "learning_rate": 4.398314689815995e-05, "loss": 0.086, "num_input_tokens_seen": 353803568, "step": 940, "train_runtime": 82523.3997, "train_tokens_per_second": 4287.312 }, { "epoch": 0.9149219411835895, "grad_norm": 0.1129658121707594, "learning_rate": 4.389122442028762e-05, "loss": 0.0792, "num_input_tokens_seen": 355770400, "step": 945, "train_runtime": 82863.0664, "train_tokens_per_second": 4293.474 }, { "epoch": 0.9197627980152487, "grad_norm": 0.12295404647246533, "learning_rate": 4.379870270118801e-05, "loss": 0.0789, "num_input_tokens_seen": 357722704, "step": 950, "train_runtime": 83211.7587, "train_tokens_per_second": 4298.944 }, { "epoch": 0.9246036548469079, "grad_norm": 0.10702291365375034, "learning_rate": 4.370558467570678e-05, "loss": 0.0784, "num_input_tokens_seen": 359625392, "step": 955, "train_runtime": 83503.0123, "train_tokens_per_second": 4306.736 }, { "epoch": 0.9294445116785671, "grad_norm": 0.10671020858504487, "learning_rate": 4.361187329760483e-05, "loss": 0.0812, "num_input_tokens_seen": 361478880, "step": 960, "train_runtime": 83796.3829, "train_tokens_per_second": 4313.777 }, { "epoch": 0.9342853685102263, "grad_norm": 0.10376407616031118, "learning_rate": 4.351757153946456e-05, "loss": 0.0831, "num_input_tokens_seen": 363401552, "step": 965, "train_runtime": 84125.1709, "train_tokens_per_second": 4319.772 }, { "epoch": 0.9391262253418855, "grad_norm": 0.09840753061510307, "learning_rate": 4.3422682392595594e-05, "loss": 0.077, "num_input_tokens_seen": 365256960, "step": 970, "train_runtime": 84396.4737, "train_tokens_per_second": 4327.87 }, { "epoch": 0.9439670821735447, "grad_norm": 0.09244245286343136, "learning_rate": 4.332720886693987e-05, "loss": 0.0781, "num_input_tokens_seen": 367167264, "step": 975, "train_runtime": 84737.6756, "train_tokens_per_second": 4332.987 }, { "epoch": 0.9488079390052039, "grad_norm": 0.10909448781256045, "learning_rate": 4.32311539909762e-05, "loss": 0.084, "num_input_tokens_seen": 368976208, "step": 980, "train_runtime": 85107.1009, "train_tokens_per_second": 4335.434 }, { "epoch": 0.9536487958368631, "grad_norm": 0.10797216762579882, "learning_rate": 4.313452081162416e-05, "loss": 0.0829, "num_input_tokens_seen": 370862208, "step": 985, "train_runtime": 85937.0527, "train_tokens_per_second": 4315.51 }, { "epoch": 0.9584896526685224, "grad_norm": 0.09730637257808095, "learning_rate": 4.303731239414749e-05, "loss": 0.0859, "num_input_tokens_seen": 372719648, "step": 990, "train_runtime": 86738.8585, "train_tokens_per_second": 4297.032 }, { "epoch": 0.9633305095001815, "grad_norm": 0.10271769317718264, "learning_rate": 4.2939531822056815e-05, "loss": 0.0869, "num_input_tokens_seen": 374624944, "step": 995, "train_runtime": 87558.6912, "train_tokens_per_second": 4278.558 }, { "epoch": 0.9681713663318408, "grad_norm": 0.10573584364906888, "learning_rate": 4.284118219701187e-05, "loss": 0.0794, "num_input_tokens_seen": 376525888, "step": 1000, "train_runtime": 88323.819, "train_tokens_per_second": 4263.016 }, { "epoch": 0.9730122231634999, "grad_norm": 0.10710167799127884, "learning_rate": 4.2742266638723096e-05, "loss": 0.0804, "num_input_tokens_seen": 378537840, "step": 1005, "train_runtime": 89245.7806, "train_tokens_per_second": 4241.521 }, { "epoch": 0.9778530799951591, "grad_norm": 0.11197233290927573, "learning_rate": 4.264278828485267e-05, "loss": 0.0845, "num_input_tokens_seen": 380397552, "step": 1010, "train_runtime": 90080.3151, "train_tokens_per_second": 4222.871 }, { "epoch": 0.9826939368268184, "grad_norm": 0.11297920761471489, "learning_rate": 4.254275029091501e-05, "loss": 0.089, "num_input_tokens_seen": 382191440, "step": 1015, "train_runtime": 90923.6166, "train_tokens_per_second": 4203.434 }, { "epoch": 0.9875347936584775, "grad_norm": 0.11866337547317915, "learning_rate": 4.2442155830176655e-05, "loss": 0.0887, "num_input_tokens_seen": 384056752, "step": 1020, "train_runtime": 91754.463, "train_tokens_per_second": 4185.701 }, { "epoch": 0.9923756504901368, "grad_norm": 0.10420016956278523, "learning_rate": 4.2341008093555604e-05, "loss": 0.0833, "num_input_tokens_seen": 385974032, "step": 1025, "train_runtime": 92603.2825, "train_tokens_per_second": 4168.038 }, { "epoch": 0.997216507321796, "grad_norm": 0.11976582014616471, "learning_rate": 4.22393102895201e-05, "loss": 0.0807, "num_input_tokens_seen": 387884720, "step": 1030, "train_runtime": 93427.3132, "train_tokens_per_second": 4151.727 }, { "epoch": 1.0019363427326637, "grad_norm": 0.12104509350691006, "learning_rate": 4.213706564398688e-05, "loss": 0.0727, "num_input_tokens_seen": 389685952, "step": 1035, "train_runtime": 94255.5336, "train_tokens_per_second": 4134.356 }, { "epoch": 1.0067771995643229, "grad_norm": 0.10790601727390471, "learning_rate": 4.203427740021884e-05, "loss": 0.0713, "num_input_tokens_seen": 391624192, "step": 1040, "train_runtime": 95169.1567, "train_tokens_per_second": 4115.033 }, { "epoch": 1.011618056395982, "grad_norm": 0.09855811615223968, "learning_rate": 4.1930948818722104e-05, "loss": 0.0671, "num_input_tokens_seen": 393660384, "step": 1045, "train_runtime": 96151.6752, "train_tokens_per_second": 4094.16 }, { "epoch": 1.0164589132276414, "grad_norm": 0.10511029272951843, "learning_rate": 4.182708317714267e-05, "loss": 0.0686, "num_input_tokens_seen": 395529008, "step": 1050, "train_runtime": 96960.8327, "train_tokens_per_second": 4079.266 }, { "epoch": 1.0212997700593005, "grad_norm": 0.10834149804451805, "learning_rate": 4.172268377016241e-05, "loss": 0.0684, "num_input_tokens_seen": 397347360, "step": 1055, "train_runtime": 97712.5511, "train_tokens_per_second": 4066.493 }, { "epoch": 1.0261406268909596, "grad_norm": 0.09606298518108944, "learning_rate": 4.161775390939454e-05, "loss": 0.0664, "num_input_tokens_seen": 399294608, "step": 1060, "train_runtime": 98616.012, "train_tokens_per_second": 4048.984 }, { "epoch": 1.030981483722619, "grad_norm": 0.10742642762254552, "learning_rate": 4.151229692327863e-05, "loss": 0.0657, "num_input_tokens_seen": 401232320, "step": 1065, "train_runtime": 99525.5766, "train_tokens_per_second": 4031.449 }, { "epoch": 1.0358223405542781, "grad_norm": 0.099355578010755, "learning_rate": 4.1406316156974965e-05, "loss": 0.0668, "num_input_tokens_seen": 403156576, "step": 1070, "train_runtime": 100492.0641, "train_tokens_per_second": 4011.825 }, { "epoch": 1.0406631973859373, "grad_norm": 0.09750969581285164, "learning_rate": 4.1299814972258466e-05, "loss": 0.0685, "num_input_tokens_seen": 404973808, "step": 1075, "train_runtime": 101227.6948, "train_tokens_per_second": 4000.623 }, { "epoch": 1.0455040542175964, "grad_norm": 0.10948901174240346, "learning_rate": 4.1192796747412046e-05, "loss": 0.0739, "num_input_tokens_seen": 406856480, "step": 1080, "train_runtime": 102069.5624, "train_tokens_per_second": 3986.071 }, { "epoch": 1.0503449110492558, "grad_norm": 0.10315111502114589, "learning_rate": 4.108526487711944e-05, "loss": 0.0743, "num_input_tokens_seen": 408659088, "step": 1085, "train_runtime": 102905.3681, "train_tokens_per_second": 3971.213 }, { "epoch": 1.055185767880915, "grad_norm": 0.12136499885985239, "learning_rate": 4.0977222772357545e-05, "loss": 0.0674, "num_input_tokens_seen": 410533120, "step": 1090, "train_runtime": 103694.499, "train_tokens_per_second": 3959.064 }, { "epoch": 1.060026624712574, "grad_norm": 0.12011224246251123, "learning_rate": 4.0868673860288196e-05, "loss": 0.0723, "num_input_tokens_seen": 412377808, "step": 1095, "train_runtime": 104483.7209, "train_tokens_per_second": 3946.814 }, { "epoch": 1.0648674815442334, "grad_norm": 0.11631933894778271, "learning_rate": 4.075962158414948e-05, "loss": 0.0659, "num_input_tokens_seen": 414390304, "step": 1100, "train_runtime": 105480.5472, "train_tokens_per_second": 3928.595 }, { "epoch": 1.0697083383758925, "grad_norm": 0.10080944032903205, "learning_rate": 4.0650069403146484e-05, "loss": 0.0671, "num_input_tokens_seen": 416413616, "step": 1105, "train_runtime": 106510.3272, "train_tokens_per_second": 3909.608 }, { "epoch": 1.0745491952075517, "grad_norm": 0.09617342271272442, "learning_rate": 4.0540020792341605e-05, "loss": 0.0723, "num_input_tokens_seen": 418267408, "step": 1110, "train_runtime": 107384.6318, "train_tokens_per_second": 3895.04 }, { "epoch": 1.079390052039211, "grad_norm": 0.09814369288730117, "learning_rate": 4.0429479242544285e-05, "loss": 0.0676, "num_input_tokens_seen": 420122000, "step": 1115, "train_runtime": 108179.5693, "train_tokens_per_second": 3883.561 }, { "epoch": 1.0842309088708701, "grad_norm": 0.09566981500512636, "learning_rate": 4.031844826020028e-05, "loss": 0.0673, "num_input_tokens_seen": 421979504, "step": 1120, "train_runtime": 108984.7112, "train_tokens_per_second": 3871.915 }, { "epoch": 1.0890717657025293, "grad_norm": 0.11481624591628409, "learning_rate": 4.020693136728047e-05, "loss": 0.0716, "num_input_tokens_seen": 423852128, "step": 1125, "train_runtime": 109849.6222, "train_tokens_per_second": 3858.476 }, { "epoch": 1.0939126225341886, "grad_norm": 0.11408237955687323, "learning_rate": 4.0094932101169116e-05, "loss": 0.066, "num_input_tokens_seen": 425865744, "step": 1130, "train_runtime": 110829.9796, "train_tokens_per_second": 3842.514 }, { "epoch": 1.0987534793658478, "grad_norm": 0.09876683127617192, "learning_rate": 3.9982454014551626e-05, "loss": 0.0671, "num_input_tokens_seen": 427825520, "step": 1135, "train_runtime": 111678.4669, "train_tokens_per_second": 3830.868 }, { "epoch": 1.103594336197507, "grad_norm": 0.08998178840723907, "learning_rate": 3.9869500675301925e-05, "loss": 0.0659, "num_input_tokens_seen": 429650464, "step": 1140, "train_runtime": 112452.9235, "train_tokens_per_second": 3820.714 }, { "epoch": 1.1084351930291663, "grad_norm": 0.09112012767906727, "learning_rate": 3.975607566636921e-05, "loss": 0.0674, "num_input_tokens_seen": 431560992, "step": 1145, "train_runtime": 113300.5965, "train_tokens_per_second": 3808.991 }, { "epoch": 1.1132760498608254, "grad_norm": 0.09193978880262148, "learning_rate": 3.964218258566436e-05, "loss": 0.0679, "num_input_tokens_seen": 433452176, "step": 1150, "train_runtime": 114121.3844, "train_tokens_per_second": 3798.168 }, { "epoch": 1.1181169066924845, "grad_norm": 0.10253244394594298, "learning_rate": 3.952782504594574e-05, "loss": 0.0683, "num_input_tokens_seen": 435371936, "step": 1155, "train_runtime": 114932.6591, "train_tokens_per_second": 3788.061 }, { "epoch": 1.1229577635241437, "grad_norm": 0.11604949971136329, "learning_rate": 3.9413006674704684e-05, "loss": 0.0696, "num_input_tokens_seen": 437191424, "step": 1160, "train_runtime": 115724.1081, "train_tokens_per_second": 3777.877 }, { "epoch": 1.127798620355803, "grad_norm": 0.09600367492582994, "learning_rate": 3.929773111405034e-05, "loss": 0.0693, "num_input_tokens_seen": 439045440, "step": 1165, "train_runtime": 116553.6206, "train_tokens_per_second": 3766.897 }, { "epoch": 1.1326394771874622, "grad_norm": 0.10626811630147584, "learning_rate": 3.9182002020594235e-05, "loss": 0.0705, "num_input_tokens_seen": 440950928, "step": 1170, "train_runtime": 117462.4396, "train_tokens_per_second": 3753.974 }, { "epoch": 1.1374803340191213, "grad_norm": 0.09282379898506365, "learning_rate": 3.906582306533418e-05, "loss": 0.0638, "num_input_tokens_seen": 442914960, "step": 1175, "train_runtime": 118455.494, "train_tokens_per_second": 3739.083 }, { "epoch": 1.1423211908507807, "grad_norm": 0.09929776152503765, "learning_rate": 3.8949197933537916e-05, "loss": 0.0689, "num_input_tokens_seen": 444832176, "step": 1180, "train_runtime": 119337.8677, "train_tokens_per_second": 3727.502 }, { "epoch": 1.1471620476824398, "grad_norm": 0.1011962511269963, "learning_rate": 3.883213032462617e-05, "loss": 0.0682, "num_input_tokens_seen": 446780448, "step": 1185, "train_runtime": 120254.4758, "train_tokens_per_second": 3715.292 }, { "epoch": 1.152002904514099, "grad_norm": 0.09174519283988156, "learning_rate": 3.871462395205531e-05, "loss": 0.071, "num_input_tokens_seen": 448550288, "step": 1190, "train_runtime": 121004.0348, "train_tokens_per_second": 3706.904 }, { "epoch": 1.1568437613457583, "grad_norm": 0.10621267015218695, "learning_rate": 3.8596682543199546e-05, "loss": 0.068, "num_input_tokens_seen": 450450928, "step": 1195, "train_runtime": 121845.7134, "train_tokens_per_second": 3696.896 }, { "epoch": 1.1616846181774174, "grad_norm": 0.09664344676790702, "learning_rate": 3.847830983923273e-05, "loss": 0.0687, "num_input_tokens_seen": 452419728, "step": 1200, "train_runtime": 122865.555, "train_tokens_per_second": 3682.234 }, { "epoch": 1.1665254750090766, "grad_norm": 0.09223910440789163, "learning_rate": 3.835950959500963e-05, "loss": 0.0719, "num_input_tokens_seen": 454259472, "step": 1205, "train_runtime": 123677.075, "train_tokens_per_second": 3672.948 }, { "epoch": 1.1713663318407357, "grad_norm": 0.10736751066204356, "learning_rate": 3.8240285578946904e-05, "loss": 0.069, "num_input_tokens_seen": 456165264, "step": 1210, "train_runtime": 124627.8308, "train_tokens_per_second": 3660.22 }, { "epoch": 1.176207188672395, "grad_norm": 0.09115327170920039, "learning_rate": 3.812064157290345e-05, "loss": 0.0666, "num_input_tokens_seen": 458044704, "step": 1215, "train_runtime": 125470.063, "train_tokens_per_second": 3650.629 }, { "epoch": 1.1810480455040542, "grad_norm": 0.10235191270929159, "learning_rate": 3.800058137206053e-05, "loss": 0.0666, "num_input_tokens_seen": 459916000, "step": 1220, "train_runtime": 126329.2794, "train_tokens_per_second": 3640.613 }, { "epoch": 1.1858889023357135, "grad_norm": 0.11706669148345339, "learning_rate": 3.788010878480139e-05, "loss": 0.0721, "num_input_tokens_seen": 461786304, "step": 1225, "train_runtime": 127251.258, "train_tokens_per_second": 3628.933 }, { "epoch": 1.1907297591673727, "grad_norm": 0.09895227308048286, "learning_rate": 3.775922763259038e-05, "loss": 0.0661, "num_input_tokens_seen": 463676192, "step": 1230, "train_runtime": 128161.1571, "train_tokens_per_second": 3617.915 }, { "epoch": 1.1955706159990318, "grad_norm": 0.11484504551320268, "learning_rate": 3.763794174985181e-05, "loss": 0.0705, "num_input_tokens_seen": 465602880, "step": 1235, "train_runtime": 129003.0382, "train_tokens_per_second": 3609.24 }, { "epoch": 1.200411472830691, "grad_norm": 0.08994605167832181, "learning_rate": 3.7516254983848277e-05, "loss": 0.0698, "num_input_tokens_seen": 467493856, "step": 1240, "train_runtime": 129744.0548, "train_tokens_per_second": 3603.201 }, { "epoch": 1.2052523296623503, "grad_norm": 0.10300174734565927, "learning_rate": 3.7394171194558646e-05, "loss": 0.0693, "num_input_tokens_seen": 469496768, "step": 1245, "train_runtime": 130459.977, "train_tokens_per_second": 3598.78 }, { "epoch": 1.2100931864940094, "grad_norm": 0.10675389319087088, "learning_rate": 3.727169425455562e-05, "loss": 0.0693, "num_input_tokens_seen": 471449056, "step": 1250, "train_runtime": 131010.3565, "train_tokens_per_second": 3598.563 }, { "epoch": 1.2149340433256686, "grad_norm": 0.09921256550808967, "learning_rate": 3.7148828048882857e-05, "loss": 0.0697, "num_input_tokens_seen": 473264848, "step": 1255, "train_runtime": 131606.2869, "train_tokens_per_second": 3596.066 }, { "epoch": 1.219774900157328, "grad_norm": 0.10462373380428001, "learning_rate": 3.702557647493177e-05, "loss": 0.069, "num_input_tokens_seen": 475144208, "step": 1260, "train_runtime": 132107.4645, "train_tokens_per_second": 3596.649 }, { "epoch": 1.224615756988987, "grad_norm": 0.10415516508324393, "learning_rate": 3.690194344231789e-05, "loss": 0.0671, "num_input_tokens_seen": 477118864, "step": 1265, "train_runtime": 132532.0609, "train_tokens_per_second": 3600.026 }, { "epoch": 1.2294566138206462, "grad_norm": 0.10859351584240985, "learning_rate": 3.677793287275687e-05, "loss": 0.0697, "num_input_tokens_seen": 478984544, "step": 1270, "train_runtime": 132834.8843, "train_tokens_per_second": 3605.864 }, { "epoch": 1.2342974706523056, "grad_norm": 0.10820393055282612, "learning_rate": 3.665354869994003e-05, "loss": 0.073, "num_input_tokens_seen": 480862384, "step": 1275, "train_runtime": 133150.7075, "train_tokens_per_second": 3611.414 }, { "epoch": 1.2391383274839647, "grad_norm": 0.09874866443912496, "learning_rate": 3.652879486940965e-05, "loss": 0.071, "num_input_tokens_seen": 482657712, "step": 1280, "train_runtime": 133461.2075, "train_tokens_per_second": 3616.464 }, { "epoch": 1.2439791843156238, "grad_norm": 0.1063684888653233, "learning_rate": 3.640367533843376e-05, "loss": 0.0694, "num_input_tokens_seen": 484527216, "step": 1285, "train_runtime": 133797.6333, "train_tokens_per_second": 3621.344 }, { "epoch": 1.2488200411472832, "grad_norm": 0.11664187966473348, "learning_rate": 3.6278194075880625e-05, "loss": 0.0664, "num_input_tokens_seen": 486424320, "step": 1290, "train_runtime": 134098.3231, "train_tokens_per_second": 3627.371 }, { "epoch": 1.2536608979789423, "grad_norm": 0.08949284897230625, "learning_rate": 3.615235506209288e-05, "loss": 0.0686, "num_input_tokens_seen": 488454832, "step": 1295, "train_runtime": 134472.7384, "train_tokens_per_second": 3632.371 }, { "epoch": 1.2585017548106014, "grad_norm": 0.08869637799926437, "learning_rate": 3.602616228876123e-05, "loss": 0.0616, "num_input_tokens_seen": 490406800, "step": 1300, "train_runtime": 134802.021, "train_tokens_per_second": 3637.978 }, { "epoch": 1.2633426116422606, "grad_norm": 0.1058762928159983, "learning_rate": 3.589961975879787e-05, "loss": 0.0675, "num_input_tokens_seen": 492319664, "step": 1305, "train_runtime": 135130.0473, "train_tokens_per_second": 3643.303 }, { "epoch": 1.26818346847392, "grad_norm": 0.10187344379977778, "learning_rate": 3.577273148620946e-05, "loss": 0.0693, "num_input_tokens_seen": 494216816, "step": 1310, "train_runtime": 135461.9383, "train_tokens_per_second": 3648.381 }, { "epoch": 1.273024325305579, "grad_norm": 0.10989990396916802, "learning_rate": 3.564550149596985e-05, "loss": 0.0699, "num_input_tokens_seen": 496046624, "step": 1315, "train_runtime": 135771.2575, "train_tokens_per_second": 3653.547 }, { "epoch": 1.2778651821372382, "grad_norm": 0.10349545274745546, "learning_rate": 3.5517933823892384e-05, "loss": 0.0755, "num_input_tokens_seen": 497876768, "step": 1320, "train_runtime": 136087.2124, "train_tokens_per_second": 3658.513 }, { "epoch": 1.2827060389688976, "grad_norm": 0.09370049601860225, "learning_rate": 3.539003251650188e-05, "loss": 0.0786, "num_input_tokens_seen": 499699664, "step": 1325, "train_runtime": 136409.5926, "train_tokens_per_second": 3663.23 }, { "epoch": 1.2875468958005567, "grad_norm": 0.09622928750516618, "learning_rate": 3.526180163090627e-05, "loss": 0.0683, "num_input_tokens_seen": 501536176, "step": 1330, "train_runtime": 136698.4884, "train_tokens_per_second": 3668.923 }, { "epoch": 1.2923877526322158, "grad_norm": 0.09942000608374545, "learning_rate": 3.5133245234667883e-05, "loss": 0.0688, "num_input_tokens_seen": 503368832, "step": 1335, "train_runtime": 137002.8399, "train_tokens_per_second": 3674.149 }, { "epoch": 1.297228609463875, "grad_norm": 0.10312231028530282, "learning_rate": 3.50043674056745e-05, "loss": 0.0675, "num_input_tokens_seen": 505269776, "step": 1340, "train_runtime": 137326.8406, "train_tokens_per_second": 3679.323 }, { "epoch": 1.3020694662955343, "grad_norm": 0.10706010123105417, "learning_rate": 3.48751722320099e-05, "loss": 0.0646, "num_input_tokens_seen": 507187600, "step": 1345, "train_runtime": 137656.0465, "train_tokens_per_second": 3684.456 }, { "epoch": 1.3069103231271935, "grad_norm": 0.11190774683646115, "learning_rate": 3.4745663811824234e-05, "loss": 0.0701, "num_input_tokens_seen": 509000368, "step": 1350, "train_runtime": 137932.7138, "train_tokens_per_second": 3690.208 }, { "epoch": 1.3117511799588528, "grad_norm": 0.09559260030218834, "learning_rate": 3.461584625320407e-05, "loss": 0.0672, "num_input_tokens_seen": 510812240, "step": 1355, "train_runtime": 138211.6321, "train_tokens_per_second": 3695.87 }, { "epoch": 1.316592036790512, "grad_norm": 0.10592839240696805, "learning_rate": 3.448572367404197e-05, "loss": 0.0753, "num_input_tokens_seen": 512638608, "step": 1360, "train_runtime": 138535.165, "train_tokens_per_second": 3700.422 }, { "epoch": 1.321432893622171, "grad_norm": 0.10547925559174358, "learning_rate": 3.435530020190598e-05, "loss": 0.0666, "num_input_tokens_seen": 514458464, "step": 1365, "train_runtime": 138805.6716, "train_tokens_per_second": 3706.322 }, { "epoch": 1.3262737504538302, "grad_norm": 0.10826688240565788, "learning_rate": 3.422457997390865e-05, "loss": 0.0675, "num_input_tokens_seen": 516280880, "step": 1370, "train_runtime": 139079.93, "train_tokens_per_second": 3712.116 }, { "epoch": 1.3311146072854896, "grad_norm": 0.10770450004766066, "learning_rate": 3.4093567136575794e-05, "loss": 0.0726, "num_input_tokens_seen": 518154288, "step": 1375, "train_runtime": 139387.3129, "train_tokens_per_second": 3717.371 }, { "epoch": 1.3359554641171487, "grad_norm": 0.1142302841713288, "learning_rate": 3.396226584571499e-05, "loss": 0.0747, "num_input_tokens_seen": 519986144, "step": 1380, "train_runtime": 139690.3762, "train_tokens_per_second": 3722.419 }, { "epoch": 1.340796320948808, "grad_norm": 0.1002246744241518, "learning_rate": 3.383068026628371e-05, "loss": 0.07, "num_input_tokens_seen": 521862240, "step": 1385, "train_runtime": 139987.7519, "train_tokens_per_second": 3727.914 }, { "epoch": 1.3456371777804672, "grad_norm": 0.09964010260687982, "learning_rate": 3.3698814572257284e-05, "loss": 0.0665, "num_input_tokens_seen": 523858656, "step": 1390, "train_runtime": 140338.2763, "train_tokens_per_second": 3732.828 }, { "epoch": 1.3504780346121263, "grad_norm": 0.09090502017139336, "learning_rate": 3.356667294649639e-05, "loss": 0.0675, "num_input_tokens_seen": 525774560, "step": 1395, "train_runtime": 140671.1613, "train_tokens_per_second": 3737.614 }, { "epoch": 1.3553188914437855, "grad_norm": 0.10302633457655123, "learning_rate": 3.343425958061447e-05, "loss": 0.0662, "num_input_tokens_seen": 527681344, "step": 1400, "train_runtime": 140994.6669, "train_tokens_per_second": 3742.562 }, { "epoch": 1.3601597482754448, "grad_norm": 0.09470780975547861, "learning_rate": 3.330157867484472e-05, "loss": 0.0717, "num_input_tokens_seen": 529406736, "step": 1405, "train_runtime": 141236.0937, "train_tokens_per_second": 3748.381 }, { "epoch": 1.365000605107104, "grad_norm": 0.08585499478944604, "learning_rate": 3.3168634437906865e-05, "loss": 0.0662, "num_input_tokens_seen": 531305872, "step": 1410, "train_runtime": 141547.5703, "train_tokens_per_second": 3753.55 }, { "epoch": 1.369841461938763, "grad_norm": 0.10732846763853142, "learning_rate": 3.303543108687364e-05, "loss": 0.0715, "num_input_tokens_seen": 533037392, "step": 1415, "train_runtime": 141806.9679, "train_tokens_per_second": 3758.894 }, { "epoch": 1.3746823187704225, "grad_norm": 0.09819540974635672, "learning_rate": 3.290197284703707e-05, "loss": 0.0687, "num_input_tokens_seen": 535085280, "step": 1420, "train_runtime": 142243.4525, "train_tokens_per_second": 3761.757 }, { "epoch": 1.3795231756020816, "grad_norm": 0.09528337700443731, "learning_rate": 3.276826395177438e-05, "loss": 0.0728, "num_input_tokens_seen": 536966576, "step": 1425, "train_runtime": 142532.1226, "train_tokens_per_second": 3767.337 }, { "epoch": 1.3843640324337407, "grad_norm": 0.09402247417532383, "learning_rate": 3.263430864241376e-05, "loss": 0.068, "num_input_tokens_seen": 538829216, "step": 1430, "train_runtime": 142855.1544, "train_tokens_per_second": 3771.857 }, { "epoch": 1.3892048892653999, "grad_norm": 0.09890632658294073, "learning_rate": 3.250011116809978e-05, "loss": 0.0678, "num_input_tokens_seen": 540632144, "step": 1435, "train_runtime": 143161.7888, "train_tokens_per_second": 3776.372 }, { "epoch": 1.3940457460970592, "grad_norm": 0.08589478003550063, "learning_rate": 3.236567578565867e-05, "loss": 0.0693, "num_input_tokens_seen": 542432288, "step": 1440, "train_runtime": 143420.8223, "train_tokens_per_second": 3782.103 }, { "epoch": 1.3988866029287184, "grad_norm": 0.09609704465838043, "learning_rate": 3.223100675946321e-05, "loss": 0.0655, "num_input_tokens_seen": 544403440, "step": 1445, "train_runtime": 143809.3419, "train_tokens_per_second": 3785.592 }, { "epoch": 1.4037274597603777, "grad_norm": 0.08644550430034957, "learning_rate": 3.209610836129755e-05, "loss": 0.0657, "num_input_tokens_seen": 546192304, "step": 1450, "train_runtime": 144057.4772, "train_tokens_per_second": 3791.489 }, { "epoch": 1.4085683165920369, "grad_norm": 0.10119133528568032, "learning_rate": 3.1960984870221596e-05, "loss": 0.065, "num_input_tokens_seen": 548030912, "step": 1455, "train_runtime": 144334.6546, "train_tokens_per_second": 3796.946 }, { "epoch": 1.413409173423696, "grad_norm": 0.09755652421098708, "learning_rate": 3.1825640572435394e-05, "loss": 0.067, "num_input_tokens_seen": 549968368, "step": 1460, "train_runtime": 144647.7235, "train_tokens_per_second": 3802.123 }, { "epoch": 1.4182500302553551, "grad_norm": 0.09202887102207528, "learning_rate": 3.169007976114311e-05, "loss": 0.0633, "num_input_tokens_seen": 551997376, "step": 1465, "train_runtime": 145066.4375, "train_tokens_per_second": 3805.135 }, { "epoch": 1.4230908870870145, "grad_norm": 0.09893125478101567, "learning_rate": 3.155430673641681e-05, "loss": 0.069, "num_input_tokens_seen": 553815552, "step": 1470, "train_runtime": 145350.8539, "train_tokens_per_second": 3810.198 }, { "epoch": 1.4279317439186736, "grad_norm": 0.11231020588498852, "learning_rate": 3.1418325805060126e-05, "loss": 0.0713, "num_input_tokens_seen": 555719392, "step": 1475, "train_runtime": 145699.1273, "train_tokens_per_second": 3814.157 }, { "epoch": 1.4327726007503327, "grad_norm": 0.10568451565662428, "learning_rate": 3.1282141280471645e-05, "loss": 0.069, "num_input_tokens_seen": 557504016, "step": 1480, "train_runtime": 145974.2182, "train_tokens_per_second": 3819.195 }, { "epoch": 1.437613457581992, "grad_norm": 0.09645253967773308, "learning_rate": 3.114575748250801e-05, "loss": 0.0678, "num_input_tokens_seen": 559452496, "step": 1485, "train_runtime": 146328.064, "train_tokens_per_second": 3823.275 }, { "epoch": 1.4424543144136512, "grad_norm": 0.09105344674720453, "learning_rate": 3.100917873734696e-05, "loss": 0.0648, "num_input_tokens_seen": 561337136, "step": 1490, "train_runtime": 146656.3116, "train_tokens_per_second": 3827.569 }, { "epoch": 1.4472951712453104, "grad_norm": 0.09122782894664445, "learning_rate": 3.087240937735008e-05, "loss": 0.0682, "num_input_tokens_seen": 563237008, "step": 1495, "train_runtime": 146953.3037, "train_tokens_per_second": 3832.762 }, { "epoch": 1.4521360280769695, "grad_norm": 0.10094808912146574, "learning_rate": 3.073545374092535e-05, "loss": 0.07, "num_input_tokens_seen": 565056704, "step": 1500, "train_runtime": 147252.9238, "train_tokens_per_second": 3837.321 }, { "epoch": 1.4569768849086289, "grad_norm": 0.09898552497334528, "learning_rate": 3.05983161723896e-05, "loss": 0.0653, "num_input_tokens_seen": 566984576, "step": 1505, "train_runtime": 147575.0281, "train_tokens_per_second": 3842.009 }, { "epoch": 1.461817741740288, "grad_norm": 0.0893045303914291, "learning_rate": 3.046100102183061e-05, "loss": 0.0664, "num_input_tokens_seen": 568931872, "step": 1510, "train_runtime": 147890.4027, "train_tokens_per_second": 3846.983 }, { "epoch": 1.4666585985719474, "grad_norm": 0.07794027254932458, "learning_rate": 3.0323512644969194e-05, "loss": 0.0644, "num_input_tokens_seen": 570766304, "step": 1515, "train_runtime": 148153.9685, "train_tokens_per_second": 3852.521 }, { "epoch": 1.4714994554036065, "grad_norm": 0.09387987802719616, "learning_rate": 3.0185855403021013e-05, "loss": 0.0697, "num_input_tokens_seen": 572692384, "step": 1520, "train_runtime": 148490.2566, "train_tokens_per_second": 3856.767 }, { "epoch": 1.4763403122352656, "grad_norm": 0.10537208908005448, "learning_rate": 3.0048033662558222e-05, "loss": 0.0731, "num_input_tokens_seen": 574478256, "step": 1525, "train_runtime": 148778.7811, "train_tokens_per_second": 3861.292 }, { "epoch": 1.4811811690669248, "grad_norm": 0.09520737006467728, "learning_rate": 2.9910051795370974e-05, "loss": 0.068, "num_input_tokens_seen": 576337936, "step": 1530, "train_runtime": 149086.1454, "train_tokens_per_second": 3865.805 }, { "epoch": 1.4860220258985841, "grad_norm": 0.09478631735802089, "learning_rate": 2.977191417832874e-05, "loss": 0.0663, "num_input_tokens_seen": 578104656, "step": 1535, "train_runtime": 149330.6478, "train_tokens_per_second": 3871.306 }, { "epoch": 1.4908628827302433, "grad_norm": 0.10196404573493925, "learning_rate": 2.9633625193241475e-05, "loss": 0.0674, "num_input_tokens_seen": 579956016, "step": 1540, "train_runtime": 149621.4808, "train_tokens_per_second": 3876.155 }, { "epoch": 1.4957037395619024, "grad_norm": 0.10883093122161902, "learning_rate": 2.9495189226720615e-05, "loss": 0.0694, "num_input_tokens_seen": 581873728, "step": 1545, "train_runtime": 149975.2694, "train_tokens_per_second": 3879.798 }, { "epoch": 1.5005445963935617, "grad_norm": 0.09938067661414723, "learning_rate": 2.935661067003994e-05, "loss": 0.0689, "num_input_tokens_seen": 583722192, "step": 1550, "train_runtime": 150285.3755, "train_tokens_per_second": 3884.092 }, { "epoch": 1.5053854532252209, "grad_norm": 0.10557220766786837, "learning_rate": 2.9217893918996285e-05, "loss": 0.0711, "num_input_tokens_seen": 585531280, "step": 1555, "train_runtime": 150574.9739, "train_tokens_per_second": 3888.636 }, { "epoch": 1.51022631005688, "grad_norm": 0.09792652092850067, "learning_rate": 2.9079043373770088e-05, "loss": 0.0657, "num_input_tokens_seen": 587465120, "step": 1560, "train_runtime": 150916.7039, "train_tokens_per_second": 3892.645 }, { "epoch": 1.5150671668885392, "grad_norm": 0.10432359138697463, "learning_rate": 2.8940063438785808e-05, "loss": 0.065, "num_input_tokens_seen": 589312272, "step": 1565, "train_runtime": 151233.2823, "train_tokens_per_second": 3896.71 }, { "epoch": 1.5199080237201985, "grad_norm": 0.10464650554060877, "learning_rate": 2.8800958522572246e-05, "loss": 0.0669, "num_input_tokens_seen": 591224496, "step": 1570, "train_runtime": 151542.6129, "train_tokens_per_second": 3901.375 }, { "epoch": 1.5247488805518576, "grad_norm": 0.08982894126025302, "learning_rate": 2.866173303762268e-05, "loss": 0.0703, "num_input_tokens_seen": 593094000, "step": 1575, "train_runtime": 151870.8088, "train_tokens_per_second": 3905.253 }, { "epoch": 1.529589737383517, "grad_norm": 0.09170883604725243, "learning_rate": 2.8522391400254887e-05, "loss": 0.0678, "num_input_tokens_seen": 595038832, "step": 1580, "train_runtime": 152202.7319, "train_tokens_per_second": 3909.515 }, { "epoch": 1.5344305942151761, "grad_norm": 0.10514358953867606, "learning_rate": 2.8382938030471112e-05, "loss": 0.0697, "num_input_tokens_seen": 596824960, "step": 1585, "train_runtime": 152452.6889, "train_tokens_per_second": 3914.821 }, { "epoch": 1.5392714510468353, "grad_norm": 0.10505102469174953, "learning_rate": 2.8243377351817755e-05, "loss": 0.0689, "num_input_tokens_seen": 598705200, "step": 1590, "train_runtime": 152787.4796, "train_tokens_per_second": 3918.549 }, { "epoch": 1.5441123078784944, "grad_norm": 0.09749265072972622, "learning_rate": 2.8103713791245178e-05, "loss": 0.0612, "num_input_tokens_seen": 600600592, "step": 1595, "train_runtime": 153095.0077, "train_tokens_per_second": 3923.058 }, { "epoch": 1.5489531647101535, "grad_norm": 0.09875360040532327, "learning_rate": 2.7963951778967197e-05, "loss": 0.0681, "num_input_tokens_seen": 602389504, "step": 1600, "train_runtime": 153336.2956, "train_tokens_per_second": 3928.551 }, { "epoch": 1.553794021541813, "grad_norm": 0.0943003245484352, "learning_rate": 2.7824095748320568e-05, "loss": 0.0697, "num_input_tokens_seen": 604104288, "step": 1605, "train_runtime": 153582.7315, "train_tokens_per_second": 3933.413 }, { "epoch": 1.5586348783734723, "grad_norm": 0.08913383289373572, "learning_rate": 2.7684150135624376e-05, "loss": 0.0701, "num_input_tokens_seen": 605951504, "step": 1610, "train_runtime": 153880.8942, "train_tokens_per_second": 3937.796 }, { "epoch": 1.5634757352051314, "grad_norm": 0.09806492578910929, "learning_rate": 2.7544119380039314e-05, "loss": 0.0661, "num_input_tokens_seen": 607990208, "step": 1615, "train_runtime": 154299.517, "train_tokens_per_second": 3940.325 }, { "epoch": 1.5683165920367905, "grad_norm": 0.09468491378509171, "learning_rate": 2.740400792342685e-05, "loss": 0.0699, "num_input_tokens_seen": 609909232, "step": 1620, "train_runtime": 154619.4097, "train_tokens_per_second": 3944.584 }, { "epoch": 1.5731574488684497, "grad_norm": 0.11209084242527069, "learning_rate": 2.726382021020833e-05, "loss": 0.0718, "num_input_tokens_seen": 611736000, "step": 1625, "train_runtime": 154922.3873, "train_tokens_per_second": 3948.661 }, { "epoch": 1.5779983057001088, "grad_norm": 0.08630941935012056, "learning_rate": 2.7123560687224013e-05, "loss": 0.0665, "num_input_tokens_seen": 613559536, "step": 1630, "train_runtime": 155208.4781, "train_tokens_per_second": 3953.132 }, { "epoch": 1.5828391625317682, "grad_norm": 0.0933989919648124, "learning_rate": 2.6983233803592022e-05, "loss": 0.0664, "num_input_tokens_seen": 615391184, "step": 1635, "train_runtime": 155485.9117, "train_tokens_per_second": 3957.858 }, { "epoch": 1.5876800193634273, "grad_norm": 0.08973133951027497, "learning_rate": 2.6842844010567177e-05, "loss": 0.0734, "num_input_tokens_seen": 617305296, "step": 1640, "train_runtime": 155858.1926, "train_tokens_per_second": 3960.686 }, { "epoch": 1.5925208761950866, "grad_norm": 0.0933153070243224, "learning_rate": 2.6702395761399844e-05, "loss": 0.0674, "num_input_tokens_seen": 619107456, "step": 1645, "train_runtime": 156128.9329, "train_tokens_per_second": 3965.36 }, { "epoch": 1.5973617330267458, "grad_norm": 0.10662169623290256, "learning_rate": 2.656189351119463e-05, "loss": 0.0705, "num_input_tokens_seen": 620950304, "step": 1650, "train_runtime": 156466.5875, "train_tokens_per_second": 3968.581 }, { "epoch": 1.602202589858405, "grad_norm": 0.10372114619240043, "learning_rate": 2.6421341716769112e-05, "loss": 0.0692, "num_input_tokens_seen": 622796352, "step": 1655, "train_runtime": 156767.9586, "train_tokens_per_second": 3972.727 }, { "epoch": 1.607043446690064, "grad_norm": 0.09132496975982439, "learning_rate": 2.6280744836512423e-05, "loss": 0.0666, "num_input_tokens_seen": 624698480, "step": 1660, "train_runtime": 157088.7012, "train_tokens_per_second": 3976.724 }, { "epoch": 1.6118843035217232, "grad_norm": 0.09728380085896536, "learning_rate": 2.6140107330243858e-05, "loss": 0.0696, "num_input_tokens_seen": 626526800, "step": 1665, "train_runtime": 157381.2207, "train_tokens_per_second": 3980.95 }, { "epoch": 1.6167251603533825, "grad_norm": 0.10083788994513473, "learning_rate": 2.599943365907138e-05, "loss": 0.0708, "num_input_tokens_seen": 628340752, "step": 1670, "train_runtime": 157653.9021, "train_tokens_per_second": 3985.571 }, { "epoch": 1.621566017185042, "grad_norm": 0.09206012606935819, "learning_rate": 2.5858728285250156e-05, "loss": 0.0628, "num_input_tokens_seen": 630294736, "step": 1675, "train_runtime": 157993.6687, "train_tokens_per_second": 3989.367 }, { "epoch": 1.626406874016701, "grad_norm": 0.11407194887653826, "learning_rate": 2.5717995672040934e-05, "loss": 0.0668, "num_input_tokens_seen": 632181360, "step": 1680, "train_runtime": 158306.6385, "train_tokens_per_second": 3993.398 }, { "epoch": 1.6312477308483602, "grad_norm": 0.10532893116463045, "learning_rate": 2.5577240283568547e-05, "loss": 0.0672, "num_input_tokens_seen": 634139120, "step": 1685, "train_runtime": 158659.1124, "train_tokens_per_second": 3996.865 }, { "epoch": 1.6360885876800193, "grad_norm": 0.09253705708958208, "learning_rate": 2.5436466584680257e-05, "loss": 0.065, "num_input_tokens_seen": 636061136, "step": 1690, "train_runtime": 158996.3128, "train_tokens_per_second": 4000.477 }, { "epoch": 1.6409294445116784, "grad_norm": 0.10270850201788718, "learning_rate": 2.529567904080416e-05, "loss": 0.0679, "num_input_tokens_seen": 637901488, "step": 1695, "train_runtime": 159300.0136, "train_tokens_per_second": 4004.403 }, { "epoch": 1.6457703013433378, "grad_norm": 0.1020950381213344, "learning_rate": 2.5154882117807503e-05, "loss": 0.0674, "num_input_tokens_seen": 639783152, "step": 1700, "train_runtime": 159641.5796, "train_tokens_per_second": 4007.622 }, { "epoch": 1.650611158174997, "grad_norm": 0.10434335557369012, "learning_rate": 2.5014080281855062e-05, "loss": 0.0718, "num_input_tokens_seen": 641601136, "step": 1705, "train_runtime": 159948.6783, "train_tokens_per_second": 4011.294 }, { "epoch": 1.6554520150066563, "grad_norm": 0.09269120806836094, "learning_rate": 2.4873277999267443e-05, "loss": 0.0682, "num_input_tokens_seen": 643464448, "step": 1710, "train_runtime": 160228.7859, "train_tokens_per_second": 4015.91 }, { "epoch": 1.6602928718383154, "grad_norm": 0.09545551967743877, "learning_rate": 2.473247973637942e-05, "loss": 0.0668, "num_input_tokens_seen": 645403984, "step": 1715, "train_runtime": 160573.5465, "train_tokens_per_second": 4019.367 }, { "epoch": 1.6651337286699746, "grad_norm": 0.09245805131119629, "learning_rate": 2.459168995939827e-05, "loss": 0.063, "num_input_tokens_seen": 647359520, "step": 1720, "train_runtime": 160916.6879, "train_tokens_per_second": 4022.948 }, { "epoch": 1.6699745855016337, "grad_norm": 0.1130522027501556, "learning_rate": 2.4450913134262077e-05, "loss": 0.0698, "num_input_tokens_seen": 649221056, "step": 1725, "train_runtime": 161212.5554, "train_tokens_per_second": 4027.112 }, { "epoch": 1.674815442333293, "grad_norm": 0.1001979401813176, "learning_rate": 2.4310153726498107e-05, "loss": 0.0713, "num_input_tokens_seen": 651074128, "step": 1730, "train_runtime": 161567.7558, "train_tokens_per_second": 4029.728 }, { "epoch": 1.6796562991649522, "grad_norm": 0.09664199798231568, "learning_rate": 2.4169416201081102e-05, "loss": 0.0697, "num_input_tokens_seen": 652978208, "step": 1735, "train_runtime": 161942.97, "train_tokens_per_second": 4032.149 }, { "epoch": 1.6844971559966115, "grad_norm": 0.10551245269293205, "learning_rate": 2.4028705022291728e-05, "loss": 0.0661, "num_input_tokens_seen": 654910336, "step": 1740, "train_runtime": 162300.8039, "train_tokens_per_second": 4035.164 }, { "epoch": 1.6893380128282707, "grad_norm": 0.0926086339529829, "learning_rate": 2.3888024653574865e-05, "loss": 0.0654, "num_input_tokens_seen": 656834320, "step": 1745, "train_runtime": 162632.724, "train_tokens_per_second": 4038.759 }, { "epoch": 1.6941788696599298, "grad_norm": 0.0997990172837255, "learning_rate": 2.374737955739814e-05, "loss": 0.0675, "num_input_tokens_seen": 658744624, "step": 1750, "train_runtime": 162977.7639, "train_tokens_per_second": 4041.929 }, { "epoch": 1.699019726491589, "grad_norm": 0.09727233240178633, "learning_rate": 2.3606774195110256e-05, "loss": 0.0688, "num_input_tokens_seen": 660657152, "step": 1755, "train_runtime": 163311.2124, "train_tokens_per_second": 4045.388 }, { "epoch": 1.703860583323248, "grad_norm": 0.09186313556651109, "learning_rate": 2.346621302679957e-05, "loss": 0.0626, "num_input_tokens_seen": 662618160, "step": 1760, "train_runtime": 163650.2492, "train_tokens_per_second": 4048.99 }, { "epoch": 1.7087014401549074, "grad_norm": 0.09815622832004638, "learning_rate": 2.3325700511152572e-05, "loss": 0.0688, "num_input_tokens_seen": 664466176, "step": 1765, "train_runtime": 163980.8046, "train_tokens_per_second": 4052.097 }, { "epoch": 1.7135422969865666, "grad_norm": 0.09851131213295354, "learning_rate": 2.3185241105312435e-05, "loss": 0.0644, "num_input_tokens_seen": 666303008, "step": 1770, "train_runtime": 164258.4151, "train_tokens_per_second": 4056.431 }, { "epoch": 1.718383153818226, "grad_norm": 0.09650945128005722, "learning_rate": 2.3044839264737695e-05, "loss": 0.0654, "num_input_tokens_seen": 668249808, "step": 1775, "train_runtime": 164606.7007, "train_tokens_per_second": 4059.676 }, { "epoch": 1.723224010649885, "grad_norm": 0.07976228580722033, "learning_rate": 2.2904499443060852e-05, "loss": 0.0618, "num_input_tokens_seen": 670142528, "step": 1780, "train_runtime": 164916.446, "train_tokens_per_second": 4063.528 }, { "epoch": 1.7280648674815442, "grad_norm": 0.09316901052660637, "learning_rate": 2.2764226091947153e-05, "loss": 0.0705, "num_input_tokens_seen": 671892176, "step": 1785, "train_runtime": 165204.9592, "train_tokens_per_second": 4067.022 }, { "epoch": 1.7329057243132033, "grad_norm": 0.10554377156841822, "learning_rate": 2.2624023660953322e-05, "loss": 0.0694, "num_input_tokens_seen": 673790848, "step": 1790, "train_runtime": 165555.898, "train_tokens_per_second": 4069.869 }, { "epoch": 1.7377465811448627, "grad_norm": 0.1000010423520895, "learning_rate": 2.2483896597386506e-05, "loss": 0.0672, "num_input_tokens_seen": 675763984, "step": 1795, "train_runtime": 165926.0539, "train_tokens_per_second": 4072.682 }, { "epoch": 1.7425874379765218, "grad_norm": 0.09335668641339016, "learning_rate": 2.2343849346163092e-05, "loss": 0.0664, "num_input_tokens_seen": 677686256, "step": 1800, "train_runtime": 166242.0359, "train_tokens_per_second": 4076.504 }, { "epoch": 1.7474282948081812, "grad_norm": 0.09333896876761687, "learning_rate": 2.2203886349667826e-05, "loss": 0.067, "num_input_tokens_seen": 679548000, "step": 1805, "train_runtime": 166529.1496, "train_tokens_per_second": 4080.655 }, { "epoch": 1.7522691516398403, "grad_norm": 0.11018536147162415, "learning_rate": 2.2064012047612796e-05, "loss": 0.0674, "num_input_tokens_seen": 681323824, "step": 1810, "train_runtime": 166764.5916, "train_tokens_per_second": 4085.542 }, { "epoch": 1.7571100084714995, "grad_norm": 0.09634536795886925, "learning_rate": 2.1924230876896684e-05, "loss": 0.0639, "num_input_tokens_seen": 683289152, "step": 1815, "train_runtime": 167115.3972, "train_tokens_per_second": 4088.726 }, { "epoch": 1.7619508653031586, "grad_norm": 0.079098965621037, "learning_rate": 2.1784547271463966e-05, "loss": 0.063, "num_input_tokens_seen": 685248960, "step": 1820, "train_runtime": 167461.2987, "train_tokens_per_second": 4091.984 }, { "epoch": 1.7667917221348177, "grad_norm": 0.09262520602448539, "learning_rate": 2.164496566216428e-05, "loss": 0.0658, "num_input_tokens_seen": 687075264, "step": 1825, "train_runtime": 167728.4786, "train_tokens_per_second": 4096.354 }, { "epoch": 1.771632578966477, "grad_norm": 0.08425246771738228, "learning_rate": 2.150549047661193e-05, "loss": 0.0706, "num_input_tokens_seen": 688930992, "step": 1830, "train_runtime": 168041.5415, "train_tokens_per_second": 4099.766 }, { "epoch": 1.7764734357981364, "grad_norm": 0.08261309626975141, "learning_rate": 2.136612613904533e-05, "loss": 0.0633, "num_input_tokens_seen": 690736016, "step": 1835, "train_runtime": 168299.5399, "train_tokens_per_second": 4104.206 }, { "epoch": 1.7813142926297956, "grad_norm": 0.08439774632959215, "learning_rate": 2.1226877070186782e-05, "loss": 0.068, "num_input_tokens_seen": 692670496, "step": 1840, "train_runtime": 168637.0439, "train_tokens_per_second": 4107.463 }, { "epoch": 1.7861551494614547, "grad_norm": 0.09065583044936652, "learning_rate": 2.108774768710215e-05, "loss": 0.0643, "num_input_tokens_seen": 694613552, "step": 1845, "train_runtime": 168980.2344, "train_tokens_per_second": 4110.62 }, { "epoch": 1.7909960062931138, "grad_norm": 0.09792211257284066, "learning_rate": 2.0948742403060838e-05, "loss": 0.0667, "num_input_tokens_seen": 696469216, "step": 1850, "train_runtime": 169280.592, "train_tokens_per_second": 4114.289 }, { "epoch": 1.795836863124773, "grad_norm": 0.09897529531743958, "learning_rate": 2.0809865627395705e-05, "loss": 0.0681, "num_input_tokens_seen": 698418624, "step": 1855, "train_runtime": 169653.3494, "train_tokens_per_second": 4116.739 }, { "epoch": 1.8006777199564323, "grad_norm": 0.09500857629505935, "learning_rate": 2.0671121765363288e-05, "loss": 0.0647, "num_input_tokens_seen": 700384480, "step": 1860, "train_runtime": 170043.8506, "train_tokens_per_second": 4118.846 }, { "epoch": 1.8055185767880915, "grad_norm": 0.10247249755147927, "learning_rate": 2.0532515218003985e-05, "loss": 0.0699, "num_input_tokens_seen": 702304704, "step": 1865, "train_runtime": 170404.8081, "train_tokens_per_second": 4121.39 }, { "epoch": 1.8103594336197508, "grad_norm": 0.09643799643651368, "learning_rate": 2.039405038200252e-05, "loss": 0.0691, "num_input_tokens_seen": 704195280, "step": 1870, "train_runtime": 170715.849, "train_tokens_per_second": 4124.955 }, { "epoch": 1.81520029045141, "grad_norm": 0.09857814992539433, "learning_rate": 2.0255731649548427e-05, "loss": 0.0656, "num_input_tokens_seen": 706059712, "step": 1875, "train_runtime": 171033.6245, "train_tokens_per_second": 4128.192 }, { "epoch": 1.820041147283069, "grad_norm": 0.09418823224621559, "learning_rate": 2.011756340819673e-05, "loss": 0.0649, "num_input_tokens_seen": 708049376, "step": 1880, "train_runtime": 171420.7357, "train_tokens_per_second": 4130.477 }, { "epoch": 1.8248820041147282, "grad_norm": 0.10339589373318703, "learning_rate": 1.9979550040728838e-05, "loss": 0.0677, "num_input_tokens_seen": 709957648, "step": 1885, "train_runtime": 171751.8988, "train_tokens_per_second": 4133.623 }, { "epoch": 1.8297228609463874, "grad_norm": 0.09532254902656445, "learning_rate": 1.9841695925013406e-05, "loss": 0.0666, "num_input_tokens_seen": 711775040, "step": 1890, "train_runtime": 172045.6813, "train_tokens_per_second": 4137.128 }, { "epoch": 1.8345637177780467, "grad_norm": 0.0880459196194006, "learning_rate": 1.9704005433867555e-05, "loss": 0.0652, "num_input_tokens_seen": 713667296, "step": 1895, "train_runtime": 172364.1363, "train_tokens_per_second": 4140.463 }, { "epoch": 1.839404574609706, "grad_norm": 0.08969514913899801, "learning_rate": 1.9566482934918135e-05, "loss": 0.0653, "num_input_tokens_seen": 715572784, "step": 1900, "train_runtime": 172680.2824, "train_tokens_per_second": 4143.917 }, { "epoch": 1.8442454314413652, "grad_norm": 0.0913565853305896, "learning_rate": 1.9429132790463172e-05, "loss": 0.0623, "num_input_tokens_seen": 717403184, "step": 1905, "train_runtime": 172952.4861, "train_tokens_per_second": 4147.978 }, { "epoch": 1.8490862882730243, "grad_norm": 0.07640087073283276, "learning_rate": 1.9291959357333495e-05, "loss": 0.0652, "num_input_tokens_seen": 719188672, "step": 1910, "train_runtime": 173218.7348, "train_tokens_per_second": 4151.91 }, { "epoch": 1.8539271451046835, "grad_norm": 0.08816630513451665, "learning_rate": 1.9154966986754568e-05, "loss": 0.0656, "num_input_tokens_seen": 721142368, "step": 1915, "train_runtime": 173569.3058, "train_tokens_per_second": 4154.781 }, { "epoch": 1.8587680019363426, "grad_norm": 0.09071657241594651, "learning_rate": 1.9018160024208406e-05, "loss": 0.0673, "num_input_tokens_seen": 722942560, "step": 1920, "train_runtime": 173839.4391, "train_tokens_per_second": 4158.68 }, { "epoch": 1.863608858768002, "grad_norm": 0.10455029456382117, "learning_rate": 1.8881542809295804e-05, "loss": 0.0659, "num_input_tokens_seen": 724852000, "step": 1925, "train_runtime": 174168.4044, "train_tokens_per_second": 4161.788 }, { "epoch": 1.868449715599661, "grad_norm": 0.08954830475655517, "learning_rate": 1.874511967559861e-05, "loss": 0.0635, "num_input_tokens_seen": 726715728, "step": 1930, "train_runtime": 174457.6742, "train_tokens_per_second": 4165.57 }, { "epoch": 1.8732905724313205, "grad_norm": 0.09323369420411323, "learning_rate": 1.860889495054231e-05, "loss": 0.0675, "num_input_tokens_seen": 728548704, "step": 1935, "train_runtime": 174741.255, "train_tokens_per_second": 4169.3 }, { "epoch": 1.8781314292629796, "grad_norm": 0.09793727462020756, "learning_rate": 1.8472872955258764e-05, "loss": 0.065, "num_input_tokens_seen": 730456800, "step": 1940, "train_runtime": 175068.9574, "train_tokens_per_second": 4172.395 }, { "epoch": 1.8829722860946387, "grad_norm": 0.09583230673463962, "learning_rate": 1.8337058004449087e-05, "loss": 0.066, "num_input_tokens_seen": 732296016, "step": 1945, "train_runtime": 175374.5228, "train_tokens_per_second": 4175.612 }, { "epoch": 1.8878131429262979, "grad_norm": 0.09835179058080876, "learning_rate": 1.8201454406246853e-05, "loss": 0.0651, "num_input_tokens_seen": 734259984, "step": 1950, "train_runtime": 175740.5312, "train_tokens_per_second": 4178.091 }, { "epoch": 1.892653999757957, "grad_norm": 0.09108637153342643, "learning_rate": 1.8066066462081365e-05, "loss": 0.0626, "num_input_tokens_seen": 736134608, "step": 1955, "train_runtime": 176053.0546, "train_tokens_per_second": 4181.323 }, { "epoch": 1.8974948565896164, "grad_norm": 0.0919055897190125, "learning_rate": 1.7930898466541278e-05, "loss": 0.0643, "num_input_tokens_seen": 737950480, "step": 1960, "train_runtime": 176349.4191, "train_tokens_per_second": 4184.593 }, { "epoch": 1.9023357134212757, "grad_norm": 0.090159941398429, "learning_rate": 1.779595470723831e-05, "loss": 0.0615, "num_input_tokens_seen": 739837152, "step": 1965, "train_runtime": 176648.1234, "train_tokens_per_second": 4188.197 }, { "epoch": 1.9071765702529349, "grad_norm": 0.09346294098357426, "learning_rate": 1.7661239464671307e-05, "loss": 0.0665, "num_input_tokens_seen": 741721280, "step": 1970, "train_runtime": 176965.8973, "train_tokens_per_second": 4191.323 }, { "epoch": 1.912017427084594, "grad_norm": 0.10241633630793719, "learning_rate": 1.7526757012090384e-05, "loss": 0.0651, "num_input_tokens_seen": 743570432, "step": 1975, "train_runtime": 177269.1725, "train_tokens_per_second": 4194.584 }, { "epoch": 1.9168582839162531, "grad_norm": 0.09173533340918434, "learning_rate": 1.7392511615361454e-05, "loss": 0.0651, "num_input_tokens_seen": 745332208, "step": 1980, "train_runtime": 177532.7109, "train_tokens_per_second": 4198.281 }, { "epoch": 1.9216991407479123, "grad_norm": 0.1008623059871893, "learning_rate": 1.7258507532830843e-05, "loss": 0.0648, "num_input_tokens_seen": 747317616, "step": 1985, "train_runtime": 177898.9302, "train_tokens_per_second": 4200.799 }, { "epoch": 1.9265399975795716, "grad_norm": 0.089928131700152, "learning_rate": 1.7124749015190245e-05, "loss": 0.0646, "num_input_tokens_seen": 749180048, "step": 1990, "train_runtime": 178198.867, "train_tokens_per_second": 4204.18 }, { "epoch": 1.9313808544112308, "grad_norm": 0.08994654227932986, "learning_rate": 1.699124030534191e-05, "loss": 0.0629, "num_input_tokens_seen": 751039536, "step": 1995, "train_runtime": 178494.6526, "train_tokens_per_second": 4207.63 }, { "epoch": 1.93622171124289, "grad_norm": 0.08906162646309361, "learning_rate": 1.6857985638263994e-05, "loss": 0.0669, "num_input_tokens_seen": 752911840, "step": 2000, "train_runtime": 178774.0756, "train_tokens_per_second": 4211.527 }, { "epoch": 1.9410625680745492, "grad_norm": 0.0871511140825133, "learning_rate": 1.6724989240876302e-05, "loss": 0.0633, "num_input_tokens_seen": 754784096, "step": 2005, "train_runtime": 179064.7161, "train_tokens_per_second": 4215.147 }, { "epoch": 1.9459034249062084, "grad_norm": 0.09886328312958423, "learning_rate": 1.6592255331906127e-05, "loss": 0.0631, "num_input_tokens_seen": 756649472, "step": 2010, "train_runtime": 179383.4531, "train_tokens_per_second": 4218.056 }, { "epoch": 1.9507442817378675, "grad_norm": 0.09426446935032833, "learning_rate": 1.64597881217545e-05, "loss": 0.0611, "num_input_tokens_seen": 758536336, "step": 2015, "train_runtime": 179682.0449, "train_tokens_per_second": 4221.548 }, { "epoch": 1.9555851385695266, "grad_norm": 0.09397155441601662, "learning_rate": 1.632759181236258e-05, "loss": 0.0653, "num_input_tokens_seen": 760377888, "step": 2020, "train_runtime": 179989.7183, "train_tokens_per_second": 4224.563 }, { "epoch": 1.960425995401186, "grad_norm": 0.10368768607789577, "learning_rate": 1.61956705970784e-05, "loss": 0.0674, "num_input_tokens_seen": 762278528, "step": 2025, "train_runtime": 180338.7056, "train_tokens_per_second": 4226.927 }, { "epoch": 1.9652668522328454, "grad_norm": 0.0931039697726731, "learning_rate": 1.606402866052382e-05, "loss": 0.0666, "num_input_tokens_seen": 764061840, "step": 2030, "train_runtime": 180601.1349, "train_tokens_per_second": 4230.659 }, { "epoch": 1.9701077090645045, "grad_norm": 0.09153549712253775, "learning_rate": 1.593267017846185e-05, "loss": 0.0652, "num_input_tokens_seen": 765994112, "step": 2035, "train_runtime": 180938.3007, "train_tokens_per_second": 4233.455 }, { "epoch": 1.9749485658961636, "grad_norm": 0.0910008764482439, "learning_rate": 1.5801599317664106e-05, "loss": 0.0655, "num_input_tokens_seen": 767872592, "step": 2040, "train_runtime": 181245.9015, "train_tokens_per_second": 4236.634 }, { "epoch": 1.9797894227278228, "grad_norm": 0.09525002435832676, "learning_rate": 1.567082023577869e-05, "loss": 0.0643, "num_input_tokens_seen": 769751856, "step": 2045, "train_runtime": 181570.7397, "train_tokens_per_second": 4239.405 }, { "epoch": 1.984630279559482, "grad_norm": 0.08882860457825476, "learning_rate": 1.554033708119832e-05, "loss": 0.0614, "num_input_tokens_seen": 771662720, "step": 2050, "train_runtime": 181876.0901, "train_tokens_per_second": 4242.794 }, { "epoch": 1.9894711363911413, "grad_norm": 0.09856363331292063, "learning_rate": 1.5410153992928685e-05, "loss": 0.0637, "num_input_tokens_seen": 773536048, "step": 2055, "train_runtime": 182203.6654, "train_tokens_per_second": 4245.447 }, { "epoch": 1.9943119932228004, "grad_norm": 0.09210117173688946, "learning_rate": 1.528027510045723e-05, "loss": 0.0647, "num_input_tokens_seen": 775448176, "step": 2060, "train_runtime": 182525.7774, "train_tokens_per_second": 4248.431 }, { "epoch": 1.9991528500544598, "grad_norm": 0.09915830664554921, "learning_rate": 1.5150704523622066e-05, "loss": 0.0633, "num_input_tokens_seen": 777334320, "step": 2065, "train_runtime": 182819.8586, "train_tokens_per_second": 4251.914 }, { "epoch": 2.0038726854653275, "grad_norm": 0.08510468955619002, "learning_rate": 1.5021446372481408e-05, "loss": 0.0456, "num_input_tokens_seen": 779269280, "step": 2070, "train_runtime": 183203.0122, "train_tokens_per_second": 4253.583 }, { "epoch": 2.0087135422969866, "grad_norm": 0.11048348864635747, "learning_rate": 1.4892504747183078e-05, "loss": 0.0469, "num_input_tokens_seen": 781110080, "step": 2075, "train_runtime": 183493.0172, "train_tokens_per_second": 4256.893 }, { "epoch": 2.0135543991286458, "grad_norm": 0.08898938959054886, "learning_rate": 1.4763883737834538e-05, "loss": 0.0424, "num_input_tokens_seen": 783014480, "step": 2080, "train_runtime": 183815.8349, "train_tokens_per_second": 4259.777 }, { "epoch": 2.018395255960305, "grad_norm": 0.0919658947768258, "learning_rate": 1.4635587424373104e-05, "loss": 0.0432, "num_input_tokens_seen": 784918160, "step": 2085, "train_runtime": 184106.3305, "train_tokens_per_second": 4263.396 }, { "epoch": 2.023236112791964, "grad_norm": 0.09232021098074011, "learning_rate": 1.4507619876436548e-05, "loss": 0.0417, "num_input_tokens_seen": 786925872, "step": 2090, "train_runtime": 184467.9736, "train_tokens_per_second": 4265.921 }, { "epoch": 2.0280769696236236, "grad_norm": 0.09272252103462124, "learning_rate": 1.4379985153233983e-05, "loss": 0.0427, "num_input_tokens_seen": 788883568, "step": 2095, "train_runtime": 184855.3063, "train_tokens_per_second": 4267.573 }, { "epoch": 2.0329178264552827, "grad_norm": 0.08716923303447359, "learning_rate": 1.4252687303417125e-05, "loss": 0.0447, "num_input_tokens_seen": 790834704, "step": 2100, "train_runtime": 185222.062, "train_tokens_per_second": 4269.657 }, { "epoch": 2.037758683286942, "grad_norm": 0.09674916113938142, "learning_rate": 1.4125730364951867e-05, "loss": 0.044, "num_input_tokens_seen": 792644976, "step": 2105, "train_runtime": 185490.7671, "train_tokens_per_second": 4273.231 }, { "epoch": 2.042599540118601, "grad_norm": 0.1018255197888474, "learning_rate": 1.3999118364990172e-05, "loss": 0.0432, "num_input_tokens_seen": 794521776, "step": 2110, "train_runtime": 185814.5824, "train_tokens_per_second": 4275.885 }, { "epoch": 2.04744039695026, "grad_norm": 0.10641267828109918, "learning_rate": 1.387285531974235e-05, "loss": 0.0465, "num_input_tokens_seen": 796324144, "step": 2115, "train_runtime": 186093.9032, "train_tokens_per_second": 4279.152 }, { "epoch": 2.0522812537819193, "grad_norm": 0.08890413404478513, "learning_rate": 1.3746945234349651e-05, "loss": 0.042, "num_input_tokens_seen": 798176336, "step": 2120, "train_runtime": 186381.3552, "train_tokens_per_second": 4282.49 }, { "epoch": 2.0571221106135784, "grad_norm": 0.1007973258825116, "learning_rate": 1.3621392102757236e-05, "loss": 0.0454, "num_input_tokens_seen": 799981648, "step": 2125, "train_runtime": 186688.6329, "train_tokens_per_second": 4285.112 }, { "epoch": 2.061962967445238, "grad_norm": 0.11828945277047198, "learning_rate": 1.3496199907587426e-05, "loss": 0.0447, "num_input_tokens_seen": 801919920, "step": 2130, "train_runtime": 187027.4012, "train_tokens_per_second": 4287.714 }, { "epoch": 2.066803824276897, "grad_norm": 0.09697782214954971, "learning_rate": 1.337137262001349e-05, "loss": 0.0422, "num_input_tokens_seen": 803830048, "step": 2135, "train_runtime": 187371.5567, "train_tokens_per_second": 4290.032 }, { "epoch": 2.0716446811085563, "grad_norm": 0.08971480039994223, "learning_rate": 1.324691419963352e-05, "loss": 0.0414, "num_input_tokens_seen": 805707856, "step": 2140, "train_runtime": 187658.5794, "train_tokens_per_second": 4293.477 }, { "epoch": 2.0764855379402154, "grad_norm": 0.09701860651683053, "learning_rate": 1.3122828594345007e-05, "loss": 0.0445, "num_input_tokens_seen": 807627408, "step": 2145, "train_runtime": 187990.6441, "train_tokens_per_second": 4296.104 }, { "epoch": 2.0813263947718745, "grad_norm": 0.10086672081682901, "learning_rate": 1.2999119740219434e-05, "loss": 0.0462, "num_input_tokens_seen": 809504384, "step": 2150, "train_runtime": 188275.3974, "train_tokens_per_second": 4299.576 }, { "epoch": 2.0861672516035337, "grad_norm": 0.10444978322400048, "learning_rate": 1.2875791561377557e-05, "loss": 0.0458, "num_input_tokens_seen": 811379696, "step": 2155, "train_runtime": 188599.6081, "train_tokens_per_second": 4302.128 }, { "epoch": 2.091008108435193, "grad_norm": 0.10161748986408596, "learning_rate": 1.2752847969864857e-05, "loss": 0.0425, "num_input_tokens_seen": 813165728, "step": 2160, "train_runtime": 188873.2381, "train_tokens_per_second": 4305.352 }, { "epoch": 2.0958489652668524, "grad_norm": 0.09934783137965864, "learning_rate": 1.2630292865527483e-05, "loss": 0.0425, "num_input_tokens_seen": 815080720, "step": 2165, "train_runtime": 189177.9808, "train_tokens_per_second": 4308.539 }, { "epoch": 2.1006898220985115, "grad_norm": 0.09905949770742634, "learning_rate": 1.2508130135888518e-05, "loss": 0.0425, "num_input_tokens_seen": 817073936, "step": 2170, "train_runtime": 189630.3128, "train_tokens_per_second": 4308.773 }, { "epoch": 2.1055306789301707, "grad_norm": 0.09067487917664219, "learning_rate": 1.2386363656024691e-05, "loss": 0.0415, "num_input_tokens_seen": 818989136, "step": 2175, "train_runtime": 189949.6186, "train_tokens_per_second": 4311.612 }, { "epoch": 2.11037153576183, "grad_norm": 0.10406493878646128, "learning_rate": 1.226499728844345e-05, "loss": 0.0425, "num_input_tokens_seen": 820972848, "step": 2180, "train_runtime": 190334.3589, "train_tokens_per_second": 4313.319 }, { "epoch": 2.115212392593489, "grad_norm": 0.08540798243716224, "learning_rate": 1.2144034882960392e-05, "loss": 0.0398, "num_input_tokens_seen": 822903904, "step": 2185, "train_runtime": 190672.7343, "train_tokens_per_second": 4315.792 }, { "epoch": 2.120053249425148, "grad_norm": 0.11741293089028593, "learning_rate": 1.2023480276577267e-05, "loss": 0.0473, "num_input_tokens_seen": 824814944, "step": 2190, "train_runtime": 191027.2891, "train_tokens_per_second": 4317.786 }, { "epoch": 2.1248941062568076, "grad_norm": 0.09836368383563045, "learning_rate": 1.1903337293360101e-05, "loss": 0.042, "num_input_tokens_seen": 826652352, "step": 2195, "train_runtime": 191295.2085, "train_tokens_per_second": 4321.344 }, { "epoch": 2.1297349630884668, "grad_norm": 0.1091691815964545, "learning_rate": 1.1783609744318057e-05, "loss": 0.0419, "num_input_tokens_seen": 828587888, "step": 2200, "train_runtime": 191647.5762, "train_tokens_per_second": 4323.498 }, { "epoch": 2.134575819920126, "grad_norm": 0.0943679758427896, "learning_rate": 1.1664301427282418e-05, "loss": 0.0449, "num_input_tokens_seen": 830391184, "step": 2205, "train_runtime": 191919.3897, "train_tokens_per_second": 4326.771 }, { "epoch": 2.139416676751785, "grad_norm": 0.10612562433762897, "learning_rate": 1.154541612678619e-05, "loss": 0.0476, "num_input_tokens_seen": 832282752, "step": 2210, "train_runtime": 192247.4177, "train_tokens_per_second": 4329.227 }, { "epoch": 2.144257533583444, "grad_norm": 0.09834530066415154, "learning_rate": 1.1426957613944039e-05, "loss": 0.0457, "num_input_tokens_seen": 834136992, "step": 2215, "train_runtime": 192568.1333, "train_tokens_per_second": 4331.646 }, { "epoch": 2.1490983904151033, "grad_norm": 0.11010939625022646, "learning_rate": 1.1308929646332658e-05, "loss": 0.0417, "num_input_tokens_seen": 836052320, "step": 2220, "train_runtime": 192890.3051, "train_tokens_per_second": 4334.341 }, { "epoch": 2.153939247246763, "grad_norm": 0.0989548927712895, "learning_rate": 1.1191335967871582e-05, "loss": 0.0439, "num_input_tokens_seen": 837885568, "step": 2225, "train_runtime": 193169.477, "train_tokens_per_second": 4337.567 }, { "epoch": 2.158780104078422, "grad_norm": 0.10316511914427559, "learning_rate": 1.1074180308704429e-05, "loss": 0.0448, "num_input_tokens_seen": 839786048, "step": 2230, "train_runtime": 193482.1859, "train_tokens_per_second": 4340.379 }, { "epoch": 2.163620960910081, "grad_norm": 0.11447985050959848, "learning_rate": 1.0957466385080578e-05, "loss": 0.0462, "num_input_tokens_seen": 841636432, "step": 2235, "train_runtime": 193776.8505, "train_tokens_per_second": 4343.328 }, { "epoch": 2.1684618177417403, "grad_norm": 0.09641942816445281, "learning_rate": 1.0841197899237282e-05, "loss": 0.0435, "num_input_tokens_seen": 843588608, "step": 2240, "train_runtime": 194131.362, "train_tokens_per_second": 4345.452 }, { "epoch": 2.1733026745733994, "grad_norm": 0.08526832186648046, "learning_rate": 1.0725378539282246e-05, "loss": 0.0435, "num_input_tokens_seen": 845476688, "step": 2245, "train_runtime": 194465.4517, "train_tokens_per_second": 4347.696 }, { "epoch": 2.1781435314050586, "grad_norm": 0.09389266894701039, "learning_rate": 1.061001197907659e-05, "loss": 0.0426, "num_input_tokens_seen": 847341792, "step": 2250, "train_runtime": 194755.117, "train_tokens_per_second": 4350.806 }, { "epoch": 2.1829843882367177, "grad_norm": 0.1091155029328961, "learning_rate": 1.0495101878118408e-05, "loss": 0.0438, "num_input_tokens_seen": 849140496, "step": 2255, "train_runtime": 195018.4821, "train_tokens_per_second": 4354.154 }, { "epoch": 2.1878252450683773, "grad_norm": 0.10639940814249475, "learning_rate": 1.0380651881426567e-05, "loss": 0.0421, "num_input_tokens_seen": 850942480, "step": 2260, "train_runtime": 195260.0607, "train_tokens_per_second": 4357.996 }, { "epoch": 2.1926661019000364, "grad_norm": 0.10716301854067384, "learning_rate": 1.0266665619425181e-05, "loss": 0.0441, "num_input_tokens_seen": 852750800, "step": 2265, "train_runtime": 195553.3537, "train_tokens_per_second": 4360.707 }, { "epoch": 2.1975069587316955, "grad_norm": 0.10343872393389002, "learning_rate": 1.0153146707828404e-05, "loss": 0.0427, "num_input_tokens_seen": 854703456, "step": 2270, "train_runtime": 195905.746, "train_tokens_per_second": 4362.83 }, { "epoch": 2.2023478155633547, "grad_norm": 0.09659246514397754, "learning_rate": 1.004009874752575e-05, "loss": 0.0438, "num_input_tokens_seen": 856635792, "step": 2275, "train_runtime": 196280.3262, "train_tokens_per_second": 4364.349 }, { "epoch": 2.207188672395014, "grad_norm": 0.10837299321798222, "learning_rate": 9.927525324467866e-06, "loss": 0.0435, "num_input_tokens_seen": 858544176, "step": 2280, "train_runtime": 196631.0128, "train_tokens_per_second": 4366.27 }, { "epoch": 2.212029529226673, "grad_norm": 0.0935449758044754, "learning_rate": 9.81543000955279e-06, "loss": 0.0412, "num_input_tokens_seen": 860409424, "step": 2285, "train_runtime": 196926.666, "train_tokens_per_second": 4369.187 }, { "epoch": 2.2168703860583325, "grad_norm": 0.08817931437627177, "learning_rate": 9.703816358512674e-06, "loss": 0.0415, "num_input_tokens_seen": 862320048, "step": 2290, "train_runtime": 197238.6013, "train_tokens_per_second": 4371.964 }, { "epoch": 2.2217112428899917, "grad_norm": 0.0980673558112737, "learning_rate": 9.592687911801001e-06, "loss": 0.0442, "num_input_tokens_seen": 864243216, "step": 2295, "train_runtime": 197594.875, "train_tokens_per_second": 4373.814 }, { "epoch": 2.226552099721651, "grad_norm": 0.10853728697423164, "learning_rate": 9.482048194480275e-06, "loss": 0.0425, "num_input_tokens_seen": 866168848, "step": 2300, "train_runtime": 197933.6478, "train_tokens_per_second": 4376.057 }, { "epoch": 2.23139295655331, "grad_norm": 0.11490821008247167, "learning_rate": 9.371900716110183e-06, "loss": 0.0431, "num_input_tokens_seen": 868041296, "step": 2305, "train_runtime": 198239.5933, "train_tokens_per_second": 4378.748 }, { "epoch": 2.236233813384969, "grad_norm": 0.08947053295422262, "learning_rate": 9.262248970636334e-06, "loss": 0.046, "num_input_tokens_seen": 869878384, "step": 2310, "train_runtime": 198532.6139, "train_tokens_per_second": 4381.539 }, { "epoch": 2.241074670216628, "grad_norm": 0.10485732709538129, "learning_rate": 9.153096436279327e-06, "loss": 0.042, "num_input_tokens_seen": 871688688, "step": 2315, "train_runtime": 198817.926, "train_tokens_per_second": 4384.357 }, { "epoch": 2.2459155270482873, "grad_norm": 0.11044109934649306, "learning_rate": 9.044446575424548e-06, "loss": 0.0431, "num_input_tokens_seen": 873728640, "step": 2320, "train_runtime": 199228.9285, "train_tokens_per_second": 4385.551 }, { "epoch": 2.250756383879947, "grad_norm": 0.09950408440238417, "learning_rate": 8.936302834512206e-06, "loss": 0.0439, "num_input_tokens_seen": 875601696, "step": 2325, "train_runtime": 199550.3994, "train_tokens_per_second": 4387.872 }, { "epoch": 2.255597240711606, "grad_norm": 0.09167335947382782, "learning_rate": 8.828668643928112e-06, "loss": 0.0419, "num_input_tokens_seen": 877559760, "step": 2330, "train_runtime": 199908.7951, "train_tokens_per_second": 4389.801 }, { "epoch": 2.260438097543265, "grad_norm": 0.10598202575372626, "learning_rate": 8.721547417894816e-06, "loss": 0.0432, "num_input_tokens_seen": 879392576, "step": 2335, "train_runtime": 200216.2117, "train_tokens_per_second": 4392.215 }, { "epoch": 2.2652789543749243, "grad_norm": 0.10543835687376972, "learning_rate": 8.614942554363326e-06, "loss": 0.0453, "num_input_tokens_seen": 881180800, "step": 2340, "train_runtime": 200501.3334, "train_tokens_per_second": 4394.887 }, { "epoch": 2.2701198112065835, "grad_norm": 0.10496418814685284, "learning_rate": 8.508857434905304e-06, "loss": 0.0423, "num_input_tokens_seen": 883048928, "step": 2345, "train_runtime": 200822.2237, "train_tokens_per_second": 4397.167 }, { "epoch": 2.2749606680382426, "grad_norm": 0.09577382338976835, "learning_rate": 8.40329542460582e-06, "loss": 0.0424, "num_input_tokens_seen": 885000208, "step": 2350, "train_runtime": 201157.0354, "train_tokens_per_second": 4399.549 }, { "epoch": 2.279801524869902, "grad_norm": 0.10452053840305, "learning_rate": 8.298259871956596e-06, "loss": 0.0432, "num_input_tokens_seen": 886939040, "step": 2355, "train_runtime": 201495.1483, "train_tokens_per_second": 4401.789 }, { "epoch": 2.2846423817015613, "grad_norm": 0.08543950174023542, "learning_rate": 8.193754108749793e-06, "loss": 0.0408, "num_input_tokens_seen": 888957616, "step": 2360, "train_runtime": 201854.2624, "train_tokens_per_second": 4403.958 }, { "epoch": 2.2894832385332204, "grad_norm": 0.10273943575259807, "learning_rate": 8.08978144997234e-06, "loss": 0.0427, "num_input_tokens_seen": 890755008, "step": 2365, "train_runtime": 202132.5142, "train_tokens_per_second": 4406.787 }, { "epoch": 2.2943240953648796, "grad_norm": 0.10320774278040343, "learning_rate": 7.986345193700726e-06, "loss": 0.0414, "num_input_tokens_seen": 892675136, "step": 2370, "train_runtime": 202472.4309, "train_tokens_per_second": 4408.873 }, { "epoch": 2.2991649521965387, "grad_norm": 0.10693768647967408, "learning_rate": 7.883448620996489e-06, "loss": 0.0427, "num_input_tokens_seen": 894455920, "step": 2375, "train_runtime": 202743.8126, "train_tokens_per_second": 4411.754 }, { "epoch": 2.304005809028198, "grad_norm": 0.10326302338915108, "learning_rate": 7.781094995802007e-06, "loss": 0.0485, "num_input_tokens_seen": 896339664, "step": 2380, "train_runtime": 203093.6325, "train_tokens_per_second": 4413.431 }, { "epoch": 2.308846665859857, "grad_norm": 0.1056092406773901, "learning_rate": 7.679287564837074e-06, "loss": 0.0417, "num_input_tokens_seen": 898166624, "step": 2385, "train_runtime": 203393.0523, "train_tokens_per_second": 4415.916 }, { "epoch": 2.3136875226915166, "grad_norm": 0.10998323658142867, "learning_rate": 7.57802955749585e-06, "loss": 0.0408, "num_input_tokens_seen": 900181312, "step": 2390, "train_runtime": 203758.3229, "train_tokens_per_second": 4417.887 }, { "epoch": 2.3185283795231757, "grad_norm": 0.1005967471563839, "learning_rate": 7.4773241857444415e-06, "loss": 0.0423, "num_input_tokens_seen": 902067280, "step": 2395, "train_runtime": 204090.2987, "train_tokens_per_second": 4419.942 }, { "epoch": 2.323369236354835, "grad_norm": 0.09602599680101273, "learning_rate": 7.3771746440190195e-06, "loss": 0.0417, "num_input_tokens_seen": 903961952, "step": 2400, "train_runtime": 204390.7435, "train_tokens_per_second": 4422.715 }, { "epoch": 2.328210093186494, "grad_norm": 0.10546331216246121, "learning_rate": 7.277584109124483e-06, "loss": 0.0439, "num_input_tokens_seen": 905764752, "step": 2405, "train_runtime": 204689.5804, "train_tokens_per_second": 4425.065 }, { "epoch": 2.333050950018153, "grad_norm": 0.10566890437995724, "learning_rate": 7.178555740133694e-06, "loss": 0.0437, "num_input_tokens_seen": 907639040, "step": 2410, "train_runtime": 205007.1634, "train_tokens_per_second": 4427.353 }, { "epoch": 2.3378918068498122, "grad_norm": 0.1078878711996976, "learning_rate": 7.080092678287262e-06, "loss": 0.0429, "num_input_tokens_seen": 909548736, "step": 2415, "train_runtime": 205363.2901, "train_tokens_per_second": 4428.974 }, { "epoch": 2.3427326636814714, "grad_norm": 0.10362678547369009, "learning_rate": 6.982198046893912e-06, "loss": 0.0403, "num_input_tokens_seen": 911369184, "step": 2420, "train_runtime": 205601.0346, "train_tokens_per_second": 4432.707 }, { "epoch": 2.347573520513131, "grad_norm": 0.1116147652899066, "learning_rate": 6.884874951231407e-06, "loss": 0.0444, "num_input_tokens_seen": 913234896, "step": 2425, "train_runtime": 205896.2768, "train_tokens_per_second": 4435.412 }, { "epoch": 2.35241437734479, "grad_norm": 0.11521220794776303, "learning_rate": 6.788126478448046e-06, "loss": 0.045, "num_input_tokens_seen": 915148176, "step": 2430, "train_runtime": 206245.7823, "train_tokens_per_second": 4437.173 }, { "epoch": 2.357255234176449, "grad_norm": 0.09281196214081752, "learning_rate": 6.691955697464722e-06, "loss": 0.0438, "num_input_tokens_seen": 917092976, "step": 2435, "train_runtime": 206645.7743, "train_tokens_per_second": 4437.995 }, { "epoch": 2.3620960910081084, "grad_norm": 0.100214030668592, "learning_rate": 6.596365658877612e-06, "loss": 0.0424, "num_input_tokens_seen": 919066720, "step": 2440, "train_runtime": 207017.7612, "train_tokens_per_second": 4439.555 }, { "epoch": 2.3669369478397675, "grad_norm": 0.10760036279429956, "learning_rate": 6.501359394861378e-06, "loss": 0.0417, "num_input_tokens_seen": 920862256, "step": 2445, "train_runtime": 207284.7764, "train_tokens_per_second": 4442.498 }, { "epoch": 2.371777804671427, "grad_norm": 0.10509350266248424, "learning_rate": 6.4069399190729965e-06, "loss": 0.045, "num_input_tokens_seen": 922629344, "step": 2450, "train_runtime": 207527.687, "train_tokens_per_second": 4445.813 }, { "epoch": 2.376618661503086, "grad_norm": 0.10714050286215131, "learning_rate": 6.3131102265561595e-06, "loss": 0.0428, "num_input_tokens_seen": 924459968, "step": 2455, "train_runtime": 207811.3596, "train_tokens_per_second": 4448.554 }, { "epoch": 2.3814595183347453, "grad_norm": 0.09927356415259088, "learning_rate": 6.219873293646275e-06, "loss": 0.0434, "num_input_tokens_seen": 926282784, "step": 2460, "train_runtime": 208098.2228, "train_tokens_per_second": 4451.181 }, { "epoch": 2.3863003751664045, "grad_norm": 0.10624426038475755, "learning_rate": 6.127232077876044e-06, "loss": 0.0423, "num_input_tokens_seen": 928104928, "step": 2465, "train_runtime": 208396.5232, "train_tokens_per_second": 4453.553 }, { "epoch": 2.3911412319980636, "grad_norm": 0.10065956898085184, "learning_rate": 6.0351895178816675e-06, "loss": 0.0414, "num_input_tokens_seen": 930034976, "step": 2470, "train_runtime": 208753.2157, "train_tokens_per_second": 4455.189 }, { "epoch": 2.3959820888297227, "grad_norm": 0.08919790357784825, "learning_rate": 5.943748533309609e-06, "loss": 0.0403, "num_input_tokens_seen": 931869680, "step": 2475, "train_runtime": 209030.5157, "train_tokens_per_second": 4458.056 }, { "epoch": 2.400822945661382, "grad_norm": 0.09926050843218681, "learning_rate": 5.8529120247239945e-06, "loss": 0.0426, "num_input_tokens_seen": 933822272, "step": 2480, "train_runtime": 209411.1645, "train_tokens_per_second": 4459.276 }, { "epoch": 2.4056638024930415, "grad_norm": 0.09286276749293895, "learning_rate": 5.762682873514605e-06, "loss": 0.041, "num_input_tokens_seen": 935677664, "step": 2485, "train_runtime": 209681.396, "train_tokens_per_second": 4462.378 }, { "epoch": 2.4105046593247006, "grad_norm": 0.11296254298586997, "learning_rate": 5.673063941805451e-06, "loss": 0.0445, "num_input_tokens_seen": 937557696, "step": 2490, "train_runtime": 209997.2312, "train_tokens_per_second": 4464.619 }, { "epoch": 2.4153455161563597, "grad_norm": 0.08754210776797587, "learning_rate": 5.584058072364032e-06, "loss": 0.0403, "num_input_tokens_seen": 939447456, "step": 2495, "train_runtime": 210255.8593, "train_tokens_per_second": 4468.115 }, { "epoch": 2.420186372988019, "grad_norm": 0.10698743548503842, "learning_rate": 5.495668088511128e-06, "loss": 0.0413, "num_input_tokens_seen": 941372928, "step": 2500, "train_runtime": 210635.05, "train_tokens_per_second": 4469.213 }, { "epoch": 2.425027229819678, "grad_norm": 0.11191046124395111, "learning_rate": 5.407896794031245e-06, "loss": 0.0435, "num_input_tokens_seen": 943222032, "step": 2505, "train_runtime": 210947.9822, "train_tokens_per_second": 4471.349 }, { "epoch": 2.429868086651337, "grad_norm": 0.11147955526705726, "learning_rate": 5.3207469730836875e-06, "loss": 0.0424, "num_input_tokens_seen": 945141472, "step": 2510, "train_runtime": 211288.3155, "train_tokens_per_second": 4473.231 }, { "epoch": 2.4347089434829963, "grad_norm": 0.11269801669995572, "learning_rate": 5.2342213901142415e-06, "loss": 0.042, "num_input_tokens_seen": 946942384, "step": 2515, "train_runtime": 211559.8316, "train_tokens_per_second": 4476.003 }, { "epoch": 2.439549800314656, "grad_norm": 0.10414158057001394, "learning_rate": 5.1483227897674736e-06, "loss": 0.0424, "num_input_tokens_seen": 948797600, "step": 2520, "train_runtime": 211859.0494, "train_tokens_per_second": 4478.438 }, { "epoch": 2.444390657146315, "grad_norm": 0.10225241274686434, "learning_rate": 5.0630538967996824e-06, "loss": 0.044, "num_input_tokens_seen": 950738128, "step": 2525, "train_runtime": 212239.9612, "train_tokens_per_second": 4479.543 }, { "epoch": 2.449231513977974, "grad_norm": 0.09486493161848354, "learning_rate": 4.978417415992459e-06, "loss": 0.0414, "num_input_tokens_seen": 952683104, "step": 2530, "train_runtime": 212560.4677, "train_tokens_per_second": 4481.939 }, { "epoch": 2.4540723708096333, "grad_norm": 0.10225202491694556, "learning_rate": 4.8944160320668925e-06, "loss": 0.044, "num_input_tokens_seen": 954565248, "step": 2535, "train_runtime": 212864.2481, "train_tokens_per_second": 4484.385 }, { "epoch": 2.4589132276412924, "grad_norm": 0.10555017743814563, "learning_rate": 4.811052409598415e-06, "loss": 0.0435, "num_input_tokens_seen": 956458752, "step": 2540, "train_runtime": 213206.0267, "train_tokens_per_second": 4486.077 }, { "epoch": 2.4637540844729515, "grad_norm": 0.10832795415710876, "learning_rate": 4.728329192932277e-06, "loss": 0.0434, "num_input_tokens_seen": 958360512, "step": 2545, "train_runtime": 213541.3209, "train_tokens_per_second": 4487.939 }, { "epoch": 2.468594941304611, "grad_norm": 0.10095120179330958, "learning_rate": 4.64624900609964e-06, "loss": 0.0433, "num_input_tokens_seen": 960237872, "step": 2550, "train_runtime": 213826.9979, "train_tokens_per_second": 4490.723 }, { "epoch": 2.4734357981362702, "grad_norm": 0.10772165766963647, "learning_rate": 4.56481445273439e-06, "loss": 0.0413, "num_input_tokens_seen": 962247008, "step": 2555, "train_runtime": 214215.3446, "train_tokens_per_second": 4491.961 }, { "epoch": 2.4782766549679294, "grad_norm": 0.09806092608971834, "learning_rate": 4.4840281159905155e-06, "loss": 0.0414, "num_input_tokens_seen": 964170272, "step": 2560, "train_runtime": 214550.1729, "train_tokens_per_second": 4493.915 }, { "epoch": 2.4831175117995885, "grad_norm": 0.09220012235718683, "learning_rate": 4.403892558460177e-06, "loss": 0.0412, "num_input_tokens_seen": 966120000, "step": 2565, "train_runtime": 214929.4897, "train_tokens_per_second": 4495.056 }, { "epoch": 2.4879583686312476, "grad_norm": 0.09514831510019058, "learning_rate": 4.324410322092415e-06, "loss": 0.0412, "num_input_tokens_seen": 967960784, "step": 2570, "train_runtime": 215212.0348, "train_tokens_per_second": 4497.708 }, { "epoch": 2.4927992254629068, "grad_norm": 0.08465241248211482, "learning_rate": 4.2455839281125256e-06, "loss": 0.0417, "num_input_tokens_seen": 969844080, "step": 2575, "train_runtime": 215514.8865, "train_tokens_per_second": 4500.126 }, { "epoch": 2.4976400822945664, "grad_norm": 0.09900416312304733, "learning_rate": 4.167415876942085e-06, "loss": 0.0417, "num_input_tokens_seen": 971694160, "step": 2580, "train_runtime": 215818.3338, "train_tokens_per_second": 4502.371 }, { "epoch": 2.5024809391262255, "grad_norm": 0.09456247404799463, "learning_rate": 4.089908648119625e-06, "loss": 0.0425, "num_input_tokens_seen": 973545824, "step": 2585, "train_runtime": 216108.9629, "train_tokens_per_second": 4504.884 }, { "epoch": 2.5073217959578846, "grad_norm": 0.1006524579480961, "learning_rate": 4.013064700221991e-06, "loss": 0.0411, "num_input_tokens_seen": 975457776, "step": 2590, "train_runtime": 216449.4006, "train_tokens_per_second": 4506.632 }, { "epoch": 2.5121626527895438, "grad_norm": 0.09861997049355122, "learning_rate": 3.936886470786347e-06, "loss": 0.0405, "num_input_tokens_seen": 977366416, "step": 2595, "train_runtime": 216770.5711, "train_tokens_per_second": 4508.76 }, { "epoch": 2.517003509621203, "grad_norm": 0.09885870437167427, "learning_rate": 3.8613763762328695e-06, "loss": 0.0432, "num_input_tokens_seen": 979260288, "step": 2600, "train_runtime": 217072.79, "train_tokens_per_second": 4511.207 }, { "epoch": 2.521844366452862, "grad_norm": 0.1044111303405238, "learning_rate": 3.7865368117880688e-06, "loss": 0.0441, "num_input_tokens_seen": 981164192, "step": 2605, "train_runtime": 217397.5953, "train_tokens_per_second": 4513.225 }, { "epoch": 2.526685223284521, "grad_norm": 0.09689580727353043, "learning_rate": 3.712370151408842e-06, "loss": 0.041, "num_input_tokens_seen": 983019680, "step": 2610, "train_runtime": 217701.7754, "train_tokens_per_second": 4515.442 }, { "epoch": 2.5315260801161807, "grad_norm": 0.07930860025277806, "learning_rate": 3.6388787477071557e-06, "loss": 0.0382, "num_input_tokens_seen": 984946480, "step": 2615, "train_runtime": 218039.1448, "train_tokens_per_second": 4517.292 }, { "epoch": 2.53636693694784, "grad_norm": 0.09701585395532773, "learning_rate": 3.5660649318754153e-06, "loss": 0.0431, "num_input_tokens_seen": 986820704, "step": 2620, "train_runtime": 218364.0414, "train_tokens_per_second": 4519.154 }, { "epoch": 2.541207793779499, "grad_norm": 0.08951915315467092, "learning_rate": 3.493931013612528e-06, "loss": 0.0404, "num_input_tokens_seen": 988683712, "step": 2625, "train_runtime": 218662.1125, "train_tokens_per_second": 4521.514 }, { "epoch": 2.546048650611158, "grad_norm": 0.10946918657325026, "learning_rate": 3.4224792810506335e-06, "loss": 0.0412, "num_input_tokens_seen": 990613776, "step": 2630, "train_runtime": 219031.7427, "train_tokens_per_second": 4522.695 }, { "epoch": 2.5508895074428173, "grad_norm": 0.09243649296411362, "learning_rate": 3.3517120006825137e-06, "loss": 0.0421, "num_input_tokens_seen": 992516752, "step": 2635, "train_runtime": 219368.5843, "train_tokens_per_second": 4524.425 }, { "epoch": 2.5557303642744764, "grad_norm": 0.11072560597436613, "learning_rate": 3.2816314172897183e-06, "loss": 0.0433, "num_input_tokens_seen": 994379968, "step": 2640, "train_runtime": 219685.0843, "train_tokens_per_second": 4526.388 }, { "epoch": 2.5605712211061356, "grad_norm": 0.1058803495743547, "learning_rate": 3.2122397538713432e-06, "loss": 0.0413, "num_input_tokens_seen": 996372960, "step": 2645, "train_runtime": 220112.659, "train_tokens_per_second": 4526.65 }, { "epoch": 2.565412077937795, "grad_norm": 0.09586921377974332, "learning_rate": 3.143539211573518e-06, "loss": 0.0445, "num_input_tokens_seen": 998147504, "step": 2650, "train_runtime": 220369.4226, "train_tokens_per_second": 4529.428 }, { "epoch": 2.5702529347694543, "grad_norm": 0.100735105625526, "learning_rate": 3.075531969619594e-06, "loss": 0.0402, "num_input_tokens_seen": 1000010032, "step": 2655, "train_runtime": 220683.995, "train_tokens_per_second": 4531.412 }, { "epoch": 2.5750937916011134, "grad_norm": 0.08157917824449015, "learning_rate": 3.00822018524099e-06, "loss": 0.0394, "num_input_tokens_seen": 1001911504, "step": 2660, "train_runtime": 220992.259, "train_tokens_per_second": 4533.695 }, { "epoch": 2.5799346484327725, "grad_norm": 0.10671166825123399, "learning_rate": 2.9416059936088176e-06, "loss": 0.0405, "num_input_tokens_seen": 1003915536, "step": 2665, "train_runtime": 221370.367, "train_tokens_per_second": 4535.004 }, { "epoch": 2.5847755052644317, "grad_norm": 0.10485178372092739, "learning_rate": 2.8756915077660816e-06, "loss": 0.042, "num_input_tokens_seen": 1005786320, "step": 2670, "train_runtime": 221680.2811, "train_tokens_per_second": 4537.103 }, { "epoch": 2.5896163620960913, "grad_norm": 0.11239886407003709, "learning_rate": 2.810478818560719e-06, "loss": 0.046, "num_input_tokens_seen": 1007566272, "step": 2675, "train_runtime": 221965.9319, "train_tokens_per_second": 4539.283 }, { "epoch": 2.59445721892775, "grad_norm": 0.10703312505536274, "learning_rate": 2.745969994579231e-06, "loss": 0.0429, "num_input_tokens_seen": 1009450528, "step": 2680, "train_runtime": 222294.4882, "train_tokens_per_second": 4541.051 }, { "epoch": 2.5992980757594095, "grad_norm": 0.09294201066454781, "learning_rate": 2.6821670820810883e-06, "loss": 0.0413, "num_input_tokens_seen": 1011300880, "step": 2685, "train_runtime": 222609.6014, "train_tokens_per_second": 4542.935 }, { "epoch": 2.6041389325910687, "grad_norm": 0.09973495575855759, "learning_rate": 2.6190721049338145e-06, "loss": 0.0406, "num_input_tokens_seen": 1013140688, "step": 2690, "train_runtime": 222887.9, "train_tokens_per_second": 4545.517 }, { "epoch": 2.608979789422728, "grad_norm": 0.100229670298609, "learning_rate": 2.5566870645487906e-06, "loss": 0.0423, "num_input_tokens_seen": 1015019296, "step": 2695, "train_runtime": 223209.5025, "train_tokens_per_second": 4547.384 }, { "epoch": 2.613820646254387, "grad_norm": 0.08644793584438773, "learning_rate": 2.495013939817767e-06, "loss": 0.0395, "num_input_tokens_seen": 1016892176, "step": 2700, "train_runtime": 223525.4721, "train_tokens_per_second": 4549.335 }, { "epoch": 2.618661503086046, "grad_norm": 0.09650950829215955, "learning_rate": 2.434054687050091e-06, "loss": 0.0405, "num_input_tokens_seen": 1018797600, "step": 2705, "train_runtime": 223827.1085, "train_tokens_per_second": 4551.717 }, { "epoch": 2.6235023599177056, "grad_norm": 0.12423581499512551, "learning_rate": 2.3738112399106644e-06, "loss": 0.0454, "num_input_tokens_seen": 1020718416, "step": 2710, "train_runtime": 224190.0696, "train_tokens_per_second": 4552.915 }, { "epoch": 2.6283432167493648, "grad_norm": 0.10989337129875472, "learning_rate": 2.314285509358566e-06, "loss": 0.0428, "num_input_tokens_seen": 1022524336, "step": 2715, "train_runtime": 224455.0174, "train_tokens_per_second": 4555.587 }, { "epoch": 2.633184073581024, "grad_norm": 0.09039073713194058, "learning_rate": 2.255479383586509e-06, "loss": 0.0406, "num_input_tokens_seen": 1024388192, "step": 2720, "train_runtime": 224743.9654, "train_tokens_per_second": 4558.023 }, { "epoch": 2.638024930412683, "grad_norm": 0.08953330757222618, "learning_rate": 2.197394727960861e-06, "loss": 0.0408, "num_input_tokens_seen": 1026219264, "step": 2725, "train_runtime": 225055.269, "train_tokens_per_second": 4559.854 }, { "epoch": 2.642865787244342, "grad_norm": 0.08090268583905104, "learning_rate": 2.1400333849625338e-06, "loss": 0.0392, "num_input_tokens_seen": 1028260624, "step": 2730, "train_runtime": 225415.8787, "train_tokens_per_second": 4561.616 }, { "epoch": 2.6477066440760013, "grad_norm": 0.10367619160660342, "learning_rate": 2.083397174128518e-06, "loss": 0.0415, "num_input_tokens_seen": 1030159584, "step": 2735, "train_runtime": 225741.8649, "train_tokens_per_second": 4563.441 }, { "epoch": 2.6525475009076604, "grad_norm": 0.11074351346114625, "learning_rate": 2.027487891994162e-06, "loss": 0.0432, "num_input_tokens_seen": 1032081920, "step": 2740, "train_runtime": 226092.6663, "train_tokens_per_second": 4564.862 }, { "epoch": 2.65738835773932, "grad_norm": 0.10039057936660997, "learning_rate": 1.9723073120361924e-06, "loss": 0.0425, "num_input_tokens_seen": 1033931904, "step": 2745, "train_runtime": 226382.3218, "train_tokens_per_second": 4567.194 }, { "epoch": 2.662229214570979, "grad_norm": 0.09363138188924082, "learning_rate": 1.9178571846164532e-06, "loss": 0.0397, "num_input_tokens_seen": 1035806768, "step": 2750, "train_runtime": 226704.6785, "train_tokens_per_second": 4568.97 }, { "epoch": 2.6670700714026383, "grad_norm": 0.08714334312836239, "learning_rate": 1.8641392369263933e-06, "loss": 0.04, "num_input_tokens_seen": 1037779856, "step": 2755, "train_runtime": 227072.2692, "train_tokens_per_second": 4570.262 }, { "epoch": 2.6719109282342974, "grad_norm": 0.10402712457251752, "learning_rate": 1.8111551729322662e-06, "loss": 0.0439, "num_input_tokens_seen": 1039606192, "step": 2760, "train_runtime": 227397.5471, "train_tokens_per_second": 4571.756 }, { "epoch": 2.6767517850659566, "grad_norm": 0.10242769460180877, "learning_rate": 1.7589066733210814e-06, "loss": 0.0473, "num_input_tokens_seen": 1041352016, "step": 2765, "train_runtime": 227680.2847, "train_tokens_per_second": 4573.747 }, { "epoch": 2.681592641897616, "grad_norm": 0.10096005023224232, "learning_rate": 1.7073953954472949e-06, "loss": 0.0432, "num_input_tokens_seen": 1043209584, "step": 2770, "train_runtime": 228004.7054, "train_tokens_per_second": 4575.386 }, { "epoch": 2.686433498729275, "grad_norm": 0.10430771290514242, "learning_rate": 1.6566229732802501e-06, "loss": 0.0423, "num_input_tokens_seen": 1045070400, "step": 2775, "train_runtime": 228328.0951, "train_tokens_per_second": 4577.056 }, { "epoch": 2.6912743555609344, "grad_norm": 0.09976539169225217, "learning_rate": 1.6065910173523101e-06, "loss": 0.0404, "num_input_tokens_seen": 1047027216, "step": 2780, "train_runtime": 228649.8742, "train_tokens_per_second": 4579.173 }, { "epoch": 2.6961152123925936, "grad_norm": 0.09965543945612194, "learning_rate": 1.5573011147078236e-06, "loss": 0.0407, "num_input_tokens_seen": 1048933616, "step": 2785, "train_runtime": 228968.4155, "train_tokens_per_second": 4581.128 }, { "epoch": 2.7009560692242527, "grad_norm": 0.09982142335711996, "learning_rate": 1.5087548288527291e-06, "loss": 0.0416, "num_input_tokens_seen": 1050730672, "step": 2790, "train_runtime": 229238.8178, "train_tokens_per_second": 4583.563 }, { "epoch": 2.705796926055912, "grad_norm": 0.10720846842326023, "learning_rate": 1.4609536997049977e-06, "loss": 0.0415, "num_input_tokens_seen": 1052607296, "step": 2795, "train_runtime": 229546.0728, "train_tokens_per_second": 4585.604 }, { "epoch": 2.710637782887571, "grad_norm": 0.1048290688754556, "learning_rate": 1.4138992435457688e-06, "loss": 0.041, "num_input_tokens_seen": 1054530832, "step": 2800, "train_runtime": 229875.6229, "train_tokens_per_second": 4587.397 }, { "epoch": 2.7154786397192305, "grad_norm": 0.10487087810225432, "learning_rate": 1.3675929529712555e-06, "loss": 0.0404, "num_input_tokens_seen": 1056509856, "step": 2805, "train_runtime": 230229.7535, "train_tokens_per_second": 4588.937 }, { "epoch": 2.7203194965508897, "grad_norm": 0.10984325530051703, "learning_rate": 1.3220362968454026e-06, "loss": 0.041, "num_input_tokens_seen": 1058388320, "step": 2810, "train_runtime": 230583.9386, "train_tokens_per_second": 4590.035 }, { "epoch": 2.725160353382549, "grad_norm": 0.09575719645821536, "learning_rate": 1.277230720253289e-06, "loss": 0.0416, "num_input_tokens_seen": 1060208912, "step": 2815, "train_runtime": 230862.7987, "train_tokens_per_second": 4592.377 }, { "epoch": 2.730001210214208, "grad_norm": 0.0952849105139362, "learning_rate": 1.2331776444552939e-06, "loss": 0.042, "num_input_tokens_seen": 1062178720, "step": 2820, "train_runtime": 231234.4358, "train_tokens_per_second": 4593.514 }, { "epoch": 2.734842067045867, "grad_norm": 0.09720163766984548, "learning_rate": 1.1898784668419927e-06, "loss": 0.0409, "num_input_tokens_seen": 1064115760, "step": 2825, "train_runtime": 231566.8629, "train_tokens_per_second": 4595.285 }, { "epoch": 2.739682923877526, "grad_norm": 0.10713711730304852, "learning_rate": 1.1473345608898789e-06, "loss": 0.0433, "num_input_tokens_seen": 1065971408, "step": 2830, "train_runtime": 231877.52, "train_tokens_per_second": 4597.131 }, { "epoch": 2.7445237807091853, "grad_norm": 0.09631173848061805, "learning_rate": 1.10554727611773e-06, "loss": 0.0433, "num_input_tokens_seen": 1067891024, "step": 2835, "train_runtime": 232238.3471, "train_tokens_per_second": 4598.254 }, { "epoch": 2.749364637540845, "grad_norm": 0.10739207946404603, "learning_rate": 1.0645179380438657e-06, "loss": 0.0433, "num_input_tokens_seen": 1069744096, "step": 2840, "train_runtime": 232579.3058, "train_tokens_per_second": 4599.481 }, { "epoch": 2.754205494372504, "grad_norm": 0.10233938837342275, "learning_rate": 1.0242478481440498e-06, "loss": 0.0395, "num_input_tokens_seen": 1071601856, "step": 2845, "train_runtime": 232856.6503, "train_tokens_per_second": 4601.981 }, { "epoch": 2.759046351204163, "grad_norm": 0.09778768115212016, "learning_rate": 9.847382838102492e-07, "loss": 0.0421, "num_input_tokens_seen": 1073480480, "step": 2850, "train_runtime": 233167.1054, "train_tokens_per_second": 4603.91 }, { "epoch": 2.7638872080358223, "grad_norm": 0.08134168754707406, "learning_rate": 9.459904983100704e-07, "loss": 0.0404, "num_input_tokens_seen": 1075401632, "step": 2855, "train_runtime": 233474.4543, "train_tokens_per_second": 4606.078 }, { "epoch": 2.7687280648674815, "grad_norm": 0.10586368696758197, "learning_rate": 9.080057207470405e-07, "loss": 0.0379, "num_input_tokens_seen": 1077335872, "step": 2860, "train_runtime": 233807.1971, "train_tokens_per_second": 4607.796 }, { "epoch": 2.7735689216991406, "grad_norm": 0.09864439204162297, "learning_rate": 8.707851560216112e-07, "loss": 0.0403, "num_input_tokens_seen": 1079310576, "step": 2865, "train_runtime": 234165.3118, "train_tokens_per_second": 4609.182 }, { "epoch": 2.7784097785307997, "grad_norm": 0.09763664283200235, "learning_rate": 8.343299847929226e-07, "loss": 0.0387, "num_input_tokens_seen": 1081245280, "step": 2870, "train_runtime": 234484.5006, "train_tokens_per_second": 4611.159 }, { "epoch": 2.7832506353624593, "grad_norm": 0.10339662193442319, "learning_rate": 7.986413634413686e-07, "loss": 0.0423, "num_input_tokens_seen": 1083177952, "step": 2875, "train_runtime": 234868.9456, "train_tokens_per_second": 4611.84 }, { "epoch": 2.7880914921941184, "grad_norm": 0.10893179032282667, "learning_rate": 7.637204240319163e-07, "loss": 0.0415, "num_input_tokens_seen": 1085074640, "step": 2880, "train_runtime": 235198.1723, "train_tokens_per_second": 4613.448 }, { "epoch": 2.7929323490257776, "grad_norm": 0.09451603653033806, "learning_rate": 7.295682742781862e-07, "loss": 0.04, "num_input_tokens_seen": 1086903600, "step": 2885, "train_runtime": 235485.5269, "train_tokens_per_second": 4615.586 }, { "epoch": 2.7977732058574367, "grad_norm": 0.09247794261791085, "learning_rate": 6.961859975073121e-07, "loss": 0.0407, "num_input_tokens_seen": 1088720960, "step": 2890, "train_runtime": 235741.8558, "train_tokens_per_second": 4618.276 }, { "epoch": 2.802614062689096, "grad_norm": 0.09819027331260509, "learning_rate": 6.635746526255981e-07, "loss": 0.0399, "num_input_tokens_seen": 1090542208, "step": 2895, "train_runtime": 236021.1683, "train_tokens_per_second": 4620.527 }, { "epoch": 2.8074549195207554, "grad_norm": 0.10508939236953255, "learning_rate": 6.317352740849048e-07, "loss": 0.0407, "num_input_tokens_seen": 1092439568, "step": 2900, "train_runtime": 236348.1264, "train_tokens_per_second": 4622.163 }, { "epoch": 2.812295776352414, "grad_norm": 0.1127436229338298, "learning_rate": 6.006688718498549e-07, "loss": 0.0425, "num_input_tokens_seen": 1094246688, "step": 2905, "train_runtime": 236640.2834, "train_tokens_per_second": 4624.093 }, { "epoch": 2.8171366331840737, "grad_norm": 0.09425104412787287, "learning_rate": 5.703764313657795e-07, "loss": 0.0426, "num_input_tokens_seen": 1096143712, "step": 2910, "train_runtime": 236969.0877, "train_tokens_per_second": 4625.682 }, { "epoch": 2.821977490015733, "grad_norm": 0.0964656395201385, "learning_rate": 5.408589135274755e-07, "loss": 0.0411, "num_input_tokens_seen": 1098111104, "step": 2915, "train_runtime": 237285.2741, "train_tokens_per_second": 4627.81 }, { "epoch": 2.826818346847392, "grad_norm": 0.10779087274875156, "learning_rate": 5.121172546487196e-07, "loss": 0.0407, "num_input_tokens_seen": 1100058224, "step": 2920, "train_runtime": 237651.5376, "train_tokens_per_second": 4628.871 }, { "epoch": 2.831659203679051, "grad_norm": 0.10847631726351929, "learning_rate": 4.841523664325581e-07, "loss": 0.0397, "num_input_tokens_seen": 1101896720, "step": 2925, "train_runtime": 237940.6724, "train_tokens_per_second": 4630.973 }, { "epoch": 2.8365000605107102, "grad_norm": 0.09624461904411019, "learning_rate": 4.5696513594240264e-07, "loss": 0.0412, "num_input_tokens_seen": 1103744592, "step": 2930, "train_runtime": 238212.5813, "train_tokens_per_second": 4633.444 }, { "epoch": 2.84134091734237, "grad_norm": 0.1121598440101494, "learning_rate": 4.305564255738831e-07, "loss": 0.0407, "num_input_tokens_seen": 1105651136, "step": 2935, "train_runtime": 238531.5307, "train_tokens_per_second": 4635.241 }, { "epoch": 2.846181774174029, "grad_norm": 0.10501345290205552, "learning_rate": 4.0492707302749176e-07, "loss": 0.0412, "num_input_tokens_seen": 1107467904, "step": 2940, "train_runtime": 238807.7005, "train_tokens_per_second": 4637.488 }, { "epoch": 2.851022631005688, "grad_norm": 0.09913381095639745, "learning_rate": 3.800778912820102e-07, "loss": 0.041, "num_input_tokens_seen": 1109281136, "step": 2945, "train_runtime": 239102.1626, "train_tokens_per_second": 4639.361 }, { "epoch": 2.8558634878373472, "grad_norm": 0.10690920603697875, "learning_rate": 3.560096685687325e-07, "loss": 0.041, "num_input_tokens_seen": 1111166496, "step": 2950, "train_runtime": 239424.9583, "train_tokens_per_second": 4640.98 }, { "epoch": 2.8607043446690064, "grad_norm": 0.10371511929667834, "learning_rate": 3.32723168346441e-07, "loss": 0.0408, "num_input_tokens_seen": 1113024464, "step": 2955, "train_runtime": 239736.7802, "train_tokens_per_second": 4642.694 }, { "epoch": 2.8655452015006655, "grad_norm": 0.09279356116922857, "learning_rate": 3.102191292772144e-07, "loss": 0.0428, "num_input_tokens_seen": 1114789808, "step": 2960, "train_runtime": 239990.4972, "train_tokens_per_second": 4645.141 }, { "epoch": 2.8703860583323246, "grad_norm": 0.10484566419466071, "learning_rate": 2.884982652029716e-07, "loss": 0.0414, "num_input_tokens_seen": 1116715504, "step": 2965, "train_runtime": 240306.4348, "train_tokens_per_second": 4647.048 }, { "epoch": 2.875226915163984, "grad_norm": 0.09371900146099076, "learning_rate": 2.6756126512285094e-07, "loss": 0.0423, "num_input_tokens_seen": 1118661392, "step": 2970, "train_runtime": 240658.0326, "train_tokens_per_second": 4648.344 }, { "epoch": 2.8800677719956433, "grad_norm": 0.08968284205417433, "learning_rate": 2.4740879317133314e-07, "loss": 0.043, "num_input_tokens_seen": 1120523488, "step": 2975, "train_runtime": 240990.9356, "train_tokens_per_second": 4649.65 }, { "epoch": 2.8849086288273025, "grad_norm": 0.1194432143401207, "learning_rate": 2.2804148859719433e-07, "loss": 0.0425, "num_input_tokens_seen": 1122329216, "step": 2980, "train_runtime": 241267.0534, "train_tokens_per_second": 4651.813 }, { "epoch": 2.8897494856589616, "grad_norm": 0.10332352909612158, "learning_rate": 2.0945996574321392e-07, "loss": 0.0423, "num_input_tokens_seen": 1124108576, "step": 2985, "train_runtime": 241512.3485, "train_tokens_per_second": 4654.456 }, { "epoch": 2.8945903424906207, "grad_norm": 0.09902903701655406, "learning_rate": 1.9166481402669856e-07, "loss": 0.0416, "num_input_tokens_seen": 1125941808, "step": 2990, "train_runtime": 241794.3455, "train_tokens_per_second": 4656.609 }, { "epoch": 2.89943119932228, "grad_norm": 0.10194619378485503, "learning_rate": 1.7465659792077484e-07, "loss": 0.0412, "num_input_tokens_seen": 1127752624, "step": 2995, "train_runtime": 242080.3868, "train_tokens_per_second": 4658.587 }, { "epoch": 2.904272056153939, "grad_norm": 0.11088193358770991, "learning_rate": 1.5843585693648967e-07, "loss": 0.0418, "num_input_tokens_seen": 1129601232, "step": 3000, "train_runtime": 242368.9222, "train_tokens_per_second": 4660.669 }, { "epoch": 2.9091129129855986, "grad_norm": 0.09014663812275092, "learning_rate": 1.4300310560570184e-07, "loss": 0.041, "num_input_tokens_seen": 1131446736, "step": 3005, "train_runtime": 242641.5856, "train_tokens_per_second": 4663.037 }, { "epoch": 2.9139537698172577, "grad_norm": 0.10599016609672014, "learning_rate": 1.2835883346474786e-07, "loss": 0.0402, "num_input_tokens_seen": 1133359440, "step": 3010, "train_runtime": 242967.2782, "train_tokens_per_second": 4664.659 }, { "epoch": 2.918794626648917, "grad_norm": 0.09666954905813178, "learning_rate": 1.1450350503892648e-07, "loss": 0.0413, "num_input_tokens_seen": 1135150272, "step": 3015, "train_runtime": 243234.1704, "train_tokens_per_second": 4666.903 }, { "epoch": 2.923635483480576, "grad_norm": 0.0952319372122157, "learning_rate": 1.014375598277495e-07, "loss": 0.0397, "num_input_tokens_seen": 1137058240, "step": 3020, "train_runtime": 243553.8042, "train_tokens_per_second": 4668.612 }, { "epoch": 2.928476340312235, "grad_norm": 0.0961571420112863, "learning_rate": 8.916141229101671e-08, "loss": 0.0423, "num_input_tokens_seen": 1138880976, "step": 3025, "train_runtime": 243833.0558, "train_tokens_per_second": 4670.741 }, { "epoch": 2.9333171971438947, "grad_norm": 0.10748828390119694, "learning_rate": 7.767545183565983e-08, "loss": 0.0417, "num_input_tokens_seen": 1140802272, "step": 3030, "train_runtime": 244185.0486, "train_tokens_per_second": 4671.876 }, { "epoch": 2.9381580539755534, "grad_norm": 0.09453143585948885, "learning_rate": 6.69800428033912e-08, "loss": 0.0408, "num_input_tokens_seen": 1142726336, "step": 3035, "train_runtime": 244534.0926, "train_tokens_per_second": 4673.076 }, { "epoch": 2.942998910807213, "grad_norm": 0.10758820416109051, "learning_rate": 5.707552445914366e-08, "loss": 0.041, "num_input_tokens_seen": 1144590976, "step": 3040, "train_runtime": 244862.1006, "train_tokens_per_second": 4674.431 }, { "epoch": 2.947839767638872, "grad_norm": 0.09221176210129879, "learning_rate": 4.7962210980317946e-08, "loss": 0.0408, "num_input_tokens_seen": 1146496640, "step": 3045, "train_runtime": 245189.2651, "train_tokens_per_second": 4675.966 }, { "epoch": 2.9526806244705313, "grad_norm": 0.10331440557683293, "learning_rate": 3.964039144680465e-08, "loss": 0.0418, "num_input_tokens_seen": 1148333472, "step": 3050, "train_runtime": 245482.1925, "train_tokens_per_second": 4677.869 }, { "epoch": 2.9575214813021904, "grad_norm": 0.08747696767377501, "learning_rate": 3.2110329831824825e-08, "loss": 0.0423, "num_input_tokens_seen": 1150262528, "step": 3055, "train_runtime": 245807.9806, "train_tokens_per_second": 4679.517 }, { "epoch": 2.9623623381338495, "grad_norm": 0.08985631847344186, "learning_rate": 2.537226499355616e-08, "loss": 0.0384, "num_input_tokens_seen": 1152112096, "step": 3060, "train_runtime": 246091.225, "train_tokens_per_second": 4681.646 }, { "epoch": 2.967203194965509, "grad_norm": 0.10286034088749602, "learning_rate": 1.942641066754458e-08, "loss": 0.0419, "num_input_tokens_seen": 1153978224, "step": 3065, "train_runtime": 246407.1175, "train_tokens_per_second": 4683.218 }, { "epoch": 2.9720440517971682, "grad_norm": 0.09984385756001556, "learning_rate": 1.4272955459937453e-08, "loss": 0.0414, "num_input_tokens_seen": 1155885360, "step": 3070, "train_runtime": 246753.738, "train_tokens_per_second": 4684.368 }, { "epoch": 2.9768849086288274, "grad_norm": 0.09194901699543462, "learning_rate": 9.912062841496705e-09, "loss": 0.0387, "num_input_tokens_seen": 1157755504, "step": 3075, "train_runtime": 247066.5954, "train_tokens_per_second": 4686.006 }, { "epoch": 2.9817257654604865, "grad_norm": 0.10514662226999139, "learning_rate": 6.343871142411306e-09, "loss": 0.042, "num_input_tokens_seen": 1159608192, "step": 3080, "train_runtime": 247360.5627, "train_tokens_per_second": 4687.927 }, { "epoch": 2.9865666222921456, "grad_norm": 0.09518313021004897, "learning_rate": 3.568493547909113e-09, "loss": 0.0407, "num_input_tokens_seen": 1161579776, "step": 3085, "train_runtime": 247715.5527, "train_tokens_per_second": 4689.168 }, { "epoch": 2.991407479123805, "grad_norm": 0.09017165474805426, "learning_rate": 1.586018094670849e-09, "loss": 0.0391, "num_input_tokens_seen": 1163408592, "step": 3090, "train_runtime": 247983.5908, "train_tokens_per_second": 4691.474 }, { "epoch": 2.996248335955464, "grad_norm": 0.1022357668004034, "learning_rate": 3.965076680351176e-10, "loss": 0.0417, "num_input_tokens_seen": 1165284400, "step": 3095, "train_runtime": 248298.2933, "train_tokens_per_second": 4693.083 }, { "epoch": 3.0, "num_input_tokens_seen": 1166793552, "step": 3099, "total_flos": 5942698242998272.0, "train_loss": 0.21310824125995556, "train_runtime": 248614.6933, "train_samples_per_second": 1.595, "train_steps_per_second": 0.012 } ], "logging_steps": 5, "max_steps": 3099, "num_input_tokens_seen": 1166793552, "num_train_epochs": 3, "save_steps": 7000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5942698242998272.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }